Update pyproject.toml
Browse files- pyproject.toml +2 -7
pyproject.toml
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
[build-system]
|
| 2 |
-
requires = ["setuptools >=
|
| 3 |
build-backend = "setuptools.build_meta"
|
| 4 |
|
| 5 |
[project]
|
|
@@ -20,13 +20,10 @@ dependencies = [
|
|
| 20 |
"omegaconf==2.3.0",
|
| 21 |
"torch==2.2.2",
|
| 22 |
"torchvision==0.17.2",
|
| 23 |
-
"edge-maxxing-pipelines @ git+https://github.com/womboai/edge-maxxing@
|
| 24 |
"huggingface-hub==0.25.2",
|
| 25 |
"oneflow",
|
| 26 |
"setuptools>=75.2.0",
|
| 27 |
-
"bitsandbytes>=0.44.1",
|
| 28 |
-
"stable-fast",
|
| 29 |
-
"tomesd>=0.1.3",
|
| 30 |
]
|
| 31 |
|
| 32 |
[tool.edge-maxxing]
|
|
@@ -37,8 +34,6 @@ models = [
|
|
| 37 |
|
| 38 |
[tool.uv.sources]
|
| 39 |
oneflow = { url = "https://github.com/siliconflow/oneflow_releases/releases/download/community_cu118/oneflow-0.9.1.dev20240802%2Bcu118-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }
|
| 40 |
-
stable-fast = { url = "https://github.com/chengzeyi/stable-fast/releases/download/v1.0.5/stable_fast-1.0.5+torch222cu121-cp310-cp310-manylinux2014_x86_64.whl" }
|
| 41 |
|
| 42 |
[project.scripts]
|
| 43 |
start_inference = "main:main"
|
| 44 |
-
|
|
|
|
| 1 |
[build-system]
|
| 2 |
+
requires = ["setuptools >= 75.0"]
|
| 3 |
build-backend = "setuptools.build_meta"
|
| 4 |
|
| 5 |
[project]
|
|
|
|
| 20 |
"omegaconf==2.3.0",
|
| 21 |
"torch==2.2.2",
|
| 22 |
"torchvision==0.17.2",
|
| 23 |
+
"edge-maxxing-pipelines @ git+https://github.com/womboai/edge-maxxing@e713a4f52ca3ea8c1d57ff63c1c08470f4fd0a60#subdirectory=pipelines",
|
| 24 |
"huggingface-hub==0.25.2",
|
| 25 |
"oneflow",
|
| 26 |
"setuptools>=75.2.0",
|
|
|
|
|
|
|
|
|
|
| 27 |
]
|
| 28 |
|
| 29 |
[tool.edge-maxxing]
|
|
|
|
| 34 |
|
| 35 |
[tool.uv.sources]
|
| 36 |
oneflow = { url = "https://github.com/siliconflow/oneflow_releases/releases/download/community_cu118/oneflow-0.9.1.dev20240802%2Bcu118-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }
|
|
|
|
| 37 |
|
| 38 |
[project.scripts]
|
| 39 |
start_inference = "main:main"
|
|
|