Act commited on
Commit
15abeec
·
verified ·
1 Parent(s): 58b4052

Initial commit with folder contents

Browse files
Files changed (5) hide show
  1. .gitattributes +0 -1
  2. pyproject.toml +7 -22
  3. src/main.py +7 -11
  4. src/pipeline.py +32 -59
  5. uv.lock +9 -51
.gitattributes CHANGED
@@ -33,4 +33,3 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
- RobertML.png filter=lfs diff=lfs merge=lfs -text
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
pyproject.toml CHANGED
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
4
 
5
  [project]
6
  name = "flux-schnell-edge-inference"
7
- description = "An edge-maxxing model submission by RobertML for the 4090 Flux contest"
8
  requires-python = ">=3.10,<3.13"
9
  version = "8"
10
  dependencies = [
@@ -15,30 +15,15 @@ dependencies = [
15
  "torch==2.5.1",
16
  "protobuf==5.28.3",
17
  "sentencepiece==0.2.0",
18
- "edge-maxxing-pipelines @ git+https://github.com/womboai/edge-maxxing@7c760ac54f6052803dadb3ade8ebfc9679a94589#subdirectory=pipelines",
19
- "gitpython>=3.1.43",
20
- "hf_transfer==0.1.8",
21
  "torchao==0.6.1",
 
 
 
22
  ]
23
 
24
  [[tool.edge-maxxing.models]]
25
- repository = "black-forest-labs/FLUX.1-schnell"
26
- revision = "741f7c3ce8b383c54771c7003378a50191e9efe9"
27
- exclude = ["transformer"]
28
-
29
- [[tool.edge-maxxing.models]]
30
- repository = "RobertML/FLUX.1-schnell-int8wo"
31
- revision = "307e0777d92df966a3c0f99f31a6ee8957a9857a"
32
-
33
- [[tool.edge-maxxing.models]]
34
- repository = "city96/t5-v1_1-xxl-encoder-bf16"
35
- revision = "1b9c856aadb864af93c1dcdc226c2774fa67bc86"
36
-
37
- [[tool.edge-maxxing.models]]
38
- repository = "proact/PRO_FLUX_012"
39
- revision = "44b2be6e38d78508fc1e6bdf768fa8a4dbf61fcd"
40
-
41
 
42
  [project.scripts]
43
- start_inference = "main:main"
44
-
 
4
 
5
  [project]
6
  name = "flux-schnell-edge-inference"
7
+ description = "An edge-maxxing model submission for the 4090 Flux contest"
8
  requires-python = ">=3.10,<3.13"
9
  version = "8"
10
  dependencies = [
 
15
  "torch==2.5.1",
16
  "protobuf==5.28.3",
17
  "sentencepiece==0.2.0",
 
 
 
18
  "torchao==0.6.1",
19
+ "hf_transfer==0.1.8",
20
+ "setuptools==75.2.0",
21
+ "edge-maxxing-pipelines @ git+https://github.com/womboai/edge-maxxing@7c760ac54f6052803dadb3ade8ebfc9679a94589#subdirectory=pipelines",
22
  ]
23
 
24
  [[tool.edge-maxxing.models]]
25
+ repository = "slobers/Flux.1.Schnella"
26
+ revision = "e34d670e44cecbbc90e4962e7aada2ac5ce8b55b"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
 
28
  [project.scripts]
29
+ start_inference = "main:main"
 
src/main.py CHANGED
@@ -1,24 +1,18 @@
1
- import atexit
2
  from io import BytesIO
3
  from multiprocessing.connection import Listener
4
  from os import chmod, remove
5
  from os.path import abspath, exists
6
  from pathlib import Path
7
- import torch
8
 
9
  from PIL.JpegImagePlugin import JpegImageFile
10
  from pipelines.models import TextToImageRequest
11
- from pipeline import load_pipeline, infer
12
- SOCKET = abspath(Path(__file__).parent.parent / "inferences.sock")
13
 
 
14
 
15
- def at_exit():
16
- torch.cuda.empty_cache()
17
 
18
 
19
  def main():
20
- atexit.register(at_exit)
21
-
22
  print(f"Loading pipeline")
23
  pipeline = load_pipeline()
24
 
@@ -33,7 +27,7 @@ def main():
33
  print(f"Awaiting connections")
34
  with listener.accept() as connection:
35
  print(f"Connected")
36
- generator = torch.Generator("cuda")
37
  while True:
38
  try:
39
  request = TextToImageRequest.model_validate_json(connection.recv_bytes().decode("utf-8"))
@@ -41,13 +35,15 @@ def main():
41
  print(f"Inference socket exiting")
42
 
43
  return
44
- image = infer(request, pipeline, generator.manual_seed(request.seed))
 
 
45
  data = BytesIO()
46
  image.save(data, format=JpegImageFile.format)
47
 
48
  packet = data.getvalue()
49
 
50
- connection.send_bytes(packet )
51
 
52
 
53
  if __name__ == '__main__':
 
 
1
  from io import BytesIO
2
  from multiprocessing.connection import Listener
3
  from os import chmod, remove
4
  from os.path import abspath, exists
5
  from pathlib import Path
 
6
 
7
  from PIL.JpegImagePlugin import JpegImageFile
8
  from pipelines.models import TextToImageRequest
 
 
9
 
10
+ from pipeline import load_pipeline, infer
11
 
12
+ SOCKET = abspath(Path(__file__).parent.parent / "inferences.sock")
 
13
 
14
 
15
  def main():
 
 
16
  print(f"Loading pipeline")
17
  pipeline = load_pipeline()
18
 
 
27
  print(f"Awaiting connections")
28
  with listener.accept() as connection:
29
  print(f"Connected")
30
+
31
  while True:
32
  try:
33
  request = TextToImageRequest.model_validate_json(connection.recv_bytes().decode("utf-8"))
 
35
  print(f"Inference socket exiting")
36
 
37
  return
38
+
39
+ image = infer(request, pipeline)
40
+
41
  data = BytesIO()
42
  image.save(data, format=JpegImageFile.format)
43
 
44
  packet = data.getvalue()
45
 
46
+ connection.send_bytes(packet)
47
 
48
 
49
  if __name__ == '__main__':
src/pipeline.py CHANGED
@@ -1,73 +1,46 @@
1
- from diffusers import FluxPipeline, AutoencoderKL, AutoencoderTiny
2
-
3
  from huggingface_hub.constants import HF_HUB_CACHE
4
- from transformers import T5EncoderModel, T5TokenizerFast, CLIPTokenizer
5
  import torch
6
  import torch._dynamo
7
  import gc
8
- from PIL import Image as img
 
9
  from PIL.Image import Image
10
  from pipelines.models import TextToImageRequest
11
  from torch import Generator
12
- import time
13
  from diffusers import FluxTransformer2DModel, DiffusionPipeline
14
  from torchao.quantization import quantize_, int8_weight_only, fpx_weight_only
15
- import os
16
  os.environ['PYTORCH_CUDA_ALLOC_CONF']="expandable_segments:True"
17
-
18
-
19
- torch.backends.cuda.matmul.allow_tf32 = True
20
- torch.backends.cudnn.enabled = True
21
- torch.backends.cudnn.benchmark = True
22
-
23
- ckpt_id = "black-forest-labs/FLUX.1-schnell"
24
- ckpt_revision = "741f7c3ce8b383c54771c7003378a50191e9efe9"
25
-
26
-
27
  Pipeline = None
28
- def empty_cache():
29
- gc.collect()
30
- torch.cuda.empty_cache()
31
- torch.cuda.reset_max_memory_allocated()
32
- torch.cuda.reset_peak_memory_stats()
33
-
34
- def load_pipeline() -> Pipeline:
35
- empty_cache()
36
-
37
- dtype, device = torch.bfloat16, "cuda"
38
-
39
- text_encoder_2 = T5EncoderModel.from_pretrained(
40
- "city96/t5-v1_1-xxl-encoder-bf16", revision = "1b9c856aadb864af93c1dcdc226c2774fa67bc86", torch_dtype=torch.bfloat16
41
- ).to(memory_format=torch.channels_last)
42
-
43
-
44
- path = os.path.join(HF_HUB_CACHE, "models--RobertML--FLUX.1-schnell-int8wo/snapshots/307e0777d92df966a3c0f99f31a6ee8957a9857a")
45
- model = FluxTransformer2DModel.from_pretrained(path, torch_dtype=dtype, use_safetensors=False).to(memory_format=torch.channels_last)
46
- pipeline = DiffusionPipeline.from_pretrained(
47
- ckpt_id,
48
- revision=ckpt_revision,
49
- transformer=model,
50
- text_encoder_2=text_encoder_2,
51
- torch_dtype=dtype,
52
- ).to(device)
53
  quantize_(pipeline.vae, int8_weight_only())
54
-
55
- pipeline(prompt="imprisonable, forechamber, monotropic, blandiloquious, blechnoid", width=1024, height=1024, guidance_scale=0.0, num_inference_steps=4, max_sequence_length=256)
56
-
57
- empty_cache()
58
  return pipeline
59
-
60
-
61
  @torch.no_grad()
62
- def infer(request: TextToImageRequest, pipeline: Pipeline, generator: Generator) -> Image:
63
-
64
- image=pipeline(request.prompt,
65
- generator=generator,
66
- guidance_scale=0.0,
67
- num_inference_steps=4,
68
- max_sequence_length=256,
69
- height=request.height,
70
- width=request.width,
71
- output_type="pil").images[0]
72
-
73
- return(image)
 
1
+ #4
2
+ #4
3
  from huggingface_hub.constants import HF_HUB_CACHE
4
+ from transformers import T5EncoderModel, T5TokenizerFast, CLIPTokenizer, CLIPTextModel
5
  import torch
6
  import torch._dynamo
7
  import gc
8
+ import os
9
+ from diffusers import FluxPipeline, AutoencoderTiny
10
  from PIL.Image import Image
11
  from pipelines.models import TextToImageRequest
12
  from torch import Generator
 
13
  from diffusers import FluxTransformer2DModel, DiffusionPipeline
14
  from torchao.quantization import quantize_, int8_weight_only, fpx_weight_only
 
15
  os.environ['PYTORCH_CUDA_ALLOC_CONF']="expandable_segments:True"
16
+ os.environ["TOKENIZERS_PARALLELISM"] = "True"
17
+ torch._dynamo.config.suppress_errors = True
 
 
 
 
 
 
 
 
18
  Pipeline = None
19
+ ids = "slobers/Flux.1.Schnella"
20
+ Revision = "e34d670e44cecbbc90e4962e7aada2ac5ce8b55b"
21
+ gc.collect()
22
+ torch.cuda.empty_cache()
23
+ torch.cuda.reset_max_memory_allocated()
24
+ torch.cuda.reset_peak_memory_stats()
25
+ def load_pipeline() -> Pipeline:
26
+ path = os.path.join(HF_HUB_CACHE, "models--slobers--Flux.1.Schnella/snapshots/e34d670e44cecbbc90e4962e7aada2ac5ce8b55b/transformer")
27
+ transformer = FluxTransformer2DModel.from_pretrained(path, torch_dtype=torch.bfloat16, use_safetensors=False).to(memory_format=torch.channels_last)
28
+ pipeline = FluxPipeline.from_pretrained(ids, revision=Revision, transformer=transformer, local_files_only=True, torch_dtype=torch.bfloat16,)
29
+ pipeline.to("cuda")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  quantize_(pipeline.vae, int8_weight_only())
31
+ pipeline.transformer = torch.compile(pipeline.transformer, options={"shape_padding": True})
32
+ for _ in range(3):
33
+ pipeline(prompt="insensible, timbale, pothery, electrovital, actinogram, taxis, intracerebellar, centrodesmus", width=1024, height=1024, guidance_scale=0.0, num_inference_steps=4, max_sequence_length=256)
 
34
  return pipeline
 
 
35
  @torch.no_grad()
36
+ def infer(request: TextToImageRequest, pipeline: Pipeline) -> Image:
37
+ generator = Generator(pipeline.device).manual_seed(request.seed)
38
+ return pipeline(
39
+ request.prompt,
40
+ generator=generator,
41
+ guidance_scale=0.0,
42
+ num_inference_steps=4,
43
+ max_sequence_length=256,
44
+ height=request.height,
45
+ width=request.width,
46
+ ).images[0]
 
uv.lock CHANGED
@@ -1,15 +1,8 @@
1
  version = 1
2
  requires-python = ">=3.10, <3.13"
3
  resolution-markers = [
4
- "python_full_version < '3.11' and platform_system == 'Darwin'",
5
- "python_full_version < '3.11' and platform_machine == 'aarch64' and platform_system == 'Linux'",
6
- "(python_full_version < '3.11' and platform_machine != 'aarch64' and platform_system != 'Darwin') or (python_full_version < '3.11' and platform_system != 'Darwin' and platform_system != 'Linux')",
7
- "python_full_version == '3.11.*' and platform_system == 'Darwin'",
8
- "python_full_version == '3.11.*' and platform_machine == 'aarch64' and platform_system == 'Linux'",
9
- "(python_full_version == '3.11.*' and platform_machine != 'aarch64' and platform_system != 'Darwin') or (python_full_version == '3.11.*' and platform_system != 'Darwin' and platform_system != 'Linux')",
10
- "python_full_version >= '3.12' and platform_system == 'Darwin'",
11
- "python_full_version >= '3.12' and platform_machine == 'aarch64' and platform_system == 'Linux'",
12
- "(python_full_version >= '3.12' and platform_machine != 'aarch64' and platform_system != 'Darwin') or (python_full_version >= '3.12' and platform_system != 'Darwin' and platform_system != 'Linux')",
13
  ]
14
 
15
  [[package]]
@@ -160,7 +153,6 @@ dependencies = [
160
  { name = "accelerate" },
161
  { name = "diffusers" },
162
  { name = "edge-maxxing-pipelines" },
163
- { name = "gitpython" },
164
  { name = "hf-transfer" },
165
  { name = "omegaconf" },
166
  { name = "protobuf" },
@@ -175,7 +167,6 @@ requires-dist = [
175
  { name = "accelerate", specifier = "==1.1.0" },
176
  { name = "diffusers", specifier = "==0.31.0" },
177
  { name = "edge-maxxing-pipelines", git = "https://github.com/womboai/edge-maxxing?subdirectory=pipelines&rev=7c760ac54f6052803dadb3ade8ebfc9679a94589#7c760ac54f6052803dadb3ade8ebfc9679a94589" },
178
- { name = "gitpython", specifier = ">=3.1.43" },
179
  { name = "hf-transfer", specifier = "==0.1.8" },
180
  { name = "omegaconf", specifier = "==2.3.0" },
181
  { name = "protobuf", specifier = "==5.28.3" },
@@ -194,30 +185,6 @@ wheels = [
194
  { url = "https://files.pythonhosted.org/packages/c6/b2/454d6e7f0158951d8a78c2e1eb4f69ae81beb8dca5fee9809c6c99e9d0d0/fsspec-2024.10.0-py3-none-any.whl", hash = "sha256:03b9a6785766a4de40368b88906366755e2819e758b83705c88cd7cb5fe81871", size = 179641 },
195
  ]
196
 
197
- [[package]]
198
- name = "gitdb"
199
- version = "4.0.11"
200
- source = { registry = "https://pypi.org/simple" }
201
- dependencies = [
202
- { name = "smmap" },
203
- ]
204
- sdist = { url = "https://files.pythonhosted.org/packages/19/0d/bbb5b5ee188dec84647a4664f3e11b06ade2bde568dbd489d9d64adef8ed/gitdb-4.0.11.tar.gz", hash = "sha256:bf5421126136d6d0af55bc1e7c1af1c397a34f5b7bd79e776cd3e89785c2b04b", size = 394469 }
205
- wheels = [
206
- { url = "https://files.pythonhosted.org/packages/fd/5b/8f0c4a5bb9fd491c277c21eff7ccae71b47d43c4446c9d0c6cff2fe8c2c4/gitdb-4.0.11-py3-none-any.whl", hash = "sha256:81a3407ddd2ee8df444cbacea00e2d038e40150acfa3001696fe0dcf1d3adfa4", size = 62721 },
207
- ]
208
-
209
- [[package]]
210
- name = "gitpython"
211
- version = "3.1.43"
212
- source = { registry = "https://pypi.org/simple" }
213
- dependencies = [
214
- { name = "gitdb" },
215
- ]
216
- sdist = { url = "https://files.pythonhosted.org/packages/b6/a1/106fd9fa2dd989b6fb36e5893961f82992cf676381707253e0bf93eb1662/GitPython-3.1.43.tar.gz", hash = "sha256:35f314a9f878467f5453cc1fee295c3e18e52f1b99f10f6cf5b1682e968a9e7c", size = 214149 }
217
- wheels = [
218
- { url = "https://files.pythonhosted.org/packages/e9/bd/cc3a402a6439c15c3d4294333e13042b915bbeab54edc457c723931fed3f/GitPython-3.1.43-py3-none-any.whl", hash = "sha256:eec7ec56b92aad751f9912a73404bc02ba212a23adb2c7098ee668417051a1ff", size = 207337 },
219
- ]
220
-
221
  [[package]]
222
  name = "hf-transfer"
223
  version = "0.1.8"
@@ -445,7 +412,7 @@ name = "nvidia-cudnn-cu12"
445
  version = "9.1.0.70"
446
  source = { registry = "https://pypi.org/simple" }
447
  dependencies = [
448
- { name = "nvidia-cublas-cu12", marker = "(platform_machine != 'aarch64' and platform_system != 'Darwin') or (platform_system != 'Darwin' and platform_system != 'Linux')" },
449
  ]
450
  wheels = [
451
  { url = "https://files.pythonhosted.org/packages/9f/fd/713452cd72343f682b1c7b9321e23829f00b842ceaedcda96e742ea0b0b3/nvidia_cudnn_cu12-9.1.0.70-py3-none-manylinux2014_x86_64.whl", hash = "sha256:165764f44ef8c61fcdfdfdbe769d687e06374059fbb388b6c89ecb0e28793a6f", size = 664752741 },
@@ -456,7 +423,7 @@ name = "nvidia-cufft-cu12"
456
  version = "11.2.1.3"
457
  source = { registry = "https://pypi.org/simple" }
458
  dependencies = [
459
- { name = "nvidia-nvjitlink-cu12", marker = "(platform_machine != 'aarch64' and platform_system != 'Darwin') or (platform_system != 'Darwin' and platform_system != 'Linux')" },
460
  ]
461
  wheels = [
462
  { url = "https://files.pythonhosted.org/packages/7a/8a/0e728f749baca3fbeffad762738276e5df60851958be7783af121a7221e7/nvidia_cufft_cu12-11.2.1.3-py3-none-manylinux2014_aarch64.whl", hash = "sha256:5dad8008fc7f92f5ddfa2101430917ce2ffacd86824914c82e28990ad7f00399", size = 211422548 },
@@ -477,9 +444,9 @@ name = "nvidia-cusolver-cu12"
477
  version = "11.6.1.9"
478
  source = { registry = "https://pypi.org/simple" }
479
  dependencies = [
480
- { name = "nvidia-cublas-cu12", marker = "(platform_machine != 'aarch64' and platform_system != 'Darwin') or (platform_system != 'Darwin' and platform_system != 'Linux')" },
481
- { name = "nvidia-cusparse-cu12", marker = "(platform_machine != 'aarch64' and platform_system != 'Darwin') or (platform_system != 'Darwin' and platform_system != 'Linux')" },
482
- { name = "nvidia-nvjitlink-cu12", marker = "(platform_machine != 'aarch64' and platform_system != 'Darwin') or (platform_system != 'Darwin' and platform_system != 'Linux')" },
483
  ]
484
  wheels = [
485
  { url = "https://files.pythonhosted.org/packages/46/6b/a5c33cf16af09166845345275c34ad2190944bcc6026797a39f8e0a282e0/nvidia_cusolver_cu12-11.6.1.9-py3-none-manylinux2014_aarch64.whl", hash = "sha256:d338f155f174f90724bbde3758b7ac375a70ce8e706d70b018dd3375545fc84e", size = 127634111 },
@@ -491,7 +458,7 @@ name = "nvidia-cusparse-cu12"
491
  version = "12.3.1.170"
492
  source = { registry = "https://pypi.org/simple" }
493
  dependencies = [
494
- { name = "nvidia-nvjitlink-cu12", marker = "(platform_machine != 'aarch64' and platform_system != 'Darwin') or (platform_system != 'Darwin' and platform_system != 'Linux')" },
495
  ]
496
  wheels = [
497
  { url = "https://files.pythonhosted.org/packages/96/a9/c0d2f83a53d40a4a41be14cea6a0bf9e668ffcf8b004bd65633f433050c0/nvidia_cusparse_cu12-12.3.1.170-py3-none-manylinux2014_aarch64.whl", hash = "sha256:9d32f62896231ebe0480efd8a7f702e143c98cfaa0e8a76df3386c1ba2b54df3", size = 207381987 },
@@ -888,15 +855,6 @@ wheels = [
888
  { url = "https://files.pythonhosted.org/packages/90/12/282ee9bce8b58130cb762fbc9beabd531549952cac11fc56add11dcb7ea0/setuptools-75.3.0-py3-none-any.whl", hash = "sha256:f2504966861356aa38616760c0f66568e535562374995367b4e69c7143cf6bcd", size = 1251070 },
889
  ]
890
 
891
- [[package]]
892
- name = "smmap"
893
- version = "5.0.1"
894
- source = { registry = "https://pypi.org/simple" }
895
- sdist = { url = "https://files.pythonhosted.org/packages/88/04/b5bf6d21dc4041000ccba7eb17dd3055feb237e7ffc2c20d3fae3af62baa/smmap-5.0.1.tar.gz", hash = "sha256:dceeb6c0028fdb6734471eb07c0cd2aae706ccaecab45965ee83f11c8d3b1f62", size = 22291 }
896
- wheels = [
897
- { url = "https://files.pythonhosted.org/packages/a7/a5/10f97f73544edcdef54409f1d839f6049a0d79df68adbc1ceb24d1aaca42/smmap-5.0.1-py3-none-any.whl", hash = "sha256:e6d8668fa5f93e706934a62d7b4db19c8d9eb8cf2adbb75ef1b675aa332b69da", size = 24282 },
898
- ]
899
-
900
  [[package]]
901
  name = "sympy"
902
  version = "1.13.1"
@@ -1055,7 +1013,7 @@ name = "triton"
1055
  version = "3.1.0"
1056
  source = { registry = "https://pypi.org/simple" }
1057
  dependencies = [
1058
- { name = "filelock", marker = "(platform_machine != 'aarch64' and platform_system != 'Darwin') or (platform_system != 'Darwin' and platform_system != 'Linux')" },
1059
  ]
1060
  wheels = [
1061
  { url = "https://files.pythonhosted.org/packages/98/29/69aa56dc0b2eb2602b553881e34243475ea2afd9699be042316842788ff5/triton-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6b0dd10a925263abbe9fa37dcde67a5e9b2383fc269fdf59f5657cac38c5d1d8", size = 209460013 },
 
1
  version = 1
2
  requires-python = ">=3.10, <3.13"
3
  resolution-markers = [
4
+ "python_full_version < '3.12'",
5
+ "python_full_version >= '3.12'",
 
 
 
 
 
 
 
6
  ]
7
 
8
  [[package]]
 
153
  { name = "accelerate" },
154
  { name = "diffusers" },
155
  { name = "edge-maxxing-pipelines" },
 
156
  { name = "hf-transfer" },
157
  { name = "omegaconf" },
158
  { name = "protobuf" },
 
167
  { name = "accelerate", specifier = "==1.1.0" },
168
  { name = "diffusers", specifier = "==0.31.0" },
169
  { name = "edge-maxxing-pipelines", git = "https://github.com/womboai/edge-maxxing?subdirectory=pipelines&rev=7c760ac54f6052803dadb3ade8ebfc9679a94589#7c760ac54f6052803dadb3ade8ebfc9679a94589" },
 
170
  { name = "hf-transfer", specifier = "==0.1.8" },
171
  { name = "omegaconf", specifier = "==2.3.0" },
172
  { name = "protobuf", specifier = "==5.28.3" },
 
185
  { url = "https://files.pythonhosted.org/packages/c6/b2/454d6e7f0158951d8a78c2e1eb4f69ae81beb8dca5fee9809c6c99e9d0d0/fsspec-2024.10.0-py3-none-any.whl", hash = "sha256:03b9a6785766a4de40368b88906366755e2819e758b83705c88cd7cb5fe81871", size = 179641 },
186
  ]
187
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
188
  [[package]]
189
  name = "hf-transfer"
190
  version = "0.1.8"
 
412
  version = "9.1.0.70"
413
  source = { registry = "https://pypi.org/simple" }
414
  dependencies = [
415
+ { name = "nvidia-cublas-cu12" },
416
  ]
417
  wheels = [
418
  { url = "https://files.pythonhosted.org/packages/9f/fd/713452cd72343f682b1c7b9321e23829f00b842ceaedcda96e742ea0b0b3/nvidia_cudnn_cu12-9.1.0.70-py3-none-manylinux2014_x86_64.whl", hash = "sha256:165764f44ef8c61fcdfdfdbe769d687e06374059fbb388b6c89ecb0e28793a6f", size = 664752741 },
 
423
  version = "11.2.1.3"
424
  source = { registry = "https://pypi.org/simple" }
425
  dependencies = [
426
+ { name = "nvidia-nvjitlink-cu12" },
427
  ]
428
  wheels = [
429
  { url = "https://files.pythonhosted.org/packages/7a/8a/0e728f749baca3fbeffad762738276e5df60851958be7783af121a7221e7/nvidia_cufft_cu12-11.2.1.3-py3-none-manylinux2014_aarch64.whl", hash = "sha256:5dad8008fc7f92f5ddfa2101430917ce2ffacd86824914c82e28990ad7f00399", size = 211422548 },
 
444
  version = "11.6.1.9"
445
  source = { registry = "https://pypi.org/simple" }
446
  dependencies = [
447
+ { name = "nvidia-cublas-cu12" },
448
+ { name = "nvidia-cusparse-cu12" },
449
+ { name = "nvidia-nvjitlink-cu12" },
450
  ]
451
  wheels = [
452
  { url = "https://files.pythonhosted.org/packages/46/6b/a5c33cf16af09166845345275c34ad2190944bcc6026797a39f8e0a282e0/nvidia_cusolver_cu12-11.6.1.9-py3-none-manylinux2014_aarch64.whl", hash = "sha256:d338f155f174f90724bbde3758b7ac375a70ce8e706d70b018dd3375545fc84e", size = 127634111 },
 
458
  version = "12.3.1.170"
459
  source = { registry = "https://pypi.org/simple" }
460
  dependencies = [
461
+ { name = "nvidia-nvjitlink-cu12" },
462
  ]
463
  wheels = [
464
  { url = "https://files.pythonhosted.org/packages/96/a9/c0d2f83a53d40a4a41be14cea6a0bf9e668ffcf8b004bd65633f433050c0/nvidia_cusparse_cu12-12.3.1.170-py3-none-manylinux2014_aarch64.whl", hash = "sha256:9d32f62896231ebe0480efd8a7f702e143c98cfaa0e8a76df3386c1ba2b54df3", size = 207381987 },
 
855
  { url = "https://files.pythonhosted.org/packages/90/12/282ee9bce8b58130cb762fbc9beabd531549952cac11fc56add11dcb7ea0/setuptools-75.3.0-py3-none-any.whl", hash = "sha256:f2504966861356aa38616760c0f66568e535562374995367b4e69c7143cf6bcd", size = 1251070 },
856
  ]
857
 
 
 
 
 
 
 
 
 
 
858
  [[package]]
859
  name = "sympy"
860
  version = "1.13.1"
 
1013
  version = "3.1.0"
1014
  source = { registry = "https://pypi.org/simple" }
1015
  dependencies = [
1016
+ { name = "filelock" },
1017
  ]
1018
  wheels = [
1019
  { url = "https://files.pythonhosted.org/packages/98/29/69aa56dc0b2eb2602b553881e34243475ea2afd9699be042316842788ff5/triton-3.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6b0dd10a925263abbe9fa37dcde67a5e9b2383fc269fdf59f5657cac38c5d1d8", size = 209460013 },