Upload black-forest-labs_FLUX.2-dev_1.txt with huggingface_hub
Browse files
black-forest-labs_FLUX.2-dev_1.txt
CHANGED
|
@@ -14,7 +14,7 @@ image = pipe(image=input_image, prompt=prompt).images[0]
|
|
| 14 |
|
| 15 |
ERROR:
|
| 16 |
Traceback (most recent call last):
|
| 17 |
-
File "/tmp/black-forest-labs_FLUX.2-
|
| 18 |
pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.2-dev", dtype=torch.bfloat16, device_map="cuda")
|
| 19 |
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
|
| 20 |
return fn(*args, **kwargs)
|
|
@@ -26,21 +26,42 @@ Traceback (most recent call last):
|
|
| 26 |
)
|
| 27 |
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/pipelines/pipeline_loading_utils.py", line 876, in load_sub_model
|
| 28 |
loaded_sub_model = load_method(os.path.join(cached_folder, name), **loading_kwargs)
|
| 29 |
-
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/
|
| 30 |
-
|
| 31 |
-
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
model,
|
| 35 |
-
|
| 36 |
-
...<
|
| 37 |
-
|
| 38 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 39 |
)
|
| 40 |
^
|
| 41 |
-
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
|
| 15 |
ERROR:
|
| 16 |
Traceback (most recent call last):
|
| 17 |
+
File "/tmp/black-forest-labs_FLUX.2-dev_135Owr3.py", line 28, in <module>
|
| 18 |
pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.2-dev", dtype=torch.bfloat16, device_map="cuda")
|
| 19 |
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
|
| 20 |
return fn(*args, **kwargs)
|
|
|
|
| 26 |
)
|
| 27 |
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/pipelines/pipeline_loading_utils.py", line 876, in load_sub_model
|
| 28 |
loaded_sub_model = load_method(os.path.join(cached_folder, name), **loading_kwargs)
|
| 29 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/modeling_utils.py", line 4109, in from_pretrained
|
| 30 |
+
load_info = cls._load_pretrained_model(model, state_dict, checkpoint_files, load_config)
|
| 31 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/modeling_utils.py", line 4231, in _load_pretrained_model
|
| 32 |
+
convert_and_load_state_dict_in_model(
|
| 33 |
+
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^
|
| 34 |
+
model=model,
|
| 35 |
+
^^^^^^^^^^^^
|
| 36 |
+
...<4 lines>...
|
| 37 |
+
disk_offload_index=disk_offload_index,
|
| 38 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 39 |
+
)
|
| 40 |
+
^
|
| 41 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 1217, in convert_and_load_state_dict_in_model
|
| 42 |
+
realized_value, conversion_errors = mapping.convert(
|
| 43 |
+
~~~~~~~~~~~~~~~^
|
| 44 |
+
first_param_name,
|
| 45 |
+
^^^^^^^^^^^^^^^^^
|
| 46 |
+
...<4 lines>...
|
| 47 |
+
conversion_errors=conversion_errors,
|
| 48 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 49 |
)
|
| 50 |
^
|
| 51 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 696, in convert
|
| 52 |
+
collected_tensors = self.materialize_tensors()
|
| 53 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 671, in materialize_tensors
|
| 54 |
+
tensors = [future.result() for future in tensors]
|
| 55 |
+
~~~~~~~~~~~~~^^
|
| 56 |
+
File "/usr/lib/python3.13/concurrent/futures/_base.py", line 456, in result
|
| 57 |
+
return self.__get_result()
|
| 58 |
+
~~~~~~~~~~~~~~~~~^^
|
| 59 |
+
File "/usr/lib/python3.13/concurrent/futures/_base.py", line 401, in __get_result
|
| 60 |
+
raise self._exception
|
| 61 |
+
File "/usr/lib/python3.13/concurrent/futures/thread.py", line 59, in run
|
| 62 |
+
result = self.fn(*self.args, **self.kwargs)
|
| 63 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 818, in _job
|
| 64 |
+
return _materialize_copy(tensor, device, dtype)
|
| 65 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 807, in _materialize_copy
|
| 66 |
+
tensor = tensor.to(device=device, dtype=dtype)
|
| 67 |
+
torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 320.00 MiB. GPU 0 has a total capacity of 22.03 GiB of which 287.12 MiB is free. Including non-PyTorch memory, this process has 21.75 GiB memory in use. Of the allocated memory 21.45 GiB is allocated by PyTorch, and 120.15 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
|