Upload black-forest-labs_FLUX.2-dev_1.txt with huggingface_hub
Browse files
black-forest-labs_FLUX.2-dev_1.txt
CHANGED
|
@@ -14,7 +14,7 @@ image = pipe(image=input_image, prompt=prompt).images[0]
|
|
| 14 |
|
| 15 |
ERROR:
|
| 16 |
Traceback (most recent call last):
|
| 17 |
-
File "/tmp/black-forest-labs_FLUX.2-
|
| 18 |
pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.2-dev", dtype=torch.bfloat16, device_map="cuda")
|
| 19 |
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
|
| 20 |
return fn(*args, **kwargs)
|
|
@@ -26,42 +26,21 @@ Traceback (most recent call last):
|
|
| 26 |
)
|
| 27 |
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/pipelines/pipeline_loading_utils.py", line 876, in load_sub_model
|
| 28 |
loaded_sub_model = load_method(os.path.join(cached_folder, name), **loading_kwargs)
|
| 29 |
-
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/
|
| 30 |
-
|
| 31 |
-
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
model
|
| 35 |
-
|
| 36 |
-
...<
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
)
|
| 40 |
-
^
|
| 41 |
-
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 1217, in convert_and_load_state_dict_in_model
|
| 42 |
-
realized_value, conversion_errors = mapping.convert(
|
| 43 |
-
~~~~~~~~~~~~~~~^
|
| 44 |
-
first_param_name,
|
| 45 |
-
^^^^^^^^^^^^^^^^^
|
| 46 |
-
...<4 lines>...
|
| 47 |
-
conversion_errors=conversion_errors,
|
| 48 |
-
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 49 |
)
|
| 50 |
^
|
| 51 |
-
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/
|
| 52 |
-
|
| 53 |
-
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
return self.__get_result()
|
| 58 |
-
~~~~~~~~~~~~~~~~~^^
|
| 59 |
-
File "/usr/lib/python3.13/concurrent/futures/_base.py", line 401, in __get_result
|
| 60 |
-
raise self._exception
|
| 61 |
-
File "/usr/lib/python3.13/concurrent/futures/thread.py", line 59, in run
|
| 62 |
-
result = self.fn(*self.args, **self.kwargs)
|
| 63 |
-
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 818, in _job
|
| 64 |
-
return _materialize_copy(tensor, device, dtype)
|
| 65 |
-
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 807, in _materialize_copy
|
| 66 |
-
tensor = tensor.to(device=device, dtype=dtype)
|
| 67 |
-
torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 320.00 MiB. GPU 0 has a total capacity of 22.03 GiB of which 265.12 MiB is free. Including non-PyTorch memory, this process has 21.77 GiB memory in use. Of the allocated memory 21.45 GiB is allocated by PyTorch, and 143.64 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
|
|
|
|
| 14 |
|
| 15 |
ERROR:
|
| 16 |
Traceback (most recent call last):
|
| 17 |
+
File "/tmp/black-forest-labs_FLUX.2-dev_1tKRE1v.py", line 28, in <module>
|
| 18 |
pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.2-dev", dtype=torch.bfloat16, device_map="cuda")
|
| 19 |
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
|
| 20 |
return fn(*args, **kwargs)
|
|
|
|
| 26 |
)
|
| 27 |
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/pipelines/pipeline_loading_utils.py", line 876, in load_sub_model
|
| 28 |
loaded_sub_model = load_method(os.path.join(cached_folder, name), **loading_kwargs)
|
| 29 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
|
| 30 |
+
return fn(*args, **kwargs)
|
| 31 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/models/modeling_utils.py", line 1296, in from_pretrained
|
| 32 |
+
) = cls._load_pretrained_model(
|
| 33 |
+
~~~~~~~~~~~~~~~~~~~~~~~~~~^
|
| 34 |
+
model,
|
| 35 |
+
^^^^^^
|
| 36 |
+
...<13 lines>...
|
| 37 |
+
is_parallel_loading_enabled=is_parallel_loading_enabled,
|
| 38 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 39 |
)
|
| 40 |
^
|
| 41 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/models/modeling_utils.py", line 1635, in _load_pretrained_model
|
| 42 |
+
_caching_allocator_warmup(model, expanded_device_map, dtype, hf_quantizer)
|
| 43 |
+
~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 44 |
+
File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/models/model_loading_utils.py", line 751, in _caching_allocator_warmup
|
| 45 |
+
_ = torch.empty(warmup_elems, dtype=dtype, device=device, requires_grad=False)
|
| 46 |
+
torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 60.02 GiB. GPU 0 has a total capacity of 22.03 GiB of which 21.84 GiB is free. Including non-PyTorch memory, this process has 186.00 MiB memory in use. Of the allocated memory 0 bytes is allocated by PyTorch, and 0 bytes is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|