runtime error
Exit code: 1. Reason: site-packages/transformers/models/auto/auto_factory.py", line 604, in from_pretrained return model_class.from_pretrained( ~~~~~~~~~~~~~~~~~~~~~~~~~~~^ pretrained_model_name_or_path, *model_args, config=config, **hub_kwargs, **kwargs ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/app/qwen_tts/core/models/modeling_qwen3_tts.py", line 1876, in from_pretrained model = super().from_pretrained( pretrained_model_name_or_path, ...<11 lines>... **kwargs, ) File "/usr/local/lib/python3.13/site-packages/transformers/modeling_utils.py", line 277, in _wrapper return func(*args, **kwargs) File "/usr/local/lib/python3.13/site-packages/transformers/modeling_utils.py", line 5048, in from_pretrained ) = cls._load_pretrained_model( ~~~~~~~~~~~~~~~~~~~~~~~~~~^ model, ^^^^^^ ...<12 lines>... weights_only=weights_only, ^^^^^^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/usr/local/lib/python3.13/site-packages/transformers/modeling_utils.py", line 5432, in _load_pretrained_model caching_allocator_warmup(model, expanded_device_map, hf_quantizer) ~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.13/site-packages/transformers/modeling_utils.py", line 6089, in caching_allocator_warmup index = device.index if device.index is not None else torch_accelerator_module.current_device() ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^ File "/usr/local/lib/python3.13/site-packages/torch/cuda/__init__.py", line 1071, in current_device _lazy_init() ~~~~~~~~~~^^ File "/usr/local/lib/python3.13/site-packages/torch/cuda/__init__.py", line 412, in _lazy_init torch._C._cuda_init() ~~~~~~~~~~~~~~~~~~~^^ RuntimeError: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx
Container logs:
Fetching error logs...