runtime error
Exit code: 1. Reason: Qwen3-4B-Thinking-2507.Q4_1.gguf: 0%| | 0.00/2.60G [00:00<?, ?B/s][A Qwen3-4B-Thinking-2507.Q4_1.gguf: 0%| | 5.91M/2.60G [00:01<08:19, 5.19MB/s][A Qwen3-4B-Thinking-2507.Q4_1.gguf: 31%|βββ | 794M/2.60G [00:02<00:05, 311MB/s] [A Qwen3-4B-Thinking-2507.Q4_1.gguf: 80%|ββββββββ | 2.07G/2.60G [00:04<00:00, 603MB/s][A Qwen3-4B-Thinking-2507.Q4_1.gguf: 100%|ββββββββββ| 2.60G/2.60G [00:04<00:00, 531MB/s] Traceback (most recent call last): File "/app/app.py", line 19, in <module> llm = Llama( model_path=model_path, ...<4 lines>... verbose=False ) File "/usr/local/lib/python3.13/site-packages/llama_cpp/llama.py", line 371, in __init__ _LlamaModel( ~~~~~~~~~~~^ path_model=self.model_path, ^^^^^^^^^^^^^^^^^^^^^^^^^^^ params=self.model_params, ^^^^^^^^^^^^^^^^^^^^^^^^^ verbose=self.verbose, ^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/usr/local/lib/python3.13/site-packages/llama_cpp/_internals.py", line 55, in __init__ raise ValueError(f"Failed to load model from file: {path_model}") ValueError: Failed to load model from file: /root/.cache/huggingface/hub/models--khazarai--Qwen3-4B-Qwen3.6-plus-Reasoning-Distilled-GGUF/snapshots/0da399ec8927319dd5137de8de20db43b6e4b2e6/Qwen3-4B-Thinking-2507.Q4_1.gguf
Container logs:
Fetching error logs...