runtime error

Exit code: 3. Reason: oader: - kv 28: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 29: tokenizer.chat_template str = {%- if tools %}\n {{- '<|start_of_r... llama_model_loader: - kv 30: tokenizer.ggml.add_space_prefix bool = false llama_model_loader: - kv 31: general.quantization_version u32 = 2 llama_model_loader: - type f32: 81 tensors llama_model_loader: - type f16: 281 tensors error loading model: unknown model architecture: 'granite' llama_load_model_from_file: failed to load model AVX = 1 | AVX2 = 1 | AVX512 = 1 | AVX512_VBMI = 1 | AVX512_VNNI = 1 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | 2025-05-17 12:42:14,181 - app - ERROR - Failed to load model: ERROR: Traceback (most recent call last): File "/app/app.py", line 59, in startup_event model = Llama( File "/usr/local/lib/python3.10/site-packages/llama_cpp/llama.py", line 923, in __init__ self._n_vocab = self.n_vocab() File "/usr/local/lib/python3.10/site-packages/llama_cpp/llama.py", line 2184, in n_vocab return self._model.n_vocab() File "/usr/local/lib/python3.10/site-packages/llama_cpp/llama.py", line 250, in n_vocab assert self.model is not None AssertionError During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 677, in lifespan async with self.lifespan_context(app) as maybe_state: File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 566, in __aenter__ await self._router.startup() File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 654, in startup await handler() File "/app/app.py", line 68, in startup_event raise RuntimeError(f"Failed to load model: {e}") RuntimeError: Failed to load model: ERROR: Application startup failed. Exiting.

Container logs:

Fetching error logs...