| # Inference backend: "ollama" (default) or "transformers" | |
| INFERENCE_BACKEND=ollama | |
| # Ollama settings | |
| OLLAMA_URL=http://localhost:11434 | |
| OLLAMA_MODEL=qwen2.5:0.5b | |
| # Transformers.js model (if used) | |
| TRANSFORMERS_MODEL=qwen2.5:0.5b | |
| # Inference backend: "ollama" (default) or "transformers" | |
| INFERENCE_BACKEND=ollama | |
| # Ollama settings | |
| OLLAMA_URL=http://localhost:11434 | |
| OLLAMA_MODEL=qwen2.5:0.5b | |
| # Transformers.js model (if used) | |
| TRANSFORMERS_MODEL=qwen2.5:0.5b | |