Update Dockerfile
Browse files- Dockerfile +5 -16
Dockerfile
CHANGED
|
@@ -1,24 +1,13 @@
|
|
| 1 |
FROM ollama/ollama:latest
|
| 2 |
|
| 3 |
-
# HF Spaces
|
| 4 |
ENV OLLAMA_HOST=0.0.0.0:7860
|
| 5 |
ENV OLLAMA_MAX_LOADED_MODELS=1
|
| 6 |
|
| 7 |
EXPOSE 7860
|
| 8 |
|
| 9 |
-
#
|
| 10 |
-
|
| 11 |
-
echo '🚀 Starting Ollama API server (Qwen3 8B - free CPU)...' && \
|
| 12 |
-
ollama serve & \
|
| 13 |
-
sleep 10 && \
|
| 14 |
-
echo '📥 Pulling Qwen3 8B (q4_K_M quantized)...' && \
|
| 15 |
-
ollama pull qwen3:8b-q4_K_M && \
|
| 16 |
-
echo '✅ Qwen3 8B is ready! API available at https://your-space.hf.space' && \
|
| 17 |
-
echo ' Model name: qwen3:8b-q4_K_M' && \
|
| 18 |
-
tail -f /dev/null \
|
| 19 |
-
"]
|
| 20 |
|
| 21 |
-
#
|
| 22 |
-
|
| 23 |
-
# qwen3:8b-q5_K_M (better quality, slightly more RAM)
|
| 24 |
-
# qwen3:8b-q4_0 (lighter)
|
|
|
|
| 1 |
FROM ollama/ollama:latest
|
| 2 |
|
| 3 |
+
# HF Spaces (free CPU) expects port 7860
|
| 4 |
ENV OLLAMA_HOST=0.0.0.0:7860
|
| 5 |
ENV OLLAMA_MAX_LOADED_MODELS=1
|
| 6 |
|
| 7 |
EXPOSE 7860
|
| 8 |
|
| 9 |
+
# Override default entrypoint so we can run normal commands
|
| 10 |
+
ENTRYPOINT []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
|
| 12 |
+
# Start Ollama server + pull Qwen3 8B (quantized for CPU) + keep alive
|
| 13 |
+
CMD ollama serve & sleep 15 && ollama pull qwen3:8b-q4_K_M && echo "✅ Qwen3 8B ready!" && tail -f /dev/null
|
|
|
|
|
|