Update Dockerfile
Browse files- Dockerfile +7 -7
Dockerfile
CHANGED
|
@@ -25,7 +25,7 @@ RUN npm install -g openclaw@latest
|
|
| 25 |
# Expose the Hugging Face web port
|
| 26 |
EXPOSE 7860
|
| 27 |
|
| 28 |
-
# Create a robust startup script with Qwen 3
|
| 29 |
RUN echo '#!/bin/bash\n\
|
| 30 |
echo "Writing default OpenClaw configuration..."\n\
|
| 31 |
cat <<EOF > /home/node/.openclaw/openclaw.json\n\
|
|
@@ -51,8 +51,8 @@ cat <<EOF > /home/node/.openclaw/openclaw.json\n\
|
|
| 51 |
"api": "openai-responses",\n\
|
| 52 |
"models": [\n\
|
| 53 |
{\n\
|
| 54 |
-
"id": "qwen3:
|
| 55 |
-
"name": "Qwen 3 (
|
| 56 |
"reasoning": false,\n\
|
| 57 |
"input": ["text"],\n\
|
| 58 |
"cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 },\n\
|
|
@@ -66,10 +66,10 @@ cat <<EOF > /home/node/.openclaw/openclaw.json\n\
|
|
| 66 |
"agents": {\n\
|
| 67 |
"defaults": {\n\
|
| 68 |
"model": {\n\
|
| 69 |
-
"primary": "ollama/qwen3:
|
| 70 |
},\n\
|
| 71 |
"models": {\n\
|
| 72 |
-
"ollama/qwen3:
|
| 73 |
"streaming": false\n\
|
| 74 |
}\n\
|
| 75 |
}\n\
|
|
@@ -86,8 +86,8 @@ while ! ollama list > /dev/null 2>&1; do\n\
|
|
| 86 |
sleep 1\n\
|
| 87 |
done\n\
|
| 88 |
\n\
|
| 89 |
-
echo "Ollama is up! Pulling Qwen 3
|
| 90 |
-
ollama pull qwen3:
|
| 91 |
\n\
|
| 92 |
echo "Model ready! Starting OpenClaw..."\n\
|
| 93 |
openclaw gateway --port 7860 --allow-unconfigured\n\
|
|
|
|
| 25 |
# Expose the Hugging Face web port
|
| 26 |
EXPOSE 7860
|
| 27 |
|
| 28 |
+
# Create a robust startup script with Qwen 3 8B configured
|
| 29 |
RUN echo '#!/bin/bash\n\
|
| 30 |
echo "Writing default OpenClaw configuration..."\n\
|
| 31 |
cat <<EOF > /home/node/.openclaw/openclaw.json\n\
|
|
|
|
| 51 |
"api": "openai-responses",\n\
|
| 52 |
"models": [\n\
|
| 53 |
{\n\
|
| 54 |
+
"id": "qwen3:8b",\n\
|
| 55 |
+
"name": "Qwen 3 (8B)",\n\
|
| 56 |
"reasoning": false,\n\
|
| 57 |
"input": ["text"],\n\
|
| 58 |
"cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 },\n\
|
|
|
|
| 66 |
"agents": {\n\
|
| 67 |
"defaults": {\n\
|
| 68 |
"model": {\n\
|
| 69 |
+
"primary": "ollama/qwen3:8b"\n\
|
| 70 |
},\n\
|
| 71 |
"models": {\n\
|
| 72 |
+
"ollama/qwen3:8b": {\n\
|
| 73 |
"streaming": false\n\
|
| 74 |
}\n\
|
| 75 |
}\n\
|
|
|
|
| 86 |
sleep 1\n\
|
| 87 |
done\n\
|
| 88 |
\n\
|
| 89 |
+
echo "Ollama is up! Pulling Qwen 3 8B..."\n\
|
| 90 |
+
ollama pull qwen3:8b\n\
|
| 91 |
\n\
|
| 92 |
echo "Model ready! Starting OpenClaw..."\n\
|
| 93 |
openclaw gateway --port 7860 --allow-unconfigured\n\
|