Update start.sh
Browse files
start.sh
CHANGED
|
@@ -2,7 +2,8 @@
|
|
| 2 |
|
| 3 |
export WORK="/home/user/app"
|
| 4 |
cd $WORK
|
| 5 |
-
unzip llama_cpp_avx512.zip > /dev/null 2>&1
|
| 6 |
-
|
|
|
|
| 7 |
# wget -O model.gguf https://huggingface.co/unsloth/Qwen3-30B-A3B-Instruct-2507-GGUF/resolve/main/Qwen3-30B-A3B-Instruct-2507-UD-TQ1_0.gguf > /dev/null 2>&1
|
| 8 |
./llama-server -m model.gguf --port 8000 --host 0.0.0.0 --threads 2 --ctx-size 4096 --mlock --jinja
|
|
|
|
| 2 |
|
| 3 |
export WORK="/home/user/app"
|
| 4 |
cd $WORK
|
| 5 |
+
unzip llama_cpp_avx512.zip > /dev/null 2>&1
|
| 6 |
+
echo "Booting up llama server..."
|
| 7 |
+
wget -O model.gguf https://huggingface.co/lmstudio-community/Qwen3-4B-Instruct-2507-GGUF/resolve/main/Qwen3-4B-Instruct-2507-Q6_K.gguf?download=true > /dev/null 2>&1
|
| 8 |
# wget -O model.gguf https://huggingface.co/unsloth/Qwen3-30B-A3B-Instruct-2507-GGUF/resolve/main/Qwen3-30B-A3B-Instruct-2507-UD-TQ1_0.gguf > /dev/null 2>&1
|
| 9 |
./llama-server -m model.gguf --port 8000 --host 0.0.0.0 --threads 2 --ctx-size 4096 --mlock --jinja
|