Spaces:
Sleeping
Sleeping
Update start.sh
Browse files
start.sh
CHANGED
|
@@ -4,34 +4,35 @@
|
|
| 4 |
export WORK="/home/user/app"
|
| 5 |
cd "$WORK" || exit 1
|
| 6 |
|
| 7 |
-
# 建立 Python 虛擬環境
|
| 8 |
python3 -m venv venv
|
| 9 |
-
|
| 10 |
-
# 啟動虛擬環境
|
| 11 |
source venv/bin/activate
|
| 12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
echo "Unzip and download model..."
|
| 14 |
unzip llama_cpp_avx512.zip > /dev/null 2>&1
|
| 15 |
#wget -O model.gguf https://huggingface.co/Intel/gpt-oss-20b-gguf-q4ks-AutoRound/resolve/main/gpt-oss-20b-32x2.4B-Q4_K_S.gguf > /dev/null 2>&1
|
| 16 |
wget -O model.gguf https://huggingface.co/unsloth/gpt-oss-20b-GGUF/resolve/main/gpt-oss-20b-Q4_K_M.gguf > /dev/null 2>&1
|
| 17 |
|
| 18 |
-
|
| 19 |
echo "Start llama.cpp server..."
|
| 20 |
-
./llama-server
|
| 21 |
-
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
|
|
|
|
|
|
| 33 |
|
| 34 |
echo "Start app.py..."
|
| 35 |
python app.py
|
| 36 |
-
|
| 37 |
-
#python3 -m http.server 7860 --bind 0.0.0.0
|
|
|
|
| 4 |
export WORK="/home/user/app"
|
| 5 |
cd "$WORK" || exit 1
|
| 6 |
|
| 7 |
+
# 建立 Python 虛擬環境 (可選)
|
| 8 |
python3 -m venv venv
|
|
|
|
|
|
|
| 9 |
source venv/bin/activate
|
| 10 |
|
| 11 |
+
# 安裝 Python 套件
|
| 12 |
+
pip install --upgrade pip
|
| 13 |
+
pip install gradio openai
|
| 14 |
+
|
| 15 |
echo "Unzip and download model..."
|
| 16 |
unzip llama_cpp_avx512.zip > /dev/null 2>&1
|
| 17 |
#wget -O model.gguf https://huggingface.co/Intel/gpt-oss-20b-gguf-q4ks-AutoRound/resolve/main/gpt-oss-20b-32x2.4B-Q4_K_S.gguf > /dev/null 2>&1
|
| 18 |
wget -O model.gguf https://huggingface.co/unsloth/gpt-oss-20b-GGUF/resolve/main/gpt-oss-20b-Q4_K_M.gguf > /dev/null 2>&1
|
| 19 |
|
|
|
|
| 20 |
echo "Start llama.cpp server..."
|
| 21 |
+
nohup ./llama-server \
|
| 22 |
+
-m model.gguf \
|
| 23 |
+
--port 8000 \
|
| 24 |
+
--host 0.0.0.0 \
|
| 25 |
+
--threads 2 \
|
| 26 |
+
--ctx-size 4096 \
|
| 27 |
+
--mlock \
|
| 28 |
+
--jinja \
|
| 29 |
+
--temp 0.2 \
|
| 30 |
+
--top-p 0.85 \
|
| 31 |
+
-v \
|
| 32 |
+
> llama.log 2>&1 &
|
| 33 |
+
|
| 34 |
+
# 持續監控 log,但不阻塞 python app.py
|
| 35 |
+
tail -n 50 -f llama.log &
|
| 36 |
|
| 37 |
echo "Start app.py..."
|
| 38 |
python app.py
|
|
|
|
|
|