Spaces:
Sleeping
Sleeping
Update start.sh
Browse files
start.sh
CHANGED
|
@@ -10,30 +10,11 @@ source venv/bin/activate
|
|
| 10 |
|
| 11 |
# 安裝 Python 套件
|
| 12 |
pip install --upgrade pip
|
| 13 |
-
pip install
|
| 14 |
|
| 15 |
-
echo "Unzip and download model..."
|
| 16 |
-
unzip llama_cpp_avx512.zip > /dev/null 2>&1
|
| 17 |
-
#wget -O model.gguf https://huggingface.co/Intel/gpt-oss-20b-gguf-q4ks-AutoRound/resolve/main/gpt-oss-20b-32x2.4B-Q4_K_S.gguf > /dev/null 2>&1
|
| 18 |
-
wget -O model.gguf https://huggingface.co/unsloth/gpt-oss-20b-GGUF/resolve/main/gpt-oss-20b-Q4_K_M.gguf > /dev/null 2>&1
|
| 19 |
|
| 20 |
-
|
| 21 |
-
nohup ./llama-server \
|
| 22 |
-
-m model.gguf \
|
| 23 |
-
--port 8000 \
|
| 24 |
-
--host 0.0.0.0 \
|
| 25 |
-
--threads 2 \
|
| 26 |
-
--ctx-size 16384 \
|
| 27 |
-
--mlock \
|
| 28 |
-
--jinja \
|
| 29 |
-
--temp 0.2 \
|
| 30 |
-
--top-p 0.85 \
|
| 31 |
-
-v \
|
| 32 |
-
> llama.log 2>&1 &
|
| 33 |
|
| 34 |
-
# 持續監控 log,但不阻塞 python app.py
|
| 35 |
-
#tail -n 50 -f llama.log &
|
| 36 |
-
#sleep 200
|
| 37 |
|
| 38 |
echo "Start app.py..."
|
| 39 |
python app.py
|
|
|
|
| 10 |
|
| 11 |
# 安裝 Python 套件
|
| 12 |
pip install --upgrade pip
|
| 13 |
+
pip install fastapi uvicornp
|
| 14 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
|
| 16 |
+
uvicorn app:app --host 0.0.0.0 --port 7860
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
|
|
|
|
|
|
|
|
|
|
| 18 |
|
| 19 |
echo "Start app.py..."
|
| 20 |
python app.py
|