LLAMACPP_WEBUI01 / start.sh
hsuwill000's picture
Update start.sh
a86d567 verified
#!/bin/bash
set -e
# ==========================
# 工作目錄設定
# ==========================
export WORK="/home/user/app"
SERVER_BIN="$WORK/llama-server"
cd "$WORK"
unzip llama_cpp_avx512_17565.zip
echo "PWD"
pwd
#wget https://huggingface.co/unsloth/Qwen3-0.6B-GGUF/resolve/main/Qwen3-0.6B-Q5_K_M.gguf
#wget https://huggingface.co/unsloth/Qwen3-0.6B-GGUF/resolve/main/Qwen3-0.6B-Q2_K.gguf
#wget https://huggingface.co/unsloth/LFM2-8B-A1B-GGUF/resolve/main/LFM2-8B-A1B-UD-Q4_K_XL.gguf
#wget https://huggingface.co/unsloth/ERNIE-4.5-21B-A3B-Thinking-GGUF/resolve/main/ERNIE-4.5-21B-A3B-Thinking-UD-IQ2_M.gguf
wget https://huggingface.co/unsloth/LFM2.5-1.2B-Instruct-GGUF/resolve/main/LFM2.5-1.2B-Instruct-Q4_0.gguf
echo "ls -l"
ls -l
# --models-dir $WORK \
$SERVER_BIN \
--model LFM2.5-1.2B-Instruct-Q4_0.gguf \
--threads 2 \
--ctx-size 4096 \
--mlock \
--temp 1.1 \
--top-p 0.95 \
--jinja \
--chat-template "" \
--host 0.0.0.0 \
--port 7860 \
--models-max 1 \
--verbose