AIencoder commited on
Commit
d4f638d
·
verified ·
1 Parent(s): b57cd88

Update start.sh

Browse files
Files changed (1) hide show
  1. start.sh +15 -22
start.sh CHANGED
@@ -1,31 +1,25 @@
1
  #!/bin/bash
2
  set -e
3
 
4
- # --- CONFIGURATION ---
5
  echo "------------------------------------------------"
6
  echo "🛠️ Setting up Environment"
7
  echo "------------------------------------------------"
8
 
9
- # Create model directory
10
  mkdir -p /app/models
11
 
12
- download_model () {
13
  REPO_ID=$1
14
  FILENAME=$2
15
- LOCAL_NAME=$3
16
 
17
- if [ -z "$LOCAL_NAME" ]; then LOCAL_NAME="$FILENAME"; fi
18
- FILE_PATH="/app/models/$LOCAL_NAME"
19
-
20
- if [ ! -f "$FILE_PATH" ]; then
21
- echo "⬇️ Downloading $LOCAL_NAME ..."
22
  python3 -c "
23
  from huggingface_hub import hf_hub_download
24
- import os, sys
25
  try:
26
- hf_hub_download(repo_id='$REPO_ID', filename='$FILENAME', local_dir='/app/models')
27
- if '$FILENAME' != '$LOCAL_NAME':
28
- os.rename('/app/models/$FILENAME', '/app/models/$LOCAL_NAME')
29
  except Exception as e:
30
  print(f'Error downloading $FILENAME: {e}')
31
  sys.exit(1)
@@ -33,23 +27,22 @@ except Exception as e:
33
  fi
34
  }
35
 
36
- # 1. TEXT: Qwen 2.5 Coder
37
- download_model "Qwen/Qwen2.5-Coder-14B-Instruct-GGUF" "qwen2.5-coder-14b-instruct-q4_k_m.gguf"
38
-
39
- # 2. IMAGE: Qwen 2.5 VL (correct case-sensitive filename)
40
- download_model "ggml-org/Qwen2.5-VL-7B-Instruct-GGUF" "Qwen2.5-VL-7B-Instruct-Q4_K_M.gguf" "qwen2.5-vl-7b-instruct-q4_k_m.gguf"
41
 
42
- # 3. AUDIO: Qwen 2 Audio
43
- download_model "gaianet/Qwen2-Audio-7B-Instruct-GGUF" "Qwen2-Audio-7B-Instruct-Q4_K_M.gguf" "qwen2-audio-7b-instruct-q4_k_m.gguf"
 
 
 
44
 
45
  echo "------------------------------------------------"
46
  echo "🚀 Starting Server in Router Mode"
47
  echo "------------------------------------------------"
48
 
49
- # Start in ROUTER MODE - serve all models from /app/models
50
  exec /usr/local/bin/llama-server \
51
  --models-dir /app/models \
52
- --models-max 3 \
53
  --host 0.0.0.0 \
54
  --port 7860 \
55
  -c 8192 \
 
1
  #!/bin/bash
2
  set -e
3
 
 
4
  echo "------------------------------------------------"
5
  echo "🛠️ Setting up Environment"
6
  echo "------------------------------------------------"
7
 
 
8
  mkdir -p /app/models
9
 
10
+ download_file () {
11
  REPO_ID=$1
12
  FILENAME=$2
13
+ DEST_PATH=$3
14
 
15
+ if [ ! -f "$DEST_PATH" ]; then
16
+ echo "⬇️ Downloading $FILENAME ..."
 
 
 
17
  python3 -c "
18
  from huggingface_hub import hf_hub_download
19
+ import shutil, sys
20
  try:
21
+ path = hf_hub_download(repo_id='$REPO_ID', filename='$FILENAME')
22
+ shutil.copy(path, '$DEST_PATH')
 
23
  except Exception as e:
24
  print(f'Error downloading $FILENAME: {e}')
25
  sys.exit(1)
 
27
  fi
28
  }
29
 
30
+ # 1. TEXT/CODE: Qwen 2.5 Coder 14B (single file, text-only)
31
+ download_file "Qwen/Qwen2.5-Coder-14B-Instruct-GGUF" "qwen2.5-coder-14b-instruct-q4_k_m.gguf" "/app/models/qwen2.5-coder-14b-instruct-q4_k_m.gguf"
 
 
 
32
 
33
+ # 2. MULTIMODAL (Image + Audio): Qwen 2.5 Omni 7B
34
+ # Needs subdirectory with model + mmproj
35
+ mkdir -p /app/models/qwen2.5-omni-7b
36
+ download_file "ggml-org/Qwen2.5-Omni-7B-GGUF" "Qwen2.5-Omni-7B-Q4_K_M.gguf" "/app/models/qwen2.5-omni-7b/Qwen2.5-Omni-7B-Q4_K_M.gguf"
37
+ download_file "ggml-org/Qwen2.5-Omni-7B-GGUF" "mmproj-Qwen2.5-Omni-7B-Q8_0.gguf" "/app/models/qwen2.5-omni-7b/mmproj-Qwen2.5-Omni-7B-Q8_0.gguf"
38
 
39
  echo "------------------------------------------------"
40
  echo "🚀 Starting Server in Router Mode"
41
  echo "------------------------------------------------"
42
 
 
43
  exec /usr/local/bin/llama-server \
44
  --models-dir /app/models \
45
+ --models-max 2 \
46
  --host 0.0.0.0 \
47
  --port 7860 \
48
  -c 8192 \