Spaces:
Sleeping
Sleeping
Create start.sh
Browse files
start.sh
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/bin/bash
|
| 2 |
+
|
| 3 |
+
# Default to CODER if no variable is set
|
| 4 |
+
# Options: CODER, VISION
|
| 5 |
+
MODE="${MODEL_TYPE:-CODER}"
|
| 6 |
+
|
| 7 |
+
echo "Selected Mode: $MODE"
|
| 8 |
+
|
| 9 |
+
if [ "$MODE" = "VISION" ]; then
|
| 10 |
+
echo "Loading Vision Model..."
|
| 11 |
+
MODEL_FILE="/app/qwen2.5-vl-7b-instruct-q4_k_m.gguf"
|
| 12 |
+
# Lower context for Vision to save RAM
|
| 13 |
+
CONTEXT=4096
|
| 14 |
+
else
|
| 15 |
+
echo "Loading Coder Model..."
|
| 16 |
+
MODEL_FILE="/app/qwen2.5-coder-14b-instruct-q4_k_m.gguf"
|
| 17 |
+
# Higher context for Coding
|
| 18 |
+
CONTEXT=8192
|
| 19 |
+
fi
|
| 20 |
+
|
| 21 |
+
# Run the server
|
| 22 |
+
exec llama-server \
|
| 23 |
+
-m "$MODEL_FILE" \
|
| 24 |
+
--host 0.0.0.0 \
|
| 25 |
+
--port 7860 \
|
| 26 |
+
-c $CONTEXT \
|
| 27 |
+
--n-gpu-layers 0
|