triflix commited on
Commit
6f4f407
·
verified ·
1 Parent(s): 16d5d1f

Create start.sh

Browse files
Files changed (1) hide show
  1. start.sh +22 -0
start.sh ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+
3
+ # Start llama-server in background
4
+ cd /llama.cpp/build
5
+ ./bin/llama-server \
6
+ --host 0.0.0.0 \
7
+ --port 8080 \
8
+ --model /models/model.gguf \
9
+ --ctx-size 32768 \
10
+ --threads 2 &
11
+
12
+ # Wait for llama server
13
+ echo "Waiting for llama.cpp server..."
14
+ until curl -s "http://localhost:8080/v1/models" >/dev/null 2>&1; do
15
+ sleep 1
16
+ done
17
+ echo "llama.cpp server is ready."
18
+
19
+ # Start FastAPI
20
+ echo "Starting FastAPI server on port 7860..."
21
+ cd /
22
+ python3 -m uvicorn app:app --host 0.0.0.0 --port 7860