Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -56,9 +56,10 @@ def setup_server():
|
|
| 56 |
f"./{BINARY_NAME}",
|
| 57 |
"-m", model_path,
|
| 58 |
"--port", SERVER_PORT,
|
| 59 |
-
"--ctx-size", "
|
| 60 |
"--n-gpu-layers", "0",
|
| 61 |
-
"--host", "127.0.0.1"
|
|
|
|
| 62 |
]
|
| 63 |
proc = subprocess.Popen(
|
| 64 |
cmd,
|
|
@@ -94,6 +95,7 @@ def predict(message, history, system_prompt, temperature):
|
|
| 94 |
"messages": messages,
|
| 95 |
"temperature": temperature,
|
| 96 |
"max_tokens": 100,
|
|
|
|
| 97 |
"stream": False
|
| 98 |
}
|
| 99 |
print("\n--- [Request Payload] ---")
|
|
|
|
| 56 |
f"./{BINARY_NAME}",
|
| 57 |
"-m", model_path,
|
| 58 |
"--port", SERVER_PORT,
|
| 59 |
+
"--ctx-size", "8192",
|
| 60 |
"--n-gpu-layers", "0",
|
| 61 |
+
"--host", "127.0.0.1",
|
| 62 |
+
"--no-cache-prompt"
|
| 63 |
]
|
| 64 |
proc = subprocess.Popen(
|
| 65 |
cmd,
|
|
|
|
| 95 |
"messages": messages,
|
| 96 |
"temperature": temperature,
|
| 97 |
"max_tokens": 100,
|
| 98 |
+
"top_p": 0.5,
|
| 99 |
"stream": False
|
| 100 |
}
|
| 101 |
print("\n--- [Request Payload] ---")
|