Spaces:
Sleeping
Sleeping
Remove --stop argument (not supported in llama.cpp CLI)
Browse files
llm_clients/qwen_translator.py
CHANGED
|
@@ -356,8 +356,7 @@ class QwenTranslatorClient(LlmClient):
|
|
| 356 |
if self.n_gpu_layers > 0:
|
| 357 |
cmd.extend(["-ngl", str(self.n_gpu_layers)])
|
| 358 |
|
| 359 |
-
#
|
| 360 |
-
cmd.extend(["--stop", "<|im_end|>", "--stop", "<|im_start|>"])
|
| 361 |
|
| 362 |
try:
|
| 363 |
# Run the binary and capture output
|
|
|
|
| 356 |
if self.n_gpu_layers > 0:
|
| 357 |
cmd.extend(["-ngl", str(self.n_gpu_layers)])
|
| 358 |
|
| 359 |
+
# Note: Stop sequences are handled in post-processing since --stop may not be available in all llama.cpp versions
|
|
|
|
| 360 |
|
| 361 |
try:
|
| 362 |
# Run the binary and capture output
|