Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -85,9 +85,12 @@ def run_gpt4(req1, req2, prompt_type, api_key):
|
|
| 85 |
return completion.choices[0].message.content.strip()
|
| 86 |
|
| 87 |
def run_deepseek(req1, req2, prompt_type):
|
|
|
|
| 88 |
tokenizer = app.state.deepseek_tokenizer
|
| 89 |
model = app.state.deepseek_model
|
|
|
|
| 90 |
prompt = build_prompt(req1, req2, prompt_type)
|
|
|
|
| 91 |
inputs = tokenizer([prompt], return_tensors="pt").to(model.device)
|
| 92 |
outputs = model.generate(inputs.input_ids, max_new_tokens=256)
|
| 93 |
return tokenizer.decode(outputs[0], skip_special_tokens=True)
|
|
@@ -122,7 +125,7 @@ def predict(request: ConflictDetectionRequest):
|
|
| 122 |
|
| 123 |
elif request.model_choice == "DeepSeek-Reasoner":
|
| 124 |
answer = run_deepseek(request.Req1, request.Req2, request.prompt_type)
|
| 125 |
-
|
| 126 |
elif request.model_choice == "LLaMA-3.1-8B-Instruct":
|
| 127 |
if not hasattr(app.state, "llama_model"):
|
| 128 |
return JSONResponse({"error": "LLaMA not loaded (missing HF_TOKEN)"}, status_code=400)
|
|
|
|
| 85 |
return completion.choices[0].message.content.strip()
|
| 86 |
|
| 87 |
def run_deepseek(req1, req2, prompt_type):
|
| 88 |
+
print("Start run deepseek")
|
| 89 |
tokenizer = app.state.deepseek_tokenizer
|
| 90 |
model = app.state.deepseek_model
|
| 91 |
+
print("Start prompt building")
|
| 92 |
prompt = build_prompt(req1, req2, prompt_type)
|
| 93 |
+
print("The prompt is ", prompt)
|
| 94 |
inputs = tokenizer([prompt], return_tensors="pt").to(model.device)
|
| 95 |
outputs = model.generate(inputs.input_ids, max_new_tokens=256)
|
| 96 |
return tokenizer.decode(outputs[0], skip_special_tokens=True)
|
|
|
|
| 125 |
|
| 126 |
elif request.model_choice == "DeepSeek-Reasoner":
|
| 127 |
answer = run_deepseek(request.Req1, request.Req2, request.prompt_type)
|
| 128 |
+
print("Deepseek answer is : ", answer)
|
| 129 |
elif request.model_choice == "LLaMA-3.1-8B-Instruct":
|
| 130 |
if not hasattr(app.state, "llama_model"):
|
| 131 |
return JSONResponse({"error": "LLaMA not loaded (missing HF_TOKEN)"}, status_code=400)
|