dxnxk commited on
Commit
02bf09e
·
verified ·
1 Parent(s): 263149c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +21 -32
app.py CHANGED
@@ -26,38 +26,27 @@ index.add(embeddings)
26
  client = InferenceClient("mistralai/Mistral-7B-Instruct-v0.2")
27
 
28
  # --- RAG pipeline ---
29
- def generate_answer(messages):
30
- try:
31
- sys.stderr.write("=== generate_answer called ===\n")
32
- sys.stderr.flush()
33
-
34
- user_query = messages[-1]["content"]
35
- query_embedding = embedding_model.encode([user_query], convert_to_numpy=True)
36
- faiss.normalize_L2(query_embedding)
37
- _, indices = index.search(query_embedding, k=5)
38
-
39
- context = "\n".join([f"{codes[i]}: {descriptions[i]}" for i in indices[0]])
40
- prompt = f"""Here are some tariff code descriptions:
41
- {context}
42
-
43
- Question: {user_query}
44
- Answer:"""
45
-
46
- sys.stderr.write(f"Prompt sent to model:\n{prompt}\n")
47
- sys.stderr.flush()
48
-
49
- response = client.text_generation(
50
- prompt,
51
- max_new_tokens=200,
52
- temperature=0.7,
53
- stop_sequences=["\n\n"]
54
- )
55
- return {"role": "assistant", "content": response.strip()}
56
-
57
- except Exception as e:
58
- sys.stderr.write(f"Error during inference: {type(e).__name__}: {e}\n")
59
- sys.stderr.flush()
60
- return {"role": "assistant", "content": "An internal error occurred. Please try again."}
61
 
62
  # --- Gradio Chat Interface ---
63
  gr.ChatInterface(
 
26
  client = InferenceClient("mistralai/Mistral-7B-Instruct-v0.2")
27
 
28
  # --- RAG pipeline ---
29
+ def generate_answer(user_query):
30
+ sys.stderr.write("=== generate_answer called ===\n")
31
+ sys.stderr.flush()
32
+
33
+ query_embedding = embedding_model.encode([user_query], convert_to_numpy=True)
34
+ faiss.normalize_L2(query_embedding)
35
+ _, indices = index.search(query_embedding, k=5)
36
+
37
+ context = "\n".join([f"{codes[i]}: {descriptions[i]}" for i in indices[0]])
38
+ prompt = f"""Here are some tariff code descriptions:\n{context}\n\nQuestion: {user_query}\nAnswer:"""
39
+
40
+ sys.stderr.write(f"Prompt sent to model:\n{prompt}\n")
41
+ sys.stderr.flush()
42
+
43
+ response = client.text_generation(
44
+ prompt,
45
+ max_new_tokens=200,
46
+ temperature=0.7,
47
+ stop_sequences=["\n\n"]
48
+ )
49
+ return response.strip()
 
 
 
 
 
 
 
 
 
 
 
50
 
51
  # --- Gradio Chat Interface ---
52
  gr.ChatInterface(