ChristopherMarais commited on
Commit
cdcbde5
·
verified ·
1 Parent(s): 2da3186

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -3
app.py CHANGED
@@ -63,7 +63,8 @@ def initialize_qa_chain(temperature, max_tokens, top_k, vector_db):
63
  raise ValueError("Missing HF_TOKEN environment variable!")
64
 
65
  llm = HuggingFaceEndpoint(
66
- repo_id="deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
 
67
  huggingfacehub_api_token=HF_TOKEN, # Only needed if the model endpoint requires authentication
68
  temperature=temperature,
69
  max_new_tokens=max_tokens,
@@ -188,8 +189,10 @@ HF_TOKEN = os.getenv("AMAbot_r", "") # use for publishing
188
  if not HF_TOKEN:
189
  raise ValueError("Missing HF_TOKEN environment variable!")
190
  # Global InferenceClient for plain chat (fallback)
191
- client = InferenceClient("deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
192
- token=HF_TOKEN)
 
 
193
 
194
  # --- Auto-load vector database and initialize QA chain at startup ---
195
  try:
 
63
  raise ValueError("Missing HF_TOKEN environment variable!")
64
 
65
  llm = HuggingFaceEndpoint(
66
+ # repo_id="deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
67
+ repo_id="Qwen/Qwen2.5-1.5B-Instruct",
68
  huggingfacehub_api_token=HF_TOKEN, # Only needed if the model endpoint requires authentication
69
  temperature=temperature,
70
  max_new_tokens=max_tokens,
 
189
  if not HF_TOKEN:
190
  raise ValueError("Missing HF_TOKEN environment variable!")
191
  # Global InferenceClient for plain chat (fallback)
192
+ client = InferenceClient(
193
+ # "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
194
+ "Qwen/Qwen2.5-1.5B-Instruct",
195
+ token=HF_TOKEN)
196
 
197
  # --- Auto-load vector database and initialize QA chain at startup ---
198
  try: