Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -63,7 +63,8 @@ def initialize_qa_chain(temperature, max_tokens, top_k, vector_db):
|
|
| 63 |
raise ValueError("Missing HF_TOKEN environment variable!")
|
| 64 |
|
| 65 |
llm = HuggingFaceEndpoint(
|
| 66 |
-
repo_id="deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
|
|
|
|
| 67 |
huggingfacehub_api_token=HF_TOKEN, # Only needed if the model endpoint requires authentication
|
| 68 |
temperature=temperature,
|
| 69 |
max_new_tokens=max_tokens,
|
|
@@ -188,8 +189,10 @@ HF_TOKEN = os.getenv("AMAbot_r", "") # use for publishing
|
|
| 188 |
if not HF_TOKEN:
|
| 189 |
raise ValueError("Missing HF_TOKEN environment variable!")
|
| 190 |
# Global InferenceClient for plain chat (fallback)
|
| 191 |
-
client = InferenceClient(
|
| 192 |
-
|
|
|
|
|
|
|
| 193 |
|
| 194 |
# --- Auto-load vector database and initialize QA chain at startup ---
|
| 195 |
try:
|
|
|
|
| 63 |
raise ValueError("Missing HF_TOKEN environment variable!")
|
| 64 |
|
| 65 |
llm = HuggingFaceEndpoint(
|
| 66 |
+
# repo_id="deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
|
| 67 |
+
repo_id="Qwen/Qwen2.5-1.5B-Instruct",
|
| 68 |
huggingfacehub_api_token=HF_TOKEN, # Only needed if the model endpoint requires authentication
|
| 69 |
temperature=temperature,
|
| 70 |
max_new_tokens=max_tokens,
|
|
|
|
| 189 |
if not HF_TOKEN:
|
| 190 |
raise ValueError("Missing HF_TOKEN environment variable!")
|
| 191 |
# Global InferenceClient for plain chat (fallback)
|
| 192 |
+
client = InferenceClient(
|
| 193 |
+
# "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
|
| 194 |
+
"Qwen/Qwen2.5-1.5B-Instruct",
|
| 195 |
+
token=HF_TOKEN)
|
| 196 |
|
| 197 |
# --- Auto-load vector database and initialize QA chain at startup ---
|
| 198 |
try:
|