Spaces:
Sleeping
Sleeping
Commit ·
db12d3a
1
Parent(s): b49169b
fix
Browse files
app.py
CHANGED
|
@@ -15,7 +15,7 @@ from transformers import (
|
|
| 15 |
import torch.nn.functional as F
|
| 16 |
from collections import defaultdict
|
| 17 |
HF_TOKEN = os.getenv("HF_token")
|
| 18 |
-
CHAT_MODEL_ID = "
|
| 19 |
EMB_MODEL_ID = "mixedbread-ai/mxbai-embed-large-v1"
|
| 20 |
MAX_PROMPT_TOKENS = 8192
|
| 21 |
|
|
@@ -160,7 +160,7 @@ def answer(system: str, context: str, question: str, user_id="demo", history="No
|
|
| 160 |
|
| 161 |
|
| 162 |
except Exception as e:
|
| 163 |
-
return f"Error in app.py: {e}, k={k}, sims.numel()={sims.numel()},
|
| 164 |
finally:
|
| 165 |
torch.cuda.empty_cache()
|
| 166 |
|
|
|
|
| 15 |
import torch.nn.functional as F
|
| 16 |
from collections import defaultdict
|
| 17 |
HF_TOKEN = os.getenv("HF_token")
|
| 18 |
+
CHAT_MODEL_ID = "NousResearch/Meta-Llama-3-8B-Instruct"
|
| 19 |
EMB_MODEL_ID = "mixedbread-ai/mxbai-embed-large-v1"
|
| 20 |
MAX_PROMPT_TOKENS = 8192
|
| 21 |
|
|
|
|
| 160 |
|
| 161 |
|
| 162 |
except Exception as e:
|
| 163 |
+
return f"Error in app.py: {e}, k={k}, sims.numel()={sims.numel()}, sims.shape={sims.shape if 'q_vec' in locals() else 'N/A'}"
|
| 164 |
finally:
|
| 165 |
torch.cuda.empty_cache()
|
| 166 |
|