Spaces:
Sleeping
Sleeping
Update model.py, change model to gemma-7b-it
Browse files
model.py
CHANGED
|
@@ -44,7 +44,7 @@ def ask(query, temperature=0.7, max_new_tokens=512):
|
|
| 44 |
|
| 45 |
# Load HF token from Secrets (set this in your Space Settings > Secrets)
|
| 46 |
HF_TOKEN = os.getenv("HF_TOKEN")
|
| 47 |
-
model_id = "google/gemma-
|
| 48 |
|
| 49 |
# Load LLM if not already loaded
|
| 50 |
if llm_model is None or tokenizer is None:
|
|
@@ -75,6 +75,7 @@ def ask(query, temperature=0.7, max_new_tokens=512):
|
|
| 75 |
# Clean up output
|
| 76 |
if "Answer:" in output_text:
|
| 77 |
output_text = output_text.split("Answer:")[-1].strip()
|
|
|
|
| 78 |
return output_text
|
| 79 |
|
| 80 |
|
|
|
|
| 44 |
|
| 45 |
# Load HF token from Secrets (set this in your Space Settings > Secrets)
|
| 46 |
HF_TOKEN = os.getenv("HF_TOKEN")
|
| 47 |
+
model_id = "google/gemma-7b-it"
|
| 48 |
|
| 49 |
# Load LLM if not already loaded
|
| 50 |
if llm_model is None or tokenizer is None:
|
|
|
|
| 75 |
# Clean up output
|
| 76 |
if "Answer:" in output_text:
|
| 77 |
output_text = output_text.split("Answer:")[-1].strip()
|
| 78 |
+
output_text = output_text.split("model")[-1].strip()
|
| 79 |
return output_text
|
| 80 |
|
| 81 |
|