Spaces:
Sleeping
Sleeping
João Lima commited on
Commit ·
2f98627
1
Parent(s): 96dd173
Fix HF Spaces compatibility (models + langchain imports)
Browse files- app.py +1 -1
- config.py +2 -2
- rag/llm.py +1 -1
app.py
CHANGED
|
@@ -36,4 +36,4 @@ with gr.Blocks(title="Tech Explainer RAG") as demo:
|
|
| 36 |
|
| 37 |
|
| 38 |
if __name__ == "__main__":
|
| 39 |
-
demo.launch(
|
|
|
|
| 36 |
|
| 37 |
|
| 38 |
if __name__ == "__main__":
|
| 39 |
+
demo.launch()
|
config.py
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
EMBEDDING_MODEL = "sentence-transformers/all-MiniLM-L6-v2"
|
| 2 |
-
PRIMARY_LLM = "
|
| 3 |
-
FALLBACK_LLM = "microsoft/phi-2"
|
|
|
|
| 1 |
EMBEDDING_MODEL = "sentence-transformers/all-MiniLM-L6-v2"
|
| 2 |
+
PRIMARY_LLM = "google/gemma-2b-it"
|
| 3 |
+
FALLBACK_LLM = "microsoft/phi-2"
|
rag/llm.py
CHANGED
|
@@ -7,7 +7,7 @@ def load_model():
|
|
| 7 |
try:
|
| 8 |
tokenizer = AutoTokenizer.from_pretrained(PRIMARY_LLM)
|
| 9 |
model = AutoModelForCausalLM.from_pretrained(
|
| 10 |
-
PRIMARY_LLM, device_map="auto"
|
| 11 |
)
|
| 12 |
except Exception:
|
| 13 |
tokenizer = AutoTokenizer.from_pretrained(FALLBACK_LLM, trust_remote_code=True)
|
|
|
|
| 7 |
try:
|
| 8 |
tokenizer = AutoTokenizer.from_pretrained(PRIMARY_LLM)
|
| 9 |
model = AutoModelForCausalLM.from_pretrained(
|
| 10 |
+
PRIMARY_LLM, device_map="auto"
|
| 11 |
)
|
| 12 |
except Exception:
|
| 13 |
tokenizer = AutoTokenizer.from_pretrained(FALLBACK_LLM, trust_remote_code=True)
|