João Lima commited on
Commit
2f98627
·
1 Parent(s): 96dd173

Fix HF Spaces compatibility (models + langchain imports)

Browse files
Files changed (3) hide show
  1. app.py +1 -1
  2. config.py +2 -2
  3. rag/llm.py +1 -1
app.py CHANGED
@@ -36,4 +36,4 @@ with gr.Blocks(title="Tech Explainer RAG") as demo:
36
 
37
 
38
  if __name__ == "__main__":
39
- demo.launch(share=True)
 
36
 
37
 
38
  if __name__ == "__main__":
39
+ demo.launch()
config.py CHANGED
@@ -1,3 +1,3 @@
1
  EMBEDDING_MODEL = "sentence-transformers/all-MiniLM-L6-v2"
2
- PRIMARY_LLM = "mistralai/Mistral-7B-Instruct-v0.2"
3
- FALLBACK_LLM = "microsoft/phi-2"
 
1
  EMBEDDING_MODEL = "sentence-transformers/all-MiniLM-L6-v2"
2
+ PRIMARY_LLM = "google/gemma-2b-it"
3
+ FALLBACK_LLM = "microsoft/phi-2"
rag/llm.py CHANGED
@@ -7,7 +7,7 @@ def load_model():
7
  try:
8
  tokenizer = AutoTokenizer.from_pretrained(PRIMARY_LLM)
9
  model = AutoModelForCausalLM.from_pretrained(
10
- PRIMARY_LLM, device_map="auto", load_in_8bit=True
11
  )
12
  except Exception:
13
  tokenizer = AutoTokenizer.from_pretrained(FALLBACK_LLM, trust_remote_code=True)
 
7
  try:
8
  tokenizer = AutoTokenizer.from_pretrained(PRIMARY_LLM)
9
  model = AutoModelForCausalLM.from_pretrained(
10
+ PRIMARY_LLM, device_map="auto"
11
  )
12
  except Exception:
13
  tokenizer = AutoTokenizer.from_pretrained(FALLBACK_LLM, trust_remote_code=True)