Subha95 commited on
Commit
9087b24
·
verified ·
1 Parent(s): 05cbae5

Update chatbot_rag.py

Browse files
Files changed (1) hide show
  1. chatbot_rag.py +7 -8
chatbot_rag.py CHANGED
@@ -12,23 +12,22 @@ from langchain_chroma import Chroma
12
  def build_qa():
13
  """Builds and returns the RAG QA pipeline."""
14
 
15
- # 1. Embeddings
16
  embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2")
17
-
18
  vectorstore = Chroma(
19
- persist_directory="db",
20
- collection_name="rag-docs",
21
- embedding_function=embeddings,
22
  )
23
 
24
- # 3. LLM
25
- model_id = "microsoft/phi-2"
26
  tokenizer = AutoTokenizer.from_pretrained(model_id)
27
  model = AutoModelForCausalLM.from_pretrained(model_id)
28
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer, max_new_tokens=512)
29
  llm = HuggingFacePipeline(pipeline=pipe)
30
 
31
- # 4. QA Chain
32
  retriever = vectorstore.as_retriever(search_kwargs={"k": 3})
33
  qa = RetrievalQA.from_chain_type(llm=llm, retriever=retriever, return_source_documents=False)
34
 
 
12
  def build_qa():
13
  """Builds and returns the RAG QA pipeline."""
14
 
15
+ # 1. Load embeddings + DB
16
  embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2")
 
17
  vectorstore = Chroma(
18
+ persist_directory="db",
19
+ collection_name="rag-docs",
20
+ embedding_function=embeddings,
21
  )
22
 
23
+ # 2. LLM (instruction-tuned preferred)
24
+ model_id = "microsoft/phi-3-mini-4k-instruct"
25
  tokenizer = AutoTokenizer.from_pretrained(model_id)
26
  model = AutoModelForCausalLM.from_pretrained(model_id)
27
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer, max_new_tokens=512)
28
  llm = HuggingFacePipeline(pipeline=pipe)
29
 
30
+ # 3. QA Chain
31
  retriever = vectorstore.as_retriever(search_kwargs={"k": 3})
32
  qa = RetrievalQA.from_chain_type(llm=llm, retriever=retriever, return_source_documents=False)
33