|
|
|
|
|
|
|
|
from langchain.text_splitter import RecursiveCharacterTextSplitter |
|
|
from langchain.vectorstores import FAISS |
|
|
from langchain.chains import RetrievalQA |
|
|
from langchain_community.embeddings import HuggingFaceEmbeddings |
|
|
from langchain_groq import ChatGroq |
|
|
|
|
|
def create_vectorstore_from_text(text: str): |
|
|
splitter = RecursiveCharacterTextSplitter(chunk_size=500, chunk_overlap=50) |
|
|
texts = splitter.split_text(text) |
|
|
|
|
|
embeddings = HuggingFaceEmbeddings( |
|
|
model_name="sentence-transformers/all-MiniLM-L6-v2", |
|
|
model_kwargs={"device": "cpu"} |
|
|
) |
|
|
|
|
|
vectorstore = FAISS.from_texts(texts, embedding=embeddings) |
|
|
return vectorstore |
|
|
|
|
|
def create_rag_chain(vectorstore): |
|
|
retriever = vectorstore.as_retriever(search_kwargs={"k": 3}) |
|
|
|
|
|
llm = ChatGroq(model_name="llama3-8b-8192", temperature=0) |
|
|
|
|
|
rag_chain = RetrievalQA.from_chain_type(llm=llm, retriever=retriever) |
|
|
return rag_chain |
|
|
|