RohanSardar commited on
Commit
9ccdf69
·
verified ·
1 Parent(s): 3d79f12

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -13,7 +13,7 @@ from langchain_community.document_loaders import PyPDFDirectoryLoader
13
  load_dotenv()
14
  groq_api_key = os.getenv("GROQ_API_KEY")
15
 
16
- llm = ChatGroq(groq_api_key=groq_api_key, model="Llama3-8b-8192")
17
 
18
  prompt = ChatPromptTemplate.from_template(
19
  """
@@ -30,7 +30,7 @@ st.title("Simple RAG Application")
30
 
31
  def create_vector_embedding():
32
  if "vectors" not in st.session_state:
33
- st.session_state.embeddings = HuggingFaceBgeEmbeddings(model_name="all-MiniLM-L6-v2")
34
  st.session_state.loader = PyPDFDirectoryLoader("documents")
35
  st.session_state.docs = st.session_state.loader.load()
36
  st.session_state.text_splitter = RecursiveCharacterTextSplitter(chunk_size=1000, chunk_overlap=200)
@@ -47,7 +47,7 @@ if "vectors" not in st.session_state:
47
  if "vectors" in st.session_state:
48
  user_prompt = st.text_input("Enter your query here")
49
  if user_prompt:
50
- document_chain = create_stuff_documents_chain(llm, prompt)
51
  retriever = st.session_state.vectors.as_retriever()
52
  retrieval_chain = create_retrieval_chain(retriever, document_chain)
53
  response = retrieval_chain.invoke({'input': user_prompt})
 
13
  load_dotenv()
14
  groq_api_key = os.getenv("GROQ_API_KEY")
15
 
16
+ model = ChatGroq(groq_api_key=groq_api_key, model="Llama3-8b-8192")
17
 
18
  prompt = ChatPromptTemplate.from_template(
19
  """
 
30
 
31
  def create_vector_embedding():
32
  if "vectors" not in st.session_state:
33
+ st.session_state.embeddings = HuggingFaceBgeEmbeddings(model_name="mxbai-embed-large-v1")
34
  st.session_state.loader = PyPDFDirectoryLoader("documents")
35
  st.session_state.docs = st.session_state.loader.load()
36
  st.session_state.text_splitter = RecursiveCharacterTextSplitter(chunk_size=1000, chunk_overlap=200)
 
47
  if "vectors" in st.session_state:
48
  user_prompt = st.text_input("Enter your query here")
49
  if user_prompt:
50
+ document_chain = create_stuff_documents_chain(model, prompt)
51
  retriever = st.session_state.vectors.as_retriever()
52
  retrieval_chain = create_retrieval_chain(retriever, document_chain)
53
  response = retrieval_chain.invoke({'input': user_prompt})