anshumanpatil commited on
Commit
ad65e68
Β·
1 Parent(s): 25005d0
Files changed (1) hide show
  1. app.py +13 -21
app.py CHANGED
@@ -20,10 +20,6 @@ docs = []
20
  db = None
21
  extracted_text = None
22
 
23
- # ------------------------------
24
- # Title
25
- # ------------------------------
26
- st.title("πŸ“š RAG For MSCI Indexes")
27
 
28
  # ------------------------------
29
  # Load Model for pretraining
@@ -47,35 +43,31 @@ def extract_text():
47
  documents = loader.load()
48
  return "\n".join([doc.page_content for doc in documents])
49
 
50
-
51
- with st.spinner("πŸ”„ Loading Model..."):
52
- generator = load_model()
53
- with st.spinner("πŸ”„ Loading Knowldge Base..."):
54
- extracted_text = extract_text()
55
-
56
- # ------------------------------
57
- # Extract Text
58
- # ------------------------------
59
-
60
-
61
-
62
- # ------------------------------
63
- # Build FAISS Index
64
- # ------------------------------
65
  @st.cache_resource
66
  def build_faiss(_docs):
67
  embeddings = HuggingFaceEmbeddings(model_name=embedding_model_name)
68
  return FAISS.from_documents(_docs, embeddings)
69
 
70
- if extracted_text:
 
 
 
 
71
  splitter = RecursiveCharacterTextSplitter(chunk_size=500, chunk_overlap=50)
72
  docs = [Document(page_content=chunk) for chunk in splitter.split_text(extracted_text)]
73
  db = build_faiss(docs)
74
  st.success("βœ… Knowledge Base ready! From :- https://www.msci.com/indexes#featured-indexes")
75
 
 
 
 
 
 
 
 
76
  query = st.text_input("πŸ’¬ Ask a question about MSCI Indexes", placeholder="MSCI World IMI Index")
77
 
78
- if query and db:
79
  retriever = db.as_retriever(search_kwargs={"k": 3})
80
  retrieved_docs = retriever.get_relevant_documents(query)
81
  context = "\n".join([doc.page_content for doc in retrieved_docs])
 
20
  db = None
21
  extracted_text = None
22
 
 
 
 
 
23
 
24
  # ------------------------------
25
  # Load Model for pretraining
 
43
  documents = loader.load()
44
  return "\n".join([doc.page_content for doc in documents])
45
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46
  @st.cache_resource
47
  def build_faiss(_docs):
48
  embeddings = HuggingFaceEmbeddings(model_name=embedding_model_name)
49
  return FAISS.from_documents(_docs, embeddings)
50
 
51
+
52
+ with st.spinner("πŸ”„ Loading Model..."):
53
+ generator = load_model()
54
+ with st.spinner("πŸ”„ Loading Knowldge Base..."):
55
+ extracted_text = extract_text()
56
  splitter = RecursiveCharacterTextSplitter(chunk_size=500, chunk_overlap=50)
57
  docs = [Document(page_content=chunk) for chunk in splitter.split_text(extracted_text)]
58
  db = build_faiss(docs)
59
  st.success("βœ… Knowledge Base ready! From :- https://www.msci.com/indexes#featured-indexes")
60
 
61
+
62
+ # ------------------------------
63
+ # Title
64
+ # ------------------------------
65
+ st.title("πŸ“š RAG For MSCI Indexes")
66
+ st.markdown("This app uses a local LLM model to answer questions about MSCI Indexes using RAG (Retrieval Augmented Generation).")
67
+
68
  query = st.text_input("πŸ’¬ Ask a question about MSCI Indexes", placeholder="MSCI World IMI Index")
69
 
70
+ if query and db and extracted_text and len(docs) > 0:
71
  retriever = db.as_retriever(search_kwargs={"k": 3})
72
  retrieved_docs = retriever.get_relevant_documents(query)
73
  context = "\n".join([doc.page_content for doc in retrieved_docs])