xoa-the-noob commited on
Commit
5aa0be0
·
1 Parent(s): 90de146

a new beginning

Browse files
.gitattributes CHANGED
@@ -1,35 +1,2 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ faiss_index/index.faiss filter=lfs diff=lfs merge=lfs -text
2
+ faiss_index/index.pkl filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.gitignore ADDED
@@ -0,0 +1,216 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Byte-compiled / optimized / DLL files
2
+ __pycache__/
3
+ *.py[codz]
4
+ *$py.class
5
+
6
+ # C extensions
7
+ *.so
8
+
9
+ # Distribution / packaging
10
+ .Python
11
+ build/
12
+ develop-eggs/
13
+ dist/
14
+ downloads/
15
+ eggs/
16
+ .eggs/
17
+ lib/
18
+ lib64/
19
+ parts/
20
+ sdist/
21
+ var/
22
+ wheels/
23
+ share/python-wheels/
24
+ *.egg-info/
25
+ .installed.cfg
26
+ *.egg
27
+ MANIFEST
28
+
29
+ # PyInstaller
30
+ # Usually these files are written by a python script from a template
31
+ # before PyInstaller builds the exe, so as to inject date/other infos into it.
32
+ *.manifest
33
+ *.spec
34
+
35
+ # Installer logs
36
+ pip-log.txt
37
+ pip-delete-this-directory.txt
38
+
39
+ # Unit test / coverage reports
40
+ htmlcov/
41
+ .tox/
42
+ .nox/
43
+ .coverage
44
+ .coverage.*
45
+ .cache
46
+ nosetests.xml
47
+ coverage.xml
48
+ *.cover
49
+ *.py.cover
50
+ .hypothesis/
51
+ .pytest_cache/
52
+ cover/
53
+
54
+ # Translations
55
+ *.mo
56
+ *.pot
57
+
58
+ # Django stuff:
59
+ *.log
60
+ local_settings.py
61
+ db.sqlite3
62
+ db.sqlite3-journal
63
+
64
+ # Flask stuff:
65
+ instance/
66
+ .webassets-cache
67
+
68
+ # Scrapy stuff:
69
+ .scrapy
70
+
71
+ # Sphinx documentation
72
+ docs/_build/
73
+
74
+ # PyBuilder
75
+ .pybuilder/
76
+ target/
77
+
78
+ # Jupyter Notebook
79
+ .ipynb_checkpoints
80
+
81
+ # IPython
82
+ profile_default/
83
+ ipython_config.py
84
+
85
+ # pyenv
86
+ # For a library or package, you might want to ignore these files since the code is
87
+ # intended to run in multiple environments; otherwise, check them in:
88
+ # .python-version
89
+
90
+ # pipenv
91
+ # According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control.
92
+ # However, in case of collaboration, if having platform-specific dependencies or dependencies
93
+ # having no cross-platform support, pipenv may install dependencies that don't work, or not
94
+ # install all needed dependencies.
95
+ # Pipfile.lock
96
+
97
+ # UV
98
+ # Similar to Pipfile.lock, it is generally recommended to include uv.lock in version control.
99
+ # This is especially recommended for binary packages to ensure reproducibility, and is more
100
+ # commonly ignored for libraries.
101
+ # uv.lock
102
+
103
+ # poetry
104
+ # Similar to Pipfile.lock, it is generally recommended to include poetry.lock in version control.
105
+ # This is especially recommended for binary packages to ensure reproducibility, and is more
106
+ # commonly ignored for libraries.
107
+ # https://python-poetry.org/docs/basic-usage/#commit-your-poetrylock-file-to-version-control
108
+ # poetry.lock
109
+ # poetry.toml
110
+
111
+ # pdm
112
+ # Similar to Pipfile.lock, it is generally recommended to include pdm.lock in version control.
113
+ # pdm recommends including project-wide configuration in pdm.toml, but excluding .pdm-python.
114
+ # https://pdm-project.org/en/latest/usage/project/#working-with-version-control
115
+ # pdm.lock
116
+ # pdm.toml
117
+ .pdm-python
118
+ .pdm-build/
119
+
120
+ # pixi
121
+ # Similar to Pipfile.lock, it is generally recommended to include pixi.lock in version control.
122
+ # pixi.lock
123
+ # Pixi creates a virtual environment in the .pixi directory, just like venv module creates one
124
+ # in the .venv directory. It is recommended not to include this directory in version control.
125
+ .pixi
126
+
127
+ # PEP 582; used by e.g. github.com/David-OConnor/pyflow and github.com/pdm-project/pdm
128
+ __pypackages__/
129
+
130
+ # Celery stuff
131
+ celerybeat-schedule
132
+ celerybeat.pid
133
+
134
+ # Redis
135
+ *.rdb
136
+ *.aof
137
+ *.pid
138
+
139
+ # RabbitMQ
140
+ mnesia/
141
+ rabbitmq/
142
+ rabbitmq-data/
143
+
144
+ # ActiveMQ
145
+ activemq-data/
146
+
147
+ # SageMath parsed files
148
+ *.sage.py
149
+
150
+ # Environments
151
+ .env
152
+ .envrc
153
+ .venv
154
+ env/
155
+ venv/
156
+ ENV/
157
+ env.bak/
158
+ venv.bak/
159
+
160
+ # Spyder project settings
161
+ .spyderproject
162
+ .spyproject
163
+
164
+ # Rope project settings
165
+ .ropeproject
166
+
167
+ # mkdocs documentation
168
+ /site
169
+
170
+ # mypy
171
+ .mypy_cache/
172
+ .dmypy.json
173
+ dmypy.json
174
+
175
+ # Pyre type checker
176
+ .pyre/
177
+
178
+ # pytype static type analyzer
179
+ .pytype/
180
+
181
+ # Cython debug symbols
182
+ cython_debug/
183
+
184
+ # PyCharm
185
+ # JetBrains specific template is maintained in a separate JetBrains.gitignore that can
186
+ # be found at https://github.com/github/gitignore/blob/main/Global/JetBrains.gitignore
187
+ # and can be added to the global gitignore or merged into this file. For a more nuclear
188
+ # option (not recommended) you can uncomment the following to ignore the entire idea folder.
189
+ # .idea/
190
+
191
+ # Abstra
192
+ # Abstra is an AI-powered process automation framework.
193
+ # Ignore directories containing user credentials, local state, and settings.
194
+ # Learn more at https://abstra.io/docs
195
+ .abstra/
196
+
197
+ # Visual Studio Code
198
+ # Visual Studio Code specific template is maintained in a separate VisualStudioCode.gitignore
199
+ # that can be found at https://github.com/github/gitignore/blob/main/Global/VisualStudioCode.gitignore
200
+ # and can be added to the global gitignore or merged into this file. However, if you prefer,
201
+ # you could uncomment the following to ignore the entire vscode folder
202
+ # .vscode/
203
+
204
+ # Ruff stuff:
205
+ .ruff_cache/
206
+
207
+ # PyPI configuration file
208
+ .pypirc
209
+
210
+ # Marimo
211
+ marimo/_static/
212
+ marimo/_lsp/
213
+ __marimo__/
214
+
215
+ # Streamlit
216
+ .streamlit/secrets.toml
.python-version ADDED
@@ -0,0 +1 @@
 
 
1
+ 3.12
README.md CHANGED
@@ -1,14 +0,0 @@
1
- ---
2
- title: Nlp Assignment
3
- emoji: 👁
4
- colorFrom: red
5
- colorTo: pink
6
- sdk: gradio
7
- sdk_version: 5.49.1
8
- app_file: app.py
9
- pinned: false
10
- license: mit
11
- short_description: A RAG based academic paper recommendation system
12
- ---
13
-
14
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
__pycache__/config.cpython-312.pyc ADDED
Binary file (652 Bytes). View file
 
__pycache__/rag_core.cpython-312.pyc ADDED
Binary file (6.5 kB). View file
 
__pycache__/utils.cpython-312.pyc ADDED
Binary file (6.24 kB). View file
 
faiss_index/index.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6272ece86301fc582e0a66a6d4e035e90b9635656fb8b72a821da8541c244ef7
3
+ size 31873745
main.py ADDED
@@ -0,0 +1,135 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ import getpass
3
+ import torch
4
+ from dotenv import load_dotenv
5
+ import gradio as gr
6
+ import faiss
7
+ from typing import List, TypedDict
8
+ from sentence_transformers import CrossEncoder
9
+ from langchain_community.vectorstores import FAISS
10
+ from langchain_community.docstore.in_memory import InMemoryDocstore
11
+ from langchain_core.prompts import ChatPromptTemplate
12
+ from tabulate import tabulate
13
+ from langchain.chat_models import init_chat_model
14
+ from langchain_huggingface import HuggingFaceEmbeddings
15
+
16
+ load_dotenv()
17
+
18
+ PROJECT_PATH = os.path.dirname(os.path.abspath(__file__))
19
+ FAISS_INDEX_DIR = os.path.join(PROJECT_PATH, "faiss_index")
20
+ os.makedirs(FAISS_INDEX_DIR, exist_ok=True)
21
+ EMBEDDING_MODEL = 'sentence-transformers/all-MiniLM-L6-v2'
22
+ LLM_MODEL = 'gemini-2.5-flash'
23
+ RERANKER_MODEL = "BAAI/bge-reranker-base"
24
+ device = "cuda" if torch.cuda.is_available() else "cpu"
25
+ reranker_model = CrossEncoder(RERANKER_MODEL, device=device)
26
+
27
+ if not os.environ.get("GOOGLE_API_KEY"):
28
+ os.environ["GOOGLE_API_KEY"] = getpass.getpass("Enter API key for Google Gemini: ")
29
+
30
+ embeddings = HuggingFaceEmbeddings(model_name=EMBEDDING_MODEL)
31
+ llm = init_chat_model(LLM_MODEL, model_provider="google_genai")
32
+
33
+ print("Loading FAISS index...")
34
+ vector_store = FAISS.load_local(
35
+ FAISS_INDEX_DIR,
36
+ embeddings=embeddings,
37
+ allow_dangerous_deserialization=True,
38
+ )
39
+
40
+ print("FAISS index loaded successfully")
41
+
42
+ class State(TypedDict):
43
+ question: str
44
+ context: List
45
+ answer: str
46
+
47
+ def retrieve(question, faiss_broad_k=30, metadata_min_keep=8, top_k=5, semantic_threshold=0.35):
48
+ docs_with_scores = vector_store.similarity_search_with_score(question, k=faiss_broad_k)
49
+ print(f"[retrieve] FAISS returned {len(docs_with_scores)} candidates")
50
+
51
+ SCOPE_B_KEYWORDS = [
52
+ "machine learning", "artificial intelligence", "deep learning", "robotics",
53
+ "data science", "neural network", "quantum computing", "automation",
54
+ "computer vision", "nlp", "natural language", "algorithm", "software",
55
+ "engineering", "big data", "reinforcement learning"
56
+ ]
57
+
58
+ meta_filtered = []
59
+ for doc, _ in docs_with_scores:
60
+ text = f"{doc.metadata.get('title','')} {doc.metadata.get('concepts','')} {doc.page_content}".lower()
61
+ if any(k in text for k in SCOPE_B_KEYWORDS):
62
+ meta_filtered.append(doc)
63
+
64
+ if len(meta_filtered) == 0:
65
+ qlow = question.lower()
66
+ if not any(k in qlow for k in SCOPE_B_KEYWORDS):
67
+ return []
68
+ meta_filtered = [doc for doc, _ in docs_with_scores[:metadata_min_keep]]
69
+
70
+ rerank_inputs = [[question, doc.page_content] for doc in meta_filtered]
71
+ scores = reranker_model.predict(rerank_inputs)
72
+ reranked = sorted(zip(meta_filtered, scores), key=lambda x: x[1], reverse=True)
73
+
74
+ filtered = [(doc, s) for doc, s in reranked if s >= semantic_threshold]
75
+ final_docs = [doc for doc, _ in filtered[:top_k]]
76
+
77
+ return final_docs
78
+
79
+ def generate_with_table(question: str, docs: List):
80
+ if not docs:
81
+ return f"I cannot find anything based on the search term **{question}**."
82
+
83
+ rows = []
84
+ for d in docs:
85
+ m = d.metadata
86
+ rows.append([
87
+ m.get("title", ""),
88
+ m.get("pub_year", ""),
89
+ m.get("authors", ""),
90
+ m.get("concepts", "")
91
+ ])
92
+ headers = ["Title", "Year", "Authors", "Concepts"]
93
+ papers_table = tabulate(rows, headers=headers, tablefmt="pipe")
94
+
95
+ docs_content = "\n\n".join(doc.page_content for doc in docs)
96
+
97
+ prompt_template = ChatPromptTemplate.from_messages([
98
+ (
99
+ "system",
100
+ "You are an expert RAG system. Answer the user's question based ONLY on the provided context. "
101
+ "After your answer, append a Markdown table of the retrieved papers. "
102
+ "If there are results: say how many you found for [Search term]. "
103
+ "If none: say 'I cannot find anything based on the search term [Search term]'. "
104
+ "Output format: [Summary] \n\n [Markdown Table]. Context: {context}"
105
+ ),
106
+ ("human", "Question: {question}. Table of papers: \n\n{papers_table}"),
107
+ ])
108
+
109
+ messages = prompt_template.invoke({
110
+ "question": question,
111
+ "context": docs_content,
112
+ "papers_table": papers_table
113
+ })
114
+
115
+ response = llm.invoke(messages)
116
+ return response.content
117
+
118
+ def rag_pipeline(question: str):
119
+ if not question.strip():
120
+ return "Please enter a question."
121
+
122
+ docs = retrieve(question)
123
+ response = generate_with_table(question, docs)
124
+ return response
125
+
126
+ demo = gr.Interface(
127
+ fn=rag_pipeline,
128
+ inputs=gr.Textbox(label="Enter your research query:", placeholder="e.g., deep learning in robotics"),
129
+ outputs=gr.Markdown(label="Response"),
130
+ title="📚 Research Paper RAG Assistant",
131
+ description="Retrieves and summarizes papers related to your query using FAISS, CrossEncoder, and Gemini."
132
+ )
133
+
134
+ if __name__ == "__main__":
135
+ demo.launch()
pyproject.toml ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [project]
2
+ name = "nlp-assesment"
3
+ version = "0.1.0"
4
+ description = "Add your description here"
5
+ readme = "README.md"
6
+ requires-python = ">=3.12"
7
+ dependencies = [
8
+ "accelerate>=1.11.0",
9
+ "bitsandbytes>=0.48.1",
10
+ "dotenv>=0.9.9",
11
+ "faiss-cpu>=1.12.0",
12
+ "gradio>=5.23.1",
13
+ "hf-xet>=1.1.10",
14
+ "langchain>=1.0.0a14",
15
+ "langchain-chroma>=0.2.6",
16
+ "langchain-community>=0.3.31",
17
+ "langchain-experimental>=0.0.42",
18
+ "langchain-google-genai>=3.0.0",
19
+ "langchain-huggingface>=1.0.0",
20
+ "libsql>=0.1.11",
21
+ "neo4j>=6.0.2",
22
+ "pandas>=2.3.3",
23
+ "pydantic>=2.12.2",
24
+ "sentence-transformers>=5.1.1",
25
+ "tabulate>=0.9.0",
26
+ "torch>=2.8.0",
27
+ "transformers>=4.57.1",
28
+ ]
uv.lock ADDED
The diff for this file is too large to render. See raw diff