Spaces:
Runtime error
Runtime error
Upload folder using huggingface_hub
Browse files- grad-dashboard.py +4 -11
- requirements.txt +8 -0
grad-dashboard.py
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
import pandas as pd
|
| 2 |
import numpy as np
|
| 3 |
from dotenv import load_dotenv
|
| 4 |
-
|
| 5 |
from langchain_community.vectorstores import FAISS
|
| 6 |
from langchain_community.document_loaders import TextLoader
|
| 7 |
# from langchain_openai import OpenAIEmbeddings
|
|
@@ -24,14 +24,7 @@ text_splitter=CharacterTextSplitter(separator="\n",chunk_size=0,chunk_overlap=0)
|
|
| 24 |
documents=text_splitter.split_documents(raw_documents)
|
| 25 |
print(f"Number of documents loaded: {len(documents)}")
|
| 26 |
|
| 27 |
-
|
| 28 |
-
# db_books = Chroma.from_documents(
|
| 29 |
-
# documents[:5],
|
| 30 |
-
# embedding=GoogleGenerativeAIEmbeddings(model="models/embedding-001",google_api_key=os.getenv("GOOGLE_API_KEY"))
|
| 31 |
-
# )
|
| 32 |
-
# print("Chroma DB created with sample documents")
|
| 33 |
-
# except Exception as e:
|
| 34 |
-
# print(f"An error occurred with sample documents: {e}")
|
| 35 |
db_books = FAISS.from_documents(
|
| 36 |
documents,
|
| 37 |
embedding=GoogleGenerativeAIEmbeddings(model="models/embedding-001")
|
|
@@ -44,7 +37,7 @@ def retrieve_semantic_recommendation(
|
|
| 44 |
initial_top_k:int =50,
|
| 45 |
final_top_k:int =16,
|
| 46 |
)-> pd.DataFrame:
|
| 47 |
-
print("rsr")
|
| 48 |
recs=db_books.similarity_search(query,k=initial_top_k)
|
| 49 |
books_list=[int(rec.page_content.strip('"').split()[0]) for rec in recs]
|
| 50 |
book_recs=books[books["isbn13"].isin(books_list)].head(final_top_k)
|
|
@@ -73,7 +66,7 @@ def recommend_books(
|
|
| 73 |
category:str,
|
| 74 |
tone:str
|
| 75 |
):
|
| 76 |
-
print("Inside recommend_books function")
|
| 77 |
recommendations= retrieve_semantic_recommendation(query,category,tone)
|
| 78 |
results=[]
|
| 79 |
|
|
|
|
| 1 |
import pandas as pd
|
| 2 |
import numpy as np
|
| 3 |
from dotenv import load_dotenv
|
| 4 |
+
|
| 5 |
from langchain_community.vectorstores import FAISS
|
| 6 |
from langchain_community.document_loaders import TextLoader
|
| 7 |
# from langchain_openai import OpenAIEmbeddings
|
|
|
|
| 24 |
documents=text_splitter.split_documents(raw_documents)
|
| 25 |
print(f"Number of documents loaded: {len(documents)}")
|
| 26 |
|
| 27 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
db_books = FAISS.from_documents(
|
| 29 |
documents,
|
| 30 |
embedding=GoogleGenerativeAIEmbeddings(model="models/embedding-001")
|
|
|
|
| 37 |
initial_top_k:int =50,
|
| 38 |
final_top_k:int =16,
|
| 39 |
)-> pd.DataFrame:
|
| 40 |
+
# print("rsr")
|
| 41 |
recs=db_books.similarity_search(query,k=initial_top_k)
|
| 42 |
books_list=[int(rec.page_content.strip('"').split()[0]) for rec in recs]
|
| 43 |
book_recs=books[books["isbn13"].isin(books_list)].head(final_top_k)
|
|
|
|
| 66 |
category:str,
|
| 67 |
tone:str
|
| 68 |
):
|
| 69 |
+
# print("Inside recommend_books function")
|
| 70 |
recommendations= retrieve_semantic_recommendation(query,category,tone)
|
| 71 |
results=[]
|
| 72 |
|
requirements.txt
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
pandas
|
| 2 |
+
numpy
|
| 3 |
+
dotenv
|
| 4 |
+
langchain_community.vectorstores
|
| 5 |
+
langchain_community.document_loaders
|
| 6 |
+
langchain_text_splitters
|
| 7 |
+
langchain_google_genai
|
| 8 |
+
gradio
|