Spaces:
Sleeping
Sleeping
Commit ·
bc9198b
1
Parent(s): 493c63b
update
Browse files- chatbot.py +8 -15
chatbot.py
CHANGED
|
@@ -5,31 +5,29 @@ from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
|
| 5 |
from llama_index.llms.groq import Groq
|
| 6 |
from llama_index.vector_stores.pinecone import PineconeVectorStore
|
| 7 |
from pinecone import Pinecone
|
| 8 |
-
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
|
| 9 |
from PyPDF2 import PdfReader
|
| 10 |
from flask_cors import CORS
|
| 11 |
from functools import wraps
|
| 12 |
from dotenv import load_dotenv
|
| 13 |
from huggingface_hub import InferenceClient
|
| 14 |
-
import re,
|
| 15 |
|
| 16 |
load_dotenv()
|
| 17 |
|
| 18 |
SECRET_KEY = os.getenv("SECRET_KEY")
|
| 19 |
|
| 20 |
# Initialize Hugging Face Inference Client for embeddings
|
| 21 |
-
client = InferenceClient(
|
| 22 |
-
provider="hf-inference",
|
| 23 |
-
api_key=os.getenv("HF_API_KEY") # Add your Hugging Face API key to .env
|
| 24 |
-
)
|
| 25 |
|
| 26 |
# Load summarization model and tokenizer
|
| 27 |
model_path = "Jurisight/legal_led"
|
| 28 |
-
model = AutoModelForSeq2SeqLM.from_pretrained(model_path,
|
| 29 |
-
tokenizer = AutoTokenizer.from_pretrained(model_path,
|
| 30 |
|
|
|
|
| 31 |
# Configure LlamaIndex settings
|
| 32 |
-
Settings.embed_model = HuggingFaceEmbedding(model_name=
|
| 33 |
Settings.llm = Groq(model="llama3-8b-8192", api_key=os.getenv("GROQ_API_KEY"))
|
| 34 |
|
| 35 |
# Initialize Pinecone
|
|
@@ -208,12 +206,7 @@ def summarize(user_id):
|
|
| 208 |
@authenticate_user
|
| 209 |
def retrieve_cases(user_id):
|
| 210 |
def generate_embedding(text):
|
| 211 |
-
|
| 212 |
-
result = client.feature_extraction(
|
| 213 |
-
model="BAAI/bge-base-en-v1.5",
|
| 214 |
-
inputs=text,
|
| 215 |
-
provider="hf-inference",
|
| 216 |
-
)
|
| 217 |
return result
|
| 218 |
|
| 219 |
def query_pinecone(query_text, top_k=10):
|
|
|
|
| 5 |
from llama_index.llms.groq import Groq
|
| 6 |
from llama_index.vector_stores.pinecone import PineconeVectorStore
|
| 7 |
from pinecone import Pinecone
|
| 8 |
+
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer, AutoModel
|
| 9 |
from PyPDF2 import PdfReader
|
| 10 |
from flask_cors import CORS
|
| 11 |
from functools import wraps
|
| 12 |
from dotenv import load_dotenv
|
| 13 |
from huggingface_hub import InferenceClient
|
| 14 |
+
import re, jwt, os, json
|
| 15 |
|
| 16 |
load_dotenv()
|
| 17 |
|
| 18 |
SECRET_KEY = os.getenv("SECRET_KEY")
|
| 19 |
|
| 20 |
# Initialize Hugging Face Inference Client for embeddings
|
| 21 |
+
client = InferenceClient(token=os.getenv("HF_API_KEY"))
|
|
|
|
|
|
|
|
|
|
| 22 |
|
| 23 |
# Load summarization model and tokenizer
|
| 24 |
model_path = "Jurisight/legal_led"
|
| 25 |
+
model = AutoModelForSeq2SeqLM.from_pretrained(model_path,token=os.getenv("HF_API_KEY"))
|
| 26 |
+
tokenizer = AutoTokenizer.from_pretrained(model_path,token=os.getenv("HF_API_KEY"))
|
| 27 |
|
| 28 |
+
embed_model = "BAAI/bge-base-en-v1.5"
|
| 29 |
# Configure LlamaIndex settings
|
| 30 |
+
Settings.embed_model = HuggingFaceEmbedding(model_name=embed_model)
|
| 31 |
Settings.llm = Groq(model="llama3-8b-8192", api_key=os.getenv("GROQ_API_KEY"))
|
| 32 |
|
| 33 |
# Initialize Pinecone
|
|
|
|
| 206 |
@authenticate_user
|
| 207 |
def retrieve_cases(user_id):
|
| 208 |
def generate_embedding(text):
|
| 209 |
+
result = client.feature_extraction(model=embed_model, text=text)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 210 |
return result
|
| 211 |
|
| 212 |
def query_pinecone(query_text, top_k=10):
|