arya123321 commited on
Commit
bc9198b
·
1 Parent(s): 493c63b
Files changed (1) hide show
  1. chatbot.py +8 -15
chatbot.py CHANGED
@@ -5,31 +5,29 @@ from llama_index.embeddings.huggingface import HuggingFaceEmbedding
5
  from llama_index.llms.groq import Groq
6
  from llama_index.vector_stores.pinecone import PineconeVectorStore
7
  from pinecone import Pinecone
8
- from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
9
  from PyPDF2 import PdfReader
10
  from flask_cors import CORS
11
  from functools import wraps
12
  from dotenv import load_dotenv
13
  from huggingface_hub import InferenceClient
14
- import re, torch, jwt, os, json, gc
15
 
16
  load_dotenv()
17
 
18
  SECRET_KEY = os.getenv("SECRET_KEY")
19
 
20
  # Initialize Hugging Face Inference Client for embeddings
21
- client = InferenceClient(
22
- provider="hf-inference",
23
- api_key=os.getenv("HF_API_KEY") # Add your Hugging Face API key to .env
24
- )
25
 
26
  # Load summarization model and tokenizer
27
  model_path = "Jurisight/legal_led"
28
- model = AutoModelForSeq2SeqLM.from_pretrained(model_path,use_auth_token=os.getenv("HF_API_KEY"))
29
- tokenizer = AutoTokenizer.from_pretrained(model_path,use_auth_token=os.getenv("HF_API_KEY"))
30
 
 
31
  # Configure LlamaIndex settings
32
- Settings.embed_model = HuggingFaceEmbedding(model_name="BAAI/bge-base-en-v1.5")
33
  Settings.llm = Groq(model="llama3-8b-8192", api_key=os.getenv("GROQ_API_KEY"))
34
 
35
  # Initialize Pinecone
@@ -208,12 +206,7 @@ def summarize(user_id):
208
  @authenticate_user
209
  def retrieve_cases(user_id):
210
  def generate_embedding(text):
211
- # Use Hugging Face Inference API for embeddings
212
- result = client.feature_extraction(
213
- model="BAAI/bge-base-en-v1.5",
214
- inputs=text,
215
- provider="hf-inference",
216
- )
217
  return result
218
 
219
  def query_pinecone(query_text, top_k=10):
 
5
  from llama_index.llms.groq import Groq
6
  from llama_index.vector_stores.pinecone import PineconeVectorStore
7
  from pinecone import Pinecone
8
+ from transformers import AutoModelForSeq2SeqLM, AutoTokenizer, AutoModel
9
  from PyPDF2 import PdfReader
10
  from flask_cors import CORS
11
  from functools import wraps
12
  from dotenv import load_dotenv
13
  from huggingface_hub import InferenceClient
14
+ import re, jwt, os, json
15
 
16
  load_dotenv()
17
 
18
  SECRET_KEY = os.getenv("SECRET_KEY")
19
 
20
  # Initialize Hugging Face Inference Client for embeddings
21
+ client = InferenceClient(token=os.getenv("HF_API_KEY"))
 
 
 
22
 
23
  # Load summarization model and tokenizer
24
  model_path = "Jurisight/legal_led"
25
+ model = AutoModelForSeq2SeqLM.from_pretrained(model_path,token=os.getenv("HF_API_KEY"))
26
+ tokenizer = AutoTokenizer.from_pretrained(model_path,token=os.getenv("HF_API_KEY"))
27
 
28
+ embed_model = "BAAI/bge-base-en-v1.5"
29
  # Configure LlamaIndex settings
30
+ Settings.embed_model = HuggingFaceEmbedding(model_name=embed_model)
31
  Settings.llm = Groq(model="llama3-8b-8192", api_key=os.getenv("GROQ_API_KEY"))
32
 
33
  # Initialize Pinecone
 
206
  @authenticate_user
207
  def retrieve_cases(user_id):
208
  def generate_embedding(text):
209
+ result = client.feature_extraction(model=embed_model, text=text)
 
 
 
 
 
210
  return result
211
 
212
  def query_pinecone(query_text, top_k=10):