mohAhmad commited on
Commit
2fa2df2
·
verified ·
1 Parent(s): 8701f1d

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -4
app.py CHANGED
@@ -1,10 +1,10 @@
1
- import numpy as np
2
  import streamlit as st
3
  from transformers import DPRQuestionEncoder, DPRQuestionEncoderTokenizer
4
  from transformers import BartForConditionalGeneration, BartTokenizer
5
  from sentence_transformers import SentenceTransformer
6
  import pdfplumber
7
  from sklearn.metrics.pairwise import cosine_similarity
 
8
  import torch
9
 
10
  # Load the Question Encoder, Context Encoder, and Tokenizers
@@ -60,9 +60,8 @@ query = st.text_input("🔍 Enter your query")
60
 
61
  if st.button("💬 Get Answer"):
62
  if query:
63
- # Step 1: Encode the query
64
- question_inputs = question_tokenizer(query, return_tensors="pt")
65
- question_embedding = question_encoder(**question_inputs).pooler_output.detach().cpu().numpy()
66
 
67
  # Step 2: Calculate Cosine Similarity
68
  similarity_scores = cosine_similarity(question_embedding, doc_embeddings)
 
 
1
  import streamlit as st
2
  from transformers import DPRQuestionEncoder, DPRQuestionEncoderTokenizer
3
  from transformers import BartForConditionalGeneration, BartTokenizer
4
  from sentence_transformers import SentenceTransformer
5
  import pdfplumber
6
  from sklearn.metrics.pairwise import cosine_similarity
7
+ import numpy as np # Import NumPy
8
  import torch
9
 
10
  # Load the Question Encoder, Context Encoder, and Tokenizers
 
60
 
61
  if st.button("💬 Get Answer"):
62
  if query:
63
+ # Step 1: Encode the query with the same SentenceTransformer model
64
+ question_embedding = sentence_model.encode([query]) # Use the same model
 
65
 
66
  # Step 2: Calculate Cosine Similarity
67
  similarity_scores = cosine_similarity(question_embedding, doc_embeddings)