Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -6,9 +6,12 @@ import re
|
|
| 6 |
from rank_bm25 import BM25Okapi
|
| 7 |
import numpy as np
|
| 8 |
# Load models
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
|
|
|
|
|
|
|
|
|
|
| 12 |
|
| 13 |
# Load data
|
| 14 |
df = pd.read_csv("cleaned1.csv")
|
|
@@ -16,19 +19,19 @@ df2 = pd.read_csv("cleaned2.csv")
|
|
| 16 |
df3 = pd.read_csv("cleaned3.csv")
|
| 17 |
|
| 18 |
# Load pre-computed embeddings
|
| 19 |
-
embeddings = torch.load("embeddings1_1.pt")
|
| 20 |
-
embeddings2 = torch.load("embeddings2_1.pt")
|
| 21 |
-
embeddings3 = torch.load("embeddings3_1.pt")
|
| 22 |
|
| 23 |
|
| 24 |
-
embeddingsa = torch.load("embeddings1.pt")
|
| 25 |
-
embeddingsa2 = torch.load("embeddings2.pt")
|
| 26 |
-
embeddingsa3 = torch.load("embeddings3.pt")
|
| 27 |
|
| 28 |
|
| 29 |
-
embeddingsb = torch.load("embeddingso1_3.pt")
|
| 30 |
-
embeddingsb2 = torch.load("embeddingso2_3.pt")
|
| 31 |
-
embeddingsb3 = torch.load("embeddingso3_3.pt")
|
| 32 |
|
| 33 |
# Extract questions and links
|
| 34 |
df_questions = df["question"].values
|
|
|
|
| 6 |
from rank_bm25 import BM25Okapi
|
| 7 |
import numpy as np
|
| 8 |
# Load models
|
| 9 |
+
|
| 10 |
+
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
|
| 11 |
+
|
| 12 |
+
model = SentenceTransformer("distilbert-base-multilingual-cased", device=device)
|
| 13 |
+
modela = SentenceTransformer("paraphrase-multilingual-MiniLM-L12-v2", device=device)
|
| 14 |
+
modelb = SentenceTransformer("Omartificial-Intelligence-Space/Arabert-all-nli-triplet-Matryoshka", device=device)
|
| 15 |
|
| 16 |
# Load data
|
| 17 |
df = pd.read_csv("cleaned1.csv")
|
|
|
|
| 19 |
df3 = pd.read_csv("cleaned3.csv")
|
| 20 |
|
| 21 |
# Load pre-computed embeddings
|
| 22 |
+
embeddings = torch.load("embeddings1_1.pt", map_location=device)
|
| 23 |
+
embeddings2 = torch.load("embeddings2_1.pt", map_location=device)
|
| 24 |
+
embeddings3 = torch.load("embeddings3_1.pt", map_location=device)
|
| 25 |
|
| 26 |
|
| 27 |
+
embeddingsa = torch.load("embeddings1.pt", map_location=device)
|
| 28 |
+
embeddingsa2 = torch.load("embeddings2.pt", map_location=device)
|
| 29 |
+
embeddingsa3 = torch.load("embeddings3.pt", map_location=device)
|
| 30 |
|
| 31 |
|
| 32 |
+
embeddingsb = torch.load("embeddingso1_3.pt", map_location=device)
|
| 33 |
+
embeddingsb2 = torch.load("embeddingso2_3.pt", map_location=device)
|
| 34 |
+
embeddingsb3 = torch.load("embeddingso3_3.pt", map_location=device)
|
| 35 |
|
| 36 |
# Extract questions and links
|
| 37 |
df_questions = df["question"].values
|