Update app.py
Browse files
app.py
CHANGED
|
@@ -162,14 +162,17 @@ def add_new_eval(
|
|
| 162 |
if id_to_eval == intervention_id:
|
| 163 |
references = gold_dataset['cqs']
|
| 164 |
reference_set = [row['cq'] for row in references[indx]]
|
|
|
|
| 165 |
for cq in line['cqs']:
|
| 166 |
# TODO: compare to each reference and get a value
|
| 167 |
cq_text = cq['cq']
|
|
|
|
| 168 |
|
| 169 |
#if args.metric == 'similarity':
|
| 170 |
sentence_embedding = similarity_model.encode(cq_text)
|
| 171 |
reference_embedding = similarity_model.encode(reference_set)
|
| 172 |
sims = similarity_model.similarity(sentence_embedding, reference_embedding).tolist()[0]
|
|
|
|
| 173 |
|
| 174 |
winner = np.argmax(sims)
|
| 175 |
# make sure the similarity of the winning reference sentence is at least 0.65
|
|
|
|
| 162 |
if id_to_eval == intervention_id:
|
| 163 |
references = gold_dataset['cqs']
|
| 164 |
reference_set = [row['cq'] for row in references[indx]]
|
| 165 |
+
print(reference_set, flush=True)
|
| 166 |
for cq in line['cqs']:
|
| 167 |
# TODO: compare to each reference and get a value
|
| 168 |
cq_text = cq['cq']
|
| 169 |
+
print(cq_text, flush=True)
|
| 170 |
|
| 171 |
#if args.metric == 'similarity':
|
| 172 |
sentence_embedding = similarity_model.encode(cq_text)
|
| 173 |
reference_embedding = similarity_model.encode(reference_set)
|
| 174 |
sims = similarity_model.similarity(sentence_embedding, reference_embedding).tolist()[0]
|
| 175 |
+
print(sims, flush=True)
|
| 176 |
|
| 177 |
winner = np.argmax(sims)
|
| 178 |
# make sure the similarity of the winning reference sentence is at least 0.65
|