FEA-Bench / testbed /embeddings-benchmark__mteb /tests /test_PairClassificationEvaluator.py
hc99's picture
Add files using upload-large-folder tool
73cc8d2 verified
from __future__ import annotations
import pytest
from mteb.evaluation.evaluators import PairClassificationEvaluator
TOL = 0.0001
class TestPairClassificationEvaluator:
def test_accuracy(self):
scores = [6.12, 5.39, 5.28, 5.94, 6.34, 6.47, 7.88, 6.62, 8.04, 5.9]
labels = [0, 0, 0, 0, 1, 0, 0, 0, 1, 0]
high_score_more_similar = True
acc, acc_threshold = PairClassificationEvaluator.find_best_acc_and_threshold(
scores, labels, high_score_more_similar
)
assert acc == pytest.approx(0.9, TOL)
assert acc_threshold == pytest.approx(7.95999, TOL)
def test_f1(self):
scores = [6.12, 5.39, 5.28, 5.94, 6.34, 6.47, 7.88, 6.62, 8.04, 5.9]
labels = [0, 0, 0, 0, 1, 0, 0, 0, 1, 0]
high_score_more_similar = True
f1, precision, recall, f1_threshold = (
PairClassificationEvaluator.find_best_f1_and_threshold(
scores, labels, high_score_more_similar
)
)
assert f1 == pytest.approx(0.66666, TOL)
assert precision == pytest.approx(1.0, TOL)
assert recall == pytest.approx(0.5, TOL)
assert f1_threshold == pytest.approx(7.95999, TOL)
def test_ap(self):
scores = [6.12, 5.39, 5.28, 5.94, 6.34, 6.47, 7.88, 6.62, 8.04, 5.9]
labels = [0, 0, 0, 0, 1, 0, 0, 0, 1, 0]
high_score_more_similar = True
ap = PairClassificationEvaluator.ap_score(
scores, labels, high_score_more_similar
)
assert ap == pytest.approx(0.7, TOL)