File size: 1,356 Bytes
161cecf
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
{
  "dataset_revision": "e8379541af4e31359cca9fbcf4b00f2671dba205",
  "mteb_dataset_name": "AmazonCounterfactualClassification",
  "mteb_version": "1.1.1",
  "test": {
    "de": {
      "accuracy": 0.6229122055674517,
      "accuracy_stderr": 0.02689391560397646,
      "ap": 0.7654273024410109,
      "ap_stderr": 0.014834546530994171,
      "f1": 0.6037703210343267,
      "f1_stderr": 0.023236308207408476,
      "main_score": 0.6229122055674517
    },
    "en": {
      "accuracy": 0.7073134328358208,
      "accuracy_stderr": 0.0383582089552239,
      "ap": 0.3235996836729783,
      "ap_stderr": 0.026484778108220294,
      "f1": 0.642137087561157,
      "f1_stderr": 0.03200155942638674,
      "main_score": 0.7073134328358208
    },
    "en-ext": {
      "accuracy": 0.6757871064467766,
      "accuracy_stderr": 0.04988162814232977,
      "ap": 0.1703033311712744,
      "ap_stderr": 0.013926829446773738,
      "f1": 0.5482175063189498,
      "f1_stderr": 0.03163298075732844,
      "main_score": 0.6757871064467766
    },
    "evaluation_time": 10.55,
    "ja": {
      "accuracy": 0.6251605995717344,
      "accuracy_stderr": 0.053002616964599826,
      "ap": 0.14367489440317666,
      "ap_stderr": 0.012527680056177723,
      "f1": 0.5048473578289779,
      "f1_stderr": 0.030994925213475993,
      "main_score": 0.6251605995717344
    }
  }
}