eduagarcia commited on
Commit
5bd407f
·
1 Parent(s): 7bde1d7

update results

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. results/Alibaba-NLP/gte-base-en-v1.5/PTT_LegalBenchCorporateLobbying.json +43 -0
  2. results/Alibaba-NLP/gte-large-en-v1.5/PTT_LegalBenchCorporateLobbying.json +43 -0
  3. results/Jechto/e5-dansk-test-0.1/PTT_LegalBenchCorporateLobbying.json +43 -0
  4. results/alfaneo/bert-base-multilingual-sts/FloresBitextMining.json +193 -0
  5. results/alfaneo/bert-base-multilingual-sts/MassiveIntentClassification.json +25 -0
  6. results/alfaneo/bert-base-multilingual-sts/MassiveScenarioClassification.json +25 -0
  7. results/alfaneo/bert-base-multilingual-sts/MintakaRetrieval.json +45 -0
  8. results/alfaneo/bert-base-multilingual-sts/MultiHateClassification.json +17 -0
  9. results/alfaneo/bert-base-multilingual-sts/MultiLongDocRetrieval.json +85 -0
  10. results/alfaneo/bert-base-multilingual-sts/NTREXBitextMining.json +92 -0
  11. results/alfaneo/bert-base-multilingual-sts/PTT_AILACasedocs.json +43 -0
  12. results/alfaneo/bert-base-multilingual-sts/PTT_AILAStatutes.json +43 -0
  13. results/alfaneo/bert-base-multilingual-sts/PTT_LegalBenchConsumerContractsQA.json +43 -0
  14. results/alfaneo/bert-base-multilingual-sts/PTT_LegalBenchCorporateLobbying.json +43 -0
  15. results/alfaneo/bertimbau-base-portuguese-sts/FloresBitextMining.json +193 -0
  16. results/alfaneo/bertimbau-base-portuguese-sts/MassiveIntentClassification.json +25 -0
  17. results/alfaneo/bertimbau-base-portuguese-sts/MassiveScenarioClassification.json +25 -0
  18. results/alfaneo/bertimbau-base-portuguese-sts/MintakaRetrieval.json +45 -0
  19. results/alfaneo/bertimbau-base-portuguese-sts/MultiHateClassification.json +17 -0
  20. results/alfaneo/bertimbau-base-portuguese-sts/MultiLongDocRetrieval.json +85 -0
  21. results/alfaneo/bertimbau-base-portuguese-sts/NTREXBitextMining.json +92 -0
  22. results/alfaneo/bertimbau-base-portuguese-sts/PTT_AILACasedocs.json +43 -0
  23. results/alfaneo/bertimbau-base-portuguese-sts/PTT_AILAStatutes.json +43 -0
  24. results/alfaneo/bertimbau-base-portuguese-sts/PTT_LegalBenchConsumerContractsQA.json +43 -0
  25. results/alfaneo/bertimbau-base-portuguese-sts/PTT_LegalBenchCorporateLobbying.json +43 -0
  26. results/alfaneo/bertimbaulaw-base-portuguese-cased/FloresBitextMining.json +193 -0
  27. results/alfaneo/bertimbaulaw-base-portuguese-cased/MassiveIntentClassification.json +25 -0
  28. results/alfaneo/bertimbaulaw-base-portuguese-cased/MassiveScenarioClassification.json +25 -0
  29. results/alfaneo/bertimbaulaw-base-portuguese-cased/MintakaRetrieval.json +45 -0
  30. results/alfaneo/bertimbaulaw-base-portuguese-cased/MultiHateClassification.json +17 -0
  31. results/alfaneo/bertimbaulaw-base-portuguese-cased/MultiLongDocRetrieval.json +85 -0
  32. results/alfaneo/bertimbaulaw-base-portuguese-cased/NTREXBitextMining.json +92 -0
  33. results/alfaneo/bertimbaulaw-base-portuguese-cased/PTT_AILACasedocs.json +43 -0
  34. results/alfaneo/bertimbaulaw-base-portuguese-cased/PTT_AILAStatutes.json +43 -0
  35. results/alfaneo/bertimbaulaw-base-portuguese-cased/PTT_LegalBenchConsumerContractsQA.json +43 -0
  36. results/alfaneo/bertimbaulaw-base-portuguese-cased/PTT_LegalBenchCorporateLobbying.json +43 -0
  37. results/alfaneo/bertimbaulaw-base-portuguese-sts/FloresBitextMining.json +193 -0
  38. results/alfaneo/bertimbaulaw-base-portuguese-sts/MassiveIntentClassification.json +25 -0
  39. results/alfaneo/bertimbaulaw-base-portuguese-sts/MassiveScenarioClassification.json +25 -0
  40. results/alfaneo/bertimbaulaw-base-portuguese-sts/MintakaRetrieval.json +45 -0
  41. results/alfaneo/bertimbaulaw-base-portuguese-sts/MultiHateClassification.json +17 -0
  42. results/alfaneo/bertimbaulaw-base-portuguese-sts/MultiLongDocRetrieval.json +85 -0
  43. results/alfaneo/bertimbaulaw-base-portuguese-sts/NTREXBitextMining.json +92 -0
  44. results/alfaneo/bertimbaulaw-base-portuguese-sts/PTT_AILACasedocs.json +43 -0
  45. results/alfaneo/bertimbaulaw-base-portuguese-sts/PTT_AILAStatutes.json +43 -0
  46. results/alfaneo/bertimbaulaw-base-portuguese-sts/PTT_LegalBenchConsumerContractsQA.json +43 -0
  47. results/alfaneo/bertimbaulaw-base-portuguese-sts/PTT_LegalBenchCorporateLobbying.json +43 -0
  48. results/alfaneo/jurisbert-base-portuguese-sts/FloresBitextMining.json +193 -0
  49. results/alfaneo/jurisbert-base-portuguese-sts/MassiveIntentClassification.json +25 -0
  50. results/alfaneo/jurisbert-base-portuguese-sts/MassiveScenarioClassification.json +25 -0
results/Alibaba-NLP/gte-base-en-v1.5/PTT_LegalBenchCorporateLobbying.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchCorporateLobbying",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 5.25,
7
+ "map_at_1": 0.51765,
8
+ "map_at_10": 0.61947,
9
+ "map_at_100": 0.62622,
10
+ "map_at_1000": 0.62652,
11
+ "map_at_20": 0.62401,
12
+ "map_at_3": 0.60098,
13
+ "map_at_5": 0.6148,
14
+ "mrr_at_1": 0.51765,
15
+ "mrr_at_10": 0.61947,
16
+ "mrr_at_100": 0.62622,
17
+ "mrr_at_1000": 0.62652,
18
+ "mrr_at_20": 0.62401,
19
+ "mrr_at_3": 0.60098,
20
+ "mrr_at_5": 0.6148,
21
+ "ndcg_at_1": 0.51765,
22
+ "ndcg_at_10": 0.66247,
23
+ "ndcg_at_100": 0.69549,
24
+ "ndcg_at_1000": 0.7022,
25
+ "ndcg_at_20": 0.6784,
26
+ "ndcg_at_3": 0.62654,
27
+ "ndcg_at_5": 0.6511,
28
+ "precision_at_1": 0.51765,
29
+ "precision_at_10": 0.07941,
30
+ "precision_at_100": 0.0095,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.04279,
33
+ "precision_at_3": 0.23333,
34
+ "precision_at_5": 0.15176,
35
+ "recall_at_1": 0.51765,
36
+ "recall_at_10": 0.79412,
37
+ "recall_at_100": 0.95,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.85588,
40
+ "recall_at_3": 0.7,
41
+ "recall_at_5": 0.75882
42
+ }
43
+ }
results/Alibaba-NLP/gte-large-en-v1.5/PTT_LegalBenchCorporateLobbying.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchCorporateLobbying",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 8.53,
7
+ "map_at_1": 0.62059,
8
+ "map_at_10": 0.71786,
9
+ "map_at_100": 0.72112,
10
+ "map_at_1000": 0.72118,
11
+ "map_at_20": 0.72015,
12
+ "map_at_3": 0.69314,
13
+ "map_at_5": 0.71225,
14
+ "mrr_at_1": 0.62059,
15
+ "mrr_at_10": 0.71786,
16
+ "mrr_at_100": 0.72112,
17
+ "mrr_at_1000": 0.72118,
18
+ "mrr_at_20": 0.72015,
19
+ "mrr_at_3": 0.69314,
20
+ "mrr_at_5": 0.71225,
21
+ "ndcg_at_1": 0.62059,
22
+ "ndcg_at_10": 0.76418,
23
+ "ndcg_at_100": 0.78085,
24
+ "ndcg_at_1000": 0.78239,
25
+ "ndcg_at_20": 0.77339,
26
+ "ndcg_at_3": 0.7154,
27
+ "ndcg_at_5": 0.7502,
28
+ "precision_at_1": 0.62059,
29
+ "precision_at_10": 0.09088,
30
+ "precision_at_100": 0.00988,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.04735,
33
+ "precision_at_3": 0.2598,
34
+ "precision_at_5": 0.17294,
35
+ "recall_at_1": 0.62059,
36
+ "recall_at_10": 0.90882,
37
+ "recall_at_100": 0.98824,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.94706,
40
+ "recall_at_3": 0.77941,
41
+ "recall_at_5": 0.86471
42
+ }
43
+ }
results/Jechto/e5-dansk-test-0.1/PTT_LegalBenchCorporateLobbying.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchCorporateLobbying",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 2.89,
7
+ "map_at_1": 0.63235,
8
+ "map_at_10": 0.73169,
9
+ "map_at_100": 0.73522,
10
+ "map_at_1000": 0.73528,
11
+ "map_at_20": 0.73376,
12
+ "map_at_3": 0.7152,
13
+ "map_at_5": 0.72314,
14
+ "mrr_at_1": 0.63235,
15
+ "mrr_at_10": 0.73169,
16
+ "mrr_at_100": 0.73522,
17
+ "mrr_at_1000": 0.73528,
18
+ "mrr_at_20": 0.73376,
19
+ "mrr_at_3": 0.7152,
20
+ "mrr_at_5": 0.72314,
21
+ "ndcg_at_1": 0.63235,
22
+ "ndcg_at_10": 0.77641,
23
+ "ndcg_at_100": 0.7931,
24
+ "ndcg_at_1000": 0.79431,
25
+ "ndcg_at_20": 0.7839,
26
+ "ndcg_at_3": 0.74226,
27
+ "ndcg_at_5": 0.75669,
28
+ "precision_at_1": 0.63235,
29
+ "precision_at_10": 0.09147,
30
+ "precision_at_100": 0.00991,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.04721,
33
+ "precision_at_3": 0.27353,
34
+ "precision_at_5": 0.17118,
35
+ "recall_at_1": 0.63235,
36
+ "recall_at_10": 0.91471,
37
+ "recall_at_100": 0.99118,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.94412,
40
+ "recall_at_3": 0.82059,
41
+ "recall_at_5": 0.85588
42
+ }
43
+ }
results/alfaneo/bert-base-multilingual-sts/FloresBitextMining.json ADDED
@@ -0,0 +1,193 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "80dc3040d19756742c9a18267ab30f54fb8e226b",
3
+ "dev": {
4
+ "arb_Arab-por_Latn": {
5
+ "accuracy": 0.007021063189568706,
6
+ "f1": 0.004580407890337679,
7
+ "main_score": 0.004580407890337679,
8
+ "precision": 0.004047857859292162,
9
+ "recall": 0.007021063189568706
10
+ },
11
+ "deu_Latn-por_Latn": {
12
+ "accuracy": 0.014042126379137413,
13
+ "f1": 0.007890035112880117,
14
+ "main_score": 0.007890035112880117,
15
+ "precision": 0.007217821457886855,
16
+ "recall": 0.014042126379137413
17
+ },
18
+ "eng_Latn-por_Latn": {
19
+ "accuracy": 0.01905717151454363,
20
+ "f1": 0.009356631370950188,
21
+ "main_score": 0.009356631370950188,
22
+ "precision": 0.00781848346158925,
23
+ "recall": 0.01905717151454363
24
+ },
25
+ "evaluation_time": 9.71,
26
+ "fra_Latn-por_Latn": {
27
+ "accuracy": 0.029087261785356068,
28
+ "f1": 0.017161033695419457,
29
+ "main_score": 0.017161033695419457,
30
+ "precision": 0.014339319063491583,
31
+ "recall": 0.029087261785356068
32
+ },
33
+ "ita_Latn-por_Latn": {
34
+ "accuracy": 0.022066198595787363,
35
+ "f1": 0.013832753672272227,
36
+ "main_score": 0.013832753672272227,
37
+ "precision": 0.01228610072642169,
38
+ "recall": 0.022066198595787363
39
+ },
40
+ "jpn_Jpan-por_Latn": {
41
+ "accuracy": 0.007021063189568706,
42
+ "f1": 0.0027312095015204346,
43
+ "main_score": 0.0027312095015204346,
44
+ "precision": 0.002119149231073238,
45
+ "recall": 0.007021063189568706
46
+ },
47
+ "kor_Hang-por_Latn": {
48
+ "accuracy": 0.011033099297893681,
49
+ "f1": 0.00460070688255242,
50
+ "main_score": 0.00460070688255242,
51
+ "precision": 0.0033417101120210447,
52
+ "recall": 0.011033099297893681
53
+ },
54
+ "pol_Latn-por_Latn": {
55
+ "accuracy": 0.012036108324974924,
56
+ "f1": 0.008043428531207658,
57
+ "main_score": 0.008043428531207658,
58
+ "precision": 0.006881003370471776,
59
+ "recall": 0.012036108324974924
60
+ },
61
+ "rus_Cyrl-por_Latn": {
62
+ "accuracy": 0.009027081243731194,
63
+ "f1": 0.0037232910853773438,
64
+ "main_score": 0.0037232910853773438,
65
+ "precision": 0.0028550811306053774,
66
+ "recall": 0.009027081243731194
67
+ },
68
+ "spa_Latn-por_Latn": {
69
+ "accuracy": 0.037111334002006016,
70
+ "f1": 0.02047695882679093,
71
+ "main_score": 0.02047695882679093,
72
+ "precision": 0.016176523799392405,
73
+ "recall": 0.037111334002006016
74
+ },
75
+ "swe_Latn-por_Latn": {
76
+ "accuracy": 0.013039117352056168,
77
+ "f1": 0.007533723381266009,
78
+ "main_score": 0.007533723381266009,
79
+ "precision": 0.00663211857795609,
80
+ "recall": 0.013039117352056168
81
+ },
82
+ "zho_Hans-por_Latn": {
83
+ "accuracy": 0.007021063189568706,
84
+ "f1": 0.004508459522437626,
85
+ "main_score": 0.004508459522437626,
86
+ "precision": 0.004033840652391959,
87
+ "recall": 0.007021063189568706
88
+ },
89
+ "zho_Hant-por_Latn": {
90
+ "accuracy": 0.011033099297893681,
91
+ "f1": 0.003610449985288787,
92
+ "main_score": 0.003610449985288787,
93
+ "precision": 0.0023533262124034444,
94
+ "recall": 0.011033099297893681
95
+ }
96
+ },
97
+ "devtest": {
98
+ "arb_Arab-por_Latn": {
99
+ "accuracy": 0.00691699604743083,
100
+ "f1": 0.002788468206088343,
101
+ "main_score": 0.002788468206088343,
102
+ "precision": 0.001913545391806261,
103
+ "recall": 0.00691699604743083
104
+ },
105
+ "deu_Latn-por_Latn": {
106
+ "accuracy": 0.015810276679841896,
107
+ "f1": 0.009644582470669426,
108
+ "main_score": 0.009644582470669426,
109
+ "precision": 0.008378306476132562,
110
+ "recall": 0.015810276679841896
111
+ },
112
+ "eng_Latn-por_Latn": {
113
+ "accuracy": 0.029644268774703556,
114
+ "f1": 0.01856283812805552,
115
+ "main_score": 0.01856283812805552,
116
+ "precision": 0.01592867222807934,
117
+ "recall": 0.029644268774703556
118
+ },
119
+ "evaluation_time": 16.27,
120
+ "fra_Latn-por_Latn": {
121
+ "accuracy": 0.03458498023715415,
122
+ "f1": 0.023496394447611424,
123
+ "main_score": 0.023496394447611424,
124
+ "precision": 0.020691778028734547,
125
+ "recall": 0.03458498023715415
126
+ },
127
+ "ita_Latn-por_Latn": {
128
+ "accuracy": 0.029644268774703556,
129
+ "f1": 0.01805006587615283,
130
+ "main_score": 0.01805006587615283,
131
+ "precision": 0.014395935640994931,
132
+ "recall": 0.029644268774703556
133
+ },
134
+ "jpn_Jpan-por_Latn": {
135
+ "accuracy": 0.004940711462450593,
136
+ "f1": 0.0015765936961589136,
137
+ "main_score": 0.0015765936961589136,
138
+ "precision": 0.0010304912478825523,
139
+ "recall": 0.004940711462450593
140
+ },
141
+ "kor_Hang-por_Latn": {
142
+ "accuracy": 0.009881422924901186,
143
+ "f1": 0.00598375054896794,
144
+ "main_score": 0.00598375054896794,
145
+ "precision": 0.004767786561264822,
146
+ "recall": 0.009881422924901186
147
+ },
148
+ "pol_Latn-por_Latn": {
149
+ "accuracy": 0.015810276679841896,
150
+ "f1": 0.010565159478202956,
151
+ "main_score": 0.010565159478202956,
152
+ "precision": 0.009140316205533596,
153
+ "recall": 0.015810276679841896
154
+ },
155
+ "rus_Cyrl-por_Latn": {
156
+ "accuracy": 0.012845849802371542,
157
+ "f1": 0.007076824142041533,
158
+ "main_score": 0.007076824142041533,
159
+ "precision": 0.005981337503076633,
160
+ "recall": 0.012845849802371542
161
+ },
162
+ "spa_Latn-por_Latn": {
163
+ "accuracy": 0.045454545454545456,
164
+ "f1": 0.030697574274649374,
165
+ "main_score": 0.030697574274649374,
166
+ "precision": 0.02646450771450771,
167
+ "recall": 0.045454545454545456
168
+ },
169
+ "swe_Latn-por_Latn": {
170
+ "accuracy": 0.01383399209486166,
171
+ "f1": 0.007234609460529993,
172
+ "main_score": 0.007234609460529993,
173
+ "precision": 0.006321303449514913,
174
+ "recall": 0.01383399209486166
175
+ },
176
+ "zho_Hans-por_Latn": {
177
+ "accuracy": 0.005928853754940711,
178
+ "f1": 0.002697785306480958,
179
+ "main_score": 0.002697785306480958,
180
+ "precision": 0.002091567852437418,
181
+ "recall": 0.005928853754940711
182
+ },
183
+ "zho_Hant-por_Latn": {
184
+ "accuracy": 0.01383399209486166,
185
+ "f1": 0.008330659788169669,
186
+ "main_score": 0.008330659788169669,
187
+ "precision": 0.007282925746655112,
188
+ "recall": 0.01383399209486166
189
+ }
190
+ },
191
+ "mteb_dataset_name": "FloresBitextMining",
192
+ "mteb_version": "1.7.41"
193
+ }
results/alfaneo/bert-base-multilingual-sts/MassiveIntentClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
3
+ "mteb_dataset_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 12.71,
7
+ "pt": {
8
+ "accuracy": 0.15487558843308674,
9
+ "accuracy_stderr": 0.0063286027470173985,
10
+ "f1": 0.15657302705205112,
11
+ "f1_stderr": 0.007300930787896792,
12
+ "main_score": 0.15487558843308674
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 123.75,
17
+ "pt": {
18
+ "accuracy": 0.14677816035415642,
19
+ "accuracy_stderr": 0.006855386675921042,
20
+ "f1": 0.14828128320850836,
21
+ "f1_stderr": 0.008537474152273315,
22
+ "main_score": 0.14677816035415642
23
+ }
24
+ }
25
+ }
results/alfaneo/bert-base-multilingual-sts/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
+ "mteb_dataset_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 7.28,
7
+ "pt": {
8
+ "accuracy": 0.16284465366509754,
9
+ "accuracy_stderr": 0.009721041949964029,
10
+ "f1": 0.1539128586537406,
11
+ "f1_stderr": 0.008679184602950123,
12
+ "main_score": 0.16284465366509754
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 7.44,
17
+ "pt": {
18
+ "accuracy": 0.1543531726512543,
19
+ "accuracy_stderr": 0.011463744845647763,
20
+ "f1": 0.14742885821928925,
21
+ "f1_stderr": 0.011332500439068817,
22
+ "main_score": 0.1543531726512543
23
+ }
24
+ }
25
+ }
results/alfaneo/bert-base-multilingual-sts/MintakaRetrieval.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "efa78cc2f74bbcd21eff2261f9e13aebe40b814e",
3
+ "mteb_dataset_name": "MintakaRetrieval",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 4.69,
7
+ "pt": {
8
+ "map_at_1": 0.00297,
9
+ "map_at_10": 0.0053,
10
+ "map_at_100": 0.00705,
11
+ "map_at_1000": 0.00847,
12
+ "map_at_20": 0.00586,
13
+ "map_at_3": 0.00425,
14
+ "map_at_5": 0.00452,
15
+ "mrr_at_1": 0.00297,
16
+ "mrr_at_10": 0.0053,
17
+ "mrr_at_100": 0.00705,
18
+ "mrr_at_1000": 0.00847,
19
+ "mrr_at_20": 0.00586,
20
+ "mrr_at_3": 0.00425,
21
+ "mrr_at_5": 0.00452,
22
+ "ndcg_at_1": 0.00297,
23
+ "ndcg_at_10": 0.00702,
24
+ "ndcg_at_100": 0.01943,
25
+ "ndcg_at_1000": 0.08112,
26
+ "ndcg_at_20": 0.00899,
27
+ "ndcg_at_3": 0.00468,
28
+ "ndcg_at_5": 0.00519,
29
+ "precision_at_1": 0.00297,
30
+ "precision_at_10": 0.00127,
31
+ "precision_at_100": 0.00081,
32
+ "precision_at_1000": 0.00062,
33
+ "precision_at_20": 0.00102,
34
+ "precision_at_3": 0.00198,
35
+ "precision_at_5": 0.00144,
36
+ "recall_at_1": 0.00297,
37
+ "recall_at_10": 0.01274,
38
+ "recall_at_100": 0.08071,
39
+ "recall_at_1000": 0.62341,
40
+ "recall_at_20": 0.02039,
41
+ "recall_at_3": 0.00595,
42
+ "recall_at_5": 0.00722
43
+ }
44
+ }
45
+ }
results/alfaneo/bert-base-multilingual-sts/MultiHateClassification.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ef137ea2b7c719183f8f60edf536b50f56d1365b",
3
+ "mteb_dataset_name": "MultiHateClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 4.07,
7
+ "por": {
8
+ "accuracy": 0.5269,
9
+ "accuracy_stderr": 0.04721747557843388,
10
+ "ap": 0.3101162757321531,
11
+ "ap_stderr": 0.012169866965014122,
12
+ "f1": 0.4933486444640828,
13
+ "f1_stderr": 0.023588982683831104,
14
+ "main_score": 0.5269
15
+ }
16
+ }
17
+ }
results/alfaneo/bert-base-multilingual-sts/MultiLongDocRetrieval.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d67138e705d963e346253a80e59676ddb418810a",
3
+ "dev": {
4
+ "evaluation_time": 27.19,
5
+ "pt": {
6
+ "map_at_1": 0.0,
7
+ "map_at_10": 0.00333,
8
+ "map_at_100": 0.00443,
9
+ "map_at_1000": 0.00508,
10
+ "map_at_20": 0.00369,
11
+ "map_at_3": 0.0025,
12
+ "map_at_5": 0.0025,
13
+ "mrr_at_1": 0.0,
14
+ "mrr_at_10": 0.00333,
15
+ "mrr_at_100": 0.00443,
16
+ "mrr_at_1000": 0.00508,
17
+ "mrr_at_20": 0.00369,
18
+ "mrr_at_3": 0.0025,
19
+ "mrr_at_5": 0.0025,
20
+ "ndcg_at_1": 0.0,
21
+ "ndcg_at_10": 0.00494,
22
+ "ndcg_at_100": 0.01373,
23
+ "ndcg_at_1000": 0.0446,
24
+ "ndcg_at_20": 0.00622,
25
+ "ndcg_at_3": 0.00315,
26
+ "ndcg_at_5": 0.00315,
27
+ "precision_at_1": 0.0,
28
+ "precision_at_10": 0.001,
29
+ "precision_at_100": 0.0006,
30
+ "precision_at_1000": 0.00034,
31
+ "precision_at_20": 0.00075,
32
+ "precision_at_3": 0.00167,
33
+ "precision_at_5": 0.001,
34
+ "recall_at_1": 0.0,
35
+ "recall_at_10": 0.01,
36
+ "recall_at_100": 0.06,
37
+ "recall_at_1000": 0.335,
38
+ "recall_at_20": 0.015,
39
+ "recall_at_3": 0.005,
40
+ "recall_at_5": 0.005
41
+ }
42
+ },
43
+ "mteb_dataset_name": "MultiLongDocRetrieval",
44
+ "mteb_version": "1.7.41",
45
+ "test": {
46
+ "evaluation_time": 26.33,
47
+ "pt": {
48
+ "map_at_1": 0.0,
49
+ "map_at_10": 0.00143,
50
+ "map_at_100": 0.00217,
51
+ "map_at_1000": 0.00278,
52
+ "map_at_20": 0.00172,
53
+ "map_at_3": 0.0,
54
+ "map_at_5": 0.0,
55
+ "mrr_at_1": 0.0,
56
+ "mrr_at_10": 0.00143,
57
+ "mrr_at_100": 0.00217,
58
+ "mrr_at_1000": 0.00278,
59
+ "mrr_at_20": 0.00172,
60
+ "mrr_at_3": 0.0,
61
+ "mrr_at_5": 0.0,
62
+ "ndcg_at_1": 0.0,
63
+ "ndcg_at_10": 0.00333,
64
+ "ndcg_at_100": 0.00815,
65
+ "ndcg_at_1000": 0.03353,
66
+ "ndcg_at_20": 0.00453,
67
+ "ndcg_at_3": 0.0,
68
+ "ndcg_at_5": 0.0,
69
+ "precision_at_1": 0.0,
70
+ "precision_at_10": 0.001,
71
+ "precision_at_100": 0.00035,
72
+ "precision_at_1000": 0.00026,
73
+ "precision_at_20": 0.00075,
74
+ "precision_at_3": 0.0,
75
+ "precision_at_5": 0.0,
76
+ "recall_at_1": 0.0,
77
+ "recall_at_10": 0.01,
78
+ "recall_at_100": 0.035,
79
+ "recall_at_1000": 0.255,
80
+ "recall_at_20": 0.015,
81
+ "recall_at_3": 0.0,
82
+ "recall_at_5": 0.0
83
+ }
84
+ }
85
+ }
results/alfaneo/bert-base-multilingual-sts/NTREXBitextMining.json ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "fd20d54141b6da952d5c68a2989472892489da0f",
3
+ "mteb_dataset_name": "NTREXBitextMining",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "arb_Arab-por_Latn": {
7
+ "accuracy": 0.005007511266900351,
8
+ "f1": 0.0027795661746588137,
9
+ "main_score": 0.0027795661746588137,
10
+ "precision": 0.0022325154398264067,
11
+ "recall": 0.005007511266900351
12
+ },
13
+ "deu_Latn-por_Latn": {
14
+ "accuracy": 0.013520280420630946,
15
+ "f1": 0.008953754847062895,
16
+ "main_score": 0.008953754847062895,
17
+ "precision": 0.007877845062970576,
18
+ "recall": 0.013520280420630946
19
+ },
20
+ "eng_Latn-por_Latn": {
21
+ "accuracy": 0.028042063094641963,
22
+ "f1": 0.017777495969784403,
23
+ "main_score": 0.017777495969784403,
24
+ "precision": 0.01506403377960714,
25
+ "recall": 0.028042063094641963
26
+ },
27
+ "evaluation_time": 28.76,
28
+ "fra_Latn-por_Latn": {
29
+ "accuracy": 0.02503755633450175,
30
+ "f1": 0.01645916505211079,
31
+ "main_score": 0.01645916505211079,
32
+ "precision": 0.014272456821878781,
33
+ "recall": 0.02503755633450175
34
+ },
35
+ "ita_Latn-por_Latn": {
36
+ "accuracy": 0.02303455182774161,
37
+ "f1": 0.01681927653384839,
38
+ "main_score": 0.01681927653384839,
39
+ "precision": 0.01494238015335451,
40
+ "recall": 0.02303455182774161
41
+ },
42
+ "jpn_Jpan-por_Latn": {
43
+ "accuracy": 0.005007511266900351,
44
+ "f1": 0.002305634053255485,
45
+ "main_score": 0.002305634053255485,
46
+ "precision": 0.0018560052964600965,
47
+ "recall": 0.005007511266900351
48
+ },
49
+ "kor_Hang-por_Latn": {
50
+ "accuracy": 0.010515773660490736,
51
+ "f1": 0.006482676449236197,
52
+ "main_score": 0.006482676449236197,
53
+ "precision": 0.005670482635930808,
54
+ "recall": 0.010515773660490736
55
+ },
56
+ "pol_Latn-por_Latn": {
57
+ "accuracy": 0.017526289434151226,
58
+ "f1": 0.012719945719444969,
59
+ "main_score": 0.012719945719444969,
60
+ "precision": 0.011316975463194793,
61
+ "recall": 0.017526289434151226
62
+ },
63
+ "rus_Cyrl-por_Latn": {
64
+ "accuracy": 0.010015022533800702,
65
+ "f1": 0.007603071273577032,
66
+ "main_score": 0.007603071273577032,
67
+ "precision": 0.006929044360191081,
68
+ "recall": 0.010015022533800702
69
+ },
70
+ "spa_Latn-por_Latn": {
71
+ "accuracy": 0.03104656985478217,
72
+ "f1": 0.02291174857524382,
73
+ "main_score": 0.02291174857524382,
74
+ "precision": 0.02052543100364833,
75
+ "recall": 0.03104656985478217
76
+ },
77
+ "swe_Latn-por_Latn": {
78
+ "accuracy": 0.014021031547320982,
79
+ "f1": 0.007495102257834592,
80
+ "main_score": 0.007495102257834592,
81
+ "precision": 0.0060554040946649975,
82
+ "recall": 0.014021031547320982
83
+ },
84
+ "zho_Hant-por_Latn": {
85
+ "accuracy": 0.011016524787180772,
86
+ "f1": 0.005830411286786639,
87
+ "main_score": 0.005830411286786639,
88
+ "precision": 0.004876362162291055,
89
+ "recall": 0.011016524787180772
90
+ }
91
+ }
92
+ }
results/alfaneo/bert-base-multilingual-sts/PTT_AILACasedocs.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_AILACasedocs",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.67,
7
+ "map_at_1": 0.0075,
8
+ "map_at_10": 0.0272,
9
+ "map_at_100": 0.05581,
10
+ "map_at_1000": 0.06403,
11
+ "map_at_20": 0.0344,
12
+ "map_at_3": 0.01028,
13
+ "map_at_5": 0.01917,
14
+ "mrr_at_1": 0.04,
15
+ "mrr_at_10": 0.08797,
16
+ "mrr_at_100": 0.10737,
17
+ "mrr_at_1000": 0.10855,
18
+ "mrr_at_20": 0.09666,
19
+ "mrr_at_3": 0.05333,
20
+ "mrr_at_5": 0.06833,
21
+ "ndcg_at_1": 0.04,
22
+ "ndcg_at_10": 0.06505,
23
+ "ndcg_at_100": 0.21787,
24
+ "ndcg_at_1000": 0.2832,
25
+ "ndcg_at_20": 0.08709,
26
+ "ndcg_at_3": 0.02816,
27
+ "ndcg_at_5": 0.04048,
28
+ "precision_at_1": 0.04,
29
+ "precision_at_10": 0.036,
30
+ "precision_at_100": 0.0262,
31
+ "precision_at_1000": 0.0039,
32
+ "precision_at_20": 0.03,
33
+ "precision_at_3": 0.02667,
34
+ "precision_at_5": 0.032,
35
+ "recall_at_1": 0.0075,
36
+ "recall_at_10": 0.11082,
37
+ "recall_at_100": 0.68473,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.17169,
40
+ "recall_at_3": 0.01583,
41
+ "recall_at_5": 0.04841
42
+ }
43
+ }
results/alfaneo/bert-base-multilingual-sts/PTT_AILAStatutes.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_AILAStatutes",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.41,
7
+ "map_at_1": 0.013,
8
+ "map_at_10": 0.06349,
9
+ "map_at_100": 0.13709,
10
+ "map_at_1000": 0.13709,
11
+ "map_at_20": 0.08775,
12
+ "map_at_3": 0.03383,
13
+ "map_at_5": 0.0456,
14
+ "mrr_at_1": 0.06,
15
+ "mrr_at_10": 0.1909,
16
+ "mrr_at_100": 0.21587,
17
+ "mrr_at_1000": 0.21587,
18
+ "mrr_at_20": 0.21219,
19
+ "mrr_at_3": 0.13667,
20
+ "mrr_at_5": 0.15967,
21
+ "ndcg_at_1": 0.06,
22
+ "ndcg_at_10": 0.1348,
23
+ "ndcg_at_100": 0.39645,
24
+ "ndcg_at_1000": 0.39645,
25
+ "ndcg_at_20": 0.21472,
26
+ "ndcg_at_3": 0.08123,
27
+ "ndcg_at_5": 0.08791,
28
+ "precision_at_1": 0.06,
29
+ "precision_at_10": 0.082,
30
+ "precision_at_100": 0.0434,
31
+ "precision_at_1000": 0.00434,
32
+ "precision_at_20": 0.083,
33
+ "precision_at_3": 0.08667,
34
+ "precision_at_5": 0.084,
35
+ "recall_at_1": 0.013,
36
+ "recall_at_10": 0.18533,
37
+ "recall_at_100": 1.0,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.38867,
40
+ "recall_at_3": 0.059,
41
+ "recall_at_5": 0.09567
42
+ }
43
+ }
results/alfaneo/bert-base-multilingual-sts/PTT_LegalBenchConsumerContractsQA.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchConsumerContractsQA",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.72,
7
+ "map_at_1": 0.0101,
8
+ "map_at_10": 0.03119,
9
+ "map_at_100": 0.04593,
10
+ "map_at_1000": 0.04828,
11
+ "map_at_20": 0.03505,
12
+ "map_at_3": 0.01768,
13
+ "map_at_5": 0.02449,
14
+ "mrr_at_1": 0.0101,
15
+ "mrr_at_10": 0.03119,
16
+ "mrr_at_100": 0.04593,
17
+ "mrr_at_1000": 0.04828,
18
+ "mrr_at_20": 0.03505,
19
+ "mrr_at_3": 0.01768,
20
+ "mrr_at_5": 0.02449,
21
+ "ndcg_at_1": 0.0101,
22
+ "ndcg_at_10": 0.04942,
23
+ "ndcg_at_100": 0.15761,
24
+ "ndcg_at_1000": 0.19987,
25
+ "ndcg_at_20": 0.06431,
26
+ "ndcg_at_3": 0.02026,
27
+ "ndcg_at_5": 0.03265,
28
+ "precision_at_1": 0.0101,
29
+ "precision_at_10": 0.01111,
30
+ "precision_at_100": 0.00705,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.00859,
33
+ "precision_at_3": 0.00926,
34
+ "precision_at_5": 0.01162,
35
+ "recall_at_1": 0.0101,
36
+ "recall_at_10": 0.11111,
37
+ "recall_at_100": 0.70455,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.17172,
40
+ "recall_at_3": 0.02778,
41
+ "recall_at_5": 0.05808
42
+ }
43
+ }
results/alfaneo/bert-base-multilingual-sts/PTT_LegalBenchCorporateLobbying.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchCorporateLobbying",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.97,
7
+ "map_at_1": 0.02941,
8
+ "map_at_10": 0.04726,
9
+ "map_at_100": 0.05966,
10
+ "map_at_1000": 0.06219,
11
+ "map_at_20": 0.05099,
12
+ "map_at_3": 0.03824,
13
+ "map_at_5": 0.04088,
14
+ "mrr_at_1": 0.02941,
15
+ "mrr_at_10": 0.04726,
16
+ "mrr_at_100": 0.05966,
17
+ "mrr_at_1000": 0.06219,
18
+ "mrr_at_20": 0.05099,
19
+ "mrr_at_3": 0.03824,
20
+ "mrr_at_5": 0.04088,
21
+ "ndcg_at_1": 0.02941,
22
+ "ndcg_at_10": 0.06193,
23
+ "ndcg_at_100": 0.14624,
24
+ "ndcg_at_1000": 0.20476,
25
+ "ndcg_at_20": 0.07635,
26
+ "ndcg_at_3": 0.04125,
27
+ "ndcg_at_5": 0.04606,
28
+ "precision_at_1": 0.02941,
29
+ "precision_at_10": 0.01118,
30
+ "precision_at_100": 0.00562,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.00853,
33
+ "precision_at_3": 0.01667,
34
+ "precision_at_5": 0.01235,
35
+ "recall_at_1": 0.02941,
36
+ "recall_at_10": 0.11176,
37
+ "recall_at_100": 0.56176,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.17059,
40
+ "recall_at_3": 0.05,
41
+ "recall_at_5": 0.06176
42
+ }
43
+ }
results/alfaneo/bertimbau-base-portuguese-sts/FloresBitextMining.json ADDED
@@ -0,0 +1,193 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "80dc3040d19756742c9a18267ab30f54fb8e226b",
3
+ "dev": {
4
+ "arb_Arab-por_Latn": {
5
+ "accuracy": 0.015045135406218655,
6
+ "f1": 0.011793628093614289,
7
+ "main_score": 0.011793628093614289,
8
+ "precision": 0.011581552758363225,
9
+ "recall": 0.015045135406218655
10
+ },
11
+ "deu_Latn-por_Latn": {
12
+ "accuracy": 0.10330992978936811,
13
+ "f1": 0.07953416222898997,
14
+ "main_score": 0.07953416222898997,
15
+ "precision": 0.07402791839673914,
16
+ "recall": 0.10330992978936811
17
+ },
18
+ "eng_Latn-por_Latn": {
19
+ "accuracy": 0.3119358074222668,
20
+ "f1": 0.26531580410474276,
21
+ "main_score": 0.26531580410474276,
22
+ "precision": 0.24905720187590388,
23
+ "recall": 0.3119358074222668
24
+ },
25
+ "evaluation_time": 19.62,
26
+ "fra_Latn-por_Latn": {
27
+ "accuracy": 0.15747241725175526,
28
+ "f1": 0.12880487417388942,
29
+ "main_score": 0.12880487417388942,
30
+ "precision": 0.12126768666788677,
31
+ "recall": 0.15747241725175526
32
+ },
33
+ "ita_Latn-por_Latn": {
34
+ "accuracy": 0.2657973921765296,
35
+ "f1": 0.22761879379999694,
36
+ "main_score": 0.22761879379999694,
37
+ "precision": 0.21697496416271717,
38
+ "recall": 0.2657973921765296
39
+ },
40
+ "jpn_Jpan-por_Latn": {
41
+ "accuracy": 0.012036108324974924,
42
+ "f1": 0.009101231992374831,
43
+ "main_score": 0.009101231992374831,
44
+ "precision": 0.008563893528488808,
45
+ "recall": 0.012036108324974924
46
+ },
47
+ "kor_Hang-por_Latn": {
48
+ "accuracy": 0.03911735205616851,
49
+ "f1": 0.03007843292731942,
50
+ "main_score": 0.03007843292731942,
51
+ "precision": 0.02807347349459479,
52
+ "recall": 0.03911735205616851
53
+ },
54
+ "pol_Latn-por_Latn": {
55
+ "accuracy": 0.05917753259779338,
56
+ "f1": 0.0472009169766556,
57
+ "main_score": 0.0472009169766556,
58
+ "precision": 0.04534829595842928,
59
+ "recall": 0.05917753259779338
60
+ },
61
+ "rus_Cyrl-por_Latn": {
62
+ "accuracy": 0.03009027081243731,
63
+ "f1": 0.026442593227535793,
64
+ "main_score": 0.026442593227535793,
65
+ "precision": 0.025591933725598683,
66
+ "recall": 0.03009027081243731
67
+ },
68
+ "spa_Latn-por_Latn": {
69
+ "accuracy": 0.7231695085255767,
70
+ "f1": 0.6705157744688099,
71
+ "main_score": 0.6705157744688099,
72
+ "precision": 0.6499257693716067,
73
+ "recall": 0.7231695085255767
74
+ },
75
+ "swe_Latn-por_Latn": {
76
+ "accuracy": 0.08726178535606821,
77
+ "f1": 0.06819934228679635,
78
+ "main_score": 0.06819934228679635,
79
+ "precision": 0.06401034922374804,
80
+ "recall": 0.08726178535606821
81
+ },
82
+ "zho_Hans-por_Latn": {
83
+ "accuracy": 0.033099297893681046,
84
+ "f1": 0.02515063776645484,
85
+ "main_score": 0.02515063776645484,
86
+ "precision": 0.023233677981002333,
87
+ "recall": 0.033099297893681046
88
+ },
89
+ "zho_Hant-por_Latn": {
90
+ "accuracy": 0.02106318956870612,
91
+ "f1": 0.015606294150362392,
92
+ "main_score": 0.015606294150362392,
93
+ "precision": 0.014406892220774873,
94
+ "recall": 0.02106318956870612
95
+ }
96
+ },
97
+ "devtest": {
98
+ "arb_Arab-por_Latn": {
99
+ "accuracy": 0.012845849802371542,
100
+ "f1": 0.008023504477999455,
101
+ "main_score": 0.008023504477999455,
102
+ "precision": 0.007397507961028132,
103
+ "recall": 0.012845849802371542
104
+ },
105
+ "deu_Latn-por_Latn": {
106
+ "accuracy": 0.08596837944664032,
107
+ "f1": 0.06763572696001201,
108
+ "main_score": 0.06763572696001201,
109
+ "precision": 0.06303862194044141,
110
+ "recall": 0.08596837944664032
111
+ },
112
+ "eng_Latn-por_Latn": {
113
+ "accuracy": 0.3241106719367589,
114
+ "f1": 0.26759484065561134,
115
+ "main_score": 0.26759484065561134,
116
+ "precision": 0.2506046267912388,
117
+ "recall": 0.3241106719367589
118
+ },
119
+ "evaluation_time": 16.96,
120
+ "fra_Latn-por_Latn": {
121
+ "accuracy": 0.11561264822134387,
122
+ "f1": 0.10160392555858946,
123
+ "main_score": 0.10160392555858946,
124
+ "precision": 0.09697654671871549,
125
+ "recall": 0.11561264822134387
126
+ },
127
+ "ita_Latn-por_Latn": {
128
+ "accuracy": 0.2598814229249012,
129
+ "f1": 0.21718396402512266,
130
+ "main_score": 0.21718396402512266,
131
+ "precision": 0.2057206254860409,
132
+ "recall": 0.2598814229249012
133
+ },
134
+ "jpn_Jpan-por_Latn": {
135
+ "accuracy": 0.014822134387351778,
136
+ "f1": 0.008742664561054147,
137
+ "main_score": 0.008742664561054147,
138
+ "precision": 0.008078388220312475,
139
+ "recall": 0.014822134387351778
140
+ },
141
+ "kor_Hang-por_Latn": {
142
+ "accuracy": 0.03359683794466403,
143
+ "f1": 0.022506107212681842,
144
+ "main_score": 0.022506107212681842,
145
+ "precision": 0.02043504654734871,
146
+ "recall": 0.03359683794466403
147
+ },
148
+ "pol_Latn-por_Latn": {
149
+ "accuracy": 0.029644268774703556,
150
+ "f1": 0.02254761333323333,
151
+ "main_score": 0.02254761333323333,
152
+ "precision": 0.021148947414929915,
153
+ "recall": 0.029644268774703556
154
+ },
155
+ "rus_Cyrl-por_Latn": {
156
+ "accuracy": 0.025691699604743084,
157
+ "f1": 0.02004301137091868,
158
+ "main_score": 0.02004301137091868,
159
+ "precision": 0.019325221781033346,
160
+ "recall": 0.025691699604743084
161
+ },
162
+ "spa_Latn-por_Latn": {
163
+ "accuracy": 0.7322134387351779,
164
+ "f1": 0.6882791651125355,
165
+ "main_score": 0.6882791651125355,
166
+ "precision": 0.6711144833427442,
167
+ "recall": 0.7322134387351779
168
+ },
169
+ "swe_Latn-por_Latn": {
170
+ "accuracy": 0.06818181818181818,
171
+ "f1": 0.05564320960184408,
172
+ "main_score": 0.05564320960184408,
173
+ "precision": 0.05251520465021167,
174
+ "recall": 0.06818181818181818
175
+ },
176
+ "zho_Hans-por_Latn": {
177
+ "accuracy": 0.029644268774703556,
178
+ "f1": 0.018883298083388417,
179
+ "main_score": 0.018883298083388417,
180
+ "precision": 0.016894074048160488,
181
+ "recall": 0.029644268774703556
182
+ },
183
+ "zho_Hant-por_Latn": {
184
+ "accuracy": 0.015810276679841896,
185
+ "f1": 0.010144081961380953,
186
+ "main_score": 0.010144081961380953,
187
+ "precision": 0.009158202951478482,
188
+ "recall": 0.015810276679841896
189
+ }
190
+ },
191
+ "mteb_dataset_name": "FloresBitextMining",
192
+ "mteb_version": "1.7.41"
193
+ }
results/alfaneo/bertimbau-base-portuguese-sts/MassiveIntentClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
3
+ "mteb_dataset_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 10.54,
7
+ "pt": {
8
+ "accuracy": 0.5967383994620039,
9
+ "accuracy_stderr": 0.010123283048100803,
10
+ "f1": 0.5772926380511152,
11
+ "f1_stderr": 0.009107249825568554,
12
+ "main_score": 0.5967383994620039
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 121.45,
17
+ "pt": {
18
+ "accuracy": 0.598425971470733,
19
+ "accuracy_stderr": 0.009239565063990995,
20
+ "f1": 0.5832960679009822,
21
+ "f1_stderr": 0.011680709704272752,
22
+ "main_score": 0.598425971470733
23
+ }
24
+ }
25
+ }
results/alfaneo/bertimbau-base-portuguese-sts/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
+ "mteb_dataset_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 6.62,
7
+ "pt": {
8
+ "accuracy": 0.6447881640887694,
9
+ "accuracy_stderr": 0.014658384764419134,
10
+ "f1": 0.648920710125899,
11
+ "f1_stderr": 0.012573043031073743,
12
+ "main_score": 0.6447881640887694
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 6.25,
17
+ "pt": {
18
+ "accuracy": 0.6377766847024102,
19
+ "accuracy_stderr": 0.016294949321647365,
20
+ "f1": 0.6455375816251432,
21
+ "f1_stderr": 0.01467022165617318,
22
+ "main_score": 0.6377766847024102
23
+ }
24
+ }
25
+ }
results/alfaneo/bertimbau-base-portuguese-sts/MintakaRetrieval.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "efa78cc2f74bbcd21eff2261f9e13aebe40b814e",
3
+ "mteb_dataset_name": "MintakaRetrieval",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 4.77,
7
+ "pt": {
8
+ "map_at_1": 0.09516,
9
+ "map_at_10": 0.14821,
10
+ "map_at_100": 0.15627,
11
+ "map_at_1000": 0.15765,
12
+ "map_at_20": 0.15216,
13
+ "map_at_3": 0.12957,
14
+ "map_at_5": 0.13959,
15
+ "mrr_at_1": 0.09516,
16
+ "mrr_at_10": 0.14821,
17
+ "mrr_at_100": 0.15627,
18
+ "mrr_at_1000": 0.15765,
19
+ "mrr_at_20": 0.15216,
20
+ "mrr_at_3": 0.12957,
21
+ "mrr_at_5": 0.13959,
22
+ "ndcg_at_1": 0.09516,
23
+ "ndcg_at_10": 0.18019,
24
+ "ndcg_at_100": 0.22572,
25
+ "ndcg_at_1000": 0.27344,
26
+ "ndcg_at_20": 0.19466,
27
+ "ndcg_at_3": 0.14101,
28
+ "ndcg_at_5": 0.15923,
29
+ "precision_at_1": 0.09516,
30
+ "precision_at_10": 0.02838,
31
+ "precision_at_100": 0.00511,
32
+ "precision_at_1000": 0.00091,
33
+ "precision_at_20": 0.01706,
34
+ "precision_at_3": 0.05806,
35
+ "precision_at_5": 0.04376,
36
+ "recall_at_1": 0.09516,
37
+ "recall_at_10": 0.28377,
38
+ "recall_at_100": 0.51147,
39
+ "recall_at_1000": 0.911,
40
+ "recall_at_20": 0.34112,
41
+ "recall_at_3": 0.17417,
42
+ "recall_at_5": 0.21878
43
+ }
44
+ }
45
+ }
results/alfaneo/bertimbau-base-portuguese-sts/MultiHateClassification.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ef137ea2b7c719183f8f60edf536b50f56d1365b",
3
+ "mteb_dataset_name": "MultiHateClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 3.95,
7
+ "por": {
8
+ "accuracy": 0.5929999999999999,
9
+ "accuracy_stderr": 0.0206929939834718,
10
+ "ap": 0.339705037731863,
11
+ "ap_stderr": 0.012308598494106447,
12
+ "f1": 0.5555685393751421,
13
+ "f1_stderr": 0.014506588828489517,
14
+ "main_score": 0.5929999999999999
15
+ }
16
+ }
17
+ }
results/alfaneo/bertimbau-base-portuguese-sts/MultiLongDocRetrieval.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d67138e705d963e346253a80e59676ddb418810a",
3
+ "dev": {
4
+ "evaluation_time": 27.21,
5
+ "pt": {
6
+ "map_at_1": 0.105,
7
+ "map_at_10": 0.17853,
8
+ "map_at_100": 0.18874,
9
+ "map_at_1000": 0.18982,
10
+ "map_at_20": 0.18421,
11
+ "map_at_3": 0.155,
12
+ "map_at_5": 0.17125,
13
+ "mrr_at_1": 0.105,
14
+ "mrr_at_10": 0.17853,
15
+ "mrr_at_100": 0.18874,
16
+ "mrr_at_1000": 0.18982,
17
+ "mrr_at_20": 0.18421,
18
+ "mrr_at_3": 0.155,
19
+ "mrr_at_5": 0.17125,
20
+ "ndcg_at_1": 0.105,
21
+ "ndcg_at_10": 0.21748,
22
+ "ndcg_at_100": 0.27442,
23
+ "ndcg_at_1000": 0.30795,
24
+ "ndcg_at_20": 0.23788,
25
+ "ndcg_at_3": 0.17167,
26
+ "ndcg_at_5": 0.20072,
27
+ "precision_at_1": 0.105,
28
+ "precision_at_10": 0.034,
29
+ "precision_at_100": 0.00625,
30
+ "precision_at_1000": 0.0009,
31
+ "precision_at_20": 0.021,
32
+ "precision_at_3": 0.07333,
33
+ "precision_at_5": 0.058,
34
+ "recall_at_1": 0.105,
35
+ "recall_at_10": 0.34,
36
+ "recall_at_100": 0.625,
37
+ "recall_at_1000": 0.9,
38
+ "recall_at_20": 0.42,
39
+ "recall_at_3": 0.22,
40
+ "recall_at_5": 0.29
41
+ }
42
+ },
43
+ "mteb_dataset_name": "MultiLongDocRetrieval",
44
+ "mteb_version": "1.7.41",
45
+ "test": {
46
+ "evaluation_time": 26.36,
47
+ "pt": {
48
+ "map_at_1": 0.13,
49
+ "map_at_10": 0.18589,
50
+ "map_at_100": 0.19602,
51
+ "map_at_1000": 0.19735,
52
+ "map_at_20": 0.19096,
53
+ "map_at_3": 0.165,
54
+ "map_at_5": 0.1775,
55
+ "mrr_at_1": 0.13,
56
+ "mrr_at_10": 0.18589,
57
+ "mrr_at_100": 0.19602,
58
+ "mrr_at_1000": 0.19735,
59
+ "mrr_at_20": 0.19096,
60
+ "mrr_at_3": 0.165,
61
+ "mrr_at_5": 0.1775,
62
+ "ndcg_at_1": 0.13,
63
+ "ndcg_at_10": 0.21989,
64
+ "ndcg_at_100": 0.27699,
65
+ "ndcg_at_1000": 0.31398,
66
+ "ndcg_at_20": 0.23787,
67
+ "ndcg_at_3": 0.17655,
68
+ "ndcg_at_5": 0.19914,
69
+ "precision_at_1": 0.13,
70
+ "precision_at_10": 0.033,
71
+ "precision_at_100": 0.00615,
72
+ "precision_at_1000": 0.00091,
73
+ "precision_at_20": 0.02,
74
+ "precision_at_3": 0.07,
75
+ "precision_at_5": 0.053,
76
+ "recall_at_1": 0.13,
77
+ "recall_at_10": 0.33,
78
+ "recall_at_100": 0.615,
79
+ "recall_at_1000": 0.91,
80
+ "recall_at_20": 0.4,
81
+ "recall_at_3": 0.21,
82
+ "recall_at_5": 0.265
83
+ }
84
+ }
85
+ }
results/alfaneo/bertimbau-base-portuguese-sts/NTREXBitextMining.json ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "fd20d54141b6da952d5c68a2989472892489da0f",
3
+ "mteb_dataset_name": "NTREXBitextMining",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "arb_Arab-por_Latn": {
7
+ "accuracy": 0.005508262393590386,
8
+ "f1": 0.004264484027866572,
9
+ "main_score": 0.004264484027866572,
10
+ "precision": 0.004177004466260877,
11
+ "recall": 0.005508262393590386
12
+ },
13
+ "deu_Latn-por_Latn": {
14
+ "accuracy": 0.06309464196294441,
15
+ "f1": 0.038357723921177596,
16
+ "main_score": 0.038357723921177596,
17
+ "precision": 0.03368747198315137,
18
+ "recall": 0.06309464196294441
19
+ },
20
+ "eng_Latn-por_Latn": {
21
+ "accuracy": 0.19929894842263396,
22
+ "f1": 0.15634043245643356,
23
+ "main_score": 0.15634043245643356,
24
+ "precision": 0.14514912631425797,
25
+ "recall": 0.19929894842263396
26
+ },
27
+ "evaluation_time": 35.86,
28
+ "fra_Latn-por_Latn": {
29
+ "accuracy": 0.07611417125688533,
30
+ "f1": 0.06618961758414593,
31
+ "main_score": 0.06618961758414593,
32
+ "precision": 0.06300018293809763,
33
+ "recall": 0.07611417125688533
34
+ },
35
+ "ita_Latn-por_Latn": {
36
+ "accuracy": 0.1872809213820731,
37
+ "f1": 0.14350574242976522,
38
+ "main_score": 0.14350574242976522,
39
+ "precision": 0.13151003415533655,
40
+ "recall": 0.1872809213820731
41
+ },
42
+ "jpn_Jpan-por_Latn": {
43
+ "accuracy": 0.007010515773660491,
44
+ "f1": 0.005602541364635209,
45
+ "main_score": 0.005602541364635209,
46
+ "precision": 0.005383716600541838,
47
+ "recall": 0.007010515773660491
48
+ },
49
+ "kor_Hang-por_Latn": {
50
+ "accuracy": 0.022533800701051578,
51
+ "f1": 0.016331084771293636,
52
+ "main_score": 0.016331084771293636,
53
+ "precision": 0.014920928669162208,
54
+ "recall": 0.022533800701051578
55
+ },
56
+ "pol_Latn-por_Latn": {
57
+ "accuracy": 0.049073610415623435,
58
+ "f1": 0.03217515331691158,
59
+ "main_score": 0.03217515331691158,
60
+ "precision": 0.028196667919373822,
61
+ "recall": 0.049073610415623435
62
+ },
63
+ "rus_Cyrl-por_Latn": {
64
+ "accuracy": 0.026039058587881823,
65
+ "f1": 0.01882465447280755,
66
+ "main_score": 0.01882465447280755,
67
+ "precision": 0.017414428809367163,
68
+ "recall": 0.026039058587881823
69
+ },
70
+ "spa_Latn-por_Latn": {
71
+ "accuracy": 0.613420130195293,
72
+ "f1": 0.5575881569803335,
73
+ "main_score": 0.5575881569803335,
74
+ "precision": 0.5377514515415447,
75
+ "recall": 0.613420130195293
76
+ },
77
+ "swe_Latn-por_Latn": {
78
+ "accuracy": 0.06059088632949424,
79
+ "f1": 0.039199158914683424,
80
+ "main_score": 0.039199158914683424,
81
+ "precision": 0.034888528150820365,
82
+ "recall": 0.06059088632949424
83
+ },
84
+ "zho_Hant-por_Latn": {
85
+ "accuracy": 0.03655483224837256,
86
+ "f1": 0.024207921492733032,
87
+ "main_score": 0.024207921492733032,
88
+ "precision": 0.021333506386573508,
89
+ "recall": 0.03655483224837256
90
+ }
91
+ }
92
+ }
results/alfaneo/bertimbau-base-portuguese-sts/PTT_AILACasedocs.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_AILACasedocs",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.64,
7
+ "map_at_1": 0.03417,
8
+ "map_at_10": 0.07817,
9
+ "map_at_100": 0.11914,
10
+ "map_at_1000": 0.1259,
11
+ "map_at_20": 0.09426,
12
+ "map_at_3": 0.06268,
13
+ "map_at_5": 0.07182,
14
+ "mrr_at_1": 0.12,
15
+ "mrr_at_10": 0.20922,
16
+ "mrr_at_100": 0.23035,
17
+ "mrr_at_1000": 0.23105,
18
+ "mrr_at_20": 0.22264,
19
+ "mrr_at_3": 0.18333,
20
+ "mrr_at_5": 0.19633,
21
+ "ndcg_at_1": 0.12,
22
+ "ndcg_at_10": 0.13195,
23
+ "ndcg_at_100": 0.30212,
24
+ "ndcg_at_1000": 0.35192,
25
+ "ndcg_at_20": 0.18044,
26
+ "ndcg_at_3": 0.12554,
27
+ "ndcg_at_5": 0.12458,
28
+ "precision_at_1": 0.12,
29
+ "precision_at_10": 0.056,
30
+ "precision_at_100": 0.0284,
31
+ "precision_at_1000": 0.0039,
32
+ "precision_at_20": 0.054,
33
+ "precision_at_3": 0.11333,
34
+ "precision_at_5": 0.088,
35
+ "recall_at_1": 0.03417,
36
+ "recall_at_10": 0.16323,
37
+ "recall_at_100": 0.76916,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.29603,
40
+ "recall_at_3": 0.08071,
41
+ "recall_at_5": 0.116
42
+ }
43
+ }
results/alfaneo/bertimbau-base-portuguese-sts/PTT_AILAStatutes.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_AILAStatutes",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.4,
7
+ "map_at_1": 0.017,
8
+ "map_at_10": 0.09252,
9
+ "map_at_100": 0.17482,
10
+ "map_at_1000": 0.17482,
11
+ "map_at_20": 0.12249,
12
+ "map_at_3": 0.05756,
13
+ "map_at_5": 0.07046,
14
+ "mrr_at_1": 0.08,
15
+ "mrr_at_10": 0.26378,
16
+ "mrr_at_100": 0.28506,
17
+ "mrr_at_1000": 0.28506,
18
+ "mrr_at_20": 0.28269,
19
+ "mrr_at_3": 0.21,
20
+ "mrr_at_5": 0.245,
21
+ "ndcg_at_1": 0.08,
22
+ "ndcg_at_10": 0.17982,
23
+ "ndcg_at_100": 0.43624,
24
+ "ndcg_at_1000": 0.43624,
25
+ "ndcg_at_20": 0.267,
26
+ "ndcg_at_3": 0.13207,
27
+ "ndcg_at_5": 0.133,
28
+ "precision_at_1": 0.08,
29
+ "precision_at_10": 0.102,
30
+ "precision_at_100": 0.0434,
31
+ "precision_at_1000": 0.00434,
32
+ "precision_at_20": 0.099,
33
+ "precision_at_3": 0.14,
34
+ "precision_at_5": 0.124,
35
+ "recall_at_1": 0.017,
36
+ "recall_at_10": 0.23567,
37
+ "recall_at_100": 1.0,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.45,
40
+ "recall_at_3": 0.09367,
41
+ "recall_at_5": 0.14467
42
+ }
43
+ }
results/alfaneo/bertimbau-base-portuguese-sts/PTT_LegalBenchConsumerContractsQA.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchConsumerContractsQA",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.74,
7
+ "map_at_1": 0.23232,
8
+ "map_at_10": 0.3618,
9
+ "map_at_100": 0.37194,
10
+ "map_at_1000": 0.3725,
11
+ "map_at_20": 0.36691,
12
+ "map_at_3": 0.32365,
13
+ "map_at_5": 0.34638,
14
+ "mrr_at_1": 0.23232,
15
+ "mrr_at_10": 0.3618,
16
+ "mrr_at_100": 0.37194,
17
+ "mrr_at_1000": 0.3725,
18
+ "mrr_at_20": 0.36691,
19
+ "mrr_at_3": 0.32365,
20
+ "mrr_at_5": 0.34638,
21
+ "ndcg_at_1": 0.23232,
22
+ "ndcg_at_10": 0.43229,
23
+ "ndcg_at_100": 0.48762,
24
+ "ndcg_at_1000": 0.49773,
25
+ "ndcg_at_20": 0.45042,
26
+ "ndcg_at_3": 0.35398,
27
+ "ndcg_at_5": 0.39527,
28
+ "precision_at_1": 0.23232,
29
+ "precision_at_10": 0.06566,
30
+ "precision_at_100": 0.00929,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.03636,
33
+ "precision_at_3": 0.14731,
34
+ "precision_at_5": 0.10859,
35
+ "recall_at_1": 0.23232,
36
+ "recall_at_10": 0.65657,
37
+ "recall_at_100": 0.92929,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.72727,
40
+ "recall_at_3": 0.44192,
41
+ "recall_at_5": 0.54293
42
+ }
43
+ }
results/alfaneo/bertimbau-base-portuguese-sts/PTT_LegalBenchCorporateLobbying.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchCorporateLobbying",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.89,
7
+ "map_at_1": 0.52353,
8
+ "map_at_10": 0.63712,
9
+ "map_at_100": 0.6426,
10
+ "map_at_1000": 0.64274,
11
+ "map_at_20": 0.64125,
12
+ "map_at_3": 0.61127,
13
+ "map_at_5": 0.62407,
14
+ "mrr_at_1": 0.52353,
15
+ "mrr_at_10": 0.63712,
16
+ "mrr_at_100": 0.6426,
17
+ "mrr_at_1000": 0.64274,
18
+ "mrr_at_20": 0.64125,
19
+ "mrr_at_3": 0.61127,
20
+ "mrr_at_5": 0.62407,
21
+ "ndcg_at_1": 0.52353,
22
+ "ndcg_at_10": 0.69335,
23
+ "ndcg_at_100": 0.71759,
24
+ "ndcg_at_1000": 0.72043,
25
+ "ndcg_at_20": 0.7078,
26
+ "ndcg_at_3": 0.63939,
27
+ "ndcg_at_5": 0.66242,
28
+ "precision_at_1": 0.52353,
29
+ "precision_at_10": 0.08706,
30
+ "precision_at_100": 0.00979,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.04632,
33
+ "precision_at_3": 0.2402,
34
+ "precision_at_5": 0.15529,
35
+ "recall_at_1": 0.52353,
36
+ "recall_at_10": 0.87059,
37
+ "recall_at_100": 0.97941,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.92647,
40
+ "recall_at_3": 0.72059,
41
+ "recall_at_5": 0.77647
42
+ }
43
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/FloresBitextMining.json ADDED
@@ -0,0 +1,193 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "80dc3040d19756742c9a18267ab30f54fb8e226b",
3
+ "dev": {
4
+ "arb_Arab-por_Latn": {
5
+ "accuracy": 0.003009027081243731,
6
+ "f1": 0.0010469126678280456,
7
+ "main_score": 0.0010469126678280456,
8
+ "precision": 0.001025406556193028,
9
+ "recall": 0.003009027081243731
10
+ },
11
+ "deu_Latn-por_Latn": {
12
+ "accuracy": 0.004012036108324975,
13
+ "f1": 0.00202331820813038,
14
+ "main_score": 0.00202331820813038,
15
+ "precision": 0.0020147257649851106,
16
+ "recall": 0.004012036108324975
17
+ },
18
+ "eng_Latn-por_Latn": {
19
+ "accuracy": 0.09729187562688064,
20
+ "f1": 0.0831264599482296,
21
+ "main_score": 0.0831264599482296,
22
+ "precision": 0.08007725235681022,
23
+ "recall": 0.09729187562688064
24
+ },
25
+ "evaluation_time": 19.75,
26
+ "fra_Latn-por_Latn": {
27
+ "accuracy": 0.015045135406218655,
28
+ "f1": 0.011790739863611715,
29
+ "main_score": 0.011790739863611715,
30
+ "precision": 0.01150594979086824,
31
+ "recall": 0.015045135406218655
32
+ },
33
+ "ita_Latn-por_Latn": {
34
+ "accuracy": 0.01905717151454363,
35
+ "f1": 0.01415186368986062,
36
+ "main_score": 0.01415186368986062,
37
+ "precision": 0.013529891562390546,
38
+ "recall": 0.01905717151454363
39
+ },
40
+ "jpn_Jpan-por_Latn": {
41
+ "accuracy": 0.006018054162487462,
42
+ "f1": 0.003679736332517009,
43
+ "main_score": 0.003679736332517009,
44
+ "precision": 0.0032106482047769324,
45
+ "recall": 0.006018054162487462
46
+ },
47
+ "kor_Hang-por_Latn": {
48
+ "accuracy": 0.01805416248746239,
49
+ "f1": 0.014351481039563718,
50
+ "main_score": 0.014351481039563718,
51
+ "precision": 0.013719287209182215,
52
+ "recall": 0.01805416248746239
53
+ },
54
+ "pol_Latn-por_Latn": {
55
+ "accuracy": 0.009027081243731194,
56
+ "f1": 0.0028148804745262096,
57
+ "main_score": 0.0028148804745262096,
58
+ "precision": 0.002447501949047557,
59
+ "recall": 0.009027081243731194
60
+ },
61
+ "rus_Cyrl-por_Latn": {
62
+ "accuracy": 0.009027081243731194,
63
+ "f1": 0.007070941576991372,
64
+ "main_score": 0.007070941576991372,
65
+ "precision": 0.007046585968015566,
66
+ "recall": 0.009027081243731194
67
+ },
68
+ "spa_Latn-por_Latn": {
69
+ "accuracy": 0.12337011033099297,
70
+ "f1": 0.11764280600335085,
71
+ "main_score": 0.11764280600335085,
72
+ "precision": 0.11700428323708087,
73
+ "recall": 0.12337011033099297
74
+ },
75
+ "swe_Latn-por_Latn": {
76
+ "accuracy": 0.012036108324974924,
77
+ "f1": 0.007383254210331279,
78
+ "main_score": 0.007383254210331279,
79
+ "precision": 0.007219174311984182,
80
+ "recall": 0.012036108324974924
81
+ },
82
+ "zho_Hans-por_Latn": {
83
+ "accuracy": 0.004012036108324975,
84
+ "f1": 0.002151345768157546,
85
+ "main_score": 0.002151345768157546,
86
+ "precision": 0.0020841939888284625,
87
+ "recall": 0.004012036108324975
88
+ },
89
+ "zho_Hant-por_Latn": {
90
+ "accuracy": 0.003009027081243731,
91
+ "f1": 0.002008032128514056,
92
+ "main_score": 0.002008032128514056,
93
+ "precision": 0.002007026103435936,
94
+ "recall": 0.003009027081243731
95
+ }
96
+ },
97
+ "devtest": {
98
+ "arb_Arab-por_Latn": {
99
+ "accuracy": 0.005928853754940711,
100
+ "f1": 0.0028198098444691564,
101
+ "main_score": 0.0028198098444691564,
102
+ "precision": 0.0023157842566320147,
103
+ "recall": 0.005928853754940711
104
+ },
105
+ "deu_Latn-por_Latn": {
106
+ "accuracy": 0.00691699604743083,
107
+ "f1": 0.0034379273013823387,
108
+ "main_score": 0.0034379273013823387,
109
+ "precision": 0.0032388407723443376,
110
+ "recall": 0.00691699604743083
111
+ },
112
+ "eng_Latn-por_Latn": {
113
+ "accuracy": 0.06324110671936758,
114
+ "f1": 0.056852859891237416,
115
+ "main_score": 0.056852859891237416,
116
+ "precision": 0.05580698833247146,
117
+ "recall": 0.06324110671936758
118
+ },
119
+ "evaluation_time": 20.97,
120
+ "fra_Latn-por_Latn": {
121
+ "accuracy": 0.029644268774703556,
122
+ "f1": 0.025230860196068784,
123
+ "main_score": 0.025230860196068784,
124
+ "precision": 0.024526671427060193,
125
+ "recall": 0.029644268774703556
126
+ },
127
+ "ita_Latn-por_Latn": {
128
+ "accuracy": 0.05434782608695652,
129
+ "f1": 0.04806680776013462,
130
+ "main_score": 0.04806680776013462,
131
+ "precision": 0.04687597453342339,
132
+ "recall": 0.05434782608695652
133
+ },
134
+ "jpn_Jpan-por_Latn": {
135
+ "accuracy": 0.00691699604743083,
136
+ "f1": 0.004415698486844732,
137
+ "main_score": 0.004415698486844732,
138
+ "precision": 0.003953568302713245,
139
+ "recall": 0.00691699604743083
140
+ },
141
+ "kor_Hang-por_Latn": {
142
+ "accuracy": 0.023715415019762844,
143
+ "f1": 0.01542338175697726,
144
+ "main_score": 0.01542338175697726,
145
+ "precision": 0.013492470082079438,
146
+ "recall": 0.023715415019762844
147
+ },
148
+ "pol_Latn-por_Latn": {
149
+ "accuracy": 0.0029644268774703555,
150
+ "f1": 0.0012098408837539272,
151
+ "main_score": 0.0012098408837539272,
152
+ "precision": 0.0011127169157278433,
153
+ "recall": 0.0029644268774703555
154
+ },
155
+ "rus_Cyrl-por_Latn": {
156
+ "accuracy": 0.009881422924901186,
157
+ "f1": 0.005244187368937715,
158
+ "main_score": 0.005244187368937715,
159
+ "precision": 0.00474476652253705,
160
+ "recall": 0.009881422924901186
161
+ },
162
+ "spa_Latn-por_Latn": {
163
+ "accuracy": 0.26976284584980237,
164
+ "f1": 0.259325070715698,
165
+ "main_score": 0.259325070715698,
166
+ "precision": 0.25644529142355227,
167
+ "recall": 0.26976284584980237
168
+ },
169
+ "swe_Latn-por_Latn": {
170
+ "accuracy": 0.010869565217391304,
171
+ "f1": 0.006199976895822091,
172
+ "main_score": 0.006199976895822091,
173
+ "precision": 0.005789734782576482,
174
+ "recall": 0.010869565217391304
175
+ },
176
+ "zho_Hans-por_Latn": {
177
+ "accuracy": 0.00691699604743083,
178
+ "f1": 0.005107392060629943,
179
+ "main_score": 0.005107392060629943,
180
+ "precision": 0.004777017191550852,
181
+ "recall": 0.00691699604743083
182
+ },
183
+ "zho_Hant-por_Latn": {
184
+ "accuracy": 0.005928853754940711,
185
+ "f1": 0.004942677914773957,
186
+ "main_score": 0.004942677914773957,
187
+ "precision": 0.0049416956679212,
188
+ "recall": 0.005928853754940711
189
+ }
190
+ },
191
+ "mteb_dataset_name": "FloresBitextMining",
192
+ "mteb_version": "1.7.41"
193
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/MassiveIntentClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
3
+ "mteb_dataset_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 12.07,
7
+ "pt": {
8
+ "accuracy": 0.5683927370544721,
9
+ "accuracy_stderr": 0.013317437842083713,
10
+ "f1": 0.5429420381694829,
11
+ "f1_stderr": 0.011420990268208564,
12
+ "main_score": 0.5683927370544721
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 125.37,
17
+ "pt": {
18
+ "accuracy": 0.5632562715199214,
19
+ "accuracy_stderr": 0.013450034782957594,
20
+ "f1": 0.5464274397903219,
21
+ "f1_stderr": 0.009554901036188113,
22
+ "main_score": 0.5632562715199214
23
+ }
24
+ }
25
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
+ "mteb_dataset_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 6.52,
7
+ "pt": {
8
+ "accuracy": 0.621318090114324,
9
+ "accuracy_stderr": 0.014989096448074427,
10
+ "f1": 0.6114759387750228,
11
+ "f1_stderr": 0.009988067774728639,
12
+ "main_score": 0.621318090114324
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 6.28,
17
+ "pt": {
18
+ "accuracy": 0.60486965076242,
19
+ "accuracy_stderr": 0.018533060622122035,
20
+ "f1": 0.6035802699729752,
21
+ "f1_stderr": 0.0141825299951036,
22
+ "main_score": 0.60486965076242
23
+ }
24
+ }
25
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/MintakaRetrieval.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "efa78cc2f74bbcd21eff2261f9e13aebe40b814e",
3
+ "mteb_dataset_name": "MintakaRetrieval",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 4.54,
7
+ "pt": {
8
+ "map_at_1": 0.04673,
9
+ "map_at_10": 0.07078,
10
+ "map_at_100": 0.07681,
11
+ "map_at_1000": 0.07824,
12
+ "map_at_20": 0.07344,
13
+ "map_at_3": 0.06167,
14
+ "map_at_5": 0.06643,
15
+ "mrr_at_1": 0.04673,
16
+ "mrr_at_10": 0.07078,
17
+ "mrr_at_100": 0.07681,
18
+ "mrr_at_1000": 0.07824,
19
+ "mrr_at_20": 0.07344,
20
+ "mrr_at_3": 0.06167,
21
+ "mrr_at_5": 0.06643,
22
+ "ndcg_at_1": 0.04673,
23
+ "ndcg_at_10": 0.0862,
24
+ "ndcg_at_100": 0.12213,
25
+ "ndcg_at_1000": 0.17657,
26
+ "ndcg_at_20": 0.09602,
27
+ "ndcg_at_3": 0.06706,
28
+ "ndcg_at_5": 0.07564,
29
+ "precision_at_1": 0.04673,
30
+ "precision_at_10": 0.01364,
31
+ "precision_at_100": 0.0032,
32
+ "precision_at_1000": 0.00079,
33
+ "precision_at_20": 0.00877,
34
+ "precision_at_3": 0.02761,
35
+ "precision_at_5": 0.02073,
36
+ "recall_at_1": 0.04673,
37
+ "recall_at_10": 0.13636,
38
+ "recall_at_100": 0.31988,
39
+ "recall_at_1000": 0.78483,
40
+ "recall_at_20": 0.17545,
41
+ "recall_at_3": 0.08284,
42
+ "recall_at_5": 0.10365
43
+ }
44
+ }
45
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/MultiHateClassification.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ef137ea2b7c719183f8f60edf536b50f56d1365b",
3
+ "mteb_dataset_name": "MultiHateClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 3.92,
7
+ "por": {
8
+ "accuracy": 0.5974,
9
+ "accuracy_stderr": 0.05656182458160272,
10
+ "ap": 0.35727142525319355,
11
+ "ap_stderr": 0.02758086672641401,
12
+ "f1": 0.5714785764992408,
13
+ "f1_stderr": 0.04506226806965137,
14
+ "main_score": 0.5974
15
+ }
16
+ }
17
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/MultiLongDocRetrieval.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d67138e705d963e346253a80e59676ddb418810a",
3
+ "dev": {
4
+ "evaluation_time": 32.88,
5
+ "pt": {
6
+ "map_at_1": 0.085,
7
+ "map_at_10": 0.11794,
8
+ "map_at_100": 0.13034,
9
+ "map_at_1000": 0.13159,
10
+ "map_at_20": 0.12405,
11
+ "map_at_3": 0.10167,
12
+ "map_at_5": 0.11042,
13
+ "mrr_at_1": 0.085,
14
+ "mrr_at_10": 0.11794,
15
+ "mrr_at_100": 0.13034,
16
+ "mrr_at_1000": 0.13159,
17
+ "mrr_at_20": 0.12405,
18
+ "mrr_at_3": 0.10167,
19
+ "mrr_at_5": 0.11042,
20
+ "ndcg_at_1": 0.085,
21
+ "ndcg_at_10": 0.14269,
22
+ "ndcg_at_100": 0.20883,
23
+ "ndcg_at_1000": 0.2471,
24
+ "ndcg_at_20": 0.16611,
25
+ "ndcg_at_3": 0.10762,
26
+ "ndcg_at_5": 0.12375,
27
+ "precision_at_1": 0.085,
28
+ "precision_at_10": 0.0225,
29
+ "precision_at_100": 0.00545,
30
+ "precision_at_1000": 0.00086,
31
+ "precision_at_20": 0.016,
32
+ "precision_at_3": 0.04167,
33
+ "precision_at_5": 0.033,
34
+ "recall_at_1": 0.085,
35
+ "recall_at_10": 0.225,
36
+ "recall_at_100": 0.545,
37
+ "recall_at_1000": 0.855,
38
+ "recall_at_20": 0.32,
39
+ "recall_at_3": 0.125,
40
+ "recall_at_5": 0.165
41
+ }
42
+ },
43
+ "mteb_dataset_name": "MultiLongDocRetrieval",
44
+ "mteb_version": "1.7.41",
45
+ "test": {
46
+ "evaluation_time": 32.21,
47
+ "pt": {
48
+ "map_at_1": 0.065,
49
+ "map_at_10": 0.10531,
50
+ "map_at_100": 0.11854,
51
+ "map_at_1000": 0.12011,
52
+ "map_at_20": 0.11278,
53
+ "map_at_3": 0.09417,
54
+ "map_at_5": 0.10017,
55
+ "mrr_at_1": 0.065,
56
+ "mrr_at_10": 0.10531,
57
+ "mrr_at_100": 0.11854,
58
+ "mrr_at_1000": 0.12011,
59
+ "mrr_at_20": 0.11278,
60
+ "mrr_at_3": 0.09417,
61
+ "mrr_at_5": 0.10017,
62
+ "ndcg_at_1": 0.065,
63
+ "ndcg_at_10": 0.12669,
64
+ "ndcg_at_100": 0.19699,
65
+ "ndcg_at_1000": 0.23863,
66
+ "ndcg_at_20": 0.15265,
67
+ "ndcg_at_3": 0.10339,
68
+ "ndcg_at_5": 0.11394,
69
+ "precision_at_1": 0.065,
70
+ "precision_at_10": 0.0195,
71
+ "precision_at_100": 0.0054,
72
+ "precision_at_1000": 0.00087,
73
+ "precision_at_20": 0.01475,
74
+ "precision_at_3": 0.04333,
75
+ "precision_at_5": 0.031,
76
+ "recall_at_1": 0.065,
77
+ "recall_at_10": 0.195,
78
+ "recall_at_100": 0.54,
79
+ "recall_at_1000": 0.87,
80
+ "recall_at_20": 0.295,
81
+ "recall_at_3": 0.13,
82
+ "recall_at_5": 0.155
83
+ }
84
+ }
85
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/NTREXBitextMining.json ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "fd20d54141b6da952d5c68a2989472892489da0f",
3
+ "mteb_dataset_name": "NTREXBitextMining",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "arb_Arab-por_Latn": {
7
+ "accuracy": 0.006009013520280421,
8
+ "f1": 0.0029705451033082636,
9
+ "main_score": 0.0029705451033082636,
10
+ "precision": 0.002769634716250188,
11
+ "recall": 0.006009013520280421
12
+ },
13
+ "deu_Latn-por_Latn": {
14
+ "accuracy": 0.006509764646970456,
15
+ "f1": 0.0042098011934932926,
16
+ "main_score": 0.0042098011934932926,
17
+ "precision": 0.0038745787773491525,
18
+ "recall": 0.006509764646970456
19
+ },
20
+ "eng_Latn-por_Latn": {
21
+ "accuracy": 0.018027040560841263,
22
+ "f1": 0.015992716063797263,
23
+ "main_score": 0.015992716063797263,
24
+ "precision": 0.015705144496998592,
25
+ "recall": 0.018027040560841263
26
+ },
27
+ "evaluation_time": 34.19,
28
+ "fra_Latn-por_Latn": {
29
+ "accuracy": 0.00801201802704056,
30
+ "f1": 0.007344856691244694,
31
+ "main_score": 0.007344856691244694,
32
+ "precision": 0.0072611448818797825,
33
+ "recall": 0.00801201802704056
34
+ },
35
+ "ita_Latn-por_Latn": {
36
+ "accuracy": 0.02303455182774161,
37
+ "f1": 0.018203763386108122,
38
+ "main_score": 0.018203763386108122,
39
+ "precision": 0.017366903840537882,
40
+ "recall": 0.02303455182774161
41
+ },
42
+ "jpn_Jpan-por_Latn": {
43
+ "accuracy": 0.006509764646970456,
44
+ "f1": 0.004266244013098337,
45
+ "main_score": 0.004266244013098337,
46
+ "precision": 0.0038525530230830115,
47
+ "recall": 0.006509764646970456
48
+ },
49
+ "kor_Hang-por_Latn": {
50
+ "accuracy": 0.018027040560841263,
51
+ "f1": 0.0119429857223451,
52
+ "main_score": 0.0119429857223451,
53
+ "precision": 0.010973116995071008,
54
+ "recall": 0.018027040560841263
55
+ },
56
+ "pol_Latn-por_Latn": {
57
+ "accuracy": 0.009514271407110666,
58
+ "f1": 0.005815280204843705,
59
+ "main_score": 0.005815280204843705,
60
+ "precision": 0.005232546224147646,
61
+ "recall": 0.009514271407110666
62
+ },
63
+ "rus_Cyrl-por_Latn": {
64
+ "accuracy": 0.013520280420630946,
65
+ "f1": 0.008146237464386944,
66
+ "main_score": 0.008146237464386944,
67
+ "precision": 0.007198219484836695,
68
+ "recall": 0.013520280420630946
69
+ },
70
+ "spa_Latn-por_Latn": {
71
+ "accuracy": 0.07310966449674512,
72
+ "f1": 0.0695840992986419,
73
+ "main_score": 0.0695840992986419,
74
+ "precision": 0.06878679191764633,
75
+ "recall": 0.07310966449674512
76
+ },
77
+ "swe_Latn-por_Latn": {
78
+ "accuracy": 0.011517275913870806,
79
+ "f1": 0.006959420380990229,
80
+ "main_score": 0.006959420380990229,
81
+ "precision": 0.006495683738748116,
82
+ "recall": 0.011517275913870806
83
+ },
84
+ "zho_Hant-por_Latn": {
85
+ "accuracy": 0.017526289434151226,
86
+ "f1": 0.011642239018597672,
87
+ "main_score": 0.011642239018597672,
88
+ "precision": 0.010552851615853988,
89
+ "recall": 0.017526289434151226
90
+ }
91
+ }
92
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/PTT_AILACasedocs.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_AILACasedocs",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.92,
7
+ "map_at_1": 0.02483,
8
+ "map_at_10": 0.05714,
9
+ "map_at_100": 0.08403,
10
+ "map_at_1000": 0.09231,
11
+ "map_at_20": 0.06401,
12
+ "map_at_3": 0.0415,
13
+ "map_at_5": 0.04367,
14
+ "mrr_at_1": 0.1,
15
+ "mrr_at_10": 0.18488,
16
+ "mrr_at_100": 0.20242,
17
+ "mrr_at_1000": 0.20333,
18
+ "mrr_at_20": 0.19432,
19
+ "mrr_at_3": 0.15,
20
+ "mrr_at_5": 0.16,
21
+ "ndcg_at_1": 0.1,
22
+ "ndcg_at_10": 0.11175,
23
+ "ndcg_at_100": 0.25009,
24
+ "ndcg_at_1000": 0.32029,
25
+ "ndcg_at_20": 0.13951,
26
+ "ndcg_at_3": 0.0849,
27
+ "ndcg_at_5": 0.08185,
28
+ "precision_at_1": 0.1,
29
+ "precision_at_10": 0.058,
30
+ "precision_at_100": 0.0248,
31
+ "precision_at_1000": 0.0039,
32
+ "precision_at_20": 0.044,
33
+ "precision_at_3": 0.07333,
34
+ "precision_at_5": 0.056,
35
+ "recall_at_1": 0.02483,
36
+ "recall_at_10": 0.14073,
37
+ "recall_at_100": 0.66009,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.22051,
40
+ "recall_at_3": 0.06539,
41
+ "recall_at_5": 0.07156
42
+ }
43
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/PTT_AILAStatutes.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_AILAStatutes",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.57,
7
+ "map_at_1": 0.038,
8
+ "map_at_10": 0.08333,
9
+ "map_at_100": 0.14184,
10
+ "map_at_1000": 0.14184,
11
+ "map_at_20": 0.09871,
12
+ "map_at_3": 0.069,
13
+ "map_at_5": 0.07575,
14
+ "mrr_at_1": 0.18,
15
+ "mrr_at_10": 0.26875,
16
+ "mrr_at_100": 0.2938,
17
+ "mrr_at_1000": 0.2938,
18
+ "mrr_at_20": 0.28565,
19
+ "mrr_at_3": 0.24333,
20
+ "mrr_at_5": 0.25333,
21
+ "ndcg_at_1": 0.18,
22
+ "ndcg_at_10": 0.14975,
23
+ "ndcg_at_100": 0.40273,
24
+ "ndcg_at_1000": 0.40273,
25
+ "ndcg_at_20": 0.2037,
26
+ "ndcg_at_3": 0.14608,
27
+ "ndcg_at_5": 0.12772,
28
+ "precision_at_1": 0.18,
29
+ "precision_at_10": 0.066,
30
+ "precision_at_100": 0.0434,
31
+ "precision_at_1000": 0.00434,
32
+ "precision_at_20": 0.063,
33
+ "precision_at_3": 0.13333,
34
+ "precision_at_5": 0.096,
35
+ "recall_at_1": 0.038,
36
+ "recall_at_10": 0.16467,
37
+ "recall_at_100": 1.0,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.30033,
40
+ "recall_at_3": 0.103,
41
+ "recall_at_5": 0.121
42
+ }
43
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/PTT_LegalBenchConsumerContractsQA.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchConsumerContractsQA",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.85,
7
+ "map_at_1": 0.12374,
8
+ "map_at_10": 0.19657,
9
+ "map_at_100": 0.21315,
10
+ "map_at_1000": 0.21432,
11
+ "map_at_20": 0.20451,
12
+ "map_at_3": 0.17088,
13
+ "map_at_5": 0.183,
14
+ "mrr_at_1": 0.12374,
15
+ "mrr_at_10": 0.19657,
16
+ "mrr_at_100": 0.21315,
17
+ "mrr_at_1000": 0.21432,
18
+ "mrr_at_20": 0.20451,
19
+ "mrr_at_3": 0.17088,
20
+ "mrr_at_5": 0.183,
21
+ "ndcg_at_1": 0.12374,
22
+ "ndcg_at_10": 0.24115,
23
+ "ndcg_at_100": 0.33332,
24
+ "ndcg_at_1000": 0.35543,
25
+ "ndcg_at_20": 0.26996,
26
+ "ndcg_at_3": 0.18663,
27
+ "ndcg_at_5": 0.20794,
28
+ "precision_at_1": 0.12374,
29
+ "precision_at_10": 0.03864,
30
+ "precision_at_100": 0.00843,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.025,
33
+ "precision_at_3": 0.07744,
34
+ "precision_at_5": 0.05657,
35
+ "recall_at_1": 0.12374,
36
+ "recall_at_10": 0.38636,
37
+ "recall_at_100": 0.84343,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.5,
40
+ "recall_at_3": 0.23232,
41
+ "recall_at_5": 0.28283
42
+ }
43
+ }
results/alfaneo/bertimbaulaw-base-portuguese-cased/PTT_LegalBenchCorporateLobbying.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchCorporateLobbying",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 1.08,
7
+ "map_at_1": 0.51176,
8
+ "map_at_10": 0.62755,
9
+ "map_at_100": 0.63245,
10
+ "map_at_1000": 0.63272,
11
+ "map_at_20": 0.63091,
12
+ "map_at_3": 0.60294,
13
+ "map_at_5": 0.62015,
14
+ "mrr_at_1": 0.51176,
15
+ "mrr_at_10": 0.62755,
16
+ "mrr_at_100": 0.63245,
17
+ "mrr_at_1000": 0.63272,
18
+ "mrr_at_20": 0.63091,
19
+ "mrr_at_3": 0.60294,
20
+ "mrr_at_5": 0.62015,
21
+ "ndcg_at_1": 0.51176,
22
+ "ndcg_at_10": 0.68121,
23
+ "ndcg_at_100": 0.70413,
24
+ "ndcg_at_1000": 0.7104,
25
+ "ndcg_at_20": 0.69273,
26
+ "ndcg_at_3": 0.63312,
27
+ "ndcg_at_5": 0.66376,
28
+ "precision_at_1": 0.51176,
29
+ "precision_at_10": 0.08471,
30
+ "precision_at_100": 0.00953,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.04456,
33
+ "precision_at_3": 0.2402,
34
+ "precision_at_5": 0.15882,
35
+ "recall_at_1": 0.51176,
36
+ "recall_at_10": 0.84706,
37
+ "recall_at_100": 0.95294,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.89118,
40
+ "recall_at_3": 0.72059,
41
+ "recall_at_5": 0.79412
42
+ }
43
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/FloresBitextMining.json ADDED
@@ -0,0 +1,193 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "80dc3040d19756742c9a18267ab30f54fb8e226b",
3
+ "dev": {
4
+ "arb_Arab-por_Latn": {
5
+ "accuracy": 0.00802407221664995,
6
+ "f1": 0.004212445713818159,
7
+ "main_score": 0.004212445713818159,
8
+ "precision": 0.0038191792771967385,
9
+ "recall": 0.00802407221664995
10
+ },
11
+ "deu_Latn-por_Latn": {
12
+ "accuracy": 0.0320962888665998,
13
+ "f1": 0.024167031204853535,
14
+ "main_score": 0.024167031204853535,
15
+ "precision": 0.022443423063328555,
16
+ "recall": 0.0320962888665998
17
+ },
18
+ "eng_Latn-por_Latn": {
19
+ "accuracy": 0.28986960882647944,
20
+ "f1": 0.24075078174897974,
21
+ "main_score": 0.24075078174897974,
22
+ "precision": 0.22870221641896415,
23
+ "recall": 0.28986960882647944
24
+ },
25
+ "evaluation_time": 14.92,
26
+ "fra_Latn-por_Latn": {
27
+ "accuracy": 0.1464393179538616,
28
+ "f1": 0.11743538476431901,
29
+ "main_score": 0.11743538476431901,
30
+ "precision": 0.11081915772467692,
31
+ "recall": 0.1464393179538616
32
+ },
33
+ "ita_Latn-por_Latn": {
34
+ "accuracy": 0.17051153460381144,
35
+ "f1": 0.13478698529047614,
36
+ "main_score": 0.13478698529047614,
37
+ "precision": 0.12733566158921636,
38
+ "recall": 0.17051153460381144
39
+ },
40
+ "jpn_Jpan-por_Latn": {
41
+ "accuracy": 0.00802407221664995,
42
+ "f1": 0.00526815464018377,
43
+ "main_score": 0.00526815464018377,
44
+ "precision": 0.004892666062980074,
45
+ "recall": 0.00802407221664995
46
+ },
47
+ "kor_Hang-por_Latn": {
48
+ "accuracy": 0.025075225677031094,
49
+ "f1": 0.01816506016119284,
50
+ "main_score": 0.01816506016119284,
51
+ "precision": 0.01735234719811439,
52
+ "recall": 0.025075225677031094
53
+ },
54
+ "pol_Latn-por_Latn": {
55
+ "accuracy": 0.0160481444332999,
56
+ "f1": 0.007968643290496916,
57
+ "main_score": 0.007968643290496916,
58
+ "precision": 0.006579144439035258,
59
+ "recall": 0.0160481444332999
60
+ },
61
+ "rus_Cyrl-por_Latn": {
62
+ "accuracy": 0.01805416248746239,
63
+ "f1": 0.012871252089001993,
64
+ "main_score": 0.012871252089001993,
65
+ "precision": 0.011790625913230086,
66
+ "recall": 0.01805416248746239
67
+ },
68
+ "spa_Latn-por_Latn": {
69
+ "accuracy": 0.6960882647943831,
70
+ "f1": 0.6394989143646881,
71
+ "main_score": 0.6394989143646881,
72
+ "precision": 0.6175866274035783,
73
+ "recall": 0.6960882647943831
74
+ },
75
+ "swe_Latn-por_Latn": {
76
+ "accuracy": 0.03610832497492478,
77
+ "f1": 0.026251764740916405,
78
+ "main_score": 0.026251764740916405,
79
+ "precision": 0.023632266709465753,
80
+ "recall": 0.03610832497492478
81
+ },
82
+ "zho_Hans-por_Latn": {
83
+ "accuracy": 0.014042126379137413,
84
+ "f1": 0.006885272447225891,
85
+ "main_score": 0.006885272447225891,
86
+ "precision": 0.00602623043200051,
87
+ "recall": 0.014042126379137413
88
+ },
89
+ "zho_Hant-por_Latn": {
90
+ "accuracy": 0.011033099297893681,
91
+ "f1": 0.007808053705525227,
92
+ "main_score": 0.007808053705525227,
93
+ "precision": 0.007298153907566529,
94
+ "recall": 0.011033099297893681
95
+ }
96
+ },
97
+ "devtest": {
98
+ "arb_Arab-por_Latn": {
99
+ "accuracy": 0.0029644268774703555,
100
+ "f1": 0.001487975201373604,
101
+ "main_score": 0.001487975201373604,
102
+ "precision": 0.001320412361602293,
103
+ "recall": 0.0029644268774703555
104
+ },
105
+ "deu_Latn-por_Latn": {
106
+ "accuracy": 0.029644268774703556,
107
+ "f1": 0.02020645655089603,
108
+ "main_score": 0.02020645655089603,
109
+ "precision": 0.01820616570383175,
110
+ "recall": 0.029644268774703556
111
+ },
112
+ "eng_Latn-por_Latn": {
113
+ "accuracy": 0.29150197628458496,
114
+ "f1": 0.23861316034275154,
115
+ "main_score": 0.23861316034275154,
116
+ "precision": 0.2245573380974679,
117
+ "recall": 0.29150197628458496
118
+ },
119
+ "evaluation_time": 14.48,
120
+ "fra_Latn-por_Latn": {
121
+ "accuracy": 0.1600790513833992,
122
+ "f1": 0.12362931998875243,
123
+ "main_score": 0.12362931998875243,
124
+ "precision": 0.11426986266426353,
125
+ "recall": 0.1600790513833992
126
+ },
127
+ "ita_Latn-por_Latn": {
128
+ "accuracy": 0.17292490118577075,
129
+ "f1": 0.1330541492769943,
130
+ "main_score": 0.1330541492769943,
131
+ "precision": 0.12501861476002218,
132
+ "recall": 0.17292490118577075
133
+ },
134
+ "jpn_Jpan-por_Latn": {
135
+ "accuracy": 0.008893280632411068,
136
+ "f1": 0.004484576740993713,
137
+ "main_score": 0.004484576740993713,
138
+ "precision": 0.003633923872296643,
139
+ "recall": 0.008893280632411068
140
+ },
141
+ "kor_Hang-por_Latn": {
142
+ "accuracy": 0.019762845849802372,
143
+ "f1": 0.011173934991914306,
144
+ "main_score": 0.011173934991914306,
145
+ "precision": 0.01005678306941476,
146
+ "recall": 0.019762845849802372
147
+ },
148
+ "pol_Latn-por_Latn": {
149
+ "accuracy": 0.008893280632411068,
150
+ "f1": 0.0034594115433908123,
151
+ "main_score": 0.0034594115433908123,
152
+ "precision": 0.0027562784220073193,
153
+ "recall": 0.008893280632411068
154
+ },
155
+ "rus_Cyrl-por_Latn": {
156
+ "accuracy": 0.010869565217391304,
157
+ "f1": 0.006987370655609849,
158
+ "main_score": 0.006987370655609849,
159
+ "precision": 0.006581893875237407,
160
+ "recall": 0.010869565217391304
161
+ },
162
+ "spa_Latn-por_Latn": {
163
+ "accuracy": 0.7272727272727273,
164
+ "f1": 0.6738879478009911,
165
+ "main_score": 0.6738879478009911,
166
+ "precision": 0.6530632411067193,
167
+ "recall": 0.7272727272727273
168
+ },
169
+ "swe_Latn-por_Latn": {
170
+ "accuracy": 0.03162055335968379,
171
+ "f1": 0.022479867721338934,
172
+ "main_score": 0.022479867721338934,
173
+ "precision": 0.02054007625373916,
174
+ "recall": 0.03162055335968379
175
+ },
176
+ "zho_Hans-por_Latn": {
177
+ "accuracy": 0.012845849802371542,
178
+ "f1": 0.007417256269267429,
179
+ "main_score": 0.007417256269267429,
180
+ "precision": 0.006569240534168946,
181
+ "recall": 0.012845849802371542
182
+ },
183
+ "zho_Hant-por_Latn": {
184
+ "accuracy": 0.007905138339920948,
185
+ "f1": 0.005272706359662881,
186
+ "main_score": 0.005272706359662881,
187
+ "precision": 0.0048102679545236416,
188
+ "recall": 0.007905138339920948
189
+ }
190
+ },
191
+ "mteb_dataset_name": "FloresBitextMining",
192
+ "mteb_version": "1.7.41"
193
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/MassiveIntentClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
3
+ "mteb_dataset_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 10.62,
7
+ "pt": {
8
+ "accuracy": 0.5568594485541359,
9
+ "accuracy_stderr": 0.009179556153328843,
10
+ "f1": 0.5363448911901963,
11
+ "f1_stderr": 0.0070186836964918255,
12
+ "main_score": 0.5568594485541359
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 125.72,
17
+ "pt": {
18
+ "accuracy": 0.5504181013280866,
19
+ "accuracy_stderr": 0.012993558435932426,
20
+ "f1": 0.5310879696675931,
21
+ "f1_stderr": 0.014503549350616414,
22
+ "main_score": 0.5504181013280866
23
+ }
24
+ }
25
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
+ "mteb_dataset_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 6.4,
7
+ "pt": {
8
+ "accuracy": 0.605850706119704,
9
+ "accuracy_stderr": 0.01762963196364774,
10
+ "f1": 0.6004388557438236,
11
+ "f1_stderr": 0.015912826773464703,
12
+ "main_score": 0.605850706119704
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 6.46,
17
+ "pt": {
18
+ "accuracy": 0.5982292179045745,
19
+ "accuracy_stderr": 0.017313754853920497,
20
+ "f1": 0.5978155432786366,
21
+ "f1_stderr": 0.013112666495972144,
22
+ "main_score": 0.5982292179045745
23
+ }
24
+ }
25
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/MintakaRetrieval.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "efa78cc2f74bbcd21eff2261f9e13aebe40b814e",
3
+ "mteb_dataset_name": "MintakaRetrieval",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 4.52,
7
+ "pt": {
8
+ "map_at_1": 0.0514,
9
+ "map_at_10": 0.08337,
10
+ "map_at_100": 0.09064,
11
+ "map_at_1000": 0.09217,
12
+ "map_at_20": 0.08698,
13
+ "map_at_3": 0.07271,
14
+ "map_at_5": 0.07809,
15
+ "mrr_at_1": 0.0514,
16
+ "mrr_at_10": 0.08337,
17
+ "mrr_at_100": 0.09064,
18
+ "mrr_at_1000": 0.09217,
19
+ "mrr_at_20": 0.08698,
20
+ "mrr_at_3": 0.07271,
21
+ "mrr_at_5": 0.07809,
22
+ "ndcg_at_1": 0.0514,
23
+ "ndcg_at_10": 0.10194,
24
+ "ndcg_at_100": 0.14331,
25
+ "ndcg_at_1000": 0.19886,
26
+ "ndcg_at_20": 0.11513,
27
+ "ndcg_at_3": 0.07958,
28
+ "ndcg_at_5": 0.08923,
29
+ "precision_at_1": 0.0514,
30
+ "precision_at_10": 0.01619,
31
+ "precision_at_100": 0.0037,
32
+ "precision_at_1000": 0.00084,
33
+ "precision_at_20": 0.01071,
34
+ "precision_at_3": 0.03314,
35
+ "precision_at_5": 0.02455,
36
+ "recall_at_1": 0.0514,
37
+ "recall_at_10": 0.16185,
38
+ "recall_at_100": 0.36958,
39
+ "recall_at_1000": 0.8407,
40
+ "recall_at_20": 0.2141,
41
+ "recall_at_3": 0.09941,
42
+ "recall_at_5": 0.12277
43
+ }
44
+ }
45
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/MultiHateClassification.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ef137ea2b7c719183f8f60edf536b50f56d1365b",
3
+ "mteb_dataset_name": "MultiHateClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 4.01,
7
+ "por": {
8
+ "accuracy": 0.5782999999999999,
9
+ "accuracy_stderr": 0.025530569911382705,
10
+ "ap": 0.33637568520089195,
11
+ "ap_stderr": 0.0170991306842142,
12
+ "f1": 0.5459394765807749,
13
+ "f1_stderr": 0.02213676770687564,
14
+ "main_score": 0.5782999999999999
15
+ }
16
+ }
17
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/MultiLongDocRetrieval.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d67138e705d963e346253a80e59676ddb418810a",
3
+ "dev": {
4
+ "evaluation_time": 26.68,
5
+ "pt": {
6
+ "map_at_1": 0.09,
7
+ "map_at_10": 0.13852,
8
+ "map_at_100": 0.14905,
9
+ "map_at_1000": 0.15054,
10
+ "map_at_20": 0.14436,
11
+ "map_at_3": 0.12083,
12
+ "map_at_5": 0.13183,
13
+ "mrr_at_1": 0.09,
14
+ "mrr_at_10": 0.13852,
15
+ "mrr_at_100": 0.14905,
16
+ "mrr_at_1000": 0.15054,
17
+ "mrr_at_20": 0.14436,
18
+ "mrr_at_3": 0.12083,
19
+ "mrr_at_5": 0.13183,
20
+ "ndcg_at_1": 0.09,
21
+ "ndcg_at_10": 0.16849,
22
+ "ndcg_at_100": 0.22331,
23
+ "ndcg_at_1000": 0.26515,
24
+ "ndcg_at_20": 0.18907,
25
+ "ndcg_at_3": 0.13208,
26
+ "ndcg_at_5": 0.1523,
27
+ "precision_at_1": 0.09,
28
+ "precision_at_10": 0.0265,
29
+ "precision_at_100": 0.0053,
30
+ "precision_at_1000": 0.00087,
31
+ "precision_at_20": 0.01725,
32
+ "precision_at_3": 0.055,
33
+ "precision_at_5": 0.043,
34
+ "recall_at_1": 0.09,
35
+ "recall_at_10": 0.265,
36
+ "recall_at_100": 0.53,
37
+ "recall_at_1000": 0.865,
38
+ "recall_at_20": 0.345,
39
+ "recall_at_3": 0.165,
40
+ "recall_at_5": 0.215
41
+ }
42
+ },
43
+ "mteb_dataset_name": "MultiLongDocRetrieval",
44
+ "mteb_version": "1.7.41",
45
+ "test": {
46
+ "evaluation_time": 26.36,
47
+ "pt": {
48
+ "map_at_1": 0.055,
49
+ "map_at_10": 0.11249,
50
+ "map_at_100": 0.12284,
51
+ "map_at_1000": 0.12421,
52
+ "map_at_20": 0.11782,
53
+ "map_at_3": 0.095,
54
+ "map_at_5": 0.10675,
55
+ "mrr_at_1": 0.055,
56
+ "mrr_at_10": 0.11249,
57
+ "mrr_at_100": 0.12284,
58
+ "mrr_at_1000": 0.12421,
59
+ "mrr_at_20": 0.11782,
60
+ "mrr_at_3": 0.095,
61
+ "mrr_at_5": 0.10675,
62
+ "ndcg_at_1": 0.055,
63
+ "ndcg_at_10": 0.14421,
64
+ "ndcg_at_100": 0.20442,
65
+ "ndcg_at_1000": 0.2451,
66
+ "ndcg_at_20": 0.16414,
67
+ "ndcg_at_3": 0.10905,
68
+ "ndcg_at_5": 0.12992,
69
+ "precision_at_1": 0.055,
70
+ "precision_at_10": 0.0245,
71
+ "precision_at_100": 0.0055,
72
+ "precision_at_1000": 0.00088,
73
+ "precision_at_20": 0.01625,
74
+ "precision_at_3": 0.05,
75
+ "precision_at_5": 0.04,
76
+ "recall_at_1": 0.055,
77
+ "recall_at_10": 0.245,
78
+ "recall_at_100": 0.55,
79
+ "recall_at_1000": 0.88,
80
+ "recall_at_20": 0.325,
81
+ "recall_at_3": 0.15,
82
+ "recall_at_5": 0.2
83
+ }
84
+ }
85
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/NTREXBitextMining.json ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "fd20d54141b6da952d5c68a2989472892489da0f",
3
+ "mteb_dataset_name": "NTREXBitextMining",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "arb_Arab-por_Latn": {
7
+ "accuracy": 0.005007511266900351,
8
+ "f1": 0.0019191616246095747,
9
+ "main_score": 0.0019191616246095747,
10
+ "precision": 0.001563586667030668,
11
+ "recall": 0.005007511266900351
12
+ },
13
+ "deu_Latn-por_Latn": {
14
+ "accuracy": 0.020530796194291438,
15
+ "f1": 0.011446898568100676,
16
+ "main_score": 0.011446898568100676,
17
+ "precision": 0.010334852754702958,
18
+ "recall": 0.020530796194291438
19
+ },
20
+ "eng_Latn-por_Latn": {
21
+ "accuracy": 0.1412118177265899,
22
+ "f1": 0.10456522893829873,
23
+ "main_score": 0.10456522893829873,
24
+ "precision": 0.09566822290719214,
25
+ "recall": 0.1412118177265899
26
+ },
27
+ "evaluation_time": 33.47,
28
+ "fra_Latn-por_Latn": {
29
+ "accuracy": 0.07461191787681522,
30
+ "f1": 0.0577583400507415,
31
+ "main_score": 0.0577583400507415,
32
+ "precision": 0.053824309178236364,
33
+ "recall": 0.07461191787681522
34
+ },
35
+ "ita_Latn-por_Latn": {
36
+ "accuracy": 0.12518778167250877,
37
+ "f1": 0.09776332045102043,
38
+ "main_score": 0.09776332045102043,
39
+ "precision": 0.09190488111155093,
40
+ "recall": 0.12518778167250877
41
+ },
42
+ "jpn_Jpan-por_Latn": {
43
+ "accuracy": 0.006509764646970456,
44
+ "f1": 0.005262960389393178,
45
+ "main_score": 0.005262960389393178,
46
+ "precision": 0.005154113954815006,
47
+ "recall": 0.006509764646970456
48
+ },
49
+ "kor_Hang-por_Latn": {
50
+ "accuracy": 0.014521782674011016,
51
+ "f1": 0.010663466716949926,
52
+ "main_score": 0.010663466716949926,
53
+ "precision": 0.009860438160143816,
54
+ "recall": 0.014521782674011016
55
+ },
56
+ "pol_Latn-por_Latn": {
57
+ "accuracy": 0.013520280420630946,
58
+ "f1": 0.008695620212577918,
59
+ "main_score": 0.008695620212577918,
60
+ "precision": 0.007871128179712195,
61
+ "recall": 0.013520280420630946
62
+ },
63
+ "rus_Cyrl-por_Latn": {
64
+ "accuracy": 0.013520280420630946,
65
+ "f1": 0.009939388742527222,
66
+ "main_score": 0.009939388742527222,
67
+ "precision": 0.009189461848450332,
68
+ "recall": 0.013520280420630946
69
+ },
70
+ "spa_Latn-por_Latn": {
71
+ "accuracy": 0.5453179769654481,
72
+ "f1": 0.4834225587171182,
73
+ "main_score": 0.4834225587171182,
74
+ "precision": 0.46125447817398857,
75
+ "recall": 0.5453179769654481
76
+ },
77
+ "swe_Latn-por_Latn": {
78
+ "accuracy": 0.02503755633450175,
79
+ "f1": 0.013948491303838344,
80
+ "main_score": 0.013948491303838344,
81
+ "precision": 0.012607973003146759,
82
+ "recall": 0.02503755633450175
83
+ },
84
+ "zho_Hant-por_Latn": {
85
+ "accuracy": 0.023535302954431646,
86
+ "f1": 0.013612183771076893,
87
+ "main_score": 0.013612183771076893,
88
+ "precision": 0.012073190680565902,
89
+ "recall": 0.023535302954431646
90
+ }
91
+ }
92
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/PTT_AILACasedocs.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_AILACasedocs",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.6,
7
+ "map_at_1": 0.01874,
8
+ "map_at_10": 0.0777,
9
+ "map_at_100": 0.11069,
10
+ "map_at_1000": 0.11698,
11
+ "map_at_20": 0.08688,
12
+ "map_at_3": 0.05062,
13
+ "map_at_5": 0.05965,
14
+ "mrr_at_1": 0.1,
15
+ "mrr_at_10": 0.22755,
16
+ "mrr_at_100": 0.23927,
17
+ "mrr_at_1000": 0.24003,
18
+ "mrr_at_20": 0.23467,
19
+ "mrr_at_3": 0.19333,
20
+ "mrr_at_5": 0.20333,
21
+ "ndcg_at_1": 0.1,
22
+ "ndcg_at_10": 0.15269,
23
+ "ndcg_at_100": 0.29301,
24
+ "ndcg_at_1000": 0.34561,
25
+ "ndcg_at_20": 0.17763,
26
+ "ndcg_at_3": 0.12411,
27
+ "ndcg_at_5": 0.12081,
28
+ "precision_at_1": 0.1,
29
+ "precision_at_10": 0.078,
30
+ "precision_at_100": 0.0286,
31
+ "precision_at_1000": 0.0039,
32
+ "precision_at_20": 0.055,
33
+ "precision_at_3": 0.12,
34
+ "precision_at_5": 0.092,
35
+ "recall_at_1": 0.01874,
36
+ "recall_at_10": 0.21169,
37
+ "recall_at_100": 0.74165,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.28188,
40
+ "recall_at_3": 0.08828,
41
+ "recall_at_5": 0.11482
42
+ }
43
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/PTT_AILAStatutes.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_AILAStatutes",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.39,
7
+ "map_at_1": 0.045,
8
+ "map_at_10": 0.14775,
9
+ "map_at_100": 0.22563,
10
+ "map_at_1000": 0.22563,
11
+ "map_at_20": 0.17742,
12
+ "map_at_3": 0.094,
13
+ "map_at_5": 0.11625,
14
+ "mrr_at_1": 0.18,
15
+ "mrr_at_10": 0.35102,
16
+ "mrr_at_100": 0.36333,
17
+ "mrr_at_1000": 0.36333,
18
+ "mrr_at_20": 0.35882,
19
+ "mrr_at_3": 0.29333,
20
+ "mrr_at_5": 0.32833,
21
+ "ndcg_at_1": 0.18,
22
+ "ndcg_at_10": 0.25501,
23
+ "ndcg_at_100": 0.48256,
24
+ "ndcg_at_1000": 0.48256,
25
+ "ndcg_at_20": 0.32943,
26
+ "ndcg_at_3": 0.18982,
27
+ "ndcg_at_5": 0.19255,
28
+ "precision_at_1": 0.18,
29
+ "precision_at_10": 0.138,
30
+ "precision_at_100": 0.0434,
31
+ "precision_at_1000": 0.00434,
32
+ "precision_at_20": 0.11,
33
+ "precision_at_3": 0.18,
34
+ "precision_at_5": 0.168,
35
+ "recall_at_1": 0.045,
36
+ "recall_at_10": 0.31667,
37
+ "recall_at_100": 1.0,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.503,
40
+ "recall_at_3": 0.134,
41
+ "recall_at_5": 0.197
42
+ }
43
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/PTT_LegalBenchConsumerContractsQA.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchConsumerContractsQA",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.71,
7
+ "map_at_1": 0.20707,
8
+ "map_at_10": 0.3102,
9
+ "map_at_100": 0.32336,
10
+ "map_at_1000": 0.32409,
11
+ "map_at_20": 0.31726,
12
+ "map_at_3": 0.2782,
13
+ "map_at_5": 0.29689,
14
+ "mrr_at_1": 0.20707,
15
+ "mrr_at_10": 0.3102,
16
+ "mrr_at_100": 0.32336,
17
+ "mrr_at_1000": 0.32409,
18
+ "mrr_at_20": 0.31726,
19
+ "mrr_at_3": 0.2782,
20
+ "mrr_at_5": 0.29689,
21
+ "ndcg_at_1": 0.20707,
22
+ "ndcg_at_10": 0.36806,
23
+ "ndcg_at_100": 0.43963,
24
+ "ndcg_at_1000": 0.45322,
25
+ "ndcg_at_20": 0.3945,
26
+ "ndcg_at_3": 0.30204,
27
+ "ndcg_at_5": 0.33604,
28
+ "precision_at_1": 0.20707,
29
+ "precision_at_10": 0.0553,
30
+ "precision_at_100": 0.00904,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.03295,
33
+ "precision_at_3": 0.12374,
34
+ "precision_at_5": 0.09091,
35
+ "recall_at_1": 0.20707,
36
+ "recall_at_10": 0.55303,
37
+ "recall_at_100": 0.90404,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.65909,
40
+ "recall_at_3": 0.37121,
41
+ "recall_at_5": 0.45455
42
+ }
43
+ }
results/alfaneo/bertimbaulaw-base-portuguese-sts/PTT_LegalBenchCorporateLobbying.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_LegalBenchCorporateLobbying",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 0.92,
7
+ "map_at_1": 0.53235,
8
+ "map_at_10": 0.65532,
9
+ "map_at_100": 0.65958,
10
+ "map_at_1000": 0.65974,
11
+ "map_at_20": 0.65837,
12
+ "map_at_3": 0.62745,
13
+ "map_at_5": 0.64363,
14
+ "mrr_at_1": 0.53235,
15
+ "mrr_at_10": 0.65532,
16
+ "mrr_at_100": 0.65958,
17
+ "mrr_at_1000": 0.65974,
18
+ "mrr_at_20": 0.65837,
19
+ "mrr_at_3": 0.62745,
20
+ "mrr_at_5": 0.64363,
21
+ "ndcg_at_1": 0.53235,
22
+ "ndcg_at_10": 0.7124,
23
+ "ndcg_at_100": 0.73138,
24
+ "ndcg_at_1000": 0.7346,
25
+ "ndcg_at_20": 0.72305,
26
+ "ndcg_at_3": 0.65539,
27
+ "ndcg_at_5": 0.6845,
28
+ "precision_at_1": 0.53235,
29
+ "precision_at_10": 0.08912,
30
+ "precision_at_100": 0.00976,
31
+ "precision_at_1000": 0.001,
32
+ "precision_at_20": 0.04662,
33
+ "precision_at_3": 0.2451,
34
+ "precision_at_5": 0.16118,
35
+ "recall_at_1": 0.53235,
36
+ "recall_at_10": 0.89118,
37
+ "recall_at_100": 0.97647,
38
+ "recall_at_1000": 1.0,
39
+ "recall_at_20": 0.93235,
40
+ "recall_at_3": 0.73529,
41
+ "recall_at_5": 0.80588
42
+ }
43
+ }
results/alfaneo/jurisbert-base-portuguese-sts/FloresBitextMining.json ADDED
@@ -0,0 +1,193 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "80dc3040d19756742c9a18267ab30f54fb8e226b",
3
+ "dev": {
4
+ "arb_Arab-por_Latn": {
5
+ "accuracy": 0.0010030090270812437,
6
+ "f1": 2.545708190561532e-06,
7
+ "main_score": 2.545708190561532e-06,
8
+ "precision": 1.274471444830043e-06,
9
+ "recall": 0.0010030090270812437
10
+ },
11
+ "deu_Latn-por_Latn": {
12
+ "accuracy": 0.012036108324974924,
13
+ "f1": 0.006762116525388376,
14
+ "main_score": 0.006762116525388376,
15
+ "precision": 0.006240284366697061,
16
+ "recall": 0.012036108324974924
17
+ },
18
+ "eng_Latn-por_Latn": {
19
+ "accuracy": 0.07422266800401203,
20
+ "f1": 0.05581339445298409,
21
+ "main_score": 0.05581339445298409,
22
+ "precision": 0.05143819760174456,
23
+ "recall": 0.07422266800401203
24
+ },
25
+ "evaluation_time": 19.5,
26
+ "fra_Latn-por_Latn": {
27
+ "accuracy": 0.05917753259779338,
28
+ "f1": 0.039680664424223036,
29
+ "main_score": 0.039680664424223036,
30
+ "precision": 0.036039091881736096,
31
+ "recall": 0.05917753259779338
32
+ },
33
+ "ita_Latn-por_Latn": {
34
+ "accuracy": 0.05315947843530592,
35
+ "f1": 0.03818445627660448,
36
+ "main_score": 0.03818445627660448,
37
+ "precision": 0.03642934426296591,
38
+ "recall": 0.05315947843530592
39
+ },
40
+ "jpn_Jpan-por_Latn": {
41
+ "accuracy": 0.006018054162487462,
42
+ "f1": 0.000686484405633528,
43
+ "main_score": 0.000686484405633528,
44
+ "precision": 0.00039242979672353374,
45
+ "recall": 0.006018054162487462
46
+ },
47
+ "kor_Hang-por_Latn": {
48
+ "accuracy": 0.007021063189568706,
49
+ "f1": 0.0021007428529578823,
50
+ "main_score": 0.0021007428529578823,
51
+ "precision": 0.001736977709991689,
52
+ "recall": 0.007021063189568706
53
+ },
54
+ "pol_Latn-por_Latn": {
55
+ "accuracy": 0.0050150451354062184,
56
+ "f1": 0.002684143700944102,
57
+ "main_score": 0.002684143700944102,
58
+ "precision": 0.002512260988025725,
59
+ "recall": 0.0050150451354062184
60
+ },
61
+ "rus_Cyrl-por_Latn": {
62
+ "accuracy": 0.0020060180541624875,
63
+ "f1": 3.696331681758791e-05,
64
+ "main_score": 3.696331681758791e-05,
65
+ "precision": 1.8786458645462513e-05,
66
+ "recall": 0.0020060180541624875
67
+ },
68
+ "spa_Latn-por_Latn": {
69
+ "accuracy": 0.2858575727181545,
70
+ "f1": 0.24169809581914078,
71
+ "main_score": 0.24169809581914078,
72
+ "precision": 0.23126978769068057,
73
+ "recall": 0.2858575727181545
74
+ },
75
+ "swe_Latn-por_Latn": {
76
+ "accuracy": 0.009027081243731194,
77
+ "f1": 0.0037790444282874425,
78
+ "main_score": 0.0037790444282874425,
79
+ "precision": 0.003451916888830422,
80
+ "recall": 0.009027081243731194
81
+ },
82
+ "zho_Hans-por_Latn": {
83
+ "accuracy": 0.004012036108324975,
84
+ "f1": 0.001158381321489241,
85
+ "main_score": 0.001158381321489241,
86
+ "precision": 0.0010843859913354879,
87
+ "recall": 0.004012036108324975
88
+ },
89
+ "zho_Hant-por_Latn": {
90
+ "accuracy": 0.003009027081243731,
91
+ "f1": 0.0013406720992163888,
92
+ "main_score": 0.0013406720992163888,
93
+ "precision": 0.0012052769604494547,
94
+ "recall": 0.003009027081243731
95
+ }
96
+ },
97
+ "devtest": {
98
+ "arb_Arab-por_Latn": {
99
+ "accuracy": 0.001976284584980237,
100
+ "f1": 0.0007822793148880105,
101
+ "main_score": 0.0007822793148880105,
102
+ "precision": 0.0005599472990777338,
103
+ "recall": 0.001976284584980237
104
+ },
105
+ "deu_Latn-por_Latn": {
106
+ "accuracy": 0.004940711462450593,
107
+ "f1": 0.00329586627140975,
108
+ "main_score": 0.00329586627140975,
109
+ "precision": 0.0029654572656794176,
110
+ "recall": 0.004940711462450593
111
+ },
112
+ "eng_Latn-por_Latn": {
113
+ "accuracy": 0.058300395256917,
114
+ "f1": 0.046148037981534606,
115
+ "main_score": 0.046148037981534606,
116
+ "precision": 0.04334043552006938,
117
+ "recall": 0.058300395256917
118
+ },
119
+ "evaluation_time": 15.32,
120
+ "fra_Latn-por_Latn": {
121
+ "accuracy": 0.05731225296442688,
122
+ "f1": 0.03978759218383724,
123
+ "main_score": 0.03978759218383724,
124
+ "precision": 0.03742564080104413,
125
+ "recall": 0.05731225296442688
126
+ },
127
+ "ita_Latn-por_Latn": {
128
+ "accuracy": 0.0533596837944664,
129
+ "f1": 0.0385258773621945,
130
+ "main_score": 0.0385258773621945,
131
+ "precision": 0.035981218460363784,
132
+ "recall": 0.0533596837944664
133
+ },
134
+ "jpn_Jpan-por_Latn": {
135
+ "accuracy": 0.0009881422924901185,
136
+ "f1": 0.0009881422924901185,
137
+ "main_score": 0.0009881422924901185,
138
+ "precision": 0.0009881422924901185,
139
+ "recall": 0.0009881422924901185
140
+ },
141
+ "kor_Hang-por_Latn": {
142
+ "accuracy": 0.004940711462450593,
143
+ "f1": 0.00203595984488433,
144
+ "main_score": 0.00203595984488433,
145
+ "precision": 0.001697648756472286,
146
+ "recall": 0.004940711462450593
147
+ },
148
+ "pol_Latn-por_Latn": {
149
+ "accuracy": 0.003952569169960474,
150
+ "f1": 0.0024724165076434198,
151
+ "main_score": 0.0024724165076434198,
152
+ "precision": 0.0023066968129144766,
153
+ "recall": 0.003952569169960474
154
+ },
155
+ "rus_Cyrl-por_Latn": {
156
+ "accuracy": 0.001976284584980237,
157
+ "f1": 1.496591170504214e-05,
158
+ "main_score": 1.496591170504214e-05,
159
+ "precision": 7.52352913770643e-06,
160
+ "recall": 0.001976284584980237
161
+ },
162
+ "spa_Latn-por_Latn": {
163
+ "accuracy": 0.29347826086956524,
164
+ "f1": 0.25703547396694676,
165
+ "main_score": 0.25703547396694676,
166
+ "precision": 0.24758520597726577,
167
+ "recall": 0.29347826086956524
168
+ },
169
+ "swe_Latn-por_Latn": {
170
+ "accuracy": 0.010869565217391304,
171
+ "f1": 0.006149167843069215,
172
+ "main_score": 0.006149167843069215,
173
+ "precision": 0.005480290969714964,
174
+ "recall": 0.010869565217391304
175
+ },
176
+ "zho_Hans-por_Latn": {
177
+ "accuracy": 0.003952569169960474,
178
+ "f1": 0.0010623746815601504,
179
+ "main_score": 0.0010623746815601504,
180
+ "precision": 0.0010259332182543962,
181
+ "recall": 0.003952569169960474
182
+ },
183
+ "zho_Hant-por_Latn": {
184
+ "accuracy": 0.0029644268774703555,
185
+ "f1": 0.0010837318149571114,
186
+ "main_score": 0.0010837318149571114,
187
+ "precision": 0.001037126090343056,
188
+ "recall": 0.0029644268774703555
189
+ }
190
+ },
191
+ "mteb_dataset_name": "FloresBitextMining",
192
+ "mteb_version": "1.7.41"
193
+ }
results/alfaneo/jurisbert-base-portuguese-sts/MassiveIntentClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
3
+ "mteb_dataset_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 10.93,
7
+ "pt": {
8
+ "accuracy": 0.4603564223268326,
9
+ "accuracy_stderr": 0.013225643904114293,
10
+ "f1": 0.4433015064975333,
11
+ "f1_stderr": 0.009241924462843686,
12
+ "main_score": 0.4603564223268326
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 125.53,
17
+ "pt": {
18
+ "accuracy": 0.4533694048204623,
19
+ "accuracy_stderr": 0.014864419406043938,
20
+ "f1": 0.44013228491597634,
21
+ "f1_stderr": 0.012570691392373753,
22
+ "main_score": 0.4533694048204623
23
+ }
24
+ }
25
+ }
results/alfaneo/jurisbert-base-portuguese-sts/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
+ "mteb_dataset_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.7.41",
5
+ "test": {
6
+ "evaluation_time": 7.27,
7
+ "pt": {
8
+ "accuracy": 0.4658372562205784,
9
+ "accuracy_stderr": 0.014830915359372508,
10
+ "f1": 0.4542453356402298,
11
+ "f1_stderr": 0.011845114475951761,
12
+ "main_score": 0.4658372562205784
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 6.35,
17
+ "pt": {
18
+ "accuracy": 0.45268076733890805,
19
+ "accuracy_stderr": 0.01885404586342209,
20
+ "f1": 0.44739289046316505,
21
+ "f1_stderr": 0.015035843367392098,
22
+ "main_score": 0.45268076733890805
23
+ }
24
+ }
25
+ }