Mira commited on
Commit
da2f4ee
·
unverified ·
1 Parent(s): ea4c24a

add for `Mira190/Euler-Legal-Embedding-V1` (#342)

Browse files

* add all results

* move to revison folder

* modified folder name

This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/AILACasedocs.json +158 -0
  2. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/AILAStatutes.json +158 -0
  3. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/AlloprofReranking.json +26 -0
  4. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/AmazonCounterfactualClassification.json +347 -0
  5. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/ArXivHierarchicalClusteringP2P.json +46 -0
  6. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/ArXivHierarchicalClusteringS2S.json +46 -0
  7. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/ArguAna.json +158 -0
  8. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/ArmenianParaphrasePC.json +58 -0
  9. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BUCC.v2.json +59 -0
  10. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BelebeleRetrieval.json +0 -0
  11. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BigPatentClustering.v2.json +34 -0
  12. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BiorxivClusteringP2P.v2.json +34 -0
  13. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BornholmBitextMining.json +22 -0
  14. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BrazilianToxicTweetsClassification.json +73 -0
  15. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BulgarianStoreReviewSentimentClassfication.json +73 -0
  16. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CEDRClassification.json +73 -0
  17. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CLSClusteringP2P.v2.json +34 -0
  18. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CSFDSKMovieReviewSentimentClassification.json +73 -0
  19. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CTKFactsNLI.json +58 -0
  20. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CataloniaTweetClassification.json +135 -0
  21. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/Core17InstructionRetrieval.json +137 -0
  22. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CyrillicTurkicLangClassification.json +81 -0
  23. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CzechProductReviewSentimentClassification.json +73 -0
  24. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/DBpediaClassification.json +73 -0
  25. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/DiaBlaBitextMining.json +35 -0
  26. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/EstonianValenceClassification.json +73 -0
  27. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/FilipinoShopeeReviewsClassification.json +73 -0
  28. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/GerDaLIRSmall.json +158 -0
  29. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/GermanSTSBenchmark.json +26 -0
  30. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/GreekLegalCodeClassification.json +73 -0
  31. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/GujaratiNewsClassification.json +73 -0
  32. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/HALClusteringS2S.v2.json +34 -0
  33. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/IN22GenBitextMining.json +0 -0
  34. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/IndicCrosslingualSTS.json +203 -0
  35. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/IndicGenBenchFloresBitextMining.json +707 -0
  36. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/IndicLangClassification.json +96 -0
  37. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/ItaCaseholdClassification.json +73 -0
  38. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/KorHateSpeechMLClassification.json +73 -0
  39. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/KorSarcasmClassification.json +95 -0
  40. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/KurdishSentimentClassification.json +95 -0
  41. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/LeCaRDv2.json +158 -0
  42. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/LegalBenchConsumerContractsQA.json +158 -0
  43. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/LegalBenchCorporateLobbying.json +158 -0
  44. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/LegalQuAD.json +158 -0
  45. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/LegalSummarization.json +158 -0
  46. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/MLQARetrieval.json +0 -0
  47. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/MacedonianTweetSentimentClassification.json +73 -0
  48. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/MalteseNewsClassification.json +73 -0
  49. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/MasakhaNEWSClassification.json +1003 -0
  50. results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/MassiveIntentClassification.json +3173 -0
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/AILACasedocs.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "4106e6bcc72e0698d714ea8b101355e3e238431a",
3
+ "task_name": "AILACasedocs",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.46,
9
+ "ndcg_at_3": 0.40688,
10
+ "ndcg_at_5": 0.43187,
11
+ "ndcg_at_10": 0.4561,
12
+ "ndcg_at_20": 0.49447,
13
+ "ndcg_at_100": 0.57662,
14
+ "ndcg_at_1000": 0.58837,
15
+ "map_at_1": 0.15317,
16
+ "map_at_3": 0.2572,
17
+ "map_at_5": 0.30968,
18
+ "map_at_10": 0.34542,
19
+ "map_at_20": 0.37002,
20
+ "map_at_100": 0.39287,
21
+ "map_at_1000": 0.39473,
22
+ "recall_at_1": 0.15317,
23
+ "recall_at_3": 0.28306,
24
+ "recall_at_5": 0.40362,
25
+ "recall_at_10": 0.52916,
26
+ "recall_at_20": 0.65004,
27
+ "recall_at_100": 0.94488,
28
+ "recall_at_1000": 1.0,
29
+ "precision_at_1": 0.46,
30
+ "precision_at_3": 0.32,
31
+ "precision_at_5": 0.284,
32
+ "precision_at_10": 0.19,
33
+ "precision_at_20": 0.118,
34
+ "precision_at_100": 0.0364,
35
+ "precision_at_1000": 0.0039,
36
+ "mrr_at_1": 0.46,
37
+ "mrr_at_3": 0.5,
38
+ "mrr_at_5": 0.528,
39
+ "mrr_at_10": 0.549246,
40
+ "mrr_at_20": 0.552064,
41
+ "mrr_at_100": 0.5557,
42
+ "mrr_at_1000": 0.555854,
43
+ "nauc_ndcg_at_1_max": 0.422959,
44
+ "nauc_ndcg_at_1_std": 0.015322,
45
+ "nauc_ndcg_at_1_diff1": 0.380095,
46
+ "nauc_ndcg_at_3_max": 0.344352,
47
+ "nauc_ndcg_at_3_std": -0.018685,
48
+ "nauc_ndcg_at_3_diff1": 0.271702,
49
+ "nauc_ndcg_at_5_max": 0.274762,
50
+ "nauc_ndcg_at_5_std": 0.048706,
51
+ "nauc_ndcg_at_5_diff1": 0.187327,
52
+ "nauc_ndcg_at_10_max": 0.265897,
53
+ "nauc_ndcg_at_10_std": 0.006697,
54
+ "nauc_ndcg_at_10_diff1": 0.21833,
55
+ "nauc_ndcg_at_20_max": 0.288313,
56
+ "nauc_ndcg_at_20_std": -0.011933,
57
+ "nauc_ndcg_at_20_diff1": 0.222793,
58
+ "nauc_ndcg_at_100_max": 0.287056,
59
+ "nauc_ndcg_at_100_std": 0.04258,
60
+ "nauc_ndcg_at_100_diff1": 0.193631,
61
+ "nauc_ndcg_at_1000_max": 0.289234,
62
+ "nauc_ndcg_at_1000_std": 0.007546,
63
+ "nauc_ndcg_at_1000_diff1": 0.218991,
64
+ "nauc_map_at_1_max": 0.318145,
65
+ "nauc_map_at_1_std": 0.063592,
66
+ "nauc_map_at_1_diff1": 0.316677,
67
+ "nauc_map_at_3_max": 0.265247,
68
+ "nauc_map_at_3_std": -0.022442,
69
+ "nauc_map_at_3_diff1": 0.294668,
70
+ "nauc_map_at_5_max": 0.24776,
71
+ "nauc_map_at_5_std": 0.004472,
72
+ "nauc_map_at_5_diff1": 0.238104,
73
+ "nauc_map_at_10_max": 0.261601,
74
+ "nauc_map_at_10_std": 0.009586,
75
+ "nauc_map_at_10_diff1": 0.245793,
76
+ "nauc_map_at_20_max": 0.283075,
77
+ "nauc_map_at_20_std": 0.002577,
78
+ "nauc_map_at_20_diff1": 0.243359,
79
+ "nauc_map_at_100_max": 0.285159,
80
+ "nauc_map_at_100_std": 0.016426,
81
+ "nauc_map_at_100_diff1": 0.238971,
82
+ "nauc_map_at_1000_max": 0.283136,
83
+ "nauc_map_at_1000_std": 0.012356,
84
+ "nauc_map_at_1000_diff1": 0.238141,
85
+ "nauc_recall_at_1_max": 0.318145,
86
+ "nauc_recall_at_1_std": 0.063592,
87
+ "nauc_recall_at_1_diff1": 0.316677,
88
+ "nauc_recall_at_3_max": 0.215717,
89
+ "nauc_recall_at_3_std": -0.038016,
90
+ "nauc_recall_at_3_diff1": 0.224594,
91
+ "nauc_recall_at_5_max": 0.118203,
92
+ "nauc_recall_at_5_std": 0.037922,
93
+ "nauc_recall_at_5_diff1": 0.065476,
94
+ "nauc_recall_at_10_max": 0.10279,
95
+ "nauc_recall_at_10_std": 0.0044,
96
+ "nauc_recall_at_10_diff1": 0.088561,
97
+ "nauc_recall_at_20_max": 0.121644,
98
+ "nauc_recall_at_20_std": -0.078601,
99
+ "nauc_recall_at_20_diff1": 0.081318,
100
+ "nauc_recall_at_100_max": -0.016471,
101
+ "nauc_recall_at_100_std": 0.756352,
102
+ "nauc_recall_at_100_diff1": -0.694251,
103
+ "nauc_recall_at_1000_max": NaN,
104
+ "nauc_recall_at_1000_std": NaN,
105
+ "nauc_recall_at_1000_diff1": NaN,
106
+ "nauc_precision_at_1_max": 0.422959,
107
+ "nauc_precision_at_1_std": 0.015322,
108
+ "nauc_precision_at_1_diff1": 0.380095,
109
+ "nauc_precision_at_3_max": 0.288745,
110
+ "nauc_precision_at_3_std": -0.045425,
111
+ "nauc_precision_at_3_diff1": 0.224114,
112
+ "nauc_precision_at_5_max": 0.2481,
113
+ "nauc_precision_at_5_std": 0.061771,
114
+ "nauc_precision_at_5_diff1": 0.100614,
115
+ "nauc_precision_at_10_max": 0.214658,
116
+ "nauc_precision_at_10_std": 0.029312,
117
+ "nauc_precision_at_10_diff1": 0.078245,
118
+ "nauc_precision_at_20_max": 0.274231,
119
+ "nauc_precision_at_20_std": 0.014762,
120
+ "nauc_precision_at_20_diff1": 0.089821,
121
+ "nauc_precision_at_100_max": 0.143387,
122
+ "nauc_precision_at_100_std": 0.047518,
123
+ "nauc_precision_at_100_diff1": -0.02353,
124
+ "nauc_precision_at_1000_max": 0.088673,
125
+ "nauc_precision_at_1000_std": -0.009857,
126
+ "nauc_precision_at_1000_diff1": -0.047083,
127
+ "nauc_mrr_at_1_max": 0.422959,
128
+ "nauc_mrr_at_1_std": 0.015322,
129
+ "nauc_mrr_at_1_diff1": 0.380095,
130
+ "nauc_mrr_at_3_max": 0.379501,
131
+ "nauc_mrr_at_3_std": -0.004308,
132
+ "nauc_mrr_at_3_diff1": 0.301086,
133
+ "nauc_mrr_at_5_max": 0.363862,
134
+ "nauc_mrr_at_5_std": 0.033015,
135
+ "nauc_mrr_at_5_diff1": 0.267587,
136
+ "nauc_mrr_at_10_max": 0.349231,
137
+ "nauc_mrr_at_10_std": 0.011555,
138
+ "nauc_mrr_at_10_diff1": 0.264636,
139
+ "nauc_mrr_at_20_max": 0.35169,
140
+ "nauc_mrr_at_20_std": 0.006435,
141
+ "nauc_mrr_at_20_diff1": 0.271316,
142
+ "nauc_mrr_at_100_max": 0.349672,
143
+ "nauc_mrr_at_100_std": 0.008858,
144
+ "nauc_mrr_at_100_diff1": 0.269838,
145
+ "nauc_mrr_at_1000_max": 0.350122,
146
+ "nauc_mrr_at_1000_std": 0.00837,
147
+ "nauc_mrr_at_1000_diff1": 0.270827,
148
+ "main_score": 0.4561,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "eng-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 60.59555220603943,
157
+ "kg_co2_emissions": null
158
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/AILAStatutes.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ebfcd844eadd3d667efa3c57fc5c8c87f5c2867e",
3
+ "task_name": "AILAStatutes",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.96,
9
+ "ndcg_at_3": 0.83289,
10
+ "ndcg_at_5": 0.80629,
11
+ "ndcg_at_10": 0.9003,
12
+ "ndcg_at_20": 0.91712,
13
+ "ndcg_at_100": 0.92007,
14
+ "ndcg_at_1000": 0.92007,
15
+ "map_at_1": 0.23167,
16
+ "map_at_3": 0.52661,
17
+ "map_at_5": 0.70358,
18
+ "map_at_10": 0.80515,
19
+ "map_at_20": 0.81847,
20
+ "map_at_100": 0.81982,
21
+ "map_at_1000": 0.81982,
22
+ "recall_at_1": 0.23167,
23
+ "recall_at_3": 0.55,
24
+ "recall_at_5": 0.77833,
25
+ "recall_at_10": 0.951,
26
+ "recall_at_20": 0.991,
27
+ "recall_at_100": 1.0,
28
+ "recall_at_1000": 1.0,
29
+ "precision_at_1": 0.96,
30
+ "precision_at_3": 0.78,
31
+ "precision_at_5": 0.664,
32
+ "precision_at_10": 0.412,
33
+ "precision_at_20": 0.215,
34
+ "precision_at_100": 0.0434,
35
+ "precision_at_1000": 0.00434,
36
+ "mrr_at_1": 0.96,
37
+ "mrr_at_3": 0.98,
38
+ "mrr_at_5": 0.98,
39
+ "mrr_at_10": 0.98,
40
+ "mrr_at_20": 0.98,
41
+ "mrr_at_100": 0.98,
42
+ "mrr_at_1000": 0.98,
43
+ "nauc_ndcg_at_1_max": 0.777077,
44
+ "nauc_ndcg_at_1_std": 0.561391,
45
+ "nauc_ndcg_at_1_diff1": 0.079365,
46
+ "nauc_ndcg_at_3_max": 0.069887,
47
+ "nauc_ndcg_at_3_std": 0.276397,
48
+ "nauc_ndcg_at_3_diff1": -0.303027,
49
+ "nauc_ndcg_at_5_max": 0.317162,
50
+ "nauc_ndcg_at_5_std": 0.245583,
51
+ "nauc_ndcg_at_5_diff1": -0.126626,
52
+ "nauc_ndcg_at_10_max": 0.38182,
53
+ "nauc_ndcg_at_10_std": 0.339945,
54
+ "nauc_ndcg_at_10_diff1": -0.286538,
55
+ "nauc_ndcg_at_20_max": 0.343719,
56
+ "nauc_ndcg_at_20_std": 0.37452,
57
+ "nauc_ndcg_at_20_diff1": -0.228703,
58
+ "nauc_ndcg_at_100_max": 0.325565,
59
+ "nauc_ndcg_at_100_std": 0.341978,
60
+ "nauc_ndcg_at_100_diff1": -0.226628,
61
+ "nauc_ndcg_at_1000_max": 0.325565,
62
+ "nauc_ndcg_at_1000_std": 0.341978,
63
+ "nauc_ndcg_at_1000_diff1": -0.226628,
64
+ "nauc_map_at_1_max": 0.134622,
65
+ "nauc_map_at_1_std": -0.111108,
66
+ "nauc_map_at_1_diff1": 0.14758,
67
+ "nauc_map_at_3_max": 0.091712,
68
+ "nauc_map_at_3_std": 0.130451,
69
+ "nauc_map_at_3_diff1": -0.161761,
70
+ "nauc_map_at_5_max": 0.297167,
71
+ "nauc_map_at_5_std": 0.245253,
72
+ "nauc_map_at_5_diff1": -0.130307,
73
+ "nauc_map_at_10_max": 0.328023,
74
+ "nauc_map_at_10_std": 0.317118,
75
+ "nauc_map_at_10_diff1": -0.252516,
76
+ "nauc_map_at_20_max": 0.303781,
77
+ "nauc_map_at_20_std": 0.326354,
78
+ "nauc_map_at_20_diff1": -0.228912,
79
+ "nauc_map_at_100_max": 0.29823,
80
+ "nauc_map_at_100_std": 0.317901,
81
+ "nauc_map_at_100_diff1": -0.228762,
82
+ "nauc_map_at_1000_max": 0.29823,
83
+ "nauc_map_at_1000_std": 0.317901,
84
+ "nauc_map_at_1000_diff1": -0.228762,
85
+ "nauc_recall_at_1_max": 0.134622,
86
+ "nauc_recall_at_1_std": -0.111108,
87
+ "nauc_recall_at_1_diff1": 0.14758,
88
+ "nauc_recall_at_3_max": -0.02119,
89
+ "nauc_recall_at_3_std": 0.077221,
90
+ "nauc_recall_at_3_diff1": -0.210049,
91
+ "nauc_recall_at_5_max": 0.301644,
92
+ "nauc_recall_at_5_std": 0.168616,
93
+ "nauc_recall_at_5_diff1": -0.061199,
94
+ "nauc_recall_at_10_max": 0.471509,
95
+ "nauc_recall_at_10_std": 0.305049,
96
+ "nauc_recall_at_10_diff1": -0.419322,
97
+ "nauc_recall_at_20_max": 0.585279,
98
+ "nauc_recall_at_20_std": 0.869281,
99
+ "nauc_recall_at_20_diff1": -0.258481,
100
+ "nauc_recall_at_100_max": NaN,
101
+ "nauc_recall_at_100_std": NaN,
102
+ "nauc_recall_at_100_diff1": NaN,
103
+ "nauc_recall_at_1000_max": NaN,
104
+ "nauc_recall_at_1000_std": NaN,
105
+ "nauc_recall_at_1000_diff1": NaN,
106
+ "nauc_precision_at_1_max": 0.777077,
107
+ "nauc_precision_at_1_std": 0.561391,
108
+ "nauc_precision_at_1_diff1": 0.079365,
109
+ "nauc_precision_at_3_max": 0.023417,
110
+ "nauc_precision_at_3_std": 0.342984,
111
+ "nauc_precision_at_3_diff1": -0.361417,
112
+ "nauc_precision_at_5_max": 0.270942,
113
+ "nauc_precision_at_5_std": 0.380535,
114
+ "nauc_precision_at_5_diff1": -0.230479,
115
+ "nauc_precision_at_10_max": 0.112217,
116
+ "nauc_precision_at_10_std": 0.317965,
117
+ "nauc_precision_at_10_diff1": -0.323634,
118
+ "nauc_precision_at_20_max": -0.013155,
119
+ "nauc_precision_at_20_std": 0.291903,
120
+ "nauc_precision_at_20_diff1": -0.231008,
121
+ "nauc_precision_at_100_max": -0.061762,
122
+ "nauc_precision_at_100_std": 0.235122,
123
+ "nauc_precision_at_100_diff1": -0.222718,
124
+ "nauc_precision_at_1000_max": -0.061762,
125
+ "nauc_precision_at_1000_std": 0.235122,
126
+ "nauc_precision_at_1000_diff1": -0.222718,
127
+ "nauc_mrr_at_1_max": 0.777077,
128
+ "nauc_mrr_at_1_std": 0.561391,
129
+ "nauc_mrr_at_1_diff1": 0.079365,
130
+ "nauc_mrr_at_3_max": 0.777077,
131
+ "nauc_mrr_at_3_std": 0.561391,
132
+ "nauc_mrr_at_3_diff1": 0.079365,
133
+ "nauc_mrr_at_5_max": 0.777077,
134
+ "nauc_mrr_at_5_std": 0.561391,
135
+ "nauc_mrr_at_5_diff1": 0.079365,
136
+ "nauc_mrr_at_10_max": 0.777077,
137
+ "nauc_mrr_at_10_std": 0.561391,
138
+ "nauc_mrr_at_10_diff1": 0.079365,
139
+ "nauc_mrr_at_20_max": 0.777077,
140
+ "nauc_mrr_at_20_std": 0.561391,
141
+ "nauc_mrr_at_20_diff1": 0.079365,
142
+ "nauc_mrr_at_100_max": 0.777077,
143
+ "nauc_mrr_at_100_std": 0.561391,
144
+ "nauc_mrr_at_100_diff1": 0.079365,
145
+ "nauc_mrr_at_1000_max": 0.777077,
146
+ "nauc_mrr_at_1000_std": 0.561391,
147
+ "nauc_mrr_at_1000_diff1": 0.079365,
148
+ "main_score": 0.9003,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "eng-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 117.04100131988525,
157
+ "kg_co2_emissions": null
158
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/AlloprofReranking.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "65393d0d7a08a10b4e348135e824f385d420b0fd",
3
+ "task_name": "AlloprofReranking",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "map": 0.832873,
9
+ "mrr": 0.842449,
10
+ "nAUC_map_max": 0.134644,
11
+ "nAUC_map_std": 0.087277,
12
+ "nAUC_map_diff1": 0.657688,
13
+ "nAUC_mrr_max": 0.142451,
14
+ "nAUC_mrr_std": 0.079582,
15
+ "nAUC_mrr_diff1": 0.66181,
16
+ "main_score": 0.832873,
17
+ "hf_subset": "default",
18
+ "languages": [
19
+ "fra-Latn"
20
+ ]
21
+ }
22
+ ]
23
+ },
24
+ "evaluation_time": 296.1569836139679,
25
+ "kg_co2_emissions": null
26
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/AmazonCounterfactualClassification.json ADDED
@@ -0,0 +1,347 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e8379541af4e31359cca9fbcf4b00f2671dba205",
3
+ "task_name": "AmazonCounterfactualClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.779385,
9
+ "f1": 0.639538,
10
+ "f1_weighted": 0.816716,
11
+ "ap": 0.241858,
12
+ "ap_weighted": 0.241858,
13
+ "scores_per_experiment": [
14
+ {
15
+ "accuracy": 0.841079,
16
+ "f1": 0.686972,
17
+ "f1_weighted": 0.860837,
18
+ "ap": 0.274726,
19
+ "ap_weighted": 0.274726
20
+ },
21
+ {
22
+ "accuracy": 0.778111,
23
+ "f1": 0.636851,
24
+ "f1_weighted": 0.816143,
25
+ "ap": 0.237245,
26
+ "ap_weighted": 0.237245
27
+ },
28
+ {
29
+ "accuracy": 0.774363,
30
+ "f1": 0.63893,
31
+ "f1_weighted": 0.813981,
32
+ "ap": 0.245232,
33
+ "ap_weighted": 0.245232
34
+ },
35
+ {
36
+ "accuracy": 0.730885,
37
+ "f1": 0.591626,
38
+ "f1_weighted": 0.780403,
39
+ "ap": 0.196825,
40
+ "ap_weighted": 0.196825
41
+ },
42
+ {
43
+ "accuracy": 0.716642,
44
+ "f1": 0.598327,
45
+ "f1_weighted": 0.770896,
46
+ "ap": 0.222774,
47
+ "ap_weighted": 0.222774
48
+ },
49
+ {
50
+ "accuracy": 0.747376,
51
+ "f1": 0.615629,
52
+ "f1_weighted": 0.793766,
53
+ "ap": 0.225928,
54
+ "ap_weighted": 0.225928
55
+ },
56
+ {
57
+ "accuracy": 0.790855,
58
+ "f1": 0.64152,
59
+ "f1_weighted": 0.824676,
60
+ "ap": 0.234151,
61
+ "ap_weighted": 0.234151
62
+ },
63
+ {
64
+ "accuracy": 0.829085,
65
+ "f1": 0.679101,
66
+ "f1_weighted": 0.852767,
67
+ "ap": 0.270674,
68
+ "ap_weighted": 0.270674
69
+ },
70
+ {
71
+ "accuracy": 0.756372,
72
+ "f1": 0.62732,
73
+ "f1_weighted": 0.800923,
74
+ "ap": 0.240349,
75
+ "ap_weighted": 0.240349
76
+ },
77
+ {
78
+ "accuracy": 0.829085,
79
+ "f1": 0.679101,
80
+ "f1_weighted": 0.852767,
81
+ "ap": 0.270674,
82
+ "ap_weighted": 0.270674
83
+ }
84
+ ],
85
+ "main_score": 0.779385,
86
+ "hf_subset": "en-ext",
87
+ "languages": [
88
+ "eng-Latn"
89
+ ]
90
+ },
91
+ {
92
+ "accuracy": 0.732985,
93
+ "f1": 0.66903,
94
+ "f1_weighted": 0.757265,
95
+ "ap": 0.352636,
96
+ "ap_weighted": 0.352636,
97
+ "scores_per_experiment": [
98
+ {
99
+ "accuracy": 0.686567,
100
+ "f1": 0.62362,
101
+ "f1_weighted": 0.717352,
102
+ "ap": 0.305859,
103
+ "ap_weighted": 0.305859
104
+ },
105
+ {
106
+ "accuracy": 0.734328,
107
+ "f1": 0.672035,
108
+ "f1_weighted": 0.759075,
109
+ "ap": 0.355321,
110
+ "ap_weighted": 0.355321
111
+ },
112
+ {
113
+ "accuracy": 0.671642,
114
+ "f1": 0.617536,
115
+ "f1_weighted": 0.705136,
116
+ "ap": 0.308447,
117
+ "ap_weighted": 0.308447
118
+ },
119
+ {
120
+ "accuracy": 0.725373,
121
+ "f1": 0.666248,
122
+ "f1_weighted": 0.751791,
123
+ "ap": 0.35231,
124
+ "ap_weighted": 0.35231
125
+ },
126
+ {
127
+ "accuracy": 0.737313,
128
+ "f1": 0.682683,
129
+ "f1_weighted": 0.76286,
130
+ "ap": 0.375471,
131
+ "ap_weighted": 0.375471
132
+ },
133
+ {
134
+ "accuracy": 0.738806,
135
+ "f1": 0.677043,
136
+ "f1_weighted": 0.763048,
137
+ "ap": 0.361322,
138
+ "ap_weighted": 0.361322
139
+ },
140
+ {
141
+ "accuracy": 0.792537,
142
+ "f1": 0.716491,
143
+ "f1_weighted": 0.805906,
144
+ "ap": 0.394723,
145
+ "ap_weighted": 0.394723
146
+ },
147
+ {
148
+ "accuracy": 0.792537,
149
+ "f1": 0.722274,
150
+ "f1_weighted": 0.80734,
151
+ "ap": 0.406313,
152
+ "ap_weighted": 0.406313
153
+ },
154
+ {
155
+ "accuracy": 0.747761,
156
+ "f1": 0.679615,
157
+ "f1_weighted": 0.769594,
158
+ "ap": 0.358113,
159
+ "ap_weighted": 0.358113
160
+ },
161
+ {
162
+ "accuracy": 0.702985,
163
+ "f1": 0.632751,
164
+ "f1_weighted": 0.730551,
165
+ "ap": 0.308482,
166
+ "ap_weighted": 0.308482
167
+ }
168
+ ],
169
+ "main_score": 0.732985,
170
+ "hf_subset": "en",
171
+ "languages": [
172
+ "eng-Latn"
173
+ ]
174
+ },
175
+ {
176
+ "accuracy": 0.735867,
177
+ "f1": 0.722828,
178
+ "f1_weighted": 0.745876,
179
+ "ap": 0.847409,
180
+ "ap_weighted": 0.847409,
181
+ "scores_per_experiment": [
182
+ {
183
+ "accuracy": 0.713062,
184
+ "f1": 0.706467,
185
+ "f1_weighted": 0.723708,
186
+ "ap": 0.853746,
187
+ "ap_weighted": 0.853746
188
+ },
189
+ {
190
+ "accuracy": 0.759101,
191
+ "f1": 0.740669,
192
+ "f1_weighted": 0.767761,
193
+ "ap": 0.845346,
194
+ "ap_weighted": 0.845346
195
+ },
196
+ {
197
+ "accuracy": 0.766595,
198
+ "f1": 0.750677,
199
+ "f1_weighted": 0.775364,
200
+ "ap": 0.856639,
201
+ "ap_weighted": 0.856639
202
+ },
203
+ {
204
+ "accuracy": 0.740899,
205
+ "f1": 0.728964,
206
+ "f1_weighted": 0.751252,
207
+ "ap": 0.852403,
208
+ "ap_weighted": 0.852403
209
+ },
210
+ {
211
+ "accuracy": 0.726981,
212
+ "f1": 0.720046,
213
+ "f1_weighted": 0.737312,
214
+ "ap": 0.862596,
215
+ "ap_weighted": 0.862596
216
+ },
217
+ {
218
+ "accuracy": 0.728051,
219
+ "f1": 0.718733,
220
+ "f1_weighted": 0.738794,
221
+ "ap": 0.85298,
222
+ "ap_weighted": 0.85298
223
+ },
224
+ {
225
+ "accuracy": 0.718415,
226
+ "f1": 0.696497,
227
+ "f1_weighted": 0.728458,
228
+ "ap": 0.813487,
229
+ "ap_weighted": 0.813487
230
+ },
231
+ {
232
+ "accuracy": 0.713062,
233
+ "f1": 0.703934,
234
+ "f1_weighted": 0.724305,
235
+ "ap": 0.843038,
236
+ "ap_weighted": 0.843038
237
+ },
238
+ {
239
+ "accuracy": 0.754818,
240
+ "f1": 0.736059,
241
+ "f1_weighted": 0.763633,
242
+ "ap": 0.841969,
243
+ "ap_weighted": 0.841969
244
+ },
245
+ {
246
+ "accuracy": 0.737687,
247
+ "f1": 0.726239,
248
+ "f1_weighted": 0.748177,
249
+ "ap": 0.851884,
250
+ "ap_weighted": 0.851884
251
+ }
252
+ ],
253
+ "main_score": 0.735867,
254
+ "hf_subset": "de",
255
+ "languages": [
256
+ "deu-Latn"
257
+ ]
258
+ },
259
+ {
260
+ "accuracy": 0.811135,
261
+ "f1": 0.673481,
262
+ "f1_weighted": 0.841256,
263
+ "ap": 0.279918,
264
+ "ap_weighted": 0.279918,
265
+ "scores_per_experiment": [
266
+ {
267
+ "accuracy": 0.756959,
268
+ "f1": 0.630081,
269
+ "f1_weighted": 0.80219,
270
+ "ap": 0.246226,
271
+ "ap_weighted": 0.246226
272
+ },
273
+ {
274
+ "accuracy": 0.835118,
275
+ "f1": 0.694226,
276
+ "f1_weighted": 0.859118,
277
+ "ap": 0.296735,
278
+ "ap_weighted": 0.296735
279
+ },
280
+ {
281
+ "accuracy": 0.799786,
282
+ "f1": 0.663038,
283
+ "f1_weighted": 0.833571,
284
+ "ap": 0.269013,
285
+ "ap_weighted": 0.269013
286
+ },
287
+ {
288
+ "accuracy": 0.846895,
289
+ "f1": 0.703955,
290
+ "f1_weighted": 0.867378,
291
+ "ap": 0.304599,
292
+ "ap_weighted": 0.304599
293
+ },
294
+ {
295
+ "accuracy": 0.892934,
296
+ "f1": 0.755725,
297
+ "f1_weighted": 0.901166,
298
+ "ap": 0.369893,
299
+ "ap_weighted": 0.369893
300
+ },
301
+ {
302
+ "accuracy": 0.864026,
303
+ "f1": 0.718991,
304
+ "f1_weighted": 0.879372,
305
+ "ap": 0.318402,
306
+ "ap_weighted": 0.318402
307
+ },
308
+ {
309
+ "accuracy": 0.79015,
310
+ "f1": 0.642702,
311
+ "f1_weighted": 0.825046,
312
+ "ap": 0.237285,
313
+ "ap_weighted": 0.237285
314
+ },
315
+ {
316
+ "accuracy": 0.739829,
317
+ "f1": 0.611358,
318
+ "f1_weighted": 0.788873,
319
+ "ap": 0.225527,
320
+ "ap_weighted": 0.225527
321
+ },
322
+ {
323
+ "accuracy": 0.739829,
324
+ "f1": 0.606997,
325
+ "f1_weighted": 0.788509,
326
+ "ap": 0.216916,
327
+ "ap_weighted": 0.216916
328
+ },
329
+ {
330
+ "accuracy": 0.845824,
331
+ "f1": 0.707739,
332
+ "f1_weighted": 0.867333,
333
+ "ap": 0.314584,
334
+ "ap_weighted": 0.314584
335
+ }
336
+ ],
337
+ "main_score": 0.811135,
338
+ "hf_subset": "ja",
339
+ "languages": [
340
+ "jpn-Jpan"
341
+ ]
342
+ }
343
+ ]
344
+ },
345
+ "evaluation_time": 716.3125834465027,
346
+ "kg_co2_emissions": null
347
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/ArXivHierarchicalClusteringP2P.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "0bbdb47bcbe3a90093699aefeed338a0f28a7ee8",
3
+ "task_name": "ArXivHierarchicalClusteringP2P",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.520411,
11
+ 0.547927,
12
+ 0.531985,
13
+ 0.507774,
14
+ 0.53944,
15
+ 0.589286,
16
+ 0.611598,
17
+ 0.588157,
18
+ 0.531446,
19
+ 0.571207
20
+ ],
21
+ "Level 1": [
22
+ 0.557943,
23
+ 0.612987,
24
+ 0.57228,
25
+ 0.578151,
26
+ 0.596098,
27
+ 0.578954,
28
+ 0.596939,
29
+ 0.583138,
30
+ 0.589699,
31
+ 0.59241
32
+ ]
33
+ },
34
+ "v_measure": 0.569891,
35
+ "v_measure_std": 0.029776,
36
+ "main_score": 0.569891,
37
+ "hf_subset": "default",
38
+ "languages": [
39
+ "eng-Latn"
40
+ ]
41
+ }
42
+ ]
43
+ },
44
+ "evaluation_time": 36.79362654685974,
45
+ "kg_co2_emissions": null
46
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/ArXivHierarchicalClusteringS2S.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "b73bd54100e5abfa6e3a23dcafb46fe4d2438dc3",
3
+ "task_name": "ArXivHierarchicalClusteringS2S",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.569424,
11
+ 0.586751,
12
+ 0.619322,
13
+ 0.582351,
14
+ 0.618507,
15
+ 0.631571,
16
+ 0.637502,
17
+ 0.61528,
18
+ 0.6074,
19
+ 0.665284
20
+ ],
21
+ "Level 1": [
22
+ 0.586,
23
+ 0.609661,
24
+ 0.614603,
25
+ 0.605424,
26
+ 0.609524,
27
+ 0.611621,
28
+ 0.598667,
29
+ 0.585651,
30
+ 0.609931,
31
+ 0.613299
32
+ ]
33
+ },
34
+ "v_measure": 0.608889,
35
+ "v_measure_std": 0.020919,
36
+ "main_score": 0.608889,
37
+ "hf_subset": "default",
38
+ "languages": [
39
+ "eng-Latn"
40
+ ]
41
+ }
42
+ ]
43
+ },
44
+ "evaluation_time": 35.75266218185425,
45
+ "kg_co2_emissions": null
46
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/ArguAna.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "c22ab2a51041ffd869aaddef7af8d8215647e41a",
3
+ "task_name": "ArguAna",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.50142,
9
+ "ndcg_at_3": 0.68188,
10
+ "ndcg_at_5": 0.72188,
11
+ "ndcg_at_10": 0.73971,
12
+ "ndcg_at_20": 0.7463,
13
+ "ndcg_at_100": 0.74773,
14
+ "ndcg_at_1000": 0.74773,
15
+ "map_at_1": 0.50142,
16
+ "map_at_3": 0.63691,
17
+ "map_at_5": 0.65921,
18
+ "map_at_10": 0.66663,
19
+ "map_at_20": 0.66851,
20
+ "map_at_100": 0.66875,
21
+ "map_at_1000": 0.66875,
22
+ "recall_at_1": 0.50142,
23
+ "recall_at_3": 0.81223,
24
+ "recall_at_5": 0.90896,
25
+ "recall_at_10": 0.96373,
26
+ "recall_at_20": 0.98933,
27
+ "recall_at_100": 0.99644,
28
+ "recall_at_1000": 0.99644,
29
+ "precision_at_1": 0.50142,
30
+ "precision_at_3": 0.27074,
31
+ "precision_at_5": 0.18179,
32
+ "precision_at_10": 0.09637,
33
+ "precision_at_20": 0.04947,
34
+ "precision_at_100": 0.00996,
35
+ "precision_at_1000": 0.001,
36
+ "mrr_at_1": 0.51138,
37
+ "mrr_at_3": 0.640351,
38
+ "mrr_at_5": 0.66215,
39
+ "mrr_at_10": 0.670048,
40
+ "mrr_at_20": 0.671924,
41
+ "mrr_at_100": 0.672163,
42
+ "mrr_at_1000": 0.672163,
43
+ "nauc_ndcg_at_1_max": -0.165929,
44
+ "nauc_ndcg_at_1_std": -0.245991,
45
+ "nauc_ndcg_at_1_diff1": 0.158499,
46
+ "nauc_ndcg_at_3_max": -0.075397,
47
+ "nauc_ndcg_at_3_std": -0.235576,
48
+ "nauc_ndcg_at_3_diff1": 0.169231,
49
+ "nauc_ndcg_at_5_max": -0.079406,
50
+ "nauc_ndcg_at_5_std": -0.242823,
51
+ "nauc_ndcg_at_5_diff1": 0.167897,
52
+ "nauc_ndcg_at_10_max": -0.086863,
53
+ "nauc_ndcg_at_10_std": -0.237635,
54
+ "nauc_ndcg_at_10_diff1": 0.17203,
55
+ "nauc_ndcg_at_20_max": -0.101264,
56
+ "nauc_ndcg_at_20_std": -0.237995,
57
+ "nauc_ndcg_at_20_diff1": 0.164356,
58
+ "nauc_ndcg_at_100_max": -0.105913,
59
+ "nauc_ndcg_at_100_std": -0.23766,
60
+ "nauc_ndcg_at_100_diff1": 0.163234,
61
+ "nauc_ndcg_at_1000_max": -0.105913,
62
+ "nauc_ndcg_at_1000_std": -0.23766,
63
+ "nauc_ndcg_at_1000_diff1": 0.163234,
64
+ "nauc_map_at_1_max": -0.165929,
65
+ "nauc_map_at_1_std": -0.245991,
66
+ "nauc_map_at_1_diff1": 0.158499,
67
+ "nauc_map_at_3_max": -0.103143,
68
+ "nauc_map_at_3_std": -0.23913,
69
+ "nauc_map_at_3_diff1": 0.164053,
70
+ "nauc_map_at_5_max": -0.107283,
71
+ "nauc_map_at_5_std": -0.24295,
72
+ "nauc_map_at_5_diff1": 0.162997,
73
+ "nauc_map_at_10_max": -0.110573,
74
+ "nauc_map_at_10_std": -0.241374,
75
+ "nauc_map_at_10_diff1": 0.164262,
76
+ "nauc_map_at_20_max": -0.113845,
77
+ "nauc_map_at_20_std": -0.241569,
78
+ "nauc_map_at_20_diff1": 0.162587,
79
+ "nauc_map_at_100_max": -0.114457,
80
+ "nauc_map_at_100_std": -0.241589,
81
+ "nauc_map_at_100_diff1": 0.162436,
82
+ "nauc_map_at_1000_max": -0.114457,
83
+ "nauc_map_at_1000_std": -0.241589,
84
+ "nauc_map_at_1000_diff1": 0.162436,
85
+ "nauc_recall_at_1_max": -0.165929,
86
+ "nauc_recall_at_1_std": -0.245991,
87
+ "nauc_recall_at_1_diff1": 0.158499,
88
+ "nauc_recall_at_3_max": 0.050787,
89
+ "nauc_recall_at_3_std": -0.219176,
90
+ "nauc_recall_at_3_diff1": 0.193997,
91
+ "nauc_recall_at_5_max": 0.14777,
92
+ "nauc_recall_at_5_std": -0.24361,
93
+ "nauc_recall_at_5_diff1": 0.209978,
94
+ "nauc_recall_at_10_max": 0.365525,
95
+ "nauc_recall_at_10_std": -0.15247,
96
+ "nauc_recall_at_10_diff1": 0.341837,
97
+ "nauc_recall_at_20_max": 0.432205,
98
+ "nauc_recall_at_20_std": 0.043037,
99
+ "nauc_recall_at_20_diff1": 0.198741,
100
+ "nauc_recall_at_100_max": 0.268327,
101
+ "nauc_recall_at_100_std": 0.743186,
102
+ "nauc_recall_at_100_diff1": -0.026984,
103
+ "nauc_recall_at_1000_max": 0.268327,
104
+ "nauc_recall_at_1000_std": 0.743186,
105
+ "nauc_recall_at_1000_diff1": -0.026984,
106
+ "nauc_precision_at_1_max": -0.165929,
107
+ "nauc_precision_at_1_std": -0.245991,
108
+ "nauc_precision_at_1_diff1": 0.158499,
109
+ "nauc_precision_at_3_max": 0.050787,
110
+ "nauc_precision_at_3_std": -0.219176,
111
+ "nauc_precision_at_3_diff1": 0.193997,
112
+ "nauc_precision_at_5_max": 0.14777,
113
+ "nauc_precision_at_5_std": -0.24361,
114
+ "nauc_precision_at_5_diff1": 0.209978,
115
+ "nauc_precision_at_10_max": 0.365525,
116
+ "nauc_precision_at_10_std": -0.15247,
117
+ "nauc_precision_at_10_diff1": 0.341837,
118
+ "nauc_precision_at_20_max": 0.432205,
119
+ "nauc_precision_at_20_std": 0.043037,
120
+ "nauc_precision_at_20_diff1": 0.198741,
121
+ "nauc_precision_at_100_max": 0.268327,
122
+ "nauc_precision_at_100_std": 0.743186,
123
+ "nauc_precision_at_100_diff1": -0.026984,
124
+ "nauc_precision_at_1000_max": 0.268327,
125
+ "nauc_precision_at_1000_std": 0.743186,
126
+ "nauc_precision_at_1000_diff1": -0.026984,
127
+ "nauc_mrr_at_1_max": -0.171375,
128
+ "nauc_mrr_at_1_std": -0.232382,
129
+ "nauc_mrr_at_1_diff1": 0.131375,
130
+ "nauc_mrr_at_3_max": -0.119916,
131
+ "nauc_mrr_at_3_std": -0.234723,
132
+ "nauc_mrr_at_3_diff1": 0.13217,
133
+ "nauc_mrr_at_5_max": -0.126585,
134
+ "nauc_mrr_at_5_std": -0.238874,
135
+ "nauc_mrr_at_5_diff1": 0.127312,
136
+ "nauc_mrr_at_10_max": -0.127398,
137
+ "nauc_mrr_at_10_std": -0.23595,
138
+ "nauc_mrr_at_10_diff1": 0.130786,
139
+ "nauc_mrr_at_20_max": -0.130821,
140
+ "nauc_mrr_at_20_std": -0.236116,
141
+ "nauc_mrr_at_20_diff1": 0.128852,
142
+ "nauc_mrr_at_100_max": -0.131452,
143
+ "nauc_mrr_at_100_std": -0.236131,
144
+ "nauc_mrr_at_100_diff1": 0.128668,
145
+ "nauc_mrr_at_1000_max": -0.131452,
146
+ "nauc_mrr_at_1000_std": -0.236131,
147
+ "nauc_mrr_at_1000_diff1": 0.128668,
148
+ "main_score": 0.73971,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "eng-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 499.8532328605652,
157
+ "kg_co2_emissions": null
158
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/ArmenianParaphrasePC.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f43b4f32987048043a8b31e5e26be4d360c2438f",
3
+ "task_name": "ArmenianParaphrasePC",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "similarity_accuracy": 0.904762,
9
+ "similarity_accuracy_threshold": 0.747987,
10
+ "similarity_f1": 0.935484,
11
+ "similarity_f1_threshold": 0.741345,
12
+ "similarity_precision": 0.883377,
13
+ "similarity_recall": 0.994123,
14
+ "similarity_ap": 0.953269,
15
+ "cosine_accuracy": 0.904762,
16
+ "cosine_accuracy_threshold": 0.747986,
17
+ "cosine_f1": 0.935484,
18
+ "cosine_f1_threshold": 0.741344,
19
+ "cosine_precision": 0.883377,
20
+ "cosine_recall": 0.994123,
21
+ "cosine_ap": 0.953269,
22
+ "manhattan_accuracy": 0.904762,
23
+ "manhattan_accuracy_threshold": 35.372028,
24
+ "manhattan_f1": 0.935424,
25
+ "manhattan_f1_threshold": 35.626312,
26
+ "manhattan_precision": 0.884045,
27
+ "manhattan_recall": 0.993144,
28
+ "manhattan_ap": 0.953077,
29
+ "euclidean_accuracy": 0.905442,
30
+ "euclidean_accuracy_threshold": 0.719254,
31
+ "euclidean_f1": 0.935915,
32
+ "euclidean_f1_threshold": 0.719254,
33
+ "euclidean_precision": 0.884146,
34
+ "euclidean_recall": 0.994123,
35
+ "euclidean_ap": 0.953237,
36
+ "dot_accuracy": 0.904762,
37
+ "dot_accuracy_threshold": 0.748643,
38
+ "dot_f1": 0.935365,
39
+ "dot_f1_threshold": 0.748643,
40
+ "dot_precision": 0.884716,
41
+ "dot_recall": 0.992165,
42
+ "dot_ap": 0.954403,
43
+ "max_accuracy": 0.905442,
44
+ "max_f1": 0.935915,
45
+ "max_precision": 0.884716,
46
+ "max_recall": 0.994123,
47
+ "max_ap": 0.954403,
48
+ "main_score": 0.954403,
49
+ "hf_subset": "default",
50
+ "languages": [
51
+ "hye-Armn"
52
+ ]
53
+ }
54
+ ]
55
+ },
56
+ "evaluation_time": 110.84050726890564,
57
+ "kg_co2_emissions": null
58
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BUCC.v2.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1739dc11ffe9b7bfccd7f3d585aeb4c544fc6677",
3
+ "task_name": "BUCC.v2",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "precision": 0.994624,
9
+ "recall": 0.996347,
10
+ "f1": 0.995198,
11
+ "accuracy": 0.996347,
12
+ "main_score": 0.995198,
13
+ "hf_subset": "de-en",
14
+ "languages": [
15
+ "deu-Latn",
16
+ "eng-Latn"
17
+ ]
18
+ },
19
+ {
20
+ "precision": 0.972417,
21
+ "recall": 0.981295,
22
+ "f1": 0.975361,
23
+ "accuracy": 0.981295,
24
+ "main_score": 0.975361,
25
+ "hf_subset": "ru-en",
26
+ "languages": [
27
+ "rus-Cyrl",
28
+ "eng-Latn"
29
+ ]
30
+ },
31
+ {
32
+ "precision": 0.990315,
33
+ "recall": 0.993506,
34
+ "f1": 0.991379,
35
+ "accuracy": 0.993506,
36
+ "main_score": 0.991379,
37
+ "hf_subset": "fr-en",
38
+ "languages": [
39
+ "fra-Latn",
40
+ "eng-Latn"
41
+ ]
42
+ },
43
+ {
44
+ "precision": 0.993681,
45
+ "recall": 0.995787,
46
+ "f1": 0.994383,
47
+ "accuracy": 0.995787,
48
+ "main_score": 0.994383,
49
+ "hf_subset": "zh-en",
50
+ "languages": [
51
+ "cmn-Hans",
52
+ "eng-Latn"
53
+ ]
54
+ }
55
+ ]
56
+ },
57
+ "evaluation_time": 5385.880857229233,
58
+ "kg_co2_emissions": null
59
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BelebeleRetrieval.json ADDED
The diff for this file is too large to render. See raw diff
 
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BigPatentClustering.v2.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "58a863a958586a5d6ba51088b94ac74a46aa864f",
3
+ "task_name": "BigPatentClustering.v2",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.281881,
11
+ 0.312428,
12
+ 0.332042,
13
+ 0.375078,
14
+ 0.300211,
15
+ 0.348814,
16
+ 0.310416,
17
+ 0.29452,
18
+ 0.301125,
19
+ 0.323222
20
+ ]
21
+ },
22
+ "v_measure": 0.317974,
23
+ "v_measure_std": 0.026415,
24
+ "main_score": 0.317974,
25
+ "hf_subset": "default",
26
+ "languages": [
27
+ "eng-Latn"
28
+ ]
29
+ }
30
+ ]
31
+ },
32
+ "evaluation_time": 21.053664207458496,
33
+ "kg_co2_emissions": null
34
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BiorxivClusteringP2P.v2.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f5dbc242e11dd8e24def4c4268607a49e02946dc",
3
+ "task_name": "BiorxivClusteringP2P.v2",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.432188,
11
+ 0.401276,
12
+ 0.436429,
13
+ 0.400495,
14
+ 0.410255,
15
+ 0.422017,
16
+ 0.417689,
17
+ 0.387773,
18
+ 0.439086,
19
+ 0.408914
20
+ ]
21
+ },
22
+ "v_measure": 0.415612,
23
+ "v_measure_std": 0.016084,
24
+ "main_score": 0.415612,
25
+ "hf_subset": "default",
26
+ "languages": [
27
+ "eng-Latn"
28
+ ]
29
+ }
30
+ ]
31
+ },
32
+ "evaluation_time": 117.01535153388977,
33
+ "kg_co2_emissions": null
34
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BornholmBitextMining.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3bc5cfb4ec514264fe2db5615fac9016f7251552",
3
+ "task_name": "BornholmBitextMining",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "precision": 0.5651,
9
+ "recall": 0.654,
10
+ "f1": 0.591697,
11
+ "accuracy": 0.654,
12
+ "main_score": 0.591697,
13
+ "hf_subset": "default",
14
+ "languages": [
15
+ "dan-Latn"
16
+ ]
17
+ }
18
+ ]
19
+ },
20
+ "evaluation_time": 74.43902254104614,
21
+ "kg_co2_emissions": null
22
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BrazilianToxicTweetsClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f333c1fcfa3ab43f008a327c8bd0140441354d34",
3
+ "task_name": "BrazilianToxicTweetsClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.209473,
9
+ "f1": 0.188782,
10
+ "lrap": 0.809945,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.17334,
14
+ "f1": 0.169217,
15
+ "lrap": 0.847738
16
+ },
17
+ {
18
+ "accuracy": 0.134277,
19
+ "f1": 0.18848,
20
+ "lrap": 0.812676
21
+ },
22
+ {
23
+ "accuracy": 0.115723,
24
+ "f1": 0.170022,
25
+ "lrap": 0.824843
26
+ },
27
+ {
28
+ "accuracy": 0.217773,
29
+ "f1": 0.152533,
30
+ "lrap": 0.843031
31
+ },
32
+ {
33
+ "accuracy": 0.194336,
34
+ "f1": 0.229408,
35
+ "lrap": 0.808431
36
+ },
37
+ {
38
+ "accuracy": 0.199219,
39
+ "f1": 0.240764,
40
+ "lrap": 0.819309
41
+ },
42
+ {
43
+ "accuracy": 0.281738,
44
+ "f1": 0.175724,
45
+ "lrap": 0.778497
46
+ },
47
+ {
48
+ "accuracy": 0.163086,
49
+ "f1": 0.198641,
50
+ "lrap": 0.821791
51
+ },
52
+ {
53
+ "accuracy": 0.287598,
54
+ "f1": 0.18755,
55
+ "lrap": 0.780396
56
+ },
57
+ {
58
+ "accuracy": 0.327637,
59
+ "f1": 0.175483,
60
+ "lrap": 0.762736
61
+ }
62
+ ],
63
+ "main_score": 0.209473,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "por-Latn"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 104.54572677612305,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/BulgarianStoreReviewSentimentClassfication.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "701984d6c6efea0e14a1c7850ef70e464c5577c0",
3
+ "task_name": "BulgarianStoreReviewSentimentClassfication",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.648352,
9
+ "f1": 0.473341,
10
+ "f1_weighted": 0.68251,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.620879,
14
+ "f1": 0.497385,
15
+ "f1_weighted": 0.663394
16
+ },
17
+ {
18
+ "accuracy": 0.604396,
19
+ "f1": 0.441179,
20
+ "f1_weighted": 0.671954
21
+ },
22
+ {
23
+ "accuracy": 0.615385,
24
+ "f1": 0.450957,
25
+ "f1_weighted": 0.656301
26
+ },
27
+ {
28
+ "accuracy": 0.626374,
29
+ "f1": 0.424899,
30
+ "f1_weighted": 0.677175
31
+ },
32
+ {
33
+ "accuracy": 0.67033,
34
+ "f1": 0.480964,
35
+ "f1_weighted": 0.690368
36
+ },
37
+ {
38
+ "accuracy": 0.543956,
39
+ "f1": 0.413541,
40
+ "f1_weighted": 0.595773
41
+ },
42
+ {
43
+ "accuracy": 0.681319,
44
+ "f1": 0.495881,
45
+ "f1_weighted": 0.717621
46
+ },
47
+ {
48
+ "accuracy": 0.703297,
49
+ "f1": 0.533786,
50
+ "f1_weighted": 0.731661
51
+ },
52
+ {
53
+ "accuracy": 0.686813,
54
+ "f1": 0.479972,
55
+ "f1_weighted": 0.694872
56
+ },
57
+ {
58
+ "accuracy": 0.730769,
59
+ "f1": 0.514844,
60
+ "f1_weighted": 0.725985
61
+ }
62
+ ],
63
+ "main_score": 0.648352,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "bul-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 99.66069555282593,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CEDRClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "c0ba03d058e3e1b2f3fd20518875a4563dd12db4",
3
+ "task_name": "CEDRClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.425292,
9
+ "f1": 0.356858,
10
+ "lrap": 0.673305,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.409139,
14
+ "f1": 0.321349,
15
+ "lrap": 0.637673
16
+ },
17
+ {
18
+ "accuracy": 0.413921,
19
+ "f1": 0.350593,
20
+ "lrap": 0.671892
21
+ },
22
+ {
23
+ "accuracy": 0.431456,
24
+ "f1": 0.337593,
25
+ "lrap": 0.648459
26
+ },
27
+ {
28
+ "accuracy": 0.449522,
29
+ "f1": 0.382532,
30
+ "lrap": 0.679809
31
+ },
32
+ {
33
+ "accuracy": 0.465994,
34
+ "f1": 0.378172,
35
+ "lrap": 0.720882
36
+ },
37
+ {
38
+ "accuracy": 0.414984,
39
+ "f1": 0.332556,
40
+ "lrap": 0.657651
41
+ },
42
+ {
43
+ "accuracy": 0.417641,
44
+ "f1": 0.367538,
45
+ "lrap": 0.680181
46
+ },
47
+ {
48
+ "accuracy": 0.392136,
49
+ "f1": 0.309033,
50
+ "lrap": 0.664878
51
+ },
52
+ {
53
+ "accuracy": 0.426142,
54
+ "f1": 0.388862,
55
+ "lrap": 0.680074
56
+ },
57
+ {
58
+ "accuracy": 0.431987,
59
+ "f1": 0.400347,
60
+ "lrap": 0.691552
61
+ }
62
+ ],
63
+ "main_score": 0.425292,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 93.12559366226196,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CLSClusteringP2P.v2.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "4b6227591c6c1a73bc76b1055f3b7f3588e72476",
3
+ "task_name": "CLSClusteringP2P.v2",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.427589,
11
+ 0.426011,
12
+ 0.425302,
13
+ 0.408956,
14
+ 0.39785,
15
+ 0.423979,
16
+ 0.406984,
17
+ 0.381813,
18
+ 0.398873,
19
+ 0.424785
20
+ ]
21
+ },
22
+ "v_measure": 0.412214,
23
+ "v_measure_std": 0.01497,
24
+ "main_score": 0.412214,
25
+ "hf_subset": "default",
26
+ "languages": [
27
+ "cmn-Hans"
28
+ ]
29
+ }
30
+ ]
31
+ },
32
+ "evaluation_time": 91.68763017654419,
33
+ "kg_co2_emissions": null
34
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CSFDSKMovieReviewSentimentClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "23a20c659d868740ef9c54854de631fe19cd5c17",
3
+ "task_name": "CSFDSKMovieReviewSentimentClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.377002,
9
+ "f1": 0.365314,
10
+ "f1_weighted": 0.367747,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.389648,
14
+ "f1": 0.38023,
15
+ "f1_weighted": 0.382672
16
+ },
17
+ {
18
+ "accuracy": 0.402344,
19
+ "f1": 0.39006,
20
+ "f1_weighted": 0.392455
21
+ },
22
+ {
23
+ "accuracy": 0.394531,
24
+ "f1": 0.381217,
25
+ "f1_weighted": 0.3837
26
+ },
27
+ {
28
+ "accuracy": 0.359375,
29
+ "f1": 0.353808,
30
+ "f1_weighted": 0.355898
31
+ },
32
+ {
33
+ "accuracy": 0.363281,
34
+ "f1": 0.351216,
35
+ "f1_weighted": 0.353627
36
+ },
37
+ {
38
+ "accuracy": 0.36377,
39
+ "f1": 0.333191,
40
+ "f1_weighted": 0.336177
41
+ },
42
+ {
43
+ "accuracy": 0.359375,
44
+ "f1": 0.345533,
45
+ "f1_weighted": 0.348155
46
+ },
47
+ {
48
+ "accuracy": 0.381348,
49
+ "f1": 0.3783,
50
+ "f1_weighted": 0.379799
51
+ },
52
+ {
53
+ "accuracy": 0.380859,
54
+ "f1": 0.370287,
55
+ "f1_weighted": 0.373841
56
+ },
57
+ {
58
+ "accuracy": 0.375488,
59
+ "f1": 0.369302,
60
+ "f1_weighted": 0.371144
61
+ }
62
+ ],
63
+ "main_score": 0.377002,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "slk-Latn"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 135.74484515190125,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CTKFactsNLI.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "387ae4582c8054cb52ef57ef0941f19bd8012abf",
3
+ "task_name": "CTKFactsNLI",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "similarity_accuracy": 0.72,
9
+ "similarity_accuracy_threshold": 0.403134,
10
+ "similarity_f1": 0.831461,
11
+ "similarity_f1_threshold": 0.399513,
12
+ "similarity_precision": 0.713499,
13
+ "similarity_recall": 0.996154,
14
+ "similarity_ap": 0.795362,
15
+ "cosine_accuracy": 0.72,
16
+ "cosine_accuracy_threshold": 0.403132,
17
+ "cosine_f1": 0.831461,
18
+ "cosine_f1_threshold": 0.399511,
19
+ "cosine_precision": 0.713499,
20
+ "cosine_recall": 0.996154,
21
+ "cosine_ap": 0.795362,
22
+ "manhattan_accuracy": 0.714667,
23
+ "manhattan_accuracy_threshold": 54.078541,
24
+ "manhattan_f1": 0.829346,
25
+ "manhattan_f1_threshold": 55.72036,
26
+ "manhattan_precision": 0.708447,
27
+ "manhattan_recall": 1.0,
28
+ "manhattan_ap": 0.789792,
29
+ "euclidean_accuracy": 0.72,
30
+ "euclidean_accuracy_threshold": 1.094939,
31
+ "euclidean_f1": 0.830918,
32
+ "euclidean_f1_threshold": 1.094939,
33
+ "euclidean_precision": 0.714681,
34
+ "euclidean_recall": 0.992308,
35
+ "euclidean_ap": 0.795524,
36
+ "dot_accuracy": 0.72,
37
+ "dot_accuracy_threshold": 0.404998,
38
+ "dot_f1": 0.831461,
39
+ "dot_f1_threshold": 0.401406,
40
+ "dot_precision": 0.713499,
41
+ "dot_recall": 0.996154,
42
+ "dot_ap": 0.795317,
43
+ "max_accuracy": 0.72,
44
+ "max_f1": 0.831461,
45
+ "max_precision": 0.714681,
46
+ "max_recall": 1.0,
47
+ "max_ap": 0.795524,
48
+ "main_score": 0.795524,
49
+ "hf_subset": "default",
50
+ "languages": [
51
+ "ces-Latn"
52
+ ]
53
+ }
54
+ ]
55
+ },
56
+ "evaluation_time": 14.797275066375732,
57
+ "kg_co2_emissions": null
58
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CataloniaTweetClassification.json ADDED
@@ -0,0 +1,135 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "cf24d44e517efa534f048e5fc5981f399ed25bee",
3
+ "task_name": "CataloniaTweetClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.525794,
9
+ "f1": 0.549233,
10
+ "f1_weighted": 0.517639,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.553571,
14
+ "f1": 0.584086,
15
+ "f1_weighted": 0.551816
16
+ },
17
+ {
18
+ "accuracy": 0.566964,
19
+ "f1": 0.587871,
20
+ "f1_weighted": 0.564099
21
+ },
22
+ {
23
+ "accuracy": 0.465278,
24
+ "f1": 0.470657,
25
+ "f1_weighted": 0.443433
26
+ },
27
+ {
28
+ "accuracy": 0.536706,
29
+ "f1": 0.56473,
30
+ "f1_weighted": 0.534761
31
+ },
32
+ {
33
+ "accuracy": 0.515873,
34
+ "f1": 0.537752,
35
+ "f1_weighted": 0.498123
36
+ },
37
+ {
38
+ "accuracy": 0.56746,
39
+ "f1": 0.602129,
40
+ "f1_weighted": 0.568323
41
+ },
42
+ {
43
+ "accuracy": 0.533234,
44
+ "f1": 0.559995,
45
+ "f1_weighted": 0.523366
46
+ },
47
+ {
48
+ "accuracy": 0.500496,
49
+ "f1": 0.510807,
50
+ "f1_weighted": 0.485268
51
+ },
52
+ {
53
+ "accuracy": 0.49504,
54
+ "f1": 0.506979,
55
+ "f1_weighted": 0.482797
56
+ },
57
+ {
58
+ "accuracy": 0.523313,
59
+ "f1": 0.567319,
60
+ "f1_weighted": 0.524407
61
+ }
62
+ ],
63
+ "main_score": 0.525794,
64
+ "hf_subset": "spanish",
65
+ "languages": [
66
+ "spa-Latn"
67
+ ]
68
+ },
69
+ {
70
+ "accuracy": 0.51,
71
+ "f1": 0.49745,
72
+ "f1_weighted": 0.507792,
73
+ "scores_per_experiment": [
74
+ {
75
+ "accuracy": 0.533831,
76
+ "f1": 0.514831,
77
+ "f1_weighted": 0.530116
78
+ },
79
+ {
80
+ "accuracy": 0.459701,
81
+ "f1": 0.451874,
82
+ "f1_weighted": 0.457143
83
+ },
84
+ {
85
+ "accuracy": 0.558209,
86
+ "f1": 0.547588,
87
+ "f1_weighted": 0.552017
88
+ },
89
+ {
90
+ "accuracy": 0.491542,
91
+ "f1": 0.474789,
92
+ "f1_weighted": 0.488983
93
+ },
94
+ {
95
+ "accuracy": 0.502985,
96
+ "f1": 0.461663,
97
+ "f1_weighted": 0.493571
98
+ },
99
+ {
100
+ "accuracy": 0.524378,
101
+ "f1": 0.516597,
102
+ "f1_weighted": 0.521844
103
+ },
104
+ {
105
+ "accuracy": 0.538806,
106
+ "f1": 0.527502,
107
+ "f1_weighted": 0.540065
108
+ },
109
+ {
110
+ "accuracy": 0.499502,
111
+ "f1": 0.495767,
112
+ "f1_weighted": 0.500189
113
+ },
114
+ {
115
+ "accuracy": 0.516915,
116
+ "f1": 0.513497,
117
+ "f1_weighted": 0.517313
118
+ },
119
+ {
120
+ "accuracy": 0.474129,
121
+ "f1": 0.470391,
122
+ "f1_weighted": 0.476676
123
+ }
124
+ ],
125
+ "main_score": 0.51,
126
+ "hf_subset": "catalan",
127
+ "languages": [
128
+ "cat-Latn"
129
+ ]
130
+ }
131
+ ]
132
+ },
133
+ "evaluation_time": 351.615517616272,
134
+ "kg_co2_emissions": null
135
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/Core17InstructionRetrieval.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e39ff896cf3efbbdeeb950e6bd7c79f266995b07",
3
+ "task_name": "Core17InstructionRetrieval",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "p-MRR": 0.11337,
9
+ "individual": {
10
+ "original": {
11
+ "ndcg_at_1": 0.275,
12
+ "ndcg_at_3": 0.29012,
13
+ "ndcg_at_5": 0.30286,
14
+ "ndcg_at_10": 0.31235,
15
+ "ndcg_at_20": 0.31111,
16
+ "ndcg_at_100": 0.41942,
17
+ "ndcg_at_1000": 0.58685,
18
+ "map_at_1": 0.01174,
19
+ "map_at_3": 0.02882,
20
+ "map_at_5": 0.04555,
21
+ "map_at_10": 0.07724,
22
+ "map_at_20": 0.11551,
23
+ "map_at_100": 0.20645,
24
+ "map_at_1000": 0.25053,
25
+ "recall_at_1": 0.01174,
26
+ "recall_at_3": 0.03531,
27
+ "recall_at_5": 0.067,
28
+ "recall_at_10": 0.12542,
29
+ "recall_at_20": 0.2092,
30
+ "recall_at_100": 0.52347,
31
+ "recall_at_1000": 1.0,
32
+ "precision_at_1": 0.35,
33
+ "precision_at_3": 0.38333,
34
+ "precision_at_5": 0.38,
35
+ "precision_at_10": 0.355,
36
+ "precision_at_20": 0.3025,
37
+ "precision_at_100": 0.166,
38
+ "precision_at_1000": 0.0327,
39
+ "naucs_at_1_max": 0.091436,
40
+ "naucs_at_1_std": 0.01068,
41
+ "naucs_at_1_diff1": 0.078977,
42
+ "naucs_at_3_max": 0.069722,
43
+ "naucs_at_3_std": 0.00617,
44
+ "naucs_at_3_diff1": 0.158235,
45
+ "naucs_at_5_max": 0.100777,
46
+ "naucs_at_5_std": -0.022436,
47
+ "naucs_at_5_diff1": 0.178238,
48
+ "naucs_at_10_max": 0.115412,
49
+ "naucs_at_10_std": -0.060505,
50
+ "naucs_at_10_diff1": 0.147285,
51
+ "naucs_at_20_max": 0.108113,
52
+ "naucs_at_20_std": -0.067626,
53
+ "naucs_at_20_diff1": 0.155401,
54
+ "naucs_at_100_max": 0.108871,
55
+ "naucs_at_100_std": -0.053601,
56
+ "naucs_at_100_diff1": 0.15436,
57
+ "naucs_at_1000_max": 0.108871,
58
+ "naucs_at_1000_std": -0.053601,
59
+ "naucs_at_1000_diff1": 0.15436,
60
+ "mrr_at_1": 0.35,
61
+ "mrr_at_3": 0.475,
62
+ "mrr_at_5": 0.4975,
63
+ "mrr_at_10": 0.509643,
64
+ "mrr_at_20": 0.513214,
65
+ "mrr_at_100": 0.516576,
66
+ "mrr_at_1000": 0.516576
67
+ },
68
+ "changed": {
69
+ "ndcg_at_1": 0.675,
70
+ "ndcg_at_3": 0.58957,
71
+ "ndcg_at_5": 0.57597,
72
+ "ndcg_at_10": 0.55538,
73
+ "ndcg_at_20": 0.53471,
74
+ "ndcg_at_100": 0.55562,
75
+ "ndcg_at_1000": 0.7441,
76
+ "map_at_1": 0.02291,
77
+ "map_at_3": 0.04486,
78
+ "map_at_5": 0.06538,
79
+ "map_at_10": 0.11322,
80
+ "map_at_20": 0.18124,
81
+ "map_at_100": 0.33106,
82
+ "map_at_1000": 0.40646,
83
+ "recall_at_1": 0.02291,
84
+ "recall_at_3": 0.04828,
85
+ "recall_at_5": 0.07631,
86
+ "recall_at_10": 0.13942,
87
+ "recall_at_20": 0.23231,
88
+ "recall_at_100": 0.54679,
89
+ "recall_at_1000": 1.0,
90
+ "precision_at_1": 0.8,
91
+ "precision_at_3": 0.66667,
92
+ "precision_at_5": 0.66,
93
+ "precision_at_10": 0.62,
94
+ "precision_at_20": 0.5525,
95
+ "precision_at_100": 0.2885,
96
+ "precision_at_1000": 0.0545,
97
+ "naucs_at_1_max": 0.001873,
98
+ "naucs_at_1_std": 0.324051,
99
+ "naucs_at_1_diff1": 0.041209,
100
+ "naucs_at_3_max": 0.041995,
101
+ "naucs_at_3_std": 0.356215,
102
+ "naucs_at_3_diff1": 0.083759,
103
+ "naucs_at_5_max": 0.041995,
104
+ "naucs_at_5_std": 0.356215,
105
+ "naucs_at_5_diff1": 0.083759,
106
+ "naucs_at_10_max": 0.041995,
107
+ "naucs_at_10_std": 0.356215,
108
+ "naucs_at_10_diff1": 0.083759,
109
+ "naucs_at_20_max": 0.041995,
110
+ "naucs_at_20_std": 0.356215,
111
+ "naucs_at_20_diff1": 0.083759,
112
+ "naucs_at_100_max": 0.041995,
113
+ "naucs_at_100_std": 0.356215,
114
+ "naucs_at_100_diff1": 0.083759,
115
+ "naucs_at_1000_max": 0.041995,
116
+ "naucs_at_1000_std": 0.356215,
117
+ "naucs_at_1000_diff1": 0.083759,
118
+ "mrr_at_1": 0.8,
119
+ "mrr_at_3": 0.875,
120
+ "mrr_at_5": 0.875,
121
+ "mrr_at_10": 0.875,
122
+ "mrr_at_20": 0.875,
123
+ "mrr_at_100": 0.875,
124
+ "mrr_at_1000": 0.875
125
+ }
126
+ },
127
+ "main_score": 0.11337,
128
+ "hf_subset": "default",
129
+ "languages": [
130
+ "eng-Latn"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 2546.0000607967377,
136
+ "kg_co2_emissions": null
137
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CyrillicTurkicLangClassification.json ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e42d330f33d65b7b72dfd408883daf1661f06f18",
3
+ "task_name": "CyrillicTurkicLangClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.819434,
9
+ "f1": 0.816873,
10
+ "f1_weighted": 0.816897,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.825684,
14
+ "f1": 0.825088,
15
+ "f1_weighted": 0.825157
16
+ },
17
+ {
18
+ "accuracy": 0.808594,
19
+ "f1": 0.805388,
20
+ "f1_weighted": 0.805459
21
+ },
22
+ {
23
+ "accuracy": 0.814941,
24
+ "f1": 0.810465,
25
+ "f1_weighted": 0.810456
26
+ },
27
+ {
28
+ "accuracy": 0.835449,
29
+ "f1": 0.833361,
30
+ "f1_weighted": 0.833358
31
+ },
32
+ {
33
+ "accuracy": 0.817871,
34
+ "f1": 0.811819,
35
+ "f1_weighted": 0.811846
36
+ },
37
+ {
38
+ "accuracy": 0.844238,
39
+ "f1": 0.843029,
40
+ "f1_weighted": 0.843048
41
+ },
42
+ {
43
+ "accuracy": 0.811523,
44
+ "f1": 0.811064,
45
+ "f1_weighted": 0.811077
46
+ },
47
+ {
48
+ "accuracy": 0.812988,
49
+ "f1": 0.81043,
50
+ "f1_weighted": 0.810438
51
+ },
52
+ {
53
+ "accuracy": 0.786621,
54
+ "f1": 0.784479,
55
+ "f1_weighted": 0.784511
56
+ },
57
+ {
58
+ "accuracy": 0.836426,
59
+ "f1": 0.833611,
60
+ "f1_weighted": 0.83362
61
+ }
62
+ ],
63
+ "main_score": 0.819434,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "bak-Cyrl",
67
+ "chv-Cyrl",
68
+ "tat-Cyrl",
69
+ "kir-Cyrl",
70
+ "rus-Cyrl",
71
+ "kaz-Cyrl",
72
+ "tyv-Cyrl",
73
+ "krc-Cyrl",
74
+ "sah-Cyrl"
75
+ ]
76
+ }
77
+ ]
78
+ },
79
+ "evaluation_time": 255.61350011825562,
80
+ "kg_co2_emissions": null
81
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/CzechProductReviewSentimentClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "2e6fedf42c9c104e83dfd95c3a453721e683e244",
3
+ "task_name": "CzechProductReviewSentimentClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.565723,
9
+ "f1": 0.553393,
10
+ "f1_weighted": 0.553343,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.578613,
14
+ "f1": 0.565857,
15
+ "f1_weighted": 0.565812
16
+ },
17
+ {
18
+ "accuracy": 0.550293,
19
+ "f1": 0.544142,
20
+ "f1_weighted": 0.544102
21
+ },
22
+ {
23
+ "accuracy": 0.543457,
24
+ "f1": 0.526513,
25
+ "f1_weighted": 0.526442
26
+ },
27
+ {
28
+ "accuracy": 0.55957,
29
+ "f1": 0.543408,
30
+ "f1_weighted": 0.54337
31
+ },
32
+ {
33
+ "accuracy": 0.56543,
34
+ "f1": 0.557029,
35
+ "f1_weighted": 0.556999
36
+ },
37
+ {
38
+ "accuracy": 0.530762,
39
+ "f1": 0.513238,
40
+ "f1_weighted": 0.513183
41
+ },
42
+ {
43
+ "accuracy": 0.600586,
44
+ "f1": 0.591779,
45
+ "f1_weighted": 0.591723
46
+ },
47
+ {
48
+ "accuracy": 0.583008,
49
+ "f1": 0.570875,
50
+ "f1_weighted": 0.570826
51
+ },
52
+ {
53
+ "accuracy": 0.583008,
54
+ "f1": 0.574911,
55
+ "f1_weighted": 0.574859
56
+ },
57
+ {
58
+ "accuracy": 0.5625,
59
+ "f1": 0.54618,
60
+ "f1_weighted": 0.546116
61
+ }
62
+ ],
63
+ "main_score": 0.565723,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "ces-Latn"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 252.52034187316895,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/DBpediaClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "9abd46cf7fc8b4c64290f26993c540b92aa145ac",
3
+ "task_name": "DBpediaClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.905273,
9
+ "f1": 0.901306,
10
+ "f1_weighted": 0.901344,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.90918,
14
+ "f1": 0.905738,
15
+ "f1_weighted": 0.905779
16
+ },
17
+ {
18
+ "accuracy": 0.909668,
19
+ "f1": 0.905101,
20
+ "f1_weighted": 0.905143
21
+ },
22
+ {
23
+ "accuracy": 0.906738,
24
+ "f1": 0.903304,
25
+ "f1_weighted": 0.903343
26
+ },
27
+ {
28
+ "accuracy": 0.893555,
29
+ "f1": 0.887717,
30
+ "f1_weighted": 0.887754
31
+ },
32
+ {
33
+ "accuracy": 0.898926,
34
+ "f1": 0.894722,
35
+ "f1_weighted": 0.894792
36
+ },
37
+ {
38
+ "accuracy": 0.908691,
39
+ "f1": 0.905254,
40
+ "f1_weighted": 0.90528
41
+ },
42
+ {
43
+ "accuracy": 0.909668,
44
+ "f1": 0.90746,
45
+ "f1_weighted": 0.907515
46
+ },
47
+ {
48
+ "accuracy": 0.902344,
49
+ "f1": 0.898902,
50
+ "f1_weighted": 0.898941
51
+ },
52
+ {
53
+ "accuracy": 0.905762,
54
+ "f1": 0.902584,
55
+ "f1_weighted": 0.90257
56
+ },
57
+ {
58
+ "accuracy": 0.908203,
59
+ "f1": 0.902274,
60
+ "f1_weighted": 0.902326
61
+ }
62
+ ],
63
+ "main_score": 0.905273,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "eng-Latn"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 307.0829930305481,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/DiaBlaBitextMining.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "5345895c56a601afe1a98519ce3199be60a27dba",
3
+ "task_name": "DiaBlaBitextMining",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "precision": 0.81731,
9
+ "recall": 0.844468,
10
+ "f1": 0.824683,
11
+ "accuracy": 0.844468,
12
+ "main_score": 0.824683,
13
+ "hf_subset": "fr-en",
14
+ "languages": [
15
+ "fra-Latn",
16
+ "eng-Latn"
17
+ ]
18
+ },
19
+ {
20
+ "precision": 0.81731,
21
+ "recall": 0.844468,
22
+ "f1": 0.824683,
23
+ "accuracy": 0.844468,
24
+ "main_score": 0.824683,
25
+ "hf_subset": "en-fr",
26
+ "languages": [
27
+ "eng-Latn",
28
+ "fra-Latn"
29
+ ]
30
+ }
31
+ ]
32
+ },
33
+ "evaluation_time": 1765.068434715271,
34
+ "kg_co2_emissions": null
35
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/EstonianValenceClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "9157397f05a127b3ac93b93dd88abf1bdf710c22",
3
+ "task_name": "EstonianValenceClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.404279,
9
+ "f1": 0.369538,
10
+ "f1_weighted": 0.423493,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.370416,
14
+ "f1": 0.343942,
15
+ "f1_weighted": 0.389314
16
+ },
17
+ {
18
+ "accuracy": 0.420538,
19
+ "f1": 0.37761,
20
+ "f1_weighted": 0.444412
21
+ },
22
+ {
23
+ "accuracy": 0.380196,
24
+ "f1": 0.33836,
25
+ "f1_weighted": 0.395387
26
+ },
27
+ {
28
+ "accuracy": 0.419315,
29
+ "f1": 0.370538,
30
+ "f1_weighted": 0.436729
31
+ },
32
+ {
33
+ "accuracy": 0.376528,
34
+ "f1": 0.354854,
35
+ "f1_weighted": 0.400578
36
+ },
37
+ {
38
+ "accuracy": 0.430318,
39
+ "f1": 0.416155,
40
+ "f1_weighted": 0.447153
41
+ },
42
+ {
43
+ "accuracy": 0.392421,
44
+ "f1": 0.357886,
45
+ "f1_weighted": 0.41233
46
+ },
47
+ {
48
+ "accuracy": 0.42665,
49
+ "f1": 0.375691,
50
+ "f1_weighted": 0.439448
51
+ },
52
+ {
53
+ "accuracy": 0.392421,
54
+ "f1": 0.361348,
55
+ "f1_weighted": 0.415867
56
+ },
57
+ {
58
+ "accuracy": 0.433985,
59
+ "f1": 0.398993,
60
+ "f1_weighted": 0.453708
61
+ }
62
+ ],
63
+ "main_score": 0.404279,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "est-Latn"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 170.2254433631897,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/FilipinoShopeeReviewsClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d096f402fdc76886458c0cfb5dedc829bea2b935",
3
+ "task_name": "FilipinoShopeeReviewsClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.369434,
9
+ "f1": 0.357316,
10
+ "f1_weighted": 0.357316,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.367188,
14
+ "f1": 0.360246,
15
+ "f1_weighted": 0.360256
16
+ },
17
+ {
18
+ "accuracy": 0.391113,
19
+ "f1": 0.371352,
20
+ "f1_weighted": 0.371374
21
+ },
22
+ {
23
+ "accuracy": 0.349121,
24
+ "f1": 0.343917,
25
+ "f1_weighted": 0.343924
26
+ },
27
+ {
28
+ "accuracy": 0.356934,
29
+ "f1": 0.352067,
30
+ "f1_weighted": 0.352079
31
+ },
32
+ {
33
+ "accuracy": 0.369629,
34
+ "f1": 0.351581,
35
+ "f1_weighted": 0.351547
36
+ },
37
+ {
38
+ "accuracy": 0.394531,
39
+ "f1": 0.358854,
40
+ "f1_weighted": 0.358812
41
+ },
42
+ {
43
+ "accuracy": 0.354004,
44
+ "f1": 0.352859,
45
+ "f1_weighted": 0.352855
46
+ },
47
+ {
48
+ "accuracy": 0.393555,
49
+ "f1": 0.382621,
50
+ "f1_weighted": 0.382625
51
+ },
52
+ {
53
+ "accuracy": 0.397949,
54
+ "f1": 0.393559,
55
+ "f1_weighted": 0.39354
56
+ },
57
+ {
58
+ "accuracy": 0.320312,
59
+ "f1": 0.306108,
60
+ "f1_weighted": 0.306148
61
+ }
62
+ ],
63
+ "main_score": 0.369434,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "fil-Latn"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 223.02259707450867,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/GerDaLIRSmall.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "48327de6ee192e9610f3069789719788957c7abd",
3
+ "task_name": "GerDaLIRSmall",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.28993,
9
+ "ndcg_at_3": 0.35023,
10
+ "ndcg_at_5": 0.37605,
11
+ "ndcg_at_10": 0.40247,
12
+ "ndcg_at_20": 0.42136,
13
+ "ndcg_at_100": 0.4466,
14
+ "ndcg_at_1000": 0.46353,
15
+ "map_at_1": 0.26187,
16
+ "map_at_3": 0.3265,
17
+ "map_at_5": 0.3421,
18
+ "map_at_10": 0.35362,
19
+ "map_at_20": 0.3592,
20
+ "map_at_100": 0.36299,
21
+ "map_at_1000": 0.36362,
22
+ "recall_at_1": 0.26187,
23
+ "recall_at_3": 0.39603,
24
+ "recall_at_5": 0.45688,
25
+ "recall_at_10": 0.53552,
26
+ "recall_at_20": 0.60657,
27
+ "recall_at_100": 0.73673,
28
+ "recall_at_1000": 0.86878,
29
+ "precision_at_1": 0.28993,
30
+ "precision_at_3": 0.15037,
31
+ "precision_at_5": 0.10517,
32
+ "precision_at_10": 0.06196,
33
+ "precision_at_20": 0.03536,
34
+ "precision_at_100": 0.00864,
35
+ "precision_at_1000": 0.00102,
36
+ "mrr_at_1": 0.28993,
37
+ "mrr_at_3": 0.348251,
38
+ "mrr_at_5": 0.361946,
39
+ "mrr_at_10": 0.372332,
40
+ "mrr_at_20": 0.377237,
41
+ "mrr_at_100": 0.380465,
42
+ "mrr_at_1000": 0.380993,
43
+ "nauc_ndcg_at_1_max": 0.366802,
44
+ "nauc_ndcg_at_1_std": 0.012569,
45
+ "nauc_ndcg_at_1_diff1": 0.501158,
46
+ "nauc_ndcg_at_3_max": 0.36801,
47
+ "nauc_ndcg_at_3_std": 0.025339,
48
+ "nauc_ndcg_at_3_diff1": 0.434079,
49
+ "nauc_ndcg_at_5_max": 0.377851,
50
+ "nauc_ndcg_at_5_std": 0.041135,
51
+ "nauc_ndcg_at_5_diff1": 0.421107,
52
+ "nauc_ndcg_at_10_max": 0.385771,
53
+ "nauc_ndcg_at_10_std": 0.057805,
54
+ "nauc_ndcg_at_10_diff1": 0.413311,
55
+ "nauc_ndcg_at_20_max": 0.391664,
56
+ "nauc_ndcg_at_20_std": 0.073678,
57
+ "nauc_ndcg_at_20_diff1": 0.410514,
58
+ "nauc_ndcg_at_100_max": 0.393259,
59
+ "nauc_ndcg_at_100_std": 0.089138,
60
+ "nauc_ndcg_at_100_diff1": 0.409156,
61
+ "nauc_ndcg_at_1000_max": 0.389774,
62
+ "nauc_ndcg_at_1000_std": 0.084716,
63
+ "nauc_ndcg_at_1000_diff1": 0.412498,
64
+ "nauc_map_at_1_max": 0.337359,
65
+ "nauc_map_at_1_std": -0.005284,
66
+ "nauc_map_at_1_diff1": 0.499537,
67
+ "nauc_map_at_3_max": 0.361528,
68
+ "nauc_map_at_3_std": 0.016623,
69
+ "nauc_map_at_3_diff1": 0.448257,
70
+ "nauc_map_at_5_max": 0.368861,
71
+ "nauc_map_at_5_std": 0.026383,
72
+ "nauc_map_at_5_diff1": 0.440851,
73
+ "nauc_map_at_10_max": 0.372413,
74
+ "nauc_map_at_10_std": 0.033655,
75
+ "nauc_map_at_10_diff1": 0.437387,
76
+ "nauc_map_at_20_max": 0.374379,
77
+ "nauc_map_at_20_std": 0.038501,
78
+ "nauc_map_at_20_diff1": 0.43665,
79
+ "nauc_map_at_100_max": 0.374842,
80
+ "nauc_map_at_100_std": 0.040889,
81
+ "nauc_map_at_100_diff1": 0.436519,
82
+ "nauc_map_at_1000_max": 0.374743,
83
+ "nauc_map_at_1000_std": 0.040826,
84
+ "nauc_map_at_1000_diff1": 0.436642,
85
+ "nauc_recall_at_1_max": 0.337359,
86
+ "nauc_recall_at_1_std": -0.005284,
87
+ "nauc_recall_at_1_diff1": 0.499537,
88
+ "nauc_recall_at_3_max": 0.364535,
89
+ "nauc_recall_at_3_std": 0.036456,
90
+ "nauc_recall_at_3_diff1": 0.385536,
91
+ "nauc_recall_at_5_max": 0.383811,
92
+ "nauc_recall_at_5_std": 0.071233,
93
+ "nauc_recall_at_5_diff1": 0.353547,
94
+ "nauc_recall_at_10_max": 0.403971,
95
+ "nauc_recall_at_10_std": 0.120695,
96
+ "nauc_recall_at_10_diff1": 0.324728,
97
+ "nauc_recall_at_20_max": 0.4274,
98
+ "nauc_recall_at_20_std": 0.187547,
99
+ "nauc_recall_at_20_diff1": 0.306169,
100
+ "nauc_recall_at_100_max": 0.448503,
101
+ "nauc_recall_at_100_std": 0.328052,
102
+ "nauc_recall_at_100_diff1": 0.267646,
103
+ "nauc_recall_at_1000_max": 0.43485,
104
+ "nauc_recall_at_1000_std": 0.427672,
105
+ "nauc_recall_at_1000_diff1": 0.22687,
106
+ "nauc_precision_at_1_max": 0.366802,
107
+ "nauc_precision_at_1_std": 0.012569,
108
+ "nauc_precision_at_1_diff1": 0.501158,
109
+ "nauc_precision_at_3_max": 0.386237,
110
+ "nauc_precision_at_3_std": 0.05932,
111
+ "nauc_precision_at_3_diff1": 0.356113,
112
+ "nauc_precision_at_5_max": 0.394031,
113
+ "nauc_precision_at_5_std": 0.095479,
114
+ "nauc_precision_at_5_diff1": 0.307314,
115
+ "nauc_precision_at_10_max": 0.399909,
116
+ "nauc_precision_at_10_std": 0.142889,
117
+ "nauc_precision_at_10_diff1": 0.263125,
118
+ "nauc_precision_at_20_max": 0.399171,
119
+ "nauc_precision_at_20_std": 0.20115,
120
+ "nauc_precision_at_20_diff1": 0.223118,
121
+ "nauc_precision_at_100_max": 0.354924,
122
+ "nauc_precision_at_100_std": 0.277068,
123
+ "nauc_precision_at_100_diff1": 0.14352,
124
+ "nauc_precision_at_1000_max": 0.25551,
125
+ "nauc_precision_at_1000_std": 0.249122,
126
+ "nauc_precision_at_1000_diff1": 0.056288,
127
+ "nauc_mrr_at_1_max": 0.366802,
128
+ "nauc_mrr_at_1_std": 0.012569,
129
+ "nauc_mrr_at_1_diff1": 0.501158,
130
+ "nauc_mrr_at_3_max": 0.375409,
131
+ "nauc_mrr_at_3_std": 0.029362,
132
+ "nauc_mrr_at_3_diff1": 0.45183,
133
+ "nauc_mrr_at_5_max": 0.378999,
134
+ "nauc_mrr_at_5_std": 0.037505,
135
+ "nauc_mrr_at_5_diff1": 0.444157,
136
+ "nauc_mrr_at_10_max": 0.381929,
137
+ "nauc_mrr_at_10_std": 0.044004,
138
+ "nauc_mrr_at_10_diff1": 0.441292,
139
+ "nauc_mrr_at_20_max": 0.383228,
140
+ "nauc_mrr_at_20_std": 0.047697,
141
+ "nauc_mrr_at_20_diff1": 0.440852,
142
+ "nauc_mrr_at_100_max": 0.383194,
143
+ "nauc_mrr_at_100_std": 0.049224,
144
+ "nauc_mrr_at_100_diff1": 0.440816,
145
+ "nauc_mrr_at_1000_max": 0.38308,
146
+ "nauc_mrr_at_1000_std": 0.049122,
147
+ "nauc_mrr_at_1000_diff1": 0.440945,
148
+ "main_score": 0.40247,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "deu-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 2493.449137210846,
157
+ "kg_co2_emissions": null
158
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/GermanSTSBenchmark.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e36907544d44c3a247898ed81540310442329e20",
3
+ "task_name": "GermanSTSBenchmark",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "pearson": 0.837071,
9
+ "spearman": 0.835312,
10
+ "cosine_pearson": 0.837071,
11
+ "cosine_spearman": 0.83531,
12
+ "manhattan_pearson": 0.821942,
13
+ "manhattan_spearman": 0.835651,
14
+ "euclidean_pearson": 0.822397,
15
+ "euclidean_spearman": 0.835403,
16
+ "main_score": 0.83531,
17
+ "hf_subset": "default",
18
+ "languages": [
19
+ "deu-Latn"
20
+ ]
21
+ }
22
+ ]
23
+ },
24
+ "evaluation_time": 134.46190094947815,
25
+ "kg_co2_emissions": null
26
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/GreekLegalCodeClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "de0fdb34424f07d1ac6f0ede23ee0ed44bd9f5d1",
3
+ "task_name": "GreekLegalCodeClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.562939,
9
+ "f1": 0.493839,
10
+ "f1_weighted": 0.553598,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.569336,
14
+ "f1": 0.499977,
15
+ "f1_weighted": 0.563701
16
+ },
17
+ {
18
+ "accuracy": 0.5625,
19
+ "f1": 0.50368,
20
+ "f1_weighted": 0.548755
21
+ },
22
+ {
23
+ "accuracy": 0.546387,
24
+ "f1": 0.485503,
25
+ "f1_weighted": 0.538462
26
+ },
27
+ {
28
+ "accuracy": 0.5625,
29
+ "f1": 0.484222,
30
+ "f1_weighted": 0.544062
31
+ },
32
+ {
33
+ "accuracy": 0.558594,
34
+ "f1": 0.492295,
35
+ "f1_weighted": 0.549866
36
+ },
37
+ {
38
+ "accuracy": 0.567383,
39
+ "f1": 0.48736,
40
+ "f1_weighted": 0.558717
41
+ },
42
+ {
43
+ "accuracy": 0.577148,
44
+ "f1": 0.511282,
45
+ "f1_weighted": 0.57444
46
+ },
47
+ {
48
+ "accuracy": 0.574707,
49
+ "f1": 0.504846,
50
+ "f1_weighted": 0.567625
51
+ },
52
+ {
53
+ "accuracy": 0.556152,
54
+ "f1": 0.485485,
55
+ "f1_weighted": 0.549158
56
+ },
57
+ {
58
+ "accuracy": 0.554688,
59
+ "f1": 0.483744,
60
+ "f1_weighted": 0.541197
61
+ }
62
+ ],
63
+ "main_score": 0.562939,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "ell-Grek"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 6566.739972114563,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/GujaratiNewsClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1a5f2fa2914bfeff4fcdc6fff4194fa8ec8fa19e",
3
+ "task_name": "GujaratiNewsClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.84742,
9
+ "f1": 0.821249,
10
+ "f1_weighted": 0.848564,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.847496,
14
+ "f1": 0.823527,
15
+ "f1_weighted": 0.851126
16
+ },
17
+ {
18
+ "accuracy": 0.825493,
19
+ "f1": 0.794508,
20
+ "f1_weighted": 0.825887
21
+ },
22
+ {
23
+ "accuracy": 0.820941,
24
+ "f1": 0.789322,
25
+ "f1_weighted": 0.81997
26
+ },
27
+ {
28
+ "accuracy": 0.858877,
29
+ "f1": 0.835056,
30
+ "f1_weighted": 0.858574
31
+ },
32
+ {
33
+ "accuracy": 0.866464,
34
+ "f1": 0.84619,
35
+ "f1_weighted": 0.866331
36
+ },
37
+ {
38
+ "accuracy": 0.879363,
39
+ "f1": 0.857318,
40
+ "f1_weighted": 0.881929
41
+ },
42
+ {
43
+ "accuracy": 0.795903,
44
+ "f1": 0.766223,
45
+ "f1_weighted": 0.796604
46
+ },
47
+ {
48
+ "accuracy": 0.871017,
49
+ "f1": 0.844217,
50
+ "f1_weighted": 0.872135
51
+ },
52
+ {
53
+ "accuracy": 0.864947,
54
+ "f1": 0.837703,
55
+ "f1_weighted": 0.865273
56
+ },
57
+ {
58
+ "accuracy": 0.843703,
59
+ "f1": 0.818423,
60
+ "f1_weighted": 0.847807
61
+ }
62
+ ],
63
+ "main_score": 0.84742,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "guj-Gujr"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 141.02200317382812,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/HALClusteringS2S.v2.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e06ebbbb123f8144bef1a5d18796f3dec9ae2915",
3
+ "task_name": "HALClusteringS2S.v2",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.284519,
11
+ 0.315454,
12
+ 0.296774,
13
+ 0.295918,
14
+ 0.293623,
15
+ 0.308484,
16
+ 0.285998,
17
+ 0.265491,
18
+ 0.286374,
19
+ 0.285411
20
+ ]
21
+ },
22
+ "v_measure": 0.291805,
23
+ "v_measure_std": 0.01316,
24
+ "main_score": 0.291805,
25
+ "hf_subset": "default",
26
+ "languages": [
27
+ "fra-Latn"
28
+ ]
29
+ }
30
+ ]
31
+ },
32
+ "evaluation_time": 87.32846474647522,
33
+ "kg_co2_emissions": null
34
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/IN22GenBitextMining.json ADDED
The diff for this file is too large to render. See raw diff
 
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/IndicCrosslingualSTS.json ADDED
@@ -0,0 +1,203 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "0ca7b87dda68ef4ebb2f50a20a62b9dbebcac3e4",
3
+ "task_name": "IndicCrosslingualSTS",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "pearson": 0.683264,
9
+ "spearman": 0.659694,
10
+ "cosine_pearson": 0.683264,
11
+ "cosine_spearman": 0.659694,
12
+ "manhattan_pearson": 0.660961,
13
+ "manhattan_spearman": 0.65489,
14
+ "euclidean_pearson": 0.665651,
15
+ "euclidean_spearman": 0.659058,
16
+ "main_score": 0.659694,
17
+ "hf_subset": "en-as",
18
+ "languages": [
19
+ "eng-Latn",
20
+ "asm-Beng"
21
+ ]
22
+ },
23
+ {
24
+ "pearson": 0.578617,
25
+ "spearman": 0.568908,
26
+ "cosine_pearson": 0.578617,
27
+ "cosine_spearman": 0.568908,
28
+ "manhattan_pearson": 0.566611,
29
+ "manhattan_spearman": 0.57055,
30
+ "euclidean_pearson": 0.565584,
31
+ "euclidean_spearman": 0.569357,
32
+ "main_score": 0.568908,
33
+ "hf_subset": "en-bn",
34
+ "languages": [
35
+ "eng-Latn",
36
+ "ben-Beng"
37
+ ]
38
+ },
39
+ {
40
+ "pearson": 0.705541,
41
+ "spearman": 0.691557,
42
+ "cosine_pearson": 0.705541,
43
+ "cosine_spearman": 0.691557,
44
+ "manhattan_pearson": 0.667256,
45
+ "manhattan_spearman": 0.689351,
46
+ "euclidean_pearson": 0.669276,
47
+ "euclidean_spearman": 0.691887,
48
+ "main_score": 0.691557,
49
+ "hf_subset": "en-gu",
50
+ "languages": [
51
+ "eng-Latn",
52
+ "guj-Gujr"
53
+ ]
54
+ },
55
+ {
56
+ "pearson": 0.682602,
57
+ "spearman": 0.62348,
58
+ "cosine_pearson": 0.682602,
59
+ "cosine_spearman": 0.62348,
60
+ "manhattan_pearson": 0.667717,
61
+ "manhattan_spearman": 0.629845,
62
+ "euclidean_pearson": 0.662544,
63
+ "euclidean_spearman": 0.622928,
64
+ "main_score": 0.62348,
65
+ "hf_subset": "en-hi",
66
+ "languages": [
67
+ "eng-Latn",
68
+ "hin-Deva"
69
+ ]
70
+ },
71
+ {
72
+ "pearson": 0.608065,
73
+ "spearman": 0.595991,
74
+ "cosine_pearson": 0.608065,
75
+ "cosine_spearman": 0.595991,
76
+ "manhattan_pearson": 0.564157,
77
+ "manhattan_spearman": 0.601641,
78
+ "euclidean_pearson": 0.558749,
79
+ "euclidean_spearman": 0.596287,
80
+ "main_score": 0.595991,
81
+ "hf_subset": "en-kn",
82
+ "languages": [
83
+ "eng-Latn",
84
+ "kan-Knda"
85
+ ]
86
+ },
87
+ {
88
+ "pearson": 0.680191,
89
+ "spearman": 0.670888,
90
+ "cosine_pearson": 0.680191,
91
+ "cosine_spearman": 0.670888,
92
+ "manhattan_pearson": 0.673219,
93
+ "manhattan_spearman": 0.670919,
94
+ "euclidean_pearson": 0.676442,
95
+ "euclidean_spearman": 0.671274,
96
+ "main_score": 0.670888,
97
+ "hf_subset": "en-ml",
98
+ "languages": [
99
+ "eng-Latn",
100
+ "mal-Mlym"
101
+ ]
102
+ },
103
+ {
104
+ "pearson": 0.614148,
105
+ "spearman": 0.617451,
106
+ "cosine_pearson": 0.614148,
107
+ "cosine_spearman": 0.617451,
108
+ "manhattan_pearson": 0.56855,
109
+ "manhattan_spearman": 0.621714,
110
+ "euclidean_pearson": 0.565619,
111
+ "euclidean_spearman": 0.617906,
112
+ "main_score": 0.617451,
113
+ "hf_subset": "en-mr",
114
+ "languages": [
115
+ "eng-Latn",
116
+ "mar-Deva"
117
+ ]
118
+ },
119
+ {
120
+ "pearson": 0.066333,
121
+ "spearman": 0.097857,
122
+ "cosine_pearson": 0.066333,
123
+ "cosine_spearman": 0.097857,
124
+ "manhattan_pearson": 0.065662,
125
+ "manhattan_spearman": 0.093712,
126
+ "euclidean_pearson": 0.070785,
127
+ "euclidean_spearman": 0.098768,
128
+ "main_score": 0.097857,
129
+ "hf_subset": "en-or",
130
+ "languages": [
131
+ "eng-Latn",
132
+ "ory-Orya"
133
+ ]
134
+ },
135
+ {
136
+ "pearson": 0.682219,
137
+ "spearman": 0.643581,
138
+ "cosine_pearson": 0.682219,
139
+ "cosine_spearman": 0.643581,
140
+ "manhattan_pearson": 0.666051,
141
+ "manhattan_spearman": 0.642137,
142
+ "euclidean_pearson": 0.665443,
143
+ "euclidean_spearman": 0.641665,
144
+ "main_score": 0.643581,
145
+ "hf_subset": "en-pa",
146
+ "languages": [
147
+ "eng-Latn",
148
+ "pan-Guru"
149
+ ]
150
+ },
151
+ {
152
+ "pearson": 0.600754,
153
+ "spearman": 0.558303,
154
+ "cosine_pearson": 0.600754,
155
+ "cosine_spearman": 0.558303,
156
+ "manhattan_pearson": 0.586908,
157
+ "manhattan_spearman": 0.55741,
158
+ "euclidean_pearson": 0.583724,
159
+ "euclidean_spearman": 0.557351,
160
+ "main_score": 0.558303,
161
+ "hf_subset": "en-ta",
162
+ "languages": [
163
+ "eng-Latn",
164
+ "tam-Taml"
165
+ ]
166
+ },
167
+ {
168
+ "pearson": 0.594697,
169
+ "spearman": 0.558566,
170
+ "cosine_pearson": 0.594697,
171
+ "cosine_spearman": 0.558566,
172
+ "manhattan_pearson": 0.580914,
173
+ "manhattan_spearman": 0.563604,
174
+ "euclidean_pearson": 0.578287,
175
+ "euclidean_spearman": 0.557491,
176
+ "main_score": 0.558566,
177
+ "hf_subset": "en-te",
178
+ "languages": [
179
+ "eng-Latn",
180
+ "tel-Telu"
181
+ ]
182
+ },
183
+ {
184
+ "pearson": 0.297804,
185
+ "spearman": 0.296898,
186
+ "cosine_pearson": 0.297804,
187
+ "cosine_spearman": 0.296898,
188
+ "manhattan_pearson": 0.296112,
189
+ "manhattan_spearman": 0.294714,
190
+ "euclidean_pearson": 0.298513,
191
+ "euclidean_spearman": 0.296943,
192
+ "main_score": 0.296898,
193
+ "hf_subset": "en-ur",
194
+ "languages": [
195
+ "eng-Latn",
196
+ "urd-Arab"
197
+ ]
198
+ }
199
+ ]
200
+ },
201
+ "evaluation_time": 330.1080505847931,
202
+ "kg_co2_emissions": null
203
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/IndicGenBenchFloresBitextMining.json ADDED
@@ -0,0 +1,707 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f8650438298df086750ff4973661bb58a201a5ee",
3
+ "task_name": "IndicGenBenchFloresBitextMining",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "precision": 1.0,
9
+ "recall": 1.0,
10
+ "f1": 1.0,
11
+ "accuracy": 1.0,
12
+ "main_score": 1.0,
13
+ "hf_subset": "ben-eng",
14
+ "languages": [
15
+ "ben-Beng",
16
+ "eng-Latn"
17
+ ]
18
+ },
19
+ {
20
+ "precision": 0.997036,
21
+ "recall": 0.998024,
22
+ "f1": 0.997365,
23
+ "accuracy": 0.998024,
24
+ "main_score": 0.997365,
25
+ "hf_subset": "eng-ben",
26
+ "languages": [
27
+ "eng-Latn",
28
+ "ben-Beng"
29
+ ]
30
+ },
31
+ {
32
+ "precision": 1.0,
33
+ "recall": 1.0,
34
+ "f1": 1.0,
35
+ "accuracy": 1.0,
36
+ "main_score": 1.0,
37
+ "hf_subset": "guj-eng",
38
+ "languages": [
39
+ "guj-Gujr",
40
+ "eng-Latn"
41
+ ]
42
+ },
43
+ {
44
+ "precision": 0.995553,
45
+ "recall": 0.997036,
46
+ "f1": 0.996047,
47
+ "accuracy": 0.997036,
48
+ "main_score": 0.996047,
49
+ "hf_subset": "eng-guj",
50
+ "languages": [
51
+ "eng-Latn",
52
+ "guj-Gujr"
53
+ ]
54
+ },
55
+ {
56
+ "precision": 1.0,
57
+ "recall": 1.0,
58
+ "f1": 1.0,
59
+ "accuracy": 1.0,
60
+ "main_score": 1.0,
61
+ "hf_subset": "hin-eng",
62
+ "languages": [
63
+ "hin-Deva",
64
+ "eng-Latn"
65
+ ]
66
+ },
67
+ {
68
+ "precision": 0.997036,
69
+ "recall": 0.998024,
70
+ "f1": 0.997365,
71
+ "accuracy": 0.998024,
72
+ "main_score": 0.997365,
73
+ "hf_subset": "eng-hin",
74
+ "languages": [
75
+ "eng-Latn",
76
+ "hin-Deva"
77
+ ]
78
+ },
79
+ {
80
+ "precision": 1.0,
81
+ "recall": 1.0,
82
+ "f1": 1.0,
83
+ "accuracy": 1.0,
84
+ "main_score": 1.0,
85
+ "hf_subset": "kan-eng",
86
+ "languages": [
87
+ "kan-Knda",
88
+ "eng-Latn"
89
+ ]
90
+ },
91
+ {
92
+ "precision": 0.995553,
93
+ "recall": 0.997036,
94
+ "f1": 0.996047,
95
+ "accuracy": 0.997036,
96
+ "main_score": 0.996047,
97
+ "hf_subset": "eng-kan",
98
+ "languages": [
99
+ "eng-Latn",
100
+ "kan-Knda"
101
+ ]
102
+ },
103
+ {
104
+ "precision": 0.998518,
105
+ "recall": 0.999012,
106
+ "f1": 0.998682,
107
+ "accuracy": 0.999012,
108
+ "main_score": 0.998682,
109
+ "hf_subset": "mal-eng",
110
+ "languages": [
111
+ "mal-Mlym",
112
+ "eng-Latn"
113
+ ]
114
+ },
115
+ {
116
+ "precision": 0.992589,
117
+ "recall": 0.995059,
118
+ "f1": 0.993412,
119
+ "accuracy": 0.995059,
120
+ "main_score": 0.993412,
121
+ "hf_subset": "eng-mal",
122
+ "languages": [
123
+ "eng-Latn",
124
+ "mal-Mlym"
125
+ ]
126
+ },
127
+ {
128
+ "precision": 1.0,
129
+ "recall": 1.0,
130
+ "f1": 1.0,
131
+ "accuracy": 1.0,
132
+ "main_score": 1.0,
133
+ "hf_subset": "mar-eng",
134
+ "languages": [
135
+ "mar-Deva",
136
+ "eng-Latn"
137
+ ]
138
+ },
139
+ {
140
+ "precision": 0.994071,
141
+ "recall": 0.996047,
142
+ "f1": 0.99473,
143
+ "accuracy": 0.996047,
144
+ "main_score": 0.99473,
145
+ "hf_subset": "eng-mar",
146
+ "languages": [
147
+ "eng-Latn",
148
+ "mar-Deva"
149
+ ]
150
+ },
151
+ {
152
+ "precision": 1.0,
153
+ "recall": 1.0,
154
+ "f1": 1.0,
155
+ "accuracy": 1.0,
156
+ "main_score": 1.0,
157
+ "hf_subset": "tam-eng",
158
+ "languages": [
159
+ "tam-Taml",
160
+ "eng-Latn"
161
+ ]
162
+ },
163
+ {
164
+ "precision": 0.992589,
165
+ "recall": 0.995059,
166
+ "f1": 0.993412,
167
+ "accuracy": 0.995059,
168
+ "main_score": 0.993412,
169
+ "hf_subset": "eng-tam",
170
+ "languages": [
171
+ "eng-Latn",
172
+ "tam-Taml"
173
+ ]
174
+ },
175
+ {
176
+ "precision": 1.0,
177
+ "recall": 1.0,
178
+ "f1": 1.0,
179
+ "accuracy": 1.0,
180
+ "main_score": 1.0,
181
+ "hf_subset": "tel-eng",
182
+ "languages": [
183
+ "tel-Telu",
184
+ "eng-Latn"
185
+ ]
186
+ },
187
+ {
188
+ "precision": 0.991107,
189
+ "recall": 0.994071,
190
+ "f1": 0.992095,
191
+ "accuracy": 0.994071,
192
+ "main_score": 0.992095,
193
+ "hf_subset": "eng-tel",
194
+ "languages": [
195
+ "eng-Latn",
196
+ "tel-Telu"
197
+ ]
198
+ },
199
+ {
200
+ "precision": 0.995553,
201
+ "recall": 0.997036,
202
+ "f1": 0.996047,
203
+ "accuracy": 0.997036,
204
+ "main_score": 0.996047,
205
+ "hf_subset": "urd-eng",
206
+ "languages": [
207
+ "urd-Arab",
208
+ "eng-Latn"
209
+ ]
210
+ },
211
+ {
212
+ "precision": 0.994071,
213
+ "recall": 0.996047,
214
+ "f1": 0.99473,
215
+ "accuracy": 0.996047,
216
+ "main_score": 0.99473,
217
+ "hf_subset": "eng-urd",
218
+ "languages": [
219
+ "eng-Latn",
220
+ "urd-Arab"
221
+ ]
222
+ },
223
+ {
224
+ "precision": 1.0,
225
+ "recall": 1.0,
226
+ "f1": 1.0,
227
+ "accuracy": 1.0,
228
+ "main_score": 1.0,
229
+ "hf_subset": "asm-eng",
230
+ "languages": [
231
+ "asm-Beng",
232
+ "eng-Latn"
233
+ ]
234
+ },
235
+ {
236
+ "precision": 0.982543,
237
+ "recall": 0.988142,
238
+ "f1": 0.984354,
239
+ "accuracy": 0.988142,
240
+ "main_score": 0.984354,
241
+ "hf_subset": "eng-asm",
242
+ "languages": [
243
+ "eng-Latn",
244
+ "asm-Beng"
245
+ ]
246
+ },
247
+ {
248
+ "precision": 0.995553,
249
+ "recall": 0.997036,
250
+ "f1": 0.996047,
251
+ "accuracy": 0.997036,
252
+ "main_score": 0.996047,
253
+ "hf_subset": "bho-eng",
254
+ "languages": [
255
+ "bho-Deva",
256
+ "eng-Latn"
257
+ ]
258
+ },
259
+ {
260
+ "precision": 0.98666,
261
+ "recall": 0.991107,
262
+ "f1": 0.988142,
263
+ "accuracy": 0.991107,
264
+ "main_score": 0.988142,
265
+ "hf_subset": "eng-bho",
266
+ "languages": [
267
+ "eng-Latn",
268
+ "bho-Deva"
269
+ ]
270
+ },
271
+ {
272
+ "precision": 0.994565,
273
+ "recall": 0.996047,
274
+ "f1": 0.995059,
275
+ "accuracy": 0.996047,
276
+ "main_score": 0.995059,
277
+ "hf_subset": "nep-eng",
278
+ "languages": [
279
+ "nep-Deva",
280
+ "eng-Latn"
281
+ ]
282
+ },
283
+ {
284
+ "precision": 0.987154,
285
+ "recall": 0.991107,
286
+ "f1": 0.988472,
287
+ "accuracy": 0.991107,
288
+ "main_score": 0.988472,
289
+ "hf_subset": "eng-nep",
290
+ "languages": [
291
+ "eng-Latn",
292
+ "nep-Deva"
293
+ ]
294
+ },
295
+ {
296
+ "precision": 1.0,
297
+ "recall": 1.0,
298
+ "f1": 1.0,
299
+ "accuracy": 1.0,
300
+ "main_score": 1.0,
301
+ "hf_subset": "ory-eng",
302
+ "languages": [
303
+ "ory-Orya",
304
+ "eng-Latn"
305
+ ]
306
+ },
307
+ {
308
+ "precision": 0.994071,
309
+ "recall": 0.996047,
310
+ "f1": 0.99473,
311
+ "accuracy": 0.996047,
312
+ "main_score": 0.99473,
313
+ "hf_subset": "eng-ory",
314
+ "languages": [
315
+ "eng-Latn",
316
+ "ory-Orya"
317
+ ]
318
+ },
319
+ {
320
+ "precision": 1.0,
321
+ "recall": 1.0,
322
+ "f1": 1.0,
323
+ "accuracy": 1.0,
324
+ "main_score": 1.0,
325
+ "hf_subset": "pan-eng",
326
+ "languages": [
327
+ "pan-Guru",
328
+ "eng-Latn"
329
+ ]
330
+ },
331
+ {
332
+ "precision": 0.995553,
333
+ "recall": 0.997036,
334
+ "f1": 0.996047,
335
+ "accuracy": 0.997036,
336
+ "main_score": 0.996047,
337
+ "hf_subset": "eng-pan",
338
+ "languages": [
339
+ "eng-Latn",
340
+ "pan-Guru"
341
+ ]
342
+ },
343
+ {
344
+ "precision": 0.959157,
345
+ "recall": 0.972332,
346
+ "f1": 0.963439,
347
+ "accuracy": 0.972332,
348
+ "main_score": 0.963439,
349
+ "hf_subset": "pus-eng",
350
+ "languages": [
351
+ "pus-Arab",
352
+ "eng-Latn"
353
+ ]
354
+ },
355
+ {
356
+ "precision": 0.907856,
357
+ "recall": 0.935771,
358
+ "f1": 0.9167,
359
+ "accuracy": 0.935771,
360
+ "main_score": 0.9167,
361
+ "hf_subset": "eng-pus",
362
+ "languages": [
363
+ "eng-Latn",
364
+ "pus-Arab"
365
+ ]
366
+ },
367
+ {
368
+ "precision": 0.994071,
369
+ "recall": 0.996047,
370
+ "f1": 0.99473,
371
+ "accuracy": 0.996047,
372
+ "main_score": 0.99473,
373
+ "hf_subset": "san-eng",
374
+ "languages": [
375
+ "san-Deva",
376
+ "eng-Latn"
377
+ ]
378
+ },
379
+ {
380
+ "precision": 0.973814,
381
+ "recall": 0.982213,
382
+ "f1": 0.976614,
383
+ "accuracy": 0.982213,
384
+ "main_score": 0.976614,
385
+ "hf_subset": "eng-san",
386
+ "languages": [
387
+ "eng-Latn",
388
+ "san-Deva"
389
+ ]
390
+ },
391
+ {
392
+ "precision": 0.988142,
393
+ "recall": 0.991107,
394
+ "f1": 0.98913,
395
+ "accuracy": 0.991107,
396
+ "main_score": 0.98913,
397
+ "hf_subset": "awa-eng",
398
+ "languages": [
399
+ "awa-Deva",
400
+ "eng-Latn"
401
+ ]
402
+ },
403
+ {
404
+ "precision": 0.981061,
405
+ "recall": 0.986166,
406
+ "f1": 0.982708,
407
+ "accuracy": 0.986166,
408
+ "main_score": 0.982708,
409
+ "hf_subset": "eng-awa",
410
+ "languages": [
411
+ "eng-Latn",
412
+ "awa-Deva"
413
+ ]
414
+ },
415
+ {
416
+ "precision": 1.0,
417
+ "recall": 1.0,
418
+ "f1": 1.0,
419
+ "accuracy": 1.0,
420
+ "main_score": 1.0,
421
+ "hf_subset": "bgc-eng",
422
+ "languages": [
423
+ "bgc-Deva",
424
+ "eng-Latn"
425
+ ]
426
+ },
427
+ {
428
+ "precision": 0.994071,
429
+ "recall": 0.996047,
430
+ "f1": 0.99473,
431
+ "accuracy": 0.996047,
432
+ "main_score": 0.99473,
433
+ "hf_subset": "eng-bgc",
434
+ "languages": [
435
+ "eng-Latn",
436
+ "bgc-Deva"
437
+ ]
438
+ },
439
+ {
440
+ "precision": 0.525323,
441
+ "recall": 0.609684,
442
+ "f1": 0.547728,
443
+ "accuracy": 0.609684,
444
+ "main_score": 0.547728,
445
+ "hf_subset": "bod-eng",
446
+ "languages": [
447
+ "bod-Tibt",
448
+ "eng-Latn"
449
+ ]
450
+ },
451
+ {
452
+ "precision": 0.531532,
453
+ "recall": 0.633399,
454
+ "f1": 0.559663,
455
+ "accuracy": 0.633399,
456
+ "main_score": 0.559663,
457
+ "hf_subset": "eng-bod",
458
+ "languages": [
459
+ "eng-Latn",
460
+ "bod-Tibt"
461
+ ]
462
+ },
463
+ {
464
+ "precision": 0.489188,
465
+ "recall": 0.572134,
466
+ "f1": 0.512236,
467
+ "accuracy": 0.572134,
468
+ "main_score": 0.512236,
469
+ "hf_subset": "boy-eng",
470
+ "languages": [
471
+ "boy-Deva",
472
+ "eng-Latn"
473
+ ]
474
+ },
475
+ {
476
+ "precision": 0.401451,
477
+ "recall": 0.521739,
478
+ "f1": 0.433523,
479
+ "accuracy": 0.521739,
480
+ "main_score": 0.433523,
481
+ "hf_subset": "eng-boy",
482
+ "languages": [
483
+ "eng-Latn",
484
+ "boy-Deva"
485
+ ]
486
+ },
487
+ {
488
+ "precision": 0.997036,
489
+ "recall": 0.998024,
490
+ "f1": 0.997365,
491
+ "accuracy": 0.998024,
492
+ "main_score": 0.997365,
493
+ "hf_subset": "gbm-eng",
494
+ "languages": [
495
+ "gbm-Deva",
496
+ "eng-Latn"
497
+ ]
498
+ },
499
+ {
500
+ "precision": 0.989625,
501
+ "recall": 0.993083,
502
+ "f1": 0.990777,
503
+ "accuracy": 0.993083,
504
+ "main_score": 0.990777,
505
+ "hf_subset": "eng-gbm",
506
+ "languages": [
507
+ "eng-Latn",
508
+ "gbm-Deva"
509
+ ]
510
+ },
511
+ {
512
+ "precision": 0.976285,
513
+ "recall": 0.98419,
514
+ "f1": 0.97892,
515
+ "accuracy": 0.98419,
516
+ "main_score": 0.97892,
517
+ "hf_subset": "gom-eng",
518
+ "languages": [
519
+ "gom-Deva",
520
+ "eng-Latn"
521
+ ]
522
+ },
523
+ {
524
+ "precision": 0.933218,
525
+ "recall": 0.954545,
526
+ "f1": 0.940119,
527
+ "accuracy": 0.954545,
528
+ "main_score": 0.940119,
529
+ "hf_subset": "eng-gom",
530
+ "languages": [
531
+ "eng-Latn",
532
+ "gom-Deva"
533
+ ]
534
+ },
535
+ {
536
+ "precision": 0.992589,
537
+ "recall": 0.995059,
538
+ "f1": 0.993412,
539
+ "accuracy": 0.995059,
540
+ "main_score": 0.993412,
541
+ "hf_subset": "hne-eng",
542
+ "languages": [
543
+ "hne-Deva",
544
+ "eng-Latn"
545
+ ]
546
+ },
547
+ {
548
+ "precision": 0.982872,
549
+ "recall": 0.988142,
550
+ "f1": 0.984519,
551
+ "accuracy": 0.988142,
552
+ "main_score": 0.984519,
553
+ "hf_subset": "eng-hne",
554
+ "languages": [
555
+ "eng-Latn",
556
+ "hne-Deva"
557
+ ]
558
+ },
559
+ {
560
+ "precision": 0.997036,
561
+ "recall": 0.998024,
562
+ "f1": 0.997365,
563
+ "accuracy": 0.998024,
564
+ "main_score": 0.997365,
565
+ "hf_subset": "raj-eng",
566
+ "languages": [
567
+ "raj-Deva",
568
+ "eng-Latn"
569
+ ]
570
+ },
571
+ {
572
+ "precision": 0.989625,
573
+ "recall": 0.993083,
574
+ "f1": 0.990777,
575
+ "accuracy": 0.993083,
576
+ "main_score": 0.990777,
577
+ "hf_subset": "eng-raj",
578
+ "languages": [
579
+ "eng-Latn",
580
+ "raj-Deva"
581
+ ]
582
+ },
583
+ {
584
+ "precision": 0.998518,
585
+ "recall": 0.999012,
586
+ "f1": 0.998682,
587
+ "accuracy": 0.999012,
588
+ "main_score": 0.998682,
589
+ "hf_subset": "mai-eng",
590
+ "languages": [
591
+ "mai-Deva",
592
+ "eng-Latn"
593
+ ]
594
+ },
595
+ {
596
+ "precision": 0.988142,
597
+ "recall": 0.992095,
598
+ "f1": 0.98946,
599
+ "accuracy": 0.992095,
600
+ "main_score": 0.98946,
601
+ "hf_subset": "eng-mai",
602
+ "languages": [
603
+ "eng-Latn",
604
+ "mai-Deva"
605
+ ]
606
+ },
607
+ {
608
+ "precision": 0.636377,
609
+ "recall": 0.712451,
610
+ "f1": 0.657314,
611
+ "accuracy": 0.712451,
612
+ "main_score": 0.657314,
613
+ "hf_subset": "mni-eng",
614
+ "languages": [
615
+ "mni-Mtei",
616
+ "eng-Latn"
617
+ ]
618
+ },
619
+ {
620
+ "precision": 0.560237,
621
+ "recall": 0.661067,
622
+ "f1": 0.588785,
623
+ "accuracy": 0.661067,
624
+ "main_score": 0.588785,
625
+ "hf_subset": "eng-mni",
626
+ "languages": [
627
+ "eng-Latn",
628
+ "mni-Mtei"
629
+ ]
630
+ },
631
+ {
632
+ "precision": 0.997036,
633
+ "recall": 0.998024,
634
+ "f1": 0.997365,
635
+ "accuracy": 0.998024,
636
+ "main_score": 0.997365,
637
+ "hf_subset": "mup-eng",
638
+ "languages": [
639
+ "mup-Deva",
640
+ "eng-Latn"
641
+ ]
642
+ },
643
+ {
644
+ "precision": 0.983696,
645
+ "recall": 0.98913,
646
+ "f1": 0.985507,
647
+ "accuracy": 0.98913,
648
+ "main_score": 0.985507,
649
+ "hf_subset": "eng-mup",
650
+ "languages": [
651
+ "eng-Latn",
652
+ "mup-Deva"
653
+ ]
654
+ },
655
+ {
656
+ "precision": 0.992589,
657
+ "recall": 0.995059,
658
+ "f1": 0.993412,
659
+ "accuracy": 0.995059,
660
+ "main_score": 0.993412,
661
+ "hf_subset": "mwr-eng",
662
+ "languages": [
663
+ "mwr-Deva",
664
+ "eng-Latn"
665
+ ]
666
+ },
667
+ {
668
+ "precision": 0.983037,
669
+ "recall": 0.988142,
670
+ "f1": 0.984684,
671
+ "accuracy": 0.988142,
672
+ "main_score": 0.984684,
673
+ "hf_subset": "eng-mwr",
674
+ "languages": [
675
+ "eng-Latn",
676
+ "mwr-Deva"
677
+ ]
678
+ },
679
+ {
680
+ "precision": 0.049279,
681
+ "recall": 0.081028,
682
+ "f1": 0.054296,
683
+ "accuracy": 0.081028,
684
+ "main_score": 0.054296,
685
+ "hf_subset": "sat-eng",
686
+ "languages": [
687
+ "sat-Olck",
688
+ "eng-Latn"
689
+ ]
690
+ },
691
+ {
692
+ "precision": 0.050783,
693
+ "recall": 0.09585,
694
+ "f1": 0.060186,
695
+ "accuracy": 0.09585,
696
+ "main_score": 0.060186,
697
+ "hf_subset": "eng-sat",
698
+ "languages": [
699
+ "eng-Latn",
700
+ "sat-Olck"
701
+ ]
702
+ }
703
+ ]
704
+ },
705
+ "evaluation_time": 8976.342182159424,
706
+ "kg_co2_emissions": null
707
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/IndicLangClassification.json ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "c54a95d9b9d62c891a03bd5da60715df7176b097",
3
+ "task_name": "IndicLangClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.717273,
9
+ "f1": 0.687329,
10
+ "f1_weighted": 0.707123,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.72559,
14
+ "f1": 0.699677,
15
+ "f1_weighted": 0.714533
16
+ },
17
+ {
18
+ "accuracy": 0.697087,
19
+ "f1": 0.670272,
20
+ "f1_weighted": 0.68476
21
+ },
22
+ {
23
+ "accuracy": 0.716648,
24
+ "f1": 0.6808,
25
+ "f1_weighted": 0.704323
26
+ },
27
+ {
28
+ "accuracy": 0.742948,
29
+ "f1": 0.714278,
30
+ "f1_weighted": 0.732524
31
+ },
32
+ {
33
+ "accuracy": 0.729568,
34
+ "f1": 0.699072,
35
+ "f1_weighted": 0.720338
36
+ },
37
+ {
38
+ "accuracy": 0.724965,
39
+ "f1": 0.699244,
40
+ "f1_weighted": 0.721679
41
+ },
42
+ {
43
+ "accuracy": 0.716747,
44
+ "f1": 0.687523,
45
+ "f1_weighted": 0.705997
46
+ },
47
+ {
48
+ "accuracy": 0.700112,
49
+ "f1": 0.670599,
50
+ "f1_weighted": 0.687215
51
+ },
52
+ {
53
+ "accuracy": 0.702939,
54
+ "f1": 0.671899,
55
+ "f1_weighted": 0.69145
56
+ },
57
+ {
58
+ "accuracy": 0.716122,
59
+ "f1": 0.679922,
60
+ "f1_weighted": 0.708407
61
+ }
62
+ ],
63
+ "main_score": 0.717273,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "asm-Beng",
67
+ "brx-Deva",
68
+ "ben-Beng",
69
+ "doi-Deva",
70
+ "gom-Deva",
71
+ "guj-Gujr",
72
+ "hin-Deva",
73
+ "kan-Knda",
74
+ "kas-Arab",
75
+ "kas-Deva",
76
+ "mai-Deva",
77
+ "mal-Mlym",
78
+ "mar-Deva",
79
+ "mni-Beng",
80
+ "mni-Mtei",
81
+ "npi-Deva",
82
+ "ory-Orya",
83
+ "pan-Guru",
84
+ "san-Deva",
85
+ "sat-Olck",
86
+ "snd-Arab",
87
+ "tam-Taml",
88
+ "tel-Telu",
89
+ "urd-Arab"
90
+ ]
91
+ }
92
+ ]
93
+ },
94
+ "evaluation_time": 2641.103290081024,
95
+ "kg_co2_emissions": null
96
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/ItaCaseholdClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "fafcfc4fee815f7017848e54b26c47ece8ff1626",
3
+ "task_name": "ItaCaseholdClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.724434,
9
+ "f1": 0.390491,
10
+ "f1_weighted": 0.674798,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.714932,
14
+ "f1": 0.369627,
15
+ "f1_weighted": 0.665844
16
+ },
17
+ {
18
+ "accuracy": 0.710407,
19
+ "f1": 0.38356,
20
+ "f1_weighted": 0.665002
21
+ },
22
+ {
23
+ "accuracy": 0.719457,
24
+ "f1": 0.376975,
25
+ "f1_weighted": 0.669053
26
+ },
27
+ {
28
+ "accuracy": 0.714932,
29
+ "f1": 0.391875,
30
+ "f1_weighted": 0.663192
31
+ },
32
+ {
33
+ "accuracy": 0.723982,
34
+ "f1": 0.417943,
35
+ "f1_weighted": 0.676158
36
+ },
37
+ {
38
+ "accuracy": 0.737557,
39
+ "f1": 0.393671,
40
+ "f1_weighted": 0.690079
41
+ },
42
+ {
43
+ "accuracy": 0.705882,
44
+ "f1": 0.375402,
45
+ "f1_weighted": 0.654425
46
+ },
47
+ {
48
+ "accuracy": 0.746606,
49
+ "f1": 0.386543,
50
+ "f1_weighted": 0.697028
51
+ },
52
+ {
53
+ "accuracy": 0.728507,
54
+ "f1": 0.413983,
55
+ "f1_weighted": 0.67273
56
+ },
57
+ {
58
+ "accuracy": 0.742081,
59
+ "f1": 0.395334,
60
+ "f1_weighted": 0.694464
61
+ }
62
+ ],
63
+ "main_score": 0.724434,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "ita-Latn"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 900.3476669788361,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/KorHateSpeechMLClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "c657d15baf277c48d467f0625f7d33c50d4352ef",
3
+ "task_name": "KorHateSpeechMLClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.107855,
9
+ "f1": 0.269836,
10
+ "lrap": 0.249959,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.083947,
14
+ "f1": 0.22588,
15
+ "lrap": 0.224254
16
+ },
17
+ {
18
+ "accuracy": 0.129602,
19
+ "f1": 0.294192,
20
+ "lrap": 0.260159
21
+ },
22
+ {
23
+ "accuracy": 0.108002,
24
+ "f1": 0.258059,
25
+ "lrap": 0.244832
26
+ },
27
+ {
28
+ "accuracy": 0.092293,
29
+ "f1": 0.274239,
30
+ "lrap": 0.250464
31
+ },
32
+ {
33
+ "accuracy": 0.142857,
34
+ "f1": 0.305781,
35
+ "lrap": 0.294142
36
+ },
37
+ {
38
+ "accuracy": 0.127148,
39
+ "f1": 0.306249,
40
+ "lrap": 0.266187
41
+ },
42
+ {
43
+ "accuracy": 0.099165,
44
+ "f1": 0.284102,
45
+ "lrap": 0.2479
46
+ },
47
+ {
48
+ "accuracy": 0.094256,
49
+ "f1": 0.280457,
50
+ "lrap": 0.232913
51
+ },
52
+ {
53
+ "accuracy": 0.108002,
54
+ "f1": 0.245503,
55
+ "lrap": 0.248882
56
+ },
57
+ {
58
+ "accuracy": 0.093274,
59
+ "f1": 0.223892,
60
+ "lrap": 0.229859
61
+ }
62
+ ],
63
+ "main_score": 0.107855,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "kor-Hang"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 122.321542263031,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/KorSarcasmClassification.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3d96e36e10a88d5b7a3f617cf8362d997504494b",
3
+ "task_name": "KorSarcasmClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.540199,
9
+ "f1": 0.537081,
10
+ "f1_weighted": 0.537559,
11
+ "ap": 0.548235,
12
+ "ap_weighted": 0.548235,
13
+ "scores_per_experiment": [
14
+ {
15
+ "accuracy": 0.61794,
16
+ "f1": 0.61794,
17
+ "f1_weighted": 0.61794,
18
+ "ap": 0.598748,
19
+ "ap_weighted": 0.598748
20
+ },
21
+ {
22
+ "accuracy": 0.511628,
23
+ "f1": 0.511283,
24
+ "f1_weighted": 0.51193,
25
+ "ap": 0.530816,
26
+ "ap_weighted": 0.530816
27
+ },
28
+ {
29
+ "accuracy": 0.521595,
30
+ "f1": 0.521463,
31
+ "f1_weighted": 0.521067,
32
+ "ap": 0.537331,
33
+ "ap_weighted": 0.537331
34
+ },
35
+ {
36
+ "accuracy": 0.574751,
37
+ "f1": 0.574709,
38
+ "f1_weighted": 0.57492,
39
+ "ap": 0.568261,
40
+ "ap_weighted": 0.568261
41
+ },
42
+ {
43
+ "accuracy": 0.51495,
44
+ "f1": 0.504845,
45
+ "f1_weighted": 0.50132,
46
+ "ap": 0.537351,
47
+ "ap_weighted": 0.537351
48
+ },
49
+ {
50
+ "accuracy": 0.554817,
51
+ "f1": 0.553036,
52
+ "f1_weighted": 0.554442,
53
+ "ap": 0.553851,
54
+ "ap_weighted": 0.553851
55
+ },
56
+ {
57
+ "accuracy": 0.55814,
58
+ "f1": 0.557964,
59
+ "f1_weighted": 0.558403,
60
+ "ap": 0.557391,
61
+ "ap_weighted": 0.557391
62
+ },
63
+ {
64
+ "accuracy": 0.55814,
65
+ "f1": 0.556887,
66
+ "f1_weighted": 0.558061,
67
+ "ap": 0.556185,
68
+ "ap_weighted": 0.556185
69
+ },
70
+ {
71
+ "accuracy": 0.478405,
72
+ "f1": 0.461059,
73
+ "f1_weighted": 0.465878,
74
+ "ap": 0.510895,
75
+ "ap_weighted": 0.510895
76
+ },
77
+ {
78
+ "accuracy": 0.511628,
79
+ "f1": 0.511628,
80
+ "f1_weighted": 0.511628,
81
+ "ap": 0.531518,
82
+ "ap_weighted": 0.531518
83
+ }
84
+ ],
85
+ "main_score": 0.540199,
86
+ "hf_subset": "default",
87
+ "languages": [
88
+ "kor-Hang"
89
+ ]
90
+ }
91
+ ]
92
+ },
93
+ "evaluation_time": 85.21761751174927,
94
+ "kg_co2_emissions": null
95
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/KurdishSentimentClassification.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f334d90a9f68cc3af78cc2a2ece6a3b69408124c",
3
+ "task_name": "KurdishSentimentClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.749623,
9
+ "f1": 0.746901,
10
+ "f1_weighted": 0.74819,
11
+ "ap": 0.715992,
12
+ "ap_weighted": 0.715992,
13
+ "scores_per_experiment": [
14
+ {
15
+ "accuracy": 0.78158,
16
+ "f1": 0.7814,
17
+ "f1_weighted": 0.780949,
18
+ "ap": 0.766885,
19
+ "ap_weighted": 0.766885
20
+ },
21
+ {
22
+ "accuracy": 0.764972,
23
+ "f1": 0.76497,
24
+ "f1_weighted": 0.764919,
25
+ "ap": 0.742425,
26
+ "ap_weighted": 0.742425
27
+ },
28
+ {
29
+ "accuracy": 0.701057,
30
+ "f1": 0.686866,
31
+ "f1_weighted": 0.691663,
32
+ "ap": 0.654586,
33
+ "ap_weighted": 0.654586
34
+ },
35
+ {
36
+ "accuracy": 0.724711,
37
+ "f1": 0.724328,
38
+ "f1_weighted": 0.725067,
39
+ "ap": 0.695243,
40
+ "ap_weighted": 0.695243
41
+ },
42
+ {
43
+ "accuracy": 0.747861,
44
+ "f1": 0.743936,
45
+ "f1_weighted": 0.746218,
46
+ "ap": 0.703048,
47
+ "ap_weighted": 0.703048
48
+ },
49
+ {
50
+ "accuracy": 0.735783,
51
+ "f1": 0.733693,
52
+ "f1_weighted": 0.735391,
53
+ "ap": 0.697414,
54
+ "ap_weighted": 0.697414
55
+ },
56
+ {
57
+ "accuracy": 0.7539,
58
+ "f1": 0.751909,
59
+ "f1_weighted": 0.753509,
60
+ "ap": 0.713683,
61
+ "ap_weighted": 0.713683
62
+ },
63
+ {
64
+ "accuracy": 0.754404,
65
+ "f1": 0.752614,
66
+ "f1_weighted": 0.754128,
67
+ "ap": 0.714885,
68
+ "ap_weighted": 0.714885
69
+ },
70
+ {
71
+ "accuracy": 0.73377,
72
+ "f1": 0.733337,
73
+ "f1_weighted": 0.732565,
74
+ "ap": 0.718621,
75
+ "ap_weighted": 0.718621
76
+ },
77
+ {
78
+ "accuracy": 0.798188,
79
+ "f1": 0.795952,
80
+ "f1_weighted": 0.797489,
81
+ "ap": 0.753135,
82
+ "ap_weighted": 0.753135
83
+ }
84
+ ],
85
+ "main_score": 0.749623,
86
+ "hf_subset": "default",
87
+ "languages": [
88
+ "kur-Arab"
89
+ ]
90
+ }
91
+ ]
92
+ },
93
+ "evaluation_time": 183.28479409217834,
94
+ "kg_co2_emissions": null
95
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/LeCaRDv2.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "b78e18688c3d012a33dc3676597c1d1b2243ce1c",
3
+ "task_name": "LeCaRDv2",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.90566,
9
+ "ndcg_at_3": 0.8563,
10
+ "ndcg_at_5": 0.82754,
11
+ "ndcg_at_10": 0.77769,
12
+ "ndcg_at_20": 0.72004,
13
+ "ndcg_at_100": 0.85082,
14
+ "ndcg_at_1000": 0.87092,
15
+ "map_at_1": 0.04275,
16
+ "map_at_3": 0.10902,
17
+ "map_at_5": 0.16619,
18
+ "map_at_10": 0.28468,
19
+ "map_at_20": 0.44816,
20
+ "map_at_100": 0.65441,
21
+ "map_at_1000": 0.66177,
22
+ "recall_at_1": 0.04275,
23
+ "recall_at_3": 0.11374,
24
+ "recall_at_5": 0.17873,
25
+ "recall_at_10": 0.32047,
26
+ "recall_at_20": 0.53173,
27
+ "recall_at_100": 0.95236,
28
+ "recall_at_1000": 1.0,
29
+ "precision_at_1": 0.90566,
30
+ "precision_at_3": 0.84486,
31
+ "precision_at_5": 0.80503,
32
+ "precision_at_10": 0.73208,
33
+ "precision_at_20": 0.63113,
34
+ "precision_at_100": 0.23321,
35
+ "precision_at_1000": 0.0245,
36
+ "mrr_at_1": 0.90566,
37
+ "mrr_at_3": 0.936059,
38
+ "mrr_at_5": 0.939203,
39
+ "mrr_at_10": 0.939203,
40
+ "mrr_at_20": 0.939727,
41
+ "mrr_at_100": 0.939853,
42
+ "mrr_at_1000": 0.939853,
43
+ "nauc_ndcg_at_1_max": 0.555651,
44
+ "nauc_ndcg_at_1_std": 0.1188,
45
+ "nauc_ndcg_at_1_diff1": 0.556263,
46
+ "nauc_ndcg_at_3_max": 0.592909,
47
+ "nauc_ndcg_at_3_std": 0.277479,
48
+ "nauc_ndcg_at_3_diff1": 0.392125,
49
+ "nauc_ndcg_at_5_max": 0.577623,
50
+ "nauc_ndcg_at_5_std": 0.234939,
51
+ "nauc_ndcg_at_5_diff1": 0.322851,
52
+ "nauc_ndcg_at_10_max": 0.507429,
53
+ "nauc_ndcg_at_10_std": 0.151749,
54
+ "nauc_ndcg_at_10_diff1": 0.199464,
55
+ "nauc_ndcg_at_20_max": 0.407266,
56
+ "nauc_ndcg_at_20_std": 0.090569,
57
+ "nauc_ndcg_at_20_diff1": 0.050211,
58
+ "nauc_ndcg_at_100_max": 0.478754,
59
+ "nauc_ndcg_at_100_std": 0.175357,
60
+ "nauc_ndcg_at_100_diff1": 0.120565,
61
+ "nauc_ndcg_at_1000_max": 0.500877,
62
+ "nauc_ndcg_at_1000_std": 0.18668,
63
+ "nauc_ndcg_at_1000_diff1": 0.169839,
64
+ "nauc_map_at_1_max": -0.044099,
65
+ "nauc_map_at_1_std": -0.259571,
66
+ "nauc_map_at_1_diff1": 0.056758,
67
+ "nauc_map_at_3_max": 0.208954,
68
+ "nauc_map_at_3_std": -0.123627,
69
+ "nauc_map_at_3_diff1": 0.247089,
70
+ "nauc_map_at_5_max": 0.253218,
71
+ "nauc_map_at_5_std": -0.128806,
72
+ "nauc_map_at_5_diff1": 0.225476,
73
+ "nauc_map_at_10_max": 0.273025,
74
+ "nauc_map_at_10_std": -0.116003,
75
+ "nauc_map_at_10_diff1": 0.146666,
76
+ "nauc_map_at_20_max": 0.298466,
77
+ "nauc_map_at_20_std": -0.052985,
78
+ "nauc_map_at_20_diff1": 0.037653,
79
+ "nauc_map_at_100_max": 0.404329,
80
+ "nauc_map_at_100_std": 0.104945,
81
+ "nauc_map_at_100_diff1": 0.030639,
82
+ "nauc_map_at_1000_max": 0.410294,
83
+ "nauc_map_at_1000_std": 0.110762,
84
+ "nauc_map_at_1000_diff1": 0.037858,
85
+ "nauc_recall_at_1_max": -0.044099,
86
+ "nauc_recall_at_1_std": -0.259571,
87
+ "nauc_recall_at_1_diff1": 0.056758,
88
+ "nauc_recall_at_3_max": 0.161653,
89
+ "nauc_recall_at_3_std": -0.144888,
90
+ "nauc_recall_at_3_diff1": 0.199489,
91
+ "nauc_recall_at_5_max": 0.138494,
92
+ "nauc_recall_at_5_std": -0.183173,
93
+ "nauc_recall_at_5_diff1": 0.129814,
94
+ "nauc_recall_at_10_max": 0.102825,
95
+ "nauc_recall_at_10_std": -0.221947,
96
+ "nauc_recall_at_10_diff1": 0.03943,
97
+ "nauc_recall_at_20_max": 0.131133,
98
+ "nauc_recall_at_20_std": -0.177483,
99
+ "nauc_recall_at_20_diff1": -0.052551,
100
+ "nauc_recall_at_100_max": 0.323564,
101
+ "nauc_recall_at_100_std": 0.144562,
102
+ "nauc_recall_at_100_diff1": -0.099001,
103
+ "nauc_recall_at_1000_max": NaN,
104
+ "nauc_recall_at_1000_std": NaN,
105
+ "nauc_recall_at_1000_diff1": NaN,
106
+ "nauc_precision_at_1_max": 0.555651,
107
+ "nauc_precision_at_1_std": 0.1188,
108
+ "nauc_precision_at_1_diff1": 0.556263,
109
+ "nauc_precision_at_3_max": 0.593484,
110
+ "nauc_precision_at_3_std": 0.310124,
111
+ "nauc_precision_at_3_diff1": 0.357814,
112
+ "nauc_precision_at_5_max": 0.557532,
113
+ "nauc_precision_at_5_std": 0.236226,
114
+ "nauc_precision_at_5_diff1": 0.272158,
115
+ "nauc_precision_at_10_max": 0.457959,
116
+ "nauc_precision_at_10_std": 0.148174,
117
+ "nauc_precision_at_10_diff1": 0.140605,
118
+ "nauc_precision_at_20_max": 0.36654,
119
+ "nauc_precision_at_20_std": 0.158632,
120
+ "nauc_precision_at_20_diff1": 0.013324,
121
+ "nauc_precision_at_100_max": 0.391925,
122
+ "nauc_precision_at_100_std": 0.469545,
123
+ "nauc_precision_at_100_diff1": 0.055541,
124
+ "nauc_precision_at_1000_max": 0.377748,
125
+ "nauc_precision_at_1000_std": 0.513194,
126
+ "nauc_precision_at_1000_diff1": 0.097968,
127
+ "nauc_mrr_at_1_max": 0.555651,
128
+ "nauc_mrr_at_1_std": 0.1188,
129
+ "nauc_mrr_at_1_diff1": 0.556263,
130
+ "nauc_mrr_at_3_max": 0.63395,
131
+ "nauc_mrr_at_3_std": 0.216457,
132
+ "nauc_mrr_at_3_diff1": 0.562933,
133
+ "nauc_mrr_at_5_max": 0.622209,
134
+ "nauc_mrr_at_5_std": 0.206255,
135
+ "nauc_mrr_at_5_diff1": 0.560352,
136
+ "nauc_mrr_at_10_max": 0.622209,
137
+ "nauc_mrr_at_10_std": 0.206255,
138
+ "nauc_mrr_at_10_diff1": 0.560352,
139
+ "nauc_mrr_at_20_max": 0.618924,
140
+ "nauc_mrr_at_20_std": 0.203236,
141
+ "nauc_mrr_at_20_diff1": 0.558947,
142
+ "nauc_mrr_at_100_max": 0.618127,
143
+ "nauc_mrr_at_100_std": 0.20157,
144
+ "nauc_mrr_at_100_diff1": 0.559371,
145
+ "nauc_mrr_at_1000_max": 0.618127,
146
+ "nauc_mrr_at_1000_std": 0.20157,
147
+ "nauc_mrr_at_1000_diff1": 0.559371,
148
+ "main_score": 0.77769,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "zho-Hans"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 1205.0223920345306,
157
+ "kg_co2_emissions": null
158
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/LegalBenchConsumerContractsQA.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "b23590301ec94e8087e2850b21d43d4956b1cca9",
3
+ "task_name": "LegalBenchConsumerContractsQA",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.69949,
9
+ "ndcg_at_3": 0.78749,
10
+ "ndcg_at_5": 0.81173,
11
+ "ndcg_at_10": 0.82996,
12
+ "ndcg_at_20": 0.83457,
13
+ "ndcg_at_100": 0.83788,
14
+ "ndcg_at_1000": 0.83788,
15
+ "map_at_1": 0.69949,
16
+ "map_at_3": 0.76557,
17
+ "map_at_5": 0.77921,
18
+ "map_at_10": 0.7869,
19
+ "map_at_20": 0.78823,
20
+ "map_at_100": 0.7887,
21
+ "map_at_1000": 0.7887,
22
+ "recall_at_1": 0.69949,
23
+ "recall_at_3": 0.85101,
24
+ "recall_at_5": 0.90909,
25
+ "recall_at_10": 0.96465,
26
+ "recall_at_20": 0.98232,
27
+ "recall_at_100": 1.0,
28
+ "recall_at_1000": 1.0,
29
+ "precision_at_1": 0.69949,
30
+ "precision_at_3": 0.28367,
31
+ "precision_at_5": 0.18182,
32
+ "precision_at_10": 0.09646,
33
+ "precision_at_20": 0.04912,
34
+ "precision_at_100": 0.01,
35
+ "precision_at_1000": 0.001,
36
+ "mrr_at_1": 0.70202,
37
+ "mrr_at_3": 0.766414,
38
+ "mrr_at_5": 0.780051,
39
+ "mrr_at_10": 0.787741,
40
+ "mrr_at_20": 0.789075,
41
+ "mrr_at_100": 0.789545,
42
+ "mrr_at_1000": 0.789545,
43
+ "nauc_ndcg_at_1_max": 0.440236,
44
+ "nauc_ndcg_at_1_std": 0.174511,
45
+ "nauc_ndcg_at_1_diff1": 0.692915,
46
+ "nauc_ndcg_at_3_max": 0.377234,
47
+ "nauc_ndcg_at_3_std": 0.207506,
48
+ "nauc_ndcg_at_3_diff1": 0.616183,
49
+ "nauc_ndcg_at_5_max": 0.3927,
50
+ "nauc_ndcg_at_5_std": 0.229283,
51
+ "nauc_ndcg_at_5_diff1": 0.618892,
52
+ "nauc_ndcg_at_10_max": 0.442882,
53
+ "nauc_ndcg_at_10_std": 0.245269,
54
+ "nauc_ndcg_at_10_diff1": 0.65406,
55
+ "nauc_ndcg_at_20_max": 0.427254,
56
+ "nauc_ndcg_at_20_std": 0.222353,
57
+ "nauc_ndcg_at_20_diff1": 0.653752,
58
+ "nauc_ndcg_at_100_max": 0.418073,
59
+ "nauc_ndcg_at_100_std": 0.211766,
60
+ "nauc_ndcg_at_100_diff1": 0.650215,
61
+ "nauc_ndcg_at_1000_max": 0.418073,
62
+ "nauc_ndcg_at_1000_std": 0.211766,
63
+ "nauc_ndcg_at_1000_diff1": 0.650215,
64
+ "nauc_map_at_1_max": 0.440236,
65
+ "nauc_map_at_1_std": 0.174511,
66
+ "nauc_map_at_1_diff1": 0.692915,
67
+ "nauc_map_at_3_max": 0.396435,
68
+ "nauc_map_at_3_std": 0.19856,
69
+ "nauc_map_at_3_diff1": 0.639708,
70
+ "nauc_map_at_5_max": 0.405326,
71
+ "nauc_map_at_5_std": 0.208256,
72
+ "nauc_map_at_5_diff1": 0.64234,
73
+ "nauc_map_at_10_max": 0.421997,
74
+ "nauc_map_at_10_std": 0.210928,
75
+ "nauc_map_at_10_diff1": 0.654401,
76
+ "nauc_map_at_20_max": 0.418279,
77
+ "nauc_map_at_20_std": 0.205469,
78
+ "nauc_map_at_20_diff1": 0.654412,
79
+ "nauc_map_at_100_max": 0.417208,
80
+ "nauc_map_at_100_std": 0.204462,
81
+ "nauc_map_at_100_diff1": 0.653893,
82
+ "nauc_map_at_1000_max": 0.417208,
83
+ "nauc_map_at_1000_std": 0.204462,
84
+ "nauc_map_at_1000_diff1": 0.653893,
85
+ "nauc_recall_at_1_max": 0.440236,
86
+ "nauc_recall_at_1_std": 0.174511,
87
+ "nauc_recall_at_1_diff1": 0.692915,
88
+ "nauc_recall_at_3_max": 0.295727,
89
+ "nauc_recall_at_3_std": 0.244825,
90
+ "nauc_recall_at_3_diff1": 0.516235,
91
+ "nauc_recall_at_5_max": 0.317249,
92
+ "nauc_recall_at_5_std": 0.367295,
93
+ "nauc_recall_at_5_diff1": 0.472661,
94
+ "nauc_recall_at_10_max": 0.881713,
95
+ "nauc_recall_at_10_std": 0.846663,
96
+ "nauc_recall_at_10_diff1": 0.728203,
97
+ "nauc_recall_at_20_max": 0.819512,
98
+ "nauc_recall_at_20_std": 0.71202,
99
+ "nauc_recall_at_20_diff1": 0.786216,
100
+ "nauc_recall_at_100_max": NaN,
101
+ "nauc_recall_at_100_std": NaN,
102
+ "nauc_recall_at_100_diff1": NaN,
103
+ "nauc_recall_at_1000_max": NaN,
104
+ "nauc_recall_at_1000_std": NaN,
105
+ "nauc_recall_at_1000_diff1": NaN,
106
+ "nauc_precision_at_1_max": 0.440236,
107
+ "nauc_precision_at_1_std": 0.174511,
108
+ "nauc_precision_at_1_diff1": 0.692915,
109
+ "nauc_precision_at_3_max": 0.295727,
110
+ "nauc_precision_at_3_std": 0.244825,
111
+ "nauc_precision_at_3_diff1": 0.516235,
112
+ "nauc_precision_at_5_max": 0.317249,
113
+ "nauc_precision_at_5_std": 0.367295,
114
+ "nauc_precision_at_5_diff1": 0.472661,
115
+ "nauc_precision_at_10_max": 0.881713,
116
+ "nauc_precision_at_10_std": 0.846663,
117
+ "nauc_precision_at_10_diff1": 0.728203,
118
+ "nauc_precision_at_20_max": 0.819512,
119
+ "nauc_precision_at_20_std": 0.71202,
120
+ "nauc_precision_at_20_diff1": 0.786216,
121
+ "nauc_precision_at_100_max": NaN,
122
+ "nauc_precision_at_100_std": NaN,
123
+ "nauc_precision_at_100_diff1": NaN,
124
+ "nauc_precision_at_1000_max": NaN,
125
+ "nauc_precision_at_1000_std": NaN,
126
+ "nauc_precision_at_1000_diff1": NaN,
127
+ "nauc_mrr_at_1_max": 0.437923,
128
+ "nauc_mrr_at_1_std": 0.146238,
129
+ "nauc_mrr_at_1_diff1": 0.68802,
130
+ "nauc_mrr_at_3_max": 0.395991,
131
+ "nauc_mrr_at_3_std": 0.181645,
132
+ "nauc_mrr_at_3_diff1": 0.637539,
133
+ "nauc_mrr_at_5_max": 0.404882,
134
+ "nauc_mrr_at_5_std": 0.190328,
135
+ "nauc_mrr_at_5_diff1": 0.640045,
136
+ "nauc_mrr_at_10_max": 0.421589,
137
+ "nauc_mrr_at_10_std": 0.19228,
138
+ "nauc_mrr_at_10_diff1": 0.652052,
139
+ "nauc_mrr_at_20_max": 0.417856,
140
+ "nauc_mrr_at_20_std": 0.18667,
141
+ "nauc_mrr_at_20_diff1": 0.652046,
142
+ "nauc_mrr_at_100_max": 0.41678,
143
+ "nauc_mrr_at_100_std": 0.185612,
144
+ "nauc_mrr_at_100_diff1": 0.651519,
145
+ "nauc_mrr_at_1000_max": 0.41678,
146
+ "nauc_mrr_at_1000_std": 0.185612,
147
+ "nauc_mrr_at_1000_diff1": 0.651519,
148
+ "main_score": 0.82996,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "eng-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 40.99130892753601,
157
+ "kg_co2_emissions": null
158
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/LegalBenchCorporateLobbying.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f69691c650464e62546d7f2a4536f8f87c891e38",
3
+ "task_name": "LegalBenchCorporateLobbying",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.88824,
9
+ "ndcg_at_3": 0.95133,
10
+ "ndcg_at_5": 0.95247,
11
+ "ndcg_at_10": 0.95332,
12
+ "ndcg_at_20": 0.95332,
13
+ "ndcg_at_100": 0.95332,
14
+ "ndcg_at_1000": 0.9541,
15
+ "map_at_1": 0.88824,
16
+ "map_at_3": 0.93824,
17
+ "map_at_5": 0.93882,
18
+ "map_at_10": 0.93912,
19
+ "map_at_20": 0.93912,
20
+ "map_at_100": 0.93912,
21
+ "map_at_1000": 0.93915,
22
+ "recall_at_1": 0.88824,
23
+ "recall_at_3": 0.98824,
24
+ "recall_at_5": 0.99118,
25
+ "recall_at_10": 0.99412,
26
+ "recall_at_20": 0.99412,
27
+ "recall_at_100": 0.99412,
28
+ "recall_at_1000": 1.0,
29
+ "precision_at_1": 0.88824,
30
+ "precision_at_3": 0.32941,
31
+ "precision_at_5": 0.19824,
32
+ "precision_at_10": 0.09941,
33
+ "precision_at_20": 0.04971,
34
+ "precision_at_100": 0.00994,
35
+ "precision_at_1000": 0.001,
36
+ "mrr_at_1": 0.888235,
37
+ "mrr_at_3": 0.938235,
38
+ "mrr_at_5": 0.938824,
39
+ "mrr_at_10": 0.939118,
40
+ "mrr_at_20": 0.939118,
41
+ "mrr_at_100": 0.939118,
42
+ "mrr_at_1000": 0.93915,
43
+ "nauc_ndcg_at_1_max": 0.421147,
44
+ "nauc_ndcg_at_1_std": -0.210212,
45
+ "nauc_ndcg_at_1_diff1": 0.886579,
46
+ "nauc_ndcg_at_3_max": 0.460239,
47
+ "nauc_ndcg_at_3_std": -0.295179,
48
+ "nauc_ndcg_at_3_diff1": 0.869232,
49
+ "nauc_ndcg_at_5_max": 0.450329,
50
+ "nauc_ndcg_at_5_std": -0.260178,
51
+ "nauc_ndcg_at_5_diff1": 0.869231,
52
+ "nauc_ndcg_at_10_max": 0.440195,
53
+ "nauc_ndcg_at_10_std": -0.254586,
54
+ "nauc_ndcg_at_10_diff1": 0.86923,
55
+ "nauc_ndcg_at_20_max": 0.440195,
56
+ "nauc_ndcg_at_20_std": -0.254586,
57
+ "nauc_ndcg_at_20_diff1": 0.86923,
58
+ "nauc_ndcg_at_100_max": 0.440195,
59
+ "nauc_ndcg_at_100_std": -0.254586,
60
+ "nauc_ndcg_at_100_diff1": 0.86923,
61
+ "nauc_ndcg_at_1000_max": 0.441291,
62
+ "nauc_ndcg_at_1000_std": -0.25108,
63
+ "nauc_ndcg_at_1000_diff1": 0.872857,
64
+ "nauc_map_at_1_max": 0.421147,
65
+ "nauc_map_at_1_std": -0.210212,
66
+ "nauc_map_at_1_diff1": 0.886579,
67
+ "nauc_map_at_3_max": 0.445579,
68
+ "nauc_map_at_3_std": -0.263315,
69
+ "nauc_map_at_3_diff1": 0.875737,
70
+ "nauc_map_at_5_max": 0.441453,
71
+ "nauc_map_at_5_std": -0.248936,
72
+ "nauc_map_at_5_diff1": 0.8758,
73
+ "nauc_map_at_10_max": 0.43872,
74
+ "nauc_map_at_10_std": -0.247397,
75
+ "nauc_map_at_10_diff1": 0.875832,
76
+ "nauc_map_at_20_max": 0.43872,
77
+ "nauc_map_at_20_std": -0.247397,
78
+ "nauc_map_at_20_diff1": 0.875832,
79
+ "nauc_map_at_100_max": 0.43872,
80
+ "nauc_map_at_100_std": -0.247397,
81
+ "nauc_map_at_100_diff1": 0.875832,
82
+ "nauc_map_at_1000_max": 0.438838,
83
+ "nauc_map_at_1000_std": -0.247085,
84
+ "nauc_map_at_1000_diff1": 0.875997,
85
+ "nauc_recall_at_1_max": 0.421147,
86
+ "nauc_recall_at_1_std": -0.210212,
87
+ "nauc_recall_at_1_diff1": 0.886579,
88
+ "nauc_recall_at_3_max": 0.674486,
89
+ "nauc_recall_at_3_std": -0.760854,
90
+ "nauc_recall_at_3_diff1": 0.77416,
91
+ "nauc_recall_at_5_max": 0.609555,
92
+ "nauc_recall_at_5_std": -0.434485,
93
+ "nauc_recall_at_5_diff1": 0.742453,
94
+ "nauc_recall_at_10_max": 0.414332,
95
+ "nauc_recall_at_10_std": -0.369981,
96
+ "nauc_recall_at_10_diff1": 0.679038,
97
+ "nauc_recall_at_20_max": 0.414332,
98
+ "nauc_recall_at_20_std": -0.369981,
99
+ "nauc_recall_at_20_diff1": 0.679038,
100
+ "nauc_recall_at_100_max": 0.414332,
101
+ "nauc_recall_at_100_std": -0.369981,
102
+ "nauc_recall_at_100_diff1": 0.679038,
103
+ "nauc_recall_at_1000_max": NaN,
104
+ "nauc_recall_at_1000_std": NaN,
105
+ "nauc_recall_at_1000_diff1": NaN,
106
+ "nauc_precision_at_1_max": 0.421147,
107
+ "nauc_precision_at_1_std": -0.210212,
108
+ "nauc_precision_at_1_diff1": 0.886579,
109
+ "nauc_precision_at_3_max": 0.674486,
110
+ "nauc_precision_at_3_std": -0.760854,
111
+ "nauc_precision_at_3_diff1": 0.77416,
112
+ "nauc_precision_at_5_max": 0.609555,
113
+ "nauc_precision_at_5_std": -0.434485,
114
+ "nauc_precision_at_5_diff1": 0.742453,
115
+ "nauc_precision_at_10_max": 0.414332,
116
+ "nauc_precision_at_10_std": -0.369981,
117
+ "nauc_precision_at_10_diff1": 0.679038,
118
+ "nauc_precision_at_20_max": 0.414332,
119
+ "nauc_precision_at_20_std": -0.369981,
120
+ "nauc_precision_at_20_diff1": 0.679038,
121
+ "nauc_precision_at_100_max": 0.414332,
122
+ "nauc_precision_at_100_std": -0.369981,
123
+ "nauc_precision_at_100_diff1": 0.679038,
124
+ "nauc_precision_at_1000_max": 1.0,
125
+ "nauc_precision_at_1000_std": 1.0,
126
+ "nauc_precision_at_1000_diff1": 1.0,
127
+ "nauc_mrr_at_1_max": 0.421147,
128
+ "nauc_mrr_at_1_std": -0.210212,
129
+ "nauc_mrr_at_1_diff1": 0.886579,
130
+ "nauc_mrr_at_3_max": 0.445579,
131
+ "nauc_mrr_at_3_std": -0.263315,
132
+ "nauc_mrr_at_3_diff1": 0.875737,
133
+ "nauc_mrr_at_5_max": 0.441453,
134
+ "nauc_mrr_at_5_std": -0.248936,
135
+ "nauc_mrr_at_5_diff1": 0.8758,
136
+ "nauc_mrr_at_10_max": 0.43872,
137
+ "nauc_mrr_at_10_std": -0.247397,
138
+ "nauc_mrr_at_10_diff1": 0.875832,
139
+ "nauc_mrr_at_20_max": 0.43872,
140
+ "nauc_mrr_at_20_std": -0.247397,
141
+ "nauc_mrr_at_20_diff1": 0.875832,
142
+ "nauc_mrr_at_100_max": 0.43872,
143
+ "nauc_mrr_at_100_std": -0.247397,
144
+ "nauc_mrr_at_100_diff1": 0.875832,
145
+ "nauc_mrr_at_1000_max": 0.438838,
146
+ "nauc_mrr_at_1000_std": -0.247085,
147
+ "nauc_mrr_at_1000_diff1": 0.875997,
148
+ "main_score": 0.95332,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "eng-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 71.93761157989502,
157
+ "kg_co2_emissions": null
158
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/LegalQuAD.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "37aa6cfb01d48960b0f8e3f17d6e3d99bf1ebc3e",
3
+ "task_name": "LegalQuAD",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.52,
9
+ "ndcg_at_3": 0.61178,
10
+ "ndcg_at_5": 0.61996,
11
+ "ndcg_at_10": 0.64761,
12
+ "ndcg_at_20": 0.65919,
13
+ "ndcg_at_100": 0.67916,
14
+ "ndcg_at_1000": 0.68768,
15
+ "map_at_1": 0.52,
16
+ "map_at_3": 0.58833,
17
+ "map_at_5": 0.59283,
18
+ "map_at_10": 0.60434,
19
+ "map_at_20": 0.60762,
20
+ "map_at_100": 0.61021,
21
+ "map_at_1000": 0.61067,
22
+ "recall_at_1": 0.52,
23
+ "recall_at_3": 0.68,
24
+ "recall_at_5": 0.7,
25
+ "recall_at_10": 0.785,
26
+ "recall_at_20": 0.83,
27
+ "recall_at_100": 0.94,
28
+ "recall_at_1000": 1.0,
29
+ "precision_at_1": 0.52,
30
+ "precision_at_3": 0.22667,
31
+ "precision_at_5": 0.14,
32
+ "precision_at_10": 0.0785,
33
+ "precision_at_20": 0.0415,
34
+ "precision_at_100": 0.0094,
35
+ "precision_at_1000": 0.001,
36
+ "mrr_at_1": 0.52,
37
+ "mrr_at_3": 0.588333,
38
+ "mrr_at_5": 0.592833,
39
+ "mrr_at_10": 0.604335,
40
+ "mrr_at_20": 0.607624,
41
+ "mrr_at_100": 0.610206,
42
+ "mrr_at_1000": 0.610669,
43
+ "nauc_ndcg_at_1_max": 0.583812,
44
+ "nauc_ndcg_at_1_std": 0.335458,
45
+ "nauc_ndcg_at_1_diff1": 0.69982,
46
+ "nauc_ndcg_at_3_max": 0.595583,
47
+ "nauc_ndcg_at_3_std": 0.394844,
48
+ "nauc_ndcg_at_3_diff1": 0.657677,
49
+ "nauc_ndcg_at_5_max": 0.583161,
50
+ "nauc_ndcg_at_5_std": 0.385431,
51
+ "nauc_ndcg_at_5_diff1": 0.649423,
52
+ "nauc_ndcg_at_10_max": 0.584675,
53
+ "nauc_ndcg_at_10_std": 0.409743,
54
+ "nauc_ndcg_at_10_diff1": 0.627446,
55
+ "nauc_ndcg_at_20_max": 0.601462,
56
+ "nauc_ndcg_at_20_std": 0.42299,
57
+ "nauc_ndcg_at_20_diff1": 0.640767,
58
+ "nauc_ndcg_at_100_max": 0.597919,
59
+ "nauc_ndcg_at_100_std": 0.414299,
60
+ "nauc_ndcg_at_100_diff1": 0.648435,
61
+ "nauc_ndcg_at_1000_max": 0.593353,
62
+ "nauc_ndcg_at_1000_std": 0.401187,
63
+ "nauc_ndcg_at_1000_diff1": 0.650685,
64
+ "nauc_map_at_1_max": 0.583812,
65
+ "nauc_map_at_1_std": 0.335458,
66
+ "nauc_map_at_1_diff1": 0.69982,
67
+ "nauc_map_at_3_max": 0.592214,
68
+ "nauc_map_at_3_std": 0.378741,
69
+ "nauc_map_at_3_diff1": 0.667851,
70
+ "nauc_map_at_5_max": 0.585697,
71
+ "nauc_map_at_5_std": 0.373739,
72
+ "nauc_map_at_5_diff1": 0.663624,
73
+ "nauc_map_at_10_max": 0.586713,
74
+ "nauc_map_at_10_std": 0.383826,
75
+ "nauc_map_at_10_diff1": 0.655177,
76
+ "nauc_map_at_20_max": 0.591351,
77
+ "nauc_map_at_20_std": 0.387445,
78
+ "nauc_map_at_20_diff1": 0.658927,
79
+ "nauc_map_at_100_max": 0.590595,
80
+ "nauc_map_at_100_std": 0.386361,
81
+ "nauc_map_at_100_diff1": 0.659601,
82
+ "nauc_map_at_1000_max": 0.590422,
83
+ "nauc_map_at_1000_std": 0.385776,
84
+ "nauc_map_at_1000_diff1": 0.659729,
85
+ "nauc_recall_at_1_max": 0.583812,
86
+ "nauc_recall_at_1_std": 0.335458,
87
+ "nauc_recall_at_1_diff1": 0.69982,
88
+ "nauc_recall_at_3_max": 0.606973,
89
+ "nauc_recall_at_3_std": 0.448898,
90
+ "nauc_recall_at_3_diff1": 0.62411,
91
+ "nauc_recall_at_5_max": 0.572239,
92
+ "nauc_recall_at_5_std": 0.424293,
93
+ "nauc_recall_at_5_diff1": 0.599252,
94
+ "nauc_recall_at_10_max": 0.572585,
95
+ "nauc_recall_at_10_std": 0.5323,
96
+ "nauc_recall_at_10_diff1": 0.492416,
97
+ "nauc_recall_at_20_max": 0.674058,
98
+ "nauc_recall_at_20_std": 0.639679,
99
+ "nauc_recall_at_20_diff1": 0.54625,
100
+ "nauc_recall_at_100_max": 0.734127,
101
+ "nauc_recall_at_100_std": 0.792678,
102
+ "nauc_recall_at_100_diff1": 0.586718,
103
+ "nauc_recall_at_1000_max": NaN,
104
+ "nauc_recall_at_1000_std": NaN,
105
+ "nauc_recall_at_1000_diff1": NaN,
106
+ "nauc_precision_at_1_max": 0.583812,
107
+ "nauc_precision_at_1_std": 0.335458,
108
+ "nauc_precision_at_1_diff1": 0.69982,
109
+ "nauc_precision_at_3_max": 0.606973,
110
+ "nauc_precision_at_3_std": 0.448898,
111
+ "nauc_precision_at_3_diff1": 0.62411,
112
+ "nauc_precision_at_5_max": 0.572239,
113
+ "nauc_precision_at_5_std": 0.424293,
114
+ "nauc_precision_at_5_diff1": 0.599252,
115
+ "nauc_precision_at_10_max": 0.572585,
116
+ "nauc_precision_at_10_std": 0.5323,
117
+ "nauc_precision_at_10_diff1": 0.492416,
118
+ "nauc_precision_at_20_max": 0.674058,
119
+ "nauc_precision_at_20_std": 0.639679,
120
+ "nauc_precision_at_20_diff1": 0.54625,
121
+ "nauc_precision_at_100_max": 0.734127,
122
+ "nauc_precision_at_100_std": 0.792678,
123
+ "nauc_precision_at_100_diff1": 0.586718,
124
+ "nauc_precision_at_1000_max": 1.0,
125
+ "nauc_precision_at_1000_std": 1.0,
126
+ "nauc_precision_at_1000_diff1": 1.0,
127
+ "nauc_mrr_at_1_max": 0.583812,
128
+ "nauc_mrr_at_1_std": 0.335458,
129
+ "nauc_mrr_at_1_diff1": 0.69982,
130
+ "nauc_mrr_at_3_max": 0.592214,
131
+ "nauc_mrr_at_3_std": 0.378741,
132
+ "nauc_mrr_at_3_diff1": 0.667851,
133
+ "nauc_mrr_at_5_max": 0.585697,
134
+ "nauc_mrr_at_5_std": 0.373739,
135
+ "nauc_mrr_at_5_diff1": 0.663624,
136
+ "nauc_mrr_at_10_max": 0.586713,
137
+ "nauc_mrr_at_10_std": 0.383826,
138
+ "nauc_mrr_at_10_diff1": 0.655177,
139
+ "nauc_mrr_at_20_max": 0.591351,
140
+ "nauc_mrr_at_20_std": 0.387445,
141
+ "nauc_mrr_at_20_diff1": 0.658927,
142
+ "nauc_mrr_at_100_max": 0.590595,
143
+ "nauc_mrr_at_100_std": 0.386361,
144
+ "nauc_mrr_at_100_diff1": 0.659601,
145
+ "nauc_mrr_at_1000_max": 0.590422,
146
+ "nauc_mrr_at_1000_std": 0.385776,
147
+ "nauc_mrr_at_1000_diff1": 0.659729,
148
+ "main_score": 0.64761,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "deu-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 72.38827276229858,
157
+ "kg_co2_emissions": null
158
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/LegalSummarization.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3bb1a05c66872889662af04c5691c14489cebd72",
3
+ "task_name": "LegalSummarization",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.55634,
9
+ "ndcg_at_3": 0.61943,
10
+ "ndcg_at_5": 0.63337,
11
+ "ndcg_at_10": 0.66211,
12
+ "ndcg_at_20": 0.68842,
13
+ "ndcg_at_100": 0.70974,
14
+ "ndcg_at_1000": 0.71543,
15
+ "map_at_1": 0.48964,
16
+ "map_at_3": 0.57414,
17
+ "map_at_5": 0.58757,
18
+ "map_at_10": 0.60417,
19
+ "map_at_20": 0.6143,
20
+ "map_at_100": 0.61892,
21
+ "map_at_1000": 0.61933,
22
+ "recall_at_1": 0.48964,
23
+ "recall_at_3": 0.66222,
24
+ "recall_at_5": 0.70342,
25
+ "recall_at_10": 0.78256,
26
+ "recall_at_20": 0.8685,
27
+ "recall_at_100": 0.9638,
28
+ "recall_at_1000": 1.0,
29
+ "precision_at_1": 0.55634,
30
+ "precision_at_3": 0.27347,
31
+ "precision_at_5": 0.17887,
32
+ "precision_at_10": 0.10669,
33
+ "precision_at_20": 0.0625,
34
+ "precision_at_100": 0.01472,
35
+ "precision_at_1000": 0.00155,
36
+ "mrr_at_1": 0.556338,
37
+ "mrr_at_3": 0.634977,
38
+ "mrr_at_5": 0.642019,
39
+ "mrr_at_10": 0.651826,
40
+ "mrr_at_20": 0.657045,
41
+ "mrr_at_100": 0.658444,
42
+ "mrr_at_1000": 0.658556,
43
+ "nauc_ndcg_at_1_max": 0.506166,
44
+ "nauc_ndcg_at_1_std": 0.023202,
45
+ "nauc_ndcg_at_1_diff1": 0.675098,
46
+ "nauc_ndcg_at_3_max": 0.562134,
47
+ "nauc_ndcg_at_3_std": 0.02505,
48
+ "nauc_ndcg_at_3_diff1": 0.643581,
49
+ "nauc_ndcg_at_5_max": 0.56945,
50
+ "nauc_ndcg_at_5_std": 0.038567,
51
+ "nauc_ndcg_at_5_diff1": 0.658221,
52
+ "nauc_ndcg_at_10_max": 0.566855,
53
+ "nauc_ndcg_at_10_std": 0.023602,
54
+ "nauc_ndcg_at_10_diff1": 0.656008,
55
+ "nauc_ndcg_at_20_max": 0.573576,
56
+ "nauc_ndcg_at_20_std": 0.043088,
57
+ "nauc_ndcg_at_20_diff1": 0.648914,
58
+ "nauc_ndcg_at_100_max": 0.572947,
59
+ "nauc_ndcg_at_100_std": 0.048256,
60
+ "nauc_ndcg_at_100_diff1": 0.653059,
61
+ "nauc_ndcg_at_1000_max": 0.56991,
62
+ "nauc_ndcg_at_1000_std": 0.044964,
63
+ "nauc_ndcg_at_1000_diff1": 0.655908,
64
+ "nauc_map_at_1_max": 0.522894,
65
+ "nauc_map_at_1_std": -0.005133,
66
+ "nauc_map_at_1_diff1": 0.688744,
67
+ "nauc_map_at_3_max": 0.545406,
68
+ "nauc_map_at_3_std": 0.004167,
69
+ "nauc_map_at_3_diff1": 0.665377,
70
+ "nauc_map_at_5_max": 0.553927,
71
+ "nauc_map_at_5_std": 0.017972,
72
+ "nauc_map_at_5_diff1": 0.67045,
73
+ "nauc_map_at_10_max": 0.561342,
74
+ "nauc_map_at_10_std": 0.012915,
75
+ "nauc_map_at_10_diff1": 0.667445,
76
+ "nauc_map_at_20_max": 0.564254,
77
+ "nauc_map_at_20_std": 0.020369,
78
+ "nauc_map_at_20_diff1": 0.66586,
79
+ "nauc_map_at_100_max": 0.565777,
80
+ "nauc_map_at_100_std": 0.023252,
81
+ "nauc_map_at_100_diff1": 0.667161,
82
+ "nauc_map_at_1000_max": 0.565698,
83
+ "nauc_map_at_1000_std": 0.023259,
84
+ "nauc_map_at_1000_diff1": 0.667352,
85
+ "nauc_recall_at_1_max": 0.522894,
86
+ "nauc_recall_at_1_std": -0.005133,
87
+ "nauc_recall_at_1_diff1": 0.688744,
88
+ "nauc_recall_at_3_max": 0.558523,
89
+ "nauc_recall_at_3_std": 0.016949,
90
+ "nauc_recall_at_3_diff1": 0.597878,
91
+ "nauc_recall_at_5_max": 0.575047,
92
+ "nauc_recall_at_5_std": 0.050273,
93
+ "nauc_recall_at_5_diff1": 0.621811,
94
+ "nauc_recall_at_10_max": 0.55155,
95
+ "nauc_recall_at_10_std": 0.000361,
96
+ "nauc_recall_at_10_diff1": 0.602664,
97
+ "nauc_recall_at_20_max": 0.590293,
98
+ "nauc_recall_at_20_std": 0.093109,
99
+ "nauc_recall_at_20_diff1": 0.531697,
100
+ "nauc_recall_at_100_max": 0.648144,
101
+ "nauc_recall_at_100_std": 0.221584,
102
+ "nauc_recall_at_100_diff1": 0.472957,
103
+ "nauc_recall_at_1000_max": NaN,
104
+ "nauc_recall_at_1000_std": NaN,
105
+ "nauc_recall_at_1000_diff1": NaN,
106
+ "nauc_precision_at_1_max": 0.506166,
107
+ "nauc_precision_at_1_std": 0.023202,
108
+ "nauc_precision_at_1_diff1": 0.675098,
109
+ "nauc_precision_at_3_max": 0.390839,
110
+ "nauc_precision_at_3_std": 0.08782,
111
+ "nauc_precision_at_3_diff1": 0.317072,
112
+ "nauc_precision_at_5_max": 0.333067,
113
+ "nauc_precision_at_5_std": 0.129522,
114
+ "nauc_precision_at_5_diff1": 0.233191,
115
+ "nauc_precision_at_10_max": 0.206733,
116
+ "nauc_precision_at_10_std": 0.102092,
117
+ "nauc_precision_at_10_diff1": 0.017177,
118
+ "nauc_precision_at_20_max": 0.084616,
119
+ "nauc_precision_at_20_std": 0.165351,
120
+ "nauc_precision_at_20_diff1": -0.132399,
121
+ "nauc_precision_at_100_max": -0.042566,
122
+ "nauc_precision_at_100_std": 0.17277,
123
+ "nauc_precision_at_100_diff1": -0.24173,
124
+ "nauc_precision_at_1000_max": -0.087542,
125
+ "nauc_precision_at_1000_std": 0.153671,
126
+ "nauc_precision_at_1000_diff1": -0.266909,
127
+ "nauc_mrr_at_1_max": 0.506166,
128
+ "nauc_mrr_at_1_std": 0.023202,
129
+ "nauc_mrr_at_1_diff1": 0.675098,
130
+ "nauc_mrr_at_3_max": 0.543067,
131
+ "nauc_mrr_at_3_std": 0.04721,
132
+ "nauc_mrr_at_3_diff1": 0.637914,
133
+ "nauc_mrr_at_5_max": 0.547038,
134
+ "nauc_mrr_at_5_std": 0.048991,
135
+ "nauc_mrr_at_5_diff1": 0.648925,
136
+ "nauc_mrr_at_10_max": 0.540656,
137
+ "nauc_mrr_at_10_std": 0.043477,
138
+ "nauc_mrr_at_10_diff1": 0.650459,
139
+ "nauc_mrr_at_20_max": 0.541523,
140
+ "nauc_mrr_at_20_std": 0.046799,
141
+ "nauc_mrr_at_20_diff1": 0.648492,
142
+ "nauc_mrr_at_100_max": 0.541163,
143
+ "nauc_mrr_at_100_std": 0.045897,
144
+ "nauc_mrr_at_100_diff1": 0.649221,
145
+ "nauc_mrr_at_1000_max": 0.541099,
146
+ "nauc_mrr_at_1000_std": 0.045833,
147
+ "nauc_mrr_at_1000_diff1": 0.64936,
148
+ "main_score": 0.66211,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "eng-Latn"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 27.97646951675415,
157
+ "kg_co2_emissions": null
158
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/MLQARetrieval.json ADDED
The diff for this file is too large to render. See raw diff
 
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/MacedonianTweetSentimentClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "957e075ba35e4417ba7837987fd7053a6533a1a2",
3
+ "task_name": "MacedonianTweetSentimentClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.608692,
9
+ "f1": 0.561092,
10
+ "f1_weighted": 0.611467,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.602283,
14
+ "f1": 0.577593,
15
+ "f1_weighted": 0.623658
16
+ },
17
+ {
18
+ "accuracy": 0.618964,
19
+ "f1": 0.549066,
20
+ "f1_weighted": 0.616055
21
+ },
22
+ {
23
+ "accuracy": 0.61194,
24
+ "f1": 0.562834,
25
+ "f1_weighted": 0.609335
26
+ },
27
+ {
28
+ "accuracy": 0.527656,
29
+ "f1": 0.473061,
30
+ "f1_weighted": 0.520736
31
+ },
32
+ {
33
+ "accuracy": 0.660228,
34
+ "f1": 0.62128,
35
+ "f1_weighted": 0.663344
36
+ },
37
+ {
38
+ "accuracy": 0.609306,
39
+ "f1": 0.574209,
40
+ "f1_weighted": 0.620241
41
+ },
42
+ {
43
+ "accuracy": 0.649693,
44
+ "f1": 0.57536,
45
+ "f1_weighted": 0.637228
46
+ },
47
+ {
48
+ "accuracy": 0.550483,
49
+ "f1": 0.526819,
50
+ "f1_weighted": 0.564473
51
+ },
52
+ {
53
+ "accuracy": 0.675154,
54
+ "f1": 0.60793,
55
+ "f1_weighted": 0.666069
56
+ },
57
+ {
58
+ "accuracy": 0.581212,
59
+ "f1": 0.542763,
60
+ "f1_weighted": 0.593537
61
+ }
62
+ ],
63
+ "main_score": 0.608692,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "mkd-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 153.21337985992432,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/MalteseNewsClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "6bb0321659c4f07c4c2176c30c98c971be6571b4",
3
+ "task_name": "MalteseNewsClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.310579,
9
+ "f1": 0.274059,
10
+ "lrap": 0.415694,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.313452,
14
+ "f1": 0.258343,
15
+ "lrap": 0.414355
16
+ },
17
+ {
18
+ "accuracy": 0.323465,
19
+ "f1": 0.289332,
20
+ "lrap": 0.426069
21
+ },
22
+ {
23
+ "accuracy": 0.29212,
24
+ "f1": 0.274935,
25
+ "lrap": 0.398857
26
+ },
27
+ {
28
+ "accuracy": 0.302133,
29
+ "f1": 0.288384,
30
+ "lrap": 0.413554
31
+ },
32
+ {
33
+ "accuracy": 0.308663,
34
+ "f1": 0.289696,
35
+ "lrap": 0.419185
36
+ },
37
+ {
38
+ "accuracy": 0.340009,
39
+ "f1": 0.259881,
40
+ "lrap": 0.450068
41
+ },
42
+ {
43
+ "accuracy": 0.349151,
44
+ "f1": 0.286167,
45
+ "lrap": 0.468343
46
+ },
47
+ {
48
+ "accuracy": 0.290379,
49
+ "f1": 0.254687,
50
+ "lrap": 0.383177
51
+ },
52
+ {
53
+ "accuracy": 0.321289,
54
+ "f1": 0.283232,
55
+ "lrap": 0.428086
56
+ },
57
+ {
58
+ "accuracy": 0.265128,
59
+ "f1": 0.255932,
60
+ "lrap": 0.355246
61
+ }
62
+ ],
63
+ "main_score": 0.310579,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "mlt-Latn"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 311.06097745895386,
72
+ "kg_co2_emissions": null
73
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/MasakhaNEWSClassification.json ADDED
@@ -0,0 +1,1003 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "18193f187b92da67168c655c9973a165ed9593dd",
3
+ "task_name": "MasakhaNEWSClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.856383,
9
+ "f1": 0.852503,
10
+ "f1_weighted": 0.854871,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.893617,
14
+ "f1": 0.8912,
15
+ "f1_weighted": 0.892974
16
+ },
17
+ {
18
+ "accuracy": 0.827128,
19
+ "f1": 0.82133,
20
+ "f1_weighted": 0.824188
21
+ },
22
+ {
23
+ "accuracy": 0.87234,
24
+ "f1": 0.868249,
25
+ "f1_weighted": 0.870728
26
+ },
27
+ {
28
+ "accuracy": 0.835106,
29
+ "f1": 0.830356,
30
+ "f1_weighted": 0.833225
31
+ },
32
+ {
33
+ "accuracy": 0.859043,
34
+ "f1": 0.852573,
35
+ "f1_weighted": 0.856243
36
+ },
37
+ {
38
+ "accuracy": 0.861702,
39
+ "f1": 0.859427,
40
+ "f1_weighted": 0.861357
41
+ },
42
+ {
43
+ "accuracy": 0.816489,
44
+ "f1": 0.811207,
45
+ "f1_weighted": 0.813919
46
+ },
47
+ {
48
+ "accuracy": 0.869681,
49
+ "f1": 0.868583,
50
+ "f1_weighted": 0.869265
51
+ },
52
+ {
53
+ "accuracy": 0.840426,
54
+ "f1": 0.835336,
55
+ "f1_weighted": 0.838983
56
+ },
57
+ {
58
+ "accuracy": 0.888298,
59
+ "f1": 0.886766,
60
+ "f1_weighted": 0.887829
61
+ }
62
+ ],
63
+ "main_score": 0.856383,
64
+ "hf_subset": "amh",
65
+ "languages": [
66
+ "amh-Ethi"
67
+ ]
68
+ },
69
+ {
70
+ "accuracy": 0.795992,
71
+ "f1": 0.786193,
72
+ "f1_weighted": 0.794274,
73
+ "scores_per_experiment": [
74
+ {
75
+ "accuracy": 0.799578,
76
+ "f1": 0.791684,
77
+ "f1_weighted": 0.798849
78
+ },
79
+ {
80
+ "accuracy": 0.805907,
81
+ "f1": 0.79715,
82
+ "f1_weighted": 0.80275
83
+ },
84
+ {
85
+ "accuracy": 0.777426,
86
+ "f1": 0.766743,
87
+ "f1_weighted": 0.777019
88
+ },
89
+ {
90
+ "accuracy": 0.791139,
91
+ "f1": 0.779665,
92
+ "f1_weighted": 0.78885
93
+ },
94
+ {
95
+ "accuracy": 0.811181,
96
+ "f1": 0.802269,
97
+ "f1_weighted": 0.807858
98
+ },
99
+ {
100
+ "accuracy": 0.780591,
101
+ "f1": 0.769794,
102
+ "f1_weighted": 0.779242
103
+ },
104
+ {
105
+ "accuracy": 0.783755,
106
+ "f1": 0.773522,
107
+ "f1_weighted": 0.782395
108
+ },
109
+ {
110
+ "accuracy": 0.795359,
111
+ "f1": 0.785907,
112
+ "f1_weighted": 0.794151
113
+ },
114
+ {
115
+ "accuracy": 0.794304,
116
+ "f1": 0.785258,
117
+ "f1_weighted": 0.793066
118
+ },
119
+ {
120
+ "accuracy": 0.820675,
121
+ "f1": 0.809938,
122
+ "f1_weighted": 0.818559
123
+ }
124
+ ],
125
+ "main_score": 0.795992,
126
+ "hf_subset": "eng",
127
+ "languages": [
128
+ "eng-Latn"
129
+ ]
130
+ },
131
+ {
132
+ "accuracy": 0.761374,
133
+ "f1": 0.723302,
134
+ "f1_weighted": 0.760553,
135
+ "scores_per_experiment": [
136
+ {
137
+ "accuracy": 0.819905,
138
+ "f1": 0.791711,
139
+ "f1_weighted": 0.809793
140
+ },
141
+ {
142
+ "accuracy": 0.777251,
143
+ "f1": 0.731544,
144
+ "f1_weighted": 0.774341
145
+ },
146
+ {
147
+ "accuracy": 0.796209,
148
+ "f1": 0.756884,
149
+ "f1_weighted": 0.79703
150
+ },
151
+ {
152
+ "accuracy": 0.71327,
153
+ "f1": 0.678977,
154
+ "f1_weighted": 0.709257
155
+ },
156
+ {
157
+ "accuracy": 0.779621,
158
+ "f1": 0.736733,
159
+ "f1_weighted": 0.773593
160
+ },
161
+ {
162
+ "accuracy": 0.727488,
163
+ "f1": 0.682118,
164
+ "f1_weighted": 0.727833
165
+ },
166
+ {
167
+ "accuracy": 0.734597,
168
+ "f1": 0.699133,
169
+ "f1_weighted": 0.744341
170
+ },
171
+ {
172
+ "accuracy": 0.741706,
173
+ "f1": 0.710782,
174
+ "f1_weighted": 0.741341
175
+ },
176
+ {
177
+ "accuracy": 0.78673,
178
+ "f1": 0.740842,
179
+ "f1_weighted": 0.780813
180
+ },
181
+ {
182
+ "accuracy": 0.736967,
183
+ "f1": 0.704296,
184
+ "f1_weighted": 0.747191
185
+ }
186
+ ],
187
+ "main_score": 0.761374,
188
+ "hf_subset": "fra",
189
+ "languages": [
190
+ "fra-Latn"
191
+ ]
192
+ },
193
+ {
194
+ "accuracy": 0.798273,
195
+ "f1": 0.781594,
196
+ "f1_weighted": 0.794272,
197
+ "scores_per_experiment": [
198
+ {
199
+ "accuracy": 0.786499,
200
+ "f1": 0.766426,
201
+ "f1_weighted": 0.783893
202
+ },
203
+ {
204
+ "accuracy": 0.805338,
205
+ "f1": 0.790579,
206
+ "f1_weighted": 0.799033
207
+ },
208
+ {
209
+ "accuracy": 0.786499,
210
+ "f1": 0.765348,
211
+ "f1_weighted": 0.779061
212
+ },
213
+ {
214
+ "accuracy": 0.810047,
215
+ "f1": 0.79476,
216
+ "f1_weighted": 0.809527
217
+ },
218
+ {
219
+ "accuracy": 0.766091,
220
+ "f1": 0.744426,
221
+ "f1_weighted": 0.756375
222
+ },
223
+ {
224
+ "accuracy": 0.821036,
225
+ "f1": 0.804882,
226
+ "f1_weighted": 0.821927
227
+ },
228
+ {
229
+ "accuracy": 0.802198,
230
+ "f1": 0.789628,
231
+ "f1_weighted": 0.800308
232
+ },
233
+ {
234
+ "accuracy": 0.783359,
235
+ "f1": 0.767015,
236
+ "f1_weighted": 0.776778
237
+ },
238
+ {
239
+ "accuracy": 0.795918,
240
+ "f1": 0.782624,
241
+ "f1_weighted": 0.792356
242
+ },
243
+ {
244
+ "accuracy": 0.825746,
245
+ "f1": 0.810249,
246
+ "f1_weighted": 0.823458
247
+ }
248
+ ],
249
+ "main_score": 0.798273,
250
+ "hf_subset": "hau",
251
+ "languages": [
252
+ "hau-Latn"
253
+ ]
254
+ },
255
+ {
256
+ "accuracy": 0.730513,
257
+ "f1": 0.709563,
258
+ "f1_weighted": 0.730738,
259
+ "scores_per_experiment": [
260
+ {
261
+ "accuracy": 0.733333,
262
+ "f1": 0.705424,
263
+ "f1_weighted": 0.736694
264
+ },
265
+ {
266
+ "accuracy": 0.746154,
267
+ "f1": 0.728672,
268
+ "f1_weighted": 0.747141
269
+ },
270
+ {
271
+ "accuracy": 0.784615,
272
+ "f1": 0.763362,
273
+ "f1_weighted": 0.78864
274
+ },
275
+ {
276
+ "accuracy": 0.733333,
277
+ "f1": 0.703728,
278
+ "f1_weighted": 0.731702
279
+ },
280
+ {
281
+ "accuracy": 0.7,
282
+ "f1": 0.681087,
283
+ "f1_weighted": 0.701022
284
+ },
285
+ {
286
+ "accuracy": 0.717949,
287
+ "f1": 0.702987,
288
+ "f1_weighted": 0.718219
289
+ },
290
+ {
291
+ "accuracy": 0.694872,
292
+ "f1": 0.672431,
293
+ "f1_weighted": 0.690924
294
+ },
295
+ {
296
+ "accuracy": 0.723077,
297
+ "f1": 0.697391,
298
+ "f1_weighted": 0.717399
299
+ },
300
+ {
301
+ "accuracy": 0.751282,
302
+ "f1": 0.737347,
303
+ "f1_weighted": 0.754802
304
+ },
305
+ {
306
+ "accuracy": 0.720513,
307
+ "f1": 0.703205,
308
+ "f1_weighted": 0.720835
309
+ }
310
+ ],
311
+ "main_score": 0.730513,
312
+ "hf_subset": "ibo",
313
+ "languages": [
314
+ "ibo-Latn"
315
+ ]
316
+ },
317
+ {
318
+ "accuracy": 0.788,
319
+ "f1": 0.782866,
320
+ "f1_weighted": 0.795898,
321
+ "scores_per_experiment": [
322
+ {
323
+ "accuracy": 0.817143,
324
+ "f1": 0.810341,
325
+ "f1_weighted": 0.82613
326
+ },
327
+ {
328
+ "accuracy": 0.76,
329
+ "f1": 0.772497,
330
+ "f1_weighted": 0.768466
331
+ },
332
+ {
333
+ "accuracy": 0.777143,
334
+ "f1": 0.786809,
335
+ "f1_weighted": 0.792228
336
+ },
337
+ {
338
+ "accuracy": 0.84,
339
+ "f1": 0.80706,
340
+ "f1_weighted": 0.838789
341
+ },
342
+ {
343
+ "accuracy": 0.697143,
344
+ "f1": 0.677004,
345
+ "f1_weighted": 0.712144
346
+ },
347
+ {
348
+ "accuracy": 0.771429,
349
+ "f1": 0.780178,
350
+ "f1_weighted": 0.782846
351
+ },
352
+ {
353
+ "accuracy": 0.851429,
354
+ "f1": 0.845194,
355
+ "f1_weighted": 0.851486
356
+ },
357
+ {
358
+ "accuracy": 0.754286,
359
+ "f1": 0.752165,
360
+ "f1_weighted": 0.768389
361
+ },
362
+ {
363
+ "accuracy": 0.794286,
364
+ "f1": 0.790383,
365
+ "f1_weighted": 0.795344
366
+ },
367
+ {
368
+ "accuracy": 0.817143,
369
+ "f1": 0.807032,
370
+ "f1_weighted": 0.823157
371
+ }
372
+ ],
373
+ "main_score": 0.788,
374
+ "hf_subset": "lin",
375
+ "languages": [
376
+ "lin-Latn"
377
+ ]
378
+ },
379
+ {
380
+ "accuracy": 0.75157,
381
+ "f1": 0.718486,
382
+ "f1_weighted": 0.75622,
383
+ "scores_per_experiment": [
384
+ {
385
+ "accuracy": 0.780269,
386
+ "f1": 0.736761,
387
+ "f1_weighted": 0.782013
388
+ },
389
+ {
390
+ "accuracy": 0.753363,
391
+ "f1": 0.742389,
392
+ "f1_weighted": 0.76312
393
+ },
394
+ {
395
+ "accuracy": 0.7713,
396
+ "f1": 0.743835,
397
+ "f1_weighted": 0.773996
398
+ },
399
+ {
400
+ "accuracy": 0.762332,
401
+ "f1": 0.726332,
402
+ "f1_weighted": 0.770705
403
+ },
404
+ {
405
+ "accuracy": 0.780269,
406
+ "f1": 0.747187,
407
+ "f1_weighted": 0.78239
408
+ },
409
+ {
410
+ "accuracy": 0.726457,
411
+ "f1": 0.71137,
412
+ "f1_weighted": 0.732457
413
+ },
414
+ {
415
+ "accuracy": 0.695067,
416
+ "f1": 0.64733,
417
+ "f1_weighted": 0.700944
418
+ },
419
+ {
420
+ "accuracy": 0.730942,
421
+ "f1": 0.686064,
422
+ "f1_weighted": 0.731029
423
+ },
424
+ {
425
+ "accuracy": 0.766816,
426
+ "f1": 0.725719,
427
+ "f1_weighted": 0.770087
428
+ },
429
+ {
430
+ "accuracy": 0.748879,
431
+ "f1": 0.71787,
432
+ "f1_weighted": 0.755462
433
+ }
434
+ ],
435
+ "main_score": 0.75157,
436
+ "hf_subset": "lug",
437
+ "languages": [
438
+ "lug-Latn"
439
+ ]
440
+ },
441
+ {
442
+ "accuracy": 0.825846,
443
+ "f1": 0.776165,
444
+ "f1_weighted": 0.832668,
445
+ "scores_per_experiment": [
446
+ {
447
+ "accuracy": 0.846154,
448
+ "f1": 0.786666,
449
+ "f1_weighted": 0.850128
450
+ },
451
+ {
452
+ "accuracy": 0.849231,
453
+ "f1": 0.789171,
454
+ "f1_weighted": 0.848755
455
+ },
456
+ {
457
+ "accuracy": 0.775385,
458
+ "f1": 0.73202,
459
+ "f1_weighted": 0.79055
460
+ },
461
+ {
462
+ "accuracy": 0.818462,
463
+ "f1": 0.783911,
464
+ "f1_weighted": 0.823707
465
+ },
466
+ {
467
+ "accuracy": 0.830769,
468
+ "f1": 0.799774,
469
+ "f1_weighted": 0.836622
470
+ },
471
+ {
472
+ "accuracy": 0.833846,
473
+ "f1": 0.78577,
474
+ "f1_weighted": 0.839326
475
+ },
476
+ {
477
+ "accuracy": 0.8,
478
+ "f1": 0.74816,
479
+ "f1_weighted": 0.815062
480
+ },
481
+ {
482
+ "accuracy": 0.8,
483
+ "f1": 0.727781,
484
+ "f1_weighted": 0.811875
485
+ },
486
+ {
487
+ "accuracy": 0.876923,
488
+ "f1": 0.82096,
489
+ "f1_weighted": 0.879211
490
+ },
491
+ {
492
+ "accuracy": 0.827692,
493
+ "f1": 0.787432,
494
+ "f1_weighted": 0.831438
495
+ }
496
+ ],
497
+ "main_score": 0.825846,
498
+ "hf_subset": "orm",
499
+ "languages": [
500
+ "orm-Ethi"
501
+ ]
502
+ },
503
+ {
504
+ "accuracy": 0.932459,
505
+ "f1": 0.899354,
506
+ "f1_weighted": 0.933408,
507
+ "scores_per_experiment": [
508
+ {
509
+ "accuracy": 0.921311,
510
+ "f1": 0.849315,
511
+ "f1_weighted": 0.923014
512
+ },
513
+ {
514
+ "accuracy": 0.914754,
515
+ "f1": 0.879243,
516
+ "f1_weighted": 0.916603
517
+ },
518
+ {
519
+ "accuracy": 0.95082,
520
+ "f1": 0.930873,
521
+ "f1_weighted": 0.951433
522
+ },
523
+ {
524
+ "accuracy": 0.960656,
525
+ "f1": 0.955732,
526
+ "f1_weighted": 0.960621
527
+ },
528
+ {
529
+ "accuracy": 0.947541,
530
+ "f1": 0.929022,
531
+ "f1_weighted": 0.947583
532
+ },
533
+ {
534
+ "accuracy": 0.944262,
535
+ "f1": 0.895088,
536
+ "f1_weighted": 0.944798
537
+ },
538
+ {
539
+ "accuracy": 0.92459,
540
+ "f1": 0.895827,
541
+ "f1_weighted": 0.926061
542
+ },
543
+ {
544
+ "accuracy": 0.914754,
545
+ "f1": 0.876787,
546
+ "f1_weighted": 0.916207
547
+ },
548
+ {
549
+ "accuracy": 0.921311,
550
+ "f1": 0.892526,
551
+ "f1_weighted": 0.921627
552
+ },
553
+ {
554
+ "accuracy": 0.92459,
555
+ "f1": 0.88913,
556
+ "f1_weighted": 0.926133
557
+ }
558
+ ],
559
+ "main_score": 0.932459,
560
+ "hf_subset": "pcm",
561
+ "languages": [
562
+ "pcm-Latn"
563
+ ]
564
+ },
565
+ {
566
+ "accuracy": 0.787888,
567
+ "f1": 0.702727,
568
+ "f1_weighted": 0.800884,
569
+ "scores_per_experiment": [
570
+ {
571
+ "accuracy": 0.754658,
572
+ "f1": 0.694294,
573
+ "f1_weighted": 0.765403
574
+ },
575
+ {
576
+ "accuracy": 0.81677,
577
+ "f1": 0.732346,
578
+ "f1_weighted": 0.834839
579
+ },
580
+ {
581
+ "accuracy": 0.807453,
582
+ "f1": 0.693552,
583
+ "f1_weighted": 0.81437
584
+ },
585
+ {
586
+ "accuracy": 0.770186,
587
+ "f1": 0.689711,
588
+ "f1_weighted": 0.787042
589
+ },
590
+ {
591
+ "accuracy": 0.795031,
592
+ "f1": 0.70602,
593
+ "f1_weighted": 0.812326
594
+ },
595
+ {
596
+ "accuracy": 0.795031,
597
+ "f1": 0.693367,
598
+ "f1_weighted": 0.803627
599
+ },
600
+ {
601
+ "accuracy": 0.791925,
602
+ "f1": 0.712968,
603
+ "f1_weighted": 0.803219
604
+ },
605
+ {
606
+ "accuracy": 0.757764,
607
+ "f1": 0.692483,
608
+ "f1_weighted": 0.778772
609
+ },
610
+ {
611
+ "accuracy": 0.785714,
612
+ "f1": 0.711774,
613
+ "f1_weighted": 0.804065
614
+ },
615
+ {
616
+ "accuracy": 0.804348,
617
+ "f1": 0.700752,
618
+ "f1_weighted": 0.805179
619
+ }
620
+ ],
621
+ "main_score": 0.787888,
622
+ "hf_subset": "run",
623
+ "languages": [
624
+ "run-Latn"
625
+ ]
626
+ },
627
+ {
628
+ "accuracy": 0.885637,
629
+ "f1": 0.888575,
630
+ "f1_weighted": 0.885092,
631
+ "scores_per_experiment": [
632
+ {
633
+ "accuracy": 0.910569,
634
+ "f1": 0.912072,
635
+ "f1_weighted": 0.910073
636
+ },
637
+ {
638
+ "accuracy": 0.886179,
639
+ "f1": 0.890975,
640
+ "f1_weighted": 0.886897
641
+ },
642
+ {
643
+ "accuracy": 0.850949,
644
+ "f1": 0.857245,
645
+ "f1_weighted": 0.850854
646
+ },
647
+ {
648
+ "accuracy": 0.910569,
649
+ "f1": 0.91116,
650
+ "f1_weighted": 0.909782
651
+ },
652
+ {
653
+ "accuracy": 0.905149,
654
+ "f1": 0.906395,
655
+ "f1_weighted": 0.903722
656
+ },
657
+ {
658
+ "accuracy": 0.869919,
659
+ "f1": 0.873018,
660
+ "f1_weighted": 0.869381
661
+ },
662
+ {
663
+ "accuracy": 0.894309,
664
+ "f1": 0.899006,
665
+ "f1_weighted": 0.89461
666
+ },
667
+ {
668
+ "accuracy": 0.875339,
669
+ "f1": 0.876877,
670
+ "f1_weighted": 0.873899
671
+ },
672
+ {
673
+ "accuracy": 0.864499,
674
+ "f1": 0.867668,
675
+ "f1_weighted": 0.863825
676
+ },
677
+ {
678
+ "accuracy": 0.888889,
679
+ "f1": 0.891332,
680
+ "f1_weighted": 0.887875
681
+ }
682
+ ],
683
+ "main_score": 0.885637,
684
+ "hf_subset": "sna",
685
+ "languages": [
686
+ "sna-Latn"
687
+ ]
688
+ },
689
+ {
690
+ "accuracy": 0.657483,
691
+ "f1": 0.618914,
692
+ "f1_weighted": 0.669745,
693
+ "scores_per_experiment": [
694
+ {
695
+ "accuracy": 0.639456,
696
+ "f1": 0.601967,
697
+ "f1_weighted": 0.648535
698
+ },
699
+ {
700
+ "accuracy": 0.656463,
701
+ "f1": 0.593663,
702
+ "f1_weighted": 0.662965
703
+ },
704
+ {
705
+ "accuracy": 0.673469,
706
+ "f1": 0.653348,
707
+ "f1_weighted": 0.685273
708
+ },
709
+ {
710
+ "accuracy": 0.687075,
711
+ "f1": 0.647539,
712
+ "f1_weighted": 0.704695
713
+ },
714
+ {
715
+ "accuracy": 0.659864,
716
+ "f1": 0.617378,
717
+ "f1_weighted": 0.678788
718
+ },
719
+ {
720
+ "accuracy": 0.62585,
721
+ "f1": 0.582826,
722
+ "f1_weighted": 0.642885
723
+ },
724
+ {
725
+ "accuracy": 0.680272,
726
+ "f1": 0.634444,
727
+ "f1_weighted": 0.692679
728
+ },
729
+ {
730
+ "accuracy": 0.656463,
731
+ "f1": 0.62666,
732
+ "f1_weighted": 0.65921
733
+ },
734
+ {
735
+ "accuracy": 0.612245,
736
+ "f1": 0.595345,
737
+ "f1_weighted": 0.62603
738
+ },
739
+ {
740
+ "accuracy": 0.683673,
741
+ "f1": 0.635973,
742
+ "f1_weighted": 0.696388
743
+ }
744
+ ],
745
+ "main_score": 0.657483,
746
+ "hf_subset": "som",
747
+ "languages": [
748
+ "som-Latn"
749
+ ]
750
+ },
751
+ {
752
+ "accuracy": 0.742437,
753
+ "f1": 0.692728,
754
+ "f1_weighted": 0.748332,
755
+ "scores_per_experiment": [
756
+ {
757
+ "accuracy": 0.741597,
758
+ "f1": 0.683676,
759
+ "f1_weighted": 0.750604
760
+ },
761
+ {
762
+ "accuracy": 0.739496,
763
+ "f1": 0.702375,
764
+ "f1_weighted": 0.747463
765
+ },
766
+ {
767
+ "accuracy": 0.754202,
768
+ "f1": 0.698329,
769
+ "f1_weighted": 0.756778
770
+ },
771
+ {
772
+ "accuracy": 0.720588,
773
+ "f1": 0.655741,
774
+ "f1_weighted": 0.726057
775
+ },
776
+ {
777
+ "accuracy": 0.766807,
778
+ "f1": 0.716732,
779
+ "f1_weighted": 0.774348
780
+ },
781
+ {
782
+ "accuracy": 0.682773,
783
+ "f1": 0.625976,
784
+ "f1_weighted": 0.68613
785
+ },
786
+ {
787
+ "accuracy": 0.760504,
788
+ "f1": 0.710471,
789
+ "f1_weighted": 0.767553
790
+ },
791
+ {
792
+ "accuracy": 0.741597,
793
+ "f1": 0.697043,
794
+ "f1_weighted": 0.748169
795
+ },
796
+ {
797
+ "accuracy": 0.754202,
798
+ "f1": 0.716088,
799
+ "f1_weighted": 0.761159
800
+ },
801
+ {
802
+ "accuracy": 0.762605,
803
+ "f1": 0.720847,
804
+ "f1_weighted": 0.765058
805
+ }
806
+ ],
807
+ "main_score": 0.742437,
808
+ "hf_subset": "swa",
809
+ "languages": [
810
+ "swa-Latn"
811
+ ]
812
+ },
813
+ {
814
+ "accuracy": 0.717279,
815
+ "f1": 0.687201,
816
+ "f1_weighted": 0.728623,
817
+ "scores_per_experiment": [
818
+ {
819
+ "accuracy": 0.742647,
820
+ "f1": 0.70081,
821
+ "f1_weighted": 0.751797
822
+ },
823
+ {
824
+ "accuracy": 0.709559,
825
+ "f1": 0.694279,
826
+ "f1_weighted": 0.714971
827
+ },
828
+ {
829
+ "accuracy": 0.731618,
830
+ "f1": 0.717875,
831
+ "f1_weighted": 0.745115
832
+ },
833
+ {
834
+ "accuracy": 0.713235,
835
+ "f1": 0.68306,
836
+ "f1_weighted": 0.725585
837
+ },
838
+ {
839
+ "accuracy": 0.779412,
840
+ "f1": 0.757559,
841
+ "f1_weighted": 0.784372
842
+ },
843
+ {
844
+ "accuracy": 0.665441,
845
+ "f1": 0.629779,
846
+ "f1_weighted": 0.681255
847
+ },
848
+ {
849
+ "accuracy": 0.716912,
850
+ "f1": 0.679172,
851
+ "f1_weighted": 0.725512
852
+ },
853
+ {
854
+ "accuracy": 0.680147,
855
+ "f1": 0.657624,
856
+ "f1_weighted": 0.699742
857
+ },
858
+ {
859
+ "accuracy": 0.727941,
860
+ "f1": 0.682614,
861
+ "f1_weighted": 0.743126
862
+ },
863
+ {
864
+ "accuracy": 0.705882,
865
+ "f1": 0.669233,
866
+ "f1_weighted": 0.714755
867
+ }
868
+ ],
869
+ "main_score": 0.717279,
870
+ "hf_subset": "tir",
871
+ "languages": [
872
+ "tir-Ethi"
873
+ ]
874
+ },
875
+ {
876
+ "accuracy": 0.87037,
877
+ "f1": 0.784756,
878
+ "f1_weighted": 0.87888,
879
+ "scores_per_experiment": [
880
+ {
881
+ "accuracy": 0.878788,
882
+ "f1": 0.793872,
883
+ "f1_weighted": 0.879955
884
+ },
885
+ {
886
+ "accuracy": 0.875421,
887
+ "f1": 0.7855,
888
+ "f1_weighted": 0.890813
889
+ },
890
+ {
891
+ "accuracy": 0.86532,
892
+ "f1": 0.779881,
893
+ "f1_weighted": 0.868837
894
+ },
895
+ {
896
+ "accuracy": 0.885522,
897
+ "f1": 0.823055,
898
+ "f1_weighted": 0.894773
899
+ },
900
+ {
901
+ "accuracy": 0.855219,
902
+ "f1": 0.755573,
903
+ "f1_weighted": 0.867076
904
+ },
905
+ {
906
+ "accuracy": 0.878788,
907
+ "f1": 0.791315,
908
+ "f1_weighted": 0.882022
909
+ },
910
+ {
911
+ "accuracy": 0.851852,
912
+ "f1": 0.754116,
913
+ "f1_weighted": 0.864692
914
+ },
915
+ {
916
+ "accuracy": 0.868687,
917
+ "f1": 0.776279,
918
+ "f1_weighted": 0.88225
919
+ },
920
+ {
921
+ "accuracy": 0.861953,
922
+ "f1": 0.772146,
923
+ "f1_weighted": 0.871506
924
+ },
925
+ {
926
+ "accuracy": 0.882155,
927
+ "f1": 0.815824,
928
+ "f1_weighted": 0.886878
929
+ }
930
+ ],
931
+ "main_score": 0.87037,
932
+ "hf_subset": "xho",
933
+ "languages": [
934
+ "xho-Latn"
935
+ ]
936
+ },
937
+ {
938
+ "accuracy": 0.853771,
939
+ "f1": 0.852724,
940
+ "f1_weighted": 0.85399,
941
+ "scores_per_experiment": [
942
+ {
943
+ "accuracy": 0.851582,
944
+ "f1": 0.84844,
945
+ "f1_weighted": 0.851253
946
+ },
947
+ {
948
+ "accuracy": 0.841849,
949
+ "f1": 0.840122,
950
+ "f1_weighted": 0.842798
951
+ },
952
+ {
953
+ "accuracy": 0.841849,
954
+ "f1": 0.842109,
955
+ "f1_weighted": 0.841234
956
+ },
957
+ {
958
+ "accuracy": 0.856448,
959
+ "f1": 0.857066,
960
+ "f1_weighted": 0.857451
961
+ },
962
+ {
963
+ "accuracy": 0.824818,
964
+ "f1": 0.82431,
965
+ "f1_weighted": 0.826693
966
+ },
967
+ {
968
+ "accuracy": 0.871046,
969
+ "f1": 0.871219,
970
+ "f1_weighted": 0.871868
971
+ },
972
+ {
973
+ "accuracy": 0.841849,
974
+ "f1": 0.840184,
975
+ "f1_weighted": 0.84123
976
+ },
977
+ {
978
+ "accuracy": 0.856448,
979
+ "f1": 0.852886,
980
+ "f1_weighted": 0.855635
981
+ },
982
+ {
983
+ "accuracy": 0.878345,
984
+ "f1": 0.880304,
985
+ "f1_weighted": 0.879114
986
+ },
987
+ {
988
+ "accuracy": 0.873479,
989
+ "f1": 0.870603,
990
+ "f1_weighted": 0.872625
991
+ }
992
+ ],
993
+ "main_score": 0.853771,
994
+ "hf_subset": "yor",
995
+ "languages": [
996
+ "yor-Latn"
997
+ ]
998
+ }
999
+ ]
1000
+ },
1001
+ "evaluation_time": 2903.7933003902435,
1002
+ "kg_co2_emissions": null
1003
+ }
results/Mira190__Euler-Legal-Embedding-V1/df607ed9e25e569514a99c27cdaaab16e76b6dd4/MassiveIntentClassification.json ADDED
@@ -0,0 +1,3173 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "4672e20407010da34463acc759c162ca9734bca6",
3
+ "task_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.38.56",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.723268,
9
+ "f1": 0.689512,
10
+ "f1_weighted": 0.71796,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.738399,
14
+ "f1": 0.702663,
15
+ "f1_weighted": 0.730207
16
+ },
17
+ {
18
+ "accuracy": 0.742771,
19
+ "f1": 0.713886,
20
+ "f1_weighted": 0.741266
21
+ },
22
+ {
23
+ "accuracy": 0.712845,
24
+ "f1": 0.68156,
25
+ "f1_weighted": 0.703197
26
+ },
27
+ {
28
+ "accuracy": 0.741762,
29
+ "f1": 0.696019,
30
+ "f1_weighted": 0.734946
31
+ },
32
+ {
33
+ "accuracy": 0.731338,
34
+ "f1": 0.683951,
35
+ "f1_weighted": 0.723475
36
+ },
37
+ {
38
+ "accuracy": 0.71419,
39
+ "f1": 0.679352,
40
+ "f1_weighted": 0.71517
41
+ },
42
+ {
43
+ "accuracy": 0.721587,
44
+ "f1": 0.688531,
45
+ "f1_weighted": 0.713287
46
+ },
47
+ {
48
+ "accuracy": 0.702421,
49
+ "f1": 0.674508,
50
+ "f1_weighted": 0.696775
51
+ },
52
+ {
53
+ "accuracy": 0.697377,
54
+ "f1": 0.68367,
55
+ "f1_weighted": 0.691308
56
+ },
57
+ {
58
+ "accuracy": 0.729993,
59
+ "f1": 0.690977,
60
+ "f1_weighted": 0.729966
61
+ }
62
+ ],
63
+ "main_score": 0.723268,
64
+ "hf_subset": "pl",
65
+ "languages": [
66
+ "pol-Latn"
67
+ ]
68
+ },
69
+ {
70
+ "accuracy": 0.673268,
71
+ "f1": 0.643382,
72
+ "f1_weighted": 0.663794,
73
+ "scores_per_experiment": [
74
+ {
75
+ "accuracy": 0.670814,
76
+ "f1": 0.635062,
77
+ "f1_weighted": 0.660073
78
+ },
79
+ {
80
+ "accuracy": 0.699731,
81
+ "f1": 0.676887,
82
+ "f1_weighted": 0.69326
83
+ },
84
+ {
85
+ "accuracy": 0.661063,
86
+ "f1": 0.630768,
87
+ "f1_weighted": 0.65112
88
+ },
89
+ {
90
+ "accuracy": 0.694015,
91
+ "f1": 0.640381,
92
+ "f1_weighted": 0.683806
93
+ },
94
+ {
95
+ "accuracy": 0.68191,
96
+ "f1": 0.639771,
97
+ "f1_weighted": 0.672521
98
+ },
99
+ {
100
+ "accuracy": 0.650303,
101
+ "f1": 0.638676,
102
+ "f1_weighted": 0.64201
103
+ },
104
+ {
105
+ "accuracy": 0.67922,
106
+ "f1": 0.652568,
107
+ "f1_weighted": 0.670118
108
+ },
109
+ {
110
+ "accuracy": 0.656355,
111
+ "f1": 0.639743,
112
+ "f1_weighted": 0.648776
113
+ },
114
+ {
115
+ "accuracy": 0.641896,
116
+ "f1": 0.628515,
117
+ "f1_weighted": 0.621914
118
+ },
119
+ {
120
+ "accuracy": 0.697377,
121
+ "f1": 0.651453,
122
+ "f1_weighted": 0.69434
123
+ }
124
+ ],
125
+ "main_score": 0.673268,
126
+ "hf_subset": "az",
127
+ "languages": [
128
+ "aze-Latn"
129
+ ]
130
+ },
131
+ {
132
+ "accuracy": 0.736617,
133
+ "f1": 0.716302,
134
+ "f1_weighted": 0.731258,
135
+ "scores_per_experiment": [
136
+ {
137
+ "accuracy": 0.760592,
138
+ "f1": 0.729,
139
+ "f1_weighted": 0.75188
140
+ },
141
+ {
142
+ "accuracy": 0.759919,
143
+ "f1": 0.731693,
144
+ "f1_weighted": 0.759367
145
+ },
146
+ {
147
+ "accuracy": 0.729657,
148
+ "f1": 0.709279,
149
+ "f1_weighted": 0.721162
150
+ },
151
+ {
152
+ "accuracy": 0.743107,
153
+ "f1": 0.717282,
154
+ "f1_weighted": 0.7411
155
+ },
156
+ {
157
+ "accuracy": 0.736382,
158
+ "f1": 0.70814,
159
+ "f1_weighted": 0.724181
160
+ },
161
+ {
162
+ "accuracy": 0.710827,
163
+ "f1": 0.70294,
164
+ "f1_weighted": 0.69996
165
+ },
166
+ {
167
+ "accuracy": 0.718897,
168
+ "f1": 0.709393,
169
+ "f1_weighted": 0.716032
170
+ },
171
+ {
172
+ "accuracy": 0.728985,
173
+ "f1": 0.707095,
174
+ "f1_weighted": 0.723857
175
+ },
176
+ {
177
+ "accuracy": 0.732347,
178
+ "f1": 0.722152,
179
+ "f1_weighted": 0.729392
180
+ },
181
+ {
182
+ "accuracy": 0.745461,
183
+ "f1": 0.726048,
184
+ "f1_weighted": 0.74565
185
+ }
186
+ ],
187
+ "main_score": 0.736617,
188
+ "hf_subset": "ja",
189
+ "languages": [
190
+ "jpn-Jpan"
191
+ ]
192
+ },
193
+ {
194
+ "accuracy": 0.641459,
195
+ "f1": 0.611658,
196
+ "f1_weighted": 0.636314,
197
+ "scores_per_experiment": [
198
+ {
199
+ "accuracy": 0.675857,
200
+ "f1": 0.643303,
201
+ "f1_weighted": 0.664132
202
+ },
203
+ {
204
+ "accuracy": 0.658709,
205
+ "f1": 0.622165,
206
+ "f1_weighted": 0.655399
207
+ },
208
+ {
209
+ "accuracy": 0.637525,
210
+ "f1": 0.603738,
211
+ "f1_weighted": 0.634935
212
+ },
213
+ {
214
+ "accuracy": 0.646604,
215
+ "f1": 0.604822,
216
+ "f1_weighted": 0.644886
217
+ },
218
+ {
219
+ "accuracy": 0.640551,
220
+ "f1": 0.601249,
221
+ "f1_weighted": 0.634616
222
+ },
223
+ {
224
+ "accuracy": 0.620377,
225
+ "f1": 0.603924,
226
+ "f1_weighted": 0.62136
227
+ },
228
+ {
229
+ "accuracy": 0.639543,
230
+ "f1": 0.624879,
231
+ "f1_weighted": 0.634369
232
+ },
233
+ {
234
+ "accuracy": 0.63887,
235
+ "f1": 0.606283,
236
+ "f1_weighted": 0.629461
237
+ },
238
+ {
239
+ "accuracy": 0.618359,
240
+ "f1": 0.603598,
241
+ "f1_weighted": 0.607557
242
+ },
243
+ {
244
+ "accuracy": 0.638198,
245
+ "f1": 0.602615,
246
+ "f1_weighted": 0.636429
247
+ }
248
+ ],
249
+ "main_score": 0.641459,
250
+ "hf_subset": "my",
251
+ "languages": [
252
+ "mya-Mymr"
253
+ ]
254
+ },
255
+ {
256
+ "accuracy": 0.506254,
257
+ "f1": 0.484532,
258
+ "f1_weighted": 0.497458,
259
+ "scores_per_experiment": [
260
+ {
261
+ "accuracy": 0.525219,
262
+ "f1": 0.49148,
263
+ "f1_weighted": 0.522196
264
+ },
265
+ {
266
+ "accuracy": 0.516476,
267
+ "f1": 0.494015,
268
+ "f1_weighted": 0.51238
269
+ },
270
+ {
271
+ "accuracy": 0.502354,
272
+ "f1": 0.484304,
273
+ "f1_weighted": 0.498428
274
+ },
275
+ {
276
+ "accuracy": 0.507734,
277
+ "f1": 0.493015,
278
+ "f1_weighted": 0.504426
279
+ },
280
+ {
281
+ "accuracy": 0.524882,
282
+ "f1": 0.473208,
283
+ "f1_weighted": 0.514943
284
+ },
285
+ {
286
+ "accuracy": 0.47848,
287
+ "f1": 0.475152,
288
+ "f1_weighted": 0.469371
289
+ },
290
+ {
291
+ "accuracy": 0.474445,
292
+ "f1": 0.463072,
293
+ "f1_weighted": 0.453341
294
+ },
295
+ {
296
+ "accuracy": 0.533289,
297
+ "f1": 0.503,
298
+ "f1_weighted": 0.524465
299
+ },
300
+ {
301
+ "accuracy": 0.494956,
302
+ "f1": 0.482163,
303
+ "f1_weighted": 0.478994
304
+ },
305
+ {
306
+ "accuracy": 0.504707,
307
+ "f1": 0.485912,
308
+ "f1_weighted": 0.496034
309
+ }
310
+ ],
311
+ "main_score": 0.506254,
312
+ "hf_subset": "sw",
313
+ "languages": [
314
+ "swa-Latn"
315
+ ]
316
+ },
317
+ {
318
+ "accuracy": 0.732952,
319
+ "f1": 0.709499,
320
+ "f1_weighted": 0.727393,
321
+ "scores_per_experiment": [
322
+ {
323
+ "accuracy": 0.754876,
324
+ "f1": 0.711558,
325
+ "f1_weighted": 0.746526
326
+ },
327
+ {
328
+ "accuracy": 0.747478,
329
+ "f1": 0.725634,
330
+ "f1_weighted": 0.744326
331
+ },
332
+ {
333
+ "accuracy": 0.724277,
334
+ "f1": 0.708342,
335
+ "f1_weighted": 0.714773
336
+ },
337
+ {
338
+ "accuracy": 0.735709,
339
+ "f1": 0.708745,
340
+ "f1_weighted": 0.728848
341
+ },
342
+ {
343
+ "accuracy": 0.728985,
344
+ "f1": 0.694911,
345
+ "f1_weighted": 0.719244
346
+ },
347
+ {
348
+ "accuracy": 0.712845,
349
+ "f1": 0.694899,
350
+ "f1_weighted": 0.708599
351
+ },
352
+ {
353
+ "accuracy": 0.732683,
354
+ "f1": 0.722799,
355
+ "f1_weighted": 0.727442
356
+ },
357
+ {
358
+ "accuracy": 0.724613,
359
+ "f1": 0.700417,
360
+ "f1_weighted": 0.723051
361
+ },
362
+ {
363
+ "accuracy": 0.719906,
364
+ "f1": 0.708759,
365
+ "f1_weighted": 0.71849
366
+ },
367
+ {
368
+ "accuracy": 0.748151,
369
+ "f1": 0.71892,
370
+ "f1_weighted": 0.74263
371
+ }
372
+ ],
373
+ "main_score": 0.732952,
374
+ "hf_subset": "zh-CN",
375
+ "languages": [
376
+ "cmo-Hans"
377
+ ]
378
+ },
379
+ {
380
+ "accuracy": 0.747377,
381
+ "f1": 0.725688,
382
+ "f1_weighted": 0.740514,
383
+ "scores_per_experiment": [
384
+ {
385
+ "accuracy": 0.761937,
386
+ "f1": 0.737749,
387
+ "f1_weighted": 0.750367
388
+ },
389
+ {
390
+ "accuracy": 0.763954,
391
+ "f1": 0.737796,
392
+ "f1_weighted": 0.757788
393
+ },
394
+ {
395
+ "accuracy": 0.734364,
396
+ "f1": 0.720062,
397
+ "f1_weighted": 0.724527
398
+ },
399
+ {
400
+ "accuracy": 0.758911,
401
+ "f1": 0.731144,
402
+ "f1_weighted": 0.751807
403
+ },
404
+ {
405
+ "accuracy": 0.750168,
406
+ "f1": 0.716329,
407
+ "f1_weighted": 0.744494
408
+ },
409
+ {
410
+ "accuracy": 0.717888,
411
+ "f1": 0.698447,
412
+ "f1_weighted": 0.715676
413
+ },
414
+ {
415
+ "accuracy": 0.742434,
416
+ "f1": 0.723775,
417
+ "f1_weighted": 0.73332
418
+ },
419
+ {
420
+ "accuracy": 0.736382,
421
+ "f1": 0.716973,
422
+ "f1_weighted": 0.733241
423
+ },
424
+ {
425
+ "accuracy": 0.730666,
426
+ "f1": 0.71848,
427
+ "f1_weighted": 0.721098
428
+ },
429
+ {
430
+ "accuracy": 0.777068,
431
+ "f1": 0.75613,
432
+ "f1_weighted": 0.772824
433
+ }
434
+ ],
435
+ "main_score": 0.747377,
436
+ "hf_subset": "ru",
437
+ "languages": [
438
+ "rus-Cyrl"
439
+ ]
440
+ },
441
+ {
442
+ "accuracy": 0.655952,
443
+ "f1": 0.658299,
444
+ "f1_weighted": 0.645753,
445
+ "scores_per_experiment": [
446
+ {
447
+ "accuracy": 0.656355,
448
+ "f1": 0.657179,
449
+ "f1_weighted": 0.645812
450
+ },
451
+ {
452
+ "accuracy": 0.683255,
453
+ "f1": 0.685714,
454
+ "f1_weighted": 0.679235
455
+ },
456
+ {
457
+ "accuracy": 0.661399,
458
+ "f1": 0.66523,
459
+ "f1_weighted": 0.652434
460
+ },
461
+ {
462
+ "accuracy": 0.650975,
463
+ "f1": 0.64856,
464
+ "f1_weighted": 0.642399
465
+ },
466
+ {
467
+ "accuracy": 0.649294,
468
+ "f1": 0.648316,
469
+ "f1_weighted": 0.634804
470
+ },
471
+ {
472
+ "accuracy": 0.64156,
473
+ "f1": 0.656905,
474
+ "f1_weighted": 0.628966
475
+ },
476
+ {
477
+ "accuracy": 0.651648,
478
+ "f1": 0.651677,
479
+ "f1_weighted": 0.636945
480
+ },
481
+ {
482
+ "accuracy": 0.651984,
483
+ "f1": 0.646089,
484
+ "f1_weighted": 0.636352
485
+ },
486
+ {
487
+ "accuracy": 0.645259,
488
+ "f1": 0.652116,
489
+ "f1_weighted": 0.63907
490
+ },
491
+ {
492
+ "accuracy": 0.667787,
493
+ "f1": 0.671202,
494
+ "f1_weighted": 0.661509
495
+ }
496
+ ],
497
+ "main_score": 0.655952,
498
+ "hf_subset": "th",
499
+ "languages": [
500
+ "tha-Thai"
501
+ ]
502
+ },
503
+ {
504
+ "accuracy": 0.520679,
505
+ "f1": 0.496969,
506
+ "f1_weighted": 0.513869,
507
+ "scores_per_experiment": [
508
+ {
509
+ "accuracy": 0.529926,
510
+ "f1": 0.509427,
511
+ "f1_weighted": 0.523087
512
+ },
513
+ {
514
+ "accuracy": 0.569267,
515
+ "f1": 0.530359,
516
+ "f1_weighted": 0.566041
517
+ },
518
+ {
519
+ "accuracy": 0.505044,
520
+ "f1": 0.47615,
521
+ "f1_weighted": 0.495422
522
+ },
523
+ {
524
+ "accuracy": 0.532952,
525
+ "f1": 0.50097,
526
+ "f1_weighted": 0.525102
527
+ },
528
+ {
529
+ "accuracy": 0.532616,
530
+ "f1": 0.502343,
531
+ "f1_weighted": 0.524565
532
+ },
533
+ {
534
+ "accuracy": 0.480498,
535
+ "f1": 0.480316,
536
+ "f1_weighted": 0.480866
537
+ },
538
+ {
539
+ "accuracy": 0.503026,
540
+ "f1": 0.486169,
541
+ "f1_weighted": 0.495852
542
+ },
543
+ {
544
+ "accuracy": 0.518494,
545
+ "f1": 0.496593,
546
+ "f1_weighted": 0.516924
547
+ },
548
+ {
549
+ "accuracy": 0.511432,
550
+ "f1": 0.486477,
551
+ "f1_weighted": 0.498653
552
+ },
553
+ {
554
+ "accuracy": 0.523537,
555
+ "f1": 0.500889,
556
+ "f1_weighted": 0.512174
557
+ }
558
+ ],
559
+ "main_score": 0.520679,
560
+ "hf_subset": "km",
561
+ "languages": [
562
+ "khm-Khmr"
563
+ ]
564
+ },
565
+ {
566
+ "accuracy": 0.646638,
567
+ "f1": 0.627778,
568
+ "f1_weighted": 0.639597,
569
+ "scores_per_experiment": [
570
+ {
571
+ "accuracy": 0.647613,
572
+ "f1": 0.626927,
573
+ "f1_weighted": 0.636177
574
+ },
575
+ {
576
+ "accuracy": 0.656019,
577
+ "f1": 0.645619,
578
+ "f1_weighted": 0.649477
579
+ },
580
+ {
581
+ "accuracy": 0.645931,
582
+ "f1": 0.63075,
583
+ "f1_weighted": 0.64205
584
+ },
585
+ {
586
+ "accuracy": 0.662071,
587
+ "f1": 0.633165,
588
+ "f1_weighted": 0.656486
589
+ },
590
+ {
591
+ "accuracy": 0.647613,
592
+ "f1": 0.617786,
593
+ "f1_weighted": 0.638483
594
+ },
595
+ {
596
+ "accuracy": 0.635171,
597
+ "f1": 0.632053,
598
+ "f1_weighted": 0.632462
599
+ },
600
+ {
601
+ "accuracy": 0.643914,
602
+ "f1": 0.616603,
603
+ "f1_weighted": 0.637667
604
+ },
605
+ {
606
+ "accuracy": 0.64425,
607
+ "f1": 0.623648,
608
+ "f1_weighted": 0.635741
609
+ },
610
+ {
611
+ "accuracy": 0.624748,
612
+ "f1": 0.61377,
613
+ "f1_weighted": 0.609831
614
+ },
615
+ {
616
+ "accuracy": 0.659045,
617
+ "f1": 0.637462,
618
+ "f1_weighted": 0.657598
619
+ }
620
+ ],
621
+ "main_score": 0.646638,
622
+ "hf_subset": "ro",
623
+ "languages": [
624
+ "ron-Latn"
625
+ ]
626
+ },
627
+ {
628
+ "accuracy": 0.685945,
629
+ "f1": 0.67826,
630
+ "f1_weighted": 0.675479,
631
+ "scores_per_experiment": [
632
+ {
633
+ "accuracy": 0.70343,
634
+ "f1": 0.687714,
635
+ "f1_weighted": 0.701321
636
+ },
637
+ {
638
+ "accuracy": 0.691997,
639
+ "f1": 0.684282,
640
+ "f1_weighted": 0.686751
641
+ },
642
+ {
643
+ "accuracy": 0.68191,
644
+ "f1": 0.680597,
645
+ "f1_weighted": 0.66994
646
+ },
647
+ {
648
+ "accuracy": 0.681574,
649
+ "f1": 0.67327,
650
+ "f1_weighted": 0.67145
651
+ },
652
+ {
653
+ "accuracy": 0.698386,
654
+ "f1": 0.673486,
655
+ "f1_weighted": 0.685877
656
+ },
657
+ {
658
+ "accuracy": 0.675521,
659
+ "f1": 0.676852,
660
+ "f1_weighted": 0.658529
661
+ },
662
+ {
663
+ "accuracy": 0.680229,
664
+ "f1": 0.681471,
665
+ "f1_weighted": 0.666843
666
+ },
667
+ {
668
+ "accuracy": 0.696032,
669
+ "f1": 0.681894,
670
+ "f1_weighted": 0.690729
671
+ },
672
+ {
673
+ "accuracy": 0.657028,
674
+ "f1": 0.659853,
675
+ "f1_weighted": 0.643241
676
+ },
677
+ {
678
+ "accuracy": 0.693342,
679
+ "f1": 0.683185,
680
+ "f1_weighted": 0.680107
681
+ }
682
+ ],
683
+ "main_score": 0.685945,
684
+ "hf_subset": "zh-TW",
685
+ "languages": [
686
+ "cmo-Hant"
687
+ ]
688
+ },
689
+ {
690
+ "accuracy": 0.684129,
691
+ "f1": 0.649181,
692
+ "f1_weighted": 0.677293,
693
+ "scores_per_experiment": [
694
+ {
695
+ "accuracy": 0.694687,
696
+ "f1": 0.656698,
697
+ "f1_weighted": 0.686527
698
+ },
699
+ {
700
+ "accuracy": 0.70881,
701
+ "f1": 0.664838,
702
+ "f1_weighted": 0.707908
703
+ },
704
+ {
705
+ "accuracy": 0.6846,
706
+ "f1": 0.652492,
707
+ "f1_weighted": 0.675461
708
+ },
709
+ {
710
+ "accuracy": 0.697714,
711
+ "f1": 0.64919,
712
+ "f1_weighted": 0.692415
713
+ },
714
+ {
715
+ "accuracy": 0.689644,
716
+ "f1": 0.648207,
717
+ "f1_weighted": 0.679673
718
+ },
719
+ {
720
+ "accuracy": 0.655346,
721
+ "f1": 0.634595,
722
+ "f1_weighted": 0.651773
723
+ },
724
+ {
725
+ "accuracy": 0.684264,
726
+ "f1": 0.650495,
727
+ "f1_weighted": 0.675686
728
+ },
729
+ {
730
+ "accuracy": 0.677539,
731
+ "f1": 0.640127,
732
+ "f1_weighted": 0.67347
733
+ },
734
+ {
735
+ "accuracy": 0.643241,
736
+ "f1": 0.627956,
737
+ "f1_weighted": 0.626495
738
+ },
739
+ {
740
+ "accuracy": 0.705447,
741
+ "f1": 0.667209,
742
+ "f1_weighted": 0.703522
743
+ }
744
+ ],
745
+ "main_score": 0.684129,
746
+ "hf_subset": "fi",
747
+ "languages": [
748
+ "fin-Latn"
749
+ ]
750
+ },
751
+ {
752
+ "accuracy": 0.736079,
753
+ "f1": 0.707163,
754
+ "f1_weighted": 0.72899,
755
+ "scores_per_experiment": [
756
+ {
757
+ "accuracy": 0.743443,
758
+ "f1": 0.713564,
759
+ "f1_weighted": 0.735703
760
+ },
761
+ {
762
+ "accuracy": 0.749832,
763
+ "f1": 0.71939,
764
+ "f1_weighted": 0.744332
765
+ },
766
+ {
767
+ "accuracy": 0.720578,
768
+ "f1": 0.695725,
769
+ "f1_weighted": 0.711962
770
+ },
771
+ {
772
+ "accuracy": 0.752522,
773
+ "f1": 0.717576,
774
+ "f1_weighted": 0.746083
775
+ },
776
+ {
777
+ "accuracy": 0.746469,
778
+ "f1": 0.705253,
779
+ "f1_weighted": 0.732862
780
+ },
781
+ {
782
+ "accuracy": 0.705111,
783
+ "f1": 0.675808,
784
+ "f1_weighted": 0.699988
785
+ },
786
+ {
787
+ "accuracy": 0.742434,
788
+ "f1": 0.721605,
789
+ "f1_weighted": 0.736793
790
+ },
791
+ {
792
+ "accuracy": 0.737727,
793
+ "f1": 0.697164,
794
+ "f1_weighted": 0.736255
795
+ },
796
+ {
797
+ "accuracy": 0.709146,
798
+ "f1": 0.698954,
799
+ "f1_weighted": 0.694881
800
+ },
801
+ {
802
+ "accuracy": 0.753531,
803
+ "f1": 0.726594,
804
+ "f1_weighted": 0.751037
805
+ }
806
+ ],
807
+ "main_score": 0.736079,
808
+ "hf_subset": "fr",
809
+ "languages": [
810
+ "fra-Latn"
811
+ ]
812
+ },
813
+ {
814
+ "accuracy": 0.487021,
815
+ "f1": 0.45674,
816
+ "f1_weighted": 0.481562,
817
+ "scores_per_experiment": [
818
+ {
819
+ "accuracy": 0.496974,
820
+ "f1": 0.461991,
821
+ "f1_weighted": 0.499481
822
+ },
823
+ {
824
+ "accuracy": 0.519166,
825
+ "f1": 0.491593,
826
+ "f1_weighted": 0.51985
827
+ },
828
+ {
829
+ "accuracy": 0.488568,
830
+ "f1": 0.458249,
831
+ "f1_weighted": 0.491625
832
+ },
833
+ {
834
+ "accuracy": 0.495965,
835
+ "f1": 0.456085,
836
+ "f1_weighted": 0.490933
837
+ },
838
+ {
839
+ "accuracy": 0.497646,
840
+ "f1": 0.44755,
841
+ "f1_weighted": 0.484882
842
+ },
843
+ {
844
+ "accuracy": 0.46772,
845
+ "f1": 0.448588,
846
+ "f1_weighted": 0.458329
847
+ },
848
+ {
849
+ "accuracy": 0.471419,
850
+ "f1": 0.450105,
851
+ "f1_weighted": 0.461555
852
+ },
853
+ {
854
+ "accuracy": 0.467384,
855
+ "f1": 0.439549,
856
+ "f1_weighted": 0.455175
857
+ },
858
+ {
859
+ "accuracy": 0.485541,
860
+ "f1": 0.461263,
861
+ "f1_weighted": 0.4748
862
+ },
863
+ {
864
+ "accuracy": 0.479825,
865
+ "f1": 0.452427,
866
+ "f1_weighted": 0.478993
867
+ }
868
+ ],
869
+ "main_score": 0.487021,
870
+ "hf_subset": "am",
871
+ "languages": [
872
+ "amh-Ethi"
873
+ ]
874
+ },
875
+ {
876
+ "accuracy": 0.681002,
877
+ "f1": 0.662274,
878
+ "f1_weighted": 0.674588,
879
+ "scores_per_experiment": [
880
+ {
881
+ "accuracy": 0.697377,
882
+ "f1": 0.67531,
883
+ "f1_weighted": 0.691329
884
+ },
885
+ {
886
+ "accuracy": 0.702757,
887
+ "f1": 0.681448,
888
+ "f1_weighted": 0.70165
889
+ },
890
+ {
891
+ "accuracy": 0.675185,
892
+ "f1": 0.655781,
893
+ "f1_weighted": 0.66587
894
+ },
895
+ {
896
+ "accuracy": 0.679556,
897
+ "f1": 0.651732,
898
+ "f1_weighted": 0.673605
899
+ },
900
+ {
901
+ "accuracy": 0.693342,
902
+ "f1": 0.661859,
903
+ "f1_weighted": 0.683319
904
+ },
905
+ {
906
+ "accuracy": 0.660054,
907
+ "f1": 0.65316,
908
+ "f1_weighted": 0.652503
909
+ },
910
+ {
911
+ "accuracy": 0.679892,
912
+ "f1": 0.667511,
913
+ "f1_weighted": 0.677912
914
+ },
915
+ {
916
+ "accuracy": 0.680229,
917
+ "f1": 0.652853,
918
+ "f1_weighted": 0.672891
919
+ },
920
+ {
921
+ "accuracy": 0.655346,
922
+ "f1": 0.66047,
923
+ "f1_weighted": 0.642594
924
+ },
925
+ {
926
+ "accuracy": 0.686281,
927
+ "f1": 0.662616,
928
+ "f1_weighted": 0.684211
929
+ }
930
+ ],
931
+ "main_score": 0.681002,
932
+ "hf_subset": "te",
933
+ "languages": [
934
+ "tel-Telu"
935
+ ]
936
+ },
937
+ {
938
+ "accuracy": 0.728447,
939
+ "f1": 0.708234,
940
+ "f1_weighted": 0.722561,
941
+ "scores_per_experiment": [
942
+ {
943
+ "accuracy": 0.745797,
944
+ "f1": 0.721343,
945
+ "f1_weighted": 0.73932
946
+ },
947
+ {
948
+ "accuracy": 0.742434,
949
+ "f1": 0.719328,
950
+ "f1_weighted": 0.736915
951
+ },
952
+ {
953
+ "accuracy": 0.720578,
954
+ "f1": 0.699465,
955
+ "f1_weighted": 0.713217
956
+ },
957
+ {
958
+ "accuracy": 0.744452,
959
+ "f1": 0.713588,
960
+ "f1_weighted": 0.742004
961
+ },
962
+ {
963
+ "accuracy": 0.73302,
964
+ "f1": 0.705034,
965
+ "f1_weighted": 0.724726
966
+ },
967
+ {
968
+ "accuracy": 0.703766,
969
+ "f1": 0.687671,
970
+ "f1_weighted": 0.700645
971
+ },
972
+ {
973
+ "accuracy": 0.727303,
974
+ "f1": 0.710824,
975
+ "f1_weighted": 0.720319
976
+ },
977
+ {
978
+ "accuracy": 0.717552,
979
+ "f1": 0.707643,
980
+ "f1_weighted": 0.706761
981
+ },
982
+ {
983
+ "accuracy": 0.709482,
984
+ "f1": 0.698837,
985
+ "f1_weighted": 0.702403
986
+ },
987
+ {
988
+ "accuracy": 0.740081,
989
+ "f1": 0.718605,
990
+ "f1_weighted": 0.739297
991
+ }
992
+ ],
993
+ "main_score": 0.728447,
994
+ "hf_subset": "de",
995
+ "languages": [
996
+ "deu-Latn"
997
+ ]
998
+ },
999
+ {
1000
+ "accuracy": 0.64267,
1001
+ "f1": 0.60778,
1002
+ "f1_weighted": 0.634122,
1003
+ "scores_per_experiment": [
1004
+ {
1005
+ "accuracy": 0.667115,
1006
+ "f1": 0.623772,
1007
+ "f1_weighted": 0.656154
1008
+ },
1009
+ {
1010
+ "accuracy": 0.671486,
1011
+ "f1": 0.64547,
1012
+ "f1_weighted": 0.672281
1013
+ },
1014
+ {
1015
+ "accuracy": 0.635171,
1016
+ "f1": 0.598352,
1017
+ "f1_weighted": 0.625034
1018
+ },
1019
+ {
1020
+ "accuracy": 0.672831,
1021
+ "f1": 0.622347,
1022
+ "f1_weighted": 0.666015
1023
+ },
1024
+ {
1025
+ "accuracy": 0.610289,
1026
+ "f1": 0.582646,
1027
+ "f1_weighted": 0.596477
1028
+ },
1029
+ {
1030
+ "accuracy": 0.629792,
1031
+ "f1": 0.601476,
1032
+ "f1_weighted": 0.622663
1033
+ },
1034
+ {
1035
+ "accuracy": 0.643914,
1036
+ "f1": 0.60434,
1037
+ "f1_weighted": 0.635887
1038
+ },
1039
+ {
1040
+ "accuracy": 0.632482,
1041
+ "f1": 0.598433,
1042
+ "f1_weighted": 0.621951
1043
+ },
1044
+ {
1045
+ "accuracy": 0.611634,
1046
+ "f1": 0.59445,
1047
+ "f1_weighted": 0.597673
1048
+ },
1049
+ {
1050
+ "accuracy": 0.651984,
1051
+ "f1": 0.606513,
1052
+ "f1_weighted": 0.647081
1053
+ }
1054
+ ],
1055
+ "main_score": 0.64267,
1056
+ "hf_subset": "lv",
1057
+ "languages": [
1058
+ "lav-Latn"
1059
+ ]
1060
+ },
1061
+ {
1062
+ "accuracy": 0.713416,
1063
+ "f1": 0.690409,
1064
+ "f1_weighted": 0.703952,
1065
+ "scores_per_experiment": [
1066
+ {
1067
+ "accuracy": 0.721587,
1068
+ "f1": 0.70151,
1069
+ "f1_weighted": 0.713165
1070
+ },
1071
+ {
1072
+ "accuracy": 0.737054,
1073
+ "f1": 0.717224,
1074
+ "f1_weighted": 0.729775
1075
+ },
1076
+ {
1077
+ "accuracy": 0.705447,
1078
+ "f1": 0.680556,
1079
+ "f1_weighted": 0.697172
1080
+ },
1081
+ {
1082
+ "accuracy": 0.732683,
1083
+ "f1": 0.695699,
1084
+ "f1_weighted": 0.722658
1085
+ },
1086
+ {
1087
+ "accuracy": 0.708137,
1088
+ "f1": 0.679096,
1089
+ "f1_weighted": 0.690052
1090
+ },
1091
+ {
1092
+ "accuracy": 0.691997,
1093
+ "f1": 0.679759,
1094
+ "f1_weighted": 0.68793
1095
+ },
1096
+ {
1097
+ "accuracy": 0.719233,
1098
+ "f1": 0.695823,
1099
+ "f1_weighted": 0.708628
1100
+ },
1101
+ {
1102
+ "accuracy": 0.710491,
1103
+ "f1": 0.68074,
1104
+ "f1_weighted": 0.70368
1105
+ },
1106
+ {
1107
+ "accuracy": 0.677539,
1108
+ "f1": 0.67384,
1109
+ "f1_weighted": 0.661528
1110
+ },
1111
+ {
1112
+ "accuracy": 0.729993,
1113
+ "f1": 0.699844,
1114
+ "f1_weighted": 0.72493
1115
+ }
1116
+ ],
1117
+ "main_score": 0.713416,
1118
+ "hf_subset": "sv",
1119
+ "languages": [
1120
+ "swe-Latn"
1121
+ ]
1122
+ },
1123
+ {
1124
+ "accuracy": 0.652017,
1125
+ "f1": 0.635476,
1126
+ "f1_weighted": 0.644408,
1127
+ "scores_per_experiment": [
1128
+ {
1129
+ "accuracy": 0.664425,
1130
+ "f1": 0.648954,
1131
+ "f1_weighted": 0.657846
1132
+ },
1133
+ {
1134
+ "accuracy": 0.669469,
1135
+ "f1": 0.65079,
1136
+ "f1_weighted": 0.668401
1137
+ },
1138
+ {
1139
+ "accuracy": 0.642569,
1140
+ "f1": 0.632603,
1141
+ "f1_weighted": 0.633788
1142
+ },
1143
+ {
1144
+ "accuracy": 0.66039,
1145
+ "f1": 0.635717,
1146
+ "f1_weighted": 0.655039
1147
+ },
1148
+ {
1149
+ "accuracy": 0.646604,
1150
+ "f1": 0.625111,
1151
+ "f1_weighted": 0.632445
1152
+ },
1153
+ {
1154
+ "accuracy": 0.631473,
1155
+ "f1": 0.616728,
1156
+ "f1_weighted": 0.623688
1157
+ },
1158
+ {
1159
+ "accuracy": 0.643578,
1160
+ "f1": 0.631581,
1161
+ "f1_weighted": 0.639157
1162
+ },
1163
+ {
1164
+ "accuracy": 0.653665,
1165
+ "f1": 0.630242,
1166
+ "f1_weighted": 0.646023
1167
+ },
1168
+ {
1169
+ "accuracy": 0.632818,
1170
+ "f1": 0.63016,
1171
+ "f1_weighted": 0.614353
1172
+ },
1173
+ {
1174
+ "accuracy": 0.675185,
1175
+ "f1": 0.652878,
1176
+ "f1_weighted": 0.67334
1177
+ }
1178
+ ],
1179
+ "main_score": 0.652017,
1180
+ "hf_subset": "kn",
1181
+ "languages": [
1182
+ "kan-Knda"
1183
+ ]
1184
+ },
1185
+ {
1186
+ "accuracy": 0.580161,
1187
+ "f1": 0.544392,
1188
+ "f1_weighted": 0.573598,
1189
+ "scores_per_experiment": [
1190
+ {
1191
+ "accuracy": 0.585071,
1192
+ "f1": 0.527179,
1193
+ "f1_weighted": 0.57385
1194
+ },
1195
+ {
1196
+ "accuracy": 0.577673,
1197
+ "f1": 0.553388,
1198
+ "f1_weighted": 0.586091
1199
+ },
1200
+ {
1201
+ "accuracy": 0.582717,
1202
+ "f1": 0.533965,
1203
+ "f1_weighted": 0.578423
1204
+ },
1205
+ {
1206
+ "accuracy": 0.591459,
1207
+ "f1": 0.556131,
1208
+ "f1_weighted": 0.587187
1209
+ },
1210
+ {
1211
+ "accuracy": 0.572966,
1212
+ "f1": 0.550562,
1213
+ "f1_weighted": 0.563138
1214
+ },
1215
+ {
1216
+ "accuracy": 0.551446,
1217
+ "f1": 0.526132,
1218
+ "f1_weighted": 0.533481
1219
+ },
1220
+ {
1221
+ "accuracy": 0.598521,
1222
+ "f1": 0.549881,
1223
+ "f1_weighted": 0.594835
1224
+ },
1225
+ {
1226
+ "accuracy": 0.588097,
1227
+ "f1": 0.555036,
1228
+ "f1_weighted": 0.578277
1229
+ },
1230
+ {
1231
+ "accuracy": 0.575656,
1232
+ "f1": 0.540914,
1233
+ "f1_weighted": 0.564224
1234
+ },
1235
+ {
1236
+ "accuracy": 0.578009,
1237
+ "f1": 0.550731,
1238
+ "f1_weighted": 0.576469
1239
+ }
1240
+ ],
1241
+ "main_score": 0.580161,
1242
+ "hf_subset": "ar",
1243
+ "languages": [
1244
+ "ara-Arab"
1245
+ ]
1246
+ },
1247
+ {
1248
+ "accuracy": 0.672999,
1249
+ "f1": 0.644536,
1250
+ "f1_weighted": 0.664091,
1251
+ "scores_per_experiment": [
1252
+ {
1253
+ "accuracy": 0.680229,
1254
+ "f1": 0.639686,
1255
+ "f1_weighted": 0.673659
1256
+ },
1257
+ {
1258
+ "accuracy": 0.69536,
1259
+ "f1": 0.661895,
1260
+ "f1_weighted": 0.689334
1261
+ },
1262
+ {
1263
+ "accuracy": 0.664761,
1264
+ "f1": 0.643661,
1265
+ "f1_weighted": 0.655536
1266
+ },
1267
+ {
1268
+ "accuracy": 0.685945,
1269
+ "f1": 0.648907,
1270
+ "f1_weighted": 0.671805
1271
+ },
1272
+ {
1273
+ "accuracy": 0.678547,
1274
+ "f1": 0.638195,
1275
+ "f1_weighted": 0.667685
1276
+ },
1277
+ {
1278
+ "accuracy": 0.662071,
1279
+ "f1": 0.633331,
1280
+ "f1_weighted": 0.658713
1281
+ },
1282
+ {
1283
+ "accuracy": 0.669805,
1284
+ "f1": 0.643201,
1285
+ "f1_weighted": 0.663079
1286
+ },
1287
+ {
1288
+ "accuracy": 0.657028,
1289
+ "f1": 0.628883,
1290
+ "f1_weighted": 0.643688
1291
+ },
1292
+ {
1293
+ "accuracy": 0.643578,
1294
+ "f1": 0.637717,
1295
+ "f1_weighted": 0.629237
1296
+ },
1297
+ {
1298
+ "accuracy": 0.69267,
1299
+ "f1": 0.669881,
1300
+ "f1_weighted": 0.688178
1301
+ }
1302
+ ],
1303
+ "main_score": 0.672999,
1304
+ "hf_subset": "hu",
1305
+ "languages": [
1306
+ "hun-Latn"
1307
+ ]
1308
+ },
1309
+ {
1310
+ "accuracy": 0.723773,
1311
+ "f1": 0.693915,
1312
+ "f1_weighted": 0.717968,
1313
+ "scores_per_experiment": [
1314
+ {
1315
+ "accuracy": 0.731338,
1316
+ "f1": 0.692884,
1317
+ "f1_weighted": 0.723121
1318
+ },
1319
+ {
1320
+ "accuracy": 0.747478,
1321
+ "f1": 0.705572,
1322
+ "f1_weighted": 0.74309
1323
+ },
1324
+ {
1325
+ "accuracy": 0.716207,
1326
+ "f1": 0.689909,
1327
+ "f1_weighted": 0.710249
1328
+ },
1329
+ {
1330
+ "accuracy": 0.743779,
1331
+ "f1": 0.698756,
1332
+ "f1_weighted": 0.736164
1333
+ },
1334
+ {
1335
+ "accuracy": 0.735037,
1336
+ "f1": 0.695828,
1337
+ "f1_weighted": 0.725718
1338
+ },
1339
+ {
1340
+ "accuracy": 0.706792,
1341
+ "f1": 0.691281,
1342
+ "f1_weighted": 0.707372
1343
+ },
1344
+ {
1345
+ "accuracy": 0.713517,
1346
+ "f1": 0.689481,
1347
+ "f1_weighted": 0.703641
1348
+ },
1349
+ {
1350
+ "accuracy": 0.713853,
1351
+ "f1": 0.685174,
1352
+ "f1_weighted": 0.707503
1353
+ },
1354
+ {
1355
+ "accuracy": 0.697041,
1356
+ "f1": 0.684665,
1357
+ "f1_weighted": 0.691905
1358
+ },
1359
+ {
1360
+ "accuracy": 0.732683,
1361
+ "f1": 0.705598,
1362
+ "f1_weighted": 0.730914
1363
+ }
1364
+ ],
1365
+ "main_score": 0.723773,
1366
+ "hf_subset": "nl",
1367
+ "languages": [
1368
+ "nld-Latn"
1369
+ ]
1370
+ },
1371
+ {
1372
+ "accuracy": 0.569839,
1373
+ "f1": 0.540126,
1374
+ "f1_weighted": 0.559746,
1375
+ "scores_per_experiment": [
1376
+ {
1377
+ "accuracy": 0.581708,
1378
+ "f1": 0.555286,
1379
+ "f1_weighted": 0.572067
1380
+ },
1381
+ {
1382
+ "accuracy": 0.565905,
1383
+ "f1": 0.538163,
1384
+ "f1_weighted": 0.558387
1385
+ },
1386
+ {
1387
+ "accuracy": 0.554472,
1388
+ "f1": 0.51573,
1389
+ "f1_weighted": 0.541189
1390
+ },
1391
+ {
1392
+ "accuracy": 0.576664,
1393
+ "f1": 0.54813,
1394
+ "f1_weighted": 0.574907
1395
+ },
1396
+ {
1397
+ "accuracy": 0.570612,
1398
+ "f1": 0.533006,
1399
+ "f1_weighted": 0.556786
1400
+ },
1401
+ {
1402
+ "accuracy": 0.555817,
1403
+ "f1": 0.537519,
1404
+ "f1_weighted": 0.552445
1405
+ },
1406
+ {
1407
+ "accuracy": 0.582381,
1408
+ "f1": 0.549933,
1409
+ "f1_weighted": 0.580289
1410
+ },
1411
+ {
1412
+ "accuracy": 0.56456,
1413
+ "f1": 0.534174,
1414
+ "f1_weighted": 0.543014
1415
+ },
1416
+ {
1417
+ "accuracy": 0.557835,
1418
+ "f1": 0.533915,
1419
+ "f1_weighted": 0.538608
1420
+ },
1421
+ {
1422
+ "accuracy": 0.588433,
1423
+ "f1": 0.555399,
1424
+ "f1_weighted": 0.579772
1425
+ }
1426
+ ],
1427
+ "main_score": 0.569839,
1428
+ "hf_subset": "is",
1429
+ "languages": [
1430
+ "isl-Latn"
1431
+ ]
1432
+ },
1433
+ {
1434
+ "accuracy": 0.662979,
1435
+ "f1": 0.636952,
1436
+ "f1_weighted": 0.657962,
1437
+ "scores_per_experiment": [
1438
+ {
1439
+ "accuracy": 0.676194,
1440
+ "f1": 0.654856,
1441
+ "f1_weighted": 0.674512
1442
+ },
1443
+ {
1444
+ "accuracy": 0.697041,
1445
+ "f1": 0.665047,
1446
+ "f1_weighted": 0.694727
1447
+ },
1448
+ {
1449
+ "accuracy": 0.647949,
1450
+ "f1": 0.606057,
1451
+ "f1_weighted": 0.643065
1452
+ },
1453
+ {
1454
+ "accuracy": 0.687626,
1455
+ "f1": 0.655421,
1456
+ "f1_weighted": 0.686045
1457
+ },
1458
+ {
1459
+ "accuracy": 0.634499,
1460
+ "f1": 0.611777,
1461
+ "f1_weighted": 0.623641
1462
+ },
1463
+ {
1464
+ "accuracy": 0.640215,
1465
+ "f1": 0.629971,
1466
+ "f1_weighted": 0.637149
1467
+ },
1468
+ {
1469
+ "accuracy": 0.666443,
1470
+ "f1": 0.648677,
1471
+ "f1_weighted": 0.662303
1472
+ },
1473
+ {
1474
+ "accuracy": 0.649294,
1475
+ "f1": 0.620246,
1476
+ "f1_weighted": 0.639949
1477
+ },
1478
+ {
1479
+ "accuracy": 0.65501,
1480
+ "f1": 0.6309,
1481
+ "f1_weighted": 0.642648
1482
+ },
1483
+ {
1484
+ "accuracy": 0.675521,
1485
+ "f1": 0.64657,
1486
+ "f1_weighted": 0.675584
1487
+ }
1488
+ ],
1489
+ "main_score": 0.662979,
1490
+ "hf_subset": "ur",
1491
+ "languages": [
1492
+ "urd-Arab"
1493
+ ]
1494
+ },
1495
+ {
1496
+ "accuracy": 0.699294,
1497
+ "f1": 0.672813,
1498
+ "f1_weighted": 0.691332,
1499
+ "scores_per_experiment": [
1500
+ {
1501
+ "accuracy": 0.709146,
1502
+ "f1": 0.676494,
1503
+ "f1_weighted": 0.698244
1504
+ },
1505
+ {
1506
+ "accuracy": 0.709482,
1507
+ "f1": 0.679792,
1508
+ "f1_weighted": 0.707249
1509
+ },
1510
+ {
1511
+ "accuracy": 0.701076,
1512
+ "f1": 0.679608,
1513
+ "f1_weighted": 0.69154
1514
+ },
1515
+ {
1516
+ "accuracy": 0.719233,
1517
+ "f1": 0.671811,
1518
+ "f1_weighted": 0.708524
1519
+ },
1520
+ {
1521
+ "accuracy": 0.70343,
1522
+ "f1": 0.681794,
1523
+ "f1_weighted": 0.692521
1524
+ },
1525
+ {
1526
+ "accuracy": 0.669805,
1527
+ "f1": 0.667619,
1528
+ "f1_weighted": 0.66864
1529
+ },
1530
+ {
1531
+ "accuracy": 0.711163,
1532
+ "f1": 0.680014,
1533
+ "f1_weighted": 0.699725
1534
+ },
1535
+ {
1536
+ "accuracy": 0.6846,
1537
+ "f1": 0.64862,
1538
+ "f1_weighted": 0.677752
1539
+ },
1540
+ {
1541
+ "accuracy": 0.677875,
1542
+ "f1": 0.662457,
1543
+ "f1_weighted": 0.66474
1544
+ },
1545
+ {
1546
+ "accuracy": 0.707128,
1547
+ "f1": 0.679923,
1548
+ "f1_weighted": 0.704386
1549
+ }
1550
+ ],
1551
+ "main_score": 0.699294,
1552
+ "hf_subset": "fa",
1553
+ "languages": [
1554
+ "fas-Arab"
1555
+ ]
1556
+ },
1557
+ {
1558
+ "accuracy": 0.642636,
1559
+ "f1": 0.612847,
1560
+ "f1_weighted": 0.64227,
1561
+ "scores_per_experiment": [
1562
+ {
1563
+ "accuracy": 0.644923,
1564
+ "f1": 0.612034,
1565
+ "f1_weighted": 0.643371
1566
+ },
1567
+ {
1568
+ "accuracy": 0.669132,
1569
+ "f1": 0.632734,
1570
+ "f1_weighted": 0.669613
1571
+ },
1572
+ {
1573
+ "accuracy": 0.643914,
1574
+ "f1": 0.609033,
1575
+ "f1_weighted": 0.642278
1576
+ },
1577
+ {
1578
+ "accuracy": 0.667451,
1579
+ "f1": 0.633831,
1580
+ "f1_weighted": 0.671122
1581
+ },
1582
+ {
1583
+ "accuracy": 0.63887,
1584
+ "f1": 0.597595,
1585
+ "f1_weighted": 0.634692
1586
+ },
1587
+ {
1588
+ "accuracy": 0.62273,
1589
+ "f1": 0.598132,
1590
+ "f1_weighted": 0.631559
1591
+ },
1592
+ {
1593
+ "accuracy": 0.637861,
1594
+ "f1": 0.611803,
1595
+ "f1_weighted": 0.63535
1596
+ },
1597
+ {
1598
+ "accuracy": 0.623403,
1599
+ "f1": 0.608466,
1600
+ "f1_weighted": 0.624853
1601
+ },
1602
+ {
1603
+ "accuracy": 0.619368,
1604
+ "f1": 0.599173,
1605
+ "f1_weighted": 0.609207
1606
+ },
1607
+ {
1608
+ "accuracy": 0.658709,
1609
+ "f1": 0.625666,
1610
+ "f1_weighted": 0.660654
1611
+ }
1612
+ ],
1613
+ "main_score": 0.642636,
1614
+ "hf_subset": "he",
1615
+ "languages": [
1616
+ "heb-Hebr"
1617
+ ]
1618
+ },
1619
+ {
1620
+ "accuracy": 0.715972,
1621
+ "f1": 0.698643,
1622
+ "f1_weighted": 0.708427,
1623
+ "scores_per_experiment": [
1624
+ {
1625
+ "accuracy": 0.737054,
1626
+ "f1": 0.71295,
1627
+ "f1_weighted": 0.725932
1628
+ },
1629
+ {
1630
+ "accuracy": 0.71688,
1631
+ "f1": 0.707235,
1632
+ "f1_weighted": 0.713866
1633
+ },
1634
+ {
1635
+ "accuracy": 0.703766,
1636
+ "f1": 0.685773,
1637
+ "f1_weighted": 0.694379
1638
+ },
1639
+ {
1640
+ "accuracy": 0.726295,
1641
+ "f1": 0.704115,
1642
+ "f1_weighted": 0.717795
1643
+ },
1644
+ {
1645
+ "accuracy": 0.722932,
1646
+ "f1": 0.699105,
1647
+ "f1_weighted": 0.710328
1648
+ },
1649
+ {
1650
+ "accuracy": 0.697714,
1651
+ "f1": 0.687283,
1652
+ "f1_weighted": 0.689616
1653
+ },
1654
+ {
1655
+ "accuracy": 0.718225,
1656
+ "f1": 0.705447,
1657
+ "f1_weighted": 0.7097
1658
+ },
1659
+ {
1660
+ "accuracy": 0.704438,
1661
+ "f1": 0.677868,
1662
+ "f1_weighted": 0.6987
1663
+ },
1664
+ {
1665
+ "accuracy": 0.703766,
1666
+ "f1": 0.69539,
1667
+ "f1_weighted": 0.700162
1668
+ },
1669
+ {
1670
+ "accuracy": 0.728648,
1671
+ "f1": 0.711261,
1672
+ "f1_weighted": 0.723788
1673
+ }
1674
+ ],
1675
+ "main_score": 0.715972,
1676
+ "hf_subset": "vi",
1677
+ "languages": [
1678
+ "vie-Latn"
1679
+ ]
1680
+ },
1681
+ {
1682
+ "accuracy": 0.701614,
1683
+ "f1": 0.687989,
1684
+ "f1_weighted": 0.694745,
1685
+ "scores_per_experiment": [
1686
+ {
1687
+ "accuracy": 0.698722,
1688
+ "f1": 0.690141,
1689
+ "f1_weighted": 0.693205
1690
+ },
1691
+ {
1692
+ "accuracy": 0.731002,
1693
+ "f1": 0.70874,
1694
+ "f1_weighted": 0.726378
1695
+ },
1696
+ {
1697
+ "accuracy": 0.705447,
1698
+ "f1": 0.69742,
1699
+ "f1_weighted": 0.699488
1700
+ },
1701
+ {
1702
+ "accuracy": 0.717216,
1703
+ "f1": 0.685667,
1704
+ "f1_weighted": 0.712119
1705
+ },
1706
+ {
1707
+ "accuracy": 0.697377,
1708
+ "f1": 0.677651,
1709
+ "f1_weighted": 0.688238
1710
+ },
1711
+ {
1712
+ "accuracy": 0.684936,
1713
+ "f1": 0.684789,
1714
+ "f1_weighted": 0.682975
1715
+ },
1716
+ {
1717
+ "accuracy": 0.694351,
1718
+ "f1": 0.685584,
1719
+ "f1_weighted": 0.686116
1720
+ },
1721
+ {
1722
+ "accuracy": 0.683927,
1723
+ "f1": 0.667004,
1724
+ "f1_weighted": 0.674774
1725
+ },
1726
+ {
1727
+ "accuracy": 0.685272,
1728
+ "f1": 0.681627,
1729
+ "f1_weighted": 0.672504
1730
+ },
1731
+ {
1732
+ "accuracy": 0.717888,
1733
+ "f1": 0.701268,
1734
+ "f1_weighted": 0.711653
1735
+ }
1736
+ ],
1737
+ "main_score": 0.701614,
1738
+ "hf_subset": "hi",
1739
+ "languages": [
1740
+ "hin-Deva"
1741
+ ]
1742
+ },
1743
+ {
1744
+ "accuracy": 0.581574,
1745
+ "f1": 0.535656,
1746
+ "f1_weighted": 0.572902,
1747
+ "scores_per_experiment": [
1748
+ {
1749
+ "accuracy": 0.589442,
1750
+ "f1": 0.528775,
1751
+ "f1_weighted": 0.580566
1752
+ },
1753
+ {
1754
+ "accuracy": 0.598521,
1755
+ "f1": 0.55802,
1756
+ "f1_weighted": 0.593675
1757
+ },
1758
+ {
1759
+ "accuracy": 0.585407,
1760
+ "f1": 0.544246,
1761
+ "f1_weighted": 0.585102
1762
+ },
1763
+ {
1764
+ "accuracy": 0.582381,
1765
+ "f1": 0.530482,
1766
+ "f1_weighted": 0.57048
1767
+ },
1768
+ {
1769
+ "accuracy": 0.574311,
1770
+ "f1": 0.52058,
1771
+ "f1_weighted": 0.56277
1772
+ },
1773
+ {
1774
+ "accuracy": 0.549765,
1775
+ "f1": 0.519758,
1776
+ "f1_weighted": 0.535436
1777
+ },
1778
+ {
1779
+ "accuracy": 0.569939,
1780
+ "f1": 0.516298,
1781
+ "f1_weighted": 0.553421
1782
+ },
1783
+ {
1784
+ "accuracy": 0.594822,
1785
+ "f1": 0.552697,
1786
+ "f1_weighted": 0.593147
1787
+ },
1788
+ {
1789
+ "accuracy": 0.569267,
1790
+ "f1": 0.536861,
1791
+ "f1_weighted": 0.556395
1792
+ },
1793
+ {
1794
+ "accuracy": 0.601883,
1795
+ "f1": 0.548843,
1796
+ "f1_weighted": 0.598033
1797
+ }
1798
+ ],
1799
+ "main_score": 0.581574,
1800
+ "hf_subset": "jv",
1801
+ "languages": [
1802
+ "jav-Latn"
1803
+ ]
1804
+ },
1805
+ {
1806
+ "accuracy": 0.697545,
1807
+ "f1": 0.684021,
1808
+ "f1_weighted": 0.69177,
1809
+ "scores_per_experiment": [
1810
+ {
1811
+ "accuracy": 0.723268,
1812
+ "f1": 0.689989,
1813
+ "f1_weighted": 0.713863
1814
+ },
1815
+ {
1816
+ "accuracy": 0.703766,
1817
+ "f1": 0.694084,
1818
+ "f1_weighted": 0.706167
1819
+ },
1820
+ {
1821
+ "accuracy": 0.692334,
1822
+ "f1": 0.67747,
1823
+ "f1_weighted": 0.686072
1824
+ },
1825
+ {
1826
+ "accuracy": 0.704438,
1827
+ "f1": 0.682842,
1828
+ "f1_weighted": 0.696445
1829
+ },
1830
+ {
1831
+ "accuracy": 0.70343,
1832
+ "f1": 0.682407,
1833
+ "f1_weighted": 0.697165
1834
+ },
1835
+ {
1836
+ "accuracy": 0.659045,
1837
+ "f1": 0.667468,
1838
+ "f1_weighted": 0.646055
1839
+ },
1840
+ {
1841
+ "accuracy": 0.692334,
1842
+ "f1": 0.683636,
1843
+ "f1_weighted": 0.687705
1844
+ },
1845
+ {
1846
+ "accuracy": 0.704102,
1847
+ "f1": 0.683478,
1848
+ "f1_weighted": 0.696546
1849
+ },
1850
+ {
1851
+ "accuracy": 0.680901,
1852
+ "f1": 0.67889,
1853
+ "f1_weighted": 0.676699
1854
+ },
1855
+ {
1856
+ "accuracy": 0.711836,
1857
+ "f1": 0.699948,
1858
+ "f1_weighted": 0.710987
1859
+ }
1860
+ ],
1861
+ "main_score": 0.697545,
1862
+ "hf_subset": "ko",
1863
+ "languages": [
1864
+ "kor-Kore"
1865
+ ]
1866
+ },
1867
+ {
1868
+ "accuracy": 0.775588,
1869
+ "f1": 0.75861,
1870
+ "f1_weighted": 0.76915,
1871
+ "scores_per_experiment": [
1872
+ {
1873
+ "accuracy": 0.789845,
1874
+ "f1": 0.76684,
1875
+ "f1_weighted": 0.783395
1876
+ },
1877
+ {
1878
+ "accuracy": 0.792872,
1879
+ "f1": 0.765188,
1880
+ "f1_weighted": 0.7855
1881
+ },
1882
+ {
1883
+ "accuracy": 0.767653,
1884
+ "f1": 0.752013,
1885
+ "f1_weighted": 0.758055
1886
+ },
1887
+ {
1888
+ "accuracy": 0.784465,
1889
+ "f1": 0.75903,
1890
+ "f1_weighted": 0.779529
1891
+ },
1892
+ {
1893
+ "accuracy": 0.777404,
1894
+ "f1": 0.756543,
1895
+ "f1_weighted": 0.765982
1896
+ },
1897
+ {
1898
+ "accuracy": 0.757566,
1899
+ "f1": 0.74771,
1900
+ "f1_weighted": 0.751965
1901
+ },
1902
+ {
1903
+ "accuracy": 0.767317,
1904
+ "f1": 0.760559,
1905
+ "f1_weighted": 0.762363
1906
+ },
1907
+ {
1908
+ "accuracy": 0.763954,
1909
+ "f1": 0.750649,
1910
+ "f1_weighted": 0.758685
1911
+ },
1912
+ {
1913
+ "accuracy": 0.756893,
1914
+ "f1": 0.750547,
1915
+ "f1_weighted": 0.751351
1916
+ },
1917
+ {
1918
+ "accuracy": 0.797915,
1919
+ "f1": 0.777027,
1920
+ "f1_weighted": 0.794677
1921
+ }
1922
+ ],
1923
+ "main_score": 0.775588,
1924
+ "hf_subset": "en",
1925
+ "languages": [
1926
+ "eng-Latn"
1927
+ ]
1928
+ },
1929
+ {
1930
+ "accuracy": 0.454909,
1931
+ "f1": 0.430406,
1932
+ "f1_weighted": 0.449208,
1933
+ "scores_per_experiment": [
1934
+ {
1935
+ "accuracy": 0.465367,
1936
+ "f1": 0.448279,
1937
+ "f1_weighted": 0.463704
1938
+ },
1939
+ {
1940
+ "accuracy": 0.46772,
1941
+ "f1": 0.434925,
1942
+ "f1_weighted": 0.465242
1943
+ },
1944
+ {
1945
+ "accuracy": 0.452589,
1946
+ "f1": 0.426374,
1947
+ "f1_weighted": 0.453014
1948
+ },
1949
+ {
1950
+ "accuracy": 0.459314,
1951
+ "f1": 0.438415,
1952
+ "f1_weighted": 0.459009
1953
+ },
1954
+ {
1955
+ "accuracy": 0.462004,
1956
+ "f1": 0.422972,
1957
+ "f1_weighted": 0.452415
1958
+ },
1959
+ {
1960
+ "accuracy": 0.424681,
1961
+ "f1": 0.422058,
1962
+ "f1_weighted": 0.42101
1963
+ },
1964
+ {
1965
+ "accuracy": 0.452925,
1966
+ "f1": 0.429245,
1967
+ "f1_weighted": 0.444626
1968
+ },
1969
+ {
1970
+ "accuracy": 0.470074,
1971
+ "f1": 0.437186,
1972
+ "f1_weighted": 0.464928
1973
+ },
1974
+ {
1975
+ "accuracy": 0.453934,
1976
+ "f1": 0.42621,
1977
+ "f1_weighted": 0.430667
1978
+ },
1979
+ {
1980
+ "accuracy": 0.440484,
1981
+ "f1": 0.418392,
1982
+ "f1_weighted": 0.437468
1983
+ }
1984
+ ],
1985
+ "main_score": 0.454909,
1986
+ "hf_subset": "cy",
1987
+ "languages": [
1988
+ "cym-Latn"
1989
+ ]
1990
+ },
1991
+ {
1992
+ "accuracy": 0.718931,
1993
+ "f1": 0.709726,
1994
+ "f1_weighted": 0.711706,
1995
+ "scores_per_experiment": [
1996
+ {
1997
+ "accuracy": 0.735373,
1998
+ "f1": 0.721965,
1999
+ "f1_weighted": 0.721597
2000
+ },
2001
+ {
2002
+ "accuracy": 0.732011,
2003
+ "f1": 0.723526,
2004
+ "f1_weighted": 0.725878
2005
+ },
2006
+ {
2007
+ "accuracy": 0.70343,
2008
+ "f1": 0.701242,
2009
+ "f1_weighted": 0.698283
2010
+ },
2011
+ {
2012
+ "accuracy": 0.738399,
2013
+ "f1": 0.718247,
2014
+ "f1_weighted": 0.733675
2015
+ },
2016
+ {
2017
+ "accuracy": 0.72495,
2018
+ "f1": 0.702411,
2019
+ "f1_weighted": 0.713614
2020
+ },
2021
+ {
2022
+ "accuracy": 0.70074,
2023
+ "f1": 0.698803,
2024
+ "f1_weighted": 0.696601
2025
+ },
2026
+ {
2027
+ "accuracy": 0.715535,
2028
+ "f1": 0.705586,
2029
+ "f1_weighted": 0.707342
2030
+ },
2031
+ {
2032
+ "accuracy": 0.712845,
2033
+ "f1": 0.697723,
2034
+ "f1_weighted": 0.706737
2035
+ },
2036
+ {
2037
+ "accuracy": 0.697041,
2038
+ "f1": 0.701828,
2039
+ "f1_weighted": 0.685584
2040
+ },
2041
+ {
2042
+ "accuracy": 0.728985,
2043
+ "f1": 0.725927,
2044
+ "f1_weighted": 0.727752
2045
+ }
2046
+ ],
2047
+ "main_score": 0.718931,
2048
+ "hf_subset": "es",
2049
+ "languages": [
2050
+ "spa-Latn"
2051
+ ]
2052
+ },
2053
+ {
2054
+ "accuracy": 0.725387,
2055
+ "f1": 0.70935,
2056
+ "f1_weighted": 0.717343,
2057
+ "scores_per_experiment": [
2058
+ {
2059
+ "accuracy": 0.725622,
2060
+ "f1": 0.706471,
2061
+ "f1_weighted": 0.71179
2062
+ },
2063
+ {
2064
+ "accuracy": 0.756893,
2065
+ "f1": 0.73009,
2066
+ "f1_weighted": 0.750692
2067
+ },
2068
+ {
2069
+ "accuracy": 0.700403,
2070
+ "f1": 0.696784,
2071
+ "f1_weighted": 0.692638
2072
+ },
2073
+ {
2074
+ "accuracy": 0.746806,
2075
+ "f1": 0.722594,
2076
+ "f1_weighted": 0.739505
2077
+ },
2078
+ {
2079
+ "accuracy": 0.731675,
2080
+ "f1": 0.69835,
2081
+ "f1_weighted": 0.716853
2082
+ },
2083
+ {
2084
+ "accuracy": 0.706456,
2085
+ "f1": 0.700559,
2086
+ "f1_weighted": 0.700778
2087
+ },
2088
+ {
2089
+ "accuracy": 0.735373,
2090
+ "f1": 0.713632,
2091
+ "f1_weighted": 0.730311
2092
+ },
2093
+ {
2094
+ "accuracy": 0.714862,
2095
+ "f1": 0.696114,
2096
+ "f1_weighted": 0.706149
2097
+ },
2098
+ {
2099
+ "accuracy": 0.709482,
2100
+ "f1": 0.71334,
2101
+ "f1_weighted": 0.699246
2102
+ },
2103
+ {
2104
+ "accuracy": 0.726295,
2105
+ "f1": 0.715568,
2106
+ "f1_weighted": 0.725472
2107
+ }
2108
+ ],
2109
+ "main_score": 0.725387,
2110
+ "hf_subset": "pt",
2111
+ "languages": [
2112
+ "por-Latn"
2113
+ ]
2114
+ },
2115
+ {
2116
+ "accuracy": 0.728783,
2117
+ "f1": 0.710386,
2118
+ "f1_weighted": 0.723751,
2119
+ "scores_per_experiment": [
2120
+ {
2121
+ "accuracy": 0.735709,
2122
+ "f1": 0.709991,
2123
+ "f1_weighted": 0.725845
2124
+ },
2125
+ {
2126
+ "accuracy": 0.73302,
2127
+ "f1": 0.713946,
2128
+ "f1_weighted": 0.731563
2129
+ },
2130
+ {
2131
+ "accuracy": 0.724277,
2132
+ "f1": 0.702855,
2133
+ "f1_weighted": 0.716037
2134
+ },
2135
+ {
2136
+ "accuracy": 0.734028,
2137
+ "f1": 0.709732,
2138
+ "f1_weighted": 0.729839
2139
+ },
2140
+ {
2141
+ "accuracy": 0.729993,
2142
+ "f1": 0.70586,
2143
+ "f1_weighted": 0.721492
2144
+ },
2145
+ {
2146
+ "accuracy": 0.717216,
2147
+ "f1": 0.701956,
2148
+ "f1_weighted": 0.714075
2149
+ },
2150
+ {
2151
+ "accuracy": 0.738736,
2152
+ "f1": 0.718459,
2153
+ "f1_weighted": 0.734481
2154
+ },
2155
+ {
2156
+ "accuracy": 0.712508,
2157
+ "f1": 0.697816,
2158
+ "f1_weighted": 0.701737
2159
+ },
2160
+ {
2161
+ "accuracy": 0.712508,
2162
+ "f1": 0.707631,
2163
+ "f1_weighted": 0.712065
2164
+ },
2165
+ {
2166
+ "accuracy": 0.749832,
2167
+ "f1": 0.735614,
2168
+ "f1_weighted": 0.750379
2169
+ }
2170
+ ],
2171
+ "main_score": 0.728783,
2172
+ "hf_subset": "it",
2173
+ "languages": [
2174
+ "ita-Latn"
2175
+ ]
2176
+ },
2177
+ {
2178
+ "accuracy": 0.691022,
2179
+ "f1": 0.67089,
2180
+ "f1_weighted": 0.681113,
2181
+ "scores_per_experiment": [
2182
+ {
2183
+ "accuracy": 0.719906,
2184
+ "f1": 0.695377,
2185
+ "f1_weighted": 0.712102
2186
+ },
2187
+ {
2188
+ "accuracy": 0.700403,
2189
+ "f1": 0.677034,
2190
+ "f1_weighted": 0.697028
2191
+ },
2192
+ {
2193
+ "accuracy": 0.667787,
2194
+ "f1": 0.65731,
2195
+ "f1_weighted": 0.655992
2196
+ },
2197
+ {
2198
+ "accuracy": 0.712508,
2199
+ "f1": 0.677148,
2200
+ "f1_weighted": 0.703707
2201
+ },
2202
+ {
2203
+ "accuracy": 0.697041,
2204
+ "f1": 0.663457,
2205
+ "f1_weighted": 0.681182
2206
+ },
2207
+ {
2208
+ "accuracy": 0.673167,
2209
+ "f1": 0.662415,
2210
+ "f1_weighted": 0.667806
2211
+ },
2212
+ {
2213
+ "accuracy": 0.688971,
2214
+ "f1": 0.666819,
2215
+ "f1_weighted": 0.681291
2216
+ },
2217
+ {
2218
+ "accuracy": 0.67922,
2219
+ "f1": 0.657808,
2220
+ "f1_weighted": 0.669787
2221
+ },
2222
+ {
2223
+ "accuracy": 0.654674,
2224
+ "f1": 0.660592,
2225
+ "f1_weighted": 0.63393
2226
+ },
2227
+ {
2228
+ "accuracy": 0.716543,
2229
+ "f1": 0.690935,
2230
+ "f1_weighted": 0.7083
2231
+ }
2232
+ ],
2233
+ "main_score": 0.691022,
2234
+ "hf_subset": "ml",
2235
+ "languages": [
2236
+ "mal-Mlym"
2237
+ ]
2238
+ },
2239
+ {
2240
+ "accuracy": 0.697445,
2241
+ "f1": 0.675855,
2242
+ "f1_weighted": 0.688646,
2243
+ "scores_per_experiment": [
2244
+ {
2245
+ "accuracy": 0.712508,
2246
+ "f1": 0.690338,
2247
+ "f1_weighted": 0.704225
2248
+ },
2249
+ {
2250
+ "accuracy": 0.72495,
2251
+ "f1": 0.702775,
2252
+ "f1_weighted": 0.721127
2253
+ },
2254
+ {
2255
+ "accuracy": 0.68729,
2256
+ "f1": 0.671605,
2257
+ "f1_weighted": 0.683692
2258
+ },
2259
+ {
2260
+ "accuracy": 0.7115,
2261
+ "f1": 0.686616,
2262
+ "f1_weighted": 0.706376
2263
+ },
2264
+ {
2265
+ "accuracy": 0.688299,
2266
+ "f1": 0.649596,
2267
+ "f1_weighted": 0.674267
2268
+ },
2269
+ {
2270
+ "accuracy": 0.665434,
2271
+ "f1": 0.658816,
2272
+ "f1_weighted": 0.657869
2273
+ },
2274
+ {
2275
+ "accuracy": 0.696369,
2276
+ "f1": 0.675006,
2277
+ "f1_weighted": 0.684924
2278
+ },
2279
+ {
2280
+ "accuracy": 0.696032,
2281
+ "f1": 0.671825,
2282
+ "f1_weighted": 0.688093
2283
+ },
2284
+ {
2285
+ "accuracy": 0.676866,
2286
+ "f1": 0.660441,
2287
+ "f1_weighted": 0.659756
2288
+ },
2289
+ {
2290
+ "accuracy": 0.715198,
2291
+ "f1": 0.691536,
2292
+ "f1_weighted": 0.706136
2293
+ }
2294
+ ],
2295
+ "main_score": 0.697445,
2296
+ "hf_subset": "ms",
2297
+ "languages": [
2298
+ "msa-Latn"
2299
+ ]
2300
+ },
2301
+ {
2302
+ "accuracy": 0.596537,
2303
+ "f1": 0.551067,
2304
+ "f1_weighted": 0.588922,
2305
+ "scores_per_experiment": [
2306
+ {
2307
+ "accuracy": 0.623739,
2308
+ "f1": 0.56683,
2309
+ "f1_weighted": 0.615492
2310
+ },
2311
+ {
2312
+ "accuracy": 0.624412,
2313
+ "f1": 0.576169,
2314
+ "f1_weighted": 0.62253
2315
+ },
2316
+ {
2317
+ "accuracy": 0.605582,
2318
+ "f1": 0.553169,
2319
+ "f1_weighted": 0.609397
2320
+ },
2321
+ {
2322
+ "accuracy": 0.605918,
2323
+ "f1": 0.540036,
2324
+ "f1_weighted": 0.591683
2325
+ },
2326
+ {
2327
+ "accuracy": 0.588769,
2328
+ "f1": 0.545355,
2329
+ "f1_weighted": 0.573855
2330
+ },
2331
+ {
2332
+ "accuracy": 0.526564,
2333
+ "f1": 0.513725,
2334
+ "f1_weighted": 0.524259
2335
+ },
2336
+ {
2337
+ "accuracy": 0.603228,
2338
+ "f1": 0.54524,
2339
+ "f1_weighted": 0.598469
2340
+ },
2341
+ {
2342
+ "accuracy": 0.586079,
2343
+ "f1": 0.548829,
2344
+ "f1_weighted": 0.577136
2345
+ },
2346
+ {
2347
+ "accuracy": 0.578346,
2348
+ "f1": 0.546686,
2349
+ "f1_weighted": 0.559307
2350
+ },
2351
+ {
2352
+ "accuracy": 0.62273,
2353
+ "f1": 0.574627,
2354
+ "f1_weighted": 0.617095
2355
+ }
2356
+ ],
2357
+ "main_score": 0.596537,
2358
+ "hf_subset": "sq",
2359
+ "languages": [
2360
+ "sqi-Latn"
2361
+ ]
2362
+ },
2363
+ {
2364
+ "accuracy": 0.557666,
2365
+ "f1": 0.542831,
2366
+ "f1_weighted": 0.545647,
2367
+ "scores_per_experiment": [
2368
+ {
2369
+ "accuracy": 0.568931,
2370
+ "f1": 0.563736,
2371
+ "f1_weighted": 0.55755
2372
+ },
2373
+ {
2374
+ "accuracy": 0.567586,
2375
+ "f1": 0.558986,
2376
+ "f1_weighted": 0.561806
2377
+ },
2378
+ {
2379
+ "accuracy": 0.559852,
2380
+ "f1": 0.544974,
2381
+ "f1_weighted": 0.546728
2382
+ },
2383
+ {
2384
+ "accuracy": 0.567922,
2385
+ "f1": 0.545523,
2386
+ "f1_weighted": 0.559809
2387
+ },
2388
+ {
2389
+ "accuracy": 0.53766,
2390
+ "f1": 0.523463,
2391
+ "f1_weighted": 0.521845
2392
+ },
2393
+ {
2394
+ "accuracy": 0.542031,
2395
+ "f1": 0.544435,
2396
+ "f1_weighted": 0.531745
2397
+ },
2398
+ {
2399
+ "accuracy": 0.565568,
2400
+ "f1": 0.551814,
2401
+ "f1_weighted": 0.551179
2402
+ },
2403
+ {
2404
+ "accuracy": 0.560861,
2405
+ "f1": 0.531281,
2406
+ "f1_weighted": 0.552972
2407
+ },
2408
+ {
2409
+ "accuracy": 0.539677,
2410
+ "f1": 0.525248,
2411
+ "f1_weighted": 0.516833
2412
+ },
2413
+ {
2414
+ "accuracy": 0.566577,
2415
+ "f1": 0.538847,
2416
+ "f1_weighted": 0.555997
2417
+ }
2418
+ ],
2419
+ "main_score": 0.557666,
2420
+ "hf_subset": "ka",
2421
+ "languages": [
2422
+ "kat-Geor"
2423
+ ]
2424
+ },
2425
+ {
2426
+ "accuracy": 0.649227,
2427
+ "f1": 0.623755,
2428
+ "f1_weighted": 0.637336,
2429
+ "scores_per_experiment": [
2430
+ {
2431
+ "accuracy": 0.661399,
2432
+ "f1": 0.634004,
2433
+ "f1_weighted": 0.651984
2434
+ },
2435
+ {
2436
+ "accuracy": 0.691997,
2437
+ "f1": 0.658667,
2438
+ "f1_weighted": 0.687398
2439
+ },
2440
+ {
2441
+ "accuracy": 0.64963,
2442
+ "f1": 0.615741,
2443
+ "f1_weighted": 0.640461
2444
+ },
2445
+ {
2446
+ "accuracy": 0.670141,
2447
+ "f1": 0.628085,
2448
+ "f1_weighted": 0.662325
2449
+ },
2450
+ {
2451
+ "accuracy": 0.64963,
2452
+ "f1": 0.617042,
2453
+ "f1_weighted": 0.634566
2454
+ },
2455
+ {
2456
+ "accuracy": 0.630128,
2457
+ "f1": 0.619937,
2458
+ "f1_weighted": 0.618841
2459
+ },
2460
+ {
2461
+ "accuracy": 0.63618,
2462
+ "f1": 0.615425,
2463
+ "f1_weighted": 0.617471
2464
+ },
2465
+ {
2466
+ "accuracy": 0.629792,
2467
+ "f1": 0.609171,
2468
+ "f1_weighted": 0.614973
2469
+ },
2470
+ {
2471
+ "accuracy": 0.618695,
2472
+ "f1": 0.613094,
2473
+ "f1_weighted": 0.601236
2474
+ },
2475
+ {
2476
+ "accuracy": 0.654674,
2477
+ "f1": 0.626378,
2478
+ "f1_weighted": 0.644099
2479
+ }
2480
+ ],
2481
+ "main_score": 0.649227,
2482
+ "hf_subset": "tl",
2483
+ "languages": [
2484
+ "tgl-Latn"
2485
+ ]
2486
+ },
2487
+ {
2488
+ "accuracy": 0.653295,
2489
+ "f1": 0.619971,
2490
+ "f1_weighted": 0.644938,
2491
+ "scores_per_experiment": [
2492
+ {
2493
+ "accuracy": 0.659045,
2494
+ "f1": 0.624727,
2495
+ "f1_weighted": 0.649021
2496
+ },
2497
+ {
2498
+ "accuracy": 0.686954,
2499
+ "f1": 0.646185,
2500
+ "f1_weighted": 0.682806
2501
+ },
2502
+ {
2503
+ "accuracy": 0.647276,
2504
+ "f1": 0.622633,
2505
+ "f1_weighted": 0.639633
2506
+ },
2507
+ {
2508
+ "accuracy": 0.662744,
2509
+ "f1": 0.615852,
2510
+ "f1_weighted": 0.660893
2511
+ },
2512
+ {
2513
+ "accuracy": 0.675521,
2514
+ "f1": 0.613898,
2515
+ "f1_weighted": 0.660338
2516
+ },
2517
+ {
2518
+ "accuracy": 0.615669,
2519
+ "f1": 0.60868,
2520
+ "f1_weighted": 0.610334
2521
+ },
2522
+ {
2523
+ "accuracy": 0.645259,
2524
+ "f1": 0.621509,
2525
+ "f1_weighted": 0.640316
2526
+ },
2527
+ {
2528
+ "accuracy": 0.649966,
2529
+ "f1": 0.614332,
2530
+ "f1_weighted": 0.63948
2531
+ },
2532
+ {
2533
+ "accuracy": 0.619032,
2534
+ "f1": 0.597357,
2535
+ "f1_weighted": 0.600454
2536
+ },
2537
+ {
2538
+ "accuracy": 0.671486,
2539
+ "f1": 0.634536,
2540
+ "f1_weighted": 0.666102
2541
+ }
2542
+ ],
2543
+ "main_score": 0.653295,
2544
+ "hf_subset": "af",
2545
+ "languages": [
2546
+ "afr-Latn"
2547
+ ]
2548
+ },
2549
+ {
2550
+ "accuracy": 0.687088,
2551
+ "f1": 0.675291,
2552
+ "f1_weighted": 0.67831,
2553
+ "scores_per_experiment": [
2554
+ {
2555
+ "accuracy": 0.697377,
2556
+ "f1": 0.676918,
2557
+ "f1_weighted": 0.687669
2558
+ },
2559
+ {
2560
+ "accuracy": 0.706792,
2561
+ "f1": 0.687296,
2562
+ "f1_weighted": 0.703316
2563
+ },
2564
+ {
2565
+ "accuracy": 0.675185,
2566
+ "f1": 0.672247,
2567
+ "f1_weighted": 0.666015
2568
+ },
2569
+ {
2570
+ "accuracy": 0.714526,
2571
+ "f1": 0.684176,
2572
+ "f1_weighted": 0.70631
2573
+ },
2574
+ {
2575
+ "accuracy": 0.687626,
2576
+ "f1": 0.66715,
2577
+ "f1_weighted": 0.674305
2578
+ },
2579
+ {
2580
+ "accuracy": 0.668796,
2581
+ "f1": 0.662125,
2582
+ "f1_weighted": 0.66209
2583
+ },
2584
+ {
2585
+ "accuracy": 0.686617,
2586
+ "f1": 0.673123,
2587
+ "f1_weighted": 0.678493
2588
+ },
2589
+ {
2590
+ "accuracy": 0.669805,
2591
+ "f1": 0.667244,
2592
+ "f1_weighted": 0.661804
2593
+ },
2594
+ {
2595
+ "accuracy": 0.653665,
2596
+ "f1": 0.663897,
2597
+ "f1_weighted": 0.638684
2598
+ },
2599
+ {
2600
+ "accuracy": 0.710491,
2601
+ "f1": 0.698734,
2602
+ "f1_weighted": 0.704413
2603
+ }
2604
+ ],
2605
+ "main_score": 0.687088,
2606
+ "hf_subset": "el",
2607
+ "languages": [
2608
+ "ell-Grek"
2609
+ ]
2610
+ },
2611
+ {
2612
+ "accuracy": 0.635709,
2613
+ "f1": 0.605936,
2614
+ "f1_weighted": 0.628267,
2615
+ "scores_per_experiment": [
2616
+ {
2617
+ "accuracy": 0.63887,
2618
+ "f1": 0.603684,
2619
+ "f1_weighted": 0.629056
2620
+ },
2621
+ {
2622
+ "accuracy": 0.645595,
2623
+ "f1": 0.622245,
2624
+ "f1_weighted": 0.643521
2625
+ },
2626
+ {
2627
+ "accuracy": 0.620377,
2628
+ "f1": 0.595589,
2629
+ "f1_weighted": 0.614765
2630
+ },
2631
+ {
2632
+ "accuracy": 0.657028,
2633
+ "f1": 0.609409,
2634
+ "f1_weighted": 0.649887
2635
+ },
2636
+ {
2637
+ "accuracy": 0.637189,
2638
+ "f1": 0.601531,
2639
+ "f1_weighted": 0.625085
2640
+ },
2641
+ {
2642
+ "accuracy": 0.618695,
2643
+ "f1": 0.58998,
2644
+ "f1_weighted": 0.619514
2645
+ },
2646
+ {
2647
+ "accuracy": 0.642569,
2648
+ "f1": 0.615642,
2649
+ "f1_weighted": 0.633977
2650
+ },
2651
+ {
2652
+ "accuracy": 0.624748,
2653
+ "f1": 0.598284,
2654
+ "f1_weighted": 0.614785
2655
+ },
2656
+ {
2657
+ "accuracy": 0.61197,
2658
+ "f1": 0.592315,
2659
+ "f1_weighted": 0.596814
2660
+ },
2661
+ {
2662
+ "accuracy": 0.660054,
2663
+ "f1": 0.630678,
2664
+ "f1_weighted": 0.655269
2665
+ }
2666
+ ],
2667
+ "main_score": 0.635709,
2668
+ "hf_subset": "sl",
2669
+ "languages": [
2670
+ "slv-Latn"
2671
+ ]
2672
+ },
2673
+ {
2674
+ "accuracy": 0.694553,
2675
+ "f1": 0.66892,
2676
+ "f1_weighted": 0.686624,
2677
+ "scores_per_experiment": [
2678
+ {
2679
+ "accuracy": 0.714862,
2680
+ "f1": 0.683374,
2681
+ "f1_weighted": 0.705545
2682
+ },
2683
+ {
2684
+ "accuracy": 0.714526,
2685
+ "f1": 0.692251,
2686
+ "f1_weighted": 0.712118
2687
+ },
2688
+ {
2689
+ "accuracy": 0.682246,
2690
+ "f1": 0.66038,
2691
+ "f1_weighted": 0.672797
2692
+ },
2693
+ {
2694
+ "accuracy": 0.71688,
2695
+ "f1": 0.671796,
2696
+ "f1_weighted": 0.711056
2697
+ },
2698
+ {
2699
+ "accuracy": 0.69805,
2700
+ "f1": 0.655524,
2701
+ "f1_weighted": 0.684545
2702
+ },
2703
+ {
2704
+ "accuracy": 0.667787,
2705
+ "f1": 0.659033,
2706
+ "f1_weighted": 0.658849
2707
+ },
2708
+ {
2709
+ "accuracy": 0.704775,
2710
+ "f1": 0.676065,
2711
+ "f1_weighted": 0.697443
2712
+ },
2713
+ {
2714
+ "accuracy": 0.683927,
2715
+ "f1": 0.65939,
2716
+ "f1_weighted": 0.678164
2717
+ },
2718
+ {
2719
+ "accuracy": 0.659045,
2720
+ "f1": 0.65943,
2721
+ "f1_weighted": 0.646794
2722
+ },
2723
+ {
2724
+ "accuracy": 0.70343,
2725
+ "f1": 0.671954,
2726
+ "f1_weighted": 0.698927
2727
+ }
2728
+ ],
2729
+ "main_score": 0.694553,
2730
+ "hf_subset": "da",
2731
+ "languages": [
2732
+ "dan-Latn"
2733
+ ]
2734
+ },
2735
+ {
2736
+ "accuracy": 0.658339,
2737
+ "f1": 0.627172,
2738
+ "f1_weighted": 0.652182,
2739
+ "scores_per_experiment": [
2740
+ {
2741
+ "accuracy": 0.666443,
2742
+ "f1": 0.634169,
2743
+ "f1_weighted": 0.658522
2744
+ },
2745
+ {
2746
+ "accuracy": 0.663753,
2747
+ "f1": 0.631552,
2748
+ "f1_weighted": 0.664054
2749
+ },
2750
+ {
2751
+ "accuracy": 0.641224,
2752
+ "f1": 0.616543,
2753
+ "f1_weighted": 0.636446
2754
+ },
2755
+ {
2756
+ "accuracy": 0.669469,
2757
+ "f1": 0.624636,
2758
+ "f1_weighted": 0.665993
2759
+ },
2760
+ {
2761
+ "accuracy": 0.661399,
2762
+ "f1": 0.618999,
2763
+ "f1_weighted": 0.646889
2764
+ },
2765
+ {
2766
+ "accuracy": 0.624748,
2767
+ "f1": 0.608532,
2768
+ "f1_weighted": 0.619884
2769
+ },
2770
+ {
2771
+ "accuracy": 0.67115,
2772
+ "f1": 0.633271,
2773
+ "f1_weighted": 0.66443
2774
+ },
2775
+ {
2776
+ "accuracy": 0.658373,
2777
+ "f1": 0.626174,
2778
+ "f1_weighted": 0.652448
2779
+ },
2780
+ {
2781
+ "accuracy": 0.66577,
2782
+ "f1": 0.645104,
2783
+ "f1_weighted": 0.657549
2784
+ },
2785
+ {
2786
+ "accuracy": 0.661063,
2787
+ "f1": 0.63274,
2788
+ "f1_weighted": 0.655604
2789
+ }
2790
+ ],
2791
+ "main_score": 0.658339,
2792
+ "hf_subset": "hy",
2793
+ "languages": [
2794
+ "hye-Armn"
2795
+ ]
2796
+ },
2797
+ {
2798
+ "accuracy": 0.707532,
2799
+ "f1": 0.675772,
2800
+ "f1_weighted": 0.70183,
2801
+ "scores_per_experiment": [
2802
+ {
2803
+ "accuracy": 0.715871,
2804
+ "f1": 0.675946,
2805
+ "f1_weighted": 0.7065
2806
+ },
2807
+ {
2808
+ "accuracy": 0.723268,
2809
+ "f1": 0.694953,
2810
+ "f1_weighted": 0.723596
2811
+ },
2812
+ {
2813
+ "accuracy": 0.702421,
2814
+ "f1": 0.662523,
2815
+ "f1_weighted": 0.696113
2816
+ },
2817
+ {
2818
+ "accuracy": 0.714862,
2819
+ "f1": 0.682766,
2820
+ "f1_weighted": 0.709688
2821
+ },
2822
+ {
2823
+ "accuracy": 0.715198,
2824
+ "f1": 0.673396,
2825
+ "f1_weighted": 0.70324
2826
+ },
2827
+ {
2828
+ "accuracy": 0.698386,
2829
+ "f1": 0.674501,
2830
+ "f1_weighted": 0.697011
2831
+ },
2832
+ {
2833
+ "accuracy": 0.703093,
2834
+ "f1": 0.670175,
2835
+ "f1_weighted": 0.694681
2836
+ },
2837
+ {
2838
+ "accuracy": 0.707128,
2839
+ "f1": 0.670842,
2840
+ "f1_weighted": 0.705715
2841
+ },
2842
+ {
2843
+ "accuracy": 0.684936,
2844
+ "f1": 0.668252,
2845
+ "f1_weighted": 0.672412
2846
+ },
2847
+ {
2848
+ "accuracy": 0.710155,
2849
+ "f1": 0.684363,
2850
+ "f1_weighted": 0.709347
2851
+ }
2852
+ ],
2853
+ "main_score": 0.707532,
2854
+ "hf_subset": "tr",
2855
+ "languages": [
2856
+ "tur-Latn"
2857
+ ]
2858
+ },
2859
+ {
2860
+ "accuracy": 0.728816,
2861
+ "f1": 0.703941,
2862
+ "f1_weighted": 0.719995,
2863
+ "scores_per_experiment": [
2864
+ {
2865
+ "accuracy": 0.748487,
2866
+ "f1": 0.719296,
2867
+ "f1_weighted": 0.737945
2868
+ },
2869
+ {
2870
+ "accuracy": 0.752186,
2871
+ "f1": 0.717815,
2872
+ "f1_weighted": 0.748628
2873
+ },
2874
+ {
2875
+ "accuracy": 0.713181,
2876
+ "f1": 0.683974,
2877
+ "f1_weighted": 0.707161
2878
+ },
2879
+ {
2880
+ "accuracy": 0.755548,
2881
+ "f1": 0.715324,
2882
+ "f1_weighted": 0.749232
2883
+ },
2884
+ {
2885
+ "accuracy": 0.730666,
2886
+ "f1": 0.702097,
2887
+ "f1_weighted": 0.714866
2888
+ },
2889
+ {
2890
+ "accuracy": 0.705447,
2891
+ "f1": 0.685211,
2892
+ "f1_weighted": 0.700809
2893
+ },
2894
+ {
2895
+ "accuracy": 0.720242,
2896
+ "f1": 0.700763,
2897
+ "f1_weighted": 0.711392
2898
+ },
2899
+ {
2900
+ "accuracy": 0.712172,
2901
+ "f1": 0.691436,
2902
+ "f1_weighted": 0.703948
2903
+ },
2904
+ {
2905
+ "accuracy": 0.692334,
2906
+ "f1": 0.693357,
2907
+ "f1_weighted": 0.671312
2908
+ },
2909
+ {
2910
+ "accuracy": 0.757902,
2911
+ "f1": 0.73014,
2912
+ "f1_weighted": 0.754654
2913
+ }
2914
+ ],
2915
+ "main_score": 0.728816,
2916
+ "hf_subset": "id",
2917
+ "languages": [
2918
+ "ind-Latn"
2919
+ ]
2920
+ },
2921
+ {
2922
+ "accuracy": 0.690585,
2923
+ "f1": 0.660243,
2924
+ "f1_weighted": 0.683327,
2925
+ "scores_per_experiment": [
2926
+ {
2927
+ "accuracy": 0.696705,
2928
+ "f1": 0.661086,
2929
+ "f1_weighted": 0.692167
2930
+ },
2931
+ {
2932
+ "accuracy": 0.709818,
2933
+ "f1": 0.677275,
2934
+ "f1_weighted": 0.704185
2935
+ },
2936
+ {
2937
+ "accuracy": 0.67922,
2938
+ "f1": 0.655812,
2939
+ "f1_weighted": 0.673753
2940
+ },
2941
+ {
2942
+ "accuracy": 0.70343,
2943
+ "f1": 0.654989,
2944
+ "f1_weighted": 0.698795
2945
+ },
2946
+ {
2947
+ "accuracy": 0.687962,
2948
+ "f1": 0.647085,
2949
+ "f1_weighted": 0.675318
2950
+ },
2951
+ {
2952
+ "accuracy": 0.67653,
2953
+ "f1": 0.648567,
2954
+ "f1_weighted": 0.670687
2955
+ },
2956
+ {
2957
+ "accuracy": 0.703093,
2958
+ "f1": 0.671673,
2959
+ "f1_weighted": 0.698751
2960
+ },
2961
+ {
2962
+ "accuracy": 0.678884,
2963
+ "f1": 0.654147,
2964
+ "f1_weighted": 0.671696
2965
+ },
2966
+ {
2967
+ "accuracy": 0.663416,
2968
+ "f1": 0.653404,
2969
+ "f1_weighted": 0.641974
2970
+ },
2971
+ {
2972
+ "accuracy": 0.706792,
2973
+ "f1": 0.678392,
2974
+ "f1_weighted": 0.705949
2975
+ }
2976
+ ],
2977
+ "main_score": 0.690585,
2978
+ "hf_subset": "nb",
2979
+ "languages": [
2980
+ "nob-Latn"
2981
+ ]
2982
+ },
2983
+ {
2984
+ "accuracy": 0.592603,
2985
+ "f1": 0.568022,
2986
+ "f1_weighted": 0.587304,
2987
+ "scores_per_experiment": [
2988
+ {
2989
+ "accuracy": 0.590451,
2990
+ "f1": 0.568634,
2991
+ "f1_weighted": 0.588625
2992
+ },
2993
+ {
2994
+ "accuracy": 0.588097,
2995
+ "f1": 0.560717,
2996
+ "f1_weighted": 0.588648
2997
+ },
2998
+ {
2999
+ "accuracy": 0.603564,
3000
+ "f1": 0.570794,
3001
+ "f1_weighted": 0.603322
3002
+ },
3003
+ {
3004
+ "accuracy": 0.62811,
3005
+ "f1": 0.592963,
3006
+ "f1_weighted": 0.623476
3007
+ },
3008
+ {
3009
+ "accuracy": 0.594149,
3010
+ "f1": 0.560619,
3011
+ "f1_weighted": 0.585212
3012
+ },
3013
+ {
3014
+ "accuracy": 0.580363,
3015
+ "f1": 0.564528,
3016
+ "f1_weighted": 0.578614
3017
+ },
3018
+ {
3019
+ "accuracy": 0.575656,
3020
+ "f1": 0.563518,
3021
+ "f1_weighted": 0.566778
3022
+ },
3023
+ {
3024
+ "accuracy": 0.596503,
3025
+ "f1": 0.558898,
3026
+ "f1_weighted": 0.590128
3027
+ },
3028
+ {
3029
+ "accuracy": 0.575656,
3030
+ "f1": 0.568616,
3031
+ "f1_weighted": 0.564214
3032
+ },
3033
+ {
3034
+ "accuracy": 0.593477,
3035
+ "f1": 0.570929,
3036
+ "f1_weighted": 0.584026
3037
+ }
3038
+ ],
3039
+ "main_score": 0.592603,
3040
+ "hf_subset": "mn",
3041
+ "languages": [
3042
+ "mon-Cyrl"
3043
+ ]
3044
+ },
3045
+ {
3046
+ "accuracy": 0.676597,
3047
+ "f1": 0.656777,
3048
+ "f1_weighted": 0.670983,
3049
+ "scores_per_experiment": [
3050
+ {
3051
+ "accuracy": 0.689644,
3052
+ "f1": 0.670022,
3053
+ "f1_weighted": 0.683783
3054
+ },
3055
+ {
3056
+ "accuracy": 0.701748,
3057
+ "f1": 0.68194,
3058
+ "f1_weighted": 0.704319
3059
+ },
3060
+ {
3061
+ "accuracy": 0.658373,
3062
+ "f1": 0.641828,
3063
+ "f1_weighted": 0.65372
3064
+ },
3065
+ {
3066
+ "accuracy": 0.688635,
3067
+ "f1": 0.653288,
3068
+ "f1_weighted": 0.6841
3069
+ },
3070
+ {
3071
+ "accuracy": 0.681574,
3072
+ "f1": 0.65479,
3073
+ "f1_weighted": 0.669087
3074
+ },
3075
+ {
3076
+ "accuracy": 0.651648,
3077
+ "f1": 0.64675,
3078
+ "f1_weighted": 0.648097
3079
+ },
3080
+ {
3081
+ "accuracy": 0.674512,
3082
+ "f1": 0.655595,
3083
+ "f1_weighted": 0.668846
3084
+ },
3085
+ {
3086
+ "accuracy": 0.662408,
3087
+ "f1": 0.64486,
3088
+ "f1_weighted": 0.653161
3089
+ },
3090
+ {
3091
+ "accuracy": 0.680229,
3092
+ "f1": 0.664174,
3093
+ "f1_weighted": 0.670404
3094
+ },
3095
+ {
3096
+ "accuracy": 0.677202,
3097
+ "f1": 0.65452,
3098
+ "f1_weighted": 0.674314
3099
+ }
3100
+ ],
3101
+ "main_score": 0.676597,
3102
+ "hf_subset": "bn",
3103
+ "languages": [
3104
+ "ben-Beng"
3105
+ ]
3106
+ },
3107
+ {
3108
+ "accuracy": 0.653329,
3109
+ "f1": 0.631046,
3110
+ "f1_weighted": 0.647876,
3111
+ "scores_per_experiment": [
3112
+ {
3113
+ "accuracy": 0.664761,
3114
+ "f1": 0.641991,
3115
+ "f1_weighted": 0.656591
3116
+ },
3117
+ {
3118
+ "accuracy": 0.662408,
3119
+ "f1": 0.641014,
3120
+ "f1_weighted": 0.659413
3121
+ },
3122
+ {
3123
+ "accuracy": 0.645259,
3124
+ "f1": 0.621774,
3125
+ "f1_weighted": 0.641575
3126
+ },
3127
+ {
3128
+ "accuracy": 0.656019,
3129
+ "f1": 0.620762,
3130
+ "f1_weighted": 0.655316
3131
+ },
3132
+ {
3133
+ "accuracy": 0.642905,
3134
+ "f1": 0.606241,
3135
+ "f1_weighted": 0.63159
3136
+ },
3137
+ {
3138
+ "accuracy": 0.63887,
3139
+ "f1": 0.631482,
3140
+ "f1_weighted": 0.634978
3141
+ },
3142
+ {
3143
+ "accuracy": 0.661063,
3144
+ "f1": 0.64097,
3145
+ "f1_weighted": 0.655563
3146
+ },
3147
+ {
3148
+ "accuracy": 0.657364,
3149
+ "f1": 0.645564,
3150
+ "f1_weighted": 0.649317
3151
+ },
3152
+ {
3153
+ "accuracy": 0.64425,
3154
+ "f1": 0.624895,
3155
+ "f1_weighted": 0.635448
3156
+ },
3157
+ {
3158
+ "accuracy": 0.66039,
3159
+ "f1": 0.635765,
3160
+ "f1_weighted": 0.658974
3161
+ }
3162
+ ],
3163
+ "main_score": 0.653329,
3164
+ "hf_subset": "ta",
3165
+ "languages": [
3166
+ "tam-Taml"
3167
+ ]
3168
+ }
3169
+ ]
3170
+ },
3171
+ "evaluation_time": 16697.124433755875,
3172
+ "kg_co2_emissions": null
3173
+ }