sergeyz-zh commited on
Commit
ec7b3fc
·
unverified ·
1 Parent(s): 8104baf

Add results for rubert-mini-frida, BERTA (#152)

Browse files

* Add results for rubert-mini-frida model

* Add results for BERTA model

Files changed (48) hide show
  1. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/CEDRClassification.json +73 -0
  2. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/GeoreviewClassification.json +73 -0
  3. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/GeoreviewClusteringP2P.json +34 -0
  4. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/HeadlineClassification.json +73 -0
  5. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/InappropriatenessClassification.json +95 -0
  6. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/KinopoiskClassification.json +73 -0
  7. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/MIRACLReranking.json +130 -0
  8. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/MIRACLRetrieval.json +158 -0
  9. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/MassiveIntentClassification.json +137 -0
  10. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/MassiveScenarioClassification.json +137 -0
  11. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RUParaPhraserSTS.json +26 -0
  12. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RiaNewsRetrieval.json +158 -0
  13. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuBQReranking.json +26 -0
  14. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuBQRetrieval.json +158 -0
  15. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuReviewsClassification.json +73 -0
  16. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuSTSBenchmarkSTS.json +26 -0
  17. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuSciBenchGRNTIClassification.json +73 -0
  18. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuSciBenchGRNTIClusteringP2P.json +34 -0
  19. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuSciBenchOECDClassification.json +73 -0
  20. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuSciBenchOECDClusteringP2P.json +34 -0
  21. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/STS22.json +26 -0
  22. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/SensitiveTopicsClassification.json +73 -0
  23. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/TERRa.json +58 -0
  24. results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/model_meta.json +1 -0
  25. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/CEDRClassification.json +73 -0
  26. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/GeoreviewClassification.json +73 -0
  27. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/GeoreviewClusteringP2P.json +34 -0
  28. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/HeadlineClassification.json +73 -0
  29. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/InappropriatenessClassification.json +95 -0
  30. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/KinopoiskClassification.json +73 -0
  31. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/MIRACLReranking.json +130 -0
  32. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/MIRACLRetrieval.json +158 -0
  33. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/MassiveIntentClassification.json +137 -0
  34. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/MassiveScenarioClassification.json +137 -0
  35. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RUParaPhraserSTS.json +26 -0
  36. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RiaNewsRetrieval.json +158 -0
  37. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuBQReranking.json +26 -0
  38. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuBQRetrieval.json +158 -0
  39. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuReviewsClassification.json +73 -0
  40. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuSTSBenchmarkSTS.json +26 -0
  41. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuSciBenchGRNTIClassification.json +73 -0
  42. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuSciBenchGRNTIClusteringP2P.json +34 -0
  43. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuSciBenchOECDClassification.json +73 -0
  44. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuSciBenchOECDClusteringP2P.json +34 -0
  45. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/STS22.json +26 -0
  46. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/SensitiveTopicsClassification.json +73 -0
  47. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/TERRa.json +58 -0
  48. results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/model_meta.json +1 -0
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/CEDRClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "c0ba03d058e3e1b2f3fd20518875a4563dd12db4",
3
+ "task_name": "CEDRClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.730074,
9
+ "f1": 0.737591,
10
+ "lrap": 0.905699,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.789586,
14
+ "f1": 0.749645,
15
+ "lrap": 0.899097
16
+ },
17
+ {
18
+ "accuracy": 0.765675,
19
+ "f1": 0.740541,
20
+ "lrap": 0.903773
21
+ },
22
+ {
23
+ "accuracy": 0.803401,
24
+ "f1": 0.754733,
25
+ "lrap": 0.910043
26
+ },
27
+ {
28
+ "accuracy": 0.766738,
29
+ "f1": 0.738748,
30
+ "lrap": 0.905367
31
+ },
32
+ {
33
+ "accuracy": 0.792774,
34
+ "f1": 0.754356,
35
+ "lrap": 0.894102
36
+ },
37
+ {
38
+ "accuracy": 0.510096,
39
+ "f1": 0.683943,
40
+ "lrap": 0.909989
41
+ },
42
+ {
43
+ "accuracy": 0.538257,
44
+ "f1": 0.701587,
45
+ "lrap": 0.919129
46
+ },
47
+ {
48
+ "accuracy": 0.81509,
49
+ "f1": 0.768223,
50
+ "lrap": 0.904251
51
+ },
52
+ {
53
+ "accuracy": 0.725292,
54
+ "f1": 0.730902,
55
+ "lrap": 0.914506
56
+ },
57
+ {
58
+ "accuracy": 0.793836,
59
+ "f1": 0.75323,
60
+ "lrap": 0.896732
61
+ }
62
+ ],
63
+ "main_score": 0.730074,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 5.003009557723999,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/GeoreviewClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3765c0d1de6b7d264bc459433c45e5a75513839c",
3
+ "task_name": "GeoreviewClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.548438,
9
+ "f1": 0.50672,
10
+ "f1_weighted": 0.506644,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.518066,
14
+ "f1": 0.450777,
15
+ "f1_weighted": 0.450671
16
+ },
17
+ {
18
+ "accuracy": 0.561035,
19
+ "f1": 0.51939,
20
+ "f1_weighted": 0.519296
21
+ },
22
+ {
23
+ "accuracy": 0.541016,
24
+ "f1": 0.4955,
25
+ "f1_weighted": 0.495427
26
+ },
27
+ {
28
+ "accuracy": 0.545898,
29
+ "f1": 0.494639,
30
+ "f1_weighted": 0.494564
31
+ },
32
+ {
33
+ "accuracy": 0.557617,
34
+ "f1": 0.510934,
35
+ "f1_weighted": 0.510836
36
+ },
37
+ {
38
+ "accuracy": 0.542969,
39
+ "f1": 0.500696,
40
+ "f1_weighted": 0.500633
41
+ },
42
+ {
43
+ "accuracy": 0.571777,
44
+ "f1": 0.556992,
45
+ "f1_weighted": 0.556939
46
+ },
47
+ {
48
+ "accuracy": 0.565918,
49
+ "f1": 0.533864,
50
+ "f1_weighted": 0.533798
51
+ },
52
+ {
53
+ "accuracy": 0.541016,
54
+ "f1": 0.506626,
55
+ "f1_weighted": 0.506555
56
+ },
57
+ {
58
+ "accuracy": 0.539062,
59
+ "f1": 0.49778,
60
+ "f1_weighted": 0.49772
61
+ }
62
+ ],
63
+ "main_score": 0.548438,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 18.44209313392639,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/GeoreviewClusteringP2P.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "97a313c8fc85b47f13f33e7e9a95c1ad888c7fec",
3
+ "task_name": "GeoreviewClusteringP2P",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.733451,
11
+ 0.735862,
12
+ 0.735497,
13
+ 0.733281,
14
+ 0.741149,
15
+ 0.741435,
16
+ 0.735931,
17
+ 0.736518,
18
+ 0.737075,
19
+ 0.745947
20
+ ]
21
+ },
22
+ "v_measure": 0.737615,
23
+ "v_measure_std": 0.003802,
24
+ "main_score": 0.737615,
25
+ "hf_subset": "default",
26
+ "languages": [
27
+ "rus-Cyrl"
28
+ ]
29
+ }
30
+ ]
31
+ },
32
+ "evaluation_time": 13.0330491065979,
33
+ "kg_co2_emissions": null
34
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/HeadlineClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "2fe05ee6b5832cda29f2ef7aaad7b7fe6a3609eb",
3
+ "task_name": "HeadlineClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.890186,
9
+ "f1": 0.889888,
10
+ "f1_weighted": 0.88988,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.888184,
14
+ "f1": 0.887833,
15
+ "f1_weighted": 0.887826
16
+ },
17
+ {
18
+ "accuracy": 0.893555,
19
+ "f1": 0.893302,
20
+ "f1_weighted": 0.893294
21
+ },
22
+ {
23
+ "accuracy": 0.889648,
24
+ "f1": 0.889417,
25
+ "f1_weighted": 0.889411
26
+ },
27
+ {
28
+ "accuracy": 0.895508,
29
+ "f1": 0.895344,
30
+ "f1_weighted": 0.895339
31
+ },
32
+ {
33
+ "accuracy": 0.894043,
34
+ "f1": 0.893852,
35
+ "f1_weighted": 0.893845
36
+ },
37
+ {
38
+ "accuracy": 0.88623,
39
+ "f1": 0.885914,
40
+ "f1_weighted": 0.885901
41
+ },
42
+ {
43
+ "accuracy": 0.887207,
44
+ "f1": 0.886793,
45
+ "f1_weighted": 0.886786
46
+ },
47
+ {
48
+ "accuracy": 0.88916,
49
+ "f1": 0.888813,
50
+ "f1_weighted": 0.888808
51
+ },
52
+ {
53
+ "accuracy": 0.887207,
54
+ "f1": 0.886801,
55
+ "f1_weighted": 0.88679
56
+ },
57
+ {
58
+ "accuracy": 0.891113,
59
+ "f1": 0.890808,
60
+ "f1_weighted": 0.890801
61
+ }
62
+ ],
63
+ "main_score": 0.890186,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 7.147644281387329,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/InappropriatenessClassification.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "601651fdc45ef243751676e62dd7a19f491c0285",
3
+ "task_name": "InappropriatenessClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.747803,
9
+ "f1": 0.745933,
10
+ "f1_weighted": 0.745933,
11
+ "ap": 0.692164,
12
+ "ap_weighted": 0.692164,
13
+ "scores_per_experiment": [
14
+ {
15
+ "accuracy": 0.769043,
16
+ "f1": 0.768545,
17
+ "f1_weighted": 0.768545,
18
+ "ap": 0.714308,
19
+ "ap_weighted": 0.714308
20
+ },
21
+ {
22
+ "accuracy": 0.721191,
23
+ "f1": 0.720687,
24
+ "f1_weighted": 0.720687,
25
+ "ap": 0.664064,
26
+ "ap_weighted": 0.664064
27
+ },
28
+ {
29
+ "accuracy": 0.749512,
30
+ "f1": 0.744422,
31
+ "f1_weighted": 0.744422,
32
+ "ap": 0.711491,
33
+ "ap_weighted": 0.711491
34
+ },
35
+ {
36
+ "accuracy": 0.674805,
37
+ "f1": 0.674561,
38
+ "f1_weighted": 0.674561,
39
+ "ap": 0.616375,
40
+ "ap_weighted": 0.616375
41
+ },
42
+ {
43
+ "accuracy": 0.753418,
44
+ "f1": 0.748479,
45
+ "f1_weighted": 0.748479,
46
+ "ap": 0.715938,
47
+ "ap_weighted": 0.715938
48
+ },
49
+ {
50
+ "accuracy": 0.757812,
51
+ "f1": 0.757257,
52
+ "f1_weighted": 0.757257,
53
+ "ap": 0.689568,
54
+ "ap_weighted": 0.689568
55
+ },
56
+ {
57
+ "accuracy": 0.786621,
58
+ "f1": 0.786615,
59
+ "f1_weighted": 0.786615,
60
+ "ap": 0.724589,
61
+ "ap_weighted": 0.724589
62
+ },
63
+ {
64
+ "accuracy": 0.745117,
65
+ "f1": 0.741262,
66
+ "f1_weighted": 0.741262,
67
+ "ap": 0.670851,
68
+ "ap_weighted": 0.670851
69
+ },
70
+ {
71
+ "accuracy": 0.750488,
72
+ "f1": 0.747494,
73
+ "f1_weighted": 0.747494,
74
+ "ap": 0.705457,
75
+ "ap_weighted": 0.705457
76
+ },
77
+ {
78
+ "accuracy": 0.77002,
79
+ "f1": 0.770007,
80
+ "f1_weighted": 0.770007,
81
+ "ap": 0.709004,
82
+ "ap_weighted": 0.709004
83
+ }
84
+ ],
85
+ "main_score": 0.747803,
86
+ "hf_subset": "default",
87
+ "languages": [
88
+ "rus-Cyrl"
89
+ ]
90
+ }
91
+ ]
92
+ },
93
+ "evaluation_time": 5.260155439376831,
94
+ "kg_co2_emissions": null
95
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/KinopoiskClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "5911f26666ac11af46cb9c6849d0dc80a378af24",
3
+ "task_name": "KinopoiskClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.677867,
9
+ "f1": 0.645153,
10
+ "f1_weighted": 0.645153,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.684667,
14
+ "f1": 0.651969,
15
+ "f1_weighted": 0.651969
16
+ },
17
+ {
18
+ "accuracy": 0.657333,
19
+ "f1": 0.607321,
20
+ "f1_weighted": 0.607321
21
+ },
22
+ {
23
+ "accuracy": 0.692667,
24
+ "f1": 0.669488,
25
+ "f1_weighted": 0.669488
26
+ },
27
+ {
28
+ "accuracy": 0.673333,
29
+ "f1": 0.627514,
30
+ "f1_weighted": 0.627514
31
+ },
32
+ {
33
+ "accuracy": 0.668667,
34
+ "f1": 0.640201,
35
+ "f1_weighted": 0.640201
36
+ },
37
+ {
38
+ "accuracy": 0.673333,
39
+ "f1": 0.645168,
40
+ "f1_weighted": 0.645168
41
+ },
42
+ {
43
+ "accuracy": 0.671333,
44
+ "f1": 0.634934,
45
+ "f1_weighted": 0.634934
46
+ },
47
+ {
48
+ "accuracy": 0.69,
49
+ "f1": 0.66326,
50
+ "f1_weighted": 0.66326
51
+ },
52
+ {
53
+ "accuracy": 0.686,
54
+ "f1": 0.652313,
55
+ "f1_weighted": 0.652313
56
+ },
57
+ {
58
+ "accuracy": 0.681333,
59
+ "f1": 0.659363,
60
+ "f1_weighted": 0.659363
61
+ }
62
+ ],
63
+ "main_score": 0.677867,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 26.355074167251587,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/MIRACLReranking.json ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "6d1962c527217f8927fca80f890f14f36b2802af",
3
+ "task_name": "MIRACLReranking",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "dev": [
7
+ {
8
+ "NDCG@1(MIRACL)": 0.61508,
9
+ "NDCG@3(MIRACL)": 0.60328,
10
+ "NDCG@5(MIRACL)": 0.61587,
11
+ "NDCG@10(MIRACL)": 0.64278,
12
+ "NDCG@20(MIRACL)": 0.66588,
13
+ "NDCG@100(MIRACL)": 0.67524,
14
+ "NDCG@1000(MIRACL)": 0.67524,
15
+ "MAP@1(MIRACL)": 0.37902,
16
+ "MAP@3(MIRACL)": 0.51553,
17
+ "MAP@5(MIRACL)": 0.55325,
18
+ "MAP@10(MIRACL)": 0.58191,
19
+ "MAP@20(MIRACL)": 0.59569,
20
+ "MAP@100(MIRACL)": 0.59941,
21
+ "MAP@1000(MIRACL)": 0.59941,
22
+ "Recall@1(MIRACL)": 0.37902,
23
+ "Recall@3(MIRACL)": 0.5629,
24
+ "Recall@5(MIRACL)": 0.63281,
25
+ "Recall@10(MIRACL)": 0.70597,
26
+ "Recall@20(MIRACL)": 0.76804,
27
+ "Recall@100(MIRACL)": 0.79952,
28
+ "Recall@1000(MIRACL)": 0.79952,
29
+ "P@1(MIRACL)": 0.61508,
30
+ "P@3(MIRACL)": 0.37049,
31
+ "P@5(MIRACL)": 0.26848,
32
+ "P@10(MIRACL)": 0.16055,
33
+ "P@20(MIRACL)": 0.09134,
34
+ "P@100(MIRACL)": 0.01953,
35
+ "P@1000(MIRACL)": 0.00195,
36
+ "nAUC_NDCG@1_max(MIRACL)": 0.600553,
37
+ "nAUC_NDCG@1_std(MIRACL)": 0.27805,
38
+ "nAUC_NDCG@1_diff1(MIRACL)": 0.405028,
39
+ "nAUC_NDCG@3_max(MIRACL)": 0.506047,
40
+ "nAUC_NDCG@3_std(MIRACL)": 0.233443,
41
+ "nAUC_NDCG@3_diff1(MIRACL)": 0.313015,
42
+ "nAUC_NDCG@5_max(MIRACL)": 0.511062,
43
+ "nAUC_NDCG@5_std(MIRACL)": 0.247068,
44
+ "nAUC_NDCG@5_diff1(MIRACL)": 0.298472,
45
+ "nAUC_NDCG@10_max(MIRACL)": 0.553836,
46
+ "nAUC_NDCG@10_std(MIRACL)": 0.284236,
47
+ "nAUC_NDCG@10_diff1(MIRACL)": 0.298458,
48
+ "nAUC_NDCG@20_max(MIRACL)": 0.577914,
49
+ "nAUC_NDCG@20_std(MIRACL)": 0.305184,
50
+ "nAUC_NDCG@20_diff1(MIRACL)": 0.284474,
51
+ "nAUC_NDCG@100_max(MIRACL)": 0.59327,
52
+ "nAUC_NDCG@100_std(MIRACL)": 0.312327,
53
+ "nAUC_NDCG@100_diff1(MIRACL)": 0.282776,
54
+ "nAUC_NDCG@1000_max(MIRACL)": 0.59327,
55
+ "nAUC_NDCG@1000_std(MIRACL)": 0.312327,
56
+ "nAUC_NDCG@1000_diff1(MIRACL)": 0.282776,
57
+ "nAUC_MAP@1_max(MIRACL)": 0.262352,
58
+ "nAUC_MAP@1_std(MIRACL)": 0.049827,
59
+ "nAUC_MAP@1_diff1(MIRACL)": 0.438564,
60
+ "nAUC_MAP@3_max(MIRACL)": 0.391016,
61
+ "nAUC_MAP@3_std(MIRACL)": 0.134972,
62
+ "nAUC_MAP@3_diff1(MIRACL)": 0.348295,
63
+ "nAUC_MAP@5_max(MIRACL)": 0.43532,
64
+ "nAUC_MAP@5_std(MIRACL)": 0.176288,
65
+ "nAUC_MAP@5_diff1(MIRACL)": 0.322674,
66
+ "nAUC_MAP@10_max(MIRACL)": 0.481519,
67
+ "nAUC_MAP@10_std(MIRACL)": 0.221083,
68
+ "nAUC_MAP@10_diff1(MIRACL)": 0.318529,
69
+ "nAUC_MAP@20_max(MIRACL)": 0.496245,
70
+ "nAUC_MAP@20_std(MIRACL)": 0.234774,
71
+ "nAUC_MAP@20_diff1(MIRACL)": 0.31199,
72
+ "nAUC_MAP@100_max(MIRACL)": 0.502056,
73
+ "nAUC_MAP@100_std(MIRACL)": 0.237664,
74
+ "nAUC_MAP@100_diff1(MIRACL)": 0.31144,
75
+ "nAUC_MAP@1000_max(MIRACL)": 0.502056,
76
+ "nAUC_MAP@1000_std(MIRACL)": 0.237664,
77
+ "nAUC_MAP@1000_diff1(MIRACL)": 0.31144,
78
+ "nAUC_Recall@1_max(MIRACL)": 0.262352,
79
+ "nAUC_Recall@1_std(MIRACL)": 0.049827,
80
+ "nAUC_Recall@1_diff1(MIRACL)": 0.438564,
81
+ "nAUC_Recall@3_max(MIRACL)": 0.38336,
82
+ "nAUC_Recall@3_std(MIRACL)": 0.149345,
83
+ "nAUC_Recall@3_diff1(MIRACL)": 0.292101,
84
+ "nAUC_Recall@5_max(MIRACL)": 0.449206,
85
+ "nAUC_Recall@5_std(MIRACL)": 0.220407,
86
+ "nAUC_Recall@5_diff1(MIRACL)": 0.233645,
87
+ "nAUC_Recall@10_max(MIRACL)": 0.5701,
88
+ "nAUC_Recall@10_std(MIRACL)": 0.328159,
89
+ "nAUC_Recall@10_diff1(MIRACL)": 0.221075,
90
+ "nAUC_Recall@20_max(MIRACL)": 0.657781,
91
+ "nAUC_Recall@20_std(MIRACL)": 0.400951,
92
+ "nAUC_Recall@20_diff1(MIRACL)": 0.168991,
93
+ "nAUC_Recall@100_max(MIRACL)": 0.744632,
94
+ "nAUC_Recall@100_std(MIRACL)": 0.444755,
95
+ "nAUC_Recall@100_diff1(MIRACL)": 0.153793,
96
+ "nAUC_Recall@1000_max(MIRACL)": 0.744632,
97
+ "nAUC_Recall@1000_std(MIRACL)": 0.444755,
98
+ "nAUC_Recall@1000_diff1(MIRACL)": 0.153793,
99
+ "nAUC_P@1_max(MIRACL)": 0.600553,
100
+ "nAUC_P@1_std(MIRACL)": 0.27805,
101
+ "nAUC_P@1_diff1(MIRACL)": 0.405028,
102
+ "nAUC_P@3_max(MIRACL)": 0.441107,
103
+ "nAUC_P@3_std(MIRACL)": 0.281702,
104
+ "nAUC_P@3_diff1(MIRACL)": -0.010734,
105
+ "nAUC_P@5_max(MIRACL)": 0.407748,
106
+ "nAUC_P@5_std(MIRACL)": 0.302857,
107
+ "nAUC_P@5_diff1(MIRACL)": -0.085616,
108
+ "nAUC_P@10_max(MIRACL)": 0.382584,
109
+ "nAUC_P@10_std(MIRACL)": 0.322036,
110
+ "nAUC_P@10_diff1(MIRACL)": -0.10017,
111
+ "nAUC_P@20_max(MIRACL)": 0.337207,
112
+ "nAUC_P@20_std(MIRACL)": 0.299126,
113
+ "nAUC_P@20_diff1(MIRACL)": -0.133249,
114
+ "nAUC_P@100_max(MIRACL)": 0.318626,
115
+ "nAUC_P@100_std(MIRACL)": 0.278223,
116
+ "nAUC_P@100_diff1(MIRACL)": -0.137392,
117
+ "nAUC_P@1000_max(MIRACL)": 0.318626,
118
+ "nAUC_P@1000_std(MIRACL)": 0.278223,
119
+ "nAUC_P@1000_diff1(MIRACL)": -0.137392,
120
+ "main_score": 0.64278,
121
+ "hf_subset": "ru",
122
+ "languages": [
123
+ "rus-Cyrl"
124
+ ]
125
+ }
126
+ ]
127
+ },
128
+ "evaluation_time": 661.112181186676,
129
+ "kg_co2_emissions": null
130
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/MIRACLRetrieval.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "task_name": "MIRACLRetrieval",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "dev": [
7
+ {
8
+ "ndcg_at_1": 0.66454,
9
+ "ndcg_at_3": 0.63146,
10
+ "ndcg_at_5": 0.64449,
11
+ "ndcg_at_10": 0.67598,
12
+ "ndcg_at_20": 0.70249,
13
+ "ndcg_at_100": 0.73078,
14
+ "ndcg_at_1000": 0.73983,
15
+ "map_at_1": 0.33797,
16
+ "map_at_3": 0.50025,
17
+ "map_at_5": 0.54478,
18
+ "map_at_10": 0.57852,
19
+ "map_at_20": 0.5934,
20
+ "map_at_100": 0.60222,
21
+ "map_at_1000": 0.60294,
22
+ "recall_at_1": 0.33797,
23
+ "recall_at_3": 0.57033,
24
+ "recall_at_5": 0.66001,
25
+ "recall_at_10": 0.75571,
26
+ "recall_at_20": 0.82446,
27
+ "recall_at_100": 0.92659,
28
+ "recall_at_1000": 0.97769,
29
+ "precision_at_1": 0.66454,
30
+ "precision_at_3": 0.42945,
31
+ "precision_at_5": 0.31853,
32
+ "precision_at_10": 0.19425,
33
+ "precision_at_20": 0.11122,
34
+ "precision_at_100": 0.02609,
35
+ "precision_at_1000": 0.00278,
36
+ "mrr_at_1": 0.664537,
37
+ "mrr_at_3": 0.748536,
38
+ "mrr_at_5": 0.75856,
39
+ "mrr_at_10": 0.764219,
40
+ "mrr_at_20": 0.765676,
41
+ "mrr_at_100": 0.766198,
42
+ "mrr_at_1000": 0.766267,
43
+ "nauc_ndcg_at_1_max": 0.454534,
44
+ "nauc_ndcg_at_1_std": 0.113134,
45
+ "nauc_ndcg_at_1_diff1": 0.472641,
46
+ "nauc_ndcg_at_3_max": 0.345453,
47
+ "nauc_ndcg_at_3_std": 0.03846,
48
+ "nauc_ndcg_at_3_diff1": 0.300801,
49
+ "nauc_ndcg_at_5_max": 0.333443,
50
+ "nauc_ndcg_at_5_std": 0.0237,
51
+ "nauc_ndcg_at_5_diff1": 0.295483,
52
+ "nauc_ndcg_at_10_max": 0.358479,
53
+ "nauc_ndcg_at_10_std": 0.048536,
54
+ "nauc_ndcg_at_10_diff1": 0.296483,
55
+ "nauc_ndcg_at_20_max": 0.375429,
56
+ "nauc_ndcg_at_20_std": 0.07551,
57
+ "nauc_ndcg_at_20_diff1": 0.30184,
58
+ "nauc_ndcg_at_100_max": 0.396757,
59
+ "nauc_ndcg_at_100_std": 0.109449,
60
+ "nauc_ndcg_at_100_diff1": 0.309037,
61
+ "nauc_ndcg_at_1000_max": 0.39949,
62
+ "nauc_ndcg_at_1000_std": 0.104072,
63
+ "nauc_ndcg_at_1000_diff1": 0.316566,
64
+ "nauc_map_at_1_max": 0.150228,
65
+ "nauc_map_at_1_std": -0.112757,
66
+ "nauc_map_at_1_diff1": 0.372127,
67
+ "nauc_map_at_3_max": 0.227972,
68
+ "nauc_map_at_3_std": -0.089535,
69
+ "nauc_map_at_3_diff1": 0.286184,
70
+ "nauc_map_at_5_max": 0.261792,
71
+ "nauc_map_at_5_std": -0.053616,
72
+ "nauc_map_at_5_diff1": 0.278609,
73
+ "nauc_map_at_10_max": 0.297156,
74
+ "nauc_map_at_10_std": -0.013764,
75
+ "nauc_map_at_10_diff1": 0.280363,
76
+ "nauc_map_at_20_max": 0.309328,
77
+ "nauc_map_at_20_std": 0.002653,
78
+ "nauc_map_at_20_diff1": 0.282283,
79
+ "nauc_map_at_100_max": 0.316597,
80
+ "nauc_map_at_100_std": 0.013545,
81
+ "nauc_map_at_100_diff1": 0.282954,
82
+ "nauc_map_at_1000_max": 0.316838,
83
+ "nauc_map_at_1000_std": 0.013567,
84
+ "nauc_map_at_1000_diff1": 0.28343,
85
+ "nauc_recall_at_1_max": 0.150228,
86
+ "nauc_recall_at_1_std": -0.112757,
87
+ "nauc_recall_at_1_diff1": 0.372127,
88
+ "nauc_recall_at_3_max": 0.182906,
89
+ "nauc_recall_at_3_std": -0.096111,
90
+ "nauc_recall_at_3_diff1": 0.222511,
91
+ "nauc_recall_at_5_max": 0.221552,
92
+ "nauc_recall_at_5_std": -0.050731,
93
+ "nauc_recall_at_5_diff1": 0.191517,
94
+ "nauc_recall_at_10_max": 0.285819,
95
+ "nauc_recall_at_10_std": 0.045897,
96
+ "nauc_recall_at_10_diff1": 0.173858,
97
+ "nauc_recall_at_20_max": 0.307503,
98
+ "nauc_recall_at_20_std": 0.110959,
99
+ "nauc_recall_at_20_diff1": 0.164484,
100
+ "nauc_recall_at_100_max": 0.416226,
101
+ "nauc_recall_at_100_std": 0.395137,
102
+ "nauc_recall_at_100_diff1": 0.131079,
103
+ "nauc_recall_at_1000_max": 0.671115,
104
+ "nauc_recall_at_1000_std": 0.65945,
105
+ "nauc_recall_at_1000_diff1": 0.168476,
106
+ "nauc_precision_at_1_max": 0.454534,
107
+ "nauc_precision_at_1_std": 0.113134,
108
+ "nauc_precision_at_1_diff1": 0.472641,
109
+ "nauc_precision_at_3_max": 0.342548,
110
+ "nauc_precision_at_3_std": 0.202324,
111
+ "nauc_precision_at_3_diff1": 0.03691,
112
+ "nauc_precision_at_5_max": 0.311811,
113
+ "nauc_precision_at_5_std": 0.25771,
114
+ "nauc_precision_at_5_diff1": -0.024771,
115
+ "nauc_precision_at_10_max": 0.288927,
116
+ "nauc_precision_at_10_std": 0.312047,
117
+ "nauc_precision_at_10_diff1": -0.048352,
118
+ "nauc_precision_at_20_max": 0.246845,
119
+ "nauc_precision_at_20_std": 0.323635,
120
+ "nauc_precision_at_20_diff1": -0.065645,
121
+ "nauc_precision_at_100_max": 0.20531,
122
+ "nauc_precision_at_100_std": 0.342819,
123
+ "nauc_precision_at_100_diff1": -0.09019,
124
+ "nauc_precision_at_1000_max": 0.173362,
125
+ "nauc_precision_at_1000_std": 0.31117,
126
+ "nauc_precision_at_1000_diff1": -0.086925,
127
+ "nauc_mrr_at_1_max": 0.454534,
128
+ "nauc_mrr_at_1_std": 0.113134,
129
+ "nauc_mrr_at_1_diff1": 0.472641,
130
+ "nauc_mrr_at_3_max": 0.496982,
131
+ "nauc_mrr_at_3_std": 0.154943,
132
+ "nauc_mrr_at_3_diff1": 0.448405,
133
+ "nauc_mrr_at_5_max": 0.501371,
134
+ "nauc_mrr_at_5_std": 0.153853,
135
+ "nauc_mrr_at_5_diff1": 0.455103,
136
+ "nauc_mrr_at_10_max": 0.502336,
137
+ "nauc_mrr_at_10_std": 0.157663,
138
+ "nauc_mrr_at_10_diff1": 0.457385,
139
+ "nauc_mrr_at_20_max": 0.500066,
140
+ "nauc_mrr_at_20_std": 0.15628,
141
+ "nauc_mrr_at_20_diff1": 0.45726,
142
+ "nauc_mrr_at_100_max": 0.499609,
143
+ "nauc_mrr_at_100_std": 0.155922,
144
+ "nauc_mrr_at_100_diff1": 0.457623,
145
+ "nauc_mrr_at_1000_max": 0.499501,
146
+ "nauc_mrr_at_1000_std": 0.155727,
147
+ "nauc_mrr_at_1000_diff1": 0.457528,
148
+ "main_score": 0.67598,
149
+ "hf_subset": "ru",
150
+ "languages": [
151
+ "rus-Cyrl"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 36008.40910077095,
157
+ "kg_co2_emissions": null
158
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/MassiveIntentClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "4672e20407010da34463acc759c162ca9734bca6",
3
+ "task_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.739203,
9
+ "f1": 0.692043,
10
+ "f1_weighted": 0.730365,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.75455,
14
+ "f1": 0.708016,
15
+ "f1_weighted": 0.747282
16
+ },
17
+ {
18
+ "accuracy": 0.761928,
19
+ "f1": 0.702271,
20
+ "f1_weighted": 0.75945
21
+ },
22
+ {
23
+ "accuracy": 0.738318,
24
+ "f1": 0.68185,
25
+ "f1_weighted": 0.729956
26
+ },
27
+ {
28
+ "accuracy": 0.750615,
29
+ "f1": 0.700633,
30
+ "f1_weighted": 0.738755
31
+ },
32
+ {
33
+ "accuracy": 0.74668,
34
+ "f1": 0.693007,
35
+ "f1_weighted": 0.735926
36
+ },
37
+ {
38
+ "accuracy": 0.713232,
39
+ "f1": 0.674349,
40
+ "f1_weighted": 0.703677
41
+ },
42
+ {
43
+ "accuracy": 0.725037,
44
+ "f1": 0.675986,
45
+ "f1_weighted": 0.710093
46
+ },
47
+ {
48
+ "accuracy": 0.715199,
49
+ "f1": 0.679388,
50
+ "f1_weighted": 0.704615
51
+ },
52
+ {
53
+ "accuracy": 0.732907,
54
+ "f1": 0.694565,
55
+ "f1_weighted": 0.723237
56
+ },
57
+ {
58
+ "accuracy": 0.753566,
59
+ "f1": 0.71036,
60
+ "f1_weighted": 0.75066
61
+ }
62
+ ],
63
+ "main_score": 0.739203,
64
+ "hf_subset": "ru",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.740081,
73
+ "f1": 0.717027,
74
+ "f1_weighted": 0.732054,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.756221,
78
+ "f1": 0.742433,
79
+ "f1_weighted": 0.747767
80
+ },
81
+ {
82
+ "accuracy": 0.757229,
83
+ "f1": 0.729741,
84
+ "f1_weighted": 0.753225
85
+ },
86
+ {
87
+ "accuracy": 0.718897,
88
+ "f1": 0.69209,
89
+ "f1_weighted": 0.71092
90
+ },
91
+ {
92
+ "accuracy": 0.758238,
93
+ "f1": 0.727514,
94
+ "f1_weighted": 0.750771
95
+ },
96
+ {
97
+ "accuracy": 0.736046,
98
+ "f1": 0.712475,
99
+ "f1_weighted": 0.723101
100
+ },
101
+ {
102
+ "accuracy": 0.713517,
103
+ "f1": 0.698443,
104
+ "f1_weighted": 0.705223
105
+ },
106
+ {
107
+ "accuracy": 0.737727,
108
+ "f1": 0.710294,
109
+ "f1_weighted": 0.729653
110
+ },
111
+ {
112
+ "accuracy": 0.726631,
113
+ "f1": 0.70469,
114
+ "f1_weighted": 0.715884
115
+ },
116
+ {
117
+ "accuracy": 0.739408,
118
+ "f1": 0.718179,
119
+ "f1_weighted": 0.731342
120
+ },
121
+ {
122
+ "accuracy": 0.756893,
123
+ "f1": 0.734409,
124
+ "f1_weighted": 0.752653
125
+ }
126
+ ],
127
+ "main_score": 0.740081,
128
+ "hf_subset": "ru",
129
+ "languages": [
130
+ "rus-Cyrl"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 39.10078191757202,
136
+ "kg_co2_emissions": null
137
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "fad2c6e8459f9e1c45d9315f4953d921437d70f8",
3
+ "task_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.84604,
9
+ "f1": 0.836355,
10
+ "f1_weighted": 0.843628,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.845548,
14
+ "f1": 0.838163,
15
+ "f1_weighted": 0.844973
16
+ },
17
+ {
18
+ "accuracy": 0.85391,
19
+ "f1": 0.843265,
20
+ "f1_weighted": 0.851264
21
+ },
22
+ {
23
+ "accuracy": 0.857846,
24
+ "f1": 0.848271,
25
+ "f1_weighted": 0.853002
26
+ },
27
+ {
28
+ "accuracy": 0.841121,
29
+ "f1": 0.833726,
30
+ "f1_weighted": 0.837569
31
+ },
32
+ {
33
+ "accuracy": 0.846532,
34
+ "f1": 0.837552,
35
+ "f1_weighted": 0.84369
36
+ },
37
+ {
38
+ "accuracy": 0.839154,
39
+ "f1": 0.825871,
40
+ "f1_weighted": 0.83579
41
+ },
42
+ {
43
+ "accuracy": 0.826365,
44
+ "f1": 0.812775,
45
+ "f1_weighted": 0.823798
46
+ },
47
+ {
48
+ "accuracy": 0.831284,
49
+ "f1": 0.822564,
50
+ "f1_weighted": 0.832204
51
+ },
52
+ {
53
+ "accuracy": 0.863256,
54
+ "f1": 0.8513,
55
+ "f1_weighted": 0.859449
56
+ },
57
+ {
58
+ "accuracy": 0.855386,
59
+ "f1": 0.850065,
60
+ "f1_weighted": 0.854544
61
+ }
62
+ ],
63
+ "main_score": 0.84604,
64
+ "hf_subset": "ru",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.845024,
73
+ "f1": 0.836857,
74
+ "f1_weighted": 0.842567,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.840955,
78
+ "f1": 0.838206,
79
+ "f1_weighted": 0.839704
80
+ },
81
+ {
82
+ "accuracy": 0.851379,
83
+ "f1": 0.839277,
84
+ "f1_weighted": 0.847187
85
+ },
86
+ {
87
+ "accuracy": 0.854405,
88
+ "f1": 0.840472,
89
+ "f1_weighted": 0.85029
90
+ },
91
+ {
92
+ "accuracy": 0.840619,
93
+ "f1": 0.832708,
94
+ "f1_weighted": 0.837378
95
+ },
96
+ {
97
+ "accuracy": 0.8423,
98
+ "f1": 0.834852,
99
+ "f1_weighted": 0.840059
100
+ },
101
+ {
102
+ "accuracy": 0.84499,
103
+ "f1": 0.83545,
104
+ "f1_weighted": 0.841
105
+ },
106
+ {
107
+ "accuracy": 0.825824,
108
+ "f1": 0.81715,
109
+ "f1_weighted": 0.82311
110
+ },
111
+ {
112
+ "accuracy": 0.834902,
113
+ "f1": 0.83237,
114
+ "f1_weighted": 0.836539
115
+ },
116
+ {
117
+ "accuracy": 0.85575,
118
+ "f1": 0.846792,
119
+ "f1_weighted": 0.851793
120
+ },
121
+ {
122
+ "accuracy": 0.859112,
123
+ "f1": 0.851289,
124
+ "f1_weighted": 0.858612
125
+ }
126
+ ],
127
+ "main_score": 0.845024,
128
+ "hf_subset": "ru",
129
+ "languages": [
130
+ "rus-Cyrl"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 9.918192625045776,
136
+ "kg_co2_emissions": null
137
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RUParaPhraserSTS.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "43265056790b8f7c59e0139acb4be0a8dad2c8f4",
3
+ "task_name": "RUParaPhraserSTS",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "pearson": 0.712547,
9
+ "spearman": 0.778353,
10
+ "cosine_pearson": 0.712547,
11
+ "cosine_spearman": 0.778352,
12
+ "manhattan_pearson": 0.756428,
13
+ "manhattan_spearman": 0.777939,
14
+ "euclidean_pearson": 0.756771,
15
+ "euclidean_spearman": 0.778352,
16
+ "main_score": 0.778352,
17
+ "hf_subset": "default",
18
+ "languages": [
19
+ "rus-Cyrl"
20
+ ]
21
+ }
22
+ ]
23
+ },
24
+ "evaluation_time": 4.62164568901062,
25
+ "kg_co2_emissions": null
26
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RiaNewsRetrieval.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "82374b0bbacda6114f39ff9c5b925fa1512ca5d7",
3
+ "task_name": "RiaNewsRetrieval",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.7102,
9
+ "ndcg_at_3": 0.79192,
10
+ "ndcg_at_5": 0.80584,
11
+ "ndcg_at_10": 0.81606,
12
+ "ndcg_at_20": 0.8211,
13
+ "ndcg_at_100": 0.82681,
14
+ "ndcg_at_1000": 0.8291,
15
+ "map_at_1": 0.7102,
16
+ "map_at_3": 0.7727,
17
+ "map_at_5": 0.78044,
18
+ "map_at_10": 0.78471,
19
+ "map_at_20": 0.78611,
20
+ "map_at_100": 0.78694,
21
+ "map_at_1000": 0.78702,
22
+ "recall_at_1": 0.7102,
23
+ "recall_at_3": 0.8472,
24
+ "recall_at_5": 0.8809,
25
+ "recall_at_10": 0.9122,
26
+ "recall_at_20": 0.9321,
27
+ "recall_at_100": 0.9623,
28
+ "recall_at_1000": 0.9806,
29
+ "precision_at_1": 0.7102,
30
+ "precision_at_3": 0.2824,
31
+ "precision_at_5": 0.17618,
32
+ "precision_at_10": 0.09122,
33
+ "precision_at_20": 0.04661,
34
+ "precision_at_100": 0.00962,
35
+ "precision_at_1000": 0.00098,
36
+ "mrr_at_1": 0.7104,
37
+ "mrr_at_3": 0.772783,
38
+ "mrr_at_5": 0.780558,
39
+ "mrr_at_10": 0.784827,
40
+ "mrr_at_20": 0.786219,
41
+ "mrr_at_100": 0.787049,
42
+ "mrr_at_1000": 0.787129,
43
+ "nauc_ndcg_at_1_max": 0.316768,
44
+ "nauc_ndcg_at_1_std": -0.207197,
45
+ "nauc_ndcg_at_1_diff1": 0.800044,
46
+ "nauc_ndcg_at_3_max": 0.378069,
47
+ "nauc_ndcg_at_3_std": -0.189967,
48
+ "nauc_ndcg_at_3_diff1": 0.762221,
49
+ "nauc_ndcg_at_5_max": 0.38219,
50
+ "nauc_ndcg_at_5_std": -0.178563,
51
+ "nauc_ndcg_at_5_diff1": 0.761678,
52
+ "nauc_ndcg_at_10_max": 0.383186,
53
+ "nauc_ndcg_at_10_std": -0.17001,
54
+ "nauc_ndcg_at_10_diff1": 0.76132,
55
+ "nauc_ndcg_at_20_max": 0.383959,
56
+ "nauc_ndcg_at_20_std": -0.163678,
57
+ "nauc_ndcg_at_20_diff1": 0.762984,
58
+ "nauc_ndcg_at_100_max": 0.382142,
59
+ "nauc_ndcg_at_100_std": -0.162127,
60
+ "nauc_ndcg_at_100_diff1": 0.764826,
61
+ "nauc_ndcg_at_1000_max": 0.379043,
62
+ "nauc_ndcg_at_1000_std": -0.165657,
63
+ "nauc_ndcg_at_1000_diff1": 0.765714,
64
+ "nauc_map_at_1_max": 0.316768,
65
+ "nauc_map_at_1_std": -0.207197,
66
+ "nauc_map_at_1_diff1": 0.800044,
67
+ "nauc_map_at_3_max": 0.360843,
68
+ "nauc_map_at_3_std": -0.195662,
69
+ "nauc_map_at_3_diff1": 0.772331,
70
+ "nauc_map_at_5_max": 0.362393,
71
+ "nauc_map_at_5_std": -0.19018,
72
+ "nauc_map_at_5_diff1": 0.772417,
73
+ "nauc_map_at_10_max": 0.362396,
74
+ "nauc_map_at_10_std": -0.187381,
75
+ "nauc_map_at_10_diff1": 0.772433,
76
+ "nauc_map_at_20_max": 0.362355,
77
+ "nauc_map_at_20_std": -0.186116,
78
+ "nauc_map_at_20_diff1": 0.772865,
79
+ "nauc_map_at_100_max": 0.362056,
80
+ "nauc_map_at_100_std": -0.186047,
81
+ "nauc_map_at_100_diff1": 0.77316,
82
+ "nauc_map_at_1000_max": 0.361975,
83
+ "nauc_map_at_1000_std": -0.186118,
84
+ "nauc_map_at_1000_diff1": 0.773189,
85
+ "nauc_recall_at_1_max": 0.316768,
86
+ "nauc_recall_at_1_std": -0.207197,
87
+ "nauc_recall_at_1_diff1": 0.800044,
88
+ "nauc_recall_at_3_max": 0.44701,
89
+ "nauc_recall_at_3_std": -0.166654,
90
+ "nauc_recall_at_3_diff1": 0.722075,
91
+ "nauc_recall_at_5_max": 0.478904,
92
+ "nauc_recall_at_5_std": -0.118778,
93
+ "nauc_recall_at_5_diff1": 0.70991,
94
+ "nauc_recall_at_10_max": 0.516027,
95
+ "nauc_recall_at_10_std": -0.048377,
96
+ "nauc_recall_at_10_diff1": 0.691644,
97
+ "nauc_recall_at_20_max": 0.56368,
98
+ "nauc_recall_at_20_std": 0.048903,
99
+ "nauc_recall_at_20_diff1": 0.687687,
100
+ "nauc_recall_at_100_max": 0.666999,
101
+ "nauc_recall_at_100_std": 0.254258,
102
+ "nauc_recall_at_100_diff1": 0.665737,
103
+ "nauc_recall_at_1000_max": 0.731923,
104
+ "nauc_recall_at_1000_std": 0.403434,
105
+ "nauc_recall_at_1000_diff1": 0.630478,
106
+ "nauc_precision_at_1_max": 0.316768,
107
+ "nauc_precision_at_1_std": -0.207197,
108
+ "nauc_precision_at_1_diff1": 0.800044,
109
+ "nauc_precision_at_3_max": 0.44701,
110
+ "nauc_precision_at_3_std": -0.166654,
111
+ "nauc_precision_at_3_diff1": 0.722075,
112
+ "nauc_precision_at_5_max": 0.478904,
113
+ "nauc_precision_at_5_std": -0.118778,
114
+ "nauc_precision_at_5_diff1": 0.70991,
115
+ "nauc_precision_at_10_max": 0.516027,
116
+ "nauc_precision_at_10_std": -0.048377,
117
+ "nauc_precision_at_10_diff1": 0.691644,
118
+ "nauc_precision_at_20_max": 0.56368,
119
+ "nauc_precision_at_20_std": 0.048903,
120
+ "nauc_precision_at_20_diff1": 0.687687,
121
+ "nauc_precision_at_100_max": 0.666999,
122
+ "nauc_precision_at_100_std": 0.254258,
123
+ "nauc_precision_at_100_diff1": 0.665737,
124
+ "nauc_precision_at_1000_max": 0.731923,
125
+ "nauc_precision_at_1000_std": 0.403434,
126
+ "nauc_precision_at_1000_diff1": 0.630478,
127
+ "nauc_mrr_at_1_max": 0.317855,
128
+ "nauc_mrr_at_1_std": -0.20907,
129
+ "nauc_mrr_at_1_diff1": 0.799709,
130
+ "nauc_mrr_at_3_max": 0.361503,
131
+ "nauc_mrr_at_3_std": -0.196845,
132
+ "nauc_mrr_at_3_diff1": 0.77208,
133
+ "nauc_mrr_at_5_max": 0.363175,
134
+ "nauc_mrr_at_5_std": -0.191237,
135
+ "nauc_mrr_at_5_diff1": 0.772149,
136
+ "nauc_mrr_at_10_max": 0.363193,
137
+ "nauc_mrr_at_10_std": -0.188458,
138
+ "nauc_mrr_at_10_diff1": 0.77216,
139
+ "nauc_mrr_at_20_max": 0.363158,
140
+ "nauc_mrr_at_20_std": -0.187201,
141
+ "nauc_mrr_at_20_diff1": 0.77259,
142
+ "nauc_mrr_at_100_max": 0.362863,
143
+ "nauc_mrr_at_100_std": -0.187136,
144
+ "nauc_mrr_at_100_diff1": 0.772884,
145
+ "nauc_mrr_at_1000_max": 0.362782,
146
+ "nauc_mrr_at_1000_std": -0.187208,
147
+ "nauc_mrr_at_1000_diff1": 0.772913,
148
+ "main_score": 0.81606,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "rus-Cyrl"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 8441.960134983063,
157
+ "kg_co2_emissions": null
158
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuBQReranking.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "2e96b8f098fa4b0950fc58eacadeb31c0d0c7fa2",
3
+ "task_name": "RuBQReranking",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "map": 0.751605,
9
+ "mrr": 0.807142,
10
+ "nAUC_map_max": 0.270127,
11
+ "nAUC_map_std": 0.137775,
12
+ "nAUC_map_diff1": 0.4362,
13
+ "nAUC_mrr_max": 0.359072,
14
+ "nAUC_mrr_std": 0.217059,
15
+ "nAUC_mrr_diff1": 0.514186,
16
+ "main_score": 0.751605,
17
+ "hf_subset": "default",
18
+ "languages": [
19
+ "rus-Cyrl"
20
+ ]
21
+ }
22
+ ]
23
+ },
24
+ "evaluation_time": 178.88464188575745,
25
+ "kg_co2_emissions": null
26
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuBQRetrieval.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e19b6ffa60b3bc248e0b41f4cc37c26a55c2a67b",
3
+ "task_name": "RuBQRetrieval",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.64007,
9
+ "ndcg_at_3": 0.64217,
10
+ "ndcg_at_5": 0.67861,
11
+ "ndcg_at_10": 0.70999,
12
+ "ndcg_at_20": 0.72578,
13
+ "ndcg_at_100": 0.74083,
14
+ "ndcg_at_1000": 0.74691,
15
+ "map_at_1": 0.45181,
16
+ "map_at_3": 0.58589,
17
+ "map_at_5": 0.6183,
18
+ "map_at_10": 0.63668,
19
+ "map_at_20": 0.64329,
20
+ "map_at_100": 0.64634,
21
+ "map_at_1000": 0.64668,
22
+ "recall_at_1": 0.45181,
23
+ "recall_at_3": 0.66017,
24
+ "recall_at_5": 0.74531,
25
+ "recall_at_10": 0.82758,
26
+ "recall_at_20": 0.87822,
27
+ "recall_at_100": 0.94707,
28
+ "recall_at_1000": 0.98719,
29
+ "precision_at_1": 0.64007,
30
+ "precision_at_3": 0.34043,
31
+ "precision_at_5": 0.23759,
32
+ "precision_at_10": 0.1344,
33
+ "precision_at_20": 0.07246,
34
+ "precision_at_100": 0.01576,
35
+ "precision_at_1000": 0.00166,
36
+ "mrr_at_1": 0.641844,
37
+ "mrr_at_3": 0.714835,
38
+ "mrr_at_5": 0.72831,
39
+ "mrr_at_10": 0.734838,
40
+ "mrr_at_20": 0.736774,
41
+ "mrr_at_100": 0.737795,
42
+ "mrr_at_1000": 0.737892,
43
+ "nauc_ndcg_at_1_max": 0.36107,
44
+ "nauc_ndcg_at_1_std": -0.107256,
45
+ "nauc_ndcg_at_1_diff1": 0.569909,
46
+ "nauc_ndcg_at_3_max": 0.343441,
47
+ "nauc_ndcg_at_3_std": -0.073555,
48
+ "nauc_ndcg_at_3_diff1": 0.43699,
49
+ "nauc_ndcg_at_5_max": 0.352176,
50
+ "nauc_ndcg_at_5_std": -0.063766,
51
+ "nauc_ndcg_at_5_diff1": 0.421732,
52
+ "nauc_ndcg_at_10_max": 0.366196,
53
+ "nauc_ndcg_at_10_std": -0.051622,
54
+ "nauc_ndcg_at_10_diff1": 0.426417,
55
+ "nauc_ndcg_at_20_max": 0.369509,
56
+ "nauc_ndcg_at_20_std": -0.047491,
57
+ "nauc_ndcg_at_20_diff1": 0.425272,
58
+ "nauc_ndcg_at_100_max": 0.36871,
59
+ "nauc_ndcg_at_100_std": -0.049856,
60
+ "nauc_ndcg_at_100_diff1": 0.437676,
61
+ "nauc_ndcg_at_1000_max": 0.368105,
62
+ "nauc_ndcg_at_1000_std": -0.055502,
63
+ "nauc_ndcg_at_1000_diff1": 0.443428,
64
+ "nauc_map_at_1_max": 0.214575,
65
+ "nauc_map_at_1_std": -0.102765,
66
+ "nauc_map_at_1_diff1": 0.465951,
67
+ "nauc_map_at_3_max": 0.307994,
68
+ "nauc_map_at_3_std": -0.078975,
69
+ "nauc_map_at_3_diff1": 0.426084,
70
+ "nauc_map_at_5_max": 0.31927,
71
+ "nauc_map_at_5_std": -0.075899,
72
+ "nauc_map_at_5_diff1": 0.414743,
73
+ "nauc_map_at_10_max": 0.329976,
74
+ "nauc_map_at_10_std": -0.069228,
75
+ "nauc_map_at_10_diff1": 0.419286,
76
+ "nauc_map_at_20_max": 0.331143,
77
+ "nauc_map_at_20_std": -0.068498,
78
+ "nauc_map_at_20_diff1": 0.418543,
79
+ "nauc_map_at_100_max": 0.331712,
80
+ "nauc_map_at_100_std": -0.068278,
81
+ "nauc_map_at_100_diff1": 0.420559,
82
+ "nauc_map_at_1000_max": 0.331692,
83
+ "nauc_map_at_1000_std": -0.068504,
84
+ "nauc_map_at_1000_diff1": 0.420889,
85
+ "nauc_recall_at_1_max": 0.214575,
86
+ "nauc_recall_at_1_std": -0.102765,
87
+ "nauc_recall_at_1_diff1": 0.465951,
88
+ "nauc_recall_at_3_max": 0.314425,
89
+ "nauc_recall_at_3_std": -0.050767,
90
+ "nauc_recall_at_3_diff1": 0.354904,
91
+ "nauc_recall_at_5_max": 0.327509,
92
+ "nauc_recall_at_5_std": -0.026655,
93
+ "nauc_recall_at_5_diff1": 0.295881,
94
+ "nauc_recall_at_10_max": 0.354432,
95
+ "nauc_recall_at_10_std": 0.028836,
96
+ "nauc_recall_at_10_diff1": 0.260809,
97
+ "nauc_recall_at_20_max": 0.378636,
98
+ "nauc_recall_at_20_std": 0.085972,
99
+ "nauc_recall_at_20_diff1": 0.210342,
100
+ "nauc_recall_at_100_max": 0.36613,
101
+ "nauc_recall_at_100_std": 0.188171,
102
+ "nauc_recall_at_100_diff1": 0.196836,
103
+ "nauc_recall_at_1000_max": 0.464743,
104
+ "nauc_recall_at_1000_std": 0.354506,
105
+ "nauc_recall_at_1000_diff1": 0.166186,
106
+ "nauc_precision_at_1_max": 0.36107,
107
+ "nauc_precision_at_1_std": -0.107256,
108
+ "nauc_precision_at_1_diff1": 0.569909,
109
+ "nauc_precision_at_3_max": 0.303941,
110
+ "nauc_precision_at_3_std": 0.007998,
111
+ "nauc_precision_at_3_diff1": 0.151599,
112
+ "nauc_precision_at_5_max": 0.247721,
113
+ "nauc_precision_at_5_std": 0.035399,
114
+ "nauc_precision_at_5_diff1": 0.042852,
115
+ "nauc_precision_at_10_max": 0.203185,
116
+ "nauc_precision_at_10_std": 0.059959,
117
+ "nauc_precision_at_10_diff1": -0.013197,
118
+ "nauc_precision_at_20_max": 0.160332,
119
+ "nauc_precision_at_20_std": 0.069668,
120
+ "nauc_precision_at_20_diff1": -0.062233,
121
+ "nauc_precision_at_100_max": 0.109211,
122
+ "nauc_precision_at_100_std": 0.073609,
123
+ "nauc_precision_at_100_diff1": -0.089026,
124
+ "nauc_precision_at_1000_max": 0.069657,
125
+ "nauc_precision_at_1000_std": 0.049009,
126
+ "nauc_precision_at_1000_diff1": -0.100702,
127
+ "nauc_mrr_at_1_max": 0.366514,
128
+ "nauc_mrr_at_1_std": -0.108169,
129
+ "nauc_mrr_at_1_diff1": 0.565746,
130
+ "nauc_mrr_at_3_max": 0.421045,
131
+ "nauc_mrr_at_3_std": -0.078613,
132
+ "nauc_mrr_at_3_diff1": 0.541953,
133
+ "nauc_mrr_at_5_max": 0.423054,
134
+ "nauc_mrr_at_5_std": -0.073103,
135
+ "nauc_mrr_at_5_diff1": 0.540457,
136
+ "nauc_mrr_at_10_max": 0.420858,
137
+ "nauc_mrr_at_10_std": -0.072461,
138
+ "nauc_mrr_at_10_diff1": 0.539844,
139
+ "nauc_mrr_at_20_max": 0.420268,
140
+ "nauc_mrr_at_20_std": -0.073062,
141
+ "nauc_mrr_at_20_diff1": 0.539993,
142
+ "nauc_mrr_at_100_max": 0.419193,
143
+ "nauc_mrr_at_100_std": -0.074401,
144
+ "nauc_mrr_at_100_diff1": 0.540785,
145
+ "nauc_mrr_at_1000_max": 0.419222,
146
+ "nauc_mrr_at_1000_std": -0.074446,
147
+ "nauc_mrr_at_1000_diff1": 0.540853,
148
+ "main_score": 0.70999,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "rus-Cyrl"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 275.88254618644714,
157
+ "kg_co2_emissions": null
158
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuReviewsClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f6d2c31f4dc6b88f468552750bfec05b4b41b05a",
3
+ "task_name": "RuReviewsClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.722949,
9
+ "f1": 0.716192,
10
+ "f1_weighted": 0.716187,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.725098,
14
+ "f1": 0.722265,
15
+ "f1_weighted": 0.722261
16
+ },
17
+ {
18
+ "accuracy": 0.695312,
19
+ "f1": 0.682509,
20
+ "f1_weighted": 0.682496
21
+ },
22
+ {
23
+ "accuracy": 0.709473,
24
+ "f1": 0.696781,
25
+ "f1_weighted": 0.696765
26
+ },
27
+ {
28
+ "accuracy": 0.735352,
29
+ "f1": 0.733723,
30
+ "f1_weighted": 0.733724
31
+ },
32
+ {
33
+ "accuracy": 0.743164,
34
+ "f1": 0.744398,
35
+ "f1_weighted": 0.744405
36
+ },
37
+ {
38
+ "accuracy": 0.713867,
39
+ "f1": 0.706354,
40
+ "f1_weighted": 0.706351
41
+ },
42
+ {
43
+ "accuracy": 0.734375,
44
+ "f1": 0.728091,
45
+ "f1_weighted": 0.728087
46
+ },
47
+ {
48
+ "accuracy": 0.737793,
49
+ "f1": 0.734114,
50
+ "f1_weighted": 0.734118
51
+ },
52
+ {
53
+ "accuracy": 0.710938,
54
+ "f1": 0.694692,
55
+ "f1_weighted": 0.694665
56
+ },
57
+ {
58
+ "accuracy": 0.724121,
59
+ "f1": 0.718993,
60
+ "f1_weighted": 0.718998
61
+ }
62
+ ],
63
+ "main_score": 0.722949,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 6.427979946136475,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuSTSBenchmarkSTS.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7cf24f325c6da6195df55bef3d86b5e0616f3018",
3
+ "task_name": "RuSTSBenchmarkSTS",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "pearson": 0.826477,
9
+ "spearman": 0.822481,
10
+ "cosine_pearson": 0.826477,
11
+ "cosine_spearman": 0.822484,
12
+ "manhattan_pearson": 0.816692,
13
+ "manhattan_spearman": 0.823602,
14
+ "euclidean_pearson": 0.816037,
15
+ "euclidean_spearman": 0.822482,
16
+ "main_score": 0.822484,
17
+ "hf_subset": "default",
18
+ "languages": [
19
+ "rus-Cyrl"
20
+ ]
21
+ }
22
+ ]
23
+ },
24
+ "evaluation_time": 2.9777777194976807,
25
+ "kg_co2_emissions": null
26
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuSciBenchGRNTIClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "673a610d6d3dd91a547a0d57ae1b56f37ebbf6a1",
3
+ "task_name": "RuSciBenchGRNTIClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.6896,
9
+ "f1": 0.679907,
10
+ "f1_weighted": 0.680016,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.69873,
14
+ "f1": 0.689083,
15
+ "f1_weighted": 0.689191
16
+ },
17
+ {
18
+ "accuracy": 0.683594,
19
+ "f1": 0.669697,
20
+ "f1_weighted": 0.669817
21
+ },
22
+ {
23
+ "accuracy": 0.683594,
24
+ "f1": 0.674345,
25
+ "f1_weighted": 0.674468
26
+ },
27
+ {
28
+ "accuracy": 0.697754,
29
+ "f1": 0.690575,
30
+ "f1_weighted": 0.69068
31
+ },
32
+ {
33
+ "accuracy": 0.699219,
34
+ "f1": 0.691508,
35
+ "f1_weighted": 0.691632
36
+ },
37
+ {
38
+ "accuracy": 0.679688,
39
+ "f1": 0.669836,
40
+ "f1_weighted": 0.66993
41
+ },
42
+ {
43
+ "accuracy": 0.689453,
44
+ "f1": 0.678762,
45
+ "f1_weighted": 0.678858
46
+ },
47
+ {
48
+ "accuracy": 0.686035,
49
+ "f1": 0.675861,
50
+ "f1_weighted": 0.675968
51
+ },
52
+ {
53
+ "accuracy": 0.694824,
54
+ "f1": 0.686413,
55
+ "f1_weighted": 0.686523
56
+ },
57
+ {
58
+ "accuracy": 0.683105,
59
+ "f1": 0.672992,
60
+ "f1_weighted": 0.673088
61
+ }
62
+ ],
63
+ "main_score": 0.6896,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 53.08151817321777,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuSciBenchGRNTIClusteringP2P.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "673a610d6d3dd91a547a0d57ae1b56f37ebbf6a1",
3
+ "task_name": "RuSciBenchGRNTIClusteringP2P",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.647954,
11
+ 0.657885,
12
+ 0.652594,
13
+ 0.65128,
14
+ 0.662589,
15
+ 0.656943,
16
+ 0.644904,
17
+ 0.628403,
18
+ 0.65101,
19
+ 0.649499
20
+ ]
21
+ },
22
+ "v_measure": 0.650306,
23
+ "v_measure_std": 0.008799,
24
+ "main_score": 0.650306,
25
+ "hf_subset": "default",
26
+ "languages": [
27
+ "rus-Cyrl"
28
+ ]
29
+ }
30
+ ]
31
+ },
32
+ "evaluation_time": 23.036219596862793,
33
+ "kg_co2_emissions": null
34
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuSciBenchOECDClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "26c88e99dcaba32bb45d0e1bfc21902337f6d471",
3
+ "task_name": "RuSciBenchOECDClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.547705,
9
+ "f1": 0.522447,
10
+ "f1_weighted": 0.522413,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.55127,
14
+ "f1": 0.524899,
15
+ "f1_weighted": 0.524888
16
+ },
17
+ {
18
+ "accuracy": 0.538086,
19
+ "f1": 0.515329,
20
+ "f1_weighted": 0.515255
21
+ },
22
+ {
23
+ "accuracy": 0.555176,
24
+ "f1": 0.530577,
25
+ "f1_weighted": 0.530631
26
+ },
27
+ {
28
+ "accuracy": 0.541016,
29
+ "f1": 0.506643,
30
+ "f1_weighted": 0.50661
31
+ },
32
+ {
33
+ "accuracy": 0.54541,
34
+ "f1": 0.518962,
35
+ "f1_weighted": 0.518892
36
+ },
37
+ {
38
+ "accuracy": 0.561523,
39
+ "f1": 0.542091,
40
+ "f1_weighted": 0.542093
41
+ },
42
+ {
43
+ "accuracy": 0.53125,
44
+ "f1": 0.500447,
45
+ "f1_weighted": 0.500464
46
+ },
47
+ {
48
+ "accuracy": 0.578613,
49
+ "f1": 0.566891,
50
+ "f1_weighted": 0.566915
51
+ },
52
+ {
53
+ "accuracy": 0.535645,
54
+ "f1": 0.503717,
55
+ "f1_weighted": 0.50358
56
+ },
57
+ {
58
+ "accuracy": 0.539062,
59
+ "f1": 0.51492,
60
+ "f1_weighted": 0.514802
61
+ }
62
+ ],
63
+ "main_score": 0.547705,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 52.34759593009949,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/RuSciBenchOECDClusteringP2P.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "26c88e99dcaba32bb45d0e1bfc21902337f6d471",
3
+ "task_name": "RuSciBenchOECDClusteringP2P",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.548058,
11
+ 0.559924,
12
+ 0.551244,
13
+ 0.555431,
14
+ 0.560125,
15
+ 0.5594,
16
+ 0.560836,
17
+ 0.551309,
18
+ 0.554706,
19
+ 0.555158
20
+ ]
21
+ },
22
+ "v_measure": 0.555619,
23
+ "v_measure_std": 0.004199,
24
+ "main_score": 0.555619,
25
+ "hf_subset": "default",
26
+ "languages": [
27
+ "rus-Cyrl"
28
+ ]
29
+ }
30
+ ]
31
+ },
32
+ "evaluation_time": 22.064496755599976,
33
+ "kg_co2_emissions": null
34
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/STS22.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3",
3
+ "task_name": "STS22",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "pearson": 0.56179,
9
+ "spearman": 0.61084,
10
+ "cosine_pearson": 0.56179,
11
+ "cosine_spearman": 0.61084,
12
+ "manhattan_pearson": 0.571254,
13
+ "manhattan_spearman": 0.611716,
14
+ "euclidean_pearson": 0.567931,
15
+ "euclidean_spearman": 0.610784,
16
+ "main_score": 0.61084,
17
+ "hf_subset": "ru",
18
+ "languages": [
19
+ "rus-Cyrl"
20
+ ]
21
+ }
22
+ ]
23
+ },
24
+ "evaluation_time": 6.6172194480896,
25
+ "kg_co2_emissions": null
26
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/SensitiveTopicsClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "416b34a802308eac30e4192afc0ff99bb8dcc7f2",
3
+ "task_name": "SensitiveTopicsClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.399121,
9
+ "f1": 0.550058,
10
+ "lrap": 0.638698,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.401855,
14
+ "f1": 0.526036,
15
+ "lrap": 0.6199
16
+ },
17
+ {
18
+ "accuracy": 0.395508,
19
+ "f1": 0.567261,
20
+ "lrap": 0.650296
21
+ },
22
+ {
23
+ "accuracy": 0.394043,
24
+ "f1": 0.52953,
25
+ "lrap": 0.612115
26
+ },
27
+ {
28
+ "accuracy": 0.395508,
29
+ "f1": 0.546454,
30
+ "lrap": 0.629205
31
+ },
32
+ {
33
+ "accuracy": 0.395996,
34
+ "f1": 0.540874,
35
+ "lrap": 0.633158
36
+ },
37
+ {
38
+ "accuracy": 0.390625,
39
+ "f1": 0.547441,
40
+ "lrap": 0.643399
41
+ },
42
+ {
43
+ "accuracy": 0.414551,
44
+ "f1": 0.566625,
45
+ "lrap": 0.651923
46
+ },
47
+ {
48
+ "accuracy": 0.417969,
49
+ "f1": 0.567175,
50
+ "lrap": 0.648397
51
+ },
52
+ {
53
+ "accuracy": 0.398926,
54
+ "f1": 0.545177,
55
+ "lrap": 0.646647
56
+ },
57
+ {
58
+ "accuracy": 0.38623,
59
+ "f1": 0.56401,
60
+ "lrap": 0.651944
61
+ }
62
+ ],
63
+ "main_score": 0.399121,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 8.444582462310791,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/TERRa.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7b58f24536063837d644aab9a023c62199b2a612",
3
+ "task_name": "TERRa",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "dev": [
7
+ {
8
+ "similarity_accuracy": 0.664495,
9
+ "similarity_accuracy_threshold": 0.561947,
10
+ "similarity_f1": 0.701754,
11
+ "similarity_f1_threshold": 0.442617,
12
+ "similarity_precision": 0.569106,
13
+ "similarity_recall": 0.915033,
14
+ "similarity_ap": 0.655055,
15
+ "cosine_accuracy": 0.664495,
16
+ "cosine_accuracy_threshold": 0.561947,
17
+ "cosine_f1": 0.701754,
18
+ "cosine_f1_threshold": 0.442617,
19
+ "cosine_precision": 0.569106,
20
+ "cosine_recall": 0.915033,
21
+ "cosine_ap": 0.655055,
22
+ "manhattan_accuracy": 0.65798,
23
+ "manhattan_accuracy_threshold": 21.30641,
24
+ "manhattan_f1": 0.698254,
25
+ "manhattan_f1_threshold": 23.322489,
26
+ "manhattan_precision": 0.564516,
27
+ "manhattan_recall": 0.915033,
28
+ "manhattan_ap": 0.657363,
29
+ "euclidean_accuracy": 0.664495,
30
+ "euclidean_accuracy_threshold": 0.936004,
31
+ "euclidean_f1": 0.701754,
32
+ "euclidean_f1_threshold": 1.055822,
33
+ "euclidean_precision": 0.569106,
34
+ "euclidean_recall": 0.915033,
35
+ "euclidean_ap": 0.655055,
36
+ "dot_accuracy": 0.664495,
37
+ "dot_accuracy_threshold": 0.561947,
38
+ "dot_f1": 0.701754,
39
+ "dot_f1_threshold": 0.442617,
40
+ "dot_precision": 0.569106,
41
+ "dot_recall": 0.915033,
42
+ "dot_ap": 0.655055,
43
+ "max_accuracy": 0.664495,
44
+ "max_f1": 0.701754,
45
+ "max_precision": 0.569106,
46
+ "max_recall": 0.915033,
47
+ "max_ap": 0.657363,
48
+ "main_score": 0.657363,
49
+ "hf_subset": "default",
50
+ "languages": [
51
+ "rus-Cyrl"
52
+ ]
53
+ }
54
+ ]
55
+ },
56
+ "evaluation_time": 1.2874946594238281,
57
+ "kg_co2_emissions": null
58
+ }
results/sergeyzh__BERTA/914c8c8aed14042ed890fc2c662d5e9e66b2faa7/model_meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"name": "sergeyzh/BERTA", "revision": "914c8c8aed14042ed890fc2c662d5e9e66b2faa7", "release_date": "2025-03-10", "languages": ["rus_Cyrl"], "n_parameters": 128000000, "memory_usage_mb": 489.0, "max_tokens": 512.0, "embed_dim": 768, "license": "mit", "open_weights": true, "public_training_code": null, "public_training_data": null, "framework": ["Sentence Transformers", "PyTorch"], "reference": "https://huggingface.co/sergeyzh/BERTA", "similarity_fn_name": "cosine", "use_instructions": true, "training_datasets": {}, "adapted_from": "sergeyzh/LaBSE-ru-turbo", "superseded_by": null, "is_cross_encoder": null, "modalities": ["text"], "loader": null}
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/CEDRClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "c0ba03d058e3e1b2f3fd20518875a4563dd12db4",
3
+ "task_name": "CEDRClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.570404,
9
+ "f1": 0.586511,
10
+ "lrap": 0.801344,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.478215,
14
+ "f1": 0.552591,
15
+ "lrap": 0.818066
16
+ },
17
+ {
18
+ "accuracy": 0.580234,
19
+ "f1": 0.597915,
20
+ "lrap": 0.8144
21
+ },
22
+ {
23
+ "accuracy": 0.426142,
24
+ "f1": 0.551842,
25
+ "lrap": 0.797768
26
+ },
27
+ {
28
+ "accuracy": 0.687035,
29
+ "f1": 0.649755,
30
+ "lrap": 0.824867
31
+ },
32
+ {
33
+ "accuracy": 0.510627,
34
+ "f1": 0.558153,
35
+ "lrap": 0.814772
36
+ },
37
+ {
38
+ "accuracy": 0.457492,
39
+ "f1": 0.587055,
40
+ "lrap": 0.825558
41
+ },
42
+ {
43
+ "accuracy": 0.616897,
44
+ "f1": 0.545462,
45
+ "lrap": 0.746015
46
+ },
47
+ {
48
+ "accuracy": 0.626461,
49
+ "f1": 0.614466,
50
+ "lrap": 0.813231
51
+ },
52
+ {
53
+ "accuracy": 0.687566,
54
+ "f1": 0.636525,
55
+ "lrap": 0.80627
56
+ },
57
+ {
58
+ "accuracy": 0.633369,
59
+ "f1": 0.571346,
60
+ "lrap": 0.752497
61
+ }
62
+ ],
63
+ "main_score": 0.570404,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 2.9182145595550537,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/GeoreviewClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3765c0d1de6b7d264bc459433c45e5a75513839c",
3
+ "task_name": "GeoreviewClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.46416,
9
+ "f1": 0.448622,
10
+ "f1_weighted": 0.448608,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.477051,
14
+ "f1": 0.456706,
15
+ "f1_weighted": 0.456699
16
+ },
17
+ {
18
+ "accuracy": 0.464844,
19
+ "f1": 0.454226,
20
+ "f1_weighted": 0.454205
21
+ },
22
+ {
23
+ "accuracy": 0.472656,
24
+ "f1": 0.446708,
25
+ "f1_weighted": 0.446697
26
+ },
27
+ {
28
+ "accuracy": 0.45166,
29
+ "f1": 0.439403,
30
+ "f1_weighted": 0.439396
31
+ },
32
+ {
33
+ "accuracy": 0.453613,
34
+ "f1": 0.440016,
35
+ "f1_weighted": 0.440025
36
+ },
37
+ {
38
+ "accuracy": 0.448242,
39
+ "f1": 0.439501,
40
+ "f1_weighted": 0.439487
41
+ },
42
+ {
43
+ "accuracy": 0.481445,
44
+ "f1": 0.46716,
45
+ "f1_weighted": 0.467125
46
+ },
47
+ {
48
+ "accuracy": 0.487793,
49
+ "f1": 0.481135,
50
+ "f1_weighted": 0.481101
51
+ },
52
+ {
53
+ "accuracy": 0.436523,
54
+ "f1": 0.418829,
55
+ "f1_weighted": 0.418844
56
+ },
57
+ {
58
+ "accuracy": 0.467773,
59
+ "f1": 0.442535,
60
+ "f1_weighted": 0.442505
61
+ }
62
+ ],
63
+ "main_score": 0.46416,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 9.837544918060303,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/GeoreviewClusteringP2P.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "97a313c8fc85b47f13f33e7e9a95c1ad888c7fec",
3
+ "task_name": "GeoreviewClusteringP2P",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.703295,
11
+ 0.69596,
12
+ 0.707849,
13
+ 0.6918,
14
+ 0.692633,
15
+ 0.696793,
16
+ 0.696891,
17
+ 0.706144,
18
+ 0.700275,
19
+ 0.692976
20
+ ]
21
+ },
22
+ "v_measure": 0.698462,
23
+ "v_measure_std": 0.00542,
24
+ "main_score": 0.698462,
25
+ "hf_subset": "default",
26
+ "languages": [
27
+ "rus-Cyrl"
28
+ ]
29
+ }
30
+ ]
31
+ },
32
+ "evaluation_time": 5.817827463150024,
33
+ "kg_co2_emissions": null
34
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/HeadlineClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "2fe05ee6b5832cda29f2ef7aaad7b7fe6a3609eb",
3
+ "task_name": "HeadlineClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.882178,
9
+ "f1": 0.88198,
10
+ "f1_weighted": 0.88197,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.879395,
14
+ "f1": 0.879417,
15
+ "f1_weighted": 0.879404
16
+ },
17
+ {
18
+ "accuracy": 0.882812,
19
+ "f1": 0.882609,
20
+ "f1_weighted": 0.882602
21
+ },
22
+ {
23
+ "accuracy": 0.885742,
24
+ "f1": 0.885603,
25
+ "f1_weighted": 0.885597
26
+ },
27
+ {
28
+ "accuracy": 0.885742,
29
+ "f1": 0.885626,
30
+ "f1_weighted": 0.885615
31
+ },
32
+ {
33
+ "accuracy": 0.881348,
34
+ "f1": 0.881395,
35
+ "f1_weighted": 0.881383
36
+ },
37
+ {
38
+ "accuracy": 0.87793,
39
+ "f1": 0.877687,
40
+ "f1_weighted": 0.877674
41
+ },
42
+ {
43
+ "accuracy": 0.880859,
44
+ "f1": 0.880588,
45
+ "f1_weighted": 0.880581
46
+ },
47
+ {
48
+ "accuracy": 0.884277,
49
+ "f1": 0.88391,
50
+ "f1_weighted": 0.883899
51
+ },
52
+ {
53
+ "accuracy": 0.879883,
54
+ "f1": 0.879456,
55
+ "f1_weighted": 0.879449
56
+ },
57
+ {
58
+ "accuracy": 0.883789,
59
+ "f1": 0.883506,
60
+ "f1_weighted": 0.883499
61
+ }
62
+ ],
63
+ "main_score": 0.882178,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 5.1978394985198975,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/InappropriatenessClassification.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "601651fdc45ef243751676e62dd7a19f491c0285",
3
+ "task_name": "InappropriatenessClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.698438,
9
+ "f1": 0.693926,
10
+ "f1_weighted": 0.693926,
11
+ "ap": 0.646455,
12
+ "ap_weighted": 0.646455,
13
+ "scores_per_experiment": [
14
+ {
15
+ "accuracy": 0.722656,
16
+ "f1": 0.720822,
17
+ "f1_weighted": 0.720822,
18
+ "ap": 0.670496,
19
+ "ap_weighted": 0.670496
20
+ },
21
+ {
22
+ "accuracy": 0.693359,
23
+ "f1": 0.693284,
24
+ "f1_weighted": 0.693284,
25
+ "ap": 0.635274,
26
+ "ap_weighted": 0.635274
27
+ },
28
+ {
29
+ "accuracy": 0.703125,
30
+ "f1": 0.692883,
31
+ "f1_weighted": 0.692883,
32
+ "ap": 0.666562,
33
+ "ap_weighted": 0.666562
34
+ },
35
+ {
36
+ "accuracy": 0.639648,
37
+ "f1": 0.638717,
38
+ "f1_weighted": 0.638717,
39
+ "ap": 0.587528,
40
+ "ap_weighted": 0.587528
41
+ },
42
+ {
43
+ "accuracy": 0.702637,
44
+ "f1": 0.692771,
45
+ "f1_weighted": 0.692771,
46
+ "ap": 0.665317,
47
+ "ap_weighted": 0.665317
48
+ },
49
+ {
50
+ "accuracy": 0.672363,
51
+ "f1": 0.664731,
52
+ "f1_weighted": 0.664731,
53
+ "ap": 0.609004,
54
+ "ap_weighted": 0.609004
55
+ },
56
+ {
57
+ "accuracy": 0.736328,
58
+ "f1": 0.736264,
59
+ "f1_weighted": 0.736264,
60
+ "ap": 0.675817,
61
+ "ap_weighted": 0.675817
62
+ },
63
+ {
64
+ "accuracy": 0.71875,
65
+ "f1": 0.716203,
66
+ "f1_weighted": 0.716203,
67
+ "ap": 0.649605,
68
+ "ap_weighted": 0.649605
69
+ },
70
+ {
71
+ "accuracy": 0.691406,
72
+ "f1": 0.682447,
73
+ "f1_weighted": 0.682447,
74
+ "ap": 0.650873,
75
+ "ap_weighted": 0.650873
76
+ },
77
+ {
78
+ "accuracy": 0.704102,
79
+ "f1": 0.701136,
80
+ "f1_weighted": 0.701136,
81
+ "ap": 0.654072,
82
+ "ap_weighted": 0.654072
83
+ }
84
+ ],
85
+ "main_score": 0.698438,
86
+ "hf_subset": "default",
87
+ "languages": [
88
+ "rus-Cyrl"
89
+ ]
90
+ }
91
+ ]
92
+ },
93
+ "evaluation_time": 2.822906017303467,
94
+ "kg_co2_emissions": null
95
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/KinopoiskClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "5911f26666ac11af46cb9c6849d0dc80a378af24",
3
+ "task_name": "KinopoiskClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.595267,
9
+ "f1": 0.55905,
10
+ "f1_weighted": 0.55905,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.591333,
14
+ "f1": 0.566599,
15
+ "f1_weighted": 0.566599
16
+ },
17
+ {
18
+ "accuracy": 0.585333,
19
+ "f1": 0.53943,
20
+ "f1_weighted": 0.53943
21
+ },
22
+ {
23
+ "accuracy": 0.604,
24
+ "f1": 0.550982,
25
+ "f1_weighted": 0.550982
26
+ },
27
+ {
28
+ "accuracy": 0.588,
29
+ "f1": 0.548112,
30
+ "f1_weighted": 0.548112
31
+ },
32
+ {
33
+ "accuracy": 0.556,
34
+ "f1": 0.537351,
35
+ "f1_weighted": 0.537351
36
+ },
37
+ {
38
+ "accuracy": 0.614,
39
+ "f1": 0.582064,
40
+ "f1_weighted": 0.582064
41
+ },
42
+ {
43
+ "accuracy": 0.595333,
44
+ "f1": 0.543102,
45
+ "f1_weighted": 0.543102
46
+ },
47
+ {
48
+ "accuracy": 0.590667,
49
+ "f1": 0.575387,
50
+ "f1_weighted": 0.575387
51
+ },
52
+ {
53
+ "accuracy": 0.62,
54
+ "f1": 0.569014,
55
+ "f1_weighted": 0.569014
56
+ },
57
+ {
58
+ "accuracy": 0.608,
59
+ "f1": 0.578459,
60
+ "f1_weighted": 0.578459
61
+ }
62
+ ],
63
+ "main_score": 0.595267,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 13.222419500350952,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/MIRACLReranking.json ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "6d1962c527217f8927fca80f890f14f36b2802af",
3
+ "task_name": "MIRACLReranking",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "dev": [
7
+ {
8
+ "NDCG@1(MIRACL)": 0.55814,
9
+ "NDCG@3(MIRACL)": 0.55156,
10
+ "NDCG@5(MIRACL)": 0.57016,
11
+ "NDCG@10(MIRACL)": 0.60118,
12
+ "NDCG@20(MIRACL)": 0.62364,
13
+ "NDCG@100(MIRACL)": 0.63844,
14
+ "NDCG@1000(MIRACL)": 0.63844,
15
+ "MAP@1(MIRACL)": 0.33306,
16
+ "MAP@3(MIRACL)": 0.46189,
17
+ "MAP@5(MIRACL)": 0.49971,
18
+ "MAP@10(MIRACL)": 0.52966,
19
+ "MAP@20(MIRACL)": 0.54266,
20
+ "MAP@100(MIRACL)": 0.54805,
21
+ "MAP@1000(MIRACL)": 0.54805,
22
+ "Recall@1(MIRACL)": 0.33306,
23
+ "Recall@3(MIRACL)": 0.52189,
24
+ "Recall@5(MIRACL)": 0.60248,
25
+ "Recall@10(MIRACL)": 0.68646,
26
+ "Recall@20(MIRACL)": 0.74725,
27
+ "Recall@100(MIRACL)": 0.79952,
28
+ "Recall@1000(MIRACL)": 0.79952,
29
+ "P@1(MIRACL)": 0.55814,
30
+ "P@3(MIRACL)": 0.34403,
31
+ "P@5(MIRACL)": 0.25357,
32
+ "P@10(MIRACL)": 0.15525,
33
+ "P@20(MIRACL)": 0.08845,
34
+ "P@100(MIRACL)": 0.01953,
35
+ "P@1000(MIRACL)": 0.00195,
36
+ "nAUC_NDCG@1_max(MIRACL)": 0.519043,
37
+ "nAUC_NDCG@1_std(MIRACL)": 0.213521,
38
+ "nAUC_NDCG@1_diff1(MIRACL)": 0.361426,
39
+ "nAUC_NDCG@3_max(MIRACL)": 0.44736,
40
+ "nAUC_NDCG@3_std(MIRACL)": 0.19118,
41
+ "nAUC_NDCG@3_diff1(MIRACL)": 0.293165,
42
+ "nAUC_NDCG@5_max(MIRACL)": 0.450322,
43
+ "nAUC_NDCG@5_std(MIRACL)": 0.209937,
44
+ "nAUC_NDCG@5_diff1(MIRACL)": 0.282382,
45
+ "nAUC_NDCG@10_max(MIRACL)": 0.494004,
46
+ "nAUC_NDCG@10_std(MIRACL)": 0.23209,
47
+ "nAUC_NDCG@10_diff1(MIRACL)": 0.277441,
48
+ "nAUC_NDCG@20_max(MIRACL)": 0.516936,
49
+ "nAUC_NDCG@20_std(MIRACL)": 0.248519,
50
+ "nAUC_NDCG@20_diff1(MIRACL)": 0.26202,
51
+ "nAUC_NDCG@100_max(MIRACL)": 0.532602,
52
+ "nAUC_NDCG@100_std(MIRACL)": 0.248543,
53
+ "nAUC_NDCG@100_diff1(MIRACL)": 0.258746,
54
+ "nAUC_NDCG@1000_max(MIRACL)": 0.532602,
55
+ "nAUC_NDCG@1000_std(MIRACL)": 0.248543,
56
+ "nAUC_NDCG@1000_diff1(MIRACL)": 0.258746,
57
+ "nAUC_MAP@1_max(MIRACL)": 0.218248,
58
+ "nAUC_MAP@1_std(MIRACL)": 0.056505,
59
+ "nAUC_MAP@1_diff1(MIRACL)": 0.387698,
60
+ "nAUC_MAP@3_max(MIRACL)": 0.341724,
61
+ "nAUC_MAP@3_std(MIRACL)": 0.122016,
62
+ "nAUC_MAP@3_diff1(MIRACL)": 0.325925,
63
+ "nAUC_MAP@5_max(MIRACL)": 0.382434,
64
+ "nAUC_MAP@5_std(MIRACL)": 0.15931,
65
+ "nAUC_MAP@5_diff1(MIRACL)": 0.309835,
66
+ "nAUC_MAP@10_max(MIRACL)": 0.422968,
67
+ "nAUC_MAP@10_std(MIRACL)": 0.183142,
68
+ "nAUC_MAP@10_diff1(MIRACL)": 0.303733,
69
+ "nAUC_MAP@20_max(MIRACL)": 0.436873,
70
+ "nAUC_MAP@20_std(MIRACL)": 0.192985,
71
+ "nAUC_MAP@20_diff1(MIRACL)": 0.295674,
72
+ "nAUC_MAP@100_max(MIRACL)": 0.443055,
73
+ "nAUC_MAP@100_std(MIRACL)": 0.193956,
74
+ "nAUC_MAP@100_diff1(MIRACL)": 0.29514,
75
+ "nAUC_MAP@1000_max(MIRACL)": 0.443055,
76
+ "nAUC_MAP@1000_std(MIRACL)": 0.193956,
77
+ "nAUC_MAP@1000_diff1(MIRACL)": 0.29514,
78
+ "nAUC_Recall@1_max(MIRACL)": 0.218248,
79
+ "nAUC_Recall@1_std(MIRACL)": 0.056505,
80
+ "nAUC_Recall@1_diff1(MIRACL)": 0.387698,
81
+ "nAUC_Recall@3_max(MIRACL)": 0.326202,
82
+ "nAUC_Recall@3_std(MIRACL)": 0.150243,
83
+ "nAUC_Recall@3_diff1(MIRACL)": 0.262242,
84
+ "nAUC_Recall@5_max(MIRACL)": 0.392855,
85
+ "nAUC_Recall@5_std(MIRACL)": 0.205062,
86
+ "nAUC_Recall@5_diff1(MIRACL)": 0.218314,
87
+ "nAUC_Recall@10_max(MIRACL)": 0.511434,
88
+ "nAUC_Recall@10_std(MIRACL)": 0.270613,
89
+ "nAUC_Recall@10_diff1(MIRACL)": 0.184093,
90
+ "nAUC_Recall@20_max(MIRACL)": 0.600383,
91
+ "nAUC_Recall@20_std(MIRACL)": 0.340011,
92
+ "nAUC_Recall@20_diff1(MIRACL)": 0.131679,
93
+ "nAUC_Recall@100_max(MIRACL)": 0.707588,
94
+ "nAUC_Recall@100_std(MIRACL)": 0.357172,
95
+ "nAUC_Recall@100_diff1(MIRACL)": 0.098075,
96
+ "nAUC_Recall@1000_max(MIRACL)": 0.707588,
97
+ "nAUC_Recall@1000_std(MIRACL)": 0.357172,
98
+ "nAUC_Recall@1000_diff1(MIRACL)": 0.098075,
99
+ "nAUC_P@1_max(MIRACL)": 0.519043,
100
+ "nAUC_P@1_std(MIRACL)": 0.213521,
101
+ "nAUC_P@1_diff1(MIRACL)": 0.361426,
102
+ "nAUC_P@3_max(MIRACL)": 0.465491,
103
+ "nAUC_P@3_std(MIRACL)": 0.23009,
104
+ "nAUC_P@3_diff1(MIRACL)": 0.044383,
105
+ "nAUC_P@5_max(MIRACL)": 0.434656,
106
+ "nAUC_P@5_std(MIRACL)": 0.260258,
107
+ "nAUC_P@5_diff1(MIRACL)": -0.019016,
108
+ "nAUC_P@10_max(MIRACL)": 0.401164,
109
+ "nAUC_P@10_std(MIRACL)": 0.241747,
110
+ "nAUC_P@10_diff1(MIRACL)": -0.051254,
111
+ "nAUC_P@20_max(MIRACL)": 0.366172,
112
+ "nAUC_P@20_std(MIRACL)": 0.222595,
113
+ "nAUC_P@20_diff1(MIRACL)": -0.09334,
114
+ "nAUC_P@100_max(MIRACL)": 0.335503,
115
+ "nAUC_P@100_std(MIRACL)": 0.189048,
116
+ "nAUC_P@100_diff1(MIRACL)": -0.099733,
117
+ "nAUC_P@1000_max(MIRACL)": 0.335503,
118
+ "nAUC_P@1000_std(MIRACL)": 0.189048,
119
+ "nAUC_P@1000_diff1(MIRACL)": -0.099733,
120
+ "main_score": 0.60118,
121
+ "hf_subset": "ru",
122
+ "languages": [
123
+ "rus-Cyrl"
124
+ ]
125
+ }
126
+ ]
127
+ },
128
+ "evaluation_time": 143.46533703804016,
129
+ "kg_co2_emissions": null
130
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/MIRACLRetrieval.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "task_name": "MIRACLRetrieval",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "dev": [
7
+ {
8
+ "ndcg_at_1": 0.55431,
9
+ "ndcg_at_3": 0.519,
10
+ "ndcg_at_5": 0.53581,
11
+ "ndcg_at_10": 0.5709,
12
+ "ndcg_at_20": 0.60307,
13
+ "ndcg_at_100": 0.63723,
14
+ "ndcg_at_1000": 0.65163,
15
+ "map_at_1": 0.27186,
16
+ "map_at_3": 0.39502,
17
+ "map_at_5": 0.43644,
18
+ "map_at_10": 0.46754,
19
+ "map_at_20": 0.48355,
20
+ "map_at_100": 0.49366,
21
+ "map_at_1000": 0.49473,
22
+ "recall_at_1": 0.27186,
23
+ "recall_at_3": 0.46269,
24
+ "recall_at_5": 0.55601,
25
+ "recall_at_10": 0.66361,
26
+ "recall_at_20": 0.75243,
27
+ "recall_at_100": 0.87659,
28
+ "recall_at_1000": 0.95853,
29
+ "precision_at_1": 0.55431,
30
+ "precision_at_3": 0.35916,
31
+ "precision_at_5": 0.27141,
32
+ "precision_at_10": 0.16957,
33
+ "precision_at_20": 0.10028,
34
+ "precision_at_100": 0.02466,
35
+ "precision_at_1000": 0.00273,
36
+ "mrr_at_1": 0.554313,
37
+ "mrr_at_3": 0.642173,
38
+ "mrr_at_5": 0.654752,
39
+ "mrr_at_10": 0.663527,
40
+ "mrr_at_20": 0.666026,
41
+ "mrr_at_100": 0.667106,
42
+ "mrr_at_1000": 0.667243,
43
+ "nauc_ndcg_at_1_max": 0.449053,
44
+ "nauc_ndcg_at_1_std": 0.177802,
45
+ "nauc_ndcg_at_1_diff1": 0.444503,
46
+ "nauc_ndcg_at_3_max": 0.398994,
47
+ "nauc_ndcg_at_3_std": 0.173248,
48
+ "nauc_ndcg_at_3_diff1": 0.323636,
49
+ "nauc_ndcg_at_5_max": 0.381131,
50
+ "nauc_ndcg_at_5_std": 0.151429,
51
+ "nauc_ndcg_at_5_diff1": 0.325224,
52
+ "nauc_ndcg_at_10_max": 0.407494,
53
+ "nauc_ndcg_at_10_std": 0.183845,
54
+ "nauc_ndcg_at_10_diff1": 0.330863,
55
+ "nauc_ndcg_at_20_max": 0.431071,
56
+ "nauc_ndcg_at_20_std": 0.214671,
57
+ "nauc_ndcg_at_20_diff1": 0.330457,
58
+ "nauc_ndcg_at_100_max": 0.451676,
59
+ "nauc_ndcg_at_100_std": 0.245317,
60
+ "nauc_ndcg_at_100_diff1": 0.340654,
61
+ "nauc_ndcg_at_1000_max": 0.450708,
62
+ "nauc_ndcg_at_1000_std": 0.239727,
63
+ "nauc_ndcg_at_1000_diff1": 0.341284,
64
+ "nauc_map_at_1_max": 0.199729,
65
+ "nauc_map_at_1_std": -0.02255,
66
+ "nauc_map_at_1_diff1": 0.359379,
67
+ "nauc_map_at_3_max": 0.291782,
68
+ "nauc_map_at_3_std": 0.062132,
69
+ "nauc_map_at_3_diff1": 0.316299,
70
+ "nauc_map_at_5_max": 0.322706,
71
+ "nauc_map_at_5_std": 0.090771,
72
+ "nauc_map_at_5_diff1": 0.314453,
73
+ "nauc_map_at_10_max": 0.352173,
74
+ "nauc_map_at_10_std": 0.124584,
75
+ "nauc_map_at_10_diff1": 0.315591,
76
+ "nauc_map_at_20_max": 0.366149,
77
+ "nauc_map_at_20_std": 0.14145,
78
+ "nauc_map_at_20_diff1": 0.315177,
79
+ "nauc_map_at_100_max": 0.374989,
80
+ "nauc_map_at_100_std": 0.153138,
81
+ "nauc_map_at_100_diff1": 0.317427,
82
+ "nauc_map_at_1000_max": 0.375246,
83
+ "nauc_map_at_1000_std": 0.153349,
84
+ "nauc_map_at_1000_diff1": 0.317505,
85
+ "nauc_recall_at_1_max": 0.199729,
86
+ "nauc_recall_at_1_std": -0.02255,
87
+ "nauc_recall_at_1_diff1": 0.359379,
88
+ "nauc_recall_at_3_max": 0.25336,
89
+ "nauc_recall_at_3_std": 0.061852,
90
+ "nauc_recall_at_3_diff1": 0.255928,
91
+ "nauc_recall_at_5_max": 0.276545,
92
+ "nauc_recall_at_5_std": 0.085246,
93
+ "nauc_recall_at_5_diff1": 0.244297,
94
+ "nauc_recall_at_10_max": 0.332597,
95
+ "nauc_recall_at_10_std": 0.179314,
96
+ "nauc_recall_at_10_diff1": 0.228538,
97
+ "nauc_recall_at_20_max": 0.387653,
98
+ "nauc_recall_at_20_std": 0.269982,
99
+ "nauc_recall_at_20_diff1": 0.217891,
100
+ "nauc_recall_at_100_max": 0.50847,
101
+ "nauc_recall_at_100_std": 0.495238,
102
+ "nauc_recall_at_100_diff1": 0.247615,
103
+ "nauc_recall_at_1000_max": 0.656631,
104
+ "nauc_recall_at_1000_std": 0.738805,
105
+ "nauc_recall_at_1000_diff1": 0.182967,
106
+ "nauc_precision_at_1_max": 0.449053,
107
+ "nauc_precision_at_1_std": 0.177802,
108
+ "nauc_precision_at_1_diff1": 0.444503,
109
+ "nauc_precision_at_3_max": 0.439964,
110
+ "nauc_precision_at_3_std": 0.306541,
111
+ "nauc_precision_at_3_diff1": 0.161587,
112
+ "nauc_precision_at_5_max": 0.394769,
113
+ "nauc_precision_at_5_std": 0.310302,
114
+ "nauc_precision_at_5_diff1": 0.099505,
115
+ "nauc_precision_at_10_max": 0.372877,
116
+ "nauc_precision_at_10_std": 0.354607,
117
+ "nauc_precision_at_10_diff1": 0.055965,
118
+ "nauc_precision_at_20_max": 0.341717,
119
+ "nauc_precision_at_20_std": 0.358702,
120
+ "nauc_precision_at_20_diff1": 0.012705,
121
+ "nauc_precision_at_100_max": 0.270786,
122
+ "nauc_precision_at_100_std": 0.342444,
123
+ "nauc_precision_at_100_diff1": -0.019592,
124
+ "nauc_precision_at_1000_max": 0.209252,
125
+ "nauc_precision_at_1000_std": 0.287951,
126
+ "nauc_precision_at_1000_diff1": -0.046472,
127
+ "nauc_mrr_at_1_max": 0.449053,
128
+ "nauc_mrr_at_1_std": 0.177802,
129
+ "nauc_mrr_at_1_diff1": 0.444503,
130
+ "nauc_mrr_at_3_max": 0.485199,
131
+ "nauc_mrr_at_3_std": 0.228999,
132
+ "nauc_mrr_at_3_diff1": 0.422257,
133
+ "nauc_mrr_at_5_max": 0.485177,
134
+ "nauc_mrr_at_5_std": 0.223536,
135
+ "nauc_mrr_at_5_diff1": 0.425592,
136
+ "nauc_mrr_at_10_max": 0.488423,
137
+ "nauc_mrr_at_10_std": 0.232735,
138
+ "nauc_mrr_at_10_diff1": 0.426943,
139
+ "nauc_mrr_at_20_max": 0.48849,
140
+ "nauc_mrr_at_20_std": 0.233615,
141
+ "nauc_mrr_at_20_diff1": 0.427241,
142
+ "nauc_mrr_at_100_max": 0.487703,
143
+ "nauc_mrr_at_100_std": 0.232691,
144
+ "nauc_mrr_at_100_diff1": 0.427563,
145
+ "nauc_mrr_at_1000_max": 0.487543,
146
+ "nauc_mrr_at_1000_std": 0.232427,
147
+ "nauc_mrr_at_1000_diff1": 0.427497,
148
+ "main_score": 0.5709,
149
+ "hf_subset": "ru",
150
+ "languages": [
151
+ "rus-Cyrl"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 7931.134541988373,
157
+ "kg_co2_emissions": null
158
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/MassiveIntentClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "4672e20407010da34463acc759c162ca9734bca6",
3
+ "task_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.647664,
9
+ "f1": 0.608599,
10
+ "f1_weighted": 0.633221,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.659616,
14
+ "f1": 0.625382,
15
+ "f1_weighted": 0.649549
16
+ },
17
+ {
18
+ "accuracy": 0.664535,
19
+ "f1": 0.621902,
20
+ "f1_weighted": 0.657157
21
+ },
22
+ {
23
+ "accuracy": 0.648795,
24
+ "f1": 0.60208,
25
+ "f1_weighted": 0.638126
26
+ },
27
+ {
28
+ "accuracy": 0.665027,
29
+ "f1": 0.620407,
30
+ "f1_weighted": 0.653705
31
+ },
32
+ {
33
+ "accuracy": 0.648303,
34
+ "f1": 0.604417,
35
+ "f1_weighted": 0.630373
36
+ },
37
+ {
38
+ "accuracy": 0.625676,
39
+ "f1": 0.592523,
40
+ "f1_weighted": 0.604256
41
+ },
42
+ {
43
+ "accuracy": 0.647319,
44
+ "f1": 0.602269,
45
+ "f1_weighted": 0.633531
46
+ },
47
+ {
48
+ "accuracy": 0.620758,
49
+ "f1": 0.581883,
50
+ "f1_weighted": 0.604521
51
+ },
52
+ {
53
+ "accuracy": 0.632563,
54
+ "f1": 0.613927,
55
+ "f1_weighted": 0.606582
56
+ },
57
+ {
58
+ "accuracy": 0.664043,
59
+ "f1": 0.621196,
60
+ "f1_weighted": 0.65441
61
+ }
62
+ ],
63
+ "main_score": 0.647664,
64
+ "hf_subset": "ru",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.645595,
73
+ "f1": 0.625037,
74
+ "f1_weighted": 0.63329,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.655346,
78
+ "f1": 0.638349,
79
+ "f1_weighted": 0.647148
80
+ },
81
+ {
82
+ "accuracy": 0.670141,
83
+ "f1": 0.63709,
84
+ "f1_weighted": 0.66458
85
+ },
86
+ {
87
+ "accuracy": 0.635844,
88
+ "f1": 0.611981,
89
+ "f1_weighted": 0.628635
90
+ },
91
+ {
92
+ "accuracy": 0.659381,
93
+ "f1": 0.626729,
94
+ "f1_weighted": 0.651412
95
+ },
96
+ {
97
+ "accuracy": 0.640888,
98
+ "f1": 0.624626,
99
+ "f1_weighted": 0.62564
100
+ },
101
+ {
102
+ "accuracy": 0.621385,
103
+ "f1": 0.608287,
104
+ "f1_weighted": 0.602768
105
+ },
106
+ {
107
+ "accuracy": 0.648285,
108
+ "f1": 0.628397,
109
+ "f1_weighted": 0.637509
110
+ },
111
+ {
112
+ "accuracy": 0.63349,
113
+ "f1": 0.618394,
114
+ "f1_weighted": 0.619325
115
+ },
116
+ {
117
+ "accuracy": 0.629455,
118
+ "f1": 0.614532,
119
+ "f1_weighted": 0.603522
120
+ },
121
+ {
122
+ "accuracy": 0.661735,
123
+ "f1": 0.641986,
124
+ "f1_weighted": 0.652357
125
+ }
126
+ ],
127
+ "main_score": 0.645595,
128
+ "hf_subset": "ru",
129
+ "languages": [
130
+ "rus-Cyrl"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 32.8644540309906,
136
+ "kg_co2_emissions": null
137
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "fad2c6e8459f9e1c45d9315f4953d921437d70f8",
3
+ "task_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "validation": [
7
+ {
8
+ "accuracy": 0.724299,
9
+ "f1": 0.718652,
10
+ "f1_weighted": 0.720962,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.749139,
14
+ "f1": 0.739381,
15
+ "f1_weighted": 0.748074
16
+ },
17
+ {
18
+ "accuracy": 0.711264,
19
+ "f1": 0.709769,
20
+ "f1_weighted": 0.707836
21
+ },
22
+ {
23
+ "accuracy": 0.73635,
24
+ "f1": 0.725275,
25
+ "f1_weighted": 0.732857
26
+ },
27
+ {
28
+ "accuracy": 0.724053,
29
+ "f1": 0.717275,
30
+ "f1_weighted": 0.722571
31
+ },
32
+ {
33
+ "accuracy": 0.743237,
34
+ "f1": 0.733784,
35
+ "f1_weighted": 0.737775
36
+ },
37
+ {
38
+ "accuracy": 0.669946,
39
+ "f1": 0.671646,
40
+ "f1_weighted": 0.658632
41
+ },
42
+ {
43
+ "accuracy": 0.706345,
44
+ "f1": 0.702719,
45
+ "f1_weighted": 0.705804
46
+ },
47
+ {
48
+ "accuracy": 0.708805,
49
+ "f1": 0.707502,
50
+ "f1_weighted": 0.707387
51
+ },
52
+ {
53
+ "accuracy": 0.749631,
54
+ "f1": 0.741891,
55
+ "f1_weighted": 0.74515
56
+ },
57
+ {
58
+ "accuracy": 0.74422,
59
+ "f1": 0.737278,
60
+ "f1_weighted": 0.74353
61
+ }
62
+ ],
63
+ "main_score": 0.724299,
64
+ "hf_subset": "ru",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ],
70
+ "test": [
71
+ {
72
+ "accuracy": 0.722327,
73
+ "f1": 0.719225,
74
+ "f1_weighted": 0.718613,
75
+ "scores_per_experiment": [
76
+ {
77
+ "accuracy": 0.729657,
78
+ "f1": 0.725178,
79
+ "f1_weighted": 0.727967
80
+ },
81
+ {
82
+ "accuracy": 0.710491,
83
+ "f1": 0.709409,
84
+ "f1_weighted": 0.704448
85
+ },
86
+ {
87
+ "accuracy": 0.744116,
88
+ "f1": 0.734157,
89
+ "f1_weighted": 0.741614
90
+ },
91
+ {
92
+ "accuracy": 0.720915,
93
+ "f1": 0.709983,
94
+ "f1_weighted": 0.71872
95
+ },
96
+ {
97
+ "accuracy": 0.729657,
98
+ "f1": 0.721675,
99
+ "f1_weighted": 0.720982
100
+ },
101
+ {
102
+ "accuracy": 0.676866,
103
+ "f1": 0.676432,
104
+ "f1_weighted": 0.666904
105
+ },
106
+ {
107
+ "accuracy": 0.713853,
108
+ "f1": 0.712989,
109
+ "f1_weighted": 0.713212
110
+ },
111
+ {
112
+ "accuracy": 0.714526,
113
+ "f1": 0.717238,
114
+ "f1_weighted": 0.714356
115
+ },
116
+ {
117
+ "accuracy": 0.743107,
118
+ "f1": 0.743454,
119
+ "f1_weighted": 0.740183
120
+ },
121
+ {
122
+ "accuracy": 0.740081,
123
+ "f1": 0.741738,
124
+ "f1_weighted": 0.737744
125
+ }
126
+ ],
127
+ "main_score": 0.722327,
128
+ "hf_subset": "ru",
129
+ "languages": [
130
+ "rus-Cyrl"
131
+ ]
132
+ }
133
+ ]
134
+ },
135
+ "evaluation_time": 6.143388509750366,
136
+ "kg_co2_emissions": null
137
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RUParaPhraserSTS.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "43265056790b8f7c59e0139acb4be0a8dad2c8f4",
3
+ "task_name": "RUParaPhraserSTS",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "pearson": 0.674719,
9
+ "spearman": 0.751039,
10
+ "cosine_pearson": 0.674719,
11
+ "cosine_spearman": 0.75104,
12
+ "manhattan_pearson": 0.724039,
13
+ "manhattan_spearman": 0.752023,
14
+ "euclidean_pearson": 0.723391,
15
+ "euclidean_spearman": 0.751041,
16
+ "main_score": 0.75104,
17
+ "hf_subset": "default",
18
+ "languages": [
19
+ "rus-Cyrl"
20
+ ]
21
+ }
22
+ ]
23
+ },
24
+ "evaluation_time": 2.470493793487549,
25
+ "kg_co2_emissions": null
26
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RiaNewsRetrieval.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "82374b0bbacda6114f39ff9c5b925fa1512ca5d7",
3
+ "task_name": "RiaNewsRetrieval",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.5914,
9
+ "ndcg_at_3": 0.68444,
10
+ "ndcg_at_5": 0.70405,
11
+ "ndcg_at_10": 0.72031,
12
+ "ndcg_at_20": 0.72958,
13
+ "ndcg_at_100": 0.7397,
14
+ "ndcg_at_1000": 0.74441,
15
+ "map_at_1": 0.5914,
16
+ "map_at_3": 0.662,
17
+ "map_at_5": 0.67293,
18
+ "map_at_10": 0.67979,
19
+ "map_at_20": 0.68236,
20
+ "map_at_100": 0.6838,
21
+ "map_at_1000": 0.68398,
22
+ "recall_at_1": 0.5914,
23
+ "recall_at_3": 0.7492,
24
+ "recall_at_5": 0.7966,
25
+ "recall_at_10": 0.8461,
26
+ "recall_at_20": 0.8826,
27
+ "recall_at_100": 0.9365,
28
+ "recall_at_1000": 0.9737,
29
+ "precision_at_1": 0.5914,
30
+ "precision_at_3": 0.24973,
31
+ "precision_at_5": 0.15932,
32
+ "precision_at_10": 0.08461,
33
+ "precision_at_20": 0.04413,
34
+ "precision_at_100": 0.00937,
35
+ "precision_at_1000": 0.00097,
36
+ "mrr_at_1": 0.5916,
37
+ "mrr_at_3": 0.6621,
38
+ "mrr_at_5": 0.673015,
39
+ "mrr_at_10": 0.679893,
40
+ "mrr_at_20": 0.682456,
41
+ "mrr_at_100": 0.683897,
42
+ "mrr_at_1000": 0.684073,
43
+ "nauc_ndcg_at_1_max": 0.293905,
44
+ "nauc_ndcg_at_1_std": -0.114295,
45
+ "nauc_ndcg_at_1_diff1": 0.731349,
46
+ "nauc_ndcg_at_3_max": 0.322969,
47
+ "nauc_ndcg_at_3_std": -0.099189,
48
+ "nauc_ndcg_at_3_diff1": 0.688411,
49
+ "nauc_ndcg_at_5_max": 0.330399,
50
+ "nauc_ndcg_at_5_std": -0.089704,
51
+ "nauc_ndcg_at_5_diff1": 0.684593,
52
+ "nauc_ndcg_at_10_max": 0.336026,
53
+ "nauc_ndcg_at_10_std": -0.081989,
54
+ "nauc_ndcg_at_10_diff1": 0.683654,
55
+ "nauc_ndcg_at_20_max": 0.337509,
56
+ "nauc_ndcg_at_20_std": -0.073882,
57
+ "nauc_ndcg_at_20_diff1": 0.683799,
58
+ "nauc_ndcg_at_100_max": 0.335683,
59
+ "nauc_ndcg_at_100_std": -0.073841,
60
+ "nauc_ndcg_at_100_diff1": 0.685885,
61
+ "nauc_ndcg_at_1000_max": 0.333306,
62
+ "nauc_ndcg_at_1000_std": -0.07774,
63
+ "nauc_ndcg_at_1000_diff1": 0.687929,
64
+ "nauc_map_at_1_max": 0.293905,
65
+ "nauc_map_at_1_std": -0.114295,
66
+ "nauc_map_at_1_diff1": 0.731349,
67
+ "nauc_map_at_3_max": 0.315264,
68
+ "nauc_map_at_3_std": -0.103336,
69
+ "nauc_map_at_3_diff1": 0.699074,
70
+ "nauc_map_at_5_max": 0.31901,
71
+ "nauc_map_at_5_std": -0.098417,
72
+ "nauc_map_at_5_diff1": 0.697469,
73
+ "nauc_map_at_10_max": 0.320966,
74
+ "nauc_map_at_10_std": -0.095819,
75
+ "nauc_map_at_10_diff1": 0.697364,
76
+ "nauc_map_at_20_max": 0.321199,
77
+ "nauc_map_at_20_std": -0.093935,
78
+ "nauc_map_at_20_diff1": 0.69755,
79
+ "nauc_map_at_100_max": 0.32095,
80
+ "nauc_map_at_100_std": -0.093909,
81
+ "nauc_map_at_100_diff1": 0.697821,
82
+ "nauc_map_at_1000_max": 0.32086,
83
+ "nauc_map_at_1000_std": -0.094024,
84
+ "nauc_map_at_1000_diff1": 0.697887,
85
+ "nauc_recall_at_1_max": 0.293905,
86
+ "nauc_recall_at_1_std": -0.114295,
87
+ "nauc_recall_at_1_diff1": 0.731349,
88
+ "nauc_recall_at_3_max": 0.350296,
89
+ "nauc_recall_at_3_std": -0.084403,
90
+ "nauc_recall_at_3_diff1": 0.650974,
91
+ "nauc_recall_at_5_max": 0.378037,
92
+ "nauc_recall_at_5_std": -0.052448,
93
+ "nauc_recall_at_5_diff1": 0.631774,
94
+ "nauc_recall_at_10_max": 0.418432,
95
+ "nauc_recall_at_10_std": -0.003025,
96
+ "nauc_recall_at_10_diff1": 0.612263,
97
+ "nauc_recall_at_20_max": 0.452591,
98
+ "nauc_recall_at_20_std": 0.080814,
99
+ "nauc_recall_at_20_diff1": 0.593166,
100
+ "nauc_recall_at_100_max": 0.509259,
101
+ "nauc_recall_at_100_std": 0.198832,
102
+ "nauc_recall_at_100_diff1": 0.560345,
103
+ "nauc_recall_at_1000_max": 0.607513,
104
+ "nauc_recall_at_1000_std": 0.324765,
105
+ "nauc_recall_at_1000_diff1": 0.515381,
106
+ "nauc_precision_at_1_max": 0.293905,
107
+ "nauc_precision_at_1_std": -0.114295,
108
+ "nauc_precision_at_1_diff1": 0.731349,
109
+ "nauc_precision_at_3_max": 0.350296,
110
+ "nauc_precision_at_3_std": -0.084403,
111
+ "nauc_precision_at_3_diff1": 0.650974,
112
+ "nauc_precision_at_5_max": 0.378037,
113
+ "nauc_precision_at_5_std": -0.052448,
114
+ "nauc_precision_at_5_diff1": 0.631774,
115
+ "nauc_precision_at_10_max": 0.418432,
116
+ "nauc_precision_at_10_std": -0.003025,
117
+ "nauc_precision_at_10_diff1": 0.612263,
118
+ "nauc_precision_at_20_max": 0.452591,
119
+ "nauc_precision_at_20_std": 0.080814,
120
+ "nauc_precision_at_20_diff1": 0.593166,
121
+ "nauc_precision_at_100_max": 0.509259,
122
+ "nauc_precision_at_100_std": 0.198832,
123
+ "nauc_precision_at_100_diff1": 0.560345,
124
+ "nauc_precision_at_1000_max": 0.607513,
125
+ "nauc_precision_at_1000_std": 0.324765,
126
+ "nauc_precision_at_1000_diff1": 0.515381,
127
+ "nauc_mrr_at_1_max": 0.294083,
128
+ "nauc_mrr_at_1_std": -0.113999,
129
+ "nauc_mrr_at_1_diff1": 0.730884,
130
+ "nauc_mrr_at_3_max": 0.315369,
131
+ "nauc_mrr_at_3_std": -0.103197,
132
+ "nauc_mrr_at_3_diff1": 0.698793,
133
+ "nauc_mrr_at_5_max": 0.319088,
134
+ "nauc_mrr_at_5_std": -0.098281,
135
+ "nauc_mrr_at_5_diff1": 0.697151,
136
+ "nauc_mrr_at_10_max": 0.321069,
137
+ "nauc_mrr_at_10_std": -0.09561,
138
+ "nauc_mrr_at_10_diff1": 0.697073,
139
+ "nauc_mrr_at_20_max": 0.321307,
140
+ "nauc_mrr_at_20_std": -0.09378,
141
+ "nauc_mrr_at_20_diff1": 0.697247,
142
+ "nauc_mrr_at_100_max": 0.321059,
143
+ "nauc_mrr_at_100_std": -0.093753,
144
+ "nauc_mrr_at_100_diff1": 0.697517,
145
+ "nauc_mrr_at_1000_max": 0.320968,
146
+ "nauc_mrr_at_1000_std": -0.093868,
147
+ "nauc_mrr_at_1000_diff1": 0.697583,
148
+ "main_score": 0.72031,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "rus-Cyrl"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 1400.8417766094208,
157
+ "kg_co2_emissions": null
158
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuBQReranking.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "2e96b8f098fa4b0950fc58eacadeb31c0d0c7fa2",
3
+ "task_name": "RuBQReranking",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "map": 0.711186,
9
+ "mrr": 0.77272,
10
+ "nAUC_map_max": 0.20826,
11
+ "nAUC_map_std": 0.115449,
12
+ "nAUC_map_diff1": 0.403793,
13
+ "nAUC_mrr_max": 0.257487,
14
+ "nAUC_mrr_std": 0.164996,
15
+ "nAUC_mrr_diff1": 0.451187,
16
+ "main_score": 0.711186,
17
+ "hf_subset": "default",
18
+ "languages": [
19
+ "rus-Cyrl"
20
+ ]
21
+ }
22
+ ]
23
+ },
24
+ "evaluation_time": 42.09844660758972,
25
+ "kg_co2_emissions": null
26
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuBQRetrieval.json ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e19b6ffa60b3bc248e0b41f4cc37c26a55c2a67b",
3
+ "task_name": "RuBQRetrieval",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "ndcg_at_1": 0.56856,
9
+ "ndcg_at_3": 0.58401,
10
+ "ndcg_at_5": 0.62177,
11
+ "ndcg_at_10": 0.65435,
12
+ "ndcg_at_20": 0.67266,
13
+ "ndcg_at_100": 0.6919,
14
+ "ndcg_at_1000": 0.69939,
15
+ "map_at_1": 0.39483,
16
+ "map_at_3": 0.52327,
17
+ "map_at_5": 0.55633,
18
+ "map_at_10": 0.57491,
19
+ "map_at_20": 0.58174,
20
+ "map_at_100": 0.58569,
21
+ "map_at_1000": 0.58612,
22
+ "recall_at_1": 0.39483,
23
+ "recall_at_3": 0.61193,
24
+ "recall_at_5": 0.70004,
25
+ "recall_at_10": 0.78502,
26
+ "recall_at_20": 0.84704,
27
+ "recall_at_100": 0.93518,
28
+ "recall_at_1000": 0.98412,
29
+ "precision_at_1": 0.56856,
30
+ "precision_at_3": 0.31659,
31
+ "precision_at_5": 0.22364,
32
+ "precision_at_10": 0.1279,
33
+ "precision_at_20": 0.06947,
34
+ "precision_at_100": 0.0155,
35
+ "precision_at_1000": 0.00165,
36
+ "mrr_at_1": 0.56974,
37
+ "mrr_at_3": 0.655831,
38
+ "mrr_at_5": 0.669041,
39
+ "mrr_at_10": 0.676821,
40
+ "mrr_at_20": 0.679916,
41
+ "mrr_at_100": 0.681177,
42
+ "mrr_at_1000": 0.681303,
43
+ "nauc_ndcg_at_1_max": 0.320745,
44
+ "nauc_ndcg_at_1_std": -0.040708,
45
+ "nauc_ndcg_at_1_diff1": 0.507783,
46
+ "nauc_ndcg_at_3_max": 0.302455,
47
+ "nauc_ndcg_at_3_std": -0.035899,
48
+ "nauc_ndcg_at_3_diff1": 0.399797,
49
+ "nauc_ndcg_at_5_max": 0.315547,
50
+ "nauc_ndcg_at_5_std": -0.037695,
51
+ "nauc_ndcg_at_5_diff1": 0.394691,
52
+ "nauc_ndcg_at_10_max": 0.335905,
53
+ "nauc_ndcg_at_10_std": -0.019011,
54
+ "nauc_ndcg_at_10_diff1": 0.397614,
55
+ "nauc_ndcg_at_20_max": 0.338677,
56
+ "nauc_ndcg_at_20_std": -0.006954,
57
+ "nauc_ndcg_at_20_diff1": 0.399567,
58
+ "nauc_ndcg_at_100_max": 0.335997,
59
+ "nauc_ndcg_at_100_std": -0.009213,
60
+ "nauc_ndcg_at_100_diff1": 0.408431,
61
+ "nauc_ndcg_at_1000_max": 0.3321,
62
+ "nauc_ndcg_at_1000_std": -0.014573,
63
+ "nauc_ndcg_at_1000_diff1": 0.410539,
64
+ "nauc_map_at_1_max": 0.2142,
65
+ "nauc_map_at_1_std": -0.058618,
66
+ "nauc_map_at_1_diff1": 0.448012,
67
+ "nauc_map_at_3_max": 0.277537,
68
+ "nauc_map_at_3_std": -0.042698,
69
+ "nauc_map_at_3_diff1": 0.400984,
70
+ "nauc_map_at_5_max": 0.292269,
71
+ "nauc_map_at_5_std": -0.039749,
72
+ "nauc_map_at_5_diff1": 0.395753,
73
+ "nauc_map_at_10_max": 0.302559,
74
+ "nauc_map_at_10_std": -0.031652,
75
+ "nauc_map_at_10_diff1": 0.396798,
76
+ "nauc_map_at_20_max": 0.304792,
77
+ "nauc_map_at_20_std": -0.028476,
78
+ "nauc_map_at_20_diff1": 0.398196,
79
+ "nauc_map_at_100_max": 0.304625,
80
+ "nauc_map_at_100_std": -0.028238,
81
+ "nauc_map_at_100_diff1": 0.399356,
82
+ "nauc_map_at_1000_max": 0.30447,
83
+ "nauc_map_at_1000_std": -0.028445,
84
+ "nauc_map_at_1000_diff1": 0.399425,
85
+ "nauc_recall_at_1_max": 0.2142,
86
+ "nauc_recall_at_1_std": -0.058618,
87
+ "nauc_recall_at_1_diff1": 0.448012,
88
+ "nauc_recall_at_3_max": 0.27315,
89
+ "nauc_recall_at_3_std": -0.026935,
90
+ "nauc_recall_at_3_diff1": 0.313983,
91
+ "nauc_recall_at_5_max": 0.29164,
92
+ "nauc_recall_at_5_std": -0.035707,
93
+ "nauc_recall_at_5_diff1": 0.287917,
94
+ "nauc_recall_at_10_max": 0.346744,
95
+ "nauc_recall_at_10_std": 0.027427,
96
+ "nauc_recall_at_10_diff1": 0.262649,
97
+ "nauc_recall_at_20_max": 0.359733,
98
+ "nauc_recall_at_20_std": 0.112853,
99
+ "nauc_recall_at_20_diff1": 0.229114,
100
+ "nauc_recall_at_100_max": 0.371893,
101
+ "nauc_recall_at_100_std": 0.195166,
102
+ "nauc_recall_at_100_diff1": 0.2245,
103
+ "nauc_recall_at_1000_max": 0.304389,
104
+ "nauc_recall_at_1000_std": 0.356072,
105
+ "nauc_recall_at_1000_diff1": 0.09331,
106
+ "nauc_precision_at_1_max": 0.320745,
107
+ "nauc_precision_at_1_std": -0.040708,
108
+ "nauc_precision_at_1_diff1": 0.507783,
109
+ "nauc_precision_at_3_max": 0.278835,
110
+ "nauc_precision_at_3_std": 0.02232,
111
+ "nauc_precision_at_3_diff1": 0.168569,
112
+ "nauc_precision_at_5_max": 0.235018,
113
+ "nauc_precision_at_5_std": 0.02516,
114
+ "nauc_precision_at_5_diff1": 0.081065,
115
+ "nauc_precision_at_10_max": 0.204528,
116
+ "nauc_precision_at_10_std": 0.058575,
117
+ "nauc_precision_at_10_diff1": 0.017179,
118
+ "nauc_precision_at_20_max": 0.181062,
119
+ "nauc_precision_at_20_std": 0.082078,
120
+ "nauc_precision_at_20_diff1": -0.015101,
121
+ "nauc_precision_at_100_max": 0.109944,
122
+ "nauc_precision_at_100_std": 0.079398,
123
+ "nauc_precision_at_100_diff1": -0.056709,
124
+ "nauc_precision_at_1000_max": 0.054119,
125
+ "nauc_precision_at_1000_std": 0.052698,
126
+ "nauc_precision_at_1000_diff1": -0.092125,
127
+ "nauc_mrr_at_1_max": 0.318271,
128
+ "nauc_mrr_at_1_std": -0.044394,
129
+ "nauc_mrr_at_1_diff1": 0.504811,
130
+ "nauc_mrr_at_3_max": 0.356277,
131
+ "nauc_mrr_at_3_std": -0.028175,
132
+ "nauc_mrr_at_3_diff1": 0.470747,
133
+ "nauc_mrr_at_5_max": 0.354792,
134
+ "nauc_mrr_at_5_std": -0.035569,
135
+ "nauc_mrr_at_5_diff1": 0.470656,
136
+ "nauc_mrr_at_10_max": 0.358809,
137
+ "nauc_mrr_at_10_std": -0.030536,
138
+ "nauc_mrr_at_10_diff1": 0.473632,
139
+ "nauc_mrr_at_20_max": 0.356713,
140
+ "nauc_mrr_at_20_std": -0.028445,
141
+ "nauc_mrr_at_20_diff1": 0.473001,
142
+ "nauc_mrr_at_100_max": 0.356391,
143
+ "nauc_mrr_at_100_std": -0.029535,
144
+ "nauc_mrr_at_100_diff1": 0.473848,
145
+ "nauc_mrr_at_1000_max": 0.356295,
146
+ "nauc_mrr_at_1000_std": -0.029684,
147
+ "nauc_mrr_at_1000_diff1": 0.473917,
148
+ "main_score": 0.65435,
149
+ "hf_subset": "default",
150
+ "languages": [
151
+ "rus-Cyrl"
152
+ ]
153
+ }
154
+ ]
155
+ },
156
+ "evaluation_time": 58.87525010108948,
157
+ "kg_co2_emissions": null
158
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuReviewsClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "f6d2c31f4dc6b88f468552750bfec05b4b41b05a",
3
+ "task_name": "RuReviewsClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.658447,
9
+ "f1": 0.654242,
10
+ "f1_weighted": 0.654252,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.686035,
14
+ "f1": 0.685663,
15
+ "f1_weighted": 0.685676
16
+ },
17
+ {
18
+ "accuracy": 0.625977,
19
+ "f1": 0.616979,
20
+ "f1_weighted": 0.617003
21
+ },
22
+ {
23
+ "accuracy": 0.66748,
24
+ "f1": 0.663276,
25
+ "f1_weighted": 0.663281
26
+ },
27
+ {
28
+ "accuracy": 0.667969,
29
+ "f1": 0.671162,
30
+ "f1_weighted": 0.671179
31
+ },
32
+ {
33
+ "accuracy": 0.674316,
34
+ "f1": 0.677712,
35
+ "f1_weighted": 0.677715
36
+ },
37
+ {
38
+ "accuracy": 0.64209,
39
+ "f1": 0.637745,
40
+ "f1_weighted": 0.637772
41
+ },
42
+ {
43
+ "accuracy": 0.650391,
44
+ "f1": 0.64232,
45
+ "f1_weighted": 0.642322
46
+ },
47
+ {
48
+ "accuracy": 0.645508,
49
+ "f1": 0.639031,
50
+ "f1_weighted": 0.639034
51
+ },
52
+ {
53
+ "accuracy": 0.641113,
54
+ "f1": 0.62666,
55
+ "f1_weighted": 0.626632
56
+ },
57
+ {
58
+ "accuracy": 0.683594,
59
+ "f1": 0.681867,
60
+ "f1_weighted": 0.681907
61
+ }
62
+ ],
63
+ "main_score": 0.658447,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 3.222393751144409,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuSTSBenchmarkSTS.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7cf24f325c6da6195df55bef3d86b5e0616f3018",
3
+ "task_name": "RuSTSBenchmarkSTS",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "pearson": 0.81294,
9
+ "spearman": 0.80262,
10
+ "cosine_pearson": 0.81294,
11
+ "cosine_spearman": 0.802618,
12
+ "manhattan_pearson": 0.797642,
13
+ "manhattan_spearman": 0.801426,
14
+ "euclidean_pearson": 0.798763,
15
+ "euclidean_spearman": 0.802622,
16
+ "main_score": 0.802618,
17
+ "hf_subset": "default",
18
+ "languages": [
19
+ "rus-Cyrl"
20
+ ]
21
+ }
22
+ ]
23
+ },
24
+ "evaluation_time": 2.0354042053222656,
25
+ "kg_co2_emissions": null
26
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuSciBenchGRNTIClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "673a610d6d3dd91a547a0d57ae1b56f37ebbf6a1",
3
+ "task_name": "RuSciBenchGRNTIClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.625342,
9
+ "f1": 0.607794,
10
+ "f1_weighted": 0.607906,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.639648,
14
+ "f1": 0.627117,
15
+ "f1_weighted": 0.62722
16
+ },
17
+ {
18
+ "accuracy": 0.625977,
19
+ "f1": 0.60801,
20
+ "f1_weighted": 0.608126
21
+ },
22
+ {
23
+ "accuracy": 0.612793,
24
+ "f1": 0.591011,
25
+ "f1_weighted": 0.591178
26
+ },
27
+ {
28
+ "accuracy": 0.637207,
29
+ "f1": 0.619847,
30
+ "f1_weighted": 0.619938
31
+ },
32
+ {
33
+ "accuracy": 0.632812,
34
+ "f1": 0.615881,
35
+ "f1_weighted": 0.615997
36
+ },
37
+ {
38
+ "accuracy": 0.612305,
39
+ "f1": 0.591041,
40
+ "f1_weighted": 0.591157
41
+ },
42
+ {
43
+ "accuracy": 0.624023,
44
+ "f1": 0.605125,
45
+ "f1_weighted": 0.605259
46
+ },
47
+ {
48
+ "accuracy": 0.630371,
49
+ "f1": 0.612698,
50
+ "f1_weighted": 0.612803
51
+ },
52
+ {
53
+ "accuracy": 0.633301,
54
+ "f1": 0.617558,
55
+ "f1_weighted": 0.617691
56
+ },
57
+ {
58
+ "accuracy": 0.60498,
59
+ "f1": 0.589651,
60
+ "f1_weighted": 0.589689
61
+ }
62
+ ],
63
+ "main_score": 0.625342,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 12.012431621551514,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuSciBenchGRNTIClusteringP2P.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "673a610d6d3dd91a547a0d57ae1b56f37ebbf6a1",
3
+ "task_name": "RuSciBenchGRNTIClusteringP2P",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.582825,
11
+ 0.579362,
12
+ 0.585397,
13
+ 0.598136,
14
+ 0.5796,
15
+ 0.586166,
16
+ 0.579628,
17
+ 0.589436,
18
+ 0.588864,
19
+ 0.587165
20
+ ]
21
+ },
22
+ "v_measure": 0.585658,
23
+ "v_measure_std": 0.00551,
24
+ "main_score": 0.585658,
25
+ "hf_subset": "default",
26
+ "languages": [
27
+ "rus-Cyrl"
28
+ ]
29
+ }
30
+ ]
31
+ },
32
+ "evaluation_time": 7.271762371063232,
33
+ "kg_co2_emissions": null
34
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuSciBenchOECDClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "26c88e99dcaba32bb45d0e1bfc21902337f6d471",
3
+ "task_name": "RuSciBenchOECDClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.491162,
9
+ "f1": 0.462461,
10
+ "f1_weighted": 0.462363,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.494141,
14
+ "f1": 0.463986,
15
+ "f1_weighted": 0.463906
16
+ },
17
+ {
18
+ "accuracy": 0.487793,
19
+ "f1": 0.459439,
20
+ "f1_weighted": 0.459315
21
+ },
22
+ {
23
+ "accuracy": 0.487793,
24
+ "f1": 0.462431,
25
+ "f1_weighted": 0.462389
26
+ },
27
+ {
28
+ "accuracy": 0.480469,
29
+ "f1": 0.4402,
30
+ "f1_weighted": 0.440081
31
+ },
32
+ {
33
+ "accuracy": 0.481934,
34
+ "f1": 0.4513,
35
+ "f1_weighted": 0.451233
36
+ },
37
+ {
38
+ "accuracy": 0.496582,
39
+ "f1": 0.473999,
40
+ "f1_weighted": 0.47387
41
+ },
42
+ {
43
+ "accuracy": 0.466309,
44
+ "f1": 0.430939,
45
+ "f1_weighted": 0.430946
46
+ },
47
+ {
48
+ "accuracy": 0.523926,
49
+ "f1": 0.511153,
50
+ "f1_weighted": 0.511123
51
+ },
52
+ {
53
+ "accuracy": 0.48877,
54
+ "f1": 0.450335,
55
+ "f1_weighted": 0.45016
56
+ },
57
+ {
58
+ "accuracy": 0.503906,
59
+ "f1": 0.480823,
60
+ "f1_weighted": 0.480604
61
+ }
62
+ ],
63
+ "main_score": 0.491162,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 12.004972219467163,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/RuSciBenchOECDClusteringP2P.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "26c88e99dcaba32bb45d0e1bfc21902337f6d471",
3
+ "task_name": "RuSciBenchOECDClusteringP2P",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "v_measures": {
9
+ "Level 0": [
10
+ 0.491972,
11
+ 0.510493,
12
+ 0.512482,
13
+ 0.509155,
14
+ 0.508742,
15
+ 0.505142,
16
+ 0.511958,
17
+ 0.513889,
18
+ 0.505867,
19
+ 0.502838
20
+ ]
21
+ },
22
+ "v_measure": 0.507254,
23
+ "v_measure_std": 0.006077,
24
+ "main_score": 0.507254,
25
+ "hf_subset": "default",
26
+ "languages": [
27
+ "rus-Cyrl"
28
+ ]
29
+ }
30
+ ]
31
+ },
32
+ "evaluation_time": 7.268473386764526,
33
+ "kg_co2_emissions": null
34
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/STS22.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3",
3
+ "task_name": "STS22",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "pearson": 0.548849,
9
+ "spearman": 0.610626,
10
+ "cosine_pearson": 0.548849,
11
+ "cosine_spearman": 0.610654,
12
+ "manhattan_pearson": 0.564739,
13
+ "manhattan_spearman": 0.610472,
14
+ "euclidean_pearson": 0.562749,
15
+ "euclidean_spearman": 0.610598,
16
+ "main_score": 0.610654,
17
+ "hf_subset": "ru",
18
+ "languages": [
19
+ "rus-Cyrl"
20
+ ]
21
+ }
22
+ ]
23
+ },
24
+ "evaluation_time": 3.1988155841827393,
25
+ "kg_co2_emissions": null
26
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/SensitiveTopicsClassification.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "416b34a802308eac30e4192afc0ff99bb8dcc7f2",
3
+ "task_name": "SensitiveTopicsClassification",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "test": [
7
+ {
8
+ "accuracy": 0.373438,
9
+ "f1": 0.49652,
10
+ "lrap": 0.598327,
11
+ "scores_per_experiment": [
12
+ {
13
+ "accuracy": 0.381348,
14
+ "f1": 0.483055,
15
+ "lrap": 0.586616
16
+ },
17
+ {
18
+ "accuracy": 0.366699,
19
+ "f1": 0.486796,
20
+ "lrap": 0.588447
21
+ },
22
+ {
23
+ "accuracy": 0.362305,
24
+ "f1": 0.479753,
25
+ "lrap": 0.587836
26
+ },
27
+ {
28
+ "accuracy": 0.365234,
29
+ "f1": 0.500353,
30
+ "lrap": 0.582248
31
+ },
32
+ {
33
+ "accuracy": 0.376465,
34
+ "f1": 0.512828,
35
+ "lrap": 0.609836
36
+ },
37
+ {
38
+ "accuracy": 0.372559,
39
+ "f1": 0.488137,
40
+ "lrap": 0.594347
41
+ },
42
+ {
43
+ "accuracy": 0.386719,
44
+ "f1": 0.503806,
45
+ "lrap": 0.612901
46
+ },
47
+ {
48
+ "accuracy": 0.386719,
49
+ "f1": 0.519459,
50
+ "lrap": 0.605516
51
+ },
52
+ {
53
+ "accuracy": 0.380859,
54
+ "f1": 0.501056,
55
+ "lrap": 0.620321
56
+ },
57
+ {
58
+ "accuracy": 0.355469,
59
+ "f1": 0.489961,
60
+ "lrap": 0.595201
61
+ }
62
+ ],
63
+ "main_score": 0.373438,
64
+ "hf_subset": "default",
65
+ "languages": [
66
+ "rus-Cyrl"
67
+ ]
68
+ }
69
+ ]
70
+ },
71
+ "evaluation_time": 4.6399986743927,
72
+ "kg_co2_emissions": null
73
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/TERRa.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7b58f24536063837d644aab9a023c62199b2a612",
3
+ "task_name": "TERRa",
4
+ "mteb_version": "1.36.16",
5
+ "scores": {
6
+ "dev": [
7
+ {
8
+ "similarity_accuracy": 0.605863,
9
+ "similarity_accuracy_threshold": 0.633979,
10
+ "similarity_f1": 0.678492,
11
+ "similarity_f1_threshold": 0.358098,
12
+ "similarity_precision": 0.513423,
13
+ "similarity_recall": 1.0,
14
+ "similarity_ap": 0.603584,
15
+ "cosine_accuracy": 0.605863,
16
+ "cosine_accuracy_threshold": 0.633979,
17
+ "cosine_f1": 0.678492,
18
+ "cosine_f1_threshold": 0.358098,
19
+ "cosine_precision": 0.513423,
20
+ "cosine_recall": 1.0,
21
+ "cosine_ap": 0.603584,
22
+ "manhattan_accuracy": 0.602606,
23
+ "manhattan_accuracy_threshold": 12.64444,
24
+ "manhattan_f1": 0.67619,
25
+ "manhattan_f1_threshold": 14.47312,
26
+ "manhattan_precision": 0.531835,
27
+ "manhattan_recall": 0.928105,
28
+ "manhattan_ap": 0.603085,
29
+ "euclidean_accuracy": 0.605863,
30
+ "euclidean_accuracy_threshold": 0.855592,
31
+ "euclidean_f1": 0.678492,
32
+ "euclidean_f1_threshold": 1.133051,
33
+ "euclidean_precision": 0.513423,
34
+ "euclidean_recall": 1.0,
35
+ "euclidean_ap": 0.603584,
36
+ "dot_accuracy": 0.605863,
37
+ "dot_accuracy_threshold": 0.633979,
38
+ "dot_f1": 0.678492,
39
+ "dot_f1_threshold": 0.358098,
40
+ "dot_precision": 0.513423,
41
+ "dot_recall": 1.0,
42
+ "dot_ap": 0.603584,
43
+ "max_accuracy": 0.605863,
44
+ "max_f1": 0.678492,
45
+ "max_precision": 0.531835,
46
+ "max_recall": 1.0,
47
+ "max_ap": 0.603584,
48
+ "main_score": 0.603584,
49
+ "hf_subset": "default",
50
+ "languages": [
51
+ "rus-Cyrl"
52
+ ]
53
+ }
54
+ ]
55
+ },
56
+ "evaluation_time": 0.6962718963623047,
57
+ "kg_co2_emissions": null
58
+ }
results/sergeyzh__rubert-mini-frida/19b279b78afd945b5ccae78f63e284909814adc2/model_meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"name": "sergeyzh/rubert-mini-frida", "revision": "19b279b78afd945b5ccae78f63e284909814adc2", "release_date": "2025-03-02", "languages": ["rus_Cyrl"], "n_parameters": 32300000, "memory_usage_mb": 123.0, "max_tokens": 2048.0, "embed_dim": 312, "license": "mit", "open_weights": true, "public_training_code": null, "public_training_data": null, "framework": ["Sentence Transformers", "PyTorch"], "reference": "https://huggingface.co/sergeyzh/rubert-mini-frida", "similarity_fn_name": "cosine", "use_instructions": true, "training_datasets": {}, "adapted_from": "sergeyzh/rubert-mini-sts", "superseded_by": null, "is_cross_encoder": null, "modalities": ["text"], "loader": null}