dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
demelin/moral_stories
gen-norm$actions+context-norm_distance
validation
1,000
norm
19
81
46.156
46
9.21513
{ "bin_edges": [ 19, 26, 33, 40, 47, 54, 61, 68, 75, 81 ], "hist": [ 16, 62, 152, 281, 275, 157, 48, 7, 2 ] }
false
demelin/moral_stories
gen-norm$actions+context-norm_distance
validation
1,000
situation
20
166
85.536
84
25.66229
{ "bin_edges": [ 20, 35, 50, 65, 80, 95, 110, 125, 140, 155, 166 ], "hist": [ 18, 47, 143, 226, 225, 159, 109, 46, 19, 8 ] }
false
demelin/moral_stories
gen-norm$actions-norm_distance
train
10,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 10000 ] }
false
demelin/moral_stories
gen-norm$actions-norm_distance
train
10,000
immoral_action
17
179
80.8544
78
22.81142
{ "bin_edges": [ 17, 34, 51, 68, 85, 102, 119, 136, 153, 170, 179 ], "hist": [ 67, 609, 2350, 3045, 2162, 1089, 481, 162, 33, 2 ] }
false
demelin/moral_stories
gen-norm$actions-norm_distance
train
10,000
moral_action
18
177
80.656
78
22.83864
{ "bin_edges": [ 18, 34, 50, 66, 82, 98, 114, 130, 146, 162, 177 ], "hist": [ 53, 544, 2106, 2945, 2197, 1252, 569, 247, 75, 12 ] }
false
demelin/moral_stories
gen-norm$actions-norm_distance
train
10,000
norm
15
82
43.8687
44
9.75162
{ "bin_edges": [ 15, 22, 29, 36, 43, 50, 57, 64, 71, 78, 82 ], "hist": [ 49, 532, 1495, 2389, 2648, 1887, 781, 187, 28, 4 ] }
false
demelin/moral_stories
gen-norm$actions-norm_distance
validation
1,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 1000 ] }
false
demelin/moral_stories
gen-norm$actions-norm_distance
validation
1,000
immoral_action
24
151
80.034
78
23.18967
{ "bin_edges": [ 24, 37, 50, 63, 76, 89, 102, 115, 128, 141, 151 ], "hist": [ 19, 50, 169, 216, 204, 160, 96, 50, 29, 7 ] }
false
demelin/moral_stories
gen-norm$actions-norm_distance
validation
1,000
moral_action
25
163
79.577
77
22.93587
{ "bin_edges": [ 25, 39, 53, 67, 81, 95, 109, 123, 137, 151, 163 ], "hist": [ 20, 77, 201, 258, 209, 130, 52, 33, 15, 5 ] }
false
demelin/moral_stories
gen-norm$actions-norm_distance
validation
1,000
norm
19
81
46.156
46
9.21513
{ "bin_edges": [ 19, 26, 33, 40, 47, 54, 61, 68, 75, 81 ], "hist": [ 16, 62, 152, 281, 275, 157, 48, 7, 2 ] }
false
nbroad/mediasum
mediasum
validation
10,000
date
8
10
9.4599
10
0.69285
{ "bin_edges": [ 8, 9, 10, 10 ], "hist": [ 1158, 3085, 5757 ] }
false
nbroad/mediasum
mediasum
validation
10,000
id
6
10
9.637
10
0.57781
{ "bin_edges": [ 6, 7, 8, 9, 10, 10 ], "hist": [ 2, 34, 399, 2722, 6843 ] }
false
nbroad/mediasum
mediasum
validation
10,000
program
0
55
15.8578
14
6.90851
{ "bin_edges": [ 0, 6, 12, 18, 24, 30, 36, 42, 48, 54, 55 ], "hist": [ 193, 1534, 4983, 2097, 693, 329, 117, 0, 33, 21 ] }
false
nbroad/mediasum
mediasum
validation
10,000
summary
19
1,100
91.0429
60
78.65461
{ "bin_edges": [ 19, 128, 237, 346, 455, 564, 673, 782, 891, 1000, 1100 ], "hist": [ 8005, 1363, 475, 111, 36, 4, 3, 1, 1, 1 ] }
false
nbroad/mediasum
mediasum
validation
10,000
title
0
89
4.883
0
15.29849
{ "bin_edges": [ 0, 9, 18, 27, 36, 45, 54, 63, 72, 81, 89 ], "hist": [ 9010, 5, 19, 90, 261, 309, 134, 103, 51, 18 ] }
false
nbroad/mediasum
mediasum
validation
10,000
url
51
121
60.4008
58
8.99148
{ "bin_edges": [ 51, 59, 67, 75, 83, 91, 99, 107, 115, 121 ], "hist": [ 6948, 2510, 34, 63, 108, 114, 103, 72, 48 ] }
false
nbroad/mediasum
mediasum
test
10,000
date
8
10
9.4649
10
0.68689
{ "bin_edges": [ 8, 9, 10, 10 ], "hist": [ 1115, 3121, 5764 ] }
false
nbroad/mediasum
mediasum
test
10,000
id
6
10
9.6269
10
0.58441
{ "bin_edges": [ 6, 7, 8, 9, 10, 10 ], "hist": [ 2, 36, 418, 2779, 6765 ] }
false
nbroad/mediasum
mediasum
test
10,000
program
0
55
15.8148
14
6.86293
{ "bin_edges": [ 0, 6, 12, 18, 24, 30, 36, 42, 48, 54, 55 ], "hist": [ 214, 1467, 5051, 2075, 704, 329, 106, 0, 37, 17 ] }
false
nbroad/mediasum
mediasum
test
10,000
summary
19
1,105
91.9013
61
78.46786
{ "bin_edges": [ 19, 128, 237, 346, 455, 564, 673, 782, 891, 1000, 1105 ], "hist": [ 7981, 1373, 476, 126, 32, 8, 3, 0, 0, 1 ] }
false
nbroad/mediasum
mediasum
test
10,000
title
0
98
5.1157
0
15.41853
{ "bin_edges": [ 0, 10, 20, 30, 40, 50, 60, 70, 80, 90, 98 ], "hist": [ 8940, 5, 43, 197, 407, 203, 132, 55, 17, 1 ] }
false
nbroad/mediasum
mediasum
test
10,000
url
41
127
60.3947
58
8.89655
{ "bin_edges": [ 41, 50, 59, 68, 77, 86, 95, 104, 113, 122, 127 ], "hist": [ 1, 6961, 2488, 37, 110, 151, 101, 97, 53, 1 ] }
false
nbroad/mediasum
mediasum
train
443,596
date
8
10
9.46865
10
0.68964
{ "bin_edges": [ 8, 9, 10, 10 ], "hist": [ 50257, 135191, 258148 ] }
false
nbroad/mediasum
mediasum
train
443,596
id
5
10
9.62974
10
0.58268
{ "bin_edges": [ 5, 6, 7, 8, 9, 10, 10 ], "hist": [ 18, 176, 1729, 17164, 123936, 300573 ] }
false
nbroad/mediasum
mediasum
train
443,596
program
0
55
15.88469
14
6.86631
{ "bin_edges": [ 0, 6, 12, 18, 24, 30, 36, 42, 48, 54, 55 ], "hist": [ 8379, 65655, 222675, 92912, 32455, 13982, 5060, 2, 1741, 735 ] }
false
nbroad/mediasum
mediasum
train
443,596
summary
15
1,202
92.31527
60
79.68302
{ "bin_edges": [ 15, 134, 253, 372, 491, 610, 729, 848, 967, 1086, 1202 ], "hist": [ 358434, 59745, 19824, 4360, 872, 229, 64, 38, 21, 9 ] }
false
nbroad/mediasum
mediasum
train
443,596
title
0
100
5.29231
0
15.87787
{ "bin_edges": [ 0, 11, 22, 33, 44, 55, 66, 77, 88, 99, 100 ], "hist": [ 396230, 292, 2733, 12851, 17791, 7286, 4694, 1610, 107, 2 ] }
false
nbroad/mediasum
mediasum
train
443,596
url
41
131
60.53698
58
9.31584
{ "bin_edges": [ 41, 51, 61, 71, 81, 91, 101, 111, 121, 131, 131 ], "hist": [ 15, 369177, 49699, 2555, 6513, 6161, 5662, 2885, 928, 1 ] }
false
Muennighoff/xwinograd
fr
test
83
option1
4
25
7.79518
6
3.90983
{ "bin_edges": [ 4, 7, 10, 13, 16, 19, 22, 25, 25 ], "hist": [ 42, 25, 6, 4, 4, 1, 0, 1 ] }
false
Muennighoff/xwinograd
fr
test
83
option2
4
25
7.93976
6
4.37941
{ "bin_edges": [ 4, 7, 10, 13, 16, 19, 22, 25, 25 ], "hist": [ 46, 17, 10, 4, 2, 3, 0, 1 ] }
false
Muennighoff/xwinograd
fr
test
83
sentence
36
165
88.15663
86
31.40976
{ "bin_edges": [ 36, 49, 62, 75, 88, 101, 114, 127, 140, 153, 165 ], "hist": [ 5, 12, 17, 9, 18, 6, 8, 0, 3, 5 ] }
false
Muennighoff/xwinograd
ru
test
315
option1
3
24
6.93968
6
3.79568
{ "bin_edges": [ 3, 6, 9, 12, 15, 18, 21, 24, 24 ], "hist": [ 146, 99, 31, 19, 13, 4, 2, 1 ] }
false
Muennighoff/xwinograd
ru
test
315
option2
3
29
6.28571
5
3.59683
{ "bin_edges": [ 3, 6, 9, 12, 15, 18, 21, 24, 27, 29 ], "hist": [ 163, 109, 18, 15, 3, 2, 2, 2, 1 ] }
false
Muennighoff/xwinograd
ru
test
315
sentence
33
291
105.35238
84
58.19077
{ "bin_edges": [ 33, 59, 85, 111, 137, 163, 189, 215, 241, 267, 291 ], "hist": [ 45, 115, 58, 32, 20, 3, 10, 19, 8, 5 ] }
false
Muennighoff/xwinograd
zh
test
504
option1
1
17
2.93651
3
1.6039
{ "bin_edges": [ 1, 3, 5, 7, 9, 11, 13, 15, 17, 17 ], "hist": [ 248, 209, 28, 13, 2, 2, 1, 0, 1 ] }
false
Muennighoff/xwinograd
zh
test
504
option2
1
12
2.89683
3
1.41747
{ "bin_edges": [ 1, 3, 5, 7, 9, 11, 12 ], "hist": [ 250, 203, 38, 4, 8, 1 ] }
false
Muennighoff/xwinograd
zh
test
504
sentence
16
147
59.39484
56
23.26246
{ "bin_edges": [ 16, 30, 44, 58, 72, 86, 100, 114, 128, 142, 147 ], "hist": [ 37, 98, 130, 102, 73, 40, 10, 7, 5, 2 ] }
false
Muennighoff/xwinograd
pt
test
263
option1
3
26
7.62738
7
3.67656
{ "bin_edges": [ 3, 6, 9, 12, 15, 18, 21, 24, 26 ], "hist": [ 76, 122, 33, 13, 15, 1, 0, 3 ] }
false
Muennighoff/xwinograd
pt
test
263
option2
2
26
7.36122
7
3.37519
{ "bin_edges": [ 2, 5, 8, 11, 14, 17, 20, 23, 26, 26 ], "hist": [ 29, 150, 53, 9, 12, 8, 1, 0, 1 ] }
false
Muennighoff/xwinograd
pt
test
263
sentence
35
169
92.78327
87
30.3792
{ "bin_edges": [ 35, 49, 63, 77, 91, 105, 119, 133, 147, 161, 169 ], "hist": [ 8, 30, 56, 52, 31, 29, 22, 18, 15, 2 ] }
false
Muennighoff/xwinograd
en
test
2,325
option1
3
37
9.16258
8
5.08527
{ "bin_edges": [ 3, 7, 11, 15, 19, 23, 27, 31, 35, 37 ], "hist": [ 846, 699, 474, 172, 88, 28, 11, 5, 2 ] }
false
Muennighoff/xwinograd
en
test
2,325
option2
2
39
9.39613
8
5.38421
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 38, 39 ], "hist": [ 699, 629, 571, 236, 100, 58, 19, 9, 2, 2 ] }
false
Muennighoff/xwinograd
en
test
2,325
sentence
26
476
85.74323
77
40.45608
{ "bin_edges": [ 26, 72, 118, 164, 210, 256, 302, 348, 394, 440, 476 ], "hist": [ 1042, 875, 328, 37, 32, 9, 0, 0, 0, 2 ] }
false
Muennighoff/xwinograd
jp
test
959
option1
1
21
4.40563
4
2.79375
{ "bin_edges": [ 1, 4, 7, 10, 13, 16, 19, 21 ], "hist": [ 458, 330, 110, 48, 7, 5, 1 ] }
false
Muennighoff/xwinograd
jp
test
959
option2
1
21
4.47758
4
2.80346
{ "bin_edges": [ 1, 4, 7, 10, 13, 16, 19, 21 ], "hist": [ 447, 335, 103, 60, 11, 2, 1 ] }
false
Muennighoff/xwinograd
jp
test
959
sentence
16
96
39.78102
37
13.58127
{ "bin_edges": [ 16, 25, 34, 43, 52, 61, 70, 79, 88, 96 ], "hist": [ 96, 279, 246, 150, 100, 58, 23, 5, 2 ] }
false
pyronear/openfire
default
train
7,143
image_url
27
840
107.01176
97
50.90907
{ "bin_edges": [ 27, 109, 191, 273, 355, 437, 519, 601, 683, 765, 840 ], "hist": [ 4408, 2427, 207, 65, 18, 11, 1, 0, 5, 1 ] }
false
pyronear/openfire
default
validation
792
image_url
35
447
107.62247
97
50.40504
{ "bin_edges": [ 35, 77, 119, 161, 203, 245, 287, 329, 371, 413, 447 ], "hist": [ 217, 342, 156, 43, 12, 9, 5, 5, 2, 1 ] }
false
linxinyuan/imdb
plain_text
test
25,000
text
32
12,988
1,293.7924
962
975.90776
{ "bin_edges": [ 32, 1328, 2624, 3920, 5216, 6512, 7808, 9104, 10400, 11696, 12988 ], "hist": [ 17044, 5634, 1550, 544, 221, 2, 1, 1, 0, 3 ] }
false
linxinyuan/amazon
default
train
40,001
text
1
15,713
349.25017
171
542.56868
{ "bin_edges": [ 1, 1573, 3145, 4717, 6289, 7861, 9433, 11005, 12577, 14149, 15713 ], "hist": [ 38714, 1039, 188, 38, 17, 2, 0, 1, 1, 1 ] }
false
linxinyuan/amazon
default
test
5,000
text
1
19,534
441.4266
229
691.84944
{ "bin_edges": [ 1, 1955, 3909, 5863, 7817, 9771, 11725, 13679, 15633, 17587, 19534 ], "hist": [ 4845, 131, 18, 2, 2, 0, 1, 0, 0, 1 ] }
false
tner/tweetner7
tweetner7
test_2020
576
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 576 ] }
false
tner/tweetner7
tweetner7
test_2020
576
id
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 576 ] }
false
tner/tweetner7
tweetner7
test_2021
2,807
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 2807 ] }
false
tner/tweetner7
tweetner7
test_2021
2,807
id
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 2807 ] }
false
tner/tweetner7
tweetner7
validation_2020
576
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 576 ] }
false
tner/tweetner7
tweetner7
validation_2020
576
id
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 576 ] }
false
tner/tweetner7
tweetner7
validation_2021
310
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 310 ] }
false
tner/tweetner7
tweetner7
validation_2021
310
id
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 310 ] }
false
tner/tweetner7
tweetner7
train_2020
4,616
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 4616 ] }
false
tner/tweetner7
tweetner7
train_2020
4,616
id
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 4616 ] }
false
tner/tweetner7
tweetner7
train_2021
2,495
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 2495 ] }
false
tner/tweetner7
tweetner7
train_2021
2,495
id
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 2495 ] }
false
tner/tweetner7
tweetner7
train_all
7,111
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 7111 ] }
false
tner/tweetner7
tweetner7
train_all
7,111
id
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 7111 ] }
false
tner/tweetner7
tweetner7
validation_random
576
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 576 ] }
false
tner/tweetner7
tweetner7
validation_random
576
id
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 576 ] }
false
tner/tweetner7
tweetner7
train_random
4,616
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 4616 ] }
false
tner/tweetner7
tweetner7
train_random
4,616
id
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 4616 ] }
false
linxinyuan/cola
default
train
8,551
text
6
231
40.69933
37
19.14304
{ "bin_edges": [ 6, 29, 52, 75, 98, 121, 144, 167, 190, 213, 231 ], "hist": [ 2260, 4512, 1262, 380, 102, 26, 6, 1, 1, 1 ] }
false
linxinyuan/cola
default
test
527
text
9
134
39.81025
35
18.32092
{ "bin_edges": [ 9, 22, 35, 48, 61, 74, 87, 100, 113, 126, 134 ], "hist": [ 57, 182, 155, 72, 27, 24, 5, 2, 1, 2 ] }
false
relbert/analogy_questions
bats
test
1,799
prefix
63
73
68.12285
68
2.49157
{ "bin_edges": [ 63, 65, 67, 69, 71, 73, 73 ], "hist": [ 135, 405, 495, 540, 134, 90 ] }
false
relbert/analogy_questions
bats
validation
199
prefix
63
73
68.10553
68
2.48915
{ "bin_edges": [ 63, 65, 67, 69, 71, 73, 73 ], "hist": [ 15, 45, 55, 60, 14, 10 ] }
false
Muennighoff/mbpp
full
test
974
code
30
1,331
181.06776
145.5
127.48223
{ "bin_edges": [ 30, 161, 292, 423, 554, 685, 816, 947, 1078, 1209, 1331 ], "hist": [ 557, 270, 101, 28, 11, 3, 3, 0, 0, 1 ] }
false
Muennighoff/mbpp
full
test
974
text
37
249
78.61602
77
21.63341
{ "bin_edges": [ 37, 59, 81, 103, 125, 147, 169, 191, 213, 235, 249 ], "hist": [ 154, 409, 309, 65, 29, 5, 1, 1, 0, 1 ] }
false
Muennighoff/mbpp
sanitized
test
427
code
36
906
153.50351
122
106.30447
{ "bin_edges": [ 36, 124, 212, 300, 388, 476, 564, 652, 740, 828, 906 ], "hist": [ 215, 141, 35, 16, 12, 3, 4, 0, 0, 1 ] }
false
Muennighoff/mbpp
sanitized
test
427
prompt
39
410
92.63232
82
40.30282
{ "bin_edges": [ 39, 77, 115, 153, 191, 229, 267, 305, 343, 381, 410 ], "hist": [ 171, 169, 51, 23, 8, 4, 0, 0, 0, 1 ] }
false
relbert/analogy_questions
semeval2012_relational_similarity
validation
79
prefix
3
4
3.07595
3
0.26661
{ "bin_edges": [ 3, 4, 4 ], "hist": [ 73, 6 ] }
false
relbert/analogy_questions
t_rex_relational_similarity
validation
496
prefix
21
52
32.10685
30
5.48263
{ "bin_edges": [ 21, 25, 29, 33, 37, 41, 45, 49, 52 ], "hist": [ 13, 110, 195, 83, 42, 23, 29, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-ayr_Latn
train
6,000
sentence_ayr_Latn
5
1,322
131.08383
117
73.68602
{ "bin_edges": [ 5, 137, 269, 401, 533, 665, 797, 929, 1061, 1193, 1322 ], "hist": [ 3648, 2049, 280, 15, 6, 1, 0, 0, 0, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-ayr_Latn
train
6,000
sentence_eng_Latn
4
1,506
127.54433
116
72.28255
{ "bin_edges": [ 4, 155, 306, 457, 608, 759, 910, 1061, 1212, 1363, 1506 ], "hist": [ 4247, 1683, 56, 7, 6, 0, 0, 0, 0, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-ayr_Latn
valid
1,309
sentence_ayr_Latn
8
1,066
133.05577
118
80.91565
{ "bin_edges": [ 8, 114, 220, 326, 432, 538, 644, 750, 856, 962, 1066 ], "hist": [ 623, 542, 113, 23, 2, 3, 2, 0, 0, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-ayr_Latn
valid
1,309
sentence_eng_Latn
3
1,168
128.49274
111
82.75026
{ "bin_edges": [ 3, 120, 237, 354, 471, 588, 705, 822, 939, 1056, 1168 ], "hist": [ 717, 496, 77, 12, 4, 0, 1, 0, 1, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-ayr_Latn
test
1,500
sentence_ayr_Latn
5
452
129.48467
118
67.2867
{ "bin_edges": [ 5, 50, 95, 140, 185, 230, 275, 320, 365, 410, 452 ], "hist": [ 105, 415, 454, 250, 144, 76, 32, 17, 5, 2 ] }
false
breakend/nllb-multi-domain
eng_Latn-ayr_Latn
test
1,500
sentence_eng_Latn
5
425
127.184
115
64.18829
{ "bin_edges": [ 5, 48, 91, 134, 177, 220, 263, 306, 349, 392, 425 ], "hist": [ 95, 382, 439, 290, 161, 74, 46, 4, 4, 5 ] }
false
breakend/nllb-multi-domain
eng_Latn-bho_Deva
train
6,000
sentence_bho_Deva
2
1,452
125.575
115
71.05374
{ "bin_edges": [ 2, 148, 294, 440, 586, 732, 878, 1024, 1170, 1316, 1452 ], "hist": [ 4115, 1760, 109, 9, 5, 1, 0, 0, 0, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-bho_Deva
train
6,000
sentence_eng_Latn
4
1,506
127.56733
116
72.33224
{ "bin_edges": [ 4, 155, 306, 457, 608, 759, 910, 1061, 1212, 1363, 1506 ], "hist": [ 4245, 1685, 56, 7, 6, 0, 0, 0, 0, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-bho_Deva
valid
1,309
sentence_bho_Deva
3
1,099
123.54011
108
80.68613
{ "bin_edges": [ 3, 113, 223, 333, 443, 553, 663, 773, 883, 993, 1099 ], "hist": [ 697, 498, 97, 10, 1, 3, 0, 2, 0, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-bho_Deva
valid
1,309
sentence_eng_Latn
3
1,168
128.54775
111
82.78879
{ "bin_edges": [ 3, 120, 237, 354, 471, 588, 705, 822, 939, 1056, 1168 ], "hist": [ 717, 495, 78, 12, 4, 0, 1, 0, 1, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-bho_Deva
test
1,500
sentence_bho_Deva
7
504
122.80933
113
62.01875
{ "bin_edges": [ 7, 57, 107, 157, 207, 257, 307, 357, 407, 457, 504 ], "hist": [ 179, 499, 459, 211, 104, 35, 7, 3, 2, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-bho_Deva
test
1,500
sentence_eng_Latn
5
425
127.02
115
64.01686
{ "bin_edges": [ 5, 48, 91, 134, 177, 220, 263, 306, 349, 392, 425 ], "hist": [ 95, 382, 440, 291, 160, 74, 45, 4, 4, 5 ] }
false
breakend/nllb-multi-domain
eng_Latn-dyu_Latn
train
6,000
sentence_dyu_Latn
5
1,513
123.49817
111
71.67745
{ "bin_edges": [ 5, 156, 307, 458, 609, 760, 911, 1062, 1213, 1364, 1513 ], "hist": [ 4444, 1460, 80, 10, 5, 0, 0, 0, 0, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-dyu_Latn
train
6,000
sentence_eng_Latn
4
1,506
127.54433
116
72.28255
{ "bin_edges": [ 4, 155, 306, 457, 608, 759, 910, 1061, 1212, 1363, 1506 ], "hist": [ 4247, 1683, 56, 7, 6, 0, 0, 0, 0, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-dyu_Latn
valid
1,309
sentence_dyu_Latn
3
877
123.31016
108
76.18036
{ "bin_edges": [ 3, 91, 179, 267, 355, 443, 531, 619, 707, 795, 877 ], "hist": [ 498, 593, 168, 35, 6, 5, 2, 0, 1, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-dyu_Latn
valid
1,309
sentence_eng_Latn
3
1,168
128.49274
111
82.75026
{ "bin_edges": [ 3, 120, 237, 354, 471, 588, 705, 822, 939, 1056, 1168 ], "hist": [ 717, 496, 77, 12, 4, 0, 1, 0, 1, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-dyu_Latn
test
1,500
sentence_dyu_Latn
7
455
122.94333
110
63.88825
{ "bin_edges": [ 7, 52, 97, 142, 187, 232, 277, 322, 367, 412, 455 ], "hist": [ 147, 438, 432, 255, 135, 58, 22, 8, 1, 4 ] }
false
breakend/nllb-multi-domain
eng_Latn-dyu_Latn
test
1,500
sentence_eng_Latn
5
425
127.184
115
64.18829
{ "bin_edges": [ 5, 48, 91, 134, 177, 220, 263, 306, 349, 392, 425 ], "hist": [ 95, 382, 439, 290, 161, 74, 46, 4, 4, 5 ] }
false
breakend/nllb-multi-domain
eng_Latn-fur_Latn
train
6,000
sentence_eng_Latn
4
1,506
127.54433
116
72.28255
{ "bin_edges": [ 4, 155, 306, 457, 608, 759, 910, 1061, 1212, 1363, 1506 ], "hist": [ 4247, 1683, 56, 7, 6, 0, 0, 0, 0, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-fur_Latn
train
6,000
sentence_fur_Latn
7
1,617
134.05933
121
76.57275
{ "bin_edges": [ 7, 169, 331, 493, 655, 817, 979, 1141, 1303, 1465, 1617 ], "hist": [ 4383, 1542, 63, 7, 4, 0, 0, 0, 0, 1 ] }
false
breakend/nllb-multi-domain
eng_Latn-fur_Latn
valid
1,309
sentence_eng_Latn
3
1,168
128.49274
111
82.75026
{ "bin_edges": [ 3, 120, 237, 354, 471, 588, 705, 822, 939, 1056, 1168 ], "hist": [ 717, 496, 77, 12, 4, 0, 1, 0, 1, 1 ] }
false