dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
hoskinson-center/proof-pile
default
test
46,251
meta
20
165
63.42217
66
7.61876
{ "bin_edges": [ 20, 35, 50, 65, 80, 95, 110, 125, 140, 155, 165 ], "hist": [ 174, 3573, 5788, 36194, 333, 136, 40, 6, 4, 3 ] }
true
hoskinson-center/proof-pile
default
test
46,251
text
94
1,279,388
12,334.74699
1,646
33,391.22509
{ "bin_edges": [ 94, 128024, 255954, 383884, 511814, 639744, 767674, 895604, 1023534, 1151464, 1279388 ], "hist": [ 45494, 669, 65, 16, 3, 0, 1, 1, 0, 2 ] }
true
UCL-DARK/ludwig
0-shot
validation
118
id
1
3
2.08475
2
0.48174
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 19 ] }
false
UCL-DARK/ludwig
0-shot
validation
118
response
6
96
32.18644
28
18.40473
{ "bin_edges": [ 6, 16, 26, 36, 46, 56, 66, 76, 86, 96, 96 ], "hist": [ 20, 30, 30, 17, 9, 3, 3, 4, 1, 1 ] }
false
UCL-DARK/ludwig
0-shot
validation
118
utterance
13
107
33.79661
30
15.9307
{ "bin_edges": [ 13, 23, 33, 43, 53, 63, 73, 83, 93, 103, 107 ], "hist": [ 28, 38, 24, 12, 9, 6, 0, 0, 0, 1 ] }
false
UCL-DARK/ludwig
0-shot
test
600
id
1
3
2.82
3
0.42178
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 501 ] }
false
UCL-DARK/ludwig
0-shot
test
600
response
4
142
32.82667
29
19.27509
{ "bin_edges": [ 4, 18, 32, 46, 60, 74, 88, 102, 116, 130, 142 ], "hist": [ 120, 222, 139, 66, 29, 14, 4, 3, 1, 2 ] }
false
UCL-DARK/ludwig
0-shot
test
600
utterance
4
150
33.84833
29
19.43102
{ "bin_edges": [ 4, 19, 34, 49, 64, 79, 94, 109, 124, 139, 150 ], "hist": [ 99, 264, 147, 47, 26, 6, 4, 2, 3, 2 ] }
false
UCL-DARK/ludwig
1-shot
validation
118
id
1
3
2.08475
2
0.48174
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 19 ] }
false
UCL-DARK/ludwig
1-shot
validation
118
response
6
96
32.18644
28
18.40473
{ "bin_edges": [ 6, 16, 26, 36, 46, 56, 66, 76, 86, 96, 96 ], "hist": [ 20, 30, 30, 17, 9, 3, 3, 4, 1, 1 ] }
false
UCL-DARK/ludwig
1-shot
validation
118
utterance
13
107
33.79661
30
15.9307
{ "bin_edges": [ 13, 23, 33, 43, 53, 63, 73, 83, 93, 103, 107 ], "hist": [ 28, 38, 24, 12, 9, 6, 0, 0, 0, 1 ] }
false
UCL-DARK/ludwig
1-shot
test
600
id
1
3
2.82
3
0.42178
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 501 ] }
false
UCL-DARK/ludwig
1-shot
test
600
response
4
142
32.82667
29
19.27509
{ "bin_edges": [ 4, 18, 32, 46, 60, 74, 88, 102, 116, 130, 142 ], "hist": [ 120, 222, 139, 66, 29, 14, 4, 3, 1, 2 ] }
false
UCL-DARK/ludwig
1-shot
test
600
utterance
4
150
33.84833
29
19.43102
{ "bin_edges": [ 4, 19, 34, 49, 64, 79, 94, 109, 124, 139, 150 ], "hist": [ 99, 264, 147, 47, 26, 6, 4, 2, 3, 2 ] }
false
UCL-DARK/ludwig
10-shot
validation
118
id
1
3
2.08475
2
0.48174
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 19 ] }
false
UCL-DARK/ludwig
10-shot
validation
118
response
6
96
32.18644
28
18.40473
{ "bin_edges": [ 6, 16, 26, 36, 46, 56, 66, 76, 86, 96, 96 ], "hist": [ 20, 30, 30, 17, 9, 3, 3, 4, 1, 1 ] }
false
UCL-DARK/ludwig
10-shot
validation
118
utterance
13
107
33.79661
30
15.9307
{ "bin_edges": [ 13, 23, 33, 43, 53, 63, 73, 83, 93, 103, 107 ], "hist": [ 28, 38, 24, 12, 9, 6, 0, 0, 0, 1 ] }
false
UCL-DARK/ludwig
10-shot
test
600
id
1
3
2.82
3
0.42178
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 501 ] }
false
UCL-DARK/ludwig
10-shot
test
600
response
4
142
32.82667
29
19.27509
{ "bin_edges": [ 4, 18, 32, 46, 60, 74, 88, 102, 116, 130, 142 ], "hist": [ 120, 222, 139, 66, 29, 14, 4, 3, 1, 2 ] }
false
UCL-DARK/ludwig
10-shot
test
600
utterance
4
150
33.84833
29
19.43102
{ "bin_edges": [ 4, 19, 34, 49, 64, 79, 94, 109, 124, 139, 150 ], "hist": [ 99, 264, 147, 47, 26, 6, 4, 2, 3, 2 ] }
false
UCL-DARK/ludwig
15-shot
validation
118
id
1
3
2.08475
2
0.48174
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 19 ] }
false
UCL-DARK/ludwig
15-shot
validation
118
response
6
96
32.18644
28
18.40473
{ "bin_edges": [ 6, 16, 26, 36, 46, 56, 66, 76, 86, 96, 96 ], "hist": [ 20, 30, 30, 17, 9, 3, 3, 4, 1, 1 ] }
false
UCL-DARK/ludwig
15-shot
validation
118
utterance
13
107
33.79661
30
15.9307
{ "bin_edges": [ 13, 23, 33, 43, 53, 63, 73, 83, 93, 103, 107 ], "hist": [ 28, 38, 24, 12, 9, 6, 0, 0, 0, 1 ] }
false
UCL-DARK/ludwig
15-shot
test
600
id
1
3
2.82
3
0.42178
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 501 ] }
false
UCL-DARK/ludwig
15-shot
test
600
response
4
142
32.82667
29
19.27509
{ "bin_edges": [ 4, 18, 32, 46, 60, 74, 88, 102, 116, 130, 142 ], "hist": [ 120, 222, 139, 66, 29, 14, 4, 3, 1, 2 ] }
false
UCL-DARK/ludwig
15-shot
test
600
utterance
4
150
33.84833
29
19.43102
{ "bin_edges": [ 4, 19, 34, 49, 64, 79, 94, 109, 124, 139, 150 ], "hist": [ 99, 264, 147, 47, 26, 6, 4, 2, 3, 2 ] }
false
UCL-DARK/ludwig
30-shot
validation
118
id
1
3
2.08475
2
0.48174
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 19 ] }
false
UCL-DARK/ludwig
30-shot
validation
118
response
6
96
32.18644
28
18.40473
{ "bin_edges": [ 6, 16, 26, 36, 46, 56, 66, 76, 86, 96, 96 ], "hist": [ 20, 30, 30, 17, 9, 3, 3, 4, 1, 1 ] }
false
UCL-DARK/ludwig
30-shot
validation
118
utterance
13
107
33.79661
30
15.9307
{ "bin_edges": [ 13, 23, 33, 43, 53, 63, 73, 83, 93, 103, 107 ], "hist": [ 28, 38, 24, 12, 9, 6, 0, 0, 0, 1 ] }
false
UCL-DARK/ludwig
30-shot
test
600
id
1
3
2.82
3
0.42178
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 501 ] }
false
UCL-DARK/ludwig
30-shot
test
600
response
4
142
32.82667
29
19.27509
{ "bin_edges": [ 4, 18, 32, 46, 60, 74, 88, 102, 116, 130, 142 ], "hist": [ 120, 222, 139, 66, 29, 14, 4, 3, 1, 2 ] }
false
UCL-DARK/ludwig
30-shot
test
600
utterance
4
150
33.84833
29
19.43102
{ "bin_edges": [ 4, 19, 34, 49, 64, 79, 94, 109, 124, 139, 150 ], "hist": [ 99, 264, 147, 47, 26, 6, 4, 2, 3, 2 ] }
false
UCL-DARK/ludwig
5-shot
validation
118
id
1
3
2.08475
2
0.48174
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 19 ] }
false
UCL-DARK/ludwig
5-shot
validation
118
response
6
96
32.18644
28
18.40473
{ "bin_edges": [ 6, 16, 26, 36, 46, 56, 66, 76, 86, 96, 96 ], "hist": [ 20, 30, 30, 17, 9, 3, 3, 4, 1, 1 ] }
false
UCL-DARK/ludwig
5-shot
validation
118
utterance
13
107
33.79661
30
15.9307
{ "bin_edges": [ 13, 23, 33, 43, 53, 63, 73, 83, 93, 103, 107 ], "hist": [ 28, 38, 24, 12, 9, 6, 0, 0, 0, 1 ] }
false
UCL-DARK/ludwig
5-shot
test
600
id
1
3
2.82
3
0.42178
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 90, 501 ] }
false
UCL-DARK/ludwig
5-shot
test
600
response
4
142
32.82667
29
19.27509
{ "bin_edges": [ 4, 18, 32, 46, 60, 74, 88, 102, 116, 130, 142 ], "hist": [ 120, 222, 139, 66, 29, 14, 4, 3, 1, 2 ] }
false
UCL-DARK/ludwig
5-shot
test
600
utterance
4
150
33.84833
29
19.43102
{ "bin_edges": [ 4, 19, 34, 49, 64, 79, 94, 109, 124, 139, 150 ], "hist": [ 99, 264, 147, 47, 26, 6, 4, 2, 3, 2 ] }
false
jakartaresearch/news-title-gen
default
train
100,000
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 100000 ] }
false
jakartaresearch/news-title-gen
default
train
100,000
link
0
184
103.58697
111
26.75114
{ "bin_edges": [ 0, 19, 38, 57, 76, 95, 114, 133, 152, 171, 184 ], "hist": [ 4, 58, 13037, 2370, 6777, 33952, 37129, 6638, 32, 3 ] }
false
jakartaresearch/news-title-gen
default
train
100,000
title
7
155
67.40294
65
16.30515
{ "bin_edges": [ 7, 22, 37, 52, 67, 82, 97, 112, 127, 142, 155 ], "hist": [ 71, 1860, 12923, 37895, 26774, 14189, 6218, 57, 10, 3 ] }
false
jakartaresearch/indoqa
default
train
3,309
answer
0
410
34.46462
22
34.57033
{ "bin_edges": [ 0, 42, 84, 126, 168, 210, 252, 294, 336, 378, 410 ], "hist": [ 2310, 659, 212, 44, 10, 9, 4, 1, 0, 1 ] }
false
jakartaresearch/indoqa
default
train
3,309
context
304
1,216
625.21306
588
189.80315
{ "bin_edges": [ 304, 396, 488, 580, 672, 764, 856, 948, 1040, 1132, 1216 ], "hist": [ 242, 727, 654, 524, 379, 309, 215, 170, 75, 14 ] }
false
jakartaresearch/indoqa
default
train
3,309
id
17
32
27.44545
28
1.45274
{ "bin_edges": [ 17, 19, 21, 23, 25, 27, 29, 31, 32 ], "hist": [ 25, 5, 30, 8, 347, 2834, 0, 60 ] }
false
jakartaresearch/indoqa
default
train
3,309
question
14
142
54.89423
54
17.41236
{ "bin_edges": [ 14, 27, 40, 53, 66, 79, 92, 105, 118, 131, 142 ], "hist": [ 109, 541, 912, 893, 544, 221, 65, 18, 4, 2 ] }
false
jakartaresearch/indoqa
default
validation
1,104
answer
2
325
34.72509
22
33.93253
{ "bin_edges": [ 2, 35, 68, 101, 134, 167, 200, 233, 266, 299, 325 ], "hist": [ 684, 253, 92, 33, 13, 6, 1, 1, 0, 1 ] }
false
jakartaresearch/indoqa
default
validation
1,104
context
308
1,216
621.36322
584
186.43764
{ "bin_edges": [ 308, 399, 490, 581, 672, 763, 854, 945, 1036, 1127, 1216 ], "hist": [ 86, 247, 217, 168, 126, 110, 72, 57, 18, 3 ] }
false
jakartaresearch/indoqa
default
validation
1,104
id
17
32
27.42572
28
1.45899
{ "bin_edges": [ 17, 19, 21, 23, 25, 27, 29, 31, 32 ], "hist": [ 8, 1, 12, 2, 117, 944, 0, 20 ] }
false
jakartaresearch/indoqa
default
validation
1,104
question
14
121
55.24728
54
17.34827
{ "bin_edges": [ 14, 25, 36, 47, 58, 69, 80, 91, 102, 113, 121 ], "hist": [ 22, 109, 251, 265, 218, 134, 69, 26, 8, 2 ] }
false
jakartaresearch/id-paraphrase-detection
default
train
4,076
sentence1
22
258
124.94087
122
37.69583
{ "bin_edges": [ 22, 46, 70, 94, 118, 142, 166, 190, 214, 238, 258 ], "hist": [ 11, 207, 713, 937, 854, 709, 432, 169, 38, 6 ] }
false
jakartaresearch/id-paraphrase-detection
default
train
4,076
sentence2
35
250
124.68621
121
38.08565
{ "bin_edges": [ 35, 57, 79, 101, 123, 145, 167, 189, 211, 233, 250 ], "hist": [ 64, 365, 786, 878, 763, 595, 369, 193, 56, 7 ] }
false
jakartaresearch/id-paraphrase-detection
default
validation
1,725
sentence1
35
246
124.10435
121
37.59706
{ "bin_edges": [ 35, 57, 79, 101, 123, 145, 167, 189, 211, 233, 246 ], "hist": [ 30, 160, 340, 363, 307, 262, 181, 60, 20, 2 ] }
false
jakartaresearch/id-paraphrase-detection
default
validation
1,725
sentence2
35
252
124.02667
120
38.05577
{ "bin_edges": [ 35, 57, 79, 101, 123, 145, 167, 189, 211, 233, 252 ], "hist": [ 23, 169, 327, 385, 304, 242, 177, 71, 24, 3 ] }
false
jakartaresearch/semeval-absa
laptop
train
3,048
id
1
4
3.64403
4
0.5521
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 9, 88, 882, 2069 ] }
false
jakartaresearch/semeval-absa
laptop
train
3,048
text
6
464
78.10039
68
48.41396
{ "bin_edges": [ 6, 52, 98, 144, 190, 236, 282, 328, 374, 420, 464 ], "hist": [ 987, 1238, 541, 191, 53, 25, 9, 1, 2, 1 ] }
false
jakartaresearch/semeval-absa
laptop
validation
800
id
3
8
5.3525
5
0.86988
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 8 ], "hist": [ 6, 48, 523, 130, 67, 26 ] }
false
jakartaresearch/semeval-absa
laptop
validation
800
text
13
324
67.57125
58
39.2502
{ "bin_edges": [ 13, 45, 77, 109, 141, 173, 205, 237, 269, 301, 324 ], "hist": [ 249, 291, 159, 59, 24, 10, 5, 1, 0, 2 ] }
false
jakartaresearch/semeval-absa
restaurant
train
3,044
id
1
4
3.70138
4
0.51131
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 6, 61, 769, 2208 ] }
false
jakartaresearch/semeval-absa
restaurant
train
3,044
text
5
357
72.841
65
42.87758
{ "bin_edges": [ 5, 41, 77, 113, 149, 185, 221, 257, 293, 329, 357 ], "hist": [ 727, 1131, 701, 313, 117, 36, 10, 4, 3, 2 ] }
false
jakartaresearch/semeval-absa
restaurant
validation
800
id
12
19
17.20125
17
1.12019
{ "bin_edges": [ 12, 13, 14, 15, 16, 17, 18, 19, 19 ], "hist": [ 29, 0, 0, 0, 0, 467, 302, 2 ] }
false
jakartaresearch/semeval-absa
restaurant
validation
800
text
21
300
75.0875
66
39.55804
{ "bin_edges": [ 21, 49, 77, 105, 133, 161, 189, 217, 245, 273, 300 ], "hist": [ 224, 260, 160, 86, 40, 16, 8, 3, 2, 1 ] }
false
cjvt/sentinews
document_level
train
10,427
content
59
15,467
2,017.36137
1,830
1,071.72716
{ "bin_edges": [ 59, 1600, 3141, 4682, 6223, 7764, 9305, 10846, 12387, 13928, 15467 ], "hist": [ 4125, 5040, 1056, 135, 37, 26, 4, 2, 1, 1 ] }
false
GateNLP/broad_twitter_corpus
broad-twitter-corpus
train
5,342
id
1
4
3.79221
4
0.4578
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 10, 90, 900, 4342 ] }
false
GateNLP/broad_twitter_corpus
broad-twitter-corpus
validation
2,002
id
1
4
3.44555
4
0.60589
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 10, 90, 900, 1002 ] }
false
GateNLP/broad_twitter_corpus
broad-twitter-corpus
test
2,002
id
1
4
3.44555
4
0.60589
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 10, 90, 900, 1002 ] }
false
jakartaresearch/indo-movie-subtitle
default
train
17,740
duration
3
30
28.98839
29
0.55213
{ "bin_edges": [ 3, 6, 9, 12, 15, 18, 21, 24, 27, 30, 30 ], "hist": [ 8, 0, 0, 0, 0, 0, 0, 0, 17730, 2 ] }
false
jakartaresearch/indo-movie-subtitle
default
train
17,740
movie_title
7
57
36.38157
36
14.32917
{ "bin_edges": [ 7, 13, 19, 25, 31, 37, 43, 49, 55, 57 ], "hist": [ 589, 1395, 2966, 2491, 2036, 1959, 3, 5120, 1181 ] }
false
jakartaresearch/indo-movie-subtitle
default
train
17,740
order
0
4
2.93196
3
0.50662
{ "bin_edges": [ 0, 1, 2, 3, 4, 4 ], "hist": [ 8, 211, 2240, 13802, 1479 ] }
false
jakartaresearch/indo-movie-subtitle
default
train
17,740
text
4
322
52.69487
52
13.5972
{ "bin_edges": [ 4, 36, 68, 100, 132, 164, 196, 228, 260, 292, 322 ], "hist": [ 1471, 14055, 2163, 34, 15, 0, 0, 0, 0, 2 ] }
false
research-backup/semeval2012_relational_similarity_v2
semeval2012_relational_similarity_v2
train
89
relation_type
1
3
1.96629
2
0.41147
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 74, 6 ] }
false
research-backup/semeval2012_relational_similarity_v2
semeval2012_relational_similarity_v2
validation
89
relation_type
1
3
1.96629
2
0.41147
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 74, 6 ] }
false
SLPL/naab
all
train
21,662,000
text
8
239,741
127.46832
45
502.66459
{ "bin_edges": [ 8, 23982, 47956, 71930, 95904, 119878, 143852, 167826, 191800, 215774, 239741 ], "hist": [ 21660808, 977, 164, 46, 4, 0, 0, 0, 0, 1 ] }
true
SLPL/naab
all
test
11,010,067
text
6
266,629
124.20202
47
491.59822
{ "bin_edges": [ 6, 26669, 53332, 79995, 106658, 133321, 159984, 186647, 213310, 239973, 266629 ], "hist": [ 11009638, 348, 75, 1, 1, 1, 0, 2, 0, 1 ] }
true
jakartaresearch/inglish
default
train
8,701
english
35
225
115.73451
114
32.98977
{ "bin_edges": [ 35, 55, 75, 95, 115, 135, 155, 175, 195, 215, 225 ], "hist": [ 113, 778, 1714, 1836, 1713, 1289, 900, 303, 50, 5 ] }
false
jakartaresearch/inglish
default
train
8,701
indonesian
22
258
124.74359
121
37.70794
{ "bin_edges": [ 22, 46, 70, 94, 118, 142, 166, 190, 214, 238, 258 ], "hist": [ 22, 449, 1522, 2025, 1828, 1472, 921, 366, 84, 12 ] }
false
jakartaresearch/inglish
default
validation
2,901
english
35
216
115.2082
112
33.18454
{ "bin_edges": [ 35, 54, 73, 92, 111, 130, 149, 168, 187, 206, 216 ], "hist": [ 33, 236, 510, 622, 530, 440, 317, 168, 42, 3 ] }
false
jakartaresearch/inglish
default
validation
2,901
indonesian
35
252
124.13616
120
38.34908
{ "bin_edges": [ 35, 57, 79, 101, 123, 145, 167, 189, 211, 233, 252 ], "hist": [ 45, 284, 542, 655, 504, 420, 276, 127, 43, 5 ] }
false
tyqiangz/multilingual-sentiments
all
train
270,399
text
2
2,377
80.89474
55
80.48124
{ "bin_edges": [ 2, 240, 478, 716, 954, 1192, 1430, 1668, 1906, 2144, 2377 ], "hist": [ 258470, 10688, 921, 211, 58, 48, 0, 0, 0, 3 ] }
false
tyqiangz/multilingual-sentiments
all
validation
10,857
text
3
1,378
93.66713
71
82.07476
{ "bin_edges": [ 3, 141, 279, 417, 555, 693, 831, 969, 1107, 1245, 1378 ], "hist": [ 9173, 1226, 379, 65, 5, 4, 3, 0, 1, 1 ] }
false
tyqiangz/multilingual-sentiments
all
test
14,465
text
2
947
84.24611
74
59.16151
{ "bin_edges": [ 2, 97, 192, 287, 382, 477, 572, 667, 762, 857, 947 ], "hist": [ 9477, 4361, 454, 114, 39, 9, 4, 2, 2, 3 ] }
false
tyqiangz/multilingual-sentiments
arabic
train
1,839
text
11
140
89.34475
90
32.76197
{ "bin_edges": [ 11, 24, 37, 50, 63, 76, 89, 102, 115, 128, 140 ], "hist": [ 35, 68, 137, 221, 233, 199, 193, 168, 271, 314 ] }
false
tyqiangz/multilingual-sentiments
arabic
validation
324
text
12
140
87.94753
93
34.6955
{ "bin_edges": [ 12, 25, 38, 51, 64, 77, 90, 103, 116, 129, 140 ], "hist": [ 11, 15, 33, 36, 34, 28, 36, 34, 48, 49 ] }
false
tyqiangz/multilingual-sentiments
arabic
test
870
text
14
140
85.75747
86
31.20746
{ "bin_edges": [ 14, 27, 40, 53, 66, 79, 92, 105, 118, 131, 140 ], "hist": [ 19, 44, 83, 119, 110, 100, 94, 115, 135, 51 ] }
false
csebuetnlp/BanglaNMT
BanglaNMT
train
2,379,749
bn
1
7,084
71.12126
51
78.45216
{ "bin_edges": [ 1, 710, 1419, 2128, 2837, 3546, 4255, 4964, 5673, 6382, 7084 ], "hist": [ 2376528, 2962, 217, 29, 7, 3, 2, 0, 0, 1 ] }
false
csebuetnlp/BanglaNMT
BanglaNMT
train
2,379,749
en
1
7,153
73.95013
52
83.81505
{ "bin_edges": [ 1, 717, 1433, 2149, 2865, 3581, 4297, 5013, 5729, 6445, 7153 ], "hist": [ 2375756, 3661, 263, 50, 11, 3, 2, 1, 1, 1 ] }
false
csebuetnlp/BanglaNMT
BanglaNMT
test
1,000
bn
33
240
100.299
91
42.0152
{ "bin_edges": [ 33, 54, 75, 96, 117, 138, 159, 180, 201, 222, 240 ], "hist": [ 118, 200, 213, 145, 132, 86, 51, 34, 12, 9 ] }
false
csebuetnlp/BanglaNMT
BanglaNMT
test
1,000
en
50
249
114.004
105
46.76617
{ "bin_edges": [ 50, 70, 90, 110, 130, 150, 170, 190, 210, 230, 249 ], "hist": [ 189, 179, 164, 143, 100, 80, 64, 36, 26, 19 ] }
false
csebuetnlp/BanglaNMT
BanglaNMT
validation
597
bn
32
243
103.55946
96
42.43983
{ "bin_edges": [ 32, 54, 76, 98, 120, 142, 164, 186, 208, 230, 243 ], "hist": [ 48, 140, 118, 104, 70, 55, 36, 14, 11, 1 ] }
false
csebuetnlp/BanglaNMT
BanglaNMT
validation
597
en
50
249
116.40369
111
45.72363
{ "bin_edges": [ 50, 70, 90, 110, 130, 150, 170, 190, 210, 230, 249 ], "hist": [ 98, 103, 93, 97, 70, 54, 29, 27, 16, 10 ] }
false
tyqiangz/multilingual-sentiments
chinese
train
120,000
text
11
2,377
51.75502
37
49.28847
{ "bin_edges": [ 11, 248, 485, 722, 959, 1196, 1433, 1670, 1907, 2144, 2377 ], "hist": [ 118914, 929, 120, 18, 6, 10, 0, 0, 0, 3 ] }
false
tyqiangz/multilingual-sentiments
chinese
validation
3,000
text
16
1,378
51.009
36
50.62178
{ "bin_edges": [ 16, 153, 290, 427, 564, 701, 838, 975, 1112, 1249, 1378 ], "hist": [ 2912, 70, 12, 4, 1, 0, 0, 0, 0, 1 ] }
false
tyqiangz/multilingual-sentiments
chinese
test
3,000
text
18
833
51.95433
36
48.91702
{ "bin_edges": [ 18, 100, 182, 264, 346, 428, 510, 592, 674, 756, 833 ], "hist": [ 2725, 203, 44, 17, 7, 1, 1, 0, 1, 1 ] }
false
tyqiangz/multilingual-sentiments
english
train
1,839
text
29
185
109.56661
115
25.67076
{ "bin_edges": [ 29, 45, 61, 77, 93, 109, 125, 141, 157, 173, 185 ], "hist": [ 24, 67, 149, 204, 308, 491, 452, 135, 8, 1 ] }
false
tyqiangz/multilingual-sentiments
english
validation
324
text
30
171
109.24383
115
26.77343
{ "bin_edges": [ 30, 45, 60, 75, 90, 105, 120, 135, 150, 165, 171 ], "hist": [ 3, 17, 23, 34, 32, 83, 68, 63, 0, 1 ] }
false
tyqiangz/multilingual-sentiments
english
test
870
text
15
142
92.83333
96
29.63527
{ "bin_edges": [ 15, 28, 41, 54, 67, 80, 93, 106, 119, 132, 142 ], "hist": [ 7, 30, 70, 92, 102, 104, 114, 173, 89, 89 ] }
false
tyqiangz/multilingual-sentiments
french
train
1,839
text
16
144
96.84122
101
25.24023
{ "bin_edges": [ 16, 29, 42, 55, 68, 81, 94, 107, 120, 133, 144 ], "hist": [ 9, 24, 85, 136, 244, 271, 283, 464, 217, 106 ] }
false
tyqiangz/multilingual-sentiments
french
validation
324
text
18
142
95.71914
98.5
25.62116
{ "bin_edges": [ 18, 31, 44, 57, 70, 83, 96, 109, 122, 135, 142 ], "hist": [ 4, 4, 21, 30, 36, 54, 45, 93, 26, 11 ] }
false
tyqiangz/multilingual-sentiments
french
test
870
text
12
140
94.46782
98
25.36225
{ "bin_edges": [ 12, 25, 38, 51, 64, 77, 90, 103, 116, 129, 140 ], "hist": [ 4, 9, 35, 73, 101, 126, 135, 160, 167, 60 ] }
false
tyqiangz/multilingual-sentiments
german
train
1,839
text
7
144
74.53562
71
34.68468
{ "bin_edges": [ 7, 21, 35, 49, 63, 77, 91, 105, 119, 133, 144 ], "hist": [ 83, 194, 224, 254, 243, 191, 188, 234, 160, 68 ] }
false
tyqiangz/multilingual-sentiments
german
test
870
text
9
144
73.9046
71
34.46516
{ "bin_edges": [ 9, 23, 37, 51, 65, 79, 93, 107, 121, 135, 144 ], "hist": [ 47, 97, 117, 118, 118, 85, 85, 119, 61, 23 ] }
false