dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
midas/semeval2010
extraction
test
100
id
3
4
3.74
4
0.44084
{ "bin_edges": [ 3, 4, 4 ], "hist": [ 26, 74 ] }
false
midas/semeval2010
raw
train
144
id
4
4
4
4
0
{ "bin_edges": [ 4, 4 ], "hist": [ 144 ] }
false
midas/semeval2010
generation
test
100
id
3
4
3.74
4
0.44084
{ "bin_edges": [ 3, 4, 4 ], "hist": [ 26, 74 ] }
false
midas/semeval2010
raw
test
100
id
3
4
3.74
4
0.44084
{ "bin_edges": [ 3, 4, 4 ], "hist": [ 26, 74 ] }
false
midas/semeval2010
extraction
train
144
id
4
4
4
4
0
{ "bin_edges": [ 4, 4 ], "hist": [ 144 ] }
false
midas/semeval2010
generation
train
144
id
4
4
4
4
0
{ "bin_edges": [ 4, 4 ], "hist": [ 144 ] }
false
arbml/PADIC
default
train
7,213
ALGIERS
4
280
34.45349
28
24.43763
{ "bin_edges": [ 4, 32, 60, 88, 116, 144, 172, 200, 228, 256, 280 ], "hist": [ 4084, 2236, 610, 192, 63, 17, 2, 6, 1, 2 ] }
false
arbml/PADIC
default
train
7,213
ANNABA
4
280
33.47803
27
24.70457
{ "bin_edges": [ 4, 32, 60, 88, 116, 144, 172, 200, 228, 256, 280 ], "hist": [ 4195, 2159, 592, 175, 65, 11, 8, 4, 2, 2 ] }
false
arbml/PADIC
default
train
7,213
MODERN-STANDARD-ARABIC
4
268
36.01275
30
25.739
{ "bin_edges": [ 4, 31, 58, 85, 112, 139, 166, 193, 220, 247, 268 ], "hist": [ 3704, 2381, 764, 225, 93, 24, 12, 5, 2, 3 ] }
false
arbml/PADIC
default
train
7,213
PALESTINIAN
4
275
33.86386
28
24.60537
{ "bin_edges": [ 4, 32, 60, 88, 116, 144, 172, 200, 228, 256, 275 ], "hist": [ 4179, 2125, 635, 177, 65, 19, 7, 4, 1, 1 ] }
false
arbml/PADIC
default
train
7,213
SYRIAN
4
246
32.15112
26
22.9777
{ "bin_edges": [ 4, 29, 54, 79, 104, 129, 154, 179, 204, 229, 246 ], "hist": [ 3959, 2187, 755, 204, 69, 23, 10, 4, 0, 2 ] }
false
midas/semeval2017
extraction
test
100
id
17
18
17.01
17
0.1
{ "bin_edges": [ 17, 18, 18 ], "hist": [ 99, 1 ] }
false
midas/semeval2017
generation
train
350
id
17
17
17
17
0
{ "bin_edges": [ 17, 17 ], "hist": [ 350 ] }
false
midas/semeval2017
extraction
train
350
id
17
17
17
17
0
{ "bin_edges": [ 17, 17 ], "hist": [ 350 ] }
false
midas/semeval2017
extraction
validation
50
id
17
18
17.02
17
0.14142
{ "bin_edges": [ 17, 18, 18 ], "hist": [ 49, 1 ] }
false
midas/semeval2017
raw
validation
50
id
17
18
17.02
17
0.14142
{ "bin_edges": [ 17, 18, 18 ], "hist": [ 49, 1 ] }
false
midas/semeval2017
generation
validation
50
id
17
18
17.02
17
0.14142
{ "bin_edges": [ 17, 18, 18 ], "hist": [ 49, 1 ] }
false
midas/semeval2017
generation
test
100
id
17
18
17.01
17
0.1
{ "bin_edges": [ 17, 18, 18 ], "hist": [ 99, 1 ] }
false
midas/semeval2017
raw
train
350
id
17
17
17
17
0
{ "bin_edges": [ 17, 17 ], "hist": [ 350 ] }
false
midas/semeval2017
raw
test
100
id
17
18
17.01
17
0.1
{ "bin_edges": [ 17, 18, 18 ], "hist": [ 99, 1 ] }
false
arize-ai/beer_reviews_label_drift_neg
default
training
9,000
beer_name
2
63
21.48333
21
9.53948
{ "bin_edges": [ 2, 9, 16, 23, 30, 37, 44, 51, 58, 63 ], "hist": [ 614, 2169, 2381, 2229, 938, 483, 136, 46, 4 ] }
false
arize-ai/beer_reviews_label_drift_neg
default
training
9,000
beer_style
4
35
17.79411
18
7.17103
{ "bin_edges": [ 4, 8, 12, 16, 20, 24, 28, 32, 35 ], "hist": [ 895, 794, 2105, 1498, 2277, 416, 548, 467 ] }
false
arize-ai/beer_reviews_label_drift_neg
default
training
9,000
text
13
4,380
684.21811
602
367.49913
{ "bin_edges": [ 13, 450, 887, 1324, 1761, 2198, 2635, 3072, 3509, 3946, 4380 ], "hist": [ 2505, 4599, 1376, 340, 131, 33, 8, 4, 1, 3 ] }
false
arize-ai/beer_reviews_label_drift_neg
default
validation
1,260
beer_name
3
75
21.53095
21
10.00928
{ "bin_edges": [ 3, 11, 19, 27, 35, 43, 51, 59, 67, 75, 75 ], "hist": [ 156, 356, 410, 192, 105, 32, 7, 1, 0, 1 ] }
false
arize-ai/beer_reviews_label_drift_neg
default
validation
1,260
beer_style
4
35
17.80873
18
7.13241
{ "bin_edges": [ 4, 8, 12, 16, 20, 24, 28, 32, 35 ], "hist": [ 114, 129, 296, 212, 301, 58, 90, 60 ] }
false
arize-ai/beer_reviews_label_drift_neg
default
validation
1,260
text
6
2,993
686.50397
593
358.27504
{ "bin_edges": [ 6, 305, 604, 903, 1202, 1501, 1800, 2099, 2398, 2697, 2993 ], "hist": [ 76, 569, 351, 155, 65, 23, 14, 5, 1, 1 ] }
false
arize-ai/beer_reviews_label_drift_neg
default
production
27,742
beer_name
2
75
21.41143
21
9.60752
{ "bin_edges": [ 2, 10, 18, 26, 34, 42, 50, 58, 66, 74, 75 ], "hist": [ 2421, 7861, 9175, 4991, 2300, 777, 194, 16, 3, 4 ] }
false
arize-ai/beer_reviews_label_drift_neg
default
production
27,742
beer_style
4
35
17.77388
18
7.09677
{ "bin_edges": [ 4, 8, 12, 16, 20, 24, 28, 32, 35 ], "hist": [ 2669, 2551, 6519, 4597, 7090, 1262, 1642, 1412 ] }
false
arize-ai/beer_reviews_label_drift_neg
default
production
27,742
text
6
4,859
684.70864
593
374.59102
{ "bin_edges": [ 6, 492, 978, 1464, 1950, 2436, 2922, 3408, 3894, 4380, 4859 ], "hist": [ 9614, 13660, 3361, 778, 208, 72, 26, 16, 3, 4 ] }
false
arize-ai/beer_reviews_label_drift_neutral
default
training
8,525
beer_name
2
75
21.23261
21
9.66598
{ "bin_edges": [ 2, 10, 18, 26, 34, 42, 50, 58, 66, 74, 75 ], "hist": [ 742, 2518, 2725, 1530, 721, 222, 61, 2, 3, 1 ] }
false
arize-ai/beer_reviews_label_drift_neutral
default
training
8,525
beer_style
4
35
17.87191
18
6.9189
{ "bin_edges": [ 4, 8, 12, 16, 20, 24, 28, 32, 35 ], "hist": [ 710, 955, 1857, 1389, 2400, 323, 476, 415 ] }
false
arize-ai/beer_reviews_label_drift_neutral
default
training
8,525
text
6
4,889
688.64106
595
389.5445
{ "bin_edges": [ 6, 495, 984, 1473, 1962, 2451, 2940, 3429, 3918, 4407, 4889 ], "hist": [ 3047, 4081, 1018, 258, 78, 29, 8, 5, 0, 1 ] }
false
arize-ai/beer_reviews_label_drift_neutral
default
validation
1,193
beer_name
3
65
21.03604
20
9.70338
{ "bin_edges": [ 3, 10, 17, 24, 31, 38, 45, 52, 59, 65 ], "hist": [ 108, 342, 308, 252, 117, 41, 21, 2, 2 ] }
false
arize-ai/beer_reviews_label_drift_neutral
default
validation
1,193
beer_style
4
35
18.16764
19
7.24618
{ "bin_edges": [ 4, 8, 12, 16, 20, 24, 28, 32, 35 ], "hist": [ 104, 136, 245, 167, 342, 53, 73, 73 ] }
false
arize-ai/beer_reviews_label_drift_neutral
default
validation
1,193
text
41
3,178
697.89438
613
373.48639
{ "bin_edges": [ 41, 355, 669, 983, 1297, 1611, 1925, 2239, 2553, 2867, 3178 ], "hist": [ 157, 530, 299, 131, 41, 18, 10, 4, 0, 3 ] }
false
arize-ai/beer_reviews_label_drift_neutral
default
production
26,278
beer_name
2
75
21.18289
21
9.65355
{ "bin_edges": [ 2, 10, 18, 26, 34, 42, 50, 58, 66, 74, 75 ], "hist": [ 2298, 7858, 8514, 4547, 2088, 774, 176, 15, 4, 4 ] }
false
arize-ai/beer_reviews_label_drift_neutral
default
production
26,278
beer_style
4
35
17.83343
18
6.96395
{ "bin_edges": [ 4, 8, 12, 16, 20, 24, 28, 32, 35 ], "hist": [ 2333, 2765, 5783, 4342, 7194, 1130, 1426, 1305 ] }
false
arize-ai/beer_reviews_label_drift_neutral
default
production
26,278
text
6
4,952
690.37617
601
381.91534
{ "bin_edges": [ 6, 501, 996, 1491, 1986, 2481, 2976, 3471, 3966, 4461, 4952 ], "hist": [ 9435, 12651, 3167, 707, 215, 65, 20, 8, 4, 6 ] }
false
research-backup/semeval2012_relational_similarity_v3
semeval2012_relational_similarity_v3
train
3,170
relation_type
1
3
2.07287
2
0.27067
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 2921, 240 ] }
false
research-backup/semeval2012_relational_similarity_v3
semeval2012_relational_similarity_v3
validation
3,768
relation_type
1
3
2.07086
2
0.26577
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 3483, 276 ] }
false
cjvt/slo_thesaurus
default
train
105,473
headword
1
50
12.97112
12
5.91335
{ "bin_edges": [ 1, 6, 11, 16, 21, 26, 31, 36, 41, 46, 50 ], "hist": [ 5208, 37011, 33372, 18934, 6846, 2700, 981, 324, 83, 14 ] }
false
cjvt/slo_thesaurus
default
train
105,473
id_headword
4
9
7.9466
8
0.42137
{ "bin_edges": [ 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 9, 90, 900, 9000, 90000, 5474 ] }
false
research-backup/semeval2012_relational_similarity_v4
semeval2012_relational_similarity_v4
train
6,821
relation_type
1
3
2.07228
2
0.26401
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 6310, 502 ] }
false
research-backup/semeval2012_relational_similarity_v4
semeval2012_relational_similarity_v4
validation
1,772
relation_type
1
3
2.07393
2
0.28047
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 1623, 140 ] }
false
drt/kqa_pro
test
test
11,797
question
19
264
89.75155
86
34.01722
{ "bin_edges": [ 19, 44, 69, 94, 119, 144, 169, 194, 219, 244, 264 ], "hist": [ 663, 2885, 3341, 2684, 1374, 586, 179, 63, 19, 3 ] }
false
drt/kqa_pro
train_val
train
94,376
answer
1
316
10.28485
9
8.66547
{ "bin_edges": [ 1, 33, 65, 97, 129, 161, 193, 225, 257, 289, 316 ], "hist": [ 91982, 2338, 45, 7, 1, 1, 0, 0, 1, 1 ] }
false
drt/kqa_pro
train_val
train
94,376
question
15
439
89.6041
86
34.27607
{ "bin_edges": [ 15, 58, 101, 144, 187, 230, 273, 316, 359, 402, 439 ], "hist": [ 17062, 45737, 24608, 6035, 816, 101, 12, 3, 1, 1 ] }
false
drt/kqa_pro
train_val
train
94,376
sparql
70
654
225.9281
220
81.6446
{ "bin_edges": [ 70, 129, 188, 247, 306, 365, 424, 483, 542, 601, 654 ], "hist": [ 11347, 23712, 24029, 18812, 11930, 2962, 922, 544, 95, 23 ] }
false
drt/kqa_pro
train_val
validation
11,797
answer
1
106
10.18454
9
8.35931
{ "bin_edges": [ 1, 12, 23, 34, 45, 56, 67, 78, 89, 100, 106 ], "hist": [ 7834, 3150, 562, 157, 81, 6, 4, 2, 0, 1 ] }
false
drt/kqa_pro
train_val
validation
11,797
question
16
305
89.67746
86
34.1252
{ "bin_edges": [ 16, 45, 74, 103, 132, 161, 190, 219, 248, 277, 305 ], "hist": [ 778, 3447, 3842, 2355, 970, 295, 88, 18, 3, 1 ] }
false
drt/kqa_pro
train_val
validation
11,797
sparql
71
616
226.51776
221
81.09923
{ "bin_edges": [ 71, 126, 181, 236, 291, 346, 401, 456, 511, 566, 616 ], "hist": [ 1323, 2601, 2752, 2596, 1612, 614, 180, 89, 25, 5 ] }
false
cjvt/slownet
default
train
117,658
en_def
3
505
59.41832
50
38.6694
{ "bin_edges": [ 3, 54, 105, 156, 207, 258, 309, 360, 411, 462, 505 ], "hist": [ 63705, 40241, 10567, 2379, 567, 139, 44, 7, 7, 2 ] }
false
cjvt/slownet
default
train
117,658
id
17
17
17
17
0
{ "bin_edges": [ 17, 17 ], "hist": [ 117658 ] }
false
cjvt/slownet
default
train
117,658
semeval07_cluster
2
36
4.24264
3
2.65806
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 36 ], "hist": [ 93555, 16155, 6723, 988, 155, 56, 17, 7, 2 ] }
false
cjvt/slownet
default
train
117,658
sl_def
1
417
5.17178
3
15.35015
{ "bin_edges": [ 1, 43, 85, 127, 169, 211, 253, 295, 337, 379, 417 ], "hist": [ 115157, 1134, 824, 367, 118, 36, 13, 6, 1, 2 ] }
false
research-backup/semeval2012_relational_similarity_v5
semeval2012_relational_similarity_v5
train
8,504
relation_type
1
3
2.07373
2
0.26537
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 9, 7859, 636 ] }
false
drt/complex_web_questions
complex_web_questions
train
27,639
ID
11
46
44.90799
45
0.79019
{ "bin_edges": [ 11, 15, 19, 23, 27, 31, 35, 39, 43, 46 ], "hist": [ 5, 0, 0, 0, 0, 0, 0, 42, 27592 ] }
false
drt/complex_web_questions
complex_web_questions
train
27,639
composition_answer
0
41
4.78878
0
5.88365
{ "bin_edges": [ 0, 5, 10, 15, 20, 25, 30, 35, 40, 41 ], "hist": [ 15196, 5799, 4393, 1940, 272, 32, 5, 1, 1 ] }
false
drt/complex_web_questions
complex_web_questions
train
27,639
created
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 27639 ] }
false
drt/complex_web_questions
complex_web_questions
train
27,639
machine_question
21
172
84.7319
83
18.38548
{ "bin_edges": [ 21, 37, 53, 69, 85, 101, 117, 133, 149, 165, 172 ], "hist": [ 16, 553, 4471, 9913, 7731, 3345, 1200, 337, 68, 5 ] }
false
drt/complex_web_questions
complex_web_questions
train
27,639
question
19
168
77.99302
76
18.21628
{ "bin_edges": [ 19, 34, 49, 64, 79, 94, 109, 124, 139, 154, 168 ], "hist": [ 22, 954, 5082, 9006, 7264, 3680, 1223, 334, 70, 4 ] }
false
drt/complex_web_questions
complex_web_questions
train
27,639
sparql
127
1,390
350.33341
317
127.69912
{ "bin_edges": [ 127, 254, 381, 508, 635, 762, 889, 1016, 1143, 1270, 1390 ], "hist": [ 1487, 20709, 3381, 241, 939, 611, 255, 0, 0, 16 ] }
false
drt/complex_web_questions
complex_web_questions
train
27,639
webqsp_ID
9
13
11.9136
12
0.64645
{ "bin_edges": [ 9, 10, 11, 12, 13, 13 ], "hist": [ 47, 409, 5563, 17486, 4134 ] }
false
drt/complex_web_questions
complex_web_questions
train
27,639
webqsp_question
15
74
36.15749
36
7.71525
{ "bin_edges": [ 15, 21, 27, 33, 39, 45, 51, 57, 63, 69, 74 ], "hist": [ 393, 2377, 6287, 8658, 6394, 2545, 623, 266, 64, 32 ] }
false
drt/complex_web_questions
complex_web_questions
validation
3,519
ID
43
46
44.91731
45
0.60717
{ "bin_edges": [ 43, 44, 45, 46, 46 ], "hist": [ 70, 596, 2408, 445 ] }
false
drt/complex_web_questions
complex_web_questions
validation
3,519
composition_answer
0
29
4.38704
0
5.6182
{ "bin_edges": [ 0, 3, 6, 9, 12, 15, 18, 21, 24, 27, 29 ], "hist": [ 1953, 173, 569, 278, 303, 168, 54, 15, 3, 3 ] }
false
drt/complex_web_questions
complex_web_questions
validation
3,519
created
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 3519 ] }
false
drt/complex_web_questions
complex_web_questions
validation
3,519
machine_question
31
162
84.37056
82
17.72175
{ "bin_edges": [ 31, 45, 59, 73, 87, 101, 115, 129, 143, 157, 162 ], "hist": [ 12, 155, 752, 1180, 824, 383, 151, 50, 11, 1 ] }
false
drt/complex_web_questions
complex_web_questions
validation
3,519
question
29
146
77.55016
76
18.33997
{ "bin_edges": [ 29, 41, 53, 65, 77, 89, 101, 113, 125, 137, 146 ], "hist": [ 31, 206, 659, 888, 826, 518, 231, 118, 33, 9 ] }
false
drt/complex_web_questions
complex_web_questions
validation
3,519
sparql
236
926
353.42029
320
129.12007
{ "bin_edges": [ 236, 306, 376, 446, 516, 586, 656, 726, 796, 866, 926 ], "hist": [ 1387, 1468, 335, 58, 1, 34, 103, 11, 58, 64 ] }
false
drt/complex_web_questions
complex_web_questions
validation
3,519
webqsp_ID
10
13
11.91731
12
0.60717
{ "bin_edges": [ 10, 11, 12, 13, 13 ], "hist": [ 70, 596, 2408, 445 ] }
false
drt/complex_web_questions
complex_web_questions
validation
3,519
webqsp_question
17
72
35.12845
35
7.57352
{ "bin_edges": [ 17, 23, 29, 35, 41, 47, 53, 59, 65, 71, 72 ], "hist": [ 138, 570, 875, 1246, 401, 225, 45, 16, 0, 3 ] }
false
drt/complex_web_questions
complexwebquestions_test
test
3,531
ID
42
46
44.8785
45
0.75887
{ "bin_edges": [ 42, 43, 44, 45, 46, 46 ], "hist": [ 2, 209, 618, 2089, 613 ] }
false
drt/complex_web_questions
complexwebquestions_test
test
3,531
created
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 3531 ] }
false
drt/complex_web_questions
complexwebquestions_test
test
3,531
machine_question
27
166
84.4789
82
18.58353
{ "bin_edges": [ 27, 41, 55, 69, 83, 97, 111, 125, 139, 153, 166 ], "hist": [ 8, 102, 553, 1110, 937, 524, 183, 80, 26, 8 ] }
false
drt/complex_web_questions
complexwebquestions_test
test
3,531
question
25
160
77.04673
75
18.1502
{ "bin_edges": [ 25, 39, 53, 67, 81, 95, 109, 123, 137, 151, 160 ], "hist": [ 11, 245, 790, 1103, 820, 373, 129, 49, 9, 2 ] }
false
drt/complex_web_questions
complexwebquestions_test
test
3,531
sparql
236
924
343.06939
316
115.06775
{ "bin_edges": [ 236, 305, 374, 443, 512, 581, 650, 719, 788, 857, 924 ], "hist": [ 1536, 1288, 400, 111, 20, 31, 43, 4, 39, 59 ] }
false
drt/complex_web_questions
complexwebquestions_test
test
3,531
webqsp_ID
9
13
11.8785
12
0.75887
{ "bin_edges": [ 9, 10, 11, 12, 13, 13 ], "hist": [ 2, 209, 618, 2089, 613 ] }
false
drt/complex_web_questions
complexwebquestions_test
test
3,531
webqsp_question
18
69
36.55452
36
7.50634
{ "bin_edges": [ 18, 24, 30, 36, 42, 48, 54, 60, 66, 69 ], "hist": [ 94, 554, 933, 1100, 552, 239, 46, 0, 13 ] }
false
matejklemen/vuamc
default
train
16,740
document_name
14
14
14
14
0
{ "bin_edges": [ 14, 14 ], "hist": [ 16740 ] }
false
vialibre/splittedspanish3bwc
DGT_12
all
88,003
text
2
1,664
118.09945
76
119.32461
{ "bin_edges": [ 2, 169, 336, 503, 670, 837, 1004, 1171, 1338, 1505, 1664 ], "hist": [ 65315, 17551, 4068, 812, 183, 40, 15, 14, 2, 3 ] }
false
evaluate/glue-ci
sst2
train
67,349
sentence
2
268
53.50559
39
43.40738
{ "bin_edges": [ 2, 29, 56, 83, 110, 137, 164, 191, 218, 245, 268 ], "hist": [ 25025, 18113, 9996, 6262, 3844, 2177, 1168, 499, 187, 78 ] }
false
evaluate/glue-ci
stsb
test
1,379
sentence1
16
215
53.73459
44
32.66204
{ "bin_edges": [ 16, 36, 56, 76, 96, 116, 136, 156, 176, 196, 215 ], "hist": [ 454, 492, 179, 101, 56, 42, 30, 19, 5, 1 ] }
false
evaluate/glue-ci
stsb
test
1,379
sentence2
13
199
53.49746
43
32.53056
{ "bin_edges": [ 13, 32, 51, 70, 89, 108, 127, 146, 165, 184, 199 ], "hist": [ 344, 537, 217, 95, 68, 46, 34, 27, 5, 6 ] }
false
evaluate/glue-ci
stsb
train
5,749
sentence1
16
367
57.70847
47
35.37775
{ "bin_edges": [ 16, 52, 88, 124, 160, 196, 232, 268, 304, 340, 367 ], "hist": [ 3494, 1355, 485, 278, 107, 21, 5, 1, 1, 2 ] }
false
evaluate/glue-ci
stsb
train
5,749
sentence2
15
311
57.53261
47
34.64733
{ "bin_edges": [ 15, 45, 75, 105, 135, 165, 195, 225, 255, 285, 311 ], "hist": [ 2564, 2033, 531, 329, 186, 83, 12, 8, 2, 1 ] }
false
evaluate/glue-ci
stsb
validation
1,500
sentence1
12
200
64.258
55
33.82119
{ "bin_edges": [ 12, 31, 50, 69, 88, 107, 126, 145, 164, 183, 200 ], "hist": [ 202, 459, 267, 217, 171, 102, 38, 27, 13, 4 ] }
false
evaluate/glue-ci
stsb
validation
1,500
sentence2
17
186
63.712
55
33.3409
{ "bin_edges": [ 17, 34, 51, 68, 85, 102, 119, 136, 153, 170, 186 ], "hist": [ 280, 387, 278, 188, 150, 104, 59, 28, 14, 12 ] }
false
evaluate/glue-ci
wnli
train
635
sentence1
36
323
104.57638
89
55.35804
{ "bin_edges": [ 36, 65, 94, 123, 152, 181, 210, 239, 268, 297, 323 ], "hist": [ 133, 227, 114, 73, 15, 14, 22, 28, 7, 2 ] }
false
evaluate/glue-ci
wnli
train
635
sentence2
10
186
44.09921
38
24.26252
{ "bin_edges": [ 10, 28, 46, 64, 82, 100, 118, 136, 154, 172, 186 ], "hist": [ 173, 221, 130, 53, 40, 7, 7, 2, 1, 1 ] }
false
evaluate/glue-ci
wnli
validation
71
sentence1
44
323
111.83099
92
59.2884
{ "bin_edges": [ 44, 72, 100, 128, 156, 184, 212, 240, 268, 296, 323 ], "hist": [ 19, 22, 12, 3, 5, 3, 2, 4, 0, 1 ] }
false
evaluate/glue-ci
wnli
validation
71
sentence2
16
109
39.46479
33
18.07906
{ "bin_edges": [ 16, 26, 36, 46, 56, 66, 76, 86, 96, 106, 109 ], "hist": [ 12, 25, 17, 6, 5, 1, 4, 0, 0, 1 ] }
false
evaluate/glue-ci
wnli
test
146
sentence1
81
292
172.16438
163
59.44862
{ "bin_edges": [ 81, 103, 125, 147, 169, 191, 213, 235, 257, 279, 292 ], "hist": [ 19, 26, 14, 15, 13, 17, 21, 12, 0, 9 ] }
false
evaluate/glue-ci
wnli
test
146
sentence2
15
177
67.34932
59.5
33.51821
{ "bin_edges": [ 15, 32, 49, 66, 83, 100, 117, 134, 151, 168, 177 ], "hist": [ 13, 43, 22, 23, 22, 13, 4, 2, 0, 4 ] }
false
feradauto/MoralExceptQA
all
test
148
context
3
19
8.94595
9
3.18753
{ "bin_edges": [ 3, 5, 7, 9, 11, 13, 15, 17, 19, 19 ], "hist": [ 14, 27, 17, 49, 23, 11, 6, 0, 1 ] }
false
feradauto/MoralExceptQA
all
test
148
scenario
171
455
275.22973
275
74.80133
{ "bin_edges": [ 171, 200, 229, 258, 287, 316, 345, 374, 403, 432, 455 ], "hist": [ 43, 9, 13, 18, 18, 13, 14, 16, 3, 1 ] }
false
feradauto/MoralExceptQA
cutting_in_line
test
66
context
4
19
7.81818
7
3.1764
{ "bin_edges": [ 4, 6, 8, 10, 12, 14, 16, 18, 19 ], "hist": [ 19, 17, 7, 15, 5, 2, 0, 1 ] }
false
feradauto/MoralExceptQA
cutting_in_line
test
66
scenario
234
455
299.43939
291
44.05386
{ "bin_edges": [ 234, 257, 280, 303, 326, 349, 372, 395, 418, 441, 455 ], "hist": [ 10, 13, 18, 10, 6, 5, 2, 1, 0, 1 ] }
false
feradauto/MoralExceptQA
property_damage
test
54
scenario
171
236
194.5
190
15.81825
{ "bin_edges": [ 171, 178, 185, 192, 199, 206, 213, 220, 227, 234, 236 ], "hist": [ 4, 9, 17, 10, 4, 2, 1, 3, 3, 1 ] }
false
taln-ls2n/kpbiomed
large
train
2,363,000
abstract
1
13,403
1,473.82214
1,481
478.88747
{ "bin_edges": [ 1, 1342, 2683, 4024, 5365, 6706, 8047, 9388, 10729, 12070, 13403 ], "hist": [ 887898, 1448050, 26025, 869, 111, 25, 11, 8, 1, 2 ] }
true