dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
taln-ls2n/kpbiomed
large
train
2,363,000
authors
2
63,913
144.52647
123
126.60294
{ "bin_edges": [ 2, 6394, 12786, 19178, 25570, 31962, 38354, 44746, 51138, 57530, 63913 ], "hist": [ 2362981, 9, 3, 1, 1, 3, 1, 0, 0, 1 ] }
true
taln-ls2n/kpbiomed
large
train
2,363,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 2363000 ] }
true
taln-ls2n/kpbiomed
large
train
2,363,000
title
1
1,062
110.08315
107
36.28907
{ "bin_edges": [ 1, 108, 215, 322, 429, 536, 643, 750, 857, 964, 1062 ], "hist": [ 1183332, 1161115, 18249, 260, 30, 8, 2, 1, 1, 2 ] }
true
taln-ls2n/kpbiomed
large
test
20,000
abstract
18
4,682
1,474.05685
1,483
478.92172
{ "bin_edges": [ 18, 485, 952, 1419, 1886, 2353, 2820, 3287, 3754, 4221, 4682 ], "hist": [ 362, 2278, 6230, 7952, 2432, 583, 118, 31, 9, 5 ] }
true
taln-ls2n/kpbiomed
large
test
20,000
authors
2
4,734
144.68
122
111.95902
{ "bin_edges": [ 2, 476, 950, 1424, 1898, 2372, 2846, 3320, 3794, 4268, 4734 ], "hist": [ 19743, 236, 11, 7, 1, 0, 0, 0, 0, 2 ] }
true
taln-ls2n/kpbiomed
large
test
20,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 20000 ] }
true
taln-ls2n/kpbiomed
large
test
20,000
title
6
388
110.71435
108
36.26139
{ "bin_edges": [ 6, 45, 84, 123, 162, 201, 240, 279, 318, 357, 388 ], "hist": [ 379, 4240, 8532, 5143, 1394, 262, 41, 7, 0, 2 ] }
true
taln-ls2n/kpbiomed
large
validation
20,000
abstract
11
5,791
1,475.03755
1,485
485.84017
{ "bin_edges": [ 11, 590, 1169, 1748, 2327, 2906, 3485, 4064, 4643, 5222, 5791 ], "hist": [ 677, 4468, 9544, 4484, 702, 92, 26, 5, 0, 2 ] }
true
taln-ls2n/kpbiomed
large
validation
20,000
authors
2
2,556
145.1623
124
102.08738
{ "bin_edges": [ 2, 258, 514, 770, 1026, 1282, 1538, 1794, 2050, 2306, 2556 ], "hist": [ 18057, 1749, 150, 23, 15, 3, 2, 0, 0, 1 ] }
true
taln-ls2n/kpbiomed
large
validation
20,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 20000 ] }
true
taln-ls2n/kpbiomed
large
validation
20,000
title
10
987
110.393
108
36.68865
{ "bin_edges": [ 10, 108, 206, 304, 402, 500, 598, 696, 794, 892, 987 ], "hist": [ 9942, 9820, 234, 3, 0, 0, 0, 0, 0, 1 ] }
true
taln-ls2n/kpbiomed
medium
train
2,000,000
abstract
1
31,166
1,472.02197
1,480
481.24955
{ "bin_edges": [ 1, 3118, 6235, 9352, 12469, 15586, 18703, 21820, 24937, 28054, 31166 ], "hist": [ 1993531, 6405, 50, 12, 1, 0, 0, 0, 0, 1 ] }
false
taln-ls2n/kpbiomed
medium
train
2,000,000
authors
2
63,913
144.47274
123
134.11166
{ "bin_edges": [ 2, 6394, 12786, 19178, 25570, 31962, 38354, 44746, 51138, 57530, 63913 ], "hist": [ 1999982, 7, 1, 2, 2, 5, 0, 0, 0, 1 ] }
false
taln-ls2n/kpbiomed
medium
train
2,000,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 2000000 ] }
false
taln-ls2n/kpbiomed
medium
train
2,000,000
title
1
1,159
110.22193
108
36.36995
{ "bin_edges": [ 1, 117, 233, 349, 465, 581, 697, 813, 929, 1045, 1159 ], "hist": [ 1197931, 794964, 6970, 104, 21, 6, 2, 1, 0, 1 ] }
false
taln-ls2n/kpbiomed
medium
test
20,000
abstract
18
4,682
1,474.05685
1,483
478.92172
{ "bin_edges": [ 18, 485, 952, 1419, 1886, 2353, 2820, 3287, 3754, 4221, 4682 ], "hist": [ 362, 2278, 6230, 7952, 2432, 583, 118, 31, 9, 5 ] }
false
taln-ls2n/kpbiomed
medium
test
20,000
authors
2
4,734
144.68
122
111.95902
{ "bin_edges": [ 2, 476, 950, 1424, 1898, 2372, 2846, 3320, 3794, 4268, 4734 ], "hist": [ 19743, 236, 11, 7, 1, 0, 0, 0, 0, 2 ] }
false
taln-ls2n/kpbiomed
medium
test
20,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 20000 ] }
false
taln-ls2n/kpbiomed
medium
test
20,000
title
6
388
110.71435
108
36.26139
{ "bin_edges": [ 6, 45, 84, 123, 162, 201, 240, 279, 318, 357, 388 ], "hist": [ 379, 4240, 8532, 5143, 1394, 262, 41, 7, 0, 2 ] }
false
taln-ls2n/kpbiomed
medium
validation
20,000
abstract
11
5,791
1,475.03755
1,485
485.84017
{ "bin_edges": [ 11, 590, 1169, 1748, 2327, 2906, 3485, 4064, 4643, 5222, 5791 ], "hist": [ 677, 4468, 9544, 4484, 702, 92, 26, 5, 0, 2 ] }
false
taln-ls2n/kpbiomed
medium
validation
20,000
authors
2
2,556
145.1623
124
102.08738
{ "bin_edges": [ 2, 258, 514, 770, 1026, 1282, 1538, 1794, 2050, 2306, 2556 ], "hist": [ 18057, 1749, 150, 23, 15, 3, 2, 0, 0, 1 ] }
false
taln-ls2n/kpbiomed
medium
validation
20,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 20000 ] }
false
taln-ls2n/kpbiomed
medium
validation
20,000
title
10
987
110.393
108
36.68865
{ "bin_edges": [ 10, 108, 206, 304, 402, 500, 598, 696, 794, 892, 987 ], "hist": [ 9942, 9820, 234, 3, 0, 0, 0, 0, 0, 1 ] }
false
taln-ls2n/kpbiomed
small
train
500,000
abstract
1
13,852
1,471.93579
1,480
481.22005
{ "bin_edges": [ 1, 1387, 2773, 4159, 5545, 6931, 8317, 9703, 11089, 12475, 13852 ], "hist": [ 208359, 287259, 4188, 166, 23, 3, 0, 0, 1, 1 ] }
false
taln-ls2n/kpbiomed
small
train
500,000
authors
2
38,137
144.48746
122
124.676
{ "bin_edges": [ 2, 3816, 7630, 11444, 15258, 19072, 22886, 26700, 30514, 34328, 38137 ], "hist": [ 499987, 8, 2, 0, 1, 1, 0, 0, 0, 1 ] }
false
taln-ls2n/kpbiomed
small
train
500,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 500000 ] }
false
taln-ls2n/kpbiomed
small
train
500,000
title
4
1,159
110.23957
108
36.38714
{ "bin_edges": [ 4, 120, 236, 352, 468, 584, 700, 816, 932, 1048, 1159 ], "hist": [ 315539, 182829, 1600, 20, 9, 1, 1, 0, 0, 1 ] }
false
taln-ls2n/kpbiomed
small
test
20,000
abstract
18
4,682
1,474.05685
1,483
478.92172
{ "bin_edges": [ 18, 485, 952, 1419, 1886, 2353, 2820, 3287, 3754, 4221, 4682 ], "hist": [ 362, 2278, 6230, 7952, 2432, 583, 118, 31, 9, 5 ] }
false
taln-ls2n/kpbiomed
small
test
20,000
authors
2
4,734
144.68
122
111.95902
{ "bin_edges": [ 2, 476, 950, 1424, 1898, 2372, 2846, 3320, 3794, 4268, 4734 ], "hist": [ 19743, 236, 11, 7, 1, 0, 0, 0, 0, 2 ] }
false
taln-ls2n/kpbiomed
small
test
20,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 20000 ] }
false
taln-ls2n/kpbiomed
small
test
20,000
title
6
388
110.71435
108
36.26139
{ "bin_edges": [ 6, 45, 84, 123, 162, 201, 240, 279, 318, 357, 388 ], "hist": [ 379, 4240, 8532, 5143, 1394, 262, 41, 7, 0, 2 ] }
false
taln-ls2n/kpbiomed
small
validation
20,000
abstract
11
5,791
1,475.03755
1,485
485.84017
{ "bin_edges": [ 11, 590, 1169, 1748, 2327, 2906, 3485, 4064, 4643, 5222, 5791 ], "hist": [ 677, 4468, 9544, 4484, 702, 92, 26, 5, 0, 2 ] }
false
taln-ls2n/kpbiomed
small
validation
20,000
authors
2
2,556
145.1623
124
102.08738
{ "bin_edges": [ 2, 258, 514, 770, 1026, 1282, 1538, 1794, 2050, 2306, 2556 ], "hist": [ 18057, 1749, 150, 23, 15, 3, 2, 0, 0, 1 ] }
false
taln-ls2n/kpbiomed
small
validation
20,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 20000 ] }
false
taln-ls2n/kpbiomed
small
validation
20,000
title
10
987
110.393
108
36.68865
{ "bin_edges": [ 10, 108, 206, 304, 402, 500, 598, 696, 794, 892, 987 ], "hist": [ 9942, 9820, 234, 3, 0, 0, 0, 0, 0, 1 ] }
false
SALT-NLP/wikisql_VALUE
AppE
test
15,878
question
17
238
64.94571
61
24.7009
{ "bin_edges": [ 17, 40, 63, 86, 109, 132, 155, 178, 201, 224, 238 ], "hist": [ 1908, 6577, 4584, 1875, 654, 192, 63, 19, 2, 4 ] }
false
SALT-NLP/wikisql_VALUE
AppE
validation
8,421
question
18
284
64.89146
61
24.78027
{ "bin_edges": [ 18, 45, 72, 99, 126, 153, 180, 207, 234, 261, 284 ], "hist": [ 1701, 3974, 1945, 592, 161, 40, 5, 1, 1, 1 ] }
false
AmazonScience/mintaka
all
train
126,000
answerText
1
266
10.91964
9
11.83667
{ "bin_edges": [ 1, 28, 55, 82, 109, 136, 163, 190, 217, 244, 266 ], "hist": [ 119403, 5535, 612, 225, 90, 81, 18, 18, 0, 18 ] }
false
AmazonScience/mintaka
all
train
126,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 126000 ] }
false
AmazonScience/mintaka
all
train
126,000
question
6
246
55.83501
53
22.41131
{ "bin_edges": [ 6, 31, 56, 81, 106, 131, 156, 181, 206, 231, 246 ], "hist": [ 12988, 56539, 40512, 12277, 2810, 649, 176, 39, 6, 4 ] }
false
AmazonScience/mintaka
all
validation
18,000
answerText
1
174
10.7755
9
11.15348
{ "bin_edges": [ 1, 19, 37, 55, 73, 91, 109, 127, 145, 163, 174 ], "hist": [ 15588, 2007, 252, 72, 45, 9, 9, 9, 0, 9 ] }
false
AmazonScience/mintaka
all
validation
18,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 18000 ] }
false
AmazonScience/mintaka
all
validation
18,000
question
10
206
56.24522
53
22.18588
{ "bin_edges": [ 10, 30, 50, 70, 90, 110, 130, 150, 170, 190, 206 ], "hist": [ 1614, 6082, 5991, 2868, 1043, 315, 53, 30, 2, 2 ] }
false
SALT-NLP/wikisql_VALUE
AppE
train
56,355
question
14
268
64.73275
61
24.32358
{ "bin_edges": [ 14, 40, 66, 92, 118, 144, 170, 196, 222, 248, 268 ], "hist": [ 6883, 26010, 16120, 5404, 1465, 360, 78, 26, 8, 1 ] }
false
SALT-NLP/wikisql_VALUE
ChcE
test
15,878
question
15
231
61.23511
57
24.05744
{ "bin_edges": [ 15, 37, 59, 81, 103, 125, 147, 169, 191, 213, 231 ], "hist": [ 1903, 6552, 4524, 1912, 666, 220, 73, 21, 4, 3 ] }
false
SALT-NLP/wikisql_VALUE
ChcE
validation
8,421
question
16
275
61.1691
57
24.14395
{ "bin_edges": [ 16, 42, 68, 94, 120, 146, 172, 198, 224, 250, 275 ], "hist": [ 1801, 3851, 1953, 594, 166, 42, 10, 2, 1, 1 ] }
false
SALT-NLP/wikisql_VALUE
ChcE
train
56,355
question
12
244
60.98593
57
23.65395
{ "bin_edges": [ 12, 36, 60, 84, 108, 132, 156, 180, 204, 228, 244 ], "hist": [ 5960, 24870, 16748, 6229, 1869, 488, 143, 35, 7, 6 ] }
false
SALT-NLP/wikisql_VALUE
CollSgE
test
15,878
question
13
258
54.926
51
22.98608
{ "bin_edges": [ 13, 38, 63, 88, 113, 138, 163, 188, 213, 238, 258 ], "hist": [ 3676, 7386, 3398, 1024, 301, 70, 16, 5, 1, 1 ] }
false
AmazonScience/mintaka
all
test
36,000
answerText
1
345
11.006
9
13.87535
{ "bin_edges": [ 1, 36, 71, 106, 141, 176, 211, 246, 281, 316, 345 ], "hist": [ 34812, 882, 207, 45, 9, 36, 0, 0, 0, 9 ] }
false
AmazonScience/mintaka
all
test
36,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 36000 ] }
false
AmazonScience/mintaka
all
test
36,000
question
8
222
55.49367
53
21.7828
{ "bin_edges": [ 8, 30, 52, 74, 96, 118, 140, 162, 184, 206, 222 ], "hist": [ 3276, 13793, 12488, 4724, 1270, 329, 81, 30, 8, 1 ] }
false
AmazonScience/mintaka
ar
train
14,000
answerText
1
266
10.91964
9
11.83705
{ "bin_edges": [ 1, 28, 55, 82, 109, 136, 163, 190, 217, 244, 266 ], "hist": [ 13267, 615, 68, 25, 10, 9, 2, 2, 0, 2 ] }
false
AmazonScience/mintaka
ar
train
14,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 14000 ] }
false
AmazonScience/mintaka
ar
train
14,000
question
12
226
52.72143
49
20.31038
{ "bin_edges": [ 12, 34, 56, 78, 100, 122, 144, 166, 188, 210, 226 ], "hist": [ 2169, 6508, 3662, 1276, 308, 58, 14, 3, 1, 1 ] }
false
AmazonScience/mintaka
ar
validation
2,000
answerText
1
174
10.7755
9
11.15595
{ "bin_edges": [ 1, 19, 37, 55, 73, 91, 109, 127, 145, 163, 174 ], "hist": [ 1732, 223, 28, 8, 5, 1, 1, 1, 0, 1 ] }
false
AmazonScience/mintaka
ar
validation
2,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 2000 ] }
false
AmazonScience/mintaka
ar
validation
2,000
question
15
168
53.292
50
20.39563
{ "bin_edges": [ 15, 31, 47, 63, 79, 95, 111, 127, 143, 159, 168 ], "hist": [ 205, 656, 574, 330, 151, 63, 12, 6, 1, 2 ] }
false
AmazonScience/mintaka
ar
test
4,000
answerText
1
345
11.006
9
13.87689
{ "bin_edges": [ 1, 36, 71, 106, 141, 176, 211, 246, 281, 316, 345 ], "hist": [ 3868, 98, 23, 5, 1, 4, 0, 0, 0, 1 ] }
false
AmazonScience/mintaka
ar
test
4,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 4000 ] }
false
AmazonScience/mintaka
ar
test
4,000
question
13
182
52.34425
49
19.87988
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 182 ], "hist": [ 373, 1363, 1262, 654, 238, 76, 23, 6, 4, 1 ] }
false
AmazonScience/mintaka
de
train
14,000
answerText
1
266
10.91964
9
11.83705
{ "bin_edges": [ 1, 28, 55, 82, 109, 136, 163, 190, 217, 244, 266 ], "hist": [ 13267, 615, 68, 25, 10, 9, 2, 2, 0, 2 ] }
false
AmazonScience/mintaka
de
train
14,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 14000 ] }
false
AmazonScience/mintaka
de
train
14,000
question
13
246
62.33536
59
22.24717
{ "bin_edges": [ 13, 37, 61, 85, 109, 133, 157, 181, 205, 229, 246 ], "hist": [ 1127, 6324, 4522, 1509, 363, 117, 27, 8, 1, 2 ] }
false
SALT-NLP/wikisql_VALUE
CollSgE
validation
8,421
question
13
243
54.67308
50
23.19449
{ "bin_edges": [ 13, 37, 61, 85, 109, 133, 157, 181, 205, 229, 243 ], "hist": [ 1855, 3800, 1924, 600, 171, 50, 17, 2, 1, 1 ] }
false
AmazonScience/mintaka
de
validation
2,000
answerText
1
174
10.7755
9
11.15595
{ "bin_edges": [ 1, 19, 37, 55, 73, 91, 109, 127, 145, 163, 174 ], "hist": [ 1732, 223, 28, 8, 5, 1, 1, 1, 0, 1 ] }
false
AmazonScience/mintaka
de
validation
2,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 2000 ] }
false
AmazonScience/mintaka
de
validation
2,000
question
19
175
62.753
59
21.70056
{ "bin_edges": [ 19, 35, 51, 67, 83, 99, 115, 131, 147, 163, 175 ], "hist": [ 95, 571, 585, 398, 210, 94, 33, 9, 4, 1 ] }
false
SALT-NLP/wikisql_VALUE
IndE
test
15,878
question
17
242
62.52588
58
25.20189
{ "bin_edges": [ 17, 40, 63, 86, 109, 132, 155, 178, 201, 224, 242 ], "hist": [ 2640, 6557, 4078, 1731, 581, 211, 58, 14, 4, 4 ] }
false
AmazonScience/mintaka
de
test
4,000
answerText
1
345
11.006
9
13.87689
{ "bin_edges": [ 1, 36, 71, 106, 141, 176, 211, 246, 281, 316, 345 ], "hist": [ 3868, 98, 23, 5, 1, 4, 0, 0, 0, 1 ] }
false
AmazonScience/mintaka
de
test
4,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 4000 ] }
false
AmazonScience/mintaka
de
test
4,000
question
17
200
61.863
59
21.4999
{ "bin_edges": [ 17, 36, 55, 74, 93, 112, 131, 150, 169, 188, 200 ], "hist": [ 278, 1382, 1362, 638, 230, 75, 21, 11, 1, 2 ] }
false
AmazonScience/mintaka
en
train
14,000
answerText
1
266
10.91964
9
11.83705
{ "bin_edges": [ 1, 28, 55, 82, 109, 136, 163, 190, 217, 244, 266 ], "hist": [ 13267, 615, 68, 25, 10, 9, 2, 2, 0, 2 ] }
false
AmazonScience/mintaka
en
train
14,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 14000 ] }
false
AmazonScience/mintaka
en
train
14,000
question
15
197
55.916
53
18.8932
{ "bin_edges": [ 15, 34, 53, 72, 91, 110, 129, 148, 167, 186, 197 ], "hist": [ 981, 5974, 4594, 1746, 494, 142, 48, 15, 4, 2 ] }
false
AmazonScience/mintaka
en
validation
2,000
answerText
1
174
10.7755
9
11.15595
{ "bin_edges": [ 1, 19, 37, 55, 73, 91, 109, 127, 145, 163, 174 ], "hist": [ 1732, 223, 28, 8, 5, 1, 1, 1, 0, 1 ] }
false
AmazonScience/mintaka
en
validation
2,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 2000 ] }
false
AmazonScience/mintaka
en
validation
2,000
question
20
157
56.377
53
18.71823
{ "bin_edges": [ 20, 34, 48, 62, 76, 90, 104, 118, 132, 146, 157 ], "hist": [ 147, 592, 589, 366, 195, 72, 26, 9, 1, 3 ] }
false
AmazonScience/mintaka
en
test
4,000
answerText
1
345
11.006
9
13.87689
{ "bin_edges": [ 1, 36, 71, 106, 141, 176, 211, 246, 281, 316, 345 ], "hist": [ 3868, 98, 23, 5, 1, 4, 0, 0, 0, 1 ] }
false
AmazonScience/mintaka
en
test
4,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 4000 ] }
false
AmazonScience/mintaka
en
test
4,000
question
15
184
55.502
53
17.90598
{ "bin_edges": [ 15, 32, 49, 66, 83, 100, 117, 134, 151, 168, 184 ], "hist": [ 190, 1367, 1477, 660, 218, 60, 21, 4, 2, 1 ] }
false
AmazonScience/mintaka
es
train
14,000
answerText
1
266
10.91964
9
11.83705
{ "bin_edges": [ 1, 28, 55, 82, 109, 136, 163, 190, 217, 244, 266 ], "hist": [ 13267, 615, 68, 25, 10, 9, 2, 2, 0, 2 ] }
false
AmazonScience/mintaka
es
train
14,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 14000 ] }
false
AmazonScience/mintaka
es
train
14,000
question
16
239
61.28686
58
21.3408
{ "bin_edges": [ 16, 39, 62, 85, 108, 131, 154, 177, 200, 223, 239 ], "hist": [ 1454, 6564, 4185, 1335, 338, 79, 35, 7, 2, 1 ] }
false
AmazonScience/mintaka
es
validation
2,000
answerText
1
174
10.7755
9
11.15595
{ "bin_edges": [ 1, 19, 37, 55, 73, 91, 109, 127, 145, 163, 174 ], "hist": [ 1732, 223, 28, 8, 5, 1, 1, 1, 0, 1 ] }
false
AmazonScience/mintaka
es
validation
2,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 2000 ] }
false
AmazonScience/mintaka
es
validation
2,000
question
20
168
61.853
58
21.01267
{ "bin_edges": [ 20, 35, 50, 65, 80, 95, 110, 125, 140, 155, 168 ], "hist": [ 107, 508, 620, 402, 216, 82, 45, 13, 6, 1 ] }
false
AmazonScience/mintaka
es
test
4,000
answerText
1
345
11.006
9
13.87689
{ "bin_edges": [ 1, 36, 71, 106, 141, 176, 211, 246, 281, 316, 345 ], "hist": [ 3868, 98, 23, 5, 1, 4, 0, 0, 0, 1 ] }
false
AmazonScience/mintaka
es
test
4,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 4000 ] }
false
AmazonScience/mintaka
es
test
4,000
question
17
181
60.9155
58
20.44281
{ "bin_edges": [ 17, 34, 51, 68, 85, 102, 119, 136, 153, 170, 181 ], "hist": [ 180, 1199, 1343, 797, 321, 103, 35, 13, 4, 5 ] }
false
AmazonScience/mintaka
fr
train
14,000
answerText
1
266
10.91964
9
11.83705
{ "bin_edges": [ 1, 28, 55, 82, 109, 136, 163, 190, 217, 244, 266 ], "hist": [ 13267, 615, 68, 25, 10, 9, 2, 2, 0, 2 ] }
false
AmazonScience/mintaka
fr
train
14,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 14000 ] }
false
AmazonScience/mintaka
fr
train
14,000
question
14
233
66.231
62
22.49741
{ "bin_edges": [ 14, 36, 58, 80, 102, 124, 146, 168, 190, 212, 233 ], "hist": [ 491, 5075, 5257, 2190, 674, 211, 72, 21, 7, 2 ] }
false
AmazonScience/mintaka
fr
validation
2,000
answerText
1
174
10.7755
9
11.15595
{ "bin_edges": [ 1, 19, 37, 55, 73, 91, 109, 127, 145, 163, 174 ], "hist": [ 1732, 223, 28, 8, 5, 1, 1, 1, 0, 1 ] }
false
AmazonScience/mintaka
fr
validation
2,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 2000 ] }
false
AmazonScience/mintaka
fr
validation
2,000
question
19
206
66.506
63
21.97345
{ "bin_edges": [ 19, 38, 57, 76, 95, 114, 133, 152, 171, 190, 206 ], "hist": [ 98, 634, 685, 362, 156, 50, 6, 7, 0, 2 ] }
false
AmazonScience/mintaka
fr
test
4,000
answerText
1
345
11.006
9
13.87689
{ "bin_edges": [ 1, 36, 71, 106, 141, 176, 211, 246, 281, 316, 345 ], "hist": [ 3868, 98, 23, 5, 1, 4, 0, 0, 0, 1 ] }
false
AmazonScience/mintaka
fr
test
4,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 4000 ] }
false
AmazonScience/mintaka
fr
test
4,000
question
18
222
65.81875
63
21.82974
{ "bin_edges": [ 18, 39, 60, 81, 102, 123, 144, 165, 186, 207, 222 ], "hist": [ 208, 1526, 1428, 580, 175, 60, 13, 6, 3, 1 ] }
false
AmazonScience/mintaka
hi
train
14,000
answerText
1
266
10.91964
9
11.83705
{ "bin_edges": [ 1, 28, 55, 82, 109, 136, 163, 190, 217, 244, 266 ], "hist": [ 13267, 615, 68, 25, 10, 9, 2, 2, 0, 2 ] }
false
AmazonScience/mintaka
hi
train
14,000
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 14000 ] }
false