dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
AmazonScience/massive
zh-TW
train
11,514
utt
1
74
9.96613
9
5.28737
{ "bin_edges": [ 1, 9, 17, 25, 33, 41, 49, 57, 65, 73, 74 ], "hist": [ 5172, 5208, 924, 159, 36, 10, 4, 0, 0, 1 ] }
false
AmazonScience/massive
zh-TW
train
11,514
worker_id
1
2
1.88171
2
0.32297
{ "bin_edges": [ 1, 2, 2 ], "hist": [ 1362, 10152 ] }
false
AmazonScience/massive
zh-TW
validation
2,033
annot_utt
1
97
24.16183
23
15.69072
{ "bin_edges": [ 1, 11, 21, 31, 41, 51, 61, 71, 81, 91, 97 ], "hist": [ 512, 348, 610, 256, 164, 87, 40, 9, 5, 2 ] }
false
AmazonScience/massive
zh-TW
validation
2,033
id
2
5
4.36153
4
0.60805
{ "bin_edges": [ 2, 3, 4, 5, 5 ], "hist": [ 13, 102, 1055, 863 ] }
false
AmazonScience/massive
zh-TW
validation
2,033
utt
1
55
9.86129
9
5.09839
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 55, 55 ], "hist": [ 545, 991, 376, 93, 21, 4, 1, 0, 1, 1 ] }
false
AmazonScience/massive
zh-TW
validation
2,033
worker_id
1
2
1.87752
2
0.32792
{ "bin_edges": [ 1, 2, 2 ], "hist": [ 249, 1784 ] }
false
AmazonScience/massive
tr-TR
validation
2,033
annot_utt
3
158
48.51943
45
24.85776
{ "bin_edges": [ 3, 19, 35, 51, 67, 83, 99, 115, 131, 147, 158 ], "hist": [ 191, 426, 598, 403, 208, 121, 51, 26, 5, 4 ] }
false
AmazonScience/massive
tr-TR
validation
2,033
id
2
5
4.36153
4
0.60805
{ "bin_edges": [ 2, 3, 4, 5, 5 ], "hist": [ 13, 102, 1055, 863 ] }
false
AmazonScience/massive
tr-TR
validation
2,033
utt
3
150
34.60551
32
16.95422
{ "bin_edges": [ 3, 18, 33, 48, 63, 78, 93, 108, 123, 138, 150 ], "hist": [ 255, 800, 600, 235, 103, 26, 8, 3, 2, 1 ] }
false
AmazonScience/massive
tr-TR
test
2,974
annot_utt
2
366
48.24143
44
25.52594
{ "bin_edges": [ 2, 39, 76, 113, 150, 187, 224, 261, 298, 335, 366 ], "hist": [ 1158, 1419, 343, 44, 9, 0, 0, 0, 0, 1 ] }
false
AmazonScience/massive
tr-TR
test
2,974
id
1
5
4.34902
4
0.61315
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 3, 16, 155, 1566, 1234 ] }
false
AmazonScience/massive
tr-TR
test
2,974
utt
2
366
34.81439
32
17.78036
{ "bin_edges": [ 2, 39, 76, 113, 150, 187, 224, 261, 298, 335, 366 ], "hist": [ 1972, 932, 64, 5, 0, 0, 0, 0, 0, 1 ] }
false
AmazonScience/massive
ur-PK
test
2,974
annot_utt
1
170
48.28278
44
24.62686
{ "bin_edges": [ 1, 18, 35, 52, 69, 86, 103, 120, 137, 154, 170 ], "hist": [ 223, 698, 919, 602, 281, 146, 68, 25, 9, 3 ] }
false
AmazonScience/massive
ur-PK
test
2,974
id
1
5
4.34902
4
0.61315
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 3, 16, 155, 1566, 1234 ] }
false
AmazonScience/massive
ur-PK
test
2,974
utt
1
137
34.86214
32
16.91644
{ "bin_edges": [ 1, 15, 29, 43, 57, 71, 85, 99, 113, 127, 137 ], "hist": [ 204, 973, 1007, 502, 175, 72, 26, 7, 4, 4 ] }
false
AmazonScience/massive
vi-VN
validation
2,033
annot_utt
1
184
52.36645
49
25.58216
{ "bin_edges": [ 1, 20, 39, 58, 77, 96, 115, 134, 153, 172, 184 ], "hist": [ 155, 484, 661, 397, 205, 86, 34, 7, 2, 2 ] }
false
AmazonScience/massive
vi-VN
validation
2,033
id
2
5
4.36153
4
0.60805
{ "bin_edges": [ 2, 3, 4, 5, 5 ], "hist": [ 13, 102, 1055, 863 ] }
false
AmazonScience/massive
vi-VN
validation
2,033
utt
1
184
38.32317
36
18.04812
{ "bin_edges": [ 1, 20, 39, 58, 77, 96, 115, 134, 153, 172, 184 ], "hist": [ 240, 902, 630, 189, 57, 11, 0, 1, 1, 2 ] }
false
AmazonScience/massive
zh-CN
test
2,974
annot_utt
1
100
25.27202
24
16.22785
{ "bin_edges": [ 1, 11, 21, 31, 41, 51, 61, 71, 81, 91, 100 ], "hist": [ 734, 402, 951, 363, 305, 127, 47, 32, 10, 3 ] }
false
AmazonScience/massive
zh-CN
test
2,974
id
1
5
4.34902
4
0.61315
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 3, 16, 155, 1566, 1234 ] }
false
AmazonScience/massive
zh-CN
test
2,974
utt
1
87
10.42804
10
5.09744
{ "bin_edges": [ 1, 10, 19, 28, 37, 46, 55, 64, 73, 82, 87 ], "hist": [ 1442, 1363, 141, 23, 3, 1, 0, 0, 0, 1 ] }
false
AmazonScience/massive
zh-TW
test
2,974
annot_utt
1
97
23.55716
23
15.68325
{ "bin_edges": [ 1, 11, 21, 31, 41, 51, 61, 71, 81, 91, 97 ], "hist": [ 785, 527, 877, 356, 240, 122, 33, 24, 7, 3 ] }
false
AmazonScience/massive
zh-TW
test
2,974
id
1
5
4.34902
4
0.61315
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 3, 16, 155, 1566, 1234 ] }
false
AmazonScience/massive
zh-TW
test
2,974
utt
1
90
9.769
9
5.33841
{ "bin_edges": [ 1, 10, 19, 28, 37, 46, 55, 64, 73, 82, 90 ], "hist": [ 1643, 1169, 131, 26, 4, 0, 0, 0, 0, 1 ] }
false
AmazonScience/massive
zh-TW
test
2,974
worker_id
1
2
1.88063
2
0.32428
{ "bin_edges": [ 1, 2, 2 ], "hist": [ 355, 2619 ] }
false
AmazonScience/massive
tr-TR
train
11,514
annot_utt
2
206
48.91801
45
25.39822
{ "bin_edges": [ 2, 23, 44, 65, 86, 107, 128, 149, 170, 191, 206 ], "hist": [ 1501, 3992, 3431, 1548, 688, 253, 71, 21, 7, 2 ] }
false
AmazonScience/massive
tr-TR
train
11,514
id
1
5
4.35001
4
0.60954
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 7, 60, 607, 6062, 4778 ] }
false
AmazonScience/massive
tr-TR
train
11,514
utt
2
206
35.03465
32
17.30408
{ "bin_edges": [ 2, 23, 44, 65, 86, 107, 128, 149, 170, 191, 206 ], "hist": [ 2698, 5877, 2241, 538, 120, 31, 6, 2, 0, 1 ] }
false
AmazonScience/massive
ur-PK
validation
2,033
annot_utt
3
175
48.69208
45
24.43986
{ "bin_edges": [ 3, 21, 39, 57, 75, 93, 111, 129, 147, 165, 175 ], "hist": [ 213, 562, 601, 371, 170, 76, 30, 7, 2, 1 ] }
false
AmazonScience/massive
ur-PK
validation
2,033
id
2
5
4.36153
4
0.60805
{ "bin_edges": [ 2, 3, 4, 5, 5 ], "hist": [ 13, 102, 1055, 863 ] }
false
AmazonScience/massive
ur-PK
validation
2,033
utt
3
175
34.65912
32
16.88615
{ "bin_edges": [ 3, 21, 39, 57, 75, 93, 111, 129, 147, 165, 175 ], "hist": [ 374, 978, 480, 150, 41, 5, 2, 2, 0, 1 ] }
false
AmazonScience/massive
vi-VN
test
2,974
annot_utt
1
426
51.36079
48
26.37486
{ "bin_edges": [ 1, 44, 87, 130, 173, 216, 259, 302, 345, 388, 426 ], "hist": [ 1254, 1456, 234, 29, 0, 0, 0, 0, 0, 1 ] }
false
AmazonScience/massive
vi-VN
test
2,974
id
1
5
4.34902
4
0.61315
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 3, 16, 155, 1566, 1234 ] }
false
AmazonScience/massive
vi-VN
test
2,974
utt
1
426
37.8655
35.5
19.16561
{ "bin_edges": [ 1, 44, 87, 130, 173, 216, 259, 302, 345, 388, 426 ], "hist": [ 2029, 895, 46, 3, 0, 0, 0, 0, 0, 1 ] }
false
ekinakyurek/ftrace
abstracts
train
1,560,453
example_uris
14
29,331
66.04817
50
67.77377
{ "bin_edges": [ 14, 2946, 5878, 8810, 11742, 14674, 17606, 20538, 23470, 26402, 29331 ], "hist": [ 1560404, 41, 3, 1, 3, 0, 0, 0, 0, 1 ] }
false
ekinakyurek/ftrace
abstracts
train
1,560,453
facts
11
20,947
131.03391
79
230.81104
{ "bin_edges": [ 11, 2105, 4199, 6293, 8387, 10481, 12575, 14669, 16763, 18857, 20947 ], "hist": [ 1558266, 2011, 55, 9, 7, 13, 45, 46, 0, 1 ] }
false
ekinakyurek/ftrace
abstracts
train
1,560,453
id
1
8
7.20051
7
0.59312
{ "bin_edges": [ 1, 2, 3, 4, 5, 6, 7, 8, 8 ], "hist": [ 7, 16, 239, 1564, 12480, 99801, 1002936, 443410 ] }
false
ekinakyurek/ftrace
abstracts
train
1,560,453
inputs_pretokenized
12
5,151
185.39355
162
134.25277
{ "bin_edges": [ 12, 526, 1040, 1554, 2068, 2582, 3096, 3610, 4124, 4638, 5151 ], "hist": [ 1541489, 15658, 1783, 511, 398, 196, 114, 0, 87, 217 ] }
false
ekinakyurek/ftrace
abstracts
train
1,560,453
masked_uri
2
9
5.08643
5
1.54047
{ "bin_edges": [ 2, 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 337, 254197, 379394, 400405, 169463, 223227, 125971, 7459 ] }
false
ekinakyurek/ftrace
abstracts
train
1,560,453
page_uri
2
9
7.83784
8
0.69952
{ "bin_edges": [ 2, 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 15, 616, 2784, 9843, 43989, 300406, 1023623, 179177 ] }
false
ekinakyurek/ftrace
abstracts
train
1,560,453
targets_pretokenized
14
71
21.75629
20
4.6508
{ "bin_edges": [ 14, 20, 26, 32, 38, 44, 50, 56, 62, 68, 71 ], "hist": [ 567676, 721118, 199923, 54562, 12129, 3528, 1142, 283, 71, 21 ] }
false
ekinakyurek/ftrace
queries
train
31,479
inputs_pretokenized
23
128
46.88205
46
9.18744
{ "bin_edges": [ 23, 34, 45, 56, 67, 78, 89, 100, 111, 122, 128 ], "hist": [ 1354, 12727, 12103, 4383, 760, 118, 28, 4, 1, 1 ] }
false
ekinakyurek/ftrace
queries
train
31,479
obj_surface
1
16
6.7511
6
2.10896
{ "bin_edges": [ 1, 3, 5, 7, 9, 11, 13, 15, 16 ], "hist": [ 20, 3795, 12295, 8643, 5218, 1358, 140, 10 ] }
false
ekinakyurek/ftrace
queries
train
31,479
obj_uri
2
9
4.98497
5
1.27314
{ "bin_edges": [ 2, 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 16, 4348, 6859, 10073, 6032, 3364, 778, 9 ] }
false
ekinakyurek/ftrace
queries
train
31,479
predicate_id
3
5
3.816
4
0.57342
{ "bin_edges": [ 3, 4, 5, 5 ], "hist": [ 8604, 20063, 2812 ] }
false
ekinakyurek/ftrace
queries
train
31,479
sub_surface
1
101
14.94168
14
6.89627
{ "bin_edges": [ 1, 12, 23, 34, 45, 56, 67, 78, 89, 100, 101 ], "hist": [ 9754, 18064, 2955, 586, 92, 23, 2, 2, 0, 1 ] }
false
ekinakyurek/ftrace
queries
train
31,479
sub_uri
3
9
7.2454
7
1.00161
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 166, 658, 1201, 2717, 12202, 13588, 947 ] }
false
ekinakyurek/ftrace
queries
train
31,479
targets_pretokenized
14
29
19.7511
19
2.10896
{ "bin_edges": [ 14, 16, 18, 20, 22, 24, 26, 28, 29 ], "hist": [ 20, 3795, 12295, 8643, 5218, 1358, 140, 10 ] }
false
ekinakyurek/ftrace
queries
train
31,479
uuid
36
36
36
36
0
{ "bin_edges": [ 36, 36 ], "hist": [ 31479 ] }
false
cpllab/syntaxgym
all-2020
test
799
suite_name
4
21
12.35544
12
4.42638
{ "bin_edges": [ 4, 6, 8, 10, 12, 14, 16, 18, 20, 21 ], "hist": [ 68, 48, 52, 205, 175, 40, 66, 99, 46 ] }
false
juliensimon/amazon-shoe-reviews
default
test
10,000
text
0
5,039
178.9897
106
234.45338
{ "bin_edges": [ 0, 504, 1008, 1512, 2016, 2520, 3024, 3528, 4032, 4536, 5039 ], "hist": [ 9299, 574, 90, 24, 6, 2, 2, 1, 1, 1 ] }
false
enoriega/odinsynth_dataset
default
validation
177,553
child
3
82
21.48286
19
13.90748
{ "bin_edges": [ 3, 11, 19, 27, 35, 43, 51, 59, 67, 75, 82 ], "hist": [ 43802, 44949, 32660, 23943, 15481, 9525, 5011, 1726, 432, 24 ] }
false
enoriega/odinsynth_dataset
default
validation
177,553
negative_child
0
92
19.40722
16
14.91759
{ "bin_edges": [ 0, 10, 20, 30, 40, 50, 60, 70, 80, 90, 92 ], "hist": [ 53367, 47967, 33551, 22209, 12776, 5906, 1535, 218, 23, 1 ] }
false
enoriega/odinsynth_dataset
default
validation
177,553
parent
1
80
18.99806
16
13.79597
{ "bin_edges": [ 1, 9, 17, 25, 33, 41, 49, 57, 65, 73, 80 ], "hist": [ 46445, 44978, 31800, 23194, 15321, 9213, 4700, 1579, 299, 24 ] }
false
juliensimon/amazon-shoe-reviews
default
train
90,000
text
0
9,231
174.72597
104
228.9613
{ "bin_edges": [ 0, 924, 1848, 2772, 3696, 4620, 5544, 6468, 7392, 8316, 9231 ], "hist": [ 88715, 1134, 106, 25, 8, 9, 1, 1, 0, 1 ] }
false
taln-ls2n/pubmed
raw
test
1,320
id
16
65
28.51364
27
5.87059
{ "bin_edges": [ 16, 21, 26, 31, 36, 41, 46, 51, 56, 61, 65 ], "hist": [ 28, 398, 514, 246, 77, 23, 30, 1, 0, 3 ] }
false
taln-ls2n/pubmed
raw
test
1,320
text
894
135,968
29,355.03333
27,088.5
15,245.81938
{ "bin_edges": [ 894, 14402, 27910, 41418, 54926, 68434, 81942, 95450, 108958, 122466, 135968 ], "hist": [ 172, 518, 406, 156, 42, 13, 5, 2, 4, 2 ] }
false
taln-ls2n/pubmed
raw
test
1,320
title
19
278
99.575
98
36.44403
{ "bin_edges": [ 19, 45, 71, 97, 123, 149, 175, 201, 227, 253, 278 ], "hist": [ 76, 220, 343, 335, 215, 100, 24, 6, 0, 1 ] }
false
meetyildiz/toqad
plain_text
train
8,308
context
31
11,089
1,028.41899
638
1,203.25886
{ "bin_edges": [ 31, 1137, 2243, 3349, 4455, 5561, 6667, 7773, 8879, 9985, 11089 ], "hist": [ 6263, 1131, 464, 330, 13, 37, 7, 34, 0, 29 ] }
false
meetyildiz/toqad
plain_text
train
8,308
id
1
4
3.86844
4
0.37741
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 9, 90, 886, 7323 ] }
false
meetyildiz/toqad
plain_text
train
8,308
question
2
257
61.79754
58
24.74813
{ "bin_edges": [ 2, 28, 54, 80, 106, 132, 158, 184, 210, 236, 257 ], "hist": [ 202, 3322, 3163, 1179, 308, 89, 30, 10, 2, 3 ] }
false
meetyildiz/toqad
plain_text
train
8,308
title
4
104
15.9988
13
8.74069
{ "bin_edges": [ 4, 15, 26, 37, 48, 59, 70, 81, 92, 103, 104 ], "hist": [ 4951, 2376, 699, 181, 55, 44, 0, 0, 0, 2 ] }
false
meetyildiz/toqad
plain_text
validation
892
context
14
10,521
1,189.23991
632
1,869.13443
{ "bin_edges": [ 14, 1065, 2116, 3167, 4218, 5269, 6320, 7371, 8422, 9473, 10521 ], "hist": [ 691, 124, 5, 0, 0, 0, 64, 0, 0, 8 ] }
false
meetyildiz/toqad
plain_text
validation
892
id
3
4
3.9843
4
0.12436
{ "bin_edges": [ 3, 4, 4 ], "hist": [ 14, 878 ] }
false
meetyildiz/toqad
plain_text
validation
892
question
14
199
64.43386
60
26.99422
{ "bin_edges": [ 14, 33, 52, 71, 90, 109, 128, 147, 166, 185, 199 ], "hist": [ 69, 255, 251, 175, 92, 23, 12, 10, 3, 2 ] }
false
meetyildiz/toqad
plain_text
validation
892
title
6
37
15.713
12
8.82785
{ "bin_edges": [ 6, 10, 14, 18, 22, 26, 30, 34, 37 ], "hist": [ 319, 173, 55, 206, 7, 7, 51, 74 ] }
false
enoriega/odinsynth_dataset
default
train
9,579,809
child
3
94
21.3174
18
13.91664
{ "bin_edges": [ 3, 13, 23, 33, 43, 53, 63, 73, 83, 93, 94 ], "hist": [ 3081041, 2733732, 1803154, 1068190, 584568, 245566, 57008, 6280, 252, 18 ] }
false
enoriega/odinsynth_dataset
default
train
9,579,809
negative_child
0
100
19.27044
16
14.87286
{ "bin_edges": [ 0, 11, 22, 33, 44, 55, 66, 77, 88, 99, 100 ], "hist": [ 3054547, 2894400, 1819723, 1046235, 537235, 193189, 31515, 2788, 173, 4 ] }
false
enoriega/odinsynth_dataset
default
train
9,579,809
parent
1
86
18.85486
16
13.78206
{ "bin_edges": [ 1, 10, 19, 28, 37, 46, 55, 64, 73, 82, 86 ], "hist": [ 3007448, 2473743, 1770014, 1135321, 676323, 358191, 131217, 25021, 2459, 72 ] }
false
enoriega/odinsynth_dataset
default
test
1,072,111
child
3
86
21.29195
18
13.81765
{ "bin_edges": [ 3, 12, 21, 30, 39, 48, 57, 66, 75, 84, 86 ], "hist": [ 316928, 280216, 203946, 132205, 78650, 41831, 15285, 2675, 336, 39 ] }
false
enoriega/odinsynth_dataset
default
test
1,072,111
negative_child
0
93
19.24878
16
14.76913
{ "bin_edges": [ 0, 10, 20, 30, 40, 50, 60, 70, 80, 90, 93 ], "hist": [ 319718, 296550, 206936, 130091, 73133, 34889, 9579, 1085, 129, 1 ] }
false
enoriega/odinsynth_dataset
default
test
1,072,111
parent
1
84
18.83027
16
13.69487
{ "bin_edges": [ 1, 10, 19, 28, 37, 46, 55, 64, 73, 82, 84 ], "hist": [ 336762, 276098, 198157, 128632, 75783, 40365, 13958, 2080, 239, 37 ] }
false
meetyildiz/toqad
plain_text
test
1,190
context
140
3,266
809.37227
724
402.19234
{ "bin_edges": [ 140, 453, 766, 1079, 1392, 1705, 2018, 2331, 2644, 2957, 3266 ], "hist": [ 97, 544, 342, 152, 29, 5, 5, 0, 6, 10 ] }
false
meetyildiz/toqad
plain_text
test
1,190
id
24
24
24
24
0
{ "bin_edges": [ 24, 24 ], "hist": [ 1190 ] }
false
meetyildiz/toqad
plain_text
test
1,190
question
13
164
60.80504
58
21.79507
{ "bin_edges": [ 13, 29, 45, 61, 77, 93, 109, 125, 141, 157, 164 ], "hist": [ 42, 248, 349, 291, 167, 58, 19, 14, 1, 1 ] }
false
meetyildiz/toqad
plain_text
test
1,190
title
5
41
14.68824
13
6.79839
{ "bin_edges": [ 5, 9, 13, 17, 21, 25, 29, 33, 37, 41, 41 ], "hist": [ 243, 275, 208, 299, 82, 25, 37, 0, 0, 21 ] }
false
Evelyn18/becasv2
v1.1.0
train
64
context
153
2,494
599.375
371.5
547.004
{ "bin_edges": [ 153, 388, 623, 858, 1093, 1328, 1563, 1798, 2033, 2268, 2494 ], "hist": [ 36, 10, 4, 2, 4, 0, 6, 0, 0, 2 ] }
false
Evelyn18/becasv2
v1.1.0
train
64
id
1
2
1.85938
2
0.35038
{ "bin_edges": [ 1, 2, 2 ], "hist": [ 9, 55 ] }
false
Evelyn18/becasv2
v1.1.0
train
64
question
13
94
49.59375
46
19.61917
{ "bin_edges": [ 13, 22, 31, 40, 49, 58, 67, 76, 85, 94, 94 ], "hist": [ 3, 7, 12, 12, 11, 5, 7, 3, 3, 1 ] }
false
sileod/movie_recommendation
default
test
500
answer_0
2
51
16.496
15
8.71193
{ "bin_edges": [ 2, 7, 12, 17, 22, 27, 32, 37, 42, 47, 51 ], "hist": [ 46, 102, 153, 73, 65, 27, 21, 5, 5, 3 ] }
false
sileod/movie_recommendation
default
test
500
answer_1
2
82
16.854
15
10.12867
{ "bin_edges": [ 2, 11, 20, 29, 38, 47, 56, 65, 74, 82 ], "hist": [ 135, 228, 83, 30, 17, 4, 1, 1, 1 ] }
false
sileod/movie_recommendation
default
test
500
answer_2
2
62
16.178
14
8.93005
{ "bin_edges": [ 2, 9, 16, 23, 30, 37, 44, 51, 58, 62 ], "hist": [ 68, 224, 119, 49, 19, 11, 9, 0, 1 ] }
false
sileod/movie_recommendation
default
test
500
answer_3
2
82
16.152
14
9.03683
{ "bin_edges": [ 2, 11, 20, 29, 38, 47, 56, 65, 74, 82 ], "hist": [ 137, 233, 83, 36, 7, 2, 0, 1, 1 ] }
false
sileod/movie_recommendation
default
test
500
question
42
151
74.832
69
21.66357
{ "bin_edges": [ 42, 53, 64, 75, 86, 97, 108, 119, 130, 141, 151 ], "hist": [ 46, 120, 153, 64, 44, 20, 25, 14, 8, 6 ] }
false
sileod/discourse_marker_qa
default
test
857
answer_0
16
221
101.08635
97
44.4751
{ "bin_edges": [ 16, 37, 58, 79, 100, 121, 142, 163, 184, 205, 221 ], "hist": [ 55, 93, 141, 161, 131, 107, 78, 55, 28, 8 ] }
false
sileod/discourse_marker_qa
default
test
857
answer_1
11
223
101.06534
97
44.55893
{ "bin_edges": [ 11, 33, 55, 77, 99, 121, 143, 165, 187, 209, 223 ], "hist": [ 40, 100, 139, 161, 136, 107, 93, 48, 27, 6 ] }
false
sileod/discourse_marker_qa
default
test
857
answer_2
9
218
100.80397
96
44.38135
{ "bin_edges": [ 9, 30, 51, 72, 93, 114, 135, 156, 177, 198, 218 ], "hist": [ 25, 88, 137, 155, 133, 119, 86, 67, 30, 17 ] }
false
sileod/discourse_marker_qa
default
test
857
answer_3
15
220
100.93349
97
44.50952
{ "bin_edges": [ 15, 36, 57, 78, 99, 120, 141, 162, 183, 204, 220 ], "hist": [ 49, 98, 138, 157, 131, 102, 85, 61, 27, 9 ] }
false
sileod/discourse_marker_qa
default
test
857
answer_4
12
215
101.11085
96
44.4677
{ "bin_edges": [ 12, 33, 54, 75, 96, 117, 138, 159, 180, 201, 215 ], "hist": [ 37, 93, 137, 150, 138, 111, 92, 56, 31, 12 ] }
false
sileod/discourse_marker_qa
default
test
857
answer_5
12
220
101.17736
96
44.39091
{ "bin_edges": [ 12, 33, 54, 75, 96, 117, 138, 159, 180, 201, 220 ], "hist": [ 39, 85, 143, 154, 127, 119, 85, 64, 30, 11 ] }
false
sileod/discourse_marker_qa
default
test
857
answer_6
11
228
101.05951
96
44.71621
{ "bin_edges": [ 11, 33, 55, 77, 99, 121, 143, 165, 187, 209, 228 ], "hist": [ 36, 104, 143, 162, 137, 103, 89, 52, 26, 5 ] }
false
sileod/discourse_marker_qa
default
test
857
answer_7
10
219
100.98716
96
44.56994
{ "bin_edges": [ 10, 31, 52, 73, 94, 115, 136, 157, 178, 199, 219 ], "hist": [ 30, 90, 131, 156, 141, 106, 92, 66, 29, 16 ] }
false
sileod/discourse_marker_qa
default
test
857
answer_8
11
221
100.89382
97
44.56549
{ "bin_edges": [ 11, 33, 55, 77, 99, 121, 143, 165, 187, 209, 221 ], "hist": [ 38, 96, 145, 165, 134, 110, 85, 50, 29, 5 ] }
false
sileod/discourse_marker_qa
default
test
857
answer_9
12
222
101.14936
96
44.54782
{ "bin_edges": [ 12, 34, 56, 78, 100, 122, 144, 166, 188, 210, 222 ], "hist": [ 40, 99, 148, 159, 137, 102, 93, 46, 26, 7 ] }
false
sileod/discourse_marker_qa
default
test
857
context
10
248
103.74796
101
46.6181
{ "bin_edges": [ 10, 34, 58, 82, 106, 130, 154, 178, 202, 226, 248 ], "hist": [ 58, 99, 130, 179, 131, 123, 83, 35, 16, 3 ] }
false
Divyanshu/indicxnli
ta
train
392,702
hypothesis
3
875
65.20997
61
29.70476
{ "bin_edges": [ 3, 91, 179, 267, 355, 443, 531, 619, 707, 795, 875 ], "hist": [ 326375, 64575, 1595, 125, 14, 7, 8, 2, 0, 1 ] }
false
Divyanshu/indicxnli
ta
train
392,702
premise
4
947
121.68004
106
81.30854
{ "bin_edges": [ 4, 99, 194, 289, 384, 479, 574, 669, 764, 859, 947 ], "hist": [ 180451, 147543, 50182, 10608, 2631, 766, 288, 146, 48, 39 ] }
false
Divyanshu/indicxnli
ta
test
5,010
hypothesis
6
241
61.5002
58
25.05292
{ "bin_edges": [ 6, 30, 54, 78, 102, 126, 150, 174, 198, 222, 241 ], "hist": [ 287, 1874, 1718, 791, 255, 59, 13, 8, 4, 1 ] }
false
Divyanshu/indicxnli
ta
test
5,010
premise
6
795
117.50419
110.5
59.98111
{ "bin_edges": [ 6, 85, 164, 243, 322, 401, 480, 559, 638, 717, 795 ], "hist": [ 1692, 2238, 972, 96, 9, 0, 0, 0, 0, 3 ] }
false
wdc/products-2017
computers_large
train
26,687
brand_left
4
59
12.29041
12
4.92675
{ "bin_edges": [ 4, 10, 16, 22, 28, 34, 40, 46, 52, 58, 59 ], "hist": [ 4124, 8554, 526, 259, 146, 24, 0, 38, 4, 9 ] }
false