dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
baber/cce-renewals
full
train
445,386
author
4
308
25.53516
18
18.30221
{ "bin_edges": [ 4, 35, 66, 97, 128, 159, 190, 221, 252, 283, 308 ], "hist": [ 202430, 36874, 8739, 2062, 433, 109, 32, 10, 2, 1 ] }
false
baber/cce-renewals
full
train
445,386
claimants
3
449
26.80059
25
13.25926
{ "bin_edges": [ 3, 48, 93, 138, 183, 228, 273, 318, 363, 408, 449 ], "hist": [ 411505, 17916, 1198, 307, 71, 18, 2, 0, 1, 42 ] }
false
baber/cce-renewals
full
train
445,386
entry_id
36
36
36
36
0
{ "bin_edges": [ 36, 36 ], "hist": [ 445386 ] }
false
baber/cce-renewals
full
train
445,386
full_text
68
1,191
163.15529
148
71.71905
{ "bin_edges": [ 68, 181, 294, 407, 520, 633, 746, 859, 972, 1085, 1191 ], "hist": [ 130832, 38279, 5168, 887, 413, 34, 32, 74, 106, 106 ] }
false
baber/cce-renewals
full
train
445,386
id
2
8
7.44548
7
0.57704
{ "bin_edges": [ 2, 3, 4, 5, 6, 7, 8, 8 ], "hist": [ 2, 1, 76, 538, 16956, 210189, 215931 ] }
false
baber/cce-renewals
full
train
445,386
new_matter
3
175
20.11128
19
13.44919
{ "bin_edges": [ 3, 21, 39, 57, 75, 93, 111, 129, 147, 165, 175 ], "hist": [ 4165, 2988, 418, 104, 39, 9, 3, 0, 1, 1 ] }
false
baber/cce-renewals
full
train
445,386
notes
4
246
41.70912
31
29.28512
{ "bin_edges": [ 4, 29, 54, 79, 104, 129, 154, 179, 204, 229, 246 ], "hist": [ 3256, 2235, 1232, 584, 188, 74, 24, 9, 3, 3 ] }
false
baber/cce-renewals
full
train
445,386
odat
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 443672 ] }
false
baber/cce-renewals
full
train
445,386
oreg
2
18
6.88794
7
0.70125
{ "bin_edges": [ 2, 4, 6, 8, 10, 12, 14, 16, 18, 18 ], "hist": [ 68, 11663, 382388, 49417, 129, 0, 1, 0, 1 ] }
false
baber/cce-renewals
full
train
445,386
rdat
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 443693 ] }
false
baber/cce-renewals
full
train
445,386
title
2
444
53.13736
48
31.93309
{ "bin_edges": [ 2, 47, 92, 137, 182, 227, 272, 317, 362, 407, 444 ], "hist": [ 204564, 186850, 31589, 8866, 2067, 430, 98, 16, 42, 9 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=1, 'temperature'=0.55, 'top_k'=10000, 'top_p'=1.0}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=1, 'temperature'=0.55, 'top_k'=10000, 'top_p'=1.0}
train
3,270
prediction
5
489
39.6844
6
62.09945
{ "bin_edges": [ 5, 54, 103, 152, 201, 250, 299, 348, 397, 446, 489 ], "hist": [ 2187, 488, 533, 6, 16, 7, 6, 6, 11, 10 ] }
false
baber/cce-renewals
matched
train
159,647
author
2
269
19.09651
18
7.94064
{ "bin_edges": [ 2, 29, 56, 83, 110, 137, 164, 191, 218, 245, 269 ], "hist": [ 140768, 8654, 667, 126, 42, 86, 3, 1, 0, 1 ] }
false
baber/cce-renewals
matched
train
159,647
author_ren
5
308
26.43582
19
19.2042
{ "bin_edges": [ 5, 36, 67, 98, 129, 160, 191, 222, 253, 284, 308 ], "hist": [ 100106, 19743, 4513, 1200, 308, 76, 20, 8, 2, 1 ] }
false
baber/cce-renewals
matched
train
159,647
claimants
4
368
25.20871
21
13.98204
{ "bin_edges": [ 4, 41, 78, 115, 152, 189, 226, 263, 300, 337, 368 ], "hist": [ 140158, 13643, 1296, 274, 60, 45, 2, 2, 0, 1 ] }
false
baber/cce-renewals
matched
train
159,647
date
8
10
9.99987
10
0.01416
{ "bin_edges": [ 8, 9, 10, 10 ], "hist": [ 6, 8, 159633 ] }
false
baber/cce-renewals
matched
train
159,647
date_ren
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 159647 ] }
false
baber/cce-renewals
matched
train
159,647
edition
2
785
10.02299
7
9.91057
{ "bin_edges": [ 2, 81, 160, 239, 318, 397, 476, 555, 634, 713, 785 ], "hist": [ 28152, 66, 4, 1, 0, 1, 0, 0, 0, 1 ] }
false
baber/cce-renewals
matched
train
159,647
entry_id
36
36
36
36
0
{ "bin_edges": [ 36, 36 ], "hist": [ 159647 ] }
false
baber/cce-renewals
matched
train
159,647
full_text
68
595
148.85197
137
51.29287
{ "bin_edges": [ 68, 121, 174, 227, 280, 333, 386, 439, 492, 545, 595 ], "hist": [ 26390, 29945, 12622, 4471, 1275, 341, 119, 20, 21, 3 ] }
false
baber/cce-renewals
matched
train
159,647
id
36
36
36
36
0
{ "bin_edges": [ 36, 36 ], "hist": [ 159647 ] }
false
baber/cce-renewals
matched
train
159,647
id_ren
2
8
7.41202
7
0.56107
{ "bin_edges": [ 2, 3, 4, 5, 6, 7, 8, 8 ], "hist": [ 1, 1, 10, 139, 5288, 82826, 71382 ] }
false
baber/cce-renewals
matched
train
159,647
new_matter
4
175
20.54638
20
13.88684
{ "bin_edges": [ 4, 22, 40, 58, 76, 94, 112, 130, 148, 166, 175 ], "hist": [ 3015, 1676, 322, 80, 29, 7, 1, 1, 0, 1 ] }
false
baber/cce-renewals
matched
train
159,647
notes
4
246
49.0251
43
31.50255
{ "bin_edges": [ 4, 29, 54, 79, 104, 129, 154, 179, 204, 229, 246 ], "hist": [ 1247, 1238, 825, 395, 147, 61, 17, 8, 3, 3 ] }
false
baber/cce-renewals
matched
train
159,647
page
1
4
3.18367
3
0.6408
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 1596, 16021, 93494, 48536 ] }
false
baber/cce-renewals
matched
train
159,647
prev-regNum
2
134
10.45256
7
12.96693
{ "bin_edges": [ 2, 16, 30, 44, 58, 72, 86, 100, 114, 128, 134 ], "hist": [ 1309, 50, 42, 31, 18, 10, 1, 2, 0, 2 ] }
false
baber/cce-renewals
matched
train
159,647
prevPub
4
354
48.64399
40
24.74532
{ "bin_edges": [ 4, 40, 76, 112, 148, 184, 220, 256, 292, 328, 354 ], "hist": [ 1195, 1037, 234, 51, 6, 4, 0, 0, 0, 1 ] }
false
baber/cce-renewals
matched
train
159,647
pubName
1
274
20.89832
19
9.91536
{ "bin_edges": [ 1, 29, 57, 85, 113, 141, 169, 197, 225, 253, 274 ], "hist": [ 130027, 18587, 1419, 209, 48, 13, 7, 1, 0, 5 ] }
false
baber/cce-renewals
matched
train
159,647
pubPlace
1
530
10.92473
8
5.34047
{ "bin_edges": [ 1, 54, 107, 160, 213, 266, 319, 372, 425, 478, 530 ], "hist": [ 86206, 28, 2, 1, 0, 0, 0, 0, 0, 1 ] }
false
baber/cce-renewals
matched
train
159,647
rdat
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 159647 ] }
false
baber/cce-renewals
matched
train
159,647
regDate
4
79
8.98225
7
3.0091
{ "bin_edges": [ 4, 12, 20, 28, 36, 44, 52, 60, 68, 76, 79 ], "hist": [ 101288, 58296, 46, 11, 3, 1, 1, 0, 0, 1 ] }
false
baber/cce-renewals
matched
train
159,647
regNum
3
224
7.34889
7
4.03019
{ "bin_edges": [ 3, 26, 49, 72, 95, 118, 141, 164, 187, 210, 224 ], "hist": [ 159300, 215, 65, 0, 16, 0, 20, 0, 0, 28 ] }
false
baber/cce-renewals
matched
train
159,647
regnum
3
8
6.67654
7
0.57055
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 8 ], "hist": [ 43, 406, 4234, 44588, 107569, 2807 ] }
false
baber/cce-renewals
matched
train
159,647
title
1
805
40.74009
29
40.00122
{ "bin_edges": [ 1, 82, 163, 244, 325, 406, 487, 568, 649, 730, 805 ], "hist": [ 145077, 11029, 2108, 897, 222, 54, 39, 12, 3, 2 ] }
false
baber/cce-renewals
matched
train
159,647
title_ren
3
444
50.70419
40
37.69678
{ "bin_edges": [ 3, 48, 93, 138, 183, 228, 273, 318, 363, 408, 444 ], "hist": [ 92726, 44258, 14042, 3985, 1350, 263, 64, 10, 4, 8 ] }
false
petrock/train_3d
default
train
1,228
Instruction
43
140
78.61889
74
21.17143
{ "bin_edges": [ 43, 53, 63, 73, 83, 93, 103, 113, 123, 133, 140 ], "hist": [ 63, 248, 270, 215, 148, 114, 71, 34, 33, 32 ] }
false
petrock/train_3d
default
train
1,228
Output
21
4,030
445.32329
295.5
460.30462
{ "bin_edges": [ 21, 422, 823, 1224, 1625, 2026, 2427, 2828, 3229, 3630, 4030 ], "hist": [ 802, 264, 88, 35, 17, 10, 7, 1, 2, 2 ] }
false
petrock/train_3d
default
train
1,228
Text
218
4,253
663.94218
512.5
465.8426
{ "bin_edges": [ 218, 622, 1026, 1430, 1834, 2238, 2642, 3046, 3450, 3854, 4253 ], "hist": [ 785, 273, 92, 38, 18, 10, 7, 1, 2, 2 ] }
false
baber/cce-renewals
unmatched
train
69,318
author
4
210
23.82734
19
15.65607
{ "bin_edges": [ 4, 25, 46, 67, 88, 109, 130, 151, 172, 193, 210 ], "hist": [ 31208, 9167, 2475, 824, 310, 110, 44, 13, 9, 7 ] }
false
baber/cce-renewals
unmatched
train
69,318
claimants
3
272
28.60196
27
12.51613
{ "bin_edges": [ 3, 30, 57, 84, 111, 138, 165, 192, 219, 246, 272 ], "hist": [ 41442, 23200, 1105, 238, 149, 38, 13, 4, 12, 4 ] }
false
baber/cce-renewals
unmatched
train
69,318
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 69318 ] }
false
baber/cce-renewals
unmatched
train
69,318
entry_id
36
36
36
36
0
{ "bin_edges": [ 36, 36 ], "hist": [ 69318 ] }
false
baber/cce-renewals
unmatched
train
69,318
full_text
72
1,191
167.43338
154
66.84454
{ "bin_edges": [ 72, 184, 296, 408, 520, 632, 744, 856, 968, 1080, 1191 ], "hist": [ 36612, 11463, 959, 190, 39, 0, 0, 40, 0, 55 ] }
false
baber/cce-renewals
unmatched
train
69,318
id
4
8
7.1346
7
0.46073
{ "bin_edges": [ 4, 5, 6, 7, 8, 8 ], "hist": [ 1, 39, 3197, 53473, 12608 ] }
false
baber/cce-renewals
unmatched
train
69,318
rdat
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 69318 ] }
false
baber/cce-renewals
unmatched
train
69,318
regnum
3
8
7.36758
7
0.71348
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 8 ], "hist": [ 13, 65, 786, 6708, 27739, 34007 ] }
false
baber/cce-renewals
unmatched
train
69,318
title
2
444
62.69433
55
37.51587
{ "bin_edges": [ 2, 47, 92, 137, 182, 227, 272, 317, 362, 407, 444 ], "hist": [ 26444, 28741, 8076, 3128, 496, 108, 22, 3, 37, 1 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=1, 'temperature'=0.9, 'top_k'=100, 'top_p'=0.05}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=1, 'temperature'=0.9, 'top_k'=100, 'top_p'=0.05}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
HaltiaAI/Her-The-Movie-Samantha-and-Theodore-Dataset
default
train
306
Prompt
3
610
45.38662
29
57.29668
{ "bin_edges": [ 3, 64, 125, 186, 247, 308, 369, 430, 491, 552, 610 ], "hist": [ 217, 38, 6, 4, 2, 1, 0, 0, 0, 1 ] }
false
HaltiaAI/Her-The-Movie-Samantha-and-Theodore-Dataset
default
train
306
Response
3
612
68.57951
39
91.36555
{ "bin_edges": [ 3, 64, 125, 186, 247, 308, 369, 430, 491, 552, 612 ], "hist": [ 194, 47, 19, 11, 4, 3, 1, 0, 1, 3 ] }
false
HaltiaAI/Her-The-Movie-Samantha-and-Theodore-Dataset
default
train
306
end time
4
8
5.90234
5
1.00695
{ "bin_edges": [ 4, 5, 6, 7, 8, 8 ], "hist": [ 2, 136, 4, 113, 1 ] }
false
HaltiaAI/Her-The-Movie-Samantha-and-Theodore-Dataset
default
train
306
start time
5
15
6.01974
5
1.25322
{ "bin_edges": [ 5, 7, 9, 11, 13, 15, 15 ], "hist": [ 163, 136, 0, 4, 0, 1 ] }
false
nataliaElv/oasst_quality_with_suggestions
default
train
5,590
metadata
155
162
160.71145
161
1.08038
{ "bin_edges": [ 155, 156, 157, 158, 159, 160, 161, 162, 162 ], "hist": [ 3, 4, 38, 151, 508, 1368, 2104, 1414 ] }
false
nataliaElv/oasst_quality_with_suggestions
default
train
5,590
prompt
2
9,785
132.98712
72
285.81012
{ "bin_edges": [ 2, 981, 1960, 2939, 3918, 4897, 5876, 6855, 7834, 8813, 9785 ], "hist": [ 5520, 57, 5, 2, 2, 2, 0, 0, 1, 1 ] }
false
nataliaElv/oasst_quality_with_suggestions
default
train
5,590
response
1
9,886
908.41735
699.5
784.40086
{ "bin_edges": [ 1, 990, 1979, 2968, 3957, 4946, 5935, 6924, 7913, 8902, 9886 ], "hist": [ 3401, 1650, 445, 73, 13, 3, 3, 0, 1, 1 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=1, 'temperature'=0.9, 'top_k'=100, 'top_p'=0.1}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=1, 'temperature'=0.9, 'top_k'=100, 'top_p'=0.1}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
losence/train1
default
train
180
input
0
18
7.24444
8
4.26185
{ "bin_edges": [ 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 18 ], "hist": [ 34, 1, 4, 37, 54, 32, 10, 3, 3, 2 ] }
false
losence/train1
default
train
180
instruction
9
23
15.08333
15
3.15498
{ "bin_edges": [ 9, 11, 13, 15, 17, 19, 21, 23, 23 ], "hist": [ 3, 40, 42, 42, 30, 12, 9, 2 ] }
false
losence/train1
default
train
180
output
25
77
48.34444
49
10.02584
{ "bin_edges": [ 25, 31, 37, 43, 49, 55, 61, 67, 73, 77 ], "hist": [ 5, 21, 23, 35, 56, 21, 10, 4, 5 ] }
false
vidhikatkoria/SGD_Movies
default
train
7,219
context
21
500
167.8799
163
63.22964
{ "bin_edges": [ 21, 69, 117, 165, 213, 261, 309, 357, 405, 453, 500 ], "hist": [ 383, 936, 2384, 2057, 875, 389, 138, 42, 9, 6 ] }
false
vidhikatkoria/SGD_Movies
default
train
7,219
response
2
220
48.57764
42
28.38188
{ "bin_edges": [ 2, 24, 46, 68, 90, 112, 134, 156, 178, 200, 220 ], "hist": [ 1347, 2637, 1624, 947, 413, 159, 66, 20, 3, 3 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=1, 'temperature'=0.9, 'top_k'=100, 'top_p'=0.2}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=1, 'temperature'=0.9, 'top_k'=100, 'top_p'=0.2}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
KevinNotSmile/nuscenes-qa-mini
day
train
1,534
question
19
184
71.97327
69
31.32114
{ "bin_edges": [ 19, 36, 53, 70, 87, 104, 121, 138, 155, 172, 184 ], "hist": [ 247, 192, 344, 250, 260, 139, 57, 38, 4, 3 ] }
true
KevinNotSmile/nuscenes-qa-mini
day
train
1,534
token
32
32
32
32
0
{ "bin_edges": [ 32, 32 ], "hist": [ 1534 ] }
true
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=1, 'temperature'=0.9, 'top_k'=1000, 'top_p'=0.05}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=1, 'temperature'=0.9, 'top_k'=1000, 'top_p'=0.05}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
vsarathy/nl-robotics-semantic-parsing-info_structure-2k-novelty-no-context-TEST
default
test
2,473
input
1
119
38.47068
36
19.62709
{ "bin_edges": [ 1, 13, 25, 37, 49, 61, 73, 85, 97, 109, 119 ], "hist": [ 98, 581, 571, 574, 319, 172, 93, 40, 18, 7 ] }
false
vsarathy/nl-robotics-semantic-parsing-info_structure-2k-novelty-no-context-TEST
default
test
2,473
text
603
769
660.28225
664
34.30579
{ "bin_edges": [ 603, 620, 637, 654, 671, 688, 705, 722, 739, 756, 769 ], "hist": [ 348, 487, 241, 323, 510, 321, 152, 60, 27, 4 ] }
false
vsarathy/nl-robotics-semantic-parsing-info_structure-2k-novelty-no-context-TEST
default
test
2,473
utterance
1
119
38.47068
36
19.62709
{ "bin_edges": [ 1, 13, 25, 37, 49, 61, 73, 85, 97, 109, 119 ], "hist": [ 98, 581, 571, 574, 319, 172, 93, 40, 18, 7 ] }
false
Kiwihead15/github-issues
default
train
4,500
body
0
227,963
1,109.89176
429.5
3,993.40912
{ "bin_edges": [ 0, 22797, 45594, 68391, 91188, 113985, 136782, 159579, 182376, 205173, 227963 ], "hist": [ 4291, 4, 0, 0, 0, 0, 0, 0, 0, 1 ] }
false
Kiwihead15/github-issues
default
train
4,500
comments_url
70
70
70
70
0
{ "bin_edges": [ 70, 70 ], "hist": [ 4500 ] }
false
Kiwihead15/github-issues
default
train
4,500
events_url
68
68
68
68
0
{ "bin_edges": [ 68, 68 ], "hist": [ 4500 ] }
false
Kiwihead15/github-issues
default
train
4,500
html_url
49
51
49.89644
49
0.99473
{ "bin_edges": [ 49, 50, 51, 51 ], "hist": [ 2483, 0, 2017 ] }
false
Kiwihead15/github-issues
default
train
4,500
labels_url
75
75
75
75
0
{ "bin_edges": [ 75, 75 ], "hist": [ 4500 ] }
false
Kiwihead15/github-issues
default
train
4,500
node_id
18
32
21.23156
19
4.94155
{ "bin_edges": [ 18, 20, 22, 24, 26, 28, 30, 32, 32 ], "hist": [ 3308, 0, 0, 491, 0, 0, 0, 701 ] }
false
Kiwihead15/github-issues
default
train
4,500
timeline_url
70
70
70
70
0
{ "bin_edges": [ 70, 70 ], "hist": [ 4500 ] }
false
Kiwihead15/github-issues
default
train
4,500
title
1
290
44.49067
41
22.72653
{ "bin_edges": [ 1, 30, 59, 88, 117, 146, 175, 204, 233, 262, 290 ], "hist": [ 1185, 2353, 774, 149, 29, 5, 1, 0, 1, 3 ] }
false
Kiwihead15/github-issues
default
train
4,500
url
61
61
61
61
0
{ "bin_edges": [ 61, 61 ], "hist": [ 4500 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
author
3
130
14.80206
14
5.06751
{ "bin_edges": [ 3, 16, 29, 42, 55, 68, 81, 94, 107, 120, 130 ], "hist": [ 29358, 12797, 827, 168, 48, 3, 2, 1, 1, 1 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
cover_url
0
87
81.09094
86
19.42314
{ "bin_edges": [ 0, 9, 18, 27, 36, 45, 54, 63, 72, 81, 87 ], "hist": [ 2323, 0, 0, 0, 0, 0, 0, 28, 885, 39970 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
date_added
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 43206 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
date_modified
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 43206 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
description
0
27,208
646.51074
509
812.77446
{ "bin_edges": [ 0, 2721, 5442, 8163, 10884, 13605, 16326, 19047, 21768, 24489, 27208 ], "hist": [ 42495, 620, 49, 31, 1, 8, 0, 1, 0, 1 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
edition
0
54
0.84706
0
3.13517
{ "bin_edges": [ 0, 6, 12, 18, 24, 30, 36, 42, 48, 54, 54 ], "hist": [ 40906, 1301, 649, 196, 95, 33, 14, 8, 3, 1 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
md5
32
32
32
32
0
{ "bin_edges": [ 32, 32 ], "hist": [ 19894 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
md5_reported
32
32
32
32
0
{ "bin_edges": [ 32, 32 ], "hist": [ 43206 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
pages
0
5
1.32678
0
1.48914
{ "bin_edges": [ 0, 1, 2, 3, 4, 5, 5 ], "hist": [ 23953, 41, 690, 18185, 336, 1 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
pilimi_torrent
28
38
37.47426
38
1.01073
{ "bin_edges": [ 28, 30, 32, 34, 36, 38, 38 ], "hist": [ 9, 0, 2, 2221, 3309, 14353 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
publisher
0
100
13.02858
12
10.66543
{ "bin_edges": [ 0, 11, 22, 33, 44, 55, 66, 77, 88, 99, 100 ], "hist": [ 20151, 12655, 9121, 1025, 173, 50, 16, 6, 4, 5 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
series
0
123
4.50069
0
12.47275
{ "bin_edges": [ 0, 13, 26, 39, 52, 65, 78, 91, 104, 117, 123 ], "hist": [ 37640, 1807, 2035, 1091, 419, 146, 45, 18, 3, 2 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
title
1
138
26.14841
22
16.81334
{ "bin_edges": [ 1, 15, 29, 43, 57, 71, 85, 99, 113, 127, 138 ], "hist": [ 10669, 17798, 8414, 3713, 1517, 714, 265, 81, 29, 6 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
volume
0
40
0.04821
0
0.66071
{ "bin_edges": [ 0, 5, 10, 15, 20, 25, 30, 35, 40, 40 ], "hist": [ 43057, 120, 16, 2, 5, 0, 5, 0, 1 ] }
false
P1ayer-1/college-texts-annas-archive-v1
default
train
43,206
year
0
4
3.45753
4
1.36796
{ "bin_edges": [ 0, 1, 2, 3, 4, 4 ], "hist": [ 5837, 8, 17, 32, 37312 ] }
false
luistakahashi/ts-classifier-5
default
train
238
prompt
30
5,537
764.15126
240
1,213.2123
{ "bin_edges": [ 30, 581, 1132, 1683, 2234, 2785, 3336, 3887, 4438, 4989, 5537 ], "hist": [ 175, 14, 10, 9, 8, 3, 7, 5, 5, 2 ] }
false
MoL2/chr1
default
train
1,244,971
sequence
3
200
199.98568
200
1.37334
{ "bin_edges": [ 3, 23, 43, 63, 83, 103, 123, 143, 163, 183, 200 ], "hist": [ 15, 20, 18, 23, 17, 21, 18, 11, 19, 1244809 ] }
false
canristiian/drug_rule_sort2
default
train
4,818
input
55
1,957
102.00498
85
71.15894
{ "bin_edges": [ 55, 246, 437, 628, 819, 1010, 1201, 1392, 1583, 1774, 1957 ], "hist": [ 4686, 107, 13, 5, 3, 1, 1, 1, 0, 1 ] }
false