dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
atmallen/quirky_sentiment_bob_easy
default
train
10,494
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 10494 ] }
false
atmallen/quirky_sentiment_bob_easy
default
train
10,494
statement
186
1,100
481.17972
423.5
229.79181
{ "bin_edges": [ 186, 278, 370, 462, 554, 646, 738, 830, 922, 1014, 1100 ], "hist": [ 2315, 1915, 1645, 1263, 895, 753, 612, 431, 396, 269 ] }
false
atmallen/quirky_sentiment_bob_easy
default
validation
948
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 948 ] }
false
atmallen/quirky_sentiment_bob_easy
default
validation
948
statement
187
1,099
496.74367
442.5
237.91075
{ "bin_edges": [ 187, 279, 371, 463, 555, 647, 739, 831, 923, 1015, 1099 ], "hist": [ 198, 152, 160, 112, 77, 85, 56, 36, 34, 38 ] }
false
wesley7137/hypothesis_reasoning
default
train
6,733
instruction
10
329
85.89811
72
45.59354
{ "bin_edges": [ 10, 42, 74, 106, 138, 170, 202, 234, 266, 298, 329 ], "hist": [ 563, 2930, 1441, 1026, 326, 246, 126, 35, 26, 14 ] }
false
wesley7137/hypothesis_reasoning
default
train
6,733
output
13
334
84.16932
66
61.1254
{ "bin_edges": [ 13, 46, 79, 112, 145, 178, 211, 244, 277, 310, 334 ], "hist": [ 2093, 1894, 1244, 566, 228, 264, 231, 137, 68, 8 ] }
false
multiIR/testing_ko
default
train
49,504
date_download
25
25
25
25
0
{ "bin_edges": [ 25, 25 ], "hist": [ 49504 ] }
false
multiIR/testing_ko
default
train
49,504
date_publish
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 47009 ] }
false
multiIR/testing_ko
default
train
49,504
description
1
486
139.78605
118
67.33148
{ "bin_edges": [ 1, 50, 99, 148, 197, 246, 295, 344, 393, 442, 486 ], "hist": [ 1853, 5985, 27013, 4407, 515, 2768, 3390, 1, 0, 3 ] }
false
multiIR/testing_ko
default
train
49,504
filename
50
333
78.24818
78
11.13172
{ "bin_edges": [ 50, 79, 108, 137, 166, 195, 224, 253, 282, 311, 333 ], "hist": [ 26511, 22772, 98, 2, 119, 0, 1, 0, 0, 1 ] }
false
multiIR/testing_ko
default
train
49,504
image_url
29
130
67.01988
64
12.75911
{ "bin_edges": [ 29, 40, 51, 62, 73, 84, 95, 106, 117, 128, 130 ], "hist": [ 5, 4017, 18656, 6525, 13666, 5540, 979, 96, 13, 7 ] }
false
multiIR/testing_ko
default
train
49,504
maintext
60
41,311
994.34387
787
968.15744
{ "bin_edges": [ 60, 4186, 8312, 12438, 16564, 20690, 24816, 28942, 33068, 37194, 41311 ], "hist": [ 47909, 400, 27, 14, 1, 2, 0, 3, 1, 4 ] }
false
multiIR/testing_ko
default
train
49,504
path
80
94
85.79577
85
2.27178
{ "bin_edges": [ 80, 82, 84, 86, 88, 90, 92, 94, 94 ], "hist": [ 1205, 106, 28213, 5679, 12831, 1318, 0, 152 ] }
false
multiIR/testing_ko
default
train
49,504
title
4
110
30.36136
30
8.49997
{ "bin_edges": [ 4, 15, 26, 37, 48, 59, 70, 81, 92, 103, 110 ], "hist": [ 893, 13381, 24676, 9049, 1173, 215, 29, 9, 1, 1 ] }
false
multiIR/testing_ko
default
train
49,504
url
35
204
59.17776
59
10.34722
{ "bin_edges": [ 35, 52, 69, 86, 103, 120, 137, 154, 171, 188, 204 ], "hist": [ 12169, 25726, 11453, 31, 2, 1, 120, 0, 0, 2 ] }
false
atmallen/quirky_squaring_increment0_alice
default
test
4,000
statement
18
29
27.02125
27
1.41078
{ "bin_edges": [ 18, 20, 22, 24, 26, 28, 29 ], "hist": [ 7, 14, 55, 669, 1269, 1986 ] }
false
aligner/aligner-20K
default
train
20,004
answer
0
7,054
471.87857
325
509.84539
{ "bin_edges": [ 0, 706, 1412, 2118, 2824, 3530, 4236, 4942, 5648, 6354, 7054 ], "hist": [ 16011, 3012, 633, 213, 78, 29, 18, 4, 5, 1 ] }
false
aligner/aligner-20K
default
train
20,004
correction
2
4,229
461.26285
378
325.61863
{ "bin_edges": [ 2, 425, 848, 1271, 1694, 2117, 2540, 2963, 3386, 3809, 4229 ], "hist": [ 11435, 6271, 1772, 360, 115, 36, 9, 3, 2, 1 ] }
false
aligner/aligner-20K
default
train
20,004
question
3
1,029
67.03424
52
50.60424
{ "bin_edges": [ 3, 106, 209, 312, 415, 518, 621, 724, 827, 930, 1029 ], "hist": [ 17155, 2446, 308, 62, 14, 3, 12, 1, 0, 3 ] }
false
atmallen/quirky_squaring_increment0_alice
default
train
92,000
statement
15
29
26.9948
27
1.44388
{ "bin_edges": [ 15, 17, 19, 21, 23, 25, 27, 29, 29 ], "hist": [ 13, 45, 123, 1633, 3976, 12653, 72446, 1111 ] }
false
atmallen/quirky_squaring_increment0_alice
default
validation
4,000
statement
16
29
27.023
27
1.4246
{ "bin_edges": [ 16, 18, 20, 22, 24, 26, 28, 29 ], "hist": [ 1, 5, 23, 43, 671, 1261, 1996 ] }
false
atmallen/quirky_hemisphere_alice_hard
default
train
938
statement
55
81
61.84542
61
3.45185
{ "bin_edges": [ 55, 58, 61, 64, 67, 70, 73, 76, 79, 81 ], "hist": [ 35, 345, 327, 145, 56, 20, 3, 4, 3 ] }
false
atmallen/quirky_hemisphere_alice_hard
default
validation
525
statement
55
85
61.88952
61
3.84072
{ "bin_edges": [ 55, 59, 63, 67, 71, 75, 79, 83, 85 ], "hist": [ 74, 280, 113, 38, 16, 2, 1, 1 ] }
false
atmallen/quirky_squaring_increment0_alice_easy
default
train
23,000
statement
15
27
24.97417
25
1.45472
{ "bin_edges": [ 15, 17, 19, 21, 23, 25, 27, 27 ], "hist": [ 13, 45, 123, 1633, 3976, 12478, 4732 ] }
false
atmallen/quirky_squaring_increment0_alice_easy
default
test
985
statement
18
27
25.03858
25
1.44778
{ "bin_edges": [ 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 27 ], "hist": [ 2, 5, 1, 13, 51, 4, 165, 503, 17, 224 ] }
false
atmallen/quirky_squaring_increment0_alice_easy
default
validation
970
statement
16
27
24.99381
25
1.45662
{ "bin_edges": [ 16, 18, 20, 22, 24, 26, 27 ], "hist": [ 1, 5, 23, 43, 671, 227 ] }
false
jahb57/bert_embeddings_BATCH_11
default
train
27,500
sentence
5
1,362
72.22676
57
53.56792
{ "bin_edges": [ 5, 141, 277, 413, 549, 685, 821, 957, 1093, 1229, 1362 ], "hist": [ 25018, 2244, 192, 34, 5, 1, 4, 0, 1, 1 ] }
true
blackhc/WizardLM_evol_instruct_70k_scored_subset
default
train
1,024
explanation
88
977
432.13672
420
154.21122
{ "bin_edges": [ 88, 177, 266, 355, 444, 533, 622, 711, 800, 889, 977 ], "hist": [ 31, 114, 188, 243, 192, 142, 68, 28, 10, 8 ] }
false
blackhc/WizardLM_evol_instruct_70k_scored_subset
default
train
1,024
instruction
22
3,424
496.26172
395.5
457.97398
{ "bin_edges": [ 22, 363, 704, 1045, 1386, 1727, 2068, 2409, 2750, 3091, 3424 ], "hist": [ 467, 324, 138, 54, 18, 11, 1, 2, 4, 5 ] }
false
blackhc/WizardLM_evol_instruct_70k_scored_subset
default
train
1,024
output
4
8,876
1,371.42578
1,178.5
1,158.95358
{ "bin_edges": [ 4, 892, 1780, 2668, 3556, 4444, 5332, 6220, 7108, 7996, 8876 ], "hist": [ 403, 327, 183, 60, 24, 15, 7, 3, 1, 1 ] }
false
nateraw/meta-vsc-2022
default
train_query
3,084
video_id
7
7
7
7
0
{ "bin_edges": [ 7, 7 ], "hist": [ 3084 ] }
true
tomekkorbak/test
default
test
3,294
text
1
1,321,344
16,643.92016
3,308
70,300.4597
{ "bin_edges": [ 1, 132136, 264271, 396406, 528541, 660676, 792811, 924946, 1057081, 1189216, 1321344 ], "hist": [ 3243, 16, 9, 7, 7, 5, 1, 2, 3, 1 ] }
false
bzantium/LongPerplexity
arxiv
test
1,643
text
27,887
483,199
60,086.21181
53,778
23,626.35702
{ "bin_edges": [ 27887, 73419, 118951, 164483, 210015, 255547, 301079, 346611, 392143, 437675, 483199 ], "hist": [ 1343, 263, 29, 6, 0, 1, 0, 0, 0, 1 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
test
23,168
answerEntity
2
10
6.75121
7
1.61584
{ "bin_edges": [ 2, 3, 4, 5, 6, 7, 8, 9, 10, 10 ], "hist": [ 48, 1059, 1406, 2190, 3720, 7120, 4622, 2601, 402 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
test
23,168
graph
78
4,217
617.0824
516
335.06725
{ "bin_edges": [ 78, 492, 906, 1320, 1734, 2148, 2562, 2976, 3390, 3804, 4217 ], "hist": [ 11047, 8358, 2759, 754, 169, 48, 19, 6, 4, 4 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
test
23,168
groundTruthAnswerEntity
3
108
7.83136
7
8.21193
{ "bin_edges": [ 3, 14, 25, 36, 47, 58, 69, 80, 91, 102, 108 ], "hist": [ 22194, 384, 217, 148, 58, 16, 77, 0, 37, 37 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
test
23,168
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 23168 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
test
23,168
question
15
184
58.42382
56
19.15909
{ "bin_edges": [ 15, 32, 49, 66, 83, 100, 117, 134, 151, 168, 184 ], "hist": [ 927, 6837, 8386, 4499, 1735, 576, 145, 31, 21, 11 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
test
23,168
questionEntity
2
59
9.60521
7
5.78199
{ "bin_edges": [ 2, 8, 14, 20, 26, 32, 38, 44, 50, 56, 59 ], "hist": [ 12177, 4604, 5163, 887, 243, 49, 10, 0, 10, 11 ] }
false
bzantium/LongPerplexity
github
test
1,797
text
16,299
895,755
76,549.90095
56,173
66,456.74543
{ "bin_edges": [ 16299, 104245, 192191, 280137, 368083, 456029, 543975, 631921, 719867, 807813, 895755 ], "hist": [ 1514, 199, 44, 26, 5, 3, 1, 2, 2, 1 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
train
91,831
answerEntity
2
10
6.78048
7
1.60982
{ "bin_edges": [ 2, 3, 4, 5, 6, 7, 8, 9, 10, 10 ], "hist": [ 198, 4031, 5347, 8648, 14246, 28864, 18145, 10663, 1689 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
train
91,831
graph
78
5,333
615.20705
514
335.3535
{ "bin_edges": [ 78, 604, 1130, 1656, 2182, 2708, 3234, 3760, 4286, 4812, 5333 ], "hist": [ 58205, 26301, 5949, 1120, 189, 41, 13, 8, 2, 3 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
train
91,831
groundTruthAnswerEntity
3
125
7.5122
7
6.15933
{ "bin_edges": [ 3, 16, 29, 42, 55, 68, 81, 94, 107, 120, 125 ], "hist": [ 88653, 1886, 636, 226, 169, 128, 83, 33, 0, 17 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
train
91,831
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 91831 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
train
91,831
question
15
197
59.15815
56
20.28201
{ "bin_edges": [ 15, 34, 53, 72, 91, 110, 129, 148, 167, 186, 197 ], "hist": [ 5241, 34059, 31374, 14497, 4682, 1321, 442, 160, 39, 16 ] }
false
hle2000/Mintaka_Subgraphs_T5_large_ssm
default
train
91,831
questionEntity
2
52
9.58905
7
5.69021
{ "bin_edges": [ 2, 8, 14, 20, 26, 32, 38, 44, 50, 52 ], "hist": [ 48060, 19385, 19400, 3370, 1081, 278, 49, 16, 8 ] }
false
bzantium/LongPerplexity
c4
test
374
text
29,537
161,929
61,477.22727
56,199.5
19,386.78163
{ "bin_edges": [ 29537, 42777, 56017, 69257, 82497, 95737, 108977, 122217, 135457, 148697, 161929 ], "hist": [ 26, 158, 108, 33, 25, 11, 7, 3, 1, 2 ] }
false
nateraw/meta-vsc-2022
default
train_reference
630
video_id
7
7
7
7
0
{ "bin_edges": [ 7, 7 ], "hist": [ 630 ] }
true
Deojoandco/reddit_ah_v1
default
train
297
body
0
7,593
453.45791
337
642.11788
{ "bin_edges": [ 0, 760, 1520, 2280, 3040, 3800, 4560, 5320, 6080, 6840, 7593 ], "hist": [ 245, 39, 8, 4, 0, 0, 0, 0, 0, 1 ] }
false
Deojoandco/reddit_ah_v1
default
train
297
id
7
7
7
7
0
{ "bin_edges": [ 7, 7 ], "hist": [ 297 ] }
false
Deojoandco/reddit_ah_v1
default
train
297
name
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 297 ] }
false
Deojoandco/reddit_ah_v1
default
train
297
title
16
300
107.21549
88
66.39384
{ "bin_edges": [ 16, 45, 74, 103, 132, 161, 190, 219, 248, 277, 300 ], "hist": [ 31, 86, 63, 37, 24, 16, 17, 6, 8, 9 ] }
false
Deojoandco/reddit_ah_v1
default
train
297
url
72
107
101.56229
103
5.4485
{ "bin_edges": [ 72, 76, 80, 84, 88, 92, 96, 100, 104, 107 ], "hist": [ 2, 3, 2, 3, 8, 9, 35, 98, 137 ] }
false
katielink/genomic-benchmarks
demo_human_or_worm
train
75,000
region
1
5
2.99571
4
1.52449
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 20221, 12019, 5252, 22877, 14631 ] }
false
katielink/genomic-benchmarks
demo_human_or_worm
test
25,000
region
1
5
2.99404
4
1.53144
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 6806, 4006, 1685, 7537, 4966 ] }
false
katielink/genomic-benchmarks
demo_coding_vs_intergenomic_seqs
train
75,000
region
4
15
9.68675
10
5.32429
{ "bin_edges": [ 4, 6, 8, 10, 12, 14, 15 ], "hist": [ 37500, 0, 0, 0, 0, 37500 ] }
false
RealTimeData/News_August_2023
default
train
5,059
date_download
25
25
25
25
0
{ "bin_edges": [ 25, 25 ], "hist": [ 5059 ] }
false
RealTimeData/News_August_2023
default
train
5,059
date_publish
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 5059 ] }
false
RealTimeData/News_August_2023
default
train
5,059
description
1
1,496
141.18937
129
112.42245
{ "bin_edges": [ 1, 151, 301, 451, 601, 751, 901, 1051, 1201, 1351, 1496 ], "hist": [ 3286, 1117, 608, 26, 18, 3, 0, 0, 0, 1 ] }
false
RealTimeData/News_August_2023
default
train
5,059
filename
52
574
131.96205
126
38.81964
{ "bin_edges": [ 52, 105, 158, 211, 264, 317, 370, 423, 476, 529, 574 ], "hist": [ 1165, 2815, 908, 130, 28, 8, 1, 1, 2, 1 ] }
false
RealTimeData/News_August_2023
default
train
5,059
image_url
29
396
128.22396
108
55.17467
{ "bin_edges": [ 29, 66, 103, 140, 177, 214, 251, 288, 325, 362, 396 ], "hist": [ 170, 2198, 853, 655, 781, 312, 71, 14, 3, 2 ] }
false
RealTimeData/News_August_2023
default
train
5,059
maintext
1
32,272
2,828.10516
2,147
2,647.87268
{ "bin_edges": [ 1, 3229, 6457, 9685, 12913, 16141, 19369, 22597, 25825, 29053, 32272 ], "hist": [ 3366, 1279, 302, 84, 17, 3, 2, 1, 4, 1 ] }
false
RealTimeData/News_August_2023
default
train
5,059
source_domain
8
39
16.55129
16
4.30522
{ "bin_edges": [ 8, 12, 16, 20, 24, 28, 32, 36, 39 ], "hist": [ 285, 1922, 1630, 915, 229, 75, 0, 3 ] }
false
RealTimeData/News_August_2023
default
train
5,059
title
1
200
67.27594
65
25.93642
{ "bin_edges": [ 1, 21, 41, 61, 81, 101, 121, 141, 161, 181, 200 ], "hist": [ 199, 489, 1370, 1614, 995, 277, 66, 28, 11, 10 ] }
false
RealTimeData/News_August_2023
default
train
5,059
url
37
391
112.63076
108
35.69623
{ "bin_edges": [ 37, 73, 109, 145, 181, 217, 253, 289, 325, 361, 391 ], "hist": [ 483, 2086, 1663, 627, 140, 39, 10, 5, 4, 2 ] }
false
katielink/genomic-benchmarks
demo_coding_vs_intergenomic_seqs
test
25,000
region
4
15
9.68268
10
5.32832
{ "bin_edges": [ 4, 6, 8, 10, 12, 14, 15 ], "hist": [ 12500, 0, 0, 0, 0, 12500 ] }
false
lewtun/my-awesome-dataset
default
train
891
Cabin
1
15
3.58824
3
2.07425
{ "bin_edges": [ 1, 3, 5, 7, 9, 11, 13, 15, 15 ], "hist": [ 28, 152, 4, 12, 0, 6, 0, 2 ] }
false
lewtun/my-awesome-dataset
default
train
891
Name
12
82
26.96521
25
9.28161
{ "bin_edges": [ 12, 20, 28, 36, 44, 52, 60, 68, 76, 82 ], "hist": [ 204, 353, 200, 62, 54, 14, 3, 0, 1 ] }
false
lewtun/my-awesome-dataset
default
train
891
Ticket
3
18
6.75084
6
2.74551
{ "bin_edges": [ 3, 5, 7, 9, 11, 13, 15, 17, 18 ], "hist": [ 103, 550, 103, 67, 18, 10, 20, 20 ] }
false
lewtun/my-awesome-dataset
default
test
418
Name
13
63
27.48325
25
9.97123
{ "bin_edges": [ 13, 19, 25, 31, 37, 43, 49, 55, 61, 63 ], "hist": [ 70, 120, 123, 29, 30, 28, 11, 4, 3 ] }
false
lewtun/my-awesome-dataset
default
test
418
Ticket
3
18
6.8756
6
2.82144
{ "bin_edges": [ 3, 5, 7, 9, 11, 13, 15, 17, 18 ], "hist": [ 49, 245, 51, 33, 16, 9, 5, 10 ] }
false
lionelchg/dolly_brainstorming
default
test
89
instruction
20
535
66.10112
57
58.43879
{ "bin_edges": [ 20, 72, 124, 176, 228, 280, 332, 384, 436, 488, 535 ], "hist": [ 68, 18, 1, 0, 1, 0, 0, 0, 0, 1 ] }
false
lionelchg/dolly_brainstorming
default
test
89
response
32
3,186
302.59551
202
410.98448
{ "bin_edges": [ 32, 348, 664, 980, 1296, 1612, 1928, 2244, 2560, 2876, 3186 ], "hist": [ 67, 15, 3, 2, 0, 1, 0, 0, 0, 1 ] }
false
lionelchg/dolly_brainstorming
default
test
89
text
100
3,327
416.69663
310
422.58236
{ "bin_edges": [ 100, 423, 746, 1069, 1392, 1715, 2038, 2361, 2684, 3007, 3327 ], "hist": [ 62, 17, 5, 3, 0, 1, 0, 0, 0, 1 ] }
false
lionelchg/dolly_brainstorming
default
train
1,677
instruction
10
677
60.42516
53
46.77811
{ "bin_edges": [ 10, 77, 144, 211, 278, 345, 412, 479, 546, 613, 677 ], "hist": [ 1408, 231, 19, 6, 2, 4, 1, 1, 2, 3 ] }
false
lionelchg/dolly_brainstorming
default
train
1,677
response
3
6,172
339.54264
215
431.45739
{ "bin_edges": [ 3, 620, 1237, 1854, 2471, 3088, 3705, 4322, 4939, 5556, 6172 ], "hist": [ 1440, 182, 40, 6, 2, 3, 1, 1, 0, 2 ] }
false
lionelchg/dolly_brainstorming
default
train
1,677
text
73
6,271
447.9678
321
440.95187
{ "bin_edges": [ 73, 693, 1313, 1933, 2553, 3173, 3793, 4413, 5033, 5653, 6271 ], "hist": [ 1412, 206, 42, 7, 3, 2, 2, 1, 0, 2 ] }
false
koakande/github-issues
default
train
6,493
body
0
227,963
988.52493
367
3,407.89334
{ "bin_edges": [ 0, 22797, 45594, 68391, 91188, 113985, 136782, 159579, 182376, 205173, 227963 ], "hist": [ 6253, 4, 0, 0, 0, 0, 0, 0, 0, 1 ] }
false
koakande/github-issues
default
train
6,493
comments_url
67
70
69.83105
70
0.42009
{ "bin_edges": [ 67, 68, 69, 70, 70 ], "hist": [ 9, 90, 890, 5504 ] }
false
koakande/github-issues
default
train
6,493
events_url
65
68
67.83105
68
0.42009
{ "bin_edges": [ 65, 66, 67, 68, 68 ], "hist": [ 9, 90, 890, 5504 ] }
false
koakande/github-issues
default
train
6,493
html_url
46
51
49.6313
49
1.08477
{ "bin_edges": [ 46, 47, 48, 49, 50, 51, 51 ], "hist": [ 4, 89, 538, 3270, 357, 2235 ] }
false
koakande/github-issues
default
train
6,493
labels_url
72
75
74.83105
75
0.42009
{ "bin_edges": [ 72, 73, 74, 75, 75 ], "hist": [ 9, 90, 890, 5504 ] }
false
koakande/github-issues
default
train
6,493
node_id
18
32
23.33482
19
5.96527
{ "bin_edges": [ 18, 20, 22, 24, 26, 28, 30, 32, 32 ], "hist": [ 3637, 0, 0, 912, 0, 0, 0, 1944 ] }
false
koakande/github-issues
default
train
6,493
timeline_url
67
70
69.83105
70
0.42009
{ "bin_edges": [ 67, 68, 69, 70, 70 ], "hist": [ 9, 90, 890, 5504 ] }
false
koakande/github-issues
default
train
6,493
title
1
290
42.02972
38
22.55949
{ "bin_edges": [ 1, 30, 59, 88, 117, 146, 175, 204, 233, 262, 290 ], "hist": [ 2080, 3182, 994, 189, 35, 7, 1, 0, 1, 4 ] }
false
koakande/github-issues
default
train
6,493
url
58
61
60.83105
61
0.42009
{ "bin_edges": [ 58, 59, 60, 61, 61 ], "hist": [ 9, 90, 890, 5504 ] }
false
pat-jj/ClinicalTrialSummary_Full
default
train
163,088
article
1
5,771
1,223.87804
981
918.89579
{ "bin_edges": [ 1, 579, 1157, 1735, 2313, 2891, 3469, 4047, 4625, 5203, 5771 ], "hist": [ 43076, 52306, 31061, 16559, 9298, 5285, 3019, 1707, 682, 95 ] }
false
pat-jj/ClinicalTrialSummary_Full
default
train
163,088
summary
3
817
306.54474
276
145.0363
{ "bin_edges": [ 3, 85, 167, 249, 331, 413, 495, 577, 659, 741, 817 ], "hist": [ 2037, 26036, 41895, 30868, 22825, 17848, 12948, 6803, 1700, 128 ] }
false
adambuttrick/360K-funding-statement-sentences-name-identifier
default
train
10,218,820
Sentence #
11
16
15.69393
16
0.52945
{ "bin_edges": [ 11, 12, 13, 14, 15, 16, 16 ], "hist": [ 9, 90, 900, 9000, 90000, 263012 ] }
false
adambuttrick/360K-funding-statement-sentences-name-identifier
default
train
10,218,820
Word
1
735
6.19954
6
3.98361
{ "bin_edges": [ 1, 75, 149, 223, 297, 371, 445, 519, 593, 667, 735 ], "hist": [ 10217461, 1165, 66, 22, 7, 3, 0, 1, 0, 1 ] }
false
pat-jj/ClinicalTrialSummary_Full
default
validation
20,386
article
8
5,614
1,252.99235
987
936.722
{ "bin_edges": [ 8, 569, 1130, 1691, 2252, 2813, 3374, 3935, 4496, 5057, 5614 ], "hist": [ 4994, 6599, 3885, 2070, 1244, 727, 420, 280, 146, 21 ] }
false
pat-jj/ClinicalTrialSummary_Full
default
validation
20,386
summary
24
813
309.27485
278
147.45135
{ "bin_edges": [ 24, 103, 182, 261, 340, 419, 498, 577, 656, 735, 813 ], "hist": [ 564, 3933, 4827, 3453, 2671, 2138, 1644, 880, 250, 26 ] }
false
assafm/cs-strategy-001
default
train
186
text
106
410
209.31183
212.5
53.58943
{ "bin_edges": [ 106, 137, 168, 199, 230, 261, 292, 323, 354, 385, 410 ], "hist": [ 16, 22, 42, 39, 39, 18, 5, 3, 1, 1 ] }
false
pat-jj/ClinicalTrialSummary_Full
default
test
20,386
article
8
5,634
1,252.23722
1,000
916.53693
{ "bin_edges": [ 8, 571, 1134, 1697, 2260, 2823, 3386, 3949, 4512, 5075, 5634 ], "hist": [ 4880, 6675, 3935, 2185, 1186, 736, 413, 252, 103, 21 ] }
false
pat-jj/ClinicalTrialSummary_Full
default
test
20,386
summary
2
807
311.26695
277
149.80296
{ "bin_edges": [ 2, 83, 164, 245, 326, 407, 488, 569, 650, 731, 807 ], "hist": [ 234, 3027, 5265, 3684, 2739, 2212, 1794, 1095, 305, 31 ] }
false
lionelchg/dolly15k
default
test
751
context
0
6,514
326.39015
0
760.49109
{ "bin_edges": [ 0, 652, 1304, 1956, 2608, 3260, 3912, 4564, 5216, 5868, 6514 ], "hist": [ 610, 88, 28, 9, 6, 2, 2, 3, 1, 2 ] }
false
lionelchg/dolly15k
default
test
751
instruction
13
1,505
68.47537
54
70.68016
{ "bin_edges": [ 13, 163, 313, 463, 613, 763, 913, 1063, 1213, 1363, 1505 ], "hist": [ 716, 32, 2, 0, 0, 0, 0, 0, 0, 1 ] }
false
lionelchg/dolly15k
default
test
751
response
1
4,804
377.86418
193
536.41191
{ "bin_edges": [ 1, 482, 963, 1444, 1925, 2406, 2887, 3368, 3849, 4330, 4804 ], "hist": [ 579, 105, 32, 17, 6, 6, 2, 1, 1, 2 ] }
false
lionelchg/dolly15k
default
test
751
text
71
7,782
820.72969
520
935.71986
{ "bin_edges": [ 71, 843, 1615, 2387, 3159, 3931, 4703, 5475, 6247, 7019, 7782 ], "hist": [ 511, 152, 46, 20, 9, 4, 2, 4, 2, 1 ] }
false
lionelchg/dolly15k
default
train
14,260
context
0
23,505
354.29488
0
965.50259
{ "bin_edges": [ 0, 2351, 4702, 7053, 9404, 11755, 14106, 16457, 18808, 21159, 23505 ], "hist": [ 13859, 294, 60, 27, 4, 8, 3, 1, 3, 1 ] }
false