dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
alancooney/relation_counterfact
default
train
6,597
prompt
7
85
32.75443
32
11.26884
{ "bin_edges": [ 7, 15, 23, 31, 39, 47, 55, 63, 71, 79, 85 ], "hist": [ 140, 1126, 1757, 1778, 1025, 522, 152, 67, 20, 10 ] }
false
alancooney/relation_counterfact
default
train
6,597
relation
2
62
16.74958
15
8.60511
{ "bin_edges": [ 2, 9, 16, 23, 30, 37, 44, 51, 58, 62 ], "hist": [ 1121, 2404, 1426, 1120, 478, 32, 4, 2, 10 ] }
false
alancooney/relation_counterfact
default
train
6,597
rome_relation_id
3
5
3.76307
4
0.56111
{ "bin_edges": [ 3, 4, 5, 5 ], "hist": [ 2005, 4150, 442 ] }
false
alancooney/relation_counterfact
default
train
6,597
subject
2
70
15.00485
14
7.54212
{ "bin_edges": [ 2, 9, 16, 23, 30, 37, 44, 51, 58, 65, 70 ], "hist": [ 954, 3373, 1472, 470, 166, 80, 58, 15, 7, 2 ] }
false
hugosousa/professor_heideltime_en
english
train
24,642
dct
10
19
15.71183
19
4.33385
{ "bin_edges": [ 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 19 ], "hist": [ 9003, 0, 0, 0, 0, 0, 0, 0, 0, 15639 ] }
false
hugosousa/professor_heideltime_en
english
train
24,642
id
12
12
12
12
0
{ "bin_edges": [ 12, 12 ], "hist": [ 24642 ] }
false
hugosousa/professor_heideltime_en
english
train
24,642
text
25
73,628
3,144.60742
2,377.5
3,212.94597
{ "bin_edges": [ 25, 7386, 14747, 22108, 29469, 36830, 44191, 51552, 58913, 66274, 73628 ], "hist": [ 22888, 1548, 131, 26, 27, 9, 8, 2, 2, 1 ] }
false
mareloraby/uk_acc_1985
default
train
245,645
accident_index
13
13
13
13
0
{ "bin_edges": [ 13, 13 ], "hist": [ 245645 ] }
false
mareloraby/uk_acc_1985
default
train
245,645
accident_reference
9
9
9
9
0
{ "bin_edges": [ 9, 9 ], "hist": [ 245645 ] }
false
mareloraby/uk_acc_1985
default
train
245,645
first_road_number
3
100
40.52185
6
46.02094
{ "bin_edges": [ 3, 13, 23, 33, 43, 53, 63, 73, 83, 93, 100 ], "hist": [ 153660, 0, 0, 0, 0, 0, 0, 0, 0, 91971 ] }
false
mareloraby/uk_acc_1985
default
train
245,645
second_road_number
3
100
82.39798
100
36.87368
{ "bin_edges": [ 3, 13, 23, 33, 43, 53, 63, 73, 83, 93, 100 ], "hist": [ 27335, 0, 0, 0, 0, 0, 0, 0, 0, 119945 ] }
false
mareloraby/uk_acc_1985
default
train
245,645
time
5
5
5
5
0
{ "bin_edges": [ 5, 5 ], "hist": [ 245645 ] }
false
vidhikatkoria/SGD_Buses
default
train
7,552
context
19
482
182.19174
178
69.52024
{ "bin_edges": [ 19, 66, 113, 160, 207, 254, 301, 348, 395, 442, 482 ], "hist": [ 314, 777, 1866, 1957, 1516, 753, 246, 88, 25, 10 ] }
false
vidhikatkoria/SGD_Buses
default
train
7,552
response
3
255
50.85395
41
31.90658
{ "bin_edges": [ 3, 29, 55, 81, 107, 133, 159, 185, 211, 237, 255 ], "hist": [ 2042, 2802, 1455, 808, 260, 111, 51, 18, 3, 2 ] }
false
HydraLM/alpaca_data_cleaned_standardized
default
train
122,650
message
1
4,522
323.16759
69
521.74152
{ "bin_edges": [ 1, 454, 907, 1360, 1813, 2266, 2719, 3172, 3625, 4078, 4522 ], "hist": [ 95907, 10089, 7135, 5864, 2785, 732, 120, 1, 0, 2 ] }
false
hugosousa/professor_heideltime_en
french
train
24,293
dct
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 24293 ] }
false
hugosousa/professor_heideltime_en
french
train
24,293
id
12
12
12
12
0
{ "bin_edges": [ 12, 12 ], "hist": [ 24293 ] }
false
hugosousa/professor_heideltime_en
french
train
24,293
text
14
1,232
353.48829
330
147.27256
{ "bin_edges": [ 14, 136, 258, 380, 502, 624, 746, 868, 990, 1112, 1232 ], "hist": [ 985, 6453, 7068, 5459, 3359, 840, 99, 19, 5, 6 ] }
false
kyleeasterly/purple-aerospace-mix-v2-300-6
default
train
2,100
text
324
13,893
4,376.51667
3,942.5
2,110.34181
{ "bin_edges": [ 324, 1681, 3038, 4395, 5752, 7109, 8466, 9823, 11180, 12537, 13893 ], "hist": [ 75, 516, 664, 422, 217, 82, 43, 68, 10, 3 ] }
false
spawn99/CornellMovieDialogCorpus
default
movie_lines
304,713
characterID
2
5
4.8849
5
0.35903
{ "bin_edges": [ 2, 3, 4, 5, 5 ], "hist": [ 618, 2267, 28684, 273144 ] }
false
spawn99/CornellMovieDialogCorpus
default
movie_lines
304,713
characterName
1
52
5.86233
5
2.37378
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 52 ], "hist": [ 219428, 77828, 6791, 585, 24, 4, 0, 0, 10 ] }
false
spawn99/CornellMovieDialogCorpus
default
movie_lines
304,713
lineID
3
7
6.84846
7
0.40367
{ "bin_edges": [ 3, 4, 5, 6, 7, 7 ], "hist": [ 27, 619, 3219, 37772, 263076 ] }
false
spawn99/CornellMovieDialogCorpus
default
movie_lines
304,713
movieID
2
4
3.82332
4
0.41473
{ "bin_edges": [ 2, 3, 4, 4 ], "hist": [ 4042, 45754, 254917 ] }
false
spawn99/CornellMovieDialogCorpus
default
movie_lines
304,713
utterance
1
3,046
55.32015
35
64.08586
{ "bin_edges": [ 1, 306, 611, 916, 1221, 1526, 1831, 2136, 2441, 2746, 3046 ], "hist": [ 301465, 2698, 215, 50, 8, 6, 2, 0, 0, 2 ] }
false
andersonbcdefg/synthetic_gptj_paraphrased
default
train
33,157
fixed_chosen
0
1,716
297.13303
286
152.0749
{ "bin_edges": [ 0, 172, 344, 516, 688, 860, 1032, 1204, 1376, 1548, 1716 ], "hist": [ 6669, 15280, 8862, 1833, 337, 113, 36, 14, 7, 6 ] }
false
andersonbcdefg/synthetic_gptj_paraphrased
default
train
33,157
fixed_rejected
0
2,851
271.04346
259
147.97969
{ "bin_edges": [ 0, 286, 572, 858, 1144, 1430, 1716, 2002, 2288, 2574, 2851 ], "hist": [ 19112, 13033, 904, 82, 19, 4, 1, 1, 0, 1 ] }
false
andersonbcdefg/synthetic_gptj_paraphrased
default
train
33,157
orig_chosen
1
3,095
380.92083
363
211.78569
{ "bin_edges": [ 1, 311, 621, 931, 1241, 1551, 1861, 2171, 2481, 2791, 3095 ], "hist": [ 13102, 16282, 3199, 470, 82, 17, 3, 1, 0, 1 ] }
false
andersonbcdefg/synthetic_gptj_paraphrased
default
train
33,157
orig_rejected
1
4,938
344.67594
296
248.46878
{ "bin_edges": [ 1, 495, 989, 1483, 1977, 2471, 2965, 3459, 3953, 4447, 4938 ], "hist": [ 25308, 7379, 422, 37, 5, 2, 1, 1, 1, 1 ] }
false
andersonbcdefg/synthetic_gptj_paraphrased
default
train
33,157
prompt
1
206
53.34506
50
18.43873
{ "bin_edges": [ 1, 22, 43, 64, 85, 106, 127, 148, 169, 190, 206 ], "hist": [ 99, 9018, 17197, 4875, 1235, 498, 175, 40, 18, 2 ] }
false
hugosousa/professor_heideltime_en
italian
train
9,619
dct
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 9619 ] }
false
hugosousa/professor_heideltime_en
italian
train
9,619
id
12
12
12
12
0
{ "bin_edges": [ 12, 12 ], "hist": [ 9619 ] }
false
hugosousa/professor_heideltime_en
italian
train
9,619
text
138
24,193
1,930.26708
1,778
1,202.30586
{ "bin_edges": [ 138, 2544, 4950, 7356, 9762, 12168, 14574, 16980, 19386, 21792, 24193 ], "hist": [ 7033, 2495, 60, 7, 5, 8, 4, 5, 1, 1 ] }
false
Talha185/Common-voice-urdu-11
default
train
10,411
path
28
28
28
28
0
{ "bin_edges": [ 28, 28 ], "hist": [ 10411 ] }
false
Talha185/Common-voice-urdu-11
default
train
10,411
sentence
2
125
35.80309
33
17.63505
{ "bin_edges": [ 2, 15, 28, 41, 54, 67, 80, 93, 106, 119, 125 ], "hist": [ 728, 2872, 3194, 2168, 896, 308, 163, 53, 25, 4 ] }
false
sidmanale643/med_QnA
default
train
24,665
text
141
1,696
415.29738
400
120.08427
{ "bin_edges": [ 141, 297, 453, 609, 765, 921, 1077, 1233, 1389, 1545, 1696 ], "hist": [ 3675, 12896, 6443, 1397, 206, 41, 5, 1, 0, 1 ] }
false
ticoAg/Medical-Dialogue-System
default
train
3,206,606
input
1
6,935
96.22716
65
108.13437
{ "bin_edges": [ 1, 695, 1389, 2083, 2777, 3471, 4165, 4859, 5553, 6247, 6935 ], "hist": [ 3192918, 12672, 736, 183, 59, 18, 8, 6, 3, 3 ] }
false
ticoAg/Medical-Dialogue-System
default
train
3,206,606
output
7
5,912
37.82622
27
38.52453
{ "bin_edges": [ 7, 598, 1189, 1780, 2371, 2962, 3553, 4144, 4735, 5326, 5912 ], "hist": [ 3205982, 528, 72, 11, 3, 3, 1, 2, 3, 1 ] }
false
jonathansuru/customer_service_information_extraction
default
train
190
completion
30
588
163.65263
124.5
115.99961
{ "bin_edges": [ 30, 86, 142, 198, 254, 310, 366, 422, 478, 534, 588 ], "hist": [ 30, 85, 38, 9, 8, 1, 1, 10, 7, 1 ] }
false
jonathansuru/customer_service_information_extraction
default
train
190
prompt
356
1,554
660.72632
654.5
206.52317
{ "bin_edges": [ 356, 476, 596, 716, 836, 956, 1076, 1196, 1316, 1436, 1554 ], "hist": [ 48, 30, 39, 35, 23, 11, 1, 1, 1, 1 ] }
false
bprateek/amazon_product_description
default
train
10,002
About Product
13
2,614
426.60746
326
323.29762
{ "bin_edges": [ 13, 274, 535, 796, 1057, 1318, 1579, 1840, 2101, 2362, 2614 ], "hist": [ 3882, 3328, 1301, 581, 443, 128, 37, 18, 7, 4 ] }
false
bprateek/amazon_product_description
default
train
10,002
Image
64
655
309.35813
284
146.48281
{ "bin_edges": [ 64, 124, 184, 244, 304, 364, 424, 484, 544, 604, 655 ], "hist": [ 125, 2691, 1811, 1268, 923, 770, 671, 26, 1713, 4 ] }
false
bprateek/amazon_product_description
default
train
10,002
Model Number
1
39
7.45905
6
3.92622
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 33, 37, 39 ], "hist": [ 761, 5404, 1401, 335, 163, 89, 52, 14, 6, 5 ] }
false
bprateek/amazon_product_description
default
train
10,002
Product Name
1
200
67.55609
58
36.93096
{ "bin_edges": [ 1, 21, 41, 61, 81, 101, 121, 141, 161, 181, 200 ], "hist": [ 191, 1994, 3141, 2131, 1085, 573, 270, 194, 175, 248 ] }
false
bprateek/amazon_product_description
default
train
10,002
Product Specification
15
642
247.94205
218
86.12486
{ "bin_edges": [ 15, 78, 141, 204, 267, 330, 393, 456, 519, 582, 642 ], "hist": [ 103, 554, 3356, 634, 2903, 482, 237, 60, 32, 9 ] }
false
bprateek/amazon_product_description
default
train
10,002
Product Url
51
126
74.09098
74
6.26915
{ "bin_edges": [ 51, 59, 67, 75, 83, 91, 99, 107, 115, 123, 126 ], "hist": [ 31, 1006, 4441, 3608, 833, 75, 4, 1, 2, 1 ] }
false
bprateek/amazon_product_description
default
train
10,002
Selling Price
5
354
6.27246
6
6.50101
{ "bin_edges": [ 5, 40, 75, 110, 145, 180, 215, 250, 285, 320, 354 ], "hist": [ 9891, 0, 0, 0, 0, 0, 0, 0, 3, 1 ] }
false
bprateek/amazon_product_description
default
train
10,002
Technical Details
8
14,326
680.48665
570
497.16055
{ "bin_edges": [ 8, 1440, 2872, 4304, 5736, 7168, 8600, 10032, 11464, 12896, 14326 ], "hist": [ 8620, 557, 30, 3, 1, 0, 0, 0, 0, 1 ] }
false
bprateek/amazon_product_description
default
train
10,002
Uniq Id
32
32
32
32
0
{ "bin_edges": [ 32, 32 ], "hist": [ 10002 ] }
false
bprateek/amazon_product_description
default
train
10,002
Variants
119
2,959
423.74132
303
362.46811
{ "bin_edges": [ 119, 404, 689, 974, 1259, 1544, 1829, 2114, 2399, 2684, 2959 ], "hist": [ 1583, 504, 211, 92, 33, 27, 14, 5, 2, 7 ] }
false
rusheeliyer/german-courts
bundesarbeitsgericht
train
117
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 117 ] }
false
rusheeliyer/german-courts
bundesarbeitsgericht
train
117
id
8
9
8.78632
9
0.41166
{ "bin_edges": [ 8, 9, 9 ], "hist": [ 25, 92 ] }
false
rusheeliyer/german-courts
bundesarbeitsgericht
train
117
judgement
11,433
190,350
46,175.00855
39,070
27,828.88341
{ "bin_edges": [ 11433, 29325, 47217, 65109, 83001, 100893, 118785, 136677, 154569, 172461, 190350 ], "hist": [ 31, 49, 16, 8, 8, 2, 1, 1, 0, 1 ] }
false
rusheeliyer/german-courts
bundesarbeitsgericht
train
117
summary
378
8,928
3,270.62393
3,106
1,193.69943
{ "bin_edges": [ 378, 1234, 2090, 2946, 3802, 4658, 5514, 6370, 7226, 8082, 8928 ], "hist": [ 3, 8, 39, 37, 18, 5, 5, 1, 0, 1 ] }
false
rusheeliyer/german-courts
bundesarbeitsgericht
test
31
id
8
9
8.80645
9
0.40161
{ "bin_edges": [ 8, 9, 9 ], "hist": [ 6, 25 ] }
false
rusheeliyer/german-courts
bundesarbeitsgericht
test
31
judgement
17,817
85,402
39,362.19355
37,521
15,819.56867
{ "bin_edges": [ 17817, 24576, 31335, 38094, 44853, 51612, 58371, 65130, 71889, 78648, 85402 ], "hist": [ 5, 6, 6, 5, 3, 2, 2, 1, 0, 1 ] }
false
rusheeliyer/german-courts
bundesarbeitsgericht
test
31
summary
1,977
6,985
3,971.77419
3,791
1,430.50193
{ "bin_edges": [ 1977, 2478, 2979, 3480, 3981, 4482, 4983, 5484, 5985, 6486, 6985 ], "hist": [ 6, 5, 2, 5, 3, 2, 2, 3, 1, 2 ] }
false
hugosousa/professor_heideltime_en
german
train
33,266
dct
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 33266 ] }
false
hugosousa/professor_heideltime_en
german
train
33,266
id
12
12
12
12
0
{ "bin_edges": [ 12, 12 ], "hist": [ 33266 ] }
false
hugosousa/professor_heideltime_en
german
train
33,266
text
40
157,293
3,940.97917
2,797
5,356.56263
{ "bin_edges": [ 40, 15766, 31492, 47218, 62944, 78670, 94396, 110122, 125848, 141574, 157293 ], "hist": [ 32547, 503, 142, 32, 18, 6, 17, 0, 0, 1 ] }
false
Vezora/Wizard_Math_Alpaca
default
train
6,319
instruction
16
2,828
201.30733
161
171.08753
{ "bin_edges": [ 16, 298, 580, 862, 1144, 1426, 1708, 1990, 2272, 2554, 2828 ], "hist": [ 5266, 853, 143, 30, 10, 8, 6, 2, 0, 1 ] }
false
Vezora/Wizard_Math_Alpaca
default
train
6,319
output
26
5,313
472.56006
352
398.58093
{ "bin_edges": [ 26, 555, 1084, 1613, 2142, 2671, 3200, 3729, 4258, 4787, 5313 ], "hist": [ 4606, 1254, 307, 105, 32, 7, 5, 2, 0, 1 ] }
false
RealTimeData/github_july_week2_2023
default
train
937
description
3
347
68.80218
55
53.68989
{ "bin_edges": [ 3, 38, 73, 108, 143, 178, 213, 248, 283, 318, 347 ], "hist": [ 211, 269, 146, 53, 18, 11, 9, 7, 3, 6 ] }
false
RealTimeData/github_july_week2_2023
default
train
937
full_name
10
67
24.85165
24
8.33015
{ "bin_edges": [ 10, 16, 22, 28, 34, 40, 46, 52, 58, 64, 67 ], "hist": [ 104, 261, 276, 168, 76, 28, 17, 4, 2, 1 ] }
false
RealTimeData/github_july_week2_2023
default
train
937
readme
0
161,737
4,743.51121
2,718
8,550.88684
{ "bin_edges": [ 0, 16174, 32348, 48522, 64696, 80870, 97044, 113218, 129392, 145566, 161737 ], "hist": [ 897, 31, 5, 1, 1, 0, 1, 0, 0, 1 ] }
false
RealTimeData/github_july_week2_2023
default
train
937
url
29
86
43.85165
43
8.33015
{ "bin_edges": [ 29, 35, 41, 47, 53, 59, 65, 71, 77, 83, 86 ], "hist": [ 104, 261, 276, 168, 76, 28, 17, 4, 2, 1 ] }
false
augtoma/usmle_step_1
default
test
94
answer
1
120
30.70213
19.5
27.1922
{ "bin_edges": [ 1, 13, 25, 37, 49, 61, 73, 85, 97, 109, 120 ], "hist": [ 22, 34, 13, 7, 5, 3, 3, 3, 1, 3 ] }
false
augtoma/usmle_step_1
default
test
94
question
236
1,508
603.05319
567.5
239.0183
{ "bin_edges": [ 236, 364, 492, 620, 748, 876, 1004, 1132, 1260, 1388, 1508 ], "hist": [ 15, 17, 26, 15, 8, 6, 5, 0, 1, 1 ] }
false
hugosousa/professor_heideltime_en
portuguese
train
27,154
dct
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 27154 ] }
false
hugosousa/professor_heideltime_en
portuguese
train
27,154
id
12
12
12
12
0
{ "bin_edges": [ 12, 12 ], "hist": [ 27154 ] }
false
drhead/laion_hd_21M_deduped
default
train
21,802,719
TEXT
2
24,527
72.37943
57
65.58343
{ "bin_edges": [ 2, 2455, 4908, 7361, 9814, 12267, 14720, 17173, 19626, 22079, 24527 ], "hist": [ 21802491, 173, 24, 8, 6, 11, 4, 0, 1, 1 ] }
false
drhead/laion_hd_21M_deduped
default
train
21,802,719
URL
23
5,802
115.98943
97
70.01716
{ "bin_edges": [ 23, 601, 1179, 1757, 2335, 2913, 3491, 4069, 4647, 5225, 5802 ], "hist": [ 21777759, 24461, 111, 331, 44, 3, 8, 0, 1, 1 ] }
false
hugosousa/professor_heideltime_en
spanish
train
19,095
dct
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 19095 ] }
false
hugosousa/professor_heideltime_en
spanish
train
19,095
id
12
12
12
12
0
{ "bin_edges": [ 12, 12 ], "hist": [ 19095 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
train
569
date
12
18
14.8348
15
1.87105
{ "bin_edges": [ 12, 13, 14, 15, 16, 17, 18, 18 ], "hist": [ 50, 152, 58, 75, 101, 82, 51 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
train
569
id
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 569 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
train
569
judgement
5,311
113,647
22,101.60281
18,441
13,301.36967
{ "bin_edges": [ 5311, 16145, 26979, 37813, 48647, 59481, 70315, 81149, 91983, 102817, 113647 ], "hist": [ 222, 207, 73, 41, 15, 4, 4, 2, 0, 1 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
train
569
summary
1,016
7,273
2,770.33919
2,690
671.93388
{ "bin_edges": [ 1016, 1642, 2268, 2894, 3520, 4146, 4772, 5398, 6024, 6650, 7273 ], "hist": [ 9, 112, 243, 143, 43, 12, 4, 1, 1, 1 ] }
false
Abdelkareem/arabic_tweets_classification
default
train
13,240
Cleaned Text
3
277
94.24322
79
50.2543
{ "bin_edges": [ 3, 31, 59, 87, 115, 143, 171, 199, 227, 255, 277 ], "hist": [ 314, 3328, 3748, 2012, 1624, 1140, 423, 277, 259, 114 ] }
false
Abdelkareem/arabic_tweets_classification
default
train
13,240
Date
8
19
17.56269
19
3.70751
{ "bin_edges": [ 8, 10, 12, 14, 16, 18, 19 ], "hist": [ 1730, 0, 0, 0, 0, 11510 ] }
false
Abdelkareem/arabic_tweets_classification
default
train
13,240
Date Time
19
30
28.56269
30
3.70751
{ "bin_edges": [ 19, 21, 23, 25, 27, 29, 30 ], "hist": [ 1730, 0, 0, 0, 0, 11510 ] }
false
Abdelkareem/arabic_tweets_classification
default
train
13,240
Time
8
19
9.43731
8
3.70751
{ "bin_edges": [ 8, 10, 12, 14, 16, 18, 19 ], "hist": [ 11510, 0, 0, 0, 0, 1730 ] }
false
Abdelkareem/arabic_tweets_classification
default
train
13,240
Tweet Text
3
277
94.2898
79
50.28334
{ "bin_edges": [ 3, 31, 59, 87, 115, 143, 171, 199, 227, 255, 277 ], "hist": [ 311, 3328, 3749, 2013, 1624, 1141, 423, 277, 259, 115 ] }
false
Abdelkareem/arabic_tweets_classification
default
train
13,240
URL
55
60
57.86202
59
1.9725
{ "bin_edges": [ 55, 56, 57, 58, 59, 60, 60 ], "hist": [ 1752, 3034, 0, 0, 3261, 3252 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
test
133
date
12
18
14.91729
15
1.83421
{ "bin_edges": [ 12, 13, 14, 15, 16, 17, 18, 18 ], "hist": [ 10, 35, 11, 18, 29, 19, 11 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
test
133
id
10
12
10.01504
10
0.17342
{ "bin_edges": [ 10, 11, 12, 12 ], "hist": [ 132, 0, 1 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
test
133
judgement
6,114
85,988
24,798.12782
20,761
14,609.11185
{ "bin_edges": [ 6114, 14102, 22090, 30078, 38066, 46054, 54042, 62030, 70018, 78006, 85988 ], "hist": [ 25, 49, 27, 14, 8, 3, 3, 1, 1, 2 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
test
133
summary
1,642
8,471
2,843.42857
2,753
814.79352
{ "bin_edges": [ 1642, 2325, 3008, 3691, 4374, 5057, 5740, 6423, 7106, 7789, 8471 ], "hist": [ 29, 64, 32, 4, 1, 1, 1, 0, 0, 1 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
validation
59
date
12
18
14.59322
15
1.92191
{ "bin_edges": [ 12, 13, 14, 15, 16, 17, 18, 18 ], "hist": [ 9, 15, 3, 13, 7, 7, 5 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
validation
59
id
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 59 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
validation
59
judgement
4,611
75,662
23,614.37288
20,368
12,201.01906
{ "bin_edges": [ 4611, 11717, 18823, 25929, 33035, 40141, 47247, 54353, 61459, 68565, 75662 ], "hist": [ 7, 18, 12, 12, 4, 4, 1, 0, 0, 1 ] }
false
rusheeliyer/german-courts
bundesfinanzhof
validation
59
summary
1,975
6,310
3,068.84746
2,975
728.59378
{ "bin_edges": [ 1975, 2409, 2843, 3277, 3711, 4145, 4579, 5013, 5447, 5881, 6310 ], "hist": [ 9, 19, 14, 8, 5, 3, 0, 0, 0, 1 ] }
false
euclaise/WritingPrompts_preferences
default
train
265,174
post_text
0
17,539
15.12991
0
101.84174
{ "bin_edges": [ 0, 1754, 3508, 5262, 7016, 8770, 10524, 12278, 14032, 15786, 17539 ], "hist": [ 265097, 52, 11, 6, 1, 3, 3, 0, 0, 1 ] }
false
euclaise/WritingPrompts_preferences
default
train
265,174
post_title
4
315
165.66489
161
78.42946
{ "bin_edges": [ 4, 36, 68, 100, 132, 164, 196, 228, 260, 292, 315 ], "hist": [ 8863, 21663, 32021, 36556, 36836, 33088, 28842, 23664, 24675, 18966 ] }
false
FreedomIntelligence/sharegpt-hindi
default
train
3,142
id
1
5
4.7683
5
0.53687
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 6, 24, 65, 502, 2545 ] }
false
Falah/2M_arabic_female_SDXL_refiner_prompts
default
train
2,000,000
prompts
484
555
521.35608
522
10.16072
{ "bin_edges": [ 484, 492, 500, 508, 516, 524, 532, 540, 548, 555 ], "hist": [ 1779, 30261, 154754, 384258, 570754, 528209, 266121, 60136, 3728 ] }
false
katielink/gpt4_bias
nursing_bias
train
192
prompt
802
1,301
982
981
123.91468
{ "bin_edges": [ 802, 852, 902, 952, 1002, 1052, 1102, 1152, 1202, 1252, 1301 ], "hist": [ 32, 32, 15, 46, 11, 15, 25, 8, 0, 8 ] }
false
katielink/gpt4_bias
nursing_bias
train
192
text
556
848
698.25
714.5
97.60214
{ "bin_edges": [ 556, 586, 616, 646, 676, 706, 736, 766, 796, 826, 848 ], "hist": [ 45, 15, 12, 0, 24, 3, 33, 15, 21, 24 ] }
false
rusheeliyer/german-courts
bundesgerichtshof
train
1,889
date
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 1889 ] }
false
rusheeliyer/german-courts
bundesgerichtshof
train
1,889
id
10
12
11.80942
12
0.43264
{ "bin_edges": [ 10, 11, 12, 12 ], "hist": [ 31, 298, 1560 ] }
false
rusheeliyer/german-courts
bundesgerichtshof
train
1,889
judgement
959
145,651
24,331.80836
20,857
14,804.4102
{ "bin_edges": [ 959, 15429, 29899, 44369, 58839, 73309, 87779, 102249, 116719, 131189, 145651 ], "hist": [ 548, 839, 338, 108, 37, 11, 3, 3, 0, 2 ] }
false