dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
unpredictable/unpredictable_unique
default
train
380,248
outputColName
1
1,129
12.35057
10
24.13294
{ "bin_edges": [ 1, 114, 227, 340, 453, 566, 679, 792, 905, 1018, 1129 ], "hist": [ 378821, 957, 169, 60, 48, 80, 37, 7, 38, 31 ] }
false
unpredictable/unpredictable_unique
default
train
380,248
pageTitle
0
1,105
45.49277
41
30.7436
{ "bin_edges": [ 0, 111, 222, 333, 444, 555, 666, 777, 888, 999, 1105 ], "hist": [ 370299, 9330, 403, 114, 24, 37, 28, 0, 0, 13 ] }
false
unpredictable/unpredictable_unique
default
train
380,248
task
12
101
49.30776
49
12.50067
{ "bin_edges": [ 12, 21, 30, 39, 48, 57, 66, 75, 84, 93, 101 ], "hist": [ 10550, 11363, 27412, 112040, 144807, 39883, 21500, 5484, 6111, 1098 ] }
false
unpredictable/unpredictable_unique
default
train
380,248
url
14
440
60.60004
55
25.24703
{ "bin_edges": [ 14, 57, 100, 143, 186, 229, 272, 315, 358, 401, 440 ], "hist": [ 200725, 151973, 23940, 2493, 833, 183, 92, 0, 0, 9 ] }
false
unpredictable/unpredictable_unique
default
train
380,248
wdcFile
66
76
72.63791
73
0.73914
{ "bin_edges": [ 66, 68, 70, 72, 74, 76, 76 ], "hist": [ 21, 649, 28400, 331546, 19626, 6 ] }
false
OxAISH-AL-LLM/pubmed_20k_rct
default
train
178,882
text
2
1,454
152.71467
140
78.70755
{ "bin_edges": [ 2, 148, 294, 440, 586, 732, 878, 1024, 1170, 1316, 1454 ], "hist": [ 97445, 72121, 8214, 907, 140, 34, 16, 3, 0, 2 ] }
false
OxAISH-AL-LLM/pubmed_20k_rct
default
validation
30,212
text
2
1,101
151.89537
138
79.74361
{ "bin_edges": [ 2, 112, 222, 332, 442, 552, 662, 772, 882, 992, 1101 ], "hist": [ 10310, 15057, 3953, 695, 139, 40, 10, 3, 2, 3 ] }
false
OxAISH-AL-LLM/pubmed_20k_rct
default
test
30,135
text
2
934
151.06411
138
78.95487
{ "bin_edges": [ 2, 96, 190, 284, 378, 472, 566, 660, 754, 848, 934 ], "hist": [ 7479, 15043, 5820, 1324, 336, 92, 24, 13, 2, 2 ] }
false
jamescalam/reddit-topics-targz
default
train
3,791
id
9
9
9
9
0
{ "bin_edges": [ 9, 9 ], "hist": [ 3791 ] }
false
jamescalam/reddit-topics-targz
default
train
3,791
selftext
3
29,985
741.81403
458
1,273.10073
{ "bin_edges": [ 3, 3002, 6001, 9000, 11999, 14998, 17997, 20996, 23995, 26994, 29985 ], "hist": [ 3689, 65, 24, 5, 4, 2, 0, 0, 1, 1 ] }
false
jamescalam/reddit-topics-targz
default
train
3,791
title
3
304
58.64389
52
35.08408
{ "bin_edges": [ 3, 34, 65, 96, 127, 158, 189, 220, 251, 282, 304 ], "hist": [ 764, 1841, 809, 231, 64, 39, 12, 5, 17, 9 ] }
false
evaluate/glue-ci
mnli
train
392,702
hypothesis
1
393
56.33299
53
25.04366
{ "bin_edges": [ 1, 41, 81, 121, 161, 201, 241, 281, 321, 361, 393 ], "hist": [ 107290, 228997, 48975, 6066, 1024, 241, 71, 25, 12, 1 ] }
false
evaluate/glue-ci
mnli
train
392,702
premise
5
1,815
113.66865
98
79.83146
{ "bin_edges": [ 5, 187, 369, 551, 733, 915, 1097, 1279, 1461, 1643, 1815 ], "hist": [ 334354, 54364, 3180, 549, 135, 51, 48, 15, 0, 6 ] }
false
evaluate/glue-ci
mnli
validation_matched
9,815
hypothesis
4
276
55.99185
53
25.306
{ "bin_edges": [ 4, 32, 60, 88, 116, 144, 172, 200, 228, 256, 276 ], "hist": [ 1409, 4715, 2704, 742, 168, 52, 17, 5, 1, 2 ] }
false
evaluate/glue-ci
mnli
validation_matched
9,815
premise
8
967
110.82588
94
80.05203
{ "bin_edges": [ 8, 104, 200, 296, 392, 488, 584, 680, 776, 872, 967 ], "hist": [ 5372, 3271, 953, 137, 47, 14, 6, 9, 0, 6 ] }
false
evaluate/glue-ci
mnli
validation_mismatched
9,832
hypothesis
6
316
62.22762
59
27.04235
{ "bin_edges": [ 6, 38, 70, 102, 134, 166, 198, 230, 262, 294, 316 ], "hist": [ 1601, 4965, 2536, 565, 123, 27, 5, 7, 1, 2 ] }
false
evaluate/glue-ci
mnli
validation_mismatched
9,832
premise
9
827
116.0241
102
78.39492
{ "bin_edges": [ 9, 91, 173, 255, 337, 419, 501, 583, 665, 747, 827 ], "hist": [ 4227, 3755, 1371, 348, 63, 29, 21, 9, 6, 3 ] }
false
evaluate/glue-ci
mnli
test_matched
9,796
hypothesis
3
291
55.98214
53
25.05101
{ "bin_edges": [ 3, 32, 61, 90, 119, 148, 177, 206, 235, 264, 291 ], "hist": [ 1390, 4807, 2758, 646, 132, 46, 4, 7, 4, 2 ] }
false
evaluate/glue-ci
mnli
test_matched
9,796
premise
8
992
112.7207
97
78.7166
{ "bin_edges": [ 8, 107, 206, 305, 404, 503, 602, 701, 800, 899, 992 ], "hist": [ 5426, 3270, 904, 131, 36, 9, 14, 3, 0, 3 ] }
false
evaluate/glue-ci
mnli
test_mismatched
9,847
hypothesis
3
403
62.44237
59
27.69715
{ "bin_edges": [ 3, 44, 85, 126, 167, 208, 249, 290, 331, 372, 403 ], "hist": [ 2335, 5874, 1377, 203, 35, 14, 6, 2, 0, 1 ] }
false
evaluate/glue-ci
mnli
test_mismatched
9,847
premise
9
1,005
115.65644
104
76.0861
{ "bin_edges": [ 9, 109, 209, 309, 409, 509, 609, 709, 809, 909, 1005 ], "hist": [ 5221, 3637, 783, 162, 32, 6, 0, 0, 0, 6 ] }
false
evaluate/glue-ci
mnli_matched
validation
9,815
hypothesis
4
276
55.99185
53
25.306
{ "bin_edges": [ 4, 32, 60, 88, 116, 144, 172, 200, 228, 256, 276 ], "hist": [ 1409, 4715, 2704, 742, 168, 52, 17, 5, 1, 2 ] }
false
evaluate/glue-ci
mnli_matched
validation
9,815
premise
8
967
110.82588
94
80.05203
{ "bin_edges": [ 8, 104, 200, 296, 392, 488, 584, 680, 776, 872, 967 ], "hist": [ 5372, 3271, 953, 137, 47, 14, 6, 9, 0, 6 ] }
false
evaluate/glue-ci
mnli_mismatched
validation
9,832
hypothesis
6
316
62.22762
59
27.04235
{ "bin_edges": [ 6, 38, 70, 102, 134, 166, 198, 230, 262, 294, 316 ], "hist": [ 1601, 4965, 2536, 565, 123, 27, 5, 7, 1, 2 ] }
false
evaluate/glue-ci
mnli_mismatched
validation
9,832
premise
9
827
116.0241
102
78.39492
{ "bin_edges": [ 9, 91, 173, 255, 337, 419, 501, 583, 665, 747, 827 ], "hist": [ 4227, 3755, 1371, 348, 63, 29, 21, 9, 6, 3 ] }
false
evaluate/glue-ci
mrpc
train
3,668
sentence1
38
226
118.48173
117
33.17004
{ "bin_edges": [ 38, 57, 76, 95, 114, 133, 152, 171, 190, 209, 226 ], "hist": [ 47, 276, 671, 724, 724, 563, 400, 203, 51, 9 ] }
false
evaluate/glue-ci
mrpc
train
3,668
sentence2
42
215
118.76908
116
33.4907
{ "bin_edges": [ 42, 60, 78, 96, 114, 132, 150, 168, 186, 204, 215 ], "hist": [ 66, 324, 614, 734, 671, 515, 399, 257, 74, 14 ] }
false
evaluate/glue-ci
mnli_mismatched
test
9,847
hypothesis
3
403
62.44237
59
27.69715
{ "bin_edges": [ 3, 44, 85, 126, 167, 208, 249, 290, 331, 372, 403 ], "hist": [ 2335, 5874, 1377, 203, 35, 14, 6, 2, 0, 1 ] }
false
evaluate/glue-ci
mnli_mismatched
test
9,847
premise
9
1,005
115.65644
104
76.0861
{ "bin_edges": [ 9, 109, 209, 309, 409, 509, 609, 709, 809, 909, 1005 ], "hist": [ 5221, 3637, 783, 162, 32, 6, 0, 0, 0, 6 ] }
false
OxAISH-AL-LLM/trec6
default
train
4,843
text
13
196
50.53459
48
20.63039
{ "bin_edges": [ 13, 32, 51, 70, 89, 108, 127, 146, 165, 184, 196 ], "hist": [ 827, 1875, 1372, 517, 189, 39, 16, 4, 3, 1 ] }
false
evaluate/glue-ci
mrpc
validation
408
sentence1
51
222
121.53431
118
34.84347
{ "bin_edges": [ 51, 69, 87, 105, 123, 141, 159, 177, 195, 213, 222 ], "hist": [ 24, 47, 73, 74, 63, 58, 46, 16, 6, 1 ] }
false
evaluate/glue-ci
mrpc
validation
408
sentence2
50
211
117.87745
116
32.5405
{ "bin_edges": [ 50, 67, 84, 101, 118, 135, 152, 169, 186, 203, 211 ], "hist": [ 13, 54, 68, 75, 72, 54, 43, 26, 2, 1 ] }
false
OxAISH-AL-LLM/trec6
default
validation
538
text
14
192
50.39591
46
21.36918
{ "bin_edges": [ 14, 32, 50, 68, 86, 104, 122, 140, 158, 176, 192 ], "hist": [ 85, 211, 138, 78, 16, 4, 5, 0, 0, 1 ] }
false
evaluate/glue-ci
mrpc
test
1,725
sentence1
37
222
118.21043
116
33.46556
{ "bin_edges": [ 37, 56, 75, 94, 113, 132, 151, 170, 189, 208, 222 ], "hist": [ 17, 151, 290, 349, 313, 264, 204, 123, 11, 3 ] }
false
evaluate/glue-ci
mrpc
test
1,725
sentence2
37
219
118.19362
115
33.51937
{ "bin_edges": [ 37, 56, 75, 94, 113, 132, 151, 170, 189, 208, 219 ], "hist": [ 17, 140, 291, 360, 313, 268, 204, 106, 24, 2 ] }
false
OxAISH-AL-LLM/trec6
default
test
500
text
13
91
35.91
33.5
14.12292
{ "bin_edges": [ 13, 21, 29, 37, 45, 53, 61, 69, 77, 85, 91 ], "hist": [ 63, 110, 112, 95, 63, 24, 15, 12, 5, 1 ] }
false
evaluate/glue-ci
qnli
train
104,743
question
1
270
58.53609
56
21.18957
{ "bin_edges": [ 1, 28, 55, 82, 109, 136, 163, 190, 217, 244, 270 ], "hist": [ 3525, 46628, 40804, 11262, 2086, 340, 82, 10, 2, 4 ] }
false
evaluate/glue-ci
qnli
train
104,743
sentence
3
2,458
165.63137
152
80.871
{ "bin_edges": [ 3, 249, 495, 741, 987, 1233, 1479, 1725, 1971, 2217, 2458 ], "hist": [ 91022, 13318, 364, 23, 11, 0, 0, 0, 0, 5 ] }
false
evaluate/glue-ci
qnli
validation
5,463
question
12
197
58.93465
56
21.57249
{ "bin_edges": [ 12, 31, 50, 69, 88, 107, 126, 145, 164, 183, 197 ], "hist": [ 311, 1656, 1993, 954, 379, 121, 33, 13, 1, 2 ] }
false
evaluate/glue-ci
qnli
validation
5,463
sentence
4
1,136
171.20886
154
86.70411
{ "bin_edges": [ 4, 118, 232, 346, 460, 574, 688, 802, 916, 1030, 1136 ], "hist": [ 1547, 2889, 790, 177, 52, 6, 1, 0, 0, 1 ] }
false
jonathanli/echr
non-anon
train
7,100
conclusion
12
1,803
104.24099
48
139.08554
{ "bin_edges": [ 12, 192, 372, 552, 732, 912, 1092, 1272, 1452, 1632, 1803 ], "hist": [ 5851, 975, 164, 52, 31, 13, 6, 5, 2, 1 ] }
false
jonathanli/echr
non-anon
train
7,100
docname
11
228
29.20986
27
11.52763
{ "bin_edges": [ 11, 33, 55, 77, 99, 121, 143, 165, 187, 209, 228 ], "hist": [ 5181, 1674, 200, 32, 9, 1, 1, 0, 0, 2 ] }
false
jonathanli/echr
non-anon
train
7,100
itemid
8
10
9.16648
9
0.509
{ "bin_edges": [ 8, 9, 10, 10 ], "hist": [ 427, 5064, 1609 ] }
false
jonathanli/echr
non-anon
train
7,100
judges
0
407
57.82563
34
54.32908
{ "bin_edges": [ 0, 41, 82, 123, 164, 205, 246, 287, 328, 369, 407 ], "hist": [ 3720, 510, 2347, 424, 0, 12, 21, 20, 33, 13 ] }
false
jonathanli/echr
non-anon
train
7,100
respondent
3
83
3.03887
3
1.20321
{ "bin_edges": [ 3, 12, 21, 30, 39, 48, 57, 66, 75, 83 ], "hist": [ 7097, 1, 0, 0, 0, 0, 1, 0, 1 ] }
false
jonathanli/echr
non-anon
test
2,998
conclusion
12
2,594
202.73215
142
198.16293
{ "bin_edges": [ 12, 271, 530, 789, 1048, 1307, 1566, 1825, 2084, 2343, 2594 ], "hist": [ 2214, 610, 128, 29, 6, 6, 2, 2, 0, 1 ] }
false
jonathanli/echr
non-anon
test
2,998
docname
11
141
31.29019
28
11.36784
{ "bin_edges": [ 11, 25, 39, 53, 67, 81, 95, 109, 123, 137, 141 ], "hist": [ 732, 1711, 423, 80, 35, 12, 1, 1, 1, 2 ] }
false
jonathanli/echr
non-anon
test
2,998
itemid
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 2998 ] }
false
jonathanli/echr
non-anon
test
2,998
judges
0
416
91.73382
91
50.62584
{ "bin_edges": [ 0, 42, 84, 126, 168, 210, 252, 294, 336, 378, 416 ], "hist": [ 375, 734, 1563, 242, 2, 10, 17, 18, 24, 13 ] }
false
jonathanli/echr
non-anon
test
2,998
respondent
3
19
3.02802
3
0.43136
{ "bin_edges": [ 3, 5, 7, 9, 11, 13, 15, 17, 19, 19 ], "hist": [ 2981, 0, 15, 0, 1, 0, 0, 0, 1 ] }
false
jonathanli/echr
non-anon
validation
1,380
conclusion
12
5,894
114.38768
49
216.77025
{ "bin_edges": [ 12, 601, 1190, 1779, 2368, 2957, 3546, 4135, 4724, 5313, 5894 ], "hist": [ 1354, 22, 3, 0, 0, 0, 0, 0, 0, 1 ] }
false
jonathanli/echr
non-anon
validation
1,380
docname
13
107
29.1087
27
10.68617
{ "bin_edges": [ 13, 23, 33, 43, 53, 63, 73, 83, 93, 103, 107 ], "hist": [ 318, 681, 267, 64, 27, 13, 6, 1, 1, 2 ] }
false
jonathanli/echr
non-anon
validation
1,380
itemid
8
10
9.18551
9
0.51272
{ "bin_edges": [ 8, 9, 10, 10 ], "hist": [ 77, 970, 333 ] }
false
jonathanli/echr
non-anon
validation
1,380
judges
0
393
57.20362
32
55.30136
{ "bin_edges": [ 0, 40, 80, 120, 160, 200, 240, 280, 320, 360, 393 ], "hist": [ 727, 99, 430, 103, 0, 2, 2, 5, 8, 4 ] }
false
jonathanli/echr
non-anon
validation
1,380
respondent
3
135
3.10435
3
3.55797
{ "bin_edges": [ 3, 17, 31, 45, 59, 73, 87, 101, 115, 129, 135 ], "hist": [ 1379, 0, 0, 0, 0, 0, 0, 0, 0, 1 ] }
false
cjvt/sentinews
sentence_level
train
168,899
content
1
2,160
123.45033
113
81.00786
{ "bin_edges": [ 1, 217, 433, 649, 865, 1081, 1297, 1513, 1729, 1945, 2160 ], "hist": [ 149017, 19210, 636, 33, 1, 0, 1, 0, 0, 1 ] }
false
cjvt/komet
default
train
13,963
document_name
14
19
17.22817
17
0.97766
{ "bin_edges": [ 14, 15, 16, 17, 18, 19, 19 ], "hist": [ 97, 490, 775, 9985, 0, 2616 ] }
false
cjvt/solar3
document_level
train
5,485
doc_title
27
46
34.48277
38
5.2723
{ "bin_edges": [ 27, 29, 31, 33, 35, 37, 39, 41, 43, 45, 46 ], "hist": [ 1661, 17, 446, 579, 0, 126, 2520, 135, 0, 1 ] }
false
cjvt/solar3
document_level
train
5,485
id_doc
6
9
8.79818
9
0.45143
{ "bin_edges": [ 6, 7, 8, 9, 9 ], "hist": [ 9, 90, 900, 4486 ] }
false
cjvt/solar3
paragraph_level
train
38,345
doc_title
27
46
33.77228
33
5.53206
{ "bin_edges": [ 27, 29, 31, 33, 35, 37, 39, 41, 43, 45, 46 ], "hist": [ 14861, 113, 2608, 2674, 0, 889, 16140, 1058, 0, 2 ] }
false
cjvt/solar3
paragraph_level
train
38,345
id_doc
6
9
8.77799
9
0.47674
{ "bin_edges": [ 6, 7, 8, 9, 9 ], "hist": [ 73, 827, 6640, 30805 ] }
false
cjvt/solar3
sentence_level
train
125,867
doc_title
27
46
33.57463
33
5.50083
{ "bin_edges": [ 27, 29, 31, 33, 35, 37, 39, 41, 43, 45, 46 ], "hist": [ 48443, 520, 10527, 11836, 0, 1807, 48386, 4321, 0, 27 ] }
false
cjvt/solar3
sentence_level
train
125,867
id_doc
6
9
8.70472
9
0.54477
{ "bin_edges": [ 6, 7, 8, 9, 9 ], "hist": [ 380, 4441, 27144, 93902 ] }
false
jonathanli/eurlex
eurlex57k
train
45,000
celex_id
10
14
10.05058
10
0.44694
{ "bin_edges": [ 10, 11, 12, 13, 14, 14 ], "hist": [ 44431, 0, 0, 0, 569 ] }
false
jonathanli/eurlex
eurlex57k
train
45,000
text
154
19,824
2,289.73169
1,787
1,676.30357
{ "bin_edges": [ 154, 2122, 4090, 6058, 8026, 9994, 11962, 13930, 15898, 17866, 19824 ], "hist": [ 28296, 12426, 2630, 895, 338, 212, 126, 72, 3, 2 ] }
false
jonathanli/eurlex
eurlex57k
train
45,000
title
9
1,277
219.06631
205
77.90162
{ "bin_edges": [ 9, 136, 263, 390, 517, 644, 771, 898, 1025, 1152, 1277 ], "hist": [ 4412, 29978, 9181, 1173, 188, 55, 10, 2, 0, 1 ] }
false
jonathanli/eurlex
eurlex57k
test
6,000
celex_id
10
14
10.052
10
0.45313
{ "bin_edges": [ 10, 11, 12, 13, 14, 14 ], "hist": [ 5922, 0, 0, 0, 78 ] }
false
jonathanli/eurlex
eurlex57k
test
6,000
text
453
15,348
2,284.4655
1,784
1,672.69015
{ "bin_edges": [ 453, 1943, 3433, 4923, 6413, 7903, 9393, 10883, 12373, 13863, 15348 ], "hist": [ 3430, 1733, 442, 204, 81, 45, 24, 18, 13, 10 ] }
false
jonathanli/eurlex
eurlex57k
test
6,000
title
9
894
217.484
205
75.56224
{ "bin_edges": [ 9, 98, 187, 276, 365, 454, 543, 632, 721, 810, 894 ], "hist": [ 54, 2316, 2514, 871, 190, 31, 12, 7, 3, 2 ] }
false
jonathanli/eurlex
eurlex57k
validation
6,000
celex_id
10
14
10.04667
10
0.42956
{ "bin_edges": [ 10, 11, 12, 13, 14, 14 ], "hist": [ 5930, 0, 0, 0, 70 ] }
false
jonathanli/eurlex
eurlex57k
validation
6,000
text
441
15,315
2,241.79617
1,755
1,622.50137
{ "bin_edges": [ 441, 1929, 3417, 4905, 6393, 7881, 9369, 10857, 12345, 13833, 15315 ], "hist": [ 3421, 1777, 456, 180, 63, 39, 26, 15, 9, 14 ] }
false
jonathanli/eurlex
eurlex57k
validation
6,000
title
63
1,532
218.97883
207
77.57782
{ "bin_edges": [ 63, 210, 357, 504, 651, 798, 945, 1092, 1239, 1386, 1532 ], "hist": [ 3094, 2608, 264, 26, 6, 1, 0, 0, 0, 1 ] }
false
THUDM/humaneval-x
cpp
test
164
canonical_solution
18
1,399
259.18293
205.5
197.36899
{ "bin_edges": [ 18, 157, 296, 435, 574, 713, 852, 991, 1130, 1269, 1399 ], "hist": [ 52, 61, 25, 19, 2, 1, 1, 2, 0, 1 ] }
false
THUDM/humaneval-x
cpp
test
164
declaration
111
515
157.84146
151
40.44885
{ "bin_edges": [ 111, 152, 193, 234, 275, 316, 357, 398, 439, 480, 515 ], "hist": [ 84, 70, 5, 2, 1, 1, 0, 0, 0, 1 ] }
false
THUDM/humaneval-x
cpp
test
164
example_test
0
679
251.43902
223.5
134.61967
{ "bin_edges": [ 0, 68, 136, 204, 272, 340, 408, 476, 544, 612, 679 ], "hist": [ 6, 27, 39, 23, 26, 21, 15, 3, 2, 2 ] }
false
THUDM/humaneval-x
cpp
test
164
prompt
133
1,349
481.18293
434.5
228.32689
{ "bin_edges": [ 133, 255, 377, 499, 621, 743, 865, 987, 1109, 1231, 1349 ], "hist": [ 21, 41, 47, 16, 17, 10, 6, 3, 1, 2 ] }
false
THUDM/humaneval-x
cpp
test
164
task_id
5
7
6.32927
6
0.58729
{ "bin_edges": [ 5, 6, 7, 7 ], "hist": [ 10, 90, 64 ] }
false
THUDM/humaneval-x
cpp
test
164
test
148
1,759
503.31098
447.5
250.63704
{ "bin_edges": [ 148, 310, 472, 634, 796, 958, 1120, 1282, 1444, 1606, 1759 ], "hist": [ 34, 56, 26, 33, 10, 1, 1, 1, 1, 1 ] }
false
THUDM/humaneval-x
go
test
164
canonical_solution
17
1,144
294.79878
266.5
200.82023
{ "bin_edges": [ 17, 130, 243, 356, 469, 582, 695, 808, 921, 1034, 1144 ], "hist": [ 30, 48, 35, 26, 14, 2, 5, 1, 2, 1 ] }
false
THUDM/humaneval-x
go
test
164
declaration
23
74
40.20732
39
9.60156
{ "bin_edges": [ 23, 29, 35, 41, 47, 53, 59, 65, 71, 74 ], "hist": [ 12, 37, 44, 38, 16, 10, 3, 2, 2 ] }
false
THUDM/humaneval-x
go
test
164
example_test
0
612
225.02439
210
104.45511
{ "bin_edges": [ 0, 62, 124, 186, 248, 310, 372, 434, 496, 558, 612 ], "hist": [ 6, 13, 45, 48, 24, 17, 5, 1, 3, 2 ] }
false
THUDM/humaneval-x
go
test
164
prompt
102
1,336
444.66463
398.5
230.63511
{ "bin_edges": [ 102, 226, 350, 474, 598, 722, 846, 970, 1094, 1218, 1336 ], "hist": [ 23, 37, 51, 14, 17, 12, 6, 1, 1, 2 ] }
false
THUDM/humaneval-x
go
test
164
task_id
4
6
5.32927
5
0.58729
{ "bin_edges": [ 4, 5, 6, 6 ], "hist": [ 10, 90, 64 ] }
false
THUDM/humaneval-x
go
test
164
test
149
1,944
495.83537
448.5
248.50577
{ "bin_edges": [ 149, 329, 509, 689, 869, 1049, 1229, 1409, 1589, 1769, 1944 ], "hist": [ 44, 54, 36, 22, 3, 2, 1, 1, 0, 1 ] }
false
THUDM/humaneval-x
java
test
164
canonical_solution
29
1,333
353.64634
302
245.8164
{ "bin_edges": [ 29, 160, 291, 422, 553, 684, 815, 946, 1077, 1208, 1333 ], "hist": [ 25, 52, 38, 25, 9, 6, 3, 1, 3, 2 ] }
false
THUDM/humaneval-x
java
test
164
declaration
86
791
121.56707
110
71.43646
{ "bin_edges": [ 86, 157, 228, 299, 370, 441, 512, 583, 654, 725, 791 ], "hist": [ 157, 3, 0, 0, 1, 2, 0, 0, 0, 1 ] }
false
THUDM/humaneval-x
java
test
164
example_test
0
1,387
450.95732
436
172.93651
{ "bin_edges": [ 0, 139, 278, 417, 556, 695, 834, 973, 1112, 1251, 1387 ], "hist": [ 6, 0, 61, 71, 20, 1, 2, 1, 0, 2 ] }
false
THUDM/humaneval-x
java
test
164
prompt
186
1,462
543.75
490
238.97422
{ "bin_edges": [ 186, 314, 442, 570, 698, 826, 954, 1082, 1210, 1338, 1462 ], "hist": [ 23, 40, 47, 18, 11, 16, 3, 3, 1, 2 ] }
false
THUDM/humaneval-x
java
test
164
task_id
6
8
7.32927
7
0.58729
{ "bin_edges": [ 6, 7, 8, 8 ], "hist": [ 10, 90, 64 ] }
false
THUDM/humaneval-x
java
test
164
test
370
2,830
792.58537
711.5
351.63569
{ "bin_edges": [ 370, 617, 864, 1111, 1358, 1605, 1852, 2099, 2346, 2593, 2830 ], "hist": [ 56, 58, 33, 8, 4, 1, 1, 2, 0, 1 ] }
false
THUDM/humaneval-x
js
test
164
canonical_solution
19
1,325
234.59756
186.5
182.2529
{ "bin_edges": [ 19, 150, 281, 412, 543, 674, 805, 936, 1067, 1198, 1325 ], "hist": [ 63, 52, 29, 10, 5, 3, 1, 0, 0, 1 ] }
false
THUDM/humaneval-x
js
test
164
declaration
20
400
38.35366
33
34.17094
{ "bin_edges": [ 20, 59, 98, 137, 176, 215, 254, 293, 332, 371, 400 ], "hist": [ 160, 0, 1, 1, 1, 0, 0, 0, 0, 1 ] }
false
THUDM/humaneval-x
js
test
164
example_test
0
655
235.84756
210.5
120.32087
{ "bin_edges": [ 0, 66, 132, 198, 264, 330, 396, 462, 528, 594, 655 ], "hist": [ 6, 16, 51, 34, 26, 15, 8, 3, 1, 4 ] }
false
THUDM/humaneval-x
js
test
164
prompt
96
1,314
421.71951
370
227.73711
{ "bin_edges": [ 96, 218, 340, 462, 584, 706, 828, 950, 1072, 1194, 1314 ], "hist": [ 28, 39, 48, 17, 12, 11, 2, 4, 1, 2 ] }
false
THUDM/humaneval-x
js
test
164
task_id
12
14
13.32927
13
0.58729
{ "bin_edges": [ 12, 13, 14, 14 ], "hist": [ 10, 90, 64 ] }
false
THUDM/humaneval-x
js
test
164
test
158
2,474
555.31098
459
323.88464
{ "bin_edges": [ 158, 390, 622, 854, 1086, 1318, 1550, 1782, 2014, 2246, 2474 ], "hist": [ 53, 56, 38, 7, 5, 2, 1, 0, 1, 1 ] }
false
THUDM/humaneval-x
python
test
164
canonical_solution
16
864
180.86585
148.5
139.52797
{ "bin_edges": [ 16, 101, 186, 271, 356, 441, 526, 611, 696, 781, 864 ], "hist": [ 53, 49, 31, 13, 10, 4, 1, 1, 1, 1 ] }
false
THUDM/humaneval-x
python
test
164
declaration
10
462
39.73171
26
47.1852
{ "bin_edges": [ 10, 56, 102, 148, 194, 240, 286, 332, 378, 424, 462 ], "hist": [ 137, 22, 1, 1, 1, 1, 0, 0, 0, 1 ] }
false
THUDM/humaneval-x
python
test
164
example_test
0
1,152
258.44512
208.5
182.37379
{ "bin_edges": [ 0, 116, 232, 348, 464, 580, 696, 812, 928, 1044, 1152 ], "hist": [ 25, 70, 32, 20, 8, 3, 3, 2, 0, 1 ] }
false
THUDM/humaneval-x
python
test
164
prompt
115
1,360
450.59756
396
230.30377
{ "bin_edges": [ 115, 240, 365, 490, 615, 740, 865, 990, 1115, 1240, 1360 ], "hist": [ 27, 40, 47, 21, 9, 11, 3, 3, 1, 2 ] }
false