dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=0.55, 'top_k'=10000, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=0.55, 'top_k'=10000, 'top_p'=1.0}
train
1,838
prediction
2
464
71.04788
66
46.73316
{ "bin_edges": [ 2, 49, 96, 143, 190, 237, 284, 331, 378, 425, 464 ], "hist": [ 708, 586, 409, 120, 10, 1, 2, 0, 1, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=1.05, 'top_k'=100, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=1.05, 'top_k'=100, 'top_p'=1.0}
train
1,838
prediction
2
461
68.68118
64
49.45141
{ "bin_edges": [ 2, 48, 94, 140, 186, 232, 278, 324, 370, 416, 461 ], "hist": [ 721, 623, 382, 93, 5, 3, 2, 2, 2, 5 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=1.05, 'top_k'=1000, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=1.05, 'top_k'=1000, 'top_p'=0.5}
train
1,838
prediction
9
477
73.61262
66
50.49249
{ "bin_edges": [ 9, 56, 103, 150, 197, 244, 291, 338, 385, 432, 477 ], "hist": [ 720, 637, 362, 106, 3, 2, 2, 1, 3, 2 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=1.05, 'top_k'=10000, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=1.05, 'top_k'=10000, 'top_p'=0.5}
train
1,838
prediction
9
463
73.53754
66
50.63486
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 463 ], "hist": [ 744, 593, 355, 129, 7, 3, 1, 1, 3, 2 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=1.05, 'top_k'=10000, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=1.05, 'top_k'=10000, 'top_p'=1.0}
train
1,838
prediction
2
489
67.25517
63
50.77925
{ "bin_edges": [ 2, 51, 100, 149, 198, 247, 296, 345, 394, 443, 489 ], "hist": [ 776, 654, 316, 74, 4, 3, 2, 2, 4, 3 ] }
false
Shubh8434/All
default
test
125
id
12
46
39.688
43
9.77358
{ "bin_edges": [ 12, 16, 20, 24, 28, 32, 36, 40, 44, 46 ], "hist": [ 8, 0, 13, 0, 0, 0, 0, 60, 44 ] }
false
Shubh8434/All
default
test
125
input
16,974
348,504
126,845.224
104,169
80,677.10054
{ "bin_edges": [ 16974, 50128, 83282, 116436, 149590, 182744, 215898, 249052, 282206, 315360, 348504 ], "hist": [ 23, 21, 26, 16, 11, 9, 4, 7, 6, 2 ] }
false
Shubh8434/All
default
train
1,000
id
12
46
37.607
43
11.65806
{ "bin_edges": [ 12, 16, 20, 24, 28, 32, 36, 40, 44, 46 ], "hist": [ 118, 0, 121, 0, 0, 0, 0, 395, 366 ] }
false
Shubh8434/All
default
train
1,000
input
3,993
449,240
134,574.48
114,361.5
88,604.5865
{ "bin_edges": [ 3993, 48518, 93043, 137568, 182093, 226618, 271143, 315668, 360193, 404718, 449240 ], "hist": [ 164, 235, 203, 123, 104, 84, 48, 16, 17, 6 ] }
false
Shubh8434/All
default
validation
125
id
12
46
36.808
43
11.98366
{ "bin_edges": [ 12, 16, 20, 24, 28, 32, 36, 40, 44, 46 ], "hist": [ 17, 0, 15, 0, 0, 0, 0, 54, 39 ] }
false
Shubh8434/All
default
validation
125
input
11,523
458,343
137,143.304
113,107
86,917.4884
{ "bin_edges": [ 11523, 56206, 100889, 145572, 190255, 234938, 279621, 324304, 368987, 413670, 458343 ], "hist": [ 18, 34, 24, 21, 12, 5, 6, 2, 2, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=1.05, 'top_k'=1000, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=10, 'temperature'=1.05, 'top_k'=1000, 'top_p'=1.0}
train
1,838
prediction
2
434
69.81121
64
51.60971
{ "bin_edges": [ 2, 46, 90, 134, 178, 222, 266, 310, 354, 398, 434 ], "hist": [ 681, 592, 417, 114, 16, 2, 3, 0, 6, 7 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=100, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=100, 'top_p'=0.5}
train
1,838
prediction
9
463
75.75952
69
47.86846
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 463 ], "hist": [ 651, 676, 373, 129, 5, 0, 1, 0, 2, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=100, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=100, 'top_p'=1.0}
train
1,838
prediction
9
463
75.51687
67
47.89967
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 463 ], "hist": [ 658, 662, 373, 140, 1, 0, 1, 0, 2, 1 ] }
false
lum-ai/metal-python-synthetic-explanations-gpt4-graphcodebert
default
train
4,021,376
id
16
22
20.67174
21
0.57179
{ "bin_edges": [ 16, 17, 18, 19, 20, 21, 22, 22 ], "hist": [ 117, 1495, 10089, 72431, 1239547, 2596517, 101180 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=1000, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=1000, 'top_p'=0.5}
train
1,838
prediction
9
463
75.86616
69
47.91001
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 463 ], "hist": [ 649, 675, 375, 130, 5, 0, 1, 0, 2, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=1000, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=1000, 'top_p'=1.0}
train
1,838
prediction
9
463
76.29434
73
47.78525
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 463 ], "hist": [ 648, 669, 377, 134, 6, 1, 1, 0, 1, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=10000, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=10000, 'top_p'=0.5}
train
1,838
prediction
9
463
75.70457
69
47.86701
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 463 ], "hist": [ 653, 673, 374, 129, 5, 0, 1, 0, 2, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=10000, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.05, 'top_k'=10000, 'top_p'=1.0}
train
1,838
prediction
9
463
76.41948
73
48.47314
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 463 ], "hist": [ 650, 662, 385, 132, 4, 0, 1, 1, 2, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=100, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=100, 'top_p'=0.5}
train
1,838
prediction
9
463
75.9679
70
47.71014
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 463 ], "hist": [ 651, 655, 400, 123, 5, 0, 2, 0, 1, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=100, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=100, 'top_p'=1.0}
train
1,838
prediction
9
459
72.91404
66
49.77151
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 459 ], "hist": [ 712, 644, 369, 92, 12, 1, 1, 1, 3, 3 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=1000, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=1000, 'top_p'=0.5}
train
1,838
prediction
9
468
75.8482
69
47.74026
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 468 ], "hist": [ 660, 654, 386, 129, 6, 0, 0, 0, 2, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=10000, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=10000, 'top_p'=0.5}
train
1,838
prediction
9
463
76.01687
72
47.96361
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 463 ], "hist": [ 654, 649, 404, 120, 7, 0, 1, 0, 2, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=1000, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=1000, 'top_p'=1.0}
train
1,838
prediction
9
486
72.57563
66
51.59642
{ "bin_edges": [ 9, 57, 105, 153, 201, 249, 297, 345, 393, 441, 486 ], "hist": [ 727, 689, 328, 74, 3, 5, 3, 5, 2, 2 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=100, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=100, 'top_p'=0.5}
train
1,838
prediction
9
462
74.38411
66
48.5079
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 462 ], "hist": [ 696, 644, 369, 114, 8, 3, 1, 0, 1, 2 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=10000, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.55, 'top_k'=10000, 'top_p'=1.0}
train
1,838
prediction
9
478
73.99075
66
51.38258
{ "bin_edges": [ 9, 56, 103, 150, 197, 244, 291, 338, 385, 432, 478 ], "hist": [ 704, 665, 343, 112, 4, 0, 1, 3, 5, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=1000, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=1000, 'top_p'=0.5}
train
1,838
prediction
9
475
74.04625
66
48.19442
{ "bin_edges": [ 9, 56, 103, 150, 197, 244, 291, 338, 385, 432, 475 ], "hist": [ 700, 647, 379, 100, 8, 1, 0, 0, 1, 2 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=100, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=100, 'top_p'=1.0}
train
1,838
prediction
2
471
70.10881
64
54.12873
{ "bin_edges": [ 2, 49, 96, 143, 190, 237, 284, 331, 378, 425, 471 ], "hist": [ 729, 634, 360, 90, 4, 1, 4, 4, 6, 6 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=1000, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=1000, 'top_p'=1.0}
train
1,838
prediction
2
484
69.14853
64
50.17667
{ "bin_edges": [ 2, 51, 100, 149, 198, 247, 296, 345, 394, 443, 484 ], "hist": [ 713, 720, 313, 68, 10, 3, 1, 5, 4, 1 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=10000, 'top_p'=0.5}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=10000, 'top_p'=0.5}
train
1,838
prediction
9
444
74.6605
66
48.30252
{ "bin_edges": [ 9, 53, 97, 141, 185, 229, 273, 317, 361, 405, 444 ], "hist": [ 703, 532, 429, 155, 13, 0, 4, 0, 0, 2 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=10000, 'top_p'=1.0}
train
1,838
id
16
20
19.36779
19
0.62229
{ "bin_edges": [ 16, 17, 18, 19, 20, 20 ], "hist": [ 2, 11, 97, 927, 801 ] }
false
automated-research-group/llama2_7b_chat-piqa-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=10000, 'top_p'=1.0}
train
1,838
prediction
2
454
68.65724
63
51.25137
{ "bin_edges": [ 2, 48, 94, 140, 186, 232, 278, 324, 370, 416, 454 ], "hist": [ 727, 635, 360, 89, 8, 2, 5, 1, 5, 6 ] }
false
gayanin/pubmed-abstracts
default
test
9,341
refs
16
753
125.19302
116
59.47267
{ "bin_edges": [ 16, 90, 164, 238, 312, 386, 460, 534, 608, 682, 753 ], "hist": [ 2786, 4600, 1522, 347, 55, 20, 6, 0, 2, 3 ] }
false
gayanin/pubmed-abstracts
default
train
74,724
refs
14
911
122.05562
112
57.02703
{ "bin_edges": [ 14, 104, 194, 284, 374, 464, 554, 644, 734, 824, 911 ], "hist": [ 32048, 34725, 6959, 827, 122, 26, 10, 3, 1, 3 ] }
false
gayanin/pubmed-abstracts
default
validation
9,341
refs
18
619
128.72027
119
59.5242
{ "bin_edges": [ 18, 79, 140, 201, 262, 323, 384, 445, 506, 567, 619 ], "hist": [ 1820, 4106, 2337, 819, 188, 47, 12, 9, 1, 2 ] }
false
Atipico1/nq-output-entities2
default
train
10,000
context
31
447
98.8191
95
30.81515
{ "bin_edges": [ 31, 73, 115, 157, 199, 241, 283, 325, 367, 409, 447 ], "hist": [ 1888, 5612, 2031, 390, 68, 6, 0, 3, 0, 2 ] }
false
Atipico1/nq-output-entities2
default
train
10,000
context_vague
69
377
165.8996
163
33.7049
{ "bin_edges": [ 69, 100, 131, 162, 193, 224, 255, 286, 317, 348, 377 ], "hist": [ 79, 1317, 3426, 3171, 1453, 455, 74, 15, 6, 4 ] }
false
Atipico1/nq-output-entities2
default
train
10,000
masked_query
7
110
44.7754
43
11.38636
{ "bin_edges": [ 7, 18, 29, 40, 51, 62, 73, 84, 95, 106, 110 ], "hist": [ 10, 468, 2741, 4564, 1452, 461, 187, 85, 30, 2 ] }
false
Atipico1/nq-output-entities2
default
train
10,000
question
28
101
48.0422
46
10.58009
{ "bin_edges": [ 28, 36, 44, 52, 60, 68, 76, 84, 92, 100, 101 ], "hist": [ 371, 3486, 3520, 1467, 581, 257, 160, 91, 56, 11 ] }
false
Atipico1/nq-output-entities2
default
train
10,000
similar_entity
2
72
13.48364
13
5.29934
{ "bin_edges": [ 2, 10, 18, 26, 34, 42, 50, 58, 66, 72 ], "hist": [ 1223, 4603, 857, 151, 26, 9, 5, 0, 1 ] }
false
Atipico1/nq-output-entities2
default
test
3,610
context
36
492
102.21551
96
33.63829
{ "bin_edges": [ 36, 82, 128, 174, 220, 266, 312, 358, 404, 450, 492 ], "hist": [ 1038, 1896, 545, 112, 13, 5, 0, 0, 0, 1 ] }
false
Atipico1/nq-output-entities2
default
test
3,610
context_vague
79
375
168.0892
164
35.18676
{ "bin_edges": [ 79, 109, 139, 169, 199, 229, 259, 289, 319, 349, 375 ], "hist": [ 71, 677, 1232, 961, 485, 133, 37, 9, 3, 2 ] }
false
Atipico1/nq-output-entities2
default
test
3,610
masked_query
14
102
45.81607
44
11.85742
{ "bin_edges": [ 14, 23, 32, 41, 50, 59, 68, 77, 86, 95, 102 ], "hist": [ 28, 225, 956, 1386, 590, 214, 112, 60, 29, 10 ] }
false
Atipico1/nq-output-entities2
default
test
3,610
question
26
101
48.74654
46
11.00275
{ "bin_edges": [ 26, 34, 42, 50, 58, 66, 74, 82, 90, 98, 101 ], "hist": [ 41, 863, 1442, 692, 281, 138, 80, 38, 26, 9 ] }
false
Atipico1/nq-output-entities2
default
test
3,610
similar_entity
2
62
13.46009
13
5.44502
{ "bin_edges": [ 2, 9, 16, 23, 30, 37, 44, 51, 58, 62 ], "hist": [ 325, 1284, 513, 74, 24, 5, 3, 1, 1 ] }
false
SarthakG/smart_contract
default
train
789
contract
2
32
11.35234
11
4.07482
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 32 ], "hist": [ 66, 140, 398, 117, 47, 19, 1, 1 ] }
false
winder-hybrids/MedicalTextbook_QA
Anatomy_Gray
test
500
original_text
2,724
4,270
3,590.22
3,624.5
320.79437
{ "bin_edges": [ 2724, 2879, 3034, 3189, 3344, 3499, 3654, 3809, 3964, 4119, 4270 ], "hist": [ 20, 30, 10, 10, 80, 130, 100, 80, 30, 10 ] }
false
winder-hybrids/MedicalTextbook_QA
Anatomy_Gray
test
500
question
32
175
78.626
75
22.36924
{ "bin_edges": [ 32, 47, 62, 77, 92, 107, 122, 137, 152, 167, 175 ], "hist": [ 20, 98, 145, 108, 69, 39, 13, 7, 0, 1 ] }
false
winder-hybrids/MedicalTextbook_QA
Biochemistry_Lippincott
test
500
original_text
1,671
3,795
3,095.56
3,211.5
499.32187
{ "bin_edges": [ 1671, 1884, 2097, 2310, 2523, 2736, 2949, 3162, 3375, 3588, 3795 ], "hist": [ 30, 0, 10, 20, 50, 20, 80, 150, 90, 50 ] }
false
winder-hybrids/MedicalTextbook_QA
Biochemistry_Lippincott
test
500
question
21
152
82.428
80
24.60943
{ "bin_edges": [ 21, 35, 49, 63, 77, 91, 105, 119, 133, 147, 152 ], "hist": [ 5, 27, 82, 115, 92, 89, 47, 27, 12, 4 ] }
false
winder-hybrids/MedicalTextbook_QA
Gynecology_Novak
test
500
original_text
2,238
4,516
3,197.84
3,233
644.94449
{ "bin_edges": [ 2238, 2466, 2694, 2922, 3150, 3378, 3606, 3834, 4062, 4290, 4516 ], "hist": [ 90, 70, 10, 40, 90, 50, 70, 30, 20, 30 ] }
false
winder-hybrids/MedicalTextbook_QA
Gynecology_Novak
test
500
question
25
187
96.438
94.5
26.65181
{ "bin_edges": [ 25, 42, 59, 76, 93, 110, 127, 144, 161, 178, 187 ], "hist": [ 2, 31, 82, 126, 114, 70, 51, 18, 5, 1 ] }
false
winder-hybrids/MedicalTextbook_QA
Cell_Biology_Alberts
test
500
original_text
1,768
4,391
3,574.54
3,632.5
465.57522
{ "bin_edges": [ 1768, 2031, 2294, 2557, 2820, 3083, 3346, 3609, 3872, 4135, 4391 ], "hist": [ 10, 0, 0, 20, 40, 60, 80, 160, 90, 40 ] }
false
winder-hybrids/MedicalTextbook_QA
Cell_Biology_Alberts
test
500
question
18
164
85.252
83
23.47447
{ "bin_edges": [ 18, 33, 48, 63, 78, 93, 108, 123, 138, 153, 164 ], "hist": [ 2, 14, 64, 122, 111, 101, 54, 20, 9, 3 ] }
false
winder-hybrids/MedicalTextbook_QA
Neurology_Adams
test
500
original_text
2,616
4,562
3,554.76
3,563
407.7498
{ "bin_edges": [ 2616, 2811, 3006, 3201, 3396, 3591, 3786, 3981, 4176, 4371, 4562 ], "hist": [ 10, 40, 80, 60, 70, 50, 130, 50, 0, 10 ] }
false
winder-hybrids/MedicalTextbook_QA
Neurology_Adams
test
500
question
29
173
94.934
93
24.31263
{ "bin_edges": [ 29, 44, 59, 74, 89, 104, 119, 134, 149, 164, 173 ], "hist": [ 4, 27, 63, 108, 132, 79, 58, 16, 8, 5 ] }
false
winder-hybrids/MedicalTextbook_QA
Immunology_Janeway
test
500
original_text
1,325
4,070
3,092.88
3,119
648.07236
{ "bin_edges": [ 1325, 1600, 1875, 2150, 2425, 2700, 2975, 3250, 3525, 3800, 4070 ], "hist": [ 10, 0, 0, 100, 50, 60, 40, 60, 100, 80 ] }
false
winder-hybrids/MedicalTextbook_QA
Immunology_Janeway
test
500
question
27
166
85.488
83
24.73413
{ "bin_edges": [ 27, 41, 55, 69, 83, 97, 111, 125, 139, 153, 166 ], "hist": [ 9, 34, 97, 105, 95, 79, 47, 25, 4, 5 ] }
false
winder-hybrids/MedicalTextbook_QA
Histology_Ross
test
500
original_text
1,188
4,199
3,409.14
3,500.5
559.72821
{ "bin_edges": [ 1188, 1490, 1792, 2094, 2396, 2698, 3000, 3302, 3604, 3906, 4199 ], "hist": [ 20, 0, 0, 10, 0, 20, 80, 150, 170, 50 ] }
false
winder-hybrids/MedicalTextbook_QA
Histology_Ross
test
500
question
20
152
80.566
79
23.1259
{ "bin_edges": [ 20, 34, 48, 62, 76, 90, 104, 118, 132, 146, 152 ], "hist": [ 3, 27, 87, 113, 100, 82, 55, 24, 6, 3 ] }
false
winder-hybrids/MedicalTextbook_QA
Obstentrics_Williams
test
500
original_text
1,304
3,897
2,973.82
3,208.5
534.2679
{ "bin_edges": [ 1304, 1564, 1824, 2084, 2344, 2604, 2864, 3124, 3384, 3644, 3897 ], "hist": [ 20, 0, 10, 20, 70, 70, 50, 150, 100, 10 ] }
false
winder-hybrids/MedicalTextbook_QA
Obstentrics_Williams
test
500
question
23
184
99.55
98.5
25.20782
{ "bin_edges": [ 23, 40, 57, 74, 91, 108, 125, 142, 159, 176, 184 ], "hist": [ 1, 18, 55, 105, 138, 113, 37, 23, 8, 2 ] }
false
winder-hybrids/MedicalTextbook_QA
Pathology_Robbins
test
500
original_text
2,017
4,103
3,273.18
3,275
455.50877
{ "bin_edges": [ 2017, 2226, 2435, 2644, 2853, 3062, 3271, 3480, 3689, 3898, 4103 ], "hist": [ 10, 10, 20, 40, 90, 70, 80, 50, 90, 40 ] }
false
winder-hybrids/MedicalTextbook_QA
Pathology_Robbins
test
500
question
20
161
85.184
84
24.46473
{ "bin_edges": [ 20, 35, 50, 65, 80, 95, 110, 125, 140, 155, 161 ], "hist": [ 5, 22, 74, 116, 125, 72, 55, 17, 12, 2 ] }
false
winder-hybrids/MedicalTextbook_QA
Pediatrics_Nelson
test
500
original_text
2,309
4,597
3,560.38
3,573
361.51734
{ "bin_edges": [ 2309, 2538, 2767, 2996, 3225, 3454, 3683, 3912, 4141, 4370, 4597 ], "hist": [ 10, 0, 20, 40, 100, 140, 130, 40, 10, 10 ] }
false
winder-hybrids/MedicalTextbook_QA
Pediatrics_Nelson
test
500
question
32
202
88.316
87
25.84174
{ "bin_edges": [ 32, 50, 68, 86, 104, 122, 140, 158, 176, 194, 202 ], "hist": [ 13, 100, 123, 140, 75, 26, 16, 5, 1, 1 ] }
false
winder-hybrids/MedicalTextbook_QA
Pharmacology_Katzung
test
500
original_text
2,254
4,500
3,276.72
3,264
556.71237
{ "bin_edges": [ 2254, 2479, 2704, 2929, 3154, 3379, 3604, 3829, 4054, 4279, 4500 ], "hist": [ 50, 40, 70, 30, 80, 70, 80, 50, 20, 10 ] }
false
winder-hybrids/MedicalTextbook_QA
Pharmacology_Katzung
test
500
question
32
177
88.93
88
28.13589
{ "bin_edges": [ 32, 47, 62, 77, 92, 107, 122, 137, 152, 167, 177 ], "hist": [ 28, 60, 84, 108, 95, 63, 34, 16, 9, 3 ] }
false
winder-hybrids/MedicalTextbook_QA
Physiology_Levy
test
500
original_text
1,946
3,959
3,261.62
3,336.5
457.93174
{ "bin_edges": [ 1946, 2148, 2350, 2552, 2754, 2956, 3158, 3360, 3562, 3764, 3959 ], "hist": [ 30, 0, 10, 20, 40, 50, 120, 100, 90, 40 ] }
false
winder-hybrids/MedicalTextbook_QA
Physiology_Levy
test
500
question
16
182
83.508
80
26.01822
{ "bin_edges": [ 16, 33, 50, 67, 84, 101, 118, 135, 152, 169, 182 ], "hist": [ 5, 21, 113, 135, 117, 56, 26, 20, 6, 1 ] }
false
winder-hybrids/MedicalTextbook_QA
Psichiatry_DSM-5
test
500
original_text
1,898
4,468
3,704.62
3,917.5
665.43553
{ "bin_edges": [ 1898, 2156, 2414, 2672, 2930, 3188, 3446, 3704, 3962, 4220, 4468 ], "hist": [ 30, 20, 0, 20, 20, 30, 40, 120, 130, 90 ] }
false
winder-hybrids/MedicalTextbook_QA
Psichiatry_DSM-5
test
500
question
29
200
96.068
92.5
28.53025
{ "bin_edges": [ 29, 47, 65, 83, 101, 119, 137, 155, 173, 191, 200 ], "hist": [ 8, 52, 121, 142, 65, 67, 29, 11, 4, 1 ] }
false
A-Bar/nl-de_non_top_cs_train
default
train
1,000,000
passage
34
1,421
355.22233
315
142.16478
{ "bin_edges": [ 34, 173, 312, 451, 590, 729, 868, 1007, 1146, 1285, 1421 ], "hist": [ 43092, 438152, 288773, 149617, 62894, 14312, 2669, 403, 77, 11 ] }
false
A-Bar/nl-de_non_top_cs_train
default
train
1,000,000
query
6
231
35.80702
34
13.47931
{ "bin_edges": [ 6, 29, 52, 75, 98, 121, 144, 167, 190, 213, 231 ], "hist": [ 309468, 586948, 88620, 11608, 2228, 644, 276, 144, 60, 4 ] }
false
atmallen/quirky_multiplication_increment0_bob_easy
default
test
986
statement
15
22
20.0213
20
1.06182
{ "bin_edges": [ 15, 16, 17, 18, 19, 20, 21, 22, 22 ], "hist": [ 3, 3, 8, 69, 160, 428, 260, 55 ] }
false
atmallen/quirky_multiplication_increment0_bob_easy
default
train
47,510
statement
15
22
20.01478
20
1.04527
{ "bin_edges": [ 15, 16, 17, 18, 19, 20, 21, 22, 22 ], "hist": [ 46, 153, 756, 2490, 8683, 20606, 12077, 2699 ] }
false
atmallen/quirky_multiplication_increment0_bob_easy
default
validation
966
statement
16
22
20.02692
20
1.02371
{ "bin_edges": [ 16, 17, 18, 19, 20, 21, 22, 22 ], "hist": [ 7, 8, 46, 182, 426, 242, 55 ] }
false