dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
lum-ai/metal-python-synthetic-explanations-gpt4-raw
default
validation
18,272
code
107
19,176
3,855.77556
2,716
3,708.6061
{ "bin_edges": [ 107, 2014, 3921, 5828, 7735, 9642, 11549, 13456, 15363, 17270, 19176 ], "hist": [ 7582, 3975, 2875, 1399, 839, 474, 501, 381, 163, 83 ] }
false
lum-ai/metal-python-synthetic-explanations-gpt4-raw
default
validation
18,272
id
16
22
19.78815
20
0.85533
{ "bin_edges": [ 16, 17, 18, 19, 20, 21, 22, 22 ], "hist": [ 10, 76, 1090, 5201, 8119, 3774, 2 ] }
false
lum-ai/metal-python-synthetic-explanations-gpt4-raw
default
validation
18,272
raw_text
425
9,295
2,447.23057
2,288
1,175.58634
{ "bin_edges": [ 425, 1313, 2201, 3089, 3977, 4865, 5753, 6641, 7529, 8417, 9295 ], "hist": [ 1917, 6496, 6345, 1995, 815, 333, 206, 0, 0, 165 ] }
false
lum-ai/metal-python-synthetic-explanations-gpt4-raw
default
validation
18,272
text
388
8,421
2,260.68307
2,104
1,094.74757
{ "bin_edges": [ 388, 1192, 1996, 2800, 3604, 4408, 5212, 6016, 6820, 7624, 8421 ], "hist": [ 1985, 6106, 6347, 2247, 883, 275, 189, 75, 0, 165 ] }
false
tomekkorbak/pile-pii
default
test
10,000
text
12
20,390
2,967.7991
1,744
3,409.08946
{ "bin_edges": [ 12, 2050, 4088, 6126, 8164, 10202, 12240, 14278, 16316, 18354, 20390 ], "hist": [ 5747, 2108, 842, 466, 330, 165, 104, 97, 77, 64 ] }
false
lum-ai/metal-python-synthetic-explanations-gpt4-raw
default
test
35,131
chunk_id
15
18
17.29
17
0.67722
{ "bin_edges": [ 15, 16, 17, 18, 18 ], "hist": [ 392, 3263, 17241, 14235 ] }
false
lum-ai/metal-python-synthetic-explanations-gpt4-raw
default
test
35,131
code
92
20,432
3,851.83428
2,601
3,775.85523
{ "bin_edges": [ 92, 2127, 4162, 6197, 8232, 10267, 12302, 14337, 16372, 18407, 20432 ], "hist": [ 15178, 8721, 4331, 2520, 1320, 1044, 1194, 486, 258, 79 ] }
false
lum-ai/metal-python-synthetic-explanations-gpt4-raw
default
test
35,131
filename
6
112
39.28741
33
21.88674
{ "bin_edges": [ 6, 17, 28, 39, 50, 61, 72, 83, 94, 105, 112 ], "hist": [ 4106, 9573, 6681, 4800, 3376, 1795, 3359, 1227, 193, 21 ] }
false
lum-ai/metal-python-synthetic-explanations-gpt4-raw
default
test
35,131
id
17
22
19.78011
20
0.84065
{ "bin_edges": [ 17, 18, 19, 20, 21, 22, 22 ], "hist": [ 205, 1831, 10387, 15784, 6909, 15 ] }
false
lum-ai/metal-python-synthetic-explanations-gpt4-raw
default
test
35,131
raw_text
313
11,868
2,441.22129
2,247
1,341.18285
{ "bin_edges": [ 313, 1469, 2625, 3781, 4937, 6093, 7249, 8405, 9561, 10717, 11868 ], "hist": [ 5424, 18583, 8411, 1713, 277, 325, 0, 0, 103, 295 ] }
false
lum-ai/metal-python-synthetic-explanations-gpt4-raw
default
test
35,131
text
293
10,518
2,270.77533
2,098
1,217.46009
{ "bin_edges": [ 293, 1316, 2339, 3362, 4385, 5408, 6431, 7454, 8477, 9500, 10518 ], "hist": [ 4401, 18110, 9516, 1995, 386, 96, 229, 0, 0, 398 ] }
false
BSC-LT/bsc-dolly-15k-en
annotated
train
15,015
context
41
22,685
1,141.70951
777
1,390.93668
{ "bin_edges": [ 41, 2306, 4571, 6836, 9101, 11366, 13631, 15896, 18161, 20426, 22685 ], "hist": [ 4106, 295, 66, 26, 5, 7, 3, 1, 1, 3 ] }
false
BSC-LT/bsc-dolly-15k-en
annotated
train
15,015
instruction
4
11,698
71.6993
54
131.28645
{ "bin_edges": [ 4, 1174, 2344, 3514, 4684, 5854, 7024, 8194, 9364, 10534, 11698 ], "hist": [ 14994, 16, 4, 0, 0, 0, 0, 0, 0, 1 ] }
false
BSC-LT/bsc-dolly-15k-en
annotated
train
15,015
response
1
9,796
333.81252
178
503.10115
{ "bin_edges": [ 1, 981, 1961, 2941, 3921, 4901, 5881, 6861, 7841, 8821, 9796 ], "hist": [ 14079, 722, 128, 48, 18, 4, 5, 4, 4, 3 ] }
false
arrafmousa/SimAPI
default
validation
314
answer
1
397
44.48243
11
64.42841
{ "bin_edges": [ 1, 41, 81, 121, 161, 201, 241, 281, 321, 361, 397 ], "hist": [ 222, 19, 26, 22, 14, 5, 2, 1, 1, 1 ] }
false
arrafmousa/SimAPI
default
validation
314
code
187
1,092
465.65287
368
180.32698
{ "bin_edges": [ 187, 278, 369, 460, 551, 642, 733, 824, 915, 1006, 1092 ], "hist": [ 28, 131, 54, 5, 26, 35, 20, 12, 2, 1 ] }
false
arrafmousa/SimAPI
default
validation
314
context
321
1,019
710.65287
691
135.06114
{ "bin_edges": [ 321, 391, 461, 531, 601, 671, 741, 811, 881, 951, 1019 ], "hist": [ 2, 3, 18, 42, 76, 53, 38, 40, 30, 12 ] }
false
arrafmousa/SimAPI
default
validation
314
prompt
393
1,271
808.40127
777.5
153.9947
{ "bin_edges": [ 393, 481, 569, 657, 745, 833, 921, 1009, 1097, 1185, 1271 ], "hist": [ 2, 9, 35, 80, 70, 39, 41, 25, 8, 5 ] }
false
arrafmousa/SimAPI
default
validation
314
question
35
341
96.74841
82
47.38744
{ "bin_edges": [ 35, 66, 97, 128, 159, 190, 221, 252, 283, 314, 341 ], "hist": [ 44, 170, 37, 32, 13, 10, 4, 2, 0, 2 ] }
false
arrafmousa/SimAPI
default
validation
314
text
446
1,693
877.87898
844
185.93076
{ "bin_edges": [ 446, 571, 696, 821, 946, 1071, 1196, 1321, 1446, 1571, 1693 ], "hist": [ 6, 37, 97, 75, 51, 29, 14, 3, 1, 1 ] }
false
arrafmousa/SimAPI
default
train
2,515
answer
1
329
46.01647
11
60.3117
{ "bin_edges": [ 1, 34, 67, 100, 133, 166, 199, 232, 265, 298, 329 ], "hist": [ 1635, 154, 254, 163, 138, 66, 41, 27, 6, 5 ] }
false
arrafmousa/SimAPI
default
train
2,515
code
188
1,075
478.96461
377
177.14561
{ "bin_edges": [ 188, 277, 366, 455, 544, 633, 722, 811, 900, 989, 1075 ], "hist": [ 163, 851, 571, 64, 215, 323, 219, 78, 25, 6 ] }
false
arrafmousa/SimAPI
default
train
2,515
context
301
1,025
713.9165
699
140.07676
{ "bin_edges": [ 301, 374, 447, 520, 593, 666, 739, 812, 885, 958, 1025 ], "hist": [ 9, 21, 147, 339, 531, 418, 355, 326, 262, 107 ] }
false
arrafmousa/SimAPI
default
train
2,515
prompt
384
1,279
812.6827
795
155.2516
{ "bin_edges": [ 384, 474, 564, 654, 744, 834, 924, 1014, 1104, 1194, 1279 ], "hist": [ 12, 73, 288, 599, 484, 392, 370, 218, 57, 22 ] }
false
arrafmousa/SimAPI
default
train
2,515
question
31
345
97.7662
82
44.81423
{ "bin_edges": [ 31, 63, 95, 127, 159, 191, 223, 255, 287, 319, 345 ], "hist": [ 199, 1422, 390, 232, 144, 73, 28, 14, 9, 4 ] }
false
arrafmousa/SimAPI
default
train
2,515
text
412
1,633
883.39324
858
180.97766
{ "bin_edges": [ 412, 535, 658, 781, 904, 1027, 1150, 1273, 1396, 1519, 1633 ], "hist": [ 19, 172, 654, 637, 449, 369, 161, 39, 10, 5 ] }
false
SkunkworksAI-shared/concatenated_1
default
train
2,527,636
text
8
606,172
1,695.13901
1,252
2,241.39225
{ "bin_edges": [ 8, 60625, 121242, 181859, 242476, 303093, 363710, 424327, 484944, 545561, 606172 ], "hist": [ 2527426, 144, 46, 12, 2, 0, 2, 2, 0, 2 ] }
false
SkunkworksAI-shared/concatenated_1
default
train
2,527,636
unique_conversation_id
19
65
45.77963
35
15.2592
{ "bin_edges": [ 19, 24, 29, 34, 39, 44, 49, 54, 59, 64, 65 ], "hist": [ 3130, 47411, 1174526, 211256, 6106, 10000, 133000, 0, 10000, 932207 ] }
false
arrafmousa/SimAPI
default
test
315
answer
1
290
44.04516
11
61.85894
{ "bin_edges": [ 1, 30, 59, 88, 117, 146, 175, 204, 233, 262, 290 ], "hist": [ 217, 6, 22, 20, 14, 10, 12, 5, 3, 1 ] }
false
arrafmousa/SimAPI
default
test
315
code
188
1,062
466.85397
369
173.65935
{ "bin_edges": [ 188, 276, 364, 452, 540, 628, 716, 804, 892, 980, 1062 ], "hist": [ 21, 97, 91, 11, 23, 27, 34, 9, 0, 2 ] }
false
arrafmousa/SimAPI
default
test
315
context
347
999
717.07302
699
134.38154
{ "bin_edges": [ 347, 413, 479, 545, 611, 677, 743, 809, 875, 941, 999 ], "hist": [ 1, 11, 16, 46, 59, 52, 40, 38, 38, 14 ] }
false
arrafmousa/SimAPI
default
test
315
prompt
382
1,338
813.14921
795
149.53393
{ "bin_edges": [ 382, 478, 574, 670, 766, 862, 958, 1054, 1150, 1246, 1338 ], "hist": [ 1, 13, 39, 81, 61, 58, 46, 15, 0, 1 ] }
false
arrafmousa/SimAPI
default
test
315
question
32
351
95.07619
82
42.99201
{ "bin_edges": [ 32, 64, 96, 128, 160, 192, 224, 256, 288, 320, 351 ], "hist": [ 27, 187, 44, 28, 19, 6, 2, 0, 1, 1 ] }
false
arrafmousa/SimAPI
default
test
315
text
412
1,368
881.68889
852
174.82535
{ "bin_edges": [ 412, 508, 604, 700, 796, 892, 988, 1084, 1180, 1276, 1368 ], "hist": [ 1, 8, 34, 73, 66, 44, 35, 44, 4, 6 ] }
false
BSC-LT/bsc-dolly-15k-en
filtered
train
10,157
instruction
9
464
66.83391
52
52.14765
{ "bin_edges": [ 9, 55, 101, 147, 193, 239, 285, 331, 377, 423, 464 ], "hist": [ 5447, 3130, 843, 365, 185, 98, 38, 25, 11, 15 ] }
false
BSC-LT/bsc-dolly-15k-en
filtered
train
10,157
response
1
1,842
318.97381
201
335.69371
{ "bin_edges": [ 1, 186, 371, 556, 741, 926, 1111, 1296, 1481, 1666, 1842 ], "hist": [ 4863, 2102, 1393, 690, 440, 246, 163, 118, 88, 54 ] }
false
Norquinal/claude_multiround_chat_1k
default
train
1,609
id
8
8
8
8
0
{ "bin_edges": [ 8, 8 ], "hist": [ 1609 ] }
false
google/mittens
default
test
3,632
example
13
1,353
178.30589
93
199.24369
{ "bin_edges": [ 13, 148, 283, 418, 553, 688, 823, 958, 1093, 1228, 1353 ], "hist": [ 2390, 654, 125, 190, 151, 62, 21, 26, 4, 9 ] }
false
google/mittens
default
test
3,632
example_hash
56
56
56
56
0
{ "bin_edges": [ 56, 56 ], "hist": [ 3632 ] }
false
google/mittens
default
test
3,632
inputs
86
1,424
250.51404
164
199.24631
{ "bin_edges": [ 86, 220, 354, 488, 622, 756, 890, 1024, 1158, 1292, 1424 ], "hist": [ 2361, 680, 124, 192, 150, 64, 21, 26, 3, 11 ] }
false
google/mittens
default
test
3,632
model_input_hash
56
56
56
56
0
{ "bin_edges": [ 56, 56 ], "hist": [ 3632 ] }
false
google/mittens
default
test
3,632
targets
13
1,277
181.35793
83
209.11123
{ "bin_edges": [ 13, 140, 267, 394, 521, 648, 775, 902, 1029, 1156, 1277 ], "hist": [ 2365, 565, 115, 203, 224, 80, 16, 48, 0, 16 ] }
false
elsheikhams/q2q_similarity_workshop
default
train
11,997
question1
10
85
30.65733
29
9.94182
{ "bin_edges": [ 10, 18, 26, 34, 42, 50, 58, 66, 74, 82, 85 ], "hist": [ 669, 3360, 3979, 2377, 1019, 400, 152, 33, 6, 2 ] }
false
elsheikhams/q2q_similarity_workshop
default
train
11,997
question2
10
187
28.18238
27
9.53482
{ "bin_edges": [ 10, 28, 46, 64, 82, 100, 118, 136, 154, 172, 187 ], "hist": [ 6444, 4906, 608, 35, 3, 0, 0, 0, 0, 1 ] }
false
srikanthsri/dhinesdataset
default
train
4,000
text
54
2,973
730.931
511.5
634.95856
{ "bin_edges": [ 54, 346, 638, 930, 1222, 1514, 1806, 2098, 2390, 2682, 2973 ], "hist": [ 1645, 592, 495, 350, 321, 271, 172, 87, 52, 15 ] }
false
jayhii/top_50_data
default
train
6,181
instruction
10
159
41.99725
40
13.65018
{ "bin_edges": [ 10, 25, 40, 55, 70, 85, 100, 115, 130, 145, 159 ], "hist": [ 279, 2775, 2142, 732, 192, 44, 12, 3, 0, 2 ] }
false
jayhii/top_50_data
default
train
6,181
output
13
609
85.33862
74
46.86061
{ "bin_edges": [ 13, 73, 133, 193, 253, 313, 373, 433, 493, 553, 609 ], "hist": [ 3011, 2362, 620, 121, 54, 9, 1, 1, 0, 2 ] }
false
jayhii/top_50_data
default
train
6,181
text
173
793
266.33587
257
49.2451
{ "bin_edges": [ 173, 236, 299, 362, 425, 488, 551, 614, 677, 740, 793 ], "hist": [ 1743, 3162, 1004, 194, 61, 12, 2, 1, 0, 2 ] }
false
0x-YuAN/NLP_explainable
default
train
31,060
q
5
12,963
324.46336
182
587.51499
{ "bin_edges": [ 5, 1301, 2597, 3893, 5189, 6485, 7781, 9077, 10373, 11669, 12963 ], "hist": [ 30112, 661, 140, 52, 32, 20, 25, 4, 9, 5 ] }
false
0x-YuAN/NLP_explainable
default
train
31,060
q'
3
2,669
95.71806
80
74.03604
{ "bin_edges": [ 3, 270, 537, 804, 1071, 1338, 1605, 1872, 2139, 2406, 2669 ], "hist": [ 30326, 642, 63, 19, 4, 2, 1, 1, 1, 1 ] }
false
0x-YuAN/NLP_explainable
default
train
31,060
r
3
35,335
381.36043
221
656.38807
{ "bin_edges": [ 3, 3537, 7071, 10605, 14139, 17673, 21207, 24741, 28275, 31809, 35335 ], "hist": [ 30902, 111, 40, 3, 2, 0, 0, 0, 0, 2 ] }
false
0x-YuAN/NLP_explainable
default
train
31,060
r'
3
6,776
94.15113
77
90.95747
{ "bin_edges": [ 3, 681, 1359, 2037, 2715, 3393, 4071, 4749, 5427, 6105, 6776 ], "hist": [ 30992, 54, 10, 3, 0, 0, 0, 0, 0, 1 ] }
false
jmgb0127/FronxOwnerManual
default
train
1,177
answer
31
1,096
186.1113
151
138.49397
{ "bin_edges": [ 31, 138, 245, 352, 459, 566, 673, 780, 887, 994, 1096 ], "hist": [ 536, 379, 144, 61, 28, 13, 5, 6, 2, 3 ] }
false
jmgb0127/FronxOwnerManual
default
train
1,177
context
47
998
845.37553
908
189.04982
{ "bin_edges": [ 47, 143, 239, 335, 431, 527, 623, 719, 815, 911, 998 ], "hist": [ 20, 14, 9, 15, 33, 34, 60, 113, 300, 579 ] }
false
jmgb0127/FronxOwnerManual
default
train
1,177
question
28
175
72.4503
69
20.74073
{ "bin_edges": [ 28, 43, 58, 73, 88, 103, 118, 133, 148, 163, 175 ], "hist": [ 31, 236, 436, 252, 116, 63, 22, 12, 7, 2 ] }
false
0x-YuAN/NLP_explainable
default
validation
3,451
q
6
12,963
318.30571
182
536.32147
{ "bin_edges": [ 6, 1302, 2598, 3894, 5190, 6486, 7782, 9078, 10374, 11670, 12963 ], "hist": [ 3353, 73, 9, 10, 2, 1, 2, 0, 0, 1 ] }
false
0x-YuAN/NLP_explainable
default
validation
3,451
q'
3
1,153
95.84323
79
75.03082
{ "bin_edges": [ 3, 119, 235, 351, 467, 583, 699, 815, 931, 1047, 1153 ], "hist": [ 2597, 717, 94, 27, 6, 3, 3, 2, 1, 1 ] }
false
0x-YuAN/NLP_explainable
default
validation
3,451
r
3
16,147
372.96899
215
609.89333
{ "bin_edges": [ 3, 1618, 3233, 4848, 6463, 8078, 9693, 11308, 12923, 14538, 16147 ], "hist": [ 3355, 81, 6, 2, 3, 3, 0, 0, 0, 1 ] }
false
0x-YuAN/NLP_explainable
default
validation
3,451
r'
3
1,454
93.16691
76
78.86486
{ "bin_edges": [ 3, 149, 295, 441, 587, 733, 879, 1025, 1171, 1317, 1454 ], "hist": [ 2962, 417, 49, 16, 3, 1, 1, 0, 0, 2 ] }
false
0x-YuAN/NLP_explainable
default
test
3,835
q
5
12,963
338.17158
188
616.68277
{ "bin_edges": [ 5, 1301, 2597, 3893, 5189, 6485, 7781, 9077, 10373, 11669, 12963 ], "hist": [ 3701, 92, 17, 9, 9, 2, 4, 0, 0, 1 ] }
false
0x-YuAN/NLP_explainable
default
test
3,835
q'
5
1,044
95.75072
81
69.49886
{ "bin_edges": [ 5, 109, 213, 317, 421, 525, 629, 733, 837, 941, 1044 ], "hist": [ 2669, 968, 155, 23, 8, 4, 4, 2, 0, 2 ] }
false
0x-YuAN/NLP_explainable
default
test
3,835
r
5
10,372
376.04537
222
611.40635
{ "bin_edges": [ 5, 1042, 2079, 3116, 4153, 5190, 6227, 7264, 8301, 9338, 10372 ], "hist": [ 3595, 187, 33, 2, 4, 3, 2, 3, 4, 2 ] }
false
0x-YuAN/NLP_explainable
default
test
3,835
r'
3
2,291
93.80156
76
92.75579
{ "bin_edges": [ 3, 232, 461, 690, 919, 1148, 1377, 1606, 1835, 2064, 2291 ], "hist": [ 3679, 136, 13, 3, 0, 0, 0, 1, 2, 1 ] }
false
jmgb0127/FronxOwnerManual
default
test
294
answer
31
1,082
178.36054
137.5
142.20252
{ "bin_edges": [ 31, 137, 243, 349, 455, 561, 667, 773, 879, 985, 1082 ], "hist": [ 146, 83, 38, 12, 8, 2, 3, 0, 1, 1 ] }
false
jmgb0127/FronxOwnerManual
default
test
294
context
82
998
856.91837
911
163.89783
{ "bin_edges": [ 82, 174, 266, 358, 450, 542, 634, 726, 818, 910, 998 ], "hist": [ 2, 2, 3, 5, 3, 15, 14, 34, 63, 153 ] }
false
jmgb0127/FronxOwnerManual
default
test
294
question
31
155
70.7551
69
18.75405
{ "bin_edges": [ 31, 44, 57, 70, 83, 96, 109, 122, 135, 148, 155 ], "hist": [ 11, 49, 95, 76, 36, 12, 12, 1, 1, 1 ] }
false
haitengzhao/molecule_property_instruction
default
esol
1,128
graph
1
98
22.51507
20
13.29348
{ "bin_edges": [ 1, 11, 21, 31, 41, 51, 61, 71, 81, 91, 98 ], "hist": [ 233, 343, 246, 191, 78, 28, 7, 0, 0, 2 ] }
false
haitengzhao/molecule_property_instruction
default
esol
1,128
label
3
5
4.8023
5
0.41587
{ "bin_edges": [ 3, 4, 5, 5 ], "hist": [ 8, 207, 913 ] }
false
haitengzhao/molecule_property_instruction
default
esol
1,128
molecule_index
1
4
3.01596
3
0.4782
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 10, 90, 900, 128 ] }
false
haitengzhao/molecule_property_instruction
default
lipo
4,200
graph
11
267
47.49786
48
15.68045
{ "bin_edges": [ 11, 37, 63, 89, 115, 141, 167, 193, 219, 245, 267 ], "hist": [ 1071, 2547, 556, 17, 5, 1, 1, 0, 0, 2 ] }
false
haitengzhao/molecule_property_instruction
default
lipo
4,200
label
3
5
3.73214
4
0.53851
{ "bin_edges": [ 3, 4, 5, 5 ], "hist": [ 1322, 2681, 197 ] }
false
haitengzhao/molecule_property_instruction
default
lipo
4,200
molecule_index
1
4
3.73571
4
0.50164
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 10, 90, 900, 3200 ] }
false
haitengzhao/molecule_property_instruction
default
freesolv
642
graph
1
82
15.3271
12
10.63398
{ "bin_edges": [ 1, 10, 19, 28, 37, 46, 55, 64, 73, 82, 82 ], "hist": [ 195, 288, 79, 52, 14, 7, 4, 2, 0, 1 ] }
false
haitengzhao/molecule_property_instruction
default
freesolv
642
label
3
6
4.7757
5
0.55254
{ "bin_edges": [ 3, 4, 5, 6, 6 ], "hist": [ 14, 144, 456, 28 ] }
false
haitengzhao/molecule_property_instruction
default
freesolv
642
molecule_index
1
3
2.82866
3
0.41642
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 10, 90, 542 ] }
false
haitengzhao/molecule_property_instruction
default
bace
1,513
graph
17
198
65.22075
62
18.94252
{ "bin_edges": [ 17, 36, 55, 74, 93, 112, 131, 150, 169, 188, 198 ], "hist": [ 38, 476, 528, 352, 103, 5, 5, 2, 3, 1 ] }
false
haitengzhao/molecule_property_instruction
default
bace
1,513
molecule_index
1
4
3.26636
3
0.59521
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 10, 90, 900, 513 ] }
false
haitengzhao/molecule_property_instruction
default
hiv
41,127
graph
4
580
45.34245
40
22.99853
{ "bin_edges": [ 4, 62, 120, 178, 236, 294, 352, 410, 468, 526, 580 ], "hist": [ 34921, 5624, 467, 78, 24, 5, 5, 2, 0, 1 ] }
false
haitengzhao/molecule_property_instruction
default
hiv
41,127
molecule_index
1
5
4.72986
5
0.50694
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 10, 90, 900, 9000, 31127 ] }
false
haitengzhao/molecule_property_instruction
default
muv
249,886
graph
9
86
40.68399
40
8.88594
{ "bin_edges": [ 9, 17, 25, 33, 41, 49, 57, 65, 73, 81, 86 ], "hist": [ 104, 5859, 41232, 80427, 72071, 39421, 10172, 559, 39, 2 ] }
false
haitengzhao/molecule_property_instruction
default
muv
249,886
molecule_index
1
5
4.87579
5
0.37027
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 40, 255, 2533, 25048, 222010 ] }
false
haitengzhao/molecule_property_instruction
default
tox21
77,946
graph
2
342
33.63854
27
27.85493
{ "bin_edges": [ 2, 37, 72, 107, 142, 177, 212, 247, 282, 317, 342 ], "hist": [ 54180, 18580, 3474, 748, 420, 281, 100, 115, 15, 33 ] }
false
haitengzhao/molecule_property_instruction
default
tox21
77,946
molecule_index
1
4
3.85676
4
0.39135
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 88, 922, 9057, 67879 ] }
false
haitengzhao/molecule_property_instruction
default
toxcast
1,490,412
graph
2
325
36.31873
30
26.44159
{ "bin_edges": [ 2, 35, 68, 101, 134, 167, 200, 233, 266, 299, 325 ], "hist": [ 870758, 487718, 100812, 15757, 5377, 5568, 1719, 1486, 752, 465 ] }
false
haitengzhao/molecule_property_instruction
default
toxcast
1,490,412
molecule_index
1
4
3.85345
4
0.39413
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 2234, 15854, 180009, 1292315 ] }
false
haitengzhao/molecule_property_instruction
default
toxcast
1,490,412
task_index
1
3
2.85173
3
0.37063
{ "bin_edges": [ 1, 2, 3, 3 ], "hist": [ 8257, 204467, 1277688 ] }
false
haitengzhao/molecule_property_instruction
default
bbbp
2,039
graph
3
400
51.56253
45
30.67422
{ "bin_edges": [ 3, 43, 83, 123, 163, 203, 243, 283, 323, 363, 400 ], "hist": [ 889, 908, 190, 27, 19, 2, 2, 0, 1, 1 ] }
false
haitengzhao/molecule_property_instruction
default
bbbp
2,039
molecule_index
1
4
3.45562
4
0.60491
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 10, 90, 900, 1039 ] }
false
haitengzhao/molecule_property_instruction
default
cyp450
53,178
graph
3
365
47.83147
43
23.30933
{ "bin_edges": [ 3, 40, 77, 114, 151, 188, 225, 262, 299, 336, 365 ], "hist": [ 20608, 27778, 3658, 862, 136, 90, 26, 10, 8, 2 ] }
false
haitengzhao/molecule_property_instruction
default
cyp450
53,178
molecule_index
1
5
4.25941
4
0.58849
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 24, 291, 3083, 32248, 17532 ] }
false
haitengzhao/molecule_property_instruction
default
chembl_zero_shot
180,229
graph
4
216
44.82724
43
14.69974
{ "bin_edges": [ 4, 26, 48, 70, 92, 114, 136, 158, 180, 202, 216 ], "hist": [ 9506, 104873, 56401, 7175, 1660, 523, 82, 1, 0, 8 ] }
false
haitengzhao/molecule_property_instruction
default
chembl_zero_shot
180,229
molecule_index
7
13
12.87021
13
0.54181
{ "bin_edges": [ 7, 8, 9, 10, 11, 12, 13, 13 ], "hist": [ 24, 196, 1372, 1038, 2610, 8446, 166543 ] }
false
haitengzhao/molecule_property_instruction
default
chembl_zero_shot
180,229
task_index
12
13
12.99931
13
0.02622
{ "bin_edges": [ 12, 13, 13 ], "hist": [ 124, 180105 ] }
false
haitengzhao/molecule_property_instruction
default
chembl_pretraining
23,874,346
graph
4
407
45.0281
43
15.11796
{ "bin_edges": [ 4, 45, 86, 127, 168, 209, 250, 291, 332, 373, 407 ], "hist": [ 12965543, 10421538, 451877, 30641, 3390, 726, 625, 3, 0, 3 ] }
false
haitengzhao/molecule_property_instruction
default
chembl_pretraining
23,874,346
label
1
7
2.43922
2
1.2173
{ "bin_edges": [ 1, 2, 3, 4, 5, 6, 7, 7 ], "hist": [ 3584103, 13302847, 3112801, 1539992, 1461182, 873360, 61 ] }
false
haitengzhao/molecule_property_instruction
default
chembl_pretraining
23,874,346
molecule_index
7
13
12.89379
13
0.47345
{ "bin_edges": [ 7, 8, 9, 10, 11, 12, 13, 13 ], "hist": [ 1974, 27474, 110623, 88165, 310088, 1059208, 22276814 ] }
false
haitengzhao/molecule_property_instruction
default
chembl_pretraining
23,874,346
task_index
12
13
12.99959
13
0.02036
{ "bin_edges": [ 12, 13, 13 ], "hist": [ 6120, 14756250 ] }
false
alancooney/relation_counterfact
default
train
6,597
attribute
3
17
6.6988
6
2.03854
{ "bin_edges": [ 3, 5, 7, 9, 11, 13, 15, 17, 17 ], "hist": [ 723, 2827, 1778, 1018, 244, 2, 0, 5 ] }
false
alancooney/relation_counterfact
default
train
6,597
counterfactual_attribute
2
31
7.13127
7
2.67828
{ "bin_edges": [ 2, 5, 8, 11, 14, 17, 20, 23, 26, 29, 31 ], "hist": [ 798, 3763, 1288, 587, 131, 22, 4, 2, 1, 1 ] }
false
alancooney/relation_counterfact
default
train
6,597
counterfactual_attribute_distant
2
22
6.81204
6
2.35685
{ "bin_edges": [ 2, 5, 8, 11, 14, 17, 20, 22 ], "hist": [ 676, 4121, 1334, 331, 114, 17, 4 ] }
false