dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
demelin/moral_stories
cls-action+context+consequence-minimal_pairs
validation
2,000
moral_consequence
13
152
43.6485
17
34.00947
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 152 ], "hist": [ 1001, 29, 133, 280, 240, 166, 94, 40, 11, 6 ] }
false
demelin/moral_stories
cls-action+context+consequence-minimal_pairs
validation
2,000
norm
18
72
44.446
45
9.52221
{ "bin_edges": [ 18, 24, 30, 36, 42, 48, 54, 60, 66, 72, 72 ], "hist": [ 22, 116, 218, 400, 482, 404, 258, 74, 24, 2 ] }
false
demelin/moral_stories
cls-action+context+consequence-minimal_pairs
validation
2,000
situation
21
165
84.93
83
24.85723
{ "bin_edges": [ 21, 36, 51, 66, 81, 96, 111, 126, 141, 156, 165 ], "hist": [ 38, 102, 298, 460, 456, 348, 166, 90, 34, 8 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
train
20,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 20000 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
train
20,000
immoral_action
13
179
46.9272
15
37.56701
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 179 ], "hist": [ 10026, 390, 1892, 3080, 2459, 1290, 597, 212, 49, 5 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
train
20,000
intention
16
127
43.2478
41
12.43246
{ "bin_edges": [ 16, 28, 40, 52, 64, 76, 88, 100, 112, 124, 127 ], "hist": [ 1230, 7516, 6826, 3098, 1002, 198, 92, 32, 2, 4 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
train
20,000
moral_action
13
177
46.828
15.5
37.48572
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 177 ], "hist": [ 10025, 386, 1953, 3092, 2389, 1314, 551, 226, 57, 7 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
train
20,000
norm
15
82
43.8687
44
9.75138
{ "bin_edges": [ 15, 22, 29, 36, 43, 50, 57, 64, 71, 78, 82 ], "hist": [ 98, 1064, 2990, 4778, 5296, 3774, 1562, 374, 56, 8 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
train
20,000
situation
18
185
85.9598
84
25.27411
{ "bin_edges": [ 18, 35, 52, 69, 86, 103, 120, 137, 154, 171, 185 ], "hist": [ 296, 1224, 3550, 5314, 4658, 2880, 1430, 550, 92, 6 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
test
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
test
2,000
immoral_action
13
156
45.6165
19.5
36.3682
{ "bin_edges": [ 13, 28, 43, 58, 73, 88, 103, 118, 133, 148, 156 ], "hist": [ 1001, 39, 130, 275, 242, 168, 88, 36, 18, 3 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
test
2,000
intention
18
102
42.995
41
12.66938
{ "bin_edges": [ 18, 27, 36, 45, 54, 63, 72, 81, 90, 99, 102 ], "hist": [ 106, 510, 596, 446, 178, 96, 46, 18, 2, 2 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
test
2,000
moral_action
13
159
45.8435
17.5
36.42991
{ "bin_edges": [ 13, 28, 43, 58, 73, 88, 103, 118, 133, 148, 159 ], "hist": [ 1001, 23, 132, 286, 261, 146, 84, 51, 14, 2 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
test
2,000
norm
19
73
47.144
48
9.30401
{ "bin_edges": [ 19, 25, 31, 37, 43, 49, 55, 61, 67, 73, 73 ], "hist": [ 14, 68, 184, 344, 468, 480, 296, 114, 26, 6 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
test
2,000
situation
22
177
85.398
84
24.93637
{ "bin_edges": [ 22, 38, 54, 70, 86, 102, 118, 134, 150, 166, 177 ], "hist": [ 38, 132, 376, 522, 454, 262, 132, 64, 16, 4 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
train
20,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 20000 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
train
20,000
immoral_action
13
179
47.1949
15
37.88703
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 179 ], "hist": [ 10030, 396, 1822, 3032, 2470, 1344, 621, 228, 52, 5 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
train
20,000
moral_action
13
177
47.09475
15.5
37.76016
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 177 ], "hist": [ 10025, 367, 1892, 3086, 2395, 1360, 581, 227, 61, 6 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
train
20,000
norm
15
82
44.2884
44
9.76838
{ "bin_edges": [ 15, 22, 29, 36, 43, 50, 57, 64, 71, 78, 82 ], "hist": [ 92, 988, 2822, 4668, 5326, 3946, 1680, 414, 54, 10 ] }
false
midas/ldkp3k
large
test
3,413
id
3
9
7.50249
8
0.733
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 2, 8, 22, 220, 1286, 1739, 136 ] }
true
midas/ldkp3k
large
train
58,000
id
3
9
7.46512
8
0.74081
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 59, 113, 480, 4013, 22789, 28567, 1979 ] }
true
midas/ldkp3k
small
train
20,000
id
3
9
7.4959
8
0.73594
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 17, 36, 168, 1296, 7487, 10266, 730 ] }
false
midas/ldkp3k
medium
validation
3,339
id
3
9
7.50404
8
0.72915
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 3, 4, 25, 196, 1303, 1663, 145 ] }
false
midas/ldkp3k
large
validation
3,339
id
3
9
7.50404
8
0.72915
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 3, 4, 25, 196, 1303, 1663, 145 ] }
true
demelin/moral_stories
cls-action+norm-minimal_pairs
test
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
test
2,000
immoral_action
13
149
44.1085
20
34.47444
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 149 ], "hist": [ 1000, 28, 129, 256, 258, 190, 77, 32, 26, 4 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
test
2,000
moral_action
13
169
44.061
16.5
34.79429
{ "bin_edges": [ 13, 29, 45, 61, 77, 93, 109, 125, 141, 157, 169 ], "hist": [ 1004, 48, 218, 306, 246, 92, 59, 13, 11, 3 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
test
2,000
norm
18
77
44.657
45
9.52359
{ "bin_edges": [ 18, 24, 30, 36, 42, 48, 54, 60, 66, 72, 77 ], "hist": [ 12, 98, 248, 408, 420, 488, 208, 80, 28, 10 ] }
false
midas/ldkp3k
medium
train
50,000
id
3
9
7.49994
8
0.73038
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 34, 84, 392, 3193, 18808, 25616, 1873 ] }
false
midas/ldkp3k
medium
test
3,413
id
3
9
7.50249
8
0.733
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 2, 8, 22, 220, 1286, 1739, 136 ] }
false
midas/ldkp3k
small
test
3,413
id
3
9
7.50249
8
0.733
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 2, 8, 22, 220, 1286, 1739, 136 ] }
false
midas/ldkp3k
small
validation
3,339
id
3
9
7.50404
8
0.72915
{ "bin_edges": [ 3, 4, 5, 6, 7, 8, 9, 9 ], "hist": [ 3, 4, 25, 196, 1303, 1663, 145 ] }
false
demelin/moral_stories
cls-action-lexical_bias
train
20,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 20000 ] }
false
demelin/moral_stories
cls-action-lexical_bias
train
20,000
immoral_action
13
179
47.6482
15.5
38.19419
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 179 ], "hist": [ 10017, 321, 1756, 3023, 2572, 1368, 651, 237, 51, 4 ] }
false
demelin/moral_stories
cls-action-lexical_bias
train
20,000
moral_action
13
177
47.50605
15.5
38.11825
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 177 ], "hist": [ 10020, 328, 1797, 3062, 2491, 1381, 606, 243, 65, 7 ] }
false
ArthurBaia/squad_v1_pt_br
default
train
87,510
context
139
3,839
788.71402
724
322.89869
{ "bin_edges": [ 139, 510, 881, 1252, 1623, 1994, 2365, 2736, 3107, 3478, 3839 ], "hist": [ 10802, 50266, 18917, 5678, 1332, 391, 78, 22, 14, 10 ] }
false
ArthurBaia/squad_v1_pt_br
default
train
87,510
id
24
24
24
24
0
{ "bin_edges": [ 24, 24 ], "hist": [ 87510 ] }
false
ArthurBaia/squad_v1_pt_br
default
train
87,510
question
1
264
61.51427
58
22.79388
{ "bin_edges": [ 1, 28, 55, 82, 109, 136, 163, 190, 217, 244, 264 ], "hist": [ 2534, 34883, 35077, 11824, 2539, 526, 100, 22, 3, 2 ] }
false
ArthurBaia/squad_v1_pt_br
default
train
87,510
title
3
59
14.07468
12
8.56206
{ "bin_edges": [ 3, 9, 15, 21, 27, 33, 39, 45, 51, 57, 59 ], "hist": [ 27004, 26576, 16478, 10957, 2496, 2281, 979, 343, 202, 194 ] }
false
ArthurBaia/squad_v1_pt_br
default
validation
10,570
context
152
4,468
815.9684
745
370.41252
{ "bin_edges": [ 152, 584, 1016, 1448, 1880, 2312, 2744, 3176, 3608, 4040, 4468 ], "hist": [ 2319, 6068, 1700, 294, 112, 33, 11, 29, 0, 4 ] }
false
ArthurBaia/squad_v1_pt_br
default
validation
10,570
id
24
24
24
24
0
{ "bin_edges": [ 24, 24 ], "hist": [ 10570 ] }
false
ArthurBaia/squad_v1_pt_br
default
validation
10,570
question
10
196
62.13132
59
22.78889
{ "bin_edges": [ 10, 29, 48, 67, 86, 105, 124, 143, 162, 181, 196 ], "hist": [ 341, 2610, 3778, 2315, 1011, 343, 116, 45, 7, 4 ] }
false
ArthurBaia/squad_v1_pt_br
default
validation
10,570
title
5
41
14.45762
13
6.89047
{ "bin_edges": [ 5, 9, 13, 17, 21, 25, 29, 33, 37, 41, 41 ], "hist": [ 2264, 2541, 2094, 2033, 635, 295, 604, 0, 0, 104 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
validation
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
validation
2,000
immoral_action
13
157
45.348
21
35.72118
{ "bin_edges": [ 13, 28, 43, 58, 73, 88, 103, 118, 133, 148, 157 ], "hist": [ 1000, 33, 140, 271, 264, 160, 82, 35, 14, 1 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
validation
2,000
moral_action
13
152
45.4035
19.5
35.66983
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 152 ], "hist": [ 1001, 19, 91, 247, 275, 189, 108, 33, 28, 9 ] }
false
demelin/moral_stories
cls-action+norm-minimal_pairs
validation
2,000
norm
18
72
44.446
45
9.52221
{ "bin_edges": [ 18, 24, 30, 36, 42, 48, 54, 60, 66, 72, 72 ], "hist": [ 22, 116, 218, 400, 482, 404, 258, 74, 24, 2 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
test
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
test
2,000
immoral_action
13
156
45.6165
19.5
36.3682
{ "bin_edges": [ 13, 28, 43, 58, 73, 88, 103, 118, 133, 148, 156 ], "hist": [ 1001, 39, 130, 275, 242, 168, 88, 36, 18, 3 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
test
2,000
moral_action
13
159
45.8435
17.5
36.42991
{ "bin_edges": [ 13, 28, 43, 58, 73, 88, 103, 118, 133, 148, 159 ], "hist": [ 1001, 23, 132, 286, 261, 146, 84, 51, 14, 2 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
test
2,000
norm
19
73
47.144
48
9.30401
{ "bin_edges": [ 19, 25, 31, 37, 43, 49, 55, 61, 67, 73, 73 ], "hist": [ 14, 68, 184, 344, 468, 480, 296, 114, 26, 6 ] }
false
demelin/moral_stories
cls-action-minimal_pairs
test
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action-minimal_pairs
test
2,000
immoral_action
13
149
44.1085
20
34.47444
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 149 ], "hist": [ 1000, 28, 129, 256, 258, 190, 77, 32, 26, 4 ] }
false
demelin/moral_stories
cls-action-minimal_pairs
test
2,000
moral_action
13
169
44.061
16.5
34.79429
{ "bin_edges": [ 13, 29, 45, 61, 77, 93, 109, 125, 141, 157, 169 ], "hist": [ 1004, 48, 218, 306, 246, 92, 59, 13, 11, 3 ] }
false
colbertv2/lotte
lifestyle
forum_dev
2,076
author
0
30
8.35168
8
4.53438
{ "bin_edges": [ 0, 4, 8, 12, 16, 20, 24, 28, 30 ], "hist": [ 248, 633, 601, 388, 56, 32, 3, 1 ] }
false
colbertv2/lotte
lifestyle
forum_dev
2,076
query
15
141
55.32033
53
18.18091
{ "bin_edges": [ 15, 28, 41, 54, 67, 80, 93, 106, 119, 132, 141 ], "hist": [ 63, 378, 627, 508, 285, 135, 53, 19, 6, 2 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
test
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
test
2,000
immoral_action
13
170
41.4085
15
32.18555
{ "bin_edges": [ 13, 29, 45, 61, 77, 93, 109, 125, 141, 157, 170 ], "hist": [ 1009, 99, 246, 296, 220, 79, 38, 8, 4, 1 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
test
2,000
moral_action
13
145
41.7025
15.5
31.84131
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 145 ], "hist": [ 1008, 36, 147, 306, 277, 126, 63, 25, 9, 3 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
test
2,000
norm
17
81
43.902
44
9.79798
{ "bin_edges": [ 17, 24, 31, 38, 45, 52, 59, 66, 73, 80, 81 ], "hist": [ 18, 188, 322, 512, 518, 312, 106, 22, 0, 2 ] }
false
colbertv2/lotte
lifestyle
search_test
661
author
0
21
4.75076
3.5
5.28229
{ "bin_edges": [ 0, 3, 6, 9, 12, 15, 18, 21, 21 ], "hist": [ 307, 67, 103, 89, 53, 23, 11, 1 ] }
false
colbertv2/lotte
lifestyle
search_test
661
query
20
88
42.44327
40
13.46922
{ "bin_edges": [ 20, 27, 34, 41, 48, 55, 62, 69, 76, 83, 88 ], "hist": [ 70, 144, 122, 99, 83, 66, 54, 20, 2, 1 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
train
20,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 20000 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
train
20,000
immoral_action
13
179
47.6482
15.5
38.19419
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 179 ], "hist": [ 10017, 321, 1756, 3023, 2572, 1368, 651, 237, 51, 4 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
train
20,000
moral_action
13
177
47.50605
15.5
38.11825
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 177 ], "hist": [ 10020, 328, 1797, 3062, 2491, 1381, 606, 243, 65, 7 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
train
20,000
norm
15
82
44.4278
44
9.70317
{ "bin_edges": [ 15, 22, 29, 36, 43, 50, 57, 64, 71, 78, 82 ], "hist": [ 94, 914, 2768, 4692, 5314, 4042, 1716, 396, 58, 6 ] }
false
colbertv2/lotte
pooled
forum_dev
10,097
query
13
148
56.26751
53
21.15868
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 148 ], "hist": [ 427, 1984, 2857, 2411, 1322, 591, 296, 129, 52, 28 ] }
false
colbertv2/lotte
lifestyle
search_dev
417
author
0
30
9.39024
9
4.2781
{ "bin_edges": [ 0, 4, 8, 12, 16, 20, 24, 28, 30 ], "hist": [ 13, 145, 134, 93, 12, 12, 0, 1 ] }
false
colbertv2/lotte
lifestyle
search_dev
417
query
16
70
37.55875
35
11.21199
{ "bin_edges": [ 16, 22, 28, 34, 40, 46, 52, 58, 64, 70, 70 ], "hist": [ 9, 69, 106, 80, 54, 47, 27, 11, 13, 1 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
train
20,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 20000 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
train
20,000
immoral_action
13
179
46.9272
15
37.56701
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 179 ], "hist": [ 10026, 390, 1892, 3080, 2459, 1290, 597, 212, 49, 5 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
train
20,000
moral_action
13
177
46.828
15.5
37.48572
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 177 ], "hist": [ 10025, 386, 1953, 3092, 2389, 1314, 551, 226, 57, 7 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
train
20,000
norm
15
82
43.8687
44
9.75138
{ "bin_edges": [ 15, 22, 29, 36, 43, 50, 57, 64, 71, 78, 82 ], "hist": [ 98, 1064, 2990, 4778, 5296, 3774, 1562, 374, 56, 8 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
validation
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
validation
2,000
immoral_action
13
151
46.517
18.5
37.31886
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 151 ], "hist": [ 1001, 24, 103, 225, 215, 206, 108, 79, 30, 9 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
validation
2,000
moral_action
13
163
46.2885
19
37.03477
{ "bin_edges": [ 13, 29, 45, 61, 77, 93, 109, 125, 141, 157, 163 ], "hist": [ 1001, 38, 161, 294, 243, 158, 59, 33, 11, 2 ] }
false
demelin/moral_stories
cls-action+norm-norm_distance
validation
2,000
norm
19
81
46.156
46
9.21282
{ "bin_edges": [ 19, 26, 33, 40, 47, 54, 61, 68, 75, 81 ], "hist": [ 32, 124, 304, 562, 550, 314, 96, 14, 4 ] }
false
demelin/moral_stories
cls-action-minimal_pairs
validation
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action-minimal_pairs
validation
2,000
immoral_action
13
157
45.348
21
35.72118
{ "bin_edges": [ 13, 28, 43, 58, 73, 88, 103, 118, 133, 148, 157 ], "hist": [ 1000, 33, 140, 271, 264, 160, 82, 35, 14, 1 ] }
false
demelin/moral_stories
cls-action-minimal_pairs
validation
2,000
moral_action
13
152
45.4035
19.5
35.66983
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 152 ], "hist": [ 1001, 19, 91, 247, 275, 189, 108, 33, 28, 9 ] }
false
demelin/moral_stories
cls-action-norm_distance
train
20,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 20000 ] }
false
demelin/moral_stories
cls-action-norm_distance
train
20,000
immoral_action
13
179
46.9272
15
37.56701
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 179 ], "hist": [ 10026, 390, 1892, 3080, 2459, 1290, 597, 212, 49, 5 ] }
false
demelin/moral_stories
cls-action-norm_distance
train
20,000
moral_action
13
177
46.828
15.5
37.48572
{ "bin_edges": [ 13, 30, 47, 64, 81, 98, 115, 132, 149, 166, 177 ], "hist": [ 10025, 386, 1953, 3092, 2389, 1314, 551, 226, 57, 7 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
validation
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
validation
2,000
immoral_action
13
151
46.517
18.5
37.31886
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 151 ], "hist": [ 1001, 24, 103, 225, 215, 206, 108, 79, 30, 9 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
validation
2,000
intention
17
112
42.733
41
12.37127
{ "bin_edges": [ 17, 27, 37, 47, 57, 67, 77, 87, 97, 107, 112 ], "hist": [ 118, 562, 646, 420, 164, 66, 18, 2, 0, 4 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
validation
2,000
moral_action
13
163
46.2885
19
37.03477
{ "bin_edges": [ 13, 29, 45, 61, 77, 93, 109, 125, 141, 157, 163 ], "hist": [ 1001, 38, 161, 294, 243, 158, 59, 33, 11, 2 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
validation
2,000
norm
19
81
46.156
46
9.21282
{ "bin_edges": [ 19, 26, 33, 40, 47, 54, 61, 68, 75, 81 ], "hist": [ 32, 124, 304, 562, 550, 314, 96, 14, 4 ] }
false
demelin/moral_stories
cls-action+context-norm_distance
validation
2,000
situation
20
166
85.536
84
25.65587
{ "bin_edges": [ 20, 35, 50, 65, 80, 95, 110, 125, 140, 155, 166 ], "hist": [ 36, 94, 286, 452, 450, 318, 218, 92, 38, 16 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
validation
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
validation
2,000
immoral_action
13
142
43.515
15.5
34.0386
{ "bin_edges": [ 13, 26, 39, 52, 65, 78, 91, 104, 117, 130, 142 ], "hist": [ 1004, 18, 107, 228, 266, 159, 112, 75, 16, 15 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
validation
2,000
moral_action
13
149
43.649
21
34.15667
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 149 ], "hist": [ 1000, 33, 133, 262, 268, 152, 97, 30, 18, 7 ] }
false
demelin/moral_stories
cls-action+norm-lexical_bias
validation
2,000
norm
18
78
43.807
44
9.88224
{ "bin_edges": [ 18, 25, 32, 39, 46, 53, 60, 67, 74, 78 ], "hist": [ 20, 214, 360, 550, 468, 258, 104, 22, 4 ] }
false
demelin/moral_stories
cls-action-lexical_bias
validation
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action-lexical_bias
validation
2,000
immoral_action
13
142
43.515
15.5
34.0386
{ "bin_edges": [ 13, 26, 39, 52, 65, 78, 91, 104, 117, 130, 142 ], "hist": [ 1004, 18, 107, 228, 266, 159, 112, 75, 16, 15 ] }
false
demelin/moral_stories
cls-action-lexical_bias
validation
2,000
moral_action
13
149
43.649
21
34.15667
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 149 ], "hist": [ 1000, 33, 133, 262, 268, 152, 97, 30, 18, 7 ] }
false
colbertv2/lotte
pooled
forum_test
10,025
query
8
150
55.09436
52
21.60609
{ "bin_edges": [ 8, 23, 38, 53, 68, 83, 98, 113, 128, 143, 150 ], "hist": [ 323, 1788, 2978, 2465, 1406, 658, 235, 90, 60, 22 ] }
false
demelin/moral_stories
cls-action-lexical_bias
test
2,000
ID
31
31
31
31
0
{ "bin_edges": [ 31, 31 ], "hist": [ 2000 ] }
false
demelin/moral_stories
cls-action-lexical_bias
test
2,000
immoral_action
13
170
41.4085
15
32.18555
{ "bin_edges": [ 13, 29, 45, 61, 77, 93, 109, 125, 141, 157, 170 ], "hist": [ 1009, 99, 246, 296, 220, 79, 38, 8, 4, 1 ] }
false
demelin/moral_stories
cls-action-lexical_bias
test
2,000
moral_action
13
145
41.7025
15.5
31.84131
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 145 ], "hist": [ 1008, 36, 147, 306, 277, 126, 63, 25, 9, 3 ] }
false