dataset stringlengths 4 115 | config stringlengths 1 121 | split stringlengths 1 228 | num_examples int64 3 341M | column_name stringlengths 1 22.7k | min int64 0 1.81M | max int64 0 981M | mean float64 0 42.2M | median float64 0 24M | std float64 0 84.2M | histogram dict | partial bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|
demelin/moral_stories | cls-consequence+action-norm_distance | validation | 4,000 | moral_consequence | 13 | 158 | 43.725 | 17.5 | 34.36509 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
158
],
"hist": [
2014,
98,
292,
606,
490,
306,
120,
42,
26,
6
]
} | false |
colbertv2/lotte | science | search_dev | 538 | author | 0 | 22 | 8.18075 | 7 | 4.97976 | {
"bin_edges": [
0,
3,
6,
9,
12,
15,
18,
21,
22
],
"hist": [
45,
108,
145,
87,
63,
36,
19,
6
]
} | false |
colbertv2/lotte | science | search_dev | 538 | query | 17 | 93 | 45.47955 | 41 | 17.20477 | {
"bin_edges": [
17,
25,
33,
41,
49,
57,
65,
73,
81,
89,
93
],
"hist": [
30,
131,
103,
67,
60,
50,
45,
39,
8,
5
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | validation | 4,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
4000
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | validation | 4,000 | immoral_action | 13 | 147 | 45.097 | 15.5 | 35.79512 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
147
],
"hist": [
2002,
64,
218,
484,
500,
362,
198,
114,
38,
20
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | validation | 4,000 | immoral_consequence | 13 | 146 | 42.9835 | 17 | 33.70304 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
146
],
"hist": [
2008,
116,
260,
516,
504,
320,
156,
80,
30,
10
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | validation | 4,000 | intention | 18 | 103 | 43 | 41 | 12.57949 | {
"bin_edges": [
18,
27,
36,
45,
54,
63,
72,
81,
90,
99,
103
],
"hist": [
192,
1024,
1220,
856,
432,
168,
64,
20,
12,
12
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | validation | 4,000 | moral_action | 13 | 157 | 44.9575 | 18.5 | 35.55384 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
157
],
"hist": [
2004,
64,
282,
568,
538,
292,
150,
64,
24,
14
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | validation | 4,000 | moral_consequence | 13 | 163 | 42.115 | 16.5 | 32.20401 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
163
],
"hist": [
2006,
142,
466,
656,
472,
190,
46,
18,
2,
2
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | validation | 4,000 | norm | 18 | 82 | 44.826 | 45 | 9.84073 | {
"bin_edges": [
18,
25,
32,
39,
46,
53,
60,
67,
74,
81,
82
],
"hist": [
32,
336,
680,
1036,
1060,
600,
200,
44,
4,
8
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | validation | 4,000 | situation | 22 | 167 | 83.007 | 81 | 24.42421 | {
"bin_edges": [
22,
37,
52,
67,
82,
97,
112,
127,
142,
157,
167
],
"hist": [
52,
284,
712,
988,
836,
628,
312,
120,
44,
24
]
} | false |
colbertv2/lotte | recreation | forum_test | 2,002 | author | 0 | 20 | 4.44737 | 3 | 5.26928 | {
"bin_edges": [
0,
3,
6,
9,
12,
15,
18,
20
],
"hist": [
972,
260,
312,
165,
178,
37,
52
]
} | false |
colbertv2/lotte | recreation | forum_test | 2,002 | query | 15 | 140 | 51.4985 | 49 | 18.4993 | {
"bin_edges": [
15,
28,
41,
54,
67,
80,
93,
106,
119,
132,
140
],
"hist": [
143,
468,
560,
449,
227,
99,
42,
9,
3,
2
]
} | false |
demelin/moral_stories | cls-consequence+action-lexical_bias | validation | 4,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
4000
]
} | false |
demelin/moral_stories | cls-consequence+action-lexical_bias | validation | 4,000 | immoral_action | 13 | 147 | 45.097 | 15.5 | 35.79512 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
147
],
"hist": [
2002,
64,
218,
484,
500,
362,
198,
114,
38,
20
]
} | false |
demelin/moral_stories | cls-consequence+action-lexical_bias | validation | 4,000 | immoral_consequence | 13 | 146 | 42.9835 | 17 | 33.70304 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
146
],
"hist": [
2008,
116,
260,
516,
504,
320,
156,
80,
30,
10
]
} | false |
demelin/moral_stories | cls-consequence+action-lexical_bias | validation | 4,000 | moral_action | 13 | 157 | 44.9575 | 18.5 | 35.55384 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
157
],
"hist": [
2004,
64,
282,
568,
538,
292,
150,
64,
24,
14
]
} | false |
demelin/moral_stories | cls-consequence+action-lexical_bias | validation | 4,000 | moral_consequence | 13 | 163 | 42.115 | 16.5 | 32.20401 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
163
],
"hist": [
2006,
142,
466,
656,
472,
190,
46,
18,
2,
2
]
} | false |
demelin/moral_stories | cls-consequence+action-lexical_bias | test | 4,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
4000
]
} | false |
demelin/moral_stories | cls-consequence+action-lexical_bias | test | 4,000 | immoral_action | 13 | 152 | 44.877 | 18 | 35.51887 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
152
],
"hist": [
2002,
76,
220,
500,
448,
392,
206,
98,
48,
10
]
} | false |
demelin/moral_stories | cls-consequence+action-lexical_bias | test | 4,000 | immoral_consequence | 13 | 165 | 42.1815 | 14 | 33.33748 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
165
],
"hist": [
2042,
176,
422,
584,
462,
194,
82,
26,
10,
2
]
} | false |
demelin/moral_stories | cls-consequence+action-lexical_bias | test | 4,000 | moral_action | 13 | 153 | 45.055 | 16 | 35.91662 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
153
],
"hist": [
2008,
54,
334,
540,
518,
266,
172,
62,
32,
14
]
} | false |
demelin/moral_stories | cls-consequence+action-lexical_bias | test | 4,000 | moral_consequence | 13 | 157 | 41.1595 | 15 | 31.73328 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
157
],
"hist": [
2022,
146,
378,
652,
426,
234,
106,
30,
2,
4
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | test | 4,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
4000
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | test | 4,000 | immoral_action | 13 | 140 | 43.708 | 17.5 | 33.99569 | {
"bin_edges": [
13,
26,
39,
52,
65,
78,
91,
104,
117,
130,
140
],
"hist": [
2004,
50,
152,
446,
560,
396,
220,
94,
42,
36
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | test | 4,000 | immoral_consequence | 13 | 166 | 40.8695 | 19 | 31.36371 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
166
],
"hist": [
2020,
182,
514,
698,
338,
156,
70,
18,
2,
2
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | test | 4,000 | moral_action | 13 | 154 | 44.2495 | 15.5 | 34.90179 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
154
],
"hist": [
2010,
62,
336,
572,
500,
288,
150,
50,
30,
2
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | test | 4,000 | moral_consequence | 13 | 169 | 40.327 | 15.5 | 30.88664 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
169
],
"hist": [
2016,
192,
582,
650,
336,
138,
58,
22,
4,
2
]
} | false |
demelin/moral_stories | cls-consequence+action-norm_distance | test | 4,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
4000
]
} | false |
demelin/moral_stories | cls-consequence+action-norm_distance | test | 4,000 | immoral_action | 13 | 156 | 45.6165 | 19.5 | 36.36365 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
156
],
"hist": [
2002,
78,
260,
550,
484,
336,
176,
72,
36,
6
]
} | false |
demelin/moral_stories | cls-consequence+action-norm_distance | test | 4,000 | immoral_consequence | 13 | 166 | 45.167 | 14 | 35.85049 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
166
],
"hist": [
2024,
94,
304,
634,
480,
290,
122,
40,
10,
2
]
} | false |
demelin/moral_stories | cls-consequence+action-norm_distance | test | 4,000 | moral_action | 13 | 159 | 45.8435 | 17.5 | 36.42536 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
159
],
"hist": [
2002,
46,
264,
572,
522,
292,
168,
102,
28,
4
]
} | false |
demelin/moral_stories | cls-consequence+action-norm_distance | test | 4,000 | moral_consequence | 13 | 167 | 43.5675 | 14.5 | 34.08445 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
167
],
"hist": [
2024,
102,
406,
626,
500,
218,
82,
22,
14,
6
]
} | false |
colbertv2/lotte | recreation | search_dev | 563 | author | 0 | 29 | 5.33577 | 5 | 5.74698 | {
"bin_edges": [
0,
3,
6,
9,
12,
15,
18,
21,
24,
27,
29
],
"hist": [
251,
27,
107,
80,
48,
28,
1,
3,
0,
3
]
} | false |
colbertv2/lotte | recreation | search_dev | 563 | query | 18 | 75 | 34.53641 | 32 | 9.90093 | {
"bin_edges": [
18,
24,
30,
36,
42,
48,
54,
60,
66,
72,
75
],
"hist": [
28,
181,
164,
85,
40,
25,
22,
14,
3,
1
]
} | false |
colbertv2/lotte | science | search_test | 617 | author | 0 | 27 | 5.26644 | 5 | 5.56502 | {
"bin_edges": [
0,
3,
6,
9,
12,
15,
18,
21,
24,
27,
27
],
"hist": [
255,
59,
119,
61,
39,
51,
8,
0,
0,
1
]
} | false |
colbertv2/lotte | science | search_test | 617 | query | 17 | 100 | 46.57699 | 46 | 15.95166 | {
"bin_edges": [
17,
26,
35,
44,
53,
62,
71,
80,
89,
98,
100
],
"hist": [
41,
127,
123,
115,
97,
70,
24,
9,
10,
1
]
} | false |
colbertv2/lotte | technology | forum_dev | 2,003 | author | 0 | 27 | 5.72012 | 6 | 5.24566 | {
"bin_edges": [
0,
3,
6,
9,
12,
15,
18,
21,
24,
27,
27
],
"hist": [
706,
267,
403,
255,
222,
83,
19,
1,
0,
2
]
} | false |
colbertv2/lotte | technology | forum_dev | 2,003 | query | 15 | 148 | 52.47279 | 50 | 18.11253 | {
"bin_edges": [
15,
29,
43,
57,
71,
85,
99,
113,
127,
141,
148
],
"hist": [
121,
519,
657,
405,
198,
62,
28,
8,
3,
2
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | train | 40,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
40000
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | train | 40,000 | immoral_action | 13 | 179 | 47.14315 | 15 | 37.78632 | {
"bin_edges": [
13,
30,
47,
64,
81,
98,
115,
132,
149,
166,
179
],
"hist": [
20052,
726,
3734,
6164,
4920,
2600,
1242,
448,
104,
10
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | train | 40,000 | immoral_consequence | 13 | 189 | 45.48185 | 13.5 | 36.33055 | {
"bin_edges": [
13,
31,
49,
67,
85,
103,
121,
139,
157,
175,
189
],
"hist": [
20174,
1368,
5062,
6502,
4004,
1948,
690,
208,
32,
12
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | train | 40,000 | intention | 16 | 127 | 43.1246 | 41 | 12.36135 | {
"bin_edges": [
16,
28,
40,
52,
64,
76,
88,
100,
112,
124,
127
],
"hist": [
2512,
15076,
13720,
6032,
2028,
412,
156,
52,
4,
8
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | train | 40,000 | moral_action | 13 | 177 | 47.03995 | 15.5 | 37.669 | {
"bin_edges": [
13,
30,
47,
64,
81,
98,
115,
132,
149,
166,
177
],
"hist": [
20044,
740,
3758,
6204,
4860,
2666,
1146,
458,
110,
14
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | train | 40,000 | moral_consequence | 13 | 184 | 44.6893 | 14.5 | 35.28362 | {
"bin_edges": [
13,
31,
49,
67,
85,
103,
121,
139,
157,
175,
184
],
"hist": [
20166,
1422,
5296,
6812,
4006,
1570,
514,
162,
50,
2
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | train | 40,000 | norm | 15 | 79 | 44.3334 | 44 | 9.70896 | {
"bin_edges": [
15,
22,
29,
36,
43,
50,
57,
64,
71,
78,
79
],
"hist": [
204,
1856,
5572,
9460,
10612,
8024,
3356,
796,
112,
8
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | train | 40,000 | situation | 18 | 185 | 86.6299 | 85 | 25.17771 | {
"bin_edges": [
18,
35,
52,
69,
86,
103,
120,
137,
154,
171,
185
],
"hist": [
540,
2240,
6984,
10664,
9424,
5820,
3012,
1100,
204,
12
]
} | false |
colbertv2/lotte | writing | search_test | 1,071 | author | 0 | 28 | 1.86879 | 0 | 4.2073 | {
"bin_edges": [
0,
3,
6,
9,
12,
15,
18,
21,
24,
27,
28
],
"hist": [
837,
84,
44,
31,
47,
11,
12,
0,
0,
1
]
} | false |
colbertv2/lotte | writing | search_test | 1,071 | query | 16 | 82 | 39.47712 | 39 | 13.7973 | {
"bin_edges": [
16,
23,
30,
37,
44,
51,
58,
65,
72,
79,
82
],
"hist": [
113,
233,
146,
123,
193,
156,
70,
28,
5,
4
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | test | 4,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
4000
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | test | 4,000 | immoral_action | 13 | 152 | 44.877 | 18 | 35.51887 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
152
],
"hist": [
2002,
76,
220,
500,
448,
392,
206,
98,
48,
10
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | test | 4,000 | immoral_consequence | 13 | 165 | 42.1815 | 14 | 33.33748 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
165
],
"hist": [
2042,
176,
422,
584,
462,
194,
82,
26,
10,
2
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | test | 4,000 | intention | 17 | 112 | 43.96 | 42 | 13.13177 | {
"bin_edges": [
17,
27,
37,
47,
57,
67,
77,
87,
97,
107,
112
],
"hist": [
204,
1024,
1332,
836,
380,
140,
40,
28,
8,
8
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | test | 4,000 | moral_action | 13 | 153 | 45.055 | 16 | 35.91662 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
153
],
"hist": [
2008,
54,
334,
540,
518,
266,
172,
62,
32,
14
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | test | 4,000 | moral_consequence | 13 | 157 | 41.1595 | 15 | 31.73328 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
157
],
"hist": [
2022,
146,
378,
652,
426,
234,
106,
30,
2,
4
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | test | 4,000 | norm | 19 | 76 | 43.827 | 44 | 9.78136 | {
"bin_edges": [
19,
25,
31,
37,
43,
49,
55,
61,
67,
73,
76
],
"hist": [
52,
340,
584,
756,
968,
756,
360,
140,
40,
4
]
} | false |
demelin/moral_stories | cls-consequence+action+context-lexical_bias | test | 4,000 | situation | 20 | 177 | 81.226 | 81 | 26.39203 | {
"bin_edges": [
20,
36,
52,
68,
84,
100,
116,
132,
148,
164,
177
],
"hist": [
140,
396,
700,
972,
812,
588,
248,
100,
36,
8
]
} | false |
demelin/moral_stories | cls-consequence+action+context-norm_distance | validation | 4,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
4000
]
} | false |
demelin/moral_stories | cls-consequence+action+context-norm_distance | validation | 4,000 | immoral_action | 13 | 151 | 46.517 | 18.5 | 37.31419 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
151
],
"hist": [
2002,
48,
206,
450,
430,
412,
216,
158,
60,
18
]
} | false |
demelin/moral_stories | cls-consequence+action+context-norm_distance | validation | 4,000 | immoral_consequence | 13 | 169 | 44.3325 | 17.5 | 35.33542 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
169
],
"hist": [
2022,
148,
322,
578,
508,
256,
112,
36,
16,
2
]
} | false |
demelin/moral_stories | cls-consequence+action+context-norm_distance | validation | 4,000 | intention | 17 | 112 | 42.733 | 41 | 12.36972 | {
"bin_edges": [
17,
27,
37,
47,
57,
67,
77,
87,
97,
107,
112
],
"hist": [
236,
1124,
1292,
840,
328,
132,
36,
4,
0,
8
]
} | false |
demelin/moral_stories | cls-consequence+action+context-norm_distance | validation | 4,000 | moral_action | 13 | 163 | 46.2885 | 19 | 37.03014 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
163
],
"hist": [
2002,
76,
322,
588,
486,
316,
118,
66,
22,
4
]
} | false |
demelin/moral_stories | cls-consequence+action+context-norm_distance | validation | 4,000 | moral_consequence | 13 | 158 | 43.725 | 17.5 | 34.36509 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
158
],
"hist": [
2014,
98,
292,
606,
490,
306,
120,
42,
26,
6
]
} | false |
demelin/moral_stories | cls-consequence+action+context-norm_distance | validation | 4,000 | norm | 19 | 81 | 46.156 | 46 | 9.21167 | {
"bin_edges": [
19,
26,
33,
40,
47,
54,
61,
68,
75,
81
],
"hist": [
64,
248,
608,
1124,
1100,
628,
192,
28,
8
]
} | false |
demelin/moral_stories | cls-consequence+action+context-norm_distance | validation | 4,000 | situation | 20 | 166 | 85.536 | 84 | 25.65266 | {
"bin_edges": [
20,
35,
50,
65,
80,
95,
110,
125,
140,
155,
166
],
"hist": [
72,
188,
572,
904,
900,
636,
436,
184,
76,
32
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | train | 40,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
40000
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | train | 40,000 | immoral_action | 13 | 179 | 47.1972 | 15.5 | 37.86793 | {
"bin_edges": [
13,
30,
47,
64,
81,
98,
115,
132,
149,
166,
179
],
"hist": [
20050,
768,
3686,
6042,
4986,
2654,
1244,
456,
104,
10
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | train | 40,000 | immoral_consequence | 13 | 189 | 45.63275 | 13.5 | 36.53428 | {
"bin_edges": [
13,
31,
49,
67,
85,
103,
121,
139,
157,
175,
189
],
"hist": [
20200,
1364,
4962,
6406,
4110,
1974,
720,
220,
32,
12
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | train | 40,000 | moral_action | 13 | 177 | 47.0382 | 17.5 | 37.69755 | {
"bin_edges": [
13,
30,
47,
64,
81,
98,
115,
132,
149,
166,
177
],
"hist": [
20044,
740,
3822,
6150,
4832,
2680,
1134,
462,
122,
14
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | train | 40,000 | moral_consequence | 13 | 184 | 44.80715 | 14.5 | 35.35996 | {
"bin_edges": [
13,
31,
49,
67,
85,
103,
121,
139,
157,
175,
184
],
"hist": [
20168,
1346,
5266,
6804,
4096,
1608,
504,
154,
52,
2
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | validation | 4,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
4000
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | validation | 4,000 | immoral_action | 13 | 165 | 45.7255 | 15 | 36.34422 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
165
],
"hist": [
2006,
74,
348,
586,
496,
264,
162,
48,
14,
2
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | validation | 4,000 | immoral_consequence | 13 | 152 | 42.7865 | 15.5 | 33.21873 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
152
],
"hist": [
2010,
92,
254,
524,
550,
316,
150,
78,
18,
8
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | validation | 4,000 | moral_action | 13 | 159 | 45.7805 | 15.5 | 36.23986 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
159
],
"hist": [
2004,
46,
256,
548,
546,
320,
164,
84,
22,
10
]
} | false |
demelin/moral_stories | cls-consequence+action-minimal_pairs | validation | 4,000 | moral_consequence | 13 | 148 | 41.769 | 15.5 | 32.05551 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
148
],
"hist": [
2006,
98,
296,
574,
534,
296,
136,
38,
10,
12
]
} | false |
demelin/moral_stories | full | train | 12,000 | ID | 30 | 30 | 30 | 30 | 0 | {
"bin_edges": [
30,
30
],
"hist": [
12000
]
} | false |
demelin/moral_stories | full | train | 12,000 | immoral_action | 17 | 179 | 80.56758 | 78 | 22.84675 | {
"bin_edges": [
17,
34,
51,
68,
85,
102,
119,
136,
153,
170,
179
],
"hist": [
85,
759,
2868,
3601,
2582,
1311,
568,
190,
34,
2
]
} | false |
demelin/moral_stories | full | train | 12,000 | immoral_consequence | 14 | 189 | 76.99725 | 74 | 22.99687 | {
"bin_edges": [
14,
32,
50,
68,
86,
104,
122,
140,
158,
176,
189
],
"hist": [
146,
979,
3243,
3847,
2238,
1041,
380,
103,
17,
6
]
} | false |
demelin/moral_stories | full | train | 12,000 | intention | 16 | 127 | 43.18383 | 41 | 12.44797 | {
"bin_edges": [
16,
28,
40,
52,
64,
76,
88,
100,
112,
124,
127
],
"hist": [
758,
4511,
4100,
1825,
607,
127,
50,
18,
2,
2
]
} | false |
demelin/moral_stories | full | train | 12,000 | moral_action | 18 | 177 | 80.402 | 78 | 22.80604 | {
"bin_edges": [
18,
34,
50,
66,
82,
98,
114,
130,
146,
162,
177
],
"hist": [
67,
674,
2540,
3562,
2623,
1470,
671,
292,
88,
13
]
} | false |
demelin/moral_stories | full | train | 12,000 | moral_consequence | 16 | 184 | 75.36125 | 73 | 21.76409 | {
"bin_edges": [
16,
33,
50,
67,
84,
101,
118,
135,
152,
169,
184
],
"hist": [
164,
985,
3203,
3860,
2342,
966,
335,
104,
34,
7
]
} | false |
demelin/moral_stories | full | train | 12,000 | norm | 15 | 82 | 44.33225 | 44 | 9.7283 | {
"bin_edges": [
15,
22,
29,
36,
43,
50,
57,
64,
71,
78,
82
],
"hist": [
54,
584,
1670,
2803,
3212,
2395,
999,
244,
34,
5
]
} | false |
demelin/moral_stories | full | train | 12,000 | situation | 18 | 185 | 85.87767 | 84 | 25.27836 | {
"bin_edges": [
18,
35,
52,
69,
86,
103,
120,
137,
154,
171,
185
],
"hist": [
178,
735,
2146,
3199,
2792,
1716,
850,
321,
59,
4
]
} | false |
demelin/moral_stories | gen-action$context+consequence-norm_distance | validation | 2,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
2000
]
} | false |
demelin/moral_stories | gen-action$context+consequence-norm_distance | validation | 2,000 | immoral_action | 13 | 151 | 46.517 | 18.5 | 37.31886 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
151
],
"hist": [
1001,
24,
103,
225,
215,
206,
108,
79,
30,
9
]
} | false |
demelin/moral_stories | gen-action$context+consequence-norm_distance | validation | 2,000 | immoral_consequence | 13 | 169 | 44.3325 | 17.5 | 35.33984 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
169
],
"hist": [
1011,
74,
161,
289,
254,
128,
56,
18,
8,
1
]
} | false |
demelin/moral_stories | gen-action$context+consequence-norm_distance | validation | 2,000 | intention | 17 | 112 | 42.733 | 41 | 12.37127 | {
"bin_edges": [
17,
27,
37,
47,
57,
67,
77,
87,
97,
107,
112
],
"hist": [
118,
562,
646,
420,
164,
66,
18,
2,
0,
4
]
} | false |
demelin/moral_stories | gen-action$context+consequence-norm_distance | validation | 2,000 | moral_action | 13 | 163 | 46.2885 | 19 | 37.03477 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
163
],
"hist": [
1001,
38,
161,
294,
243,
158,
59,
33,
11,
2
]
} | false |
demelin/moral_stories | gen-action$context+consequence-norm_distance | validation | 2,000 | moral_consequence | 13 | 158 | 43.725 | 17.5 | 34.36939 | {
"bin_edges": [
13,
28,
43,
58,
73,
88,
103,
118,
133,
148,
158
],
"hist": [
1007,
49,
146,
303,
245,
153,
60,
21,
13,
3
]
} | false |
demelin/moral_stories | gen-action$context+consequence-norm_distance | validation | 2,000 | norm | 19 | 81 | 46.156 | 46 | 9.21282 | {
"bin_edges": [
19,
26,
33,
40,
47,
54,
61,
68,
75,
81
],
"hist": [
32,
124,
304,
562,
550,
314,
96,
14,
4
]
} | false |
demelin/moral_stories | gen-action$context+consequence-norm_distance | validation | 2,000 | situation | 20 | 166 | 85.536 | 84 | 25.65587 | {
"bin_edges": [
20,
35,
50,
65,
80,
95,
110,
125,
140,
155,
166
],
"hist": [
36,
94,
286,
452,
450,
318,
218,
92,
38,
16
]
} | false |
demelin/moral_stories | gen-action$context-norm_distance | train | 20,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
20000
]
} | false |
demelin/moral_stories | gen-action$context-norm_distance | train | 20,000 | immoral_action | 13 | 179 | 46.9272 | 15 | 37.56701 | {
"bin_edges": [
13,
30,
47,
64,
81,
98,
115,
132,
149,
166,
179
],
"hist": [
10026,
390,
1892,
3080,
2459,
1290,
597,
212,
49,
5
]
} | false |
demelin/moral_stories | gen-action$context-norm_distance | train | 20,000 | intention | 16 | 127 | 43.2478 | 41 | 12.43246 | {
"bin_edges": [
16,
28,
40,
52,
64,
76,
88,
100,
112,
124,
127
],
"hist": [
1230,
7516,
6826,
3098,
1002,
198,
92,
32,
2,
4
]
} | false |
demelin/moral_stories | gen-action$context-norm_distance | train | 20,000 | moral_action | 13 | 177 | 46.828 | 15.5 | 37.48572 | {
"bin_edges": [
13,
30,
47,
64,
81,
98,
115,
132,
149,
166,
177
],
"hist": [
10025,
386,
1953,
3092,
2389,
1314,
551,
226,
57,
7
]
} | false |
demelin/moral_stories | gen-action$context-norm_distance | train | 20,000 | norm | 15 | 82 | 43.8687 | 44 | 9.75138 | {
"bin_edges": [
15,
22,
29,
36,
43,
50,
57,
64,
71,
78,
82
],
"hist": [
98,
1064,
2990,
4778,
5296,
3774,
1562,
374,
56,
8
]
} | false |
demelin/moral_stories | gen-action$context-norm_distance | train | 20,000 | situation | 18 | 185 | 85.9598 | 84 | 25.27411 | {
"bin_edges": [
18,
35,
52,
69,
86,
103,
120,
137,
154,
171,
185
],
"hist": [
296,
1224,
3550,
5314,
4658,
2880,
1430,
550,
92,
6
]
} | false |
demelin/moral_stories | gen-consequence$action+context-norm_distance | validation | 2,000 | ID | 31 | 31 | 31 | 31 | 0 | {
"bin_edges": [
31,
31
],
"hist": [
2000
]
} | false |
demelin/moral_stories | gen-consequence$action+context-norm_distance | validation | 2,000 | immoral_action | 13 | 151 | 46.517 | 18.5 | 37.31886 | {
"bin_edges": [
13,
27,
41,
55,
69,
83,
97,
111,
125,
139,
151
],
"hist": [
1001,
24,
103,
225,
215,
206,
108,
79,
30,
9
]
} | false |
demelin/moral_stories | gen-consequence$action+context-norm_distance | validation | 2,000 | immoral_consequence | 13 | 169 | 44.3325 | 17.5 | 35.33984 | {
"bin_edges": [
13,
29,
45,
61,
77,
93,
109,
125,
141,
157,
169
],
"hist": [
1011,
74,
161,
289,
254,
128,
56,
18,
8,
1
]
} | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.