JuncheolK commited on
Commit
2109723
·
verified ·
1 Parent(s): 9b2f732

Upload dataset_info.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. dataset_info.json +296 -0
dataset_info.json ADDED
@@ -0,0 +1,296 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "mncai/ai2_arc": [
3
+ {
4
+ "name": "ARC-Easy",
5
+ "size": 1000
6
+ },
7
+ {
8
+ "name": "ARC-Easy_ko",
9
+ "size": 1000
10
+ },
11
+ {
12
+ "name": "ARC-Challenge",
13
+ "size": 1000
14
+ },
15
+ {
16
+ "name": "ARC-Challenge_ko",
17
+ "size": 1000
18
+ }
19
+ ],
20
+ "mncai/hellaswag": [
21
+ {
22
+ "name": "default",
23
+ "size": 1000
24
+ },
25
+ {
26
+ "name": "default_ko",
27
+ "size": 1000
28
+ }
29
+ ],
30
+ "mncai/commonsense_qa": [
31
+ {
32
+ "name": "default",
33
+ "size": 1000
34
+ },
35
+ {
36
+ "name": "default_ko",
37
+ "size": 1000
38
+ }
39
+ ],
40
+ "mncai/mbpp": [
41
+ {
42
+ "name": "sanitized",
43
+ "size": 1000
44
+ },
45
+ {
46
+ "name": "sanitized_ko",
47
+ "size": 1000
48
+ }
49
+ ],
50
+ "mncai/trivia_qa": [
51
+ {
52
+ "name": "rc",
53
+ "size": 100
54
+ },
55
+ {
56
+ "name": "rc_ko",
57
+ "size": 100
58
+ }
59
+ ],
60
+ "mncai/winogrande": [
61
+ {
62
+ "name": "winogrande_debiased",
63
+ "size": 1000
64
+ },
65
+ {
66
+ "name": "winogrande_debiased_ko",
67
+ "size": 1000
68
+ }
69
+ ],
70
+ "mncai/openbookqa": [
71
+ {
72
+ "name": "additional",
73
+ "size": 1000
74
+ },
75
+ {
76
+ "name": "additional_ko",
77
+ "size": 1000
78
+ }
79
+ ],
80
+ "mncai/KMMLU": [
81
+ {
82
+ "name": "Accounting",
83
+ "size": 1000
84
+ },
85
+ {
86
+ "name": "Agricultural-Sciences",
87
+ "size": 1000
88
+ },
89
+ {
90
+ "name": "Aviation-Engineering-and-Maintenance",
91
+ "size": 1000
92
+ },
93
+ {
94
+ "name": "Biology",
95
+ "size": 1000
96
+ },
97
+ {
98
+ "name": "Chemical-Engineering",
99
+ "size": 1000
100
+ },
101
+ {
102
+ "name": "Chemistry",
103
+ "size": 1000
104
+ },
105
+ {
106
+ "name": "Civil-Engineering",
107
+ "size": 1000
108
+ },
109
+ {
110
+ "name": "Computer-Science",
111
+ "size": 1000
112
+ },
113
+ {
114
+ "name": "Construction",
115
+ "size": 1000
116
+ },
117
+ {
118
+ "name": "Criminal-Law",
119
+ "size": 1000
120
+ },
121
+ {
122
+ "name": "Ecology",
123
+ "size": 1000
124
+ },
125
+ {
126
+ "name": "Economics",
127
+ "size": 1000
128
+ },
129
+ {
130
+ "name": "Education",
131
+ "size": 1000
132
+ },
133
+ {
134
+ "name": "Electrical-Engineering",
135
+ "size": 1000
136
+ },
137
+ {
138
+ "name": "Electronics-Engineering",
139
+ "size": 1000
140
+ },
141
+ {
142
+ "name": "Energy-Management",
143
+ "size": 1000
144
+ },
145
+ {
146
+ "name": "Environmental-Science",
147
+ "size": 1000
148
+ },
149
+ {
150
+ "name": "Fashion",
151
+ "size": 1000
152
+ },
153
+ {
154
+ "name": "Food-Processing",
155
+ "size": 1000
156
+ },
157
+ {
158
+ "name": "Gas-Technology-and-Engineering",
159
+ "size": 1000
160
+ },
161
+ {
162
+ "name": "Geomatics",
163
+ "size": 1000
164
+ },
165
+ {
166
+ "name": "Health",
167
+ "size": 1000
168
+ },
169
+ {
170
+ "name": "Industrial-Engineer",
171
+ "size": 1000
172
+ },
173
+ {
174
+ "name": "Information-Technology",
175
+ "size": 1000
176
+ },
177
+ {
178
+ "name": "Interior-Architecture-and-Design",
179
+ "size": 1000
180
+ },
181
+ {
182
+ "name": "Law",
183
+ "size": 1000
184
+ },
185
+ {
186
+ "name": "Machine-Design-and-Manufacturing",
187
+ "size": 1000
188
+ },
189
+ {
190
+ "name": "Management",
191
+ "size": 1000
192
+ },
193
+ {
194
+ "name": "Maritime-Engineering",
195
+ "size": 1000
196
+ },
197
+ {
198
+ "name": "Marketing",
199
+ "size": 1000
200
+ },
201
+ {
202
+ "name": "Materials-Engineering",
203
+ "size": 1000
204
+ },
205
+ {
206
+ "name": "Mechanical-Engineering",
207
+ "size": 1000
208
+ },
209
+ {
210
+ "name": "Nondestructive-Testing",
211
+ "size": 1000
212
+ },
213
+ {
214
+ "name": "Patent",
215
+ "size": 1000
216
+ },
217
+ {
218
+ "name": "Political-Science-and-Sociology",
219
+ "size": 1000
220
+ },
221
+ {
222
+ "name": "Psychology",
223
+ "size": 1000
224
+ },
225
+ {
226
+ "name": "Public-Safety",
227
+ "size": 1000
228
+ },
229
+ {
230
+ "name": "Railway-and-Automotive-Engineering",
231
+ "size": 1000
232
+ },
233
+ {
234
+ "name": "Real-Estate",
235
+ "size": 1000
236
+ },
237
+ {
238
+ "name": "Refrigerating-Machinery",
239
+ "size": 1000
240
+ },
241
+ {
242
+ "name": "Social-Welfare",
243
+ "size": 1000
244
+ },
245
+ {
246
+ "name": "Taxation",
247
+ "size": 1000
248
+ },
249
+ {
250
+ "name": "Telecommunications-and-Wireless-Technology",
251
+ "size": 1000
252
+ },
253
+ {
254
+ "name": "Korean-History",
255
+ "size": 1000
256
+ },
257
+ {
258
+ "name": "Math",
259
+ "size": 1000
260
+ }
261
+ ],
262
+ "mncai/common-sense-generated": [
263
+ {
264
+ "name": "topics_en",
265
+ "size": 100
266
+ },
267
+ {
268
+ "name": "topics_ko",
269
+ "size": 100
270
+ },
271
+ {
272
+ "name": "keywords_en",
273
+ "size": 100
274
+ },
275
+ {
276
+ "name": "keywords_ko",
277
+ "size": 100
278
+ },
279
+ {
280
+ "name": "inputs_en",
281
+ "size": 100
282
+ },
283
+ {
284
+ "name": "inputs_ko",
285
+ "size": 100
286
+ },
287
+ {
288
+ "name": "outputs_en",
289
+ "size": 100
290
+ },
291
+ {
292
+ "name": "outputs_ko",
293
+ "size": 100
294
+ }
295
+ ]
296
+ }