Convert dataset to Parquet

#1
by ygorg - opened
Files changed (33) hide show
  1. Basque_clinical/test-00000-of-00001.parquet +3 -0
  2. Basque_clinical/train-00000-of-00001.parquet +3 -0
  3. Basque_clinical/validation-00000-of-00001.parquet +3 -0
  4. Basque_temporal/test-00000-of-00001.parquet +3 -0
  5. Basque_temporal/train-00000-of-00001.parquet +3 -0
  6. Basque_temporal/validation-00000-of-00001.parquet +3 -0
  7. English_clinical/test-00000-of-00001.parquet +3 -0
  8. English_clinical/train-00000-of-00001.parquet +3 -0
  9. English_clinical/validation-00000-of-00001.parquet +3 -0
  10. English_temporal/test-00000-of-00001.parquet +3 -0
  11. English_temporal/train-00000-of-00001.parquet +3 -0
  12. English_temporal/validation-00000-of-00001.parquet +3 -0
  13. French_clinical/test-00000-of-00001.parquet +3 -0
  14. French_clinical/train-00000-of-00001.parquet +3 -0
  15. French_clinical/validation-00000-of-00001.parquet +3 -0
  16. French_temporal/test-00000-of-00001.parquet +3 -0
  17. French_temporal/train-00000-of-00001.parquet +3 -0
  18. French_temporal/validation-00000-of-00001.parquet +3 -0
  19. Italian_clinical/test-00000-of-00001.parquet +3 -0
  20. Italian_clinical/train-00000-of-00001.parquet +3 -0
  21. Italian_clinical/validation-00000-of-00001.parquet +3 -0
  22. Italian_temporal/test-00000-of-00001.parquet +3 -0
  23. Italian_temporal/train-00000-of-00001.parquet +3 -0
  24. Italian_temporal/validation-00000-of-00001.parquet +3 -0
  25. README.md +395 -0
  26. Spanish_clinical/test-00000-of-00001.parquet +3 -0
  27. Spanish_clinical/train-00000-of-00001.parquet +3 -0
  28. Spanish_clinical/validation-00000-of-00001.parquet +3 -0
  29. Spanish_temporal/test-00000-of-00001.parquet +3 -0
  30. Spanish_temporal/train-00000-of-00001.parquet +3 -0
  31. Spanish_temporal/validation-00000-of-00001.parquet +3 -0
  32. E3C.py → _attic/E3C.py +0 -0
  33. test_e3c.py +0 -8
Basque_clinical/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5303c94b814f7df47f6670c2117fac715befcff134058c7c6cdf6ca131c5846f
3
+ size 315707
Basque_clinical/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0587ecc130e88894d8fa0bd3f91c3489e9d7c5117d6e6fe4e40c55c08e474f9b
3
+ size 161073
Basque_clinical/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6389a5498dcedb63c76482310f1945e31ff27bac305b07cf64bd9309d6c734ae
3
+ size 28482
Basque_temporal/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3be5e14eee193f1f53dabf46b290f7edcd1ce8c647428e1cf1c3e16fe5228b2
3
+ size 71994
Basque_temporal/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e906c0a91f4ad77c8c0cc1233e2aaad4fa30c60a7daddf4764763c49aaf2065b
3
+ size 232201
Basque_temporal/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d55d9c48e45d26c1b7e5b941fd56c6abe5aa6de6a1efbf45a4fb501e8ef42e42
3
+ size 41753
English_clinical/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac9d4a7d36eec5e6ae9858826621c95276e0f41c39dec838ad8dce1bce87a547
3
+ size 223863
English_clinical/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfe8d124c138d902bf9b9356d8f2ec94d585ec7d6be26d565685058b8260ee55
3
+ size 385142
English_clinical/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4266d963007306eca7f1920694f8f96f76a86d44de0f0cee1bec99636dc1871c
3
+ size 63191
English_temporal/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:076af91ba34be5765fa29a673c3fee9b142a5cadf4e0ef362cd6a5753b9bca3e
3
+ size 52858
English_temporal/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e63ad7f2e4d877abc901b8840616485fc58b9517d734da36a752737935033b74
3
+ size 167723
English_temporal/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:433081305e597968fafa988481a7ab771fd947d6e174e4a1a87b864dd6cbfa29
3
+ size 31038
French_clinical/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7980f3f513a8b3db62186d03b6330682b4fe8257b5d1c0a6dd7d993e30af1a92
3
+ size 211991
French_clinical/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fff6f6aa8cc8efe0c7f1441630d9914ba3851be36f5f327b940a6b4ff906bb8
3
+ size 378733
French_clinical/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db31d31ccb2767ae361edad2f4e2081fe1c107db71dbb3d783c828da55a1562e
3
+ size 66856
French_temporal/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5385e610c7a74c8dcf9c5835bda1ef32e1b8610f5f9ba118eb294d014a6cbae8
3
+ size 51399
French_temporal/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9fc6442f5365bc97ecadc517818573971b0d8052adaa85868e3c82789b2c22b
3
+ size 154387
French_temporal/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5fb23da355c62bcc817d87df33c37bdf5c10344d997dff733cc8fa2248e368e
3
+ size 29589
Italian_clinical/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6f4d58f22a4ce635793fc9b1658b178618bffd65856efec63d89e1a14de955e
3
+ size 227792
Italian_clinical/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a19c8ab559e965d7c36884a395d36139146e9c5135d4267871960aa39c0aed0d
3
+ size 413833
Italian_clinical/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f74bb6544e2a2d300068126a922fcd582ef86aa110c698eb9dd9de82a886b4a1
3
+ size 68398
Italian_temporal/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:391d975321765435d57d791e38353a3c23124d56a656c4a067ad77fe770f2736
3
+ size 54713
Italian_temporal/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d14a3cdd6c27a2410921bc9b556038ff08c73f6b194d45b4d851da5dd77ba909
3
+ size 163704
Italian_temporal/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb9e93fffc1becbfb38d5dda1e5fa038f1515f2b069efd9f16334a4ca1ae044f
3
+ size 32118
README.md ADDED
@@ -0,0 +1,395 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ dataset_info:
3
+ - config_name: Basque_clinical
4
+ features:
5
+ - name: id
6
+ dtype: string
7
+ - name: text
8
+ dtype: string
9
+ - name: tokens
10
+ sequence: string
11
+ - name: ner_tags
12
+ sequence:
13
+ class_label:
14
+ names:
15
+ '0': O
16
+ '1': B-CLINENTITY
17
+ '2': I-CLINENTITY
18
+ splits:
19
+ - name: train
20
+ num_bytes: 414353
21
+ num_examples: 1392
22
+ - name: validation
23
+ num_bytes: 57192
24
+ num_examples: 202
25
+ - name: test
26
+ num_bytes: 848051
27
+ num_examples: 3126
28
+ download_size: 505262
29
+ dataset_size: 1319596
30
+ - config_name: Basque_temporal
31
+ features:
32
+ - name: id
33
+ dtype: string
34
+ - name: text
35
+ dtype: string
36
+ - name: tokens
37
+ sequence: string
38
+ - name: ner_tags
39
+ sequence:
40
+ class_label:
41
+ names:
42
+ '0': O
43
+ '1': B-EVENT
44
+ '2': B-ACTOR
45
+ '3': B-BODYPART
46
+ '4': B-TIMEX3
47
+ '5': B-RML
48
+ '6': I-EVENT
49
+ '7': I-ACTOR
50
+ '8': I-BODYPART
51
+ '9': I-TIMEX3
52
+ '10': I-RML
53
+ splits:
54
+ - name: train
55
+ num_bytes: 590623
56
+ num_examples: 2188
57
+ - name: validation
58
+ num_bytes: 87750
59
+ num_examples: 312
60
+ - name: test
61
+ num_bytes: 169678
62
+ num_examples: 626
63
+ download_size: 345948
64
+ dataset_size: 848051
65
+ - config_name: English_clinical
66
+ features:
67
+ - name: id
68
+ dtype: string
69
+ - name: text
70
+ dtype: string
71
+ - name: tokens
72
+ sequence: string
73
+ - name: ner_tags
74
+ sequence:
75
+ class_label:
76
+ names:
77
+ '0': O
78
+ '1': B-CLINENTITY
79
+ '2': I-CLINENTITY
80
+ splits:
81
+ - name: train
82
+ num_bytes: 1209779
83
+ num_examples: 2510
84
+ - name: validation
85
+ num_bytes: 168864
86
+ num_examples: 363
87
+ - name: test
88
+ num_bytes: 689360
89
+ num_examples: 1520
90
+ download_size: 672196
91
+ dataset_size: 2068003
92
+ - config_name: English_temporal
93
+ features:
94
+ - name: id
95
+ dtype: string
96
+ - name: text
97
+ dtype: string
98
+ - name: tokens
99
+ sequence: string
100
+ - name: ner_tags
101
+ sequence:
102
+ class_label:
103
+ names:
104
+ '0': O
105
+ '1': B-EVENT
106
+ '2': B-ACTOR
107
+ '3': B-BODYPART
108
+ '4': B-TIMEX3
109
+ '5': B-RML
110
+ '6': I-EVENT
111
+ '7': I-ACTOR
112
+ '8': I-BODYPART
113
+ '9': I-TIMEX3
114
+ '10': I-RML
115
+ splits:
116
+ - name: train
117
+ num_bytes: 486527
118
+ num_examples: 1064
119
+ - name: validation
120
+ num_bytes: 68753
121
+ num_examples: 152
122
+ - name: test
123
+ num_bytes: 134080
124
+ num_examples: 304
125
+ download_size: 251619
126
+ dataset_size: 689360
127
+ - config_name: French_clinical
128
+ features:
129
+ - name: id
130
+ dtype: string
131
+ - name: text
132
+ dtype: string
133
+ - name: tokens
134
+ sequence: string
135
+ - name: ner_tags
136
+ sequence:
137
+ class_label:
138
+ names:
139
+ '0': O
140
+ '1': B-CLINENTITY
141
+ '2': I-CLINENTITY
142
+ splits:
143
+ - name: train
144
+ num_bytes: 1237658
145
+ num_examples: 2087
146
+ - name: validation
147
+ num_bytes: 189900
148
+ num_examples: 302
149
+ - name: test
150
+ num_bytes: 682748
151
+ num_examples: 1109
152
+ download_size: 657580
153
+ dataset_size: 2110306
154
+ - config_name: French_temporal
155
+ features:
156
+ - name: id
157
+ dtype: string
158
+ - name: text
159
+ dtype: string
160
+ - name: tokens
161
+ sequence: string
162
+ - name: ner_tags
163
+ sequence:
164
+ class_label:
165
+ names:
166
+ '0': O
167
+ '1': B-EVENT
168
+ '2': B-ACTOR
169
+ '3': B-BODYPART
170
+ '4': B-TIMEX3
171
+ '5': B-RML
172
+ '6': I-EVENT
173
+ '7': I-ACTOR
174
+ '8': I-BODYPART
175
+ '9': I-TIMEX3
176
+ '10': I-RML
177
+ splits:
178
+ - name: train
179
+ num_bytes: 480645
180
+ num_examples: 776
181
+ - name: validation
182
+ num_bytes: 67127
183
+ num_examples: 111
184
+ - name: test
185
+ num_bytes: 134976
186
+ num_examples: 222
187
+ download_size: 235375
188
+ dataset_size: 682748
189
+ - config_name: Italian_clinical
190
+ features:
191
+ - name: id
192
+ dtype: string
193
+ - name: text
194
+ dtype: string
195
+ - name: tokens
196
+ sequence: string
197
+ - name: ner_tags
198
+ sequence:
199
+ class_label:
200
+ names:
201
+ '0': O
202
+ '1': B-CLINENTITY
203
+ '2': I-CLINENTITY
204
+ splits:
205
+ - name: train
206
+ num_bytes: 1256397
207
+ num_examples: 2128
208
+ - name: validation
209
+ num_bytes: 178790
210
+ num_examples: 308
211
+ - name: test
212
+ num_bytes: 691524
213
+ num_examples: 1146
214
+ download_size: 710023
215
+ dataset_size: 2126711
216
+ - config_name: Italian_temporal
217
+ features:
218
+ - name: id
219
+ dtype: string
220
+ - name: text
221
+ dtype: string
222
+ - name: tokens
223
+ sequence: string
224
+ - name: ner_tags
225
+ sequence:
226
+ class_label:
227
+ names:
228
+ '0': O
229
+ '1': B-EVENT
230
+ '2': B-ACTOR
231
+ '3': B-BODYPART
232
+ '4': B-TIMEX3
233
+ '5': B-RML
234
+ '6': I-EVENT
235
+ '7': I-ACTOR
236
+ '8': I-BODYPART
237
+ '9': I-TIMEX3
238
+ '10': I-RML
239
+ splits:
240
+ - name: train
241
+ num_bytes: 484067
242
+ num_examples: 802
243
+ - name: validation
244
+ num_bytes: 71242
245
+ num_examples: 114
246
+ - name: test
247
+ num_bytes: 136215
248
+ num_examples: 230
249
+ download_size: 250535
250
+ dataset_size: 691524
251
+ - config_name: Spanish_clinical
252
+ features:
253
+ - name: id
254
+ dtype: string
255
+ - name: text
256
+ dtype: string
257
+ - name: tokens
258
+ sequence: string
259
+ - name: ner_tags
260
+ sequence:
261
+ class_label:
262
+ names:
263
+ '0': O
264
+ '1': B-CLINENTITY
265
+ '2': I-CLINENTITY
266
+ splits:
267
+ - name: train
268
+ num_bytes: 1187917
269
+ num_examples: 2050
270
+ - name: validation
271
+ num_bytes: 167939
272
+ num_examples: 297
273
+ - name: test
274
+ num_bytes: 680981
275
+ num_examples: 1134
276
+ download_size: 668173
277
+ dataset_size: 2036837
278
+ - config_name: Spanish_temporal
279
+ features:
280
+ - name: id
281
+ dtype: string
282
+ - name: text
283
+ dtype: string
284
+ - name: tokens
285
+ sequence: string
286
+ - name: ner_tags
287
+ sequence:
288
+ class_label:
289
+ names:
290
+ '0': O
291
+ '1': B-EVENT
292
+ '2': B-ACTOR
293
+ '3': B-BODYPART
294
+ '4': B-TIMEX3
295
+ '5': B-RML
296
+ '6': I-EVENT
297
+ '7': I-ACTOR
298
+ '8': I-BODYPART
299
+ '9': I-TIMEX3
300
+ '10': I-RML
301
+ splits:
302
+ - name: train
303
+ num_bytes: 474711
304
+ num_examples: 793
305
+ - name: validation
306
+ num_bytes: 68540
307
+ num_examples: 114
308
+ - name: test
309
+ num_bytes: 137730
310
+ num_examples: 227
311
+ download_size: 246575
312
+ dataset_size: 680981
313
+ configs:
314
+ - config_name: Basque_clinical
315
+ data_files:
316
+ - split: train
317
+ path: Basque_clinical/train-*
318
+ - split: validation
319
+ path: Basque_clinical/validation-*
320
+ - split: test
321
+ path: Basque_clinical/test-*
322
+ - config_name: Basque_temporal
323
+ data_files:
324
+ - split: train
325
+ path: Basque_temporal/train-*
326
+ - split: validation
327
+ path: Basque_temporal/validation-*
328
+ - split: test
329
+ path: Basque_temporal/test-*
330
+ - config_name: English_clinical
331
+ data_files:
332
+ - split: train
333
+ path: English_clinical/train-*
334
+ - split: validation
335
+ path: English_clinical/validation-*
336
+ - split: test
337
+ path: English_clinical/test-*
338
+ - config_name: English_temporal
339
+ data_files:
340
+ - split: train
341
+ path: English_temporal/train-*
342
+ - split: validation
343
+ path: English_temporal/validation-*
344
+ - split: test
345
+ path: English_temporal/test-*
346
+ - config_name: French_clinical
347
+ data_files:
348
+ - split: train
349
+ path: French_clinical/train-*
350
+ - split: validation
351
+ path: French_clinical/validation-*
352
+ - split: test
353
+ path: French_clinical/test-*
354
+ default: true
355
+ - config_name: French_temporal
356
+ data_files:
357
+ - split: train
358
+ path: French_temporal/train-*
359
+ - split: validation
360
+ path: French_temporal/validation-*
361
+ - split: test
362
+ path: French_temporal/test-*
363
+ - config_name: Italian_clinical
364
+ data_files:
365
+ - split: train
366
+ path: Italian_clinical/train-*
367
+ - split: validation
368
+ path: Italian_clinical/validation-*
369
+ - split: test
370
+ path: Italian_clinical/test-*
371
+ - config_name: Italian_temporal
372
+ data_files:
373
+ - split: train
374
+ path: Italian_temporal/train-*
375
+ - split: validation
376
+ path: Italian_temporal/validation-*
377
+ - split: test
378
+ path: Italian_temporal/test-*
379
+ - config_name: Spanish_clinical
380
+ data_files:
381
+ - split: train
382
+ path: Spanish_clinical/train-*
383
+ - split: validation
384
+ path: Spanish_clinical/validation-*
385
+ - split: test
386
+ path: Spanish_clinical/test-*
387
+ - config_name: Spanish_temporal
388
+ data_files:
389
+ - split: train
390
+ path: Spanish_temporal/train-*
391
+ - split: validation
392
+ path: Spanish_temporal/validation-*
393
+ - split: test
394
+ path: Spanish_temporal/test-*
395
+ ---
Spanish_clinical/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4b88cfaefbf6f0ecfb2c41fc52931df3938fdbf7c717688574cb8bd1537a78d
3
+ size 222398
Spanish_clinical/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:989f7aa4141524ecd0eb6982b43435237449e7142c5f746523b5c89817285d6f
3
+ size 382838
Spanish_clinical/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8614c03c6d2ef3831c78058d9b273dbd002358c434c3ddfc646c4b4eb525fd27
3
+ size 62937
Spanish_temporal/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07e3863984f03b2ce72c2bb6985579224b58a7594f2165715a341684eb370d29
3
+ size 55677
Spanish_temporal/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e56f7b1b4bc69d40cc276f772a549a495d3cc0b6c4c7ab1e256cddc89faf1c89
3
+ size 160007
Spanish_temporal/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4eb7cac63c03ced32bd37e866e9323d82c8f5fae479a9f9acdd4c9b3c53e108
3
+ size 30891
E3C.py → _attic/E3C.py RENAMED
File without changes
test_e3c.py DELETED
@@ -1,8 +0,0 @@
1
- import json
2
-
3
- from datasets import load_dataset
4
-
5
- dataset = load_dataset("./E3C.py", name="French")
6
- print(dataset)
7
- # print(dataset["train"][0])
8
- print(json.dumps(dataset["train"][0], sort_keys=True, indent=4))