Datasets:

Modalities:
Text
ArXiv:
Libraries:
Datasets
License:
jilee commited on
Commit
68de5a1
·
1 Parent(s): 8678238

Upload TexPrax.py

Browse files
Files changed (1) hide show
  1. TexPrax.py +14 -44
TexPrax.py CHANGED
@@ -73,7 +73,8 @@ class TexPraxDataset(datasets.GeneratorBasedBuilder):
73
  # Note: ID consists of <dialog-id_sentence-id_turn-id>
74
  "id": datasets.Value("string"),
75
  "sentence": datasets.Value("string"),
76
- "label": datasets.Value("string")
 
77
  # These are the features of your dataset like images, labels ...
78
  }
79
  )
@@ -119,7 +120,8 @@ class TexPraxDataset(datasets.GeneratorBasedBuilder):
119
  "B-LOC",
120
  ]
121
  )
122
- )
 
123
  }
124
  )
125
  return datasets.DatasetInfo(
@@ -152,32 +154,16 @@ class TexPraxDataset(datasets.GeneratorBasedBuilder):
152
  name=datasets.Split.TRAIN,
153
  # These kwargs will be passed to _generate_examples
154
  gen_kwargs={
155
- "filepath": os.path.join(data_dir, "industrie_sents_batch_1.csv"),
156
- "split": "batch-1-industrie",
157
- },
158
- ),
159
- datasets.SplitGenerator(
160
- name=datasets.Split.TRAIN,
161
- # These kwargs will be passed to _generate_examples
162
- gen_kwargs={
163
- "filepath": os.path.join(data_dir, "zerspanung_sents_batch_1.csv"),
164
- "split": "batch-1-zerspanung",
165
- },
166
- ),
167
- datasets.SplitGenerator(
168
- name=datasets.Split.TRAIN,
169
- # These kwargs will be passed to _generate_examples
170
- gen_kwargs={
171
- "filepath": os.path.join(data_dir, "sents_batch_2.csv"),
172
- "split": "batch-2",
173
  },
174
  ),
175
  datasets.SplitGenerator(
176
  name=datasets.Split.TEST,
177
  # These kwargs will be passed to _generate_examples
178
  gen_kwargs={
179
- "filepath": os.path.join(data_dir, "sents_batch_3.csv"),
180
- "split": "batch-3"
181
  },
182
  ),
183
  ]
@@ -189,32 +175,15 @@ class TexPraxDataset(datasets.GeneratorBasedBuilder):
189
  name=datasets.Split.TRAIN,
190
  # These kwargs will be passed to _generate_examples
191
  gen_kwargs={
192
- "filepath": os.path.join(data_dir, "industrie_entities_batch_1.csv"),
193
- "split": "batch-1-industrie",
194
  },
195
- ),
196
- datasets.SplitGenerator(
197
- name=datasets.Split.TRAIN,
198
- # These kwargs will be passed to _generate_examples
199
- gen_kwargs={
200
- "filepath": os.path.join(data_dir, "zerspanung_entities_batch_1.csv"),
201
- "split": "batch-1-zerspanung",
202
- },
203
- ),
204
- datasets.SplitGenerator(
205
- name=datasets.Split.TRAIN,
206
- # These kwargs will be passed to _generate_examples
207
- gen_kwargs={
208
- "filepath": os.path.join(data_dir, "entities_batch_2.csv"),
209
- "split": "batch-2",
210
- },
211
- ),
212
  datasets.SplitGenerator(
213
  name=datasets.Split.TEST,
214
  # These kwargs will be passed to _generate_examples
215
  gen_kwargs={
216
- "filepath": os.path.join(data_dir, "entities_batch_3.csv"),
217
- "split": "batch-3"
218
  },
219
  ),
220
  ]
@@ -228,7 +197,6 @@ class TexPraxDataset(datasets.GeneratorBasedBuilder):
228
  creader = csv.reader(f, delimiter=';', quotechar='"')
229
  next(creader) # skip header
230
  for key, row in enumerate(creader):
231
- print(key, row)
232
  if self.config.name == "sentence_cl":
233
  idx, sentence, label = row
234
  # Yields examples as (key, example) tuples
@@ -236,6 +204,7 @@ class TexPraxDataset(datasets.GeneratorBasedBuilder):
236
  "id": idx,
237
  "sentence": sentence,
238
  "label": label,
 
239
  }
240
  else:
241
  idx, sentence, labels = row
@@ -244,6 +213,7 @@ class TexPraxDataset(datasets.GeneratorBasedBuilder):
244
  "id": idx,
245
  "tokens": [t.strip() for t in ast.literal_eval(sentence)],
246
  "entities": [l.strip() for l in ast.literal_eval(labels)],
 
247
  }
248
 
249
 
 
73
  # Note: ID consists of <dialog-id_sentence-id_turn-id>
74
  "id": datasets.Value("string"),
75
  "sentence": datasets.Value("string"),
76
+ "label": datasets.Value("string"),
77
+ "subsplit": datasets.Value("string"),
78
  # These are the features of your dataset like images, labels ...
79
  }
80
  )
 
120
  "B-LOC",
121
  ]
122
  )
123
+ ),
124
+ "subsplit": datasets.Value("string"),
125
  }
126
  )
127
  return datasets.DatasetInfo(
 
154
  name=datasets.Split.TRAIN,
155
  # These kwargs will be passed to _generate_examples
156
  gen_kwargs={
157
+ "filepath": os.path.join(data_dir, "sents_train.csv"),
158
+ "split": "train",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
159
  },
160
  ),
161
  datasets.SplitGenerator(
162
  name=datasets.Split.TEST,
163
  # These kwargs will be passed to _generate_examples
164
  gen_kwargs={
165
+ "filepath": os.path.join(data_dir, "sents_test.csv"),
166
+ "split": "test"
167
  },
168
  ),
169
  ]
 
175
  name=datasets.Split.TRAIN,
176
  # These kwargs will be passed to _generate_examples
177
  gen_kwargs={
178
+ "filepath": os.path.join(data_dir, "entities_train.csv"),
179
+ "split": "train",
180
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
181
  datasets.SplitGenerator(
182
  name=datasets.Split.TEST,
183
  # These kwargs will be passed to _generate_examples
184
  gen_kwargs={
185
+ "filepath": os.path.join(data_dir, "entities_test.csv"),
186
+ "split": "test"
187
  },
188
  ),
189
  ]
 
197
  creader = csv.reader(f, delimiter=';', quotechar='"')
198
  next(creader) # skip header
199
  for key, row in enumerate(creader):
 
200
  if self.config.name == "sentence_cl":
201
  idx, sentence, label = row
202
  # Yields examples as (key, example) tuples
 
204
  "id": idx,
205
  "sentence": sentence,
206
  "label": label,
207
+ "subsplit": split,
208
  }
209
  else:
210
  idx, sentence, labels = row
 
213
  "id": idx,
214
  "tokens": [t.strip() for t in ast.literal_eval(sentence)],
215
  "entities": [l.strip() for l in ast.literal_eval(labels)],
216
+ "subsplit": split,
217
  }
218
 
219