Update openwebtext_split.py
Browse files- openwebtext_split.py +5 -6
openwebtext_split.py
CHANGED
|
@@ -54,10 +54,7 @@ def custom_iter_archive(path_or_buf, _filter=lambda x: True):
|
|
| 54 |
if not file_path.endswith('xz'):
|
| 55 |
continue
|
| 56 |
file_obj = stream.extractfile(tarinfo)
|
| 57 |
-
|
| 58 |
-
for txt_file in file_obj:
|
| 59 |
-
if txt_file:
|
| 60 |
-
yield file_path, file_obj
|
| 61 |
stream.members = []
|
| 62 |
del stream
|
| 63 |
if hasattr(path_or_buf, "read"):
|
|
@@ -97,8 +94,10 @@ class Openwebtext(datasets.GeneratorBasedBuilder):
|
|
| 97 |
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"files": custom_iter_archive(archive, test_filter)}),
|
| 98 |
]
|
| 99 |
|
| 100 |
-
def _generate_examples(self,
|
| 101 |
"""Yields examples."""
|
| 102 |
-
for idx, filepath in enumerate(
|
|
|
|
|
|
|
| 103 |
with open(filepath, encoding="utf-8") as f:
|
| 104 |
yield idx, {"text": re.sub("\n\n\n+", "\n\n", f.read()).strip()}
|
|
|
|
| 54 |
if not file_path.endswith('xz'):
|
| 55 |
continue
|
| 56 |
file_obj = stream.extractfile(tarinfo)
|
| 57 |
+
yield file_path, file_obj
|
|
|
|
|
|
|
|
|
|
| 58 |
stream.members = []
|
| 59 |
del stream
|
| 60 |
if hasattr(path_or_buf, "read"):
|
|
|
|
| 94 |
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"files": custom_iter_archive(archive, test_filter)}),
|
| 95 |
]
|
| 96 |
|
| 97 |
+
def _generate_examples(self, files):
|
| 98 |
"""Yields examples."""
|
| 99 |
+
for idx, filepath in enumerate(files):
|
| 100 |
+
if not filepath.endswith('txt'):
|
| 101 |
+
continue
|
| 102 |
with open(filepath, encoding="utf-8") as f:
|
| 103 |
yield idx, {"text": re.sub("\n\n\n+", "\n\n", f.read()).strip()}
|