Upload nle_hf_dataset.py
Browse files- nle_hf_dataset.py +1 -1
nle_hf_dataset.py
CHANGED
|
@@ -111,6 +111,7 @@ class NleHfDataset(datasets.GeneratorBasedBuilder):
|
|
| 111 |
else:
|
| 112 |
urls = self.config.data_files["train"]
|
| 113 |
|
|
|
|
| 114 |
filepaths = [dl_manager.download(url) for url in urls]
|
| 115 |
return [
|
| 116 |
datasets.SplitGenerator(
|
|
@@ -121,7 +122,6 @@ class NleHfDataset(datasets.GeneratorBasedBuilder):
|
|
| 121 |
for i, filepath in enumerate(filepaths):
|
| 122 |
if self.config.name == "metadata":
|
| 123 |
with open(filepath, "r") as f:
|
| 124 |
-
print(filepath)
|
| 125 |
data = json.load(f)
|
| 126 |
yield i, data
|
| 127 |
else:
|
|
|
|
| 111 |
else:
|
| 112 |
urls = self.config.data_files["train"]
|
| 113 |
|
| 114 |
+
print(urls)
|
| 115 |
filepaths = [dl_manager.download(url) for url in urls]
|
| 116 |
return [
|
| 117 |
datasets.SplitGenerator(
|
|
|
|
| 122 |
for i, filepath in enumerate(filepaths):
|
| 123 |
if self.config.name == "metadata":
|
| 124 |
with open(filepath, "r") as f:
|
|
|
|
| 125 |
data = json.load(f)
|
| 126 |
yield i, data
|
| 127 |
else:
|