Update hit.py
Browse files
hit.py
CHANGED
|
@@ -111,6 +111,7 @@ class NewDataset(datasets.GeneratorBasedBuilder):
|
|
| 111 |
# It can accept any type or nested list/dict and will give back the same structure with the url replaced with path to local files.
|
| 112 |
# By default the archives will be extracted and a path to a cached folder where they are extracted is returned instead of the archive
|
| 113 |
rel_path = _PATHS[self.config.name]
|
|
|
|
| 114 |
return [
|
| 115 |
datasets.SplitGenerator(
|
| 116 |
name=datasets.Split.TRAIN,
|
|
@@ -144,6 +145,7 @@ class NewDataset(datasets.GeneratorBasedBuilder):
|
|
| 144 |
# The `key` is for legacy reasons (tfds) and is not important in itself, but must be unique for each example.
|
| 145 |
# List all files in the path .gz
|
| 146 |
files = glob(filepath)
|
|
|
|
| 147 |
for subject_path in files:
|
| 148 |
with gzip.open(subject_path, 'rb') as f:
|
| 149 |
data = pickle.load(f)
|
|
|
|
| 111 |
# It can accept any type or nested list/dict and will give back the same structure with the url replaced with path to local files.
|
| 112 |
# By default the archives will be extracted and a path to a cached folder where they are extracted is returned instead of the archive
|
| 113 |
rel_path = _PATHS[self.config.name]
|
| 114 |
+
print(os.path.join(rel_path, "train", "*.gz"))
|
| 115 |
return [
|
| 116 |
datasets.SplitGenerator(
|
| 117 |
name=datasets.Split.TRAIN,
|
|
|
|
| 145 |
# The `key` is for legacy reasons (tfds) and is not important in itself, but must be unique for each example.
|
| 146 |
# List all files in the path .gz
|
| 147 |
files = glob(filepath)
|
| 148 |
+
print(files)
|
| 149 |
for subject_path in files:
|
| 150 |
with gzip.open(subject_path, 'rb') as f:
|
| 151 |
data = pickle.load(f)
|