Updated loading logic
Browse files- AstroM3Dataset.py +4 -12
AstroM3Dataset.py
CHANGED
|
@@ -5,10 +5,8 @@ import pandas as pd
|
|
| 5 |
import numpy as np
|
| 6 |
import json
|
| 7 |
from astropy.io import fits
|
| 8 |
-
from tqdm import tqdm
|
| 9 |
|
| 10 |
from utils.parallelzipfile import ParallelZipFile as ZipFile
|
| 11 |
-
from datasets.utils.tqdm import enable_progress_bars, disable_progress_bars
|
| 12 |
|
| 13 |
_DESCRIPTION = (
|
| 14 |
"AstroM3 is a time-series astronomy dataset containing photometry, spectra, "
|
|
@@ -39,8 +37,6 @@ _CITATION = """
|
|
| 39 |
class AstroM3Dataset(datasets.GeneratorBasedBuilder):
|
| 40 |
"""Hugging Face dataset for AstroM3 with configurable subsets and seeds."""
|
| 41 |
|
| 42 |
-
# HF_DATASETS_DISABLE_PROGRESS_BARS = True
|
| 43 |
-
|
| 44 |
DEFAULT_CONFIG_NAME = "full_42"
|
| 45 |
BUILDER_CONFIGS = [
|
| 46 |
datasets.BuilderConfig(name=f"{sub}_{seed}", version=_VERSION, data_dir=None)
|
|
@@ -118,25 +114,21 @@ class AstroM3Dataset(datasets.GeneratorBasedBuilder):
|
|
| 118 |
"info": f"{_URL}/splits/{sub}/{seed}/info.json",
|
| 119 |
}
|
| 120 |
extracted_path = dl_manager.download(urls)
|
| 121 |
-
# print("Downloaded train.csv val.csv test.csv info.json")
|
| 122 |
|
| 123 |
-
# Load all spectra files
|
| 124 |
-
spectra = {}
|
| 125 |
df1 = pd.read_csv(extracted_path["train"])
|
| 126 |
df2 = pd.read_csv(extracted_path["val"])
|
| 127 |
df3 = pd.read_csv(extracted_path["test"])
|
| 128 |
df_combined = pd.concat([df1, df2, df3], ignore_index=True)
|
| 129 |
|
|
|
|
|
|
|
| 130 |
for _, row in df_combined.iterrows():
|
| 131 |
-
|
| 132 |
-
|
| 133 |
-
|
| 134 |
-
# print("Downloaded spectra files")
|
| 135 |
|
| 136 |
# Load photometry and init reader
|
| 137 |
photometry_path = dl_manager.download(f"{_URL}/photometry.zip")
|
| 138 |
self.reader_v = ZipFile(photometry_path)
|
| 139 |
-
# print("Downloaded photometry")
|
| 140 |
|
| 141 |
return [
|
| 142 |
datasets.SplitGenerator(
|
|
|
|
| 5 |
import numpy as np
|
| 6 |
import json
|
| 7 |
from astropy.io import fits
|
|
|
|
| 8 |
|
| 9 |
from utils.parallelzipfile import ParallelZipFile as ZipFile
|
|
|
|
| 10 |
|
| 11 |
_DESCRIPTION = (
|
| 12 |
"AstroM3 is a time-series astronomy dataset containing photometry, spectra, "
|
|
|
|
| 37 |
class AstroM3Dataset(datasets.GeneratorBasedBuilder):
|
| 38 |
"""Hugging Face dataset for AstroM3 with configurable subsets and seeds."""
|
| 39 |
|
|
|
|
|
|
|
| 40 |
DEFAULT_CONFIG_NAME = "full_42"
|
| 41 |
BUILDER_CONFIGS = [
|
| 42 |
datasets.BuilderConfig(name=f"{sub}_{seed}", version=_VERSION, data_dir=None)
|
|
|
|
| 114 |
"info": f"{_URL}/splits/{sub}/{seed}/info.json",
|
| 115 |
}
|
| 116 |
extracted_path = dl_manager.download(urls)
|
|
|
|
| 117 |
|
|
|
|
|
|
|
| 118 |
df1 = pd.read_csv(extracted_path["train"])
|
| 119 |
df2 = pd.read_csv(extracted_path["val"])
|
| 120 |
df3 = pd.read_csv(extracted_path["test"])
|
| 121 |
df_combined = pd.concat([df1, df2, df3], ignore_index=True)
|
| 122 |
|
| 123 |
+
# Load all spectra files
|
| 124 |
+
spectra_urls = {}
|
| 125 |
for _, row in df_combined.iterrows():
|
| 126 |
+
spectra_urls[row["spec_filename"]] = f"{_URL}/spectra/{row['target']}/{row['spec_filename']}"
|
| 127 |
+
spectra = dl_manager.download(spectra_urls)
|
|
|
|
|
|
|
| 128 |
|
| 129 |
# Load photometry and init reader
|
| 130 |
photometry_path = dl_manager.download(f"{_URL}/photometry.zip")
|
| 131 |
self.reader_v = ZipFile(photometry_path)
|
|
|
|
| 132 |
|
| 133 |
return [
|
| 134 |
datasets.SplitGenerator(
|