|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""P and S phase arrivals dataset for Alaska""" |
|
|
|
|
|
import h5py |
|
|
import csv |
|
|
import os |
|
|
|
|
|
import datasets |
|
|
|
|
|
_PSAlaska_DESCRIPTION = """ |
|
|
|
|
|
""" |
|
|
|
|
|
_ManualPick_CITATION = """\ |
|
|
@InProceedings{huggingface:dataset, |
|
|
title = {A great new dataset}, |
|
|
author={huggingface, Inc. |
|
|
}, |
|
|
year={2020} |
|
|
} |
|
|
""" |
|
|
|
|
|
_ManualPick_DESCRIPTION = """\ |
|
|
This dataset includes P and S phases recorded by the broadband stations in the Alaska Peninsula |
|
|
""" |
|
|
|
|
|
_PNTFIter1_CITATION = """ |
|
|
""" |
|
|
|
|
|
_PNTFIter1_DESCRIPTION = """ |
|
|
This dataset includes P and S phases predicted by the PhaseNet-TF using model trained by the manualpick dataset |
|
|
""" |
|
|
|
|
|
_PNTFIter1Combined_CITATION = """ |
|
|
""" |
|
|
|
|
|
_PNTFIter1Combined_DESCRIPTION = """ |
|
|
This dataset includes all P and S phases from PNTFiter1 dataset and all false negative arrivals of manualpick dataset |
|
|
""" |
|
|
|
|
|
_Data_URL = "/mnt/scratch/jieyaqi/alaska/final/PS_Alaska" |
|
|
|
|
|
|
|
|
class PSAlaskaConfig(datasets.BuilderConfig): |
|
|
|
|
|
def __init__(self, description, data_url, citation, **kwargs): |
|
|
"""BuilderConfig for PS_Alaska. |
|
|
Args: |
|
|
features: `list[string]`, list of the features that will appear in the |
|
|
feature dict. Should not include "label". |
|
|
data_url: `string`, url to download the zip file from. |
|
|
citation: `string`, citation for the data set. |
|
|
url: `string`, url for information about the data set. |
|
|
label_classes: `list[string]`, the list of classes for the label if the |
|
|
label is present as a string. Non-string labels will be cast to either |
|
|
'False' or 'True'. |
|
|
**kwargs: keyword arguments forwarded to super. |
|
|
""" |
|
|
super(PSAlaskaConfig, self).__init__( |
|
|
version = datasets.Version("1.0.0"), |
|
|
**kwargs) |
|
|
self.description = description |
|
|
self.data_url = data_url |
|
|
self.citation = citation |
|
|
|
|
|
|
|
|
class PSAlaskaDataset(datasets.GeneratorBasedBuilder): |
|
|
"""P and S phase arrivals dataset for Alaska""" |
|
|
|
|
|
VERSION = datasets.Version("1.0.0") |
|
|
|
|
|
BUILDER_CONFIGS = [ |
|
|
PSAlaskaConfig( |
|
|
name="ManualPick", |
|
|
description=_ManualPick_DESCRIPTION, |
|
|
data_url=_Data_URL+"/ManualPick", |
|
|
citation=_ManualPick_CITATION, |
|
|
), |
|
|
PSAlaskaConfig( |
|
|
name="PNTFIter1", |
|
|
description=_PNTFIter1_DESCRIPTION, |
|
|
data_url=_Data_URL+"/PNTFIter1", |
|
|
citation=_PNTFIter1_CITATION, |
|
|
), |
|
|
PSAlaskaConfig( |
|
|
name="PNTFIter1_combined", |
|
|
description=_PNTFIter1Combined_DESCRIPTION, |
|
|
data_url=_Data_URL+"/PNTFIter1_combined", |
|
|
citation=_PNTFIter1Combined_CITATION, |
|
|
), |
|
|
] |
|
|
|
|
|
DEFAULT_CONFIG_NAME = "PNTFIter1_combined" |
|
|
|
|
|
|
|
|
def _info(self): |
|
|
|
|
|
return datasets.DatasetInfo( |
|
|
description=_PSAlaska_DESCRIPTION + self.config.description, |
|
|
features=datasets.Features( |
|
|
{ |
|
|
"begin_time": datasets.Value("string"), |
|
|
"end_time": datasets.Value("string"), |
|
|
"component": datasets.Sequence(datasets.Value('string')), |
|
|
"dt_s": datasets.Value("float"), |
|
|
"event_id": datasets.Value("string"), |
|
|
"station": datasets.Value("string"), |
|
|
"network": datasets.Value("string"), |
|
|
"phase_index": datasets.Sequence(datasets.Value('int32')), |
|
|
"phase_time": datasets.Sequence(datasets.Value('string')), |
|
|
"phase_type": datasets.Sequence(datasets.Value('string')), |
|
|
"waveform": datasets.Array2D(shape=(3, 24000), dtype='float32'), |
|
|
} |
|
|
), |
|
|
supervised_keys=("waveform", "phase_type"), |
|
|
citation=self.config.citation, |
|
|
) |
|
|
|
|
|
|
|
|
def _split_generators(self, dl_manager): |
|
|
|
|
|
urls = self.config.data_url |
|
|
data_dir = dl_manager.download_and_extract(urls) |
|
|
stationf = dl_manager.download_and_extract(_Data_URL, 'stations.csv') |
|
|
stationl = [] |
|
|
eventl = [] |
|
|
waveform_files = {} |
|
|
with open(stationf, newline='') as csvfile: |
|
|
r = csv.reader(csvfile, delimiter=',') |
|
|
next(r) |
|
|
for row in r: |
|
|
stationl.append(row[-1]) |
|
|
|
|
|
with open(os.path.join(data_dir, 'catalogs.csv'), newline='') as csvfile: |
|
|
r = csv.reader(csvfile, delimiter=',') |
|
|
next(r) |
|
|
for row in r: |
|
|
eventl.append(row[3]) |
|
|
|
|
|
for e in eventl: |
|
|
waveform_files[e] = os.path.join(data_dir, 'waveform', f'{e}.h5') |
|
|
|
|
|
return [ |
|
|
datasets.SplitGenerator( |
|
|
name="full", |
|
|
gen_kwargs={ |
|
|
"stations": stationl, |
|
|
"events": eventl, |
|
|
"waveform_files": waveform_files |
|
|
}, |
|
|
), |
|
|
] |
|
|
|
|
|
|
|
|
def _generate_examples(self, stations, events, waveform_files): |
|
|
|
|
|
for e in events: |
|
|
f = h5py.File(waveform_files[e], 'r') |
|
|
for sta in f[e].keys(): |
|
|
key = f'{e}_{sta}' |
|
|
meta = f[e][sta].attrs |
|
|
yield key, { |
|
|
"begin_time": meta['begin_time'], |
|
|
"end_time": meta['end_time'], |
|
|
"component": meta['component'], |
|
|
"dt_s": meta['dt_s'], |
|
|
"event_id": meta['event_id'], |
|
|
"station": meta['station'], |
|
|
"network": meta['network'], |
|
|
"phase_index": meta['phase_index'], |
|
|
"phase_time": meta['phase_time'], |
|
|
"phase_type": meta['phase_type'], |
|
|
"waveform": f[e][sta] |
|
|
} |
|
|
f.close() |
|
|
|