|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""QuakeFlow_NC: A dataset of earthquake waveforms organized by earthquake events and based on the HDF5 format.""" |
|
|
|
|
|
import h5py |
|
import numpy as np |
|
import torch |
|
from typing import Dict, List, Optional, Tuple, Union |
|
import fsspec |
|
|
|
import datasets |
|
|
|
|
|
|
|
|
|
_CITATION = """\ |
|
@InProceedings{huggingface:dataset, |
|
title = {NCEDC dataset for QuakeFlow}, |
|
author={Zhu et al.}, |
|
year={2023} |
|
} |
|
""" |
|
|
|
|
|
|
|
_DESCRIPTION = """\ |
|
A dataset of earthquake waveforms organized by earthquake events and based on the HDF5 format. |
|
""" |
|
|
|
|
|
_HOMEPAGE = "" |
|
|
|
|
|
_LICENSE = "" |
|
|
|
|
|
|
|
|
|
_REPO = "https://huggingface.co/datasets/AI4EPS/quakeflow_nc/resolve/main/data" |
|
_FILES = [ |
|
"NC1970-1989.h5", |
|
"NC1990-1994.h5", |
|
"NC1995-1999.h5", |
|
"NC2000-2004.h5", |
|
"NC2005-2009.h5", |
|
"NC2010.h5", |
|
"NC2011.h5", |
|
"NC2012.h5", |
|
"NC2013.h5", |
|
"NC2014.h5", |
|
"NC2015.h5", |
|
"NC2016.h5", |
|
"NC2017.h5", |
|
"NC2018.h5", |
|
"NC2019.h5", |
|
"NC2020.h5", |
|
] |
|
_URLS = { |
|
"station": [f"{_REPO}/{x}" for x in _FILES], |
|
"event": [f"{_REPO}/{x}" for x in _FILES], |
|
"station_train": [f"{_REPO}/{x}" for x in _FILES[:-1]], |
|
"event_train": [f"{_REPO}/{x}" for x in _FILES[:-1]], |
|
"station_test": [f"{_REPO}/{x}" for x in _FILES[-1:]], |
|
"event_test": [f"{_REPO}/{x}" for x in _FILES[-1:]], |
|
} |
|
|
|
|
|
class BatchBuilderConfig(datasets.BuilderConfig): |
|
""" |
|
yield a batch of event-based sample, so the number of sample stations can vary among batches |
|
Batch Config for QuakeFlow_NC |
|
:param batch_size: number of samples in a batch |
|
:param num_stations_list: possible number of stations in a batch |
|
""" |
|
|
|
def __init__(self, batch_size: int, num_stations_list: List, **kwargs): |
|
super().__init__(**kwargs) |
|
self.batch_size = batch_size |
|
self.num_stations_list = num_stations_list |
|
|
|
|
|
|
|
class QuakeFlow_NC(datasets.GeneratorBasedBuilder): |
|
"""QuakeFlow_NC: A dataset of earthquake waveforms organized by earthquake events and based on the HDF5 format.""" |
|
|
|
VERSION = datasets.Version("1.1.0") |
|
|
|
degree2km = 111.32 |
|
nt = 8192 |
|
feature_nt = 512 |
|
feature_scale = int(nt / feature_nt) |
|
sampling_rate = 100.0 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
BUILDER_CONFIGS = [ |
|
datasets.BuilderConfig( |
|
name="station", version=VERSION, description="yield station-based samples one by one of whole dataset" |
|
), |
|
datasets.BuilderConfig( |
|
name="event", version=VERSION, description="yield event-based samples one by one of whole dataset" |
|
), |
|
datasets.BuilderConfig( |
|
name="station_train", |
|
version=VERSION, |
|
description="yield station-based samples one by one of training dataset", |
|
), |
|
datasets.BuilderConfig( |
|
name="event_train", version=VERSION, description="yield event-based samples one by one of training dataset" |
|
), |
|
datasets.BuilderConfig( |
|
name="station_test", version=VERSION, description="yield station-based samples one by one of test dataset" |
|
), |
|
datasets.BuilderConfig( |
|
name="event_test", version=VERSION, description="yield event-based samples one by one of test dataset" |
|
), |
|
] |
|
|
|
DEFAULT_CONFIG_NAME = ( |
|
"station_test" |
|
) |
|
|
|
def _info(self): |
|
|
|
if ( |
|
(self.config.name == "station") |
|
or (self.config.name == "station_train") |
|
or (self.config.name == "station_test") |
|
): |
|
features = datasets.Features( |
|
{ |
|
"waveform": datasets.Array2D(shape=(3, self.nt), dtype="float32"), |
|
"phase_pick": datasets.Array2D(shape=(3, self.nt), dtype="float32"), |
|
"event_location": datasets.Sequence(datasets.Value("float32")), |
|
"station_location": datasets.Sequence(datasets.Value("float32")), |
|
} |
|
) |
|
|
|
elif (self.config.name == "event") or (self.config.name == "event_train") or (self.config.name == "event_test"): |
|
features = datasets.Features( |
|
{ |
|
"waveform": datasets.Array3D(shape=(None, 3, self.nt), dtype="float32"), |
|
"phase_pick": datasets.Array3D(shape=(None, 3, self.nt), dtype="float32"), |
|
"event_location": datasets.Sequence(datasets.Value("float32")), |
|
"station_location": datasets.Array2D(shape=(None, 3), dtype="float32"), |
|
} |
|
) |
|
|
|
return datasets.DatasetInfo( |
|
|
|
description=_DESCRIPTION, |
|
|
|
features=features, |
|
|
|
|
|
|
|
|
|
homepage=_HOMEPAGE, |
|
|
|
license=_LICENSE, |
|
|
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
|
|
|
|
|
|
|
|
|
|
|
|
urls = _URLS[self.config.name] |
|
|
|
files = dl_manager.download_and_extract(urls) |
|
print(files) |
|
|
|
if self.config.name == "station" or self.config.name == "event": |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
|
|
gen_kwargs={ |
|
"filepath": files[:-1], |
|
"split": "train", |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
gen_kwargs={"filepath": files[-1:], "split": "test"}, |
|
), |
|
] |
|
elif self.config.name == "station_train" or self.config.name == "event_train": |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs={ |
|
"filepath": files, |
|
"split": "train", |
|
}, |
|
), |
|
] |
|
elif self.config.name == "station_test" or self.config.name == "event_test": |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
gen_kwargs={"filepath": files, "split": "test"}, |
|
), |
|
] |
|
else: |
|
raise ValueError("config.name is not in BUILDER_CONFIGS") |
|
|
|
|
|
def _generate_examples(self, filepath, split): |
|
|
|
|
|
|
|
for file in filepath: |
|
with fsspec.open(file, "rb") as fs: |
|
with h5py.File(fs, "r") as fp: |
|
|
|
event_ids = list(fp.keys()) |
|
for event_id in event_ids: |
|
event = fp[event_id] |
|
station_ids = list(event.keys()) |
|
if ( |
|
(self.config.name == "station") |
|
or (self.config.name == "station_train") |
|
or (self.config.name == "station_test") |
|
): |
|
waveforms = np.zeros([3, self.nt], dtype="float32") |
|
phase_pick = np.zeros_like(waveforms) |
|
attrs = event.attrs |
|
event_location = [ |
|
attrs["longitude"], |
|
attrs["latitude"], |
|
attrs["depth_km"], |
|
attrs["event_time_index"], |
|
] |
|
|
|
for i, sta_id in enumerate(station_ids): |
|
waveforms[:, : self.nt] = event[sta_id][:, : self.nt] |
|
|
|
attrs = event[sta_id].attrs |
|
p_picks = attrs["phase_index"][attrs["phase_type"] == "P"] |
|
s_picks = attrs["phase_index"][attrs["phase_type"] == "S"] |
|
|
|
station_location = [attrs["longitude"], attrs["latitude"], -attrs["elevation_m"] / 1e3] |
|
|
|
yield f"{event_id}/{sta_id}", { |
|
"waveform": torch.from_numpy(waveforms).float(), |
|
"phase_pick": torch.from_numpy(phase_pick).float(), |
|
"event_location": torch.from_numpy(np.array(event_location)).float(), |
|
"station_location": torch.from_numpy(np.array(station_location)).float(), |
|
} |
|
|
|
elif ( |
|
(self.config.name == "event") |
|
or (self.config.name == "event_train") |
|
or (self.config.name == "event_test") |
|
): |
|
waveforms = np.zeros([len(station_ids), 3, self.nt], dtype="float32") |
|
phase_pick = np.zeros_like(waveforms) |
|
attrs = event.attrs |
|
event_location = [ |
|
attrs["longitude"], |
|
attrs["latitude"], |
|
attrs["depth_km"], |
|
attrs["event_time_index"], |
|
] |
|
station_location = [] |
|
|
|
for i, sta_id in enumerate(station_ids): |
|
waveforms[i, :, : self.nt] = event[sta_id][:, : self.nt] |
|
|
|
attrs = event[sta_id].attrs |
|
p_picks = attrs["phase_index"][attrs["phase_type"] == "P"] |
|
s_picks = attrs["phase_index"][attrs["phase_type"] == "S"] |
|
phase_pick[i, :, :] = generate_label([p_picks, s_picks], nt=self.nt) |
|
station_location.append( |
|
[attrs["longitude"], attrs["latitude"], -attrs["elevation_m"] / 1e3] |
|
) |
|
|
|
yield event_id, { |
|
"waveform": torch.from_numpy(waveforms).float(), |
|
"phase_pick": torch.from_numpy(phase_pick).float(), |
|
"event_location": torch.from_numpy(np.array(event_location)).float(), |
|
"station_location": torch.from_numpy(np.array(station_location)).float(), |
|
} |
|
|
|
|
|
def generate_label(phase_list, label_width=[150, 150], nt=8192): |
|
target = np.zeros([len(phase_list) + 1, nt], dtype=np.float32) |
|
|
|
for i, (picks, w) in enumerate(zip(phase_list, label_width)): |
|
for phase_time in picks: |
|
t = np.arange(nt) - phase_time |
|
gaussian = np.exp(-(t**2) / (2 * (w / 6) ** 2)) |
|
gaussian[gaussian < 0.1] = 0.0 |
|
target[i + 1, :] += gaussian |
|
|
|
target[0:1, :] = np.maximum(0, 1 - np.sum(target[1:, :], axis=0, keepdims=True)) |
|
|
|
return target |
|
|