The viewer is disabled because this dataset repo requires arbitrary Python code execution. Please consider
removing the
loading script
and relying on
automated data support
(you can use
convert_to_parquet
from the datasets library). If this is not possible, please
open a discussion
for direct help.
Super-resolution of Velocity Fields in Three-dimensional Fluid Dynamics
This dataset loader attempts to reproduce the data of Wang et al. (2024)'s experiments on Super-resolution of 3D Turbulence.
References:
- Wang et al. (2024): "Discovering Symmetry Breaking in Physical Systems with Relaxed Group Convolution"
Usage
For a given configuration (e.g. large_50):
>>> ds = datasets.load_dataset("dl2-g32/jhtdb", name="large_50")
>>> ds
DatasetDict({
train: Dataset({
features: ['lrs', 'hr'],
num_rows: 40
})
validation: Dataset({
features: ['lrs', 'hr'],
num_rows: 5
})
test: Dataset({
features: ['lrs', 'hr'],
num_rows: 5
})
})
Each split contains the input lrs which corresponds on a sequence of low resolution samples from time t - ws/2, ..., t, ... ts + ws/2 (ws = window size) and hr corresponds to the high resolution sample at time t. All the parameters per data point are specified in the corresponding metadata_*.csv.
Specifically, for the default configuration, for each datapoint we have 3 low resolution samples and 1 high resolution sample. Each of the former have shapes (3, 16, 16, 16) and the latter has shape (3, 64, 64, 64).
Replication
This dataset is entirely generated by scripts/generate.py and each configuration is fully specified in their corresponding scripts/*.yaml.
Usage
python -m scripts.generate --config scripts/small_100.yaml --token edu.jhu.pha.turbulence.testing-201311
This will create two folders on datasets/jhtdb:
- A
tmpfolder that will store all samples accross runs to serve as a cache. - The corresponding subset,
small_50for example. This folder will contain ametadata_*.csvand data*.zipfor each split.
Note:
- For the small variants, the default token is enough, but for the large variants a token has to be requested. More details here.
- For reference, the
large_100takes ~15 minutes to generate for a total of ~300MB.
- Downloads last month
- 115