The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
Error code: DatasetGenerationError
Exception: ChunkedEncodingError
Message: ('Connection broken: IncompleteRead(0 bytes read, 5242888 more expected)', IncompleteRead(0 bytes read, 5242888 more expected))
Traceback: Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 779, in _error_catcher
yield
File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 925, in _raw_read
raise IncompleteRead(self._fp_bytes_read, self.length_remaining)
urllib3.exceptions.IncompleteRead: IncompleteRead(0 bytes read, 5242888 more expected)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 820, in generate
yield from self.raw.stream(chunk_size, decode_content=True)
File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 1091, in stream
data = self.read(amt=amt, decode_content=decode_content)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 980, in read
data = self._raw_read(amt)
^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 903, in _raw_read
with self._error_catcher():
^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/contextlib.py", line 158, in __exit__
self.gen.throw(value)
File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 803, in _error_catcher
raise ProtocolError(arg, e) from e
urllib3.exceptions.ProtocolError: ('Connection broken: IncompleteRead(0 bytes read, 5242888 more expected)', IncompleteRead(0 bytes read, 5242888 more expected))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1779, in _prepare_split_single
for key, table in generator:
^^^^^^^^^
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 609, in wrapped
for item in generator(*args, **kwargs):
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/hdf5/hdf5.py", line 76, in _generate_tables
with h5py.File(f, "r") as h5:
^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/h5py/_hl/files.py", line 564, in __init__
fid = make_fid(name, mode, userblock_size, fapl, fcpl, swmr=swmr)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/h5py/_hl/files.py", line 238, in make_fid
fid = h5f.open(name, flags, fapl=fapl)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "h5py/_objects.pyx", line 56, in h5py._objects.with_phil.wrapper
File "h5py/_objects.pyx", line 57, in h5py._objects.with_phil.wrapper
File "h5py/h5f.pyx", line 102, in h5py.h5f.open
File "h5py/h5fd.pyx", line 162, in h5py.h5fd.H5FD_fileobj_read
File "/usr/local/lib/python3.12/site-packages/fsspec/spec.py", line 1856, in readinto
data = self.read(out.nbytes)
^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/utils/file_utils.py", line 844, in read_with_retries
out = read(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 728, in track_read
out = f_read(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 1015, in read
return super().read(length)
^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/fsspec/spec.py", line 1846, in read
out = self.cache._fetch(self.loc, self.loc + length)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/fsspec/caching.py", line 189, in _fetch
self.cache = self.fetcher(start, end) # new block replaces old
^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 969, in _fetch_range
r = http_backoff(
^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 310, in http_backoff
response = session.request(method=method, url=url, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/requests/sessions.py", line 589, in request
resp = self.send(prep, **send_kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/requests/sessions.py", line 724, in send
history = [resp for resp in gen]
^^^
File "/usr/local/lib/python3.12/site-packages/requests/sessions.py", line 265, in resolve_redirects
resp = self.send(
^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/requests/sessions.py", line 746, in send
r.content
File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 902, in content
self._content = b"".join(self.iter_content(CONTENT_CHUNK_SIZE)) or b""
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 822, in generate
raise ChunkedEncodingError(e)
requests.exceptions.ChunkedEncodingError: ('Connection broken: IncompleteRead(0 bytes read, 5242888 more expected)', IncompleteRead(0 bytes read, 5242888 more expected))
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1342, in compute_config_parquet_and_info_response
parquet_operations, partial, estimated_dataset_info = stream_convert_to_parquet(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 907, in stream_convert_to_parquet
builder._prepare_split(split_generator=splits_generators[split], file_format="parquet")
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1646, in _prepare_split
for job_id, done, content in self._prepare_split_single(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1832, in _prepare_split_single
raise DatasetGenerationError("An error occurred while generating the dataset") from e
datasets.exceptions.DatasetGenerationError: An error occurred while generating the datasetNeed help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
a_1 float32 | a_2 float32 | chirp_mass float32 | dec float32 | distance float32 | inclination float32 | mass_1 float32 | mass_2 float32 | mass_ratio float32 | phi float32 | phi_12 float32 | phi_jl float32 | phic float32 | psi float32 | s1x float32 | s1y float32 | s1z float32 | s2x float32 | s2y float32 | s2z float32 | snr float32 | tilt_1 float32 | tilt_2 float32 | whitened_bkg array 2D | whitened_injected array 2D | whitened_signal array 2D |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0.132638 | 0.212598 | 1.476183 | 0.440216 | 694.887573 | 2.280191 | 2.335073 | 1.254969 | 0.537443 | -1.775144 | 5.953039 | 0.403361 | 4.037373 | 2.853332 | 0.120349 | 0.05136 | 0.021703 | 0.147711 | 0.133125 | -0.075212 | 11.220628 | 1.406429 | 1.932402 | [[0.5108880538651458,-0.47651083522487014,-0.31207012392851186,-0.04660678774945981,0.19580924185522(...TRUNCATED) | [[0.5231142049577889,-0.4740286709518906,-0.3099992869563174,-0.045210956977267354,0.196248839565143(...TRUNCATED) | [[0.012111886547223685,0.0024845281082506096,0.0020715929507103457,0.0013962815292797815,0.000438950(...TRUNCATED) |
0.014281 | 0.383475 | 1.886027 | -0.239368 | 814.183289 | 1.554106 | 2.210183 | 2.123803 | 0.960917 | -1.025218 | 0.665693 | 6.281428 | 6.028352 | 2.547857 | 0.012859 | -0.000023 | -0.006213 | 0.224465 | -0.176904 | -0.255682 | 12.462122 | 2.020871 | 2.300636 | [[-0.42334735280668756,0.07345079832356109,-0.1585301175102272,0.018150542151466727,-0.2291691968021(...TRUNCATED) | [[-0.402928563780769,0.07647090138222057,-0.1578304337230037,0.018050963765232495,-0.229500035139664(...TRUNCATED) | [[0.020498005534614255,0.003017056289381034,0.0007001588870790571,-0.00009800748984697289,-0.0003325(...TRUNCATED) |
0.256894 | 0.04948 | 1.61549 | -0.106339 | 970.483215 | 1.311422 | 2.260245 | 1.534959 | 0.679112 | 0.212646 | 1.864316 | 2.430209 | 1.505919 | 2.200867 | -0.146133 | 0.125957 | 0.169631 | 0.021748 | 0.013814 | 0.042243 | 12.113312 | 0.849559 | 0.547675 | [[0.5665224202274873,0.03597173186093768,0.09746331650601193,0.14616306867471182,-0.253655749550361,(...TRUNCATED) | [[0.5659381692920017,0.03490227328154931,0.09541330159681029,0.15057429888656876,-0.2497453246319728(...TRUNCATED) | [[-0.0007099313695994291,-0.0010663080359196147,-0.002048561818766557,0.004410047200299345,0.0039098(...TRUNCATED) |
0.03257 | 0.151314 | 1.332667 | 0.258392 | 907.442749 | 2.235384 | 1.634382 | 1.435055 | 0.878042 | 0.342371 | 1.385909 | 3.813955 | 1.668176 | 1.315749 | -0.019596 | -0.015601 | 0.020818 | -0.084723 | 0.073346 | 0.101677 | 12.69169 | 0.877369 | 0.833945 | [[-1.251544117638249,0.10098490950102347,0.030766283227287035,-0.11846407026248132,0.432213509447886(...TRUNCATED) | [[-1.2532487108606971,0.10259120560647035,0.033343054042583614,-0.11991083631334422,0.42909295430293(...TRUNCATED) | [[-0.0016575319401292694,0.0016044133193437095,0.0025774177534775796,-0.0014466628085057186,-0.00312(...TRUNCATED) |
0.373605 | 0.163436 | 1.689224 | -0.258468 | 934.636719 | 1.812189 | 2.46042 | 1.54673 | 0.628645 | 0.51651 | 1.411628 | 4.844299 | 2.687889 | 2.132041 | 0.027368 | -0.206266 | 0.3103 | -0.121324 | -0.036347 | 0.103298 | 11.275496 | 0.59069 | 0.886611 | [[-0.9451473918793599,-0.010154334666425851,-0.03130436460620256,-0.05956653843705813,0.026939975579(...TRUNCATED) | [[-0.9448960404649935,-0.010713043172117494,-0.031981808593155844,-0.06068782456977415,0.02950181275(...TRUNCATED) | [[0.00004458375747419148,-0.0005531690069304632,-0.0006741621028094778,-0.001122362771532646,0.00255(...TRUNCATED) |
0.228106 | 0.340577 | 1.310226 | 0.102883 | 606.432068 | 0.873127 | 1.611998 | 1.406513 | 0.872528 | -0.554062 | 5.477046 | 5.476353 | 3.30203 | 0.873803 | 0.145616 | -0.151996 | -0.087896 | 0.260237 | -0.000181 | -0.219703 | 11.545363 | 1.966361 | 2.271937 | [[0.928858090799905,-0.19433427762304312,0.09615794545177847,0.27092492089295084,0.24371138895977706(...TRUNCATED) | [[0.9228856400827439,-0.19592318677778356,0.09486439359339208,0.27036071455179966,0.2440101661149028(...TRUNCATED) | [[-0.00607008913943698,-0.001590661728062555,-0.0012962453942596103,-0.00056233180790942,0.000303117(...TRUNCATED) |
0.16326 | 0.189812 | 1.562275 | 0.963337 | 893.692139 | 2.159919 | 1.801761 | 1.787439 | 0.992051 | -1.248401 | 1.911827 | 0.327613 | 0.960607 | 3.069793 | 0.12302 | 0.04181 | -0.098853 | -0.000506 | -0.037716 | -0.186026 | 14.45129 | 2.221182 | 2.941541 | [[-1.6371117911760338,0.2533690397849764,-0.1926435936009726,-0.08498662225352203,-0.174275751059941(...TRUNCATED) | [[-1.655611405228846,0.24995592821410162,-0.19510987374321356,-0.08654308637805225,-0.17475460010407(...TRUNCATED) | [[-0.018530100613245767,-0.0034130609585147492,-0.002465104604546704,-0.0015563478291564302,-0.00047(...TRUNCATED) |
0.176649 | 0.087061 | 1.434963 | 1.200544 | 899.533569 | 0.954407 | 2.159362 | 1.275607 | 0.590733 | 0.225789 | 4.809042 | 2.57957 | 3.348222 | 1.60893 | -0.118046 | 0.074342 | 0.108366 | -0.053179 | -0.068708 | -0.005553 | 11.511329 | 0.910368 | 1.634617 | [[-0.05906546787115546,-0.1578955978829909,0.19486455101166614,0.04735822612665856,-0.11164285808723(...TRUNCATED) | [[-0.0373178887380724,-0.1532120197936032,0.19696549645956918,0.047642125228227195,-0.11280314214476(...TRUNCATED) | [[0.021705662912164744,0.0046848606497519926,0.0021009312793212626,0.0002833554955397278,-0.00116034(...TRUNCATED) |
0.350441 | 0.31102 | 1.601859 | 0.196206 | 649.574341 | 2.266205 | 2.028328 | 1.67236 | 0.824502 | -1.539207 | 2.586654 | 1.396045 | 3.119108 | 2.918881 | 0.034879 | 0.197556 | 0.28734 | 0.111531 | -0.279086 | -0.08003 | 14.235178 | 0.609493 | 1.831039 | [[-1.4370049241295024,-0.32183395970277046,0.07884177025867768,-0.023059451606906693,-0.135956798767(...TRUNCATED) | [[-1.4395669576032024,-0.32193690615072307,0.08033539833375343,-0.021341303101032678,-0.134572542451(...TRUNCATED) | [[-0.0025563014233071393,-0.00010185691891051266,0.0014909226205316456,0.001721398413998949,0.001383(...TRUNCATED) |
0.050068 | 0.335339 | 1.478567 | 1.110939 | 928.172729 | 2.388863 | 2.008337 | 1.444137 | 0.719071 | 1.630054 | 2.751683 | 3.808259 | 4.741426 | 2.198338 | -0.014778 | -0.011628 | 0.046403 | 0.13659 | 0.24179 | -0.187971 | 10.266898 | 0.385007 | 2.165835 | [[-1.0085911678869346,-0.3139898766747979,-0.0867876532708081,0.012932721741354776,-0.05387678981226(...TRUNCATED) | [[-1.0137946608575312,-0.31384719124066557,-0.0847058162793768,0.015096833868175775,-0.0524856914764(...TRUNCATED) | [[-0.005185916565675295,0.00013910455745910118,0.002083269725772033,0.002167080430832048,0.001391153(...TRUNCATED) |
PhyTS Dataset
Abstract
We introduce PhyTS, a benchmark suite of precision scientific time series datasets for machine learning, spanning experiments in gravitational-wave detection, dark matter searches, neutrino mass determination, and stellar variability detection. Despite their diverse scientific goals, these domains share a common challenge: recovering weak, structured signals and estimating underlying physical parameters from noise-dominated measurements. Unlike standard sequence modeling benchmarks such as audio and speech, these data exhibit non-Gaussian and nonstationary noise, long-range temporal correlations, detector-specific systematics, irregular sampling, and signals that are sparse, weak, or only partially modeled. As a result, they provide a challenging testbed for evaluating whether modern AI methods can support downstream scientific inference. We provide standardized tasks, data splits, and evaluation protocols for denoising, signal recovery, and parameter inference across physics domains, along with baseline results. By unifying diverse weak-signal inference problems under a common framework, this benchmark aims to enable reproducible evaluation and accelerate the development of more robust, interpretable, and physically grounded methods for scientific time series analysis.
📌 Github repo: kyoon-mit/TimeSeriesPhysics
LIGO
LIGO Dataset Field Description
TESS
The TESS dataset is broken into training, validation and test data, which are identically constructed. There is a set for classification '/tess/split/tess_classification_{train,val,test}.parquet' and one for regression '/tess/split/tess_regression_{train,val,test}.parquet. The dataset was split 80/10/10 into training, validation, and test sets. Since multiple TESS light curves, from different TESS sectors, are available per unique Gaia Data Release 3 (DR3) identifier (ID), the split ensures that no Gaia DR3 ID appears in more than one split to avoid label leakage, with the exception of when different light curves for the same Gaia DR3 ID have different labels due to variations in dominant variability detected in separate TESS sectors. The data includes the following
TESS Dataset Field Description
| Parameter | Type / Size | Description |
|---|---|---|
GaiaID |
dtype=int | Unique Gaia DR3 Identifier |
TIC |
dtype=int | TESS Input Catalog Identifier |
sector |
dtype=int | TESS Observing Sector (https://tess.mit.edu/observations/) |
label |
dtype=string | Class label [FOR THE CLASSIFICATION TASK] |
frot |
dtype=double | Near-core rotation rate of the star [FOR THE REGRESSION TASK] |
frot_err |
dtype=double | Near-core rotation rate error [FOR THE REGRESSION TASK] |
time |
dtype=list | Time stamps in BTJD (Barycentric TESS Julian Date. Corrected for light travel time to solar system barycenter) |
flux |
dtype=list | Time series measurements in normalized flux |
TIDMAD
The TIDMAD dataset is broken into training and validation data which are identically constructed differing only by the time at which the detector data was taken. This temporal difference results in variable detector noise conditions. There are 20 hdf5 files (~ 4 GB/file) in both training and validation datasets. The total dataset size is 163 GB. Each file contains ~ 200 examples and the associated meta data. Each example is a set of ch1 time series, ch2 time series, and injected frequency. This data in each hdf5 file includes the following
TIDMAD Dataset Field Description
| Parameter | Type / Size | Description |
|---|---|---|
chunk_length |
dtype=int | Number of samples per example |
n_chunks |
dtype=int | Number of examples per file |
sample_rate_hz |
dtype=int | Data taking sample rate in Hz |
signal_freq_choices |
size=(618,1), dtype=int | Array of signal frequencies injected across the full dataset, all files |
signal_frequency |
size=(200,1), dtype=int | Injected frequency for the example in Hz (signal_frequency[i] is the injected signal frequency for example i) |
time_series_ch1 |
size=(200,10M), dtype=int | Noisy time series for the example in DAQ units. Conversion factor from DAQ units to mV is 40mV/128. (time_series_ch1[i] is the ch1 time series for example i) |
time_series_ch2 |
size=(200,10M), dtype=int | Signal time series for the example in DAQ units. Conversion factor from DAQ units to mV is 40mV/128. (time_series_ch2[i] is the ch2 time series for example i) |
Project 8
The Project 8 dataset is broken into training, validation, and test datasets which are identically constructed.
All contain a set of electron events with randomized energy, pitch angle, and radius parameters.
The data are divided into hdf5 files with ~5000 electrons per file; each file is about 5 GB, with the exception of one, which has fewer electrons and is smaller.
There are about 50,000 electrons total in the dataset. The total dataset size is about 50 GB.
The train dataset contains 8 hdf5 files, the validation dataset has 1 hdf5 file, and the test dataset has 2 (including the smaller file).
The experiments presented in the publication use output_ts_I + output_ts_I_cav_noise and output_ts_Q + output_ts_Q_cav_noise as inputs (noisy in-phase and quadrature time series).
The target for regression is energy_eV (truth value for energy of the electron, in eV).
This data in each hdf5 file includes the following
Project 8 Dataset Field Description
| Parameter | Type / Size | Description |
|---|---|---|
output_ts_I |
size=(24576,1), dtype=float32 | In-phase signal time series |
output_ts_Q |
size=(24576,1), dtype=float32 | Quadrature signal time series |
output_ts_I_cav_noise |
size=(24576,1), dtype=float32 | In-phase cavity (frequency-dependent) noise time series |
output_ts_Q_cav_noise |
size=(24576,1), dtype=float32 | Quadrature cavity (frequency-dependent) noise time series |
output_ts_I_gauss_noise |
size=(24576,1), dtype=float32 | In-phase Gaussian noise time series |
output_ts_Q_gauss_noise |
size=(24576,1), dtype=float32 | Quadrature Gaussian noise time series |
energy_eV |
dtype=float32 | Truth value for energy of the electron (target for regression), in eV |
avg_carrier_frequency_Hz |
dtype=float32 | Truth value for average cyclotron frequency of the electron, in Hz |
start_carrier_frequency_Hz |
dtype=float32 | Truth value for starting cyclotron frequency of the electron, in Hz |
avg_axial_frequency_Hz |
dtype=float32 | Truth value for average axial frequency of the electron, in Hz |
pitch_angle_deg |
dtype=float32 | Truth value for pitch angle (direction of the electron’s momentum with respect to the magnetic field) of the electron, in degrees |
radius_input_m |
dtype=float32 | Truth value for starting radius of the electron before one trajectory step (input value), in meters |
radius_m |
dtype=float32 | Truth value for starting radius of the electron after one trajectory step, in meters |
radius_phase |
dtype=float32 | Truth value for instantaneous polar angle in the X-Y plane of the electron’s starting position, in meters |
slope_Hz |
dtype=float32 | Truth value for slope of the track in spectrogram space, in Hz/s |
- Downloads last month
- 243