Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
kappa
array 2D
theta
list
[[0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,-0.000217914581298(...TRUNCATED)
[ 0.30000001192092896, 0.800000011920929, 7.58281946182251, 0.0037049006205052137, -0.006268441677093506 ]
[[0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0072593688964843(...TRUNCATED)
[0.19499999284744263,0.9552162885665894,7.58281946182251,0.0037049006205052137,-0.006268441677093506(...TRUNCATED)
[[0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,-0.014442443847656(...TRUNCATED)
[0.39259999990463257,0.7758402228355408,7.58281946182251,0.0037049006205052137,-0.006268441677093506(...TRUNCATED)
[[0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0072288513183593(...TRUNCATED)
[ 0.2791000008583069, 0.8265125155448914, 7.58281946182251, 0.0037049006205052137, -0.006268441677093506 ]
[[0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,-0.008018493652343(...TRUNCATED)
[ 0.301800012588501, 0.8269698023796082, 7.58281946182251, 0.0037049006205052137, -0.006268441677093506 ]
[[0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,-0.001457214355468(...TRUNCATED)
[0.12110000103712082,0.8809731006622314,7.58281946182251,0.0037049006205052137,-0.006268441677093506(...TRUNCATED)
[[0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,-0.011161804199218(...TRUNCATED)
[ 0.3142000138759613, 0.8107319474220276, 7.58281946182251, 0.0037049006205052137, -0.006268441677093506 ]
[[0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,-0.001120567321777(...TRUNCATED)
[0.31529998779296875,0.8621788024902344,7.58281946182251,0.0037049006205052137,-0.006268441677093506(...TRUNCATED)
[[0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0400390625,0.029(...TRUNCATED)
[ 0.5393999814987183, 0.7002154588699341, 7.58281946182251, 0.0037049006205052137, -0.006268441677093506 ]
[[0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,-0.011909484863281(...TRUNCATED)
[ 0.5023999810218811, 0.8639339208602905, 7.58281946182251, 0.0037049006205052137, -0.006268441677093506 ]
End of preview. Expand in Data Studio

FAIR Universe - NeurIPS 2025 Weak Lensing Uncertainty Challenge

This dataset is a HF mirror of the official challenge training data for this challenge:

https://www.codabench.org/competitions/8934/

This repo contains a preprocessed train/validation split.

The original data has shape ncosmo, np, .... This dataset is splitted along the np dimension with a fraction of 0.8, and reordered as

ncosmo, np, ... -> np, ncosmo, ... 

To get started:

import datasets

dset_train = datasets.load_dataset("b-remy/neurips-wl-challenge-split",split="train", streaming=True)
dset_val = datasets.load_dataset("b-remy/neurips-wl-challenge-split", split="validation", streaming=True)

dset_train = dset_train.with_format('torch')
dset_val = dset_val.with_format('torch')

example = dset_train['train'][0]
Downloads last month
3