Dataset Viewer
Duplicate
The dataset viewer is not available for this split.
Cannot load the dataset split (in streaming mode) to extract the first rows.
Error code:   StreamingRowsError
Exception:    ImportError
Message:      To support decoding NIfTI files, please install 'nibabel'.
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/src/worker/utils.py", line 99, in get_rows_or_raise
                  return get_rows(
                         ^^^^^^^^^
                File "/src/libs/libcommon/src/libcommon/utils.py", line 272, in decorator
                  return func(*args, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^^^
                File "/src/services/worker/src/worker/utils.py", line 77, in get_rows
                  rows_plus_one = list(itertools.islice(ds, rows_max_number + 1))
                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2690, in __iter__
                  for key, example in ex_iterable:
                                      ^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2240, in __iter__
                  example = _apply_feature_types_on_example(
                            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2159, in _apply_feature_types_on_example
                  decoded_example = features.decode_example(encoded_example, token_per_repo_id=token_per_repo_id)
                                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/features/features.py", line 2204, in decode_example
                  column_name: decode_nested_example(feature, value, token_per_repo_id=token_per_repo_id)
                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/features/features.py", line 1508, in decode_nested_example
                  return schema.decode_example(obj, token_per_repo_id=token_per_repo_id) if obj is not None else None
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/features/nifti.py", line 172, in decode_example
                  raise ImportError("To support decoding NIfTI files, please install 'nibabel'.")
              ImportError: To support decoding NIfTI files, please install 'nibabel'.

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

Dartbrains Localizer Dataset

A subset of the Brainomics/Localizer functional MRI dataset, prepared for the Dartbrains neuroimaging course at Dartmouth College.

Quick Start

Load beta maps (recommended for most exercises)

from datasets import load_dataset

ds = load_dataset("dartbrains/localizer", "betas")
img = ds[0]["nifti"]       # nibabel.Nifti1Image
subject = ds[0]["subject"]  # "S01"
condition = ds[0]["condition"]  # "audio_computation"

Load event files as a table

ds = load_dataset("dartbrains/localizer", "events")
# Convert to Polars
import polars as pl
df = pl.from_arrow(ds["train"].to_arrow())

Load a single file directly (for nibabel/nltools workflows)

from huggingface_hub import hf_hub_download

path = hf_hub_download(
    repo_id="dartbrains/localizer",
    filename="derivatives/betas/S01_betas.nii.gz",
    repo_type="dataset",
)

# Use with nibabel
import nibabel as nib
img = nib.load(path)

# Use with nltools
from nltools.data import Brain_Data
brain = Brain_Data(path)

Load specific subjects (selective download)

from huggingface_hub import snapshot_download

path = snapshot_download(
    repo_id="dartbrains/localizer",
    repo_type="dataset",
    allow_patterns=["derivatives/fmriprep/sub-S01/**", "sub-S01/**"],
)

Load tabular data with Polars

import polars as pl

events = pl.read_csv(
    "hf://datasets/dartbrains/localizer/sub-S01/func/sub-S01_task-localizer_events.tsv",
    separator="\t",
)

Dataset Structure

dartbrains/localizer/
β”œβ”€β”€ dataset_description.json
β”œβ”€β”€ participants.tsv
β”œβ”€β”€ participants.json
β”œβ”€β”€ task-localizer_bold.json
β”œβ”€β”€ README
β”œβ”€β”€ sub-S01/
β”‚   └── func/
β”‚       └── sub-S01_task-localizer_events.tsv
β”œβ”€β”€ sub-S02/
β”‚   └── ...
β”œβ”€β”€ derivatives/
β”‚   β”œβ”€β”€ betas/
β”‚   β”‚   β”œβ”€β”€ S01_betas.nii.gz              # all conditions stacked
β”‚   β”‚   β”œβ”€β”€ S01_beta_audio_computation.nii.gz
β”‚   β”‚   β”œβ”€β”€ S01_beta_audio_left_hand.nii.gz
β”‚   β”‚   └── ...
β”‚   └── fmriprep/
β”‚       β”œβ”€β”€ sub-S01/
β”‚       β”‚   β”œβ”€β”€ anat/     # T1w preprocessed, transforms
β”‚       β”‚   β”œβ”€β”€ figures/   # QC reports
β”‚       β”‚   └── func/     # preprocessed BOLD, confounds, masks
β”‚       └── ...

Conditions

The localizer task includes the following conditions:

  • audio_computation / video_computation
  • audio_sentence / video_sentence
  • audio_left_hand / audio_right_hand
  • video_left_hand / video_right_hand
  • horizontal_checkerboard / vertical_checkerboard

Citation

@article{papadopoulos2017brainomics,
  title={The Brainomics/Localizer database},
  author={Papadopoulos Orfanos, Dimitri and Michel, Vincent and Schwartz, Yannick and Pinel, Philippe and Moreno, Antonio and Le Bihan, Denis and Frouin, Vincent},
  journal={NeuroImage},
  volume={144},
  pages={309--314},
  year={2017},
  doi={10.1016/j.neuroimage.2015.09.052}
}
Downloads last month
54