sebasmos's picture
adding gated yaml
6503d06
metadata
license: other
license_name: physionet-credentialed-health-data-license-150
license_link: https://physionet.org/content/mimic-iv-echo/view-license/0.1/
extra_gated_heading: Access to MIMIC-IV-Echo V-JEPA2 Embeddings
extra_gated_description: >
  This dataset contains metadata derived from MIMIC-IV-Echo, which is a
  credentialed dataset on PhysioNet. To access this dataset, you must have an
  active PhysioNet credentialed account with signed Data Use Agreement (DUA) for
  MIMIC-IV-Echo.
extra_gated_fields:
  PhysioNet username: text
  I have an active credentialed account on PhysioNet: checkbox
  I have signed the MIMIC-IV-Echo Data Use Agreement: checkbox
  I agree to not redistribute this data: checkbox
  Affiliation: text
tags:
  - medical-imaging
  - echocardiography
  - embeddings
  - mimic-iv-echo
  - v-jepa2
  - video-embeddings
  - self-supervised-learning

MIMIC-IV-Echo V-JEPA2 Embeddings

Pre-computed video embeddings for MIMIC-IV-Echo echocardiography videos, extracted using V-JEPA2 (Meta's self-supervised video encoder).

Access requirement: This dataset includes metadata from MIMIC-IV-Echo (subject IDs, study IDs, timestamps, clinical note references). You must have PhysioNet credentialed access and a signed DUA for MIMIC-IV-Echo before requesting access.

Dataset

Videos 525,328 echocardiography clips
Subjects ~4,800 patients
Studies ~7,200 echo studies
Embedding model V-JEPA2 ViT-L (300M params)
Embedding dim 1024 (float32)
Format Sharded Parquet (10 shards, ~2.1 GB total)

Structure

mimic-iv-echo-jepa-embeddings/
└── jepa-l-embeddings/
    ├── train-00000-of-00010.parquet   (p10, 51K rows, 202 MB)
    ├── train-00001-of-00010.parquet   (p11, 52K rows, 205 MB)
    ├── ...
    └── train-00009-of-00010.parquet   (p19, 53K rows, 208 MB)

Each shard corresponds to one MIMIC-IV patient folder (p10-p19).

Columns

Column Type Source Description
subject_id int64 echo-record-list.csv MIMIC patient ID
study_id int64 echo-record-list.csv Echo study ID
dicom_id str filename Original DICOM identifier (e.g. 94106955_0001)
file_path str embedding key Relative path to source MP4
acquisition_datetime str echo-record-list.csv Per-video acquisition timestamp
study_datetime str echo-study-list.csv Per-study timestamp
note_id str echo-study-list.csv Clinical note reference (nullable)
note_seq str echo-study-list.csv Note sequence number (nullable)
note_charttime str echo-study-list.csv Note chart time (nullable)
embedding list[float32] V-JEPA2 ViT-L 1024-dim video embedding

Metadata is joined from two MIMIC-IV-Echo CSV files so each row is self-contained:

  • echo-record-list.csv (525K rows) — per-video: subject, study, acquisition time
  • echo-study-list.csv (7K rows) — per-study: study time, clinical notes

Usage

from datasets import load_dataset

ds = load_dataset("MITCriticalData/mimic-iv-echo-jepa-embeddings")
print(ds["train"][0])  # {'subject_id': 10002221, 'embedding': [...], ...}
import pyarrow.parquet as pq

table = pq.read_table("jepa-l-embeddings/")
print(table.num_rows)   # 525328
print(table.schema)

Extraction

Embeddings were extracted using the pipeline described in readme-embeddings.md.

Step Command
Extract (SLURM) sbatch extract_slurm.sh vitl
Merge .pt files python merge_embeddings.py --model vitl
Convert to Parquet python to_parquet.py --model vitl

Config: L40S GPU, batch=256, 8 DataLoader workers, ~30 min per folder.

Citation

If you use this dataset, please cite both the original MIMIC-IV-Echo dataset and V-JEPA2:

@article{mimic-iv-echo,
  title={MIMIC-IV-Echo: A Large-Scale Echocardiography Dataset},
  note={PhysioNet, https://physionet.org/content/mimic-iv-echo/}
}

@article{bardes2025vjepa2,
  title={Revisiting Feature Prediction for Learning Visual Representations from Video},
  author={Bardes, Adrien and Garrido, Quentin and Ponce, Jean and Chen, Xinlei and Rabbat, Michael and LeCun, Yann and Assran, Mahmoud and Ballas, Nicolas},
  year={2025}
}