File size: 4,617 Bytes
86f9633
6503d06
 
 
 
 
 
 
 
 
 
 
 
 
 
86f9633
 
 
 
 
 
 
 
 
 
 
 
 
 
6503d06
 
86f9633
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6503d06
86f9633
 
 
 
 
 
 
 
 
 
 
 
 
e646b4d
86f9633
 
 
 
 
 
 
 
6503d06
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
---
license: other
license_name: physionet-credentialed-health-data-license-150
license_link: https://physionet.org/content/mimic-iv-echo/view-license/0.1/
extra_gated_heading: "Access to MIMIC-IV-Echo V-JEPA2 Embeddings"
extra_gated_description: >
  This dataset contains metadata derived from MIMIC-IV-Echo, which is a credentialed dataset on PhysioNet.
  To access this dataset, you must have an active PhysioNet credentialed account with signed
  Data Use Agreement (DUA) for MIMIC-IV-Echo.
extra_gated_fields:
  PhysioNet username: text
  I have an active credentialed account on PhysioNet: checkbox
  I have signed the MIMIC-IV-Echo Data Use Agreement: checkbox
  I agree to not redistribute this data: checkbox
  Affiliation: text
tags:
- medical-imaging
- echocardiography
- embeddings
- mimic-iv-echo
- v-jepa2
- video-embeddings
- self-supervised-learning
---

# MIMIC-IV-Echo V-JEPA2 Embeddings

Pre-computed video embeddings for [MIMIC-IV-Echo](https://physionet.org/content/mimic-iv-echo/) echocardiography videos, extracted using [V-JEPA2](https://github.com/facebookresearch/jepa) (Meta's self-supervised video encoder).

> **Access requirement:** This dataset includes metadata from MIMIC-IV-Echo (subject IDs, study IDs, timestamps, clinical note references). You must have [PhysioNet credentialed access](https://physionet.org/settings/credentialing/) and a signed DUA for [MIMIC-IV-Echo](https://physionet.org/content/mimic-iv-echo/) before requesting access.

## Dataset

| | |
|---|---|
| **Videos** | 525,328 echocardiography clips |
| **Subjects** | ~4,800 patients |
| **Studies** | ~7,200 echo studies |
| **Embedding model** | V-JEPA2 ViT-L (300M params) |
| **Embedding dim** | 1024 (float32) |
| **Format** | Sharded Parquet (10 shards, ~2.1 GB total) |

## Structure

```
mimic-iv-echo-jepa-embeddings/
└── jepa-l-embeddings/
    ├── train-00000-of-00010.parquet   (p10, 51K rows, 202 MB)
    ├── train-00001-of-00010.parquet   (p11, 52K rows, 205 MB)
    ├── ...
    └── train-00009-of-00010.parquet   (p19, 53K rows, 208 MB)
```

Each shard corresponds to one MIMIC-IV patient folder (p10-p19).

## Columns

| Column | Type | Source | Description |
|--------|------|--------|-------------|
| `subject_id` | int64 | echo-record-list.csv | MIMIC patient ID |
| `study_id` | int64 | echo-record-list.csv | Echo study ID |
| `dicom_id` | str | filename | Original DICOM identifier (e.g. `94106955_0001`) |
| `file_path` | str | embedding key | Relative path to source MP4 |
| `acquisition_datetime` | str | echo-record-list.csv | Per-video acquisition timestamp |
| `study_datetime` | str | echo-study-list.csv | Per-study timestamp |
| `note_id` | str | echo-study-list.csv | Clinical note reference (nullable) |
| `note_seq` | str | echo-study-list.csv | Note sequence number (nullable) |
| `note_charttime` | str | echo-study-list.csv | Note chart time (nullable) |
| `embedding` | list[float32] | V-JEPA2 ViT-L | 1024-dim video embedding |

Metadata is joined from two MIMIC-IV-Echo CSV files so each row is self-contained:
- **echo-record-list.csv** (525K rows) — per-video: subject, study, acquisition time
- **echo-study-list.csv** (7K rows) — per-study: study time, clinical notes

## Usage

```python
from datasets import load_dataset

ds = load_dataset("MITCriticalData/mimic-iv-echo-jepa-embeddings")
print(ds["train"][0])  # {'subject_id': 10002221, 'embedding': [...], ...}
```

```python
import pyarrow.parquet as pq

table = pq.read_table("jepa-l-embeddings/")
print(table.num_rows)   # 525328
print(table.schema)
```

## Extraction

Embeddings were extracted using the pipeline described in [readme-embeddings.md](https://github.com/sebasmos/EchoJEPA-VE/blob/main/readme-embeddings.md).

| Step | Command |
|------|---------|
| Extract (SLURM) | `sbatch extract_slurm.sh vitl` |
| Merge .pt files | `python merge_embeddings.py --model vitl` |
| Convert to Parquet | `python to_parquet.py --model vitl` |

Config: L40S GPU, batch=256, 8 DataLoader workers, ~30 min per folder.

## Citation

If you use this dataset, please cite both the original MIMIC-IV-Echo dataset and V-JEPA2:

```bibtex
@article{mimic-iv-echo,
  title={MIMIC-IV-Echo: A Large-Scale Echocardiography Dataset},
  note={PhysioNet, https://physionet.org/content/mimic-iv-echo/}
}

@article{bardes2025vjepa2,
  title={Revisiting Feature Prediction for Learning Visual Representations from Video},
  author={Bardes, Adrien and Garrido, Quentin and Ponce, Jean and Chen, Xinlei and Rabbat, Michael and LeCun, Yann and Assran, Mahmoud and Ballas, Nicolas},
  year={2025}
}
```