File size: 5,176 Bytes
8fbc5e5 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 |
# LibriBrain MEG Preprocessed Dataset
Preprocessed magnetoencephalography (MEG) recordings with phoneme labels from the LibriBrain dataset, optimized for fast loading during machine learning model training.
This dataset was created for the [LibriBrain 2025 Competition](https://neural-processing-lab.github.io/2025-libribrain-competition/) (now concluded).
## Dataset Overview
### MEG Recording Specifications
- **Channels**: 306 total (102 magnetometers + 204 gradiometers)
- **Sampling Rate**: 250 Hz
- **Duration**: ~52 hours of recordings
- **Subject**: Single English speaker listening to Sherlock Holmes audiobooks
- **Phoneme Instances**: ~1.5 million
### Phoneme Inventory
39 ARPAbet phonemes with position encoding:
- **Vowels** (15): aa, ae, ah, ao, aw, ay, eh, er, ey, ih, iy, ow, oy, uh, uw
- **Consonants** (24): b, ch, d, dh, f, g, hh, jh, k, l, m, n, ng, p, r, s, sh, t, th, v, w, y, z, zh
- **Special**: oov (out-of-vocabulary)
Position markers: B (beginning), I (inside), E (end), S (singleton)
### Signal Processing
All MEG data has been preprocessed through the following pipeline:
1. Bad channel removal
2. Signal Space Separation (SSS) for noise reduction
3. Notch filtering for powerline noise removal
4. Bandpass filtering (0.1-125 Hz)
5. Downsampling to 250 Hz
## Preprocessing and Grouping
This dataset contains pre-grouped and averaged MEG samples for significantly faster data loading during training. Instead of grouping samples on-the-fly (which is computationally expensive), samples have been pre-grouped at various levels.
### Available Grouping Configurations
- `grouped_5`: 5 samples averaged together
- `grouped_10`: 10 samples averaged together
- `grouped_15`: 15 samples averaged together
- `grouped_20`: 20 samples averaged together
- `grouped_25`: 25 samples averaged together
- `grouped_30`: 30 samples averaged together
- `grouped_35`: 35 samples averaged together (partial - train only)
- `grouped_45`: 45 samples averaged together
- `grouped_50`: 50 samples averaged together
- `grouped_55`: 55 samples averaged together
- `grouped_60`: 60 samples averaged together
- `grouped_100`: 100 samples averaged together
Each configuration contains:
- `train_grouped.h5`: Training data
- `validation_grouped.h5`: Validation data
- `test_grouped.h5`: Test data
- `paths.yaml`: File path references
### Why Use Grouped Data?
- **Faster Loading**: Pre-computed grouping eliminates runtime averaging overhead
- **Memory Efficient**: Smaller file sizes for higher grouping levels
- **Flexible**: Choose grouping level based on your accuracy vs. speed requirements
- **Standardized**: Consistent preprocessing across all configurations
## Installation
This dataset requires the modified pnpl library for loading:
```bash
pip install git+https://github.com/September-Labs/pnpl.git
```
## Usage
```python
from pnpl.datasets import GroupedDataset
# Load preprocessed data with 100-sample grouping
train_dataset = GroupedDataset(
preprocessed_path="data/grouped_100/train_grouped.h5",
load_to_memory=True # Optional: load entire dataset to memory for faster access
)
val_dataset = GroupedDataset(
preprocessed_path="data/grouped_100/validation_grouped.h5",
load_to_memory=True
)
# Get a sample
sample = train_dataset[0]
meg_data = sample['meg'] # Shape: (306, time_points)
phoneme_label = sample['phoneme'] # Phoneme class index
# Use with PyTorch DataLoader
from torch.utils.data import DataLoader
dataloader = DataLoader(
train_dataset,
batch_size=32,
shuffle=True,
num_workers=4
)
```
## Data Structure
```
data/
├── grouped_5/
│ ├── train_grouped.h5
│ ├── validation_grouped.h5
│ ├── test_grouped.h5
│ └── paths.yaml
├── grouped_10/
│ ├── train_grouped.h5
│ ├── validation_grouped.h5
│ ├── test_grouped.h5
│ └── paths.yaml
├── ...
└── grouped_100/
├── train_grouped.h5
├── validation_grouped.h5
├── test_grouped.h5
└── paths.yaml
```
## File Sizes
| Grouping | Train | Validation | Test | Total |
|----------|-------|------------|------|-------|
| grouped_5 | 45.6 GB | 425 MB | 456 MB | ~47 GB |
| grouped_10 | 22.8 GB | 213 MB | 228 MB | ~24 GB |
| grouped_20 | 11.4 GB | 106 MB | 114 MB | ~12 GB |
| grouped_50 | 4.6 GB | 37 MB | 42 MB | ~4.7 GB |
| grouped_100 | 2.3 GB | 19 MB | 21 MB | ~2.4 GB |
## Dataset Splits
- **Train**: 88 sessions (~51 hours)
- **Validation**: 1 session (~0.36 hours)
- **Test**: 1 session (~0.38 hours)
## Citation
If you use this dataset, please cite the LibriBrain competition:
```bibtex
@misc{libribrain2025,
title={LibriBrain: A Dataset for Speech Decoding from Brain Signals},
author={Neural Processing Lab},
year={2025},
url={https://neural-processing-lab.github.io/2025-libribrain-competition/}
}
```
## License
Please refer to the original LibriBrain dataset license terms.
## Acknowledgments
This preprocessed version was created to facilitate faster training for the LibriBrain 2025 Competition. The original dataset and competition were organized by the Neural Processing Lab. |