Datasets:

Tasks:
Other
Modalities:
Audio
Languages:
English
ArXiv:
License:

Add paper link, GitHub link, task category and sample usage

#6
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +72 -3
README.md CHANGED
@@ -1,12 +1,68 @@
1
  ---
2
- license: cc-by-nc-4.0
3
  language:
4
- - en
 
 
 
5
  ---
6
 
7
  # LibriBrain (Sherlock Holmes 1–7)
8
 
9
- This repository contains the LibriBrain data organised by book: MEG recordings (`.h5`), event annotations (`.tsv`), and the audiobook stimulus audio (`.wav`).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
 
11
  ## Repository structure
12
 
@@ -50,6 +106,8 @@ sox "INPUT_FROM_LIBRIVOX.mp3" -c 1 -r 22050 -b 16 "OUTPUT.wav"
50
 
51
  ### Citation
52
 
 
 
53
  ```bibtex
54
  @article{ozdogan2025libribrain,
55
  author = {Özdogan, Miran and Landau, Gilad and Elvers, Gereon and Jayalath, Dulhan and Somaiya, Pratik and Mantegna, Francesco and Woolrich, Mark and Parker Jones, Oiwi},
@@ -59,3 +117,14 @@ sox "INPUT_FROM_LIBRIVOX.mp3" -c 1 -r 22050 -b 16 "OUTPUT.wav"
59
  url = {https://arxiv.org/abs/2506.02098},
60
  }
61
  ```
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
2
  language:
3
+ - en
4
+ license: cc-by-nc-4.0
5
+ task_categories:
6
+ - other
7
  ---
8
 
9
  # LibriBrain (Sherlock Holmes 1–7)
10
 
11
+ [Paper](https://huggingface.co/papers/2506.02098) | [Code](https://github.com/neural-processing-lab/pnpl)
12
+
13
+ This repository contains the LibriBrain data organised by book: MEG recordings (`.h5`), event annotations (`.tsv`), and the audiobook stimulus audio (`.wav`).
14
+
15
+ LibriBrain was first open-sourced as part of the [2025 PNPL Competition](https://libribrain.com/).
16
+
17
+ In addition, LibriBrain is used as a fine-tuning dataset in the paper ["MEG-XL: Data-Efficient Brain-to-Text via Long-Context Pre-Training"](https://huggingface.co/papers/2602.02494) to evaluate word decoding from brain data.
18
+
19
+ ## Sample Usage
20
+ The easiest way to get started with the dataset is using the [pnpl Python library](https://github.com/neural-processing-lab/pnpl). There, the following two datasets are available:
21
+
22
+ ### LibriBrainSpeech
23
+ This wraps the LibriBrain dataset for use in speech detection problems.
24
+ ```python
25
+ from pnpl.datasets import LibriBrainSpeech
26
+
27
+ speech_example_data = LibriBrainSpeech(
28
+ data_path="./data/",
29
+ partition="train"
30
+ )
31
+
32
+ sample_data, label = speech_example_data[0]
33
+
34
+ # Print out some basic info about the sample
35
+ print("Sample data shape:", sample_data.shape)
36
+ print("Label shape:", label.shape)
37
+ ```
38
+
39
+ ### LibriBrainPhoneme
40
+ This wraps the LibriBrain dataset for use in phoneme classification problems.
41
+ ```python
42
+ from pnpl.datasets import LibriBrainPhoneme
43
+
44
+ phoneme_example_data = LibriBrainPhoneme(
45
+ data_path="./data/",
46
+ partition="train"
47
+ )
48
+ sample_data, label = phoneme_example_data[0]
49
+
50
+ # Print out some basic info about the sample
51
+ print("Sample data shape:", sample_data.shape)
52
+ print("Label shape:", label.shape)
53
+ ```
54
+
55
+ ### Usage in MEG-XL
56
+ To fine-tune the MEG-XL model on the LibriBrain dataset for word decoding, you can use the following command from the [official repository](https://github.com/neural-processing-lab/MEG-XL):
57
+
58
+ ```bash
59
+ python -m brainstorm.evaluate_criss_cross_word_classification \
60
+ --config-name=eval_criss_cross_word_classification_libribrain \
61
+ model.criss_cross_checkpoint=/path/to/your/checkpoint.ckpt
62
+ ```
63
+
64
+ Note: For the MEG-XL repo, you will need to adjust the dataset paths in the configuration files to point to your local download of the data. The pnpl library includes automatic downloads from HuggingFace.
65
+
66
 
67
  ## Repository structure
68
 
 
106
 
107
  ### Citation
108
 
109
+ If you use this dataset, please cite the LibriBrain paper:
110
+
111
  ```bibtex
112
  @article{ozdogan2025libribrain,
113
  author = {Özdogan, Miran and Landau, Gilad and Elvers, Gereon and Jayalath, Dulhan and Somaiya, Pratik and Mantegna, Francesco and Woolrich, Mark and Parker Jones, Oiwi},
 
117
  url = {https://arxiv.org/abs/2506.02098},
118
  }
119
  ```
120
+
121
+ If you use this data with the MEG-XL framework, please also cite:
122
+
123
+ ```bibtex
124
+ @article{jayalath2026megxl,
125
+ title={{MEG-XL}: Data-Efficient Brain-to-Text via Long-Context Pre-Training},
126
+ author={Jayalath, Dulhan and Jones, Oiwi Parker},
127
+ journal={arXiv preprint arXiv:2602.02494},
128
+ year={2026}
129
+ }
130
+ ```