sajjad5221 commited on
Commit
4d6a44b
·
verified ·
1 Parent(s): 2aee8cf

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +57 -191
README.md CHANGED
@@ -5,68 +5,31 @@ tags:
5
  - brain-signals
6
  - neuroscience
7
  - reading
8
- - zuco
9
  - brain-to-text
10
  - cognitive-neuroscience
11
- license: mit
 
12
  task_categories:
13
  - text-generation
14
- pretty_name: ZuCo EEG Reading Dataset
15
  size_categories:
16
- - 1K<n<10K
17
- dataset_info:
18
- features:
19
- - name: participant_id
20
- dtype: string
21
- - name: sentence_id
22
- dtype: string
23
- - name: sentence_text
24
- dtype: string
25
- - name: word_eeg_segments
26
- list:
27
- list:
28
- list: float32
29
- - name: num_words
30
- dtype: int32
31
- - name: eeg_shape
32
- dtype: string
33
- - name: has_gaze_alignment
34
- dtype: bool
35
- splits:
36
- - name: train
37
- num_bytes: 15090427
38
- num_examples: 576
39
- - name: validation
40
- num_bytes: 3469583
41
- num_examples: 128
42
- - name: test
43
- num_bytes: 4327666
44
- num_examples: 160
45
- download_size: 19869836
46
- dataset_size: 22887676
47
- configs:
48
- - config_name: default
49
- data_files:
50
- - split: train
51
- path: data/train-*
52
- - split: validation
53
- path: data/validation-*
54
- - split: test
55
- path: data/test-*
56
  ---
57
 
58
- # ZuCo EEG Reading Dataset
59
 
60
- This dataset contains preprocessed EEG brain signals recorded while participants read English sentences. It is derived from the [ZuCo (Zurich Cognitive Language Processing Corpus)](https://osf.io/q3zws/) dataset.
61
 
62
  ## Dataset Description
63
 
64
- - **Task**: EEG-to-Text (Brain-to-Text) Generation
65
- - **Modality**: EEG (Electroencephalography)
66
- - **Channels**: 4 (RAW_TP9, RAW_AF7, RAW_AF8, RAW_TP10)
67
  - **Sampling Rate**: 256 Hz
68
- - **Language**: English
69
- - **Participants**: Multiple subjects reading natural sentences
 
70
 
71
  ## Dataset Structure
72
 
@@ -74,175 +37,78 @@ This dataset contains preprocessed EEG brain signals recorded while participants
74
 
75
  | Split | Samples | Description |
76
  |-------|---------|-------------|
77
- | Train | 576 | 70% of participants for training |
78
- | Validation | 128 | 15% of participants for validation |
79
- | Test | 160 | 15% of participants for testing |
80
 
81
- **Important**: Splits are done by **participant**, ensuring no data leakage. Different participants read different sentences, so validation and test sets contain both unseen participants and unseen sentence groups.
82
 
83
  ### Data Fields
84
 
85
  Each sample contains:
86
 
87
- - **`participant`** (string): Participant ID (e.g., "P01", "P02")
88
- - **`sentence_id`** (string): Sentence identifier (e.g., "S001", "S002")
89
- - **`sentence`** (string): The actual text that was being read
90
  - **`word_eeg_segments`** (3D array): Word-level EEG data
91
  - Shape: `[num_words, time_samples, 4_channels]`
92
  - Each word has its own EEG segment extracted using gaze-timestamp alignment
93
- - Time samples vary per word (typically 50-200 samples = 200-800ms)
94
  - **`num_words`** (int): Number of words in the sentence
95
- - **`eeg_shape`** (string): Shape information (e.g., "15x128x4" = 15 words, 128 time samples, 4 channels)
 
96
 
97
- ### Preprocessing
98
 
99
- The EEG data has been preprocessed with:
100
 
101
- 1. **Notch Filter**: 50Hz (removes power line noise)
102
- 2. **Bandpass Filter**: 0.5-50Hz (keeps relevant brain frequencies)
103
- 3. **Robust Scaling**: Normalization robust to eye blinks and artifacts
104
  4. **Gaze-Timestamp Alignment**: EEG segments aligned to actual word reading times using eye-tracking fixations
105
 
106
- ## Usage
107
-
108
- ### Loading the Dataset
109
-
110
- ```python
111
- from datasets import load_dataset
112
-
113
- # Load the dataset
114
- dataset = load_dataset("sajjad5221/eeg2text-emmt-dataset")
115
-
116
- # Access splits
117
- train_data = dataset['train']
118
- val_data = dataset['validation']
119
- test_data = dataset['test']
120
-
121
- # Examine a sample
122
- sample = train_data[0]
123
- print(f"Participant: {sample['participant']}")
124
- print(f"Sentence: {sample['sentence']}")
125
- print(f"EEG shape: {sample['eeg_shape']}")
126
- print(f"Number of words: {sample['num_words']}")
127
-
128
- # Access word-level EEG data
129
- word_eeg = sample['word_eeg_segments'] # List of [time, channels] arrays
130
- first_word_eeg = word_eeg[0] # EEG for first word: [time_samples, 4]
131
- print(f"First word EEG shape: {len(first_word_eeg)}x{len(first_word_eeg[0])}")
132
- ```
133
-
134
- ### Example: Training a Model
135
-
136
- ```python
137
- import torch
138
- from torch.utils.data import DataLoader
139
- from datasets import load_dataset
140
-
141
- # Load dataset
142
- dataset = load_dataset("sajjad5221/eeg2text-emmt-dataset")
143
-
144
- # Convert to PyTorch tensors
145
- def collate_fn(batch):
146
- # Your custom collation logic here
147
- # Handle variable-length sequences
148
- pass
149
-
150
- train_loader = DataLoader(
151
- dataset['train'],
152
- batch_size=8,
153
- shuffle=True,
154
- collate_fn=collate_fn
155
- )
156
-
157
- # Training loop
158
- for batch in train_loader:
159
- eeg_data = batch['word_eeg_segments']
160
- text = batch['sentence']
161
- # Your training code here
162
- ```
163
 
164
  ## Dataset Statistics
165
 
 
 
 
 
 
 
166
  ### Training Set
167
- - Samples: 576
168
- - Unique participants: ~70% of total
169
- - Average words per sentence: ~10.9
170
 
171
  ### Validation Set
172
- - Samples: 128
173
- - Unique participants: ~15% of total (different from training)
174
- - Used for: Model selection, hyperparameter tuning
175
 
176
  ### Test Set
177
- - Samples: 160
178
- - Unique participants: ~15% of total (different from training & validation)
179
- - Used for: Final performance evaluation on unseen participants
180
 
181
  ## EEG Channel Information
182
 
183
- The 4 EEG channels capture different brain regions:
184
-
185
- | Channel | Location | Brain Region |
186
- |---------|----------|--------------|
187
- | TP9 | Left temporal-parietal | Language processing, memory |
188
- | AF7 | Left anterior-frontal | Attention, working memory |
189
- | AF8 | Right anterior-frontal | Attention, executive function |
190
- | TP10 | Right temporal-parietal | Visual processing, attention |
191
-
192
- ## Citation
193
-
194
- If you use this dataset, please cite the original ZuCo paper:
195
-
196
- ```bibtex
197
- @article{hollenstein2018zuco,
198
- title={ZuCo, a simultaneous EEG and eye-tracking resource for natural sentence reading},
199
- author={Hollenstein, Nora and Rotsztejn, Jonathan and Troendle, Marius and Pedroni, Andreas and Zhang, Ce and Langer, Nicolas},
200
- journal={Scientific Data},
201
- volume={5},
202
- number={1},
203
- pages={180291},
204
- year={2018},
205
- publisher={Nature Publishing Group}
206
- }
207
- ```
208
-
209
- ## Related Papers
210
-
211
- ### ZuCo Dataset
212
- - [ZuCo 1.0](https://www.nature.com/articles/sdata2018291) - Original dataset paper
213
- - [ZuCo 2.0](https://www.nature.com/articles/s41597-022-01387-w) - Extended dataset
214
-
215
- ### EEG-to-Text Models
216
- - [Brain2Text](https://arxiv.org/abs/2112.02685) - Neural decoding of text from brain signals
217
- - [DeWave](https://arxiv.org/abs/2309.14030) - Discrete EEG waves for text generation
218
-
219
- ## Limitations
220
-
221
- - **Small vocabulary**: Limited to sentences in the ZuCo corpus
222
- - **Controlled setting**: Lab environment, not naturalistic reading
223
- - **Limited participants**: Dataset size constrained by number of participants
224
- - **Hardware dependency**: Recorded with specific EEG device (Muse headband)
225
- - **Individual differences**: Brain signals vary significantly across individuals
226
-
227
- ## Ethics
228
-
229
- - All participants provided informed consent
230
- - Data has been anonymized (participant IDs are pseudonyms)
231
- - No sensitive personal information is included
232
- - Original study approved by ethics committee
233
-
234
- ## License
235
 
236
- This dataset is released under the MIT License. Please refer to the original ZuCo dataset terms of use for any additional restrictions.
 
 
 
 
 
237
 
238
- ## Contact
239
 
240
- For questions or issues:
241
- - Open an issue on the [repository](https://huggingface.co/datasets/sajjad5221/eeg2text-emmt-dataset)
242
- - Refer to the [original ZuCo project](https://osf.io/q3zws/) for dataset-specific questions
243
 
244
- ## Acknowledgments
 
 
 
245
 
246
- - ZuCo dataset creators at University of Zurich
247
- - All participants who contributed their brain signals
248
- - Research groups advancing brain-computer interfaces
 
5
  - brain-signals
6
  - neuroscience
7
  - reading
8
+ - emmt
9
  - brain-to-text
10
  - cognitive-neuroscience
11
+ - eye-tracking
12
+ license: cc-by-4.0
13
  task_categories:
14
  - text-generation
15
+ pretty_name: EMMT EEG-to-Text Dataset
16
  size_categories:
17
+ - n<1K
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ---
19
 
20
+ # EMMT EEG-to-Text Dataset
21
 
22
+ This dataset contains preprocessed EEG brain signals recorded while participants read English sentences and translated them to Czech. It is derived from the **EMMT (Eyetracked Multi-Modal Translation)** corpus.
23
 
24
  ## Dataset Description
25
 
26
+ - **Task**: EEG-to-Text Generation (Brain-Computer Interface)
27
+ - **Modality**: EEG (Electroencephalography) + Eye-tracking
28
+ - **Channels**: 4 EEG channels (RAW_TP9, RAW_AF7, RAW_AF8, RAW_TP10)
29
  - **Sampling Rate**: 256 Hz
30
+ - **Language**: English sentences (to be translated to Czech)
31
+ - **Participants**: 21 participants
32
+ - **Total Sentences**: 200 unique sentences
33
 
34
  ## Dataset Structure
35
 
 
37
 
38
  | Split | Samples | Description |
39
  |-------|---------|-------------|
40
+ | Train | 576 | Training set (70% of participants) |
41
+ | Validation | 128 | Validation set (15% of participants) |
42
+ | Test | 160 | Test set (15% of participants) |
43
 
44
+ **Important**: Splits are done by **participant** to prevent data leakage. Different participants may read the same sentences, ensuring the model learns to generalize across individuals.
45
 
46
  ### Data Fields
47
 
48
  Each sample contains:
49
 
50
+ - **`participant_id`** (string): Anonymized participant identifier (e.g., "P01", "P02")
51
+ - **`sentence_id`** (string): Sentence identifier (e.g., "S001", "S062")
52
+ - **`sentence_text`** (string): The English sentence that was being read
53
  - **`word_eeg_segments`** (3D array): Word-level EEG data
54
  - Shape: `[num_words, time_samples, 4_channels]`
55
  - Each word has its own EEG segment extracted using gaze-timestamp alignment
56
+ - Time samples vary per word (typically 50-200 samples = 200-800ms @ 256Hz)
57
  - **`num_words`** (int): Number of words in the sentence
58
+ - **`eeg_shape`** (string): Human-readable shape info (e.g., "15words_x_128time_x_4channels")
59
+ - **`has_gaze_alignment`** (bool): Whether gaze-based alignment was used
60
 
61
+ ### Preprocessing Pipeline
62
 
63
+ The EEG data has been preprocessed with standard neuroscience techniques:
64
 
65
+ 1. **Notch Filter**: 50Hz (removes electrical power line noise)
66
+ 2. **Bandpass Filter**: 0.5-50Hz (keeps relevant brain frequencies, removes drift and high-freq noise)
67
+ 3. **Robust Scaling**: Normalization robust to eye blinks and movement artifacts
68
  4. **Gaze-Timestamp Alignment**: EEG segments aligned to actual word reading times using eye-tracking fixations
69
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
70
 
71
  ## Dataset Statistics
72
 
73
+ ### Overall Statistics
74
+ - **Total Samples**: 864
75
+ - **Unique Participants**: 21
76
+ - **Unique Sentences**: 200
77
+ - **Average Words per Sentence**: 11.0
78
+
79
  ### Training Set
80
+ - **Samples**: 576
81
+ - **Participants**: ~70% of total
82
+ - **Purpose**: Model training
83
 
84
  ### Validation Set
85
+ - **Samples**: 128
86
+ - **Participants**: ~15% of total (different from training)
87
+ - **Purpose**: Hyperparameter tuning, model selection
88
 
89
  ### Test Set
90
+ - **Samples**: 160
91
+ - **Participants**: ~15% of total (different from training & validation)
92
+ - **Purpose**: Final performance evaluation on unseen participants
93
 
94
  ## EEG Channel Information
95
 
96
+ The 4 EEG channels capture different brain regions relevant to language processing:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
97
 
98
+ | Channel | Location | Brain Region | Function |
99
+ |---------|----------|--------------|----------|
100
+ | **RAW_TP9** | Left temporal-parietal | Language processing | Semantic understanding, memory |
101
+ | **RAW_AF7** | Left anterior-frontal | Executive function | Attention, working memory |
102
+ | **RAW_AF8** | Right anterior-frontal | Executive function | Attention, cognitive control |
103
+ | **RAW_TP10** | Right temporal-parietal | Visual-spatial | Visual processing, attention |
104
 
105
+ ## Original EMMT Experiment
106
 
107
+ The data comes from the **EMMT (Eyetracked Multi-Modal Translation)** corpus, which recorded participants performing a multimodal translation task:
 
 
108
 
109
+ 1. **READ**: Read English sentence aloud
110
+ 2. **TRANSLATE**: Translate to Czech aloud
111
+ 3. **SEE**: View accompanying image
112
+ 4. **UPDATE**: Update or repeat translation
113
 
114
+ This dataset contains EEG data from the **READ** stage, aligned to word-level using eye-tracking fixations.