Singh2020 / README.md
jalauer's picture
Add link to EEG-Bench
7a78668 verified
---
license: pddl
tags:
- eeg
- medical
- clinical
- classification
- parkinson
- pedaling
---
# Singh2020: EEG Parkinson's Classification Dataset with Pedaling Task
The Singh2020 dataset contains EEG recordings collected during a lower-limb pedaling task designed to assess motor control in individuals with Parkinson's disease (PD), with a particular focus on freezing of gait (FOG) symptoms. A total of 39 participants were included: 13 PD patients with FOG (PDFOG+), 13 PD patients without FOG (PDFOG-), and 13 demographically matched healthy controls.
Participants completed a lower-limb motor task in which they pedaled one rotation in response to a visual "GO" cue, designed to minimize fall risk and reduce EEG artifacts from movement. Each subject completed at least two blocks of either 30 or 50 trials, with PDFOG+ participants performing fewer trials due to symptom severity. A tri-axial accelerometer mounted on the ankle measured pedaling kinematics.
EEG was recorded using a 64-channel cap with a sampling rate of 500 Hz.
## Paper
Singh, A., Cole, R. C., Espinoza, A. I., Brown, D., Cavanagh, J. F., & Narayanan, N. S. (2020). **Frontal theta and beta oscillations during lower-limb movement in Parkinson’s disease**. _Clinical Neurophysiology_, 131(3), 694-702.
DISCLAIMER: We (DISCO) are NOT the owners or creators of this dataset, but we merely uploaded it here, to support our's ([EEG-Bench](https://github.com/ETH-DISCO/EEG-Bench)) and other's work on EEG benchmarking.
## Dataset Structure
- `raw_data/` contains the unprocessed recording EEG and accelerometer data.
- `processed_data/` contains pre- (`_Ped_Processed.mat`) and post-processed (`_ANALYZED.mat`) experiment data, produced from the `RAW_DATA/` using `Step1_EEG_Pedaling_PreProcess.m` and `Step2_EEG_Pedaling_PostProcess.m` scripts, respectively.
- `ALL_data_Modeling.csv` contains participant information, such as age and MOCA test scores.
- `scripts/` contains various scripts that can be used to reproduce the results from the paper.
- `ORIGINAL_README.txt`: the original README with perhaps some more helpful information.
### Filename Format
In `raw_data/`, a recording consists of 3 files:
- `[GROUP][PID].vhdr`: The header file with meta information
- `[GROUP][PID].eeg`: contains the EEG (and accelerometer) data
- `[GROUP][PID].vmrk`: contains event information
where GROUP is either `Control` or `PD` (Parkinson's Disease) and PID is the participant's ID (also used in `ALL_data_Modeling.csv`).
Similarly, in `processed_data/`,
- `[GROUP][PID]_Ped_Processed.mat` contains pre-processed data epoched around trials
- `[GROUP][PID]_ANALYZED.mat` contains the time-frequency analysis of the pre-processed data
where, again, GROUP is either `Control` or `PD` (Parkinson's Disease) and PID is the participant's ID.
### Reading RAW_DATA Files in Python
In python, the 3 files that make up a raw recording can be read via:
```python
import mne
raw = mne.io.read_raw_brainvision("path_to/[GROUP][PID].vhdr")
```
Now, `raw.get_data(units='uV')` yields a numpy array of shape `(#channels, time_len)` in micro-Volt units.
Some general info can be inspected with `raw.info`, such as the sampling rate (`raw.info["sfreq"]`).
The channel names (in their correct order) can be seen via `raw.ch_names`. Note that `X`, `Y` and `Z` denote the accelerometer X-, Y- and Z-axis outputs.
Events can be read with
```python
events_list, events_dict = mne.events_from_annotations(raw)
```
where `events_dict` contains the mapping of the original event types (where `"S 1"` stands for the black smaller circle "warning" cue, and `"S 2"` stands for the green bigger circle "GO" cue) to event IDs in `[1,2]`, the latter of which are used in `events_list`.
`events_list` is a list of events, ordered by time. Each entry `e = [timestamp, (not important), event ID]` consists of the time of the event onset `timestamp` that refers to the `time_len` dimension in the `raw.get_data()` EEG array, as well as the event-ID.
(See the `https://mne.tools/stable/generated/mne.io.Raw.html` documentation for more details.)
### Reading Processed Files in Python
In python, the `[GROUP][PID]_Ped_Processed.mat` files can be read e.g. with the `mat73` package (to read MATLAB v7.3 files in HDF5 format):
```python
import mat73
mat = mat73.loadmat("path_to/[GROUP][PID]_Ped_Processed.mat")
```
Then `mat` contains (among others) the following fields and subfields
- `EEG`
- `data`: EEG data of shape `(#channels, trial_len, #trials)`. E.g. a recording of 50 trials/epochs with 59 channels, each trial having a duration of 4 seconds and a sampling rate of 500 Hz will have shape `(59, 2000, 50)`. Recall that here, trials were epoched from -1000ms to +3000ms around the "GO" cue event.
- `event`: Contains a list of dictionaries, each entry (each event) having the following description:
- `latency`: The onset of the event, measured as the index in the merged time-dimension `#trials x trial_len` (note `#trials` being the _outer_ and `trial_len` being the _inner_ array when merging).
- `type`: The type of event. Here, only `"S 2"` (larger green circle "GO" cue) events are stored.
- `TimeDiff`: The time in seconds that has passed between the `"S 1"` and the `"S 2"` cue.
- `chanlocs`: A list of channel descriptors
- `nbchan`: Number of channels
- `trials`: Number of trials/epochs in this recording
- `srate`: Sampling Rate (Hz)
Additionally, the field and `bad_chans` lists bad channels of this recording.
## License
By the original authors of this work, this work has been licensed under the PDDL v1.0 license (see LICENSE.txt).