--- license: pddl tags: - eeg - medical - clinical - classification - parkinson - pedaling --- # Singh2020: EEG Parkinson's Classification Dataset with Pedaling Task The Singh2020 dataset contains EEG recordings collected during a lower-limb pedaling task designed to assess motor control in individuals with Parkinson's disease (PD), with a particular focus on freezing of gait (FOG) symptoms. A total of 39 participants were included: 13 PD patients with FOG (PDFOG+), 13 PD patients without FOG (PDFOG-), and 13 demographically matched healthy controls. Participants completed a lower-limb motor task in which they pedaled one rotation in response to a visual "GO" cue, designed to minimize fall risk and reduce EEG artifacts from movement. Each subject completed at least two blocks of either 30 or 50 trials, with PDFOG+ participants performing fewer trials due to symptom severity. A tri-axial accelerometer mounted on the ankle measured pedaling kinematics. EEG was recorded using a 64-channel cap with a sampling rate of 500 Hz. ## Paper Singh, A., Cole, R. C., Espinoza, A. I., Brown, D., Cavanagh, J. F., & Narayanan, N. S. (2020). **Frontal theta and beta oscillations during lower-limb movement in Parkinson’s disease**. _Clinical Neurophysiology_, 131(3), 694-702. DISCLAIMER: We (DISCO) are NOT the owners or creators of this dataset, but we merely uploaded it here, to support our's ([EEG-Bench](https://github.com/ETH-DISCO/EEG-Bench)) and other's work on EEG benchmarking. ## Dataset Structure - `raw_data/` contains the unprocessed recording EEG and accelerometer data. - `processed_data/` contains pre- (`_Ped_Processed.mat`) and post-processed (`_ANALYZED.mat`) experiment data, produced from the `RAW_DATA/` using `Step1_EEG_Pedaling_PreProcess.m` and `Step2_EEG_Pedaling_PostProcess.m` scripts, respectively. - `ALL_data_Modeling.csv` contains participant information, such as age and MOCA test scores. - `scripts/` contains various scripts that can be used to reproduce the results from the paper. - `ORIGINAL_README.txt`: the original README with perhaps some more helpful information. ### Filename Format In `raw_data/`, a recording consists of 3 files: - `[GROUP][PID].vhdr`: The header file with meta information - `[GROUP][PID].eeg`: contains the EEG (and accelerometer) data - `[GROUP][PID].vmrk`: contains event information where GROUP is either `Control` or `PD` (Parkinson's Disease) and PID is the participant's ID (also used in `ALL_data_Modeling.csv`). Similarly, in `processed_data/`, - `[GROUP][PID]_Ped_Processed.mat` contains pre-processed data epoched around trials - `[GROUP][PID]_ANALYZED.mat` contains the time-frequency analysis of the pre-processed data where, again, GROUP is either `Control` or `PD` (Parkinson's Disease) and PID is the participant's ID. ### Reading RAW_DATA Files in Python In python, the 3 files that make up a raw recording can be read via: ```python import mne raw = mne.io.read_raw_brainvision("path_to/[GROUP][PID].vhdr") ``` Now, `raw.get_data(units='uV')` yields a numpy array of shape `(#channels, time_len)` in micro-Volt units. Some general info can be inspected with `raw.info`, such as the sampling rate (`raw.info["sfreq"]`). The channel names (in their correct order) can be seen via `raw.ch_names`. Note that `X`, `Y` and `Z` denote the accelerometer X-, Y- and Z-axis outputs. Events can be read with ```python events_list, events_dict = mne.events_from_annotations(raw) ``` where `events_dict` contains the mapping of the original event types (where `"S 1"` stands for the black smaller circle "warning" cue, and `"S 2"` stands for the green bigger circle "GO" cue) to event IDs in `[1,2]`, the latter of which are used in `events_list`. `events_list` is a list of events, ordered by time. Each entry `e = [timestamp, (not important), event ID]` consists of the time of the event onset `timestamp` that refers to the `time_len` dimension in the `raw.get_data()` EEG array, as well as the event-ID. (See the `https://mne.tools/stable/generated/mne.io.Raw.html` documentation for more details.) ### Reading Processed Files in Python In python, the `[GROUP][PID]_Ped_Processed.mat` files can be read e.g. with the `mat73` package (to read MATLAB v7.3 files in HDF5 format): ```python import mat73 mat = mat73.loadmat("path_to/[GROUP][PID]_Ped_Processed.mat") ``` Then `mat` contains (among others) the following fields and subfields - `EEG` - `data`: EEG data of shape `(#channels, trial_len, #trials)`. E.g. a recording of 50 trials/epochs with 59 channels, each trial having a duration of 4 seconds and a sampling rate of 500 Hz will have shape `(59, 2000, 50)`. Recall that here, trials were epoched from -1000ms to +3000ms around the "GO" cue event. - `event`: Contains a list of dictionaries, each entry (each event) having the following description: - `latency`: The onset of the event, measured as the index in the merged time-dimension `#trials x trial_len` (note `#trials` being the _outer_ and `trial_len` being the _inner_ array when merging). - `type`: The type of event. Here, only `"S 2"` (larger green circle "GO" cue) events are stored. - `TimeDiff`: The time in seconds that has passed between the `"S 1"` and the `"S 2"` cue. - `chanlocs`: A list of channel descriptors - `nbchan`: Number of channels - `trials`: Number of trials/epochs in this recording - `srate`: Sampling Rate (Hz) Additionally, the field and `bad_chans` lists bad channels of this recording. ## License By the original authors of this work, this work has been licensed under the PDDL v1.0 license (see LICENSE.txt).