File size: 2,398 Bytes
244a141 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 | ---
pretty_name: "BigP3BCI Study N — 9x8 dry/wet electrode comparison (8 ALS subjects)"
license: cc-by-4.0
tags:
- eeg
- neuroscience
- eegdash
- brain-computer-interface
- pytorch
- visual
- attention
- other
size_categories:
- n<1K
task_categories:
- other
---
# BigP3BCI Study N — 9x8 dry/wet electrode comparison (8 ALS subjects)
**Dataset ID:** `nm000187`
_Mainsah2025_BigP3BCI_N_
**Canonical aliases:** `BigP3BCI_StudyN`
> **At a glance:** EEG · Visual attention · other · 8 subjects · 160 recordings · CC-BY-4.0
## Load this dataset
This repo is a **pointer**. The raw EEG data lives at its canonical source
(OpenNeuro / NEMAR); [EEGDash](https://github.com/eegdash/EEGDash) streams it
on demand and returns a PyTorch / braindecode dataset.
```python
# pip install eegdash
from eegdash import EEGDashDataset
ds = EEGDashDataset(dataset="nm000187", cache_dir="./cache")
print(len(ds), "recordings")
```
You can also load it by canonical alias — these are registered classes in `eegdash.dataset`:
```python
from eegdash.dataset import BigP3BCI_StudyN
ds = BigP3BCI_StudyN(cache_dir="./cache")
```
If the dataset has been mirrored to the HF Hub in braindecode's Zarr layout,
you can also pull it directly:
```python
from braindecode.datasets import BaseConcatDataset
ds = BaseConcatDataset.pull_from_hub("EEGDash/nm000187")
```
## Dataset metadata
| | |
|---|---|
| **Subjects** | 8 |
| **Recordings** | 160 |
| **Tasks (count)** | 1 |
| **Channels** | 16 (×160) |
| **Sampling rate (Hz)** | 256 (×160) |
| **Total duration (h)** | 8.2 |
| **Size on disk** | 353.2 MB |
| **Recording type** | EEG |
| **Experimental modality** | Visual |
| **Paradigm type** | Attention |
| **Population** | Other |
| **Source** | nemar |
| **License** | CC-BY-4.0 |
## Links
- **NEMAR:** [nm000187](https://nemar.org/dataexplorer/detail?dataset_id=nm000187)
- **Browse 700+ datasets:** [EEGDash catalog](https://huggingface.co/spaces/EEGDash/catalog)
- **Docs:** <https://eegdash.org>
- **Code:** <https://github.com/eegdash/EEGDash>
---
_Auto-generated from [dataset_summary.csv](https://github.com/eegdash/EEGDash/blob/main/eegdash/dataset/dataset_summary.csv) and the [EEGDash API](https://data.eegdash.org/api/eegdash/datasets/summary/nm000187). Do not edit this file by hand — update the upstream source and re-run `scripts/push_metadata_stubs.py`._
|