Upload BNCI2014_001 RawDataset
Browse files- .gitattributes +10 -0
- README.md +71 -0
- dataset.zarr/recording_0/data/c/0/0 +3 -0
- dataset.zarr/recording_0/data/c/0/1 +3 -0
- dataset.zarr/recording_0/data/c/0/2 +3 -0
- dataset.zarr/recording_0/data/c/0/3 +3 -0
- dataset.zarr/recording_0/data/c/0/4 +3 -0
- dataset.zarr/recording_0/data/c/0/5 +3 -0
- dataset.zarr/recording_0/data/c/0/6 +3 -0
- dataset.zarr/recording_0/data/c/0/7 +3 -0
- dataset.zarr/recording_0/data/c/0/8 +3 -0
- dataset.zarr/recording_0/data/c/0/9 +3 -0
- dataset.zarr/recording_0/data/zarr.json +45 -0
- dataset.zarr/recording_0/zarr.json +9 -0
- dataset.zarr/zarr.json +10 -0
- format_info.json +9 -0
.gitattributes
CHANGED
|
@@ -57,3 +57,13 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 57 |
# Video files - compressed
|
| 58 |
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 59 |
*.webm filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 57 |
# Video files - compressed
|
| 58 |
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 59 |
*.webm filter=lfs diff=lfs merge=lfs -text
|
| 60 |
+
dataset.zarr/recording_0/data/c/0/0 filter=lfs diff=lfs merge=lfs -text
|
| 61 |
+
dataset.zarr/recording_0/data/c/0/1 filter=lfs diff=lfs merge=lfs -text
|
| 62 |
+
dataset.zarr/recording_0/data/c/0/2 filter=lfs diff=lfs merge=lfs -text
|
| 63 |
+
dataset.zarr/recording_0/data/c/0/3 filter=lfs diff=lfs merge=lfs -text
|
| 64 |
+
dataset.zarr/recording_0/data/c/0/4 filter=lfs diff=lfs merge=lfs -text
|
| 65 |
+
dataset.zarr/recording_0/data/c/0/5 filter=lfs diff=lfs merge=lfs -text
|
| 66 |
+
dataset.zarr/recording_0/data/c/0/6 filter=lfs diff=lfs merge=lfs -text
|
| 67 |
+
dataset.zarr/recording_0/data/c/0/7 filter=lfs diff=lfs merge=lfs -text
|
| 68 |
+
dataset.zarr/recording_0/data/c/0/8 filter=lfs diff=lfs merge=lfs -text
|
| 69 |
+
dataset.zarr/recording_0/data/c/0/9 filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
|
@@ -0,0 +1,71 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
tags:
|
| 3 |
+
- braindecode
|
| 4 |
+
- eeg
|
| 5 |
+
- neuroscience
|
| 6 |
+
- brain-computer-interface
|
| 7 |
+
license: unknown
|
| 8 |
+
---
|
| 9 |
+
|
| 10 |
+
# EEG Dataset
|
| 11 |
+
|
| 12 |
+
This dataset was created using [braindecode](https://braindecode.org), a library for deep learning with EEG/MEG/ECoG signals.
|
| 13 |
+
|
| 14 |
+
## Dataset Information
|
| 15 |
+
|
| 16 |
+
| Property | Value |
|
| 17 |
+
|---|---:|
|
| 18 |
+
| Number of recordings | 1 |
|
| 19 |
+
| Dataset type | Continuous (Raw) |
|
| 20 |
+
| Number of channels | 26 |
|
| 21 |
+
| Sampling frequency | 250 Hz |
|
| 22 |
+
| Number of windows / samples | 96735 |
|
| 23 |
+
| Total size | 19.22 MB |
|
| 24 |
+
| Storage format | zarr |
|
| 25 |
+
|
| 26 |
+
## Usage
|
| 27 |
+
|
| 28 |
+
To load this dataset::
|
| 29 |
+
|
| 30 |
+
.. code-block:: python
|
| 31 |
+
|
| 32 |
+
from braindecode.datasets import BaseConcatDataset
|
| 33 |
+
|
| 34 |
+
# Load dataset from Hugging Face Hub
|
| 35 |
+
dataset = BaseConcatDataset.from_pretrained("username/dataset-name")
|
| 36 |
+
|
| 37 |
+
# Access data
|
| 38 |
+
X, y, metainfo = dataset[0]
|
| 39 |
+
# X: EEG data (n_channels, n_times)
|
| 40 |
+
# y: label/target
|
| 41 |
+
# metainfo: window indices
|
| 42 |
+
|
| 43 |
+
## Using with PyTorch DataLoader
|
| 44 |
+
|
| 45 |
+
::
|
| 46 |
+
|
| 47 |
+
from torch.utils.data import DataLoader
|
| 48 |
+
|
| 49 |
+
# Create DataLoader for training
|
| 50 |
+
train_loader = DataLoader(
|
| 51 |
+
dataset,
|
| 52 |
+
batch_size=32,
|
| 53 |
+
shuffle=True,
|
| 54 |
+
num_workers=4
|
| 55 |
+
)
|
| 56 |
+
|
| 57 |
+
# Training loop
|
| 58 |
+
for X, y, metainfo in train_loader:
|
| 59 |
+
# X shape: [batch_size, n_channels, n_times]
|
| 60 |
+
# y shape: [batch_size]
|
| 61 |
+
# metainfo shape: [batch_size, 2] (start and end indices)
|
| 62 |
+
# Process your batch...
|
| 63 |
+
|
| 64 |
+
## Dataset Format
|
| 65 |
+
|
| 66 |
+
This dataset is stored in **Zarr** format, optimized for:
|
| 67 |
+
- Fast random access during training (critical for PyTorch DataLoader)
|
| 68 |
+
- Efficient compression with blosc
|
| 69 |
+
- Cloud-native storage compatibility
|
| 70 |
+
|
| 71 |
+
For more information about braindecode, visit: https://braindecode.org
|
dataset.zarr/recording_0/data/c/0/0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5252c8158e65a01d012077cc8bcd32a3437f8bf8c57a04ad481b5cd0ff9d2844
|
| 3 |
+
size 772956
|
dataset.zarr/recording_0/data/c/0/1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fa9d4044f778e149b99d7fdf36c74c8beb592da9d6cfca4820c7cc3bc25989a3
|
| 3 |
+
size 772139
|
dataset.zarr/recording_0/data/c/0/2
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a65d165176a2a79de76c7a61f67a6449e4af94159364625ea7222a9eea0e73a9
|
| 3 |
+
size 772348
|
dataset.zarr/recording_0/data/c/0/3
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:24365d69d1564b6752dc48b9f6e918e8a9070a09e4d89ba7a3532ad840f52a06
|
| 3 |
+
size 774500
|
dataset.zarr/recording_0/data/c/0/4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9e82e5adea4c1eb2c353f439667bc66f25e62061383d45a123929e039892e4b2
|
| 3 |
+
size 778745
|
dataset.zarr/recording_0/data/c/0/5
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e1205f9a26dc6b754e7a2634894a61fe94c530024a219469cc8fece5c6dc283
|
| 3 |
+
size 776102
|
dataset.zarr/recording_0/data/c/0/6
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f2d3ebcce273d3f4c9048a85f83ec6f9c6c9b881609053746f54cc5cb631a63f
|
| 3 |
+
size 774672
|
dataset.zarr/recording_0/data/c/0/7
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6f0a52d770b1e4794ef6ba47f1ea22036f2412861d85ff465668c3748f10d3ac
|
| 3 |
+
size 776431
|
dataset.zarr/recording_0/data/c/0/8
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a3be20c058630b49a4a55546679bc61a7aaff7f2620aaefee74974557c36ee39
|
| 3 |
+
size 775441
|
dataset.zarr/recording_0/data/c/0/9
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f419dcd371e62c4a7f5f8760b96856514a270094240f297cac4848ec3698f2b6
|
| 3 |
+
size 527352
|
dataset.zarr/recording_0/data/zarr.json
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"shape": [
|
| 3 |
+
26,
|
| 4 |
+
96735
|
| 5 |
+
],
|
| 6 |
+
"data_type": "float32",
|
| 7 |
+
"chunk_grid": {
|
| 8 |
+
"name": "regular",
|
| 9 |
+
"configuration": {
|
| 10 |
+
"chunk_shape": [
|
| 11 |
+
26,
|
| 12 |
+
10000
|
| 13 |
+
]
|
| 14 |
+
}
|
| 15 |
+
},
|
| 16 |
+
"chunk_key_encoding": {
|
| 17 |
+
"name": "default",
|
| 18 |
+
"configuration": {
|
| 19 |
+
"separator": "/"
|
| 20 |
+
}
|
| 21 |
+
},
|
| 22 |
+
"fill_value": 0.0,
|
| 23 |
+
"codecs": [
|
| 24 |
+
{
|
| 25 |
+
"name": "bytes",
|
| 26 |
+
"configuration": {
|
| 27 |
+
"endian": "little"
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
{
|
| 31 |
+
"name": "blosc",
|
| 32 |
+
"configuration": {
|
| 33 |
+
"typesize": 4,
|
| 34 |
+
"cname": "zstd",
|
| 35 |
+
"clevel": 5,
|
| 36 |
+
"shuffle": "shuffle",
|
| 37 |
+
"blocksize": 0
|
| 38 |
+
}
|
| 39 |
+
}
|
| 40 |
+
],
|
| 41 |
+
"attributes": {},
|
| 42 |
+
"zarr_format": 3,
|
| 43 |
+
"node_type": "array",
|
| 44 |
+
"storage_transformers": []
|
| 45 |
+
}
|
dataset.zarr/recording_0/zarr.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"attributes": {
|
| 3 |
+
"description": "{\"subject\":1,\"session\":\"0train\",\"run\":\"0\"}",
|
| 4 |
+
"info": "{\"ch_names\": [\"Fz\", \"FC3\", \"FC1\", \"FCz\", \"FC2\", \"FC4\", \"C5\", \"C3\", \"C1\", \"Cz\", \"C2\", \"C4\", \"C6\", \"CP3\", \"CP1\", \"CPz\", \"CP2\", \"CP4\", \"P1\", \"Pz\", \"P2\", \"POz\", \"EOG1\", \"EOG2\", \"EOG3\", \"stim\"], \"sfreq\": 250.0, \"ch_types\": [\"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eeg\", \"eog\", \"eog\", \"eog\", \"stim\"], \"lowpass\": 125.0, \"highpass\": 0.0}"
|
| 5 |
+
},
|
| 6 |
+
"zarr_format": 3,
|
| 7 |
+
"consolidated_metadata": null,
|
| 8 |
+
"node_type": "group"
|
| 9 |
+
}
|
dataset.zarr/zarr.json
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"attributes": {
|
| 3 |
+
"n_datasets": 1,
|
| 4 |
+
"dataset_type": "RawDataset",
|
| 5 |
+
"braindecode_version": "1.3.0"
|
| 6 |
+
},
|
| 7 |
+
"zarr_format": 3,
|
| 8 |
+
"consolidated_metadata": null,
|
| 9 |
+
"node_type": "group"
|
| 10 |
+
}
|
format_info.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"format": "zarr",
|
| 3 |
+
"compression": "blosc",
|
| 4 |
+
"compression_level": 5,
|
| 5 |
+
"braindecode_version": "1.3.0",
|
| 6 |
+
"n_recordings": 1,
|
| 7 |
+
"total_samples": 96735,
|
| 8 |
+
"total_size_mb": 19.22
|
| 9 |
+
}
|