cheedeh-IMU-data / README.md
Krzysztof Kołodziński
fixed category
855dcb7
---
license: mit
task_categories:
- other
pretty_name: Cheedeh IMU Gesture Data
size_categories:
- n<1K
configs:
- config_name: default
data_files:
- split: train
path: train/*.json
- split: test
path: test/*.json
---
# Cheedeh IMU Gesture Data
Phone IMU recordings of mid-air gestures drawn in the air with the phone. Collected to train a real-time gesture classifier running on Android.
## Gestures (labels)
| Label | Gesture |
|--------|---------------------|
| `z` | Letter Z |
| `m` | Letter M |
| `s` | Letter S |
| `o` | Letter O (circle) |
| `none` | No gesture / idle |
## Splits
| Split | Samples | m | none | o | s | z |
|-------|---------|-----|------|----|----|----|
| train | 266 | 31 | 111 | 43 | 42 | 39 |
| test | 54 | 15 | 20 | 8 | 4 | 7 |
## Data Format
Each sample is a separate JSON file named `gesture_{label}_{YYYYMMDD}_{HHMMSS}.json`.
```json
{
"gesture_id": "m_20260210_160623",
"gesture_name": "m",
"timestamp": "2026-02-10T15:06:23.660296Z",
"duration_ms": 3246,
"sample_rate_hz": 57,
"data": [
{
"t": 1,
"x": -0.072, "y": -0.143, "z": 0.456,
"gx": -0.011, "gy": -0.043, "gz": 0.020,
"grx": 0.522, "gry": 4.386, "grz": 8.759,
"qx": 0.152, "qy": -0.175, "qz": -0.653, "qw": 0.721
},
...
]
}
```
### Fields in `data` array
| Field | Sensor | Unit |
|--------------|-------------------------------|--------|
| `t` | Elapsed time | ms |
| `x`, `y`, `z` | Linear acceleration (`TYPE_LINEAR_ACCELERATION`) | m/s² |
| `gx`, `gy`, `gz` | Gyroscope (angular velocity) | rad/s |
| `grx`, `gry`, `grz` | Gravity vector (rotation frame) | m/s² |
| `qx`, `qy`, `qz`, `qw` | Orientation quaternion | — |
Sample rate is approximately 50–57 Hz. Duration varies by gesture (typically 1–4 s).
## Loading
```python
from datasets import load_dataset
ds = load_dataset("ravenwing/cheedeh-IMU-data")
sample = ds["train"][0]
# sample["gesture_name"] → "m"
# sample["data"] → list of dicts with t, x, y, z, ...
```
## Collection
Recorded with an Android app ([gather-data](https://github.com/ravenwing/cheedeh)) using `TYPE_LINEAR_ACCELERATION` and gyroscope sensors. Each session was a manual recording of one gesture, labeled at collection time.
## Related
- Training code: [learn/](https://github.com/ravenwing/cheedeh/tree/main/learn) — SVM/RF classifier achieving ~0.76 test accuracy, exported to ONNX for on-device inference.