The dataset viewer is not available because its heuristics could not detect any supported data files. You can try uploading some data files, or configuring the data files location manually.
Gym288-skeleton Dataset
License: CC-BY-4.0
Overview
The Gym288-skeleton dataset is a human skeleton-based action recognition benchmark derived from the Gym288 subset of the FineGym dataset. It provides temporally precise, fine-grained annotations of gymnastic actions along with 2D human pose sequences extracted from original video frames.
This dataset is designed to support research in:
- Fine-grained action recognition
- Temporally corrupted or incomplete action modeling
- Skeleton-based representation learning
- Physics-aware motion understanding
The dataset was introduced and used in the paper "FineTec: Fine-grained Action Recognition under Temporal Corruption", which has been accepted to AAAI 2026. In this work, the dataset serves as the primary evaluation benchmark for recognizing fine-grained actions from temporally corrupted skeleton sequences.
Dataset Structure
The dataset is distributed as a single Python dictionary with two top-level keys: split and annotations.
Top-Level Keys
split: Dictionary containing train/test splits.train: List of 28,739 sample IDs (strings)test: List of 9,484 sample IDs (strings)
annotations: List of 38,223 dictionaries, each representing one action instance with the following fields:
| Key | Type | Shape / Example | Description |
|---|---|---|---|
frame_dir |
str |
"A0xAXXysHUo_002184_002237_0035_0036" |
Unique identifier for the action clip |
label |
int |
268 |
Class label (0–287, corresponding to 288 fine-grained gymnastic elements) |
img_shape |
tuple |
(720, 1280) |
Height and width of original video frames |
original_shape |
tuple |
(720, 1280) |
Same as img_shape (for compatibility) |
total_frames |
int |
48 |
Number of frames in the action sequence |
keypoint |
np.ndarray (float16) |
(1, T, 17, 2) |
2D joint coordinates (x, y) for 17 COCO-style keypoints over T frames |
keypoint_score |
np.ndarray (float16) |
(1, T, 17) |
Confidence scores for each keypoint |
kp_wo_gt |
np.ndarray (float32) |
(T, 17, 3) |
Placeholder array (all zeros); originally intended for corrupted/noisy poses without ground truth |
kp_w_gt |
np.ndarray (float32) |
(T, 17, 3) |
Ground-truth 2D poses with confidence as third channel (x, y, score) |
Note: The first dimension (
1) inkeypointandkeypoint_scorecorresponds to the number of persons (always 1 in this dataset).
Action Classes
The dataset contains 288 distinct gymnastic elements across four apparatuses:
- Floor Exercise (FX)
- Balance Beam (BB)
- Uneven Bars (UB)
- Vault – Women (VT)
Each class represents a highly specific movement (e.g., "Switch leap with 0.5 turn", "Clear hip circle backward with 1 turn to handstand"), reflecting the fine-grained nature of competitive gymnastics scoring.
For the full list of class names and mappings, please refer to the website and paper of FineGym.
Usage Example
import numpy as np
# Load the dataset (e.g., using pickle or torch.load)
with open("gym288_skeleton.pkl", "rb") as f:
data = pickle.load(f)
# Access training samples
train_ids = data["split"]["train"] # list of strings
# Access annotations
sample = data["annotations"][0]
print("Label:", sample["label"])
print("Frames:", sample["total_frames"])
print("Keypoints shape:", sample["keypoint"].shape) # (1, T, 17, 2)
# Extract skeleton sequence for model input
skeleton_seq = sample["keypoint"][0] # (T, 17, 2)
Citation
If you use this dataset in your research, please cite both the FineTec and FineGym papers. FineTec's citation information will be updated upon publication.
@inproceedings{shao2020finegym,
title={FineGym: A Hierarchical Video Dataset for Fine-grained Action Understanding},
author={Shao, Dian and Zhao, Yue and Dai, Bo and Lin, Dahua},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
pages={2616--2625},
year={2020}
}
License
This dataset is licensed under Creative Commons Attribution 4.0 International (CC-BY-4.0).
You are free to share and adapt the material, even commercially, as long as appropriate credit is given.
Note: The underlying video data remains the property of its original sources (e.g., YouTube). This dataset only distributes extracted pose annotations, not raw videos.
Acknowledgements
- Skeletons were extracted using pose estimators HRNet on the FineGym video corpus.
- We thank the authors of FineGym for their foundational work in fine-grained action recognition.
- Downloads last month
- 4