HusseinLezzaik commited on
Commit
b8c2bd0
·
verified ·
1 Parent(s): 59d8a82

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +134 -0
README.md ADDED
@@ -0,0 +1,134 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ dataset_info:
3
+ features:
4
+ - name: id
5
+ dtype: string
6
+ - name: game
7
+ dtype: string
8
+ - name: trial_id
9
+ dtype: int32
10
+ - name: episode_id
11
+ dtype: int32
12
+ - name: chunk_idx
13
+ dtype: int32
14
+ - name: frame_start
15
+ dtype: int32
16
+ - name: action
17
+ dtype: string
18
+ - name: action_ints
19
+ dtype: string
20
+ - name: score
21
+ dtype: int32
22
+ - name: reward_sum
23
+ dtype: int32
24
+ - name: gaze_positions
25
+ dtype: string
26
+ - name: image_bytes
27
+ dtype: binary
28
+ license: mit
29
+ task_categories:
30
+ - robotics
31
+ - reinforcement-learning
32
+ tags:
33
+ - atari
34
+ - vla
35
+ - vision-language-action
36
+ - imitation-learning
37
+ - human-demonstrations
38
+ - action-chunking
39
+ size_categories:
40
+ - 1M<n<10M
41
+ ---
42
+
43
+ # TESS-Atari Stage 1 (15Hz)
44
+
45
+ Human gameplay demonstrations from Atari games with **action chunking**, formatted for Vision-Language-Action (VLA) model training.
46
+
47
+ ## Overview
48
+
49
+ | Metric | Value |
50
+ |--------|-------|
51
+ | Source | [Atari-HEAD](https://zenodo.org/records/3451402) |
52
+ | Games | 11 (overlapping with DIAMOND benchmark) |
53
+ | Samples | ~1.3M |
54
+ | Observation Rate | 5 Hz |
55
+ | Action Rate | 15 Hz (3 actions per observation) |
56
+ | Format | Lumine-style action tokens |
57
+
58
+ ## Why Action Chunking?
59
+
60
+ VLA models run at ~5 Hz inference speed, but Atari runs at 15 Hz (with frame_skip=4). Action chunking predicts 3 actions at once, matching the game's effective action rate while accommodating slower model inference.
61
+
62
+ ```
63
+ Observation (5 Hz) → VLA → 3 Actions (executed at 15 Hz)
64
+ ```
65
+
66
+ ## Games Included
67
+
68
+ Alien, Asterix, BankHeist, Breakout, DemonAttack, Freeway, Frostbite, Hero, MsPacman, RoadRunner, Seaquest
69
+
70
+ ## Action Format
71
+
72
+ ```
73
+ <|action_start|> RIGHT ; RIGHT ; FIRE <|action_end|>
74
+ <|action_start|> LEFT ; LEFT ; LEFT <|action_end|>
75
+ <|action_start|> NOOP ; UP ; UPFIRE <|action_end|>
76
+ ```
77
+
78
+ ## Schema
79
+
80
+ | Field | Type | Description |
81
+ |-------|------|-------------|
82
+ | `id` | string | Unique sample ID: `{game}_{trial}_{chunk}` |
83
+ | `game` | string | Game name (lowercase) |
84
+ | `trial_id` | int | Human player trial number |
85
+ | `episode_id` | int | Episode within trial (-1 if unknown) |
86
+ | `chunk_idx` | int | Chunk sequence number |
87
+ | `frame_start` | int | First frame index of this chunk |
88
+ | `action` | string | Lumine-style chunked action token |
89
+ | `action_ints` | string | Raw ALE codes comma-separated: "4,4,1" |
90
+ | `score` | int | Score at chunk start |
91
+ | `reward_sum` | int | Total reward over 3 frames |
92
+ | `gaze_positions` | string | Eye tracking from first frame |
93
+ | `image_bytes` | bytes | PNG of first frame in chunk |
94
+
95
+ ## Usage
96
+
97
+ ```python
98
+ from datasets import load_dataset
99
+
100
+ ds = load_dataset("TESS-Computer/atari-vla-stage1-15hz")
101
+
102
+ # Get a sample
103
+ sample = ds["train"][0]
104
+ print(sample["action"]) # <|action_start|> RIGHT ; RIGHT ; FIRE <|action_end|>
105
+
106
+ # Parse individual actions
107
+ actions = sample["action_ints"].split(",") # ["4", "4", "1"]
108
+
109
+ # Decode image
110
+ from PIL import Image
111
+ from io import BytesIO
112
+ img = Image.open(BytesIO(sample["image_bytes"]))
113
+ ```
114
+
115
+ ## Evaluation
116
+
117
+ Use with [DIAMOND](https://diamond-wm.github.io/) world models (frame_skip=4). Execute the 3 predicted actions sequentially at each observation step.
118
+
119
+ ## Related
120
+
121
+ - [5Hz variant](https://huggingface.co/datasets/TESS-Computer/atari-vla-stage1-5hz) - Single action per observation (simpler but slower)
122
+ - [Lumine AI](https://www.lumine-ai.org/) - Inspiration for VLA architecture and action chunking
123
+ - [DIAMOND](https://diamond-wm.github.io/) - World model for evaluation
124
+
125
+ ## Citation
126
+
127
+ ```bibtex
128
+ @misc{atarihead2019,
129
+ title={Atari-HEAD: Atari Human Eye-Tracking and Demonstration Dataset},
130
+ author={Zhang, Ruohan and others},
131
+ year={2019},
132
+ url={https://zenodo.org/records/3451402}
133
+ }
134
+ ```