File size: 8,322 Bytes
46a2a44
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d7bd929
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
---
dataset_info:
  features:
  - name: video_id
    dtype: string
  - name: title
    dtype: string
  - name: os
    dtype: string
  - name: num_scenes
    dtype: int64
  - name: scene_timestamps_in_sec
    sequence: float64
  - name: screen_bboxes
    sequence:
      sequence: int64
  - name: ui_element_bboxes
    sequence:
      sequence:
        sequence: float64
  - name: raw_actions
    list:
      list:
      - name: box_id
        dtype: int64
      - name: details
        dtype: string
      - name: type
        dtype: string
  - name: actions
    list:
      list:
      - name: action_type_id
        dtype: int64
      - name: action_type_text
        dtype: string
      - name: annot_position
        sequence: float64
      - name: lift
        sequence: float64
      - name: touch
        sequence: float64
      - name: type_text
        dtype: string
  - name: video_fps
    dtype: float64
  - name: video_width
    dtype: int64
  - name: video_height
    dtype: int64
  splits:
  - name: train
    num_bytes: 69622260
    num_examples: 19725
  - name: validation
    num_bytes: 1641036
    num_examples: 495
  - name: test
    num_bytes: 565401
    num_examples: 100
  - name: test_unseen_os
    num_bytes: 169823
    num_examples: 50
  download_size: 18085770
  dataset_size: 71998520
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
  - split: validation
    path: data/validation-*
  - split: test
    path: data/test-*
  - split: test_unseen_os
    path: data/test_unseen_os-*
---

[Paper](https://arxiv.org/abs/2505.12632) |
[Code](https://github.com/runamu/monday) |
[Dataset](https://huggingface.co/datasets/runamu/MONDAY) |
[Project](https://monday-dataset.github.io)

# Dataset Card for MONDAY

_MONDAY_ (Mobile OS Navigation Task Dataset for Agents from YouTube) is a cross-platform mobile navigation dataset for training vision-language models. This dataset contains
- **20K** curated list of videos of mobile navigation tasks from YouTube, including Android and iOS devices.
- **333K** detected scenes, each representing a temporally segmented step within a mobile navigation task.
- **313K** identified actions, including touch, scroll, hardware, typing, long press, multi touch and zoom.

Please visit our [project page](https://monday-dataset.github.io/) for more details.

## Data Fields
- **video_id (str)**: Unique identifier for the video.
- **title (str)**: Title of the video.
- **os (str)**: Operating system of the mobile device used in the video.
- **num_scenes (int)**: Number of detected scenes in the video.
- **scene_timestamps_in_sec (list)**: A list of timestamps of the detected scenes in seconds. The list has a length of `num_scenes`.
- **screen_bboxes (list)**: A list of bounding boxes for the detected phone screen in each scene, given as (left, top, right, bottom) pixel coordinates. The list has a length of `num_scenes`.
- **ui_element_bboxes (list)**: A list of bounding boxes for the detected user interface (UI) elements in each scene, given as (left, top, right, bottom) coordinates normalized to the [0, 1] range. The list has a length of `num_scenes - 1`.
  ```python
  # example
  ui_element_bboxes = [
      [ui_bbox1_scene1, ui_bbox2_scene1, ...], # UI elements in scene 1
      [ui_bbox1_scene2, ui_bbox2_scene2, ...], # UI elements in scene 2
      ...
  ]
  ```
- **raw_actions (list)**: A list of raw actions identified from the video for each scene. The list has a length of `num_scenes - 1`. Multiple actions can be annotated within a single scene, and all are considered valid. Each element is a list of actions annotated in that scene, with each action represented as a dictionary containing the following keys:
  - **box_id (int)**: The index of the UI element's bounding box (from `ui_element_bboxes[scene_id]`) associated with the action. If the action does not correspond to any UI element, the value is -1.
  - **details (str)**: A detailed description of the action, either automatically generated or manually annotated during the identification process.
  - **type (str)**: A text label describing the action type.  Possible values include `"touch"`, `"scroll"`, `"hardware"`, `"typing"`, `"long press"`, `"multi touch"` and `"zoom"`, listed in order of frequency in the dataset.
  ```python
  # example
  raw_actions = [
    [
        {"box_id": 0, "details": "...", "type": "touch"}, # First action in scene 1
        {"box_id": 1, "details": "...", "type": "touch"}, # Second action in scene 1
    ],
    [
        {"box_id": -1, "details": "...", "type": "typing"}, # First action in scene 2
    ],
    ...
  ]
  ```
  Note: The `box_id` is -1 for actions that do not correspond to any UI element.

- **actions (list)**: A list of actions in each scene, processed for mobile navigation agent training and evaluation. The list has a length of `num_scenes - 1`. Multiple actions can be annotated within a single scene, and all are considered valid. Each element is a list of actions annotated in that scene, with each action represented as a dictionary containing the following keys:
  - **action_type_id (int)**: An integer identifier for the action type, based on the action type sets used in [SeeClick](https://github.com/njucckevin/SeeClick/blob/main/agent_tasks/action_type.py) and [AitW](https://github.com/google-research/google-research/blob/master/android_in_the_wild/action_type.py).
  - **action_type_text (str)**: A text label describing the action type.  Possible values include `"click"`, `"scroll down"`, `"press home"`, `"type"`, `"scroll up"`, `"other hardware"`, `"scroll left"`, `"zoom or multi-touch"`, `"press power"`, `"scroll right"`, and `"press back"`, listed in order of frequency in the dataset.
  - **annot_position (array)**: A flattened array of bounding box coordinates for detected UI elements, formatted as (top, left, height, width), normalized to the [0, 1] range, and rounded to three decimal places. If applicable, the length of this array is `4 * num_ui_elements` per scene; otherwise, it is an empty list.
  - **lift (array)**: Lift coordinates in (x, y) format, normalized to the [0, 1] range and rounded to three decimal places. If not applicable, the value is (-1, -1).
  - **touch (array)**: Touch coordinates in (x, y) format, normalized to the [0, 1] range and rounded to three decimal places. If not applicable, the value is (-1, -1).
  - **type_text (str)**: The entered text, if the action type is `"type"`; otherwise, this is an empty string.
  ```python
  # example
  actions = [
    [
        {"action_type_id": 4, "action_type_text": "click", "annot_position": annot_position, "lift": lift_point_action1, "touch": touch_point_action1, "type_text": ""}, # First action in scene 1
        {"action_type_id": 4, "action_type_text": "click", "annot_position": annot_position, "lift": lift_point_action2, "touch": touch_point_action2, "type_text": ""}, # Second action in scene 1
    ],
    [
        {"action_type_id": 3, "action_type_text": "type", "annot_position": [], "lift": [-1, -1], "touch": [-1, -1], "type_text": "..."}, # First action in scene 2
    ],
    ...
  ]
  ```
  Note: The data format of `actions` is derived from [SeeClick](https://github.com/njucckevin/SeeClick/blob/main/agent_tasks/readme_agent.md) and [AitW](https://github.com/google-research/google-research/tree/master/android_in_the_wild#dataset-format).

- **video_fps (float)**: Frames per second of the video. _This value must be preserved when downloading the video to ensure consistency with `scene_timestamps_in_sec`._
- **video_width (int)**: Width of the video in pixels. _This value must be preserved when downloading the video to ensure consistency with `screen_bboxes`._
- **video_height (int)**: Height of the video in pixels. _This value must be preserved when downloading the video to ensure consistency with `screen_bboxes`._

## Citation
```bibtex
@inproceedings{jang2025_monday,
    title={{Scalable Video-to-Dataset Generation for Cross-Platform Mobile Agents}},
    author={Jang, Yunseok and Song, Yeda and Sohn, Sungryull and Logeswaran, Lajanugen and Luo, Tiange and Kim, Dong-Ki and Bae, Kyunghoon and Lee, Honglak},
    booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    year={2025}
}
````