Android-Control-84k / README.md
OfficerChul's picture
Upload folder using huggingface_hub
0248027 verified
---
license: apache-2.0
task_categories:
- visual-question-answering
- image-to-text
language:
- en
tags:
- mobile-ui
- gui-grounding
- android
- ui-automation
- multimodal
size_categories:
- 10K<n<100K
pretty_name: Android Control Dataset for LLaMA-Factory
---
# Android Control Dataset
## Overview
This directory contains two dataset files (`and_ctrl_train.json` and `and_ctrl_test.json`) derived from the [Android Control](https://github.com/google-research/google-research/tree/master/android_control) project by Google Research. These datasets have been formatted specifically for GUI grounding training in LLaMA-Factory.
## Dataset Description
The Android Control dataset consists of episodes where each episode contains multiple steps. Each step includes:
- **Step instructions**: Natural language instructions for UI interactions
- **Actions**: The type of action to perform (click, scroll, input text, etc.)
- **Coordinates**: Precise x, y coordinates for the action
The data has been extracted and formatted to train models for mobile UI understanding and interaction tasks.
## Files
- `and_ctrl_train.json`: Training dataset
- `and_ctrl_test.json`: Test/evaluation dataset
- `download_android_control.ipynb`: Jupyter notebook for downloading images and processing the original data
## Data Format
Each entry in the JSON files follows the LLaMA-Factory conversation format:
```json
{
"messages": [
{
"role": "system",
"content": "You are a helpful assistant that can identify what action to perform on mobile UI Screenshot given the user instruction."
},
{
"role": "user",
"content": "<image>Click on the Recording 2"
},
{
"role": "assistant",
"content": "{\"action_type\": \"click\", \"x\": 561, \"y\": 535}"
}
],
"images": ["and_ctrl/out_episode_18557_step_001.png"]
}
```
## Setup Instructions
To use these datasets in LLaMA-Factory:
1. **Create the image directory**:
```bash
mkdir -p data/and_ctrl
```
2. **Download images**:
Run the provided `download_android_control.ipynb` notebook to download and process the original images. The notebook will:
- Download TFRecord files from Google Storage (`gs://gresearch/android_control/`)
- Extract images and save them directly to `and_ctrl/` directory
- Automatically organize images with the naming convention: `out_episode_{episode_id}_step_{step_number}.png`
- Generate an `and_ctrl.json` file with the processed data
3. **Dataset files**:
- Images: Stored in `data/and_ctrl/` folder
- Training dataset: `and_ctrl_train.json` in `data/datasets/`
- Test dataset: `and_ctrl_test.json` in `data/datasets/`
## Dataset Statistics
**Total samples**: Train: 82,944 | Test: 904
| Action Type | Train | Test |
|-------------|-------|------|
| click | 51,793 (62.44%) | 125 (13.83%) |
| scroll | 11,005 (13.27%) | 125 (13.83%) |
| input_text | 5,966 (7.19%) | 125 (13.83%) |
| wait | 5,657 (6.82%) | 125 (13.83%) |
| open_app | 5,572 (6.72%) | 125 (13.83%) |
| navigate_back | 2,909 (3.51%) | 125 (13.83%) |
| long_press | 42 (0.05%) | 125 (13.83%) |
| navigate_home | 0 (0.00%) | 29 (3.21%) |
**Note**: The training dataset shows a natural distribution with click actions being dominant (62.44%), while the test dataset is intentionally balanced with most action types having equal representation (~13.83% each). The `navigate_home` action appears only in the test set.
## Training Usage
These datasets are specifically formatted for training multimodal language models to:
- Understand mobile UI screenshots
- Ground natural language instructions to specific UI elements
- Generate precise action coordinates for UI automation
- Learn mobile app interaction patterns
## Source and Attribution
Original dataset: [Google Research Android Control](https://github.com/google-research/google-research/tree/master/android_control)
The Android Control dataset was created by Google Research for advancing mobile UI understanding and automation research.
### License
This dataset is derived from Google Research's Android Control dataset, which is licensed under the [Apache License 2.0](https://www.apache.org/licenses/LICENSE-2.0). The reformatted version for LLaMA-Factory maintains the same Apache 2.0 license terms.
Copyright for the original dataset belongs to Google LLC. Any modifications or reformatting for LLaMA-Factory compatibility are also provided under Apache License 2.0.
## Notes
- The images are referenced with relative paths starting with `and_ctrl/`
- Each action includes the action type and necessary parameters (coordinates, text, direction, etc.)
- The test set can be used for evaluating model performance on unseen mobile UI interactions