ENACT / README.md
Inevitablevalor's picture
Update README.md
5c07d8f verified
---
pretty_name: ENACT
language:
- en
task_categories:
- visual-question-answering
configs:
- config_name: default
data_files:
- QA.zip
dataset_info:
features:
- name: id
dtype: string
- name: type
dtype: string
- name: task_name
dtype: string
- name: key_frame_ids
sequence: string
- name: images
sequence: string
- name: question
dtype: string
- name: options
sequence: string
- name: gt_answer
sequence: int32
license: mit
tags:
- agent
size_categories:
- 1K<n<10K
---
# ENACT: Evaluating Embodied Cognition with World Modeling of Egocentric Interaction
ENACT is a benchmark dataset for evaluating **embodied cognition** in vision–language models via **egocentric world modeling**. It probes whether models can reason about how the world changes under sequences of actions, using long-horizon household activities in a mobile manipulation setting.
- **Project page:** https://enact-embodied-cognition.github.io/
- **Code & evaluation:** https://github.com/mll-lab-nu/ENACT
- **Paper** https://arxiv.org/abs/2511.20937
## Dataset Summary
Each ENACT example is a **multi-image, multi-step reasoning problem** built from robot trajectories:
- **Forward world modeling**
- Input: one **current state image**, several **future state images** (shuffled), and a list of **actions in correct order**.
- Task: output a Python list of integers giving the **correct chronological order of future images** (e.g., `[1, 3, 2]`).
- **Inverse world modeling**
- Input: an **ordered sequence of images** showing state changes, plus **actions in shuffled order**.
- Task: output a Python list of integers giving the **correct chronological order of actions** (e.g., `[2, 3, 1]`).
All images are egocentric RGB observations rendered from long-horizon household tasks (e.g., assembling gift baskets, bringing water, preparing lunch boxes, cleaning up a desk).
## File Structure
After unpacking, the dataset has the following structure:
```text
.
├── enact_ordering.jsonl # All QA examples (one JSON per line)
└── images/
├── forward_world_modeling_3_steps/
├── forward_world_modeling_4_steps/
├── ...
├── forward_world_modeling_10_steps/
├── inverse_world_modeling_3_steps/
├── ...
└── inverse_world_modeling_10_steps/
````
Each task folder (e.g., `forward_world_modeling_3_steps/`) contains one subfolder per activity, such as:
```text
images/forward_world_modeling_3_steps/
├── assembling_gift_baskets_1749468508582193/
├── bringing_water_1750844141719178/
├── ...
```
Inside each activity folder are the PNGs for that trajectory (current state and future states, or ordered states in the inverse setting).
## JSONL Format
Each line in `enact_ordering.jsonl` is a JSON object:
```json
{
"id": "assembling_gift_baskets_1749468508582193_forward_world_modeling_3_steps_cfbcc15c",
"type": "forward_world_modeling_3_steps",
"task_name": "assembling_gift_baskets_1749468508582193",
"key_frame_ids": ["4150", "11360", "11834"],
"images": [
"QA/images/forward_world_modeling_3_steps/..._cur_state.png",
"QA/images/forward_world_modeling_3_steps/..._next_state_1.png",
"QA/images/forward_world_modeling_3_steps/..._next_state_2.png"
],
"question": "...natural language instructions and actions...",
"options": [],
"gt_answer": [1, 2]
}
```
* **`id`** – unique identifier for this QA instance.
* **`type`** – question type and horizon, e.g. `forward_world_modeling_3_steps` or `inverse_world_modeling_4_steps`.
* **`task_name`** – underlying household task instance.
* **`key_frame_ids`** – frame indices of selected key frames in the trajectory.
* **`images`** – relative paths to PNG images:
* index 0 is the **current state**;
* subsequent entries are **future states** (forward) or later states (inverse).
* **`question`** – natural language prompt specifying the task setup, actions, and the required output as a Python list of integers.
* **`gt_answer`** – ground-truth ordering of image or action labels (list of integers, e.g. `[1, 3, 2]`).
## Usage
To evaluate, follow the scripts in the code repository: [https://github.com/mll-lab-nu/ENACT](https://github.com/mll-lab-nu/ENACT)
## Citation
If you use ENACT, please cite the paper:
```
@article{wang2025enact,
title={ENACT: Evaluating Embodied Cognition with World Modeling of Egocentric Interaction},
author={Wang, Qineng and Huang, Wenlong and Zhou, Yu and Yin, Hang
and Bao, Tianwei and Lyu, Jianwen and Liu, Weiyu and Zhang, Ruohan
and Wu, Jiajun and Li, Fei-Fei and Li, Manling},
journal={arXiv preprint arXiv:2511.20937},
year={2025}
}
```