Datasets:
File size: 6,056 Bytes
ee2c9ba 36548b3 ee2c9ba 36548b3 ee2c9ba 7a2006e ee2c9ba cd30884 ee2c9ba 03cfdf5 ee2c9ba 03cfdf5 ee2c9ba 1ac6c2b ee2c9ba 1ac6c2b ee2c9ba 1ac6c2b ee2c9ba 7a2006e ee2c9ba | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 | ---
task_categories:
- any-to-any
language:
- en
configs:
- config_name: main
data_files:
- split: ballgame
path: ballgame/*.parquet
- split: cube
path: cube/*.parquet
- split: maze
path: maze/*.parquet
- split: mmsi
path: mmsi/*.parquet
- split: multihop
path: multihop/*.parquet
- split: paperfolding
path: paperfolding/*.parquet
- split: sokoban
path: sokoban/*.parquet
tags:
- multimodal
- reasoning
- world-models
---
# VisWorld-Eval: Task Suite for Reasoning with Visual World Modeling ๐
[](https://thuml.github.io/Reasoning-Visual-World/)
[](https://arxiv.org/abs/2601.19834)
[](https://github.com/thuml/Reasoning-Visual-World)
[](https://huggingface.co/datasets/thuml/VisWorld-Eval)

## ๐ Introduction
The VisWorld-Eval suite is for assessing multimodal reasoning with visual world modeling. It comprises seven tasks spanning both synthetic and real-world domains, each designed to isolate and demand specific atomic world-model capabilities.
| Task | Capability | Domain | Test Samples | Source / Reference |
|------------------------------|-----------------|-------------|--------------|--------------------------------|
| Paper folding | Simulation | Synthetic | 480 | [SpatialViz](https://github.com/wangst0181/Spatial-Visualization-Benchmark) |
| Multi-hop manipulation | Simulation | Synthetic | 480 | [ZebraCoT](https://github.com/multimodal-reasoning-lab/Bagel-Zebra-CoT), [CLEVR](https://github.com/facebookresearch/clevr-dataset-gen) |
| Ball tracking | Simulation | Synthetic | 1,024 | [RBench-V](https://github.com/CHEN-Xinsheng/VLMEvalKit_RBench-V) |
| Maze | Simulation | Synthetic | 480 | [maze-dataset](https://github.com/understanding-search/maze-dataset) |
| Sokoban | Simulation | Synthetic | 480 | [Game-RL](https://github.com/tongjingqi/Game-RL) |
| Cube 3-view projection | Reconstruction | Synthetic | 480 | [SpatialViz](https://github.com/wangst0181/Spatial-Visualization-Benchmark) |
| Real-world spatial reasoning | Reconstruction | Real-world | 522 | [MMSI-Bench](https://github.com/InternRobotics/MMSI-Bench) |
## โ๏ธ Load Data
Load from ๐ค HuggingFace:
```python
from datasets import load_dataset
ds = load_dataset("thuml/VisWorld-Eval")
```
## ๐งช Evaluation
We evaluate different models through API servers. Set your API key and server address in `evaluate.py` (for the evaluated model) and `task_verifiers.py` (for the model judge), then run the evaluation script.
```bash
python eval/evaluate.py --task ballgame --model gemini3pro
```
## ๐ Leaderboard
Zero-shot evaluation of advanced VLMs on VisWorld-Eval: We report the average accuracy over five tasks (excluding Maze and Sokoban) and over all seven tasks.
| Models | Paper Folding | Multi-Hop Manip. | Ball Tracking | Cube 3-View | MMSI (Pos. Rel.) | Maze | Sokoban | Overall (5 tasks) | Overall (7 tasks) |
|------------------------|---------------|------------------|---------------|------------|------------------|------|---------|-------------------|-------------------|
| Gemini 3 Flash | 25.6 | **75.4** | **55.3** | 52.7 | 41.3 | 73.9 | **99.3** | **50.0** | **60.5** |
| Gemini 3 Pro | **27.0** | 74.5 | 44.7 | **53.3** | **49.6** | 33.5 | 90.2 | 49.8 | 53.2 |
| Seed 1.8 | 10.6 | 75.2 | 24.4 | 42.5 | 38.8 | **83.9** | 68.3 | 38.3 | 49.1 |
| GPT 5.1 | 6.4 | 73.9 | 34.8 | 44.5 | 44.8 | 0.6 | 62.8 | 40.8 | 38.2 |
| o3 | 13.5 | 68.1 | 24.7 | 37.7 | 44.4 | 0.0 | 36.0 | 37.6 | 32.0 |
| Qwen3-VL-8B-Thinking | 11.0 | 49.3 | 17.8 | 21.2 | 27.7 | 0.0 | 5.8 | 25.4 | 18.9 |
| BAGEL-7B-MoT | 11.2 | 31.6 | 19.4 | 26.8 | 27.2 | 0.0 | 0.2 | 23.2 | 16.6 |
## ๐ Release Progress
- [x] VisWorld-Eval data
- [x] VisWorld-Eval evaluation scripts
## ๐ Citation
If you find this project useful, please cite our paper as:
```
@article{wu2026visual,
title={Visual Generation Unlocks Human-Like Reasoning through Multimodal World Models},
author={Jialong Wu and Xiaoying Zhang and Hongyi Yuan and Xiangcheng Zhang and Tianhao
Huang and Changjing He and Chaoyi Deng and Renrui Zhang and Youbin Wu and Mingsheng
Long},
journal={arXiv preprint arXiv:2601.19834},
year={2026},
}
```
## ๐ค Contact
If you have any questions, please contact wujialong0229@gmail.com.
## ๐ก Acknowledgement
We sincerely appreciate the following projects for their valuable codebase and task design: [SpatialViz](https://github.com/wangst0181/Spatial-Visualization-Benchmark), [RBench-V](https://github.com/CHEN-Xinsheng/VLMEvalKit_RBench-V), [maze-dataset](https://github.com/understanding-search/maze-dataset), [Game-RL](https://github.com/tongjingqi/Game-RL), [clevr-dataset-gen](https://github.com/facebookresearch/clevr-dataset-gen), [MMSI-Bench](https://github.com/InternRobotics/MMSI-Bench). |