File size: 2,517 Bytes
3b65786
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
---
pretty_name: VisPhyBench
license: mit
language:
  - en
---

# VisPhyBench

To evaluate how well models reconstruct appearance and reproduce physically plausible motion, we introduce VisPhyBench, a unified evaluation protocol comprising 209 scenes derived from 108 physical templates that assesses physical understanding through the lens of code-driven resimulation in both 2D and 3D scenes, integrating metrics from different aspects. Each scene is also annotated with a coarse difficulty label (easy/medium/hard).

# Dataset Details

- **Created by:** Jiarong Liang
- **Language(s) (NLP):** English
- **License:** MIT
- **Repository:** `https://github.com/TIGER-AI-Lab/VisPhyWorld`

# Uses

The dataset is used to evaluate how well models reconstruct appearance and reproduce physically plausible motion.

# Dataset Structure

The difficulty of the two sets of VisPhyBench splits, **sub (209)** and **test (49)**, are as follows: **sub** is **114/67/28** (**54.5%/32.1%/13.4%**), and **test** is **29/17/3** (**59.2%/34.7%/6.1%**) (Easy/Medium/Hard).

## What each sample contains

VisPhyBench is provided as two splits:

- `sub`: a larger split intended for evaluation and analysis.
- `test`: a smaller split subsampled from `sub` for quick sanity checks.

For each sample, we provide:

1. **A short video** of a synthetic physical scene.
2. **A detection JSON** (per sample) that describes the scene in the first frame.
3. **A difficulty label** (easy/medium/hard) derived from the mean of eight annotators’ 1–5 ratings.

## Detection JSON format

Each detection JSON includes:

- `image_size`: the image width/height.
- `coordinate_system`: conventions for coordinates (e.g., origin and axis directions).
- `objects`: a list of detected objects. Each object includes:
  - `id`: unique identifier.
  - `category`: coarse geometry category.
  - `color_rgb`: RGB color triplet.
  - `position`: object position (e.g., center coordinates).
  - `bbox`: bounding box coordinates and size.
  - `size`: coarse size fields (e.g., radius/length/thickness).

These fields specify object locations and attributes precisely, which helps an LLM initialize objects correctly when generating executable simulation code.

# BibTeX

```bibtex
@misc{visphybench2026,
  title        = {VisPhyBench},
  author       = {Liang, Jiarong and Ku, Max and Hui, Ka-Hei and Nie, Ping and Chen, Wenhu},
  howpublished = {GitHub repository},
  year         = {2026},
  url          = {https://github.com/TIGER-AI-Lab/VisPhyWorld}
}
```