File size: 3,978 Bytes
3235399
 
 
52e0b6f
 
3235399
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
---
license: cc-by-4.0
task_categories:
  - image-to-video
  - text-to-video
language:
  - en
size_categories:
  - n<1K
pretty_name: VBVR-MultiStep-Bench
tags:
  - video-reasoning
  - multi-step
  - long-horizon
  - image-to-video
  - evaluation
  - benchmark
---

# VBVR-MultiStep-Bench

The frozen **180-instance public evaluation split** released alongside the [VBVR-MultiStep](https://huggingface.co/datasets/Video-Reason/VBVR-MultiStep) training corpus. Designed for long-horizon multi-step image-to-video (I2V) reasoning evaluation.

This dataset is part of the **VBVR (Very Big Video Reasoning Suite)** project. See the parent suite at <https://video-reason.com> and the suite paper [VBVR: A Very Big Video Reasoning Suite (Wang et al., ICML 2026)](https://icml.cc/virtual/2026/poster/65709).

## At a glance

| Property | Value |
|---|---|
| Tasks | **36** parameterized tasks (`Multi-01``Multi-36`) |
| Reasoning families | Navigation, Planning, CSP, Execution, Geometry, Physics |
| Instances | **180** (5 per task × 36) |
| Per-instance artifacts | 5 (see below) |
| License | CC-BY-4.0 |

## Five-artifact data contract

Every instance lives at:

```
Multi-XX_<name>_data-generator/Multi-XX_<name>_data-generator_task/Multi-XX_<name>_data-generator_<id>/
```

and contains exactly:

| File | Role |
|---|---|
| `first_frame.png` | Model conditioning image (the only visual input the model receives at inference) |
| `prompt.txt` | Natural-language task contract |
| `final_frame.png` | Target endpoint (held out from the model) |
| `ground_truth.mp4` | Reference rollout demonstrating the correct trajectory |
| `question_metadata.json` | Seed, version, tolerances, task-specific fields |

A top-level `metadata.parquet` indexes every instance with the task id, family, seed, and per-instance metadata for fast filtering.

## Reasoning families

| Family | Characteristic | Released tasks |
|---|---|---|
| Navigation | Discrete motion under adjacency / obstacle constraints | 6 |
| Planning | Operator-based state transformation | 6 |
| CSP | Incremental labeling under global consistency | 6 (3 used for human judging) |
| Execution | Clocked deterministic update rules | 6 |
| Geometry | Ordered constructive geometry | 6 |
| Physics | Continuous dynamics with contact / conservation | 6 |

Tasks `Multi-13`, `Multi-14`, `Multi-15` (CSP) are excluded from the human-judging pool described in the paper but are included in this release for completeness.

## Intended use

- **Primary use**: trajectory-level evaluation of I2V systems under a fixed five-artifact contract.
- **Comparison protocol**: blind human pairwise judging on three independent axes — process correctness, reference fidelity, render quality.
- **Companion training corpus**: [Video-Reason/VBVR-MultiStep](https://huggingface.co/datasets/Video-Reason/VBVR-MultiStep) (~360k samples).

## Loading

```python
import pandas as pd
meta = pd.read_parquet("hf://datasets/Video-Reason/VBVR-MultiStep-Bench/metadata.parquet")
```

Or pull a single instance:

```python
from huggingface_hub import hf_hub_download
prompt_path = hf_hub_download(
    "Video-Reason/VBVR-MultiStep-Bench",
    "Multi-01_maze_shortest_path_data-generator/Multi-01_maze_shortest_path_data-generator_task/Multi-01_maze_shortest_path_data-generator_00000000/prompt.txt",
    repo_type="dataset",
)
```

## License

Released under **CC-BY-4.0**. The reference rollouts are produced from generators that consume only released task definitions; no third-party copyrighted content is embedded.

Wan2.2-I2V-A14B (Apache-2.0) is referenced as a baseline model and a fine-tuning ancestor for `VBVR-Wan2.2`; this dataset does not redistribute Wan2.2 weights.

## Responsible AI

This dataset is fully synthetic — generators produce every instance from controlled parameters. There are no human subjects, no scraped media, and no personal information. See the [Croissant file](./croissant.json) for the complete RAI metadata.