--- license: cc-by-4.0 task_categories: - image-to-video - text-to-video language: - en size_categories: - 100K. See [Wang et al., ICML 2026](https://icml.cc/virtual/2026/poster/65709) for the parent suite. ## At a glance | Property | Value | |---|---| | Tasks | **36** parameterized tasks (`Multi-01` … `Multi-36`) | | Reasoning families | Navigation, Planning, CSP, Execution, Geometry, Physics | | Total samples | **~360,000** (≈10k per task) | | Total size | **~164 GB** | | Format | Tar.gz shards (nested per-sample folders) + Parquet metadata | | Shards | 7,200 (≈50 samples per shard) | | License | CC-BY-4.0 | ## Repository layout ``` . ├── README.md ├── croissant.json # Croissant + RAI metadata ├── data/ │ ├── metadata.parquet # global index of all 360k samples │ └── metadata_shards/ │ └── Multi-XX_.parquet # per-task metadata (36 files) ├── questions/ # WebDataset shards │ └── Multi-XX__NNNNN-NNNNN.tar.gz │ └── (50 samples per shard, 5 files per sample, see "Sample format" below) └── sample/ # ~5 GB representative subset for quick inspection ├── data/metadata_shards/... └── questions/ # 6 shards × 36 tasks = 216 shards ``` The `sample/` subdirectory is a 5 GB pre-curated subset (the first 300 samples of every task) for reviewers and quick experimentation. To pull it: ```bash huggingface-cli download Video-Reason/VBVR-MultiStep \ --repo-type dataset \ --include "sample/**" \ --local-dir ./vbvr-multistep-sample ``` ## Sample format (inside each `.tar.gz` shard) Each shard expands to a nested folder tree, identical in shape to the evaluation split: ``` Multi-XX__data-generator/ └── Multi-XX__data-generator_task/ └── Multi-XX__data-generator_/ ├── first_frame.png # conditioning frame ├── prompt.txt # natural-language task contract ├── final_frame.png # target endpoint (held-out at inference) ├── ground_truth.mp4 # reference rollout └── question_metadata.json # seed, version, tolerances, task fields ``` Each shard contains 50 such instance folders. The five-artifact contract is identical to the evaluation split. To extract: ```bash tar xzf Multi-01_maze_shortest_path_data-generator_00000-00049.tar.gz ``` ## Loading ### Per-task metadata (recommended entry point) ```python import pandas as pd m = pd.read_parquet( "hf://datasets/Video-Reason/VBVR-MultiStep/data/metadata_shards/Multi-01_maze_shortest_path_data-generator.parquet" ) print(m.head()) ``` ### Direct shard download ```python from huggingface_hub import hf_hub_download import tarfile shard = hf_hub_download( "Video-Reason/VBVR-MultiStep", "questions/Multi-01_maze_shortest_path_data-generator_00000-00049.tar.gz", repo_type="dataset", ) with tarfile.open(shard) as t: t.extractall("./extracted") ``` ### Pull only the 5 GB sample ```bash huggingface-cli download Video-Reason/VBVR-MultiStep \ --repo-type dataset \ --include "sample/**" \ --local-dir ./vbvr-multistep-sample ``` ## Splits and seeds The training corpus is partitioned into disjoint seed bands: | Band | Seed range | Samples per task | Total samples | |---|---|---|---| | First-half | 1–5,000 | 5,000 | ~170k (across 34 trained tasks) | | Second-half | 5,001–10,000 | 5,000 | ~170k (across 34 trained tasks) | Both bands are disjoint from the **180-instance evaluation seeds** in `VBVR-MultiStep-Bench`. The submitted paper trains on 34 of 36 tasks; the released corpus contains all 36 task families. ## Reasoning families See the [bench dataset card](https://huggingface.co/datasets/Video-Reason/VBVR-MultiStep-Bench) for the family taxonomy. Each family contributes 6 tasks, for 36 total. ## Intended use and out-of-scope - **Primary use**: training I2V systems on long-horizon multi-step reasoning under explicit per-step rules. - **Out-of-scope**: this corpus is fully synthetic and stylized; transfer to unconstrained open-world video is not validated by this release. - **Not validated for**: production VLM pretraining at scale, real-world video generation, or any safety-critical use. ## License Released under **CC-BY-4.0**. Generators consume only released task definitions; no third-party copyrighted content is embedded. Derivatives of `Wan2.2-I2V-A14B` (Apache-2.0) referenced in the companion paper comply with the upstream license. This dataset does not redistribute model weights. ## Responsible AI The dataset is fully synthetic. There are no human subjects, no scraped media, and no personal or sensitive information. Known biases inherit from the deterministic generators — every task family covers a deliberately narrow conceptual slice, and visual style is controlled by a fixed renderer family (no demographic content). See [`croissant.json`](./croissant.json) for the complete RAI metadata.