File size: 6,349 Bytes
7cb2657
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cb9ba6b
7cb2657
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
{
  "@context": {
    "@language": "en",
    "@vocab": "https://schema.org/",
    "citeAs": "cr:citeAs",
    "column": "cr:column",
    "conformsTo": "dct:conformsTo",
    "cr": "http://mlcommons.org/croissant/",
    "rai": "http://mlcommons.org/croissant/RAI/",
    "data": {"@id": "cr:data", "@type": "@json"},
    "dataType": {"@id": "cr:dataType", "@type": "@vocab"},
    "dct": "http://purl.org/dc/terms/",
    "examples": {"@id": "cr:examples", "@type": "@json"},
    "extract": "cr:extract",
    "field": "cr:field",
    "fileProperty": "cr:fileProperty",
    "fileObject": "cr:fileObject",
    "fileSet": "cr:fileSet",
    "format": "cr:format",
    "includes": "cr:includes",
    "isLiveDataset": "cr:isLiveDataset",
    "jsonPath": "cr:jsonPath",
    "key": "cr:key",
    "md5": "cr:md5",
    "parentField": "cr:parentField",
    "path": "cr:path",
    "recordSet": "cr:recordSet",
    "references": "cr:references",
    "regex": "cr:regex",
    "repeated": "cr:repeated",
    "replace": "cr:replace",
    "sc": "https://schema.org/",
    "separator": "cr:separator",
    "source": "cr:source",
    "subField": "cr:subField",
    "transform": "cr:transform"
  },
  "@type": "sc:Dataset",
  "name": "VBVR-MultiStep",
  "conformsTo": "http://mlcommons.org/croissant/1.0",
  "description": "The ~360,000-sample programmatic training corpus for long-horizon multi-step image-to-video reasoning. 36 parameterized tasks across six reasoning families (Navigation, Planning, CSP, Execution, Geometry, Physics). Distributed as 7,200 tar.gz shards (≈50 samples per shard) plus Parquet metadata; each instance follows a five-artifact contract identical to the VBVR-MultiStep-Bench evaluation split.",
  "alternateName": ["VBVR-MultiStep Training Corpus"],
  "creator": {
    "@type": "sc:Organization",
    "name": "Video-Reason",
    "url": "https://video-reason.com"
  },
  "datePublished": "2026-05-06",
  "keywords": ["video reasoning", "multi-step reasoning", "long-horizon", "image-to-video", "training", "synthetic", "tar.gz", "parquet"],
  "license": "https://creativecommons.org/licenses/by/4.0/",
  "url": "https://huggingface.co/datasets/Video-Reason/VBVR-MultiStep",
  "version": "1.0.0",
  "isLiveDataset": false,
  "rai:dataCollection": "Fully synthetic. Each of the 36 tasks ships a deterministic generator that emits the five-artifact contract (first_frame.png, prompt.txt, final_frame.png, ground_truth.mp4, question_metadata.json) from a (task, seed) pair. No scraping, no human subjects, no third-party media, no manual annotation. The training corpus is partitioned into disjoint seed bands (1–5,000 and 5,001–10,000 per task) that are themselves disjoint from the evaluation seeds released in VBVR-MultiStep-Bench.",
  "rai:dataCollectionType": ["Synthetic"],
  "rai:hasSyntheticData": true,
  "rai:dataPreprocessingProtocol": "Per-task generator output is grouped into 50-sample batches and packed into tar.gz shards under questions/. Per-task Parquet metadata files (data/metadata_shards/) and a global metadata.parquet index every instance with task id, family, seed, and per-task fields. No sample is filtered, dropped, or transformed after generation.",
  "rai:dataAnnotationProtocol": "No human annotation. ground_truth.mp4 is rendered by each task's deterministic ground-truth solver.",
  "rai:dataAnnotationPlatform": "N/A (no annotation).",
  "rai:dataReleaseMaintenancePlan": "Versioned releases on the Hugging Face Hub. The Croissant file in this repository is the canonical long-term record. A 5 GB representative subset is provided under sample/ for quick inspection and reviewer convenience.",
  "rai:dataLimitations": [
    "Synthetic and stylized: transfer to unconstrained open-world video is not validated.",
    "Visual rendering is intentionally simplified to keep the symbolic state recoverable from frames; this is not a photorealism corpus.",
    "Per-task generator parameter ranges are bounded (e.g., maze sizes, planning horizons, physics regimes); the corpus does not span the long tail of any single family.",
    "Reference rollouts encode one valid trajectory per instance; alternative valid trajectories are not enumerated.",
    "Although 36 tasks ship, only 34 are used in the training experiments described in the companion paper; this release contains all 36 task families."
  ],
  "rai:dataBiases": [
    "Family balance is uniform (6 tasks per family) by design and does not reflect natural prevalence of these reasoning patterns.",
    "Generator parameters bias the difficulty distribution toward bounded and seed-controlled regimes that are amenable to symbolic ground truth; rare or open-ended cases are out of scope.",
    "Visual style is monocular, planar, and rendered by a fixed family of renderers; appearance distribution does not approximate any real-world video corpus and does not contain demographic content.",
    "No human demographic information is generated; bias along human demographic axes does not apply."
  ],
  "rai:personalSensitiveInformation": "None. The dataset contains no personal information, no biometric data, no demographic information, and no human subjects. All visual content is procedurally generated geometric, symbolic, or physical scenes.",
  "rai:dataUseCases": "Training image-to-video systems on long-horizon multi-step reasoning under explicit per-step rules. Validated use case in the companion paper: fine-tuning Wan2.2-I2V-A14B (Apache-2.0) with Dual-DiT two-phase LoRA. Out-of-scope: production VLM pretraining at scale, real-world video generation, or any safety-critical use.",
  "rai:dataSocialImpact": "Intended for academic research on reasoning evaluation in video generation. Risks are minimal: the dataset is synthetic, free of personal content, and rendered in a stylized regime not representative of any real population. The most plausible concern is research-direction effects (e.g., over-investing in stylized synthetic benchmarks), which we mitigate by positioning this corpus as a complement to (not a replacement for) appearance-centric and real-world video corpora.",
  "rai:dataReleaseUpdate": "If post-release errors are discovered, fixes will be published as additive shards or replacement Parquet entries in a new dataset version, with the prior version retained at its commit hash for backwards reproducibility."
}