File size: 6,087 Bytes
2322ac4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3cc4082
2322ac4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
{
  "@context": {
    "@language": "en",
    "@vocab": "https://schema.org/",
    "citeAs": "cr:citeAs",
    "column": "cr:column",
    "conformsTo": "dct:conformsTo",
    "cr": "http://mlcommons.org/croissant/",
    "rai": "http://mlcommons.org/croissant/RAI/",
    "data": {
      "@id": "cr:data",
      "@type": "@json"
    },
    "dataType": {
      "@id": "cr:dataType",
      "@type": "@vocab"
    },
    "dct": "http://purl.org/dc/terms/",
    "examples": {
      "@id": "cr:examples",
      "@type": "@json"
    },
    "extract": "cr:extract",
    "field": "cr:field",
    "fileProperty": "cr:fileProperty",
    "fileObject": "cr:fileObject",
    "fileSet": "cr:fileSet",
    "format": "cr:format",
    "includes": "cr:includes",
    "isLiveDataset": "cr:isLiveDataset",
    "jsonPath": "cr:jsonPath",
    "key": "cr:key",
    "md5": "cr:md5",
    "parentField": "cr:parentField",
    "path": "cr:path",
    "recordSet": "cr:recordSet",
    "references": "cr:references",
    "regex": "cr:regex",
    "repeated": "cr:repeated",
    "replace": "cr:replace",
    "sc": "https://schema.org/",
    "separator": "cr:separator",
    "source": "cr:source",
    "subField": "cr:subField",
    "transform": "cr:transform"
  },
  "@type": "sc:Dataset",
  "name": "VBVR-MultiStep-Bench",
  "conformsTo": "http://mlcommons.org/croissant/1.0",
  "description": "The frozen 180-instance public evaluation split of the VBVR-MultiStep benchmark for long-horizon multi-step image-to-video reasoning. 36 parameterized tasks across six reasoning families (Navigation, Planning, CSP, Execution, Geometry, Physics). Each instance follows a five-artifact contract: first_frame.png, prompt.txt, final_frame.png, ground_truth.mp4, question_metadata.json.",
  "alternateName": ["VBVR-MultiStep Evaluation Split"],
  "creator": {
    "@type": "sc:Organization",
    "name": "Video-Reason",
    "url": "https://video-reason.com"
  },
  "datePublished": "2026-05-06",
  "keywords": ["video reasoning", "multi-step reasoning", "long-horizon", "image-to-video", "benchmark", "synthetic"],
  "license": "https://creativecommons.org/licenses/by/4.0/",
  "url": "https://huggingface.co/datasets/Video-Reason/VBVR-MultiStep-Bench",
  "version": "1.0.0",
  "isLiveDataset": false,
  "rai:dataCollection": "Fully synthetic. Every instance is procedurally produced by a deterministic per-task generator that consumes only released task definitions and a seed. There is no scraping, no human subjects, no third-party media, and no annotation of pre-existing content. Generators run as released code; instance manifests record the seed, the generator version, and per-task tolerances for exact reproduction.",
  "rai:dataCollectionType": ["Synthetic"],
  "rai:hasSyntheticData": true,
  "rai:dataPreprocessingProtocol": "No post-collection preprocessing is applied. Each task's generator emits the five-artifact contract (first_frame.png, prompt.txt, final_frame.png, ground_truth.mp4, question_metadata.json) directly. Sample index, seed, and version are recorded in question_metadata.json so any instance is regenerable from the released code.",
  "rai:dataAnnotationProtocol": "No human annotation. The reference rollout (ground_truth.mp4) is computed by the generator's deterministic ground-truth solver and is not produced by humans.",
  "rai:dataAnnotationPlatform": "N/A (no annotation).",
  "rai:dataReleaseMaintenancePlan": "Versioned releases on the Hugging Face Hub. The Croissant file in this repository is the canonical long-term record. Issues and breaking changes will be documented in repository commits and the dataset version field.",
  "rai:dataLimitations": [
    "Tasks are stylized and synthetic; transfer to unconstrained open-world video is not validated.",
    "Visual style is generator-controlled and intentionally simplified to keep symbolic state recoverable; this is not a photorealism benchmark.",
    "The CSP family ships 6 tasks but only 3 (Multi-13/14/15 excluded) are used in the human-judging pool described in the companion paper.",
    "Reference rollouts encode one valid trajectory per instance; alternative valid trajectories are not enumerated."
  ],
  "rai:dataBiases": [
    "Distribution biases inherited from generator parameter ranges: e.g., maze sizes, planning horizons, physics regimes are drawn from fixed bounded distributions and do not span the long tail of real-world settings.",
    "Family balance is uniform (6 tasks per family) by design; this is a deliberate evaluation choice and does not reflect natural prevalence of these reasoning patterns.",
    "Visual rendering is monocular, planar, and stylized; appearance distribution does not approximate any real-world video corpus.",
    "No demographic content is generated; bias along human demographic axes does not apply to this dataset."
  ],
  "rai:personalSensitiveInformation": "None. The dataset contains no personal information, no biometric data, no demographic information, and no human subjects. All visual content is procedurally generated geometric / symbolic / physical scenes.",
  "rai:dataUseCases": "Trajectory-level evaluation of image-to-video systems on long-horizon, rule-grounded tasks. Suitable for: (a) blind human pairwise comparison across model families on process correctness, fidelity, and render-quality axes; (b) automated reference-trajectory comparison once parsers for dense video state become reliable; (c) family-localized failure-mode diagnosis. The companion paper validates use case (a). Use cases (b) and (c) are not validated by this release.",
  "rai:dataSocialImpact": "Intended for research on reasoning evaluation in video generation. Risks are minimal: the dataset is synthetic, free of personal content, and designed for evaluation rather than deployment. Potential for misuse is low; the most plausible negative impact is overfitting research on stylized rule-checked tasks at the expense of unconstrained video understanding, which we mitigate by clearly scoping the contribution as a complement to (not a replacement for) appearance-centric evaluation."
}