File size: 6,087 Bytes
2322ac4 3cc4082 2322ac4 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 | {
"@context": {
"@language": "en",
"@vocab": "https://schema.org/",
"citeAs": "cr:citeAs",
"column": "cr:column",
"conformsTo": "dct:conformsTo",
"cr": "http://mlcommons.org/croissant/",
"rai": "http://mlcommons.org/croissant/RAI/",
"data": {
"@id": "cr:data",
"@type": "@json"
},
"dataType": {
"@id": "cr:dataType",
"@type": "@vocab"
},
"dct": "http://purl.org/dc/terms/",
"examples": {
"@id": "cr:examples",
"@type": "@json"
},
"extract": "cr:extract",
"field": "cr:field",
"fileProperty": "cr:fileProperty",
"fileObject": "cr:fileObject",
"fileSet": "cr:fileSet",
"format": "cr:format",
"includes": "cr:includes",
"isLiveDataset": "cr:isLiveDataset",
"jsonPath": "cr:jsonPath",
"key": "cr:key",
"md5": "cr:md5",
"parentField": "cr:parentField",
"path": "cr:path",
"recordSet": "cr:recordSet",
"references": "cr:references",
"regex": "cr:regex",
"repeated": "cr:repeated",
"replace": "cr:replace",
"sc": "https://schema.org/",
"separator": "cr:separator",
"source": "cr:source",
"subField": "cr:subField",
"transform": "cr:transform"
},
"@type": "sc:Dataset",
"name": "VBVR-MultiStep-Bench",
"conformsTo": "http://mlcommons.org/croissant/1.0",
"description": "The frozen 180-instance public evaluation split of the VBVR-MultiStep benchmark for long-horizon multi-step image-to-video reasoning. 36 parameterized tasks across six reasoning families (Navigation, Planning, CSP, Execution, Geometry, Physics). Each instance follows a five-artifact contract: first_frame.png, prompt.txt, final_frame.png, ground_truth.mp4, question_metadata.json.",
"alternateName": ["VBVR-MultiStep Evaluation Split"],
"creator": {
"@type": "sc:Organization",
"name": "Video-Reason",
"url": "https://video-reason.com"
},
"datePublished": "2026-05-06",
"keywords": ["video reasoning", "multi-step reasoning", "long-horizon", "image-to-video", "benchmark", "synthetic"],
"license": "https://creativecommons.org/licenses/by/4.0/",
"url": "https://huggingface.co/datasets/Video-Reason/VBVR-MultiStep-Bench",
"version": "1.0.0",
"isLiveDataset": false,
"rai:dataCollection": "Fully synthetic. Every instance is procedurally produced by a deterministic per-task generator that consumes only released task definitions and a seed. There is no scraping, no human subjects, no third-party media, and no annotation of pre-existing content. Generators run as released code; instance manifests record the seed, the generator version, and per-task tolerances for exact reproduction.",
"rai:dataCollectionType": ["Synthetic"],
"rai:hasSyntheticData": true,
"rai:dataPreprocessingProtocol": "No post-collection preprocessing is applied. Each task's generator emits the five-artifact contract (first_frame.png, prompt.txt, final_frame.png, ground_truth.mp4, question_metadata.json) directly. Sample index, seed, and version are recorded in question_metadata.json so any instance is regenerable from the released code.",
"rai:dataAnnotationProtocol": "No human annotation. The reference rollout (ground_truth.mp4) is computed by the generator's deterministic ground-truth solver and is not produced by humans.",
"rai:dataAnnotationPlatform": "N/A (no annotation).",
"rai:dataReleaseMaintenancePlan": "Versioned releases on the Hugging Face Hub. The Croissant file in this repository is the canonical long-term record. Issues and breaking changes will be documented in repository commits and the dataset version field.",
"rai:dataLimitations": [
"Tasks are stylized and synthetic; transfer to unconstrained open-world video is not validated.",
"Visual style is generator-controlled and intentionally simplified to keep symbolic state recoverable; this is not a photorealism benchmark.",
"The CSP family ships 6 tasks but only 3 (Multi-13/14/15 excluded) are used in the human-judging pool described in the companion paper.",
"Reference rollouts encode one valid trajectory per instance; alternative valid trajectories are not enumerated."
],
"rai:dataBiases": [
"Distribution biases inherited from generator parameter ranges: e.g., maze sizes, planning horizons, physics regimes are drawn from fixed bounded distributions and do not span the long tail of real-world settings.",
"Family balance is uniform (6 tasks per family) by design; this is a deliberate evaluation choice and does not reflect natural prevalence of these reasoning patterns.",
"Visual rendering is monocular, planar, and stylized; appearance distribution does not approximate any real-world video corpus.",
"No demographic content is generated; bias along human demographic axes does not apply to this dataset."
],
"rai:personalSensitiveInformation": "None. The dataset contains no personal information, no biometric data, no demographic information, and no human subjects. All visual content is procedurally generated geometric / symbolic / physical scenes.",
"rai:dataUseCases": "Trajectory-level evaluation of image-to-video systems on long-horizon, rule-grounded tasks. Suitable for: (a) blind human pairwise comparison across model families on process correctness, fidelity, and render-quality axes; (b) automated reference-trajectory comparison once parsers for dense video state become reliable; (c) family-localized failure-mode diagnosis. The companion paper validates use case (a). Use cases (b) and (c) are not validated by this release.",
"rai:dataSocialImpact": "Intended for research on reasoning evaluation in video generation. Risks are minimal: the dataset is synthetic, free of personal content, and designed for evaluation rather than deployment. Potential for misuse is low; the most plausible negative impact is overfitting research on stylized rule-checked tasks at the expense of unconstrained video understanding, which we mitigate by clearly scoping the contribution as a complement to (not a replacement for) appearance-centric evaluation."
}
|