| { |
| "@context": { |
| "@language": "en", |
| "@vocab": "https://schema.org/", |
| "citeAs": "cr:citeAs", |
| "column": "cr:column", |
| "conformsTo": "dct:conformsTo", |
| "cr": "http://mlcommons.org/croissant/", |
| "rai": "http://mlcommons.org/croissant/RAI/", |
| "data": {"@id": "cr:data", "@type": "@json"}, |
| "dataType": {"@id": "cr:dataType", "@type": "@vocab"}, |
| "dct": "http://purl.org/dc/terms/", |
| "examples": {"@id": "cr:examples", "@type": "@json"}, |
| "extract": "cr:extract", |
| "field": "cr:field", |
| "fileProperty": "cr:fileProperty", |
| "fileObject": "cr:fileObject", |
| "fileSet": "cr:fileSet", |
| "format": "cr:format", |
| "includes": "cr:includes", |
| "isLiveDataset": "cr:isLiveDataset", |
| "jsonPath": "cr:jsonPath", |
| "key": "cr:key", |
| "md5": "cr:md5", |
| "parentField": "cr:parentField", |
| "path": "cr:path", |
| "recordSet": "cr:recordSet", |
| "references": "cr:references", |
| "regex": "cr:regex", |
| "repeated": "cr:repeated", |
| "replace": "cr:replace", |
| "sc": "https://schema.org/", |
| "separator": "cr:separator", |
| "source": "cr:source", |
| "subField": "cr:subField", |
| "transform": "cr:transform" |
| }, |
| "@type": "sc:Dataset", |
| "name": "VBVR-MultiStep", |
| "conformsTo": "http://mlcommons.org/croissant/1.0", |
| "description": "The ~360,000-sample programmatic training corpus for long-horizon multi-step image-to-video reasoning. 36 parameterized tasks across six reasoning families (Navigation, Planning, CSP, Execution, Geometry, Physics). Distributed as 7,200 tar.gz shards (≈50 samples per shard) plus Parquet metadata; each instance follows a five-artifact contract identical to the VBVR-MultiStep-Bench evaluation split.", |
| "alternateName": ["VBVR-MultiStep Training Corpus"], |
| "creator": { |
| "@type": "sc:Organization", |
| "name": "Video-Reason", |
| "url": "https://video-reason.com" |
| }, |
| "datePublished": "2026-05-06", |
| "keywords": ["video reasoning", "multi-step reasoning", "long-horizon", "image-to-video", "training", "synthetic", "tar.gz", "parquet"], |
| "license": "https://creativecommons.org/licenses/by/4.0/", |
| "url": "https://huggingface.co/datasets/Video-Reason/VBVR-MultiStep", |
| "version": "1.0.0", |
| "isLiveDataset": false, |
| "rai:dataCollection": "Fully synthetic. Each of the 36 tasks ships a deterministic generator that emits the five-artifact contract (first_frame.png, prompt.txt, final_frame.png, ground_truth.mp4, question_metadata.json) from a (task, seed) pair. No scraping, no human subjects, no third-party media, no manual annotation. The training corpus is partitioned into disjoint seed bands (1–5,000 and 5,001–10,000 per task) that are themselves disjoint from the evaluation seeds released in VBVR-MultiStep-Bench.", |
| "rai:dataCollectionType": ["Synthetic"], |
| "rai:hasSyntheticData": true, |
| "rai:dataPreprocessingProtocol": "Per-task generator output is grouped into 50-sample batches and packed into tar.gz shards under questions/. Per-task Parquet metadata files (data/metadata_shards/) and a global metadata.parquet index every instance with task id, family, seed, and per-task fields. No sample is filtered, dropped, or transformed after generation.", |
| "rai:dataAnnotationProtocol": "No human annotation. ground_truth.mp4 is rendered by each task's deterministic ground-truth solver.", |
| "rai:dataAnnotationPlatform": "N/A (no annotation).", |
| "rai:dataReleaseMaintenancePlan": "Versioned releases on the Hugging Face Hub. The Croissant file in this repository is the canonical long-term record. A 5 GB representative subset is provided under sample/ for quick inspection and reviewer convenience.", |
| "rai:dataLimitations": [ |
| "Synthetic and stylized: transfer to unconstrained open-world video is not validated.", |
| "Visual rendering is intentionally simplified to keep the symbolic state recoverable from frames; this is not a photorealism corpus.", |
| "Per-task generator parameter ranges are bounded (e.g., maze sizes, planning horizons, physics regimes); the corpus does not span the long tail of any single family.", |
| "Reference rollouts encode one valid trajectory per instance; alternative valid trajectories are not enumerated.", |
| "Although 36 tasks ship, only 34 are used in the training experiments described in the companion paper; this release contains all 36 task families." |
| ], |
| "rai:dataBiases": [ |
| "Family balance is uniform (6 tasks per family) by design and does not reflect natural prevalence of these reasoning patterns.", |
| "Generator parameters bias the difficulty distribution toward bounded and seed-controlled regimes that are amenable to symbolic ground truth; rare or open-ended cases are out of scope.", |
| "Visual style is monocular, planar, and rendered by a fixed family of renderers; appearance distribution does not approximate any real-world video corpus and does not contain demographic content.", |
| "No human demographic information is generated; bias along human demographic axes does not apply." |
| ], |
| "rai:personalSensitiveInformation": "None. The dataset contains no personal information, no biometric data, no demographic information, and no human subjects. All visual content is procedurally generated geometric, symbolic, or physical scenes.", |
| "rai:dataUseCases": "Training image-to-video systems on long-horizon multi-step reasoning under explicit per-step rules. Validated use case in the companion paper: fine-tuning Wan2.2-I2V-A14B (Apache-2.0) with Dual-DiT two-phase LoRA. Out-of-scope: production VLM pretraining at scale, real-world video generation, or any safety-critical use.", |
| "rai:dataSocialImpact": "Intended for academic research on reasoning evaluation in video generation. Risks are minimal: the dataset is synthetic, free of personal content, and rendered in a stylized regime not representative of any real population. The most plausible concern is research-direction effects (e.g., over-investing in stylized synthetic benchmarks), which we mitigate by positioning this corpus as a complement to (not a replacement for) appearance-centric and real-world video corpora.", |
| "rai:dataReleaseUpdate": "If post-release errors are discovered, fixes will be published as additive shards or replacement Parquet entries in a new dataset version, with the prior version retained at its commit hash for backwards reproducibility." |
| } |
|
|