Datasets:
File size: 3,689 Bytes
f2fb1f5 382bf9f f2fb1f5 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 | ---
license: mit
task_categories:
- image-to-text
- image-text-to-text
language:
- en
tags:
- synthetic
- multimodal
- scene-graph
- spatial-reasoning
- json
- computer-vision
- 3d
pretty_name: TEDRASIM Dataset
size_categories:
- 10K<n<100K
configs:
- config_name: default
drop_labels: true
---
# TEDRASIM Dataset
## Dataset Summary
This dataset is a multimodal training corpus for fine-tuning vision-language models to generate structured JSON scene-graph descriptions from rendered images.
This dataset contains:
- a synthetic dataset of 10,000 scenes, each rendered from 2 views (20,000 images total),
- a small real-world dataset of 120 images.
Each example consists of:
- one image showing a object assembly,
- a multi-turn chat-style prompt structure,
- a target JSON string describing the scene in a canonical structured format.
It is intended for research and development on structured visual reasoning, spatial reasoning, scene understanding, and image-to-JSON generation.
---
## Data Description
The dataset contains images of a very specific class of toy-like 3D objects.
These objects are:
- composed of geometric primitives such as cubes, sphere, cones,...
- arranged in simple spatial configurations
- rendered from multiple viewpoints
---
## Task Definition
The model is expected to generate a JSON scene graph describing the object by relative spatial relationships between a finite set of known fixed primitives
Relationships are defined locally between touching primitives, for example:
- "the blue cube is behind the green cone"
- "the red cube is left of the orange cylinder"
These relations are encoded explicitly in the JSON structure.
---
## Example Target Representation
A simplified example of a scene description:
```json
{
"primitive_counts": {
"red_cube": 2,
"yellow_sphere": 1
},
"primitives": [
{
"id": "P1",
"color": "red",
"shape": "cube",
"neighbors": {
"front": "P2",
"back": "Empty Space",
"left": "Empty Space",
"right": "Empty Space",
"up": "Empty Space",
"down": "Empty Space"
}
},
{
"id": "P2",
"color": "yellow",
"shape": "sphere",
"neighbors": {
"front": "Empty Space",
"back": "P1",
"left": "Empty Space",
"right": "Empty Space",
"up": "Empty Space",
"down": "Empty Space"
}
}
]
}
```
---
## Dataset Structure
### Data Instances
Each record in the JSONL files has the following structure:
```json
{
"image": "train/shard_0000/scene_000000_00.png",
"messages": [
{"role": "system", "content": "..."},
{"role": "user", "content": "..."},
{"role": "assistant", "content": "{\"primitive_counts\": ..., \"primitives\": ...}"}
],
"meta": {
"scene_id": "scene_000000",
"scene_hash": "...",
"split": "train",
"shard": "shard_0000",
"view_id": "view_00",
"num_primitives_in_scene": 4,
"min_primitives": 1,
"max_primitives": 6,
"seed": 42,
"attempt_index": 1,
"accepted_index": 0
}
}
```
### Data Fields
- image: relative path to the rendered image
- messages: chat-style training structure
- system: task instruction
- user: input prompt
- assistant: target JSON
- meta: auxiliary metadata for traceability
---
## Splits
The synthetic dataset is divided into:
- train.jsonl
- val.jsonl
- test.jsonl
The real dataset contains validation data only:
- val.jsonl
---
## Repository Layout
- synthetic/: synthetic dataset described here
- real/: real-world dataset component
- train.jsonl / val.jsonl / test.jsonl: split manifests
---
## License
MIT
|