File size: 3,689 Bytes
f2fb1f5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
382bf9f
 
 
f2fb1f5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
---
license: mit
task_categories:
- image-to-text
- image-text-to-text
language:
- en
tags:
- synthetic
- multimodal
- scene-graph
- spatial-reasoning
- json
- computer-vision
- 3d
pretty_name: TEDRASIM Dataset
size_categories:
- 10K<n<100K
configs:
  - config_name: default
    drop_labels: true
---


# TEDRASIM Dataset

## Dataset Summary

This dataset is a multimodal training corpus for fine-tuning vision-language models to generate structured JSON scene-graph descriptions from rendered images.

This dataset contains:

- a synthetic dataset of 10,000 scenes, each rendered from 2 views (20,000 images total),
- a small real-world dataset of 120 images.

Each example consists of:
- one image showing a object assembly,
- a multi-turn chat-style prompt structure,
- a target JSON string describing the scene in a canonical structured format.

It is intended for research and development on structured visual reasoning, spatial reasoning, scene understanding, and image-to-JSON generation.

--- 

## Data Description

The dataset contains images of a very specific class of toy-like 3D objects.

These objects are:

- composed of geometric primitives such as cubes, sphere, cones,...
- arranged in simple spatial configurations
- rendered from multiple viewpoints

---

## Task Definition

The model is expected to generate a JSON scene graph describing the object by relative spatial relationships between a finite set of known fixed primitives

Relationships are defined locally between touching primitives, for example:

- "the blue cube is behind the green cone"
- "the red cube is left of the orange cylinder"

These relations are encoded explicitly in the JSON structure.

---

## Example Target Representation

A simplified example of a scene description:

```json
{
  "primitive_counts": {
    "red_cube": 2,
    "yellow_sphere": 1
  },
  "primitives": [
    {
      "id": "P1",
      "color": "red",
      "shape": "cube",
      "neighbors": {
        "front": "P2",
        "back": "Empty Space",
        "left": "Empty Space",
        "right": "Empty Space",
        "up": "Empty Space",
        "down": "Empty Space"
      }
    },
    {
      "id": "P2",
      "color": "yellow",
      "shape": "sphere",
      "neighbors": {
        "front": "Empty Space",
        "back": "P1",
        "left": "Empty Space",
        "right": "Empty Space",
        "up": "Empty Space",
        "down": "Empty Space"
      }
    }
  ]
}
```

---

## Dataset Structure

### Data Instances

Each record in the JSONL files has the following structure:

```json
{
  "image": "train/shard_0000/scene_000000_00.png",
  "messages": [
    {"role": "system", "content": "..."},
    {"role": "user", "content": "..."},
    {"role": "assistant", "content": "{\"primitive_counts\": ..., \"primitives\": ...}"}
  ],
  "meta": {
    "scene_id": "scene_000000",
    "scene_hash": "...",
    "split": "train",
    "shard": "shard_0000",
    "view_id": "view_00",
    "num_primitives_in_scene": 4,
    "min_primitives": 1,
    "max_primitives": 6,
    "seed": 42,
    "attempt_index": 1,
    "accepted_index": 0
  }
}
```

### Data Fields

- image: relative path to the rendered image
- messages: chat-style training structure
  - system: task instruction
  - user: input prompt
  - assistant: target JSON 
- meta: auxiliary metadata for traceability

---

## Splits

The synthetic dataset is divided into:

- train.jsonl
- val.jsonl
- test.jsonl

The real dataset contains validation data only:

- val.jsonl


---

## Repository Layout

- synthetic/: synthetic dataset described here
- real/: real-world dataset component 
- train.jsonl / val.jsonl / test.jsonl: split manifests

---

## License

MIT