Datasets:
File size: 5,922 Bytes
dc7ac93 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 | ---
configs:
- config_name: all
default: true
data_files:
- split: train
path: data/all/train*.parquet
- split: test
path: data/all/test*.parquet
- config_name: qrr
data_files:
- split: train
path: data/qrr/train*.parquet
- split: test
path: data/qrr/test*.parquet
- config_name: trr
data_files:
- split: train
path: data/trr/train*.parquet
- split: test
path: data/trr/test*.parquet
- config_name: fdr
data_files:
- split: train
path: data/fdr/train*.parquet
- split: test
path: data/fdr/test*.parquet
task_categories:
- visual-question-answering
language:
- en
license: mit
tags:
- spatial-reasoning
- vlm-benchmark
- ordinal-relations
- 3d-scenes
- multi-view
size_categories:
- 100K<n<1M
---
# ORDINARY-BENCH Multi-View Dataset
A multi-view version of the ORDINARY-BENCH benchmark for evaluating Vision-Language Models (VLMs) on **ordinal spatial reasoning** in 3D scenes. Each sample includes **4 camera views** of the same scene.
> Single-view version: [TYTSTQ/ordinary-bench](https://huggingface.co/datasets/TYTSTQ/ordinary-bench)
>
> Source code & evaluation pipeline: [GitHub - tasd12-ty/ordinary-bench-core](https://github.com/tasd12-ty/ordinary-bench-core)
## Overview
| | |
|---|---|
| Scenes | 700 synthetic 3D scenes (Blender, CLEVR-style) |
| Complexity | 7 levels: 4 to 10 objects per scene (100 each) |
| Questions | 332,857 total across 3 reasoning types |
| Images | 4 views per scene (480 x 320 PNG each) |
## Question Types
### QRR (Quantitative Relation Reasoning) -- 130,557 questions
Compare 3D distances between object pairs. Two variants:
- **Disjoint**: Is `dist(A,B)` less than, approximately equal to, or greater than `dist(C,D)`?
- **Shared anchor**: From anchor A, is `dist(A,B)` less/equal/greater than `dist(A,C)`?
- **Answer format**: `<`, `~=`, or `>`
### TRR (Ternary Relation Reasoning) -- 197,400 questions
Clock-face direction reasoning:
- Standing at object `ref1`, facing toward object `ref2` (12 o'clock direction)
- What clock hour (1-12) is the `target` object at?
- **Answer format**: integer 1-12
### FDR (Full Distance Ranking) -- 4,900 questions
Given an anchor object, rank all other objects by 3D distance, nearest to farthest.
- **Answer format**: ordered JSON array of object IDs, e.g., `["obj_2", "obj_1", "obj_3"]`
## Quick Start
```python
from datasets import load_dataset
# Load QRR questions (test split)
ds = load_dataset("TYTSTQ/ordinary-bench-multiview", "qrr", split="test")
sample = ds[0]
sample["view_0"] # PIL Image (480x320) - camera view 0
sample["view_1"] # PIL Image - camera view 1
sample["view_2"] # PIL Image - camera view 2
sample["view_3"] # PIL Image - camera view 3
sample["question_text"] # "Compare the distance between obj_0 and obj_1 vs ..."
sample["qrr_gt_comparator"] # Ground truth: "<", "~=", or ">"
# Load all question types
ds_all = load_dataset("TYTSTQ/ordinary-bench-multiview", split="test")
```
## Configs
| Config | Description | Questions |
|--------|-------------|-----------|
| `all` (default) | All 3 question types | 332,857 |
| `qrr` | Distance comparison only | 130,557 |
| `trr` | Clock direction only | 197,400 |
| `fdr` | Distance ranking only | 4,900 |
## Data Splits
| Split | Scenes per complexity | Total scenes | Total questions |
|-------|----------------------|--------------|-----------------|
| train | 80 | 560 | 266,261 |
| test | 20 | 140 | 66,596 |
## Column Schema
### Common columns (all configs)
| Column | Type | Description |
|--------|------|-------------|
| `scene_id` | string | Scene identifier, e.g., `n04_000080` |
| `n_objects` | int | Number of objects in scene (4-10) |
| `split` | string | Complexity split: `n04` through `n10` |
| `view_0` | Image | Camera view 0 (480x320 PNG) |
| `view_1` | Image | Camera view 1 (480x320 PNG) |
| `view_2` | Image | Camera view 2 (480x320 PNG) |
| `view_3` | Image | Camera view 3 (480x320 PNG) |
| `objects` | string | JSON array: `[{"id": "obj_0", "desc": "large brown rubber sphere"}, ...]` |
| `question_type` | string | `qrr`, `trr`, or `fdr` |
| `qid` | string | Question ID, e.g., `qrr_0001` |
| `question_text` | string | Natural language question |
| `scene_metadata` | string | Full scene JSON (3D coordinates, camera parameters, etc.) |
### QRR-specific columns
| Column | Type | Description |
|--------|------|-------------|
| `qrr_variant` | string | `disjoint` or `shared_anchor` |
| `qrr_pair1` | string | JSON: `["obj_0", "obj_1"]` |
| `qrr_pair2` | string | JSON: `["obj_2", "obj_3"]` |
| `qrr_metric` | string | Distance metric, e.g., `dist3D` |
| `qrr_gt_comparator` | string | Ground truth: `<`, `~=`, or `>` |
### TRR-specific columns
| Column | Type | Description |
|--------|------|-------------|
| `trr_target` | string | Target object ID |
| `trr_ref1` | string | Standing position object |
| `trr_ref2` | string | 12 o'clock facing direction object |
| `trr_gt_hour` | int | Ground truth clock hour (1-12) |
| `trr_gt_quadrant` | int | Ground truth quadrant (1-4) |
| `trr_gt_angle_deg` | float | Ground truth angle in degrees |
### FDR-specific columns
| Column | Type | Description |
|--------|------|-------------|
| `fdr_anchor` | string | Anchor object ID |
| `fdr_n_ranked` | int | Number of objects to rank |
| `fdr_gt_ranking` | string | JSON: `["obj_2", "obj_1", "obj_3"]` (nearest to farthest) |
| `fdr_gt_distances` | string | JSON: `[3.006, 3.553, 3.882]` |
| `fdr_gt_tie_groups` | string | JSON: `[["obj_2"], ["obj_1", "obj_3"]]` |
## Prompt Templates
System prompts for VLM evaluation are included in `prompts/system_prompts.json`.
## Source Code
**[github.com/tasd12-ty/ordinary-bench-core](https://github.com/tasd12-ty/ordinary-bench-core)**
## License
MIT
|