File size: 2,579 Bytes
40ea7d4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3ebb330
40ea7d4
 
3ebb330
40ea7d4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
---
license: cc-by-nc-nd-4.0
---

# 🧠 Causal3D: A Benchmark for Visual Causal Reasoning

**Causal3D** is a comprehensive benchmark designed to evaluate models’ abilities to uncover *latent causal relations* from structured and visual data. This dataset integrates **3D-rendered scenes** with **tabular causal annotations**, providing a unified testbed for advancing *causal discovery*, *causal representation learning*, and *causal reasoning* with **vision-language models (VLMs)** and **large language models (LLMs)**.

---

## 📌 Overview

While recent progress in AI and computer vision has been remarkable, there remains a major gap in evaluating causal reasoning over complex visual inputs. **Causal3D** bridges this gap by providing:

- **19 curated 3D-scene datasets** simulating diverse real-world causal phenomena.
- Paired **tabular causal graphs** and **image observations** across multiple views and backgrounds.
- Benchmarks for evaluating models in both **structured** (tabular) and **unstructured** (image) modalities.

---

## 🧩 Dataset Structure

Each sub-dataset (scene) contains:

- `images/`: Rendered images under different camera views and backgrounds.
- `metadata.csv`: Instance-level annotations including object attributes and positions.
<!-- - `causal_graph.json`: Ground-truth causal structure (as adjacency matrix or graph).
- `view_info.json`: Camera/viewpoint metadata.
- `split.json`: Recommended train/val/test splits for benchmarking.
 -->
---

## 🎯 Evaluation Tasks

Causal3D supports a range of causal reasoning tasks, including:

- **Causal discovery** from image sequences or tables
- **Intervention prediction** under modified object states or backgrounds
- **Counterfactual reasoning** across views
- **VLM-based causal inference** given multimodal prompts

---

## 📊 Benchmark Results

We evaluate a diverse set of methods:

- **Classical causal discovery**: PC, GES, NOTEARS
- **Causal representation learning**: CausalVAE, ICM-based encoders
- **Vision-Language and Large Language Models**: GPT-4V, Claude-3.5, Gemini-1.5

**Key Findings**:

- As causal structures grow more complex, **model performance drops significantly** without strong prior assumptions.
- A noticeable performance gap exists between models trained on structured data and those applied directly to visual inputs.

---

<!-- ## 🔍 Example Use Case

```python
from causal3d import load_scene_data

scene = "SpringPendulum"
data = load_scene_data(scene, split="train")
images = data["images"]
metadata = data["table"]
graph = data["causal_graph"] -->