DsLiuCase commited on
Commit
40ea7d4
·
verified ·
1 Parent(s): d42895f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +68 -3
README.md CHANGED
@@ -1,3 +1,68 @@
1
- ---
2
- license: cc-by-nc-nd-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-nd-4.0
3
+ ---
4
+
5
+ # 🧠 Causal3D: A Benchmark for Visual Causal Reasoning
6
+
7
+ **Causal3D** is a comprehensive benchmark designed to evaluate models’ abilities to uncover *latent causal relations* from structured and visual data. This dataset integrates **3D-rendered scenes** with **tabular causal annotations**, providing a unified testbed for advancing *causal discovery*, *causal representation learning*, and *causal reasoning* with **vision-language models (VLMs)** and **large language models (LLMs)**.
8
+
9
+ ---
10
+
11
+ ## 📌 Overview
12
+
13
+ While recent progress in AI and computer vision has been remarkable, there remains a major gap in evaluating causal reasoning over complex visual inputs. **Causal3D** bridges this gap by providing:
14
+
15
+ - **19 curated 3D-scene datasets** simulating diverse real-world causal phenomena.
16
+ - Paired **tabular causal graphs** and **image observations** across multiple views and backgrounds.
17
+ - Benchmarks for evaluating models in both **structured** (tabular) and **unstructured** (image) modalities.
18
+
19
+ ---
20
+
21
+ ## 🧩 Dataset Structure
22
+
23
+ Each sub-dataset (scene) contains:
24
+
25
+ - `images/`: Rendered images under different camera views and backgrounds.
26
+ - `metadata.csv`: Instance-level annotations including object attributes and positions.
27
+ - `causal_graph.json`: Ground-truth causal structure (as adjacency matrix or graph).
28
+ - `view_info.json`: Camera/viewpoint metadata.
29
+ - `split.json`: Recommended train/val/test splits for benchmarking.
30
+
31
+ ---
32
+
33
+ ## 🎯 Evaluation Tasks
34
+
35
+ Causal3D supports a range of causal reasoning tasks, including:
36
+
37
+ - **Causal discovery** from image sequences or tables
38
+ - **Intervention prediction** under modified object states or backgrounds
39
+ - **Counterfactual reasoning** across views
40
+ - **VLM-based causal inference** given multimodal prompts
41
+
42
+ ---
43
+
44
+ ## 📊 Benchmark Results
45
+
46
+ We evaluate a diverse set of methods:
47
+
48
+ - **Classical causal discovery**: PC, GES, NOTEARS
49
+ - **Causal representation learning**: CausalVAE, ICM-based encoders
50
+ - **Vision-Language and Large Language Models**: GPT-4V, Claude-3.5, Gemini-1.5
51
+
52
+ **Key Findings**:
53
+
54
+ - As causal structures grow more complex, **model performance drops significantly** without strong prior assumptions.
55
+ - A noticeable performance gap exists between models trained on structured data and those applied directly to visual inputs.
56
+
57
+ ---
58
+
59
+ <!-- ## 🔍 Example Use Case
60
+
61
+ ```python
62
+ from causal3d import load_scene_data
63
+
64
+ scene = "SpringPendulum"
65
+ data = load_scene_data(scene, split="train")
66
+ images = data["images"]
67
+ metadata = data["table"]
68
+ graph = data["causal_graph"] -->