DsL commited on
Commit
8e499bb
·
1 Parent(s): 44c6896

update readme

Browse files
Files changed (1) hide show
  1. README.md +144 -1
README.md CHANGED
@@ -4,4 +4,147 @@ language:
4
  - en
5
  size_categories:
6
  - 10K<n<100K
7
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  - en
5
  size_categories:
6
  - 10K<n<100K
7
+ ---
8
+
9
+ # 🧠 Causal3D: A Benchmark for Visual Causal Reasoning
10
+
11
+ **Causal3D** is a comprehensive benchmark designed to evaluate models’ abilities to uncover *latent causal relations* from structured and visual data. This dataset integrates **3D-rendered scenes** with **tabular causal annotations**, providing a unified testbed for advancing *causal discovery*, *causal representation learning*, and *causal reasoning* with **vision-language models (VLMs)** and **large language models (LLMs)**.
12
+
13
+ ## 🖼️ Visual Previews
14
+
15
+ Below are example images from different Causal3D scenes:
16
+
17
+ <table>
18
+ <tr>
19
+ <td align="center">
20
+ <img src="https://huggingface.co/datasets/LLDDSS/Causal3D/resolve/main/preview/parabola.png" width="250"/><br/>parabola
21
+ </td>
22
+ <td align="center">
23
+ <img src="https://huggingface.co/datasets/LLDDSS/Causal3D/resolve/main/preview/convex.png" width="250"/><br/>convex
24
+ </td>
25
+ </tr>
26
+ <tr>
27
+ <td align="center">
28
+ <img src="https://huggingface.co/datasets/LLDDSS/Causal3D/resolve/main/preview/magnetic.png" width="200"/><br/>magnetic
29
+ </td>
30
+ <td align="center">
31
+ <img src="https://huggingface.co/datasets/LLDDSS/Causal3D/resolve/main/preview/pendulum.png" width="200"/><br/>pendulum
32
+ </td>
33
+ <td align="center">
34
+ <img src="https://huggingface.co/datasets/LLDDSS/Causal3D/resolve/main/preview/reflection.png" width="200"/><br/>reflection
35
+ </td>
36
+ </tr>
37
+ <tr>
38
+ <td align="center">
39
+ <img src="https://huggingface.co/datasets/LLDDSS/Causal3D/resolve/main/preview/seesaw.png" width="200"/><br/>seesaw
40
+ </td>
41
+ <td align="center">
42
+ <img src="https://huggingface.co/datasets/LLDDSS/Causal3D/resolve/main/preview/spring.png" width="200"/><br/>spring
43
+ </td>
44
+ <td align="center">
45
+ <img src="https://huggingface.co/datasets/LLDDSS/Causal3D/resolve/main/preview/water_flow.png" width="200"/><br/>water_flow
46
+ </td>
47
+ </tr>
48
+ </table>
49
+
50
+ <!-- - `causal_graph.json`: Ground-truth causal structure (as adjacency matrix or graph).
51
+ - `view_info.json`: Camera/viewpoint metadata.
52
+ - `split.json`: Recommended train/val/test splits for benchmarking. -->
53
+
54
+
55
+ ## 📚 Usage
56
+
57
+ #### 🔹 Option 1: Load from Hugging Face
58
+
59
+ You can easily load a specific scene using the Hugging Face `datasets` library:
60
+
61
+ ```python
62
+ from datasets import load_dataset
63
+
64
+ dataset = load_dataset(
65
+ "LLDDSS/Causal3D",
66
+ name="real_scenes_Real_Parabola",
67
+ download_mode="force_redownload", # Optional: force re-download
68
+ trust_remote_code=True # Required for custom dataset loading
69
+ )
70
+
71
+ print(dataset)
72
+ ```
73
+
74
+ #### 🔹 Option 2: Download via [**Kaggle**](https://www.kaggle.com/datasets/dsliu0011/causal3d-image-dataset) + Croissant
75
+ ```python
76
+ import mlcroissant as mlc
77
+ import pandas as pd
78
+
79
+ # Load the dataset metadata from Kaggle
80
+ croissant_dataset = mlc.Dataset(
81
+ "https://www.kaggle.com/datasets/dsliu0011/causal3d-image-dataset/croissant/download"
82
+ )
83
+
84
+ record_sets = croissant_dataset.metadata.record_sets
85
+ print(record_sets)
86
+
87
+ df = pd.DataFrame(croissant_dataset.records(record_set=record_sets[0].uuid))
88
+ print(df.head())
89
+ ```
90
+
91
+ ---
92
+
93
+ ## 📌 Overview
94
+
95
+ While recent progress in AI and computer vision has been remarkable, there remains a major gap in evaluating causal reasoning over complex visual inputs. **Causal3D** bridges this gap by providing:
96
+
97
+ - **19 curated 3D-scene datasets** simulating diverse real-world causal phenomena.
98
+ - Paired **tabular causal graphs** and **image observations** across multiple views and backgrounds.
99
+ - Benchmarks for evaluating models in both **structured** (tabular) and **unstructured** (image) modalities.
100
+
101
+ ---
102
+
103
+ ## 🧩 Dataset Structure
104
+
105
+ Each sub-dataset (scene) contains:
106
+
107
+
108
+ - `images/`: Rendered images under different camera views and backgrounds.
109
+ - `tabular.csv`: Instance-level annotations including object attributes in causal graph.
110
+
111
+ ---
112
+
113
+ ## 🎯 Evaluation Tasks
114
+
115
+ Causal3D supports a range of causal reasoning tasks, including:
116
+
117
+ - **Causal discovery** from image sequences or tables
118
+ - **Intervention prediction** under modified object states or backgrounds
119
+ - **Counterfactual reasoning** across views
120
+ - **VLM-based causal inference** given multimodal prompts
121
+
122
+ ---
123
+
124
+ ## 📊 Benchmark Results
125
+
126
+ We evaluate a diverse set of methods:
127
+
128
+ - **Classical causal discovery**: PC, GES, NOTEARS
129
+ - **Causal representation learning**: CausalVAE, ICM-based encoders
130
+ - **Vision-Language and Large Language Models**: GPT-4V, Claude-3.5, Gemini-1.5
131
+
132
+ **Key Findings**:
133
+
134
+ - As causal structures grow more complex, **model performance drops significantly** without strong prior assumptions.
135
+ - A noticeable performance gap exists between models trained on structured data and those applied directly to visual inputs.
136
+
137
+ ---
138
+
139
+
140
+
141
+ <!-- ## 🔍 Example Use Case
142
+
143
+ ```python
144
+ from causal3d import load_scene_data
145
+
146
+ scene = "SpringPendulum"
147
+ data = load_scene_data(scene, split="train")
148
+ images = data["images"]
149
+ metadata = data["table"]
150
+ graph = data["causal_graph"] -->