Improve dataset card: Add image-to-3d task, update paper link, and enhance usage guide

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +140 -19
README.md CHANGED
@@ -1,32 +1,46 @@
1
  ---
2
  license: cc-by-4.0
3
- tags:
4
- - computer-vision,
5
- - inverse-rendering,
6
- - photometric-stereo,
7
- - computer-graphics,
8
- - display,
9
- - polarization,
10
- - stereo,
11
- - multi-light,
12
- - illumination-multiplexing,
13
- pretty_name: Display Inverse Rendering Dataset
14
  size_categories:
15
  - n<1K
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  papers:
17
- - title: "A Real-world Display Inverse Rendering Dataset"
18
- url: "https://michaelcsj.github.io/DIR/"
19
- homepage: "https://michaelcsj.github.io/DIR/"
20
- repository: "https://github.com/MichaelCSJ/DIR"
21
  ---
 
22
  # Display Inverse Rendering Dataset
23
- - πŸ“„ [Paper (ArXiv)]()
24
  - 🌐 [Project Page](https://michaelcsj.github.io/DIR/)
25
  - πŸ’» [GitHub Repository](https://github.com/MichaelCSJ/DIR)
26
-
27
  ## Introduction
28
  This dataset is created for display inverse rendering, including multi-light stereo images captured by polarization cameras, and GT geometry (pixel-aligned point cloud and surface normals) scanned by high-precision 3D scanner.
29
 
 
 
 
 
 
 
 
 
 
 
 
30
  ## Structure
31
  - DIR-basic: The basic version of the dataset released with the paper. It includes stereo polarized RAW images, RGB images from a reference view, and ground-truth surface normals and point clouds. All images are captured under a multi-light configuration projected through 16Γ—9 superpixels on the display.
32
 
@@ -59,7 +73,61 @@ This dataset is created for display inverse rendering, including multi-light ste
59
  ```
60
 
61
  - DIR-pms: This dataset follows the DiLiGeNT format and has the same composition as **DIR-basic**. It provides multi-light RGB images from the reference view along with related information and the ground-truth normal maps.
62
- -
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
63
  ```
64
  β”œβ”€β”€ A [Suffix (default "PNG")]
65
  β”‚ β”œβ”€β”€'000 - 143.png',
@@ -70,8 +138,61 @@ This dataset is created for display inverse rendering, including multi-light ste
70
  β”‚ β”œβ”€β”€'Normal_gt.mat'
71
  ```
72
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
73
  ## TODO
74
  - [x] ~~Release training code.~~
75
  - [x] ~~Release `Display Inverse Rendering (DIR)` dataset.~~
76
  - [ ] Release EXPANDED version of DIR datset (HDR).
77
- - [ ] Release EXPANDED version of DIR datset (multi-distance).
 
 
 
1
  ---
2
  license: cc-by-4.0
 
 
 
 
 
 
 
 
 
 
 
3
  size_categories:
4
  - n<1K
5
+ pretty_name: Display Inverse Rendering Dataset
6
+ tags:
7
+ - computer-vision
8
+ - inverse-rendering
9
+ - photometric-stereo
10
+ - computer-graphics
11
+ - display
12
+ - polarization
13
+ - stereo
14
+ - multi-light
15
+ - illumination-multiplexing
16
+ task_categories:
17
+ - image-to-3d
18
  papers:
19
+ - title: A Real-world Display Inverse Rendering Dataset
20
+ url: https://huggingface.co/papers/2508.14411
21
+ homepage: https://michaelcsj.github.io/DIR/
22
+ repository: https://github.com/MichaelCSJ/DIR
23
  ---
24
+
25
  # Display Inverse Rendering Dataset
26
+ - πŸ“„ [Paper](https://huggingface.co/papers/2508.14411)
27
  - 🌐 [Project Page](https://michaelcsj.github.io/DIR/)
28
  - πŸ’» [GitHub Repository](https://github.com/MichaelCSJ/DIR)
29
+
30
  ## Introduction
31
  This dataset is created for display inverse rendering, including multi-light stereo images captured by polarization cameras, and GT geometry (pixel-aligned point cloud and surface normals) scanned by high-precision 3D scanner.
32
 
33
+ `DIR dataset` is a dataset for `Display Inverse Rendering (DIR)`. It contains assets captured from an LCD & polarization-camera system.
34
+ * **OLAT Images:** are captured under display superpixels, and can be used to simulate arbitrary display patterns.
35
+ * **GT Geometry:** is scanned with a high-precision 3D scanner.
36
+ * **Lighting information:** We carefully calibrated light direction, non-linearity, and backlight.
37
+ * **Stereo Imaging:** is an optional feature to initialize rough geometry.
38
+
39
+ **Why Display Inverse Rendering?**
40
+ Display inverse rendering uses a monitor as a per-pixel, programmable light source to reconstruct object geometry and reflectance from captured images. Key features include:
41
+ * **Illumination Multiplexing:** encodes multiple lights and reduces demanded a number of inputs.
42
+ * **Leveraging Polarization:** enables diffuse-specular separation based on optics.
43
+
44
  ## Structure
45
  - DIR-basic: The basic version of the dataset released with the paper. It includes stereo polarized RAW images, RGB images from a reference view, and ground-truth surface normals and point clouds. All images are captured under a multi-light configuration projected through 16Γ—9 superpixels on the display.
46
 
 
73
  ```
74
 
75
  - DIR-pms: This dataset follows the DiLiGeNT format and has the same composition as **DIR-basic**. It provides multi-light RGB images from the reference view along with related information and the ground-truth normal maps.
76
+ -
77
+ ```
78
+ β”œβ”€β”€ A [Suffix (default "PNG")]
79
+ β”‚ β”œβ”€β”€'000 - 143.png',
80
+ β”‚ β”œβ”€β”€'filenames.txt',
81
+ β”‚ β”œβ”€β”€'light_directions.txt'
82
+ β”‚ β”œβ”€β”€'light_intensities.txt',
83
+ β”‚ β”œβ”€β”€'mask.png'
84
+ β”‚ β”œβ”€β”€'Normal_gt.mat'
85
+ ```
86
+
87
+ ## Getting Started
88
+
89
+ ### βš™οΈ Installation
90
+
91
+ ```bash
92
+ git clone https://github.com/MichaelCSJ/DIR.git
93
+ cd DIR
94
+ conda env create -f environment.yml
95
+ conda activate DIR
96
+ ```
97
+
98
+ ### πŸ—‚οΈ Dataset Preparation
99
+ Download the [DIR dataset](https://huggingface.co/datasets/SeokjunChoi/display-inverse-rendering-dataset) for perform our display inverse rendering baseline. It consists of 16 real-world objects with diverse shapes and materials under precisely calibrated directional lighting. There are some versions of dataset as **'DIR-basic'**, **'DIR-pms'**, **'DIR-hdr'**, and **'DIR-multi-distance'**.
100
+ * **DIR-basic**: The basic version of the dataset released with the paper. It includes stereo polarized RAW images, RGB images from a reference view, and ground-truth surface normals and point clouds. All images are captured under a multi-light configuration projected through 16Γ—9 superpixels on the display.
101
+ ```
102
+ β”œβ”€β”€ A
103
+ β”‚ β”œβ”€β”€GT_geometry (for reference(main) view)
104
+ β”‚ β”‚ β”œβ”€β”€'normal.npy',
105
+ β”‚ β”‚ β”œβ”€β”€'normal.png',
106
+ β”‚ β”‚ β”œβ”€β”€'point_cloud_gt.npy'
107
+ β”‚ β”œβ”€β”€main
108
+ β”‚ β”‚ β”œβ”€β”€diffuseNspecular
109
+ β”‚ β”‚ β”‚ β”œβ”€β”€'000 - 143.png',
110
+ β”‚ β”‚ β”‚ β”œβ”€β”€'black.png',
111
+ β”‚ β”‚ β”‚ β”œβ”€β”€'white.png',
112
+ β”‚ β”‚ β”œβ”€β”€RAW_polar
113
+ β”‚ β”‚ β”‚ β”œβ”€β”€'000 - 143_[SHUTTER_TIME(us)].png',
114
+ β”‚ β”‚ β”‚ β”œβ”€β”€'black_[SHUTTER_TIME(us)].png',
115
+ β”‚ β”‚ β”‚ β”œβ”€β”€'white_[SHUTTER_TIME(us)].png',
116
+ β”‚ β”œβ”€β”€side
117
+ β”‚ β”‚ β”œβ”€β”€diffuseNspecular
118
+ β”‚ β”‚ β”‚ β”œβ”€β”€'000 - 143.png',
119
+ β”‚ β”‚ β”‚ β”œβ”€β”€'black.png',
120
+ β”‚ β”‚ β”‚ β”œβ”€β”€'white.png',
121
+ β”‚ β”‚ β”œβ”€β”€RAW_polar
122
+ β”‚ β”‚ β”‚ β”œβ”€β”€'000 - 143_[SHUTTER_TIME(us)].png',
123
+ β”‚ β”‚ β”‚ β”œβ”€β”€'black_[SHUTTER_TIME(us)].png',
124
+ β”‚ β”‚ β”‚ β”œβ”€β”€'white_[SHUTTER_TIME(us)].png',
125
+ β”‚ β”œβ”€β”€'mask.png'
126
+ β”‚ β”œβ”€β”€'point_cloud.npy' (unprojected pixel w.r.t. depth & focal length)
127
+ ```
128
+
129
+ * **DIR-pms**: This dataset follows the [DiLiGenT](https://sites.google.com/site/photometricstereodata/single) format and has the same composition as **DIR-basic**. It provides multi-light RGB images from the reference view along with related information and the ground-truth normal maps.
130
+
131
  ```
132
  β”œβ”€β”€ A [Suffix (default "PNG")]
133
  β”‚ β”œβ”€β”€'000 - 143.png',
 
138
  β”‚ β”œβ”€β”€'Normal_gt.mat'
139
  ```
140
 
141
+ * **DIR-hdr**: TBD.
142
+ * **DIR-multi-distance**: TBD.
143
+
144
+ After downloading, place them under `data/` as the following directory tree.
145
+
146
+ ### πŸ”₯ Normal and basis BRDF Recovery
147
+ To run the baseline, execute `train.py` with the following command:
148
+ ```bash
149
+ python train.py --name YOUR_SESSION_NAME --dataset_root YOUR_DATASET_PATH
150
+ ```
151
+ By default, this code performs inverse rendering using multi-light images captured with an OLAT pattern.
152
+ If you want to use a small number of multi-light images with a multiplexed display pattern, run the code as follows:
153
+ ```bash
154
+ python train.py --name YOUR_SESSION_NAME --dataset_root YOUR_DATASET_PATH --use_multiplexing True --initial_light_pattern YOUR_DISPLAY_PATTERNS
155
+ ```
156
+ You can use [display patterns](https://github.com/MichaelCSJ/DIR/tree/main/patterns) provided by `DDPS` for `YOUR_DISPLAY_PATTERNS`.
157
+ Place display patterns under `patterns/` as the following directory tree.
158
+
159
+ **Lighting Patterns (Initial)**:
160
+ <p align="center">
161
+ <img src="https://github.com/MichaelCSJ/DIR/blob/main/assets/learned_illum_initial.png?raw=true" width="400px">
162
+ </p>
163
+
164
+ **Lighting Patterns (Learned)**:
165
+ <p align="center">
166
+ <img src="https://github.com/MichaelCSJ/DIR/blob/main/assets/learned_illum_optimized.png?raw=true" width="400px">
167
+ </p>
168
+
169
+ Once training is completed, a folder named `YYYYMMDD_HHMMSS` will be created inside the `/results/SESSION` directory, containing the TensorBoard logs, OLAT rendering results, and the fitted parameters for each object.
170
+
171
+ ### πŸ–ΌοΈ Novel Relighting (Optional)
172
+ Run `relighting.py` to render images under novel directional lightings based on recovered normal map and BRDF parameter maps.
173
+ To output .avi video:
174
+ ```bash
175
+ python relighting.py --datadir ./results/YOUR_SESSION_NAME/OBJECT_NAME --format avi
176
+ ```
177
+
178
+ ## Citation
179
+
180
+ If you find this repository useful, please consider citing this paper:
181
+
182
+ ```bibtex
183
+ @inproceedings{choi2025realworld,
184
+ title={A Real-world Display Inverse Rendering Dataset},
185
+ author={Seokjun Choi and Hoon-Gyu Chung and Yujin Jeon and Giljoo Nam and Seung-Hwan Baek},
186
+ booktitle={IEEE/CVF International Conference on Computer Vision (ICCV)},
187
+ year={2025},
188
+ url={https://huggingface.co/papers/2508.14411}
189
+ }
190
+ ```
191
+
192
  ## TODO
193
  - [x] ~~Release training code.~~
194
  - [x] ~~Release `Display Inverse Rendering (DIR)` dataset.~~
195
  - [ ] Release EXPANDED version of DIR datset (HDR).
196
+ - [ ] Release EXPANDED version of DIR datset (multi-distance).
197
+ - [ ] Release additional visualization tools.
198
+ - [ ] Release raw image processing code.