Datasets:

Modalities:
Image
Text
Formats:
text
ArXiv:
Tags:
image
Libraries:
Datasets
License:
Files changed (1) hide show
  1. README.md +111 -0
README.md ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ tags:
4
+ - image
5
+ annotations_creators:
6
+ - expert-generated
7
+ pretty_name: MVSCPS
8
+ size_categories:
9
+ - n<1K
10
+ task_categories:
11
+ - image-to-3d
12
+ ---
13
+
14
+ # Neural Multi-View Self-Calibrated Photometric Stereo without Photometric Stereo Cues
15
+
16
+ This dataset contains multi-view One-Light-at-a-Time (OLAT) images captured for multi-view 3D reconstruction and inverse rendering tasks.
17
+
18
+ It is first introduced and used for qualitative evaluation in our ICCV 2025 paper.
19
+
20
+ ## Dataset Structure
21
+ The dataset contains 6 scenes. Each scene directory is organized as follows:
22
+
23
+ ```
24
+ {capture_date}_{material}_{name}/
25
+ ├── ARW/
26
+ ├── JPG/
27
+ ├── mask/
28
+ ├── CAM/
29
+ ├── images_overview.jpeg
30
+ └── mesh_RC.obj
31
+
32
+ ```
33
+
34
+ ### File and Folder Descriptions
35
+
36
+ - 🏞️ **ARW/** and **JPG/**
37
+
38
+ Contain raw images in ARW format and corresponding JPEGs captured by a SONY A7R5 camera.
39
+ No post-processing (e.g., tone mapping or white balance) has been applied.
40
+ File names follow the format `V{view_idx}L{light_idx}`, where:
41
+ - `view_idx`: index of the camera view
42
+ - `light_idx`: index of the light direction
43
+
44
+ Images with the same `light_idx` are captured under the same light source **fixed with respect to the camera**.
45
+
46
+ - ◘ **mask/**
47
+ Contains per-view foreground masks, automatically segmented using [SAM2](https://github.com/facebookresearch/sam2).
48
+
49
+ - 📸 **CAM/**
50
+ Includes camera intrinsic and extrinsic parameters as 3×4 projection matrices $P=K[R \mid t]$ for each view.
51
+ Calibrated using [RealityCapture](https://www.realityscan.com/en-US) (Now renamed RealityScan).
52
+
53
+ - 📱**images_overview.jpeg**
54
+ Provides a visual overview of all images in a scene.
55
+ Each sub-image is cropped from the jpeg image and gamma-corrected for improved visibility.
56
+
57
+ - 🧱 **mesh_RC.obj**
58
+ A reference 3D mesh reconstructed from all JPEG images using [RealityCapture](https://www.realityscan.com/en-US).
59
+ ⚠️ **Note:** This mesh is only intended for qualitative reference and should not be used for quantitative evaluation.
60
+ Because the photogrammetry process was applied to images captured under varying directional lighting, it violates the photo-consistency assumption and introduce reconstruction artifacts.
61
+
62
+
63
+ ### Scene Descriptions
64
+
65
+ - `20250205_ceramic_buddha`
66
+
67
+ The first scene captured using the camera-light setup shown in main paper's Fig. 10, and is used in paper's teaser image.
68
+ 143 out of 144 images were successfully calibrated using RealityCapture; thus, only 143 images were used in our method.
69
+
70
+ - `20250303 ~ 20250304`
71
+
72
+ Captured right before ICCV submission deadline for additional qualitative results. Each scene contains 144 successfully calibrated images.
73
+
74
+ - `20250515_ceramic_buddha`
75
+
76
+ Captured during the rebuttal phase to demonstrate that a completely dark room is not strictly required for our method.
77
+ All images were taken with a 1/200-second exposure under ambient lighting, in contrast to previous scenes captured at 1/8-second exposure in darkness.
78
+ Because the strobe flashlight emits a short burst of high-intensity light,
79
+ reducing the exposure time from 1/8 sec to 1/200 sec does not affect the flashlight’s contribution to image appearance.
80
+ However, it substantially attenuates the influence of ambient light, making it negligible.
81
+
82
+
83
+ #### Data Collection & Attribution
84
+
85
+ All images were captured, calibrated, segmented, and curated by [Xu Cao](https://xucao-42.github.io/homepage/).
86
+
87
+ ## How to Download
88
+ Step 1: Install the Hugging Face CLI
89
+ ```
90
+ pip install -U huggingface_hub
91
+ ```
92
+ Step 2: Download the dataset
93
+ ```
94
+ huggingface-cli download cyberagent/mvscps --repo-type dataset --local-dir ./mvscps_data
95
+ ```
96
+
97
+
98
+ ## License
99
+ This dataset is available under the **Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) License**.
100
+ You may use, share, and adapt the material for non-commercial purposes with appropriate credit.
101
+
102
+ ## Citation
103
+ If you use this dataset in your work, please cite:
104
+ ```
105
+ @inproceedings{mvscps2025cao,
106
+ title = {Neural Multi-View Self-Calibrated Photometric Stereo without Photometric Stereo Cues},
107
+ author = {Cao, Xu and Taketomi, Takafumi},
108
+ year = {2025},
109
+ booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
110
+ }
111
+ ```