Update README.md
#1
by
sokumura
- opened
README.md
CHANGED
|
@@ -1,53 +1,53 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: cc-by-nc-sa-4.0
|
| 3 |
-
---
|
| 4 |
-
|
| 5 |
-
# ToF-360 Dataset
|
| 6 |
-
|
| 7 |
-

|
| 8 |
-
|
| 9 |
-
## Overview
|
| 10 |
-
The ToF-360 dataset consists of spherical RGB-D images with instance-level semantic and room ayout annotations, which include 4 unique scenes. It contains 179 equirectangular RGB images along with the corresponding depths, surface normals, XYZ images, and HHA images, labeled with building-defining object categories and image based layout boundaries (ceiling-wall, wall-floor). The dataset enables development of scene understanding tasks based on single-shot reconstruction without the need for global alignment in indoor spaces.
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
## Dataset Modalities
|
| 14 |
-
Each scenes in the dataset has its own folder in the dataset. All the modalities and metadata for each area are contained in that folder as `<scene>/<modality>`.
|
| 15 |
-
|
| 16 |
-
**HHA images:**
|
| 17 |
-
We followed [[Depth2HHA-python]](https://github.com/charlesCXK/Depth2HHA-python) to create it.
|
| 18 |
-
|
| 19 |
-
**RGB images:**
|
| 20 |
-
RGB images contain equirectangular 24-bit color images converted from raw dual fisheye image.
|
| 21 |
-
|
| 22 |
-
**Manhattan aligned RGB images:**
|
| 23 |
-
We followed [[LGT-Net]](https://github.com/zhigangjiang/LGT-Net) to create Manhattan aligned RGB images.
|
| 24 |
-
|
| 25 |
-
**XYZ images:**
|
| 26 |
-
XYZ images are saved as `.npy` binary file format in NumPy. It contains pixel-aligned set of data points in space with a sensitivity of mm. It must be the size of (Height, Width, 3[xyz]).
|
| 27 |
-
|
| 28 |
-
**Annotation:**
|
| 29 |
-
|
| 30 |
-
**depth:**
|
| 31 |
-
Depth images are stored as 16-bit PNGs having a maximum depth of 128m and a sensitivity of 1/512m. Missing values are encoded with the value 0. Note that while depth is defined as the distance from the point-center of the camera in the panoramics.
|
| 32 |
-
|
| 33 |
-
**Room layout annotation:**
|
| 34 |
-
Room layout annotations are stored as same json format as [PanoAnnotator](https://github.com/SunDaDenny/PanoAnnotator). Please refer to this repo for more details.
|
| 35 |
-
|
| 36 |
-
**Normal images:**
|
| 37 |
-
Normals are 127.5-centered per-channel surface normal images. The normal vector is saved as 24-bit RGB PNGs where Red is the horizontal value (more red to the right), Green is vertical (more green downwards), and Blue is towards the camera. It is computed by [normal estimation function](https://www.open3d.org/docs/0.7.0/python_api/open3d.geometry.estimate_normals.html) in [Open3D](https://github.com/isl-org/Open3D). The tool for creating normal images from 3D is located in the `assets/compute_normal.py`
|
| 38 |
-
|
| 39 |
-
## Tools
|
| 40 |
-
This repository provides some basic tools for interacting with the dataset and how to get preprocessed data. The tools are located in the `assets/preprocessing` folder.
|
| 41 |
-
|
| 42 |
-
## Evaluation
|
| 43 |
-
**Semantic segmentation (image-based):**
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
**Semantic segmentation (pointcloud-based):**
|
| 47 |
-
|
| 48 |
-
|
| 49 |
-
**Layout estimation:**
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
## Citations
|
| 53 |
Coming soon...
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: cc-by-nc-sa-4.0
|
| 3 |
+
---
|
| 4 |
+
|
| 5 |
+
# ToF-360 Dataset (test)
|
| 6 |
+
|
| 7 |
+

|
| 8 |
+
|
| 9 |
+
## Overview
|
| 10 |
+
The ToF-360 dataset consists of spherical RGB-D images with instance-level semantic and room ayout annotations, which include 4 unique scenes. It contains 179 equirectangular RGB images along with the corresponding depths, surface normals, XYZ images, and HHA images, labeled with building-defining object categories and image based layout boundaries (ceiling-wall, wall-floor). The dataset enables development of scene understanding tasks based on single-shot reconstruction without the need for global alignment in indoor spaces.
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
## Dataset Modalities
|
| 14 |
+
Each scenes in the dataset has its own folder in the dataset. All the modalities and metadata for each area are contained in that folder as `<scene>/<modality>`.
|
| 15 |
+
|
| 16 |
+
**HHA images:**
|
| 17 |
+
We followed [[Depth2HHA-python]](https://github.com/charlesCXK/Depth2HHA-python) to create it.
|
| 18 |
+
|
| 19 |
+
**RGB images:**
|
| 20 |
+
RGB images contain equirectangular 24-bit color images converted from raw dual fisheye image.
|
| 21 |
+
|
| 22 |
+
**Manhattan aligned RGB images:**
|
| 23 |
+
We followed [[LGT-Net]](https://github.com/zhigangjiang/LGT-Net) to create Manhattan aligned RGB images.
|
| 24 |
+
|
| 25 |
+
**XYZ images:**
|
| 26 |
+
XYZ images are saved as `.npy` binary file format in NumPy. It contains pixel-aligned set of data points in space with a sensitivity of mm. It must be the size of (Height, Width, 3[xyz]).
|
| 27 |
+
|
| 28 |
+
**Annotation:**
|
| 29 |
+
|
| 30 |
+
**depth:**
|
| 31 |
+
Depth images are stored as 16-bit PNGs having a maximum depth of 128m and a sensitivity of 1/512m. Missing values are encoded with the value 0. Note that while depth is defined as the distance from the point-center of the camera in the panoramics.
|
| 32 |
+
|
| 33 |
+
**Room layout annotation:**
|
| 34 |
+
Room layout annotations are stored as same json format as [PanoAnnotator](https://github.com/SunDaDenny/PanoAnnotator). Please refer to this repo for more details.
|
| 35 |
+
|
| 36 |
+
**Normal images:**
|
| 37 |
+
Normals are 127.5-centered per-channel surface normal images. The normal vector is saved as 24-bit RGB PNGs where Red is the horizontal value (more red to the right), Green is vertical (more green downwards), and Blue is towards the camera. It is computed by [normal estimation function](https://www.open3d.org/docs/0.7.0/python_api/open3d.geometry.estimate_normals.html) in [Open3D](https://github.com/isl-org/Open3D). The tool for creating normal images from 3D is located in the `assets/compute_normal.py`
|
| 38 |
+
|
| 39 |
+
## Tools
|
| 40 |
+
This repository provides some basic tools for interacting with the dataset and how to get preprocessed data. The tools are located in the `assets/preprocessing` folder.
|
| 41 |
+
|
| 42 |
+
## Evaluation
|
| 43 |
+
**Semantic segmentation (image-based):**
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
**Semantic segmentation (pointcloud-based):**
|
| 47 |
+
|
| 48 |
+
|
| 49 |
+
**Layout estimation:**
|
| 50 |
+
|
| 51 |
+
|
| 52 |
+
## Citations
|
| 53 |
Coming soon...
|