| | |
| |
|
| | Depth estimation datasets are used to train a model to approximate the relative distance of every pixel in an |
| | image from the camera, also known as depth. The applications enabled by these datasets primarily lie in areas like visual machine |
| | perception and perception in robotics. Example applications include mapping streets for self-driving cars. This guide will show you how to apply transformations |
| | to a depth estimation dataset. |
| |
|
| | Before you start, make sure you have up-to-date versions of `albumentations` installed: |
| |
|
| | ```bash |
| | pip install -U albumentations |
| | ``` |
| |
|
| | [Albumentations](https://albumentations.ai/) is a Python library for performing data augmentation |
| | for computer vision. It supports various computer vision tasks such as image classification, object |
| | detection, segmentation, and keypoint estimation. |
| |
|
| | This guide uses the [NYU Depth V2](https://huggingface.co/datasets/sayakpaul/nyu_depth_v2) dataset which is |
| | comprised of video sequences from various indoor scenes, recorded by RGB and depth cameras. The dataset consists of scenes from 3 cities and provides images along with |
| | their depth maps as labels. |
| |
|
| | Load the `train` split of the dataset and take a look at an example: |
| |
|
| | ```py |
| | >>> from datasets import load_dataset |
| |
|
| | >>> train_dataset = load_dataset("sayakpaul/nyu_depth_v2", split="train") |
| | >>> index = 17 |
| | >>> example = train_dataset[index] |
| | >>> example |
| | {'image': <PIL.PngImagePlugin.PngImageFile image mode=RGB size=640x480>, |
| | 'depth_map': <PIL.TiffImagePlugin.TiffImageFile image mode=F size=640x480>} |
| | ``` |
| |
|
| | The dataset has two fields: |
| |
|
| | * `image`: a PIL PNG image object with `uint8` data type. |
| | * `depth_map`: a PIL Tiff image object with `float32` data type which is the depth map of the image. |
| |
|
| | It is mention-worthy that JPEG/PNG format can only store `uint8` or `uint16` data. As the depth map is `float32` data, it can't be stored using PNG/JPEG. However, we can save the depth map using TIFF format as it supports a wider range of data types, including `float32` data. |
| | |
| | Next, check out an image with: |
| | |
| | ```py |
| | >>> example["image"] |
| | ``` |
| | |
| | <div class="flex justify-center"> |
| | <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets/depth_est_sample.png"> |
| | </div> |
| | |
| | Before we look at the depth map, we need to first convert its data type to `uint8` using `.convert('RGB')` as PIL can't display `float32` images. Now take a look at its corresponding depth map: |
| |
|
| | ```py |
| | >>> example["depth_map"].convert("RGB") |
| | ``` |
| |
|
| | <div class="flex justify-center"> |
| | <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets/depth_est_target.png"> |
| | </div> |
| |
|
| | It's all black! You'll need to add some color to the depth map to visualize it properly. To do that, either we can apply color automatically during display using `plt.imshow()` or create a colored depth map using `plt.cm` and then display it. In this example, we have used the latter one, as we can save/write the colored depth map later. (the utility below is taken from the [FastDepth repository](https://github.com/dwofk/fast-depth/blob/master/utils.py)). |
| |
|
| | ```py |
| | >>> import numpy as np |
| | >>> import matplotlib.pyplot as plt |
| |
|
| | >>> cmap = plt.cm.viridis |
| |
|
| | >>> def colored_depthmap(depth, d_min=None, d_max=None): |
| | ... if d_min is None: |
| | ... d_min = np.min(depth) |
| | ... if d_max is None: |
| | ... d_max = np.max(depth) |
| | ... depth_relative = (depth - d_min) / (d_max - d_min) |
| | ... return 255 * cmap(depth_relative)[:,:,:3] |
| |
|
| | >>> def show_depthmap(depth_map): |
| | ... if not isinstance(depth_map, np.ndarray): |
| | ... depth_map = np.array(depth_map) |
| | ... if depth_map.ndim == 3: |
| | ... depth_map = depth_map.squeeze() |
| |
|
| | ... d_min = np.min(depth_map) |
| | ... d_max = np.max(depth_map) |
| | ... depth_map = colored_depthmap(depth_map, d_min, d_max) |
| |
|
| | ... plt.imshow(depth_map.astype("uint8")) |
| | ... plt.axis("off") |
| | ... plt.show() |
| |
|
| | >>> show_depthmap(example["depth_map"]) |
| | ``` |
| |
|
| | <div class="flex justify-center"> |
| | <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets/depth_est_target_viz.png"> |
| | </div> |
| |
|
| | You can also visualize several different images and their corresponding depth maps. |
| |
|
| | ```py |
| | >>> def merge_into_row(input_image, depth_target): |
| | ... if not isinstance(input_image, np.ndarray): |
| | ... input_image = np.array(input_image) |
| | ... |
| | ... d_min = np.min(depth_target) |
| | ... d_max = np.max(depth_target) |
| | ... depth_target_col = colored_depthmap(depth_target, d_min, d_max) |
| | ... img_merge = np.hstack([input_image, depth_target_col]) |
| | ... |
| | ... return img_merge |
| |
|
| | >>> random_indices = np.random.choice(len(train_dataset), 9).tolist() |
| | >>> plt.figure(figsize=(15, 6)) |
| | >>> for i, idx in enumerate(random_indices): |
| | ... example = train_dataset[idx] |
| | ... ax = plt.subplot(3, 3, i + 1) |
| | ... image_viz = merge_into_row( |
| | ... example["image"], example["depth_map"] |
| | ... ) |
| | ... plt.imshow(image_viz.astype("uint8")) |
| | ... plt.axis("off") |
| | ``` |
| |
|
| | <div class="flex justify-center"> |
| | <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets/depth_est_collage.png"> |
| | </div> |
| |
|
| | Now apply some augmentations with `albumentations`. The augmentation transformations include: |
| |
|
| | * Random horizontal flipping |
| | * Random cropping |
| | * Random brightness and contrast |
| | * Random gamma correction |
| | * Random hue saturation |
| |
|
| | ```py |
| | >>> import albumentations as A |
| |
|
| | >>> crop_size = (448, 576) |
| | >>> transforms = [ |
| | ... A.HorizontalFlip(p=0.5), |
| | ... A.RandomCrop(crop_size[0], crop_size[1]), |
| | ... A.RandomBrightnessContrast(), |
| | ... A.RandomGamma(), |
| | ... A.HueSaturationValue() |
| | ... ] |
| | ``` |
| |
|
| | Additionally, define a mapping to better reflect the target key name. |
| |
|
| | ```py |
| | >>> additional_targets = {"depth": "mask"} |
| | >>> aug = A.Compose(transforms=transforms, additional_targets=additional_targets) |
| | ``` |
| |
|
| | With `additional_targets` defined, you can pass the target depth maps to the `depth` argument of `aug` instead of `mask`. You'll notice this change |
| | in the `apply_transforms()` function defined below. |
| | |
| | Create a function to apply the transformation to the images as well as their depth maps: |
| | |
| | ```py |
| | >>> def apply_transforms(examples): |
| | ... transformed_images, transformed_maps = [], [] |
| | ... for image, depth_map in zip(examples["image"], examples["depth_map"]): |
| | ... image, depth_map = np.array(image), np.array(depth_map) |
| | ... transformed = aug(image=image, depth=depth_map) |
| | ... transformed_images.append(transformed["image"]) |
| | ... transformed_maps.append(transformed["depth"]) |
| | ... |
| | ... examples["pixel_values"] = transformed_images |
| | ... examples["labels"] = transformed_maps |
| | ... return examples |
| | ``` |
| | |
| | Use the [`~Dataset.set_transform`] function to apply the transformation on-the-fly to batches of the dataset to consume less disk space: |
| | |
| | ```py |
| | >>> train_dataset.set_transform(apply_transforms) |
| | ``` |
| | |
| | You can verify the transformation worked by indexing into the `pixel_values` and `labels` of an example image: |
| | |
| | ```py |
| | >>> example = train_dataset[index] |
| | |
| | >>> plt.imshow(example["pixel_values"]) |
| | >>> plt.axis("off") |
| | >>> plt.show() |
| | ``` |
| | |
| | <div class="flex justify-center"> |
| | <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets/depth_est_sample_aug.png"> |
| | </div> |
| | |
| | Visualize the same transformation on the image's corresponding depth map: |
| |
|
| | ```py |
| | >>> show_depthmap(example["labels"]) |
| | ``` |
| |
|
| | <div class="flex justify-center"> |
| | <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets/depth_est_target_aug.png"> |
| | </div> |
| |
|
| | You can also visualize multiple training samples reusing the previous `random_indices`: |
| |
|
| | ```py |
| | >>> plt.figure(figsize=(15, 6)) |
| |
|
| | >>> for i, idx in enumerate(random_indices): |
| | ... ax = plt.subplot(3, 3, i + 1) |
| | ... example = train_dataset[idx] |
| | ... image_viz = merge_into_row( |
| | ... example["pixel_values"], example["labels"] |
| | ... ) |
| | ... plt.imshow(image_viz.astype("uint8")) |
| | ... plt.axis("off") |
| | ``` |
| |
|
| | <div class="flex justify-center"> |
| | <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets/depth_est_aug_collage.png"> |
| | </div> |