|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| """A set of functions that are used for visualization.
|
|
|
| These functions often receive an image, perform some visualization on the image.
|
| The functions do not return a value, instead they modify the image itself.
|
| """
|
| import collections
|
| import functools
|
| from typing import Any, Dict, Optional, List, Union
|
|
|
| from absl import logging
|
|
|
| import matplotlib
|
| matplotlib.use('Agg')
|
| import matplotlib.pyplot as plt
|
| import numpy as np
|
| from PIL import Image
|
| from PIL import ImageColor
|
| from PIL import ImageDraw
|
| from PIL import ImageFont
|
| import six
|
| import tensorflow as tf, tf_keras
|
|
|
| from official.vision.ops import box_ops
|
| from official.vision.ops import preprocess_ops
|
| from official.vision.utils.object_detection import shape_utils
|
|
|
| _TITLE_LEFT_MARGIN = 10
|
| _TITLE_TOP_MARGIN = 10
|
| STANDARD_COLORS = [
|
| 'AliceBlue', 'Chartreuse', 'Aqua', 'Aquamarine', 'Azure', 'Beige', 'Bisque',
|
| 'BlanchedAlmond', 'BlueViolet', 'BurlyWood', 'CadetBlue', 'AntiqueWhite',
|
| 'Chocolate', 'Coral', 'CornflowerBlue', 'Cornsilk', 'Crimson', 'Cyan',
|
| 'DarkCyan', 'DarkGoldenRod', 'DarkGrey', 'DarkKhaki', 'DarkOrange',
|
| 'DarkOrchid', 'DarkSalmon', 'DarkSeaGreen', 'DarkTurquoise', 'DarkViolet',
|
| 'DeepPink', 'DeepSkyBlue', 'DodgerBlue', 'FireBrick', 'FloralWhite',
|
| 'ForestGreen', 'Fuchsia', 'Gainsboro', 'GhostWhite', 'Gold', 'GoldenRod',
|
| 'Salmon', 'Tan', 'HoneyDew', 'HotPink', 'IndianRed', 'Ivory', 'Khaki',
|
| 'Lavender', 'LavenderBlush', 'LawnGreen', 'LemonChiffon', 'LightBlue',
|
| 'LightCoral', 'LightCyan', 'LightGoldenRodYellow', 'LightGray', 'LightGrey',
|
| 'LightGreen', 'LightPink', 'LightSalmon', 'LightSeaGreen', 'LightSkyBlue',
|
| 'LightSlateGray', 'LightSlateGrey', 'LightSteelBlue', 'LightYellow', 'Lime',
|
| 'LimeGreen', 'Linen', 'Magenta', 'MediumAquaMarine', 'MediumOrchid',
|
| 'MediumPurple', 'MediumSeaGreen', 'MediumSlateBlue', 'MediumSpringGreen',
|
| 'MediumTurquoise', 'MediumVioletRed', 'MintCream', 'MistyRose', 'Moccasin',
|
| 'NavajoWhite', 'OldLace', 'Olive', 'OliveDrab', 'Orange', 'OrangeRed',
|
| 'Orchid', 'PaleGoldenRod', 'PaleGreen', 'PaleTurquoise', 'PaleVioletRed',
|
| 'PapayaWhip', 'PeachPuff', 'Peru', 'Pink', 'Plum', 'PowderBlue', 'Purple',
|
| 'Red', 'RosyBrown', 'RoyalBlue', 'SaddleBrown', 'Green', 'SandyBrown',
|
| 'SeaGreen', 'SeaShell', 'Sienna', 'Silver', 'SkyBlue', 'SlateBlue',
|
| 'SlateGray', 'SlateGrey', 'Snow', 'SpringGreen', 'SteelBlue', 'GreenYellow',
|
| 'Teal', 'Thistle', 'Tomato', 'Turquoise', 'Violet', 'Wheat', 'White',
|
| 'WhiteSmoke', 'Yellow', 'YellowGreen'
|
| ]
|
|
|
|
|
| def save_image_array_as_png(image, output_path):
|
| """Saves an image (represented as a numpy array) to PNG.
|
|
|
| Args:
|
| image: a numpy array with shape [height, width, 3].
|
| output_path: path to which image should be written.
|
| """
|
| image_pil = Image.fromarray(np.uint8(image)).convert('RGB')
|
| with tf.io.gfile.GFile(output_path, 'w') as fid:
|
| image_pil.save(fid, 'PNG')
|
|
|
|
|
| def encode_image_array_as_png_str(image):
|
| """Encodes a numpy array into a PNG string.
|
|
|
| Args:
|
| image: a numpy array with shape [height, width, 3].
|
|
|
| Returns:
|
| PNG encoded image string.
|
| """
|
| image_pil = Image.fromarray(np.uint8(image))
|
| output = six.BytesIO()
|
| image_pil.save(output, format='PNG')
|
| png_string = output.getvalue()
|
| output.close()
|
| return png_string
|
|
|
|
|
| def visualize_images_with_bounding_boxes(images, box_outputs, step,
|
| summary_writer):
|
| """Records subset of evaluation images with bounding boxes."""
|
| if not isinstance(images, list):
|
| logging.warning(
|
| 'visualize_images_with_bounding_boxes expects list of '
|
| 'images but received type: %s and value: %s', type(images), images)
|
| return
|
|
|
| image_shape = tf.shape(images[0])
|
| image_height = tf.cast(image_shape[0], tf.float32)
|
| image_width = tf.cast(image_shape[1], tf.float32)
|
| normalized_boxes = box_ops.normalize_boxes(box_outputs,
|
| [image_height, image_width])
|
|
|
| bounding_box_color = tf.constant([[1.0, 1.0, 0.0, 1.0]])
|
| image_summary = tf.image.draw_bounding_boxes(
|
| tf.cast(images, tf.float32), normalized_boxes, bounding_box_color)
|
| with summary_writer.as_default():
|
| tf.summary.image('bounding_box_summary', image_summary, step=step)
|
| summary_writer.flush()
|
|
|
|
|
| def draw_bounding_box_on_image_array(image,
|
| ymin,
|
| xmin,
|
| ymax,
|
| xmax,
|
| color='red',
|
| thickness=4,
|
| display_str_list=(),
|
| use_normalized_coordinates=True):
|
| """Adds a bounding box to an image (numpy array).
|
|
|
| Bounding box coordinates can be specified in either absolute (pixel) or
|
| normalized coordinates by setting the use_normalized_coordinates argument.
|
|
|
| Args:
|
| image: a numpy array with shape [height, width, 3].
|
| ymin: ymin of bounding box.
|
| xmin: xmin of bounding box.
|
| ymax: ymax of bounding box.
|
| xmax: xmax of bounding box.
|
| color: color to draw bounding box. Default is red.
|
| thickness: line thickness. Default value is 4.
|
| display_str_list: list of strings to display in box (each to be shown on its
|
| own line).
|
| use_normalized_coordinates: If True (default), treat coordinates ymin, xmin,
|
| ymax, xmax as relative to the image. Otherwise treat coordinates as
|
| absolute.
|
| """
|
| image_pil = Image.fromarray(np.uint8(image)).convert('RGB')
|
| draw_bounding_box_on_image(image_pil, ymin, xmin, ymax, xmax, color,
|
| thickness, display_str_list,
|
| use_normalized_coordinates)
|
| np.copyto(image, np.array(image_pil))
|
|
|
|
|
| def draw_bounding_box_on_image(image,
|
| ymin,
|
| xmin,
|
| ymax,
|
| xmax,
|
| color='red',
|
| thickness=4,
|
| display_str_list=(),
|
| use_normalized_coordinates=True):
|
| """Adds a bounding box to an image.
|
|
|
| Bounding box coordinates can be specified in either absolute (pixel) or
|
| normalized coordinates by setting the use_normalized_coordinates argument.
|
|
|
| Each string in display_str_list is displayed on a separate line above the
|
| bounding box in black text on a rectangle filled with the input 'color'.
|
| If the top of the bounding box extends to the edge of the image, the strings
|
| are displayed below the bounding box.
|
|
|
| Args:
|
| image: a PIL.Image object.
|
| ymin: ymin of bounding box.
|
| xmin: xmin of bounding box.
|
| ymax: ymax of bounding box.
|
| xmax: xmax of bounding box.
|
| color: color to draw bounding box. Default is red.
|
| thickness: line thickness. Default value is 4.
|
| display_str_list: list of strings to display in box (each to be shown on its
|
| own line).
|
| use_normalized_coordinates: If True (default), treat coordinates ymin, xmin,
|
| ymax, xmax as relative to the image. Otherwise treat coordinates as
|
| absolute.
|
| """
|
| draw = ImageDraw.Draw(image)
|
| im_width, im_height = image.size
|
| if use_normalized_coordinates:
|
| (left, right, top, bottom) = (xmin * im_width, xmax * im_width,
|
| ymin * im_height, ymax * im_height)
|
| else:
|
| (left, right, top, bottom) = (xmin, xmax, ymin, ymax)
|
| draw.line([(left, top), (left, bottom), (right, bottom), (right, top),
|
| (left, top)],
|
| width=thickness,
|
| fill=color)
|
| try:
|
| font = ImageFont.truetype('arial.ttf', 24)
|
| except IOError:
|
| font = ImageFont.load_default()
|
|
|
|
|
|
|
|
|
| if hasattr(font, 'getsize'):
|
| display_str_heights = [font.getsize(ds)[1] for ds in display_str_list]
|
| else:
|
| display_str_heights = [font.getbbox(ds)[3] for ds in display_str_list]
|
|
|
| total_display_str_height = (1 + 2 * 0.05) * sum(display_str_heights)
|
|
|
| if top > total_display_str_height:
|
| text_bottom = top
|
| else:
|
| text_bottom = bottom + total_display_str_height
|
|
|
| for display_str in display_str_list[::-1]:
|
| try:
|
| if hasattr(font, 'getsize'):
|
| text_width, text_height = font.getsize(display_str)
|
| else:
|
| text_width, text_height = font.getbbox(display_str)[2:4]
|
| margin = np.ceil(0.05 * text_height)
|
| draw.rectangle(
|
| [
|
| (left, text_bottom - text_height - 2 * margin),
|
| (left + text_width, text_bottom),
|
| ],
|
| fill=color,
|
| )
|
| draw.text(
|
| (left + margin, text_bottom - text_height - margin),
|
| display_str,
|
| fill='black',
|
| font=font,
|
| )
|
| except ValueError:
|
| pass
|
| text_bottom -= text_height - 2 * margin
|
|
|
|
|
| def draw_bounding_boxes_on_image_array(image,
|
| boxes,
|
| color='red',
|
| thickness=4,
|
| display_str_list_list=()):
|
| """Draws bounding boxes on image (numpy array).
|
|
|
| Args:
|
| image: a numpy array object.
|
| boxes: a 2 dimensional numpy array of [N, 4]: (ymin, xmin, ymax, xmax). The
|
| coordinates are in normalized format between [0, 1].
|
| color: color to draw bounding box. Default is red.
|
| thickness: line thickness. Default value is 4.
|
| display_str_list_list: list of list of strings. a list of strings for each
|
| bounding box. The reason to pass a list of strings for a bounding box is
|
| that it might contain multiple labels.
|
|
|
| Raises:
|
| ValueError: if boxes is not a [N, 4] array
|
| """
|
| image_pil = Image.fromarray(image)
|
| draw_bounding_boxes_on_image(image_pil, boxes, color, thickness,
|
| display_str_list_list)
|
| np.copyto(image, np.array(image_pil))
|
|
|
|
|
| def draw_bounding_boxes_on_image(image,
|
| boxes,
|
| color='red',
|
| thickness=4,
|
| display_str_list_list=()):
|
| """Draws bounding boxes on image.
|
|
|
| Args:
|
| image: a PIL.Image object.
|
| boxes: a 2 dimensional numpy array of [N, 4]: (ymin, xmin, ymax, xmax). The
|
| coordinates are in normalized format between [0, 1].
|
| color: color to draw bounding box. Default is red.
|
| thickness: line thickness. Default value is 4.
|
| display_str_list_list: list of list of strings. a list of strings for each
|
| bounding box. The reason to pass a list of strings for a bounding box is
|
| that it might contain multiple labels.
|
|
|
| Raises:
|
| ValueError: if boxes is not a [N, 4] array
|
| """
|
| boxes_shape = boxes.shape
|
| if not boxes_shape:
|
| return
|
| if len(boxes_shape) != 2 or boxes_shape[1] != 4:
|
| raise ValueError('Input must be of size [N, 4]')
|
| for i in range(boxes_shape[0]):
|
| display_str_list = ()
|
| if display_str_list_list:
|
| display_str_list = display_str_list_list[i]
|
| draw_bounding_box_on_image(image, boxes[i, 0], boxes[i, 1], boxes[i, 2],
|
| boxes[i, 3], color, thickness, display_str_list)
|
|
|
|
|
| def _visualize_boxes(image, boxes, classes, scores, category_index, **kwargs):
|
| return visualize_boxes_and_labels_on_image_array(
|
| image, boxes, classes, scores, category_index=category_index, **kwargs)
|
|
|
|
|
| def _visualize_boxes_and_masks(image, boxes, classes, scores, masks,
|
| category_index, **kwargs):
|
| return visualize_boxes_and_labels_on_image_array(
|
| image,
|
| boxes,
|
| classes,
|
| scores,
|
| category_index=category_index,
|
| instance_masks=masks,
|
| **kwargs)
|
|
|
|
|
| def _visualize_boxes_and_keypoints(image, boxes, classes, scores, keypoints,
|
| category_index, **kwargs):
|
| return visualize_boxes_and_labels_on_image_array(
|
| image,
|
| boxes,
|
| classes,
|
| scores,
|
| category_index=category_index,
|
| keypoints=keypoints,
|
| **kwargs)
|
|
|
|
|
| def _visualize_boxes_and_masks_and_keypoints(image, boxes, classes, scores,
|
| masks, keypoints, category_index,
|
| **kwargs):
|
| return visualize_boxes_and_labels_on_image_array(
|
| image,
|
| boxes,
|
| classes,
|
| scores,
|
| category_index=category_index,
|
| instance_masks=masks,
|
| keypoints=keypoints,
|
| **kwargs)
|
|
|
|
|
| def _resize_original_image(image, image_shape):
|
| image = tf.expand_dims(image, 0)
|
| image = tf.image.resize(
|
| image, image_shape, method=tf.image.ResizeMethod.NEAREST_NEIGHBOR)
|
| return tf.cast(tf.squeeze(image, 0), tf.uint8)
|
|
|
|
|
| def visualize_outputs(
|
| logs,
|
| task_config,
|
| original_image_spatial_shape=None,
|
| true_image_shape=None,
|
| max_boxes_to_draw=20,
|
| min_score_thresh=0.2,
|
| use_normalized_coordinates=False,
|
| image_mean: Optional[Union[float, List[float]]] = None,
|
| image_std: Optional[Union[float, List[float]]] = None,
|
| key: str = 'image/validation_outputs',
|
| ) -> Dict[str, Any]:
|
| """Visualizes the detection outputs.
|
|
|
| It extracts images and predictions from logs and draws visualization on input
|
| images. By default, it requires `detection_boxes`, `detection_classes` and
|
| `detection_scores` in the prediction, and optionally accepts
|
| `detection_keypoints` and `detection_masks`.
|
|
|
| Args:
|
| logs: A dictionaty of log that contains images and predictions.
|
| task_config: A task config.
|
| original_image_spatial_shape: A [N, 2] tensor containing the spatial size of
|
| the original image.
|
| true_image_shape: A [N, 3] tensor containing the spatial size of unpadded
|
| original_image.
|
| max_boxes_to_draw: The maximum number of boxes to draw on an image. Default
|
| 20.
|
| min_score_thresh: The minimum score threshold for visualization. Default
|
| 0.2.
|
| use_normalized_coordinates: Whether to assume boxes and kepoints are in
|
| normalized coordinates (as opposed to absolute coordiantes). Default is
|
| False.
|
| image_mean: An optional float or list of floats used as the mean pixel value
|
| to normalize images.
|
| image_std: An optional float or list of floats used as the std to normalize
|
| images.
|
| key: A string specifying the key of the returned dictionary.
|
|
|
| Returns:
|
| A dictionary of images with visualization drawn on it. Each key corresponds
|
| to a 4D tensor with predictions (boxes, segments and/or keypoints) drawn
|
| on each image.
|
| """
|
| images = logs['image']
|
| boxes = logs['detection_boxes']
|
| classes = tf.cast(logs['detection_classes'], dtype=tf.int32)
|
| scores = logs['detection_scores']
|
| num_classes = task_config.model.num_classes
|
|
|
| keypoints = (
|
| logs['detection_keypoints'] if 'detection_keypoints' in logs else None
|
| )
|
| instance_masks = (
|
| logs['detection_masks'] if 'detection_masks' in logs else None
|
| )
|
|
|
| category_index = {}
|
| for i in range(1, num_classes + 1):
|
| category_index[i] = {'id': i, 'name': str(i)}
|
|
|
| def _denormalize_images(images: tf.Tensor) -> tf.Tensor:
|
| if image_mean is None and image_std is None:
|
| images *= tf.constant(
|
| preprocess_ops.STDDEV_RGB, shape=[1, 1, 3], dtype=images.dtype
|
| )
|
| images += tf.constant(
|
| preprocess_ops.MEAN_RGB, shape=[1, 1, 3], dtype=images.dtype
|
| )
|
| elif image_mean is not None and image_std is not None:
|
| if isinstance(image_mean, float) and isinstance(image_std, float):
|
| images = images * image_std + image_mean
|
| elif isinstance(image_mean, list) and isinstance(image_std, list):
|
| images *= tf.constant(image_std, shape=[1, 1, 3], dtype=images.dtype)
|
| images += tf.constant(image_mean, shape=[1, 1, 3], dtype=images.dtype)
|
| else:
|
| raise ValueError(
|
| '`image_mean` and `image_std` should be the same type.'
|
| )
|
| else:
|
| raise ValueError(
|
| 'Both `image_mean` and `image_std` should be set or None at the same '
|
| 'time.'
|
| )
|
| return tf.cast(images, dtype=tf.uint8)
|
|
|
| images = tf.nest.map_structure(
|
| tf.identity,
|
| tf.map_fn(
|
| _denormalize_images,
|
| elems=images,
|
| fn_output_signature=tf.TensorSpec(
|
| shape=images.shape.as_list()[1:], dtype=tf.uint8
|
| ),
|
| parallel_iterations=32,
|
| ),
|
| )
|
|
|
| images_with_boxes = draw_bounding_boxes_on_image_tensors(
|
| images,
|
| boxes,
|
| classes,
|
| scores,
|
| category_index,
|
| original_image_spatial_shape,
|
| true_image_shape,
|
| instance_masks,
|
| keypoints,
|
| max_boxes_to_draw,
|
| min_score_thresh,
|
| use_normalized_coordinates,
|
| )
|
|
|
| outputs = {}
|
| for i, image in enumerate(images_with_boxes):
|
| outputs[key + f'/{i}'] = image[None, ...]
|
|
|
| return outputs
|
|
|
|
|
| def draw_bounding_boxes_on_image_tensors(images,
|
| boxes,
|
| classes,
|
| scores,
|
| category_index,
|
| original_image_spatial_shape=None,
|
| true_image_shape=None,
|
| instance_masks=None,
|
| keypoints=None,
|
| max_boxes_to_draw=20,
|
| min_score_thresh=0.2,
|
| use_normalized_coordinates=True):
|
| """Draws bounding boxes, masks, and keypoints on batch of image tensors.
|
|
|
| Args:
|
| images: A 4D uint8 image tensor of shape [N, H, W, C]. If C > 3, additional
|
| channels will be ignored. If C = 1, then we convert the images to RGB
|
| images.
|
| boxes: [N, max_detections, 4] float32 tensor of detection boxes.
|
| classes: [N, max_detections] int tensor of detection classes. Note that
|
| classes are 1-indexed.
|
| scores: [N, max_detections] float32 tensor of detection scores.
|
| category_index: a dict that maps integer ids to category dicts. e.g.
|
| {1: {1: 'dog'}, 2: {2: 'cat'}, ...}
|
| original_image_spatial_shape: [N, 2] tensor containing the spatial size of
|
| the original image.
|
| true_image_shape: [N, 3] tensor containing the spatial size of unpadded
|
| original_image.
|
| instance_masks: A 4D uint8 tensor of shape [N, max_detection, H, W] with
|
| instance masks.
|
| keypoints: A 4D float32 tensor of shape [N, max_detection, num_keypoints, 2]
|
| with keypoints.
|
| max_boxes_to_draw: Maximum number of boxes to draw on an image. Default 20.
|
| min_score_thresh: Minimum score threshold for visualization. Default 0.2.
|
| use_normalized_coordinates: Whether to assume boxes and kepoints are in
|
| normalized coordinates (as opposed to absolute coordiantes). Default is
|
| True.
|
|
|
| Returns:
|
| 4D image tensor of type uint8, with boxes drawn on top.
|
| """
|
|
|
| if images.shape[3] > 3:
|
| images = images[:, :, :, 0:3]
|
| elif images.shape[3] == 1:
|
| images = tf.image.grayscale_to_rgb(images)
|
| visualization_keyword_args = {
|
| 'use_normalized_coordinates': use_normalized_coordinates,
|
| 'max_boxes_to_draw': max_boxes_to_draw,
|
| 'min_score_thresh': min_score_thresh,
|
| 'agnostic_mode': False,
|
| 'line_thickness': 4
|
| }
|
| if true_image_shape is None:
|
| true_shapes = tf.constant(-1, shape=[images.shape.as_list()[0], 3])
|
| else:
|
| true_shapes = true_image_shape
|
| if original_image_spatial_shape is None:
|
| original_shapes = tf.constant(-1, shape=[images.shape.as_list()[0], 2])
|
| else:
|
| original_shapes = original_image_spatial_shape
|
|
|
| if instance_masks is not None and keypoints is None:
|
| visualize_boxes_fn = functools.partial(
|
| _visualize_boxes_and_masks,
|
| category_index=category_index,
|
| **visualization_keyword_args)
|
| elems = [
|
| true_shapes, original_shapes, images, boxes, classes, scores,
|
| instance_masks
|
| ]
|
| elif instance_masks is None and keypoints is not None:
|
| visualize_boxes_fn = functools.partial(
|
| _visualize_boxes_and_keypoints,
|
| category_index=category_index,
|
| **visualization_keyword_args)
|
| elems = [
|
| true_shapes, original_shapes, images, boxes, classes, scores, keypoints
|
| ]
|
| elif instance_masks is not None and keypoints is not None:
|
| visualize_boxes_fn = functools.partial(
|
| _visualize_boxes_and_masks_and_keypoints,
|
| category_index=category_index,
|
| **visualization_keyword_args)
|
| elems = [
|
| true_shapes, original_shapes, images, boxes, classes, scores,
|
| instance_masks, keypoints
|
| ]
|
| else:
|
| visualize_boxes_fn = functools.partial(
|
| _visualize_boxes,
|
| category_index=category_index,
|
| **visualization_keyword_args)
|
| elems = [true_shapes, original_shapes, images, boxes, classes, scores]
|
|
|
| def draw_boxes(image_and_detections):
|
| """Draws boxes on image."""
|
| true_shape = image_and_detections[0]
|
| original_shape = image_and_detections[1]
|
| if true_image_shape is not None:
|
| image = shape_utils.pad_or_clip_nd(image_and_detections[2],
|
| [true_shape[0], true_shape[1], 3])
|
| if original_image_spatial_shape is not None:
|
| image_and_detections[2] = _resize_original_image(image, original_shape)
|
|
|
| image_with_boxes = tf.compat.v1.py_func(visualize_boxes_fn,
|
| image_and_detections[2:], tf.uint8)
|
| return image_with_boxes
|
|
|
| images = tf.map_fn(draw_boxes, elems, dtype=tf.uint8, back_prop=False)
|
| return images
|
|
|
|
|
| def draw_keypoints_on_image_array(image,
|
| keypoints,
|
| color='red',
|
| radius=2,
|
| use_normalized_coordinates=True):
|
| """Draws keypoints on an image (numpy array).
|
|
|
| Args:
|
| image: a numpy array with shape [height, width, 3].
|
| keypoints: a numpy array with shape [num_keypoints, 2].
|
| color: color to draw the keypoints with. Default is red.
|
| radius: keypoint radius. Default value is 2.
|
| use_normalized_coordinates: if True (default), treat keypoint values as
|
| relative to the image. Otherwise treat them as absolute.
|
| """
|
| image_pil = Image.fromarray(np.uint8(image)).convert('RGB')
|
| draw_keypoints_on_image(image_pil, keypoints, color, radius,
|
| use_normalized_coordinates)
|
| np.copyto(image, np.array(image_pil))
|
|
|
|
|
| def draw_keypoints_on_image(image,
|
| keypoints,
|
| color='red',
|
| radius=2,
|
| use_normalized_coordinates=True):
|
| """Draws keypoints on an image.
|
|
|
| Args:
|
| image: a PIL.Image object.
|
| keypoints: a numpy array with shape [num_keypoints, 2].
|
| color: color to draw the keypoints with. Default is red.
|
| radius: keypoint radius. Default value is 2.
|
| use_normalized_coordinates: if True (default), treat keypoint values as
|
| relative to the image. Otherwise treat them as absolute.
|
| """
|
| draw = ImageDraw.Draw(image)
|
| im_width, im_height = image.size
|
| keypoints_x = [k[1] for k in keypoints]
|
| keypoints_y = [k[0] for k in keypoints]
|
| if use_normalized_coordinates:
|
| keypoints_x = tuple([im_width * x for x in keypoints_x])
|
| keypoints_y = tuple([im_height * y for y in keypoints_y])
|
| for keypoint_x, keypoint_y in zip(keypoints_x, keypoints_y):
|
| draw.ellipse([(keypoint_x - radius, keypoint_y - radius),
|
| (keypoint_x + radius, keypoint_y + radius)],
|
| outline=color,
|
| fill=color)
|
|
|
|
|
| def draw_mask_on_image_array(image, mask, color='red', alpha=0.4):
|
| """Draws mask on an image.
|
|
|
| Args:
|
| image: uint8 numpy array with shape (img_height, img_height, 3)
|
| mask: a uint8 numpy array of shape (img_height, img_height) with values
|
| between either 0 or 1.
|
| color: color to draw the keypoints with. Default is red.
|
| alpha: transparency value between 0 and 1. (default: 0.4)
|
|
|
| Raises:
|
| ValueError: On incorrect data type for image or masks.
|
| """
|
| if image.dtype != np.uint8:
|
| raise ValueError('`image` not of type np.uint8')
|
| if mask.dtype != np.uint8:
|
| raise ValueError('`mask` not of type np.uint8')
|
| if np.any(np.logical_and(mask != 1, mask != 0)):
|
| raise ValueError('`mask` elements should be in [0, 1]')
|
| if image.shape[:2] != mask.shape:
|
| raise ValueError('The image has spatial dimensions %s but the mask has '
|
| 'dimensions %s' % (image.shape[:2], mask.shape))
|
| rgb = ImageColor.getrgb(color)
|
| pil_image = Image.fromarray(image)
|
|
|
| solid_color = np.expand_dims(
|
| np.ones_like(mask), axis=2) * np.reshape(list(rgb), [1, 1, 3])
|
| pil_solid_color = Image.fromarray(np.uint8(solid_color)).convert('RGBA')
|
| pil_mask = Image.fromarray(np.uint8(255.0 * alpha * mask)).convert('L')
|
| pil_image = Image.composite(pil_solid_color, pil_image, pil_mask)
|
| np.copyto(image, np.array(pil_image.convert('RGB')))
|
|
|
|
|
| def visualize_boxes_and_labels_on_image_array(
|
| image,
|
| boxes,
|
| classes,
|
| scores,
|
| category_index,
|
| instance_masks=None,
|
| instance_boundaries=None,
|
| keypoints=None,
|
| use_normalized_coordinates=False,
|
| max_boxes_to_draw=20,
|
| min_score_thresh=.5,
|
| agnostic_mode=False,
|
| line_thickness=4,
|
| groundtruth_box_visualization_color='black',
|
| skip_scores=False,
|
| skip_labels=False):
|
| """Overlay labeled boxes on an image with formatted scores and label names.
|
|
|
| This function groups boxes that correspond to the same location
|
| and creates a display string for each detection and overlays these
|
| on the image. Note that this function modifies the image in place, and returns
|
| that same image.
|
|
|
| Args:
|
| image: uint8 numpy array with shape (img_height, img_width, 3)
|
| boxes: a numpy array of shape [N, 4]
|
| classes: a numpy array of shape [N]. Note that class indices are 1-based,
|
| and match the keys in the label map.
|
| scores: a numpy array of shape [N] or None. If scores=None, then this
|
| function assumes that the boxes to be plotted are groundtruth boxes and
|
| plot all boxes as black with no classes or scores.
|
| category_index: a dict containing category dictionaries (each holding
|
| category index `id` and category name `name`) keyed by category indices.
|
| instance_masks: a numpy array of shape [N, image_height, image_width] with
|
| values ranging between 0 and 1, can be None.
|
| instance_boundaries: a numpy array of shape [N, image_height, image_width]
|
| with values ranging between 0 and 1, can be None.
|
| keypoints: a numpy array of shape [N, num_keypoints, 2], can be None
|
| use_normalized_coordinates: whether boxes is to be interpreted as normalized
|
| coordinates or not.
|
| max_boxes_to_draw: maximum number of boxes to visualize. If None, draw all
|
| boxes.
|
| min_score_thresh: minimum score threshold for a box to be visualized
|
| agnostic_mode: boolean (default: False) controlling whether to evaluate in
|
| class-agnostic mode or not. This mode will display scores but ignore
|
| classes.
|
| line_thickness: integer (default: 4) controlling line width of the boxes.
|
| groundtruth_box_visualization_color: box color for visualizing groundtruth
|
| boxes
|
| skip_scores: whether to skip score when drawing a single detection
|
| skip_labels: whether to skip label when drawing a single detection
|
|
|
| Returns:
|
| uint8 numpy array with shape (img_height, img_width, 3) with overlaid boxes.
|
| """
|
|
|
|
|
| box_to_display_str_map = collections.defaultdict(list)
|
| box_to_color_map = collections.defaultdict(str)
|
| box_to_instance_masks_map = {}
|
| box_to_instance_boundaries_map = {}
|
| box_to_keypoints_map = collections.defaultdict(list)
|
| if not max_boxes_to_draw:
|
| max_boxes_to_draw = boxes.shape[0]
|
| for i in range(min(max_boxes_to_draw, boxes.shape[0])):
|
| if scores is None or scores[i] > min_score_thresh:
|
| box = tuple(boxes[i].tolist())
|
| if instance_masks is not None:
|
| box_to_instance_masks_map[box] = instance_masks[i]
|
| if instance_boundaries is not None:
|
| box_to_instance_boundaries_map[box] = instance_boundaries[i]
|
| if keypoints is not None:
|
| box_to_keypoints_map[box].extend(keypoints[i])
|
| if scores is None:
|
| box_to_color_map[box] = groundtruth_box_visualization_color
|
| else:
|
| display_str = ''
|
| if not skip_labels:
|
| if not agnostic_mode:
|
| if classes[i] in category_index.keys():
|
| class_name = category_index[classes[i]]['name']
|
| else:
|
| class_name = 'N/A'
|
| display_str = str(class_name)
|
| if not skip_scores:
|
| if not display_str:
|
| display_str = '{}%'.format(int(100 * scores[i]))
|
| else:
|
| display_str = '{}: {}%'.format(display_str, int(100 * scores[i]))
|
| box_to_display_str_map[box].append(display_str)
|
| if agnostic_mode:
|
| box_to_color_map[box] = 'DarkOrange'
|
| else:
|
| box_to_color_map[box] = STANDARD_COLORS[classes[i] %
|
| len(STANDARD_COLORS)]
|
|
|
|
|
| for box, color in box_to_color_map.items():
|
| ymin, xmin, ymax, xmax = box
|
| if instance_masks is not None:
|
| draw_mask_on_image_array(
|
| image, box_to_instance_masks_map[box], color=color)
|
| if instance_boundaries is not None:
|
| draw_mask_on_image_array(
|
| image, box_to_instance_boundaries_map[box], color='red', alpha=1.0)
|
| draw_bounding_box_on_image_array(
|
| image,
|
| ymin,
|
| xmin,
|
| ymax,
|
| xmax,
|
| color=color,
|
| thickness=line_thickness,
|
| display_str_list=box_to_display_str_map[box],
|
| use_normalized_coordinates=use_normalized_coordinates)
|
| if keypoints is not None:
|
| draw_keypoints_on_image_array(
|
| image,
|
| box_to_keypoints_map[box],
|
| color=color,
|
| radius=line_thickness / 2,
|
| use_normalized_coordinates=use_normalized_coordinates)
|
|
|
| return image
|
|
|
|
|
| def add_cdf_image_summary(values, name):
|
| """Adds a tf.summary.image for a CDF plot of the values.
|
|
|
| Normalizes `values` such that they sum to 1, plots the cumulative distribution
|
| function and creates a tf image summary.
|
|
|
| Args:
|
| values: a 1-D float32 tensor containing the values.
|
| name: name for the image summary.
|
| """
|
|
|
| def cdf_plot(values):
|
| """Numpy function to plot CDF."""
|
| normalized_values = values / np.sum(values)
|
| sorted_values = np.sort(normalized_values)
|
| cumulative_values = np.cumsum(sorted_values)
|
| fraction_of_examples = (
|
| np.arange(cumulative_values.size, dtype=np.float32) /
|
| cumulative_values.size)
|
| fig = plt.figure(frameon=False)
|
| ax = fig.add_subplot(1, 1, 1)
|
| ax.plot(fraction_of_examples, cumulative_values)
|
| ax.set_ylabel('cumulative normalized values')
|
| ax.set_xlabel('fraction of examples')
|
| fig.canvas.draw()
|
| width, height = fig.get_size_inches() * fig.get_dpi()
|
| image = np.fromstring(
|
| fig.canvas.tostring_rgb(),
|
| dtype='uint8').reshape(1, int(height), int(width), 3)
|
| return image
|
|
|
| cdf_plot = tf.compat.v1.py_func(cdf_plot, [values], tf.uint8)
|
| tf.compat.v1.summary.image(name, cdf_plot)
|
|
|
|
|
| def add_hist_image_summary(values, bins, name):
|
| """Adds a tf.summary.image for a histogram plot of the values.
|
|
|
| Plots the histogram of values and creates a tf image summary.
|
|
|
| Args:
|
| values: a 1-D float32 tensor containing the values.
|
| bins: bin edges which will be directly passed to np.histogram.
|
| name: name for the image summary.
|
| """
|
|
|
| def hist_plot(values, bins):
|
| """Numpy function to plot hist."""
|
| fig = plt.figure(frameon=False)
|
| ax = fig.add_subplot(1, 1, 1)
|
| y, x = np.histogram(values, bins=bins)
|
| ax.plot(x[:-1], y)
|
| ax.set_ylabel('count')
|
| ax.set_xlabel('value')
|
| fig.canvas.draw()
|
| width, height = fig.get_size_inches() * fig.get_dpi()
|
| image = np.fromstring(
|
| fig.canvas.tostring_rgb(),
|
| dtype='uint8').reshape(1, int(height), int(width), 3)
|
| return image
|
|
|
| hist_plot = tf.compat.v1.py_func(hist_plot, [values, bins], tf.uint8)
|
| tf.compat.v1.summary.image(name, hist_plot)
|
|
|
|
|
| def update_detection_state(step_outputs=None) -> Dict[str, Any]:
|
| """Updates detection state to optionally add input image and predictions."""
|
| state = {}
|
| if step_outputs:
|
| state['image'] = tf.concat(step_outputs['visualization'][0], axis=0)
|
| state['detection_boxes'] = tf.concat(
|
| step_outputs['visualization'][1]['detection_boxes'], axis=0
|
| )
|
| state['detection_classes'] = tf.concat(
|
| step_outputs['visualization'][1]['detection_classes'], axis=0
|
| )
|
| state['detection_scores'] = tf.concat(
|
| step_outputs['visualization'][1]['detection_scores'], axis=0
|
| )
|
|
|
| if 'detection_kpts' in step_outputs['visualization'][1]:
|
| detection_keypoints = step_outputs['visualization'][1]['detection_kpts']
|
| elif 'detection_keypoints' in step_outputs['visualization'][1]:
|
| detection_keypoints = step_outputs['visualization'][1][
|
| 'detection_keypoints'
|
| ]
|
| else:
|
| detection_keypoints = None
|
|
|
| if detection_keypoints is not None:
|
| state['detection_keypoints'] = tf.concat(detection_keypoints, axis=0)
|
|
|
| detection_masks = step_outputs['visualization'][1].get(
|
| 'detection_masks', None
|
| )
|
| if detection_masks:
|
| state['detection_masks'] = tf.concat(detection_masks, axis=0)
|
|
|
| return state
|
|
|
|
|
| def update_segmentation_state(step_outputs=None) -> Dict[str, Any]:
|
| """Updates segmentation state to optionally add input image and predictions."""
|
| state = {}
|
| if step_outputs:
|
| state['image'] = tf.concat(step_outputs['visualization'][0], axis=0)
|
| state['logits'] = tf.concat(
|
| step_outputs['visualization'][1]['logits'], axis=0
|
| )
|
| return state
|
|
|
|
|
| def visualize_segmentation_outputs(
|
| logs,
|
| task_config,
|
| original_image_spatial_shape=None,
|
| true_image_shape=None,
|
| image_mean: Optional[Union[float, List[float]]] = None,
|
| image_std: Optional[Union[float, List[float]]] = None,
|
| key: str = 'image/validation_outputs',
|
| ) -> Dict[str, Any]:
|
| """Visualizes the detection outputs.
|
|
|
| It extracts images and predictions from logs and draws visualization on input
|
| images. By default, it requires `detection_boxes`, `detection_classes` and
|
| `detection_scores` in the prediction, and optionally accepts
|
| `detection_keypoints` and `detection_masks`.
|
|
|
| Args:
|
| logs: A dictionaty of log that contains images and predictions.
|
| task_config: A task config.
|
| original_image_spatial_shape: A [N, 2] tensor containing the spatial size of
|
| the original image.
|
| true_image_shape: A [N, 3] tensor containing the spatial size of unpadded
|
| original_image.
|
| image_mean: An optional float or list of floats used as the mean pixel value
|
| to normalize images.
|
| image_std: An optional float or list of floats used as the std to normalize
|
| images.
|
| key: A string specifying the key of the returned dictionary.
|
|
|
| Returns:
|
| A dictionary of images with visualization drawn on it. Each key corresponds
|
| to a 4D tensor with segments drawn on each image.
|
| """
|
| images = logs['image']
|
| masks = np.argmax(logs['logits'], axis=-1)
|
| num_classes = task_config.model.num_classes
|
|
|
| def _denormalize_images(images: tf.Tensor) -> tf.Tensor:
|
| if image_mean is None and image_std is None:
|
| images *= tf.constant(
|
| preprocess_ops.STDDEV_RGB, shape=[1, 1, 3], dtype=images.dtype
|
| )
|
| images += tf.constant(
|
| preprocess_ops.MEAN_RGB, shape=[1, 1, 3], dtype=images.dtype
|
| )
|
| elif image_mean is not None and image_std is not None:
|
| if isinstance(image_mean, float) and isinstance(image_std, float):
|
| images = images * image_std + image_mean
|
| elif isinstance(image_mean, list) and isinstance(image_std, list):
|
| images *= tf.constant(image_std, shape=[1, 1, 3], dtype=images.dtype)
|
| images += tf.constant(image_mean, shape=[1, 1, 3], dtype=images.dtype)
|
| else:
|
| raise ValueError(
|
| '`image_mean` and `image_std` should be the same type.'
|
| )
|
| else:
|
| raise ValueError(
|
| 'Both `image_mean` and `image_std` should be set or None at the same '
|
| 'time.'
|
| )
|
| return tf.cast(images, dtype=tf.uint8)
|
|
|
| if images.shape[3] > 3:
|
| images = images[:, :, :, 0:3]
|
| elif images.shape[3] == 1:
|
| images = tf.image.grayscale_to_rgb(images)
|
|
|
| images = tf.nest.map_structure(
|
| tf.identity,
|
| tf.map_fn(
|
| _denormalize_images,
|
| elems=images,
|
| fn_output_signature=tf.TensorSpec(
|
| shape=images.shape.as_list()[1:], dtype=tf.uint8
|
| ),
|
| parallel_iterations=32,
|
| ),
|
| )
|
|
|
| if true_image_shape is None:
|
| true_shapes = tf.constant(-1, shape=[images.shape.as_list()[0], 3])
|
| else:
|
| true_shapes = true_image_shape
|
| if original_image_spatial_shape is None:
|
| original_shapes = tf.constant(-1, shape=[images.shape.as_list()[0], 2])
|
| else:
|
| original_shapes = original_image_spatial_shape
|
|
|
| visualize_fn = functools.partial(_visualize_masks, num_classes=num_classes)
|
| elems = [true_shapes, original_shapes, images, masks]
|
|
|
| def draw_segments(image_and_segments):
|
| """Draws boxes on image."""
|
| true_shape = image_and_segments[0]
|
| original_shape = image_and_segments[1]
|
| if true_image_shape is not None:
|
| image = shape_utils.pad_or_clip_nd(
|
| image_and_segments[2], [true_shape[0], true_shape[1], 3]
|
| )
|
| if original_image_spatial_shape is not None:
|
| image_and_segments[2] = _resize_original_image(image, original_shape)
|
|
|
| image_with_boxes = tf.compat.v1.py_func(
|
| visualize_fn, image_and_segments[2:], tf.uint8
|
| )
|
| return image_with_boxes
|
|
|
| images_with_segments = tf.map_fn(
|
| draw_segments, elems, dtype=tf.uint8, back_prop=False
|
| )
|
|
|
| outputs = {}
|
| for i, image in enumerate(images_with_segments):
|
| outputs[key + f'/{i}'] = image[None, ...]
|
|
|
| return outputs
|
|
|
|
|
| def _visualize_masks(image, mask, num_classes, alpha=0.4):
|
| """Visualizes semantic segmentation masks."""
|
| solid_color = np.repeat(
|
| np.expand_dims(np.zeros_like(mask), axis=2), 3, axis=2
|
| )
|
| for i in range(num_classes):
|
| color = STANDARD_COLORS[i % len(STANDARD_COLORS)]
|
| rgb = ImageColor.getrgb(color)
|
| one_class_mask = np.where(mask == i, 1, 0)
|
| solid_color = solid_color + np.expand_dims(
|
| one_class_mask, axis=2
|
| ) * np.reshape(list(rgb), [1, 1, 3])
|
|
|
| pil_image = Image.fromarray(image)
|
| pil_solid_color = (
|
| Image.fromarray(np.uint8(solid_color))
|
| .convert('RGBA')
|
| .resize(pil_image.size)
|
| )
|
| pil_mask = (
|
| Image.fromarray(np.uint8(255.0 * alpha * np.ones_like(mask)))
|
| .convert('L')
|
| .resize(pil_image.size)
|
| )
|
| pil_image = Image.composite(pil_solid_color, pil_image, pil_mask)
|
| np.copyto(image, np.array(pil_image.convert('RGB')))
|
| return image
|
|
|