| from collections import defaultdict |
| from typing import Optional |
|
|
| from ..image_utils import load_image |
| from ..utils import ( |
| add_end_docstrings, |
| is_torch_available, |
| logging, |
| requires_backends, |
| ) |
| from .base import PIPELINE_INIT_ARGS, ChunkPipeline |
|
|
|
|
| if is_torch_available(): |
| import torch |
|
|
| from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING_NAMES |
|
|
| logger = logging.get_logger(__name__) |
|
|
|
|
| @add_end_docstrings(PIPELINE_INIT_ARGS) |
| class MaskGenerationPipeline(ChunkPipeline): |
| """ |
| Automatic mask generation for images using `SamForMaskGeneration`. This pipeline predicts binary masks for an |
| image, given an image. It is a `ChunkPipeline` because you can seperate the points in a mini-batch in order to |
| avoid OOM issues. Use the `points_per_batch` argument to control the number of points that will be processed at the |
| same time. Default is `64`. |
| |
| The pipeline works in 3 steps: |
| 1. `preprocess`: A grid of 1024 points evenly separated is generated along with bounding boxes and point |
| labels. |
| For more details on how the points and bounding boxes are created, check the `_generate_crop_boxes` |
| function. The image is also preprocessed using the `image_processor`. This function `yields` a minibatch of |
| `points_per_batch`. |
| |
| 2. `forward`: feeds the outputs of `preprocess` to the model. The image embedding is computed only once. |
| Calls both `self.model.get_image_embeddings` and makes sure that the gradients are not computed, and the |
| tensors and models are on the same device. |
| |
| 3. `postprocess`: The most important part of the automatic mask generation happens here. Three steps |
| are induced: |
| - image_processor.postprocess_masks (run on each minibatch loop): takes in the raw output masks, |
| resizes them according |
| to the image size, and transforms there to binary masks. |
| - image_processor.filter_masks (on each minibatch loop): uses both `pred_iou_thresh` and |
| `stability_scores`. Also |
| applies a variety of filters based on non maximum suppression to remove bad masks. |
| - image_processor.postprocess_masks_for_amg applies the NSM on the mask to only keep relevant ones. |
| |
| Arguments: |
| model ([`PreTrainedModel`] or [`TFPreTrainedModel`]): |
| The model that will be used by the pipeline to make predictions. This needs to be a model inheriting from |
| [`PreTrainedModel`] for PyTorch and [`TFPreTrainedModel`] for TensorFlow. |
| tokenizer ([`PreTrainedTokenizer`]): |
| The tokenizer that will be used by the pipeline to encode data for the model. This object inherits from |
| [`PreTrainedTokenizer`]. |
| feature_extractor ([`SequenceFeatureExtractor`]): |
| The feature extractor that will be used by the pipeline to encode the input. |
| points_per_batch (*optional*, int, default to 64): |
| Sets the number of points run simultaneously by the model. Higher numbers may be faster but use more GPU |
| memory. |
| output_bboxes_mask (`bool`, *optional*, default to `False`): |
| Whether or not to output the bounding box predictions. |
| output_rle_masks (`bool`, *optional*, default to `False`): |
| Whether or not to output the masks in `RLE` format |
| |
| Example: |
| |
| ```python |
| >>> from transformers import pipeline |
| |
| >>> generator = pipeline(model="facebook/sam-vit-base", task="mask-generation") |
| >>> outputs = generator( |
| ... "http://images.cocodataset.org/val2017/000000039769.jpg", |
| ... ) |
| |
| >>> outputs = generator( |
| ... "https://huggingface.co/datasets/Narsil/image_dummy/raw/main/parrots.png", points_per_batch=128 |
| ... ) |
| ``` |
| |
| Learn more about the basics of using a pipeline in the [pipeline tutorial](../pipeline_tutorial) |
| |
| This segmentation pipeline can currently be loaded from [`pipeline`] using the following task identifier: |
| `"mask-generation"`. |
| |
| See the list of available models on [huggingface.co/models](https://huggingface.co/models?filter=mask-generation). |
| """ |
|
|
| def __init__(self, **kwargs): |
| super().__init__(**kwargs) |
| requires_backends(self, "vision") |
| requires_backends(self, "torch") |
|
|
| if self.framework != "pt": |
| raise ValueError(f"The {self.__class__} is only available in PyTorch.") |
|
|
| self.check_model_type(MODEL_FOR_MASK_GENERATION_MAPPING_NAMES) |
|
|
| def _sanitize_parameters(self, **kwargs): |
| preprocess_kwargs = {} |
| postprocess_kwargs = {} |
| forward_params = {} |
| |
| if "points_per_batch" in kwargs: |
| preprocess_kwargs["points_per_batch"] = kwargs["points_per_batch"] |
| if "points_per_crop" in kwargs: |
| preprocess_kwargs["points_per_crop"] = kwargs["points_per_crop"] |
| if "crops_n_layers" in kwargs: |
| preprocess_kwargs["crops_n_layers"] = kwargs["crops_n_layers"] |
| if "crop_overlap_ratio" in kwargs: |
| preprocess_kwargs["crop_overlap_ratio"] = kwargs["crop_overlap_ratio"] |
| if "crop_n_points_downscale_factor" in kwargs: |
| preprocess_kwargs["crop_n_points_downscale_factor"] = kwargs["crop_n_points_downscale_factor"] |
| if "timeout" in kwargs: |
| preprocess_kwargs["timeout"] = kwargs["timeout"] |
| |
| if "pred_iou_thresh" in kwargs: |
| forward_params["pred_iou_thresh"] = kwargs["pred_iou_thresh"] |
| if "stability_score_offset" in kwargs: |
| forward_params["stability_score_offset"] = kwargs["stability_score_offset"] |
| if "mask_threshold" in kwargs: |
| forward_params["mask_threshold"] = kwargs["mask_threshold"] |
| if "stability_score_thresh" in kwargs: |
| forward_params["stability_score_thresh"] = kwargs["stability_score_thresh"] |
| if "crops_nms_thresh" in kwargs: |
| postprocess_kwargs["crops_nms_thresh"] = kwargs["crops_nms_thresh"] |
| if "output_rle_mask" in kwargs: |
| postprocess_kwargs["output_rle_mask"] = kwargs["output_rle_mask"] |
| if "output_bboxes_mask" in kwargs: |
| postprocess_kwargs["output_bboxes_mask"] = kwargs["output_bboxes_mask"] |
| return preprocess_kwargs, forward_params, postprocess_kwargs |
|
|
| def __call__(self, image, *args, num_workers=None, batch_size=None, **kwargs): |
| """ |
| Generates binary segmentation masks |
| |
| Args: |
| inputs (`np.ndarray` or `bytes` or `str` or `dict`): |
| Image or list of images. |
| mask_threshold (`float`, *optional*, defaults to 0.0): |
| Threshold to use when turning the predicted masks into binary values. |
| pred_iou_thresh (`float`, *optional*, defaults to 0.88): |
| A filtering threshold in `[0,1]` applied on the model's predicted mask quality. |
| stability_score_thresh (`float`, *optional*, defaults to 0.95): |
| A filtering threshold in `[0,1]`, using the stability of the mask under changes to the cutoff used to |
| binarize the model's mask predictions. |
| stability_score_offset (`int`, *optional*, defaults to 1): |
| The amount to shift the cutoff when calculated the stability score. |
| crops_nms_thresh (`float`, *optional*, defaults to 0.7): |
| The box IoU cutoff used by non-maximal suppression to filter duplicate masks. |
| crops_n_layers (`int`, *optional*, defaults to 0): |
| If `crops_n_layers>0`, mask prediction will be run again on crops of the image. Sets the number of |
| layers to run, where each layer has 2**i_layer number of image crops. |
| crop_overlap_ratio (`float`, *optional*, defaults to `512 / 1500`): |
| Sets the degree to which crops overlap. In the first crop layer, crops will overlap by this fraction of |
| the image length. Later layers with more crops scale down this overlap. |
| crop_n_points_downscale_factor (`int`, *optional*, defaults to `1`): |
| The number of points-per-side sampled in layer n is scaled down by crop_n_points_downscale_factor**n. |
| timeout (`float`, *optional*, defaults to None): |
| The maximum time in seconds to wait for fetching images from the web. If None, no timeout is set and |
| the call may block forever. |
| |
| Return: |
| `Dict`: A dictionary with the following keys: |
| - **mask** (`PIL.Image`) -- A binary mask of the detected object as a PIL Image of shape `(width, |
| height)` of the original image. Returns a mask filled with zeros if no object is found. |
| - **score** (*optional* `float`) -- Optionally, when the model is capable of estimating a confidence of |
| the "object" described by the label and the mask. |
| |
| """ |
| return super().__call__(image, *args, num_workers=num_workers, batch_size=batch_size, **kwargs) |
|
|
| def preprocess( |
| self, |
| image, |
| points_per_batch=64, |
| crops_n_layers: int = 0, |
| crop_overlap_ratio: float = 512 / 1500, |
| points_per_crop: Optional[int] = 32, |
| crop_n_points_downscale_factor: Optional[int] = 1, |
| timeout: Optional[float] = None, |
| ): |
| image = load_image(image, timeout=timeout) |
| target_size = self.image_processor.size["longest_edge"] |
| crop_boxes, grid_points, cropped_images, input_labels = self.image_processor.generate_crop_boxes( |
| image, target_size, crops_n_layers, crop_overlap_ratio, points_per_crop, crop_n_points_downscale_factor |
| ) |
| model_inputs = self.image_processor(images=cropped_images, return_tensors="pt") |
|
|
| with self.device_placement(): |
| if self.framework == "pt": |
| inference_context = self.get_inference_context() |
| with inference_context(): |
| model_inputs = self._ensure_tensor_on_device(model_inputs, device=self.device) |
| image_embeddings = self.model.get_image_embeddings(model_inputs.pop("pixel_values")) |
| model_inputs["image_embeddings"] = image_embeddings |
|
|
| n_points = grid_points.shape[1] |
| points_per_batch = points_per_batch if points_per_batch is not None else n_points |
|
|
| if points_per_batch <= 0: |
| raise ValueError( |
| "Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. " |
| "To return all points at once, set points_per_batch to None" |
| ) |
|
|
| for i in range(0, n_points, points_per_batch): |
| batched_points = grid_points[:, i : i + points_per_batch, :, :] |
| labels = input_labels[:, i : i + points_per_batch] |
| is_last = i == n_points - points_per_batch |
| yield { |
| "input_points": batched_points, |
| "input_labels": labels, |
| "input_boxes": crop_boxes, |
| "is_last": is_last, |
| **model_inputs, |
| } |
|
|
| def _forward( |
| self, |
| model_inputs, |
| pred_iou_thresh=0.88, |
| stability_score_thresh=0.95, |
| mask_threshold=0, |
| stability_score_offset=1, |
| ): |
| input_boxes = model_inputs.pop("input_boxes") |
| is_last = model_inputs.pop("is_last") |
| original_sizes = model_inputs.pop("original_sizes").tolist() |
| reshaped_input_sizes = model_inputs.pop("reshaped_input_sizes").tolist() |
|
|
| model_outputs = self.model(**model_inputs) |
|
|
| |
| low_resolution_masks = model_outputs["pred_masks"] |
| masks = self.image_processor.post_process_masks( |
| low_resolution_masks, original_sizes, reshaped_input_sizes, mask_threshold, binarize=False |
| ) |
| iou_scores = model_outputs["iou_scores"] |
| masks, iou_scores, boxes = self.image_processor.filter_masks( |
| masks[0], |
| iou_scores[0], |
| original_sizes[0], |
| input_boxes[0], |
| pred_iou_thresh, |
| stability_score_thresh, |
| mask_threshold, |
| stability_score_offset, |
| ) |
| return { |
| "masks": masks, |
| "is_last": is_last, |
| "boxes": boxes, |
| "iou_scores": iou_scores, |
| } |
|
|
| def postprocess( |
| self, |
| model_outputs, |
| output_rle_mask=False, |
| output_bboxes_mask=False, |
| crops_nms_thresh=0.7, |
| ): |
| all_scores = [] |
| all_masks = [] |
| all_boxes = [] |
| for model_output in model_outputs: |
| all_scores.append(model_output.pop("iou_scores")) |
| all_masks.extend(model_output.pop("masks")) |
| all_boxes.append(model_output.pop("boxes")) |
|
|
| all_scores = torch.cat(all_scores) |
| all_boxes = torch.cat(all_boxes) |
| output_masks, iou_scores, rle_mask, bounding_boxes = self.image_processor.post_process_for_mask_generation( |
| all_masks, all_scores, all_boxes, crops_nms_thresh |
| ) |
|
|
| extra = defaultdict(list) |
| for output in model_outputs: |
| for k, v in output.items(): |
| extra[k].append(v) |
|
|
| optional = {} |
| if output_rle_mask: |
| optional["rle_mask"] = rle_mask |
|
|
| if output_bboxes_mask: |
| optional["bounding_boxes"] = bounding_boxes |
|
|
| return {"masks": output_masks, "scores": iou_scores, **optional, **extra} |
|
|