id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
178,946
import torch import torch.nn as nn import torch.nn.functional as F from ..builder import LOSSES from .utils import weight_reduce_loss def weight_reduce_loss(loss, weight=None, reduction='mean', avg_factor=None): """Apply element-wise weight and reduce loss. Args: loss (Tensor): Element-wise loss. weight (Tensor): Element-wise weights. reduction (str): Same as built-in losses of PyTorch. avg_factor (float): Avarage factor when computing the mean of losses. Returns: Tensor: Processed loss values. """ # if weight is specified, apply element-wise weight if weight is not None: assert weight.dim() == loss.dim() if weight.dim() > 1: assert weight.size(1) == 1 or weight.size(1) == loss.size(1) loss = loss * weight # if avg_factor is not specified, just reduce the loss if avg_factor is None: loss = reduce_loss(loss, reduction) else: # if reduction is mean, then average the loss by avg_factor if reduction == 'mean': loss = loss.sum() / avg_factor # if reduction is 'none', then do nothing, otherwise raise an error elif reduction != 'none': raise ValueError('avg_factor can not be used with reduction="sum"') return loss The provided code snippet includes necessary dependencies for implementing the `cross_entropy` function. Write a Python function `def cross_entropy(pred, label, weight=None, class_weight=None, reduction='mean', avg_factor=None, ignore_index=-100)` to solve the following problem: The wrapper function for :func:`F.cross_entropy` Here is the function: def cross_entropy(pred, label, weight=None, class_weight=None, reduction='mean', avg_factor=None, ignore_index=-100): """The wrapper function for :func:`F.cross_entropy`""" # class_weight is a manual rescaling weight given to each class. # If given, has to be a Tensor of size C element-wise losses loss = F.cross_entropy( pred, label, weight=class_weight, reduction='none', ignore_index=ignore_index) # apply weights and do the reduction if weight is not None: weight = weight.float() loss = weight_reduce_loss( loss, weight=weight, reduction=reduction, avg_factor=avg_factor) return loss
The wrapper function for :func:`F.cross_entropy`
178,947
import torch import torch.nn as nn import torch.nn.functional as F from ..builder import LOSSES from .utils import weight_reduce_loss def _expand_onehot_labels(labels, label_weights, target_shape, ignore_index): """Expand onehot labels to match the size of prediction.""" bin_labels = labels.new_zeros(target_shape) valid_mask = (labels >= 0) & (labels != ignore_index) inds = torch.nonzero(valid_mask, as_tuple=True) if inds[0].numel() > 0: if labels.dim() == 3: bin_labels[inds[0], labels[valid_mask], inds[1], inds[2]] = 1 else: bin_labels[inds[0], labels[valid_mask]] = 1 valid_mask = valid_mask.unsqueeze(1).expand(target_shape).float() if label_weights is None: bin_label_weights = valid_mask else: bin_label_weights = label_weights.unsqueeze(1).expand(target_shape) bin_label_weights *= valid_mask return bin_labels, bin_label_weights def weight_reduce_loss(loss, weight=None, reduction='mean', avg_factor=None): """Apply element-wise weight and reduce loss. Args: loss (Tensor): Element-wise loss. weight (Tensor): Element-wise weights. reduction (str): Same as built-in losses of PyTorch. avg_factor (float): Avarage factor when computing the mean of losses. Returns: Tensor: Processed loss values. """ # if weight is specified, apply element-wise weight if weight is not None: assert weight.dim() == loss.dim() if weight.dim() > 1: assert weight.size(1) == 1 or weight.size(1) == loss.size(1) loss = loss * weight # if avg_factor is not specified, just reduce the loss if avg_factor is None: loss = reduce_loss(loss, reduction) else: # if reduction is mean, then average the loss by avg_factor if reduction == 'mean': loss = loss.sum() / avg_factor # if reduction is 'none', then do nothing, otherwise raise an error elif reduction != 'none': raise ValueError('avg_factor can not be used with reduction="sum"') return loss The provided code snippet includes necessary dependencies for implementing the `binary_cross_entropy` function. Write a Python function `def binary_cross_entropy(pred, label, weight=None, reduction='mean', avg_factor=None, class_weight=None, ignore_index=255)` to solve the following problem: Calculate the binary CrossEntropy loss. Args: pred (torch.Tensor): The prediction with shape (N, 1). label (torch.Tensor): The learning label of the prediction. weight (torch.Tensor, optional): Sample-wise loss weight. reduction (str, optional): The method used to reduce the loss. Options are "none", "mean" and "sum". avg_factor (int, optional): Average factor that is used to average the loss. Defaults to None. class_weight (list[float], optional): The weight for each class. ignore_index (int | None): The label index to be ignored. Default: 255 Returns: torch.Tensor: The calculated loss Here is the function: def binary_cross_entropy(pred, label, weight=None, reduction='mean', avg_factor=None, class_weight=None, ignore_index=255): """Calculate the binary CrossEntropy loss. Args: pred (torch.Tensor): The prediction with shape (N, 1). label (torch.Tensor): The learning label of the prediction. weight (torch.Tensor, optional): Sample-wise loss weight. reduction (str, optional): The method used to reduce the loss. Options are "none", "mean" and "sum". avg_factor (int, optional): Average factor that is used to average the loss. Defaults to None. class_weight (list[float], optional): The weight for each class. ignore_index (int | None): The label index to be ignored. Default: 255 Returns: torch.Tensor: The calculated loss """ if pred.dim() != label.dim(): assert (pred.dim() == 2 and label.dim() == 1) or ( pred.dim() == 4 and label.dim() == 3), \ 'Only pred shape [N, C], label shape [N] or pred shape [N, C, ' \ 'H, W], label shape [N, H, W] are supported' label, weight = _expand_onehot_labels(label, weight, pred.shape, ignore_index) # weighted element-wise losses if weight is not None: weight = weight.float() loss = F.binary_cross_entropy_with_logits( pred, label.float(), pos_weight=class_weight, reduction='none') # do the reduction for the weighted loss loss = weight_reduce_loss( loss, weight, reduction=reduction, avg_factor=avg_factor) return loss
Calculate the binary CrossEntropy loss. Args: pred (torch.Tensor): The prediction with shape (N, 1). label (torch.Tensor): The learning label of the prediction. weight (torch.Tensor, optional): Sample-wise loss weight. reduction (str, optional): The method used to reduce the loss. Options are "none", "mean" and "sum". avg_factor (int, optional): Average factor that is used to average the loss. Defaults to None. class_weight (list[float], optional): The weight for each class. ignore_index (int | None): The label index to be ignored. Default: 255 Returns: torch.Tensor: The calculated loss
178,948
import torch import torch.nn as nn import torch.nn.functional as F from ..builder import LOSSES from .utils import weight_reduce_loss The provided code snippet includes necessary dependencies for implementing the `mask_cross_entropy` function. Write a Python function `def mask_cross_entropy(pred, target, label, reduction='mean', avg_factor=None, class_weight=None, ignore_index=None)` to solve the following problem: Calculate the CrossEntropy loss for masks. Args: pred (torch.Tensor): The prediction with shape (N, C), C is the number of classes. target (torch.Tensor): The learning label of the prediction. label (torch.Tensor): ``label`` indicates the class label of the mask' corresponding object. This will be used to select the mask in the of the class which the object belongs to when the mask prediction if not class-agnostic. reduction (str, optional): The method used to reduce the loss. Options are "none", "mean" and "sum". avg_factor (int, optional): Average factor that is used to average the loss. Defaults to None. class_weight (list[float], optional): The weight for each class. ignore_index (None): Placeholder, to be consistent with other loss. Default: None. Returns: torch.Tensor: The calculated loss Here is the function: def mask_cross_entropy(pred, target, label, reduction='mean', avg_factor=None, class_weight=None, ignore_index=None): """Calculate the CrossEntropy loss for masks. Args: pred (torch.Tensor): The prediction with shape (N, C), C is the number of classes. target (torch.Tensor): The learning label of the prediction. label (torch.Tensor): ``label`` indicates the class label of the mask' corresponding object. This will be used to select the mask in the of the class which the object belongs to when the mask prediction if not class-agnostic. reduction (str, optional): The method used to reduce the loss. Options are "none", "mean" and "sum". avg_factor (int, optional): Average factor that is used to average the loss. Defaults to None. class_weight (list[float], optional): The weight for each class. ignore_index (None): Placeholder, to be consistent with other loss. Default: None. Returns: torch.Tensor: The calculated loss """ assert ignore_index is None, 'BCE loss does not support ignore_index' # TODO: handle these two reserved arguments assert reduction == 'mean' and avg_factor is None num_rois = pred.size()[0] inds = torch.arange(0, num_rois, dtype=torch.long, device=pred.device) pred_slice = pred[inds, label].squeeze(1) return F.binary_cross_entropy_with_logits( pred_slice, target, weight=class_weight, reduction='mean')[None]
Calculate the CrossEntropy loss for masks. Args: pred (torch.Tensor): The prediction with shape (N, C), C is the number of classes. target (torch.Tensor): The learning label of the prediction. label (torch.Tensor): ``label`` indicates the class label of the mask' corresponding object. This will be used to select the mask in the of the class which the object belongs to when the mask prediction if not class-agnostic. reduction (str, optional): The method used to reduce the loss. Options are "none", "mean" and "sum". avg_factor (int, optional): Average factor that is used to average the loss. Defaults to None. class_weight (list[float], optional): The weight for each class. ignore_index (None): Placeholder, to be consistent with other loss. Default: None. Returns: torch.Tensor: The calculated loss
178,949
import warnings from mmcv.utils import Registry, build_from_cfg from torch import nn BACKBONES = Registry('backbone') def build(cfg, registry, default_args=None): """Build a module. Args: cfg (dict, list[dict]): The config of modules, is is either a dict or a list of configs. registry (:obj:`Registry`): A registry the module belongs to. default_args (dict, optional): Default arguments to build the module. Defaults to None. Returns: nn.Module: A built nn module. """ if isinstance(cfg, list): modules = [ build_from_cfg(cfg_, registry, default_args) for cfg_ in cfg ] return nn.Sequential(*modules) else: return build_from_cfg(cfg, registry, default_args) The provided code snippet includes necessary dependencies for implementing the `build_backbone` function. Write a Python function `def build_backbone(cfg)` to solve the following problem: Build backbone. Here is the function: def build_backbone(cfg): """Build backbone.""" return build(cfg, BACKBONES)
Build backbone.
178,950
import warnings from mmcv.utils import Registry, build_from_cfg from torch import nn NECKS = Registry('neck') def build(cfg, registry, default_args=None): """Build a module. Args: cfg (dict, list[dict]): The config of modules, is is either a dict or a list of configs. registry (:obj:`Registry`): A registry the module belongs to. default_args (dict, optional): Default arguments to build the module. Defaults to None. Returns: nn.Module: A built nn module. """ if isinstance(cfg, list): modules = [ build_from_cfg(cfg_, registry, default_args) for cfg_ in cfg ] return nn.Sequential(*modules) else: return build_from_cfg(cfg, registry, default_args) The provided code snippet includes necessary dependencies for implementing the `build_neck` function. Write a Python function `def build_neck(cfg)` to solve the following problem: Build neck. Here is the function: def build_neck(cfg): """Build neck.""" return build(cfg, NECKS)
Build neck.
178,951
import warnings from mmcv.utils import Registry, build_from_cfg from torch import nn HEADS = Registry('head') def build(cfg, registry, default_args=None): """Build a module. Args: cfg (dict, list[dict]): The config of modules, is is either a dict or a list of configs. registry (:obj:`Registry`): A registry the module belongs to. default_args (dict, optional): Default arguments to build the module. Defaults to None. Returns: nn.Module: A built nn module. """ if isinstance(cfg, list): modules = [ build_from_cfg(cfg_, registry, default_args) for cfg_ in cfg ] return nn.Sequential(*modules) else: return build_from_cfg(cfg, registry, default_args) The provided code snippet includes necessary dependencies for implementing the `build_head` function. Write a Python function `def build_head(cfg)` to solve the following problem: Build head. Here is the function: def build_head(cfg): """Build head.""" return build(cfg, HEADS)
Build head.
178,952
import warnings from mmcv.utils import Registry, build_from_cfg from torch import nn LOSSES = Registry('loss') def build(cfg, registry, default_args=None): """Build a module. Args: cfg (dict, list[dict]): The config of modules, is is either a dict or a list of configs. registry (:obj:`Registry`): A registry the module belongs to. default_args (dict, optional): Default arguments to build the module. Defaults to None. Returns: nn.Module: A built nn module. """ if isinstance(cfg, list): modules = [ build_from_cfg(cfg_, registry, default_args) for cfg_ in cfg ] return nn.Sequential(*modules) else: return build_from_cfg(cfg, registry, default_args) The provided code snippet includes necessary dependencies for implementing the `build_loss` function. Write a Python function `def build_loss(cfg)` to solve the following problem: Build loss. Here is the function: def build_loss(cfg): """Build loss.""" return build(cfg, LOSSES)
Build loss.
178,953
import warnings from mmcv.utils import Registry, build_from_cfg from torch import nn SEGMENTORS = Registry('segmentor') def build(cfg, registry, default_args=None): """Build a module. Args: cfg (dict, list[dict]): The config of modules, is is either a dict or a list of configs. registry (:obj:`Registry`): A registry the module belongs to. default_args (dict, optional): Default arguments to build the module. Defaults to None. Returns: nn.Module: A built nn module. """ if isinstance(cfg, list): modules = [ build_from_cfg(cfg_, registry, default_args) for cfg_ in cfg ] return nn.Sequential(*modules) else: return build_from_cfg(cfg, registry, default_args) The provided code snippet includes necessary dependencies for implementing the `build_segmentor` function. Write a Python function `def build_segmentor(cfg, train_cfg=None, test_cfg=None)` to solve the following problem: Build segmentor. Here is the function: def build_segmentor(cfg, train_cfg=None, test_cfg=None): """Build segmentor.""" if train_cfg is not None or test_cfg is not None: warnings.warn( 'train_cfg and test_cfg is deprecated, ' 'please specify them in model', UserWarning) assert cfg.get('train_cfg') is None or train_cfg is None, \ 'train_cfg specified in both outer field and model field ' assert cfg.get('test_cfg') is None or test_cfg is None, \ 'test_cfg specified in both outer field and model field ' return build(cfg, SEGMENTORS, dict(train_cfg=train_cfg, test_cfg=test_cfg))
Build segmentor.
178,954
import torch import torch.nn as nn from mmcv.cnn import ConvModule, normal_init from mmcv.ops import point_sample from mmseg.models.builder import HEADS from mmseg.ops import resize from ..losses import accuracy from .cascade_decode_head import BaseCascadeDecodeHead The provided code snippet includes necessary dependencies for implementing the `calculate_uncertainty` function. Write a Python function `def calculate_uncertainty(seg_logits)` to solve the following problem: Estimate uncertainty based on seg logits. For each location of the prediction ``seg_logits`` we estimate uncertainty as the difference between top first and top second predicted logits. Args: seg_logits (Tensor): Semantic segmentation logits, shape (batch_size, num_classes, height, width). Returns: scores (Tensor): T uncertainty scores with the most uncertain locations having the highest uncertainty score, shape ( batch_size, 1, height, width) Here is the function: def calculate_uncertainty(seg_logits): """Estimate uncertainty based on seg logits. For each location of the prediction ``seg_logits`` we estimate uncertainty as the difference between top first and top second predicted logits. Args: seg_logits (Tensor): Semantic segmentation logits, shape (batch_size, num_classes, height, width). Returns: scores (Tensor): T uncertainty scores with the most uncertain locations having the highest uncertainty score, shape ( batch_size, 1, height, width) """ top2_scores = torch.topk(seg_logits, k=2, dim=1)[0] return (top2_scores[:, 1] - top2_scores[:, 0]).unsqueeze(1)
Estimate uncertainty based on seg logits. For each location of the prediction ``seg_logits`` we estimate uncertainty as the difference between top first and top second predicted logits. Args: seg_logits (Tensor): Semantic segmentation logits, shape (batch_size, num_classes, height, width). Returns: scores (Tensor): T uncertainty scores with the most uncertain locations having the highest uncertainty score, shape ( batch_size, 1, height, width)
178,955
import math import torch import torch.distributed as dist import torch.nn as nn import torch.nn.functional as F from mmcv.cnn import ConvModule from ..builder import HEADS from .decode_head import BaseDecodeHead The provided code snippet includes necessary dependencies for implementing the `reduce_mean` function. Write a Python function `def reduce_mean(tensor)` to solve the following problem: Reduce mean when distributed training. Here is the function: def reduce_mean(tensor): """Reduce mean when distributed training.""" if not (dist.is_available() and dist.is_initialized()): return tensor tensor = tensor.clone() dist.all_reduce(tensor.div_(dist.get_world_size()), op=dist.ReduceOp.SUM) return tensor
Reduce mean when distributed training.
178,958
import torch import torch.nn as nn import torch.nn.functional as F The provided code snippet includes necessary dependencies for implementing the `drop_block_fast_2d` function. Write a Python function `def drop_block_fast_2d( x: torch.Tensor, drop_prob: float = 0.1, block_size: int = 7, gamma_scale: float = 1.0, with_noise: bool = False, inplace: bool = False, batchwise: bool = False)` to solve the following problem: DropBlock. See https://arxiv.org/pdf/1810.12890.pdf DropBlock with an experimental gaussian noise option. Simplied from above without concern for valid block mask at edges. Here is the function: def drop_block_fast_2d( x: torch.Tensor, drop_prob: float = 0.1, block_size: int = 7, gamma_scale: float = 1.0, with_noise: bool = False, inplace: bool = False, batchwise: bool = False): """ DropBlock. See https://arxiv.org/pdf/1810.12890.pdf DropBlock with an experimental gaussian noise option. Simplied from above without concern for valid block mask at edges. """ B, C, H, W = x.shape total_size = W * H clipped_block_size = min(block_size, min(W, H)) gamma = gamma_scale * drop_prob * total_size / clipped_block_size ** 2 / ( (W - block_size + 1) * (H - block_size + 1)) if batchwise: # one mask for whole batch, quite a bit faster block_mask = torch.rand((1, C, H, W), dtype=x.dtype, device=x.device) < gamma else: # mask per batch element block_mask = torch.rand_like(x) < gamma block_mask = F.max_pool2d( block_mask.to(x.dtype), kernel_size=clipped_block_size, stride=1, padding=clipped_block_size // 2) if with_noise: normal_noise = torch.randn((1, C, H, W), dtype=x.dtype, device=x.device) if batchwise else torch.randn_like(x) if inplace: x.mul_(1. - block_mask).add_(normal_noise * block_mask) else: x = x * (1. - block_mask) + normal_noise * block_mask else: block_mask = 1 - block_mask normalize_scale = (block_mask.numel() / block_mask.to(dtype=torch.float32).sum().add(1e-7)).to(dtype=x.dtype) if inplace: x.mul_(block_mask * normalize_scale) else: x = x * block_mask * normalize_scale return x
DropBlock. See https://arxiv.org/pdf/1810.12890.pdf DropBlock with an experimental gaussian noise option. Simplied from above without concern for valid block mask at edges.
178,960
import torch import math import warnings def _no_grad_trunc_normal_(tensor, mean, std, a, b): # Cut & paste from PyTorch official master until it's in a few official releases - RW # Method based on https://people.sc.fsu.edu/~jburkardt/presentations/truncated_normal.pdf def norm_cdf(x): # Computes standard normal cumulative distribution function return (1. + math.erf(x / math.sqrt(2.))) / 2. if (mean < a - 2 * std) or (mean > b + 2 * std): warnings.warn("mean is more than 2 std from [a, b] in nn.init.trunc_normal_. " "The distribution of values may be incorrect.", stacklevel=2) with torch.no_grad(): # Values are generated by using a truncated uniform distribution and # then using the inverse CDF for the normal distribution. # Get upper and lower cdf values l = norm_cdf((a - mean) / std) u = norm_cdf((b - mean) / std) # Uniformly fill tensor with values from [l, u], then translate to # [2l-1, 2u-1]. tensor.uniform_(2 * l - 1, 2 * u - 1) # Use inverse cdf transform for normal distribution to get truncated # standard normal tensor.erfinv_() # Transform to proper mean, std tensor.mul_(std * math.sqrt(2.)) tensor.add_(mean) # Clamp to ensure it's in the proper range tensor.clamp_(min=a, max=b) return tensor The provided code snippet includes necessary dependencies for implementing the `trunc_normal_` function. Write a Python function `def trunc_normal_(tensor, mean=0., std=1., a=-2., b=2.)` to solve the following problem: r"""Fills the input Tensor with values drawn from a truncated normal distribution. The values are effectively drawn from the normal distribution :math:`\mathcal{N}(\text{mean}, \text{std}^2)` with values outside :math:`[a, b]` redrawn until they are within the bounds. The method used for generating the random values works best when :math:`a \leq \text{mean} \leq b`. Args: tensor: an n-dimensional `torch.Tensor` mean: the mean of the normal distribution std: the standard deviation of the normal distribution a: the minimum cutoff value b: the maximum cutoff value Examples: >>> w = torch.empty(3, 5) >>> nn.init.trunc_normal_(w) Here is the function: def trunc_normal_(tensor, mean=0., std=1., a=-2., b=2.): # type: (Tensor, float, float, float, float) -> Tensor r"""Fills the input Tensor with values drawn from a truncated normal distribution. The values are effectively drawn from the normal distribution :math:`\mathcal{N}(\text{mean}, \text{std}^2)` with values outside :math:`[a, b]` redrawn until they are within the bounds. The method used for generating the random values works best when :math:`a \leq \text{mean} \leq b`. Args: tensor: an n-dimensional `torch.Tensor` mean: the mean of the normal distribution std: the standard deviation of the normal distribution a: the minimum cutoff value b: the maximum cutoff value Examples: >>> w = torch.empty(3, 5) >>> nn.init.trunc_normal_(w) """ return _no_grad_trunc_normal_(tensor, mean, std, a, b)
r"""Fills the input Tensor with values drawn from a truncated normal distribution. The values are effectively drawn from the normal distribution :math:`\mathcal{N}(\text{mean}, \text{std}^2)` with values outside :math:`[a, b]` redrawn until they are within the bounds. The method used for generating the random values works best when :math:`a \leq \text{mean} \leq b`. Args: tensor: an n-dimensional `torch.Tensor` mean: the mean of the normal distribution std: the standard deviation of the normal distribution a: the minimum cutoff value b: the maximum cutoff value Examples: >>> w = torch.empty(3, 5) >>> nn.init.trunc_normal_(w)
178,961
from mmcv.utils import collect_env as collect_base_env from mmcv.utils import get_git_hash import mmseg The provided code snippet includes necessary dependencies for implementing the `collect_env` function. Write a Python function `def collect_env()` to solve the following problem: Collect the information of the running environments. Here is the function: def collect_env(): """Collect the information of the running environments.""" env_info = collect_base_env() env_info['MMSegmentation'] = f'{mmseg.__version__}+{get_git_hash()[:7]}' return env_info
Collect the information of the running environments.
178,962
import logging from mmcv.utils import get_logger def get_root_logger(log_file=None, log_level=logging.INFO): """Get the root logger. The logger will be initialized if it has not been initialized. By default a StreamHandler will be added. If `log_file` is specified, a FileHandler will also be added. The name of the root logger is the top-level package name, e.g., "mmseg". Args: log_file (str | None): The log filename. If specified, a FileHandler will be added to the root logger. log_level (int): The root logger level. Note that only the process of rank 0 is affected, while other processes will set the level to "Error" and be silent most of the time. Returns: logging.Logger: The root logger. """ logger = get_logger(name='mmseg', log_file=log_file, log_level=log_level) return logger The provided code snippet includes necessary dependencies for implementing the `print_log` function. Write a Python function `def print_log(msg, logger=None, level=logging.INFO)` to solve the following problem: Print a log message. Args: msg (str): The message to be logged. logger (logging.Logger | str | None): The logger to be used. Some special loggers are: - "root": the root logger obtained with `get_root_logger()`. - "silent": no message will be printed. - None: The `print()` method will be used to print log messages. level (int): Logging level. Only available when `logger` is a Logger object or "root". Here is the function: def print_log(msg, logger=None, level=logging.INFO): """Print a log message. Args: msg (str): The message to be logged. logger (logging.Logger | str | None): The logger to be used. Some special loggers are: - "root": the root logger obtained with `get_root_logger()`. - "silent": no message will be printed. - None: The `print()` method will be used to print log messages. level (int): Logging level. Only available when `logger` is a Logger object or "root". """ if logger is None: print(msg) elif logger == 'root': _logger = get_root_logger() _logger.log(level, msg) elif isinstance(logger, logging.Logger): logger.log(level, msg) elif logger != 'silent': raise TypeError( 'logger should be either a logging.Logger object, "root", ' '"silent" or None, but got {}'.format(logger))
Print a log message. Args: msg (str): The message to be logged. logger (logging.Logger | str | None): The logger to be used. Some special loggers are: - "root": the root logger obtained with `get_root_logger()`. - "silent": no message will be printed. - None: The `print()` method will be used to print log messages. level (int): Logging level. Only available when `logger` is a Logger object or "root".
178,963
import argparse import copy import os import os.path as osp import time import mmcv import torch from mmcv.runner import init_dist from mmcv.utils import Config, DictAction, get_git_hash from mmseg import __version__ from mmseg.apis import set_random_seed, train_segmentor from mmseg.datasets import build_dataset from mmseg.models import build_segmentor from mmseg.utils import collect_env, get_root_logger def parse_args(): parser = argparse.ArgumentParser(description='Train a segmentor') parser.add_argument('config', help='train config file path') parser.add_argument('--work-dir', help='the dir to save logs and models') parser.add_argument( '--load-from', help='the checkpoint file to load weights from') parser.add_argument( '--resume-from', help='the checkpoint file to resume from') parser.add_argument( '--no-validate', action='store_true', help='whether not to evaluate the checkpoint during training') group_gpus = parser.add_mutually_exclusive_group() group_gpus.add_argument( '--gpus', type=int, help='number of gpus to use ' '(only applicable to non-distributed training)') group_gpus.add_argument( '--gpu-ids', type=int, nargs='+', help='ids of gpus to use ' '(only applicable to non-distributed training)') parser.add_argument('--seed', type=int, default=None, help='random seed') parser.add_argument( '--deterministic', action='store_true', help='whether to set deterministic options for CUDNN backend.') parser.add_argument( '--options', nargs='+', action=DictAction, help='custom options') parser.add_argument( '--launcher', choices=['none', 'pytorch', 'slurm', 'mpi'], default='none', help='job launcher') parser.add_argument('--local_rank', type=int, default=0) args = parser.parse_args() if 'LOCAL_RANK' not in os.environ: os.environ['LOCAL_RANK'] = str(args.local_rank) return args
null
178,964
import argparse from mmcv import Config from mmcv.cnn import get_model_complexity_info from mmcv.cnn.utils.flops_counter import flops_to_string, params_to_string from mmseg.models import build_segmentor import torch def parse_args(): parser = argparse.ArgumentParser(description='Train a segmentor') parser.add_argument('config', help='train config file path') parser.add_argument( '--shape', type=int, nargs='+', default=[2048, 1024], help='input image size') args = parser.parse_args() return args
null
178,965
import argparse from mmcv import Config from mmcv.cnn import get_model_complexity_info from mmcv.cnn.utils.flops_counter import flops_to_string, params_to_string from mmseg.models import build_segmentor import torch def sra_flops(h, w, r, dim, num_heads): def get_tr_flops(net, input_shape): flops, params = get_model_complexity_info(net, input_shape, as_strings=False) _, H, W = input_shape net = net.backbone try: stage1 = sra_flops(H // 4, W // 4, net.block1[0].attn.sr_ratio, net.block1[0].attn.dim, net.block1[0].attn.num_heads) * len(net.block1) stage2 = sra_flops(H // 8, W // 8, net.block2[0].attn.sr_ratio, net.block2[0].attn.dim, net.block2[0].attn.num_heads) * len(net.block2) stage3 = sra_flops(H // 16, W // 16, net.block3[0].attn.sr_ratio, net.block3[0].attn.dim, net.block3[0].attn.num_heads) * len(net.block3) stage4 = sra_flops(H // 32, W // 32, net.block4[0].attn.sr_ratio, net.block4[0].attn.dim, net.block4[0].attn.num_heads) * len(net.block4) except: stage1 = sra_flops(H // 4, W // 4, net.block1[0].attn.squeeze_ratio, 64, net.block1[0].attn.num_heads) * len(net.block1) stage2 = sra_flops(H // 8, W // 8, net.block2[0].attn.squeeze_ratio, 128, net.block2[0].attn.num_heads) * len(net.block2) stage3 = sra_flops(H // 16, W // 16, net.block3[0].attn.squeeze_ratio, 320, net.block3[0].attn.num_heads) * len(net.block3) stage4 = sra_flops(H // 32, W // 32, net.block4[0].attn.squeeze_ratio, 512, net.block4[0].attn.num_heads) * len(net.block4) print(stage1 + stage2 + stage3 + stage4) flops += stage1 + stage2 + stage3 + stage4 return flops_to_string(flops), params_to_string(params)
null
178,966
import argparse import copy import os import os.path as osp import time import mmcv import torch from mmcv.runner import init_dist from mmcv.utils import Config, DictAction, get_git_hash from IPython import embed from collections import OrderedDict def parse_args(): parser = argparse.ArgumentParser(description='Train a segmentor') parser.add_argument('oldmodel', help='train config file path') parser.add_argument('newmodel', help='train config file path') args = parser.parse_args() return args
null
178,967
import argparse from functools import partial import mmcv import numpy as np import onnxruntime as rt import torch import torch._C import torch.serialization from mmcv.onnx import register_extra_symbolics from mmcv.runner import load_checkpoint from torch import nn from mmseg.models import build_segmentor torch.manual_seed(3) def _convert_batchnorm(module): module_output = module if isinstance(module, torch.nn.SyncBatchNorm): module_output = torch.nn.BatchNorm2d(module.num_features, module.eps, module.momentum, module.affine, module.track_running_stats) if module.affine: module_output.weight.data = module.weight.data.clone().detach() module_output.bias.data = module.bias.data.clone().detach() # keep requires_grad unchanged module_output.weight.requires_grad = module.weight.requires_grad module_output.bias.requires_grad = module.bias.requires_grad module_output.running_mean = module.running_mean module_output.running_var = module.running_var module_output.num_batches_tracked = module.num_batches_tracked for name, child in module.named_children(): module_output.add_module(name, _convert_batchnorm(child)) del module return module_output
null
178,968
import argparse from functools import partial import mmcv import numpy as np import onnxruntime as rt import torch import torch._C import torch.serialization from mmcv.onnx import register_extra_symbolics from mmcv.runner import load_checkpoint from torch import nn from mmseg.models import build_segmentor torch.manual_seed(3) def _demo_mm_inputs(input_shape, num_classes): """Create a superset of inputs needed to run test or train batches. Args: input_shape (tuple): input batch dimensions num_classes (int): number of semantic classes """ (N, C, H, W) = input_shape rng = np.random.RandomState(0) imgs = rng.rand(*input_shape) segs = rng.randint( low=0, high=num_classes - 1, size=(N, 1, H, W)).astype(np.uint8) img_metas = [{ 'img_shape': (H, W, C), 'ori_shape': (H, W, C), 'pad_shape': (H, W, C), 'filename': '<demo>.png', 'scale_factor': 1.0, 'flip': False, } for _ in range(N)] mm_inputs = { 'imgs': torch.FloatTensor(imgs).requires_grad_(True), 'img_metas': img_metas, 'gt_semantic_seg': torch.LongTensor(segs) } return mm_inputs The provided code snippet includes necessary dependencies for implementing the `pytorch2onnx` function. Write a Python function `def pytorch2onnx(model, input_shape, opset_version=11, show=False, output_file='tmp.onnx', verify=False)` to solve the following problem: Export Pytorch model to ONNX model and verify the outputs are same between Pytorch and ONNX. Args: model (nn.Module): Pytorch model we want to export. input_shape (tuple): Use this input shape to construct the corresponding dummy input and execute the model. opset_version (int): The onnx op version. Default: 11. show (bool): Whether print the computation graph. Default: False. output_file (string): The path to where we store the output ONNX model. Default: `tmp.onnx`. verify (bool): Whether compare the outputs between Pytorch and ONNX. Default: False. Here is the function: def pytorch2onnx(model, input_shape, opset_version=11, show=False, output_file='tmp.onnx', verify=False): """Export Pytorch model to ONNX model and verify the outputs are same between Pytorch and ONNX. Args: model (nn.Module): Pytorch model we want to export. input_shape (tuple): Use this input shape to construct the corresponding dummy input and execute the model. opset_version (int): The onnx op version. Default: 11. show (bool): Whether print the computation graph. Default: False. output_file (string): The path to where we store the output ONNX model. Default: `tmp.onnx`. verify (bool): Whether compare the outputs between Pytorch and ONNX. Default: False. """ model.cpu().eval() if isinstance(model.decode_head, nn.ModuleList): num_classes = model.decode_head[-1].num_classes else: num_classes = model.decode_head.num_classes mm_inputs = _demo_mm_inputs(input_shape, num_classes) imgs = mm_inputs.pop('imgs') img_metas = mm_inputs.pop('img_metas') img_list = [img[None, :] for img in imgs] img_meta_list = [[img_meta] for img_meta in img_metas] # replace original forward function origin_forward = model.forward model.forward = partial( model.forward, img_metas=img_meta_list, return_loss=False) register_extra_symbolics(opset_version) with torch.no_grad(): torch.onnx.export( model, (img_list, ), output_file, export_params=True, keep_initializers_as_inputs=True, verbose=show, opset_version=opset_version) print(f'Successfully exported ONNX model: {output_file}') model.forward = origin_forward if verify: # check by onnx import onnx onnx_model = onnx.load(output_file) onnx.checker.check_model(onnx_model) # check the numerical value # get pytorch output pytorch_result = model(img_list, img_meta_list, return_loss=False)[0] # get onnx output input_all = [node.name for node in onnx_model.graph.input] input_initializer = [ node.name for node in onnx_model.graph.initializer ] net_feed_input = list(set(input_all) - set(input_initializer)) assert (len(net_feed_input) == 1) sess = rt.InferenceSession(output_file) onnx_result = sess.run( None, {net_feed_input[0]: img_list[0].detach().numpy()})[0] if not np.allclose(pytorch_result, onnx_result): raise ValueError( 'The outputs are different between Pytorch and ONNX') print('The outputs are same between Pytorch and ONNX')
Export Pytorch model to ONNX model and verify the outputs are same between Pytorch and ONNX. Args: model (nn.Module): Pytorch model we want to export. input_shape (tuple): Use this input shape to construct the corresponding dummy input and execute the model. opset_version (int): The onnx op version. Default: 11. show (bool): Whether print the computation graph. Default: False. output_file (string): The path to where we store the output ONNX model. Default: `tmp.onnx`. verify (bool): Whether compare the outputs between Pytorch and ONNX. Default: False.
178,969
import argparse from functools import partial import mmcv import numpy as np import onnxruntime as rt import torch import torch._C import torch.serialization from mmcv.onnx import register_extra_symbolics from mmcv.runner import load_checkpoint from torch import nn from mmseg.models import build_segmentor def parse_args(): parser = argparse.ArgumentParser(description='Convert MMSeg to ONNX') parser.add_argument('config', help='test config file path') parser.add_argument('--checkpoint', help='checkpoint file', default=None) parser.add_argument('--show', action='store_true', help='show onnx graph') parser.add_argument( '--verify', action='store_true', help='verify the onnx model') parser.add_argument('--output-file', type=str, default='tmp.onnx') parser.add_argument('--opset-version', type=int, default=11) parser.add_argument( '--shape', type=int, nargs='+', default=[256, 256], help='input image size') args = parser.parse_args() return args
null
178,984
import argparse import time import torch from mmcv import Config from mmcv.parallel import MMDataParallel from mmcv.runner import load_checkpoint from mmseg.datasets import build_dataloader, build_dataset from mmseg.models import build_segmentor def parse_args(): parser = argparse.ArgumentParser(description='MMSeg benchmark a model') parser.add_argument('config', help='test config file path') # parser.add_argument('checkpoint', help='checkpoint file') parser.add_argument( '--log-interval', type=int, default=50, help='interval of logging') args = parser.parse_args() return args
null
178,989
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import time import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * args = parse_args() description = 'Welcome to Our TinyGPT-V Chatbot Demo!' def parse_args(): parser = argparse.ArgumentParser(description="Demo") parser.add_argument("--cfg-path", default='eval_configs/minigptv2_eval.yaml', help="path to configuration file.") parser.add_argument("--gpu-id", type=int, default=0, help="specify the gpu to load the model.") parser.add_argument( "--options", nargs="+", help="override some settings in the used config, the key-value pair " "in xxx=yyy format will be merged into config file (deprecate), " "change to --cfg-options instead.", ) args = parser.parse_args() return args
null
178,990
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import time import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * with gr.Blocks() as demo: gr.Markdown(title) # gr.Markdown(description) gr.Markdown(article) with gr.Row(): with gr.Column(scale=0.5): image = gr.Image(type="pil", tool='sketch', brush_radius=20) temperature = gr.Slider( minimum=0.1, maximum=1.5, value=0.6, step=0.1, interactive=True, label="Temperature", ) clear = gr.Button("Restart") gr.Markdown(introduction) with gr.Column(): chat_state = gr.State(value=None) img_list = gr.State(value=[]) chatbot = gr.Chatbot(label='MiniGPT-v2') dataset = gr.Dataset( components=[gr.Textbox(visible=False)], samples=[['No Tag'], ['VQA']], type="index", label='Task Shortcuts', ) task_inst = gr.Markdown('**Hint:** Upload your image and chat') with gr.Row(): text_input.render() send = gr.Button("Send", variant='primary', size='sm', scale=1) upload_flag = gr.State(value=0) replace_flag = gr.State(value=0) image.upload(image_upload_trigger, [upload_flag, replace_flag, img_list], [upload_flag, replace_flag]) with gr.Row(): # with gr.Column(): # # gr.Examples(examples=[ # # ["examples_v2/office.jpg", "[grounding] describe this image in detail", upload_flag, replace_flag, # # img_list], # # ["examples_v2/sofa.jpg", "[detection] sofas", upload_flag, replace_flag, img_list], # # ["examples_v2/2000x1372_wmkn_0012149409555.jpg", "[refer] the world cup", upload_flag, replace_flag, # # img_list], # # ["examples_v2/KFC-20-for-20-Nuggets.jpg", "[identify] what is this {<4><50><30><65>}", upload_flag, # # replace_flag, img_list], # # ], inputs=[image, text_input, upload_flag, replace_flag, img_list], fn=example_trigger, # # outputs=[upload_flag, replace_flag]) with gr.Column(): gr.Examples(examples=[ ["examples_v2/glip_test.jpg", "[vqa] where should I hide in this room when playing hide and seek", upload_flag, replace_flag, img_list], ["examples_v2/float.png", "Please write a poem about the image", upload_flag, replace_flag, img_list], # ["examples_v2/thief.png", "Is the weapon fateful", upload_flag, replace_flag, img_list], ["examples_v2/cockdial.png", "What might happen in this image in the next second", upload_flag, replace_flag, img_list], ], inputs=[image, text_input, upload_flag, replace_flag, img_list], fn=example_trigger, outputs=[upload_flag, replace_flag]) dataset.click( gradio_taskselect, inputs=[dataset], outputs=[text_input, task_inst], show_progress="hidden", postprocess=False, queue=False, ) text_input.submit( gradio_ask, [text_input, chatbot, chat_state, image, img_list, upload_flag, replace_flag], [text_input, chatbot, chat_state, img_list, upload_flag, replace_flag], queue=False ).success( gradio_stream_answer, [chatbot, chat_state, img_list, temperature], [chatbot, chat_state] ).success( gradio_visualize, [chatbot, image], [chatbot], queue=False, ) send.click( gradio_ask, [text_input, chatbot, chat_state, image, img_list, upload_flag, replace_flag], [text_input, chatbot, chat_state, img_list, upload_flag, replace_flag], queue=False ).success( gradio_stream_answer, [chatbot, chat_state, img_list, temperature], [chatbot, chat_state] ).success( gradio_visualize, [chatbot, image], [chatbot], queue=False, ) clear.click(gradio_reset, [chat_state, img_list], [chatbot, image, text_input, chat_state, img_list], queue=False) def gradio_reset(chat_state, img_list): if chat_state is not None: chat_state.messages = [] if img_list is not None: img_list = [] return None, gr.update(value=None, interactive=True), gr.update(placeholder='Upload your image and chat', interactive=True), chat_state, img_list
null
178,991
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import time import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * def image_upload_trigger(upload_flag, replace_flag, img_list): # set the upload flag to true when receive a new image. # if there is an old image (and old conversation), set the replace flag to true to reset the conv later. upload_flag = 1 if img_list: replace_flag = 1 return upload_flag, replace_flag
null
178,992
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import time import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * def example_trigger(text_input, image, upload_flag, replace_flag, img_list): # set the upload flag to true when receive a new image. # if there is an old image (and old conversation), set the replace flag to true to reset the conv later. upload_flag = 1 if img_list or replace_flag == 1: replace_flag = 1 return upload_flag, replace_flag
null
178,993
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import time import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * CONV_VISION = Conversation( system="", roles=(r"<s>[INST] ", r" [/INST]"), messages=[], offset=2, sep_style=SeparatorStyle.SINGLE, sep="", ) def save_tmp_img(visual_img): file_name = "".join([str(random.randint(0, 9)) for _ in range(5)]) + ".jpg" file_path = "/tmp/gradio" + file_name visual_img.save(file_path) return file_path def mask2bbox(mask): if mask is None: return '' mask = mask.resize([100, 100], resample=Image.NEAREST) mask = np.array(mask)[:, :, 0] rows = np.any(mask, axis=1) cols = np.any(mask, axis=0) if rows.sum(): # Get the top, bottom, left, and right boundaries rmin, rmax = np.where(rows)[0][[0, -1]] cmin, cmax = np.where(cols)[0][[0, -1]] bbox = '{{<{}><{}><{}><{}>}}'.format(cmin, rmin, cmax, rmax) else: bbox = '' return bbox def visualize_all_bbox_together(image, generation): if image is None: return None, '' generation = html.unescape(generation) image_width, image_height = image.size image = image.resize([500, int(500 / image_width * image_height)]) image_width, image_height = image.size string_list = extract_substrings(generation) if string_list: # it is grounding or detection mode = 'all' entities = defaultdict(list) i = 0 j = 0 for string in string_list: try: obj, string = string.split('</p>') except ValueError: print('wrong string: ', string) continue bbox_list = string.split('<delim>') flag = False for bbox_string in bbox_list: integers = re.findall(r'-?\d+', bbox_string) if len(integers) == 4: x0, y0, x1, y1 = int(integers[0]), int(integers[1]), int(integers[2]), int(integers[3]) left = x0 / bounding_box_size * image_width bottom = y0 / bounding_box_size * image_height right = x1 / bounding_box_size * image_width top = y1 / bounding_box_size * image_height entities[obj].append([left, bottom, right, top]) j += 1 flag = True if flag: i += 1 else: integers = re.findall(r'-?\d+', generation) if len(integers) == 4: # it is refer mode = 'single' entities = list() x0, y0, x1, y1 = int(integers[0]), int(integers[1]), int(integers[2]), int(integers[3]) left = x0 / bounding_box_size * image_width bottom = y0 / bounding_box_size * image_height right = x1 / bounding_box_size * image_width top = y1 / bounding_box_size * image_height entities.append([left, bottom, right, top]) else: # don't detect any valid bbox to visualize return None, '' if len(entities) == 0: return None, '' if isinstance(image, Image.Image): image_h = image.height image_w = image.width image = np.array(image) elif isinstance(image, str): if os.path.exists(image): pil_img = Image.open(image).convert("RGB") image = np.array(pil_img)[:, :, [2, 1, 0]] image_h = pil_img.height image_w = pil_img.width else: raise ValueError(f"invaild image path, {image}") elif isinstance(image, torch.Tensor): image_tensor = image.cpu() reverse_norm_mean = torch.tensor([0.48145466, 0.4578275, 0.40821073])[:, None, None] reverse_norm_std = torch.tensor([0.26862954, 0.26130258, 0.27577711])[:, None, None] image_tensor = image_tensor * reverse_norm_std + reverse_norm_mean pil_img = T.ToPILImage()(image_tensor) image_h = pil_img.height image_w = pil_img.width image = np.array(pil_img)[:, :, [2, 1, 0]] else: raise ValueError(f"invaild image format, {type(image)} for {image}") indices = list(range(len(entities))) new_image = image.copy() previous_bboxes = [] # size of text text_size = 0.5 # thickness of text text_line = 1 # int(max(1 * min(image_h, image_w) / 512, 1)) box_line = 2 (c_width, text_height), _ = cv2.getTextSize("F", cv2.FONT_HERSHEY_COMPLEX, text_size, text_line) base_height = int(text_height * 0.675) text_offset_original = text_height - base_height text_spaces = 2 # num_bboxes = sum(len(x[-1]) for x in entities) used_colors = colors # random.sample(colors, k=num_bboxes) color_id = -1 for entity_idx, entity_name in enumerate(entities): if mode == 'single' or mode == 'identify': bboxes = entity_name bboxes = [bboxes] else: bboxes = entities[entity_name] color_id += 1 for bbox_id, (x1_norm, y1_norm, x2_norm, y2_norm) in enumerate(bboxes): skip_flag = False orig_x1, orig_y1, orig_x2, orig_y2 = int(x1_norm), int(y1_norm), int(x2_norm), int(y2_norm) color = used_colors[entity_idx % len(used_colors)] # tuple(np.random.randint(0, 255, size=3).tolist()) new_image = cv2.rectangle(new_image, (orig_x1, orig_y1), (orig_x2, orig_y2), color, box_line) if mode == 'all': l_o, r_o = box_line // 2 + box_line % 2, box_line // 2 + box_line % 2 + 1 x1 = orig_x1 - l_o y1 = orig_y1 - l_o if y1 < text_height + text_offset_original + 2 * text_spaces: y1 = orig_y1 + r_o + text_height + text_offset_original + 2 * text_spaces x1 = orig_x1 + r_o # add text background (text_width, text_height), _ = cv2.getTextSize(f" {entity_name}", cv2.FONT_HERSHEY_COMPLEX, text_size, text_line) text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2 = x1, y1 - ( text_height + text_offset_original + 2 * text_spaces), x1 + text_width, y1 for prev_bbox in previous_bboxes: if computeIoU((text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2), prev_bbox['bbox']) > 0.95 and \ prev_bbox['phrase'] == entity_name: skip_flag = True break while is_overlapping((text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2), prev_bbox['bbox']): text_bg_y1 += (text_height + text_offset_original + 2 * text_spaces) text_bg_y2 += (text_height + text_offset_original + 2 * text_spaces) y1 += (text_height + text_offset_original + 2 * text_spaces) if text_bg_y2 >= image_h: text_bg_y1 = max(0, image_h - (text_height + text_offset_original + 2 * text_spaces)) text_bg_y2 = image_h y1 = image_h break if not skip_flag: alpha = 0.5 for i in range(text_bg_y1, text_bg_y2): for j in range(text_bg_x1, text_bg_x2): if i < image_h and j < image_w: if j < text_bg_x1 + 1.35 * c_width: # original color bg_color = color else: # white bg_color = [255, 255, 255] new_image[i, j] = (alpha * new_image[i, j] + (1 - alpha) * np.array(bg_color)).astype( np.uint8) cv2.putText( new_image, f" {entity_name}", (x1, y1 - text_offset_original - 1 * text_spaces), cv2.FONT_HERSHEY_COMPLEX, text_size, (0, 0, 0), text_line, cv2.LINE_AA ) previous_bboxes.append( {'bbox': (text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2), 'phrase': entity_name}) if mode == 'all': def color_iterator(colors): while True: for color in colors: yield color color_gen = color_iterator(colors) # Add colors to phrases and remove <p></p> def colored_phrases(match): phrase = match.group(1) color = next(color_gen) return f'<span style="color:rgb{color}">{phrase}</span>' generation = re.sub(r'{<\d+><\d+><\d+><\d+>}|<delim>', '', generation) generation_colored = re.sub(r'<p>(.*?)</p>', colored_phrases, generation) else: generation_colored = '' pil_image = Image.fromarray(new_image) return pil_image, generation_colored chat = Chat(model, vis_processor, device=device) def gradio_ask(user_message, chatbot, chat_state, gr_img, img_list, upload_flag, replace_flag): if len(user_message) == 0: text_box_show = 'Input should not be empty!' else: text_box_show = '' if isinstance(gr_img, dict): gr_img, mask = gr_img['image'], gr_img['mask'] else: mask = None if '[identify]' in user_message: # check if user provide bbox in the text input integers = re.findall(r'-?\d+', user_message) if len(integers) != 4: # no bbox in text bbox = mask2bbox(mask) user_message = user_message + bbox if chat_state is None: chat_state = CONV_VISION.copy() if upload_flag: if replace_flag: chat_state = CONV_VISION.copy() # new image, reset everything replace_flag = 0 chatbot = [] img_list = [] llm_message = chat.upload_img(gr_img, chat_state, img_list) upload_flag = 0 chat.ask(user_message, chat_state) chatbot = chatbot + [[user_message, None]] if '[identify]' in user_message: visual_img, _ = visualize_all_bbox_together(gr_img, user_message) if visual_img is not None: file_path = save_tmp_img(visual_img) chatbot = chatbot + [[(file_path,), None]] return text_box_show, chatbot, chat_state, img_list, upload_flag, replace_flag
null
178,994
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import time import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * chat = Chat(model, vis_processor, device=device) def gradio_answer(chatbot, chat_state, img_list, temperature): llm_message = chat.answer(conv=chat_state, img_list=img_list, temperature=temperature, max_new_tokens=500, max_length=2000)[0] chatbot[-1][1] = llm_message return chatbot, chat_state
null
178,995
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import time import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * torch.manual_seed(42) print('Initializing Chat') def escape_markdown(text): # List of Markdown special characters that need to be escaped md_chars = ['<', '>'] # Escape each special character for char in md_chars: text = text.replace(char, '\\' + char) return text chat = Chat(model, vis_processor, device=device) def gradio_stream_answer(chatbot, chat_state, img_list, temperature): if len(img_list) > 0: if not isinstance(img_list[0], torch.Tensor): chat.encode_img(img_list) streamer = chat.stream_answer(conv=chat_state, img_list=img_list, temperature=temperature, max_new_tokens=500, max_length=2000) output = '' previous_time = time.time() for new_output in streamer: if '###' in new_output: current_time = time.time() inference_time = current_time - previous_time new_output = new_output.split('###')[0] output += escape_markdown(new_output) chatbot[-1][1] = output print(inference_time) yield chatbot, chat_state break escapped = escape_markdown(new_output) output += escapped chatbot[-1][1] = output yield chatbot, chat_state return chatbot, chat_state
null
178,996
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import time import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * def save_tmp_img(visual_img): file_name = "".join([str(random.randint(0, 9)) for _ in range(5)]) + ".jpg" file_path = "/tmp/gradio" + file_name visual_img.save(file_path) return file_path def reverse_escape(text): md_chars = ['\\<', '\\>'] for char in md_chars: text = text.replace(char, char[1:]) return text def visualize_all_bbox_together(image, generation): if image is None: return None, '' generation = html.unescape(generation) image_width, image_height = image.size image = image.resize([500, int(500 / image_width * image_height)]) image_width, image_height = image.size string_list = extract_substrings(generation) if string_list: # it is grounding or detection mode = 'all' entities = defaultdict(list) i = 0 j = 0 for string in string_list: try: obj, string = string.split('</p>') except ValueError: print('wrong string: ', string) continue bbox_list = string.split('<delim>') flag = False for bbox_string in bbox_list: integers = re.findall(r'-?\d+', bbox_string) if len(integers) == 4: x0, y0, x1, y1 = int(integers[0]), int(integers[1]), int(integers[2]), int(integers[3]) left = x0 / bounding_box_size * image_width bottom = y0 / bounding_box_size * image_height right = x1 / bounding_box_size * image_width top = y1 / bounding_box_size * image_height entities[obj].append([left, bottom, right, top]) j += 1 flag = True if flag: i += 1 else: integers = re.findall(r'-?\d+', generation) if len(integers) == 4: # it is refer mode = 'single' entities = list() x0, y0, x1, y1 = int(integers[0]), int(integers[1]), int(integers[2]), int(integers[3]) left = x0 / bounding_box_size * image_width bottom = y0 / bounding_box_size * image_height right = x1 / bounding_box_size * image_width top = y1 / bounding_box_size * image_height entities.append([left, bottom, right, top]) else: # don't detect any valid bbox to visualize return None, '' if len(entities) == 0: return None, '' if isinstance(image, Image.Image): image_h = image.height image_w = image.width image = np.array(image) elif isinstance(image, str): if os.path.exists(image): pil_img = Image.open(image).convert("RGB") image = np.array(pil_img)[:, :, [2, 1, 0]] image_h = pil_img.height image_w = pil_img.width else: raise ValueError(f"invaild image path, {image}") elif isinstance(image, torch.Tensor): image_tensor = image.cpu() reverse_norm_mean = torch.tensor([0.48145466, 0.4578275, 0.40821073])[:, None, None] reverse_norm_std = torch.tensor([0.26862954, 0.26130258, 0.27577711])[:, None, None] image_tensor = image_tensor * reverse_norm_std + reverse_norm_mean pil_img = T.ToPILImage()(image_tensor) image_h = pil_img.height image_w = pil_img.width image = np.array(pil_img)[:, :, [2, 1, 0]] else: raise ValueError(f"invaild image format, {type(image)} for {image}") indices = list(range(len(entities))) new_image = image.copy() previous_bboxes = [] # size of text text_size = 0.5 # thickness of text text_line = 1 # int(max(1 * min(image_h, image_w) / 512, 1)) box_line = 2 (c_width, text_height), _ = cv2.getTextSize("F", cv2.FONT_HERSHEY_COMPLEX, text_size, text_line) base_height = int(text_height * 0.675) text_offset_original = text_height - base_height text_spaces = 2 # num_bboxes = sum(len(x[-1]) for x in entities) used_colors = colors # random.sample(colors, k=num_bboxes) color_id = -1 for entity_idx, entity_name in enumerate(entities): if mode == 'single' or mode == 'identify': bboxes = entity_name bboxes = [bboxes] else: bboxes = entities[entity_name] color_id += 1 for bbox_id, (x1_norm, y1_norm, x2_norm, y2_norm) in enumerate(bboxes): skip_flag = False orig_x1, orig_y1, orig_x2, orig_y2 = int(x1_norm), int(y1_norm), int(x2_norm), int(y2_norm) color = used_colors[entity_idx % len(used_colors)] # tuple(np.random.randint(0, 255, size=3).tolist()) new_image = cv2.rectangle(new_image, (orig_x1, orig_y1), (orig_x2, orig_y2), color, box_line) if mode == 'all': l_o, r_o = box_line // 2 + box_line % 2, box_line // 2 + box_line % 2 + 1 x1 = orig_x1 - l_o y1 = orig_y1 - l_o if y1 < text_height + text_offset_original + 2 * text_spaces: y1 = orig_y1 + r_o + text_height + text_offset_original + 2 * text_spaces x1 = orig_x1 + r_o # add text background (text_width, text_height), _ = cv2.getTextSize(f" {entity_name}", cv2.FONT_HERSHEY_COMPLEX, text_size, text_line) text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2 = x1, y1 - ( text_height + text_offset_original + 2 * text_spaces), x1 + text_width, y1 for prev_bbox in previous_bboxes: if computeIoU((text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2), prev_bbox['bbox']) > 0.95 and \ prev_bbox['phrase'] == entity_name: skip_flag = True break while is_overlapping((text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2), prev_bbox['bbox']): text_bg_y1 += (text_height + text_offset_original + 2 * text_spaces) text_bg_y2 += (text_height + text_offset_original + 2 * text_spaces) y1 += (text_height + text_offset_original + 2 * text_spaces) if text_bg_y2 >= image_h: text_bg_y1 = max(0, image_h - (text_height + text_offset_original + 2 * text_spaces)) text_bg_y2 = image_h y1 = image_h break if not skip_flag: alpha = 0.5 for i in range(text_bg_y1, text_bg_y2): for j in range(text_bg_x1, text_bg_x2): if i < image_h and j < image_w: if j < text_bg_x1 + 1.35 * c_width: # original color bg_color = color else: # white bg_color = [255, 255, 255] new_image[i, j] = (alpha * new_image[i, j] + (1 - alpha) * np.array(bg_color)).astype( np.uint8) cv2.putText( new_image, f" {entity_name}", (x1, y1 - text_offset_original - 1 * text_spaces), cv2.FONT_HERSHEY_COMPLEX, text_size, (0, 0, 0), text_line, cv2.LINE_AA ) previous_bboxes.append( {'bbox': (text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2), 'phrase': entity_name}) if mode == 'all': def color_iterator(colors): while True: for color in colors: yield color color_gen = color_iterator(colors) # Add colors to phrases and remove <p></p> def colored_phrases(match): phrase = match.group(1) color = next(color_gen) return f'<span style="color:rgb{color}">{phrase}</span>' generation = re.sub(r'{<\d+><\d+><\d+><\d+>}|<delim>', '', generation) generation_colored = re.sub(r'<p>(.*?)</p>', colored_phrases, generation) else: generation_colored = '' pil_image = Image.fromarray(new_image) return pil_image, generation_colored def gradio_visualize(chatbot, gr_img): if isinstance(gr_img, dict): gr_img, mask = gr_img['image'], gr_img['mask'] unescaped = reverse_escape(chatbot[-1][1]) visual_img, generation_color = visualize_all_bbox_together(gr_img, unescaped) if visual_img is not None: if len(generation_color): chatbot[-1][1] = generation_color file_path = save_tmp_img(visual_img) chatbot = chatbot + [[None, (file_path,)]] return chatbot
null
178,997
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import time import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * def gradio_taskselect(idx): prompt_list = [ '', # '[grounding] describe this image in detail', # '[refer] ', # '[detection] ', # '[identify] what is this ', '[vqa] ' ] instruct_list = [ '**Hint:** Type in whatever you want', # '**Hint:** Send the command to generate a grounded image description', # '**Hint:** Type in a phrase about an object in the image and send the command', # '**Hint:** Type in a caption or phrase, and see object locations in the image', # '**Hint:** Draw a bounding box on the uploaded image then send the command. Click the "clear" botton on the top right of the image before redraw', '**Hint:** Send a question to get a short answer', ] return prompt_list[idx], instruct_list[idx]
null
179,003
import math from typing import List, Optional, Tuple, Union import torch import torch.nn.functional as F import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACT2FN from ...cache_utils import Cache, DynamicCache from ...modeling_attn_mask_utils import _prepare_4d_causal_attention_mask from ...modeling_outputs import ( BaseModelOutputWithPast, CausalLMOutputWithPast, SequenceClassifierOutputWithPast, TokenClassifierOutput, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, is_flash_attn_2_available, is_flash_attn_greater_or_equal_2_10, logging, replace_return_docstrings, ) from .configuration_phi import PhiConfig def _get_unpad_data(attention_mask): seqlens_in_batch = attention_mask.sum(dim=-1, dtype=torch.int32) indices = torch.nonzero(attention_mask.flatten(), as_tuple=False).flatten() max_seqlen_in_batch = seqlens_in_batch.max().item() cu_seqlens = F.pad(torch.cumsum(seqlens_in_batch, dim=0, dtype=torch.torch.int32), (1, 0)) return ( indices, cu_seqlens, max_seqlen_in_batch, )
null
179,004
import math from typing import List, Optional, Tuple, Union import torch import torch.nn.functional as F import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACT2FN from ...cache_utils import Cache, DynamicCache from ...modeling_attn_mask_utils import _prepare_4d_causal_attention_mask from ...modeling_outputs import ( BaseModelOutputWithPast, CausalLMOutputWithPast, SequenceClassifierOutputWithPast, TokenClassifierOutput, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, is_flash_attn_2_available, is_flash_attn_greater_or_equal_2_10, logging, replace_return_docstrings, ) from .configuration_phi import PhiConfig def rotate_half(x): """Rotates half the hidden dims of the input.""" x1 = x[..., : x.shape[-1] // 2] x2 = x[..., x.shape[-1] // 2 :] return torch.cat((-x2, x1), dim=-1) The provided code snippet includes necessary dependencies for implementing the `apply_rotary_pos_emb` function. Write a Python function `def apply_rotary_pos_emb(q, k, cos, sin, position_ids, unsqueeze_dim=1)` to solve the following problem: Applies Rotary Position Embedding to the query and key tensors. Args: q (`torch.Tensor`): The query tensor. k (`torch.Tensor`): The key tensor. cos (`torch.Tensor`): The cosine part of the rotary embedding. sin (`torch.Tensor`): The sine part of the rotary embedding. position_ids (`torch.Tensor`): The position indices of the tokens corresponding to the query and key tensors. For example, this can be used to pass offsetted position ids when working with a KV-cache. unsqueeze_dim (`int`, *optional*, defaults to 1): The 'unsqueeze_dim' argument specifies the dimension along which to unsqueeze cos[position_ids] and sin[position_ids] so that they can be properly broadcasted to the dimensions of q and k. For example, note that cos[position_ids] and sin[position_ids] have the shape [batch_size, seq_len, head_dim]. Then, if q and k have the shape [batch_size, heads, seq_len, head_dim], then setting unsqueeze_dim=1 makes cos[position_ids] and sin[position_ids] broadcastable to the shapes of q and k. Similarly, if q and k have the shape [batch_size, seq_len, heads, head_dim], then set unsqueeze_dim=2. Returns: `tuple(torch.Tensor)` comprising of the query and key tensors rotated using the Rotary Position Embedding. Here is the function: def apply_rotary_pos_emb(q, k, cos, sin, position_ids, unsqueeze_dim=1): """Applies Rotary Position Embedding to the query and key tensors. Args: q (`torch.Tensor`): The query tensor. k (`torch.Tensor`): The key tensor. cos (`torch.Tensor`): The cosine part of the rotary embedding. sin (`torch.Tensor`): The sine part of the rotary embedding. position_ids (`torch.Tensor`): The position indices of the tokens corresponding to the query and key tensors. For example, this can be used to pass offsetted position ids when working with a KV-cache. unsqueeze_dim (`int`, *optional*, defaults to 1): The 'unsqueeze_dim' argument specifies the dimension along which to unsqueeze cos[position_ids] and sin[position_ids] so that they can be properly broadcasted to the dimensions of q and k. For example, note that cos[position_ids] and sin[position_ids] have the shape [batch_size, seq_len, head_dim]. Then, if q and k have the shape [batch_size, heads, seq_len, head_dim], then setting unsqueeze_dim=1 makes cos[position_ids] and sin[position_ids] broadcastable to the shapes of q and k. Similarly, if q and k have the shape [batch_size, seq_len, heads, head_dim], then set unsqueeze_dim=2. Returns: `tuple(torch.Tensor)` comprising of the query and key tensors rotated using the Rotary Position Embedding. """ cos = cos[position_ids].unsqueeze(unsqueeze_dim) sin = sin[position_ids].unsqueeze(unsqueeze_dim) q_embed = (q * cos) + (rotate_half(q) * sin) k_embed = (k * cos) + (rotate_half(k) * sin) return q_embed, k_embed
Applies Rotary Position Embedding to the query and key tensors. Args: q (`torch.Tensor`): The query tensor. k (`torch.Tensor`): The key tensor. cos (`torch.Tensor`): The cosine part of the rotary embedding. sin (`torch.Tensor`): The sine part of the rotary embedding. position_ids (`torch.Tensor`): The position indices of the tokens corresponding to the query and key tensors. For example, this can be used to pass offsetted position ids when working with a KV-cache. unsqueeze_dim (`int`, *optional*, defaults to 1): The 'unsqueeze_dim' argument specifies the dimension along which to unsqueeze cos[position_ids] and sin[position_ids] so that they can be properly broadcasted to the dimensions of q and k. For example, note that cos[position_ids] and sin[position_ids] have the shape [batch_size, seq_len, head_dim]. Then, if q and k have the shape [batch_size, heads, seq_len, head_dim], then setting unsqueeze_dim=1 makes cos[position_ids] and sin[position_ids] broadcastable to the shapes of q and k. Similarly, if q and k have the shape [batch_size, seq_len, heads, head_dim], then set unsqueeze_dim=2. Returns: `tuple(torch.Tensor)` comprising of the query and key tensors rotated using the Rotary Position Embedding.
179,012
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * CONV_VISION = Conversation( system="", roles=(r"<s>[INST] ", r" [/INST]"), messages=[], offset=2, sep_style=SeparatorStyle.SINGLE, sep="", ) def save_tmp_img(visual_img): file_name = "".join([str(random.randint(0, 9)) for _ in range(5)]) + ".jpg" file_path = "/tmp/gradio" + file_name visual_img.save(file_path) return file_path def mask2bbox(mask): if mask is None: return '' mask = mask.resize([100, 100], resample=Image.NEAREST) mask = np.array(mask)[:, :, 0] rows = np.any(mask, axis=1) cols = np.any(mask, axis=0) if rows.sum(): # Get the top, bottom, left, and right boundaries rmin, rmax = np.where(rows)[0][[0, -1]] cmin, cmax = np.where(cols)[0][[0, -1]] bbox = '{{<{}><{}><{}><{}>}}'.format(cmin, rmin, cmax, rmax) else: bbox = '' return bbox def visualize_all_bbox_together(image, generation): if image is None: return None, '' generation = html.unescape(generation) image_width, image_height = image.size image = image.resize([500, int(500 / image_width * image_height)]) image_width, image_height = image.size string_list = extract_substrings(generation) if string_list: # it is grounding or detection mode = 'all' entities = defaultdict(list) i = 0 j = 0 for string in string_list: try: obj, string = string.split('</p>') except ValueError: print('wrong string: ', string) continue bbox_list = string.split('<delim>') flag = False for bbox_string in bbox_list: integers = re.findall(r'-?\d+', bbox_string) if len(integers) == 4: x0, y0, x1, y1 = int(integers[0]), int(integers[1]), int(integers[2]), int(integers[3]) left = x0 / bounding_box_size * image_width bottom = y0 / bounding_box_size * image_height right = x1 / bounding_box_size * image_width top = y1 / bounding_box_size * image_height entities[obj].append([left, bottom, right, top]) j += 1 flag = True if flag: i += 1 else: integers = re.findall(r'-?\d+', generation) if len(integers) == 4: # it is refer mode = 'single' entities = list() x0, y0, x1, y1 = int(integers[0]), int(integers[1]), int(integers[2]), int(integers[3]) left = x0 / bounding_box_size * image_width bottom = y0 / bounding_box_size * image_height right = x1 / bounding_box_size * image_width top = y1 / bounding_box_size * image_height entities.append([left, bottom, right, top]) else: # don't detect any valid bbox to visualize return None, '' if len(entities) == 0: return None, '' if isinstance(image, Image.Image): image_h = image.height image_w = image.width image = np.array(image) elif isinstance(image, str): if os.path.exists(image): pil_img = Image.open(image).convert("RGB") image = np.array(pil_img)[:, :, [2, 1, 0]] image_h = pil_img.height image_w = pil_img.width else: raise ValueError(f"invaild image path, {image}") elif isinstance(image, torch.Tensor): image_tensor = image.cpu() reverse_norm_mean = torch.tensor([0.48145466, 0.4578275, 0.40821073])[:, None, None] reverse_norm_std = torch.tensor([0.26862954, 0.26130258, 0.27577711])[:, None, None] image_tensor = image_tensor * reverse_norm_std + reverse_norm_mean pil_img = T.ToPILImage()(image_tensor) image_h = pil_img.height image_w = pil_img.width image = np.array(pil_img)[:, :, [2, 1, 0]] else: raise ValueError(f"invaild image format, {type(image)} for {image}") indices = list(range(len(entities))) new_image = image.copy() previous_bboxes = [] # size of text text_size = 0.5 # thickness of text text_line = 1 # int(max(1 * min(image_h, image_w) / 512, 1)) box_line = 2 (c_width, text_height), _ = cv2.getTextSize("F", cv2.FONT_HERSHEY_COMPLEX, text_size, text_line) base_height = int(text_height * 0.675) text_offset_original = text_height - base_height text_spaces = 2 # num_bboxes = sum(len(x[-1]) for x in entities) used_colors = colors # random.sample(colors, k=num_bboxes) color_id = -1 for entity_idx, entity_name in enumerate(entities): if mode == 'single' or mode == 'identify': bboxes = entity_name bboxes = [bboxes] else: bboxes = entities[entity_name] color_id += 1 for bbox_id, (x1_norm, y1_norm, x2_norm, y2_norm) in enumerate(bboxes): skip_flag = False orig_x1, orig_y1, orig_x2, orig_y2 = int(x1_norm), int(y1_norm), int(x2_norm), int(y2_norm) color = used_colors[entity_idx % len(used_colors)] # tuple(np.random.randint(0, 255, size=3).tolist()) new_image = cv2.rectangle(new_image, (orig_x1, orig_y1), (orig_x2, orig_y2), color, box_line) if mode == 'all': l_o, r_o = box_line // 2 + box_line % 2, box_line // 2 + box_line % 2 + 1 x1 = orig_x1 - l_o y1 = orig_y1 - l_o if y1 < text_height + text_offset_original + 2 * text_spaces: y1 = orig_y1 + r_o + text_height + text_offset_original + 2 * text_spaces x1 = orig_x1 + r_o # add text background (text_width, text_height), _ = cv2.getTextSize(f" {entity_name}", cv2.FONT_HERSHEY_COMPLEX, text_size, text_line) text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2 = x1, y1 - ( text_height + text_offset_original + 2 * text_spaces), x1 + text_width, y1 for prev_bbox in previous_bboxes: if computeIoU((text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2), prev_bbox['bbox']) > 0.95 and \ prev_bbox['phrase'] == entity_name: skip_flag = True break while is_overlapping((text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2), prev_bbox['bbox']): text_bg_y1 += (text_height + text_offset_original + 2 * text_spaces) text_bg_y2 += (text_height + text_offset_original + 2 * text_spaces) y1 += (text_height + text_offset_original + 2 * text_spaces) if text_bg_y2 >= image_h: text_bg_y1 = max(0, image_h - (text_height + text_offset_original + 2 * text_spaces)) text_bg_y2 = image_h y1 = image_h break if not skip_flag: alpha = 0.5 for i in range(text_bg_y1, text_bg_y2): for j in range(text_bg_x1, text_bg_x2): if i < image_h and j < image_w: if j < text_bg_x1 + 1.35 * c_width: # original color bg_color = color else: # white bg_color = [255, 255, 255] new_image[i, j] = (alpha * new_image[i, j] + (1 - alpha) * np.array(bg_color)).astype( np.uint8) cv2.putText( new_image, f" {entity_name}", (x1, y1 - text_offset_original - 1 * text_spaces), cv2.FONT_HERSHEY_COMPLEX, text_size, (0, 0, 0), text_line, cv2.LINE_AA ) previous_bboxes.append( {'bbox': (text_bg_x1, text_bg_y1, text_bg_x2, text_bg_y2), 'phrase': entity_name}) if mode == 'all': def color_iterator(colors): while True: for color in colors: yield color color_gen = color_iterator(colors) # Add colors to phrases and remove <p></p> def colored_phrases(match): phrase = match.group(1) color = next(color_gen) return f'<span style="color:rgb{color}">{phrase}</span>' generation = re.sub(r'{<\d+><\d+><\d+><\d+>}|<delim>', '', generation) generation_colored = re.sub(r'<p>(.*?)</p>', colored_phrases, generation) else: generation_colored = '' pil_image = Image.fromarray(new_image) return pil_image, generation_colored chat = Chat(model, vis_processor, device=device) def gradio_ask(user_message, chatbot, chat_state, gr_img, img_list, upload_flag, replace_flag): if len(user_message) == 0: text_box_show = 'Input should not be empty!' else: text_box_show = '' if isinstance(gr_img, dict): gr_img, mask = gr_img['image'], gr_img['mask'] else: mask = None if '[identify]' in user_message: # check if user provide bbox in the text input integers = re.findall(r'-?\d+', user_message) if len(integers) != 4: # no bbox in text bbox = mask2bbox(mask) user_message = user_message + bbox if chat_state is None: chat_state = CONV_VISION.copy() if upload_flag: if replace_flag: chat_state = CONV_VISION.copy() # new image, reset everything replace_flag = 0 chatbot = [] img_list = [] llm_message = chat.upload_img(gr_img, chat_state, img_list) upload_flag = 0 chat.ask(user_message, chat_state) chatbot = chatbot + [[user_message, None]] if '[identify]' in user_message: visual_img, _ = visualize_all_bbox_together(gr_img, user_message) if visual_img is not None: file_path = save_tmp_img(visual_img) chatbot = chatbot + [[(file_path,), None]] return text_box_show, chatbot, chat_state, img_list, upload_flag, replace_flag
null
179,014
import argparse import os import random from collections import defaultdict import cv2 import re import numpy as np from PIL import Image import torch import html import gradio as gr import torchvision.transforms as T import torch.backends.cudnn as cudnn from minigpt4.common.config import Config from minigpt4.common.registry import registry from minigpt4.conversation.conversation import Conversation, SeparatorStyle, Chat from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * torch.manual_seed(42) def escape_markdown(text): # List of Markdown special characters that need to be escaped md_chars = ['<', '>'] # Escape each special character for char in md_chars: text = text.replace(char, '\\' + char) return text chat = Chat(model, vis_processor, device=device) def gradio_stream_answer(chatbot, chat_state, img_list, temperature): if len(img_list) > 0: if not isinstance(img_list[0], torch.Tensor): chat.encode_img(img_list) streamer = chat.stream_answer(conv=chat_state, img_list=img_list, temperature=temperature, max_new_tokens=500, max_length=2000) output = '' for new_output in streamer: if '###' in new_output: # 如果在输出中发现 '###',则截取至 '###' 之前的内容 new_output = new_output.split('###')[0] output += escape_markdown(new_output) chatbot[-1][1] = output yield chatbot, chat_state break # 停止循环,不再生成新的输出 escapped = escape_markdown(new_output) output += escapped chatbot[-1][1] = output yield chatbot, chat_state chat_state.messages[-1][1] = '</s>' return chatbot, chat_state
null
179,051
import gzip import logging import os import random as rnd import tarfile import zipfile import random from typing import List from tqdm import tqdm import decord from decord import VideoReader import webdataset as wds import numpy as np import torch from torch.utils.data.dataset import IterableDataset from minigpt4.common.registry import registry from minigpt4.datasets.datasets.base_dataset import ConcatDataset def move_to_cuda(sample): def prepare_sample(samples, cuda_enabled=True): if cuda_enabled: samples = move_to_cuda(samples) # TODO fp16 support return samples
null
179,055
import os import json import pickle import random import time import itertools import numpy as np from PIL import Image import skimage.io as io import matplotlib.pyplot as plt from matplotlib.collections import PatchCollection from matplotlib.patches import Polygon, Rectangle from torch.utils.data import Dataset import webdataset as wds from minigpt4.datasets.datasets.base_dataset import BaseDataset from minigpt4.datasets.datasets.caption_datasets import CaptionDataset def __getitem__(self, index): while True: try: sample = self.data[index] image_path = os.path.join(self.vis_root, sample['image_path']) image = Image.open(image_path).convert("RGB") image = self.vis_processor(image) question = self.text_processor(sample["question"]) answer = self.text_processor(sample["answer"]) instruction = random.choice(self.instruction_pool).format(question) instruction = "<Img><ImageHere></Img> {} ".format(instruction) return { "image": image, "instruction_input": instruction, "answer": answer, "image_id": sample['image_id'] } except FileNotFoundError: print(f'File {image_path} not found. Skip to next.') index = (index + 1) % len(self.data) # 确保index不会超出范围
null
179,088
import os import logging import contextlib from omegaconf import OmegaConf import numpy as np import torch import torch.nn as nn from transformers import AutoTokenizer from peft import ( LoraConfig, get_peft_model, prepare_model_for_int8_training, ) from minigpt4.common.dist_utils import download_cached_file from minigpt4.common.utils import get_abs_path, is_url from minigpt4.models.eva_vit import create_eva_vit_g from transformers import PhiForCausalLM The provided code snippet includes necessary dependencies for implementing the `disabled_train` function. Write a Python function `def disabled_train(self, mode=True)` to solve the following problem: Overwrite model.train with this function to make sure train/eval mode does not change anymore. Here is the function: def disabled_train(self, mode=True): """Overwrite model.train with this function to make sure train/eval mode does not change anymore.""" return self
Overwrite model.train with this function to make sure train/eval mode does not change anymore.
179,089
from __future__ import annotations import math from dataclasses import dataclass, field from typing import Any, Dict, Optional, Tuple, Union import torch import torch.nn as nn from einops import rearrange, repeat from transformers import PretrainedConfig, PreTrainedModel from transformers.activations import ACT2FN from transformers.modeling_outputs import CausalLMOutputWithPast from .configuration_phi import PhiConfig def _apply_rotary_emb( x: torch.FloatTensor, cos: torch.FloatTensor, sin: torch.FloatTensor, ) -> torch.FloatTensor: _, seqlen, _, _ = x.shape _, rotary_dim = cos.shape rotary_dim *= 2 x_rot = x[:, :, :, :rotary_dim] x_pass = x[:, :, :, rotary_dim:] x1, x2 = x_rot.chunk(2, dim=-1) c, s = rearrange(cos[:seqlen], "s d -> s 1 d"), rearrange(sin[:seqlen], "s d -> s 1 d") x1, x2, c, s = [t.to(dtype=torch.float32) for t in [x1, x2, c, s]] x_rot = torch.cat([x1 * c - x2 * s, x1 * s + x2 * c], axis=-1).to(x.dtype) return torch.cat([x_rot, x_pass], axis=-1)
null
179,090
from __future__ import annotations import math from dataclasses import dataclass, field from typing import Any, Dict, Optional, Tuple, Union import torch import torch.nn as nn from einops import rearrange, repeat from transformers import PretrainedConfig, PreTrainedModel from transformers.activations import ACT2FN from transformers.modeling_outputs import CausalLMOutputWithPast from .configuration_phi import PhiConfig def _apply_rotary_emb_kv( kv: torch.FloatTensor, cos: torch.FloatTensor, sin: torch.FloatTensor, cos_k: Optional[torch.FloatTensor] = None, sin_k: Optional[torch.FloatTensor] = None, ) -> torch.FloatTensor: _, seqlen, _, _, _ = kv.shape _, rotary_dim = cos.shape rotary_dim *= 2 k_rot = kv[:, :, 0, :, :rotary_dim] k_pass = kv[:, :, 0, :, rotary_dim:] k1, k2 = k_rot.chunk(2, dim=-1) c, s = rearrange(cos[:seqlen], "s d -> s 1 d"), rearrange(sin[:seqlen], "s d -> s 1 d") k1, k2, c, s = [t.to(dtype=torch.float32) for t in [k1, k2, c, s]] k_rot = torch.cat([k1 * c - k2 * s, k1 * s + k2 * c], axis=-1).to(kv.dtype) return torch.cat( [ torch.cat([k_rot, k_pass], axis=-1).unsqueeze(2), kv[:, :, 1:2, :, :], ], axis=2, )
null
179,091
from __future__ import annotations import math from dataclasses import dataclass, field from typing import Any, Dict, Optional, Tuple, Union import torch import torch.nn as nn from einops import rearrange, repeat from transformers import PretrainedConfig, PreTrainedModel from transformers.activations import ACT2FN from transformers.modeling_outputs import CausalLMOutputWithPast from .configuration_phi import PhiConfig def _apply_rotary_emb_qkv( qkv: torch.FloatTensor, cos: torch.FloatTensor, sin: torch.FloatTensor, cos_k: Optional[torch.FloatTensor] = None, sin_k: Optional[torch.FloatTensor] = None, ) -> torch.FloatTensor: _, seqlen, _, _, _ = qkv.shape _, rotary_dim = cos.shape rotary_dim *= 2 q_rot = qkv[:, :, 0, :, :rotary_dim] q_pass = qkv[:, :, 0, :, rotary_dim:] k_rot = qkv[:, :, 1, :, :rotary_dim] k_pass = qkv[:, :, 1, :, rotary_dim:] q1, q2 = q_rot.chunk(2, dim=-1) k1, k2 = k_rot.chunk(2, dim=-1) c, s = rearrange(cos[:seqlen], "s d -> s 1 d"), rearrange(sin[:seqlen], "s d -> s 1 d") q1, q2, k1, k2, c, s = [t.to(dtype=torch.float32) for t in [q1, q2, k1, k2, c, s]] q_rot = torch.cat([q1 * c - q2 * s, q1 * s + q2 * c], axis=-1).to(qkv.dtype) k_rot = torch.cat([k1 * c - k2 * s, k1 * s + k2 * c], axis=-1).to(qkv.dtype) return torch.cat( [ torch.cat([q_rot, q_pass], axis=-1).unsqueeze(2), torch.cat([k_rot, k_pass], axis=-1).unsqueeze(2), qkv[:, :, 2:3, :, :], ], axis=2, )
null
179,092
from __future__ import annotations import math from dataclasses import dataclass, field from typing import Any, Dict, Optional, Tuple, Union import torch import torch.nn as nn from einops import rearrange, repeat from transformers import PretrainedConfig, PreTrainedModel from transformers.activations import ACT2FN from transformers.modeling_outputs import CausalLMOutputWithPast from .configuration_phi import PhiConfig def _find_mha_dims( config: PretrainedConfig, n_head: Optional[int] = None, n_head_kv: Optional[int] = None, head_dim: Optional[int] = None, ) -> Tuple[int, int]: if n_head is None and head_dim is None: head_dim = config.n_embd // config.n_head n_head = config.n_head elif n_head is None or head_dim is None: raise ValueError("`n_head` and `head_dim` must be both specified or `None`.") if n_head_kv is None: n_head_kv = getattr(config, "n_head_kv", None) or n_head return n_head, n_head_kv, head_dim
null
179,093
from __future__ import annotations import math from dataclasses import dataclass, field from typing import Any, Dict, Optional, Tuple, Union import torch import torch.nn as nn from einops import rearrange, repeat from transformers import PretrainedConfig, PreTrainedModel from transformers.activations import ACT2FN from transformers.modeling_outputs import CausalLMOutputWithPast from .configuration_phi import PhiConfig class InferenceParams: """Inference parameters passed to model to efficiently calculate and store context during inference. Reference: https://github.com/Dao-AILab/flash-attention/blob/main/flash_attn/utils/generation.py. Args: max_seqlen: Maximum sequence length. max_batch_size: Maximum batch size. seqlen_offset: Sequence length offset. batch_size_offset: Batch size offset. key_value_memory_dict: Key value memory dictionary. lengths_per_sample: Lengths per sample. """ max_seqlen: int = field(metadata={"help": "Maximum sequence length."}) max_batch_size: int = field(metadata={"help": "Maximum batch size."}) seqlen_offset: int = field(default=0, metadata={"help": "Sequence length offset."}) batch_size_offset: int = field(default=0, metadata={"help": "Batch size offset."}) key_value_memory_dict: Dict[str, Any] = field( default_factory=dict, metadata={"help": "Key value memory dictionary."} ) lengths_per_sample: torch.Tensor = field(default=None, metadata={"help": "Lengths per sample."}) def _update_kv_cache(kv: torch.FloatTensor, inference_params: InferenceParams, layer_idx: int) -> torch.FloatTensor: num_heads, head_dim = kv.shape[-2:] if layer_idx not in inference_params.key_value_memory_dict: inference_params.key_value_memory_dict[layer_idx] = torch.empty( inference_params.max_batch_size, inference_params.max_seqlen, 2, num_heads, head_dim, dtype=kv.dtype, device=kv.device, ) batch_start = inference_params.batch_size_offset batch_end = batch_start + kv.shape[0] sequence_start = inference_params.seqlen_offset sequence_end = sequence_start + kv.shape[1] # When the current sequence length is equal to or larger than the maximum sequence length, # we need to concatenate the current `kv` with the cached `kv` to expand its length if sequence_end >= inference_params.max_seqlen: inference_params.key_value_memory_dict[layer_idx] = torch.concatenate((inference_params.key_value_memory_dict[layer_idx], kv), dim=1) inference_params.key_value_memory_dict[layer_idx][batch_start:batch_end, sequence_start:sequence_end, ...] = kv kv = inference_params.key_value_memory_dict[layer_idx][batch_start:batch_end, :sequence_end, ...] return kv
null
179,152
import argparse import numpy as np from nltk.translate.bleu_score import sentence_bleu from minigpt4.common.registry import registry from minigpt4.common.config import Config from minigpt4.datasets.builders import * from minigpt4.models import * from minigpt4.processors import * from minigpt4.runners import * from minigpt4.tasks import * registry = Registry() class Config: def __init__(self, args): def _validate_runner_config(self, runner_config): def _build_opt_list(self, opts): def build_model_config(config, **kwargs): def build_runner_config(config): def build_dataset_config(config): def build_evaluation_dataset_config(config): def _convert_to_dot_list(self, opts): def get_config(self): def run_cfg(self): def datasets_cfg(self): def evaluation_datasets_cfg(self): def model_cfg(self): def pretty_print(self): def _convert_node_to_json(self, node): def to_dict(self): def init_model(args): print('Initialization Model') cfg = Config(args) # cfg.model_cfg.ckpt = args.ckpt # cfg.model_cfg.lora_r = args.lora_r # cfg.model_cfg.lora_alpha = args.lora_alpha model_config = cfg.model_cfg model_cls = registry.get_model_class(model_config.arch) model = model_cls.from_config(model_config).to('cuda:0') # import pudb; pudb.set_trace() key = list(cfg.datasets_cfg.keys())[0] vis_processor_cfg = cfg.datasets_cfg.get(key).vis_processor.train vis_processor = registry.get_processor_class(vis_processor_cfg.name).from_config(vis_processor_cfg) print('Initialization Finished') return model, vis_processor
null
179,174
import cv2 import os import torch import torch.nn as nn from torchvision.transforms import Compose from .midas.dpt_depth import DPTDepthModel from .midas.midas_net import MidasNet from .midas.midas_net_custom import MidasNet_small from .midas.transforms import Resize, NormalizeImage, PrepareForNet from annotator.util import annotator_ckpts_path ISL_PATHS = { "dpt_large": os.path.join(annotator_ckpts_path, "dpt_large-midas-2f21e586.pt"), "dpt_hybrid": os.path.join(annotator_ckpts_path, "dpt_hybrid-midas-501f0c75.pt"), "midas_v21": "", "midas_v21_small": "", } remote_model_path = "https://huggingface.co/lllyasviel/ControlNet/resolve/main/annotator/ckpts/dpt_hybrid-midas-501f0c75.pt" class DPTDepthModel(DPT): def __init__(self, path=None, non_negative=True, **kwargs): def forward(self, x): class MidasNet(BaseModel): def __init__(self, path=None, features=256, non_negative=True): def forward(self, x): class MidasNet_small(BaseModel): def __init__(self, path=None, features=64, backbone="efficientnet_lite3", non_negative=True, exportable=True, channels_last=False, align_corners=True, blocks={'expand': True}): def forward(self, x): class Resize(object): def __init__( self, width, height, resize_target=True, keep_aspect_ratio=False, ensure_multiple_of=1, resize_method="lower_bound", image_interpolation_method=cv2.INTER_AREA, ): def constrain_to_multiple_of(self, x, min_val=0, max_val=None): def get_size(self, width, height): def __call__(self, sample): class NormalizeImage(object): def __init__(self, mean, std): def __call__(self, sample): class PrepareForNet(object): def __init__(self): def __call__(self, sample): annotator_ckpts_path = os.path.join(os.path.dirname(__file__), 'ckpts') def load_model(model_type): # https://github.com/isl-org/MiDaS/blob/master/run.py # load network model_path = ISL_PATHS[model_type] if model_type == "dpt_large": # DPT-Large model = DPTDepthModel( path=model_path, backbone="vitl16_384", non_negative=True, ) net_w, net_h = 384, 384 resize_mode = "minimal" normalization = NormalizeImage(mean=[0.5, 0.5, 0.5], std=[0.5, 0.5, 0.5]) elif model_type == "dpt_hybrid": # DPT-Hybrid if not os.path.exists(model_path): from basicsr.utils.download_util import load_file_from_url load_file_from_url(remote_model_path, model_dir=annotator_ckpts_path) model = DPTDepthModel( path=model_path, backbone="vitb_rn50_384", non_negative=True, ) net_w, net_h = 384, 384 resize_mode = "minimal" normalization = NormalizeImage(mean=[0.5, 0.5, 0.5], std=[0.5, 0.5, 0.5]) elif model_type == "midas_v21": model = MidasNet(model_path, non_negative=True) net_w, net_h = 384, 384 resize_mode = "upper_bound" normalization = NormalizeImage( mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225] ) elif model_type == "midas_v21_small": model = MidasNet_small(model_path, features=64, backbone="efficientnet_lite3", exportable=True, non_negative=True, blocks={'expand': True}) net_w, net_h = 256, 256 resize_mode = "upper_bound" normalization = NormalizeImage( mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225] ) else: print(f"model_type '{model_type}' not implemented, use: --model_type large") assert False transform = Compose( [ Resize( net_w, net_h, resize_target=None, keep_aspect_ratio=True, ensure_multiple_of=32, resize_method=resize_mode, image_interpolation_method=cv2.INTER_CUBIC, ), normalization, PrepareForNet(), ] ) return model.eval(), transform
null
179,296
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP def load_video(folder: str, resize=(432, 768), num_frames=70): resy, resx = resize folder = Path(folder) input_files = sorted(list(folder.glob("*.jpg")) + list(folder.glob("*.png")))[:num_frames] video_tensor = torch.zeros((len(input_files), 3, resy, resx)) for i, file in enumerate(input_files): video_tensor[i] = transforms.ToTensor()(Image.open(str(file)).resize((resx, resy), Image.LANCZOS)) return video_tensor
null
179,297
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP class IMLP(nn.Module): def __init__( self, input_dim, output_dim, hidden_dim=256, use_positional=True, positional_dim=10, skip_layers=[4, 6], num_layers=8, # includes the output layer verbose=True, use_tanh=True, apply_softmax=False, ): super(IMLP, self).__init__() self.verbose = verbose self.use_tanh = use_tanh self.apply_softmax = apply_softmax if apply_softmax: self.softmax = nn.Softmax() if use_positional: encoding_dimensions = 2 * input_dim * positional_dim self.b = torch.tensor([(2 ** j) * np.pi for j in range(positional_dim)], requires_grad=False) else: encoding_dimensions = input_dim self.hidden = nn.ModuleList() for i in range(num_layers): if i == 0: input_dims = encoding_dimensions elif i in skip_layers: input_dims = hidden_dim + encoding_dimensions else: input_dims = hidden_dim if i == num_layers - 1: # last layer self.hidden.append(nn.Linear(input_dims, output_dim, bias=True)) else: self.hidden.append(nn.Linear(input_dims, hidden_dim, bias=True)) self.skip_layers = skip_layers self.num_layers = num_layers self.positional_dim = positional_dim self.use_positional = use_positional if self.verbose: print(f"Model has {count_parameters(self)} params") def forward(self, x): if self.use_positional: if self.b.device != x.device: self.b = self.b.to(x.device) pos = positionalEncoding_vec(x, self.b) x = pos input = x.detach().clone() for i, layer in enumerate(self.hidden): if i > 0: x = F.relu(x) if i in self.skip_layers: x = torch.cat((x, input), 1) x = layer(x) if self.use_tanh: x = torch.tanh(x) if self.apply_softmax: x = self.softmax(x) return x def load_neural_atlases_models(config): foreground_mapping = IMLP( input_dim=3, output_dim=2, hidden_dim=256, use_positional=False, num_layers=6, skip_layers=[], ).to(config["device"]) background_mapping = IMLP( input_dim=3, output_dim=2, hidden_dim=256, use_positional=False, num_layers=4, skip_layers=[], ).to(config["device"]) foreground_atlas_model = IMLP( input_dim=2, output_dim=3, hidden_dim=256, use_positional=True, positional_dim=10, num_layers=8, skip_layers=[4, 7], ).to(config["device"]) background_atlas_model = IMLP( input_dim=2, output_dim=3, hidden_dim=256, use_positional=True, positional_dim=10, num_layers=8, skip_layers=[4, 7], ).to(config["device"]) alpha_model = IMLP( input_dim=3, output_dim=1, hidden_dim=256, use_positional=True, positional_dim=5, num_layers=8, skip_layers=[], ).to(config["device"]) checkpoint = torch.load(config["checkpoint_path"]) foreground_mapping.load_state_dict(checkpoint["model_F_mapping1_state_dict"]) background_mapping.load_state_dict(checkpoint["model_F_mapping2_state_dict"]) foreground_atlas_model.load_state_dict(checkpoint["F_atlas_state_dict"]) background_atlas_model.load_state_dict(checkpoint["F_atlas_state_dict"]) alpha_model.load_state_dict(checkpoint["model_F_alpha_state_dict"]) foreground_mapping = foreground_mapping.eval().requires_grad_(False) background_mapping = background_mapping.eval().requires_grad_(False) foreground_atlas_model = foreground_atlas_model.eval().requires_grad_(False) background_atlas_model = background_atlas_model.eval().requires_grad_(False) alpha_model = alpha_model.eval().requires_grad_(False) return foreground_mapping, background_mapping, foreground_atlas_model, background_atlas_model, alpha_model
null
179,298
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP def get_grid_indices(x_start, y_start, h_crop, w_crop, t=None): def get_frames_data(config, foreground_mapping, background_mapping, alpha_model): max_size = max(config["resx"], config["resy"]) normalizing_factor = torch.tensor([max_size / 2, max_size / 2, config["maximum_number_of_frames"] / 2]) background_uv_values = torch.zeros( size=(config["maximum_number_of_frames"], config["resy"], config["resx"], 2), device=config["device"] ) foreground_uv_values = torch.zeros( size=(config["maximum_number_of_frames"], config["resy"], config["resx"], 2), device=config["device"] ) alpha = torch.zeros( size=(config["maximum_number_of_frames"], config["resy"], config["resx"], 1), device=config["device"] ) for frame in tqdm(range(config["maximum_number_of_frames"]), leave=False): indices = get_grid_indices(0, 0, config["resy"], config["resx"], t=torch.tensor(frame)) normalized_chunk = (indices / normalizing_factor - 1).to(config["device"]) # get the atlas UV coordinates from the two mapping networks; with torch.no_grad(): current_background_uv_values = background_mapping(normalized_chunk) current_foreground_uv_values = foreground_mapping(normalized_chunk) current_alpha = alpha_model(normalized_chunk) background_uv_values[frame, indices[:, 1], indices[:, 0]] = current_background_uv_values * 0.5 - 0.5 foreground_uv_values[frame, indices[:, 1], indices[:, 0]] = current_foreground_uv_values * 0.5 + 0.5 current_alpha = 0.5 * (current_alpha + 1.0) current_alpha = 0.99 * current_alpha + 0.001 alpha[frame, indices[:, 1], indices[:, 0]] = current_alpha # config["return_atlas_alpha"] = True if config["return_atlas_alpha"]: # this should take a few minutes foreground_atlas_alpha = torch.zeros( size=( config["maximum_number_of_frames"], config["grid_atlas_resolution"], config["grid_atlas_resolution"], 1, ), ) # foreground_uv_values: 70 x 432 x 768 x 2 foreground_uv_values_grid = foreground_uv_values * config["grid_atlas_resolution"] # indices: 4000000 x 2 indices = get_grid_indices(0, 0, config["grid_atlas_resolution"], config["grid_atlas_resolution"]) for frame in tqdm(range(config["maximum_number_of_frames"]), leave=False): interpolated = scipy.interpolate.griddata( foreground_uv_values_grid[frame].reshape(-1, 2).cpu().numpy(), # 432 x 768 x 2 -> -1 x 2 alpha[frame] .reshape( -1, ) .cpu() .numpy(), indices.reshape(-1, 2).cpu().numpy(), method="linear", ).reshape(config["grid_atlas_resolution"], config["grid_atlas_resolution"], 1) foreground_atlas_alpha[frame] = torch.from_numpy(interpolated) foreground_atlas_alpha[foreground_atlas_alpha.isnan()] = 0.0 foreground_atlas_alpha = ( torch.median(foreground_atlas_alpha, dim=0, keepdim=True).values.to(config["device"]).permute(0, 3, 2, 1) ) else: foreground_atlas_alpha = None return background_uv_values, foreground_uv_values, alpha.permute(0, 3, 1, 2), foreground_atlas_alpha
null
179,299
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP def reconstruct_video_layer(uv_values, atlas_model): t, h, w, _ = uv_values.shape reconstruction = torch.zeros(size=(t, h, w, 3), device=uv_values.device) for frame in range(t): rgb = (atlas_model(uv_values[frame].reshape(-1, 2)) + 1) * 0.5 reconstruction[frame] = rgb.reshape(h, w, 3) return reconstruction.permute(0, 3, 1, 2)
null
179,300
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP def get_grid_indices(x_start, y_start, h_crop, w_crop, t=None): crop_indices = torch.meshgrid(torch.arange(w_crop) + x_start, torch.arange(h_crop) + y_start) crop_indices = torch.stack(crop_indices, dim=-1) crop_indices = crop_indices.reshape(h_crop * w_crop, crop_indices.shape[-1]) if t is not None: crop_indices = torch.cat([crop_indices, t.repeat(h_crop * w_crop, 1)], dim=1) return crop_indices def create_uv_mask(config, mapping_model, min_u, min_v, max_u, max_v, uv_shift=-0.5, resolution_shift=1): max_size = max(config["resx"], config["resy"]) normalizing_factor = torch.tensor([max_size / 2, max_size / 2, config["maximum_number_of_frames"] / 2]) resolution = config["grid_atlas_resolution"] uv_mask = torch.zeros(size=(resolution, resolution), device=config["device"]) for frame in tqdm(range(config["maximum_number_of_frames"]), leave=False): indices = get_grid_indices(0, 0, config["resy"], config["resx"], t=torch.tensor(frame)) for chunk in indices.split(50000, dim=0): normalized_chunk = (chunk / normalizing_factor - 1).to(config["device"]) # get the atlas UV coordinates from the two mapping networks; with torch.no_grad(): uv_values = mapping_model(normalized_chunk) uv_values = uv_values * 0.5 + uv_shift uv_values = ((uv_values + resolution_shift) * resolution).clip(0, resolution - 1) uv_mask[uv_values[:, 1].floor().long(), uv_values[:, 0].floor().long()] = 1 uv_mask[uv_values[:, 1].floor().long(), uv_values[:, 0].ceil().long()] = 1 uv_mask[uv_values[:, 1].ceil().long(), uv_values[:, 0].floor().long()] = 1 uv_mask[uv_values[:, 1].ceil().long(), uv_values[:, 0].ceil().long()] = 1 uv_mask = crop(uv_mask.unsqueeze(0).unsqueeze(0), min_v, min_u, max_v, max_u) return uv_mask.detach().cpu() # shape [1, 1, resolution, resolution]
null
179,301
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP def get_grid_indices(x_start, y_start, h_crop, w_crop, t=None): crop_indices = torch.meshgrid(torch.arange(w_crop) + x_start, torch.arange(h_crop) + y_start) crop_indices = torch.stack(crop_indices, dim=-1) crop_indices = crop_indices.reshape(h_crop * w_crop, crop_indices.shape[-1]) if t is not None: crop_indices = torch.cat([crop_indices, t.repeat(h_crop * w_crop, 1)], dim=1) return crop_indices def get_high_res_atlas(atlas_model, min_v, min_u, max_v, max_u, resolution, device="cuda", layer="background"): inds_grid = get_grid_indices(0, 0, resolution, resolution) inds_grid_chunks = inds_grid.split(50000, dim=0) if layer == "background": shift = -1 else: shift = 0 rendered_atlas = torch.zeros((resolution, resolution, 3)).to(device) # resy, resx, 3 with torch.no_grad(): # reconstruct image row by row for chunk in inds_grid_chunks: normalized_chunk = torch.stack( [ (chunk[:, 0] / resolution) + shift, (chunk[:, 1] / resolution) + shift, ], dim=-1, ).to(device) rgb_output = atlas_model(normalized_chunk) rendered_atlas[chunk[:, 1], chunk[:, 0], :] = rgb_output # move colors to RGB color domain (0,1) rendered_atlas = 0.5 * (rendered_atlas + 1) rendered_atlas = rendered_atlas.permute(2, 0, 1).unsqueeze(0) # shape (1, 3, resy, resx) cropped_atlas = crop( rendered_atlas, min_v, min_u, max_v, max_u, ) return cropped_atlas
null
179,302
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP def get_atlas_crops(uv_values, grid_atlas, augmentation=None): if len(uv_values.shape) == 3: dims = [0, 1] elif len(uv_values.shape) == 4: dims = [0, 1, 2] else: raise ValueError("uv_values should be of shape of len 3 or 4") min_u, min_v = uv_values.amin(dim=dims).long() max_u, max_v = uv_values.amax(dim=dims).ceil().long() # min_u, min_v = uv_values.min(dim=0).values # max_u, max_v = uv_values.max(dim=0).values h_v = max_v - min_v w_u = max_u - min_u atlas_crop = crop(grid_atlas, min_v, min_u, h_v, w_u) if augmentation is not None: atlas_crop = augmentation(atlas_crop) return atlas_crop, torch.stack([min_u, min_v]), torch.stack([max_u, max_v])
null
179,303
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP def get_random_crop_params(input_size, output_size): w, h = input_size th, tw = output_size if h + 1 < th or w + 1 < tw: raise ValueError(f"Required crop size {(th, tw)} is larger then input image size {(h, w)}") if w == tw and h == th: return 0, 0, h, w i = torch.randint(0, h - th + 1, size=(1,)).item() j = torch.randint(0, w - tw + 1, size=(1,)).item() return i, j, th, tw
null
179,304
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP def get_masks_boundaries(alpha_video, border=20, threshold=0.95, min_crop_size=2 ** 7 + 1): resy, resx = alpha_video.shape[-2:] num_frames = alpha_video.shape[0] masks_borders = torch.zeros((num_frames, 4), dtype=torch.int64) for i, file in enumerate(range(num_frames)): mask_im = alpha_video[i] mask_im[mask_im >= threshold] = 1 mask_im[mask_im < threshold] = 0 all_ones = mask_im.squeeze().nonzero() min_y, min_x = torch.maximum(all_ones.min(dim=0).values - border, torch.tensor([0, 0])) max_y, max_x = torch.minimum(all_ones.max(dim=0).values + border, torch.tensor([resy, resx])) h = max_y - min_y w = max_x - min_x if h < min_crop_size: pad = min_crop_size - h if max_y + pad > resy: min_y -= pad else: max_y += pad h = max_y - min_y if w < min_crop_size: pad = min_crop_size - w if max_x + pad > resx: min_x -= pad else: max_x += pad w = max_x - min_x masks_borders[i] = torch.tensor([min_y, min_x, h, w]) return masks_borders
null
179,305
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP def get_atlas_bounding_box(mask_boundaries, grid_atlas, video_uvs): min_uv = torch.tensor(grid_atlas.shape[-2:], device=video_uvs.device) max_uv = torch.tensor([0, 0], device=video_uvs.device) for boundary, frame in zip(mask_boundaries, video_uvs): cropped_uvs = crop(frame.permute(2, 0, 1).unsqueeze(0), *list(boundary)) # 1,2,h,w min_uv = torch.minimum(cropped_uvs.amin(dim=[0, 2, 3]), min_uv).floor().int() max_uv = torch.maximum(cropped_uvs.amax(dim=[0, 2, 3]), max_uv).ceil().int() hw = max_uv - min_uv crop_data = [*list(min_uv)[::-1], *list(hw)[::-1]] return crop(grid_atlas, *crop_data), crop_data
null
179,306
from PIL import Image from pathlib import Path import scipy.interpolate import torch from torchvision import transforms from torchvision.transforms.functional import crop from tqdm import tqdm import numpy as np import cv2 from stablevideo.implicit_neural_networks import IMLP def tensor2im(input_image, imtype=np.uint8): if not isinstance(input_image, np.ndarray): if isinstance(input_image, torch.Tensor): # get the data from a variable image_tensor = input_image.data else: return input_image image_numpy = image_tensor[0].clamp(0.0, 1.0).cpu().float().numpy() # convert it into a numpy array image_numpy = np.transpose(image_numpy, (1, 2, 0)) * 255.0 # post-processing: tranpose and scaling else: # if it is a numpy array, do nothing image_numpy = input_image return image_numpy.astype(imtype)
null
179,307
import torch import torch.nn as nn import torch.nn.functional as F import numpy as np def count_parameters(model): return sum(p.numel() for p in model.parameters() if p.requires_grad)
null
179,308
import torch import torch.nn as nn import torch.nn.functional as F import numpy as np def positionalEncoding_vec(in_tensor, b): proj = torch.einsum("ij, k -> ijk", in_tensor, b) # shape (batch, in_tensor.size(1), freqNum) mapped_coords = torch.cat((torch.sin(proj), torch.cos(proj)), dim=1) # shape (batch, 2*in_tensor.size(1), freqNum) output = mapped_coords.transpose(2, 1).contiguous().view(mapped_coords.size(0), -1) return output
null
179,309
The provided code snippet includes necessary dependencies for implementing the `replace_with_custom_fn_if_matches_filter` function. Write a Python function `def replace_with_custom_fn_if_matches_filter( model, replacement_fn, filter_fn, cur_fqn='' ) -> None` to solve the following problem: For each `child` in `model`, replaces it with `replacement_fn(child)` if `filter_fn(child)` is `True` Here is the function: def replace_with_custom_fn_if_matches_filter( model, replacement_fn, filter_fn, cur_fqn='' ) -> None: """ For each `child` in `model`, replaces it with `replacement_fn(child)` if `filter_fn(child)` is `True` """ name_to_child = dict(model.named_children()) for name, child in name_to_child.items(): if cur_fqn == '': new_fqn = name else: new_fqn = f'{cur_fqn}.{name}' if filter_fn(child, new_fqn): new_child = replacement_fn(child) setattr(model, name, new_child) else: replace_with_custom_fn_if_matches_filter( child, replacement_fn, filter_fn, new_fqn)
For each `child` in `model`, replaces it with `replacement_fn(child)` if `filter_fn(child)` is `True`
179,310
import torch import torch.nn as nn import torch.nn.functional as F from typing import Optional, Tuple, Type from .common import LayerNorm2d, MLPBlock from segment_anything_fast.flash_4 import _attention_rel_h_rel_w The provided code snippet includes necessary dependencies for implementing the `window_partition` function. Write a Python function `def window_partition(x: torch.Tensor, window_size: int) -> Tuple[torch.Tensor, Tuple[int, int]]` to solve the following problem: Partition into non-overlapping windows with padding if needed. Args: x (tensor): input tokens with [B, H, W, C]. window_size (int): window size. Returns: windows: windows after partition with [B * num_windows, window_size, window_size, C]. (Hp, Wp): padded height and width before partition Here is the function: def window_partition(x: torch.Tensor, window_size: int) -> Tuple[torch.Tensor, Tuple[int, int]]: """ Partition into non-overlapping windows with padding if needed. Args: x (tensor): input tokens with [B, H, W, C]. window_size (int): window size. Returns: windows: windows after partition with [B * num_windows, window_size, window_size, C]. (Hp, Wp): padded height and width before partition """ B, H, W, C = x.shape pad_h = (window_size - H % window_size) % window_size pad_w = (window_size - W % window_size) % window_size if pad_h > 0 or pad_w > 0: x = F.pad(x, (0, 0, 0, pad_w, 0, pad_h)) Hp, Wp = H + pad_h, W + pad_w x = x.view(B, Hp // window_size, window_size, Wp // window_size, window_size, C) windows = x.permute(0, 1, 3, 2, 4, 5).contiguous().view(-1, window_size, window_size, C) return windows, (Hp, Wp)
Partition into non-overlapping windows with padding if needed. Args: x (tensor): input tokens with [B, H, W, C]. window_size (int): window size. Returns: windows: windows after partition with [B * num_windows, window_size, window_size, C]. (Hp, Wp): padded height and width before partition
179,311
import torch import torch.nn as nn import torch.nn.functional as F from typing import Optional, Tuple, Type from .common import LayerNorm2d, MLPBlock from segment_anything_fast.flash_4 import _attention_rel_h_rel_w The provided code snippet includes necessary dependencies for implementing the `window_unpartition` function. Write a Python function `def window_unpartition( windows: torch.Tensor, window_size: int, pad_hw: Tuple[int, int], hw: Tuple[int, int] ) -> torch.Tensor` to solve the following problem: Window unpartition into original sequences and removing padding. Args: windows (tensor): input tokens with [B * num_windows, window_size, window_size, C]. window_size (int): window size. pad_hw (Tuple): padded height and width (Hp, Wp). hw (Tuple): original height and width (H, W) before padding. Returns: x: unpartitioned sequences with [B, H, W, C]. Here is the function: def window_unpartition( windows: torch.Tensor, window_size: int, pad_hw: Tuple[int, int], hw: Tuple[int, int] ) -> torch.Tensor: """ Window unpartition into original sequences and removing padding. Args: windows (tensor): input tokens with [B * num_windows, window_size, window_size, C]. window_size (int): window size. pad_hw (Tuple): padded height and width (Hp, Wp). hw (Tuple): original height and width (H, W) before padding. Returns: x: unpartitioned sequences with [B, H, W, C]. """ Hp, Wp = pad_hw H, W = hw B = windows.shape[0] // (Hp * Wp // window_size // window_size) x = windows.view(B, Hp // window_size, Wp // window_size, window_size, window_size, -1) x = x.permute(0, 1, 3, 2, 4, 5).contiguous().view(B, Hp, Wp, -1) if Hp > H or Wp > W: x = x[:, :H, :W, :].contiguous() return x
Window unpartition into original sequences and removing padding. Args: windows (tensor): input tokens with [B * num_windows, window_size, window_size, C]. window_size (int): window size. pad_hw (Tuple): padded height and width (Hp, Wp). hw (Tuple): original height and width (H, W) before padding. Returns: x: unpartitioned sequences with [B, H, W, C].
179,312
import torch import torch.nn as nn import torch.nn.functional as F from typing import Optional, Tuple, Type from .common import LayerNorm2d, MLPBlock from segment_anything_fast.flash_4 import _attention_rel_h_rel_w def get_rel_pos(q_size: int, k_size: int, rel_pos: torch.Tensor) -> torch.Tensor: """ Get relative positional embeddings according to the relative positions of query and key sizes. Args: q_size (int): size of query q. k_size (int): size of key k. rel_pos (Tensor): relative position embeddings (L, C). Returns: Extracted positional embeddings according to relative positions. """ max_rel_dist = int(2 * max(q_size, k_size) - 1) # Interpolate rel pos if needed. if rel_pos.shape[0] != max_rel_dist: # Interpolate rel pos. rel_pos_resized = F.interpolate( rel_pos.reshape(1, rel_pos.shape[0], -1).permute(0, 2, 1), size=max_rel_dist, mode="linear", ) rel_pos_resized = rel_pos_resized.reshape(-1, max_rel_dist).permute(1, 0) else: rel_pos_resized = rel_pos # Scale the coords with short length if shapes for q and k are different. q_coords = torch.arange(q_size, device=rel_pos.device)[:, None] * max(k_size / q_size, 1.0) k_coords = torch.arange(k_size, device=rel_pos.device)[None, :] * max(q_size / k_size, 1.0) relative_coords = (q_coords - k_coords) + (k_size - 1) * max(q_size / k_size, 1.0) return rel_pos_resized[relative_coords.long()] The provided code snippet includes necessary dependencies for implementing the `add_decomposed_rel_pos` function. Write a Python function `def add_decomposed_rel_pos( q: torch.Tensor, rel_pos_h: torch.Tensor, rel_pos_w: torch.Tensor, q_size: Tuple[int, int], k_size: Tuple[int, int], ) -> torch.Tensor` to solve the following problem: Calculate decomposed Relative Positional Embeddings from :paper:`mvitv2`. https://github.com/facebookresearch/mvit/blob/19786631e330df9f3622e5402b4a419a263a2c80/mvit/models/attention.py # noqa B950 Args: attn (Tensor): attention map. q (Tensor): query q in the attention layer with shape (B, q_h * q_w, C). rel_pos_h (Tensor): relative position embeddings (Lh, C) for height axis. rel_pos_w (Tensor): relative position embeddings (Lw, C) for width axis. q_size (Tuple): spatial sequence size of query q with (q_h, q_w). k_size (Tuple): spatial sequence size of key k with (k_h, k_w). Returns: attn (Tensor): attention map with added relative positional embeddings. Here is the function: def add_decomposed_rel_pos( q: torch.Tensor, rel_pos_h: torch.Tensor, rel_pos_w: torch.Tensor, q_size: Tuple[int, int], k_size: Tuple[int, int], ) -> torch.Tensor: """ Calculate decomposed Relative Positional Embeddings from :paper:`mvitv2`. https://github.com/facebookresearch/mvit/blob/19786631e330df9f3622e5402b4a419a263a2c80/mvit/models/attention.py # noqa B950 Args: attn (Tensor): attention map. q (Tensor): query q in the attention layer with shape (B, q_h * q_w, C). rel_pos_h (Tensor): relative position embeddings (Lh, C) for height axis. rel_pos_w (Tensor): relative position embeddings (Lw, C) for width axis. q_size (Tuple): spatial sequence size of query q with (q_h, q_w). k_size (Tuple): spatial sequence size of key k with (k_h, k_w). Returns: attn (Tensor): attention map with added relative positional embeddings. """ q_h, q_w = q_size k_h, k_w = k_size Rh = get_rel_pos(q_h, k_h, rel_pos_h) Rw = get_rel_pos(q_w, k_w, rel_pos_w) B, _, dim = q.shape r_q = q.reshape(B, q_h, q_w, dim) rel_h = torch.einsum("bhwc,hkc->bhwk", r_q, Rh) rel_w = torch.einsum("bhwc,wkc->bhwk", r_q, Rw) rel_h = rel_h.unsqueeze(-1) rel_w = rel_w.unsqueeze(-2) rel_h = rel_h.reshape(B, q_h * q_w, k_h, 1) rel_w = rel_w.reshape(B, q_h * q_w, 1, k_w) return rel_h, rel_w
Calculate decomposed Relative Positional Embeddings from :paper:`mvitv2`. https://github.com/facebookresearch/mvit/blob/19786631e330df9f3622e5402b4a419a263a2c80/mvit/models/attention.py # noqa B950 Args: attn (Tensor): attention map. q (Tensor): query q in the attention layer with shape (B, q_h * q_w, C). rel_pos_h (Tensor): relative position embeddings (Lh, C) for height axis. rel_pos_w (Tensor): relative position embeddings (Lw, C) for width axis. q_size (Tuple): spatial sequence size of query q with (q_h, q_w). k_size (Tuple): spatial sequence size of key k with (k_h, k_w). Returns: attn (Tensor): attention map with added relative positional embeddings.
179,313
import torch from functools import partial from .modeling import ImageEncoderViT, MaskDecoder, PromptEncoder, Sam, TwoWayTransformer def build_sam_vit_h(checkpoint=None): def _apply_eval_dtype_sam(model, dtype): def build_sam_fast_vit_h(checkpoint=None, compile_mode='max-autotune', dtype=torch.bfloat16): sam = build_sam_vit_h(checkpoint) sam = _apply_eval_dtype_sam(sam, dtype) sam.image_encoder = torch.compile(sam.image_encoder, mode=compile_mode) return sam
null
179,314
import torch from functools import partial from .modeling import ImageEncoderViT, MaskDecoder, PromptEncoder, Sam, TwoWayTransformer def build_sam_vit_l(checkpoint=None): return _build_sam( encoder_embed_dim=1024, encoder_depth=24, encoder_num_heads=16, encoder_global_attn_indexes=[5, 11, 17, 23], checkpoint=checkpoint, ) def _apply_eval_dtype_sam(model, dtype): def prep_model(model, dtype): if dtype is not None: return model.eval().to(dtype) return model.eval() model.image_encoder = prep_model(model.image_encoder, dtype) model.prompt_encoder = prep_model(model.prompt_encoder, dtype) model.mask_decoder = prep_model(model.mask_decoder, dtype) return model def build_sam_fast_vit_l(checkpoint=None, compile_mode='max-autotune', dtype=torch.bfloat16): sam = build_sam_vit_l(checkpoint) sam = _apply_eval_dtype_sam(sam, dtype) sam.image_encoder = torch.compile(sam.image_encoder, mode=compile_mode) return sam
null
179,315
import torch from functools import partial from .modeling import ImageEncoderViT, MaskDecoder, PromptEncoder, Sam, TwoWayTransformer def build_sam_vit_b(checkpoint=None): return _build_sam( encoder_embed_dim=768, encoder_depth=12, encoder_num_heads=12, encoder_global_attn_indexes=[2, 5, 8, 11], checkpoint=checkpoint, ) def _apply_eval_dtype_sam(model, dtype): def prep_model(model, dtype): if dtype is not None: return model.eval().to(dtype) return model.eval() model.image_encoder = prep_model(model.image_encoder, dtype) model.prompt_encoder = prep_model(model.prompt_encoder, dtype) model.mask_decoder = prep_model(model.mask_decoder, dtype) return model def build_sam_fast_vit_b(checkpoint=None, compile_mode='max-autotune', dtype=torch.bfloat16): sam = build_sam_vit_b(checkpoint) sam = _apply_eval_dtype_sam(sam, dtype) sam.image_encoder = torch.compile(sam.image_encoder, mode=compile_mode) return sam
null
179,316
import torch import triton import triton.language as tl import os import pathlib def _attention_rel_h_rel_w_kernel_aligned_meta(q, k, v, rel_h_w, sm_scale): return q.contiguous()
null
179,317
import torch import triton import triton.language as tl import os import pathlib def _autotune(configs, function): import torch.utils.benchmark as benchmark def benchmark_torch_function_in_microseconds(f, *args, **kwargs): try: f(*args, **kwargs) t0 = benchmark.Timer( stmt="f(*args, **kwargs)", globals={"args": args, "kwargs": kwargs, "f": f} ) except: return None return t0.blocked_autorange().mean * 1e6 best = None best_config = None for config in configs: BLOCK_M, BLOCK_N, num_warps, num_stages = config t_config = benchmark_torch_function_in_microseconds( function, BLOCK_M, BLOCK_N, num_warps, num_stages) if t_config is not None: if best is not None: if t_config < best: best = t_config best_config = config else: best = t_config best_config = config print(str(config), " :", str(t_config)) return best, best_config def _attention_rel_h_rel_w_kernel_aligned_device(q, k, v, rel_h_w, sm_scale, o, BLOCK_M, BLOCK_N, num_warps, num_stages): _, Lk, _ = q.shape[-1], k.shape[-1], v.shape[-1] assert q.size() == k.size() assert q.size() == v.size() assert q.size(-2) == rel_h_w.size(-2) assert (q.dtype == torch.bfloat16 or q.dtype == torch.float16) assert k.dtype == q.dtype assert v.dtype == k.dtype assert o.dtype == v.dtype assert rel_h_w.dtype == q.dtype assert rel_h_w.size(-1) == 128 # assert rel_h_w.size(-1) == 2 * BLOCK_N grid = (triton.cdiv(q.shape[2], BLOCK_M), q.shape[0] * q.shape[1], 1) # print("q.shape[0] * q.shape[1]: ", q.shape[0] * q.shape[1]) P_SEQ = 0 if q.shape[-2] == k.shape[-2] else k.shape[-2] - q.shape[-2] assert P_SEQ == 0 assert rel_h_w.is_contiguous(), str(rel_h_w.stride()) OUT_DTYPE = tl.float16 if q.dtype == torch.float16 else tl.bfloat16 _fwd_kernel_aligned[grid]( q, k, v, rel_h_w, sm_scale, o, q.stride(1), q.stride(2), q.stride(3), k.stride(1), k.stride(2), k.stride(3), v.stride(1), v.stride(2), v.stride(3), o.stride(1), o.stride(2), o.stride(3), rel_h_w.stride(1), rel_h_w.stride(2), q.shape[0], q.shape[1], q.shape[2], P_SEQ, OUT_DTYPE=OUT_DTYPE, BIAS_LAST_SIZE=(rel_h_w.size(-1) // 2), B0_NUMEL=rel_h_w.size(-1), BLOCK_M=BLOCK_M, BLOCK_N=BLOCK_N, BLOCK_DMODEL=Lk, num_warps=num_warps, num_stages=num_stages) def _load_best_configs(): device_name = torch.cuda.get_device_name() if not device_name.startswith('NVIDIA A100'): print("Warning: Custom flash attention kernels were written specifically for A100.") import importlib saved_configs = importlib.resources.files("segment_anything_fast") saved_configs = saved_configs / "configs" / "flash_4_configs_a100.p" if not device_name.startswith('NVIDIA A100'): cwd = pathlib.Path.cwd() saved_configs = cwd / "flash_4_configs.p" print(f"We will try to read previously created kernel configurations from {saved_configs}.") print("You can disable this kernel by setting SEGMENT_ANYTHING_FAST_USE_FLASH_4=0") if saved_configs.is_file(): import pickle with open(saved_configs, 'rb') as f: print(f"Loading best configs from file {saved_configs}") return pickle.load(f) def _save_best_configs(best_configs): import importlib saved_configs = importlib.resources.files("segment_anything_fast") saved_configs = saved_configs / "configs" / "flash_4_configs_a100.p" device_name = torch.cuda.get_device_name() if not device_name.startswith('NVIDIA A100'): saved_configs = pathlib.Path.cwd() / "flash_4_configs.p" print("Warning: Custom flash attention kernels were written specifically for A100.") print(f"Storing configs for {device_name} locally under {saved_configs}") with open(saved_configs, 'wb') as f: import pickle print(f"Saving best configs to file {saved_configs}") pickle.dump(best_configs, f) def _create_best_configs_key(q, k, v, rel_h_w, o): key = (q.size(), k.size(), v.size(), rel_h_w.size(), o.size(), q.stride(), k.stride(), v.stride(), rel_h_w.stride(), o.stride()) return key BEST_CONFIGS = None def _attention_rel_h_rel_w_kernel_aligned(q, k, v, rel_h_w, sm_scale): # This is likely not needed, but without it the kernel # is guaranteed to fail. If the inputs are already contiguous # these are cheap checks via is_contiguous and do nothing. q = q.contiguous() k = k.contiguous() v = v.contiguous() # shape constraints Lq, Lk, Lv = q.shape[-1], k.shape[-1], v.shape[-1] assert Lq == Lk and Lk == Lv assert Lk in {16, 32, 64, 128} o = torch.empty_like(q, memory_format=torch.contiguous_format) global BEST_CONFIGS if BEST_CONFIGS is None: BEST_CONFIGS = _load_best_configs() # Loading must have not been successful. Let's create a new dictionary. if BEST_CONFIGS is None: BEST_CONFIGS = {} key = _create_best_configs_key(q, k, v, rel_h_w, o) if key not in BEST_CONFIGS: print("key ", key, " not found. Running autotune. This might take a while.") import functools import itertools configs = [] for (BLOCK_M, BLOCK_N, num_warps) in itertools.product([64, 128], [64, 128], [1, 2, 4, 8]): for num_stages in range(1, num_warps + 1): configs.append((BLOCK_M, BLOCK_N, num_warps, num_stages)) print("all configs len: ", len(configs)) best, best_config = _autotune(configs, functools.partial(_attention_rel_h_rel_w_kernel_aligned_device, q, k, v, rel_h_w, sm_scale, o)) BEST_CONFIGS[key] = best_config print("Found best_config ", best_config, " with time ", best, " for key ", key) _save_best_configs(BEST_CONFIGS) best_config = BEST_CONFIGS[key] if best_config is None: return torch.tensor([]) _attention_rel_h_rel_w_kernel_aligned_device(q, k, v, rel_h_w, sm_scale, o, best_config[0], best_config[1], best_config[2], best_config[3]) return o
null
179,318
import torch import triton import triton.language as tl import os import pathlib USE_CUSTOM_KERNEL = bool(int(os.environ.get('SEGMENT_ANYTHING_FAST_USE_FLASH_4', 1))) The provided code snippet includes necessary dependencies for implementing the `_attention_rel_h_rel_w` function. Write a Python function `def _attention_rel_h_rel_w(q_, k_, v_, rel_h_, rel_w_)` to solve the following problem: Writing this as a composite allows torch.compile to fuse the needed padding into previous operations and memory allocations. Here is the function: def _attention_rel_h_rel_w(q_, k_, v_, rel_h_, rel_w_): """ Writing this as a composite allows torch.compile to fuse the needed padding into previous operations and memory allocations. """ import math sm_scale = 1. / math.sqrt(q_.size(-1)) # Check if second last dimension is multiple of 256 q_size_2_padded = (((q_.size(-2) + 256 - 1) // 256) * 256) - q_.size(-2) def kernel_guards(q_, k_, v_): return (q_.dtype == torch.bfloat16 or q_.dtype == torch.float16) and q_.dtype == k_.dtype and k_.dtype == v_.dtype and USE_CUSTOM_KERNEL # vit_b and vit_l # TODO: This kernel currently does not produce correct results for batch size 1 for this case if q_.size(0) > 1 and q_size_2_padded == 0 and q_.size(-1) == 64 and kernel_guards(q_, k_, v_): rel_h_w = torch.cat([rel_h_.squeeze(-1), rel_w_.squeeze(-2)], dim=-1) o = torch.ops.customflash.custom_flash_aligned( q_, k_, v_, rel_h_w, sm_scale) if o.numel() > 0: return o # vit_h if q_size_2_padded == 0 and q_.size(-1) == 80 and kernel_guards(q_, k_, v_): # Only support multiples of 64, so need to pad q = torch.nn.functional.pad(q_, (0, 128 - 80, 0, 0), "constant", 0) k = torch.nn.functional.pad(k_, (0, 128 - 80, 0, 0), "constant", 0) v = torch.nn.functional.pad(v_, (0, 128 - 80, 0, 0), "constant", 0) rel_h_w = torch.cat([rel_h_.squeeze(-1), rel_w_.squeeze(-2)], dim=-1) o = torch.ops.customflash.custom_flash_aligned( q, k, v, rel_h_w, sm_scale) if o.numel() > 0: return o[:, :, :, :80] attn_bias = (rel_h_ + rel_w_).view(q_.size(0), q_.size(1), rel_h_.size(2), rel_h_.size(3) * rel_w_.size(4)) return torch.nn.functional.scaled_dot_product_attention(q_, k_, v_, attn_mask=attn_bias)
Writing this as a composite allows torch.compile to fuse the needed padding into previous operations and memory allocations.
179,322
import numpy as np import torch import math from copy import deepcopy from itertools import product from typing import Any, Dict, Generator, ItemsView, List, Tuple The provided code snippet includes necessary dependencies for implementing the `mask_to_rle_pytorch_2` function. Write a Python function `def mask_to_rle_pytorch_2(tensor: torch.Tensor) -> List[Dict[str, Any]]` to solve the following problem: Encodes masks to an uncompressed RLE, in the format expected by pycoco tools. Here is the function: def mask_to_rle_pytorch_2(tensor: torch.Tensor) -> List[Dict[str, Any]]: """ Encodes masks to an uncompressed RLE, in the format expected by pycoco tools. """ # Put in fortran order and flatten h,w b, h, w = tensor.shape tensor = tensor.permute(0, 2, 1).flatten(1) # Compute change indices diff = tensor[:, 1:] ^ tensor[:, :-1] a = torch.tensor([[True]]) if diff.is_cuda: a = a.pin_memory().cuda() a = a.expand_as(diff.narrow(1, 0, 1)) diff = torch.cat([a, diff, a], dim=1) change_indices = diff.nonzero() alt_lens = diff.sum(dim=1).tolist() all_cur_idx = change_indices[:, 1] all_btw_idx = torch.cat([all_cur_idx[1:], all_cur_idx[:1]]) - all_cur_idx all_btw_idx = all_btw_idx.detach().cpu().tolist() # Encode run length out = [] counts_init = (tensor[:, 0] == 0).tolist() offset = 0 for i, ci in zip(range(b), counts_init): btw_idxs = all_btw_idx[offset:offset + alt_lens[i]][:-1] offset += alt_lens[i] counts = [] if ci else [0] counts.extend(btw_idxs) out.append({"size": [h, w], "counts": counts}) return out
Encodes masks to an uncompressed RLE, in the format expected by pycoco tools.
179,334
import pandas as pd import fire import matplotlib.pyplot as plt import matplotlib def make_row_chart(batch_size_idx, techniques, df, value_column, ax1, ax2, label, ylim_low, ylim_high, va, title="", relative=False, data_format=None): category_column = "technique" if not isinstance(ylim_low, tuple): ylim_low = (ylim_low, ylim_low) if not isinstance(ylim_high, tuple): ylim_high = (ylim_high, ylim_high) def helper(sam_model_type, ax1, ylim_low, ylim_high, va): vit_b_df = df[df['sam_model_type'] == sam_model_type] vit_b_df = vit_b_df.copy() if relative: vit_b_df[value_column] = vit_b_df[value_column].div( vit_b_df[value_column].iloc[0]) make_sub_chart(batch_size_idx, techniques, vit_b_df, ax1, f"{title} for {sam_model_type}", category_column, value_column, ylim_low, ylim_high, data_format, label, va) helper("vit_b", ax1, ylim_low[0], ylim_high[0], va) helper("vit_h", ax2, ylim_low[1], ylim_high[1], va) def run(csv_file, fig_format): matplotlib.rcParams.update({'font.size': 12}) mdf_ = pd.read_csv(csv_file) mdf = mdf_.dropna(subset=["batch_size"]) techniques = {'fp32': 0, 'bf16': 1, 'compile': 2, 'SDPA': 3, 'Triton': 4, 'NT': 5, 'int8': 6, 'sparse': 7} print("techniques: ", techniques) fig, axs = plt.subplots(3, 2, figsize=(20, 20)) for batch_size_idx, (batch_size, hlim, va) in enumerate(zip([32, 8], [100, 100], ["bottom", "top"])): df = mdf[mdf["batch_size"] == batch_size] make_row_chart(batch_size_idx, techniques, df, "img_s(avg)", *axs[0], f"Batch size {batch_size}", (0.0, 0.0), (100.0, 25.0), va, "Images per second", data_format="{:.2f}") make_row_chart(batch_size_idx, techniques, df, "memory(MiB)", *axs[1], f"Batch size {batch_size}", 0, 40000, va, title="Memory savings", data_format="{:.0f}") make_row_chart(batch_size_idx, techniques, df, "mIoU", *axs[2], f"Batch size {batch_size}", 0.0, 1.0, va, title="Accuracy", data_format="{:.2f}") for ax in axs: ax[0].legend() ax[1].legend() # plt.tick_params(axis='both', which='both', length=10) plt.tight_layout() fig.savefig(f'bar_chart.{fig_format}', format=fig_format)
null
179,335
import tqdm import torch import fire from metrics import calculate_miou, create_result_entry from data import build_data, setup_coco_img_ids import math import segment_anything_fast def pad_to_batch_size(batch, batch_size, device): assert batch.dim() == 4 # assert batch.is_pinned() global PADDED_TENSOR if PADDED_TENSOR is None: batch = batch.to(device=device, non_blocking=True) full_batch_size = (batch_size, batch.size(1), batch.size(2), batch.size(3)) first_entry = batch[0].unsqueeze(0) repeat_first_entry = first_entry.expand(full_batch_size) padded_batch = torch.cat([batch, repeat_first_entry[batch.size(0):batch_size]], dim=0) assert padded_batch.size() == full_batch_size PADDED_TENSOR = padded_batch PADDED_TENSOR[:batch.size(0)].copy_(batch, non_blocking=True) return PADDED_TENSOR def get_features_batch(encoder, input_image_batch, pad_input_image_batch, batch_size, device): if pad_input_image_batch: features_batch = encoder(pad_to_batch_size(input_image_batch, batch_size, device)) return features_batch[:input_image_batch.size(0)] return encoder(input_image_batch)
null
179,336
import tqdm import torch import fire from metrics import calculate_miou, create_result_entry from data import build_data, setup_coco_img_ids import math import segment_anything_fast torch._dynamo.config.cache_size_limit = 50000 def build_results(batched_data_iter, predictor, mask_debug_out_dir, batch_size, use_compile, use_compile_decoder, use_nested_tensor, pad_input_image_batch, use_fullgraph=False): # TODO: Re-enable this for datapoints assert not use_compile_decoder batch_runner = None if use_nested_tensor: batch_runner = build_results_batch_nested else: batch_runner = build_results_batch results = [] batch_idx = 0 num_images = 0 num_batches = 0 elapsed_time = 0 partial_batch = False for batch in tqdm.tqdm(batched_data_iter): with torch.no_grad(): if batch_idx == 0: with torch.autograd.profiler.record_function("compilation and warmup"): if str(use_compile) != "False": predictor.model.image_encoder = torch.compile(predictor.model.image_encoder, mode=use_compile, fullgraph=use_fullgraph) # Run first batch a few times for warmup and exclude it from the final timings for _ in range(3): _ = batch_runner(predictor, batch, batch_size, pad_input_image_batch) result_batch, num_datapoints, kernel_time = batch_runner(predictor, batch, batch_size, pad_input_image_batch) if result_batch is not None: results += result_batch # We expect a partial batch to only happens once at the end assert not partial_batch # Only measure timing on full batches if num_datapoints == batch_size: num_images += num_datapoints num_batches += 1 # We consistently exclude the last (512 - filtered) images # Since batch sizes must be powers of two and less than # or equal 512 this ensures consistent timing across varying # batch sizes. if num_images <= 4488: elapsed_time += kernel_time else: partial_batch = True batch_idx += 1 avg_ms_per_img = None if num_images > 0: avg_ms_per_img = elapsed_time avg_ms_per_img = avg_ms_per_img / num_images return results, avg_ms_per_img, num_batches, num_images def identity_runner(fn, *args, **kwargs): return fn(*args, **kwargs) def profiler_runner(path, fn, *args, **kwargs): with torch.profiler.profile( activities=[torch.profiler.ProfilerActivity.CPU, torch.profiler.ProfilerActivity.CUDA], record_shapes=True) as prof: result = fn(*args, **kwargs) prof.export_chrome_trace(path) return result def profile_top_runner(fn, *args, **kwargs): with torch.profiler.profile( activities=[torch.profiler.ProfilerActivity.CPU, torch.profiler.ProfilerActivity.CUDA], record_shapes=True) as prof: result = fn(*args, **kwargs) print(prof.key_averages().table(sort_by="self_cuda_time_total", row_limit=-1)) return result def memory_runner(path, fn, *args, **kwargs): print("Start memory recording") torch.cuda.synchronize() torch.cuda.memory._record_memory_history(True, trace_alloc_max_entries=100000, trace_alloc_record_context=True) result = fn(*args, **kwargs) torch.cuda.synchronize() snapshot = torch.cuda.memory._snapshot() print("Finish memory recording") import pickle with open(path, 'wb') as f: pickle.dump(snapshot, f) # Use to convert pickle file into html # python torch/cuda/_memory_viz.py trace_plot <snapshot>.pickle -o <snapshot>.html return result def calculate_miou(results, mask_debug_out_dir, silent, cat_id_to_cat): df = pd.DataFrame(results, columns=['img_id', 'ann_id', 'cat_id', 'iou']) df.to_csv(f'{mask_debug_out_dir}/df.csv') df['supercategory'] = df['cat_id'].map( lambda cat_id: cat_id_to_cat[cat_id]['supercategory']) df['category'] = df['cat_id'].map( lambda cat_id: cat_id_to_cat[cat_id]['name']) # TODO: cross reference the specifics of how we calculate mIoU with # the SAM folks (should it be per dataset, per category, per image, etc) # currently, just calculate them all # TODO: QOL save the summaries to file # per category per_category = pd.pivot_table( df, values='iou', index=['cat_id', 'supercategory', 'category'], aggfunc=('mean', 'count')) if not silent: print('\nmIoU averaged per category') print(per_category) # per super-category per_supercategory = pd.pivot_table( df, values='iou', index=['supercategory'], aggfunc=('mean', 'count')) if not silent: print('\nmIoU averaged per supercategory') print(per_supercategory) # per all selected masks per_all_masks_agg = df['iou'].agg(['mean', 'count']) if not silent: print('\nmIoU averaged per all selected masks') print(per_all_masks_agg) return df['iou'].agg(['mean', 'count'])['mean'] def build_data(coco_img_ids, coco, catIds, coco_root_dir, coco_slice_name, point_sampling_cache_dir, predictor, use_half, use_nested_tensor, pad_input_image_batch): cache = diskcache.Cache(point_sampling_cache_dir) # make sure you clear the cache if you change the point sampling algorithm # cache.clear() pixel_mean = predictor.model.pixel_mean.cpu() pixel_std = predictor.model.pixel_std.cpu() def build_batch(indicies): batch = [[], [], [], [], [], [], [], [], [], [], []] batch[3] = [0] batch[6] = [0] for img_idx in indicies: imgId = coco_img_ids[img_idx] datapoint = build_datapoint(imgId, coco, pixel_mean, pixel_std, coco_root_dir, coco_slice_name, catIds, cache, predictor, pad_input_image_batch) I, coords_list, gt_masks_list, anns, x, predictor_input_size = datapoint if len(coords_list) == 0: continue batch[0].append(x) # batch[0].append(x[0]) coords_list = predictor.transform.apply_coords( np.array(coords_list), I.shape[:2]) coords_list = torch.tensor(coords_list, dtype=torch.float) batch[1].append(coords_list.reshape(-1)) batch[2].append(coords_list.size()) batch[3].append(coords_list.numel() + batch[3][-1]) batch[4].append(gt_masks_list.reshape(-1)) batch[5].append(gt_masks_list.size()) batch[6].append(gt_masks_list.numel() + batch[6][-1]) batch[7].append(anns) batch[8].append(I) batch[9].append(predictor_input_size) batch[10].append(img_idx) def cat_and_cast(b, use_half): b = torch.cat(b) if len(b) > 0 else None if use_half is not None and b is not None: return b.to(use_half) return b def to_nested_tensor(data, sizes=None, use_half=None): if len(data) == 0: return None dtype = use_half if use_half is not None else torch.float32 if sizes is not None: data = [d.view(s) for (d, s) in zip(data, sizes)] return torch.nested.nested_tensor(data, dtype=dtype, layout=torch.jagged) if pad_input_image_batch: batch[0] = cat_and_cast(batch[0], use_half) else: batch[0] = to_nested_tensor(batch[0], use_half=use_half) if use_nested_tensor: batch[1] = to_nested_tensor(batch[1], batch[2], use_half) batch[2] = None batch[3] = None else: batch[1] = cat_and_cast(batch[1], use_half) batch[4] = cat_and_cast(batch[4], False) return batch return build_batch def setup_coco_img_ids(coco_root_dir, coco_slice_name, coco_category_names, img_id): annFile = '{}/annotations/instances_{}.json'.format( coco_root_dir, coco_slice_name) # initialize COCO api for instance annotations coco = COCO(annFile) # display COCO categories and supercategories cats = coco.loadCats(coco.getCatIds()) cat_id_to_cat = {cat['id']: cat for cat in cats} nms = [cat['name'] for cat in cats] # print('COCO categories: \n{}\n'.format(' '.join(nms))) # nms = set([cat['supercategory'] for cat in cats]) # print('COCO supercategories: \n{}'.format(' '.join(nms))) if coco_category_names is not None: catIds = coco.getCatIds(catNms=coco_category_names) else: catIds = coco.getCatIds() if img_id is not None: coco_img_ids = [img_id] elif coco_category_names is None: coco_img_ids = coco.getImgIds() else: coco_img_ids = coco.getImgIds(catIds=catIds) return coco_img_ids, cat_id_to_cat, catIds, coco def apply_sparse(model): apply_fake_sparsity(model) for name, mod in model.named_modules(): if isinstance(mod, torch.nn.Linear): mod.weight = torch.nn.Parameter(to_sparse_semi_structured(mod.weight)) def run( coco_root_dir, coco_slice_name, sam_checkpoint_base_path, sam_model_type, point_sampling_cache_dir, mask_debug_out_dir, batch_size=1, print_header=False, coco_category_names=None, limit=None, img_id=None, use_half=None, use_compile="False", use_compile_decoder=False, compress=None, epilogue_fusion_first=False, num_workers=0, use_nested_tensor=False, use_rel_pos=True, pad_input_image_batch=True, profile_path=None, profile_top=False, memory_path=None, use_local_sam_fork=False, use_compiler_settings=False, ): from torch._inductor import config as inductorconfig inductorconfig.triton.unique_kernel_names = True inductorconfig.epilogue_fusion_first = epilogue_fusion_first if use_compiler_settings: # inductorconfig.fx_graph_cache = True # seems to slow performance inductorconfig.epilogue_fusion = False inductorconfig.coordinate_descent_tuning = True inductorconfig.coordinate_descent_check_all_directions = True if use_half is not None: if use_half == "float16": use_half = torch.float16 elif use_half == "bfloat16": use_half = torch.bfloat16 else: raise ValueError("Expected one of float16 or bfloat for specified {use_half}") # Batch size needs to be a multiple of two and at most 512. assert math.log2(batch_size).is_integer() assert batch_size <= 512 # https://github.com/facebookresearch/segment-anything/tree/main#model-checkpoints # largest to smallest: vit_h, vit_l, vit_b model_type_to_checkpoint = { 'vit_h': f'{sam_checkpoint_base_path}/sam_vit_h_4b8939.pth', 'vit_l': f'{sam_checkpoint_base_path}/sam_vit_l_0b3195.pth', 'vit_b': f'{sam_checkpoint_base_path}/sam_vit_b_01ec64.pth', } if use_local_sam_fork: from segment_anything_fast import sam_model_registry, SamPredictor else: from segment_anything import sam_model_registry, SamPredictor checkpoint_path = model_type_to_checkpoint[sam_model_type] sam = sam_model_registry[sam_model_type](checkpoint=checkpoint_path).cuda() predictor = SamPredictor(sam) from segment_anything_fast import tools tools.apply_eval_dtype_predictor(predictor, use_half) for block in predictor.model.image_encoder.blocks: block.attn.use_rel_pos = use_rel_pos if compress == "dynamic_quant": from torchao.quantization import apply_dynamic_quant apply_dynamic_quant(predictor.model.image_encoder) inductorconfig.force_fuse_int_mm_with_mul = True elif compress == "static_quant": from segment_anything_fast.static_quant import apply_static_quant apply_static_quant(predictor.model.image_encoder) from pathlib import Path weights_path = Path(f"static_quant_scalars/{sam_model_type}_{batch_size}_static_quant_weights.ptk") if weights_path.exists() and weights_path.is_file(): print("Loading static quantization weights") weights = torch.load(f"static_quant_scalars/{sam_model_type}_{batch_size}_static_quant_weights.ptk") from static_quant import set_x_absmax set_x_absmax(predictor.model.image_encoder, weights) elif compress == "sparse": from segment_anything_fast.sparse import apply_sparse apply_sparse(predictor.model.image_encoder) elif compress == "int4_dynamic_quant_sparse": from segment_anything_fast.dynamic_quant_sparse import apply_int4_dynamic_quant_sparse apply_int4_dynamic_quant_sparse(predictor.model.image_encoder) elif compress == "static_quant_sparse": raise NotImplementedError(f"Unsupported compress {compress}") elif compress == "sparse": raise NotImplementedError(f"Unsupported compress {compress}") else: assert compress is None, f"Unsupported compress mode {compress}" coco_img_ids_, cat_id_to_cat, catIds, coco = setup_coco_img_ids( coco_root_dir, coco_slice_name, coco_category_names, img_id) coco_img_ids = [] for imgId in coco_img_ids_: img = coco.loadImgs(imgId)[0] annIds = coco.getAnnIds(imgIds=img['id'], catIds=catIds, iscrowd=None) anns = coco.loadAnns(annIds) if len(anns) != 0: coco_img_ids.append(imgId) build_batch = build_data(coco_img_ids, coco, catIds, coco_root_dir, coco_slice_name, point_sampling_cache_dir, predictor, use_half, use_nested_tensor, pad_input_image_batch) limit = len(coco_img_ids) if limit is None else limit batched_data_iter = torch.utils.data.DataLoader(list(range(limit)), batch_size=batch_size, collate_fn=build_batch, num_workers=num_workers, pin_memory=False) runner = identity_runner if profile_path is not None: import functools runner = functools.partial(profiler_runner, profile_path) if profile_top: runner = profile_top_runner if memory_path is not None: assert use_compile != "max-autotune", f"Memory path does not support {use_compile}" import functools runner = functools.partial(memory_runner, memory_path) results, avg_ms_per_img, num_batches, num_images = runner(build_results, batched_data_iter, predictor, mask_debug_out_dir, batch_size, use_compile, use_compile_decoder, use_nested_tensor, pad_input_image_batch) if compress == "static_quant": from static_quant import get_x_absmax weights = get_x_absmax(predictor.model.image_encoder) print("Saving static quantization weights") torch.save(weights, f"static_quant_scalars/{sam_model_type}_{batch_size}_static_quant_weights.ptk") results = [[r[0], r[1], r[2], r[3].item()] for r in results] img_s, batch_ms_batch_size = None, None if avg_ms_per_img is not None: img_s = 1000 / avg_ms_per_img batch_ms_batch_size = (avg_ms_per_img * num_images) / num_batches / batch_size mIoU = calculate_miou(results, mask_debug_out_dir, True, cat_id_to_cat) max_memory_allocated_bytes = torch.cuda.max_memory_allocated() _, total_memory = torch.cuda.mem_get_info() max_memory_allocated_percentage = int(100 * (max_memory_allocated_bytes / total_memory)) max_memory_allocated_bytes = max_memory_allocated_bytes >> 20 if print_header: print(",".join(["sam_model_type", "batch_size", "memory(MiB)", "memory(%)", "img_s(avg)", "batch_ms(avg)/batch_size", "mIoU", "use_compile", "use_half", "compress", "epilogue_fusion_first", "use_compile_decoder", "use_nested_tensor", "use_rel_pos", "pad_input_image_batch", "num_workers", "num_batches", "num_images", "profile_path", "memory_path"])) print(",".join(map(str, [sam_model_type, batch_size, max_memory_allocated_bytes, max_memory_allocated_percentage, img_s, batch_ms_batch_size, mIoU, use_compile, use_half, compress, epilogue_fusion_first, use_compile_decoder, use_nested_tensor, use_rel_pos, pad_input_image_batch, num_workers, num_batches, num_images, profile_path, memory_path])))
null
179,337
import numpy as np import torch import matplotlib.pyplot as plt import cv2 import torch.utils.benchmark as benchmark from segment_anything_fast import sam_model_registry, sam_model_fast_registry, SamAutomaticMaskGenerator torch.cuda.synchronize() torch.cuda.synchronize() print(start_event.elapsed_time(end_event) / 10.) print(f"memory(MiB): {max_memory_allocated_bytes} memory(%): {max_memory_allocated_percentage}") def profiler_runner(path, fn, *args, **kwargs): with torch.profiler.profile( activities=[torch.profiler.ProfilerActivity.CPU, torch.profiler.ProfilerActivity.CUDA], record_shapes=True) as prof: result = fn(*args, **kwargs) print(f"Saving trace under {path}") prof.export_chrome_trace(path) return result
null
179,338
import numpy as np import torch import matplotlib.pyplot as plt import cv2 import torch.utils.benchmark as benchmark from segment_anything_fast import sam_model_registry, sam_model_fast_registry, SamAutomaticMaskGenerator plt.figure(figsize=(image.shape[1]/100., image.shape[0]/100.), dpi=100) plt.imshow(image) plt.axis('off') plt.tight_layout() plt.savefig('dog_mask_fast.png', format='png') def show_anns(anns): if len(anns) == 0: return sorted_anns = sorted(anns, key=(lambda x: x['area']), reverse=True) ax = plt.gca() ax.set_autoscale_on(False) img = np.ones((sorted_anns[0]['segmentation'].shape[0], sorted_anns[0]['segmentation'].shape[1], 4)) img[:,:,3] = 0 for ann in sorted_anns: m = ann['segmentation'] color_mask = np.concatenate([np.random.random(3), [0.35]]) img[m] = color_mask ax.imshow(img)
null
179,339
import torch from torch import nn from typing import Optional, Tuple, Union import transformers from transformers.models.llama.modeling_llama import apply_rotary_pos_emb, rotate_half import math STORE_KV_BEFORE_ROPE = False USE_MEM_EFF_ATTENTION = False def xformers_forward( self, hidden_states: torch.Tensor, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, past_key_value: Optional[Tuple[torch.Tensor]] = None, output_attentions: bool = False, use_cache: bool = False, padding_mask=None, ) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]: bsz, q_len, _ = hidden_states.size() query_states = self.q_proj(hidden_states).view(bsz, q_len, self.num_heads, self.head_dim).transpose(1, 2) key_states = self.k_proj(hidden_states).view(bsz, q_len, self.num_heads, self.head_dim).transpose(1, 2) value_states = self.v_proj(hidden_states).view(bsz, q_len, self.num_heads, self.head_dim).transpose(1, 2) kv_seq_len = key_states.shape[-2] past_kv_len = 0 if past_key_value is not None: past_kv_len = past_key_value[0].shape[-2] kv_seq_len += past_kv_len if STORE_KV_BEFORE_ROPE is False: cos, sin = self.rotary_emb(value_states, seq_len=kv_seq_len) query_states, key_states = apply_rotary_pos_emb(query_states, key_states, cos, sin, position_ids) # [bsz, nh, t, hd] if past_key_value is not None: # reuse k, v, self_attention key_states = torch.cat([past_key_value[0], key_states], dim=2) value_states = torch.cat([past_key_value[1], value_states], dim=2) past_key_value = (key_states, value_states) if use_cache else None else: if past_key_value is not None: # reuse k, v, self_attention key_states = torch.cat([past_key_value[0], key_states], dim=2) value_states = torch.cat([past_key_value[1], value_states], dim=2) past_key_value = (key_states, value_states) if use_cache else None cos, sin = self.rotary_emb(value_states, seq_len=kv_seq_len) query_states = apply_rotary_pos_emb_single(query_states, cos, sin, position_ids) position_ids = torch.arange(kv_seq_len, dtype=torch.long, device=cos.device) position_ids = position_ids.unsqueeze(0).view(-1, kv_seq_len) key_states = apply_rotary_pos_emb_single(key_states, cos, sin, position_ids) pad_query = False if xops is not None and USE_MEM_EFF_ATTENTION: attn_weights = None query_states = query_states.transpose(1, 2) key_states = key_states.transpose(1, 2) value_states = value_states.transpose(1, 2) if query_states.size(1)==1 and key_states.size(1)>1: attn_bias = None elif query_states.size(1)<key_states.size(1) and key_states.size(1)>1 and past_kv_len > 0: attn_bias = xops.LowerTriangularMask() query_states = torch.cat( ( torch.full( (bsz, past_kv_len, self.num_heads, self.head_dim), 0.0, dtype=query_states.dtype, device=query_states.device, ), query_states, ), dim=1, ) pad_query = True else: attn_bias = xops.LowerTriangularMask() attn_output = xops.memory_efficient_attention( query_states, key_states, value_states, attn_bias=attn_bias, p=0) else: attn_weights = torch.matmul(query_states, key_states.transpose(2, 3)) / math.sqrt(self.head_dim) if attn_weights.size() != (bsz, self.num_heads, q_len, kv_seq_len): raise ValueError( f"Attention weights should be of size {(bsz * self.num_heads, q_len, kv_seq_len)}, but is" f" {attn_weights.size()}" ) if attention_mask is not None: if attention_mask.size() != (bsz, 1, q_len, kv_seq_len): raise ValueError( f"Attention mask should be of size {(bsz, 1, q_len, kv_seq_len)}, but is {attention_mask.size()}" ) attn_weights = attn_weights + attention_mask attn_weights = torch.max( attn_weights, torch.tensor(torch.finfo(attn_weights.dtype).min, device=attn_weights.device) ) # upcast attention to fp32 attn_weights = nn.functional.softmax(attn_weights, dim=-1, dtype=torch.float32).to(query_states.dtype) attn_output = torch.matmul(attn_weights, value_states) if attn_output.size() != (bsz, self.num_heads, q_len, self.head_dim): raise ValueError( f"`attn_output` should be of size {(bsz, self.num_heads, q_len, self.head_dim)}, but is" f" {attn_output.size()}" ) attn_output = attn_output.transpose(1, 2) if pad_query: attn_output = attn_output[:,past_kv_len:] attn_output = attn_output.reshape(bsz, q_len, self.hidden_size) attn_output = self.o_proj(attn_output) if not output_attentions: attn_weights = None return attn_output, attn_weights, past_key_value def apply_attention_patch( use_memory_efficient_attention=False, store_kv_before_rope=False ): global USE_MEM_EFF_ATTENTION, STORE_KV_BEFORE_ROPE if use_memory_efficient_attention is True and xops is not None: USE_MEM_EFF_ATTENTION = use_memory_efficient_attention print("USE_XFORMERS_ATTENTION: ", USE_MEM_EFF_ATTENTION) STORE_KV_BEFORE_ROPE = store_kv_before_rope print("STORE_KV_BEFORE_ROPE:", STORE_KV_BEFORE_ROPE) transformers.models.llama.modeling_llama.LlamaAttention.forward = xformers_forward
null
179,340
import torch from torch import nn from typing import Optional, Tuple, Union import transformers from transformers.models.llama.modeling_llama import apply_rotary_pos_emb, rotate_half import math ALPHA = 1.0 SCALING_FACTOR = None def adaptive_ntk_init(self, dim, max_position_embeddings=2048, base=10000, device=None, scaling_factor=None): self.alpha = ALPHA if SCALING_FACTOR is None: self.scaling_factor = scaling_factor or 1.0 else: self.scaling_factor = SCALING_FACTOR if isinstance(ALPHA,(float,int)): base = base * ALPHA ** (dim / (dim-2)) self.base = base elif ALPHA=='auto': self.base = base else: raise ValueError(ALPHA) old_init(self, dim, max_position_embeddings, base, device) self.ntk_inv_freq = 1.0 / (base ** (torch.arange(0, dim, 2).float().to(device) / dim)) self._set_cos_sin_cache = _set_cos_sin_cache self._set_cos_sin_cache( self, seq_len=max_position_embeddings, device=self.ntk_inv_freq.device, dtype=torch.get_default_dtype() ) def adaptive_ntk_forward(self, x, seq_len=None): if seq_len > self.max_seq_len_cached: if isinstance(self.alpha,(float,int)): self._set_cos_sin_cache(self, seq_len=seq_len, device=x.device, dtype=x.dtype) elif self.alpha=='auto': t = torch.arange(seq_len, device=x.device, dtype=torch.float32) t = t / self.scaling_factor dim = self.dim alpha = (seq_len / (self.max_position_embeddings/2) - 1) * AUTO_COEFF base = self.base * alpha ** (dim / (dim-2)) ntk_inv_freq = 1.0 / (base ** (torch.arange(0, dim, 2).float().to(x.device) / dim )) freqs = torch.einsum("i,j->ij", t, ntk_inv_freq) emb = torch.cat((freqs, freqs), dim=-1).to(x.device) cos_cached = emb.cos() sin_cached = emb.sin() return ( cos_cached[:seq_len].to(dtype=x.dtype), sin_cached[:seq_len].to(dtype=x.dtype) ) return ( self.cos_cached[:seq_len].to(dtype=x.dtype), self.sin_cached[:seq_len].to(dtype=x.dtype) ) def apply_ntk_scaling_patch(alpha: Union[float,str], scaling_factor: Optional[float] = None): global ALPHA global SCALING_FACTOR ALPHA = alpha SCALING_FACTOR = scaling_factor try: ALPHA = float(ALPHA) except ValueError: if ALPHA!="auto": raise ValueError(f"Alpha can only be a float or 'auto', but given {ALPHA}") print(f"Apply NTK scaling with ALPHA={ALPHA}") if scaling_factor is None: print(f"The value of scaling factor will be read from model config file, or set to 1.") else: print(f"Warning: scaling factor is set to {SCALING_FACTOR}. \ If you set the value by hand, do not forget to update \ max_position_embeddings in the model config file.") transformers.models.llama.modeling_llama.LlamaRotaryEmbedding.__init__ = adaptive_ntk_init if hasattr(transformers.models.llama.modeling_llama,'LlamaLinearScalingRotaryEmbedding'): transformers.models.llama.modeling_llama.LlamaLinearScalingRotaryEmbedding.__init__ = adaptive_ntk_init transformers.models.llama.modeling_llama.LlamaRotaryEmbedding.forward = adaptive_ntk_forward
null
179,341
from typing import Optional, Tuple import torch import transformers from einops import rearrange def forward( self, hidden_states: torch.Tensor, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.Tensor] = None, past_key_value: Optional[Tuple[torch.Tensor]] = None, output_attentions: bool = False, use_cache: bool = False, padding_mask=None, ) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]: """Input shape: Batch x Time x Channel attention_mask: [bsz, q_len] """ bsz, q_len, _ = hidden_states.size() query_states = ( self.q_proj(hidden_states) .view(bsz, q_len, self.num_heads, self.head_dim) ) key_states = ( self.k_proj(hidden_states) .view(bsz, q_len, self.num_heads, self.head_dim) ) value_states = ( self.v_proj(hidden_states) .view(bsz, q_len, self.num_heads, self.head_dim) ) kv_seq_len = key_states.shape[1] past_kv_len = 0 if past_key_value is not None: past_kv_len = past_key_value[0].shape[-2] kv_seq_len += past_kv_len cos, sin = self.rotary_emb(value_states, seq_len=kv_seq_len) rotary_dim = cos.shape[-1] cos, sin = cos.squeeze(0,1)[:,:rotary_dim//2].contiguous(), sin.squeeze(0,1)[:,:rotary_dim//2].contiguous() if past_key_value is not None: key_cache = torch.cat([past_key_value[0].transpose(1, 2), key_states], dim=1) value_cache = torch.cat([past_key_value[1].transpose(1, 2), value_states], dim=1) else: key_cache = key_states value_cache = value_states assert not output_attentions, "output_attentions is not supported" q = query_states # [bsz, q_len, nh, hd] k, v = key_states, value_states # [bsz, q_len, nh, hd] output = flash_attn_with_kvcache( q, key_cache, value_cache, k, v, rotary_cos=cos, rotary_sin=sin, cache_seqlens=past_kv_len, softmax_scale=None, causal=True, rotary_interleaved=False ) output = rearrange(output, "b s h d -> b s (h d)", b=bsz) past_key_value = (key_cache[:,:kv_seq_len].transpose(1,2), value_cache[:,:kv_seq_len].transpose(1,2)) if use_cache else None output = self.o_proj(output) return output, None, past_key_value def _prepare_decoder_attention_mask( self, attention_mask, input_shape, inputs_embeds, past_key_values_length ): return attention_mask def replace_llama_attn_with_flash_attn(): if flash_attn_with_kvcache != None: print("USE_FLASH_ATTENTION: ", True) transformers.models.llama.modeling_llama.LlamaModel._prepare_decoder_attention_mask = _prepare_decoder_attention_mask transformers.models.llama.modeling_llama.LlamaAttention.forward = forward else: print("USE_FLASH_ATTENTION: ", False)
null
179,342
import argparse import json, os DEFAULT_SYSTEM_PROMPT = """You are a helpful assistant. 你是一个乐于助人的助手。""" TEMPLATE = ( "[INST] <<SYS>>\n" "{system_prompt}\n" "<</SYS>>\n\n" "{instruction} [/INST]" ) import torch from transformers import AutoModelForCausalLM, LlamaForCausalLM, LlamaTokenizer from transformers import GenerationConfig from transformers import BitsAndBytesConfig from peft import PeftModel import sys def generate_prompt(instruction, system_prompt=DEFAULT_SYSTEM_PROMPT): return TEMPLATE.format_map({'instruction': instruction,'system_prompt': system_prompt})
null
179,343
import torch from transformers import ( AutoModelForCausalLM, LlamaForCausalLM, LlamaTokenizer, StoppingCriteria, BitsAndBytesConfig, GenerationConfig ) import gradio as gr import argparse import os from queue import Queue from threading import Thread import traceback import gc import json import requests from typing import Iterable, List import subprocess import re args = parser.parse_args() if args.use_vllm is True: print("CFG sampling is disabled when using vLLM.") ENABLE_CFG_SAMPLING = False if args.only_cpu is True: args.gpus = "" if args.load_in_8bit or args.load_in_4bit: raise ValueError("Quantization is unavailable on CPU.") if args.load_in_8bit and args.load_in_4bit: raise ValueError("Only one quantization method can be chosen for inference. Please check your arguments") import sys if not args.only_cpu: if args.use_flash_attention_2: from flash_attn_patch_for_inference import replace_llama_attn_with_flash_attn replace_llama_attn_with_flash_attn() else: from attn_and_long_ctx_patches import apply_attention_patch apply_attention_patch(use_memory_efficient_attention=True) from attn_and_long_ctx_patches import apply_ntk_scaling_patch if args.use_ntk: apply_ntk_scaling_patch(args.alpha) if args.speculative_sampling: if args.draft_base_model == None: raise ValueError("Speculative sampling requires a draft model. Please specify the draft model.") if args.draft_model_load_in_8bit and args.draft_model_load_in_4bit: raise ValueError("Only one quantization method can be chosen for inference. Please check your arguments") from speculative_sample import speculative_sample from peft import PeftModel def setup(): global tokenizer, model, device, share, port, max_memory if args.speculative_sampling: global draft_model if args.use_vllm: # global share, port, max_memory max_memory = args.max_memory port = args.port share = args.share == 'True' or args.share is True if args.lora_model is not None: raise ValueError("vLLM currently does not support LoRA, please merge the LoRA weights to the base model.") if args.load_in_8bit or args.load_in_4bit: raise ValueError("vLLM currently does not support quantization, please use fp16 (default) or unuse --use_vllm.") if args.only_cpu: raise ValueError("vLLM requires GPUs with compute capability not less than 7.0. If you want to run only on CPU, please unuse --use_vllm.") if args.speculative_sampling: raise ValueError("speculative_sampling is set, but vLLM does not support speculative sampling. Please unset speculative_sampling. ") if args.tokenizer_path is None: args.tokenizer_path = args.base_model tokenizer = LlamaTokenizer.from_pretrained(args.tokenizer_path, legacy=True) print("Start launch vllm server.") cmd = f"python -m vllm.entrypoints.api_server \ --model={args.base_model} \ --tokenizer={args.tokenizer_path} \ --tokenizer-mode=slow \ --tensor-parallel-size={len(args.gpus.split(','))} \ --host {args.post_host} \ --port {args.post_port} \ &" subprocess.check_call(cmd, shell=True) else: max_memory = args.max_memory port = args.port share = args.share == 'True' or args.share is True load_type = torch.float16 if torch.cuda.is_available(): device = torch.device(0) else: device = torch.device('cpu') if args.tokenizer_path is None: args.tokenizer_path = args.lora_model if args.lora_model is None: args.tokenizer_path = args.base_model tokenizer = LlamaTokenizer.from_pretrained(args.tokenizer_path, legacy=True) if args.load_in_4bit or args.load_in_8bit: quantization_config = BitsAndBytesConfig( load_in_4bit=args.load_in_4bit, load_in_8bit=args.load_in_8bit, bnb_4bit_compute_dtype=load_type, ) base_model = AutoModelForCausalLM.from_pretrained( args.base_model, torch_dtype=load_type, low_cpu_mem_usage=True, device_map='auto', load_in_4bit=args.load_in_4bit, load_in_8bit=args.load_in_8bit, quantization_config=quantization_config if (args.load_in_4bit or args.load_in_8bit) else None, trust_remote_code=True ) if args.speculative_sampling: if args.load_in_4bit or args.load_in_8bit: draft_quantization_config = BitsAndBytesConfig( load_in_4bit=args.draft_model_load_in_4bit, load_in_8bit=args.draft_model_load_in_8bit, bnb_4bit_compute_dtype=load_type, ) draft_base_model = LlamaForCausalLM.from_pretrained( args.draft_base_model, torch_dtype=load_type, low_cpu_mem_usage=True, device_map='auto', load_in_4bit=args.draft_model_load_in_4bit, load_in_8bit=args.draft_model_load_in_8bit, quantization_config=draft_quantization_config if (args.draft_model_load_in_4bit or args.draft_model_load_in_8bit) else None ) model_vocab_size = base_model.get_input_embeddings().weight.size(0) tokenizer_vocab_size = len(tokenizer) print(f"Vocab of the base model: {model_vocab_size}") print(f"Vocab of the tokenizer: {tokenizer_vocab_size}") if model_vocab_size != tokenizer_vocab_size: print("Resize model embeddings to fit tokenizer") base_model.resize_token_embeddings(tokenizer_vocab_size) if args.speculative_sampling: draft_model_vocab_size = draft_base_model.get_input_embeddings().weight.size(0) print(f"Vocab of the draft base model: {draft_model_vocab_size}") if draft_model_vocab_size!=tokenizer_vocab_size: print("Resize draft model embeddings to fit tokenizer") draft_base_model.resize_token_embeddings(tokenizer_vocab_size) if args.lora_model is not None: print("loading peft model") model = PeftModel.from_pretrained( base_model, args.lora_model, torch_dtype=load_type, device_map='auto', ).half() else: model = base_model if args.speculative_sampling: if args.draft_lora_model is not None: print("loading peft draft model") draft_model = PeftModel.from_pretrained(draft_base_model, args.draft_lora_model,torch_dtype=load_type,device_map='auto',).half() else: draft_model = draft_base_model if device == torch.device('cpu'): model.float() model.eval() if args.speculative_sampling: if device==torch.device('cpu'): draft_model.float() draft_model.eval()
null
179,344
import torch from transformers import ( AutoModelForCausalLM, LlamaForCausalLM, LlamaTokenizer, StoppingCriteria, BitsAndBytesConfig, GenerationConfig ) import gradio as gr import argparse import os from queue import Queue from threading import Thread import traceback import gc import json import requests from typing import Iterable, List import subprocess import re import sys from attn_and_long_ctx_patches import apply_ntk_scaling_patch from peft import PeftModel with gr.Blocks() as demo: github_banner_path = 'https://raw.githubusercontent.com/ymcui/Chinese-LLaMA-Alpaca-2/main/pics/banner.png' gr.HTML(f'<p align="center"><a href="https://github.com/ymcui/Chinese-LLaMA-Alpaca-2"><img src={github_banner_path} width="700"/></a></p>') chatbot = gr.Chatbot() with gr.Row(): with gr.Column(scale=4): with gr.Column(scale=3): system_prompt_input = gr.Textbox( show_label=True, label="系统提示语(仅在对话开始前或清空历史后修改有效,对话过程中修改无效)", placeholder=DEFAULT_SYSTEM_PROMPT, lines=1).style( container=True) negative_prompt_input = gr.Textbox( show_label=True, label="反向提示语(仅在对话开始前或清空历史后修改有效,对话过程中修改无效)", placeholder="(可选,默认为空)", lines=1, visible=ENABLE_CFG_SAMPLING).style( container=True) with gr.Column(scale=12): user_input = gr.Textbox( show_label=True, label="用户指令", placeholder="Shift + Enter发送消息...", lines=10).style( container=True) with gr.Column(min_width=32, scale=1): submitBtn = gr.Button("Submit", variant="primary") with gr.Column(scale=1): emptyBtn = gr.Button("Clear History") max_new_token = gr.Slider( 0, 4096, value=512, step=1.0, label="Maximum New Token Length", interactive=True) top_p = gr.Slider(0, 1, value=0.9, step=0.01, label="Top P", interactive=True) temperature = gr.Slider( 0, 1, value=0.2, step=0.01, label="Temperature", interactive=True) top_k = gr.Slider(1, 40, value=40, step=1, label="Top K", interactive=True) do_sample = gr.Checkbox( value=True, label="Do Sample", info="use random sample strategy", interactive=True) repetition_penalty = gr.Slider( 1.0, 3.0, value=1.1, step=0.1, label="Repetition Penalty", interactive=True, visible=False if args.use_vllm else True) guidance_scale = gr.Slider( 1.0, 3.0, value=1.0, step=0.1, label="Guidance Scale", interactive=True, visible=ENABLE_CFG_SAMPLING) presence_penalty = gr.Slider( -2.0, 2.0, value=1.0, step=0.1, label="Presence Penalty", interactive=True, visible=True if args.use_vllm else False) draft_k = gr.Slider( 0, 10, value=0, step=1.0, label="Draft K", interactive=True, visible=args.speculative_sampling==True) params = [user_input, chatbot] predict_params = [ chatbot, system_prompt_input, negative_prompt_input, max_new_token, top_p, temperature, top_k, do_sample, repetition_penalty, guidance_scale, presence_penalty, draft_k] submitBtn.click( user, params, params, queue=False).then( predict, predict_params, chatbot).then( lambda: gr.update( interactive=True), None, [user_input], queue=False) user_input.submit( user, params, params, queue=False).then( predict, predict_params, chatbot).then( lambda: gr.update( interactive=True), None, [user_input], queue=False) submitBtn.click(reset_user_input, [], [user_input]) emptyBtn.click(reset_state, outputs=[chatbot], show_progress=True) def reset_user_input(): return gr.update(value='')
null
179,345
import torch from transformers import ( AutoModelForCausalLM, LlamaForCausalLM, LlamaTokenizer, StoppingCriteria, BitsAndBytesConfig, GenerationConfig ) import gradio as gr import argparse import os from queue import Queue from threading import Thread import traceback import gc import json import requests from typing import Iterable, List import subprocess import re import sys from attn_and_long_ctx_patches import apply_ntk_scaling_patch from peft import PeftModel def reset_state(): return []
null
179,346
import torch from transformers import ( AutoModelForCausalLM, LlamaForCausalLM, LlamaTokenizer, StoppingCriteria, BitsAndBytesConfig, GenerationConfig ) import gradio as gr import argparse import os from queue import Queue from threading import Thread import traceback import gc import json import requests from typing import Iterable, List import subprocess import re import sys from attn_and_long_ctx_patches import apply_ntk_scaling_patch from peft import PeftModel with gr.Blocks() as demo: github_banner_path = 'https://raw.githubusercontent.com/ymcui/Chinese-LLaMA-Alpaca-2/main/pics/banner.png' gr.HTML(f'<p align="center"><a href="https://github.com/ymcui/Chinese-LLaMA-Alpaca-2"><img src={github_banner_path} width="700"/></a></p>') chatbot = gr.Chatbot() with gr.Row(): with gr.Column(scale=4): with gr.Column(scale=3): system_prompt_input = gr.Textbox( show_label=True, label="系统提示语(仅在对话开始前或清空历史后修改有效,对话过程中修改无效)", placeholder=DEFAULT_SYSTEM_PROMPT, lines=1).style( container=True) negative_prompt_input = gr.Textbox( show_label=True, label="反向提示语(仅在对话开始前或清空历史后修改有效,对话过程中修改无效)", placeholder="(可选,默认为空)", lines=1, visible=ENABLE_CFG_SAMPLING).style( container=True) with gr.Column(scale=12): user_input = gr.Textbox( show_label=True, label="用户指令", placeholder="Shift + Enter发送消息...", lines=10).style( container=True) with gr.Column(min_width=32, scale=1): submitBtn = gr.Button("Submit", variant="primary") with gr.Column(scale=1): emptyBtn = gr.Button("Clear History") max_new_token = gr.Slider( 0, 4096, value=512, step=1.0, label="Maximum New Token Length", interactive=True) top_p = gr.Slider(0, 1, value=0.9, step=0.01, label="Top P", interactive=True) temperature = gr.Slider( 0, 1, value=0.2, step=0.01, label="Temperature", interactive=True) top_k = gr.Slider(1, 40, value=40, step=1, label="Top K", interactive=True) do_sample = gr.Checkbox( value=True, label="Do Sample", info="use random sample strategy", interactive=True) repetition_penalty = gr.Slider( 1.0, 3.0, value=1.1, step=0.1, label="Repetition Penalty", interactive=True, visible=False if args.use_vllm else True) guidance_scale = gr.Slider( 1.0, 3.0, value=1.0, step=0.1, label="Guidance Scale", interactive=True, visible=ENABLE_CFG_SAMPLING) presence_penalty = gr.Slider( -2.0, 2.0, value=1.0, step=0.1, label="Presence Penalty", interactive=True, visible=True if args.use_vllm else False) draft_k = gr.Slider( 0, 10, value=0, step=1.0, label="Draft K", interactive=True, visible=args.speculative_sampling==True) params = [user_input, chatbot] predict_params = [ chatbot, system_prompt_input, negative_prompt_input, max_new_token, top_p, temperature, top_k, do_sample, repetition_penalty, guidance_scale, presence_penalty, draft_k] submitBtn.click( user, params, params, queue=False).then( predict, predict_params, chatbot).then( lambda: gr.update( interactive=True), None, [user_input], queue=False) user_input.submit( user, params, params, queue=False).then( predict, predict_params, chatbot).then( lambda: gr.update( interactive=True), None, [user_input], queue=False) submitBtn.click(reset_user_input, [], [user_input]) emptyBtn.click(reset_state, outputs=[chatbot], show_progress=True) def user(user_message, history): return gr.update(value="", interactive=False), history + \ [[user_message, None]]
null
179,347
import torch from transformers import ( AutoModelForCausalLM, LlamaForCausalLM, LlamaTokenizer, StoppingCriteria, BitsAndBytesConfig, GenerationConfig ) import gradio as gr import argparse import os from queue import Queue from threading import Thread import traceback import gc import json import requests from typing import Iterable, List import subprocess import re DEFAULT_SYSTEM_PROMPT = """You are a helpful assistant. 你是一个乐于助人的助手。""" args = parser.parse_args() ENABLE_CFG_SAMPLING = True if args.use_vllm is True: print("CFG sampling is disabled when using vLLM.") ENABLE_CFG_SAMPLING = False if args.only_cpu is True: args.gpus = "" if args.load_in_8bit or args.load_in_4bit: raise ValueError("Quantization is unavailable on CPU.") if args.load_in_8bit and args.load_in_4bit: raise ValueError("Only one quantization method can be chosen for inference. Please check your arguments") import sys if not args.only_cpu: if args.use_flash_attention_2: from flash_attn_patch_for_inference import replace_llama_attn_with_flash_attn replace_llama_attn_with_flash_attn() else: from attn_and_long_ctx_patches import apply_attention_patch apply_attention_patch(use_memory_efficient_attention=True) from attn_and_long_ctx_patches import apply_ntk_scaling_patch if args.use_ntk: apply_ntk_scaling_patch(args.alpha) if args.speculative_sampling: if args.draft_base_model == None: raise ValueError("Speculative sampling requires a draft model. Please specify the draft model.") if args.draft_model_load_in_8bit and args.draft_model_load_in_4bit: raise ValueError("Only one quantization method can be chosen for inference. Please check your arguments") from speculative_sample import speculative_sample from peft import PeftModel def generate_prompt(instruction, response="", with_system_prompt=True, system_prompt=DEFAULT_SYSTEM_PROMPT): if with_system_prompt is True: prompt = TEMPLATE_WITH_SYSTEM_PROMPT.format_map({'instruction': instruction,'system_prompt': system_prompt}) else: prompt = TEMPLATE_WITHOUT_SYSTEM_PROMPT.format_map({'instruction': instruction}) if len(response)>0: prompt += " " + response return prompt class Stream(StoppingCriteria): def __init__(self, callback_func=None): self.callback_func = callback_func def __call__(self, input_ids, scores) -> bool: if self.callback_func is not None: self.callback_func(input_ids[0]) return False class Iteratorize: """ Transforms a function that takes a callback into a lazy iterator (generator). Adapted from: https://stackoverflow.com/a/9969000 """ def __init__(self, func, kwargs=None, callback=None): self.mfunc = func self.c_callback = callback self.q = Queue() self.sentinel = object() self.kwargs = kwargs or {} self.stop_now = False def _callback(val): if self.stop_now: raise ValueError self.q.put(val) def gentask(): try: ret = self.mfunc(callback=_callback, **self.kwargs) except ValueError: pass except Exception: traceback.print_exc() clear_torch_cache() self.q.put(self.sentinel) if self.c_callback: self.c_callback(ret) self.thread = Thread(target=gentask) self.thread.start() def __iter__(self): return self def __next__(self): obj = self.q.get(True, None) if obj is self.sentinel: raise StopIteration else: return obj def __del__(self): clear_torch_cache() def __enter__(self): return self def __exit__(self, exc_type, exc_val, exc_tb): self.stop_now = True clear_torch_cache() def clear_torch_cache(): gc.collect() if torch.cuda.device_count() > 0: torch.cuda.empty_cache() def post_http_request(prompt: str, api_url: str, n: int = 1, top_p: float = 0.9, top_k: int = 40, temperature: float = 0.2, max_tokens: int = 512, presence_penalty: float = 1.0, use_beam_search: bool = False, stream: bool = False) -> requests.Response: headers = {"User-Agent": "Test Client"} pload = { "prompt": prompt, "n": n, "top_p": 1 if use_beam_search else top_p, "top_k": -1 if use_beam_search else top_k, "temperature": 0 if use_beam_search else temperature, "max_tokens": max_tokens, "use_beam_search": use_beam_search, "best_of": 5 if use_beam_search else n, "presence_penalty": presence_penalty, "stream": stream, } print(pload) response = requests.post(api_url, headers=headers, json=pload, stream=True) return response def get_streaming_response(response: requests.Response) -> Iterable[List[str]]: for chunk in response.iter_lines(chunk_size=8192, decode_unicode=False, delimiter=b"\0"): if chunk: data = json.loads(chunk.decode("utf-8")) output = data["text"] yield output def speculative_sample( input_ids, target_model: Optional["PreTrainedModel"], draft_model: Optional["PreTrainedModel"], generation_config: GenerationConfig, logits_processor: Optional[LogitsProcessorList] = None, stopping_criteria: Optional[StoppingCriteriaList] = None, draft_k: int = 4, negative_prompt_ids: Optional[torch.Tensor] = None, negative_prompt_attention_mask: Optional[torch.Tensor] = None, streamer: Optional["BaseStreamer"] = None, **kwargs, ): generation_config = copy.deepcopy(generation_config) model_kwargs = generation_config.update(**kwargs) # All unused kwargs must be model kwargs generation_config.validate() logits_processor = logits_processor if logits_processor is not None else LogitsProcessorList() stopping_criteria = stopping_criteria if stopping_criteria is not None else StoppingCriteriaList() inputs_tensor, _, model_kwargs = target_model._prepare_model_inputs( input_ids, generation_config.bos_token_id, model_kwargs ) model_kwargs["use_cache"] = generation_config.use_cache input_ids_seq_length = input_ids.shape[-1] has_default_max_length = kwargs.get("max_length") is None and generation_config.max_length is not None if has_default_max_length and generation_config.max_new_tokens is None: # warnings.warn( # f"Using `max_length`'s default ({generation_config.max_length}) to control the generation length. " # "This behaviour is deprecated and will be removed from the config in v5 of Transformers -- we" # " recommend using `max_new_tokens` to control the maximum length of the generation.", # UserWarning, # ) pass elif generation_config.max_new_tokens is not None: # if not has_default_max_length: # logger.warning( # f"Both `max_new_tokens` (={generation_config.max_new_tokens}) and `max_length`(=" # f"{generation_config.max_length}) seem to have been set. `max_new_tokens` will take precedence. " # "Please refer to the documentation for more information. " # "(https://huggingface.co/docs/transformers/main/en/main_classes/text_generation)" # ) generation_config.max_length = generation_config.max_new_tokens + input_ids_seq_length if generation_config.min_length is not None and generation_config.min_length > generation_config.max_length: raise ValueError( f"Unfeasible length constraints: the minimum length ({generation_config.min_length}) is larger than" f" the maximum length ({generation_config.max_length})" ) if input_ids_seq_length >= generation_config.max_length: # input_ids_string = "decoder_input_ids" if target_model.config.is_encoder_decoder else "input_ids" # logger.warning( # f"Input length of {input_ids_string} is {input_ids_seq_length}, but `max_length` is set to" # f" {generation_config.max_length}. This can lead to unexpected behavior. You should consider" # " increasing `max_new_tokens`." # ) pass # prepare logis_processor, stopping_criteria, logits_warper try: logits_processor = target_model._get_logits_processor( generation_config=generation_config, input_ids_seq_length=input_ids_seq_length, encoder_input_ids=inputs_tensor, prefix_allowed_tokens_fn=None, logits_processor=logits_processor, model_kwargs=model_kwargs, negative_prompt_ids=negative_prompt_ids, negative_prompt_attention_mask=negative_prompt_attention_mask, ) except TypeError: # Please install the latest transformers (commit equal or later than d533465) to enable CFG sampling. logits_processor = target_model._get_logits_processor( generation_config=generation_config, input_ids_seq_length=input_ids_seq_length, encoder_input_ids=inputs_tensor, prefix_allowed_tokens_fn=None, logits_processor=logits_processor, ) stopping_criteria = target_model._get_stopping_criteria( generation_config=generation_config, stopping_criteria=stopping_criteria ) logits_warper=target_model._get_logits_warper(generation_config) if generation_config.do_sample else None outputs = _speculative_sampling( prefix=input_ids, target_model=target_model, draft_model=draft_model, max_new_tokens=generation_config.max_new_tokens, draft_k=draft_k, logits_processor=logits_processor, logits_warper=logits_warper, do_sample=generation_config.do_sample, eos_token_id=generation_config.eos_token_id, stopping_criteria=stopping_criteria, streamer=streamer, ) return outputs def predict( history, system_prompt, negative_prompt, max_new_tokens=128, top_p=0.9, temperature=0.2, top_k=40, do_sample=True, repetition_penalty=1.1, guidance_scale=1.0, presence_penalty=0.0, draft_k=0, ): if len(system_prompt) == 0: system_prompt = DEFAULT_SYSTEM_PROMPT while True: print("len(history):", len(history)) print("history: ", history) history[-1][1] = "" if len(history) == 1: input = history[0][0] prompt = generate_prompt(input,response="", with_system_prompt=True, system_prompt=system_prompt) else: input = history[0][0] response = history[0][1] prompt = generate_prompt(input, response=response, with_system_prompt=True, system_prompt=system_prompt)+'</s>' for hist in history[1:-1]: input = hist[0] response = hist[1] prompt = prompt + '<s>'+generate_prompt(input, response=response, with_system_prompt=False)+'</s>' input = history[-1][0] prompt = prompt + '<s>'+generate_prompt(input, response="", with_system_prompt=False) input_length = len(tokenizer.encode(prompt, add_special_tokens=True)) print(f"Input length: {input_length}") if input_length > max_memory and len(history) > 1: print(f"The input length ({input_length}) exceeds the max memory ({max_memory}). The earlier history will be discarded.") history = history[1:] print("history: ", history) else: break if args.use_vllm: generate_params = { 'max_tokens': max_new_tokens, 'top_p': top_p, 'temperature': temperature, 'top_k': top_k, "use_beam_search": not do_sample, 'presence_penalty': presence_penalty, } api_url = f"http://{args.post_host}:{args.post_port}/generate" response = post_http_request(prompt, api_url, **generate_params, stream=True) for h in get_streaming_response(response): for line in h: line = line.replace(prompt, '') history[-1][1] = line yield history else: negative_text = None if len(negative_prompt) != 0: negative_text = re.sub(r"<<SYS>>\n(.*)\n<</SYS>>", f"<<SYS>>\n{negative_prompt}\n<</SYS>>", prompt) inputs = tokenizer(prompt, return_tensors="pt") input_ids = inputs["input_ids"].to(device) if negative_text is None: negative_prompt_ids = None negative_prompt_attention_mask = None else: negative_inputs = tokenizer(negative_text,return_tensors="pt") negative_prompt_ids = negative_inputs["input_ids"].to(device) negative_prompt_attention_mask = negative_inputs["attention_mask"].to(device) generate_params = { 'input_ids': input_ids, 'max_new_tokens': max_new_tokens, 'top_p': top_p, 'temperature': temperature, 'top_k': top_k, 'do_sample': do_sample, 'repetition_penalty': repetition_penalty, 'eos_token_id': tokenizer.eos_token_id, } if ENABLE_CFG_SAMPLING is True: generate_params['guidance_scale'] = guidance_scale generate_params['negative_prompt_ids'] = negative_prompt_ids generate_params['negative_prompt_attention_mask'] = negative_prompt_attention_mask if args.speculative_sampling: generate_params['target_model'] = model generate_params['draft_model'] = draft_model generate_params['draft_k'] = draft_k generate_params['generation_config'] = GenerationConfig() def generate_with_callback(callback=None, **kwargs): if 'stopping_criteria' in kwargs: kwargs['stopping_criteria'].append(Stream(callback_func=callback)) else: kwargs['stopping_criteria'] = [Stream(callback_func=callback)] clear_torch_cache() with torch.no_grad(): if not args.speculative_sampling: model.generate(**kwargs) else: # enable speculative sampling speculative_sample(**kwargs) def generate_with_streaming(**kwargs): return Iteratorize(generate_with_callback, kwargs, callback=None) with generate_with_streaming(**generate_params) as generator: for output in generator: next_token_ids = output[len(input_ids[0]):] if next_token_ids[0] == tokenizer.eos_token_id: break new_tokens = tokenizer.decode( next_token_ids, skip_special_tokens=True) if isinstance(tokenizer, LlamaTokenizer) and len(next_token_ids) > 0: if tokenizer.convert_ids_to_tokens(int(next_token_ids[0])).startswith('▁'): new_tokens = ' ' + new_tokens history[-1][1] = new_tokens yield history if len(next_token_ids) >= max_new_tokens: break
null
179,348
import argparse import asyncio from http import HTTPStatus import json import time from typing import AsyncGenerator, Dict, List, Optional from packaging import version import fastapi from fastapi import BackgroundTasks, Request from fastapi.exceptions import RequestValidationError from fastapi.middleware.cors import CORSMiddleware from fastapi.responses import JSONResponse, StreamingResponse from fastchat.conversation import Conversation, SeparatorStyle from fastchat.model.model_adapter import get_conversation_template import uvicorn from vllm.engine.arg_utils import AsyncEngineArgs from vllm.engine.async_llm_engine import AsyncLLMEngine import sys import os from openai_api_protocol_vllm import ( CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, DeltaMessage, ErrorResponse, LogProbs, ModelCard, ModelList, ModelPermission, UsageInfo) from vllm.logger import init_logger from vllm.outputs import RequestOutput from vllm.sampling_params import SamplingParams from vllm.transformers_utils.tokenizer import get_tokenizer from vllm.utils import random_uuid from fastchat.conversation import register_conv_template, get_conv_template from fastchat.model.model_adapter import BaseModelAdapter, model_adapters import fastchat def compare_version(version1, version2): # if v1 >= v2, return True, else return False v1 = version.parse(version1) v2 = version.parse(version2) return v1 >= v2
null
179,349
import argparse import asyncio from http import HTTPStatus import json import time from typing import AsyncGenerator, Dict, List, Optional from packaging import version import fastapi from fastapi import BackgroundTasks, Request from fastapi.exceptions import RequestValidationError from fastapi.middleware.cors import CORSMiddleware from fastapi.responses import JSONResponse, StreamingResponse from fastchat.conversation import Conversation, SeparatorStyle from fastchat.model.model_adapter import get_conversation_template import uvicorn from vllm.engine.arg_utils import AsyncEngineArgs from vllm.engine.async_llm_engine import AsyncLLMEngine import sys import os from openai_api_protocol_vllm import ( CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, DeltaMessage, ErrorResponse, LogProbs, ModelCard, ModelList, ModelPermission, UsageInfo) from vllm.logger import init_logger from vllm.outputs import RequestOutput from vllm.sampling_params import SamplingParams from vllm.transformers_utils.tokenizer import get_tokenizer from vllm.utils import random_uuid from fastchat.conversation import register_conv_template, get_conv_template from fastchat.model.model_adapter import BaseModelAdapter, model_adapters import fastchat def create_error_response(status_code: HTTPStatus, message: str) -> JSONResponse: return JSONResponse(ErrorResponse(message=message, type="invalid_request_error").dict(), status_code=status_code.value) async def validation_exception_handler(request, exc): # pylint: disable=unused-argument return create_error_response(HTTPStatus.BAD_REQUEST, str(exc))
null
179,350
import argparse import asyncio from http import HTTPStatus import json import time from typing import AsyncGenerator, Dict, List, Optional from packaging import version import fastapi from fastapi import BackgroundTasks, Request from fastapi.exceptions import RequestValidationError from fastapi.middleware.cors import CORSMiddleware from fastapi.responses import JSONResponse, StreamingResponse from fastchat.conversation import Conversation, SeparatorStyle from fastchat.model.model_adapter import get_conversation_template import uvicorn from vllm.engine.arg_utils import AsyncEngineArgs from vllm.engine.async_llm_engine import AsyncLLMEngine import sys import os from openai_api_protocol_vllm import ( CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, DeltaMessage, ErrorResponse, LogProbs, ModelCard, ModelList, ModelPermission, UsageInfo) from vllm.logger import init_logger from vllm.outputs import RequestOutput from vllm.sampling_params import SamplingParams from vllm.transformers_utils.tokenizer import get_tokenizer from vllm.utils import random_uuid served_model = None from fastchat.conversation import register_conv_template, get_conv_template from fastchat.model.model_adapter import BaseModelAdapter, model_adapters import fastchat class ModelPermission(BaseModel): id: str = Field(default_factory=lambda: f"modelperm-{random_uuid()}") object: str = "model_permission" created: int = Field(default_factory=lambda: int(time.time())) allow_create_engine: bool = False allow_sampling: bool = True allow_logprobs: bool = True allow_search_indices: bool = False allow_view: bool = True allow_fine_tuning: bool = False organization: str = "*" group: Optional[str] = None is_blocking: str = False class ModelCard(BaseModel): id: str object: str = "model" created: int = Field(default_factory=lambda: int(time.time())) owned_by: str = "vllm" root: Optional[str] = None parent: Optional[str] = None permission: List[ModelPermission] = Field(default_factory=list) class ModelList(BaseModel): object: str = "list" data: List[ModelCard] = Field(default_factory=list) The provided code snippet includes necessary dependencies for implementing the `show_available_models` function. Write a Python function `async def show_available_models()` to solve the following problem: Show available models. Right now we only have one model. Here is the function: async def show_available_models(): """Show available models. Right now we only have one model.""" model_cards = [ ModelCard(id=served_model, root=served_model, permission=[ModelPermission()]) ] return ModelList(data=model_cards)
Show available models. Right now we only have one model.
179,351
import argparse import asyncio from http import HTTPStatus import json import time from typing import AsyncGenerator, Dict, List, Optional from packaging import version import fastapi from fastapi import BackgroundTasks, Request from fastapi.exceptions import RequestValidationError from fastapi.middleware.cors import CORSMiddleware from fastapi.responses import JSONResponse, StreamingResponse from fastchat.conversation import Conversation, SeparatorStyle from fastchat.model.model_adapter import get_conversation_template import uvicorn from vllm.engine.arg_utils import AsyncEngineArgs from vllm.engine.async_llm_engine import AsyncLLMEngine import sys import os from openai_api_protocol_vllm import ( CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, DeltaMessage, ErrorResponse, LogProbs, ModelCard, ModelList, ModelPermission, UsageInfo) from vllm.logger import init_logger from vllm.outputs import RequestOutput from vllm.sampling_params import SamplingParams from vllm.transformers_utils.tokenizer import get_tokenizer from vllm.utils import random_uuid logger = init_logger(__name__) from fastchat.conversation import register_conv_template, get_conv_template from fastchat.model.model_adapter import BaseModelAdapter, model_adapters import fastchat def create_error_response(status_code: HTTPStatus, message: str) -> JSONResponse: return JSONResponse(ErrorResponse(message=message, type="invalid_request_error").dict(), status_code=status_code.value) async def check_model(request) -> Optional[JSONResponse]: if request.model == served_model: return ret = create_error_response( HTTPStatus.NOT_FOUND, f"The model `{request.model}` does not exist.", ) return ret async def get_gen_prompt(request) -> str: conv = get_conversation_template(request.model) conv = getConversation( name=conv.name, system=conv.system_message if not use_old_conversation else conv.system, roles=conv.roles, messages=list(conv.messages), # prevent in-place modification offset=conv.offset, sep_style=SeparatorStyle(conv.sep_style), sep=conv.sep, sep2=conv.sep2, stop_str=conv.stop_str, stop_token_ids=conv.stop_token_ids, ) if isinstance(request.messages, str): prompt = request.messages else: for message in request.messages: msg_role = message["role"] if msg_role == "system": if not use_old_conversation: conv.system_message = message["content"] else: conv.system = message["content"] elif msg_role == "user": conv.append_message(conv.roles[0], message["content"]) elif msg_role == "assistant": conv.append_message(conv.roles[1], message["content"]) else: raise ValueError(f"Unknown role: {msg_role}") # Add a blank message for the assistant. conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() return prompt async def check_length(request, prompt, model_config): if hasattr(model_config.hf_config, "max_sequence_length"): context_len = model_config.hf_config.max_sequence_length elif hasattr(model_config.hf_config, "seq_length"): context_len = model_config.hf_config.seq_length elif hasattr(model_config.hf_config, "max_position_embeddings"): context_len = model_config.hf_config.max_position_embeddings elif hasattr(model_config.hf_config, "seq_length"): context_len = model_config.hf_config.seq_length else: context_len = 2048 input_ids = tokenizer(prompt).input_ids token_num = len(input_ids) if token_num + request.max_tokens > context_len: return create_error_response( HTTPStatus.BAD_REQUEST, f"This model's maximum context length is {context_len} tokens. " f"However, you requested {request.max_tokens + token_num} tokens " f"({token_num} in the messages, " f"{request.max_tokens} in the completion). " f"Please reduce the length of the messages or completion.", ) else: return None class UsageInfo(BaseModel): prompt_tokens: int = 0 total_tokens: int = 0 completion_tokens: Optional[int] = 0 class ChatCompletionRequest(BaseModel): model: str messages: Union[str, List[Dict[str, str]]] temperature: Optional[float] = 0.2 top_p: Optional[float] = 0.9 n: Optional[int] = 1 max_tokens: Optional[int] = 512 stop: Optional[Union[str, List[str]]] = Field(default_factory=list) stream: Optional[bool] = False presence_penalty: Optional[float] = 1.0 frequency_penalty: Optional[float] = 0.0 logit_bias: Optional[Dict[str, float]] = None user: Optional[str] = None # Additional parameters supported by vLLM best_of: Optional[int] = None top_k: Optional[int] = 40 ignore_eos: Optional[bool] = False use_beam_search: Optional[bool] = False class ChatMessage(BaseModel): role: str content: str class ChatCompletionResponseChoice(BaseModel): index: int message: ChatMessage finish_reason: Optional[Literal["stop", "length"]] = None class ChatCompletionResponse(BaseModel): id: str = Field(default_factory=lambda: f"chatcmpl-{random_uuid()}") object: str = "chat.completion" created: int = Field(default_factory=lambda: int(time.time())) model: str choices: List[ChatCompletionResponseChoice] usage: UsageInfo class DeltaMessage(BaseModel): role: Optional[str] = None content: Optional[str] = None class ChatCompletionResponseStreamChoice(BaseModel): index: int delta: DeltaMessage finish_reason: Optional[Literal["stop", "length"]] = None class ChatCompletionStreamResponse(BaseModel): id: str = Field(default_factory=lambda: f"chatcmpl-{random_uuid()}") object: str = "chat.completion.chunk" created: int = Field(default_factory=lambda: int(time.time())) model: str choices: List[ChatCompletionResponseStreamChoice] The provided code snippet includes necessary dependencies for implementing the `create_chat_completion` function. Write a Python function `async def create_chat_completion(raw_request: Request)` to solve the following problem: Completion API similar to OpenAI's API. See https://platform.openai.com/docs/api-reference/chat/create for the API specification. This API mimics the OpenAI ChatCompletion API. NOTE: Currently we do not support the following features: - function_call (Users should implement this by themselves) - logit_bias (to be supported by vLLM engine) Here is the function: async def create_chat_completion(raw_request: Request): """Completion API similar to OpenAI's API. See https://platform.openai.com/docs/api-reference/chat/create for the API specification. This API mimics the OpenAI ChatCompletion API. NOTE: Currently we do not support the following features: - function_call (Users should implement this by themselves) - logit_bias (to be supported by vLLM engine) """ request = ChatCompletionRequest(**await raw_request.json()) logger.info(f"Received chat completion request: {request}") error_check_ret = await check_model(request) if error_check_ret is not None: return error_check_ret if request.logit_bias is not None: # TODO: support logit_bias in vLLM engine. return create_error_response(HTTPStatus.BAD_REQUEST, "logit_bias is not currently supported") prompt = await get_gen_prompt(request) error_check_ret = await check_length(request, prompt, engine_model_config) if error_check_ret is not None: return error_check_ret model_name = request.model request_id = f"cmpl-{random_uuid()}" created_time = int(time.time()) try: sampling_params = SamplingParams( n=request.n, presence_penalty=request.presence_penalty, frequency_penalty=request.frequency_penalty, temperature=request.temperature, top_p=request.top_p, stop=request.stop, max_tokens=request.max_tokens, best_of=request.best_of, top_k=request.top_k, ignore_eos=request.ignore_eos, use_beam_search=request.use_beam_search, ) except ValueError as e: return create_error_response(HTTPStatus.BAD_REQUEST, str(e)) result_generator = engine.generate(prompt, sampling_params, request_id) async def abort_request() -> None: await engine.abort(request_id) def create_stream_response_json( index: int, text: str, finish_reason: Optional[str] = None, ) -> str: choice_data = ChatCompletionResponseStreamChoice( index=index, delta=DeltaMessage(content=text), finish_reason=finish_reason, ) response = ChatCompletionStreamResponse( id=request_id, created=created_time, model=model_name, choices=[choice_data], ) response_json = response.json(ensure_ascii=False) return response_json async def completion_stream_generator() -> AsyncGenerator[str, None]: # First chunk with role for i in range(request.n): choice_data = ChatCompletionResponseStreamChoice( index=i, delta=DeltaMessage(role="assistant"), finish_reason=None, ) chunk = ChatCompletionStreamResponse(id=request_id, choices=[choice_data], model=model_name) data = chunk.json(exclude_unset=True, ensure_ascii=False) yield f"data: {data}\n\n" previous_texts = [""] * request.n previous_num_tokens = [0] * request.n async for res in result_generator: res: RequestOutput for output in res.outputs: i = output.index delta_text = output.text[len(previous_texts[i]):] previous_texts[i] = output.text previous_num_tokens[i] = len(output.token_ids) response_json = create_stream_response_json( index=i, text=delta_text, ) yield f"data: {response_json}\n\n" if output.finish_reason is not None: response_json = create_stream_response_json( index=i, text="", finish_reason=output.finish_reason, ) yield f"data: {response_json}\n\n" yield "data: [DONE]\n\n" # Streaming response if request.stream: background_tasks = BackgroundTasks() # Abort the request if the client disconnects. background_tasks.add_task(abort_request) return StreamingResponse(completion_stream_generator(), media_type="text/event-stream", background=background_tasks) # Non-streaming response final_res: RequestOutput = None async for res in result_generator: if await raw_request.is_disconnected(): # Abort the request if the client disconnects. await abort_request() return create_error_response(HTTPStatus.BAD_REQUEST, "Client disconnected") final_res = res assert final_res is not None choices = [] for output in final_res.outputs: choice_data = ChatCompletionResponseChoice( index=output.index, message=ChatMessage(role="assistant", content=output.text), finish_reason=output.finish_reason, ) choices.append(choice_data) num_prompt_tokens = len(final_res.prompt_token_ids) num_generated_tokens = sum( len(output.token_ids) for output in final_res.outputs) usage = UsageInfo( prompt_tokens=num_prompt_tokens, completion_tokens=num_generated_tokens, total_tokens=num_prompt_tokens + num_generated_tokens, ) response = ChatCompletionResponse( id=request_id, created=created_time, model=model_name, choices=choices, usage=usage, ) if request.stream: # When user requests streaming but we don't stream, we still need to # return a streaming response with a single event. response_json = response.json(ensure_ascii=False) async def fake_stream_generator() -> AsyncGenerator[str, None]: yield f"data: {response_json}\n\n" yield "data: [DONE]\n\n" return StreamingResponse(fake_stream_generator(), media_type="text/event-stream") return response
Completion API similar to OpenAI's API. See https://platform.openai.com/docs/api-reference/chat/create for the API specification. This API mimics the OpenAI ChatCompletion API. NOTE: Currently we do not support the following features: - function_call (Users should implement this by themselves) - logit_bias (to be supported by vLLM engine)
179,352
import argparse import asyncio from http import HTTPStatus import json import time from typing import AsyncGenerator, Dict, List, Optional from packaging import version import fastapi from fastapi import BackgroundTasks, Request from fastapi.exceptions import RequestValidationError from fastapi.middleware.cors import CORSMiddleware from fastapi.responses import JSONResponse, StreamingResponse from fastchat.conversation import Conversation, SeparatorStyle from fastchat.model.model_adapter import get_conversation_template import uvicorn from vllm.engine.arg_utils import AsyncEngineArgs from vllm.engine.async_llm_engine import AsyncLLMEngine import sys import os from openai_api_protocol_vllm import ( CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, DeltaMessage, ErrorResponse, LogProbs, ModelCard, ModelList, ModelPermission, UsageInfo) from vllm.logger import init_logger from vllm.outputs import RequestOutput from vllm.sampling_params import SamplingParams from vllm.transformers_utils.tokenizer import get_tokenizer from vllm.utils import random_uuid logger = init_logger(__name__) from fastchat.conversation import register_conv_template, get_conv_template from fastchat.model.model_adapter import BaseModelAdapter, model_adapters import fastchat def create_error_response(status_code: HTTPStatus, message: str) -> JSONResponse: return JSONResponse(ErrorResponse(message=message, type="invalid_request_error").dict(), status_code=status_code.value) async def check_model(request) -> Optional[JSONResponse]: if request.model == served_model: return ret = create_error_response( HTTPStatus.NOT_FOUND, f"The model `{request.model}` does not exist.", ) return ret async def get_gen_prompt_nochat(request) -> str: conv = get_conversation_template(request.model) conv = getConversation( name=conv.name, system=conv.system_message if not use_old_conversation else conv.system, roles=conv.roles, messages=list(conv.messages), # prevent in-place modification offset=conv.offset, sep_style=SeparatorStyle(conv.sep_style), sep=conv.sep, sep2=conv.sep2, stop_str=conv.stop_str, stop_token_ids=conv.stop_token_ids, ) prompt = request.prompt conv.append_message(conv.roles[0], prompt) conv.append_message(conv.roles[1], None) prompt = conv.get_prompt() return prompt def create_logprobs(token_ids: List[int], id_logprobs: List[Dict[int, float]], initial_text_offset: int = 0) -> LogProbs: """Create OpenAI-style logprobs.""" logprobs = LogProbs() last_token_len = 0 for token_id, id_logprob in zip(token_ids, id_logprobs): token = tokenizer.convert_ids_to_tokens(token_id) logprobs.tokens.append(token) logprobs.token_logprobs.append(id_logprob[token_id]) if len(logprobs.text_offset) == 0: logprobs.text_offset.append(initial_text_offset) else: logprobs.text_offset.append(logprobs.text_offset[-1] + last_token_len) last_token_len = len(token) logprobs.top_logprobs.append({ tokenizer.convert_ids_to_tokens(i): p for i, p in id_logprob.items() }) return logprobs class UsageInfo(BaseModel): prompt_tokens: int = 0 total_tokens: int = 0 completion_tokens: Optional[int] = 0 class CompletionRequest(BaseModel): model: str prompt: Union[str, List[str]] suffix: Optional[str] = None max_tokens: Optional[int] = 512 temperature: Optional[float] = 0.2 top_p: Optional[float] = 0.9 n: Optional[int] = 1 stream: Optional[bool] = False logprobs: Optional[int] = None echo: Optional[bool] = False stop: Optional[Union[str, List[str]]] = Field(default_factory=list) presence_penalty: Optional[float] = 1.0 frequency_penalty: Optional[float] = 0.0 best_of: Optional[int] = None logit_bias: Optional[Dict[str, float]] = None user: Optional[str] = None # Additional parameters supported by vLLM top_k: Optional[int] = 40 ignore_eos: Optional[bool] = False use_beam_search: Optional[bool] = False class LogProbs(BaseModel): text_offset: List[int] = Field(default_factory=list) token_logprobs: List[Optional[float]] = Field(default_factory=list) tokens: List[str] = Field(default_factory=list) top_logprobs: List[Optional[Dict[str, float]]] = Field(default_factory=list) class CompletionResponseChoice(BaseModel): index: int text: str logprobs: Optional[LogProbs] = None finish_reason: Optional[Literal["stop", "length"]] = None class CompletionResponse(BaseModel): id: str = Field(default_factory=lambda: f"cmpl-{random_uuid()}") object: str = "text_completion" created: int = Field(default_factory=lambda: int(time.time())) model: str choices: List[CompletionResponseChoice] usage: UsageInfo class CompletionResponseStreamChoice(BaseModel): index: int text: str logprobs: Optional[LogProbs] = None finish_reason: Optional[Literal["stop", "length"]] = None class CompletionStreamResponse(BaseModel): id: str = Field(default_factory=lambda: f"cmpl-{random_uuid()}") object: str = "text_completion" created: int = Field(default_factory=lambda: int(time.time())) model: str choices: List[CompletionResponseStreamChoice] The provided code snippet includes necessary dependencies for implementing the `create_completion` function. Write a Python function `async def create_completion(raw_request: Request)` to solve the following problem: Completion API similar to OpenAI's API. See https://platform.openai.com/docs/api-reference/completions/create for the API specification. This API mimics the OpenAI Completion API. NOTE: Currently we do not support the following features: - echo (since the vLLM engine does not currently support getting the logprobs of prompt tokens) - suffix (the language models we currently support do not support suffix) - logit_bias (to be supported by vLLM engine) Here is the function: async def create_completion(raw_request: Request): """Completion API similar to OpenAI's API. See https://platform.openai.com/docs/api-reference/completions/create for the API specification. This API mimics the OpenAI Completion API. NOTE: Currently we do not support the following features: - echo (since the vLLM engine does not currently support getting the logprobs of prompt tokens) - suffix (the language models we currently support do not support suffix) - logit_bias (to be supported by vLLM engine) """ request = CompletionRequest(**await raw_request.json()) logger.info(f"Received completion request: {request}") error_check_ret = await check_model(request) if error_check_ret is not None: return error_check_ret if request.echo: # We do not support echo since the vLLM engine does not # currently support getting the logprobs of prompt tokens. return create_error_response(HTTPStatus.BAD_REQUEST, "echo is not currently supported") if request.suffix is not None: # The language models we currently support do not support suffix. return create_error_response(HTTPStatus.BAD_REQUEST, "suffix is not currently supported") if request.logit_bias is not None: # TODO: support logit_bias in vLLM engine. return create_error_response(HTTPStatus.BAD_REQUEST, "logit_bias is not currently supported") model_name = request.model request_id = f"cmpl-{random_uuid()}" if isinstance(request.prompt, list): if len(request.prompt) == 0: return create_error_response(HTTPStatus.BAD_REQUEST, "please provide at least one prompt") if len(request.prompt) > 1: return create_error_response( HTTPStatus.BAD_REQUEST, "multiple prompts in a batch is not currently supported") prompt = request.prompt[0] else: prompt = request.prompt request.prompt = prompt prompt = await get_gen_prompt_nochat(request) created_time = int(time.time()) try: sampling_params = SamplingParams( n=request.n, best_of=request.best_of, presence_penalty=request.presence_penalty, frequency_penalty=request.frequency_penalty, temperature=request.temperature, top_p=request.top_p, top_k=request.top_k, stop=request.stop, ignore_eos=request.ignore_eos, max_tokens=request.max_tokens, logprobs=request.logprobs, use_beam_search=request.use_beam_search, ) except ValueError as e: return create_error_response(HTTPStatus.BAD_REQUEST, str(e)) result_generator = engine.generate(prompt, sampling_params, request_id) # Similar to the OpenAI API, when n != best_of, we do not stream the # results. In addition, we do not stream the results when use beam search. stream = (request.stream and (request.best_of is None or request.n == request.best_of) and not request.use_beam_search) async def abort_request() -> None: await engine.abort(request_id) def create_stream_response_json( index: int, text: str, logprobs: Optional[LogProbs] = None, finish_reason: Optional[str] = None, ) -> str: choice_data = CompletionResponseStreamChoice( index=index, text=text, logprobs=logprobs, finish_reason=finish_reason, ) response = CompletionStreamResponse( id=request_id, created=created_time, model=model_name, choices=[choice_data], ) response_json = response.json(ensure_ascii=False) return response_json async def completion_stream_generator() -> AsyncGenerator[str, None]: previous_texts = [""] * request.n previous_num_tokens = [0] * request.n async for res in result_generator: res: RequestOutput for output in res.outputs: i = output.index delta_text = output.text[len(previous_texts[i]):] if request.logprobs is not None: logprobs = create_logprobs( output.token_ids[previous_num_tokens[i]:], output.logprobs[previous_num_tokens[i]:], len(previous_texts[i])) else: logprobs = None previous_texts[i] = output.text previous_num_tokens[i] = len(output.token_ids) response_json = create_stream_response_json( index=i, text=delta_text, logprobs=logprobs, ) yield f"data: {response_json}\n\n" if output.finish_reason is not None: logprobs = (LogProbs() if request.logprobs is not None else None) response_json = create_stream_response_json( index=i, text="", logprobs=logprobs, finish_reason=output.finish_reason, ) yield f"data: {response_json}\n\n" yield "data: [DONE]\n\n" # Streaming response if stream: background_tasks = BackgroundTasks() # Abort the request if the client disconnects. background_tasks.add_task(abort_request) return StreamingResponse(completion_stream_generator(), media_type="text/event-stream", background=background_tasks) # Non-streaming response final_res: RequestOutput = None async for res in result_generator: if await raw_request.is_disconnected(): # Abort the request if the client disconnects. await abort_request() return create_error_response(HTTPStatus.BAD_REQUEST, "Client disconnected") final_res = res assert final_res is not None choices = [] for output in final_res.outputs: if request.logprobs is not None: logprobs = create_logprobs(output.token_ids, output.logprobs) else: logprobs = None choice_data = CompletionResponseChoice( index=output.index, text=output.text, logprobs=logprobs, finish_reason=output.finish_reason, ) choices.append(choice_data) num_prompt_tokens = len(final_res.prompt_token_ids) num_generated_tokens = sum( len(output.token_ids) for output in final_res.outputs) usage = UsageInfo( prompt_tokens=num_prompt_tokens, completion_tokens=num_generated_tokens, total_tokens=num_prompt_tokens + num_generated_tokens, ) response = CompletionResponse( id=request_id, created=created_time, model=model_name, choices=choices, usage=usage, ) if request.stream: # When user requests streaming but we don't stream, we still need to # return a streaming response with a single event. response_json = response.json(ensure_ascii=False) async def fake_stream_generator() -> AsyncGenerator[str, None]: yield f"data: {response_json}\n\n" yield "data: [DONE]\n\n" return StreamingResponse(fake_stream_generator(), media_type="text/event-stream") return response
Completion API similar to OpenAI's API. See https://platform.openai.com/docs/api-reference/completions/create for the API specification. This API mimics the OpenAI Completion API. NOTE: Currently we do not support the following features: - echo (since the vLLM engine does not currently support getting the logprobs of prompt tokens) - suffix (the language models we currently support do not support suffix) - logit_bias (to be supported by vLLM engine)
179,353
import argparse import os from fastapi import FastAPI from fastapi.middleware.cors import CORSMiddleware import uvicorn from threading import Thread from sse_starlette.sse import EventSourceResponse import torch import torch.nn.functional as F from transformers import ( AutoModelForCausalLM, LlamaTokenizer, GenerationConfig, TextIteratorStreamer, BitsAndBytesConfig ) from peft import PeftModel import sys from attn_and_long_ctx_patches import apply_attention_patch, apply_ntk_scaling_patch from openai_api_protocol import ( ChatCompletionRequest, ChatCompletionResponse, ChatMessage, ChatCompletionResponseChoice, CompletionRequest, CompletionResponse, CompletionResponseChoice, EmbeddingsRequest, EmbeddingsResponse, ChatCompletionResponseStreamChoice, DeltaMessage, ) def predict( input, max_new_tokens=128, top_p=0.9, temperature=0.2, top_k=40, num_beams=1, repetition_penalty=1.1, do_sample=True, **kwargs, ): """ Main inference method type(input) == str -> /v1/completions type(input) == list -> /v1/chat/completions """ if isinstance(input, str): prompt = generate_completion_prompt(input) else: prompt = generate_chat_prompt(input) inputs = tokenizer(prompt, return_tensors="pt") input_ids = inputs["input_ids"].to(device) generation_config = GenerationConfig( temperature=temperature, top_p=top_p, top_k=top_k, num_beams=num_beams, do_sample=do_sample, **kwargs, ) generation_config.return_dict_in_generate = True generation_config.output_scores = False generation_config.max_new_tokens = max_new_tokens generation_config.repetition_penalty = float(repetition_penalty) with torch.no_grad(): generation_output = model.generate( input_ids=input_ids, generation_config=generation_config, ) s = generation_output.sequences[0] output = tokenizer.decode(s, skip_special_tokens=True) output = output.split("[/INST]")[-1].strip() return output def stream_predict( input, max_new_tokens=128, top_p=0.75, temperature=0.1, top_k=40, num_beams=4, repetition_penalty=1.0, do_sample=True, model_id="chinese-llama-alpaca-2", **kwargs, ): choice_data = ChatCompletionResponseStreamChoice( index=0, delta=DeltaMessage(role="assistant"), finish_reason=None ) chunk = ChatCompletionResponse( model=model_id, choices=[choice_data], object="chat.completion.chunk", ) yield "{}".format(chunk.json(exclude_unset=True, ensure_ascii=False)) if isinstance(input, str): prompt = generate_completion_prompt(input) else: prompt = generate_chat_prompt(input) inputs = tokenizer(prompt, return_tensors="pt") input_ids = inputs["input_ids"].to(device) generation_config = GenerationConfig( temperature=temperature, top_p=top_p, top_k=top_k, num_beams=num_beams, do_sample=do_sample, **kwargs, ) streamer = TextIteratorStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True) generation_kwargs = dict( streamer=streamer, input_ids=input_ids, generation_config=generation_config, return_dict_in_generate=True, output_scores=False, max_new_tokens=max_new_tokens, repetition_penalty=float(repetition_penalty), ) Thread(target=model.generate, kwargs=generation_kwargs).start() for new_text in streamer: choice_data = ChatCompletionResponseStreamChoice( index=0, delta=DeltaMessage(content=new_text), finish_reason=None ) chunk = ChatCompletionResponse( model=model_id, choices=[choice_data], object="chat.completion.chunk" ) yield "{}".format(chunk.json(exclude_unset=True, ensure_ascii=False)) choice_data = ChatCompletionResponseStreamChoice( index=0, delta=DeltaMessage(), finish_reason="stop" ) chunk = ChatCompletionResponse( model=model_id, choices=[choice_data], object="chat.completion.chunk" ) yield "{}".format(chunk.json(exclude_unset=True, ensure_ascii=False)) yield "[DONE]" class ChatCompletionRequest(BaseModel): model: str = "chinese-llama-alpaca-2" messages: Union[str, List[Dict[str, str]]] temperature: Optional[float] = 0.2 top_p: Optional[float] = 0.9 top_k: Optional[int] = 40 n: Optional[int] = 1 max_tokens: Optional[int] = 512 num_beams: Optional[int] = 1 stop: Optional[Union[str, List[str]]] = None stream: Optional[bool] = False repetition_penalty: Optional[float] = 1.1 user: Optional[str] = None do_sample: Optional[bool] = True class ChatMessage(BaseModel): role: str content: str class ChatCompletionResponseChoice(BaseModel): index: int message: ChatMessage class ChatCompletionResponse(BaseModel): id: str = Field(default_factory=lambda: f"chatcmpl-{shortuuid.random()}") object: str = "chat.completion" created: int = Field(default_factory=lambda: int(time.time())) model: str = "chinese-llama-alpaca-2" choices: List[ Union[ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice] ] The provided code snippet includes necessary dependencies for implementing the `create_chat_completion` function. Write a Python function `async def create_chat_completion(request: ChatCompletionRequest)` to solve the following problem: Creates a completion for the chat message Here is the function: async def create_chat_completion(request: ChatCompletionRequest): """Creates a completion for the chat message""" msgs = request.messages if isinstance(msgs, str): msgs = [ChatMessage(role="user", content=msgs)] else: msgs = [ChatMessage(role=x["role"], content=x["content"]) for x in msgs] if request.stream: generate = stream_predict( input=msgs, max_new_tokens=request.max_tokens, top_p=request.top_p, top_k=request.top_k, temperature=request.temperature, num_beams=request.num_beams, repetition_penalty=request.repetition_penalty, do_sample=request.do_sample, ) return EventSourceResponse(generate, media_type="text/event-stream") output = predict( input=msgs, max_new_tokens=request.max_tokens, top_p=request.top_p, top_k=request.top_k, temperature=request.temperature, num_beams=request.num_beams, repetition_penalty=request.repetition_penalty, do_sample=request.do_sample, ) choices = [ ChatCompletionResponseChoice(index=i, message=msg) for i, msg in enumerate(msgs) ] choices += [ ChatCompletionResponseChoice( index=len(choices), message=ChatMessage(role="assistant", content=output) ) ] return ChatCompletionResponse(choices=choices)
Creates a completion for the chat message
179,354
import argparse import os from fastapi import FastAPI from fastapi.middleware.cors import CORSMiddleware import uvicorn from threading import Thread from sse_starlette.sse import EventSourceResponse import torch import torch.nn.functional as F from transformers import ( AutoModelForCausalLM, LlamaTokenizer, GenerationConfig, TextIteratorStreamer, BitsAndBytesConfig ) from peft import PeftModel import sys from attn_and_long_ctx_patches import apply_attention_patch, apply_ntk_scaling_patch from openai_api_protocol import ( ChatCompletionRequest, ChatCompletionResponse, ChatMessage, ChatCompletionResponseChoice, CompletionRequest, CompletionResponse, CompletionResponseChoice, EmbeddingsRequest, EmbeddingsResponse, ChatCompletionResponseStreamChoice, DeltaMessage, ) def predict( input, max_new_tokens=128, top_p=0.9, temperature=0.2, top_k=40, num_beams=1, repetition_penalty=1.1, do_sample=True, **kwargs, ): """ Main inference method type(input) == str -> /v1/completions type(input) == list -> /v1/chat/completions """ if isinstance(input, str): prompt = generate_completion_prompt(input) else: prompt = generate_chat_prompt(input) inputs = tokenizer(prompt, return_tensors="pt") input_ids = inputs["input_ids"].to(device) generation_config = GenerationConfig( temperature=temperature, top_p=top_p, top_k=top_k, num_beams=num_beams, do_sample=do_sample, **kwargs, ) generation_config.return_dict_in_generate = True generation_config.output_scores = False generation_config.max_new_tokens = max_new_tokens generation_config.repetition_penalty = float(repetition_penalty) with torch.no_grad(): generation_output = model.generate( input_ids=input_ids, generation_config=generation_config, ) s = generation_output.sequences[0] output = tokenizer.decode(s, skip_special_tokens=True) output = output.split("[/INST]")[-1].strip() return output class CompletionRequest(BaseModel): prompt: Union[str, List[Any]] temperature: Optional[float] = 0.2 n: Optional[int] = 1 max_tokens: Optional[int] = 512 stop: Optional[Union[str, List[str]]] = None stream: Optional[bool] = False top_p: Optional[float] = 0.9 top_k: Optional[int] = 40 num_beams: Optional[int] = 1 logprobs: Optional[int] = None echo: Optional[bool] = False repetition_penalty: Optional[float] = 1.1 user: Optional[str] = None do_sample: Optional[bool] = True class CompletionResponseChoice(BaseModel): index: int text: str class CompletionResponse(BaseModel): id: Optional[str] = Field(default_factory=lambda: f"cmpl-{shortuuid.random()}") object: Optional[str] = "text_completion" created: Optional[int] = Field(default_factory=lambda: int(time.time())) model: Optional[str] = "chinese-llama-alpaca-2" choices: List[CompletionResponseChoice] The provided code snippet includes necessary dependencies for implementing the `create_completion` function. Write a Python function `async def create_completion(request: CompletionRequest)` to solve the following problem: Creates a completion Here is the function: async def create_completion(request: CompletionRequest): """Creates a completion""" output = predict( input=request.prompt, max_new_tokens=request.max_tokens, top_p=request.top_p, top_k=request.top_k, temperature=request.temperature, num_beams=request.num_beams, repetition_penalty=request.repetition_penalty, do_sample=request.do_sample, ) choices = [CompletionResponseChoice(index=0, text=output)] return CompletionResponse(choices=choices)
Creates a completion
179,355
import argparse import os from fastapi import FastAPI from fastapi.middleware.cors import CORSMiddleware import uvicorn from threading import Thread from sse_starlette.sse import EventSourceResponse import torch import torch.nn.functional as F from transformers import ( AutoModelForCausalLM, LlamaTokenizer, GenerationConfig, TextIteratorStreamer, BitsAndBytesConfig ) from peft import PeftModel import sys from attn_and_long_ctx_patches import apply_attention_patch, apply_ntk_scaling_patch from openai_api_protocol import ( ChatCompletionRequest, ChatCompletionResponse, ChatMessage, ChatCompletionResponseChoice, CompletionRequest, CompletionResponse, CompletionResponseChoice, EmbeddingsRequest, EmbeddingsResponse, ChatCompletionResponseStreamChoice, DeltaMessage, ) def get_embedding(input): """Get embedding main function""" with torch.no_grad(): encoding = tokenizer(input, padding=True, return_tensors="pt") input_ids = encoding["input_ids"].to(device) attention_mask = encoding["attention_mask"].to(device) model_output = model(input_ids, attention_mask, output_hidden_states=True) data = model_output.hidden_states[-1] mask = attention_mask.unsqueeze(-1).expand(data.size()).float() masked_embeddings = data * mask sum_embeddings = torch.sum(masked_embeddings, dim=1) seq_length = torch.sum(mask, dim=1) embedding = sum_embeddings / seq_length normalized_embeddings = F.normalize(embedding, p=2, dim=1) ret = normalized_embeddings.squeeze(0).tolist() return ret class EmbeddingsRequest(BaseModel): input: Union[str, List[Any]] user: Optional[str] = None class EmbeddingsResponse(BaseModel): object: str = "list" data: List[Dict[str, Any]] model: str = "chinese-llama-alpaca-2" The provided code snippet includes necessary dependencies for implementing the `create_embeddings` function. Write a Python function `async def create_embeddings(request: EmbeddingsRequest)` to solve the following problem: Creates text embedding Here is the function: async def create_embeddings(request: EmbeddingsRequest): """Creates text embedding""" embedding = get_embedding(request.input) data = [{"object": "embedding", "embedding": embedding, "index": 0}] return EmbeddingsResponse(data=data)
Creates text embedding
179,356
from dotenv import load_dotenv from langchain.chains import RetrievalQA from langchain.embeddings import HuggingFaceEmbeddings from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler from langchain.vectorstores import Chroma from langchain.llms import GPT4All, LlamaCpp import os import argparse import time from constants import CHROMA_SETTINGS def parse_arguments(): parser = argparse.ArgumentParser(description='privateGPT: Ask questions to your documents without an internet connection, ' 'using the power of LLMs.') parser.add_argument("--hide-source", "-S", action='store_true', help='Use this flag to disable printing of source documents used for answers.') parser.add_argument("--mute-stream", "-M", action='store_true', help='Use this flag to disable the streaming StdOut callback for LLMs.') return parser.parse_args()
null
179,358
import logging import numpy as np import math import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, List, Dict, Any, Mapping from pathlib import Path import datasets import torch from datasets import load_dataset, concatenate_datasets import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_CAUSAL_LM_MAPPING, AutoConfig, AutoModelForCausalLM, LlamaForCausalLM, LlamaTokenizer, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, is_torch_tpu_available, set_seed, BitsAndBytesConfig ) from transformers.testing_utils import CaptureLogger from transformers.trainer_utils import get_last_checkpoint from transformers.utils import send_example_telemetry from transformers.utils.versions import require_version from sklearn.metrics import accuracy_score from peft import LoraConfig, TaskType, get_peft_model, PeftModel, get_peft_model_state_dict from peft.tuners.lora import LoraLayer from transformers.trainer_utils import PREFIX_CHECKPOINT_DIR The provided code snippet includes necessary dependencies for implementing the `prepare_model_for_kbit_training` function. Write a Python function `def prepare_model_for_kbit_training(model, use_gradient_checkpointing=True)` to solve the following problem: r""" This method wraps the entire protocol for preparing a model before running a training. This includes: 1- Cast the layernorm in fp32 2- making output embedding layer require grads 3- Add the upcasting of the lm head to fp32 Args: model, (`transformers.PreTrainedModel`): The loaded model from `transformers` Here is the function: def prepare_model_for_kbit_training(model, use_gradient_checkpointing=True): r""" This method wraps the entire protocol for preparing a model before running a training. This includes: 1- Cast the layernorm in fp32 2- making output embedding layer require grads 3- Add the upcasting of the lm head to fp32 Args: model, (`transformers.PreTrainedModel`): The loaded model from `transformers` """ loaded_in_kbit = getattr(model, "is_loaded_in_8bit", False) or getattr(model, "is_loaded_in_4bit", False) for name, param in model.named_parameters(): # freeze base model's layers param.requires_grad = False # cast all non INT8/INT4 parameters to fp32 for param in model.parameters(): if ((param.dtype == torch.float16) or (param.dtype == torch.bfloat16)) and loaded_in_kbit: param.data = param.data.to(torch.float32) for name, module in model.named_modules(): if 'norm' in name: module = module.to(torch.float32) if loaded_in_kbit and use_gradient_checkpointing: # For backward compatibility if hasattr(model, "enable_input_require_grads"): model.enable_input_require_grads() else: def make_inputs_require_grad(module, _input, output): output.requires_grad_(True) model.get_input_embeddings().register_forward_hook(make_inputs_require_grad) # enable gradient checkpointing for memory efficiency model.gradient_checkpointing_enable() return model
r""" This method wraps the entire protocol for preparing a model before running a training. This includes: 1- Cast the layernorm in fp32 2- making output embedding layer require grads 3- Add the upcasting of the lm head to fp32 Args: model, (`transformers.PreTrainedModel`): The loaded model from `transformers`
179,359
import logging import numpy as np import math import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, List, Dict, Any, Mapping from pathlib import Path import datasets import torch from datasets import load_dataset, concatenate_datasets import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_CAUSAL_LM_MAPPING, AutoConfig, AutoModelForCausalLM, LlamaForCausalLM, LlamaTokenizer, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, is_torch_tpu_available, set_seed, BitsAndBytesConfig ) from transformers.testing_utils import CaptureLogger from transformers.trainer_utils import get_last_checkpoint from transformers.utils import send_example_telemetry from transformers.utils.versions import require_version from sklearn.metrics import accuracy_score from peft import LoraConfig, TaskType, get_peft_model, PeftModel, get_peft_model_state_dict from peft.tuners.lora import LoraLayer from transformers.trainer_utils import PREFIX_CHECKPOINT_DIR def accuracy(predictions, references, normalize=True, sample_weight=None): return { "accuracy": float( accuracy_score(references, predictions, normalize=normalize, sample_weight=sample_weight) ) } def compute_metrics(eval_preds): preds, labels = eval_preds # preds have the same shape as the labels, after the argmax(-1) has been calculated # by preprocess_logits_for_metrics but we need to shift the labels labels = labels[:, 1:].reshape(-1) preds = preds[:, :-1].reshape(-1) return accuracy(predictions=preds, references=labels)
null
179,360
import logging import numpy as np import math import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, List, Dict, Any, Mapping from pathlib import Path import datasets import torch from datasets import load_dataset, concatenate_datasets import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_CAUSAL_LM_MAPPING, AutoConfig, AutoModelForCausalLM, LlamaForCausalLM, LlamaTokenizer, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, is_torch_tpu_available, set_seed, BitsAndBytesConfig ) from transformers.testing_utils import CaptureLogger from transformers.trainer_utils import get_last_checkpoint from transformers.utils import send_example_telemetry from transformers.utils.versions import require_version from sklearn.metrics import accuracy_score from peft import LoraConfig, TaskType, get_peft_model, PeftModel, get_peft_model_state_dict from peft.tuners.lora import LoraLayer from transformers.trainer_utils import PREFIX_CHECKPOINT_DIR def preprocess_logits_for_metrics(logits, labels): if isinstance(logits, tuple): # Depending on the model and config, logits may contain extra tensors, # like past_key_values, but logits always come first logits = logits[0] return logits.argmax(dim=-1)
null
179,361
import logging import numpy as np import math import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, List, Dict, Any, Mapping from pathlib import Path import datasets import torch from datasets import load_dataset, concatenate_datasets import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_CAUSAL_LM_MAPPING, AutoConfig, AutoModelForCausalLM, LlamaForCausalLM, LlamaTokenizer, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, is_torch_tpu_available, set_seed, BitsAndBytesConfig ) from transformers.testing_utils import CaptureLogger from transformers.trainer_utils import get_last_checkpoint from transformers.utils import send_example_telemetry from transformers.utils.versions import require_version from sklearn.metrics import accuracy_score from peft import LoraConfig, TaskType, get_peft_model, PeftModel, get_peft_model_state_dict from peft.tuners.lora import LoraLayer from transformers.trainer_utils import PREFIX_CHECKPOINT_DIR def fault_tolerance_data_collator(features: List) -> Dict[str, Any]: if not isinstance(features[0], Mapping): features = [vars(f) for f in features] first = features[0] batch = {} # Special handling for labels. # Ensure that tensor is created with the correct type # (it should be automatically the case, but let's make sure of it.) if "label" in first and first["label"] is not None: label = first["label"].item() if isinstance(first["label"], torch.Tensor) else first["label"] dtype = torch.long if isinstance(label, int) else torch.float batch["labels"] = torch.tensor([f["label"] for f in features], dtype=dtype) elif "label_ids" in first and first["label_ids"] is not None: if isinstance(first["label_ids"], torch.Tensor): batch["labels"] = torch.stack([f["label_ids"] for f in features]) else: dtype = torch.long if isinstance(first["label_ids"][0], int) else torch.float batch["labels"] = torch.tensor([f["label_ids"] for f in features], dtype=dtype) # Handling of all other possible keys. # Again, we will use the first element to figure out which key/values are not None for this model. try: for k, v in first.items(): if k not in ("label", "label_ids") and v is not None and not isinstance(v, str): if isinstance(v, torch.Tensor): batch[k] = torch.stack([f[k] for f in features]) elif isinstance(v, np.ndarray): batch[k] = torch.tensor(np.stack([f[k] for f in features])) else: batch[k] = torch.tensor([f[k] for f in features]) except ValueError: # quick fix by simply take the first example for k, v in first.items(): if k not in ("label", "label_ids") and v is not None and not isinstance(v, str): if isinstance(v, torch.Tensor): batch[k] = torch.stack([features[0][k]] * len(features)) elif isinstance(v, np.ndarray): batch[k] = torch.tensor(np.stack([features[0][k]] * len(features))) else: batch[k] = torch.tensor([features[0][k]] * len(features)) return batch
null
179,364
from .peft_model import ( PeftModel, PeftModelForCausalLM, PeftModelForSeq2SeqLM, PeftModelForSequenceClassification, PeftModelForTokenClassification, ) from .tuners import LoraConfig, PrefixTuningConfig, PromptEncoderConfig, PromptTuningConfig from .utils import PromptLearningConfig PEFT_TYPE_TO_CONFIG_MAPPING = { "PROMPT_TUNING": PromptTuningConfig, "PREFIX_TUNING": PrefixTuningConfig, "P_TUNING": PromptEncoderConfig, "LORA": LoraConfig, } The provided code snippet includes necessary dependencies for implementing the `get_peft_config` function. Write a Python function `def get_peft_config(config_dict)` to solve the following problem: Returns a Peft config object from a dictionary. Args: config_dict (`Dict[str, Any]`): Dictionary containing the configuration parameters. Here is the function: def get_peft_config(config_dict): """ Returns a Peft config object from a dictionary. Args: config_dict (`Dict[str, Any]`): Dictionary containing the configuration parameters. """ return PEFT_TYPE_TO_CONFIG_MAPPING[config_dict["peft_type"]](**config_dict)
Returns a Peft config object from a dictionary. Args: config_dict (`Dict[str, Any]`): Dictionary containing the configuration parameters.
179,365
from .peft_model import ( PeftModel, PeftModelForCausalLM, PeftModelForSeq2SeqLM, PeftModelForSequenceClassification, PeftModelForTokenClassification, ) from .tuners import LoraConfig, PrefixTuningConfig, PromptEncoderConfig, PromptTuningConfig from .utils import PromptLearningConfig MODEL_TYPE_TO_PEFT_MODEL_MAPPING = { "SEQ_CLS": PeftModelForSequenceClassification, "SEQ_2_SEQ_LM": PeftModelForSeq2SeqLM, "CAUSAL_LM": PeftModelForCausalLM, "TOKEN_CLS": PeftModelForTokenClassification, } def _prepare_prompt_learning_config(peft_config, model_config): if peft_config.num_layers is None: if "num_hidden_layers" in model_config: num_layers = model_config["num_hidden_layers"] elif "num_layers" in model_config: num_layers = model_config["num_layers"] elif "n_layer" in model_config: num_layers = model_config["n_layer"] else: raise ValueError("Please specify `num_layers` in `peft_config`") peft_config.num_layers = num_layers if peft_config.token_dim is None: if "hidden_size" in model_config: token_dim = model_config["hidden_size"] elif "n_embd" in model_config: token_dim = model_config["n_embd"] elif "d_model" in model_config: token_dim = model_config["d_model"] else: raise ValueError("Please specify `token_dim` in `peft_config`") peft_config.token_dim = token_dim if peft_config.num_attention_heads is None: if "num_attention_heads" in model_config: num_attention_heads = model_config["num_attention_heads"] elif "n_head" in model_config: num_attention_heads = model_config["n_head"] elif "num_heads" in model_config: num_attention_heads = model_config["num_heads"] elif "encoder_attention_heads" in model_config: num_attention_heads = model_config["encoder_attention_heads"] else: raise ValueError("Please specify `num_attention_heads` in `peft_config`") peft_config.num_attention_heads = num_attention_heads if getattr(peft_config, "encoder_hidden_size", None) is None: setattr(peft_config, "encoder_hidden_size", token_dim) return peft_config def _prepare_lora_config(peft_config, model_config): if peft_config.target_modules is None: if model_config["model_type"] not in TRANSFORMERS_MODELS_TO_LORA_TARGET_MODULES_MAPPING: raise ValueError("Please specify `target_modules` in `peft_config`") peft_config.target_modules = TRANSFORMERS_MODELS_TO_LORA_TARGET_MODULES_MAPPING[model_config["model_type"]] if len(peft_config.target_modules) == 1: peft_config.fan_in_fan_out = True peft_config.enable_lora = [True, False, True] if peft_config.inference_mode: peft_config.merge_weights = True return peft_config class PeftModel(PushToHubMixin, torch.nn.Module): """ Parameter-Efficient Fine-Tuning Model. Base model encompassing various Peft methods. Args: model ([`PreTrainedModel`]): The base transformer model used for Peft. peft_config ([`PeftConfig`]): The configuration of the Peft model. **Attributes**: - **base_model** ([`PreTrainedModel`]) -- The base transformer model used for Peft. - **peft_config** ([`PeftConfig`]) -- The configuration of the Peft model. - **modules_to_save** (`list` of `str`) -- The list of sub-module names to save when saving the model. - **prompt_encoder** ([`PromptEncoder`]) -- The prompt encoder used for Peft if `isinstance(self.peft_config, PromptLearningConfig)`. - **prompt_tokens** (`torch.Tensor`) -- The virtual prompt tokens used for Peft if `isinstance(self.peft_config, PromptLearningConfig)`. - **transformer_backbone_name** (`str`) -- The name of the transformer backbone in the base model if `isinstance(self.peft_config, PromptLearningConfig)`. - **word_embeddings** (`torch.nn.Embedding`) -- The word embeddings of the transformer backbone in the base model if `isinstance(self.peft_config, PromptLearningConfig)`. """ def __init__(self, model, peft_config: PeftConfig): super().__init__() self.peft_config = peft_config self.base_model = model self.config = self.base_model.config self.modules_to_save = None if isinstance(self.peft_config, PromptLearningConfig): self._setup_prompt_encoder() else: self.base_model = LoraModel(peft_config, model) if getattr(self.peft_config, "modules_to_save", None) is not None: self.modules_to_save = self.peft_config.modules_to_save _set_trainable(self) self.device = torch.device("cuda" if torch.cuda.is_available() else "cpu") def save_pretrained(self, save_directory, **kwargs): r""" Args: This function saves the adapter model and the adapter configuration files to a directory, so that it can be re-loaded using the `LoraModel.from_pretrained` class method, and also used by the `LoraModel.push_to_hub` method. save_directory (`str`): Directory where the adapter model and configuration files will be saved (will be created if it does not exist). **kwargs: Additional keyword arguments passed along to the `push_to_hub` method. """ if os.path.isfile(save_directory): raise ValueError(f"Provided path ({save_directory}) should be a directory, not a file") os.makedirs(save_directory, exist_ok=True) # save only the trainable weights output_state_dict = get_peft_model_state_dict(self, kwargs.get("state_dict", None)) torch.save(output_state_dict, os.path.join(save_directory, WEIGHTS_NAME)) # save the config and change the inference mode to `True` if self.peft_config.base_model_name_or_path is None: self.peft_config.base_model_name_or_path = ( self.base_model.__dict__.get("name_or_path", None) if isinstance(self.peft_config, PromptLearningConfig) else self.base_model.model.__dict__.get("name_or_path", None) ) inference_mode = self.peft_config.inference_mode self.peft_config.inference_mode = True self.peft_config.save_pretrained(save_directory) self.peft_config.inference_mode = inference_mode def from_pretrained(cls, model, model_id, **kwargs): r""" Args: Instantiate a `LoraModel` from a pretrained Lora configuration and weights. model (`transformers.PreTrainedModel`): The model to be adapted. The model should be initialized with the `from_pretrained` method. from `transformers` library. model_id (`str`): The name of the Lora configuration to use. Can be either: - A string, the `model id` of a Lora configuration hosted inside a model repo on huggingface Hub - A path to a directory containing a Lora configuration file saved using the `save_pretrained` method, e.g., ``./my_lora_config_directory/``. """ from .mapping import MODEL_TYPE_TO_PEFT_MODEL_MAPPING, PEFT_TYPE_TO_CONFIG_MAPPING # load the config config = PEFT_TYPE_TO_CONFIG_MAPPING[PeftConfig.from_pretrained(model_id).peft_type].from_pretrained(model_id) if getattr(model, "hf_device_map", None) is not None: remove_hook_from_submodules(model) if config.task_type not in MODEL_TYPE_TO_PEFT_MODEL_MAPPING.keys(): model = cls(model, config) else: model = MODEL_TYPE_TO_PEFT_MODEL_MAPPING[config.task_type](model, config) # load weights if any if os.path.exists(os.path.join(model_id, WEIGHTS_NAME)): filename = os.path.join(model_id, WEIGHTS_NAME) else: try: filename = hf_hub_download(model_id, WEIGHTS_NAME) except: # noqa raise ValueError( f"Can't find weights for {model_id} in {model_id} or in the Hugging Face Hub. " f"Please check that the file {WEIGHTS_NAME} is present at {model_id}." ) adapters_weights = torch.load( filename, map_location=torch.device("cuda" if torch.cuda.is_available() else "cpu") ) # load the weights into the model model = set_peft_model_state_dict(model, adapters_weights) if getattr(model, "hf_device_map", None) is not None: device_map = kwargs.get("device_map", "auto") max_memory = kwargs.get("max_memory", None) no_split_module_classes = model._no_split_modules if device_map != "sequential": max_memory = get_balanced_memory( model, max_memory=max_memory, no_split_module_classes=no_split_module_classes, low_zero=(device_map == "balanced_low_0"), ) if isinstance(device_map, str): device_map = infer_auto_device_map( model, max_memory=max_memory, no_split_module_classes=no_split_module_classes ) model = dispatch_model(model, device_map=device_map) hook = AlignDevicesHook(io_same_device=True) if model.peft_config.peft_type == PeftType.LORA: add_hook_to_module(model.base_model.model, hook) else: remove_hook_from_submodules(model.prompt_encoder) add_hook_to_module(model.base_model, hook) return model def _setup_prompt_encoder(self): transformer_backbone = None for name, module in self.base_model.named_children(): for param in module.parameters(): param.requires_grad = False if isinstance(module, PreTrainedModel): # Make sure to freeze Tranformers model if transformer_backbone is None: transformer_backbone = module self.transformer_backbone_name = name if self.peft_config.num_transformer_submodules is None: self.peft_config.num_transformer_submodules = ( 2 if self.peft_config.task_type == TaskType.SEQ_2_SEQ_LM else 1 ) for named_param, value in list(transformer_backbone.named_parameters()): if value.shape[0] == self.base_model.config.vocab_size: self.word_embeddings = transformer_backbone.get_submodule(named_param.replace(".weight", "")) break if self.peft_config.peft_type == PeftType.PROMPT_TUNING: prompt_encoder = PromptEmbedding(self.peft_config, self.word_embeddings) elif self.peft_config.peft_type == PeftType.P_TUNING: prompt_encoder = PromptEncoder(self.peft_config) elif self.peft_config.peft_type == PeftType.PREFIX_TUNING: prompt_encoder = PrefixEncoder(self.peft_config) else: raise ValueError("Not supported") self.prompt_encoder = prompt_encoder self.prompt_tokens = torch.arange( self.peft_config.num_virtual_tokens * self.peft_config.num_transformer_submodules ).long() def get_prompt_embedding_to_save(self): """ Returns the prompt embedding to save when saving the model. Only applicable when `peft_config.peft_type != PeftType.LORA`. """ prompt_tokens = self.prompt_tokens.unsqueeze(0).expand(1, -1).to(self.device) if self.peft_config.peft_type == PeftType.PREFIX_TUNING: prompt_tokens = prompt_tokens[:, : self.peft_config.num_virtual_tokens] prompt_embeddings = self.prompt_encoder(prompt_tokens) return prompt_embeddings[0].detach().cpu() def get_prompt(self, batch_size): """ Returns the virtual prompts to use for Peft. Only applicable when `peft_config.peft_type != PeftType.LORA`. """ prompt_tokens = self.prompt_tokens.unsqueeze(0).expand(batch_size, -1).to(self.device) if self.peft_config.peft_type == PeftType.PREFIX_TUNING: prompt_tokens = prompt_tokens[:, : self.peft_config.num_virtual_tokens] if self.peft_config.inference_mode: past_key_values = self.prompt_encoder.embedding.weight.repeat(batch_size, 1, 1) else: past_key_values = self.prompt_encoder(prompt_tokens) past_key_values = past_key_values.view( batch_size, self.peft_config.num_virtual_tokens, self.peft_config.num_layers * 2, self.peft_config.num_attention_heads, self.peft_config.token_dim // self.peft_config.num_attention_heads, ) if self.peft_config.num_transformer_submodules == 2: past_key_values = torch.cat([past_key_values, past_key_values], dim=2) past_key_values = past_key_values.permute([2, 0, 3, 1, 4]).split( self.peft_config.num_transformer_submodules * 2 ) if TRANSFORMERS_MODELS_TO_PREFIX_TUNING_POSTPROCESS_MAPPING.get(self.config.model_type, None) is not None: post_process_fn = TRANSFORMERS_MODELS_TO_PREFIX_TUNING_POSTPROCESS_MAPPING[self.config.model_type] past_key_values = post_process_fn(past_key_values) return past_key_values else: if self.peft_config.inference_mode: prompts = self.prompt_encoder.embedding.weight.repeat(batch_size, 1, 1) else: prompts = self.prompt_encoder(prompt_tokens) return prompts def print_trainable_parameters(self): """ Prints the number of trainable parameters in the model. """ trainable_params = 0 all_param = 0 for _, param in self.named_parameters(): num_params = param.numel() # if using DS Zero 3 and the weights are initialized empty if num_params == 0 and hasattr(param, "ds_numel"): num_params = param.ds_numel all_param += num_params if param.requires_grad: trainable_params += num_params print( f"trainable params: {trainable_params} || all params: {all_param} || trainable%: {100 * trainable_params / all_param}" ) def __getattr__(self, name: str): """Forward missing attributes to the wrapped module.""" try: return super().__getattr__(name) # defer to nn.Module's logic except AttributeError: return getattr(self.base_model, name) def forward(self, *args, **kwargs): # pylint: disable=E0202 """ Forward pass of the model. """ return self.get_base_model()(*args, **kwargs) def disable_adapter(self): """ Disables the adapter module. """ if isinstance(self.peft_config, PromptLearningConfig): old_forward = self.forward self.forward = self.base_model.forward else: self.base_model.disable_adapter_layers() yield if isinstance(self.peft_config, PromptLearningConfig): self.forward = old_forward else: self.base_model.enable_adapter_layers() def get_base_model(self): """ Returns the base model. """ return self.base_model if isinstance(self.peft_config, PromptLearningConfig) else self.base_model.model The provided code snippet includes necessary dependencies for implementing the `get_peft_model` function. Write a Python function `def get_peft_model(model, peft_config)` to solve the following problem: Returns a Peft model object from a model and a config. Args: model ([`transformers.PreTrainedModel`]): Model to be wrapped. peft_config ([`PeftConfig`]): Configuration object containing the parameters of the Peft model. Here is the function: def get_peft_model(model, peft_config): """ Returns a Peft model object from a model and a config. Args: model ([`transformers.PreTrainedModel`]): Model to be wrapped. peft_config ([`PeftConfig`]): Configuration object containing the parameters of the Peft model. """ model_config = model.config.to_dict() peft_config.base_model_name_or_path = model.__dict__.get("name_or_path", None) if peft_config.task_type not in MODEL_TYPE_TO_PEFT_MODEL_MAPPING.keys(): peft_config = _prepare_lora_config(peft_config, model_config) return PeftModel(model, peft_config) if not isinstance(peft_config, PromptLearningConfig): peft_config = _prepare_lora_config(peft_config, model_config) else: peft_config = _prepare_prompt_learning_config(peft_config, model_config) return MODEL_TYPE_TO_PEFT_MODEL_MAPPING[peft_config.task_type](model, peft_config)
Returns a Peft model object from a model and a config. Args: model ([`transformers.PreTrainedModel`]): Model to be wrapped. peft_config ([`PeftConfig`]): Configuration object containing the parameters of the Peft model.
179,366
from .config import PeftType class PeftType(str, enum.Enum): PROMPT_TUNING = "PROMPT_TUNING" P_TUNING = "P_TUNING" PREFIX_TUNING = "PREFIX_TUNING" LORA = "LORA" The provided code snippet includes necessary dependencies for implementing the `get_peft_model_state_dict` function. Write a Python function `def get_peft_model_state_dict(model, state_dict=None)` to solve the following problem: Get the state dict of the Peft model. Args: model ([`PeftModel`]): The Peft model. When using torch.nn.DistributedDataParallel, DeepSpeed or FSDP, the model should be the underlying model/unwrapped model (i.e. model.module). state_dict (`dict`, *optional*, defaults to `None`): The state dict of the model. If not provided, the state dict of the model will be used. Here is the function: def get_peft_model_state_dict(model, state_dict=None): """ Get the state dict of the Peft model. Args: model ([`PeftModel`]): The Peft model. When using torch.nn.DistributedDataParallel, DeepSpeed or FSDP, the model should be the underlying model/unwrapped model (i.e. model.module). state_dict (`dict`, *optional*, defaults to `None`): The state dict of the model. If not provided, the state dict of the model will be used. """ if state_dict is None: state_dict = model.state_dict() if model.peft_config.peft_type == PeftType.LORA: # to_return = lora_state_dict(model, bias=model.peft_config.bias) # adapted from `https://github.com/microsoft/LoRA/blob/main/loralib/utils.py` # to directly with the state dict which is necessary when using DeepSpeed or FSDP bias = model.peft_config.bias if bias == "none": to_return = {k: state_dict[k] for k in state_dict if "lora_" in k} elif bias == "all": to_return = {k: state_dict[k] for k in state_dict if "lora_" in k or "bias" in k} elif bias == "lora_only": to_return = {} for k in state_dict: if "lora_" in k: to_return[k] = state_dict[k] bias_name = k.split("lora_")[0] + "bias" if bias_name in state_dict: to_return[bias_name] = state_dict[bias_name] else: raise NotImplementedError else: to_return = {} if model.peft_config.inference_mode: prompt_embeddings = model.prompt_encoder.embedding.weight else: prompt_embeddings = model.get_prompt_embedding_to_save() to_return["prompt_embeddings"] = prompt_embeddings if model.modules_to_save is not None: for key, value in state_dict.items(): if any(module_name in key for module_name in model.modules_to_save): to_return[key] = value return to_return
Get the state dict of the Peft model. Args: model ([`PeftModel`]): The Peft model. When using torch.nn.DistributedDataParallel, DeepSpeed or FSDP, the model should be the underlying model/unwrapped model (i.e. model.module). state_dict (`dict`, *optional*, defaults to `None`): The state dict of the model. If not provided, the state dict of the model will be used.
179,367
from .config import PeftType class PeftType(str, enum.Enum): PROMPT_TUNING = "PROMPT_TUNING" P_TUNING = "P_TUNING" PREFIX_TUNING = "PREFIX_TUNING" LORA = "LORA" The provided code snippet includes necessary dependencies for implementing the `set_peft_model_state_dict` function. Write a Python function `def set_peft_model_state_dict(model, peft_model_state_dict)` to solve the following problem: Set the state dict of the Peft model. Args: model ([`PeftModel`]): The Peft model. peft_model_state_dict (`dict`): The state dict of the Peft model. Here is the function: def set_peft_model_state_dict(model, peft_model_state_dict): """ Set the state dict of the Peft model. Args: model ([`PeftModel`]): The Peft model. peft_model_state_dict (`dict`): The state dict of the Peft model. """ model.load_state_dict(peft_model_state_dict, strict=False) if model.peft_config.peft_type != PeftType.LORA: model.prompt_encoder.embedding.load_state_dict( {"weight": peft_model_state_dict["prompt_embeddings"]}, strict=True ) return model
Set the state dict of the Peft model. Args: model ([`PeftModel`]): The Peft model. peft_model_state_dict (`dict`): The state dict of the Peft model.