| import importlib | |
| import math | |
| from typing import TYPE_CHECKING, Optional, Tuple, Union, Callable, List, Any, Generator | |
| import torch | |
| import torch.nn.functional as F | |
| import torch.utils.checkpoint | |
| from torch.cuda.amp import autocast | |
| from torch.nn import CrossEntropyLoss | |
| from transformers import PreTrainedTokenizer, GenerationConfig, StoppingCriteriaList | |
| from transformers.generation.logits_process import LogitsProcessorList | |
| if TYPE_CHECKING: | |
| from transformers.generation.streamers import BaseStreamer | |
| from transformers.generation.utils import GenerateOutput | |
| from transformers.modeling_outputs import ( | |
| BaseModelOutputWithPast, | |
| CausalLMOutputWithPast, | |
| ) | |
| from transformers.modeling_utils import PreTrainedModel | |
| from transformers.utils import logging | |
| from peft import LoraConfig, get_peft_model, prepare_model_for_kbit_training | |
| try: | |
| from einops import rearrange | |
| except ImportError: | |
| rearrange = None | |
| from torch import nn | |
| from .modeling_qwen import QWenModel,QWenPreTrainedModel,QWenLMHeadModel | |
| SUPPORT_CUDA = torch.cuda.is_available() | |
| SUPPORT_BF16 = SUPPORT_CUDA and torch.cuda.is_bf16_supported() | |
| SUPPORT_FP16 = SUPPORT_CUDA and torch.cuda.get_device_capability(0)[0] >= 7 | |
| logger = logging.get_logger(__name__) | |
| class MonkeyModel(QWenModel): | |
| def __init__(self, config): | |
| super().__init__(config) | |
| def forward( | |
| self, | |
| input_ids: Optional[torch.LongTensor] = None, | |
| past_key_values: Optional[Tuple[Tuple[torch.Tensor]]] = None, | |
| attention_mask: Optional[torch.FloatTensor] = None, | |
| token_type_ids: Optional[torch.LongTensor] = None, | |
| position_ids: Optional[torch.LongTensor] = None, | |
| head_mask: Optional[torch.FloatTensor] = None, | |
| inputs_embeds: Optional[torch.FloatTensor] = None, | |
| encoder_hidden_states: Optional[torch.Tensor] = None, | |
| encoder_attention_mask: Optional[torch.FloatTensor] = None, | |
| use_cache: Optional[bool] = None, | |
| output_attentions: Optional[bool] = None, | |
| output_hidden_states: Optional[bool] = None, | |
| return_dict: Optional[bool] = None, | |
| ): | |
| if past_key_values is None and torch.any(input_ids == self.config.visual['image_start_id']): | |
| bos_pos = torch.where(input_ids == self.config.visual['image_start_id']) | |
| eos_pos = torch.where(input_ids == self.config.visual['image_start_id'] + 1) | |
| assert (bos_pos[0] == eos_pos[0]).all() | |
| img_pos = torch.stack((bos_pos[0], bos_pos[1], eos_pos[1]), dim=1) | |
| images = [] | |
| for i, a, b in img_pos: | |
| image = input_ids[i][a + 1 : b - 1].tolist() | |
| image = image[ : image.index(self.config.visual['image_start_id'] + 2)] | |
| images.append(bytes(image).decode('utf-8')) | |
| windows,images_448 = self.visual.encode(images) | |
| patch_list = [] | |
| lora_idx = 0 | |
| for col in windows: | |
| for image_patch in col: | |
| patch_list.append(self.visual(image_patch,idx=lora_idx)) | |
| lora_idx += 1 | |
| global_feat = self.visual(images_448) | |
| local_feat = torch.cat(patch_list,dim=1) | |
| images = torch.cat([local_feat,global_feat],dim=1) | |
| assert images.shape[0] == len(images) | |
| else: | |
| images = None | |
| return super().forward(input_ids, | |
| past_key_values, | |
| attention_mask, | |
| token_type_ids, | |
| position_ids, | |
| head_mask,inputs_embeds, | |
| encoder_hidden_states, | |
| encoder_attention_mask, | |
| use_cache, | |
| output_attentions, | |
| output_hidden_states, | |
| return_dict, | |
| images) | |
| class MonkeyLMHeadModel(QWenLMHeadModel): | |
| _keys_to_ignore_on_load_missing = [r"h\.\d+\.attn\.rotary_emb\.inv_freq"] | |
| _keys_to_ignore_on_load_unexpected = [r"h\.\d+\.attn\.masked_bias"] | |
| def __init__(self, config): | |
| super().__init__(config) | |
| assert ( | |
| config.bf16 + config.fp16 + config.fp32 <= 1 | |
| ), "Only one of \"bf16\", \"fp16\", \"fp32\" can be true" | |
| autoset_precision = config.bf16 + config.fp16 + config.fp32 == 0 | |
| if autoset_precision: | |
| if SUPPORT_BF16: | |
| logger.warn( | |
| "The model is automatically converting to bf16 for faster inference. " | |
| "If you want to disable the automatic precision, please manually add bf16/fp16/fp32=True to \"AutoModelForCausalLM.from_pretrained\"." | |
| ) | |
| config.bf16 = True | |
| elif SUPPORT_FP16: | |
| logger.warn( | |
| "The model is automatically converting to fp16 for faster inference. " | |
| "If you want to disable the automatic precision, please manually add bf16/fp16/fp32=True to \"AutoModelForCausalLM.from_pretrained\"." | |
| ) | |
| config.fp16 = True | |
| else: | |
| config.fp32 = True | |
| if config.bf16 and SUPPORT_CUDA and not SUPPORT_BF16: | |
| logger.warn("Your device does NOT seem to support bf16, you can switch to fp16 or fp32 by by passing fp16/fp32=True in \"AutoModelForCausalLM.from_pretrained\".") | |
| if config.fp16 and SUPPORT_CUDA and not SUPPORT_FP16: | |
| logger.warn("Your device does NOT support faster inference with fp16, please switch to fp32 which is likely to be faster") | |
| if config.fp32: | |
| if SUPPORT_BF16: | |
| logger.warn("Your device support faster inference by passing bf16=True in \"AutoModelForCausalLM.from_pretrained\".") | |
| elif SUPPORT_FP16: | |
| logger.warn("Your device support faster inference by passing fp16=True in \"AutoModelForCausalLM.from_pretrained\".") | |
| self.transformer = MonkeyModel(config) | |
| self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False) | |
| if config.bf16: | |
| self.transformer.bfloat16() | |
| self.lm_head.bfloat16() | |
| if config.fp16: | |
| self.transformer.half() | |
| self.lm_head.half() | |
| self.post_init() | |