| | from transformers.models.qwen2_vl.modeling_qwen2_vl import Qwen2VLModelOutputWithPast |
| | from transformers.models.qwen2_5_vl.modeling_qwen2_5_vl import Qwen2_5_VLModelOutputWithPast |
| | from transformers.models.qwen3_vl.modeling_qwen3_vl import Qwen3VLModelOutputWithPast |
| | from transformers.models.qwen3_vl_moe.modeling_qwen3_vl_moe import Qwen3VLMoeModelOutputWithPast |
| | import torch |
| | from typing import Optional, List, Union, Tuple |
| | import transformers.models.qwen2_vl.modeling_qwen2_vl |
| | import transformers.models.qwen2_5_vl.modeling_qwen2_5_vl |
| | import transformers.models.qwen3_vl_moe.modeling_qwen3_vl_moe |
| | from transformers.utils import TransformersKwargs |
| | from transformers.processing_utils import Unpack |
| | from transformers.cache_utils import Cache |
| | from transformers.utils import is_torchdynamo_compiling |
| |
|
| | def replace_qwen_2_with_mixed_modality_forward(): |
| | transformers.models.qwen2_vl.modeling_qwen2_vl.Qwen2VLModel.forward = qwen2_mixed_modality_forward |
| |
|
| | def replace_qwen2_5_with_mixed_modality_forward(): |
| | transformers.models.qwen2_5_vl.modeling_qwen2_5_vl.Qwen2_5_VLModel.forward = qwen2_5_mixed_modality_forward |
| |
|
| | def replace_qwen3_with_mixed_modality_forward(): |
| | transformers.models.qwen3_vl.modeling_qwen3_vl.Qwen3VLModel.forward = qwen3_vl_mixed_modality_forward |
| |
|
| | def replace_qwen3_vl_moe_with_mixed_modality_forward(): |
| | transformers.models.qwen3_vl_moe.modeling_qwen3_vl_moe.Qwen3VLMoeModel.forward = qwen3_vl_moe_mixed_modality_forward |
| |
|
| | def qwen3_vl_moe_mixed_modality_forward( |
| | self, |
| | input_ids: torch.LongTensor = None, |
| | attention_mask: Optional[torch.Tensor] = None, |
| | position_ids: Optional[torch.LongTensor] = None, |
| | past_key_values: Optional[Cache] = None, |
| | inputs_embeds: Optional[torch.FloatTensor] = None, |
| | pixel_values: Optional[torch.Tensor] = None, |
| | pixel_values_videos: Optional[torch.FloatTensor] = None, |
| | image_grid_thw: Optional[torch.LongTensor] = None, |
| | video_grid_thw: Optional[torch.LongTensor] = None, |
| | cache_position: Optional[torch.LongTensor] = None, |
| | second_per_grid_ts: Optional[torch.Tensor] = None, |
| | **kwargs: Unpack[TransformersKwargs], |
| | ) -> Union[tuple, Qwen3VLMoeModelOutputWithPast]: |
| | |
| | if (input_ids is None) ^ (inputs_embeds is not None): |
| | raise ValueError("You must specify exactly one of input_ids or inputs_embeds") |
| |
|
| | if inputs_embeds is None: |
| | inputs_embeds = self.get_input_embeddings()(input_ids) |
| |
|
| | image_mask = None |
| | video_mask = None |
| | |
| | if pixel_values is None and pixel_values_videos is None: |
| | |
| | dummy_pixel = torch.zeros(1024, 1536).to(self.visual.device) |
| | dummy_grid = torch.tensor([[1, 32, 32]]).to(self.visual.device) |
| |
|
| | image_embeds, dummy_deepstack = self.get_image_features(dummy_pixel, dummy_grid) |
| | image_embeds = torch.cat(image_embeds, dim=0).to(inputs_embeds.device, inputs_embeds.dtype) |
| | |
| | inputs_embeds += image_embeds.mean() * 0 |
| |
|
| | if pixel_values is not None: |
| | image_embeds, deepstack_image_embeds = self.get_image_features(pixel_values, image_grid_thw) |
| | image_embeds = torch.cat(image_embeds, dim=0).to(inputs_embeds.device, inputs_embeds.dtype) |
| | image_mask, _ = self.get_placeholder_mask( |
| | input_ids, inputs_embeds=inputs_embeds, image_features=image_embeds |
| | ) |
| | inputs_embeds = inputs_embeds.masked_scatter(image_mask, image_embeds) |
| |
|
| | if pixel_values_videos is not None: |
| | video_embeds, deepstack_video_embeds = self.get_video_features(pixel_values_videos, video_grid_thw) |
| | video_embeds = torch.cat(video_embeds, dim=0).to(inputs_embeds.device, inputs_embeds.dtype) |
| | _, video_mask = self.get_placeholder_mask( |
| | input_ids, inputs_embeds=inputs_embeds, video_features=video_embeds |
| | ) |
| | inputs_embeds = inputs_embeds.masked_scatter(video_mask, video_embeds) |
| |
|
| | visual_pos_masks = None |
| | deepstack_visual_embeds = None |
| | if image_mask is not None and video_mask is not None: |
| | |
| | image_mask = image_mask[..., 0] |
| | video_mask = video_mask[..., 0] |
| | visual_pos_masks = image_mask | video_mask |
| | deepstack_visual_embeds = [] |
| | image_mask_joint = image_mask[visual_pos_masks] |
| | video_mask_joint = video_mask[visual_pos_masks] |
| | for img_embed, vid_embed in zip(deepstack_image_embeds, deepstack_video_embeds): |
| | embed_joint = img_embed.new_zeros(visual_pos_masks.sum(), img_embed.shape[-1]).to(img_embed.device) |
| | embed_joint[image_mask_joint, :] = img_embed |
| | embed_joint[video_mask_joint, :] = vid_embed |
| | deepstack_visual_embeds.append(embed_joint) |
| | elif image_mask is not None: |
| | image_mask = image_mask[..., 0] |
| | visual_pos_masks = image_mask |
| | deepstack_visual_embeds = deepstack_image_embeds |
| | elif video_mask is not None: |
| | video_mask = video_mask[..., 0] |
| | visual_pos_masks = video_mask |
| | deepstack_visual_embeds = deepstack_video_embeds |
| |
|
| | if visual_pos_masks is None: |
| | B, S, H = inputs_embeds.shape |
| | visual_pos_masks = torch.zeros((B, S), dtype=torch.bool, device=inputs_embeds.device) |
| | L = len(self.visual.deepstack_visual_indexes) |
| | deepstack_visual_embeds = [t.narrow(0, 0, 0) for t in dummy_deepstack] |
| |
|
| | if position_ids is None: |
| | attention_mask_tensor = ( |
| | attention_mask if not isinstance(attention_mask, dict) else attention_mask["full_attention"] |
| | ) |
| | if attention_mask_tensor is not None and attention_mask_tensor.ndim == 4: |
| | attention_mask_tensor = torch.diagonal(attention_mask_tensor[:, 0], dim1=1, dim2=2) |
| | |
| | if attention_mask_tensor.dtype.is_floating_point: |
| | attention_mask_tensor = attention_mask_tensor / torch.finfo(attention_mask_tensor.dtype).min |
| | attention_mask_tensor = (1.0 - attention_mask_tensor).int() |
| |
|
| | |
| | |
| | |
| | |
| | prefill_compiled_stage = is_torchdynamo_compiling() and ( |
| | (input_ids is not None and input_ids.shape[1] != 1) |
| | or (inputs_embeds is not None and inputs_embeds.shape[1] != 1) |
| | ) |
| | prefill_noncompiled_stage = not is_torchdynamo_compiling() and ( |
| | (cache_position is not None and cache_position[0] == 0) |
| | or (past_key_values is None or past_key_values.get_seq_length() == 0) |
| | ) |
| | if (prefill_compiled_stage or prefill_noncompiled_stage) or self.rope_deltas is None: |
| | position_ids, rope_deltas = self.get_rope_index( |
| | input_ids, |
| | image_grid_thw, |
| | video_grid_thw, |
| | attention_mask=attention_mask_tensor, |
| | ) |
| | self.rope_deltas = rope_deltas |
| | |
| | else: |
| | batch_size, seq_length, _ = inputs_embeds.shape |
| | delta = ( |
| | (cache_position[0] + self.rope_deltas).to(inputs_embeds.device) |
| | if cache_position is not None |
| | else 0 |
| | ) |
| | position_ids = torch.arange(seq_length, device=inputs_embeds.device) |
| | position_ids = position_ids.view(1, -1).expand(batch_size, -1) |
| | if cache_position is not None: |
| | delta = delta.repeat_interleave(batch_size // delta.shape[0], dim=0) |
| | position_ids = position_ids.add(delta) |
| | position_ids = position_ids.unsqueeze(0).expand(3, -1, -1) |
| |
|
| | outputs = self.language_model( |
| | input_ids=None, |
| | position_ids=position_ids, |
| | attention_mask=attention_mask, |
| | past_key_values=past_key_values, |
| | inputs_embeds=inputs_embeds, |
| | cache_position=cache_position, |
| | visual_pos_masks=visual_pos_masks, |
| | deepstack_visual_embeds=deepstack_visual_embeds, |
| | **kwargs, |
| | ) |
| |
|
| | return Qwen3VLMoeModelOutputWithPast( |
| | last_hidden_state=outputs.last_hidden_state, |
| | past_key_values=outputs.past_key_values, |
| | rope_deltas=self.rope_deltas, |
| | ) |
| |
|
| |
|
| | def qwen3_vl_mixed_modality_forward( |
| | self, |
| | input_ids: torch.LongTensor = None, |
| | attention_mask: Optional[torch.Tensor] = None, |
| | position_ids: Optional[torch.LongTensor] = None, |
| | past_key_values: Optional[Cache] = None, |
| | inputs_embeds: Optional[torch.FloatTensor] = None, |
| | pixel_values: Optional[torch.Tensor] = None, |
| | pixel_values_videos: Optional[torch.FloatTensor] = None, |
| | image_grid_thw: Optional[torch.LongTensor] = None, |
| | video_grid_thw: Optional[torch.LongTensor] = None, |
| | cache_position: Optional[torch.LongTensor] = None, |
| | second_per_grid_ts: Optional[torch.Tensor] = None, |
| | **kwargs: Unpack[TransformersKwargs], |
| | ) -> Union[tuple, Qwen3VLModelOutputWithPast]: |
| | r""" |
| | image_grid_thw (`torch.LongTensor` of shape `(num_images, 3)`, *optional*): |
| | The temporal, height and width of feature shape of each image in LLM. |
| | video_grid_thw (`torch.LongTensor` of shape `(num_videos, 3)`, *optional*): |
| | The temporal, height and width of feature shape of each video in LLM. |
| | """ |
| | if (input_ids is None) ^ (inputs_embeds is not None): |
| | raise ValueError("You must specify exactly one of input_ids or inputs_embeds") |
| |
|
| | if inputs_embeds is None: |
| | inputs_embeds = self.get_input_embeddings()(input_ids) |
| |
|
| | image_mask = None |
| | video_mask = None |
| |
|
| | if pixel_values is None and pixel_values_videos is None: |
| | |
| | dummy_pixel = torch.zeros(1024, 1536).to(self.visual.device) |
| | dummy_grid = torch.tensor([[1, 32, 32]]).to(self.visual.device) |
| |
|
| | image_embeds, dummy_deepstack = self.get_image_features(dummy_pixel, dummy_grid) |
| | image_embeds = torch.cat(image_embeds, dim=0).to(inputs_embeds.device, inputs_embeds.dtype) |
| | |
| | inputs_embeds += image_embeds.mean() * 0 |
| |
|
| | if pixel_values is not None: |
| | image_embeds, deepstack_image_embeds = self.get_image_features(pixel_values, image_grid_thw) |
| | image_embeds = torch.cat(image_embeds, dim=0).to(inputs_embeds.device, inputs_embeds.dtype) |
| | image_mask, _ = self.get_placeholder_mask( |
| | input_ids, inputs_embeds=inputs_embeds, image_features=image_embeds |
| | ) |
| | inputs_embeds = inputs_embeds.masked_scatter(image_mask, image_embeds) |
| |
|
| | if pixel_values_videos is not None: |
| | video_embeds, deepstack_video_embeds = self.get_video_features(pixel_values_videos, video_grid_thw) |
| | video_embeds = torch.cat(video_embeds, dim=0).to(inputs_embeds.device, inputs_embeds.dtype) |
| | _, video_mask = self.get_placeholder_mask( |
| | input_ids, inputs_embeds=inputs_embeds, video_features=video_embeds |
| | ) |
| | inputs_embeds = inputs_embeds.masked_scatter(video_mask, video_embeds) |
| |
|
| | visual_pos_masks = None |
| | deepstack_visual_embeds = None |
| | if image_mask is not None and video_mask is not None: |
| | |
| | image_mask = image_mask[..., 0] |
| | video_mask = video_mask[..., 0] |
| | visual_pos_masks = image_mask | video_mask |
| | deepstack_visual_embeds = [] |
| | image_mask_joint = image_mask[visual_pos_masks] |
| | video_mask_joint = video_mask[visual_pos_masks] |
| | for img_embed, vid_embed in zip(deepstack_image_embeds, deepstack_video_embeds): |
| | embed_joint = img_embed.new_zeros(visual_pos_masks.sum(), img_embed.shape[-1]).to(img_embed.device) |
| | embed_joint[image_mask_joint, :] = img_embed |
| | embed_joint[video_mask_joint, :] = vid_embed |
| | deepstack_visual_embeds.append(embed_joint) |
| | elif image_mask is not None: |
| | image_mask = image_mask[..., 0] |
| | visual_pos_masks = image_mask |
| | deepstack_visual_embeds = deepstack_image_embeds |
| | elif video_mask is not None: |
| | video_mask = video_mask[..., 0] |
| | visual_pos_masks = video_mask |
| | deepstack_visual_embeds = deepstack_video_embeds |
| |
|
| | if visual_pos_masks is None: |
| | B, S, H = inputs_embeds.shape |
| | visual_pos_masks = torch.zeros((B, S), dtype=torch.bool, device=inputs_embeds.device) |
| | L = len(self.visual.deepstack_visual_indexes) |
| | deepstack_visual_embeds = [t.narrow(0, 0, 0) for t in dummy_deepstack] |
| |
|
| | if position_ids is None: |
| | attention_mask_tensor = ( |
| | attention_mask if not isinstance(attention_mask, dict) else attention_mask["full_attention"] |
| | ) |
| | if attention_mask_tensor is not None and attention_mask_tensor.ndim == 4: |
| | attention_mask_tensor = torch.diagonal(attention_mask_tensor[:, 0], dim1=1, dim2=2) |
| | |
| | if attention_mask_tensor.dtype.is_floating_point: |
| | attention_mask_tensor = attention_mask_tensor / torch.finfo(attention_mask_tensor.dtype).min |
| | attention_mask_tensor = (1.0 - attention_mask_tensor).int() |
| |
|
| | |
| | |
| | |
| | |
| | prefill_compiled_stage = is_torchdynamo_compiling() and ( |
| | (input_ids is not None and input_ids.shape[1] != 1) |
| | or (inputs_embeds is not None and inputs_embeds.shape[1] != 1) |
| | ) |
| | prefill_noncompiled_stage = not is_torchdynamo_compiling() and ( |
| | (cache_position is not None and cache_position[0] == 0) |
| | or (past_key_values is None or past_key_values.get_seq_length() == 0) |
| | ) |
| | if (prefill_compiled_stage or prefill_noncompiled_stage) or self.rope_deltas is None: |
| | position_ids, rope_deltas = self.get_rope_index( |
| | input_ids, |
| | image_grid_thw, |
| | video_grid_thw, |
| | attention_mask=attention_mask_tensor, |
| | ) |
| | self.rope_deltas = rope_deltas |
| | |
| | else: |
| | batch_size, seq_length, _ = inputs_embeds.shape |
| | delta = ( |
| | (cache_position[0] + self.rope_deltas).to(inputs_embeds.device) |
| | if cache_position is not None |
| | else 0 |
| | ) |
| | position_ids = torch.arange(seq_length, device=inputs_embeds.device) |
| | position_ids = position_ids.view(1, -1).expand(batch_size, -1) |
| | if cache_position is not None: |
| | delta = delta.repeat_interleave(batch_size // delta.shape[0], dim=0) |
| | position_ids = position_ids.add(delta) |
| | position_ids = position_ids.unsqueeze(0).expand(3, -1, -1) |
| |
|
| | outputs = self.language_model( |
| | input_ids=None, |
| | position_ids=position_ids, |
| | attention_mask=attention_mask, |
| | past_key_values=past_key_values, |
| | inputs_embeds=inputs_embeds, |
| | cache_position=cache_position, |
| | visual_pos_masks=visual_pos_masks, |
| | deepstack_visual_embeds=deepstack_visual_embeds, |
| | **kwargs, |
| | ) |
| |
|
| | return Qwen3VLModelOutputWithPast( |
| | last_hidden_state=outputs.last_hidden_state, |
| | past_key_values=outputs.past_key_values, |
| | rope_deltas=self.rope_deltas, |
| | ) |
| |
|
| | def qwen2_5_mixed_modality_forward( |
| | self, |
| | input_ids: torch.LongTensor = None, |
| | attention_mask: Optional[torch.Tensor] = None, |
| | position_ids: Optional[torch.LongTensor] = None, |
| | past_key_values: Optional[Cache] = None, |
| | inputs_embeds: Optional[torch.FloatTensor] = None, |
| | use_cache: Optional[bool] = None, |
| | output_attentions: Optional[bool] = None, |
| | output_hidden_states: Optional[bool] = None, |
| | return_dict: Optional[bool] = None, |
| | pixel_values: Optional[torch.Tensor] = None, |
| | pixel_values_videos: Optional[torch.FloatTensor] = None, |
| | image_grid_thw: Optional[torch.LongTensor] = None, |
| | video_grid_thw: Optional[torch.LongTensor] = None, |
| | rope_deltas: Optional[torch.LongTensor] = None, |
| | cache_position: Optional[torch.LongTensor] = None, |
| | second_per_grid_ts: Optional[torch.Tensor] = None, |
| | **kwargs: Unpack[TransformersKwargs], |
| | ) -> Union[tuple, Qwen2_5_VLModelOutputWithPast]: |
| | r""" |
| | image_grid_thw (`torch.LongTensor` of shape `(num_images, 3)`, *optional*): |
| | The temporal, height and width of feature shape of each image in LLM. |
| | video_grid_thw (`torch.LongTensor` of shape `(num_videos, 3)`, *optional*): |
| | The temporal, height and width of feature shape of each video in LLM. |
| | rope_deltas (`torch.LongTensor` of shape `(batch_size, )`, *optional*): |
| | The rope index difference between sequence length and multimodal rope. |
| | second_per_grid_ts (`torch.Tensor` of shape `(num_videos)`, *optional*): |
| | The time interval (in seconds) for each grid along the temporal dimension in the 3D position IDs. |
| | """ |
| |
|
| | output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions |
| | output_hidden_states = ( |
| | output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states |
| | ) |
| | return_dict = return_dict if return_dict is not None else self.config.use_return_dict |
| |
|
| | if inputs_embeds is None: |
| | inputs_embeds = self.get_input_embeddings()(input_ids) |
| |
|
| | if pixel_values is None and pixel_values_videos is None: |
| | |
| | dummy_pixel = torch.zeros(784, 1176).to(self.visual.device) |
| | dummy_grid = torch.tensor([[1, 28, 28]]).to(self.visual.device) |
| |
|
| | image_embeds = self.get_image_features(dummy_pixel, dummy_grid) |
| | |
| | |
| | |
| | if isinstance(image_embeds, (tuple, list)): |
| | image_embeds = torch.cat(list(image_embeds), dim=0) |
| | inputs_embeds += image_embeds.mean() * 0 |
| |
|
| | if pixel_values is not None: |
| | image_embeds = self.get_image_features(pixel_values, image_grid_thw) |
| | image_embeds = torch.cat(image_embeds, dim=0).to(inputs_embeds.device, inputs_embeds.dtype) |
| | image_mask, _ = self.get_placeholder_mask( |
| | input_ids, inputs_embeds=inputs_embeds, image_features=image_embeds |
| | ) |
| | inputs_embeds = inputs_embeds.masked_scatter(image_mask, image_embeds) |
| |
|
| | if pixel_values_videos is not None: |
| | video_embeds = self.get_video_features(pixel_values_videos, video_grid_thw) |
| | video_embeds = torch.cat(video_embeds, dim=0).to(inputs_embeds.device, inputs_embeds.dtype) |
| | _, video_mask = self.get_placeholder_mask( |
| | input_ids, inputs_embeds=inputs_embeds, video_features=video_embeds |
| | ) |
| | inputs_embeds = inputs_embeds.masked_scatter(video_mask, video_embeds) |
| |
|
| | if position_ids is None: |
| | |
| | |
| | |
| | |
| | prefill_compiled_stage = is_torchdynamo_compiling() and ( |
| | (input_ids is not None and input_ids.shape[1] != 1) |
| | or (inputs_embeds is not None and inputs_embeds.shape[1] != 1) |
| | ) |
| | prefill_noncompiled_stage = not is_torchdynamo_compiling() and ( |
| | (cache_position is not None and cache_position[0] == 0) |
| | or (past_key_values is None or past_key_values.get_seq_length() == 0) |
| | ) |
| | if (prefill_compiled_stage or prefill_noncompiled_stage) or self.rope_deltas is None: |
| | position_ids, rope_deltas = self.get_rope_index( |
| | input_ids, |
| | image_grid_thw, |
| | video_grid_thw, |
| | second_per_grid_ts=second_per_grid_ts, |
| | attention_mask=attention_mask, |
| | ) |
| | self.rope_deltas = rope_deltas |
| | else: |
| | batch_size, seq_length, _ = inputs_embeds.shape |
| | position_ids = torch.arange(seq_length, device=inputs_embeds.device) |
| | position_ids = position_ids.view(1, 1, -1).expand(3, batch_size, -1) |
| | if cache_position is not None: |
| | delta = (cache_position[0] + self.rope_deltas).to(inputs_embeds.device) |
| | else: |
| | delta = torch.zeros((batch_size, seq_length), device=inputs_embeds.device) |
| | delta = delta.repeat_interleave(batch_size // delta.shape[0], dim=1) |
| | position_ids += delta.to(position_ids.device) |
| |
|
| | outputs = self.language_model( |
| | input_ids=None, |
| | position_ids=position_ids, |
| | attention_mask=attention_mask, |
| | past_key_values=past_key_values, |
| | inputs_embeds=inputs_embeds, |
| | use_cache=use_cache, |
| | output_attentions=output_attentions, |
| | output_hidden_states=output_hidden_states, |
| | return_dict=True, |
| | cache_position=cache_position, |
| | **kwargs, |
| | ) |
| |
|
| | output = Qwen2_5_VLModelOutputWithPast( |
| | last_hidden_state=outputs.last_hidden_state, |
| | past_key_values=outputs.past_key_values, |
| | hidden_states=outputs.hidden_states, |
| | attentions=outputs.attentions, |
| | rope_deltas=self.rope_deltas, |
| | ) |
| | return output if return_dict else output.to_tuple() |
| |
|
| |
|
| | def qwen2_mixed_modality_forward( |
| | self, |
| | input_ids: torch.LongTensor = None, |
| | attention_mask: Optional[torch.Tensor] = None, |
| | position_ids: Optional[torch.LongTensor] = None, |
| | past_key_values: Optional[Cache] = None, |
| | inputs_embeds: Optional[torch.FloatTensor] = None, |
| | use_cache: Optional[bool] = None, |
| | output_attentions: Optional[bool] = None, |
| | output_hidden_states: Optional[bool] = None, |
| | return_dict: Optional[bool] = None, |
| | pixel_values: Optional[torch.Tensor] = None, |
| | pixel_values_videos: Optional[torch.FloatTensor] = None, |
| | image_grid_thw: Optional[torch.LongTensor] = None, |
| | video_grid_thw: Optional[torch.LongTensor] = None, |
| | rope_deltas: Optional[torch.LongTensor] = None, |
| | cache_position: Optional[torch.LongTensor] = None, |
| | second_per_grid_ts: Optional[torch.Tensor] = None, |
| | **kwargs: Unpack[TransformersKwargs], |
| | ) -> Union[tuple, Qwen2VLModelOutputWithPast]: |
| | r""" |
| | image_grid_thw (`torch.LongTensor` of shape `(num_images, 3)`, *optional*): |
| | The temporal, height and width of feature shape of each image in LLM. |
| | video_grid_thw (`torch.LongTensor` of shape `(num_videos, 3)`, *optional*): |
| | The temporal, height and width of feature shape of each video in LLM. |
| | rope_deltas (`torch.LongTensor` of shape `(batch_size, )`, *optional*): |
| | The rope index difference between sequence length and multimodal rope. |
| | """ |
| |
|
| | output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions |
| | output_hidden_states = ( |
| | output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states |
| | ) |
| | return_dict = return_dict if return_dict is not None else self.config.use_return_dict |
| |
|
| | if inputs_embeds is None: |
| | inputs_embeds = self.get_input_embeddings()(input_ids) |
| |
|
| | if pixel_values is None and pixel_values_videos is None: |
| | |
| | dummy_pixel = torch.zeros(784, 1176).to(self.visual.get_device()) |
| | dummy_grid = torch.tensor([[1, 28, 28]]).to(self.visual.get_device()) |
| |
|
| | image_embeds = self.get_image_features(dummy_pixel, dummy_grid) |
| | |
| | |
| | |
| | if isinstance(image_embeds, (tuple, list)): |
| | image_embeds = torch.cat(list(image_embeds), dim=0) |
| | inputs_embeds += image_embeds.mean() * 0 |
| |
|
| | if pixel_values is not None: |
| | image_embeds = self.get_image_features(pixel_values, image_grid_thw) |
| | image_embeds = torch.cat(image_embeds, dim=0).to(inputs_embeds.device, inputs_embeds.dtype) |
| | image_mask, _ = self.get_placeholder_mask( |
| | input_ids, inputs_embeds=inputs_embeds, image_features=image_embeds |
| | ) |
| | inputs_embeds = inputs_embeds.masked_scatter(image_mask, image_embeds) |
| |
|
| | if pixel_values_videos is not None: |
| | video_embeds = self.get_video_features(pixel_values_videos, video_grid_thw) |
| | video_embeds = torch.cat(video_embeds, dim=0).to(inputs_embeds.device, inputs_embeds.dtype) |
| | _, video_mask = self.get_placeholder_mask( |
| | input_ids, inputs_embeds=inputs_embeds, video_features=video_embeds |
| | ) |
| | inputs_embeds = inputs_embeds.masked_scatter(video_mask, video_embeds) |
| |
|
| | if position_ids is None: |
| | if self.rope_deltas is None or cache_position is None or cache_position[0] == 0: |
| | position_ids, rope_deltas = self.get_rope_index( |
| | input_ids, image_grid_thw, video_grid_thw, attention_mask |
| | ) |
| | self.rope_deltas = rope_deltas |
| | |
| | else: |
| | batch_size, seq_length, _ = inputs_embeds.shape |
| | position_ids = torch.arange(seq_length, device=inputs_embeds.device) |
| | position_ids = position_ids.view(1, 1, -1).expand(3, batch_size, -1) |
| | if cache_position is not None: |
| | delta = (cache_position[0] + self.rope_deltas).to(inputs_embeds.device) |
| | else: |
| | delta = torch.zeros((batch_size, seq_length), device=inputs_embeds.device) |
| | delta = delta.repeat_interleave(batch_size // delta.shape[0], dim=0) |
| | position_ids += delta.to(position_ids.device) |
| |
|
| | outputs = self.language_model( |
| | input_ids=None, |
| | position_ids=position_ids, |
| | attention_mask=attention_mask, |
| | past_key_values=past_key_values, |
| | inputs_embeds=inputs_embeds, |
| | use_cache=use_cache, |
| | output_attentions=output_attentions, |
| | output_hidden_states=output_hidden_states, |
| | return_dict=True, |
| | cache_position=cache_position, |
| | **kwargs, |
| | ) |
| |
|
| | output = Qwen2VLModelOutputWithPast( |
| | last_hidden_state=outputs.last_hidden_state, |
| | past_key_values=outputs.past_key_values, |
| | hidden_states=outputs.hidden_states, |
| | attentions=outputs.attentions, |
| | rope_deltas=self.rope_deltas, |
| | ) |
| | return output if return_dict else output.to_tuple() |