import torch import torch.nn.functional as F import unicodedata import numpy as np import logging from PIL import Image from dataclasses import dataclass from typing import Optional, List, Union, Dict, Any from transformers.models.qwen3_vl.modeling_qwen3_vl import Qwen3VLPreTrainedModel, Qwen3VLModel, Qwen3VLConfig from transformers.models.qwen3_vl.processing_qwen3_vl import Qwen3VLProcessor from transformers.modeling_outputs import ModelOutput from transformers.processing_utils import Unpack from transformers.utils import TransformersKwargs from transformers.cache_utils import Cache from transformers.utils.generic import check_model_inputs from qwen_vl_utils.vision_process import process_vision_info logger = logging.getLogger(__name__) # Constants for configuration MAX_LENGTH = 8192 IMAGE_BASE_FACTOR = 16 IMAGE_FACTOR = IMAGE_BASE_FACTOR * 2 MIN_PIXELS = 4 * IMAGE_FACTOR * IMAGE_FACTOR MAX_PIXELS = 1800 * IMAGE_FACTOR * IMAGE_FACTOR FPS = 1 MAX_FRAMES = 64 FRAME_MAX_PIXELS = 768 * IMAGE_FACTOR * IMAGE_FACTOR MAX_TOTAL_PIXELS = 10 * FRAME_MAX_PIXELS PAD_TOKEN = "<|endoftext|>" # Define output structure for embeddings @dataclass class Qwen3VLForEmbeddingOutput(ModelOutput): last_hidden_state: Optional[torch.FloatTensor] = None attention_mask: Optional[torch.Tensor] = None # Define model class to compute embeddings class Qwen3VLForEmbedding(Qwen3VLPreTrainedModel): _checkpoint_conversion_mapping = {} accepts_loss_kwargs = False config: Qwen3VLConfig def __init__(self, config): super().__init__(config) self.model = Qwen3VLModel(config) self.post_init() def get_input_embeddings(self): return self.model.get_input_embeddings() def set_input_embeddings(self, value): self.model.set_input_embeddings(value) def set_decoder(self, decoder): self.model.set_decoder(decoder) def get_decoder(self): return self.model.get_decoder() # Extract video features from model def get_video_features(self, pixel_values_videos: torch.FloatTensor, video_grid_thw: Optional[torch.LongTensor] = None): return self.model.get_video_features(pixel_values_videos, video_grid_thw) # Extract image features from model def get_image_features(self, pixel_values: torch.FloatTensor, image_grid_thw: Optional[torch.LongTensor] = None): return self.model.get_image_features(pixel_values, image_grid_thw) # Make modules accessible through properties @property def language_model(self): return self.model.language_model @property def visual(self): return self.model.visual # Forward pass through model with input parameters # @check_model_inputs def forward(self, input_ids: torch.LongTensor = None, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, past_key_values: Optional[Cache] = None, inputs_embeds: Optional[torch.FloatTensor] = None, pixel_values: Optional[torch.Tensor] = None, pixel_values_videos: Optional[torch.FloatTensor] = None, image_grid_thw: Optional[torch.LongTensor] = None, video_grid_thw: Optional[torch.LongTensor] = None, cache_position: Optional[torch.LongTensor] = None, logits_to_keep: Union[int, torch.Tensor] = 0, **kwargs: Unpack[TransformersKwargs], ) -> Union[tuple, Qwen3VLForEmbeddingOutput]: # Pass inputs through the model outputs = self.model( input_ids=input_ids, pixel_values=pixel_values, pixel_values_videos=pixel_values_videos, image_grid_thw=image_grid_thw, video_grid_thw=video_grid_thw, position_ids=position_ids, attention_mask=attention_mask, past_key_values=past_key_values, inputs_embeds=inputs_embeds, cache_position=cache_position, **kwargs, ) # Return the model output return Qwen3VLForEmbeddingOutput( last_hidden_state=outputs.last_hidden_state, attention_mask=attention_mask, ) def sample_frames(frames: List[Union[str, Image.Image]], num_segments: int, max_segments: int) -> List[str]: duration = len(frames) frame_id_array = np.linspace(0, duration - 1, num_segments, dtype=int) frame_id_list = frame_id_array.tolist() last_frame_id = frame_id_list[-1] # Create a list of sampled frames sampled_frames = [] for frame_idx in frame_id_list: try: sampled_frames.append(frames[frame_idx]) except: break # Ensure the sampled list meets the required segment count while len(sampled_frames) < num_segments: sampled_frames.append(frames[last_frame_id]) return sampled_frames[:max_segments] # Define embedder class for processing inputs and generating embeddings class Qwen3VLEmbedder(): def __init__( self, model_name_or_path: str, max_length: int = MAX_LENGTH, min_pixels: int = MIN_PIXELS, max_pixels: int = MAX_PIXELS, total_pixels: int = MAX_TOTAL_PIXELS, fps: float = FPS, num_frames: int = MAX_FRAMES, max_frames: int = MAX_FRAMES, default_instruction: str = "Represent the user's input.", **kwargs ): device = torch.device("cuda" if torch.cuda.is_available() else "cpu") self.max_length = max_length self.min_pixels = min_pixels self.max_pixels = max_pixels self.total_pixels = total_pixels self.fps = fps self.num_frames = num_frames self.max_frames = max_frames self.default_instruction = default_instruction self.model = Qwen3VLForEmbedding.from_pretrained( model_name_or_path, trust_remote_code=True, **kwargs ).to(device) self.processor = Qwen3VLProcessor.from_pretrained( model_name_or_path, padding_side='right' ) self.model.eval() @torch.no_grad() def forward(self, inputs: Dict[str, Any]) -> Dict[str, torch.Tensor]: outputs = self.model(**inputs) return { 'last_hidden_state': outputs.last_hidden_state, 'attention_mask': inputs.get('attention_mask') } # Truncate token sequence to a specified max length def _truncate_tokens(self, token_ids: List[int], max_length: int) -> List[int]: if len(token_ids) <= max_length: return token_ids special_token_ids = set(self.processor.tokenizer.all_special_ids) num_special = sum(1 for token_idx in token_ids if token_idx in special_token_ids) num_non_special_to_keep = max_length - num_special final_token_ids = [] non_special_kept_count = 0 # Ensure retention of special tokens while truncating the rest for token_idx in token_ids: if token_idx in special_token_ids: final_token_ids.append(token_idx) elif non_special_kept_count < num_non_special_to_keep: final_token_ids.append(token_idx) non_special_kept_count += 1 return final_token_ids # Format input based on provided text, image, video, and instruction def format_model_input( self, text: Optional[str] = None, image: Optional[Union[str, Image.Image]] = None, video: Optional[Union[str, List[Union[str, Image.Image]]]] = None, instruction: Optional[str] = None, fps: Optional[float] = None, max_frames: Optional[int] = None ) -> List[Dict]: # Ensure instruction ends with punctuation if instruction: instruction = instruction.strip() if instruction and not unicodedata.category(instruction[-1]).startswith('P'): instruction = instruction + '.' # Initialize conversation with system prompts content = [] conversation = [ {"role": "system", "content": [{"type": "text", "text": instruction or self.default_instruction}]}, {"role": "user", "content": content} ] # Add text, image, or video content to conversation if not text and not image and not video: content.append({'type': 'text', 'text': "NULL"}) return conversation if video: video_content = None video_kwargs = { 'total_pixels': self.total_pixels } if isinstance(video, list): video_content = video if self.num_frames is not None or self.max_frames is not None: video_content = sample_frames(video_content, self.num_frames, self.max_frames) video_content = [ ('file://' + ele if isinstance(ele, str) else ele) for ele in video_content ] elif isinstance(video, str): video_content = video if video.startswith(('http://', 'https://')) else 'file://' + video video_kwargs = {'fps': fps or self.fps, 'max_frames': max_frames or self.max_frames,} else: raise TypeError(f"Unrecognized video type: {type(video)}") # Add video input details to content if video_content: content.append({ 'type': 'video', 'video': video_content, **video_kwargs }) if image: image_content = None if isinstance(image, Image.Image): image_content = image elif isinstance(image, str): image_content = image if image.startswith(('http', 'oss')) else 'file://' + image else: raise TypeError(f"Unrecognized image type: {type(image)}") # Add image input details to content if image_content: content.append({ 'type': 'image', 'image': image_content, "min_pixels": self.min_pixels, "max_pixels": self.max_pixels }) if text: content.append({'type': 'text', 'text': text}) return conversation # Preprocess input conversations for model consumption def _preprocess_inputs(self, conversations: List[List[Dict]]) -> Dict[str, torch.Tensor]: text = self.processor.apply_chat_template( conversations, add_generation_prompt=True, tokenize=False ) try: images, video_inputs, video_kwargs = process_vision_info( conversations, image_patch_size=16, return_video_metadata=True, return_video_kwargs=True ) except Exception as e: logger.error(f"Error in processing vision info: {e}") images = None video_inputs = None video_kwargs = {'do_sample_frames': False} text = self.processor.apply_chat_template( [{'role': 'user', 'content': [{'type': 'text', 'text': 'NULL'}]}], add_generation_prompt=True, tokenize=False ) if video_inputs is not None: videos, video_metadata = zip(*video_inputs) videos = list(videos) video_metadata = list(video_metadata) else: videos, video_metadata = None, None inputs = self.processor( text=text, images=images, videos=videos, video_metadata=video_metadata, truncation=True, max_length=self.max_length, padding=True, do_resize=False, return_tensors='pt', **video_kwargs ) return inputs # Pool the last hidden state by attention mask for embeddings @staticmethod def _pooling_last(hidden_state: torch.Tensor, attention_mask: torch.Tensor) -> torch.Tensor: flipped_tensor = attention_mask.flip(dims=[1]) last_one_positions = flipped_tensor.argmax(dim=1) col = attention_mask.shape[1] - last_one_positions - 1 row = torch.arange(hidden_state.shape[0], device=hidden_state.device) return hidden_state[row, col] # Process inputs to generate normalized embeddings def process(self, inputs: List[Dict[str, Any]], normalize: bool = True) -> tuple: conversations = [self.format_model_input( text=ele.get('text'), image=ele.get('image'), video=ele.get('video'), instruction=ele.get('instruction'), fps=ele.get('fps'), max_frames=ele.get('max_frames') ) for ele in inputs] processed_inputs = self._preprocess_inputs(conversations) processed_inputs = {k: v.to(self.model.device) for k, v in processed_inputs.items()} outputs = self.forward(processed_inputs) embeddings = self._pooling_last(outputs['last_hidden_state'], outputs['attention_mask']) # Normalize the embeddings if specified if normalize: embeddings = F.normalize(embeddings, p=2, dim=-1) return embeddings