sample_id
stringlengths
21
196
text
stringlengths
105
936k
metadata
dict
category
stringclasses
6 values
vllm-project/vllm:vllm/model_executor/models/audioflamingo3.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright 2025 The vLLM team. # Copyright 2025 NVIDIA CORPORATION and the HuggingFace Inc. team. All rights # reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from collections.abc import Iterable, Mapping, Sequence from typing import Annotated, Any, Literal, TypeAlias import torch import torch.nn as nn from transformers import BatchFeature, PretrainedConfig from transformers.models.audioflamingo3 import ( AudioFlamingo3Config, AudioFlamingo3Processor, ) from transformers.models.qwen2_audio import Qwen2AudioEncoder from vllm.config import VllmConfig from vllm.config.multimodal import BaseDummyOptions from vllm.model_executor.layers.activation import get_act_fn from vllm.model_executor.models.module_mapping import MultiModelKeys from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import ( MultiModalDataDict, MultiModalFieldConfig, MultiModalKwargsItems, ) from vllm.multimodal.parse import ( DictEmbeddingItems, ModalityData, ModalityDataItems, MultiModalDataItems, MultiModalDataParser, ) from vllm.multimodal.processing import ( BaseDummyInputsBuilder, BaseMultiModalProcessor, BaseProcessingInfo, PromptReplacement, PromptUpdate, PromptUpdateDetails, ) from vllm.sequence import IntermediateTensors from vllm.utils.tensor_schema import TensorSchema, TensorShape from .interfaces import ( MultiModalEmbeddings, SupportsLoRA, SupportsMultiModal, SupportsPP, ) from .utils import ( AutoWeightsLoader, init_vllm_registered_model, maybe_prefix, ) MAX_AUDIO_LEN = 10 * 60 # === Audio Inputs === # class AudioFlamingo3FeatureInputs(TensorSchema): """ Dimensions: - num_chunks: Number of audio chunks (flattened) - nmb: Number of mel bins - num_audios: Number of original audio files """ type: Literal["audio_features"] input_features: Annotated[ torch.Tensor | list[torch.Tensor], TensorShape("num_chunks", "nmb", 3000), ] feature_attention_mask: Annotated[ torch.Tensor, TensorShape("num_chunks", 3000), ] chunk_counts: Annotated[ torch.Tensor, TensorShape("num_audios"), ] class AudioFlamingo3EmbeddingInputs(TensorSchema): """ Dimensions: - bn: Batch size - naf: Number of audio features - hs: Hidden size (must match the hidden size of language model backbone) """ type: Literal["audio_embeds"] = "audio_embeds" audio_embeds: Annotated[ list[torch.Tensor], TensorShape("bn", "naf", "hs", dynamic_dims={"naf"}), ] AudioFlamingo3Inputs: TypeAlias = ( AudioFlamingo3FeatureInputs | AudioFlamingo3EmbeddingInputs ) class AudioFlamingo3Encoder(Qwen2AudioEncoder): def __init__( self, config: PretrainedConfig, ): super().__init__(config) self.avg_pooler = nn.AvgPool1d(kernel_size=2, stride=2) # self.layer_norm is already initialized in super().__init__ # Keep a dummy freqs parameter for MusicFlamingo checkpoints. self.pos_emb = nn.Module() freqs = torch.empty(getattr(config, "num_mel_bins", 128)) self.pos_emb.register_parameter( "freqs", nn.Parameter(freqs, requires_grad=False) ) def forward( self, input_features: torch.Tensor | list[torch.Tensor], attention_mask: torch.Tensor = None, ): # input_features: (batch, num_mel_bins, seq_len) if isinstance(input_features, list): input_features = torch.stack(input_features) hidden_states = nn.functional.gelu(self.conv1(input_features)) hidden_states = nn.functional.gelu(self.conv2(hidden_states)) hidden_states = hidden_states.transpose(-1, -2) hidden_states = ( hidden_states + self.embed_positions.weight[: hidden_states.size(-2), :] ).to(hidden_states.dtype) for layer in self.layers: # Qwen2AudioEncoderLayer expects layer_head_mask as third arg. layer_outputs = layer(hidden_states, attention_mask, None) hidden_states = layer_outputs[0] # AvgPool (time/2) + LayerNorm # hidden_states: (batch, seq_len, hidden_size) hidden_states = hidden_states.permute(0, 2, 1) # (batch, hidden_size, seq_len) hidden_states = self.avg_pooler(hidden_states) hidden_states = hidden_states.permute( 0, 2, 1 ) # (batch, seq_len/2, hidden_size) hidden_states = self.layer_norm(hidden_states) return hidden_states def _get_feat_extract_output_lengths(self, input_lengths: torch.Tensor): """ Computes the output length of the convolutional layers and the output length of the audio encoder """ input_lengths = (input_lengths - 1) // 2 + 1 output_lengths = (input_lengths - 2) // 2 + 1 return input_lengths, output_lengths class AudioFlamingo3MultiModalProjector(nn.Module): def __init__(self, config: PretrainedConfig): super().__init__() self.linear_1 = nn.Linear( config.audio_config.hidden_size, config.text_config.hidden_size, bias=config.projector_bias, ) self.act = get_act_fn(config.projector_hidden_act) self.linear_2 = nn.Linear( config.text_config.hidden_size, config.text_config.hidden_size, bias=config.projector_bias, ) def forward(self, audio_features): hidden_states = self.linear_1(audio_features) hidden_states = self.act(hidden_states) hidden_states = self.linear_2(hidden_states) return hidden_states class AudioFlamingo3MultiModalDataParser(MultiModalDataParser): def _parse_audio_data( self, data: dict[str, torch.Tensor] | ModalityData[Any], ) -> ModalityDataItems[Any, Any] | None: if isinstance(data, dict): return DictEmbeddingItems( data, modality="audio", required_fields={"audio_embeds"}, fields_factory=_audioflamingo3_field_config, ) return super()._parse_audio_data(data) class AudioFlamingo3ProcessingInfo(BaseProcessingInfo): def get_hf_config(self): return self.ctx.get_hf_config(AudioFlamingo3Config) def get_hf_processor(self, **kwargs: object): return self.ctx.get_hf_processor(AudioFlamingo3Processor, **kwargs) def get_feature_extractor(self, **kwargs: object): hf_processor = self.get_hf_processor(**kwargs) feature_extractor = hf_processor.feature_extractor return feature_extractor def get_data_parser(self): feature_extractor = self.get_feature_extractor() return AudioFlamingo3MultiModalDataParser( target_sr=feature_extractor.sampling_rate, expected_hidden_size=self._get_expected_hidden_size(), ) def get_supported_mm_limits(self) -> Mapping[str, int | None]: return {"audio": 1} class AudioFlamingo3DummyInputsBuilder( BaseDummyInputsBuilder[AudioFlamingo3ProcessingInfo] ): def get_dummy_text(self, mm_counts: Mapping[str, int]) -> str: num_audios = mm_counts.get("audio", 0) hf_processor = self.info.get_hf_processor() audio_token = hf_processor.audio_token return audio_token * num_audios def get_dummy_mm_data( self, seq_len: int, mm_counts: Mapping[str, int], mm_options: Mapping[str, BaseDummyOptions], ) -> MultiModalDataDict: feature_extractor = self.info.get_feature_extractor() sampling_rate = feature_extractor.sampling_rate audio_len = MAX_AUDIO_LEN * sampling_rate num_audios = mm_counts.get("audio", 0) audio_overrides = mm_options.get("audio") return { "audio": self._get_dummy_audios( length=audio_len, num_audios=num_audios, overrides=audio_overrides, ) } def _audioflamingo3_field_config(hf_inputs: Mapping[str, torch.Tensor]): chunk_counts = hf_inputs.get("chunk_counts") if chunk_counts is not None: return dict( audio_embeds=MultiModalFieldConfig.batched("audio"), input_features=MultiModalFieldConfig.flat_from_sizes( "audio", chunk_counts, dim=0 ), feature_attention_mask=MultiModalFieldConfig.flat_from_sizes( "audio", chunk_counts, dim=0 ), chunk_counts=MultiModalFieldConfig.batched("audio"), ) return dict( audio_embeds=MultiModalFieldConfig.batched("audio"), input_features=MultiModalFieldConfig.batched("audio"), feature_attention_mask=MultiModalFieldConfig.batched("audio"), chunk_counts=MultiModalFieldConfig.batched("audio"), ) class AudioFlamingo3MultiModalProcessor( BaseMultiModalProcessor[AudioFlamingo3ProcessingInfo] ): def _call_hf_processor( self, prompt: str, mm_data: dict[str, object], mm_kwargs: Mapping[str, Any], tok_kwargs: Mapping[str, object], ) -> BatchFeature: audios = mm_data.pop("audios", []) if audios: mm_data["audio"] = audios if not mm_data.get("audio", []): prompt_ids = self.info.get_tokenizer().encode(prompt) prompt_ids = self._apply_hf_processor_tokens_only(prompt_ids) return BatchFeature(dict(input_ids=[prompt_ids]), tensor_type="pt") feature_extractor = self.info.get_feature_extractor(**mm_kwargs) mm_kwargs = dict( **mm_kwargs, sampling_rate=feature_extractor.sampling_rate, ) # Calculate chunk counts audio_list = mm_data.get("audio") if not isinstance(audio_list, list): audio_list = [audio_list] chunk_counts = [] sampling_rate = feature_extractor.sampling_rate chunk_length = feature_extractor.chunk_length window_size = int(sampling_rate * chunk_length) # MAX_AUDIO_LEN is 10 * 60 in HF processor. max_windows = int(MAX_AUDIO_LEN // chunk_length) for audio in audio_list: # audio is numpy array or list n_samples = len(audio) if isinstance(audio, list) else audio.shape[0] n_win = max(1, (n_samples + window_size - 1) // window_size) if n_win > max_windows: n_win = max_windows chunk_counts.append(n_win) outputs = super()._call_hf_processor( prompt=prompt, mm_data=mm_data, mm_kwargs=mm_kwargs, tok_kwargs=tok_kwargs, ) if "input_features_mask" in outputs: outputs["feature_attention_mask"] = outputs.pop("input_features_mask") outputs["chunk_counts"] = torch.tensor(chunk_counts, dtype=torch.long) return outputs def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: return _audioflamingo3_field_config(hf_inputs) def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, object], out_mm_kwargs: MultiModalKwargsItems, ) -> Sequence[PromptUpdate]: processor = self.info.get_hf_processor(**hf_processor_mm_kwargs) tokenizer = self.info.get_tokenizer() vocab = tokenizer.get_vocab() audio_token = getattr(processor, "audio_token", "<sound>") audio_token_id = vocab.get(audio_token) if audio_token_id is None: # Fallback if not found, though it should be there audio_token_id = processor.audio_token_id out_mm_data = out_mm_kwargs.get_data() feature_attention_mask = out_mm_data.get("feature_attention_mask") chunk_counts = out_mm_data.get("chunk_counts") def get_replacement_audioflamingo3(item_idx: int): if feature_attention_mask is not None: if chunk_counts is not None: counts = ( chunk_counts.tolist() if isinstance(chunk_counts, torch.Tensor) else chunk_counts ) start_idx = sum(counts[:item_idx]) count = counts[item_idx] end_idx = start_idx + count if isinstance(feature_attention_mask, list): mask_list = feature_attention_mask[start_idx:end_idx] if len(mask_list) > 0 and isinstance( mask_list[0], torch.Tensor ): mask = torch.stack(mask_list) else: mask = torch.tensor(mask_list) else: mask = feature_attention_mask[start_idx:end_idx] else: # feature_attention_mask is list[Tensor] or Tensor if isinstance(feature_attention_mask, list): mask = feature_attention_mask[item_idx] else: mask = feature_attention_mask[item_idx].unsqueeze(0) # mask shape: (num_chunks, 3000) input_lengths = mask.sum(-1) conv_lengths = (input_lengths - 1) // 2 + 1 audio_output_lengths = (conv_lengths - 2) // 2 + 1 num_features = audio_output_lengths.sum().item() else: audio_embeds = out_mm_data["audio_embeds"][item_idx] num_features = audio_embeds.shape[0] if num_features == 0: raise ValueError("Audio is too short") audio_tokens = [audio_token_id] * int(num_features) return PromptUpdateDetails.select_token_id( audio_tokens, embed_token_id=audio_token_id, ) return [ PromptReplacement( modality="audio", target=audio_token, replacement=get_replacement_audioflamingo3, ) ] @MULTIMODAL_REGISTRY.register_processor( AudioFlamingo3MultiModalProcessor, info=AudioFlamingo3ProcessingInfo, dummy_inputs=AudioFlamingo3DummyInputsBuilder, ) class AudioFlamingo3ForConditionalGeneration( nn.Module, SupportsMultiModal, SupportsPP, SupportsLoRA ): """ AudioFlamingo3 model for conditional generation. This model integrates a Whisper-based audio encoder with a Qwen2 language model. It supports multi-chunk audio processing. """ packed_modules_mapping = { "qkv_proj": ["q_proj", "k_proj", "v_proj"], "gate_up_proj": ["gate_proj", "up_proj"], } def get_mm_mapping(self) -> MultiModelKeys: """ Get the module prefix in multimodal models """ return MultiModelKeys.from_string_field( language_model="language_model.", connector="multi_modal_projector.", tower_model="audio_tower.", ) def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config multimodal_config = vllm_config.model_config.multimodal_config self.config = config self.multimodal_config = multimodal_config self.quant_config = quant_config with self._mark_tower_model(vllm_config, "audio"): self.audio_tower = AudioFlamingo3Encoder( config.audio_config, ) self.multi_modal_projector = AudioFlamingo3MultiModalProjector(config) with self._mark_language_model(vllm_config): self.language_model = init_vllm_registered_model( vllm_config=vllm_config, hf_config=config.text_config, prefix=maybe_prefix(prefix, "language_model"), architectures=["Qwen2ForCausalLM"], ) self.make_empty_intermediate_tensors = ( self.language_model.make_empty_intermediate_tensors ) def _parse_and_validate_audio_input( self, **kwargs: object ) -> AudioFlamingo3Inputs | None: input_features = kwargs.pop("input_features", None) audio_embeds = kwargs.pop("audio_embeds", None) feature_attention_mask = kwargs.pop("feature_attention_mask", None) chunk_counts = kwargs.pop("chunk_counts", None) if input_features is None and audio_embeds is None: return None if audio_embeds is not None: return AudioFlamingo3EmbeddingInputs( type="audio_embeds", audio_embeds=audio_embeds ) if input_features is not None: return AudioFlamingo3FeatureInputs( type="audio_features", input_features=input_features, feature_attention_mask=feature_attention_mask, chunk_counts=chunk_counts, ) raise AssertionError("This line should be unreachable.") def _process_audio_input( self, audio_input: AudioFlamingo3Inputs ) -> torch.Tensor | tuple[torch.Tensor, ...]: if audio_input["type"] == "audio_embeds": audio_embeds = audio_input["audio_embeds"] return tuple(audio_embeds) input_features = audio_input["input_features"] feature_attention_mask = audio_input["feature_attention_mask"] chunk_counts = audio_input.get("chunk_counts") if isinstance(input_features, list): input_features = torch.cat(input_features, dim=0) feature_attention_mask = torch.cat(feature_attention_mask, dim=0) if chunk_counts is None: chunk_counts = [1] * input_features.shape[0] elif isinstance(chunk_counts, torch.Tensor): chunk_counts = chunk_counts.tolist() elif ( isinstance(chunk_counts, list) and chunk_counts and isinstance(chunk_counts[0], torch.Tensor) ): chunk_counts = [c.item() for c in chunk_counts] # Calculate output lengths input_lengths = feature_attention_mask.sum(-1) # Conv downsampling conv_lengths = (input_lengths - 1) // 2 + 1 # AvgPool downsampling audio_output_lengths = (conv_lengths - 2) // 2 + 1 batch_size, _, max_mel_seq_len = input_features.shape # Calculate max_seq_len after convs (before pooling) for attention mask max_seq_len = (max_mel_seq_len - 1) // 2 + 1 # Create a sequence tensor of shape (batch_size, max_seq_len) seq_range = ( torch.arange( 0, max_seq_len, dtype=conv_lengths.dtype, device=conv_lengths.device, ) .unsqueeze(0) .expand(batch_size, max_seq_len) ) lengths_expand = conv_lengths.unsqueeze(-1).expand(batch_size, max_seq_len) # Create mask padding_mask = seq_range >= lengths_expand audio_attention_mask_ = padding_mask.view(batch_size, 1, 1, max_seq_len).expand( batch_size, 1, max_seq_len, max_seq_len ) audio_attention_mask = audio_attention_mask_.to( dtype=self.audio_tower.conv1.weight.dtype, device=self.audio_tower.conv1.weight.device, ) audio_attention_mask[audio_attention_mask_] = float("-inf") # Forward pass audio_features = self.audio_tower( input_features, attention_mask=audio_attention_mask ) # Project audio_features = self.multi_modal_projector(audio_features) # Masking after pooling num_audios, max_audio_tokens, embed_dim = audio_features.shape audio_output_lengths = audio_output_lengths.unsqueeze(1) audio_features_mask = ( torch.arange(max_audio_tokens) .expand(num_audios, max_audio_tokens) .to(audio_output_lengths.device) < audio_output_lengths ) masked_audio_features = audio_features[audio_features_mask].view(-1, embed_dim) # Split to tuple of embeddings for individual audio input. chunk_embeddings = torch.split( masked_audio_features, audio_output_lengths.flatten().tolist() ) grouped_embeddings = [] current_idx = 0 for count in chunk_counts: audio_chunks = chunk_embeddings[current_idx : current_idx + count] grouped_embeddings.append(torch.cat(audio_chunks, dim=0)) current_idx += count return tuple(grouped_embeddings) def embed_multimodal(self, **kwargs: object) -> MultiModalEmbeddings: audio_input = self._parse_and_validate_audio_input(**kwargs) if audio_input is None: return [] masked_audio_features = self._process_audio_input(audio_input) return masked_audio_features def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs: object, ) -> torch.Tensor | IntermediateTensors: if intermediate_tensors is not None: inputs_embeds = None hidden_states = self.language_model.model( input_ids, positions, intermediate_tensors, inputs_embeds=inputs_embeds, ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: return self.language_model.compute_logits(hidden_states) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader(self) return loader.load_weights(weights)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/audioflamingo3.py", "license": "Apache License 2.0", "lines": 548, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/quantization/test_scaled_mm_kernel_selection.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests for ScaledMM kernel selection logic (CPU-only) Run `pytest tests/kernels/quantization/test_scaled_mm_kernel_selection.py`. """ import inspect from abc import ABC import pytest from vllm.model_executor.kernels.linear import ( AiterInt8ScaledMMLinearKernel, CPUInt8ScaledMMLinearKernel, Int8ScaledMMLinearLayerConfig, ScaledMMLinearKernel, ) pytestmark = pytest.mark.cpu_test def test_is_supported_is_abstract(): """Test that is_supported() is properly defined as abstract.""" assert issubclass(ScaledMMLinearKernel, ABC) assert hasattr(ScaledMMLinearKernel, "is_supported") def test_cpu_kernel_implements_is_supported(): """Test that CPUInt8ScaledMMLinearKernel implements is_supported() method.""" assert hasattr(CPUInt8ScaledMMLinearKernel, "is_supported"), ( "CPUInt8ScaledMMLinearKernel missing is_supported() method" ) # Verify it's a classmethod by checking if it can be called with the class # and by checking the method type assert inspect.ismethod( CPUInt8ScaledMMLinearKernel.is_supported ) or inspect.isfunction(CPUInt8ScaledMMLinearKernel.is_supported), ( "CPUInt8ScaledMMLinearKernel.is_supported() should be a classmethod" ) # Verify it can be called as a classmethod result, reason = CPUInt8ScaledMMLinearKernel.is_supported() assert isinstance(result, bool), "is_supported() should return a bool" assert reason is None or isinstance(reason, str), "reason should be str or None" def test_aiter_kernel_implements_is_supported(): """Test that AiterInt8ScaledMMLinearKernel implements is_supported() method.""" assert hasattr(AiterInt8ScaledMMLinearKernel, "is_supported"), ( "AiterInt8ScaledMMLinearKernel missing is_supported() method" ) # Verify it's a classmethod by checking if it can be called with the class # and by checking the method type assert inspect.ismethod( AiterInt8ScaledMMLinearKernel.is_supported ) or inspect.isfunction(AiterInt8ScaledMMLinearKernel.is_supported), ( "AiterInt8ScaledMMLinearKernel.is_supported() should be a classmethod" ) # Verify it can be called as a classmethod # (will return False on CPU, which is expected) result, reason = AiterInt8ScaledMMLinearKernel.is_supported() assert isinstance(result, bool), "is_supported() should return a bool" assert reason is None or isinstance(reason, str), "reason should be str or None" # On CPU, it should return False with a reason about requiring ROCm # This validates the method works correctly even on non-ROCm platforms def test_cpu_kernel_accepts_all_configs(): """Test that CPUInt8ScaledMMLinearKernel accepts all config combinations.""" configs = [ Int8ScaledMMLinearLayerConfig( is_channelwise=False, is_static_input_scheme=True, input_symmetric=True, ), Int8ScaledMMLinearLayerConfig( is_channelwise=True, is_static_input_scheme=False, input_symmetric=False, ), ] for config in configs: can_impl, reason = CPUInt8ScaledMMLinearKernel.can_implement(config) assert can_impl, ( f"CPUInt8ScaledMMLinearKernel should accept config {config}: {reason}" )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/quantization/test_scaled_mm_kernel_selection.py", "license": "Apache License 2.0", "lines": 73, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/v1/worker/workspace.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import inspect import os from itertools import accumulate from math import prod import torch import vllm.envs as envs from vllm.logger import init_logger from vllm.utils.math_utils import round_up from vllm.v1.worker.ubatching import dbo_current_ubatch_id logger = init_logger(__name__) def _compute_bytes(shape: tuple[int, ...], dtype: torch.dtype) -> int: return prod(shape) * dtype.itemsize # Constants _MB = 1024**2 _GiB = 1024**3 # Global workspace manager instance _manager: "WorkspaceManager | None" = None class WorkspaceManager: """Manager for workspace allocation. Manages workspace buffers for DBO (Dual Batch Overlap) execution. Can be locked to prevent further growth during execution. """ def __init__(self, device: torch.device, num_ubatches: int | None = None): self._device = device # Cache num ubatches at init based on configuration (default to 1) self._num_ubatches = num_ubatches if num_ubatches is not None else 1 self._current_workspaces: list[torch.Tensor | None] = [None, None] self._locked: bool = False @staticmethod def _workspace_size_bytes(workspace: torch.Tensor | None) -> int: """Get size of workspace in bytes.""" if workspace is None: return 0 return workspace.numel() * workspace.element_size() def lock(self) -> None: """Lock the workspace to prevent further growth. After locking, any attempt to allocate a larger workspace will raise an assertion error. This ensures workspace size is fixed during execution. """ self._locked = True if envs.VLLM_DEBUG_WORKSPACE: logger.info( "[WORKSPACE DEBUG] Workspace locked. Current sizes: %s", [ self._workspace_size_bytes(ws) / _MB for ws in self._current_workspaces if ws is not None ], ) def unlock(self) -> None: """Unlock the workspace to allow growth. This is used during elastic EP scaling when the workspace size needs to grow due to changes in the number of experts. """ self._locked = False if envs.VLLM_DEBUG_WORKSPACE: logger.info( "[WORKSPACE DEBUG] Workspace unlocked. Current sizes: %s", [ self._workspace_size_bytes(ws) / _MB for ws in self._current_workspaces if ws is not None ], ) def is_locked(self) -> bool: """Check if workspace is locked.""" return self._locked def get_simultaneous( self, *shapes_and_dtypes: tuple[tuple[int, ...], torch.dtype] ) -> list[torch.Tensor]: """Get multiple workspace tensors simultaneously from a single allocation. Args: *shapes_and_dtypes: One or more (shape, dtype) tuples. Returns: List of tensor views into the workspace buffer, one per shape/dtype pair. """ actual_bytes = [_compute_bytes(s, d) for s, d in shapes_and_dtypes] aligned_bytes = [round_up(actual, 256) for actual in actual_bytes] total_bytes = sum(aligned_bytes) # Calculate cumulative offsets using itertools.accumulate offsets = list(accumulate([0] + aligned_bytes[:-1])) current_workspace = self._ensure_workspace_size(total_bytes) return [ current_workspace[offsets[i] : offsets[i] + actual_bytes[i]] .view(shapes_and_dtypes[i][1]) .reshape(shapes_and_dtypes[i][0]) for i in range(len(shapes_and_dtypes)) ] def _ensure_workspace_size(self, required_bytes: int) -> torch.Tensor: """Ensure workspace is allocated and large enough, return current workspace. Args: required_bytes: The number of bytes required. Returns: The current workspace tensor. """ ubatch_id = dbo_current_ubatch_id() current_workspace = self._current_workspaces[ubatch_id] current_size = self._workspace_size_bytes(current_workspace) if current_size < required_bytes: def get_caller_info() -> str: """Find first frame outside WorkspaceManager.""" curr_frame = inspect.currentframe() if curr_frame is None: return "unknown" # Walk up the stack skipping WorkspaceManager frames curr_frame = curr_frame.f_back while curr_frame is not None: # TODO: This only catches instance methods (self), missing # classmethods and staticmethods. Once Python 3.11+ is the # minimum supported version, use co_qualname instead: # qualname = curr_frame.f_code.co_qualname # if qualname.startswith("WorkspaceManager."): if isinstance(curr_frame.f_locals.get("self"), WorkspaceManager): curr_frame = curr_frame.f_back continue filename = os.path.basename(curr_frame.f_code.co_filename) return ( f"{filename}:{curr_frame.f_lineno}:{curr_frame.f_code.co_name}" ) return "unknown" if self._locked: raise AssertionError( f"Workspace is locked but allocation from '{get_caller_info()}' " f"requires {required_bytes / _MB:.2f} MB, current size is " f"{current_size / _MB:.2f} MB. " "Workspace growth is not allowed after locking." ) for ubatch_id in range(self._num_ubatches): current_workspace = self._current_workspaces[ubatch_id] if ( current_workspace is None or self._workspace_size_bytes(current_workspace) < required_bytes ): # Delete old tensor before allocating new one to avoid # memory spike from resize_(). resize_() allocates new # memory before freeing old, which can cause OOM. # Must clear the list reference first since local var # is just a copy of the reference. self._current_workspaces[ubatch_id] = None del current_workspace self._current_workspaces[ubatch_id] = torch.empty( (required_bytes,), dtype=torch.uint8, device=self._device ) if envs.VLLM_DEBUG_WORKSPACE: logger.info( "[WORKSPACE DEBUG] Resized workspace from '%s': %.2f MB -> " "%.2f MB (%d ubatches, total memory %.2f MB)", get_caller_info(), current_size / _MB, required_bytes / _MB, self._num_ubatches, required_bytes * self._num_ubatches / _MB, ) current_workspace = self._current_workspaces[dbo_current_ubatch_id()] return current_workspace def is_workspace_manager_initialized() -> bool: """Check if workspace manager has been initialized. Returns: True if workspace manager is initialized, False otherwise. """ return _manager is not None def current_workspace_manager() -> "WorkspaceManager": """Get the current workspace manager instance. Raises: AssertionError: If workspace manager has not been initialized. """ assert _manager is not None, ( "WorkspaceManager not initialized. Call init_workspace_manager() " "with a device before using workspace functions." ) return _manager def init_workspace_manager( device: torch.device, num_ubatches: int | None = None ) -> None: """Initialize the workspace manager with a device. Must be called before using any workspace functions. Typically called from GPUModelRunner.__init__. Args: device: The device to allocate workspace on. num_ubatches: Number of micro-batches. Defaults to 1. """ global _manager if _manager is not None: logger.warning( "WorkspaceManager already initialized on device %s, " "reinitializing on device %s", _manager._device, device, ) _manager = WorkspaceManager(device, num_ubatches) def lock_workspace() -> None: """Lock the workspace to prevent further growth. After calling this function, any attempt to allocate a workspace larger than the current size will raise an AssertionError. This ensures that workspace size is fixed during execution and prevents unexpected memory allocations in the hot path. Example: # During initialization init_workspace_manager(device) reserve_workspace(shape1, dtype1) reserve_workspace(shape2, dtype2) # Lock after warmup/profiling lock_workspace() # Now all get_workspace calls must fit in pre-allocated size """ current_workspace_manager().lock() def unlock_workspace() -> None: """Unlock the workspace to allow growth. This is used during elastic EP scaling when the workspace size needs to grow due to changes in the number of experts. After scaling operations complete, lock_workspace() should be called again to prevent unexpected allocations. """ current_workspace_manager().unlock() def reset_workspace_manager() -> None: """Reset the workspace manager to uninitialized state. This is primarily intended for testing purposes to allow tests to reinitialize the workspace manager cleanly. """ global _manager _manager = None
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/worker/workspace.py", "license": "Apache License 2.0", "lines": 223, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/models/multimodal/processing/test_gemma3.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from vllm.model_executor.models.gemma3n_audio_utils import ( adjust_audio_features_to_expected_length, ) from vllm.multimodal import MULTIMODAL_REGISTRY from ....conftest import ImageTestAssets from ...utils import build_model_context # Gemma3 (image) model GEMMA3_MODEL_ID = "google/gemma-3-4b-it" # Gemma3n (multimodal with audio) model GEMMA3N_MODEL_ID = "google/gemma-3n-E2B-it" # Expected audio tokens for Gemma3n (audio_soft_tokens_per_image) GEMMA3N_EXPECTED_AUDIO_TOKENS = 188 class TestGemma3nAudioTensorLogic: """CPU-based tests for Gemma3n audio feature tensor manipulation. These tests validate the padding/truncation logic in adjust_audio_features_to_expected_length() which fixes the integer overflow in _process_audio_input when audio_seq_len > 188. """ def test_padding_when_audio_short(self): """Test that short audio is padded to expected length.""" batch_size, seq_len, embed_dim = 1, 100, 256 expected_tokens = GEMMA3N_EXPECTED_AUDIO_TOKENS audio_features = torch.randn(batch_size, seq_len, embed_dim) padding_embs = torch.zeros(1, 1, embed_dim) result, tokens_truncated = adjust_audio_features_to_expected_length( audio_features, expected_tokens, padding_embs ) assert result.shape == (batch_size, expected_tokens, embed_dim) assert tokens_truncated == 0 # First 100 tokens should be original, rest should be padding (zeros) assert torch.allclose(result[:, :seq_len, :], audio_features) assert torch.allclose( result[:, seq_len:, :], torch.zeros(batch_size, expected_tokens - seq_len, embed_dim), ) def test_truncation_when_audio_long(self): """Test that long audio is truncated to expected length. This is the key test for the overflow fix. Previously, when audio_seq_len > expected_tokens, the code would compute a negative padding value causing: RuntimeError: numel: integer multiplication overflow """ batch_size, seq_len, embed_dim = 1, 192, 256 # 192 > 188 expected_tokens = GEMMA3N_EXPECTED_AUDIO_TOKENS audio_features = torch.randn(batch_size, seq_len, embed_dim) padding_embs = torch.zeros(1, 1, embed_dim) result, tokens_truncated = adjust_audio_features_to_expected_length( audio_features, expected_tokens, padding_embs ) assert result.shape == (batch_size, expected_tokens, embed_dim) assert tokens_truncated == seq_len - expected_tokens # 192 - 188 = 4 # Result should be first 188 tokens of original assert torch.allclose(result, audio_features[:, :expected_tokens, :]) def test_no_change_when_exact_length(self): """Test that exact-length audio passes through unchanged.""" batch_size, embed_dim = 1, 256 expected_tokens = GEMMA3N_EXPECTED_AUDIO_TOKENS audio_features = torch.randn(batch_size, expected_tokens, embed_dim) padding_embs = torch.zeros(1, 1, embed_dim) result, tokens_truncated = adjust_audio_features_to_expected_length( audio_features, expected_tokens, padding_embs ) assert result.shape == audio_features.shape assert tokens_truncated == 0 assert torch.allclose(result, audio_features) def test_original_bug_would_fail(self): """Verify the original buggy implementation would cause overflow. The original code always tried to pad, which fails when audio_seq_len > expected_tokens because expand() gets negative size. """ batch_size, seq_len, embed_dim = 1, 192, 256 expected_tokens = GEMMA3N_EXPECTED_AUDIO_TOKENS padding_embs = torch.zeros(1, 1, embed_dim) # Original buggy logic (always pads, never truncates) extra_padding_tokens = expected_tokens - seq_len # = -4 (negative!) with pytest.raises(RuntimeError): # This should fail with negative size error padding_embs.expand(batch_size, extra_padding_tokens, embed_dim) @pytest.mark.parametrize( "seq_len", [50, 100, 150, 187, 188, 189, 192, 200, 300], ) def test_various_audio_lengths(self, seq_len: int): """Test padding/truncation with various audio lengths.""" batch_size, embed_dim = 1, 256 expected_tokens = GEMMA3N_EXPECTED_AUDIO_TOKENS audio_features = torch.randn(batch_size, seq_len, embed_dim) padding_embs = torch.zeros(1, 1, embed_dim) # Should not raise any errors result, tokens_truncated = adjust_audio_features_to_expected_length( audio_features, expected_tokens, padding_embs ) # Output should always be expected_tokens length assert result.shape == (batch_size, expected_tokens, embed_dim) # Verify truncation count is correct if seq_len > expected_tokens: assert tokens_truncated == seq_len - expected_tokens else: assert tokens_truncated == 0 def test_batch_processing(self): """Test that batch processing works correctly.""" batch_size, seq_len, embed_dim = 4, 192, 256 expected_tokens = GEMMA3N_EXPECTED_AUDIO_TOKENS audio_features = torch.randn(batch_size, seq_len, embed_dim) padding_embs = torch.zeros(1, 1, embed_dim) result, tokens_truncated = adjust_audio_features_to_expected_length( audio_features, expected_tokens, padding_embs ) assert result.shape == (batch_size, expected_tokens, embed_dim) assert tokens_truncated == seq_len - expected_tokens @pytest.mark.parametrize("model_id", [GEMMA3_MODEL_ID]) @pytest.mark.parametrize("mm_processor_kwargs", [{}]) def test_get_image_size_with_most_features( image_assets: ImageTestAssets, model_id: str, mm_processor_kwargs: dict[str, object], ): ctx = build_model_context( model_id, mm_processor_kwargs={"do_pan_and_scan": True}, limit_mm_per_prompt={"image": 1}, ) processor = MULTIMODAL_REGISTRY.create_processor(ctx.model_config) hf_processor = processor.info.get_hf_processor(**mm_processor_kwargs) max_image_size = processor.info.get_image_size_with_most_features() max_tokens = processor.info.get_num_image_tokens( image_width=max_image_size.width, image_height=max_image_size.height, processor=hf_processor, mm_kwargs=mm_processor_kwargs, ) prompt = "<start_of_image>" image_seq_length = hf_processor.image_seq_length for asset in image_assets: mm_data = {"image": [asset.pil_image]} processed_inputs = processor( prompt, mm_items=processor.info.parse_mm_data(mm_data), hf_processor_mm_kwargs=mm_processor_kwargs, ) mm_kwargs_data = processed_inputs["mm_kwargs"].get_data() num_patches_tensor = mm_kwargs_data["num_patches"] tokens = int(num_patches_tensor.item()) * image_seq_length assert tokens <= max_tokens
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/multimodal/processing/test_gemma3.py", "license": "Apache License 2.0", "lines": 147, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/entrypoints/openai/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json from collections.abc import AsyncGenerator from typing import Any from vllm.entrypoints.openai.chat_completion.protocol import ( ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionStreamResponse, ChatMessage, ) from vllm.entrypoints.openai.engine.protocol import ( UsageInfo, ) async def accumulate_streaming_response( stream_generator: AsyncGenerator[str, None], ) -> ChatCompletionResponse: """ Accumulate streaming SSE chunks into a complete ChatCompletionResponse. This helper parses the SSE format and builds up the complete response by combining all the delta chunks. """ accumulated_content = "" accumulated_reasoning = None accumulated_tool_calls: list[dict[str, Any]] = [] role = None finish_reason = None response_id = None created = None model = None index = 0 async for chunk_str in stream_generator: # Skip empty lines and [DONE] marker if not chunk_str.strip() or chunk_str.strip() == "data: [DONE]": continue # Parse SSE format: "data: {json}\n\n" if chunk_str.startswith("data: "): json_str = chunk_str[6:].strip() try: chunk_data = json.loads(json_str) # print(f"DEBUG: Parsed chunk_data: {chunk_data}") chunk = ChatCompletionStreamResponse(**chunk_data) # Store metadata from first chunk if response_id is None: response_id = chunk.id created = chunk.created model = chunk.model # Process each choice in the chunk for choice in chunk.choices: if choice.delta.role: role = choice.delta.role if choice.delta.content: accumulated_content += choice.delta.content if choice.delta.reasoning: if accumulated_reasoning is None: accumulated_reasoning = "" accumulated_reasoning += choice.delta.reasoning if choice.delta.tool_calls: # Accumulate tool calls for tool_call_delta in choice.delta.tool_calls: # Find or create the tool call at this index while len(accumulated_tool_calls) <= tool_call_delta.index: accumulated_tool_calls.append( { "id": None, "type": "function", "function": {"name": "", "arguments": ""}, } ) if tool_call_delta.id: accumulated_tool_calls[tool_call_delta.index]["id"] = ( tool_call_delta.id ) if tool_call_delta.function: if tool_call_delta.function.name: accumulated_tool_calls[tool_call_delta.index][ "function" ]["name"] += tool_call_delta.function.name if tool_call_delta.function.arguments: accumulated_tool_calls[tool_call_delta.index][ "function" ]["arguments"] += tool_call_delta.function.arguments if choice.finish_reason: finish_reason = choice.finish_reason if choice.index is not None: index = choice.index except json.JSONDecodeError: continue # Build the final message message_kwargs = { "role": role or "assistant", "content": accumulated_content if accumulated_content else None, "reasoning": accumulated_reasoning, } # Only include tool_calls if there are any if accumulated_tool_calls: message_kwargs["tool_calls"] = [ {"id": tc["id"], "type": tc["type"], "function": tc["function"]} for tc in accumulated_tool_calls ] message = ChatMessage(**message_kwargs) # Build the final response choice = ChatCompletionResponseChoice( index=index, message=message, finish_reason=finish_reason or "stop", ) # Create usage info (with dummy values for tests) usage = UsageInfo( prompt_tokens=0, completion_tokens=0, total_tokens=0, ) response = ChatCompletionResponse( id=response_id or "chatcmpl-test", object="chat.completion", created=created or 0, model=model or "test-model", choices=[choice], usage=usage, ) return response def verify_harmony_messages( messages: list[Any], expected_messages: list[dict[str, Any]] ): assert len(messages) == len(expected_messages) for msg, expected in zip(messages, expected_messages): if "role" in expected: assert msg.author.role == expected["role"] if "author_name" in expected: assert msg.author.name == expected["author_name"] if "channel" in expected: assert msg.channel == expected["channel"] if "recipient" in expected: assert msg.recipient == expected["recipient"] if "content" in expected: assert msg.content[0].text == expected["content"] if "content_type" in expected: assert msg.content_type == expected["content_type"] if "tool_definitions" in expected: # Check that the tool definitions match the expected list of tool names actual_tools = [t.name for t in msg.content[0].tools["functions"].tools] assert actual_tools == expected["tool_definitions"] def verify_chat_response( response: ChatCompletionResponse, content: str | None = None, reasoning: str | None = None, tool_calls: list[tuple[str, str]] | None = None, ): assert len(response.choices) == 1 message = response.choices[0].message if content is not None: assert message.content == content else: assert not message.content if reasoning is not None: assert message.reasoning == reasoning else: assert not message.reasoning if tool_calls: assert message.tool_calls is not None assert len(message.tool_calls) == len(tool_calls) for tc, (expected_name, expected_args) in zip(message.tool_calls, tool_calls): assert tc.function.name == expected_name assert tc.function.arguments == expected_args else: assert not message.tool_calls
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/utils.py", "license": "Apache License 2.0", "lines": 167, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/distributed/test_eplb_fused_moe_layer_dep_nvfp4.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Test that the interaction between EPLB and FusedMoE Layer is okay for DP w/ NVFP4 from dataclasses import dataclass import pytest import torch from tests.kernels.moe.utils import make_test_quant_config from vllm.config import VllmConfig, set_current_vllm_config from vllm.distributed.eplb.rebalance_execute import rearrange_expert_weights_inplace from vllm.distributed.parallel_state import ( ensure_model_parallel_initialized, get_dp_group, ) from vllm.forward_context import set_forward_context from vllm.model_executor.layers.fused_moe.layer import FusedMoE from vllm.model_executor.layers.quantization.modelopt import ( ModelOptNvFp4Config, ModelOptNvFp4FusedMoE, ) from .eplb_utils import distributed_run, set_env_vars_and_device @dataclass class TestConfig: num_layers: int num_experts: int num_local_experts: int num_topk: int hidden_size: int intermediate_size: int num_tokens: int def make_fused_moe_layer( rank: int, layer_idx: int, test_config: TestConfig, ) -> FusedMoE: quant_config = None device = torch.device(f"cuda:{rank}") quant_config = ModelOptNvFp4Config( is_checkpoint_nvfp4_serialized=True, kv_cache_quant_algo=None, exclude_modules=[], ) fml = FusedMoE( num_experts=test_config.num_experts, top_k=test_config.num_topk, hidden_size=test_config.hidden_size, intermediate_size=test_config.intermediate_size, prefix=f"dummy_layer_{layer_idx}", activation="silu", is_act_and_mul=True, params_dtype=torch.bfloat16, quant_config=quant_config, ) nvfp4_fused_moe = ModelOptNvFp4FusedMoE(quant_config, fml) nvfp4_fused_moe.create_weights( fml, test_config.num_local_experts, test_config.hidden_size, test_config.intermediate_size, params_dtype=torch.uint8, global_num_experts=test_config.num_experts, ) fml = fml.to(device) w1_q, w2_q, quant_config = make_test_quant_config( test_config.num_local_experts, test_config.intermediate_size, test_config.hidden_size, in_dtype=torch.bfloat16, quant_dtype="nvfp4", block_shape=None, per_act_token_quant=False, ) fml.w13_weight.data = w1_q fml.w2_weight.data = w2_q fml.w2_input_scale.data = torch.randn_like(fml.w2_input_scale.data) / 5 fml.w13_input_scale.data = torch.randn_like(fml.w13_input_scale.data) / 5 fml.w2_weight_scale_2.data = torch.randn_like(fml.w2_weight_scale_2.data) / 5 fml.w13_weight_scale_2.data = torch.randn_like(fml.w13_weight_scale_2.data) / 5 fml.w2_weight_scale.data = ( torch.randn(fml.w2_weight_scale.data.shape, device=device) / 5 ).to(fml.w2_weight_scale.data.dtype) fml.w13_weight_scale.data = ( torch.randn(fml.w13_weight_scale.data.shape, device=device) / 5 ).to(fml.w13_weight_scale.data.dtype) nvfp4_fused_moe.process_weights_after_loading(fml) fml.maybe_init_modular_kernel() return fml def _test_eplb_fml(env, world_size: int, test_config: TestConfig): set_env_vars_and_device(env) vllm_config = VllmConfig() vllm_config.parallel_config.data_parallel_size = world_size vllm_config.parallel_config.enable_expert_parallel = True with set_current_vllm_config(vllm_config): ensure_model_parallel_initialized( tensor_model_parallel_size=1, pipeline_model_parallel_size=1 ) ep_group = get_dp_group().cpu_group ep_rank = torch.distributed.get_rank() device = torch.device(f"cuda:{ep_rank}") fml_layers = [ make_fused_moe_layer(ep_rank, layer_idx, test_config).to(device) for layer_idx in range(test_config.num_layers) ] rank_expert_weights = [fml.get_expert_weights() for fml in fml_layers] hidden_states = [] router_logits = [] for layer_idx in range(test_config.num_layers): hidden_states.append( torch.randn( (test_config.num_tokens, test_config.hidden_size), dtype=torch.bfloat16, device=device, ) ) router_logits.append( torch.randn( (test_config.num_tokens, test_config.num_experts), dtype=torch.bfloat16, device=device, ) ) out_before_shuffle = [] with set_forward_context( {}, num_tokens=test_config.num_tokens, num_tokens_across_dp=torch.tensor( [test_config.num_tokens] * world_size, device="cpu", dtype=torch.int ), vllm_config=vllm_config, ): for lidx, fml in enumerate(fml_layers): out_before_shuffle.append( fml(hidden_states[lidx].clone(), router_logits[lidx].clone()) ) indices = torch.zeros( test_config.num_layers, test_config.num_experts, dtype=torch.long ) for lidx in range(test_config.num_layers): indices[lidx] = torch.Tensor(range(test_config.num_experts)) shuffled_indices = torch.zeros_like(indices) for lidx in range(test_config.num_layers): shuffled_indices[lidx] = torch.randperm(test_config.num_experts) rearrange_expert_weights_inplace( indices, shuffled_indices, rank_expert_weights, ep_group, is_profile=False, ) num_global_experts = test_config.num_experts logical_to_physical_map_list = [] for lidx, fml in enumerate(fml_layers): physical_to_logical_map = shuffled_indices[lidx].to(device) logical_to_physical_map = torch.empty( (num_global_experts,), dtype=torch.int32, device=device ) logical_to_physical_map[physical_to_logical_map] = torch.arange( 0, num_global_experts, dtype=torch.int32, device=device ) logical_to_physical_map_list.append( logical_to_physical_map.reshape(num_global_experts, 1) ) logical_to_physical_map = torch.stack(logical_to_physical_map_list) for lidx, fml in enumerate(fml_layers): logical_replica_count = torch.ones( (test_config.num_layers, num_global_experts), dtype=torch.int32, device=device, ) fml.enable_eplb = True fml.set_eplb_state( lidx, torch.zeros( (test_config.num_layers, num_global_experts), dtype=torch.int32, device=device, ), logical_to_physical_map, logical_replica_count, ) out_after_shuffle = [] with set_forward_context( {}, num_tokens=test_config.num_tokens, num_tokens_across_dp=torch.tensor( [test_config.num_tokens] * world_size, device="cpu", dtype=torch.int ), vllm_config=vllm_config, ): for lidx, fml in enumerate(fml_layers): out_after_shuffle.append( fml(hidden_states[lidx].clone(), router_logits[lidx].clone()) ) for lidx in range(test_config.num_layers): torch.testing.assert_close( out_before_shuffle[lidx], out_after_shuffle[lidx], atol=1e-1, rtol=1e-1 ) @pytest.mark.parametrize("world_size", [2, 4]) @pytest.mark.parametrize("num_layers", [8]) @pytest.mark.parametrize("num_experts", [32]) @pytest.mark.parametrize("hidden_size", [256]) @pytest.mark.parametrize("intermediate_size", [256]) @pytest.mark.parametrize("num_tokens", [256]) @pytest.mark.parametrize("backend", ["latency", "throughput"]) def test_eplb_fml( world_size: int, num_layers: int, num_experts: int, hidden_size: int, intermediate_size: int, num_tokens: int, backend: str, monkeypatch, ): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_FP4", "1") monkeypatch.setenv("VLLM_FLASHINFER_MOE_BACKEND", backend) if torch.cuda.device_count() < world_size: pytest.skip(f"Need at least {world_size} GPUs to run the test") num_local_experts = num_experts // world_size num_topk = 4 test_config = TestConfig( num_layers=num_layers, num_experts=num_experts, num_local_experts=num_local_experts, num_topk=num_topk, hidden_size=hidden_size, intermediate_size=intermediate_size, num_tokens=num_tokens, ) distributed_run( _test_eplb_fml, world_size, test_config, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/distributed/test_eplb_fused_moe_layer_dep_nvfp4.py", "license": "Apache License 2.0", "lines": 233, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/kv_connector/unit/test_lmcache_connector.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from unittest.mock import MagicMock import pytest from vllm.distributed.kv_events import BlockStored from vllm.distributed.kv_transfer.kv_connector.v1.lmcache_connector import ( LMCacheConnectorV1, LMCacheKVEvents, ) from vllm.v1.outputs import KVConnectorOutput @pytest.fixture def mock_lmcache_engine_event(): """Create a mock event object that mimics what the lmcache engine returns.""" class MockEvent: def __init__( self, block_hashes, parent_block_hash, token_ids, lora_id, block_size, medium, lora_name, ): self.block_hashes = block_hashes self.parent_block_hash = parent_block_hash self.token_ids = token_ids self.lora_id = lora_id self.block_size = block_size self.medium = medium self.lora_name = lora_name return MockEvent( block_hashes=["hash1", "hash2"], parent_block_hash="parent_hash", token_ids=[1, 2, 3, 4], lora_id=None, block_size=16, medium="GPU", lora_name=None, ) @pytest.fixture def mock_connector(): """Create a mock LMCacheConnectorV1 instance with mocked dependencies.""" connector = MagicMock(spec=LMCacheConnectorV1) connector._kv_cache_events = None connector._lmcache_engine = MagicMock() # Make the methods use the real implementation connector.get_kv_connector_kv_cache_events = ( LMCacheConnectorV1.get_kv_connector_kv_cache_events.__get__( connector, LMCacheConnectorV1 ) ) connector.update_connector_output = ( LMCacheConnectorV1.update_connector_output.__get__( connector, LMCacheConnectorV1 ) ) connector.take_events = LMCacheConnectorV1.take_events.__get__( connector, LMCacheConnectorV1 ) return connector class TestGetKVConnectorKVCacheEvents: """Test get_kv_connector_kv_cache_events method.""" def test_returns_none_when_no_events(self, mock_connector): """Test that None is returned when lmcache engine has no events.""" mock_connector._lmcache_engine.get_kv_events.return_value = None result = mock_connector.get_kv_connector_kv_cache_events() assert result is None mock_connector._lmcache_engine.get_kv_events.assert_called_once() def test_returns_none_when_empty_list(self, mock_connector): """Test that None is returned when lmcache engine returns empty list.""" mock_connector._lmcache_engine.get_kv_events.return_value = [] result = mock_connector.get_kv_connector_kv_cache_events() assert result is None def test_converts_single_event(self, mock_connector, mock_lmcache_engine_event): """Test conversion of a single event from lmcache engine format.""" mock_connector._lmcache_engine.get_kv_events.return_value = [ mock_lmcache_engine_event ] result = mock_connector.get_kv_connector_kv_cache_events() assert result is not None assert isinstance(result, LMCacheKVEvents) assert result.get_number_of_workers() == 1 events = result.get_all_events() assert len(events) == 1 assert isinstance(events[0], BlockStored) assert events[0].block_hashes == ["hash1", "hash2"] assert events[0].parent_block_hash == "parent_hash" assert events[0].token_ids == [1, 2, 3, 4] assert events[0].lora_id is None assert events[0].block_size == 16 assert events[0].medium == "GPU" assert events[0].lora_name is None def test_converts_multiple_events(self, mock_connector): """Test conversion of multiple events from lmcache engine format.""" class MockEvent: def __init__(self, i): self.block_hashes = [f"hash{i}"] self.parent_block_hash = f"parent{i}" self.token_ids = [i] self.lora_id = None self.block_size = 16 self.medium = "GPU" self.lora_name = None events = [MockEvent(i) for i in range(5)] mock_connector._lmcache_engine.get_kv_events.return_value = events result = mock_connector.get_kv_connector_kv_cache_events() assert result is not None assert isinstance(result, LMCacheKVEvents) converted_events = result.get_all_events() assert len(converted_events) == 5 for i, event in enumerate(converted_events): assert isinstance(event, BlockStored) assert event.block_hashes == [f"hash{i}"] assert event.parent_block_hash == f"parent{i}" assert event.token_ids == [i] def test_preserves_event_attributes(self, mock_connector): """Test that all event attributes are correctly preserved.""" class MockEventWithLora: def __init__(self): self.block_hashes = ["hash_a", "hash_b", "hash_c"] self.parent_block_hash = "parent_xyz" self.token_ids = [100, 200, 300] self.lora_id = 42 self.block_size = 32 self.medium = "DISK" self.lora_name = "lora_example" mock_connector._lmcache_engine.get_kv_events.return_value = [ MockEventWithLora() ] result = mock_connector.get_kv_connector_kv_cache_events() events = result.get_all_events() event = events[0] assert event.block_hashes == ["hash_a", "hash_b", "hash_c"] assert event.parent_block_hash == "parent_xyz" assert event.token_ids == [100, 200, 300] assert event.lora_id == 42 assert event.block_size == 32 assert event.medium == "DISK" assert event.lora_name == "lora_example" def test_handles_none_parent_block_hash(self, mock_connector): """Test handling of events with None parent_block_hash.""" class MockEventNoParent: def __init__(self): self.block_hashes = ["hash1"] self.parent_block_hash = None self.token_ids = [1, 2] self.lora_id = None self.block_size = 16 self.medium = "GPU" self.lora_name = None mock_connector._lmcache_engine.get_kv_events.return_value = [ MockEventNoParent() ] result = mock_connector.get_kv_connector_kv_cache_events() events = result.get_all_events() assert events[0].parent_block_hash is None class TestUpdateConnectorOutput: """Test update_connector_output method.""" def test_does_nothing_when_kv_cache_events_is_none(self, mock_connector): """Test that method returns early when kv_cache_events is None.""" connector_output = KVConnectorOutput(kv_cache_events=None) mock_connector.update_connector_output(connector_output) assert mock_connector._kv_cache_events is None def test_does_nothing_when_kv_cache_events_is_not_lmcache_kv_events( self, mock_connector ): """Test that method returns early when kv_cache_events is not LMCacheKVEvents.""" # Create a mock object that is not LMCacheKVEvents fake_events = MagicMock() connector_output = KVConnectorOutput(kv_cache_events=fake_events) mock_connector.update_connector_output(connector_output) assert mock_connector._kv_cache_events is None def test_sets_kv_cache_events_when_none(self, mock_connector): """Test that _kv_cache_events is set when it was None.""" kv_events = LMCacheKVEvents(num_workers=1) event = BlockStored( block_hashes=["hash1"], parent_block_hash=None, token_ids=[1, 2], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) kv_events.add_events([event]) connector_output = KVConnectorOutput(kv_cache_events=kv_events) mock_connector.update_connector_output(connector_output) assert mock_connector._kv_cache_events is kv_events def test_adds_events_when_kv_cache_events_already_exists(self, mock_connector): """Test that events are added when _kv_cache_events already exists.""" # Set up existing events existing_events = LMCacheKVEvents(num_workers=2) event1 = BlockStored( block_hashes=["hash1"], parent_block_hash=None, token_ids=[1], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) existing_events.add_events([event1]) existing_events.add_events([event1]) # Simulate 2 workers reporting mock_connector._kv_cache_events = existing_events # Create new events to add new_events = LMCacheKVEvents(num_workers=1) event2 = BlockStored( block_hashes=["hash2"], parent_block_hash=None, token_ids=[2], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) new_events.add_events([event2]) connector_output = KVConnectorOutput(kv_cache_events=new_events) mock_connector.update_connector_output(connector_output) # Check that events were added all_events = mock_connector._kv_cache_events.get_all_events() assert len(all_events) == 3 # 2 from existing + 1 from new assert event1 in all_events assert event2 in all_events def test_increments_workers_when_kv_cache_events_already_exists( self, mock_connector ): """Test that worker count is incremented correctly.""" # Set up existing events with 2 workers existing_events = LMCacheKVEvents(num_workers=2) mock_connector._kv_cache_events = existing_events # Create new events from 3 workers new_events = LMCacheKVEvents(num_workers=3) event = BlockStored( block_hashes=["hash1"], parent_block_hash=None, token_ids=[1], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) new_events.add_events([event]) connector_output = KVConnectorOutput(kv_cache_events=new_events) mock_connector.update_connector_output(connector_output) # Worker count should be 2 + 3 = 5 assert mock_connector._kv_cache_events.get_number_of_workers() == 5 def test_multiple_updates(self, mock_connector): """Test multiple consecutive updates.""" # First update events1 = LMCacheKVEvents(num_workers=1) event1 = BlockStored( block_hashes=["hash1"], parent_block_hash=None, token_ids=[1], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) events1.add_events([event1]) output1 = KVConnectorOutput(kv_cache_events=events1) mock_connector.update_connector_output(output1) # Second update events2 = LMCacheKVEvents(num_workers=2) event2 = BlockStored( block_hashes=["hash2"], parent_block_hash=None, token_ids=[2], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) events2.add_events([event2]) output2 = KVConnectorOutput(kv_cache_events=events2) mock_connector.update_connector_output(output2) # Third update events3 = LMCacheKVEvents(num_workers=1) event3 = BlockStored( block_hashes=["hash3"], parent_block_hash=None, token_ids=[3], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) events3.add_events([event3]) output3 = KVConnectorOutput(kv_cache_events=events3) mock_connector.update_connector_output(output3) # Check final state all_events = mock_connector._kv_cache_events.get_all_events() assert len(all_events) == 3 assert mock_connector._kv_cache_events.get_number_of_workers() == 4 # 1+2+1 def test_updates_with_empty_events(self, mock_connector): """Test updating with empty event lists.""" # First update with actual events events1 = LMCacheKVEvents(num_workers=1) event1 = BlockStored( block_hashes=["hash1"], parent_block_hash=None, token_ids=[1], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) events1.add_events([event1]) output1 = KVConnectorOutput(kv_cache_events=events1) mock_connector.update_connector_output(output1) # Second update with empty events events2 = LMCacheKVEvents(num_workers=2) # No events added output2 = KVConnectorOutput(kv_cache_events=events2) mock_connector.update_connector_output(output2) # Should still have the original event all_events = mock_connector._kv_cache_events.get_all_events() assert len(all_events) == 1 assert mock_connector._kv_cache_events.get_number_of_workers() == 3 class TestTakeEvents: """Test take_events method.""" def test_yields_nothing_when_kv_cache_events_is_none(self, mock_connector): """Test that nothing is yielded when _kv_cache_events is None.""" mock_connector._kv_cache_events = None events = list(mock_connector.take_events()) assert events == [] def test_yields_events_and_clears(self, mock_connector): """Test that events are yielded and then cleared.""" # Set up events kv_events = LMCacheKVEvents(num_workers=1) event1 = BlockStored( block_hashes=["hash1"], parent_block_hash=None, token_ids=[1], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) event2 = BlockStored( block_hashes=["hash2"], parent_block_hash=None, token_ids=[2], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) kv_events.add_events([event1, event2]) mock_connector._kv_cache_events = kv_events # Take events events = list(mock_connector.take_events()) # Check that events were yielded assert len(events) == 2 assert event1 in events assert event2 in events # Check that _kv_cache_events was cleared assert mock_connector._kv_cache_events is None def test_aggregates_before_yielding(self, mock_connector): """Test that events are aggregated before yielding.""" # Set up events from multiple workers kv_events = LMCacheKVEvents(num_workers=3) common_event = BlockStored( block_hashes=["hash_common"], parent_block_hash=None, token_ids=[1], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) uncommon_event = BlockStored( block_hashes=["hash_uncommon"], parent_block_hash=None, token_ids=[2], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) # All 3 workers report common_event kv_events.add_events([common_event]) kv_events.add_events([common_event]) kv_events.add_events([common_event]) # Only 1 worker reports uncommon_event kv_events.add_events([uncommon_event]) mock_connector._kv_cache_events = kv_events # Take events events = list(mock_connector.take_events()) # Only the common event should be yielded assert len(events) == 1 assert events[0] == common_event def test_multiple_take_events_calls(self, mock_connector): """Test calling take_events multiple times.""" # First call with events kv_events1 = LMCacheKVEvents(num_workers=1) event1 = BlockStored( block_hashes=["hash1"], parent_block_hash=None, token_ids=[1], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) kv_events1.add_events([event1]) mock_connector._kv_cache_events = kv_events1 events1 = list(mock_connector.take_events()) assert len(events1) == 1 assert events1[0] == event1 assert mock_connector._kv_cache_events is None # Second call with no events events2 = list(mock_connector.take_events()) assert events2 == [] # Third call after adding new events kv_events2 = LMCacheKVEvents(num_workers=1) event2 = BlockStored( block_hashes=["hash2"], parent_block_hash=None, token_ids=[2], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) kv_events2.add_events([event2]) mock_connector._kv_cache_events = kv_events2 events3 = list(mock_connector.take_events()) assert len(events3) == 1 assert events3[0] == event2 def test_yields_empty_after_aggregation_removes_all(self, mock_connector): """Test that nothing is yielded if aggregation removes all events.""" # Set up events from 2 workers with no common events kv_events = LMCacheKVEvents(num_workers=2) event1 = BlockStored( block_hashes=["hash1"], parent_block_hash=None, token_ids=[1], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) event2 = BlockStored( block_hashes=["hash2"], parent_block_hash=None, token_ids=[2], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) # Worker 1 reports event1 kv_events.add_events([event1]) # Worker 2 reports event2 kv_events.add_events([event2]) mock_connector._kv_cache_events = kv_events # Take events events = list(mock_connector.take_events()) # No common events, so nothing should be yielded assert events == [] assert mock_connector._kv_cache_events is None class TestIntegrationScenarios: """Test integration scenarios.""" def test_full_workflow(self, mock_connector, mock_lmcache_engine_event): """Test a complete workflow from getting events to taking them.""" # Step 1: Get events from lmcache engine mock_connector._lmcache_engine.get_kv_events.return_value = [ mock_lmcache_engine_event ] kv_events = mock_connector.get_kv_connector_kv_cache_events() assert kv_events is not None assert len(kv_events.get_all_events()) == 1 # Step 2: Update connector output (simulate receiving from worker) output1 = KVConnectorOutput(kv_cache_events=kv_events) mock_connector.update_connector_output(output1) assert mock_connector._kv_cache_events is not None # Step 3: Take events taken_events = list(mock_connector.take_events()) assert len(taken_events) == 1 assert mock_connector._kv_cache_events is None def test_multiple_workers_workflow(self, mock_connector): """Test workflow with multiple workers.""" class MockEvent: def __init__(self, hash_val): self.block_hashes = [hash_val] self.parent_block_hash = None self.token_ids = [1] self.lora_id = None self.block_size = 16 self.medium = "GPU" self.lora_name = None # Worker 1 mock_connector._lmcache_engine.get_kv_events.return_value = [ MockEvent("hash_common"), MockEvent("hash_worker1"), ] kv_events1 = mock_connector.get_kv_connector_kv_cache_events() output1 = KVConnectorOutput(kv_cache_events=kv_events1) mock_connector.update_connector_output(output1) # Worker 2 mock_connector._lmcache_engine.get_kv_events.return_value = [ MockEvent("hash_common"), MockEvent("hash_worker2"), ] kv_events2 = mock_connector.get_kv_connector_kv_cache_events() output2 = KVConnectorOutput(kv_cache_events=kv_events2) mock_connector.update_connector_output(output2) # Take events (should only get common events) taken_events = list(mock_connector.take_events()) # With aggregation, only events reported by both workers should be present # In this case, hash_common was reported by both event_hashes = [e.block_hashes[0] for e in taken_events] assert "hash_common" in event_hashes def test_empty_workflow(self, mock_connector): """Test workflow when there are no events at any stage.""" # Get events returns None mock_connector._lmcache_engine.get_kv_events.return_value = None kv_events = mock_connector.get_kv_connector_kv_cache_events() assert kv_events is None # Update with None output = KVConnectorOutput(kv_cache_events=None) mock_connector.update_connector_output(output) # Take events taken_events = list(mock_connector.take_events()) assert taken_events == [] assert mock_connector._kv_cache_events is None def test_repeated_cycles(self, mock_connector): """Test multiple cycles of the complete workflow.""" class MockEvent: def __init__(self, cycle_num): self.block_hashes = [f"hash_cycle_{cycle_num}"] self.parent_block_hash = None self.token_ids = [cycle_num] self.lora_id = None self.block_size = 16 self.medium = "GPU" self.lora_name = None for cycle in range(3): # Get events mock_connector._lmcache_engine.get_kv_events.return_value = [ MockEvent(cycle) ] kv_events = mock_connector.get_kv_connector_kv_cache_events() # Update output = KVConnectorOutput(kv_cache_events=kv_events) mock_connector.update_connector_output(output) # Take taken_events = list(mock_connector.take_events()) # Verify assert len(taken_events) == 1 assert taken_events[0].block_hashes[0] == f"hash_cycle_{cycle}" assert mock_connector._kv_cache_events is None def test_lmcache_kv_events_aggregation(self): """ Test LMCacheKVEvents aggregation across TP ranks using KVOutputAggregator (used by MultiprocExecutor). """ from vllm.distributed.kv_transfer.kv_connector.utils import KVOutputAggregator from vllm.v1.outputs import ModelRunnerOutput # Create KVOutputAggregator for 3 workers (simulating TP=3) aggregator = KVOutputAggregator(expected_finished_count=3) # Define common and unique events common_event = BlockStored( block_hashes=["hash_common"], parent_block_hash="parent_common", token_ids=[1, 2, 3], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) worker1_unique_event = BlockStored( block_hashes=["hash_worker1"], parent_block_hash="parent_w1", token_ids=[4, 5], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) worker2_unique_event = BlockStored( block_hashes=["hash_worker2"], parent_block_hash="parent_w2", token_ids=[6, 7], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) worker3_unique_event = BlockStored( block_hashes=["hash_worker3"], parent_block_hash="parent_w3", token_ids=[8, 9], block_size=16, lora_id=None, medium="GPU", lora_name=None, ) # Create events for each worker # Worker 0: reports common event and its unique event worker0_events = LMCacheKVEvents(num_workers=1) worker0_events.add_events([common_event, worker1_unique_event]) # Worker 1: reports common event and its unique event worker1_events = LMCacheKVEvents(num_workers=1) worker1_events.add_events([common_event, worker2_unique_event]) # Worker 2: reports common event and its unique event worker2_events = LMCacheKVEvents(num_workers=1) worker2_events.add_events([common_event, worker3_unique_event]) # Create ModelRunnerOutput instances for each worker worker_outputs = [] for i, worker_events in enumerate( [worker0_events, worker1_events, worker2_events] ): output = ModelRunnerOutput( req_ids=[f"req_{i}"], req_id_to_index={f"req_{i}": 0}, sampled_token_ids=[[123]], # dummy token logprobs=None, prompt_logprobs_dict={}, pooler_output=[None], kv_connector_output=KVConnectorOutput( finished_sending=set([f"req_{i}_send"]) if i < 2 else None, # Workers 0,1 finished sending finished_recving=set([f"req_{i}_recv"]) if i > 0 else None, # Workers 1,2 finished receiving kv_cache_events=worker_events, ), ) worker_outputs.append(output) # Use the real aggregation mechanism (like MultiprocExecutor.execute_model) aggregated_output = aggregator.aggregate(worker_outputs, output_rank=0) kv_cache_events = aggregated_output.kv_connector_output.kv_cache_events assert isinstance(kv_cache_events, LMCacheKVEvents) # After aggregation, events should be combined from all workers # The aggregator doesn't automatically aggregate events, so we need to call # aggregate() to get only common events kv_cache_events.aggregate() aggregated_events = kv_cache_events.get_all_events() # Only the common event should remain after aggregation # because it's the only event reported by all 3 workers assert len(aggregated_events) == 1 assert aggregated_events[0] == common_event # Verify the common event properties assert aggregated_events[0].block_hashes == ["hash_common"] assert aggregated_events[0].parent_block_hash == "parent_common" assert aggregated_events[0].token_ids == [1, 2, 3]
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/kv_connector/unit/test_lmcache_connector.py", "license": "Apache License 2.0", "lines": 651, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/v1/worker/cp_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import TYPE_CHECKING, Any, cast from vllm.config import VllmConfig, get_layers_from_vllm_config from vllm.distributed import get_dcp_group, get_pcp_group if TYPE_CHECKING: from vllm.model_executor.layers.attention_layer_base import AttentionLayerBase else: AttentionLayerBase = object def check_attention_cp_compatibility(vllm_config: VllmConfig) -> None: pcp_size = vllm_config.parallel_config.prefill_context_parallel_size dcp_size = vllm_config.parallel_config.decode_context_parallel_size interleave_size = vllm_config.parallel_config.cp_kv_cache_interleave_size if pcp_size * dcp_size > 1: layer_type = cast(type[Any], AttentionLayerBase) layers = get_layers_from_vllm_config(vllm_config, layer_type) for layer in layers.values(): layer_impl = getattr(layer, "impl", None) if layer_impl is None: continue if vllm_config.speculative_config is not None and interleave_size > 1: assert layer_impl.supports_mtp_with_cp_non_trivial_interleave_size, ( "MTP with cp_kv_cache_interleave_size > 1 is not " f"supported in {layer_impl.__class__.__name__}." ) if dcp_size > 1: assert layer_impl.need_to_return_lse_for_decode, ( "DCP requires attention impls to return" " the softmax lse for decode, but the impl " f"{layer_impl.__class__.__name__} " "does not return the softmax lse for decode." ) if pcp_size > 1: assert layer_impl.supports_pcp, ( "PCP requires attention impls' support, " f"but the impl {layer_impl.__class__.__name__} " "does not support PCP." ) def get_total_cp_world_size(): try: pcp_world_size = get_pcp_group().world_size except AssertionError: # PCP might not be initialized in testing pcp_world_size = 1 try: dcp_world_size = get_dcp_group().world_size except AssertionError: # DCP might not be initialized in testing dcp_world_size = 1 return dcp_world_size * pcp_world_size
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/worker/cp_utils.py", "license": "Apache License 2.0", "lines": 50, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/reasoning/test_minimax_m2_append_reasoning_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from transformers import AutoTokenizer from tests.reasoning.utils import run_reasoning_extraction from vllm.reasoning import ReasoningParser, ReasoningParserManager parser_name = "minimax_m2_append_think" end_token = "</think>" # MiniMax M2 model path REASONING_MODEL_NAME = "MiniMaxAI/MiniMax-M2" @pytest.fixture(scope="module") def minimax_m2_tokenizer(): return AutoTokenizer.from_pretrained(REASONING_MODEL_NAME) # ============================================================================= # MiniMaxM2AppendThinkReasoningParser behavior: # - Prepends <think> to the beginning of the output # - Does NOT separate reasoning and content # - Returns everything as content (with <think> prepended) # - reasoning is always None # # This parser is used when you want to keep the raw output with <think> added # ============================================================================= # Case: simple output with end token SIMPLE_OUTPUT = { "output": "This is reasoning</think>This is response", "reasoning": None, "content": "<think>This is reasoning</think>This is response", "is_reasoning_end": True, } # Case: output without end token (reasoning in progress) NO_END_TOKEN = { "output": "This is reasoning in progress", "reasoning": None, "content": "<think>This is reasoning in progress", "is_reasoning_end": False, } # Case: only end token ONLY_END_TOKEN = { "output": "</think>This is response", "reasoning": None, "content": "<think></think>This is response", "is_reasoning_end": True, } # Case: multiple lines MULTIPLE_LINES = { "output": "Line 1\nLine 2</think>Response 1\nResponse 2", "reasoning": None, "content": "<think>Line 1\nLine 2</think>Response 1\nResponse 2", "is_reasoning_end": True, } # Case: empty output (non-streaming prepends <think>) EMPTY = { "output": "", "reasoning": None, "content": "<think>", "is_reasoning_end": False, } # Case: empty output streaming (no tokens = no output) EMPTY_STREAMING = { "output": "", "reasoning": None, "content": None, "is_reasoning_end": False, } # Case: special characters SPECIAL_CHARS = { "output": "Let me think... 1+1=2</think>Yes!", "reasoning": None, "content": "<think>Let me think... 1+1=2</think>Yes!", "is_reasoning_end": True, } # Case: code in output CODE_OUTPUT = { "output": "```python\nprint('hi')\n```</think>Here's the code.", "reasoning": None, "content": "<think>```python\nprint('hi')\n```</think>Here's the code.", "is_reasoning_end": True, } TEST_CASES = [ pytest.param( False, SIMPLE_OUTPUT, id="simple_output", ), pytest.param( True, SIMPLE_OUTPUT, id="simple_output_streaming", ), pytest.param( False, NO_END_TOKEN, id="no_end_token", ), pytest.param( True, NO_END_TOKEN, id="no_end_token_streaming", ), pytest.param( False, ONLY_END_TOKEN, id="only_end_token", ), pytest.param( True, ONLY_END_TOKEN, id="only_end_token_streaming", ), pytest.param( False, MULTIPLE_LINES, id="multiple_lines", ), pytest.param( True, MULTIPLE_LINES, id="multiple_lines_streaming", ), pytest.param( False, EMPTY, id="empty", ), pytest.param( True, EMPTY_STREAMING, id="empty_streaming", ), pytest.param( False, SPECIAL_CHARS, id="special_chars", ), pytest.param( True, SPECIAL_CHARS, id="special_chars_streaming", ), pytest.param( False, CODE_OUTPUT, id="code_output", ), pytest.param( True, CODE_OUTPUT, id="code_output_streaming", ), ] @pytest.mark.parametrize("streaming, param_dict", TEST_CASES) def test_reasoning( streaming: bool, param_dict: dict, minimax_m2_tokenizer, ): output = minimax_m2_tokenizer.tokenize(param_dict["output"]) # decode everything to tokens output_tokens: list[str] = [ minimax_m2_tokenizer.convert_tokens_to_string([token]) for token in output ] parser: ReasoningParser = ReasoningParserManager.get_reasoning_parser(parser_name)( minimax_m2_tokenizer ) reasoning, content = run_reasoning_extraction( parser, output_tokens, streaming=streaming ) assert reasoning == param_dict["reasoning"] assert content == param_dict["content"] # Test is_reasoning_end output_ids = minimax_m2_tokenizer.convert_tokens_to_ids(output) is_reasoning_end = parser.is_reasoning_end(output_ids) assert is_reasoning_end == param_dict["is_reasoning_end"]
{ "repo_id": "vllm-project/vllm", "file_path": "tests/reasoning/test_minimax_m2_append_reasoning_parser.py", "license": "Apache License 2.0", "lines": 173, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/reasoning/test_minimax_m2_reasoning_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from transformers import AutoTokenizer from tests.reasoning.utils import run_reasoning_extraction from vllm.reasoning import ReasoningParser, ReasoningParserManager parser_name = "minimax_m2" end_token = "</think>" # MiniMax M2 model path REASONING_MODEL_NAME = "MiniMaxAI/MiniMax-M2" @pytest.fixture(scope="module") def minimax_m2_tokenizer(): return AutoTokenizer.from_pretrained(REASONING_MODEL_NAME) # ============================================================================= # MiniMax M2 specific behavior: # - Model does NOT generate <think> start token # - Model only generates </think> end token # - All content before </think> is reasoning # - All content after </think> is the actual response (content) # ============================================================================= # Case: reasoning + end token + content (typical case) SIMPLE_REASONING = { "output": "This is a reasoning section</think>This is the rest", "reasoning": "This is a reasoning section", "content": "This is the rest", "is_reasoning_end": True, } # Case: reasoning + end token only (no content after) COMPLETE_REASONING = { "output": "This is a reasoning section</think>", "reasoning": "This is a reasoning section", "content": None, "is_reasoning_end": True, } # Case: no end token yet (streaming in progress, all is reasoning) NO_END_TOKEN = { "output": "This is reasoning in progress", "reasoning": "This is reasoning in progress", "content": None, "is_reasoning_end": False, } # Case: multiple lines of reasoning MULTIPLE_LINES = { "output": "First line\nSecond line</think>Response first line\nResponse second", "reasoning": "First line\nSecond line", "content": "Response first line\nResponse second", "is_reasoning_end": True, } # Case: only end token (empty reasoning, immediate response) SHORTEST_REASONING_NO_STREAMING = { "output": "</think>This is the response", "reasoning": "", "content": "This is the response", "is_reasoning_end": True, } # Case: only end token streaming (reasoning is None because it's just the token) SHORTEST_REASONING_STREAMING = { "output": "</think>This is the response", "reasoning": None, "content": "This is the response", "is_reasoning_end": True, } # Case: empty output EMPTY = { "output": "", "reasoning": "", "content": None, "is_reasoning_end": False, } # Case: empty streaming EMPTY_STREAMING = { "output": "", "reasoning": None, "content": None, "is_reasoning_end": False, } # Case: long reasoning with special characters SPECIAL_CHARS = { "output": "Let me think... 1+1=2, right?</think>Yes, 1+1=2.", "reasoning": "Let me think... 1+1=2, right?", "content": "Yes, 1+1=2.", "is_reasoning_end": True, } # Case: reasoning with code blocks CODE_IN_REASONING = { "output": "```python\nprint('hello')\n```</think>Here is the code.", "reasoning": "```python\nprint('hello')\n```", "content": "Here is the code.", "is_reasoning_end": True, } TEST_CASES = [ # Core cases: no start token (MiniMax M2 actual behavior) pytest.param( False, SIMPLE_REASONING, id="simple_reasoning", ), pytest.param( True, SIMPLE_REASONING, id="simple_reasoning_streaming", ), pytest.param( False, COMPLETE_REASONING, id="complete_reasoning", ), pytest.param( True, COMPLETE_REASONING, id="complete_reasoning_streaming", ), pytest.param( False, NO_END_TOKEN, id="no_end_token", ), pytest.param( True, NO_END_TOKEN, id="no_end_token_streaming", ), pytest.param( False, MULTIPLE_LINES, id="multiple_lines", ), pytest.param( True, MULTIPLE_LINES, id="multiple_lines_streaming", ), pytest.param( False, SHORTEST_REASONING_NO_STREAMING, id="shortest_reasoning", ), pytest.param( True, SHORTEST_REASONING_STREAMING, id="shortest_reasoning_streaming", ), pytest.param( False, EMPTY, id="empty", ), pytest.param( True, EMPTY_STREAMING, id="empty_streaming", ), pytest.param( False, SPECIAL_CHARS, id="special_chars", ), pytest.param( True, SPECIAL_CHARS, id="special_chars_streaming", ), pytest.param( False, CODE_IN_REASONING, id="code_in_reasoning", ), pytest.param( True, CODE_IN_REASONING, id="code_in_reasoning_streaming", ), ] @pytest.mark.parametrize("streaming, param_dict", TEST_CASES) def test_reasoning( streaming: bool, param_dict: dict, minimax_m2_tokenizer, ): output = minimax_m2_tokenizer.tokenize(param_dict["output"]) # decode everything to tokens output_tokens: list[str] = [ minimax_m2_tokenizer.convert_tokens_to_string([token]) for token in output ] parser: ReasoningParser = ReasoningParserManager.get_reasoning_parser(parser_name)( minimax_m2_tokenizer ) reasoning, content = run_reasoning_extraction( parser, output_tokens, streaming=streaming ) assert reasoning == param_dict["reasoning"] assert content == param_dict["content"] # Test is_reasoning_end output_ids = minimax_m2_tokenizer.convert_tokens_to_ids(output) is_reasoning_end = parser.is_reasoning_end(output_ids) assert is_reasoning_end == param_dict["is_reasoning_end"] # Test extract_content if param_dict["content"] is not None: content = parser.extract_content_ids(output_ids) assert content == minimax_m2_tokenizer.convert_tokens_to_ids( minimax_m2_tokenizer.tokenize(param_dict["content"]) ) else: content = parser.extract_content_ids(output) assert content == []
{ "repo_id": "vllm-project/vllm", "file_path": "tests/reasoning/test_minimax_m2_reasoning_parser.py", "license": "Apache License 2.0", "lines": 205, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:benchmarks/kernels/benchmark_mla_k_concat.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Benchmark script comparing torch.cat vs direct copy for k_nope/k_pe concatenation in MLA (Multi-head Latent Attention) prefill. This validates that the optimization from commit 8d4142bd is beneficial across various batch sizes, not just the originally tested batch size of 32768. """ import time from collections.abc import Callable import torch # DeepSeek-V3 MLA dimensions NUM_HEADS = 128 QK_NOPE_HEAD_DIM = 128 PE_DIM = 64 def cat_method(k_nope: torch.Tensor, k_pe: torch.Tensor) -> torch.Tensor: """Original torch.cat approach with expand.""" return torch.cat((k_nope, k_pe.expand((*k_nope.shape[:-1], -1))), dim=-1) def direct_copy_method(k_nope: torch.Tensor, k_pe: torch.Tensor) -> torch.Tensor: """Optimized direct copy approach (avoids expand + cat overhead).""" k = torch.empty( (*k_nope.shape[:-1], k_nope.shape[-1] + k_pe.shape[-1]), dtype=k_nope.dtype, device=k_nope.device, ) k[..., : k_nope.shape[-1]] = k_nope k[..., k_nope.shape[-1] :] = k_pe return k def benchmark_method( method: Callable, k_nope: torch.Tensor, k_pe: torch.Tensor, num_warmup: int = 10, num_iters: int = 100, ) -> float: """Benchmark a concatenation method and return mean latency in ms.""" # Warmup for _ in range(num_warmup): _ = method(k_nope, k_pe) torch.cuda.synchronize() # Benchmark start = time.perf_counter() for _ in range(num_iters): _ = method(k_nope, k_pe) torch.cuda.synchronize() end = time.perf_counter() return (end - start) / num_iters * 1000 # Convert to ms @torch.inference_mode() def run_benchmark(dtype: torch.dtype, dtype_name: str): """Run benchmark for a specific dtype.""" torch.set_default_device("cuda") # Batch sizes to test (powers of 2 from 32 to 65536) batch_sizes = [32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536] print("=" * 80) print("Benchmark: torch.cat vs direct copy for MLA k_nope/k_pe concatenation") print("=" * 80) print( f"Tensor shapes: k_nope=[B, {NUM_HEADS}, {QK_NOPE_HEAD_DIM}], " f"k_pe=[B, 1, {PE_DIM}]" ) print(f"dtype: {dtype_name}") print() print( f"{'Batch Size':>12} | {'cat (ms)':>10} | {'direct (ms)':>12} | " f"{'Speedup':>8} | {'Reduction':>10}" ) print("-" * 70) results = [] for batch_size in batch_sizes: # Create input tensors (generate in float32 then convert for FP8 compatibility) k_nope = torch.randn( batch_size, NUM_HEADS, QK_NOPE_HEAD_DIM, dtype=torch.float32, device="cuda" ).to(dtype) k_pe = torch.randn( batch_size, 1, PE_DIM, dtype=torch.float32, device="cuda" ).to(dtype) # Benchmark both methods cat_time = benchmark_method(cat_method, k_nope, k_pe) direct_time = benchmark_method(direct_copy_method, k_nope, k_pe) speedup = cat_time / direct_time reduction = (1 - direct_time / cat_time) * 100 results.append((batch_size, cat_time, direct_time, speedup, reduction)) print( f"{batch_size:>12} | {cat_time:>10.3f} | {direct_time:>12.3f} | " f"{speedup:>7.2f}x | {reduction:>9.1f}%" ) print("=" * 80) # Summary statistics speedups = [r[3] for r in results] print("\nSpeedup summary:") print(f" Min: {min(speedups):.2f}x") print(f" Max: {max(speedups):.2f}x") print(f" Mean: {sum(speedups) / len(speedups):.2f}x") # Find crossover point crossover_batch = None for batch_size, _, _, speedup, _ in results: if speedup >= 1.0: crossover_batch = batch_size break print("\nConclusion:") if crossover_batch: print(f" - Direct copy becomes beneficial at batch size >= {crossover_batch}") # Filter for large batches (>= 512 which is typical for prefill) large_batch_speedups = [r[3] for r in results if r[0] >= 512] if large_batch_speedups: avg_large = sum(large_batch_speedups) / len(large_batch_speedups) print(f" - For batch sizes >= 512: avg speedup = {avg_large:.2f}x") print(" - MLA prefill typically uses large batches, so optimization is effective") return results @torch.inference_mode() def main(): # Test bfloat16 print("\n") run_benchmark(torch.bfloat16, "bfloat16") # Test float8_e4m3fn print("\n") run_benchmark(torch.float8_e4m3fn, "float8_e4m3fn") if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/kernels/benchmark_mla_k_concat.py", "license": "Apache License 2.0", "lines": 119, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/engine/test_init_error_messaging.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from vllm.v1.core.kv_cache_utils import check_enough_kv_cache_memory from vllm.v1.kv_cache_interface import FullAttentionSpec def test_kv_cache_oom_no_memory(): from unittest.mock import MagicMock config = MagicMock() config.model_config.max_model_len = 2048 spec = { "layer_0": FullAttentionSpec( block_size=16, num_kv_heads=8, head_size=128, dtype="float16", ) } with pytest.raises(ValueError): check_enough_kv_cache_memory(config, spec, 0) def test_kv_cache_oom_insufficient_memory(monkeypatch): from unittest.mock import MagicMock config = MagicMock() config.model_config.max_model_len = 2048 config.cache_config.block_size = 16 config.parallel_config.tensor_parallel_size = 1 config.parallel_config.pipeline_parallel_size = 1 config.parallel_config.decode_context_parallel_size = 1 monkeypatch.setattr( "vllm.v1.core.kv_cache_utils.max_memory_usage_bytes", lambda c, s: 100 * 1024**3, # 100 GiB ) spec = { "layer_0": FullAttentionSpec( block_size=16, num_kv_heads=8, head_size=128, dtype="float16", ) } with pytest.raises(ValueError): check_enough_kv_cache_memory(config, spec, 1024**3) # 1 GiB
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/engine/test_init_error_messaging.py", "license": "Apache License 2.0", "lines": 41, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/entrypoints/openai/test_chat_error.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass, field from http import HTTPStatus from typing import Any from unittest.mock import AsyncMock, MagicMock, patch import pytest from vllm.config.multimodal import MultiModalConfig from vllm.entrypoints.openai.chat_completion.protocol import ChatCompletionRequest from vllm.entrypoints.openai.chat_completion.serving import OpenAIServingChat from vllm.entrypoints.openai.engine.protocol import ErrorResponse from vllm.entrypoints.openai.models.protocol import BaseModelPath from vllm.entrypoints.openai.models.serving import OpenAIServingModels from vllm.outputs import CompletionOutput, RequestOutput from vllm.renderers.hf import HfRenderer from vllm.tokenizers.registry import tokenizer_args_from_config from vllm.v1.engine.async_llm import AsyncLLM MODEL_NAME = "openai-community/gpt2" MODEL_NAME_SHORT = "gpt2" BASE_MODEL_PATHS = [ BaseModelPath(name=MODEL_NAME, model_path=MODEL_NAME), BaseModelPath(name=MODEL_NAME_SHORT, model_path=MODEL_NAME_SHORT), ] @dataclass class MockHFConfig: model_type: str = "any" @dataclass class MockModelConfig: task = "generate" runner_type = "generate" model = MODEL_NAME tokenizer = MODEL_NAME trust_remote_code = False tokenizer_mode = "auto" max_model_len = 100 tokenizer_revision = None multimodal_config = MultiModalConfig() hf_config = MockHFConfig() hf_text_config = MockHFConfig() logits_processors: list[str] | None = None diff_sampling_param: dict | None = None allowed_local_media_path: str = "" allowed_media_domains: list[str] | None = None encoder_config = None generation_config: str = "auto" media_io_kwargs: dict[str, dict[str, Any]] = field(default_factory=dict) skip_tokenizer_init = False is_encoder_decoder: bool = False is_multimodal_model: bool = False def get_diff_sampling_param(self): return self.diff_sampling_param or {} @dataclass class MockVllmConfig: model_config: MockModelConfig def _build_renderer(model_config: MockModelConfig): _, tokenizer_name, _, kwargs = tokenizer_args_from_config(model_config) return HfRenderer.from_config( MockVllmConfig(model_config), tokenizer_kwargs={**kwargs, "tokenizer_name": tokenizer_name}, ) def _build_serving_chat(engine: AsyncLLM) -> OpenAIServingChat: models = OpenAIServingModels( engine_client=engine, base_model_paths=BASE_MODEL_PATHS, ) serving_chat = OpenAIServingChat( engine, models, response_role="assistant", request_logger=None, chat_template=None, chat_template_content_format="auto", ) async def _fake_preprocess_chat(*args, **kwargs): # return conversation, engine_prompts return ( [{"role": "user", "content": "Test"}], [{"prompt_token_ids": [1, 2, 3]}], ) serving_chat._preprocess_chat = AsyncMock(side_effect=_fake_preprocess_chat) return serving_chat @pytest.mark.asyncio async def test_chat_error_non_stream(): """test finish_reason='error' returns 500 InternalServerError (non-streaming)""" mock_engine = MagicMock(spec=AsyncLLM) mock_engine.errored = False mock_engine.model_config = MockModelConfig() mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() mock_engine.renderer = _build_renderer(mock_engine.model_config) serving_chat = _build_serving_chat(mock_engine) completion_output = CompletionOutput( index=0, text="", token_ids=[], cumulative_logprob=None, logprobs=None, finish_reason="error", ) request_output = RequestOutput( request_id="test-id", prompt="Test prompt", prompt_token_ids=[1, 2, 3], prompt_logprobs=None, outputs=[completion_output], finished=True, metrics=None, lora_request=None, encoder_prompt=None, encoder_prompt_token_ids=None, ) async def mock_generate(*args, **kwargs): yield request_output mock_engine.generate = MagicMock(side_effect=mock_generate) request = ChatCompletionRequest( model=MODEL_NAME, messages=[{"role": "user", "content": "Test prompt"}], max_tokens=10, stream=False, ) response = await serving_chat.create_chat_completion(request) assert isinstance(response, ErrorResponse) assert response.error.type == "InternalServerError" assert response.error.message == "Internal server error" assert response.error.code == HTTPStatus.INTERNAL_SERVER_ERROR @pytest.mark.asyncio async def test_chat_error_stream(): """test finish_reason='error' returns 500 InternalServerError (streaming)""" mock_engine = MagicMock(spec=AsyncLLM) mock_engine.errored = False mock_engine.model_config = MockModelConfig() mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() mock_engine.renderer = _build_renderer(mock_engine.model_config) serving_chat = _build_serving_chat(mock_engine) completion_output_1 = CompletionOutput( index=0, text="Hello", token_ids=[100], cumulative_logprob=None, logprobs=None, finish_reason=None, ) request_output_1 = RequestOutput( request_id="test-id", prompt="Test prompt", prompt_token_ids=[1, 2, 3], prompt_logprobs=None, outputs=[completion_output_1], finished=False, metrics=None, lora_request=None, encoder_prompt=None, encoder_prompt_token_ids=None, ) completion_output_2 = CompletionOutput( index=0, text="Hello", token_ids=[100], cumulative_logprob=None, logprobs=None, finish_reason="error", ) request_output_2 = RequestOutput( request_id="test-id", prompt="Test prompt", prompt_token_ids=[1, 2, 3], prompt_logprobs=None, outputs=[completion_output_2], finished=True, metrics=None, lora_request=None, encoder_prompt=None, encoder_prompt_token_ids=None, ) async def mock_generate(*args, **kwargs): yield request_output_1 yield request_output_2 mock_engine.generate = MagicMock(side_effect=mock_generate) request = ChatCompletionRequest( model=MODEL_NAME, messages=[{"role": "user", "content": "Test prompt"}], max_tokens=10, stream=True, ) response = await serving_chat.create_chat_completion(request) chunks = [] async for chunk in response: chunks.append(chunk) assert len(chunks) >= 2 assert any("Internal server error" in chunk for chunk in chunks), ( f"Expected error message in chunks: {chunks}" ) assert chunks[-1] == "data: [DONE]\n\n" @pytest.mark.parametrize( "image_content", [ [{"type": "image_url", "image_url": {"url": "https://example.com/image.jpg"}}], [{"image_url": {"url": "https://example.com/image.jpg"}}], ], ) def test_system_message_warns_on_image(image_content): """Test that system messages with image content trigger a warning.""" with patch( "vllm.entrypoints.openai.chat_completion.protocol.logger" ) as mock_logger: ChatCompletionRequest( model=MODEL_NAME, messages=[ { "role": "system", "content": image_content, } ], ) mock_logger.warning_once.assert_called() call_args = str(mock_logger.warning_once.call_args) assert "System messages should only contain text" in call_args assert "image_url" in call_args def test_system_message_accepts_text(): """Test that system messages can contain text content.""" # Should not raise an exception request = ChatCompletionRequest( model=MODEL_NAME, messages=[ {"role": "system", "content": "You are a helpful assistant."}, ], ) assert request.messages[0]["role"] == "system" def test_system_message_accepts_text_array(): """Test that system messages can contain an array with text content.""" # Should not raise an exception request = ChatCompletionRequest( model=MODEL_NAME, messages=[ { "role": "system", "content": [{"type": "text", "text": "You are a helpful assistant."}], }, ], ) assert request.messages[0]["role"] == "system" def test_user_message_accepts_image(): """Test that user messages can still contain image content.""" # Should not raise an exception request = ChatCompletionRequest( model=MODEL_NAME, messages=[ { "role": "user", "content": [ {"type": "text", "text": "What's in this image?"}, { "type": "image_url", "image_url": {"url": "https://example.com/image.jpg"}, }, ], }, ], ) assert request.messages[0]["role"] == "user" @pytest.mark.parametrize( "audio_content", [ [ { "type": "input_audio", "input_audio": {"data": "base64data", "format": "wav"}, } ], [{"input_audio": {"data": "base64data", "format": "wav"}}], ], ) def test_system_message_warns_on_audio(audio_content): """Test that system messages with audio content trigger a warning.""" with patch( "vllm.entrypoints.openai.chat_completion.protocol.logger" ) as mock_logger: ChatCompletionRequest( model=MODEL_NAME, messages=[ { "role": "system", "content": audio_content, } ], ) mock_logger.warning_once.assert_called() call_args = str(mock_logger.warning_once.call_args) assert "System messages should only contain text" in call_args assert "input_audio" in call_args @pytest.mark.parametrize( "video_content", [ [{"type": "video_url", "video_url": {"url": "https://example.com/video.mp4"}}], [{"video_url": {"url": "https://example.com/video.mp4"}}], ], ) def test_system_message_warns_on_video(video_content): """Test that system messages with video content trigger a warning.""" with patch( "vllm.entrypoints.openai.chat_completion.protocol.logger" ) as mock_logger: ChatCompletionRequest( model=MODEL_NAME, messages=[ { "role": "system", "content": video_content, } ], ) mock_logger.warning_once.assert_called() call_args = str(mock_logger.warning_once.call_args) assert "System messages should only contain text" in call_args assert "video_url" in call_args def test_json_schema_response_format_missing_schema(): """When response_format type is 'json_schema' but the json_schema field is not provided, request construction should raise a validation error so the API returns 400 instead of 500.""" with pytest.raises(Exception, match="json_schema.*must be provided"): ChatCompletionRequest( model=MODEL_NAME, messages=[{"role": "user", "content": "hello"}], response_format={"type": "json_schema"}, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/test_chat_error.py", "license": "Apache License 2.0", "lines": 326, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/entrypoints/openai/test_completion_error.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass, field from http import HTTPStatus from typing import Any from unittest.mock import MagicMock import pytest from vllm.config.multimodal import MultiModalConfig from vllm.entrypoints.openai.completion.protocol import CompletionRequest from vllm.entrypoints.openai.completion.serving import OpenAIServingCompletion from vllm.entrypoints.openai.engine.protocol import ErrorResponse from vllm.entrypoints.openai.models.protocol import BaseModelPath from vllm.entrypoints.openai.models.serving import OpenAIServingModels from vllm.outputs import CompletionOutput, RequestOutput from vllm.renderers.hf import HfRenderer from vllm.tokenizers.registry import tokenizer_args_from_config from vllm.v1.engine.async_llm import AsyncLLM MODEL_NAME = "openai-community/gpt2" MODEL_NAME_SHORT = "gpt2" BASE_MODEL_PATHS = [ BaseModelPath(name=MODEL_NAME, model_path=MODEL_NAME), BaseModelPath(name=MODEL_NAME_SHORT, model_path=MODEL_NAME_SHORT), ] @dataclass class MockHFConfig: model_type: str = "any" @dataclass class MockModelConfig: task = "generate" runner_type = "generate" model = MODEL_NAME tokenizer = MODEL_NAME trust_remote_code = False tokenizer_mode = "auto" max_model_len = 100 tokenizer_revision = None multimodal_config = MultiModalConfig() hf_config = MockHFConfig() logits_processors: list[str] | None = None diff_sampling_param: dict | None = None allowed_local_media_path: str = "" allowed_media_domains: list[str] | None = None encoder_config = None generation_config: str = "auto" media_io_kwargs: dict[str, dict[str, Any]] = field(default_factory=dict) skip_tokenizer_init = False is_encoder_decoder: bool = False is_multimodal_model: bool = False def get_diff_sampling_param(self): return self.diff_sampling_param or {} @dataclass class MockVllmConfig: model_config: MockModelConfig def _build_serving_completion(engine: AsyncLLM) -> OpenAIServingCompletion: models = OpenAIServingModels( engine_client=engine, base_model_paths=BASE_MODEL_PATHS, ) return OpenAIServingCompletion( engine, models, request_logger=None, ) def _build_renderer(model_config: MockModelConfig): _, tokenizer_name, _, kwargs = tokenizer_args_from_config(model_config) return HfRenderer.from_config( MockVllmConfig(model_config), tokenizer_kwargs={**kwargs, "tokenizer_name": tokenizer_name}, ) @pytest.mark.asyncio async def test_completion_error_non_stream(): """test finish_reason='error' returns 500 InternalServerError (non-streaming)""" mock_engine = MagicMock(spec=AsyncLLM) mock_engine.errored = False mock_engine.model_config = MockModelConfig() mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() mock_engine.renderer = _build_renderer(mock_engine.model_config) serving_completion = _build_serving_completion(mock_engine) completion_output = CompletionOutput( index=0, text="", token_ids=[], cumulative_logprob=None, logprobs=None, finish_reason="error", ) request_output = RequestOutput( request_id="test-id", prompt="Test prompt", prompt_token_ids=[1, 2, 3], prompt_logprobs=None, outputs=[completion_output], finished=True, metrics=None, lora_request=None, encoder_prompt=None, encoder_prompt_token_ids=None, ) async def mock_generate(*args, **kwargs): yield request_output mock_engine.generate = MagicMock(side_effect=mock_generate) request = CompletionRequest( model=MODEL_NAME, prompt="Test prompt", max_tokens=10, stream=False, ) response = await serving_completion.create_completion(request) assert isinstance(response, ErrorResponse) assert response.error.type == "InternalServerError" assert response.error.message == "Internal server error" assert response.error.code == HTTPStatus.INTERNAL_SERVER_ERROR @pytest.mark.asyncio async def test_completion_error_stream(): """test finish_reason='error' returns 500 InternalServerError (streaming)""" mock_engine = MagicMock(spec=AsyncLLM) mock_engine.errored = False mock_engine.model_config = MockModelConfig() mock_engine.input_processor = MagicMock() mock_engine.io_processor = MagicMock() mock_engine.renderer = _build_renderer(mock_engine.model_config) serving_completion = _build_serving_completion(mock_engine) completion_output_1 = CompletionOutput( index=0, text="Hello", token_ids=[100], cumulative_logprob=None, logprobs=None, finish_reason=None, ) request_output_1 = RequestOutput( request_id="test-id", prompt="Test prompt", prompt_token_ids=[1, 2, 3], prompt_logprobs=None, outputs=[completion_output_1], finished=False, metrics=None, lora_request=None, encoder_prompt=None, encoder_prompt_token_ids=None, ) completion_output_2 = CompletionOutput( index=0, text="Hello", token_ids=[100], cumulative_logprob=None, logprobs=None, finish_reason="error", ) request_output_2 = RequestOutput( request_id="test-id", prompt="Test prompt", prompt_token_ids=[1, 2, 3], prompt_logprobs=None, outputs=[completion_output_2], finished=True, metrics=None, lora_request=None, encoder_prompt=None, encoder_prompt_token_ids=None, ) async def mock_generate(*args, **kwargs): yield request_output_1 yield request_output_2 mock_engine.generate = MagicMock(side_effect=mock_generate) request = CompletionRequest( model=MODEL_NAME, prompt="Test prompt", max_tokens=10, stream=True, ) response = await serving_completion.create_completion(request) chunks = [] async for chunk in response: chunks.append(chunk) assert len(chunks) >= 2 assert any("Internal server error" in chunk for chunk in chunks), ( f"Expected error message in chunks: {chunks}" ) assert chunks[-1] == "data: [DONE]\n\n" def test_json_schema_response_format_missing_schema(): """When response_format type is 'json_schema' but the json_schema field is not provided, request construction should raise a validation error so the API returns 400 instead of 500.""" with pytest.raises(Exception, match="json_schema.*must be provided"): CompletionRequest( model=MODEL_NAME, prompt="Test prompt", max_tokens=10, response_format={"type": "json_schema"}, ) def test_negative_prompt_token_ids_nested(): """Negative token IDs in prompt (nested list) should raise validation error.""" with pytest.raises(Exception, match="greater than or equal to 0"): CompletionRequest( model=MODEL_NAME, prompt=[[-1]], max_tokens=10, ) def test_negative_prompt_token_ids_flat(): """Negative token IDs in prompt (flat list) should raise validation error.""" with pytest.raises(Exception, match="greater than or equal to 0"): CompletionRequest( model=MODEL_NAME, prompt=[-1], max_tokens=10, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/test_completion_error.py", "license": "Apache License 2.0", "lines": 209, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/kv_connector/unit/test_cache_pollution_prevention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ test that invalid blocks are evicted from prefix cache to prevent pollution. verifies that when sync-loading fails, invalid blocks are removed from the prefix cache hash table so future requests cannot match and reuse corrupted data. """ from collections.abc import Callable from unittest.mock import Mock import pytest from vllm.v1.core.sched.scheduler import Scheduler from vllm.v1.request import Request, RequestStatus from .utils import ( create_model_runner_output, create_request, create_scheduler, create_vllm_config, ) pytestmark = pytest.mark.cpu_test def _make_get_num_new_matched_tokens( req_num_new_matched_tokens: dict[str, int], async_load: bool, ) -> Callable[[Request, int], tuple[int, bool]]: def get_num_new_matched_tokens(request: Request, _: int) -> tuple[int, bool]: value = req_num_new_matched_tokens.get(request.request_id, 0) return value, async_load return get_num_new_matched_tokens @pytest.fixture def fail_scheduler(): """scheduler with kv_load_failure_policy='fail'""" vllm_config = create_vllm_config() vllm_config.kv_transfer_config.kv_load_failure_policy = "fail" return create_scheduler(vllm_config) def test_invalid_blocks_evicted_prevents_cache_pollution( fail_scheduler: Scheduler, ): """ verify invalid blocks are evicted to prevent future cache hits. scenario: 1. request 1 loads externally-computed blocks (sync mode) 2. some blocks fail to load and are marked invalid 3. with fail policy, invalid blocks should be evicted from prefix cache 4. request is marked as FINISHED_ERROR """ num_prompt_blocks = 100 num_external_computed_blocks = 99 invalid_block_idx = 50 num_prompt_tokens = num_prompt_blocks * fail_scheduler.block_size num_external_computed_tokens = ( num_external_computed_blocks * fail_scheduler.block_size ) # request 1: will have invalid blocks request1 = create_request(num_tokens=num_prompt_tokens, request_id=1) fail_scheduler.add_request(request=request1) req_num_new_matched_tokens = { request1.request_id: num_external_computed_tokens, } # mock connector indicating sync load fail_scheduler.connector = Mock() fail_scheduler.connector.get_num_new_matched_tokens.side_effect = ( _make_get_num_new_matched_tokens(req_num_new_matched_tokens, False) ) fail_scheduler.connector.request_finished.return_value = (False, None) fail_scheduler.connector.take_events.return_value = () scheduler_output = fail_scheduler.schedule() # request should be running with sync KV load assert len(fail_scheduler.running) == 1 assert request1.status == RequestStatus.RUNNING # get allocated block IDs req_block_ids = scheduler_output.scheduled_new_reqs[0].block_ids[0] invalid_block_id = req_block_ids[invalid_block_idx] invalid_block_ids = {invalid_block_id} # get the block object to verify eviction later block = fail_scheduler.kv_cache_manager.block_pool.blocks[invalid_block_id] # cache the blocks to simulate they've been computed and cached # (in real scenario blocks would be cached after compute) fail_scheduler.kv_cache_manager.cache_blocks(request1, num_external_computed_tokens) # verify block has a hash (is cached) before reporting invalid blocks assert block.block_hash is not None, ( f"block {invalid_block_id} should be cached (have a hash) before " f"eviction test, but hash is None" ) # report invalid blocks model_runner_output = create_model_runner_output( [request1], invalid_block_ids=invalid_block_ids, use_eos=False, ) fail_scheduler.update_from_output(scheduler_output, model_runner_output) # verify request finished with error (fail policy) assert request1.status == RequestStatus.FINISHED_ERROR # critical assertion: invalid block and all subsequent blocks should be evicted # all blocks from invalid_block_idx onwards become invalid since they were # computed based on the failed block for idx in range(invalid_block_idx, len(req_block_ids)): block_id = req_block_ids[idx] block_obj = fail_scheduler.kv_cache_manager.block_pool.blocks[block_id] assert block_obj.block_hash is None, ( f"block {block_id} at index {idx} should have been evicted " f"(hash reset to None), but hash is {block_obj.block_hash}. " f"All blocks from index {invalid_block_idx} onwards should be evicted " f"since they depend on the invalid block at index {invalid_block_idx}." ) # verify cache contains exactly the valid blocks (before first affected block) # and none of the invalid blocks (from first affected block onwards) # valid blocks: all blocks before invalid_block_idx should be cached for idx in range(invalid_block_idx): block_id = req_block_ids[idx] block_obj = fail_scheduler.kv_cache_manager.block_pool.blocks[block_id] assert block_obj.block_hash is not None, ( f"valid block {block_id} at index {idx} should still be cached " f"(have a hash), but hash is None. Only blocks from index " f"{invalid_block_idx} onwards should be evicted." ) # invalid blocks: verify they're not in the cached_block_hash_to_block map cached_blocks = ( fail_scheduler.kv_cache_manager.block_pool.cached_block_hash_to_block ) cached_block_ids = { b.block_id for blocks_val in cached_blocks._cache.values() for b in ( [blocks_val] if not isinstance(blocks_val, dict) else blocks_val.values() ) } for idx in range(invalid_block_idx, len(req_block_ids)): block_id = req_block_ids[idx] assert block_id not in cached_block_ids, ( f"invalid block {block_id} at index {idx} should not be in cache hash table" )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/kv_connector/unit/test_cache_pollution_prevention.py", "license": "Apache License 2.0", "lines": 130, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/kv_connector/unit/test_error_propagation.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable from unittest.mock import Mock import pytest from vllm.v1.core.sched.scheduler import Scheduler from vllm.v1.request import FinishReason, Request, RequestStatus from .utils import ( create_model_runner_output, create_request, create_scheduler, create_vllm_config, ) pytestmark = pytest.mark.cpu_test def _make_get_num_new_matched_tokens( req_num_new_matched_tokens: dict[str, int], async_load: bool, ) -> Callable[[Request, int], tuple[int, bool]]: def get_num_new_matched_tokens(request: Request, _: int) -> tuple[int, bool]: value = req_num_new_matched_tokens.get(request.request_id, 0) return value, async_load return get_num_new_matched_tokens @pytest.fixture def fail_scheduler(): """scheduler with kv_load_failure_policy='fail'""" vllm_config = create_vllm_config() vllm_config.kv_transfer_config.kv_load_failure_policy = "fail" return create_scheduler(vllm_config) def test_error_propagation_sync_load(fail_scheduler: Scheduler): """test invalid_block_ids with fail policy -> FINISHED_ERROR (sync load)""" num_prompt_blocks = 100 num_external_computed_blocks = 99 invalid_block_idx = 50 num_prompt_tokens = num_prompt_blocks * fail_scheduler.block_size num_external_computed_tokens = ( num_external_computed_blocks * fail_scheduler.block_size ) request = create_request(num_tokens=num_prompt_tokens) fail_scheduler.add_request(request=request) req_num_new_matched_tokens = { request.request_id: num_external_computed_tokens, } fail_scheduler.connector = Mock() fail_scheduler.connector.get_num_new_matched_tokens.side_effect = ( _make_get_num_new_matched_tokens(req_num_new_matched_tokens, False) ) fail_scheduler.connector.request_finished.return_value = (False, None) fail_scheduler.connector.take_events.return_value = () scheduler_output = fail_scheduler.schedule() assert len(fail_scheduler.running) == 1 assert len(scheduler_output.scheduled_new_reqs) == 1 assert fail_scheduler.connector.get_num_new_matched_tokens.call_count == 1 req_block_ids = scheduler_output.scheduled_new_reqs[0].block_ids[0] invalid_block_ids = {req_block_ids[invalid_block_idx]} model_runner_output = create_model_runner_output( [request], invalid_block_ids=invalid_block_ids, use_eos=True, ) outputs = fail_scheduler.update_from_output(scheduler_output, model_runner_output) assert request.status == RequestStatus.FINISHED_ERROR assert request.get_finished_reason() == FinishReason.ERROR assert len(outputs) == 1 engine_outputs = next(iter(outputs.values())) assert len(engine_outputs.outputs) == 1 output = engine_outputs.outputs[0] assert output.request_id == request.request_id assert output.finish_reason == FinishReason.ERROR assert len(fail_scheduler.running) == 0 def test_error_propagation_async_load(fail_scheduler: Scheduler): """test invalid_block_ids with fail policy -> FINISHED_ERROR (async load)""" num_prompt_blocks = 100 num_external_computed_blocks = 99 invalid_block_idx = 50 num_prompt_tokens = num_prompt_blocks * fail_scheduler.block_size num_external_computed_tokens = ( num_external_computed_blocks * fail_scheduler.block_size ) request = create_request(num_tokens=num_prompt_tokens) fail_scheduler.add_request(request=request) req_num_new_matched_tokens = { request.request_id: num_external_computed_tokens, } fail_scheduler.connector = Mock() fail_scheduler.connector.get_num_new_matched_tokens.side_effect = ( _make_get_num_new_matched_tokens(req_num_new_matched_tokens, True) ) fail_scheduler.connector.request_finished.return_value = (False, None) fail_scheduler.connector.take_events.return_value = () scheduler_output = fail_scheduler.schedule() assert len(fail_scheduler.waiting) == 1 assert request.status == RequestStatus.WAITING_FOR_REMOTE_KVS assert request.num_computed_tokens == 0 (req_block_ids,) = fail_scheduler.kv_cache_manager.get_block_ids(request.request_id) invalid_block_ids = {req_block_ids[invalid_block_idx]} model_runner_output = create_model_runner_output( reqs=[], finished_recving=set(), invalid_block_ids=invalid_block_ids, use_eos=True, ) outputs = fail_scheduler.update_from_output(scheduler_output, model_runner_output) assert request.status == RequestStatus.FINISHED_ERROR assert request.get_finished_reason() == FinishReason.ERROR assert len(outputs) == 1 engine_outputs = next(iter(outputs.values())) assert len(engine_outputs.outputs) == 1 output = engine_outputs.outputs[0] assert output.request_id == request.request_id assert output.finish_reason == FinishReason.ERROR assert len(fail_scheduler.waiting) == 0
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/kv_connector/unit/test_error_propagation.py", "license": "Apache License 2.0", "lines": 111, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/kv_connector/unit/test_invalid_blocks_correctness.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Tests for correctness in invalid block handling. These tests verify correct behavior in three scenarios: 1. Sync recompute case: Blocks should not be freed for running requests that need to recompute invalid blocks 2. Sync fail case: Invalid blocks must be evicted from cache when request fails 3. Async recompute case: Invalid blocks should not be cached after transfer """ from collections.abc import Callable from unittest.mock import Mock import pytest from vllm.v1.core.sched.scheduler import Scheduler from vllm.v1.request import FinishReason, Request, RequestStatus from .utils import ( create_model_runner_output, create_request, create_scheduler, create_vllm_config, ) pytestmark = pytest.mark.cpu_test def _make_get_num_new_matched_tokens( req_num_new_matched_tokens: dict[str, int], async_load: bool, ) -> Callable[[Request, int], tuple[int, bool]]: def get_num_new_matched_tokens(request: Request, _: int) -> tuple[int, bool]: value = req_num_new_matched_tokens.get(request.request_id, 0) return value, async_load return get_num_new_matched_tokens @pytest.fixture def fail_scheduler(): """scheduler with kv_load_failure_policy='fail'""" vllm_config = create_vllm_config() vllm_config.kv_transfer_config.kv_load_failure_policy = "fail" return create_scheduler(vllm_config) @pytest.fixture def recompute_scheduler(): """scheduler with kv_load_failure_policy='recompute'""" vllm_config = create_vllm_config() vllm_config.kv_transfer_config.kv_load_failure_policy = "recompute" return create_scheduler(vllm_config) def test_sync_recompute_blocks_not_freed_for_running_requests( recompute_scheduler: Scheduler, ): """ Test sync recompute case - blocks must not be freed for running requests. When a running request has invalid blocks and retry_policy is 'recompute': 1. Request should remain in RUNNING state 2. num_computed_tokens should be truncated to invalid block boundary 3. Blocks should NOT be freed (request still needs them for recomputation) 4. Request should remain in scheduler.requests and scheduler.running """ num_prompt_blocks = 100 num_external_computed_blocks = 99 invalid_block_idx = 50 num_prompt_tokens = num_prompt_blocks * recompute_scheduler.block_size num_external_computed_tokens = ( num_external_computed_blocks * recompute_scheduler.block_size ) request = create_request(num_tokens=num_prompt_tokens) recompute_scheduler.add_request(request=request) req_num_new_matched_tokens = { request.request_id: num_external_computed_tokens, } # mock connector indicating sync load recompute_scheduler.connector = Mock() recompute_scheduler.connector.get_num_new_matched_tokens.side_effect = ( _make_get_num_new_matched_tokens(req_num_new_matched_tokens, False) ) recompute_scheduler.connector.request_finished.return_value = (False, None) recompute_scheduler.connector.take_events.return_value = () scheduler_output = recompute_scheduler.schedule() # request should be running with sync KV load assert len(recompute_scheduler.running) == 1 assert len(scheduler_output.scheduled_new_reqs) == 1 assert request.status == RequestStatus.RUNNING # get the allocated block IDs before invalid blocks are reported req_block_ids = scheduler_output.scheduled_new_reqs[0].block_ids[0] invalid_block_ids = {req_block_ids[invalid_block_idx]} # store original num_computed_tokens for comparison original_num_computed_tokens = request.num_computed_tokens model_runner_output = create_model_runner_output( [request], invalid_block_ids=invalid_block_ids, use_eos=False, # not finished - should continue running ) outputs = recompute_scheduler.update_from_output( scheduler_output, model_runner_output ) # critical assertions for recompute case: # 1. request should still be RUNNING (not finished, not aborted) assert request.status == RequestStatus.RUNNING, ( f"Request should remain RUNNING for recompute, got {request.status}" ) # 2. num_computed_tokens should be truncated to first invalid block expected_truncated_tokens = invalid_block_idx * recompute_scheduler.block_size assert request.num_computed_tokens == expected_truncated_tokens, ( f"num_computed_tokens should be truncated to {expected_truncated_tokens}, " f"got {request.num_computed_tokens}" ) assert request.num_computed_tokens < original_num_computed_tokens, ( "num_computed_tokens should be reduced after invalid block detection" ) # 3. no output should be generated (request is still running) # the request should be skipped in the output loop assert len(outputs) == 0 or request.request_id not in [ out.request_id for outs in outputs.values() for out in outs.outputs ], "No output should be generated for recompute requests" # 4. request should still be in running queue assert request in recompute_scheduler.running, ( "Request should remain in running queue for recomputation" ) # 5. request should still be in scheduler.requests (not deleted) assert request.request_id in recompute_scheduler.requests, ( "Request should not be deleted from scheduler.requests" ) # 6. blocks should NOT be freed - verify blocks are still allocated try: allocated_blocks = recompute_scheduler.kv_cache_manager.get_block_ids( request.request_id ) assert allocated_blocks is not None assert len(allocated_blocks[0]) > 0, ( "Blocks should still be allocated for recomputation" ) except KeyError: pytest.fail( "Blocks were freed incorrectly! Running requests need their blocks " "to recompute invalid portions." ) # 7. verify request can be rescheduled in next step scheduler_output_2 = recompute_scheduler.schedule() # request should appear in the new schedule to recompute invalid blocks scheduled_req_ids = [ req.request_id for req in scheduler_output_2.scheduled_new_reqs ] if scheduler_output_2.num_scheduled_tokens: scheduled_req_ids.extend(scheduler_output_2.num_scheduled_tokens.keys()) assert ( request.request_id in scheduled_req_ids or len(recompute_scheduler.running) > 0 ), "Request should be reschedulable for recomputation" def test_sync_fail_invalid_blocks_evicted(fail_scheduler: Scheduler): """ Test sync fail case - invalid blocks must be evicted from cache. When a request fails with policy='fail' and has invalid blocks from sync loading: 1. Request should be finished with FINISHED_ERROR 2. Invalid blocks should be evicted from the KV cache 3. Valid blocks (if shared) should remain in cache 4. Future requests should not reuse the invalid blocks This test verifies that invalid blocks are properly evicted to prevent cache corruption and reuse of invalid data. """ num_prompt_blocks = 100 num_external_computed_blocks = 99 invalid_block_idx = 50 num_prompt_tokens = num_prompt_blocks * fail_scheduler.block_size num_external_computed_tokens = ( num_external_computed_blocks * fail_scheduler.block_size ) request = create_request(num_tokens=num_prompt_tokens) fail_scheduler.add_request(request=request) req_num_new_matched_tokens = { request.request_id: num_external_computed_tokens, } # mock connector indicating sync load fail_scheduler.connector = Mock() fail_scheduler.connector.get_num_new_matched_tokens.side_effect = ( _make_get_num_new_matched_tokens(req_num_new_matched_tokens, False) ) fail_scheduler.connector.request_finished.return_value = (False, None) fail_scheduler.connector.take_events.return_value = () scheduler_output = fail_scheduler.schedule() # request should be running with sync KV load assert len(fail_scheduler.running) == 1 assert request.status == RequestStatus.RUNNING # get allocated block IDs req_block_ids = scheduler_output.scheduled_new_reqs[0].block_ids[0] invalid_block_id = req_block_ids[invalid_block_idx] invalid_block_ids = {invalid_block_id} # verify the block is in the block pool before we report it as invalid block = fail_scheduler.kv_cache_manager.block_pool.blocks[invalid_block_id] assert block is not None # report invalid blocks - request should fail model_runner_output = create_model_runner_output( [request], invalid_block_ids=invalid_block_ids, use_eos=True, ) outputs = fail_scheduler.update_from_output(scheduler_output, model_runner_output) # verify request is finished with error assert request.status == RequestStatus.FINISHED_ERROR assert request.get_finished_reason() == FinishReason.ERROR # verify output is generated assert len(outputs) == 1 engine_outputs = next(iter(outputs.values())) assert len(engine_outputs.outputs) == 1 output = engine_outputs.outputs[0] assert output.request_id == request.request_id assert output.finish_reason == FinishReason.ERROR # verify the request was removed from scheduler assert request.request_id not in fail_scheduler.requests assert len(fail_scheduler.running) == 0 # critical: verify invalid block was actually freed from cache # this is the key assertion - the invalid block should no longer be # tracked by the KV cache manager for this request # if it's still there, a future request could reuse the invalid data try: block_ids = fail_scheduler.kv_cache_manager.get_block_ids(request.request_id) # if we get here, check if blocks were actually freed if block_ids is not None and len(block_ids[0]) > 0: pytest.fail( f"Invalid blocks still tracked for finished request! " f"Request {request.request_id} should have been freed but " f"still has {len(block_ids[0])} blocks allocated." ) # blocks list exists but is empty - this is fine, they were freed except KeyError: # expected - request completely removed from tracking pass # critical: verify invalid block was evicted from prefix cache # the block should no longer have a hash (hash is reset on eviction) assert block.block_hash is None, ( f"Invalid block {invalid_block_id} should have been evicted from cache " f"(hash should be None), but hash is still {block.block_hash}" ) # Verify connector prefix cache stats: # - queries = num_prompt_tokens (total tokens not in local cache) # - hits = num_external_computed_tokens (tokens loaded externally) assert engine_outputs.scheduler_stats is not None stats = engine_outputs.scheduler_stats assert stats.connector_prefix_cache_stats is not None conn_stats = stats.connector_prefix_cache_stats assert conn_stats.requests == 1 assert conn_stats.queries == num_prompt_tokens assert conn_stats.hits == num_external_computed_tokens def test_async_recompute_blocks_not_cached_when_invalid( recompute_scheduler: Scheduler, ): """ Test async recompute case - invalid blocks not cached after transfer. When async KV loading has invalid blocks and retry_policy is 'recompute': 1. Blocks are allocated but not cached yet 2. When async transfer completes, only valid blocks should be cached 3. Invalid blocks should never enter the prefix cache This test verifies correctness, the failed_recving_kv_req_ids protection ensures only valid blocks are cached when the transfer completes, and we only evict blocks from cache that are already hashed in the block table. """ from unittest.mock import patch num_prompt_blocks = 100 num_external_computed_blocks = 99 invalid_block_idx = 50 num_prompt_tokens = num_prompt_blocks * recompute_scheduler.block_size num_external_computed_tokens = ( num_external_computed_blocks * recompute_scheduler.block_size ) request = create_request(num_tokens=num_prompt_tokens) recompute_scheduler.add_request(request=request) req_num_new_matched_tokens = { request.request_id: num_external_computed_tokens, } # mock connector indicating async load recompute_scheduler.connector = Mock() recompute_scheduler.connector.get_num_new_matched_tokens.side_effect = ( _make_get_num_new_matched_tokens(req_num_new_matched_tokens, True) ) recompute_scheduler.connector.request_finished.return_value = (False, None) recompute_scheduler.connector.take_events.return_value = () scheduler_output = recompute_scheduler.schedule() # request should be waiting for remote KVs assert len(recompute_scheduler.waiting) == 1 assert request.status == RequestStatus.WAITING_FOR_REMOTE_KVS assert request.num_computed_tokens == 0 # get the allocated block IDs (req_block_ids,) = recompute_scheduler.kv_cache_manager.get_block_ids( request.request_id ) invalid_block_id = req_block_ids[invalid_block_idx] invalid_block_ids = {invalid_block_id} # get the block object to verify it's not cached yet and stays uncached block = recompute_scheduler.kv_cache_manager.block_pool.blocks[invalid_block_id] # verify block has no hash before invalid blocks are reported assert block.block_hash is None, ( "Async loading blocks should not be cached yet (no hash)" ) # report invalid blocks (transfer not finished yet) model_runner_output = create_model_runner_output( reqs=[], finished_recving=None, # transfer NOT finished invalid_block_ids=invalid_block_ids, use_eos=False, ) # critical: spy on evict_blocks to verify it's NOT called for async blocks original_evict_blocks = recompute_scheduler.kv_cache_manager.evict_blocks evict_blocks_calls = [] def evict_blocks_spy(block_ids): evict_blocks_calls.append(set(block_ids)) return original_evict_blocks(block_ids) with patch.object( recompute_scheduler.kv_cache_manager, "evict_blocks", evict_blocks_spy ): outputs = recompute_scheduler.update_from_output( scheduler_output, model_runner_output ) # verify evict_blocks was NOT called (async blocks excluded from eviction) assert len(evict_blocks_calls) == 0, ( f"evict_blocks should not be called for async-only invalid blocks, " f"but was called {len(evict_blocks_calls)} time(s) with {evict_blocks_calls}" ) # request should still be waiting (not finished with error due to recompute policy) assert request.status == RequestStatus.WAITING_FOR_REMOTE_KVS assert request.request_id in recompute_scheduler.failed_recving_kv_req_ids # verify num_computed_tokens was truncated to before invalid block expected_valid_tokens = invalid_block_idx * recompute_scheduler.block_size assert request.num_computed_tokens == expected_valid_tokens # verify invalid block still has no hash (was not evicted) assert block.block_hash is None, ( f"Async loading blocks shouldn't be cached or evicted. " f"Block {invalid_block_id} hash should be None but is {block.block_hash}" ) # Verify connector prefix cache stats: # - queries = num_prompt_tokens (total tokens not in local cache) # - hits = num_external_computed_tokens (tokens loaded externally) assert len(outputs) == 1 engine_outputs = next(iter(outputs.values())) assert engine_outputs.scheduler_stats is not None stats = engine_outputs.scheduler_stats assert stats.connector_prefix_cache_stats is not None conn_stats = stats.connector_prefix_cache_stats assert conn_stats.requests == 1 assert conn_stats.queries == num_prompt_tokens assert conn_stats.hits == num_external_computed_tokens # now simulate async transfer completing model_runner_output_2 = create_model_runner_output( reqs=[], finished_recving={request.request_id}, invalid_block_ids=None, use_eos=False, ) recompute_scheduler.update_from_output(scheduler_output, model_runner_output_2) # verify request is now marked as finished receiving and ready to be processed assert request.request_id in recompute_scheduler.finished_recving_kv_req_ids assert request.request_id in recompute_scheduler.failed_recving_kv_req_ids # critical: verify invalid block still has no hash before recompute # the async transfer invalid data was never cached assert block.block_hash is None, ( f"Invalid block {invalid_block_id} should not be cached before recompute " f"(hash should be None), but hash is {block.block_hash}" ) # critical end-to-end test: spy on cache_blocks to verify it's called with # the truncated num_computed_tokens value original_cache_blocks = recompute_scheduler.kv_cache_manager.cache_blocks cache_blocks_calls = [] def cache_blocks_spy(req, num_tokens): cache_blocks_calls.append((req.request_id, num_tokens)) return original_cache_blocks(req, num_tokens) with patch.object( recompute_scheduler.kv_cache_manager, "cache_blocks", cache_blocks_spy ): # call schedule() again - this triggers _update_waiting_for_remote_kv() # which should call cache_blocks with the truncated value recompute_scheduler.schedule() # verify cache_blocks was called with the truncated value assert len(cache_blocks_calls) == 1, ( f"cache_blocks should be called exactly once, " f"got {len(cache_blocks_calls)} calls" ) cached_req_id, cached_num_tokens = cache_blocks_calls[0] assert cached_req_id == request.request_id assert cached_num_tokens == expected_valid_tokens, ( f"cache_blocks should be called with truncated value {expected_valid_tokens}, " f"but was called with {cached_num_tokens}" ) # request should now be RUNNING (scheduled immediately after transfer completes) # the flow is: WAITING_FOR_REMOTE_KVS -> WAITING -> RUNNING in same schedule() call assert request.status == RequestStatus.RUNNING # num_computed_tokens should be >= expected_valid_tokens because the scheduler # will schedule additional new tokens (up to max_num_batched_tokens) for the request assert request.num_computed_tokens >= expected_valid_tokens, ( f"num_computed_tokens should be at least {expected_valid_tokens}, " f"got {request.num_computed_tokens}" ) # request should no longer be in the failed/finished receiving sets assert request.request_id not in recompute_scheduler.failed_recving_kv_req_ids assert request.request_id not in recompute_scheduler.finished_recving_kv_req_ids # request should be in the running queue assert request in recompute_scheduler.running
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/kv_connector/unit/test_invalid_blocks_correctness.py", "license": "Apache License 2.0", "lines": 388, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:docs/mkdocs/hooks/generate_metrics.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import ast import logging from pathlib import Path from typing import Literal logger = logging.getLogger("mkdocs") ROOT_DIR = Path(__file__).parent.parent.parent.parent DOCS_DIR = ROOT_DIR / "docs" GENERATED_METRICS_DIR = DOCS_DIR / "generated" / "metrics" # Files to scan for metric definitions - each will generate a separate table METRIC_SOURCE_FILES = [ {"path": "vllm/v1/metrics/loggers.py", "output": "general.inc.md"}, { "path": "vllm/v1/spec_decode/metrics.py", "output": "spec_decode.inc.md", }, { "path": "vllm/distributed/kv_transfer/kv_connector/v1/nixl_connector.py", "output": "nixl_connector.inc.md", }, {"path": "vllm/v1/metrics/perf.py", "output": "perf.inc.md"}, ] class MetricExtractor(ast.NodeVisitor): """AST visitor to extract metric definitions.""" def __init__(self): self.metrics: list[dict[str, str]] = [] def visit_Call(self, node: ast.Call) -> None: """Visit function calls to find metric class instantiations.""" metric_type = self._get_metric_type(node) if metric_type: name = self._extract_kwarg(node, "name") documentation = self._extract_kwarg(node, "documentation") if name: self.metrics.append( { "name": name, "type": metric_type, "documentation": documentation or "", } ) self.generic_visit(node) def _get_metric_type(self, node: ast.Call) -> str | None: """Determine if this call creates a metric and return its type.""" metric_type_map = { "_gauge_cls": "gauge", "_counter_cls": "counter", "_histogram_cls": "histogram", } if isinstance(node.func, ast.Attribute): return metric_type_map.get(node.func.attr) return None def _extract_kwarg(self, node: ast.Call, key: str) -> str | None: """Extract a keyword argument value from a function call.""" for keyword in node.keywords: if keyword.arg == key: return self._get_string_value(keyword.value) return None def _get_string_value(self, node: ast.AST) -> str | None: """Extract string value from an AST node.""" if isinstance(node, ast.Constant): return str(node.value) if node.value is not None else None return None def extract_metrics_from_file(filepath: Path) -> list[dict[str, str]]: """Parse a Python file and extract all metric definitions.""" try: with open(filepath, encoding="utf-8") as f: source = f.read() tree = ast.parse(source, filename=str(filepath)) extractor = MetricExtractor() extractor.visit(tree) return extractor.metrics except Exception as e: raise RuntimeError(f"Failed to parse {filepath}: {e}") from e def generate_markdown_table(metrics: list[dict[str, str]]) -> str: """Generate a markdown table from extracted metrics.""" if not metrics: return "No metrics found.\n" # Sort by type, then by name metrics_sorted = sorted(metrics, key=lambda m: (m["type"], m["name"])) lines = [] lines.append("| Metric Name | Type | Description |") lines.append("|-------------|------|-------------|") for metric in metrics_sorted: name = metric["name"] metric_type = metric["type"].capitalize() doc = metric["documentation"].replace("\n", " ").strip() lines.append(f"| `{name}` | {metric_type} | {doc} |") return "\n".join(lines) + "\n" def on_startup(command: Literal["build", "gh-deploy", "serve"], dirty: bool): """Generate metrics documentation tables from source files.""" logger.info("Generating metrics documentation") # Create generated directory if it doesn't exist GENERATED_METRICS_DIR.mkdir(parents=True, exist_ok=True) total_metrics = 0 for source_config in METRIC_SOURCE_FILES: source_path = source_config["path"] output_file = source_config["output"] filepath = ROOT_DIR / source_path if not filepath.exists(): raise FileNotFoundError(f"Metrics source file not found: {filepath}") logger.debug("Extracting metrics from: %s", source_path) metrics = extract_metrics_from_file(filepath) logger.debug("Found %d metrics in %s", len(metrics), source_path) # Generate and write the markdown table for this source table_content = generate_markdown_table(metrics) output_path = GENERATED_METRICS_DIR / output_file with open(output_path, "w", encoding="utf-8") as f: f.write(table_content) total_metrics += len(metrics) logger.info( "Generated metrics table: %s (%d metrics)", output_path.relative_to(ROOT_DIR), len(metrics), ) logger.info( "Total metrics generated: %d across %d files", total_metrics, len(METRIC_SOURCE_FILES), )
{ "repo_id": "vllm-project/vllm", "file_path": "docs/mkdocs/hooks/generate_metrics.py", "license": "Apache License 2.0", "lines": 120, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/e2e/test_async_spec_decode.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Test that verifies no implicit GPU-CPU synchronization occurs during speculative decoding generation under expected conditions. """ import multiprocessing import sys import traceback import pytest import torch @pytest.fixture def sync_tracker(): """ Fixture that patches CommonAttentionMetadata.seq_lens_cpu to detect lazy init syncs. Prints stack traces immediately when syncs occur. """ from vllm.v1.attention.backend import CommonAttentionMetadata # Shared counter for cross-process communication (inherited by fork) sync_count = multiprocessing.Value("i", 0) # Save original property original_prop = CommonAttentionMetadata.seq_lens_cpu original_fget = original_prop.fget # Create tracking wrapper def tracking_seq_lens_cpu(self): if self._seq_lens_cpu is None: # Increment counter with sync_count.get_lock(): sync_count.value += 1 count = sync_count.value # Print stack trace immediately (shows in subprocess output) print(f"\n{'=' * 60}", file=sys.stderr) print(f"SYNC #{count}: seq_lens_cpu lazy init triggered!", file=sys.stderr) print(f"{'=' * 60}", file=sys.stderr) traceback.print_stack(file=sys.stderr) print(f"{'=' * 60}\n", file=sys.stderr) sys.stderr.flush() return original_fget(self) # Apply patch CommonAttentionMetadata.seq_lens_cpu = property(tracking_seq_lens_cpu) class SyncTracker: @property def count(self) -> int: return sync_count.value def assert_no_sync(self, msg: str = ""): count = sync_count.value assert count == 0, ( f"Unexpected GPU-CPU sync: seq_lens_cpu lazy init triggered " f"{count} times. See stack traces above. {msg}" ) yield SyncTracker() # Restore original property CommonAttentionMetadata.seq_lens_cpu = original_prop torch._dynamo.reset() # Test configurations: (model, spec_model, method, num_spec_tokens, backend_env) SPEC_DECODE_CONFIGS = [ pytest.param( "meta-llama/Llama-3.2-1B-Instruct", "nm-testing/Llama3_2_1B_speculator.eagle3", "eagle3", 2, id="eagle3-llama", ), pytest.param( "eagle618/deepseek-v3-random", "eagle618/eagle-deepseek-v3-random", "eagle", 2, id="eagle-mla-deepseek", ), ] @pytest.mark.parametrize( "model,spec_model,method,num_spec_tokens", SPEC_DECODE_CONFIGS, ) def test_no_sync_with_spec_decode( sync_tracker, model: str, spec_model: str, method: str, num_spec_tokens: int, ): """ Test that no implicit GPU-CPU sync occurs during speculative decoding generation. """ # Import vLLM AFTER sync_tracker fixture has applied the patch from vllm import LLM, SamplingParams from vllm.distributed import cleanup_dist_env_and_memory llm = LLM( model=model, max_model_len=256, speculative_config={ "method": method, "num_speculative_tokens": num_spec_tokens, "model": spec_model, }, enforce_eager=True, async_scheduling=True, ) outputs = llm.generate( ["Hello, my name is"], SamplingParams(temperature=0, max_tokens=10), ) assert len(outputs) == 1 assert len(outputs[0].outputs[0].text) > 0 del llm torch.cuda.empty_cache() cleanup_dist_env_and_memory() sync_tracker.assert_no_sync()
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/e2e/test_async_spec_decode.py", "license": "Apache License 2.0", "lines": 110, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/config/profiler.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os from typing import Any, Literal from pydantic import Field, model_validator from typing_extensions import Self from vllm.config.utils import config from vllm.logger import init_logger from vllm.utils.hashing import safe_hash logger = init_logger(__name__) ProfilerKind = Literal["torch", "cuda"] def _is_uri_path(path: str) -> bool: """Check if path is a URI (scheme://...), excluding Windows drive letters. Supports custom URI schemes like gs://, s3://, hdfs://, etc. These paths should not be converted to absolute paths. """ if "://" in path: scheme = path.split("://")[0] # Windows drive letters are single characters (e.g., C://) # Valid URI schemes have more than one character return len(scheme) > 1 return False @config class ProfilerConfig: """Dataclass which contains profiler config for the engine.""" profiler: ProfilerKind | None = None """Which profiler to use. Defaults to None. Options are: - 'torch': Use PyTorch profiler.\n - 'cuda': Use CUDA profiler.""" torch_profiler_dir: str = "" """Directory to save torch profiler traces. Both AsyncLLM's CPU traces and worker's traces (CPU & GPU) will be saved under this directory. Note that it must be an absolute path.""" torch_profiler_with_stack: bool = True """If `True`, enables stack tracing in the torch profiler. Enabled by default.""" torch_profiler_with_flops: bool = False """If `True`, enables FLOPS counting in the torch profiler. Disabled by default.""" torch_profiler_use_gzip: bool = True """If `True`, saves torch profiler traces in gzip format. Enabled by default""" torch_profiler_dump_cuda_time_total: bool = True """If `True`, dumps total CUDA time in torch profiler traces. Enabled by default.""" torch_profiler_record_shapes: bool = False """If `True`, records tensor shapes in the torch profiler. Disabled by default.""" torch_profiler_with_memory: bool = False """If `True`, enables memory profiling in the torch profiler. Disabled by default.""" ignore_frontend: bool = False """If `True`, disables the front-end profiling of AsyncLLM when using the 'torch' profiler. This is needed to reduce overhead when using delay/limit options, since the front-end profiling does not track iterations and will capture the entire range. """ delay_iterations: int = Field(default=0, ge=0) """Number of engine iterations to skip before starting profiling. Defaults to 0, meaning profiling starts immediately after receiving /start_profile. """ max_iterations: int = Field(default=0, ge=0) """Maximum number of engine iterations to profile after starting profiling. Defaults to 0, meaning no limit. """ def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # no factors to consider. # this config will not affect the computation graph. factors: list[Any] = [] hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str @model_validator(mode="after") def _validate_profiler_config(self) -> Self: has_delay_or_limit = self.delay_iterations > 0 or self.max_iterations > 0 if self.profiler == "torch" and has_delay_or_limit and not self.ignore_frontend: logger.warning_once( "Using 'torch' profiler with delay_iterations or max_iterations " "while ignore_frontend is False may result in high overhead." ) profiler_dir = self.torch_profiler_dir if profiler_dir and self.profiler != "torch": raise ValueError( "torch_profiler_dir is only applicable when profiler is set to 'torch'" ) if self.profiler == "torch" and not profiler_dir: raise ValueError("torch_profiler_dir must be set when profiler is 'torch'") # Support any URI scheme (gs://, s3://, hdfs://, etc.) # These paths should not be converted to absolute paths if profiler_dir and not _is_uri_path(profiler_dir): self.torch_profiler_dir = os.path.abspath(os.path.expanduser(profiler_dir)) return self
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/config/profiler.py", "license": "Apache License 2.0", "lines": 96, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/worker/gpu/metrics/logits.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from torch._inductor.runtime.triton_helpers import libdevice from vllm.triton_utils import tl, triton @triton.jit def _num_nans_kernel( logits_ptr, logits_stride, num_nans_ptr, vocab_size, BLOCK_SIZE: tl.constexpr, ): req_idx = tl.program_id(0) num_nans = 0 for i in range(0, vocab_size, BLOCK_SIZE): block = i + tl.arange(0, BLOCK_SIZE) mask = block < vocab_size logits = tl.load( logits_ptr + req_idx * logits_stride + block, mask=mask, other=0 ) logits = logits.to(tl.float32) is_nan = libdevice.isnan(logits).to(tl.int1) num_nans += tl.sum(is_nan).to(tl.int32) tl.store(num_nans_ptr + req_idx, num_nans) def get_num_nans(logits: torch.Tensor) -> torch.Tensor: num_reqs, vocab_size = logits.shape BLOCK_SIZE = 8192 num_nans = torch.empty(num_reqs, dtype=torch.int32, device=logits.device) _num_nans_kernel[(num_reqs,)]( logits, logits.stride(0), num_nans, vocab_size, BLOCK_SIZE=BLOCK_SIZE, ) return num_nans
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/worker/gpu/metrics/logits.py", "license": "Apache License 2.0", "lines": 37, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/worker/gpu/sample/output.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass import torch from vllm.v1.outputs import LogprobsTensors @dataclass class SamplerOutput: sampled_token_ids: torch.Tensor logprobs_tensors: LogprobsTensors | None num_nans: torch.Tensor | None
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/worker/gpu/sample/output.py", "license": "Apache License 2.0", "lines": 10, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/quantization/test_cutlass_w4a8_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Tests for the CUTLASS-based W4A8 grouped GEMM kernel and the full MoE layer. """ import random from dataclasses import dataclass import pytest import torch from vllm import _custom_ops as ops from vllm.model_executor.layers.quantization.utils.quant_utils import ( pack_rows, quantize_weights, ) from vllm.platforms import current_platform from vllm.scalar_type import ScalarType, scalar_types from vllm.utils.torch_utils import set_random_seed IS_SUPPORTED_BY_GPU = ( current_platform.is_cuda() and current_platform.get_device_capability()[0] >= 9 ) def to_fp8(tensor: torch.Tensor) -> torch.Tensor: finfo = torch.finfo(torch.float8_e4m3fn) return tensor.clamp(min=finfo.min, max=finfo.max).to(dtype=torch.float8_e4m3fn) def cutlass_quantize( atype: torch.dtype, w: torch.Tensor, wtype: ScalarType, stype: torch.dtype | None, group_size: int | None, zero_points: bool = False, ): """ Quantize weights into W4 and compute reference dequantized weights. Encoding/reordering of weights and packing of scales is deferred until after all experts are combined. """ assert wtype.is_integer(), "TODO: support floating point weights" w_ref, w_q, w_s, w_zp = quantize_weights( w, wtype, group_size=group_size, zero_points=zero_points ) # Since scales are later cast to fp8, recompute w_ref in atype here. w_ref = ( w_q.to(torch.float32) * w_s.to(atype).to(torch.float32).repeat_interleave(group_size, dim=0) ).to(atype) # Bit mask prevents sign extension of int4 when packing. w_q = pack_rows(w_q & 0x0F, wtype.size_bits, *w_q.shape) # Make weights row-major (N, K). w_q = w_q.t().contiguous() return w_ref, w_q, w_s.to(atype), w_zp def cutlass_preprocess( w_q_experts: list[torch.Tensor], w_s_experts: list[torch.Tensor] ): """ Reorder/encode expert weights and pack scales. Returns: w_q_packed: Packed/encoded int4 weights for all experts. w_s_packed: Packed fp8 scales for all experts. packed_layout: Layout/stride metadata for grouped GEMM. """ w_s_packed = ops.cutlass_pack_scale_fp8(torch.stack(w_s_experts)) w_q_packed, packed_layout = ops.cutlass_encode_and_reorder_int4b_grouped( torch.stack(w_q_experts) ) # expects dim 3 return w_q_packed, w_s_packed, packed_layout GROUP_SIZE = 128 # (num_experts, N, K) TEST_SHAPES = [ (8, 512, 2048), (8, 2048, 2048), (64, 512, 1024), (64, 2048, 2048), (4, 2048, 768), (8, 768, 2048), (64, 1536, 2048), (128, 8192, 4096), # test overflow int32 ] ALIGNMENT = 16 # torch._scaled_mm alignment for M, needed for reference check @dataclass class MoETestSetup: num_experts: int K: int N: int Ms: list[int] M_full: int a: torch.Tensor a_ref: torch.Tensor a_strides: torch.Tensor out: torch.Tensor c_strides: torch.Tensor per_tok_scales: torch.Tensor per_chan_scales: torch.Tensor w_refs: list[torch.Tensor] w_q_packed: torch.Tensor w_s_packed: torch.Tensor problem_sizes: torch.Tensor expert_offsets: torch.Tensor b_strides: torch.Tensor group_scale_strides: torch.Tensor def make_moe_test_setup( num_experts: int, K: int, N: int, *, alignment: int = ALIGNMENT, max_blocks: int = 64, device: str = "cuda", random_zero: bool = False, ) -> MoETestSetup: """Create a full set of tensors for testing cutlass_w4a8_moe_mm.""" assert K % GROUP_SIZE == 0 # Token counts per expert (multiples of `alignment`). Ms = [alignment * random.randint(1, max_blocks) for _ in range(num_experts)] # set random experts to 0 tokens if random_zero and num_experts > 1: num_zero = max(1, num_experts // 8) zero_indices = random.sample(range(num_experts), k=num_zero) for idx in zero_indices: Ms[idx] = 0 M_full = sum(Ms) assert M_full > 0 # Activations. a = to_fp8(torch.randn((M_full, K), device=device)) a_ref = a.to(torch.float32) a_strides = torch.full((num_experts,), K, dtype=torch.int64, device=device) # Output buffer. out = torch.empty((M_full, N), dtype=torch.bfloat16, device=device) c_strides = torch.full((num_experts,), N, dtype=torch.int64, device=device) # Channel/token scales. per_tok_scales = torch.randn((M_full, 1), dtype=torch.float32, device=device) per_chan_scales = torch.randn( (num_experts, N, 1), dtype=torch.float32, device=device ) # Expert weights and scales. wtype = scalar_types.int4 atype = stype = torch.float8_e4m3fn w_refs, w_qs, w_ss = [], [], [] for _ in range(num_experts): b = to_fp8(torch.randn((K, N), device=device)) w_ref, w_q, w_s, _ = cutlass_quantize( atype, b.to(torch.float16), wtype, stype, GROUP_SIZE, zero_points=False ) w_refs.append(w_ref) w_qs.append(w_q) w_ss.append(w_s) w_q_packed, w_s_packed, packed_layout = cutlass_preprocess(w_qs, w_ss) problem_sizes = torch.tensor( [[N, M, K] for M in Ms], dtype=torch.int32, device=device ) expert_offsets = torch.cat( [ torch.tensor([0], dtype=torch.int64), torch.cumsum(torch.tensor(Ms, dtype=torch.int64), dim=0)[:-1], ] ).to(device=device) # B strides and group scale strides. b_strides = packed_layout group_scale_strides = torch.zeros( (num_experts, 2), dtype=torch.int64, device=device ) group_scale_strides[:, 0] = N return MoETestSetup( num_experts=num_experts, K=K, N=N, Ms=Ms, M_full=M_full, a=a, a_ref=a_ref, a_strides=a_strides, out=out, c_strides=c_strides, per_tok_scales=per_tok_scales, per_chan_scales=per_chan_scales, w_refs=w_refs, w_q_packed=w_q_packed, w_s_packed=w_s_packed, problem_sizes=problem_sizes, expert_offsets=expert_offsets, b_strides=b_strides, group_scale_strides=group_scale_strides, ) def compute_moe_reference_output(setup: MoETestSetup) -> torch.Tensor: """Compute reference output using torch._scaled_mm per expert.""" out_ref = torch.empty_like(setup.out) ends = torch.cumsum(torch.tensor(setup.Ms), 0).tolist() starts = setup.expert_offsets.cpu().tolist() for i in range(setup.num_experts): start, end = starts[i], ends[i] if start == end: continue out_ref_i = torch._scaled_mm( setup.a_ref[start:end].to(torch.float8_e4m3fn), setup.w_refs[i].to(torch.float8_e4m3fn).t().contiguous().t(), setup.per_tok_scales[start:end], # (M, 1) setup.per_chan_scales[i].reshape(1, -1), # (1, N) out_dtype=torch.bfloat16, use_fast_accum=True, ) out_ref[start:end] = out_ref_i return out_ref @pytest.mark.skipif( not IS_SUPPORTED_BY_GPU, reason="W4A8 Grouped GEMM is not supported on this GPU type.", ) @pytest.mark.parametrize("shape", TEST_SHAPES) @pytest.mark.parametrize("random_zero", [True, False]) def test_cutlass_w4a8_moe_mm_end_to_end(shape, random_zero): num_experts, N, K = shape set_random_seed(42) setup = make_moe_test_setup( num_experts=num_experts, K=K, N=N, max_blocks=64, random_zero=random_zero ) ops.cutlass_w4a8_moe_mm( setup.out, setup.a, setup.w_q_packed, setup.per_tok_scales, setup.per_chan_scales, setup.w_s_packed, GROUP_SIZE, setup.expert_offsets, setup.problem_sizes, setup.a_strides, setup.b_strides, setup.c_strides, setup.group_scale_strides, ) torch.cuda.synchronize() out_ref = compute_moe_reference_output(setup) torch.testing.assert_close(setup.out, out_ref, rtol=1e-2, atol=1e-2) class W4A8MoELayer(torch.nn.Module): """ Minimal wrapper module to test cuda graphs """ def __init__(self, setup: MoETestSetup): super().__init__() self.setup = setup def forward(self, a: torch.Tensor) -> torch.Tensor: s = self.setup ops.cutlass_w4a8_moe_mm( s.out, a, s.w_q_packed, s.per_tok_scales, s.per_chan_scales, s.w_s_packed, GROUP_SIZE, s.expert_offsets, s.problem_sizes, s.a_strides, s.b_strides, s.c_strides, s.group_scale_strides, ) return s.out @pytest.mark.skipif( not IS_SUPPORTED_BY_GPU, reason="W4A8 Grouped GEMM is not supported on this GPU type.", ) def test_cutlass_w4a8_moe_mm_cuda_graph(): set_random_seed(42) # Fixed config for CUDA graph test (single parameter point). num_experts = 8 K = 512 N = 2048 setup = make_moe_test_setup( num_experts=num_experts, K=K, N=N, max_blocks=32, ) # Construct model that calls the grouped GEMM kernel. model = W4A8MoELayer(setup) # Build reference output once. out_ref = compute_moe_reference_output(setup) # Capture and run the model in a CUDA graph. a_static = setup.a.clone() # static input tensor for graph replay stream = torch.cuda.Stream() with torch.cuda.stream(stream): g = torch.cuda.CUDAGraph() with torch.cuda.graph(g): out_static = model(a_static) out_static.zero_() g.replay() torch.testing.assert_close(out_static, out_ref, rtol=1e-2, atol=1e-2)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/quantization/test_cutlass_w4a8_moe.py", "license": "Apache License 2.0", "lines": 286, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/lora/lora_model.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os import safetensors import torch from vllm.logger import init_logger from vllm.lora.lora_weights import LoRALayerWeights from vllm.lora.peft_helper import PEFTHelper from vllm.lora.utils import ( get_lora_id, is_base_embeddding_weights, parse_fine_tuned_lora_name, ) from vllm.model_executor.model_loader.tensorizer import TensorizerConfig from vllm.model_executor.models.utils import WeightsMapper from vllm.utils.platform_utils import is_pin_memory_available logger = init_logger(__name__) class LoRAModel: """A LoRA fine-tuned model.""" def __init__( self, lora_model_id: int, rank: int, loras: dict[str, LoRALayerWeights], ) -> None: """ Args: lora_model_id: The integer id for the lora model. rank: lora rank. loras: module name -> weights for lora-replaced layers. """ self.id = lora_model_id assert lora_model_id > 0, ( f"a valid lora id should be greater than 0, got {self.id}" ) self.rank = rank self.loras: dict[str, LoRALayerWeights] = loras def clone(self, lora_model_id: int) -> "LoRAModel": """Return a copy of the object with different ids. Will share the underlying tensors.""" return self.__class__( lora_model_id, rank=self.rank, loras=self.loras.copy(), ) def get_lora(self, module_name: str) -> LoRALayerWeights | None: """Get LoRA for a given module by name""" return self.loras.get(module_name, None) def check_lora_name(self, lora_name: str) -> bool: return lora_name in self.loras @staticmethod def _should_skip_module(module_name: str, skip_prefixes: list[str]) -> bool: """Check if a module should be skipped based on skip prefixes""" for prefix in skip_prefixes: if f".{prefix}" in module_name or module_name.startswith(prefix): return True return False @classmethod def from_lora_tensors( cls, lora_model_id: int, tensors: dict[str, torch.Tensor], peft_helper: PEFTHelper, device: str = "cuda", dtype: torch.dtype | None = None, model_vocab_size: int | None = None, weights_mapper: WeightsMapper | None = None, skip_prefixes: list[str] | None = None, ) -> "LoRAModel": """Create a LoRAModel from a dictionary of tensors.""" pin_memory = str(device) == "cpu" and is_pin_memory_available() loras: dict[str, LoRALayerWeights] = {} for tensor_name, tensor in tensors.items(): if is_base_embeddding_weights(tensor_name): continue # Skip modules based on model-defined prefixes (e.g., MTP layers) if skip_prefixes and cls._should_skip_module(tensor_name, skip_prefixes): continue module_name, is_lora_a = parse_fine_tuned_lora_name( tensor_name, weights_mapper ) if module_name not in loras: loras[module_name] = LoRALayerWeights.from_config( module_name, peft_helper ) if is_lora_a: if ( "lora_embedding_A" in tensor_name and model_vocab_size is not None and model_vocab_size != tensor.shape[1] ): raise RuntimeError( f"The embedding LoRA size({tensor.shape[1]}) must be consistent" f" with the base model's vocabulary size({model_vocab_size})." ) loras[module_name].lora_a = tensor.to(device=device, dtype=dtype) if pin_memory: loras[module_name].lora_a = loras[module_name].lora_a.pin_memory() else: loras[module_name].lora_b = tensor.to(device=device, dtype=dtype) if pin_memory: loras[module_name].lora_b = loras[module_name].lora_b.pin_memory() return cls(lora_model_id, peft_helper.r, loras) @classmethod def from_local_checkpoint( cls, lora_dir: str, expected_lora_modules: set[str], peft_helper: PEFTHelper, *, lora_model_id: int | None = None, device: str = "cuda", dtype: torch.dtype | None = None, model_vocab_size: int | None = None, weights_mapper: WeightsMapper | None = None, tensorizer_config_dict: dict | None = None, skip_prefixes: list[str] | None = None, ) -> "LoRAModel": """Create a LoRAModel from a local checkpoint. Args: lora_dir: The local path that has lora data. expected_lora_modules: Name of modules that are expected to be replaced by lora. peft_helper: Loaded lora configuration information. lora_model_id: LoRA model id. If not given, automatically set by a global counter. device: Device where the lora model is loaded. dtype: dtype of the lora model weights. skip_prefixes: List of module name prefixes to skip during loading. Models can define this to skip modules not used in inference (e.g., MTP layers). Format: ["mtp."] Returns: Loaded LoRA Model. """ lora_tensor_path = os.path.join(lora_dir, "adapter_model.safetensors") lora_bin_file_path = os.path.join(lora_dir, "adapter_model.bin") lora_pt_file_path = os.path.join(lora_dir, "adapter_model.pt") tensors: dict[str, torch.Tensor] = {} unexpected_modules: list[list[str] | str] = [] def check_unexpected_modules(modules: dict): for lora_module in modules.keys(): # noqa if is_base_embeddding_weights(lora_module): continue # Handle PEFT file format where experts.base_layer is the # gate_up_proj and experts is the down_proj if "base_layer" in lora_module: continue # Skip modules based on model-defined prefixes if skip_prefixes and cls._should_skip_module( lora_module, skip_prefixes ): continue module_name, _ = parse_fine_tuned_lora_name(lora_module, weights_mapper) # Case for expert lora weights if ".experts" in module_name: expert_idx = module_name.find(".experts") expert_suffix = module_name[expert_idx + 1 :] if expert_suffix not in expected_lora_modules: unexpected_modules.append(module_name) elif module_name.rsplit(".", 1)[-1] not in expected_lora_modules: unexpected_modules.append(module_name) if unexpected_modules: raise ValueError( f"While loading {lora_dir}, expected" f" target modules in {expected_lora_modules}" f" but received {unexpected_modules}." f" Please verify that the loaded LoRA module is correct" ) if tensorizer_config_dict: from tensorizer import TensorDeserializer tensorizer_config = TensorizerConfig(**tensorizer_config_dict) lora_tensor_path = os.path.join( tensorizer_config.tensorizer_dir, "adapter_model.tensors" ) tensorizer_args = tensorizer_config._construct_tensorizer_args() tensors = TensorDeserializer( lora_tensor_path, dtype=tensorizer_config.dtype, **tensorizer_args.deserialization_kwargs, ) check_unexpected_modules(tensors) elif os.path.isfile(lora_tensor_path): # Find unexpected modules. # Use safetensor key as a source of truth to find expected modules. # in peft if you have target_modules A, B, C and C does not exist # in the model it won’t error and model will be trained with A, B # loraified. C won’t exist in the safetensor but it will exist in # the target_modules of the adapter_config.json. unexpected_modules = [] with safetensors.safe_open(lora_tensor_path, framework="pt") as f: # type: ignore # Load tensors if there are only expected modules. check_unexpected_modules(f) for module in f.keys(): # noqa tensors[module] = f.get_tensor(module) elif os.path.isfile(lora_bin_file_path) or os.path.isfile(lora_pt_file_path): lora_file_path = ( lora_bin_file_path if os.path.isfile(lora_bin_file_path) else lora_pt_file_path ) tensors = torch.load(lora_file_path, map_location=device, weights_only=True) check_unexpected_modules(tensors) else: raise ValueError(f"{lora_dir} doesn't contain tensors") return cls.from_lora_tensors( lora_model_id=get_lora_id() if lora_model_id is None else lora_model_id, tensors=tensors, peft_helper=peft_helper, device=device, dtype=dtype, model_vocab_size=model_vocab_size, weights_mapper=weights_mapper, skip_prefixes=skip_prefixes, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/lora/lora_model.py", "license": "Apache License 2.0", "lines": 214, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/entrypoints/openai/tool_parsers/test_gigachat3_tool_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json import pytest from tests.entrypoints.openai.tool_parsers.utils import ( run_tool_extraction, run_tool_extraction_streaming, ) from vllm.entrypoints.openai.engine.protocol import FunctionCall from vllm.tokenizers import TokenizerLike from vllm.tool_parsers import ToolParser, ToolParserManager SIMPLE_ARGS_DICT = { "action": "create", "id": "preferences", } SIMPLE_FUNCTION_JSON = json.dumps( { "name": "manage_user_memory", "arguments": SIMPLE_ARGS_DICT, }, ensure_ascii=False, ) SIMPLE_FUNCTION_OUTPUT = "function call" + SIMPLE_FUNCTION_JSON SIMPLE_FUNCTION_CALL = FunctionCall( name="manage_user_memory", arguments=json.dumps(SIMPLE_ARGS_DICT, ensure_ascii=False), ) PARAMETERLESS_FUNCTION_JSON = json.dumps( { "name": "manage_user_memory", "arguments": {}, }, ensure_ascii=False, ) PARAMETERLESS_FUNCTION_OUTPUT = "function call" + PARAMETERLESS_FUNCTION_JSON PARAMETERLESS_FUNCTION_CALL = FunctionCall( name="manage_user_memory", arguments=json.dumps({}, ensure_ascii=False), ) COMPLEX_ARGS_DICT = { "action": "create", "id": "preferences", "content": { "short_answers": True, "hate_emojis": True, "english_ui": False, "russian_math_explanations": True, }, } COMPLEX_FUNCTION_JSON = json.dumps( { "name": "manage_user_memory", "arguments": COMPLEX_ARGS_DICT, }, ensure_ascii=False, ) COMPLEX_FUNCTION_OUTPUT = "function call" + COMPLEX_FUNCTION_JSON COMPLEX_FUNCTION_CALL = FunctionCall( name="manage_user_memory", arguments=json.dumps(COMPLEX_ARGS_DICT, ensure_ascii=False), ) @pytest.mark.parametrize("streaming", [True, False]) def test_no_tool_call(streaming: bool, default_tokenizer: TokenizerLike): tool_parser: ToolParser = ToolParserManager.get_tool_parser("gigachat3")( default_tokenizer ) model_output = "How can I help you today?" content, tool_calls = run_tool_extraction( tool_parser, model_output, streaming=streaming ) assert content == model_output assert len(tool_calls) == 0 TEST_CASES = [ pytest.param( True, SIMPLE_FUNCTION_OUTPUT, [SIMPLE_FUNCTION_CALL], None, id="simple_streaming", ), pytest.param( False, SIMPLE_FUNCTION_OUTPUT, [SIMPLE_FUNCTION_CALL], None, id="simple_nonstreaming", ), pytest.param( True, PARAMETERLESS_FUNCTION_OUTPUT, [PARAMETERLESS_FUNCTION_CALL], None, id="parameterless_streaming", ), pytest.param( False, PARAMETERLESS_FUNCTION_OUTPUT, [PARAMETERLESS_FUNCTION_CALL], None, id="parameterless_nonstreaming", ), pytest.param( True, COMPLEX_FUNCTION_OUTPUT, [COMPLEX_FUNCTION_CALL], None, id="complex_streaming", ), pytest.param( False, COMPLEX_FUNCTION_OUTPUT, [COMPLEX_FUNCTION_CALL], None, id="complex_nonstreaming", ), ] @pytest.mark.parametrize( "streaming, model_output, expected_tool_calls, expected_content", TEST_CASES ) def test_tool_call( streaming: bool, model_output: str, expected_tool_calls: list[FunctionCall], expected_content: str | None, default_tokenizer: TokenizerLike, ): tool_parser: ToolParser = ToolParserManager.get_tool_parser("gigachat3")( default_tokenizer ) content, tool_calls = run_tool_extraction( tool_parser, model_output, streaming=streaming ) assert content == expected_content assert len(tool_calls) == len(expected_tool_calls) for actual, expected in zip(tool_calls, expected_tool_calls): assert actual.type == "function" assert actual.function.name == expected.name actual_args = json.loads(actual.function.arguments) expected_args = json.loads(expected.arguments) assert actual_args == expected_args def test_streaming_tool_call_with_large_steps(default_tokenizer: TokenizerLike): tool_parser: ToolParser = ToolParserManager.get_tool_parser("gigachat3")( default_tokenizer ) model_output_deltas = [ "function call", COMPLEX_FUNCTION_JSON[:40], COMPLEX_FUNCTION_JSON[40:], ] reconstructor = run_tool_extraction_streaming( tool_parser, model_output_deltas, assert_one_tool_per_delta=False, ) assert len(reconstructor.tool_calls) == 1 call = reconstructor.tool_calls[0] assert call.type == "function" assert call.function.name == "manage_user_memory" args_dict = json.loads(call.function.arguments) assert args_dict == COMPLEX_ARGS_DICT
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/tool_parsers/test_gigachat3_tool_parser.py", "license": "Apache License 2.0", "lines": 160, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/v1/worker/gpu/sample/min_p.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.triton_utils import tl, triton @triton.jit def _min_p_kernel( logits_ptr, logits_stride, idx_mapping_ptr, min_p_ptr, vocab_size, BLOCK_SIZE: tl.constexpr, ): req_idx = tl.program_id(0) req_state_idx = tl.load(idx_mapping_ptr + req_idx) min_p = tl.load(min_p_ptr + req_state_idx).to(tl.float32) if min_p == 0.0: return max_val = float("-inf") for i in range(0, vocab_size, BLOCK_SIZE): block = i + tl.arange(0, BLOCK_SIZE) mask = block < vocab_size logits = tl.load( logits_ptr + req_idx * logits_stride + block, mask=mask, other=float("-inf") ) max_val = tl.max(tl.maximum(logits, max_val)) max_val = max_val.to(tl.float32) # type: ignore threshold = max_val + tl.log(min_p) for i in range(0, vocab_size, BLOCK_SIZE): block = i + tl.arange(0, BLOCK_SIZE) mask = block < vocab_size logits = tl.load( logits_ptr + req_idx * logits_stride + block, mask=mask, other=float("-inf") ) logits = tl.where(logits < threshold, float("-inf"), logits) tl.store(logits_ptr + req_idx * logits_stride + block, logits, mask=mask) def apply_min_p( logits: torch.Tensor, idx_mapping: torch.Tensor, min_p: torch.Tensor ) -> None: num_reqs, vocab_size = logits.shape BLOCK_SIZE = 1024 _min_p_kernel[(num_reqs,)]( logits, logits.stride(0), idx_mapping, min_p, vocab_size, BLOCK_SIZE=BLOCK_SIZE, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/worker/gpu/sample/min_p.py", "license": "Apache License 2.0", "lines": 49, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/entrypoints/openai/tool_parsers/test_openai_tool_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json import jsonschema import openai import pytest import pytest_asyncio from rapidfuzz import fuzz from ....utils import RemoteOpenAIServer MODEL_NAME = "openai/gpt-oss-20b" @pytest.fixture(scope="module") def server(): args = [ "--max-model-len", "8192", "--enforce-eager", "--enable-auto-tool-choice", "--tool-call-parser", "openai", ] with RemoteOpenAIServer(MODEL_NAME, args) as remote_server: yield remote_server @pytest_asyncio.fixture async def client(server): """Async fixture providing an OpenAI-compatible vLLM client.""" async with server.get_async_client() as async_client: yield async_client # ========================================================== # Tool Definitions # ========================================================== TOOLS = [ { "type": "function", "function": { "name": "calculator", "description": "Performs basic arithmetic calculations.", "parameters": { "type": "object", "properties": { "expression": { "type": "string", "description": ( "Arithmetic expression to evaluate, e.g. '123 + 456'." ), } }, "required": ["expression"], }, }, }, { "type": "function", "function": { "name": "get_time", "description": "Retrieves the current local time for a given city.", "parameters": { "type": "object", "properties": { "city": { "type": "string", "description": "City name, e.g. 'New York'.", } }, "required": ["city"], }, }, }, ] # ========================================================== # Message Examples # ========================================================== MESSAGES_CALC = [ {"role": "user", "content": "Calculate 123 + 456 using the calculator."} ] MESSAGES_GET_TIME = [ {"role": "user", "content": "What is the current time in New York?"} ] MESSAGES_MULTIPLE_CALLS = [ { "role": "system", "content": ( "You can call multiple tools. " "When using more than one, return single JSON object with tool_calls array" "containing each tool call with its function name and arguments. " "Do not output multiple JSON objects separately." ), }, { "role": "user", "content": "First, calculate 7 * 8 using the calculator. " "Then, use get_time to tell me the current time in New York.", }, ] MESSAGES_INVALID_CALL = [ { "role": "user", "content": "Can you help with something, " "but don’t actually perform any calculation?", } ] # Expected outputs FUNC_CALC = "calculator" FUNC_ARGS_CALC = '{"expression":"123 + 456"}' FUNC_TIME = "get_time" FUNC_ARGS_TIME = '{"city": "New York"}' # ========================================================== # Utility to extract reasoning and tool calls # ========================================================== def extract_reasoning_and_calls(chunks: list) -> tuple[str, list[str], list[str]]: """ Extract accumulated reasoning text and tool call arguments from streaming chunks. """ reasoning: str = "" tool_calls: dict[int, dict[str, str]] = {} for chunk in chunks: choice = getattr(chunk.choices[0], "delta", None) if not choice: continue if hasattr(choice, "reasoning") and choice.reasoning: reasoning += choice.reasoning for tc in getattr(choice, "tool_calls", []) or []: idx = getattr(tc, "index", 0) tool_entry = tool_calls.setdefault(idx, {"name": "", "arguments": ""}) if getattr(tc, "function", None): func = tc.function if getattr(func, "name", None): tool_entry["name"] = func.name if getattr(func, "arguments", None): tool_entry["arguments"] += func.arguments function_names: list[str] = [v["name"] for _, v in sorted(tool_calls.items())] arguments: list[str] = [v["arguments"] for _, v in sorted(tool_calls.items())] return reasoning, arguments, function_names # ========================================================== # Test Scenarios # ========================================================== @pytest.mark.asyncio async def test_calculator_tool_call_and_argument_accuracy(client: openai.AsyncOpenAI): """Verify calculator tool call is made and arguments are accurate.""" response = await client.chat.completions.create( model=MODEL_NAME, messages=MESSAGES_CALC, tools=TOOLS, temperature=0.0, stream=False, ) message = response.choices[0].message tool_calls = getattr(message, "tool_calls", []) assert tool_calls, "No tool calls detected" calc_call = next((c for c in tool_calls if c.function.name == FUNC_CALC), None) assert calc_call, "Calculator function not called" raw_args = calc_call.function.arguments assert raw_args, "Calculator arguments missing" assert "123" in raw_args and "456" in raw_args, ( f"Expected values not in raw arguments: {raw_args}" ) try: parsed_args = json.loads(raw_args) except json.JSONDecodeError: pytest.fail(f"Invalid JSON in calculator arguments: {raw_args}") expected_expr = "123 + 456" actual_expr = parsed_args.get("expression", "") similarity = fuzz.ratio(actual_expr, expected_expr) assert similarity > 90, ( f"Expression mismatch: expected '{expected_expr}' " f"got '{actual_expr}' (similarity={similarity}%)" ) @pytest.mark.asyncio async def test_streaming_tool_call_get_time_with_reasoning(client: openai.AsyncOpenAI): """Verify streamed reasoning and tool call behavior for get_time.""" stream = await client.chat.completions.create( model=MODEL_NAME, messages=MESSAGES_GET_TIME, tools=TOOLS, temperature=0.0, stream=True, ) chunks = [chunk async for chunk in stream] reasoning, arguments, function_names = extract_reasoning_and_calls(chunks) assert FUNC_TIME in function_names, "get_time function not called" assert any("New York" in arg for arg in arguments), ( f"Expected get_time arguments for New York not found in {arguments}" ) assert len(reasoning) > 0, "Expected reasoning content missing" assert any(keyword in reasoning for keyword in ["New York", "time", "current"]), ( f"Reasoning is not relevant to the request: {reasoning}" ) @pytest.mark.asyncio async def test_streaming_multiple_tools(client: openai.AsyncOpenAI): """Test streamed multi-tool response with reasoning.""" stream = await client.chat.completions.create( model=MODEL_NAME, messages=MESSAGES_MULTIPLE_CALLS, tools=TOOLS, temperature=0.0, stream=True, ) chunks = [chunk async for chunk in stream] reasoning, arguments, function_names = extract_reasoning_and_calls(chunks) try: assert FUNC_CALC in function_names, ( f"Calculator tool missing — found {function_names}" ) assert FUNC_TIME in function_names, ( f"Time tool missing — found {function_names}" ) assert len(reasoning) > 0, "Expected reasoning content in streamed response" except AssertionError as e: print(f"ERROR: {e}") @pytest.mark.asyncio async def test_invalid_tool_call(client: openai.AsyncOpenAI): """ Verify that ambiguous instructions that should not trigger a tool do not produce any tool calls. """ response = await client.chat.completions.create( model=MODEL_NAME, messages=MESSAGES_INVALID_CALL, tools=TOOLS, temperature=0.0, stream=False, ) message = response.choices[0].message assert message is not None, "Expected message in response" assert hasattr(message, "content"), "Expected 'content' field in message" tool_calls = getattr(message, "tool_calls", []) assert not tool_calls, ( f"Model unexpectedly attempted a tool call on invalid input: {tool_calls}" ) @pytest.mark.asyncio async def test_tool_call_with_temperature(client: openai.AsyncOpenAI): """ Verify model produces valid tool or text output under non-deterministic sampling. """ response = await client.chat.completions.create( model=MODEL_NAME, messages=MESSAGES_CALC, tools=TOOLS, temperature=0.7, stream=False, ) message = response.choices[0].message assert message is not None, "Expected non-empty message in response" assert message.tool_calls or message.content, ( "Response missing both text and tool calls" ) print(f"\nTool calls: {message.tool_calls}") print(f"Text: {message.content}") @pytest.mark.asyncio async def test_tool_response_schema_accuracy(client: openai.AsyncOpenAI): """Validate that tool call arguments adhere to their declared JSON schema.""" response = await client.chat.completions.create( model=MODEL_NAME, messages=MESSAGES_MULTIPLE_CALLS, tools=TOOLS, temperature=0.0, ) calls = response.choices[0].message.tool_calls assert calls, "No tool calls produced" for call in calls: func_name = call.function.name args = json.loads(call.function.arguments) schema: dict[str, object] | None = None for tool_entry in TOOLS: function_def = tool_entry.get("function") if ( function_def and isinstance(function_def, dict) and function_def.get("name") == func_name ): schema = function_def.get("parameters") break assert schema is not None, f"No matching tool schema found for {func_name}" jsonschema.validate(instance=args, schema=schema) @pytest.mark.asyncio async def test_semantic_consistency_with_temperature(client: openai.AsyncOpenAI): """Test that temperature variation doesn't cause contradictory reasoning.""" responses = [] for temp in [0.0, 0.5, 1.0]: resp = await client.chat.completions.create( model=MODEL_NAME, messages=MESSAGES_CALC, tools=TOOLS, temperature=temp, ) text = (resp.choices[0].message.content or "").strip() responses.append(text) # Compare fuzzy similarity between low- and mid-temperature outputs low_mid_sim = fuzz.ratio(responses[0], responses[1]) assert low_mid_sim > 60, ( f"Semantic drift too large between T=0.0 and T=0.5 ({low_mid_sim}%)" )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/tool_parsers/test_openai_tool_parser.py", "license": "Apache License 2.0", "lines": 293, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/compile/test_compile_ranges.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any import torch from torch import fx as fx from torch import nn # This import automatically registers `torch.ops.silly.attention` import tests.compile.silly_attention # noqa from vllm.compilation.counter import compilation_counter from vllm.compilation.decorators import support_torch_compile from vllm.compilation.passes.inductor_pass import ( InductorPass, get_pass_context, ) from vllm.config import ( VllmConfig, set_current_vllm_config, ) from vllm.config.compilation import CompilationConfig, CompilationMode from vllm.config.scheduler import SchedulerConfig from vllm.config.utils import Range from vllm.forward_context import set_forward_context BATCH_SIZE = 64 MLP_SIZE = 128 @support_torch_compile class TestModel(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = "", **kwargs) -> None: super().__init__() def forward(self, x: torch.Tensor) -> torch.Tensor: x = x + x attn_output = torch.empty_like(x) torch.ops.silly.attention(x, x, x, attn_output) x = attn_output x = x * 3 return x @torch.inference_mode def run_model(vllm_config: VllmConfig, model: nn.Module, batch_sizes: list[int]): with set_forward_context({}, vllm_config=vllm_config): model(torch.randn(BATCH_SIZE, MLP_SIZE)) for batch_size in batch_sizes: model(torch.randn(batch_size, MLP_SIZE)) class PostGradRangeChecker(InductorPass): def __init__(self, ranges: list[Range]): self.ranges = ranges self.num_calls = 0 def __call__(self, graph: fx.Graph): compile_range = get_pass_context().compile_range assert compile_range in self.ranges, ( f"Compile range {compile_range} not in {self.ranges}" ) self.num_calls += 1 def uuid(self) -> str: state: dict[str, Any] = {} return InductorPass.hash_dict(state) def test_compile_ranges(use_fresh_inductor_cache): post_grad_range_checker = PostGradRangeChecker( [ Range(start=1, end=8), Range(start=16, end=16), Range(start=9, end=32), Range(start=64, end=64), Range(start=33, end=8192), ] ) torch.set_default_device("cuda") vllm_config = VllmConfig( scheduler_config=SchedulerConfig( max_num_batched_tokens=8192, max_model_len=8192, is_encoder_decoder=False, ), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, compile_ranges_split_points=[8, 32], compile_sizes=[16, 64, 128], inductor_compile_config={ "post_grad_custom_post_pass": post_grad_range_checker, }, ), ) with set_current_vllm_config(vllm_config): model = TestModel(vllm_config=vllm_config, prefix="").eval() # Number of compilations: 3 for each compile range + 2 compile sizes batch_sizes = [1, 4, 16, 24, 48, 64, 8192] with compilation_counter.expect( num_graphs_seen=1, num_piecewise_graphs_seen=1, num_backend_compilations=5, ): run_model(vllm_config, model, batch_sizes) assert post_grad_range_checker.num_calls == 5 def test_compile_config_get_compile_ranges(): compilation_config = CompilationConfig( compile_ranges_split_points=[8, 32], ) VllmConfig( scheduler_config=SchedulerConfig( max_num_batched_tokens=8192, max_model_len=8192, is_encoder_decoder=False, ), compilation_config=compilation_config, ) assert compilation_config.get_compile_ranges() == [ Range(start=1, end=8), Range(start=9, end=32), Range(start=33, end=8192), ] def test_inductor_cache_compile_ranges(monkeypatch, use_fresh_inductor_cache): # To force multiple compilations, we disable the compile cache monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", "1") post_grad_range_checker = PostGradRangeChecker( ranges=[ Range(start=1, end=8), Range(start=9, end=8192), ] ) scheduler_config = SchedulerConfig( max_num_batched_tokens=8192, max_model_len=8192, is_encoder_decoder=False, ) torch.set_default_device("cuda") def create_vllm_config(): return VllmConfig( scheduler_config=scheduler_config, compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, compile_ranges_split_points=[8], inductor_compile_config={ "post_grad_custom_post_pass": post_grad_range_checker, }, ), ) vllm_config_1 = create_vllm_config() with set_current_vllm_config(vllm_config_1): model1 = TestModel(vllm_config=vllm_config_1, prefix="").eval() batch_sizes = [1, 16] run_model(vllm_config_1, model1, batch_sizes) assert post_grad_range_checker.num_calls == 2 post_grad_range_checker.num_calls = 0 # Create a new vllm config with the new pass context vllm_config_2 = create_vllm_config() with set_current_vllm_config(vllm_config_2): model2 = TestModel(vllm_config=vllm_config_2, prefix="").eval() batch_sizes = [4, 32] run_model(vllm_config_2, model2, batch_sizes) # Check that cache is used, so the number of calls # should be 0 assert post_grad_range_checker.num_calls == 0
{ "repo_id": "vllm-project/vllm", "file_path": "tests/compile/test_compile_ranges.py", "license": "Apache License 2.0", "lines": 150, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/config/attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any, Literal from pydantic import field_validator from vllm.config.utils import config from vllm.v1.attention.backends.registry import AttentionBackendEnum @config class AttentionConfig: """Configuration for attention mechanisms in vLLM.""" backend: AttentionBackendEnum | None = None """Attention backend to use. If None, will be selected automatically.""" flash_attn_version: Literal[2, 3, 4] | None = None """Force vllm to use a specific flash-attention version (2, 3, or 4). Only valid when using the flash-attention backend.""" use_prefill_decode_attention: bool = False """Use separate prefill and decode kernels for attention instead of the unified triton kernel.""" flash_attn_max_num_splits_for_cuda_graph: int = 32 """Flash Attention max number splits for cuda graph decode.""" use_cudnn_prefill: bool = False """Whether to use cudnn prefill.""" use_trtllm_ragged_deepseek_prefill: bool = True """Whether to use TRTLLM ragged deepseek prefill.""" use_trtllm_attention: bool | None = None """If set to True/False, use or don't use the TRTLLM attention backend in flashinfer. If None, auto-detect the attention backend in flashinfer.""" disable_flashinfer_prefill: bool = False """Whether to disable flashinfer prefill.""" disable_flashinfer_q_quantization: bool = False """If set, when using fp8 kv, do not quantize Q to fp8.""" use_prefill_query_quantization: bool = False """If set, quantize query for attention in prefill.""" def compute_hash(self) -> str: """ Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ from vllm.config.utils import get_hash_factors, hash_factors ignored_factors: list[str] = [] factors = get_hash_factors(self, ignored_factors) return hash_factors(factors) @field_validator("backend", mode="before") @classmethod def validate_backend_before(cls, value: Any) -> Any: """Enable parsing of the `backend` enum type from string.""" if isinstance(value, str): return AttentionBackendEnum[value.upper()] return value
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/config/attention.py", "license": "Apache License 2.0", "lines": 51, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/engine/test_abort_final_step.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Test for the fix in PR #29987: Eagerly abort cancelled final-step requests. This test verifies that when a request is aborted during its final execution step (when it would naturally complete), it is properly marked as aborted rather than being treated as normally completed. The test uses a dummy KV connector to verify that the connector receives the correct finish status (FINISHED_ABORTED, not FINISHED_LENGTH_CAPPED). """ import asyncio import tempfile import time from pathlib import Path from typing import Any from unittest.mock import patch import pytest from vllm import SamplingParams from vllm.config import KVTransferConfig, VllmConfig from vllm.distributed.kv_transfer.kv_connector.factory import KVConnectorFactory from vllm.distributed.kv_transfer.kv_connector.v1.base import ( KVConnectorBase_V1, KVConnectorMetadata, KVConnectorRole, ) from vllm.engine.arg_utils import AsyncEngineArgs from vllm.platforms import current_platform from vllm.sampling_params import RequestOutputKind from vllm.utils.torch_utils import set_default_torch_num_threads from vllm.v1.core.sched.output import SchedulerOutput from vllm.v1.engine.async_llm import AsyncLLM from vllm.v1.kv_cache_interface import KVCacheConfig from vllm.v1.request import Request if not current_platform.is_cuda(): pytest.skip(reason="V1 currently only supported on CUDA.", allow_module_level=True) TEXT_PROMPT = "Hello" class DummyKVConnectorMetadata(KVConnectorMetadata): """Dummy metadata for the test connector.""" def __init__(self): self.requests: list = [] class DummyKVConnector(KVConnectorBase_V1): """ Dummy KV connector that captures request finish statuses to a file. This is used to verify the fix - without the fix, a request aborted during its final step would be captured as FINISHED_LENGTH_CAPPED instead of FINISHED_ABORTED. The connector runs in a separate process, so we write statuses to a file that can be read by the test process. """ def __init__( self, vllm_config: VllmConfig, role: KVConnectorRole, kv_cache_config: KVCacheConfig | None = None, ): super().__init__(vllm_config, role, kv_cache_config) # Get the status file path from extra config extra_config = vllm_config.kv_transfer_config.kv_connector_extra_config or {} self.status_file = extra_config.get("status_file") # Log that we were initialized if self.status_file: try: with open(self.status_file, "a") as f: f.write(f"INIT:{role.name}\n") except Exception: pass def get_num_new_matched_tokens( self, request: Request, num_computed_tokens: int, ) -> tuple[int | None, bool]: return (0, False) def update_state_after_alloc( self, request: Request, blocks: Any, num_external_tokens: int, ): pass def build_connector_meta( self, scheduler_output: SchedulerOutput ) -> KVConnectorMetadata: return DummyKVConnectorMetadata() def request_finished( self, request: Request, block_ids: list[int], ) -> tuple[bool, dict[str, Any] | None]: """Capture the request status when finished by writing to a file.""" if self.status_file: try: with open(self.status_file, "a") as f: # Write the status name (e.g., "FINISHED_ABORTED") f.write(f"{request.status.name}\n") except Exception as e: # Log but don't fail - this is just test instrumentation print(f"[DummyKVConnector] Failed to write status: {e}") return False, None def start_load_kv(self, forward_context: Any, **kwargs: Any) -> None: pass def wait_for_layer_load(self, layer_name: str) -> None: pass def save_kv_layer( self, layer_name: str, kv_layer: Any, attn_metadata: Any, **kwargs: Any, ) -> None: pass def wait_for_save(self): pass # Register the dummy connector KVConnectorFactory.register_connector( "DummyKVConnector", __name__, DummyKVConnector.__name__ ) @pytest.mark.parametrize("async_scheduling", [False, True]) @pytest.mark.asyncio async def test_abort_during_final_step(async_scheduling: bool): """ Test that a request aborted during its final execution step is treated as aborted rather than completed. This test: 1. Monkeypatches execute_model to wait for a file to be deleted 2. Configures a dummy KV connector to capture finish statuses 3. Starts a request with max_tokens=1 (will complete on first decode step) 4. Aborts the request, then deletes the file to unblock execute_model 5. Verifies the KV connector received FINISHED_ABORTED not FINISHED_LENGTH_CAPPED See https://github.com/vllm-project/vllm/pull/29987. Without the fix, the KV connector would see FINISHED_LENGTH_CAPPED because update_from_output() would mark the request as completed before processing the abort. This causes KV cache blocks to not be freed properly in disaggregated prefill scenarios. With the fix, _process_aborts_queue() runs before update_from_output(), so the abort takes precedence and the KV connector sees FINISHED_ABORTED. """ # Create three temporary files: # 1. ready_file: deleted by execute_model to signal it has started # 2. block_file: execute_model waits for this to be deleted # 3. status_file: KV connector writes finish statuses here with tempfile.NamedTemporaryFile(delete=False) as f: ready_file = Path(f.name) with tempfile.NamedTemporaryFile(delete=False) as f2: block_file = Path(f2.name) with tempfile.NamedTemporaryFile(delete=False, mode="w") as f3: status_file = Path(f3.name) try: # Get the original execute_model method from vllm.v1.worker.gpu_worker import Worker original_execute_model = Worker.execute_model def execute_model_with_wait(self, scheduler_output): # Signal that execute_model has been called by deleting ready_file if ready_file.exists(): ready_file.unlink() # Wait for the block file to be deleted (triggered from test after abort) # This runs in the worker process (after fork), so we poll the filesystem while block_file.exists(): time.sleep(0.01) return original_execute_model(self, scheduler_output) # Patch execute_model to inject the wait # This happens before the worker process is forked, so the patch applies there with patch.object(Worker, "execute_model", execute_model_with_wait): request_id = "test-abort-final-step" # Configure engine with dummy KV connector # Pass the status file path so the connector can write to it kv_transfer_config = KVTransferConfig( kv_connector="DummyKVConnector", kv_role="kv_both", kv_connector_extra_config={"status_file": str(status_file)}, ) engine_args = AsyncEngineArgs( model="meta-llama/Llama-3.2-1B-Instruct", enforce_eager=True, async_scheduling=async_scheduling, kv_transfer_config=kv_transfer_config, ) with set_default_torch_num_threads(1): engine = AsyncLLM.from_engine_args(engine_args) try: # Create a request that will complete after just 1 token sampling_params = SamplingParams( max_tokens=1, ignore_eos=True, output_kind=RequestOutputKind.DELTA, ) # Start generation in a task outputs = [] async def generate(): async for output in engine.generate( request_id=request_id, prompt=TEXT_PROMPT, sampling_params=sampling_params, ): outputs.append(output) gen_task = asyncio.create_task(generate()) # Wait for execute_model to signal it has started (with timeout) timeout = 5.0 # 5 second timeout start_time = time.time() while ready_file.exists(): if time.time() - start_time > timeout: raise TimeoutError( "Timeout waiting for execute_model to start. " "The monkeypatch may not be working correctly, " "for example if spawn was used instead of fork." ) await asyncio.sleep(0.01) # Abort the request while execute_model is blocked await engine.abort(request_id) # Now unblock execute_model by deleting the file # The abort should be processed before the model output block_file.unlink() # Wait for generation to complete await gen_task # Give the scheduler a moment to finish cleanup await asyncio.sleep(0.1) # Verify we got output assert len(outputs) > 0, "Should have received at least one output" # The final output should have finish_reason="abort" final_output = outputs[-1] assert final_output.finished, ( "Final output should be marked as finished" ) assert final_output.outputs[0].finish_reason == "abort", ( f"Expected finish_reason='abort' but got " f"'{final_output.outputs[0].finish_reason}'. " ) with open(status_file) as f4: status_lines = f4.read().strip().split("\n") # Filter for actual finish statuses (not INIT or empty lines) captured_statuses = [ line for line in status_lines if line and line.startswith("FINISHED_") ] assert len(captured_statuses) >= 1, ( f"Expected at least 1 captured finish status, got " f"{len(captured_statuses)}. File content: {status_lines}" ) assert "FINISHED_ABORTED" in captured_statuses, ( f"KV connector should see FINISHED_ABORTED but got " f"{captured_statuses}. " ) # Verify cleanup assert not engine.output_processor.has_unfinished_requests() finally: # Shutdown the engine engine.shutdown() finally: # Clean up temporary files if they still exist if ready_file.exists(): ready_file.unlink() if block_file.exists(): block_file.unlink() if status_file.exists(): status_file.unlink()
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/engine/test_abort_final_step.py", "license": "Apache License 2.0", "lines": 255, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/utils/nvtx_pytorch_hooks.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from contextlib import contextmanager import torch import torch.cuda.nvtx as nvtx def print_tensor(tensor_obj, prefix, tensor_list=None): """Descends iterators that contains Tensors and prints the Tensor. Recursive function that descends iterator type arguments until it finds a Tensor object. """ if tensor_list is None: tensor_list = [] if isinstance(tensor_obj, (list, tuple)): for ten in tensor_obj: tensor_list = print_tensor(ten, prefix, tensor_list) elif isinstance(tensor_obj, torch.Tensor): tensor_dims = list(tensor_obj.size()) tensor_list.append(tensor_dims) return tensor_list def process_layer_params(module_obj): """Extract the static parameters from LLM and VLM relevant layer types""" param_info = {} # Extract parameters for layers commonly used in LLMs and VLMs if isinstance(module_obj, (torch.nn.Conv1d, torch.nn.Conv2d, torch.nn.Conv3d)): conv_params = {} conv_params["in_chan"] = module_obj.in_channels conv_params["out_chan"] = module_obj.out_channels conv_params["filter_dim"] = module_obj.kernel_size conv_params["stride"] = module_obj.stride conv_params["padding"] = module_obj.padding conv_params["dilation"] = module_obj.dilation conv_params["transposed"] = module_obj.transposed conv_params["output_padding"] = module_obj.output_padding conv_params["groups"] = module_obj.groups conv_params["padding_mode"] = module_obj.padding_mode param_info = conv_params elif isinstance( module_obj, ( torch.nn.ConvTranspose1d, torch.nn.ConvTranspose2d, torch.nn.ConvTranspose3d, ), ): convtranspose_params = {} convtranspose_params["in_chan"] = module_obj.in_channels convtranspose_params["out_chan"] = module_obj.out_channels convtranspose_params["filter_dim"] = module_obj.kernel_size convtranspose_params["stride"] = module_obj.stride convtranspose_params["padding"] = module_obj.padding convtranspose_params["dilation"] = module_obj.dilation convtranspose_params["transposed"] = module_obj.transposed convtranspose_params["output_padding"] = module_obj.output_padding convtranspose_params["groups"] = module_obj.groups convtranspose_params["padding_mode"] = module_obj.padding_mode param_info = convtranspose_params elif isinstance( module_obj, (torch.nn.MaxPool1d, torch.nn.MaxPool2d, torch.nn.MaxPool3d) ): def _handle_int_or_tuple(parameter): if isinstance(parameter, tuple): return list(parameter) elif isinstance(parameter, int): return [parameter, parameter] pooling_params = {} pooling_params["filter_dim"] = _handle_int_or_tuple(module_obj.kernel_size) pooling_params["stride"] = _handle_int_or_tuple(module_obj.stride) pooling_params["padding"] = _handle_int_or_tuple(module_obj.padding) pooling_params["dilation"] = _handle_int_or_tuple(module_obj.dilation) param_info = pooling_params elif isinstance( module_obj, (torch.nn.AvgPool1d, torch.nn.AvgPool2d, torch.nn.AvgPool3d) ): pooling_params = {} pooling_params["filter_dim"] = [ module_obj.kernel_size, module_obj.kernel_size, ] pooling_params["stride"] = [module_obj.stride, module_obj.stride] pooling_params["padding"] = [module_obj.padding, module_obj.padding] pooling_params["ceil_mode"] = module_obj.ceil_mode pooling_params["count_include_pad"] = module_obj.count_include_pad param_info = pooling_params elif isinstance( module_obj, ( torch.nn.AdaptiveAvgPool1d, torch.nn.AdaptiveAvgPool2d, torch.nn.AdaptiveAvgPool3d, ), ): pooling_params = {} pooling_params["output_size"] = [ module_obj.output_size, module_obj.output_size, ] param_info = pooling_params elif isinstance(module_obj, torch.nn.Linear): param_info["in_features"] = module_obj.in_features param_info["out_features"] = module_obj.out_features elif isinstance( module_obj, (torch.nn.BatchNorm1d, torch.nn.BatchNorm2d, torch.nn.BatchNorm3d), ): param_info["num_features"] = module_obj.num_features param_info["epsilon"] = module_obj.eps param_info["momentum"] = module_obj.momentum elif isinstance(module_obj, torch.nn.ReLU): param_info["in_place"] = module_obj.inplace elif isinstance(module_obj, torch.nn.Dropout): param_info["p"] = module_obj.p param_info["in_place"] = module_obj.inplace elif isinstance(module_obj, torch.nn.Embedding): param_info["num_embeddings"] = module_obj.num_embeddings param_info["embedding_dim"] = module_obj.embedding_dim elif isinstance( module_obj, ( torch.nn.Upsample, torch.nn.UpsamplingNearest2d, torch.nn.UpsamplingBilinear2d, ), ): param_info["scale_factor"] = module_obj.scale_factor return param_info def construct_marker_dict_and_push( module_name, module_obj, in_tensor, kwargs=None, out_tensor=None ): marker_dict = {} marker_dict["Module"] = module_name ## Get trainable parameters like weights and bias module_params = module_obj.named_parameters(recurse=False) for idx, (param_name, param_obj) in enumerate(module_params): if idx == 0: marker_dict["TrainableParams"] = {} marker_dict["TrainableParams"][param_name] = list(param_obj.size()) in_tensor_list = print_tensor(in_tensor, "Input") if in_tensor_list: marker_dict["Inputs"] = in_tensor_list out_tensor_list = print_tensor(out_tensor, "Output") if out_tensor_list: marker_dict["Outputs"] = out_tensor_list ## Get Kwargs like input_ids and positions for the top module if kwargs: for key, value in kwargs.items(): if isinstance(value, (torch.Tensor, list, tuple)): tensor_list = print_tensor(value, key) if tensor_list: marker_dict[key] = tensor_list param_info = process_layer_params(module_obj) if param_info: marker_dict["StaticParams"] = param_info nvtx.range_push("{}".format(marker_dict)) class ResultHolder: """Holder for storing results from within a context manager.""" result = None @contextmanager def layerwise_nvtx_marker_context(module_name, module_obj, in_tensor=None, kwargs=None): """Context manager for NVTX markers that automatically pushes on enter and pops on exit. Example: with nvtx_marker_context("Module:MyModule", module, in_tensor=args, kwargs=kwargs) as ctx: ctx.result = module(*args, **kwargs) return ctx.result """ holder = ResultHolder() # Push input marker construct_marker_dict_and_push( module_name, module_obj, in_tensor=in_tensor, kwargs=kwargs, ) try: yield holder finally: # Pop input marker nvtx.range_pop() # Push and pop output marker output_name = module_name.replace("(input)", "(output)") construct_marker_dict_and_push( output_name, module_obj, in_tensor=None, kwargs=None, out_tensor=holder.result, ) nvtx.range_pop() class PytHooks: """This module contains all the code needed to enable forward hooks in a pytorch network. To register the hooks for a given network, the user needs to instantiate a PytHook object. Then call the register_hooks method. Example: my_hook = PytHook() my_hook.register_hooks(my_network_model) """ def __init__(self): """Initialize module variables.""" super().__init__() self.module_to_name_map = {} def _process_layer_params(self, module_obj): return process_layer_params(module_obj) def module_fwd_hook(self, module_obj, in_tensor, out_tensor): """Callback function that ends the NVTX marker. Records the module name and tensor information. Called after the module executes the forward method. """ nvtx.range_pop() module_name = self.module_to_name_map.get(module_obj, "unknown") construct_marker_dict_and_push( module_name, module_obj, in_tensor=None, kwargs=None, out_tensor=out_tensor ) nvtx.range_pop() return def module_fwd_pre_hook(self, module_obj, in_tensor, kwargs): """Creates an NVTX marker with the module name in it. This function is called before the module executes. """ module_name = self.module_to_name_map.get(module_obj, "unknown") construct_marker_dict_and_push( module_name, module_obj, in_tensor=in_tensor, kwargs=kwargs, out_tensor=None ) return def register_hooks(self, network_model, module_prefix="top"): """User level function that activates all the hooks. The user needs to call this method from the network source code. The code descends all the modules in the network and registers their respective hooks. """ # Module types to skip (simple operations that don't need detailed profiling) skip_types = ( torch.nn.Identity, torch.nn.Dropout, torch.nn.Dropout1d, torch.nn.Dropout2d, torch.nn.Dropout3d, ) for name, module in network_model.named_modules(prefix=module_prefix): # Skip certain module types to reduce profiling overhead if isinstance(module, skip_types): continue module.register_forward_pre_hook(self.module_fwd_pre_hook, with_kwargs=True) module.register_forward_hook(self.module_fwd_hook) if module not in self.module_to_name_map: self.module_to_name_map[module] = name else: raise ValueError("Module instance {} is not unique ".format(module)) return
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/utils/nvtx_pytorch_hooks.py", "license": "Apache License 2.0", "lines": 250, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/reasoning/test_holo2_reasoning_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from transformers import AutoTokenizer from tests.reasoning.utils import run_reasoning_extraction from vllm.reasoning import ReasoningParser, ReasoningParserManager from vllm.reasoning.deepseek_r1_reasoning_parser import DeepSeekR1ReasoningParser from vllm.reasoning.deepseek_v3_reasoning_parser import ( DeepSeekV3ReasoningWithThinkingParser as Holo2ReasoningParser, ) from vllm.reasoning.identity_reasoning_parser import IdentityReasoningParser REASONING_MODEL_NAME = "HCompany/Holo2-4B" @pytest.fixture(scope="module") def tokenizer(): return AutoTokenizer.from_pretrained(REASONING_MODEL_NAME) @pytest.mark.parametrize( "thinking,expected_parser_type", [ (True, DeepSeekR1ReasoningParser), (False, IdentityReasoningParser), ], ) def test_parser_selection(tokenizer, thinking, expected_parser_type): parser = Holo2ReasoningParser( tokenizer, chat_template_kwargs={ "thinking": thinking, }, ) assert isinstance(parser._parser, expected_parser_type) def test_holo2_default_parser_is_deepseekr1(tokenizer): parser = Holo2ReasoningParser(tokenizer) assert isinstance(parser._parser, DeepSeekR1ReasoningParser) def test_holo2_supports_structured_output(tokenizer): # Structured output manager uses the reasoning parser to check if the # reasoning content is ended before applying the grammar. The main function # used is is_reasoning_end. This test checks if the parser is able to # correctly identify the end of the reasoning content. # important to not pass chat_template_kwargs here as it is done in the # StructuredOutputManager parser = Holo2ReasoningParser(tokenizer) end_token_id = tokenizer.encode("</think>", add_special_tokens=False)[0] assert parser.is_reasoning_end([1, 2, 4, end_token_id]) assert not parser.is_reasoning_end([1, 2, 4]) assert parser.is_reasoning_end([1, 2, 4, end_token_id, 5]) # thinking is True, non-streaming WITH_THINK = { "output": "This is a reasoning section</think>This is the rest", "reasoning": "This is a reasoning section", "content": "This is the rest", } # thinking is True, streaming WITH_THINK_STREAM = { "output": "This is a reasoning section</think>This is the rest", "reasoning": "This is a reasoning section", "content": "This is the rest", } # thinking is False, non-streaming THINKING_DISABLED = { "output": "This is the rest", "reasoning": None, "content": "This is the rest", } # thinking is False, streaming THINKING_DISABLED_STREAM = { "output": "This is the rest", "reasoning": None, "content": "This is the rest", } # thinking is False but the model output </think>, non-streaming THINKING_DISABLED_WITH_CLOSE_TAG = { "output": "</think>This is the rest", "reasoning": None, "content": "</think>This is the rest", } # thinking is False but the model output </think>, streaming THINKING_DISABLED_WITH_CLOSE_TAG_STREAM = { "output": "some text</think>This is the rest", "reasoning": None, "content": "some text</think>This is the rest", } COMPLETE_REASONING = { "output": "This is a reasoning section</think>", "reasoning": "This is a reasoning section", "content": None, } TEST_CASES = [ pytest.param( False, WITH_THINK, None, id="with_think", ), pytest.param( True, WITH_THINK_STREAM, None, id="with_think_stream", ), pytest.param( False, WITH_THINK, {"thinking": True}, id="with_think_enabled", ), pytest.param( True, WITH_THINK_STREAM, {"thinking": True}, id="with_think_stream_enabled", ), pytest.param( False, THINKING_DISABLED, {"thinking": False}, id="thinking_disabled", ), pytest.param( True, THINKING_DISABLED_STREAM, {"thinking": False}, id="thinking_disabled_stream", ), pytest.param( False, THINKING_DISABLED_WITH_CLOSE_TAG, {"thinking": False}, id="thinking_disabled_with_close_tag", ), pytest.param( True, THINKING_DISABLED_WITH_CLOSE_TAG_STREAM, {"thinking": False}, id="thinking_disabled_with_close_tag_stream", ), pytest.param( False, COMPLETE_REASONING, None, id="complete_reasoning", ), pytest.param( True, COMPLETE_REASONING, None, id="complete_reasoning_stream", ), ] @pytest.mark.parametrize("streaming, param_dict, chat_template_kwargs", TEST_CASES) def test_reasoning( streaming: bool, param_dict: dict, chat_template_kwargs: dict | None, tokenizer, ): output = tokenizer.tokenize(param_dict["output"]) output_tokens: list[str] = [ tokenizer.convert_tokens_to_string([token]) for token in output ] parser: ReasoningParser = ReasoningParserManager.get_reasoning_parser("holo2")( tokenizer, chat_template_kwargs=chat_template_kwargs, ) reasoning, content = run_reasoning_extraction( parser, output_tokens, streaming=streaming ) assert reasoning == param_dict["reasoning"] assert content == param_dict["content"]
{ "repo_id": "vllm-project/vllm", "file_path": "tests/reasoning/test_holo2_reasoning_parser.py", "license": "Apache License 2.0", "lines": 168, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/distributed/eplb/policy/abstract.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import ABC, abstractmethod import torch class AbstractEplbPolicy(ABC): @classmethod @abstractmethod def rebalance_experts( cls, weight: torch.Tensor, num_replicas: int, num_groups: int, num_nodes: int, num_ranks: int, old_global_expert_indices: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Entry point for expert-parallelism load balancer. Parameters: weight: [layers, num_logical_experts], the load statistics for all logical experts num_replicas: number of physical experts, must be a multiple of `num_ranks` num_groups: number of expert groups num_nodes: number of server nodes num_ranks: number of ranks, must be a multiple of `num_nodes` old_global_expert_indices: [layers, num_logical_experts], the old global expert indices. Used to avoid unnecessary weight copying for experts moving within one rank. Returns: physical_to_logical_map: [layers, num_replicas], the expert index of each replica logical_to_physical_map: [layers, num_logical_experts, X], the replica indices for each expert expert_count: [layers, num_logical_experts], number of physical replicas for each logical expert """ raise NotImplementedError
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/distributed/eplb/policy/abstract.py", "license": "Apache License 2.0", "lines": 38, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/distributed/eplb/policy/default.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Expert parallelism load balancer (EPLB) for vLLM. This module implements the core rearrangement algorithm. The rearrangement algorithm is adapted from [DeepSeek EPLB](https://github.com/deepseek-ai/eplb). Please find at [#12](https://github.com/deepseek-ai/EPLB/issues/12) an example on how the EPLB algorithm works. """ import numpy as np import torch from .abstract import AbstractEplbPolicy class DefaultEplbPolicy(AbstractEplbPolicy): @classmethod def balanced_packing( cls, weight: np.ndarray, num_packs: int ) -> tuple[np.ndarray, np.ndarray]: """ Pack n weighted objects to m packs, such that each bin contains exactly n/m objects and the weights of all packs are as balanced as possible. Parameters: weight: [X, n], the weight of each item num_packs: number of packs Returns: pack_index: [X, n], the pack index of each item rank_in_pack: [X, n], the rank of the item in the pack """ num_layers, num_groups = weight.shape assert num_groups % num_packs == 0 groups_per_pack = num_groups // num_packs if groups_per_pack == 1: pack_index = np.tile(np.arange(num_groups, dtype=np.int64), (num_layers, 1)) rank_in_pack = np.zeros_like(pack_index, dtype=np.int64) return pack_index, rank_in_pack # Sort and get indices in decending order indices = np.argsort(-weight, axis=-1) pack_index = np.full((num_layers, num_groups), -1, dtype=np.int64) rank_in_pack = np.full((num_layers, num_groups), -1, dtype=np.int64) pack_weights = np.zeros((num_layers, num_packs), dtype=np.float64) pack_items = np.zeros((num_layers, num_packs), dtype=np.int64) # Run the packing algorithm for layer_idx in range(num_layers): weights_row = pack_weights[layer_idx] items_row = pack_items[layer_idx] for group in indices[layer_idx]: # Pick the lightest pack; full packs are masked out by inf. pack = int(np.argmin(weights_row)) pack_index[layer_idx, group] = pack rank_in_pack[layer_idx, group] = items_row[pack] weights_row[pack] += weight[layer_idx, group] items_row[pack] += 1 if items_row[pack] == groups_per_pack: # Mark as unavailable for future selections. weights_row[pack] = np.inf return pack_index, rank_in_pack @classmethod def replicate_experts( cls, weight: np.ndarray, num_phy: int ) -> tuple[np.ndarray, np.ndarray, np.ndarray]: """ Replicate `num_log` experts to `num_phy` replicas, such that the maximum load of all replicas is minimized. Parameters: weight: [X, num_log] num_phy: total number of experts after replication Returns: phy2log: [X, num_phy], logical expert id of each physical expert replica_idx: [X, num_phy], the index of the replica for each logical expert logcnt: [X, num_log], number of replicas for each logical expert """ n, num_log = weight.shape num_redundant = num_phy - num_log assert num_redundant >= 0 phy2log = np.tile(np.arange(num_phy, dtype=np.int64), (n, 1)) replica_idx = np.zeros((n, num_phy), dtype=np.int64) logcnt = np.ones((n, num_log), dtype=np.int64) arangen = np.arange(n, dtype=np.int64) for i in range(num_log, num_phy): redundant_indices = np.argmax(weight / logcnt, axis=-1) phy2log[:, i] = redundant_indices replica_idx[:, i] = logcnt[arangen, redundant_indices] logcnt[arangen, redundant_indices] += 1 return phy2log, replica_idx, logcnt @classmethod def rebalance_experts_hierarchical( cls, weight: np.ndarray, num_physical_experts: int, num_groups: int, num_nodes: int, num_gpus: int, ) -> tuple[np.ndarray, np.ndarray, np.ndarray]: """ Parameters: weight: [num_moe_layers, num_logical_experts] num_physical_experts: number of physical experts after replication num_groups: number of expert groups num_nodes: number of server nodes, where the intra-node network (e.g, NVLink) is faster num_gpus: number of GPUs, must be a multiple of `num_nodes` Returns: phy2log: [layers, num_replicas], the expert index of each replica pphy_replicas_idx: [layers, num_logical_experts, X], the replica indices for each expert logcnt: [layers, num_logical_experts], number of physical replicas for each logical expert """ num_layers, num_logical_experts = weight.shape assert num_logical_experts % num_groups == 0 group_size = num_logical_experts // num_groups assert num_groups % num_nodes == 0 groups_per_node = num_groups // num_nodes assert num_gpus % num_nodes == 0 assert num_physical_experts % num_gpus == 0 phy_experts_per_gpu = num_physical_experts // num_gpus def inverse(perm: np.ndarray) -> np.ndarray: inv = np.empty_like(perm) row_idx = np.arange(perm.shape[0])[:, None] col_idx = np.arange(perm.shape[1], dtype=np.int64) inv[row_idx, perm] = col_idx return inv # Step 1: pack groups to nodes tokens_per_group = weight.reshape(num_layers, num_groups, group_size).sum( axis=-1 ) group_pack_index, group_rank_in_pack = cls.balanced_packing( tokens_per_group, num_nodes ) # Map each logical expert into a node-local ordering based on packed groups. log2mlog = ( ( (group_pack_index * groups_per_node + group_rank_in_pack)[..., None] * group_size ) + np.arange(group_size, dtype=np.int64) ).reshape(num_layers, num_logical_experts) mlog2log = inverse(log2mlog) # Step 2: construct redundant experts within nodes # Reorder weights into the node-local layout so replication is done per node. tokens_per_mlog = np.take_along_axis(weight, mlog2log, axis=1).reshape( -1, num_logical_experts // num_nodes ) phy2mlog, replicas_idx, mlogcnt = cls.replicate_experts( tokens_per_mlog, num_physical_experts // num_nodes ) # Step 3: pack physical_experts to GPUs # Effective per-physical load = logical load divided by replica count. tokens_per_phy = np.take_along_axis(tokens_per_mlog / mlogcnt, phy2mlog, axis=1) pack_index, rank_in_pack = cls.balanced_packing( tokens_per_phy, num_gpus // num_nodes ) phy2pphy = pack_index * phy_experts_per_gpu + rank_in_pack pphy2phy = inverse(phy2pphy) # Reorder node-local logical indices into the post-packing physical order. pphy2mlog = np.take_along_axis(phy2mlog, pphy2phy, axis=1) pphy2mlog = ( pphy2mlog.reshape(num_layers, num_nodes, -1) + np.arange( 0, num_logical_experts, num_logical_experts // num_nodes, dtype=np.int64, )[None, :, None] ).reshape(num_layers, -1) # Map node-local logical indices back to global logical expert ids. pphy2log = np.take_along_axis(mlog2log, pphy2mlog, axis=1) # Reorder replica ranks to the post-packing physical ordering. pphy_replicas_idx = np.take_along_axis(replicas_idx, pphy2phy, axis=1).reshape( num_layers, -1 ) # Convert replica counts back to the original logical ordering. logcnt = np.take_along_axis(mlogcnt.reshape(num_layers, -1), log2mlog, axis=1) return pphy2log, pphy_replicas_idx, logcnt @classmethod def preserve_intragpu_slots( cls, phy2log: np.ndarray, phy_replicas_idx: np.ndarray, num_ranks: int, old_phy2log: np.ndarray, ) -> tuple[np.ndarray, np.ndarray]: """ Reorder the new mapping per GPU so that experts that remain on the same GPU keep their previous slot positions when possible. Incoming experts to that GPU fill any remaining available slots. This is applied only when the number of GPUs is unchanged and the slots per GPU remain the same between the old and new mappings. """ num_phy_experts = phy2log.shape[1] if num_ranks <= 0 or num_phy_experts % num_ranks != 0: return phy2log, phy_replicas_idx # Move to CPU and convert to NumPy for processing slots_per_gpu = num_phy_experts // num_ranks num_layers = phy2log.shape[0] post_phy2log = phy2log.copy() post_phy_replicas_idx = phy_replicas_idx.copy() for gpu_idx in range(num_ranks): start = gpu_idx * slots_per_gpu end = start + slots_per_gpu # Experts across all layers for this GPU old_local = old_phy2log[:, start:end] # [layers, slots] new_local = phy2log[:, start:end] # [layers, slots] new_ridx = phy_replicas_idx[:, start:end] # [layers, slots] used_new_indices = np.zeros((num_layers, slots_per_gpu), dtype=bool) preserved_positions = np.zeros((num_layers, slots_per_gpu), dtype=bool) # First pass: preserve same-logical experts in their previous slots for slot_idx in range(slots_per_gpu): # matches: [layers, slots], True where new local experts have # the same logical value as the old from 'slot_idx' and not checked yet matches = (new_local == old_local[:, slot_idx][:, None]) & ( ~used_new_indices ) has_any = matches.any(axis=1) if np.any(has_any): first_idx = np.argmax(matches, axis=1) layer_indices = np.nonzero(has_any)[0] matched_new_positions = first_idx[layer_indices] post_phy2log[layer_indices, start + slot_idx] = new_local[ layer_indices, matched_new_positions ] post_phy_replicas_idx[layer_indices, start + slot_idx] = new_ridx[ layer_indices, matched_new_positions ] used_new_indices[layer_indices, matched_new_positions] = True preserved_positions[layer_indices, slot_idx] = True # Second pass: fill remaining slots with remaining new experts remaining_mask = ~used_new_indices # [layers, slots] fill_mask = ~preserved_positions # [layers, slots] if remaining_mask.any() and fill_mask.any(): idx_base = np.tile(np.arange(slots_per_gpu), (num_layers, 1)) # Sentinel value for unavailable positions. large = slots_per_gpu + 1 # Priorities: keep original index for available spots, set sentinel # for unavailable; lower is earlier. remaining_priority = np.where(remaining_mask, idx_base, large) fill_priority = np.where(fill_mask, idx_base, large) # Sort to get ordered indices of available src/dst positions per layer. remaining_indices = np.argsort(remaining_priority, axis=1) fill_indices = np.argsort(fill_priority, axis=1) # Fill count per layer (cannot exceed either side). remaining_counts = remaining_mask.sum(axis=1) fill_counts = fill_mask.sum(axis=1) take_counts = np.minimum(remaining_counts, fill_counts) # Assign remaining new experts to remaining slots per layer. for layer_idx in range(num_layers): k = int(take_counts[layer_idx]) if k <= 0: continue src_pos = remaining_indices[layer_idx, :k] dst_pos = fill_indices[layer_idx, :k] post_phy2log[layer_idx, start + dst_pos] = new_local[ layer_idx, src_pos ] post_phy_replicas_idx[layer_idx, start + dst_pos] = new_ridx[ layer_idx, src_pos ] return post_phy2log, post_phy_replicas_idx @classmethod def rebalance_experts( cls, weight: torch.Tensor, num_replicas: int, num_groups: int, num_nodes: int, num_ranks: int, old_global_expert_indices: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Entry point for expert-parallelism load balancer. Parameters: weight: [layers, num_logical_experts], the load statistics for all logical experts num_replicas: number of physical experts, must be a multiple of `num_gpus` num_groups: number of expert groups num_nodes: number of server nodes, where the intra-node network (e.g, NVLink) is faster num_ranks: number of ranks, must be a multiple of `num_nodes` old_global_expert_indices: [layers, num_logical_experts], the old global expert indices. Used to avoid unnecessary weight copying for experts moving within one rank. Returns: phy2log: [layers, num_replicas], the expert index of each replica log2phy: [layers, num_logical_experts, X], the replica indices for each expert logcnt: [layers, num_logical_experts], number of physical replicas for each logical expert """ device = weight.device num_layers, num_logical_experts = weight.shape weight_np = weight.float().cpu().numpy() old_phy2log_np = ( old_global_expert_indices.cpu().numpy() if old_global_expert_indices is not None else None ) if num_groups % num_nodes == 0: # use hierarchical load-balance policy phy2log_np, phy_replicas_idx_np, logcnt_np = ( cls.rebalance_experts_hierarchical( weight_np, num_replicas, num_groups, num_nodes, num_ranks ) ) else: # use global load-balance policy phy2log_np, phy_replicas_idx_np, logcnt_np = ( cls.rebalance_experts_hierarchical( weight_np, num_replicas, 1, 1, num_ranks ) ) # Optional postprocessing to preserve slots for experts moving # within the same GPU # Only apply when the number of GPUs and slots per GPU remain unchanged. # Helps to avoid unnecessary weight copying when experts move # within the same GPU. if old_global_expert_indices is not None: phy2log_np, phy_replicas_idx_np = cls.preserve_intragpu_slots( phy2log_np, phy_replicas_idx_np, num_ranks, old_phy2log_np ) num_redundant_experts = num_replicas - num_logical_experts maxlogcnt = num_redundant_experts + 1 log2phy_np = np.full( (num_layers, num_logical_experts, maxlogcnt), -1, dtype=np.int64 ) layer_indices = np.arange(num_layers)[:, None] replica_indices = np.tile( np.arange(num_replicas, dtype=np.int64), (num_layers, 1) ) log2phy_np[layer_indices, phy2log_np, phy_replicas_idx_np] = replica_indices phy2log = torch.from_numpy(phy2log_np).to(device) log2phy = torch.from_numpy(log2phy_np).to(device) logcnt = torch.from_numpy(logcnt_np).to(device) return phy2log, log2phy, logcnt
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/distributed/eplb/policy/default.py", "license": "Apache License 2.0", "lines": 336, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/models/language/pooling/test_all_pooling_plus_chunked_prefill.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from transformers import AutoModel from tests.models.utils import check_embeddings_close from vllm import TokensPrompt @pytest.mark.parametrize( "model", ["Qwen/Qwen3-Embedding-0.6B"], ) @torch.inference_mode def test_embed_models(hf_runner, vllm_runner, model: str): chunk_size = 10 n_prompt_tokens = [55, 56, 57] token_prompts = [[1024 + i for i in range(n)] for n in n_prompt_tokens] with vllm_runner( model, runner="pooling", max_model_len=128, max_num_batched_tokens=chunk_size, enforce_eager=True, # `enable_chunked_prefill`: Set to `False` instead of `None` in VllmRunner enable_chunked_prefill=True, enable_prefix_caching=True, ) as vllm_model: vllm_outputs = vllm_model.token_embed( [TokensPrompt(prompt_token_ids=t) for t in token_prompts], ) with hf_runner( model, auto_cls=AutoModel, ) as hf_model: hf_outputs = [] for token_prompt in token_prompts: inputs = hf_model.wrap_device({"input_ids": torch.tensor([token_prompt])}) input_ids = inputs["input_ids"] output = hf_model.model(input_ids) hf_outputs.append(output.last_hidden_state.cpu().float()[0]) for hf_output, vllm_output in zip(hf_outputs, vllm_outputs): check_embeddings_close( embeddings_0_lst=hf_output, embeddings_1_lst=vllm_output, name_0="hf", name_1="vllm", tol=1e-2, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/language/pooling/test_all_pooling_plus_chunked_prefill.py", "license": "Apache License 2.0", "lines": 47, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/model_executor/model_loader/runai_streamer_loader/test_runai_model_streamer_s3.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from pathlib import Path from huggingface_hub import snapshot_download from runai_model_streamer.safetensors_streamer.streamer_mock import StreamerPatcher from vllm.engine.arg_utils import EngineArgs from .conftest import RunaiDummyExecutor load_format = "runai_streamer" test_model = "openai-community/gpt2" def test_runai_model_loader_download_files_s3_mocked_with_patch( vllm_runner, tmp_path: Path, monkeypatch, ): patcher = StreamerPatcher(str(tmp_path)) test_mock_s3_model = "s3://my-mock-bucket/gpt2/" # Download model from HF mock_model_dir = f"{tmp_path}/gpt2" snapshot_download(repo_id=test_model, local_dir=mock_model_dir) monkeypatch.setattr( "vllm.transformers_utils.runai_utils.runai_list_safetensors", patcher.shim_list_safetensors, ) monkeypatch.setattr( "vllm.transformers_utils.runai_utils.runai_pull_files", patcher.shim_pull_files, ) monkeypatch.setattr( "vllm.model_executor.model_loader.weight_utils.SafetensorsStreamer", patcher.create_mock_streamer, ) engine_args = EngineArgs( model=test_mock_s3_model, load_format=load_format, tensor_parallel_size=1, ) vllm_config = engine_args.create_engine_config() executor = RunaiDummyExecutor(vllm_config) executor.driver_worker.load_model()
{ "repo_id": "vllm-project/vllm", "file_path": "tests/model_executor/model_loader/runai_streamer_loader/test_runai_model_streamer_s3.py", "license": "Apache License 2.0", "lines": 39, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:benchmarks/kernels/benchmark_2d_silu_mul_fp8_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass from enum import Enum from itertools import product from typing import Any import torch import torch.utils.benchmark as TBenchmark from torch.utils.benchmark import Measurement as TMeasurement from vllm.model_executor.layers.quantization.utils.fp8_utils import ( _per_token_group_quant_fp8_colmajor, silu_mul_per_token_group_quant_fp8_colmajor, ) from vllm.triton_utils import triton from vllm.utils.deep_gemm import is_deep_gemm_e8m0_used from .utils import ArgPool, Bench, CudaGraphBenchParams GROUP_SIZE = 128 FLOAT8_T = torch.float8_e4m3fn def print_timers(timers: list[TMeasurement], cuda_graph_nops: int): print( f"Note : The timings reported above is for {cuda_graph_nops} " "consecutive invocations of the benchmarking functions. " f"Please divide by {cuda_graph_nops} for single invocation " "timings." ) compare = TBenchmark.Compare(timers) compare.print() class ImplType(Enum): SILU_MUL_PER_TOKEN_GROUP_QUANT_FP8_COLMAJOR = 1 REFERENCE = 2 def get_impl(self): if self == ImplType.SILU_MUL_PER_TOKEN_GROUP_QUANT_FP8_COLMAJOR: return silu_mul_per_token_group_quant_fp8_colmajor elif self == ImplType.REFERENCE: return reference raise ValueError(f"Unrecognized ImplType {self}") @dataclass class BenchmarkTensors: input: torch.Tensor output: torch.Tensor # Reference act output tensor ref_act_out: torch.Tensor ref_quant_out: torch.Tensor @staticmethod def make(T: int, N: int) -> "BenchmarkTensors": assert T % GROUP_SIZE == 0 assert N % (GROUP_SIZE * 2) == 0 input = torch.rand((T, N), dtype=torch.bfloat16, device="cuda") # silu_mul_per_token_group_quant_fp8_colmajor output. output = torch.rand((T, N // 2), dtype=torch.bfloat16, device="cuda").to( FLOAT8_T ) # reference output. ref_act_out = torch.empty((T, N // 2), dtype=torch.bfloat16, device="cuda") ref_quant_out = torch.empty( (T, N // 2), dtype=torch.bfloat16, device="cuda" ).to(FLOAT8_T) return BenchmarkTensors( input=input, output=output, ref_act_out=ref_act_out, ref_quant_out=ref_quant_out, ) @property def T(self): return self.input.size(0) @property def N(self): return self.input.size(1) def make_impl_kwargs(self, impl_type: ImplType) -> dict[str, Any]: if impl_type == ImplType.SILU_MUL_PER_TOKEN_GROUP_QUANT_FP8_COLMAJOR: return { "input": self.input, "output": self.output, "use_ue8m0": is_deep_gemm_e8m0_used(), } elif impl_type == ImplType.REFERENCE: return { "input": self.input, "act_out": self.ref_act_out, "quant_out": self.ref_quant_out, "use_ue8m0": is_deep_gemm_e8m0_used(), } raise ValueError(f"Unrecognized impl_type {impl_type}") def reference_quant(x: torch.Tensor, quant_out: torch.Tensor, use_ue8m0: bool): """ Reference triton quant kernel from, vllm.model_executor.layers.quantization.utils.fp8_utils """ assert quant_out.size() == x.size() # Allocate the scale tensor column-major format. shape = (x.shape[-1] // GROUP_SIZE,) + x.shape[:-1] x_q = quant_out x_s = torch.empty(shape, device=x.device, dtype=torch.float32).permute(-1, -2) M = x.numel() // GROUP_SIZE N = GROUP_SIZE BLOCK = triton.next_power_of_2(N) # heuristics for number of warps num_warps = min(max(BLOCK // 256, 1), 8) num_stages = 1 finfo = torch.finfo(FLOAT8_T) fp8_min = finfo.min fp8_max = finfo.max _per_token_group_quant_fp8_colmajor[(M,)]( x, x_q, x_s, GROUP_SIZE, x.shape[1], x.stride(0), x_s.stride(1), eps=1e-10, fp8_min=fp8_min, fp8_max=fp8_max, use_ue8m0=use_ue8m0, BLOCK=BLOCK, num_warps=num_warps, num_stages=num_stages, ) return x_q, x_s def reference( input: torch.Tensor, act_out: torch.Tensor, quant_out: torch.Tensor, use_ue8m0: bool, ) -> tuple[torch.Tensor, torch.Tensor]: torch.ops._C.silu_and_mul(act_out, input) return reference_quant(act_out, quant_out, use_ue8m0) def bench_impl( bench_tensors: list[BenchmarkTensors], impl_type: ImplType ) -> TMeasurement: T = bench_tensors[0].T N = bench_tensors[0].N arg_pool_size = len(bench_tensors) kwargs_list = [bt.make_impl_kwargs(impl_type) for bt in bench_tensors] # warmup for kwargs in kwargs_list: impl_type.get_impl()(**kwargs) torch.cuda.synchronize() # Merge into a single kwargs and qualify arguments as ArgPool kwargs = {k: ArgPool([]) for k in kwargs_list[0]} for _kwargs in kwargs_list: for k, v in _kwargs.items(): kwargs[k].values.append(v) cuda_graph_params = None cuda_graph_params = CudaGraphBenchParams(arg_pool_size) timer = None with Bench( cuda_graph_params, "silu-mul-quant", f"num_tokens={T}, N={N}", impl_type.name, impl_type.get_impl(), **kwargs, ) as bench: timer = bench.run() return timer def test_correctness(T: int, N: int): print(f"Testing num_tokens={T}, N={N} ...") bench_tensor = BenchmarkTensors.make(T, N) def output_from_impl(impl: ImplType) -> tuple[torch.Tensor, torch.Tensor]: return impl.get_impl()(**bench_tensor.make_impl_kwargs(impl)) # reference output ref_out_q, ref_out_s = output_from_impl(ImplType.REFERENCE) # test ouptut out_q, out_s = output_from_impl( ImplType.SILU_MUL_PER_TOKEN_GROUP_QUANT_FP8_COLMAJOR ) torch.testing.assert_close(ref_out_q.to(torch.float32), out_q.to(torch.float32)) torch.testing.assert_close(ref_out_s, out_s) def run(Ts: list[int], Ns: list[int], arg_pool_size: int) -> list[TMeasurement]: timers = [] for N, T in product(Ns, Ts): test_correctness(T, N) bench_tensors: list[BenchmarkTensors] = [ BenchmarkTensors.make(T, N) for _ in range(arg_pool_size) ] silu_mul_quant_timer = bench_impl( bench_tensors, ImplType.SILU_MUL_PER_TOKEN_GROUP_QUANT_FP8_COLMAJOR ) timers.append(silu_mul_quant_timer) reference_timer = bench_impl(bench_tensors, ImplType.REFERENCE) timers.append(reference_timer) print_timers( [silu_mul_quant_timer, reference_timer], cuda_graph_nops=arg_pool_size ) print_timers(timers, cuda_graph_nops=arg_pool_size) return timers if __name__ == "__main__": T = [128 * i for i in range(1, 16)] + [2048 * i for i in range(1, 65)] N = [2048, 4096, 8192] print(f"T = {T}, N = {N}") run(T, N, arg_pool_size=8)
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/kernels/benchmark_2d_silu_mul_fp8_quant.py", "license": "Apache License 2.0", "lines": 193, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/moe/test_silu_mul_per_token_group_quant_fp8_colmajor.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from vllm.model_executor.layers.quantization.utils.fp8_utils import ( _per_token_group_quant_fp8_colmajor, silu_mul_per_token_group_quant_fp8_colmajor, ) from vllm.platforms import current_platform from vllm.triton_utils import triton from vllm.utils.deep_gemm import is_deep_gemm_e8m0_used from vllm.utils.torch_utils import set_random_seed FLOAT8_DTYPE = torch.float8_e4m3fn GROUP_SIZE = 128 def reference_quant(x: torch.Tensor, use_ue8m0: bool): """ Reference triton quant kernel from, vllm.model_executor.layers.quantization.utils.fp8_utils """ x_q = torch.empty_like(x, device=x.device, dtype=FLOAT8_DTYPE) # Allocate the scale tensor in column-major format. shape = (x.shape[-1] // GROUP_SIZE,) + x.shape[:-1] x_s = torch.empty(shape, device=x.device, dtype=torch.float32).permute(-1, -2) M = x.numel() // GROUP_SIZE N = GROUP_SIZE BLOCK = triton.next_power_of_2(N) # heuristics for number of warps num_warps = min(max(BLOCK // 256, 1), 8) num_stages = 1 finfo = torch.finfo(FLOAT8_DTYPE) fp8_min = finfo.min fp8_max = finfo.max _per_token_group_quant_fp8_colmajor[(M,)]( x, x_q, x_s, GROUP_SIZE, x.shape[1], x.stride(0), x_s.stride(1), eps=1e-10, fp8_min=fp8_min, fp8_max=fp8_max, use_ue8m0=use_ue8m0, BLOCK=BLOCK, num_warps=num_warps, num_stages=num_stages, ) return x_q, x_s def reference(x: torch.Tensor, use_ue8m0: bool) -> tuple[torch.Tensor, torch.Tensor]: T, N = x.size() ref_act_out = torch.empty((T, N // 2), dtype=torch.bfloat16, device="cuda") torch.ops._C.silu_and_mul(ref_act_out, x) return reference_quant(ref_act_out, use_ue8m0) @pytest.mark.parametrize("T", [128, 256, 512]) @pytest.mark.parametrize("N", [128 * 2, 256 * 2, 768 * 2, 2048 * 2, 7168 * 2]) @pytest.mark.skipif( current_platform.is_rocm(), reason="ROCm does not support DeepGemm.", ) def test_silu_mul_fp8_quant_deep_gemm(T: int, N: int): set_random_seed(42) input = torch.rand((T, N), dtype=torch.bfloat16, device="cuda") use_ue8m0 = is_deep_gemm_e8m0_used() # Test output, output_scales = silu_mul_per_token_group_quant_fp8_colmajor( input, use_ue8m0=use_ue8m0 ) # Reference ref_output, ref_output_scales = reference(input, use_ue8m0) torch.testing.assert_close(output.to(torch.float32), ref_output.to(torch.float32)) torch.testing.assert_close(output_scales, ref_output_scales)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/moe/test_silu_mul_per_token_group_quant_fp8_colmajor.py", "license": "Apache License 2.0", "lines": 72, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:benchmarks/benchmark_hash.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Micro benchmark comparing built-in hash(), SHA-256, and xxHash. This focuses on a single test payload shaped like the prefix-cache hash input: (32-byte bytes object, 32-int tuple) Usage: python benchmarks/hash_micro_benchmark.py --iterations 20000 """ from __future__ import annotations import argparse import random import statistics import time from collections.abc import Callable, Iterable from vllm.utils.hashing import sha256, xxhash def _generate_test_data(seed: int) -> tuple[bytes, tuple[int, ...]]: """Generate a deterministic test payload.""" random.seed(seed) bytes_data = bytes(random.getrandbits(8) for _ in range(32)) int_tuple = tuple(random.randint(1, 1_000_000) for _ in range(32)) return (bytes_data, int_tuple) def _benchmark_func(func: Callable[[tuple], object], data: tuple, iterations: int): """Return (avg_seconds, std_seconds) for hashing `data` `iterations` times.""" times: list[float] = [] # Warm-up to avoid first-run noise. for _ in range(200): func(data) for _ in range(iterations): start = time.perf_counter() func(data) end = time.perf_counter() times.append(end - start) avg = statistics.mean(times) std = statistics.stdev(times) if len(times) > 1 else 0.0 return avg, std def _run_benchmarks( benchmarks: Iterable[tuple[str, Callable[[tuple], object]]], data: tuple, iterations: int, ): """Yield (name, avg, std) for each benchmark, skipping unavailable ones.""" for name, func in benchmarks: try: avg, std = _benchmark_func(func, data, iterations) except ModuleNotFoundError as exc: print(f"Skipping {name}: {exc}") continue yield name, avg, std def builtin_hash(data: tuple) -> int: """Wrapper for Python's built-in hash().""" return hash(data) def main() -> None: parser = argparse.ArgumentParser(description=__doc__) parser.add_argument( "--iterations", type=int, default=10_000, help="Number of measured iterations per hash function.", ) parser.add_argument( "--seed", type=int, default=42, help="Random seed for test payload." ) args = parser.parse_args() data = _generate_test_data(args.seed) benchmarks = ( ("SHA256 (pickle)", sha256), ("xxHash (pickle)", xxhash), ("built-in hash()", builtin_hash), ) print("=" * 60) print("HASH FUNCTION MICRO BENCHMARK") print("=" * 60) print("Test data: (32-byte bytes object, 32-int tuple)") print(f"Iterations: {args.iterations:,}") print("=" * 60) results = list(_run_benchmarks(benchmarks, data, args.iterations)) builtin_entry = next((r for r in results if r[0] == "built-in hash()"), None) print("\nResults:") for name, avg, std in results: print(f" {name:16s}: {avg * 1e6:8.2f} ± {std * 1e6:6.2f} μs") if builtin_entry: _, builtin_avg, _ = builtin_entry print("\n" + "=" * 60) print("SUMMARY (relative to built-in hash())") print("=" * 60) for name, avg, _ in results: if name == "built-in hash()": continue speed_ratio = avg / builtin_avg print(f"• {name} is {speed_ratio:.1f}x slower than built-in hash()") else: print("\nBuilt-in hash() result missing; cannot compute speed ratios.") if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/benchmark_hash.py", "license": "Apache License 2.0", "lines": 95, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:benchmarks/benchmark_prefix_block_hash.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Simple benchmark to compare prefix-cache block hashing algorithms. Example: python benchmark_prefix_block_hash.py --num-blocks 20000 --block-size 32 """ from __future__ import annotations import argparse import random import statistics import sys import time from collections.abc import Callable, Iterable, Sequence from vllm.utils.hashing import get_hash_fn_by_name from vllm.v1.core.kv_cache_utils import BlockHash, hash_block_tokens, init_none_hash SUPPORTED_ALGOS = ("sha256", "sha256_cbor", "xxhash", "xxhash_cbor") def _generate_blocks( num_blocks: int, block_size: int, vocab_size: int, seed: int ) -> list[list[int]]: rng = random.Random(seed) return [ [rng.randrange(vocab_size) for _ in range(block_size)] for _ in range(num_blocks) ] def _hash_all_blocks( hash_fn: Callable[[object], bytes], blocks: Iterable[Sequence[int]], ) -> float: parent_hash: BlockHash | None = None start = time.perf_counter() for block in blocks: parent_hash = hash_block_tokens(hash_fn, parent_hash, block, extra_keys=None) end = time.perf_counter() return end - start def _benchmark( hash_algo: str, blocks: list[list[int]], trials: int, ) -> tuple[float, float, float] | None: try: hash_fn = get_hash_fn_by_name(hash_algo) init_none_hash(hash_fn) timings = [_hash_all_blocks(hash_fn, blocks) for _ in range(trials)] except ModuleNotFoundError as exc: print(f"Skipping {hash_algo}: {exc}", file=sys.stderr) return None avg = statistics.mean(timings) best = min(timings) # throughput: tokens / second tokens_hashed = len(blocks) * len(blocks[0]) throughput = tokens_hashed / best return avg, best, throughput def main() -> None: parser = argparse.ArgumentParser(description=__doc__) parser.add_argument("--num-blocks", type=int, default=10000, help="Block count.") parser.add_argument("--block-size", type=int, default=32, help="Tokens per block.") parser.add_argument( "--vocab-size", type=int, default=32000, help="Token id range [0, vocab_size)." ) parser.add_argument("--seed", type=int, default=0, help="Random seed.") parser.add_argument( "--trials", type=int, default=5, help="Number of timed trials per algorithm." ) parser.add_argument( "--algorithms", nargs="+", default=SUPPORTED_ALGOS, choices=SUPPORTED_ALGOS, help="Hash algorithms to benchmark.", ) args = parser.parse_args() blocks = _generate_blocks( args.num_blocks, args.block_size, args.vocab_size, args.seed ) print( f"Benchmarking {len(args.algorithms)} algorithms on " f"{args.num_blocks} blocks (block size={args.block_size})." ) for algo in args.algorithms: result = _benchmark(algo, blocks, args.trials) if result is None: continue avg, best, throughput = result print( f"{algo:14s} avg: {avg:.6f}s best: {best:.6f}s " f"throughput: {throughput / 1e6:.2f}M tokens/s" ) if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/benchmark_prefix_block_hash.py", "license": "Apache License 2.0", "lines": 90, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/tokenizers/deepseek_v32_encoding.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # copy from https://huggingface.co/deepseek-ai/DeepSeek-V3.2/blob/main/encoding/encoding_dsv32.py import copy import json from typing import Any import regex as re # flake8: noqa: E501 TOOLS_SYSTEM_TEMPLATE = """## Tools You have access to a set of tools you can use to answer the user's question. You can invoke functions by writing a "<{dsml_token}function_calls>" block like the following as part of your reply to the user: <{dsml_token}function_calls> <{dsml_token}invoke name="$FUNCTION_NAME"> <{dsml_token}parameter name="$PARAMETER_NAME" string="true|false">$PARAMETER_VALUE</{dsml_token}parameter> ... </{dsml_token}invoke> <{dsml_token}invoke name="$FUNCTION_NAME2"> ... </{dsml_token}invoke> </{dsml_token}function_calls> String and scalar parameters should be specified as is without any escaping or quotes, while lists and objects should use JSON format. The "string" attribute should be set to "true" for string type parameters and "false" for other types (numbers, booleans, arrays, objects). If the thinking_mode is enabled, then after function results you should strongly consider outputting a thinking block. Here is an example: <{dsml_token}function_calls> ... </{dsml_token}function_calls> <function_results> ... </function_results> {thinking_start_token}...thinking about results{thinking_end_token} Here are the functions available in JSONSchema format: <functions> {tool_schemas} </functions> """ bos_token: str = "<|begin▁of▁sentence|>" eos_token: str = "<|end▁of▁sentence|>" thinking_start_token: str = "<think>" thinking_end_token: str = "</think>" dsml_token: str = "|DSML|" system_msg_template: str = "{content}" user_msg_template: str = "<|User|>{content}<|Assistant|>" assistant_msg_template: str = "{reasoning}{content}{tool_calls}<|end▁of▁sentence|>" thinking_template = "{reasoning}" response_format_template: str = "## Response Format:\n\nYou MUST strictly adhere to the following schema to reply:\n{schema}" tool_call_template: str = ( '<{dsml_token}invoke name="{name}">\n{arguments}\n</{dsml_token}invoke>' ) tool_calls_template = ( "<{dsml_token}function_calls>\n{tool_calls}\n</{dsml_token}function_calls>" ) tool_output_template: str = "\n<result>{content}</result>" def to_json(value: Any) -> str: try: return json.dumps(value, ensure_ascii=False) except Exception: return json.dumps(value, ensure_ascii=True) def tools_from_openai_format(tools): return [tool["function"] for tool in tools] def tool_calls_from_openai_format(tool_calls): return [ { "name": tool_call["function"]["name"], "arguments": tool_call["function"]["arguments"], } for tool_call in tool_calls ] def tool_calls_to_openai_format(tool_calls): return [ { "type": "function", "function": { "name": tool_call["name"], "arguments": tool_call["arguments"], }, } for tool_call in tool_calls ] def encode_arguments_to_dsml(tool_call: dict[str, str]) -> str: p_dsml_template = """<{dsml_token}parameter name="{key}" string="{is_str}">{value}</{dsml_token}parameter>""" P_dsml_strs = [] if isinstance(tool_call["arguments"], str): arguments = json.loads(tool_call["arguments"]) else: arguments = tool_call["arguments"] for k, v in arguments.items(): p_dsml_str = p_dsml_template.format( dsml_token=dsml_token, key=k, is_str="true" if isinstance(v, str) else "false", value=v if isinstance(v, str) else to_json(v), ) P_dsml_strs.append(p_dsml_str) return "\n".join(P_dsml_strs) def decode_dsml_to_arguments( tool_name: str, tool_args: dict[str, tuple[str, str]] ) -> dict[str, str]: def _decode_value(key: str, value: str, string: str): if string == "true": value = to_json(value) return f"{to_json(key)}: {value}" tool_args_json = ( "{" + ", ".join( [_decode_value(k, v, string=is_str) for k, (v, is_str) in tool_args.items()] ) + "}" ) return dict(name=tool_name, arguments=tool_args_json) def render_tools(tools: list[dict[str, str | dict[str, Any]]]) -> str: tools_json = [to_json(t) for t in tools] return TOOLS_SYSTEM_TEMPLATE.format( tool_schemas="\n".join(tools_json), dsml_token=dsml_token, thinking_start_token=thinking_start_token, thinking_end_token=thinking_end_token, ) def find_last_user_index(messages: list[dict[str, Any]]) -> int: last_user_index = -1 for idx in range(len(messages) - 1, -1, -1): if messages[idx].get("role") in ["user", "developer"]: last_user_index = idx break return last_user_index def render_message( index: int, messages: list[dict[str, Any]], thinking_mode: str ) -> str: if not (0 <= index < len(messages)): raise ValueError( f"Index {index} out of range for messages list of length {len(messages)}" ) if thinking_mode not in ["chat", "thinking"]: raise ValueError(f"Invalid thinking_mode `{thinking_mode}`") prompt = "" msg = messages[index] last_user_idx = find_last_user_index(messages) role = msg.get("role") content = msg.get("content") tools = msg.get("tools") response_format = msg.get("response_format") tool_calls = msg.get("tool_calls") reasoning = msg.get("reasoning") is_prefix = msg.get("prefix", False) if tools: tools = tools_from_openai_format(tools) if tool_calls: tool_calls = tool_calls_from_openai_format(tool_calls) if role == "system": prompt += system_msg_template.format(content=content or "") if tools: prompt += "\n\n" + render_tools(tools) if response_format: prompt += "\n\n" + response_format_template.format( schema=to_json(response_format) ) elif role == "developer": if not content: raise ValueError(f"Invalid message for role `{role}`: {msg}") content_developer = "" if tools: content_developer += "\n\n" + render_tools(tools) if response_format: content_developer += "\n\n" + response_format_template.format( schema=to_json(response_format) ) content_developer += "\n\n# The user's message is: {}".format(content) prompt += user_msg_template.format(content=content_developer) if index == last_user_idx and thinking_mode == "thinking": prompt += thinking_start_token else: prompt += thinking_end_token elif role == "user": prompt += user_msg_template.format(content=content) if index == last_user_idx and thinking_mode == "thinking": prompt += thinking_start_token else: prompt += thinking_end_token elif role == "tool": prev_assistant_idx = index - 1 assistant_msg = messages[prev_assistant_idx] while prev_assistant_idx >= 0 and assistant_msg.get("role") == "tool": prev_assistant_idx -= 1 assistant_msg = messages[prev_assistant_idx] if not ( index == 0 or prev_assistant_idx >= 0 and assistant_msg.get("role") == "assistant" ): raise ValueError(f"Invalid messages at {index}:\n{assistant_msg}") tool_call_order = index - prev_assistant_idx assistant_tool_calls = assistant_msg.get("tool_calls") if not (assistant_tool_calls and len(assistant_tool_calls) >= tool_call_order): raise ValueError("No tool calls but found tool output") if tool_call_order == 1: prompt += "\n\n<function_results>" prompt += tool_output_template.format(content=content) if tool_call_order == len(assistant_tool_calls): prompt += "\n</function_results>" if index >= last_user_idx and thinking_mode == "thinking": prompt += "\n\n" + thinking_start_token else: prompt += "\n\n" + thinking_end_token elif role == "assistant": prev_assistant_idx = index thinking_part = "" tool_calls_content = "" if tool_calls: tool_calls = [ tool_call_template.format( dsml_token=dsml_token, name=tool_call.get("name"), arguments=encode_arguments_to_dsml(tool_call), ) for tool_call in tool_calls ] tool_calls_content += "\n\n" + tool_calls_template.format( dsml_token=dsml_token, tool_calls="\n".join(tool_calls) ) summary_content = content or "" if thinking_mode == "thinking" and index > last_user_idx: if not (reasoning or tool_calls): raise ValueError( f"ThinkingMode: {thinking_mode}, invalid message without reasoning/tool_calls `{msg}` after last user message" ) thinking_part = ( thinking_template.format(reasoning=reasoning or "") + thinking_end_token ) if not tool_calls and is_prefix: prompt += summary_content else: prompt += assistant_msg_template.format( reasoning=thinking_part, content=summary_content, tool_calls=tool_calls_content, ) else: raise NotImplementedError(f"Unknown role: {role}") return prompt def drop_thinking_messages( messages: list[dict[str, Any]], last_user_idx: int | None = None ) -> list[dict[str, Any]]: messages_wo_thinking: list[dict[str, Any]] = [] last_user_idx = ( find_last_user_index(messages) if last_user_idx is None else last_user_idx ) for idx, msg in enumerate(messages): role = msg.get("role") if role in ["user", "system", "tool"] or idx >= last_user_idx: messages_wo_thinking.append(msg) continue elif role == "assistant": msg_wo_thinking = copy.copy(msg) msg_wo_thinking.pop("reasoning", None) messages_wo_thinking.append(msg_wo_thinking) return messages_wo_thinking def encode_messages( messages: list[dict[str, Any]], thinking_mode: str, context: list[dict[str, Any]] | None = None, drop_thinking: bool = True, add_default_bos_token: bool = True, ) -> str: context = context if context else [] full_messages = context + messages prompt = bos_token if add_default_bos_token and len(context) == 0 else "" if thinking_mode == "thinking" and drop_thinking: full_messages = drop_thinking_messages(full_messages) for idx in range(len(messages)): prompt += render_message( idx + len(context), full_messages, thinking_mode=thinking_mode ) return prompt def _read_until_stop( index: int, text: str, stop: list[str] ) -> tuple[int, str, None | str]: min_pos = len(text) matched_stop = None for s in stop: pos = text.find(s, index) if pos != -1 and pos < min_pos: min_pos = pos matched_stop = s if matched_stop: content = text[index:min_pos] return min_pos + len(matched_stop), content, matched_stop else: content = text[index:] return len(text), content, None def parse_tool_calls(index: int, text: str): tool_calls: list[dict[str, Any]] = [] stop_token = None tool_calls_end_token = f"</{dsml_token}function_calls>" while index < len(text): index, _, stop_token = _read_until_stop( index, text, [f"<{dsml_token}invoke", tool_calls_end_token] ) if _ != ">\n": raise RuntimeError("Tool call format error") if stop_token == tool_calls_end_token: break if stop_token is None: raise RuntimeError("Missing special token") index, tool_name_content, stop_token = _read_until_stop( index, text, [f"<{dsml_token}parameter", f"</{dsml_token}invoke"] ) p_tool_name = re.findall( r'^\s*name="(.*?)">\n$', tool_name_content, flags=re.DOTALL ) if len(p_tool_name) != 1: raise RuntimeError("Tool name format error") tool_name = p_tool_name[0] tool_args: dict[str, tuple[str, str]] = {} while stop_token == f"<{dsml_token}parameter": index, param_content, stop_token = _read_until_stop( index, text, [f"/{dsml_token}parameter"] ) param_kv = re.findall( r'^ name="(.*?)" string="(true|false)">(.*?)<$', param_content, flags=re.DOTALL, ) if len(param_kv) != 1: raise RuntimeError("Parameter format error") param_name, string, param_value = param_kv[0] if param_name in tool_args: raise RuntimeError("Duplicate parameter name") tool_args[param_name] = (param_value, string) index, content, stop_token = _read_until_stop( index, text, [f"<{dsml_token}parameter", f"</{dsml_token}invoke"] ) if content != ">\n": raise RuntimeError("Parameter format error") tool_call = decode_dsml_to_arguments(tool_name=tool_name, tool_args=tool_args) tool_calls.append(tool_call) return index, stop_token, tool_calls # NOTE: This function is designed to parse only correctly # formatted string and will not attempt to correct malformed output # that may be generated by the model. def parse_message_from_completion_text(text: str, thinking_mode: str): summary_content, reasoning, tool_calls = "", "", [] index, stop_token = 0, None tool_calls_start_token = f"\n\n<{dsml_token}function_calls" is_thinking, is_tool_calling = thinking_mode == "thinking", False if is_thinking: index, content_delta, stop_token = _read_until_stop( index, text, [thinking_end_token, tool_calls_start_token] ) reasoning = content_delta if stop_token != thinking_end_token: raise RuntimeError("Invalid thinking format") index, content_delta, stop_token = _read_until_stop( index, text, [eos_token, tool_calls_start_token] ) summary_content = content_delta if stop_token == tool_calls_start_token: is_tool_calling = True else: if stop_token != eos_token: raise RuntimeError("Invalid summary format") if is_tool_calling: index, stop_token, tool_calls = parse_tool_calls(index, text) index, tool_ends_text, stop_token = _read_until_stop(index, text, [eos_token]) if tool_ends_text: raise RuntimeError("Unexpected content after tool calls") if not (len(text) == index and stop_token in [eos_token, None]): raise RuntimeError("Unexpected content at end") for sp_token in [ bos_token, eos_token, thinking_start_token, thinking_end_token, dsml_token, ]: if sp_token in summary_content or sp_token in reasoning: raise RuntimeError("Unexpected special token in content") return { "role": "assistant", "content": summary_content, "reasoning": reasoning, "tool_calls": tool_calls_to_openai_format(tool_calls), }
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/tokenizers/deepseek_v32_encoding.py", "license": "Apache License 2.0", "lines": 380, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/serve/disagg/api_router.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import asyncio import json from http import HTTPStatus from fastapi import APIRouter, Depends, FastAPI, HTTPException, Request, Response from fastapi.responses import JSONResponse, StreamingResponse from vllm.engine.protocol import EngineClient from vllm.entrypoints.openai.engine.protocol import ( ErrorResponse, ) from vllm.entrypoints.openai.utils import validate_json_request from vllm.entrypoints.serve.disagg.protocol import ( GenerateRequest, GenerateResponse, ) from vllm.entrypoints.serve.disagg.serving import ( ServingTokens, ) from vllm.entrypoints.serve.tokenize.serving import OpenAIServingTokenization from vllm.entrypoints.utils import ( load_aware_call, with_cancellation, ) from vllm.logger import init_logger logger = init_logger(__name__) def tokenization(request: Request) -> OpenAIServingTokenization: return request.app.state.openai_serving_tokenization def generate_tokens(request: Request) -> ServingTokens | None: return request.app.state.serving_tokens def engine_client(request: Request) -> EngineClient: return request.app.state.engine_client router = APIRouter() @router.post( "/inference/v1/generate", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.OK.value: {"content": {"text/event-stream": {}}}, HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.NOT_FOUND.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, }, ) @with_cancellation @load_aware_call async def generate(request: GenerateRequest, raw_request: Request): handler = generate_tokens(raw_request) if handler is None: return tokenization(raw_request).create_error_response( message="The model does not support generate tokens API" ) try: generator = await handler.serve_tokens(request, raw_request) except Exception as e: generator = handler.create_error_response(e) if isinstance(generator, ErrorResponse): return JSONResponse( content=generator.model_dump(), status_code=generator.error.code ) elif isinstance(generator, GenerateResponse): return JSONResponse(content=generator.model_dump()) return StreamingResponse(content=generator, media_type="text/event-stream") def attach_router(app: FastAPI): if getattr(app.state.args, "tokens_only", False): @router.post("/abort_requests") async def abort_requests(raw_request: Request): """ Abort one or more requests. To be used in a Disaggregated Everything setup. """ try: body = await raw_request.json() except json.JSONDecodeError as e: raise HTTPException( status_code=HTTPStatus.BAD_REQUEST.value, detail=f"JSON decode error: {e}", ) from e request_ids = body.get("request_ids") if request_ids is None: raise HTTPException( status_code=HTTPStatus.BAD_REQUEST.value, detail="Missing 'request_ids' in request body", ) # Abort requests in background asyncio.create_task(engine_client(raw_request).abort(request_ids)) return Response(status_code=200) app.include_router(router)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/serve/disagg/api_router.py", "license": "Apache License 2.0", "lines": 87, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/serve/disagg/protocol.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any from pydantic import BaseModel, Field from vllm.config import ModelConfig from vllm.entrypoints.openai.chat_completion.protocol import ChatCompletionLogProbs from vllm.entrypoints.openai.engine.protocol import ( SamplingParams, StreamOptions, ) from vllm.logprobs import Logprob from vllm.renderers import TokenizeParams from vllm.utils import random_uuid ####### Tokens IN <> Tokens OUT ####### class GenerateRequest(BaseModel): request_id: str = Field( default_factory=lambda: f"{random_uuid()}", description=( "The request_id related to this request. If the caller does " "not set it, a random_uuid will be generated. This id is used " "through out the inference process and return in response." ), ) token_ids: list[int] """The token ids to generate text from.""" # features: MultiModalFeatureSpec # TODO (NickLucche): implement once Renderer work is completed features: str | None = None """The processed MM inputs for the model.""" sampling_params: SamplingParams """The sampling parameters for the model.""" model: str | None = None stream: bool | None = False stream_options: StreamOptions | None = None cache_salt: str | None = Field( default=None, description=( "If specified, the prefix cache will be salted with the provided " "string to prevent an attacker to guess prompts in multi-user " "environments. The salt should be random, protected from " "access by 3rd parties, and long enough to be " "unpredictable (e.g., 43 characters base64-encoded, corresponding " "to 256 bit)." ), ) priority: int = Field( default=0, description=( "The priority of the request (lower means earlier handling; " "default: 0). Any priority other than 0 will raise an error " "if the served model does not use priority scheduling." ), ) kv_transfer_params: dict[str, Any] | None = Field( default=None, description="KVTransfer parameters used for disaggregated serving.", ) def build_tok_params(self, model_config: ModelConfig) -> TokenizeParams: return TokenizeParams( max_total_tokens=None, max_output_tokens=0, ) class GenerateResponseChoice(BaseModel): index: int logprobs: ChatCompletionLogProbs | None = None # per OpenAI spec this is the default finish_reason: str | None = "stop" token_ids: list[int] | None = None class GenerateResponse(BaseModel): request_id: str = Field( default_factory=lambda: f"{random_uuid()}", description=( "The request_id related to this request. If the caller does " "not set it, a random_uuid will be generated. This id is used " "through out the inference process and return in response." ), ) choices: list[GenerateResponseChoice] prompt_logprobs: list[dict[int, Logprob] | None] | None = None kv_transfer_params: dict[str, Any] | None = Field( default=None, description="KVTransfer parameters used for disaggregated serving.", )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/serve/disagg/protocol.py", "license": "Apache License 2.0", "lines": 83, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/serve/elastic_ep/api_router.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json from http import HTTPStatus from fastapi import APIRouter, Depends, FastAPI, HTTPException, Request from fastapi.responses import JSONResponse from vllm.engine.protocol import EngineClient from vllm.entrypoints.openai.engine.protocol import ( ErrorResponse, ) from vllm.entrypoints.openai.utils import validate_json_request from vllm.entrypoints.serve.elastic_ep.middleware import ( get_scaling_elastic_ep, set_scaling_elastic_ep, ) from vllm.logger import init_logger logger = init_logger(__name__) def engine_client(request: Request) -> EngineClient: return request.app.state.engine_client router = APIRouter() @router.post( "/scale_elastic_ep", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.OK.value: {"model": dict}, HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.REQUEST_TIMEOUT.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, }, ) async def scale_elastic_ep(raw_request: Request): try: body = await raw_request.json() except json.JSONDecodeError as e: raise HTTPException(status_code=400, detail="Invalid JSON format") from e new_data_parallel_size = body.get("new_data_parallel_size") drain_timeout = body.get("drain_timeout", 120) # Default 2 minutes if new_data_parallel_size is None: raise HTTPException( status_code=400, detail="new_data_parallel_size is required" ) if not isinstance(new_data_parallel_size, int) or new_data_parallel_size <= 0: raise HTTPException( status_code=400, detail="new_data_parallel_size must be a positive integer", ) if not isinstance(drain_timeout, int) or drain_timeout <= 0: raise HTTPException( status_code=400, detail="drain_timeout must be a positive integer" ) # Set scaling flag to prevent new requests set_scaling_elastic_ep(True) client = engine_client(raw_request) try: await client.scale_elastic_ep(new_data_parallel_size, drain_timeout) return JSONResponse( { "message": f"Scaled to {new_data_parallel_size} data parallel engines", } ) except TimeoutError as e: raise HTTPException( status_code=408, detail="Scale failed due to request drain timeout " f"after {drain_timeout} seconds", ) from e except Exception as e: logger.error("Scale failed: %s", e) raise HTTPException(status_code=500, detail="Scale failed") from e finally: set_scaling_elastic_ep(False) @router.post("/is_scaling_elastic_ep") async def is_scaling_elastic_ep(raw_request: Request): return JSONResponse({"is_scaling_elastic_ep": get_scaling_elastic_ep()}) def attach_router(app: FastAPI): app.include_router(router)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/serve/elastic_ep/api_router.py", "license": "Apache License 2.0", "lines": 76, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/serve/elastic_ep/middleware.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Awaitable from fastapi.responses import JSONResponse from starlette.types import ASGIApp, Receive, Scope, Send # Global variable to track scaling state _scaling_elastic_ep = False def get_scaling_elastic_ep(): return _scaling_elastic_ep def set_scaling_elastic_ep(value): global _scaling_elastic_ep _scaling_elastic_ep = value class ScalingMiddleware: """ Middleware that checks if the model is currently scaling and returns a 503 Service Unavailable response if it is. This middleware applies to all HTTP requests and prevents processing when the model is in a scaling state. """ def __init__(self, app: ASGIApp) -> None: self.app = app def __call__(self, scope: Scope, receive: Receive, send: Send) -> Awaitable[None]: if scope["type"] != "http": return self.app(scope, receive, send) # Check global scaling state if get_scaling_elastic_ep(): # Return 503 Service Unavailable response response = JSONResponse( content={ "error": "The model is currently scaling. Please try again later." }, status_code=503, ) return response(scope, receive, send) return self.app(scope, receive, send)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/serve/elastic_ep/middleware.py", "license": "Apache License 2.0", "lines": 35, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/serve/instrumentator/health.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from fastapi import APIRouter, Request from fastapi.responses import Response from vllm.engine.protocol import EngineClient from vllm.logger import init_logger from vllm.v1.engine.exceptions import EngineDeadError logger = init_logger(__name__) router = APIRouter() def engine_client(request: Request) -> EngineClient: return request.app.state.engine_client @router.get("/health", response_class=Response) async def health(raw_request: Request) -> Response: """Health check.""" try: await engine_client(raw_request).check_health() return Response(status_code=200) except EngineDeadError: return Response(status_code=503)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/serve/instrumentator/health.py", "license": "Apache License 2.0", "lines": 19, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/serve/instrumentator/metrics.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import prometheus_client import regex as re from fastapi import FastAPI, Response from prometheus_client import make_asgi_app from prometheus_fastapi_instrumentator import Instrumentator from starlette.routing import Mount from vllm.v1.metrics.prometheus import get_prometheus_registry class PrometheusResponse(Response): media_type = prometheus_client.CONTENT_TYPE_LATEST def attach_router(app: FastAPI): """Mount prometheus metrics to a FastAPI app.""" registry = get_prometheus_registry() # `response_class=PrometheusResponse` is needed to return an HTTP response # with header "Content-Type: text/plain; version=0.0.4; charset=utf-8" # instead of the default "application/json" which is incorrect. # See https://github.com/trallnag/prometheus-fastapi-instrumentator/issues/163#issue-1296092364 Instrumentator( excluded_handlers=[ "/metrics", "/health", "/load", "/ping", "/version", "/server_info", ], registry=registry, ).add().instrument(app).expose(app, response_class=PrometheusResponse) # Add prometheus asgi middleware to route /metrics requests metrics_route = Mount("/metrics", make_asgi_app(registry=registry)) # Workaround for 307 Redirect for /metrics metrics_route.path_regex = re.compile("^/metrics(?P<path>.*)$") app.routes.append(metrics_route)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/serve/instrumentator/metrics.py", "license": "Apache License 2.0", "lines": 34, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/serve/profile/api_router.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from fastapi import APIRouter, FastAPI, Request from fastapi.responses import Response from vllm.config import ProfilerConfig from vllm.engine.protocol import EngineClient from vllm.logger import init_logger logger = init_logger(__name__) router = APIRouter() def engine_client(request: Request) -> EngineClient: return request.app.state.engine_client @router.post("/start_profile") async def start_profile(raw_request: Request): logger.info("Starting profiler...") await engine_client(raw_request).start_profile() logger.info("Profiler started.") return Response(status_code=200) @router.post("/stop_profile") async def stop_profile(raw_request: Request): logger.info("Stopping profiler...") await engine_client(raw_request).stop_profile() logger.info("Profiler stopped.") return Response(status_code=200) def attach_router(app: FastAPI): profiler_config = getattr(app.state.args, "profiler_config", None) assert profiler_config is None or isinstance(profiler_config, ProfilerConfig) if profiler_config is not None and profiler_config.profiler is not None: logger.warning_once( "Profiler with mode '%s' is enabled in the " "API server. This should ONLY be used for local development!", profiler_config.profiler, ) app.include_router(router)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/serve/profile/api_router.py", "license": "Apache License 2.0", "lines": 33, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/serve/rlhf/api_router.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json from http import HTTPStatus from typing import Annotated from fastapi import APIRouter, FastAPI, HTTPException, Query, Request from fastapi.responses import JSONResponse import vllm.envs as envs from vllm.distributed.weight_transfer.base import ( WeightTransferInitRequest, WeightTransferUpdateRequest, ) from vllm.engine.protocol import EngineClient from vllm.logger import init_logger from vllm.v1.engine import PauseMode logger = init_logger(__name__) def engine_client(request: Request) -> EngineClient: return request.app.state.engine_client router = APIRouter() @router.post("/pause") async def pause_generation( raw_request: Request, mode: Annotated[PauseMode, Query()] = "abort", wait_for_inflight_requests: bool = Query(False), clear_cache: Annotated[bool, Query()] = True, ) -> JSONResponse: """Pause generation requests to allow weight updates. Args: mode: How to handle in-flight requests: - ``"abort"``: Abort all in-flight requests immediately (default). - ``"wait"``: Wait for in-flight requests to complete. - ``"keep"``: Freeze requests in queue; they resume on /resume. wait_for_inflight_requests: DEPRECATED. Use ``mode="wait"`` instead. clear_cache: DEPRECATED. Whether to clear KV/prefix caches after draining. Ignored when mode="keep". """ engine = engine_client(raw_request) try: await engine.pause_generation( mode=mode, clear_cache=clear_cache, wait_for_inflight_requests=wait_for_inflight_requests, ) return JSONResponse( content={"status": "paused"}, status_code=HTTPStatus.OK.value, ) except ValueError as err: return JSONResponse( content={"error": str(err)}, status_code=HTTPStatus.BAD_REQUEST.value, ) except Exception as err: # pragma: no cover - defensive logger.exception("Failed to pause generation") return JSONResponse( content={"error": f"Failed to pause generation: {err}"}, status_code=HTTPStatus.INTERNAL_SERVER_ERROR.value, ) @router.post("/resume") async def resume_generation(raw_request: Request) -> JSONResponse: """Resume generation after a pause.""" engine = engine_client(raw_request) try: await engine.resume_generation() return JSONResponse( content={"status": "resumed"}, status_code=HTTPStatus.OK.value, ) except Exception as err: # pragma: no cover - defensive logger.exception("Failed to resume generation") return JSONResponse( content={"error": f"Failed to resume generation: {err}"}, status_code=HTTPStatus.INTERNAL_SERVER_ERROR.value, ) @router.get("/is_paused") async def is_paused(raw_request: Request) -> JSONResponse: """Return the current pause status.""" engine = engine_client(raw_request) try: paused = await engine.is_paused() except Exception as err: # pragma: no cover - defensive logger.exception("Failed to fetch pause status") return JSONResponse( content={"error": f"Failed to fetch pause status: {err}"}, status_code=HTTPStatus.INTERNAL_SERVER_ERROR.value, ) return JSONResponse(content={"is_paused": paused}) @router.post("/init_weight_transfer_engine") async def init_weight_transfer_engine(raw_request: Request): try: body = await raw_request.json() except json.JSONDecodeError as e: raise HTTPException(status_code=400, detail="Invalid JSON format") from e # noqa: B904 init_info = body.get("init_info") if init_info is None: raise HTTPException( status_code=HTTPStatus.BAD_REQUEST.value, detail="Missing 'init_info' in request body", ) await engine_client(raw_request).init_weight_transfer_engine( WeightTransferInitRequest(init_info=init_info) ) return JSONResponse(content={"message": "Weight transfer initialized"}) @router.post("/update_weights") async def update_weights(raw_request: Request): try: body = await raw_request.json() except json.JSONDecodeError as e: raise HTTPException(status_code=400, detail="Invalid JSON format") from e # noqa: B904 update_info = body.get("update_info") if update_info is None: raise HTTPException( status_code=HTTPStatus.BAD_REQUEST.value, detail="Missing 'update_info' in request body", ) await engine_client(raw_request).update_weights( request=WeightTransferUpdateRequest(update_info=update_info) ) return JSONResponse(content={"message": "Weights updated"}) @router.get("/get_world_size") async def get_world_size( raw_request: Request, include_dp: bool = Query(True), ): """Get the world size from the parallel config. Args: include_dp: If True (default), returns the world size including data parallelism (TP * PP * DP). If False, returns the world size without data parallelism (TP * PP). """ parallel_config = engine_client(raw_request).vllm_config.parallel_config if include_dp: world_size = parallel_config.world_size_across_dp else: world_size = parallel_config.world_size return JSONResponse(content={"world_size": world_size}) def attach_router(app: FastAPI): if not envs.VLLM_SERVER_DEV_MODE: return app.include_router(router)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/serve/rlhf/api_router.py", "license": "Apache License 2.0", "lines": 140, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/serve/sleep/api_router.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from fastapi import APIRouter, FastAPI, Request from fastapi.responses import JSONResponse, Response import vllm.envs as envs from vllm.engine.protocol import EngineClient from vllm.logger import init_logger logger = init_logger(__name__) def engine_client(request: Request) -> EngineClient: return request.app.state.engine_client router = APIRouter() @router.post("/sleep") async def sleep(raw_request: Request): # get POST params level = raw_request.query_params.get("level", "1") mode = raw_request.query_params.get("mode", "abort") await engine_client(raw_request).sleep(int(level), mode) # FIXME: in v0 with frontend multiprocessing, the sleep command # is sent but does not finish yet when we return a response. return Response(status_code=200) @router.post("/wake_up") async def wake_up(raw_request: Request): tags = raw_request.query_params.getlist("tags") if tags == []: # set to None to wake up all tags if no tags are provided tags = None logger.info("wake up the engine with tags: %s", tags) await engine_client(raw_request).wake_up(tags) # FIXME: in v0 with frontend multiprocessing, the wake-up command # is sent but does not finish yet when we return a response. return Response(status_code=200) @router.get("/is_sleeping") async def is_sleeping(raw_request: Request): logger.info("check whether the engine is sleeping") is_sleeping = await engine_client(raw_request).is_sleeping() return JSONResponse(content={"is_sleeping": is_sleeping}) def attach_router(app: FastAPI): if not envs.VLLM_SERVER_DEV_MODE: return app.include_router(router)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/serve/sleep/api_router.py", "license": "Apache License 2.0", "lines": 40, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/serve/tokenize/api_router.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from http import HTTPStatus from fastapi import APIRouter, Depends, FastAPI, HTTPException, Request from fastapi.exceptions import RequestValidationError from fastapi.responses import JSONResponse from typing_extensions import assert_never from vllm.entrypoints.openai.engine.protocol import ( ErrorResponse, ) from vllm.entrypoints.openai.utils import validate_json_request from vllm.entrypoints.serve.tokenize.protocol import ( DetokenizeRequest, DetokenizeResponse, TokenizeRequest, TokenizeResponse, ) from vllm.entrypoints.serve.tokenize.serving import OpenAIServingTokenization from vllm.entrypoints.utils import ( with_cancellation, ) from vllm.logger import init_logger logger = init_logger(__name__) def tokenization(request: Request) -> OpenAIServingTokenization: return request.app.state.openai_serving_tokenization router = APIRouter() @router.post( "/tokenize", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.NOT_FOUND.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, HTTPStatus.NOT_IMPLEMENTED.value: {"model": ErrorResponse}, }, ) @with_cancellation async def tokenize(request: TokenizeRequest, raw_request: Request): handler = tokenization(raw_request) try: generator = await handler.create_tokenize(request, raw_request) except Exception as e: generator = handler.create_error_response(e) if isinstance(generator, ErrorResponse): return JSONResponse( content=generator.model_dump(), status_code=generator.error.code ) elif isinstance(generator, TokenizeResponse): return JSONResponse(content=generator.model_dump()) assert_never(generator) @router.post( "/detokenize", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.NOT_FOUND.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, }, ) @with_cancellation async def detokenize(request: DetokenizeRequest, raw_request: Request): handler = tokenization(raw_request) try: generator = await handler.create_detokenize(request, raw_request) except OverflowError as e: raise RequestValidationError(errors=[str(e)]) from e except Exception as e: raise HTTPException( status_code=HTTPStatus.INTERNAL_SERVER_ERROR.value, detail=str(e) ) from e if isinstance(generator, ErrorResponse): return JSONResponse( content=generator.model_dump(), status_code=generator.error.code ) elif isinstance(generator, DetokenizeResponse): return JSONResponse(content=generator.model_dump()) assert_never(generator) def attach_router(app: FastAPI): if getattr(app.state.args, "enable_tokenizer_info_endpoint", False): """Conditionally register the tokenizer info endpoint if enabled.""" @router.get("/tokenizer_info") async def get_tokenizer_info(raw_request: Request): """Get comprehensive tokenizer information.""" result = await tokenization(raw_request).get_tokenizer_info() return JSONResponse( content=result.model_dump(), status_code=result.error.code if isinstance(result, ErrorResponse) else 200, ) app.include_router(router)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/serve/tokenize/api_router.py", "license": "Apache License 2.0", "lines": 91, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/transformers_utils/configs/tarsier2.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from transformers import Qwen2VLConfig class Tarsier2Config(Qwen2VLConfig): """ Tarsier2's config.json is written such that AutoConfig.from_pretrained will create a deeply nested config consisting of: - LlavaConfig - Qwen2VLConfig - Qwen2VLTextConfig - Qwen2VLVisionConfig - Qwen2VLConfig - Qwen2VLTextConfig - Qwen2VLVisionConfig When it should really just be a single Qwen2VLConfig. This class is a hack to stop AutoConfig from creating the nested config structure. """ model_type = "tarsier2"
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/configs/tarsier2.py", "license": "Apache License 2.0", "lines": 18, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/benchmarks/test_plot_filters.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pandas as pd import pytest from vllm.benchmarks.sweep.plot import ( PlotEqualTo, PlotFilterBase, PlotFilters, PlotGreaterThan, PlotGreaterThanOrEqualTo, PlotLessThan, PlotLessThanOrEqualTo, PlotNotEqualTo, ) class TestPlotFilters: """Test PlotFilter functionality including 'inf' edge case.""" def setup_method(self): """Create sample DataFrames for testing.""" # DataFrame with numeric values self.df_numeric = pd.DataFrame( { "request_rate": [1.0, 5.0, 10.0, 50.0, 100.0], "value": [10, 20, 30, 40, 50], } ) # DataFrame with float('inf') - note: string "inf" values are coerced # to float when loading data, so we only test with float('inf') self.df_inf_float = pd.DataFrame( { "request_rate": [1.0, 5.0, 10.0, float("inf"), float("inf")], "value": [10, 20, 30, 40, 50], } ) @pytest.mark.parametrize( "target,expected_count", [ ("5.0", 1), ("10.0", 1), ("1.0", 1), ], ) def test_equal_to_numeric(self, target, expected_count): """Test PlotEqualTo with numeric values.""" filter_obj = PlotEqualTo("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count def test_equal_to_inf_float(self): """Test PlotEqualTo with float('inf').""" filter_obj = PlotEqualTo("request_rate", "inf") result = filter_obj.apply(self.df_inf_float) # Should match both float('inf') entries because float('inf') == float('inf') assert len(result) == 2 @pytest.mark.parametrize( "target,expected_count", [ ("5.0", 4), # All except 5.0 ("1.0", 4), # All except 1.0 ], ) def test_not_equal_to_numeric(self, target, expected_count): """Test PlotNotEqualTo with numeric values.""" filter_obj = PlotNotEqualTo("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count def test_not_equal_to_inf_float(self): """Test PlotNotEqualTo with float('inf').""" filter_obj = PlotNotEqualTo("request_rate", "inf") result = filter_obj.apply(self.df_inf_float) # Should exclude float('inf') entries assert len(result) == 3 @pytest.mark.parametrize( "target,expected_count", [ ("10.0", 2), # 1.0, 5.0 ("50.0", 3), # 1.0, 5.0, 10.0 ("5.0", 1), # 1.0 ], ) def test_less_than(self, target, expected_count): """Test PlotLessThan with numeric values.""" filter_obj = PlotLessThan("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count @pytest.mark.parametrize( "target,expected_count", [ ("10.0", 3), # 1.0, 5.0, 10.0 ("5.0", 2), # 1.0, 5.0 ], ) def test_less_than_or_equal_to(self, target, expected_count): """Test PlotLessThanOrEqualTo with numeric values.""" filter_obj = PlotLessThanOrEqualTo("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count @pytest.mark.parametrize( "target,expected_count", [ ("10.0", 2), # 50.0, 100.0 ("5.0", 3), # 10.0, 50.0, 100.0 ], ) def test_greater_than(self, target, expected_count): """Test PlotGreaterThan with numeric values.""" filter_obj = PlotGreaterThan("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count @pytest.mark.parametrize( "target,expected_count", [ ("10.0", 3), # 10.0, 50.0, 100.0 ("5.0", 4), # 5.0, 10.0, 50.0, 100.0 ], ) def test_greater_than_or_equal_to(self, target, expected_count): """Test PlotGreaterThanOrEqualTo with numeric values.""" filter_obj = PlotGreaterThanOrEqualTo("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count @pytest.mark.parametrize( "filter_str,expected_var,expected_target,expected_type", [ ("request_rate==5.0", "request_rate", "5.0", PlotEqualTo), ("request_rate!=10.0", "request_rate", "10.0", PlotNotEqualTo), ("request_rate<50.0", "request_rate", "50.0", PlotLessThan), ("request_rate<=50.0", "request_rate", "50.0", PlotLessThanOrEqualTo), ("request_rate>10.0", "request_rate", "10.0", PlotGreaterThan), ("request_rate>=10.0", "request_rate", "10.0", PlotGreaterThanOrEqualTo), ("request_rate==inf", "request_rate", "inf", PlotEqualTo), ("request_rate!='inf'", "request_rate", "inf", PlotNotEqualTo), ], ) def test_parse_str(self, filter_str, expected_var, expected_target, expected_type): """Test parsing filter strings.""" filter_obj = PlotFilterBase.parse_str(filter_str) assert isinstance(filter_obj, expected_type) assert filter_obj.var == expected_var assert filter_obj.target == expected_target def test_parse_str_inf_edge_case(self): """Test parsing 'inf' string in filter.""" filter_obj = PlotFilterBase.parse_str("request_rate==inf") assert isinstance(filter_obj, PlotEqualTo) assert filter_obj.var == "request_rate" assert filter_obj.target == "inf" def test_parse_multiple_filters(self): """Test parsing multiple filters.""" filters = PlotFilters.parse_str("request_rate>5.0,value<=40") assert len(filters) == 2 assert isinstance(filters[0], PlotGreaterThan) assert isinstance(filters[1], PlotLessThanOrEqualTo) def test_parse_empty_filter(self): """Test parsing empty filter string.""" filters = PlotFilters.parse_str("") assert len(filters) == 0
{ "repo_id": "vllm-project/vllm", "file_path": "tests/benchmarks/test_plot_filters.py", "license": "Apache License 2.0", "lines": 154, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/entrypoints/openai/parser/responses_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import logging from collections.abc import Callable from openai.types.responses import ResponseFunctionToolCall, ResponseOutputItem from openai.types.responses.response_function_tool_call_output_item import ( ResponseFunctionToolCallOutputItem, ) from openai.types.responses.response_output_item import McpCall from openai.types.responses.response_output_message import ResponseOutputMessage from openai.types.responses.response_output_text import ResponseOutputText from openai.types.responses.response_reasoning_item import ( Content, ResponseReasoningItem, ) from vllm.entrypoints.constants import MCP_PREFIX from vllm.entrypoints.openai.responses.protocol import ( ResponseInputOutputItem, ResponsesRequest, ) from vllm.outputs import CompletionOutput from vllm.reasoning.abs_reasoning_parsers import ReasoningParser from vllm.tokenizers import TokenizerLike from vllm.tool_parsers.abstract_tool_parser import ToolParser from vllm.utils import random_uuid logger = logging.getLogger(__name__) class ResponsesParser: """Incremental parser over completion tokens with reasoning support.""" def __init__( self, *, tokenizer: TokenizerLike, reasoning_parser_cls: Callable[[TokenizerLike], ReasoningParser], response_messages: list[ResponseInputOutputItem], request: ResponsesRequest, tool_parser_cls: Callable[[TokenizerLike], ToolParser] | None, ): self.response_messages: list[ResponseInputOutputItem] = ( # TODO: initial messages may not be properly typed response_messages ) self.num_init_messages = len(response_messages) self.tokenizer = tokenizer self.request = request self.reasoning_parser_instance = reasoning_parser_cls(tokenizer) self.tool_parser_instance = None if tool_parser_cls is not None: self.tool_parser_instance = tool_parser_cls(tokenizer) # Store the last finish_reason to determine response status self.finish_reason: str | None = None def process(self, output: CompletionOutput) -> "ResponsesParser": # Store the finish_reason from the output self.finish_reason = output.finish_reason reasoning_content, content = self.reasoning_parser_instance.extract_reasoning( output.text, request=self.request ) if reasoning_content: self.response_messages.append( ResponseReasoningItem( type="reasoning", id=f"rs_{random_uuid()}", summary=[], content=[ Content( type="reasoning_text", text=reasoning_content, ) ], ) ) function_calls: list[ResponseFunctionToolCall] = [] if self.tool_parser_instance is not None: tool_call_info = self.tool_parser_instance.extract_tool_calls( content if content is not None else "", request=self.request, # type: ignore ) if tool_call_info is not None and tool_call_info.tools_called: # extract_tool_calls() returns a list of tool calls. function_calls.extend( ResponseFunctionToolCall( id=f"fc_{random_uuid()}", call_id=f"call_{random_uuid()}", type="function_call", status="completed", name=tool_call.function.name, arguments=tool_call.function.arguments, ) for tool_call in tool_call_info.tool_calls ) content = tool_call_info.content if content and content.strip() == "": content = None if content: self.response_messages.append( ResponseOutputMessage( type="message", id=f"msg_{random_uuid()}", status="completed", role="assistant", content=[ ResponseOutputText( annotations=[], # TODO type="output_text", text=content, logprobs=None, # TODO ) ], ) ) if len(function_calls) > 0: self.response_messages.extend(function_calls) return self def make_response_output_items_from_parsable_context( self, ) -> list[ResponseOutputItem]: """Given a list of sentences, construct ResponseOutput Items.""" response_messages = self.response_messages[self.num_init_messages :] output_messages: list[ResponseOutputItem] = [] for message in response_messages: if not isinstance(message, ResponseFunctionToolCallOutputItem): output_messages.append(message) else: if len(output_messages) == 0: raise ValueError( "Cannot have a FunctionToolCallOutput before FunctionToolCall." ) if isinstance(output_messages[-1], ResponseFunctionToolCall): mcp_message = McpCall( id=f"{MCP_PREFIX}{random_uuid()}", arguments=output_messages[-1].arguments, name=output_messages[-1].name, server_label=output_messages[ -1 ].name, # TODO: store the server label type="mcp_call", status="completed", output=message.output, # TODO: support error output ) output_messages[-1] = mcp_message return output_messages def get_responses_parser_for_simple_context( *, tokenizer: TokenizerLike, reasoning_parser_cls: Callable[[TokenizerLike], ReasoningParser], response_messages: list[ResponseInputOutputItem], request: ResponsesRequest, tool_parser_cls, ) -> ResponsesParser: """Factory function to create a ResponsesParser with optional reasoning parser. Returns: ResponsesParser instance configured with the provided parser """ return ResponsesParser( tokenizer=tokenizer, reasoning_parser_cls=reasoning_parser_cls, response_messages=response_messages, request=request, tool_parser_cls=tool_parser_cls, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/openai/parser/responses_parser.py", "license": "Apache License 2.0", "lines": 161, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/mistral_large_3.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Iterable import regex as re import torch from vllm.model_executor.models.deepseek_v2 import DeepseekV3ForCausalLM class MistralLarge3ForCausalLM(DeepseekV3ForCausalLM): # fmt: off remapping = { r"layers\.(\d+)\.attention_norm\.weight": r"model.layers.\1.input_layernorm.weight", # noqa: E501 r"layers\.(\d+)\.attention\.wq_a\.(\w+)": r"model.layers.\1.self_attn.q_a_proj.\2", # noqa: E501 r"layers\.(\d+)\.attention\.q_a_norm\.weight": r"model.layers.\1.self_attn.q_a_layernorm.weight", # noqa: E501 r"layers\.(\d+)\.attention\.wq_b\.(\w+)": r"model.layers.\1.self_attn.q_b_proj.\2", # noqa: E501 r"layers\.(\d+)\.attention\.wkv_a_with_mqa\.(\w+)": r"model.layers.\1.self_attn.kv_a_proj_with_mqa.\2", # noqa: E501 r"layers\.(\d+)\.attention\.kv_a_norm\.weight": r"model.layers.\1.self_attn.kv_a_layernorm.weight", # noqa: E501 r"layers\.(\d+)\.attention\.wkv_b\.(\w+)": r"model.layers.\1.self_attn.kv_b_proj.\2", # noqa: E501 r"layers\.(\d+)\.attention\.wo\.(\w+)": r"model.layers.\1.self_attn.o_proj.\2", # noqa: E501 r"layers\.(\d+)\.ffn_norm\.weight": r"model.layers.\1.post_attention_layernorm.weight", # noqa: E501 r"layers\.(\d+)\.feed_forward\.w1\.(\w+)": r"model.layers.\1.mlp.gate_proj.\2", # noqa: E501 r"layers\.(\d+)\.feed_forward\.w2\.(\w+)": r"model.layers.\1.mlp.down_proj.\2", # noqa: E501 r"layers\.(\d+)\.feed_forward\.w3\.(\w+)": r"model.layers.\1.mlp.up_proj.\2", # noqa: E501 r"layers\.(\d+)\.gate\.weight": r"model.layers.\1.mlp.gate.weight", # noqa: E501 r"layers\.(\d+)\.shared_experts\.w1\.(\w+)": r"model.layers.\1.mlp.shared_experts.gate_proj.\2", # noqa: E501 r"layers\.(\d+)\.shared_experts\.w2\.(\w+)": r"model.layers.\1.mlp.shared_experts.down_proj.\2", # noqa: E501 r"layers\.(\d+)\.shared_experts\.w3\.(\w+)": r"model.layers.\1.mlp.shared_experts.up_proj.\2", # noqa: E501 r"layers\.(\d+)\.experts\.(\d+)\.w1\.(\w+)": r"model.layers.\1.mlp.experts.\2.gate_proj.\3", # noqa: E501 r"layers\.(\d+)\.experts\.(\d+)\.w2\.(\w+)": r"model.layers.\1.mlp.experts.\2.down_proj.\3", # noqa: E501 r"layers\.(\d+)\.experts\.(\d+)\.w3\.(\w+)": r"model.layers.\1.mlp.experts.\2.up_proj.\3", # noqa: E501 r"norm\.weight": "model.norm.weight", # noqa: E501 r"tok_embeddings\.weight": "model.embed_tokens.weight", # noqa: E501 r"output\.weight": "lm_head.weight", # noqa: E501 } # fmt: on def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: return super().load_weights(map(self._remap_mistral_to_ds, weights)) def _remap_mistral_to_ds( self, weight: tuple[str, torch.Tensor] ) -> tuple[str, torch.Tensor]: """Remap Mistral parameters to DeepseekV2 parameters.""" name, loaded_weight = weight for k, v in self.remapping.items(): match = re.fullmatch(k, name) if match: name = re.sub(k, v, name) break else: raise ValueError(f"Cannot remap {name}") # Remapping scale names. We could do this in the regex above but it # would triple the number of lines for most layers. if name.endswith(".qscale_act"): name = re.sub(r"\.qscale_act$", ".input_scale", name) elif name.endswith(".qscale_weight"): name = re.sub(r"\.qscale_weight$", ".weight_scale", name) return name, loaded_weight
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/mistral_large_3.py", "license": "Apache License 2.0", "lines": 54, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/mistral_large_3_eagle.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Iterable from functools import partial import torch import torch.nn as nn from vllm.compilation.decorators import support_torch_compile from vllm.config import VllmConfig from vllm.distributed.parallel_state import get_pp_group from vllm.logger import init_logger from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import RowParallelLinear from vllm.model_executor.layers.vocab_parallel_embedding import VocabParallelEmbedding from vllm.model_executor.models.deepseek_v2 import ( DeepseekV2DecoderLayer, DeepseekV2Model, ) from vllm.model_executor.models.mistral_large_3 import MistralLarge3ForCausalLM from .interfaces import SupportsMultiModal from .utils import make_empty_intermediate_tensors_factory, maybe_prefix logger = init_logger(__name__) @support_torch_compile class EagleMistralLarge3Model(DeepseekV2Model): def __init__( self, *, vllm_config: VllmConfig, prefix: str = "", start_layer_id: int = 0 ): nn.Module.__init__(self) config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config self.config = config self.vllm_config = vllm_config self.vocab_size = config.vocab_size assert get_pp_group().world_size == 1 self.embed_tokens = VocabParallelEmbedding( config.vocab_size, config.hidden_size, quant_config=quant_config, prefix=f"{prefix}.embed_tokens", ) self.layers = nn.ModuleList( [ DeepseekV2DecoderLayer( vllm_config=vllm_config, prefix=maybe_prefix(prefix, f"layers.{i + start_layer_id}"), ) for i in range(self.config.num_hidden_layers) ] ) self.start_layer = 0 self.end_layer = self.config.num_hidden_layers self.fc = RowParallelLinear( self.config.hidden_size * 2, self.config.hidden_size, bias=False, input_is_parallel=False, quant_config=quant_config, return_bias=False, prefix=maybe_prefix(prefix, "fc"), ) self.norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.make_empty_intermediate_tensors = make_empty_intermediate_tensors_factory( ["hidden_states", "residual"], config.hidden_size ) def forward( self, input_ids: torch.Tensor, positions: torch.Tensor, hidden_states: torch.Tensor, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor: if inputs_embeds is None: inputs_embeds = self.embed_input_ids(input_ids) inputs_embeds = self.fc(torch.cat((inputs_embeds, hidden_states), dim=-1)) output = super().forward( input_ids, positions, intermediate_tensors=None, inputs_embeds=inputs_embeds ) assert isinstance(output, torch.Tensor) return output class EagleMistralLarge3ForCausalLM(MistralLarge3ForCausalLM): remapping = MistralLarge3ForCausalLM.remapping | { r"eagle_linear\.weight": r"model.fc.weight", r"eagle_linear\.qscale_act": r"model.fc.input_scale", r"eagle_linear\.qscale_weight": r"model.fc.weight_scale", } def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): target_layer_num = vllm_config.model_config.get_num_layers( vllm_config.parallel_config ) vllm_config.model_config = vllm_config.speculative_config.draft_model_config # draft model quantization config may differ from target model self.quant_config = VllmConfig.get_quantization_config( vllm_config.speculative_config.draft_model_config, vllm_config.load_config ) vllm_config.quant_config = self.quant_config self.model_cls = partial( EagleMistralLarge3Model, start_layer_id=target_layer_num ) super().__init__(vllm_config=vllm_config, prefix=prefix) def get_language_model(self) -> torch.nn.Module: return self.model embed_input_ids = SupportsMultiModal.embed_input_ids # type: ignore def forward( self, input_ids: torch.Tensor, positions: torch.Tensor, hidden_states: torch.Tensor, inputs_embeds: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: hidden_states = self.model(input_ids, positions, hidden_states, inputs_embeds) return hidden_states, hidden_states def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: # Pretend we've loaded the embedding and lm_head weights # (later copied from target model) return super().load_weights(weights) | { "model.embed_tokens.weight", "lm_head.weight", }
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/mistral_large_3_eagle.py", "license": "Apache License 2.0", "lines": 117, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:.buildkite/scripts/generate-nightly-index.py
#!/usr/bin/env python3 # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # do not complain about line length (for docstring) # ruff: noqa: E501 import argparse import json import sys from dataclasses import asdict, dataclass from datetime import datetime from pathlib import Path from typing import Any from urllib.parse import quote import regex as re def normalize_package_name(name: str) -> str: """ Normalize package name according to PEP 503. https://peps.python.org/pep-0503/#normalized-names Replace runs of underscores, hyphens, and periods with a single hyphen, and lowercase the result. """ return re.sub(r"[-_.]+", "-", name).lower() if not sys.version_info >= (3, 12): raise RuntimeError("This script requires Python 3.12 or higher.") INDEX_HTML_TEMPLATE = """<!DOCTYPE html> <html> <!-- {comment} --> <meta name="pypi:repository-version" content="1.0"> <body> {items} </body> </html> """ @dataclass class WheelFileInfo: package_name: str version: str build_tag: str | None python_tag: str abi_tag: str platform_tag: str variant: str | None filename: str def parse_from_filename(file: str) -> WheelFileInfo: """ Parse wheel file name to extract metadata. The format of wheel names: {package_name}-{version}(-{build_tag})?-{python_tag}-{abi_tag}-{platform_tag}.whl All versions could contain a variant like '+cu129' or '.cpu' or `.rocm` (or not). Example: vllm-0.11.0-cp38-abi3-manylinux1_x86_64.whl vllm-0.10.2rc2+cu129-cp38-abi3-manylinux2014_aarch64.whl vllm-0.11.1rc8.dev14+gaa384b3c0-cp38-abi3-manylinux2014_aarch64.whl vllm-0.11.1rc8.dev14+gaa384b3c0.cu130-cp38-abi3-manylinux1_x86_64.whl """ wheel_file_re = re.compile( r"^(?P<package_name>.+)-(?P<version>[^-]+?)(-(?P<build_tag>[^-]+))?-(?P<python_tag>[^-]+)-(?P<abi_tag>[^-]+)-(?P<platform_tag>[^-]+)\.whl$" ) match = wheel_file_re.match(file) if not match: raise ValueError(f"Invalid wheel file name: {file}") package_name = match.group("package_name") version = match.group("version") build_tag = match.group("build_tag") python_tag = match.group("python_tag") abi_tag = match.group("abi_tag") platform_tag = match.group("platform_tag") # extract variant from version variant = None if "dev" in version: ver_after_dev = version.split("dev")[-1] if "." in ver_after_dev: variant = ver_after_dev.split(".")[-1] version = version.removesuffix("." + variant) else: if "+" in version: version_part, suffix = version.split("+", 1) # Only treat known patterns as variants (rocmXXX, cuXXX, cpu) # Git hashes and other suffixes are NOT variants if suffix.startswith(("rocm", "cu", "cpu")): variant = suffix version = version_part # Otherwise keep the full version string (variant stays None) return WheelFileInfo( package_name=package_name, version=version, build_tag=build_tag, python_tag=python_tag, abi_tag=abi_tag, platform_tag=platform_tag, variant=variant, filename=file, ) def generate_project_list(subdir_names: list[str], comment: str = "") -> str: """ Generate project list HTML content linking to each project & variant subdirectory. """ href_tags = [] for name in sorted(subdir_names): name = name.strip("/").strip(".") href_tags.append(f' <a href="{name}/">{name}/</a><br/>') return INDEX_HTML_TEMPLATE.format(items="\n".join(href_tags), comment=comment) def generate_package_index_and_metadata( wheel_files: list[WheelFileInfo], wheel_base_dir: Path, index_base_dir: Path, comment: str = "", ) -> tuple[str, str]: """ Generate package index HTML content for a specific package, linking to actual wheel files. """ href_tags = [] metadata = [] for file in sorted(wheel_files, key=lambda x: x.filename): relative_path = ( wheel_base_dir.relative_to(index_base_dir, walk_up=True) / file.filename ) # handle with '+' in URL, and avoid double-encoding '/' and already-encoded '%2B' # NOTE: this is AWS S3 specific behavior! file_path_quoted = quote(relative_path.as_posix(), safe=":%/") href_tags.append(f' <a href="{file_path_quoted}">{file.filename}</a><br/>') file_meta = asdict(file) file_meta["path"] = file_path_quoted metadata.append(file_meta) index_str = INDEX_HTML_TEMPLATE.format(items="\n".join(href_tags), comment=comment) metadata_str = json.dumps(metadata, indent=2) return index_str, metadata_str def generate_index_and_metadata( whl_files: list[str], wheel_base_dir: Path, index_base_dir: Path, default_variant: str | None = None, alias_to_default: str | None = None, comment: str = "", ): """ Generate index for all wheel files. Args: whl_files (list[str]): List of wheel files (must be directly under `wheel_base_dir`). wheel_base_dir (Path): Base directory for wheel files. index_base_dir (Path): Base directory to store index files. default_variant (str | None): The default variant name, if any. alias_to_default (str | None): Alias variant name for the default variant, if any. comment (str | None): Optional comment to include in the generated HTML files. First, parse all wheel files to extract metadata. We need to collect all wheel files for each variant, and generate an index for it (in a subdirectory). The index for the default variant (if any) is generated in the root index directory. If `default_variant` is provided, all wheels must have variant suffixes, and the default variant index is purely a copy of the corresponding variant index, with only the links adjusted. Otherwise, all wheels without variant suffixes are treated as the default variant. If `alias_to_default` is provided, an additional alias subdirectory is created, it has the same content as the default variant index, but the links are adjusted accordingly. Index directory structure: index_base_dir/ (hosted at wheels.vllm.ai/{nightly,$commit,$version}/) index.html # project list, linking to "vllm/" and other packages, and all variant subdirectories vllm/ index.html # package index, pointing to actual files in wheel_base_dir (relative path) metadata.json # machine-readable metadata for all wheels in this package cpu/ # cpu variant subdirectory index.html vllm/ index.html metadata.json cu129/ # cu129 is actually the alias to default variant index.html vllm/ index.html metadata.json cu130/ # cu130 variant subdirectory index.html vllm/ index.html metadata.json ... metadata.json stores a dump of all wheel files' metadata in a machine-readable format: [ { "package_name": "vllm", "version": "0.10.2rc2", "build_tag": null, "python_tag": "cp38", "abi_tag": "abi3", "platform_tag": "manylinux2014_aarch64", "variant": "cu129", "filename": "vllm-0.10.2rc2+cu129-cp38-abi3-manylinux2014_aarch64.whl", "path": "../vllm-0.10.2rc2%2Bcu129-cp38-abi3-manylinux2014_aarch64.whl" # to be concatenated with the directory URL and URL-encoded }, ... ] """ parsed_files = [parse_from_filename(f) for f in whl_files] if not parsed_files: print("No wheel files found, skipping index generation.") return # For ROCm builds: inherit variant from vllm wheel # All ROCm wheels should share the same variant as vllm rocm_variant = None for file in parsed_files: if ( file.package_name == "vllm" and file.variant and file.variant.startswith("rocm") ): rocm_variant = file.variant print(f"Detected ROCm variant from vllm: {rocm_variant}") break # Apply ROCm variant to all wheels without a variant if rocm_variant: for file in parsed_files: if file.variant is None: file.variant = rocm_variant print(f"Inherited variant '{rocm_variant}' for {file.filename}") # Group by variant variant_to_files: dict[str, list[WheelFileInfo]] = {} for file in parsed_files: variant = file.variant or "default" if variant not in variant_to_files: variant_to_files[variant] = [] variant_to_files[variant].append(file) print(f"Found variants: {list(variant_to_files.keys())}") # sanity check for default variant if default_variant: if "default" in variant_to_files: raise ValueError( "All wheel files must have variant suffixes when `default_variant` is specified." ) if default_variant not in variant_to_files: raise ValueError( f"Default variant '{default_variant}' not found among wheel files." ) if alias_to_default: if "default" not in variant_to_files: # e.g. only some wheels are uploaded to S3 currently print( "[WARN] Alias to default variant specified, but no default variant found." ) elif alias_to_default in variant_to_files: raise ValueError( f"Alias variant name '{alias_to_default}' already exists among wheel files." ) else: variant_to_files[alias_to_default] = variant_to_files["default"].copy() print(f"Alias variant '{alias_to_default}' created for default variant.") # Generate comment in HTML header comment_str = f" ({comment})" if comment else "" comment_tmpl = f"Generated on {datetime.now().isoformat()}{comment_str}" # Generate index for each variant subdir_names = set() for variant, files in variant_to_files.items(): if variant == "default": variant_dir = index_base_dir else: variant_dir = index_base_dir / variant subdir_names.add(variant) variant_dir.mkdir(parents=True, exist_ok=True) # gather all package names in this variant (normalized per PEP 503) packages = set(normalize_package_name(f.package_name) for f in files) if variant == "default": # these packages should also appear in the "project list" # generate after all variants are processed subdir_names = subdir_names.union(packages) else: # generate project list for this variant directly project_list_str = generate_project_list(sorted(packages), comment_tmpl) with open(variant_dir / "index.html", "w") as f: f.write(project_list_str) for package in packages: # filter files belonging to this package only (compare normalized names) package_files = [ f for f in files if normalize_package_name(f.package_name) == package ] package_dir = variant_dir / package package_dir.mkdir(parents=True, exist_ok=True) index_str, metadata_str = generate_package_index_and_metadata( package_files, wheel_base_dir, package_dir, comment ) with open(package_dir / "index.html", "w") as f: f.write(index_str) with open(package_dir / "metadata.json", "w") as f: f.write(metadata_str) # Generate top-level project list index project_list_str = generate_project_list(sorted(subdir_names), comment_tmpl) with open(index_base_dir / "index.html", "w") as f: f.write(project_list_str) if __name__ == "__main__": """ Arguments: --version <version> : version string for the current build (e.g., commit hash) --wheel-dir <wheel_directory> : directory containing wheel files (default to be same as `version`) --current-objects <path_to_json> : path to JSON file containing current S3 objects listing in this version directory --output-dir <output_directory> : directory to store generated index files --alias-to-default <alias_variant_name> : (optional) alias variant name for the default variant --comment <comment_string> : (optional) comment string to include in generated HTML files """ parser = argparse.ArgumentParser( description="Process nightly build wheel files to generate indices." ) parser.add_argument( "--version", type=str, required=True, help="Version string for the current build (e.g., commit hash)", ) parser.add_argument( "--current-objects", type=str, required=True, help="Path to JSON file containing current S3 objects listing in this version directory", ) parser.add_argument( "--output-dir", type=str, required=True, help="Directory to store generated index files", ) parser.add_argument( "--wheel-dir", type=str, default=None, help="Directory containing wheel files (default to be same as `version`)", ) parser.add_argument( "--alias-to-default", type=str, default=None, help="Alias variant name for the default variant", ) parser.add_argument( "--comment", type=str, default="", help="Optional comment string to include in generated HTML files", ) args = parser.parse_args() version = args.version # Allow rocm/ prefix, reject other slashes and all backslashes if "\\" in version: raise ValueError("Version string must not contain backslashes.") if "/" in version and not version.startswith("rocm/"): raise ValueError( "Version string must not contain slashes (except for 'rocm/' prefix)." ) current_objects_path = Path(args.current_objects) output_dir = Path(args.output_dir) if not output_dir.exists(): output_dir.mkdir(parents=True, exist_ok=True) # Read current objects JSON with open(current_objects_path) as f: current_objects: dict[str, list[dict[str, Any]]] = json.load(f) # current_objects looks like from list_objects_v2 S3 API: """ "Contents": [ { "Key": "e2f56c309d2a28899c68975a7e104502d56deb8f/vllm-0.11.2.dev363+ge2f56c309-cp38-abi3-manylinux1_x86_64.whl", "LastModified": "2025-11-28T14:00:32+00:00", "ETag": "\"37a38339c7cdb61ca737021b968075df-52\"", "ChecksumAlgorithm": [ "CRC64NVME" ], "ChecksumType": "FULL_OBJECT", "Size": 435649349, "StorageClass": "STANDARD" }, ... ] """ # Extract wheel file keys wheel_files = [] for item in current_objects.get("Contents", []): key: str = item["Key"] if key.endswith(".whl"): wheel_files.append(key.split("/")[-1]) # only the filename is used print(f"Found {len(wheel_files)} wheel files for version {version}: {wheel_files}") # keep only "official" files for a non-nightly version (specified by cli args) PY_VERSION_RE = re.compile(r"^\d+\.\d+\.\d+([a-zA-Z0-9.+-]*)?$") if PY_VERSION_RE.match(version): # upload-wheels.sh ensures no "dev" is in args.version wheel_files = list( filter(lambda x: version in x and "dev" not in x, wheel_files) ) print(f"Non-nightly version detected, wheel files used: {wheel_files}") else: print("Nightly version detected, keeping all wheel files.") # Generate index and metadata, assuming wheels and indices are stored as: # s3://vllm-wheels/{wheel_dir}/<wheel files> # s3://vllm-wheels/<anything>/<index files> # # For ROCm builds, version is "rocm/{commit}" and indices are uploaded to: # - rocm/{commit}/ (same as wheels) # - rocm/nightly/ # - rocm/{version}/ # All these are under the "rocm/" prefix, so relative paths should be # relative to "rocm/", not the bucket root. if args.wheel_dir: # Explicit wheel-dir provided (e.g., for version-specific indices pointing to commit dir) wheel_dir = args.wheel_dir.strip().rstrip("/") elif version.startswith("rocm/"): # For rocm/commit, wheel_base_dir should be just the commit part # so relative path from rocm/0.12.0/rocm710/vllm/ -> ../../../{commit}/ wheel_dir = version.split("/", 1)[1] else: wheel_dir = version wheel_base_dir = Path(output_dir).parent / wheel_dir index_base_dir = Path(output_dir) generate_index_and_metadata( whl_files=wheel_files, wheel_base_dir=wheel_base_dir, index_base_dir=index_base_dir, default_variant=None, alias_to_default=args.alias_to_default, comment=args.comment.strip(), ) print(f"Successfully generated index and metadata in {output_dir}")
{ "repo_id": "vllm-project/vllm", "file_path": ".buildkite/scripts/generate-nightly-index.py", "license": "Apache License 2.0", "lines": 413, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:examples/offline_inference/llm_engine_reset_kv.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ This file demonstrates preempt requests when using the `LLMEngine` for processing prompts with various sampling parameters. """ import argparse from vllm import EngineArgs, LLMEngine, RequestOutput, SamplingParams from vllm.utils.argparse_utils import FlexibleArgumentParser def create_test_prompts() -> list[tuple[str, SamplingParams]]: """Create a list of test prompts with their sampling parameters.""" return [ ( "A robot may not injure a human being " * 50, SamplingParams( temperature=0.0, logprobs=1, prompt_logprobs=1, max_tokens=16 ), ), ( "A robot may not injure a human being " * 50, SamplingParams( temperature=0.0, logprobs=1, prompt_logprobs=1, max_tokens=16 ), ), ( "To be or not to be,", SamplingParams( temperature=0.8, top_k=5, presence_penalty=0.2, max_tokens=128 ), ), ( "What is the meaning of life?", SamplingParams( n=2, temperature=0.8, top_p=0.95, frequency_penalty=0.1, max_tokens=128 ), ), ] def process_requests(engine: LLMEngine, test_prompts: list[tuple[str, SamplingParams]]): """Continuously process a list of prompts and handle the outputs.""" request_id = 0 print("-" * 50) step_id = 0 while test_prompts or engine.has_unfinished_requests(): print("-" * 50) import os print(f"Step {step_id} (pid={os.getpid()})") if test_prompts: prompt, sampling_params = test_prompts.pop(0) engine.add_request(str(request_id), prompt, sampling_params) request_id += 1 if step_id == 10: print(f"Resetting prefix cache at {step_id}") engine.reset_prefix_cache(reset_running_requests=True) request_outputs: list[RequestOutput] = engine.step() for request_output in request_outputs: if request_output.finished: print("-" * 50) print(request_output) print("-" * 50) step_id += 1 def initialize_engine(args: argparse.Namespace) -> LLMEngine: """Initialize the LLMEngine from the command line arguments.""" engine_args = EngineArgs.from_cli_args(args) return LLMEngine.from_engine_args(engine_args) def parse_args(): parser = FlexibleArgumentParser( description="Demo on using the LLMEngine class directly" ) parser = EngineArgs.add_cli_args(parser) return parser.parse_args() def main(args: argparse.Namespace): """Main function that sets up and runs the prompt processing.""" engine = initialize_engine(args) test_prompts = create_test_prompts() process_requests(engine, test_prompts) if __name__ == "__main__": args = parse_args() main(args)
{ "repo_id": "vllm-project/vllm", "file_path": "examples/offline_inference/llm_engine_reset_kv.py", "license": "Apache License 2.0", "lines": 78, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/core/test_reset_prefix_cache_e2e.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from vllm import EngineArgs, LLMEngine, SamplingParams PROMPTS = [ "A robot may not injure a human being ", "To be or not to be,", "What is the meaning of life?", "What does the fox say? " * 20, # Test long prompt ] def test_reset_prefix_cache_e2e(monkeypatch): # "spawn" is required for test to be deterministic monkeypatch.setenv("VLLM_WORKER_MULTIPROC_METHOD", "spawn") engine_args = EngineArgs( model="Qwen/Qwen3-0.6B", gpu_memory_utilization=0.2, async_scheduling=True, max_num_batched_tokens=32, max_model_len=2048, compilation_config={"mode": 0}, dtype="float16", ) engine = LLMEngine.from_engine_args(engine_args) sampling_params = SamplingParams( temperature=0.0, max_tokens=16, ) # No preempt case: for i, prompt in enumerate(PROMPTS): engine.add_request("ground_truth_" + str(i), prompt, sampling_params) ground_truth_results = {} while engine.has_unfinished_requests(): request_outputs = engine.step() for request_output in request_outputs: if request_output.finished: ground_truth_results[request_output.request_id] = request_output # Preempt case: for i, prompt in enumerate(PROMPTS): engine.add_request("preempted_" + str(i), prompt, sampling_params) step_id = 0 preempted_results = {} while engine.has_unfinished_requests(): if step_id == 10: engine.reset_prefix_cache(reset_running_requests=True) request_outputs = engine.step() for request_output in request_outputs: if request_output.finished: preempted_results[request_output.request_id] = request_output step_id += 1 for i in range(len(PROMPTS)): assert ( ground_truth_results["ground_truth_" + str(i)].outputs[0].text == preempted_results["preempted_" + str(i)].outputs[0].text ), ( f"ground_truth_results['ground_truth_{i}'].outputs[0].text=" f"{ground_truth_results['ground_truth_' + str(i)].outputs[0].text} " f"preempted_results['preempted_{i}'].outputs[0].text=" f"{preempted_results['preempted_' + str(i)].outputs[0].text}" )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/core/test_reset_prefix_cache_e2e.py", "license": "Apache License 2.0", "lines": 58, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/v1/core/kv_cache_metrics.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """KV cache metrics tracking.""" import random import time from collections import deque from typing import TYPE_CHECKING if TYPE_CHECKING: from vllm.v1.core.kv_cache_utils import KVCacheBlock from vllm.v1.metrics.stats import KVCacheEvictionEvent class BlockMetricsState: """Tracks lifecycle metrics for a single KV cache block.""" def __init__(self): now_ns = time.monotonic_ns() self.birth_time_ns = now_ns self.last_access_ns = now_ns # Bounded to prevent unbounded growth if a block is accessed many times. self.access_history: deque[int] = deque(maxlen=4) def record_access(self) -> None: now_ns = time.monotonic_ns() self.last_access_ns = now_ns self.access_history.append(now_ns) def get_lifetime_seconds(self) -> float: now_ns = time.monotonic_ns() return (now_ns - self.birth_time_ns) / 1e9 def get_idle_time_seconds(self) -> float: now_ns = time.monotonic_ns() return (now_ns - self.last_access_ns) / 1e9 def get_reuse_gaps_seconds(self) -> list[float]: if len(self.access_history) < 2: return [] history = list(self.access_history) return [(history[i] - history[i - 1]) / 1e9 for i in range(1, len(history))] class KVCacheMetricsCollector: """Collects KV cache residency metrics with sampling.""" def __init__(self, sample_rate: float = 0.01): assert 0 < sample_rate <= 1.0, ( f"sample_rate must be in (0, 1.0], got {sample_rate}" ) self.sample_rate = sample_rate self.block_metrics: dict[int, BlockMetricsState] = {} self._eviction_events: list[KVCacheEvictionEvent] = [] def should_sample_block(self) -> bool: return random.random() < self.sample_rate def on_block_allocated(self, block: "KVCacheBlock") -> None: if self.should_sample_block(): self.block_metrics[block.block_id] = BlockMetricsState() def on_block_accessed(self, block: "KVCacheBlock") -> None: metrics = self.block_metrics.get(block.block_id) if metrics: metrics.record_access() def on_block_evicted(self, block: "KVCacheBlock") -> None: metrics = self.block_metrics.pop(block.block_id, None) if not metrics: return lifetime = metrics.get_lifetime_seconds() idle_time = metrics.get_idle_time_seconds() reuse_gaps = tuple(metrics.get_reuse_gaps_seconds()) self._eviction_events.append( KVCacheEvictionEvent( lifetime_seconds=lifetime, idle_seconds=idle_time, reuse_gaps_seconds=reuse_gaps, ) ) def reset(self) -> None: """Clear all state on cache reset.""" self.block_metrics.clear() self._eviction_events.clear() def drain_events(self) -> list[KVCacheEvictionEvent]: events = self._eviction_events self._eviction_events = [] return events
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/core/kv_cache_metrics.py", "license": "Apache License 2.0", "lines": 73, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:examples/online_serving/openai_responses_client_with_mcp_tools.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Example demonstrating MCP (Model Context Protocol) tools with the Responses API. This example shows how to use MCP tools with different allowed_tools configurations: 1. No filter (allows all tools from the MCP server) 2. Wildcard "*" (explicitly allows all tools) 3. Specific tool names (filters to only those tools) Set up this example by starting a vLLM OpenAI-compatible server with MCP tools enabled. For example: vllm serve openai/gpt-oss-20b --enforce-eager --tool-server demo Environment variables: - VLLM_ENABLE_RESPONSES_API_STORE=1 - VLLM_GPT_OSS_SYSTEM_TOOL_MCP_LABELS=code_interpreter,container - VLLM_GPT_OSS_HARMONY_SYSTEM_INSTRUCTIONS=1 """ from openai import OpenAI from utils import get_first_model def example_no_filter(): """Example with no allowed_tools filter - allows all tools.""" print("=" * 60) print("Example 1: No allowed_tools filter (allows all tools)") print("=" * 60) base_url = "http://0.0.0.0:8000/v1" client = OpenAI(base_url=base_url, api_key="empty") model = get_first_model(client) response = client.responses.create( model=model, input="Execute this code: print('Hello from Python!')", instructions="Use the Python tool to execute code.", tools=[ { "type": "mcp", "server_label": "code_interpreter", "server_url": "http://localhost:8888", # No allowed_tools specified - all tools are available } ], ) print(f"Status: {response.status}") print(f"Output: {response.output_text}") print() def example_wildcard(): """Example with allowed_tools=['*'] - explicitly allows all tools.""" print("=" * 60) print("Example 2: allowed_tools=['*'] (select all tools)") print("=" * 60) base_url = "http://0.0.0.0:8000/v1" client = OpenAI(base_url=base_url, api_key="empty") model = get_first_model(client) response = client.responses.create( model=model, input="Execute this code: print('Hello from Python with wildcard!')", instructions="Use the Python tool to execute code.", tools=[ { "type": "mcp", "server_label": "code_interpreter", "server_url": "http://localhost:8888", # Using "*" to explicitly allow all tools from this MCP server # This is equivalent to not specifying allowed_tools "allowed_tools": ["*"], } ], ) print(f"Status: {response.status}") print(f"Output: {response.output_text}") print() def example_specific_tools(): """Example with specific allowed_tools list - filters available tools. Note: This example uses 'web_search_preview' (browser) which has multiple sub-tools: 'search', 'open', 'find'. The code_interpreter (python) doesn't have sub-tools, so filtering doesn't apply there. """ print("=" * 60) print("Example 3: allowed_tools=['search'] (filter browser to specific tools)") print("=" * 60) base_url = "http://0.0.0.0:8000/v1" client = OpenAI(base_url=base_url, api_key="empty") model = get_first_model(client) response = client.responses.create( model=model, input="Search for 'Python programming tutorials'", instructions="Use the browser tool to search.", tools=[ { "type": "mcp", "server_label": "web_search_preview", "server_url": "http://localhost:8888", # Browser has tools: 'search', 'open', 'find' # Only allow 'search' - blocks 'open' and 'find' "allowed_tools": ["search"], } ], ) print(f"Status: {response.status}") print(f"Output: {response.output_text}") print() def example_object_format(): """Example using object format for allowed_tools with browser tools.""" print("=" * 60) print("Example 4: allowed_tools with object format") print("=" * 60) base_url = "http://0.0.0.0:8000/v1" client = OpenAI(base_url=base_url, api_key="empty") model = get_first_model(client) response = client.responses.create( model=model, input="Search for 'machine learning' and open the first result", instructions="Use the browser tool.", tools=[ { "type": "mcp", "server_label": "web_search_preview", "server_url": "http://localhost:8888", # Object format with tool_names field # Can also include read_only and other fields # Browser has tools: 'search', 'open', 'find' "allowed_tools": { "tool_names": [ "search", "open", ], # Allow search and open, block find "read_only": False, }, } ], ) print(f"Status: {response.status}") print(f"Output: {response.output_text}") print() def main(): """Run all examples.""" print("\n" + "=" * 60) print("MCP Tools with allowed_tools Examples") print("=" * 60 + "\n") # Run all examples example_no_filter() example_wildcard() example_specific_tools() example_object_format() print("=" * 60) print("Summary:") print(" - No filter or '*' → All tools available from server") print(" - Specific list → Only those sub-tools available") print(" - Object format → More control with tool_names field") print("") print("Note: allowed_tools filters SUB-TOOLS within an MCP server:") print(" - code_interpreter (python): No sub-tools to filter") print(" - web_search_preview (browser): Has 'search', 'open', 'find'") print("=" * 60) if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "examples/online_serving/openai_responses_client_with_mcp_tools.py", "license": "Apache License 2.0", "lines": 153, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/pooling/classify/api_router.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from fastapi import APIRouter, Depends, Request from starlette.responses import JSONResponse from typing_extensions import assert_never from vllm.entrypoints.openai.engine.protocol import ErrorResponse from vllm.entrypoints.openai.utils import validate_json_request from vllm.entrypoints.pooling.classify.protocol import ( ClassificationRequest, ClassificationResponse, ) from vllm.entrypoints.pooling.classify.serving import ServingClassification from vllm.entrypoints.utils import load_aware_call, with_cancellation router = APIRouter() def classify(request: Request) -> ServingClassification | None: return request.app.state.openai_serving_classification @router.post("/classify", dependencies=[Depends(validate_json_request)]) @with_cancellation @load_aware_call async def create_classify(request: ClassificationRequest, raw_request: Request): handler = classify(raw_request) if handler is None: base_server = raw_request.app.state.openai_serving_tokenization return base_server.create_error_response( message="The model does not support Classification API" ) try: generator = await handler.create_classify(request, raw_request) except Exception as e: generator = handler.create_error_response(e) if isinstance(generator, ErrorResponse): return JSONResponse( content=generator.model_dump(), status_code=generator.error.code ) elif isinstance(generator, ClassificationResponse): return JSONResponse(content=generator.model_dump()) assert_never(generator)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/pooling/classify/api_router.py", "license": "Apache License 2.0", "lines": 37, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/pooling/classify/protocol.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import time from typing import TypeAlias from pydantic import Field from vllm import PoolingParams from vllm.config import ModelConfig from vllm.entrypoints.openai.engine.protocol import OpenAIBaseModel, UsageInfo from vllm.entrypoints.pooling.base.protocol import ( ChatRequestMixin, ClassifyRequestMixin, CompletionRequestMixin, PoolingBasicRequestMixin, ) from vllm.logger import init_logger from vllm.renderers import TokenizeParams from vllm.utils import random_uuid logger = init_logger(__name__) class ClassificationCompletionRequest( PoolingBasicRequestMixin, CompletionRequestMixin, ClassifyRequestMixin ): def build_tok_params(self, model_config: ModelConfig) -> TokenizeParams: encoder_config = model_config.encoder_config or {} return TokenizeParams( max_total_tokens=model_config.max_model_len, max_output_tokens=0, truncate_prompt_tokens=self.truncate_prompt_tokens, do_lower_case=encoder_config.get("do_lower_case", False), add_special_tokens=self.add_special_tokens, max_total_tokens_param="max_model_len", ) def to_pooling_params(self): return PoolingParams( task="classify", use_activation=self.use_activation, ) class ClassificationChatRequest( PoolingBasicRequestMixin, ChatRequestMixin, ClassifyRequestMixin ): def build_tok_params(self, model_config: ModelConfig) -> TokenizeParams: encoder_config = model_config.encoder_config or {} return TokenizeParams( max_total_tokens=model_config.max_model_len, max_output_tokens=0, truncate_prompt_tokens=self.truncate_prompt_tokens, do_lower_case=encoder_config.get("do_lower_case", False), add_special_tokens=self.add_special_tokens, max_total_tokens_param="max_model_len", ) def to_pooling_params(self): return PoolingParams( task="classify", use_activation=self.use_activation, ) ClassificationRequest: TypeAlias = ( ClassificationCompletionRequest | ClassificationChatRequest ) class ClassificationData(OpenAIBaseModel): index: int label: str | None probs: list[float] num_classes: int class ClassificationResponse(OpenAIBaseModel): id: str = Field(default_factory=lambda: f"classify-{random_uuid()}") object: str = "list" created: int = Field(default_factory=lambda: int(time.time())) model: str data: list[ClassificationData] usage: UsageInfo
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/pooling/classify/protocol.py", "license": "Apache License 2.0", "lines": 69, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/pooling/embed/api_router.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import importlib.util from functools import lru_cache from http import HTTPStatus from fastapi import APIRouter, Depends, Request from fastapi.responses import JSONResponse, StreamingResponse from typing_extensions import assert_never from vllm.entrypoints.openai.engine.protocol import ErrorResponse from vllm.entrypoints.openai.utils import validate_json_request from vllm.entrypoints.pooling.embed.protocol import ( EmbeddingBytesResponse, EmbeddingRequest, EmbeddingResponse, ) from vllm.entrypoints.pooling.embed.serving import OpenAIServingEmbedding from vllm.entrypoints.utils import load_aware_call, with_cancellation from vllm.logger import init_logger router = APIRouter() logger = init_logger(__name__) @lru_cache(maxsize=1) def _get_json_response_cls(): if importlib.util.find_spec("orjson") is not None: from fastapi.responses import ORJSONResponse return ORJSONResponse logger.warning_once( "To make v1/embeddings API fast, please install orjson by `pip install orjson`" ) return JSONResponse def embedding(request: Request) -> OpenAIServingEmbedding | None: return request.app.state.openai_serving_embedding @router.post( "/v1/embeddings", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, }, ) @with_cancellation @load_aware_call async def create_embedding( request: EmbeddingRequest, raw_request: Request, ): handler = embedding(raw_request) if handler is None: base_server = raw_request.app.state.openai_serving_tokenization return base_server.create_error_response( message="The model does not support Embeddings API" ) try: generator = await handler.create_embedding(request, raw_request) except Exception as e: generator = handler.create_error_response(e) if isinstance(generator, ErrorResponse): return JSONResponse( content=generator.model_dump(), status_code=generator.error.code ) elif isinstance(generator, EmbeddingResponse): return _get_json_response_cls()(content=generator.model_dump()) elif isinstance(generator, EmbeddingBytesResponse): return StreamingResponse( content=generator.content, headers=generator.headers, media_type=generator.media_type, ) assert_never(generator)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/pooling/embed/api_router.py", "license": "Apache License 2.0", "lines": 68, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/pooling/embed/protocol.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import time from typing import TypeAlias from pydantic import Field from vllm import PoolingParams from vllm.config import ModelConfig from vllm.entrypoints.openai.engine.protocol import OpenAIBaseModel, UsageInfo from vllm.entrypoints.pooling.base.protocol import ( ChatRequestMixin, CompletionRequestMixin, EmbedRequestMixin, PoolingBasicRequestMixin, ) from vllm.renderers import TokenizeParams from vllm.utils import random_uuid def _get_max_total_output_tokens( model_config: ModelConfig, ) -> tuple[int | None, int]: max_total_tokens = model_config.max_model_len pooler_config = model_config.pooler_config if pooler_config is None: return max_total_tokens, 0 if pooler_config.enable_chunked_processing: return None, 0 max_embed_len = pooler_config.max_embed_len or max_total_tokens max_output_tokens = max_total_tokens - max_embed_len return max_total_tokens, max_output_tokens class EmbeddingCompletionRequest( PoolingBasicRequestMixin, CompletionRequestMixin, EmbedRequestMixin ): def build_tok_params(self, model_config: ModelConfig) -> TokenizeParams: encoder_config = model_config.encoder_config or {} ( max_total_tokens, max_output_tokens, ) = _get_max_total_output_tokens(model_config) return TokenizeParams( max_total_tokens=max_total_tokens, max_output_tokens=max_output_tokens, truncate_prompt_tokens=self.truncate_prompt_tokens, do_lower_case=encoder_config.get("do_lower_case", False), add_special_tokens=self.add_special_tokens, max_total_tokens_param="max_model_len", max_output_tokens_param="max_model_len - max_embed_len", ) def to_pooling_params(self): return PoolingParams( task="embed", dimensions=self.dimensions, use_activation=self.use_activation, ) class EmbeddingChatRequest( PoolingBasicRequestMixin, ChatRequestMixin, EmbedRequestMixin ): def build_tok_params(self, model_config: ModelConfig) -> TokenizeParams: encoder_config = model_config.encoder_config or {} ( max_total_tokens, max_output_tokens, ) = _get_max_total_output_tokens(model_config) return TokenizeParams( max_total_tokens=max_total_tokens, max_output_tokens=max_output_tokens, truncate_prompt_tokens=self.truncate_prompt_tokens, do_lower_case=encoder_config.get("do_lower_case", False), add_special_tokens=self.add_special_tokens, max_total_tokens_param="max_model_len", max_output_tokens_param="max_model_len - max_embed_len", ) def to_pooling_params(self): return PoolingParams( task="embed", dimensions=self.dimensions, use_activation=self.use_activation, ) EmbeddingRequest: TypeAlias = EmbeddingCompletionRequest | EmbeddingChatRequest class EmbeddingResponseData(OpenAIBaseModel): index: int object: str = "embedding" embedding: list[float] | str class EmbeddingResponse(OpenAIBaseModel): id: str = Field(default_factory=lambda: f"embd-{random_uuid()}") object: str = "list" created: int = Field(default_factory=lambda: int(time.time())) model: str data: list[EmbeddingResponseData] usage: UsageInfo class EmbeddingBytesResponse(OpenAIBaseModel): content: list[bytes] headers: dict[str, str] | None = None media_type: str = "application/octet-stream"
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/pooling/embed/protocol.py", "license": "Apache License 2.0", "lines": 92, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/pooling/pooling/api_router.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from http import HTTPStatus from fastapi import APIRouter, Depends, Request from fastapi.responses import JSONResponse, StreamingResponse from typing_extensions import assert_never from vllm.entrypoints.openai.engine.protocol import ErrorResponse from vllm.entrypoints.openai.utils import validate_json_request from vllm.entrypoints.pooling.pooling.protocol import ( IOProcessorResponse, PoolingBytesResponse, PoolingRequest, PoolingResponse, ) from vllm.entrypoints.pooling.pooling.serving import OpenAIServingPooling from vllm.entrypoints.utils import load_aware_call, with_cancellation router = APIRouter() def pooling(request: Request) -> OpenAIServingPooling | None: return request.app.state.openai_serving_pooling @router.post( "/pooling", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, }, ) @with_cancellation @load_aware_call async def create_pooling(request: PoolingRequest, raw_request: Request): handler = pooling(raw_request) if handler is None: base_server = raw_request.app.state.openai_serving_tokenization return base_server.create_error_response( message="The model does not support Pooling API" ) try: generator = await handler.create_pooling(request, raw_request) except Exception as e: generator = handler.create_error_response(e) if isinstance(generator, ErrorResponse): return JSONResponse( content=generator.model_dump(), status_code=generator.error.code ) elif isinstance(generator, (PoolingResponse, IOProcessorResponse)): return JSONResponse(content=generator.model_dump()) elif isinstance(generator, PoolingBytesResponse): return StreamingResponse( content=generator.content, headers=generator.headers, media_type=generator.media_type, ) assert_never(generator)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/pooling/pooling/api_router.py", "license": "Apache License 2.0", "lines": 53, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/pooling/pooling/protocol.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import time from typing import Generic, TypeAlias, TypeVar from pydantic import Field from vllm import PoolingParams from vllm.config import ModelConfig from vllm.entrypoints.openai.engine.protocol import OpenAIBaseModel, UsageInfo from vllm.entrypoints.pooling.base.protocol import ( ChatRequestMixin, ClassifyRequestMixin, CompletionRequestMixin, EmbedRequestMixin, EncodingRequestMixin, PoolingBasicRequestMixin, ) from vllm.renderers import TokenizeParams from vllm.tasks import PoolingTask from vllm.utils import random_uuid class PoolingCompletionRequest( PoolingBasicRequestMixin, CompletionRequestMixin, EmbedRequestMixin, ClassifyRequestMixin, ): task: PoolingTask | None = None def build_tok_params(self, model_config: ModelConfig) -> TokenizeParams: encoder_config = model_config.encoder_config or {} return TokenizeParams( max_total_tokens=model_config.max_model_len, max_output_tokens=0, truncate_prompt_tokens=self.truncate_prompt_tokens, do_lower_case=encoder_config.get("do_lower_case", False), add_special_tokens=self.add_special_tokens, max_total_tokens_param="max_model_len", ) def to_pooling_params(self): return PoolingParams( task=self.task, use_activation=self.use_activation, dimensions=self.dimensions, ) class PoolingChatRequest( PoolingBasicRequestMixin, ChatRequestMixin, EmbedRequestMixin, ClassifyRequestMixin ): task: PoolingTask | None = None def build_tok_params(self, model_config: ModelConfig) -> TokenizeParams: encoder_config = model_config.encoder_config or {} return TokenizeParams( max_total_tokens=model_config.max_model_len, max_output_tokens=0, truncate_prompt_tokens=self.truncate_prompt_tokens, do_lower_case=encoder_config.get("do_lower_case", False), add_special_tokens=self.add_special_tokens, max_total_tokens_param="max_model_len", ) def to_pooling_params(self): return PoolingParams( task=self.task, use_activation=self.use_activation, dimensions=self.dimensions, ) T = TypeVar("T") class IOProcessorRequest(PoolingBasicRequestMixin, EncodingRequestMixin, Generic[T]): data: T task: PoolingTask = "plugin" def build_tok_params(self, model_config: ModelConfig) -> TokenizeParams: encoder_config = model_config.encoder_config or {} return TokenizeParams( max_total_tokens=model_config.max_model_len, max_output_tokens=0, truncate_prompt_tokens=self.truncate_prompt_tokens, do_lower_case=encoder_config.get("do_lower_case", False), add_special_tokens=not model_config.is_encoder_decoder, max_total_tokens_param="max_model_len", ) class IOProcessorResponse(OpenAIBaseModel, Generic[T]): request_id: str | None = None """ The request_id associated with this response """ created_at: int = Field(default_factory=lambda: int(time.time())) data: T """ When using plugins IOProcessor plugins, the actual output is generated by the plugin itself. Hence, we use a generic type for the response data """ PoolingRequest: TypeAlias = ( PoolingCompletionRequest | PoolingChatRequest | IOProcessorRequest ) class PoolingResponseData(OpenAIBaseModel): index: int object: str = "pooling" data: list[list[float]] | list[float] | str class PoolingResponse(OpenAIBaseModel): id: str = Field(default_factory=lambda: f"pool-{random_uuid()}") object: str = "list" created: int = Field(default_factory=lambda: int(time.time())) model: str data: list[PoolingResponseData] usage: UsageInfo class PoolingBytesResponse(OpenAIBaseModel): content: list[bytes] headers: dict[str, str] | None = None media_type: str = "application/octet-stream"
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/pooling/pooling/protocol.py", "license": "Apache License 2.0", "lines": 105, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/pooling/score/api_router.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from http import HTTPStatus from fastapi import APIRouter, Depends, Request from fastapi.responses import JSONResponse from typing_extensions import assert_never from vllm.entrypoints.openai.engine.protocol import ErrorResponse from vllm.entrypoints.openai.utils import validate_json_request from vllm.entrypoints.pooling.score.protocol import ( RerankRequest, RerankResponse, ScoreRequest, ScoreResponse, ) from vllm.entrypoints.pooling.score.serving import ServingScores from vllm.entrypoints.utils import load_aware_call, with_cancellation from vllm.logger import init_logger router = APIRouter() logger = init_logger(__name__) def score(request: Request) -> ServingScores | None: return request.app.state.openai_serving_scores def rerank(request: Request) -> ServingScores | None: return request.app.state.openai_serving_scores @router.post( "/score", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, }, ) @with_cancellation @load_aware_call async def create_score(request: ScoreRequest, raw_request: Request): handler = score(raw_request) if handler is None: base_server = raw_request.app.state.openai_serving_tokenization return base_server.create_error_response( message="The model does not support Score API" ) try: generator = await handler.create_score(request, raw_request) except Exception as e: generator = handler.create_error_response(e) if isinstance(generator, ErrorResponse): return JSONResponse( content=generator.model_dump(), status_code=generator.error.code ) elif isinstance(generator, ScoreResponse): return JSONResponse(content=generator.model_dump()) assert_never(generator) @router.post( "/v1/score", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, }, ) @with_cancellation @load_aware_call async def create_score_v1(request: ScoreRequest, raw_request: Request): logger.warning( "To indicate that Score API is not part of standard OpenAI API, we " "have moved it to `/score`. Please update your client accordingly." ) return await create_score(request, raw_request) @router.post( "/rerank", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, }, ) @with_cancellation @load_aware_call async def do_rerank(request: RerankRequest, raw_request: Request): handler = rerank(raw_request) if handler is None: base_server = raw_request.app.state.openai_serving_tokenization return base_server.create_error_response( message="The model does not support Rerank (Score) API" ) try: generator = await handler.do_rerank(request, raw_request) except Exception as e: generator = handler.create_error_response(e) if isinstance(generator, ErrorResponse): return JSONResponse( content=generator.model_dump(), status_code=generator.error.code ) elif isinstance(generator, RerankResponse): return JSONResponse(content=generator.model_dump()) assert_never(generator) @router.post( "/v1/rerank", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, }, ) @with_cancellation async def do_rerank_v1(request: RerankRequest, raw_request: Request): logger.warning_once( "To indicate that the rerank API is not part of the standard OpenAI" " API, we have located it at `/rerank`. Please update your client " "accordingly. (Note: Conforms to JinaAI rerank API)" ) return await do_rerank(request, raw_request) @router.post( "/v2/rerank", dependencies=[Depends(validate_json_request)], responses={ HTTPStatus.BAD_REQUEST.value: {"model": ErrorResponse}, HTTPStatus.INTERNAL_SERVER_ERROR.value: {"model": ErrorResponse}, }, ) @with_cancellation async def do_rerank_v2(request: RerankRequest, raw_request: Request): return await do_rerank(request, raw_request)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/pooling/score/api_router.py", "license": "Apache License 2.0", "lines": 122, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/entrypoints/pooling/score/protocol.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import time from typing import TypeAlias from pydantic import BaseModel, Field from vllm import PoolingParams from vllm.config import ModelConfig from vllm.entrypoints.openai.engine.protocol import OpenAIBaseModel, UsageInfo from vllm.entrypoints.pooling.base.protocol import ( ClassifyRequestMixin, PoolingBasicRequestMixin, ) from vllm.entrypoints.pooling.score.utils import ( ScoreContentPartParam, ScoreInput, ScoreInputs, ) from vllm.renderers import TokenizeParams from vllm.tasks import PoolingTask from vllm.utils import random_uuid class ScoreRequestMixin(PoolingBasicRequestMixin, ClassifyRequestMixin): def build_tok_params(self, model_config: ModelConfig) -> TokenizeParams: encoder_config = model_config.encoder_config or {} return TokenizeParams( max_total_tokens=model_config.max_model_len, max_output_tokens=0, truncate_prompt_tokens=self.truncate_prompt_tokens, do_lower_case=encoder_config.get("do_lower_case", False), max_total_tokens_param="max_model_len", ) def to_pooling_params(self, task: PoolingTask = "score"): return PoolingParams( task=task, use_activation=self.use_activation, ) class ScoreDataRequest(ScoreRequestMixin): data_1: ScoreInputs data_2: ScoreInputs class ScoreQueriesDocumentsRequest(ScoreRequestMixin): queries: ScoreInputs documents: ScoreInputs @property def data_1(self): return self.queries @property def data_2(self): return self.documents class ScoreQueriesItemsRequest(ScoreRequestMixin): queries: ScoreInputs items: ScoreInputs @property def data_1(self): return self.queries @property def data_2(self): return self.items class ScoreTextRequest(ScoreRequestMixin): text_1: ScoreInputs text_2: ScoreInputs @property def data_1(self): return self.text_1 @property def data_2(self): return self.text_2 ScoreRequest: TypeAlias = ( ScoreQueriesDocumentsRequest | ScoreQueriesItemsRequest | ScoreDataRequest | ScoreTextRequest ) class RerankRequest(PoolingBasicRequestMixin, ClassifyRequestMixin): query: ScoreInput documents: ScoreInputs top_n: int = Field(default_factory=lambda: 0) def build_tok_params(self, model_config: ModelConfig) -> TokenizeParams: encoder_config = model_config.encoder_config or {} return TokenizeParams( max_total_tokens=model_config.max_model_len, max_output_tokens=0, truncate_prompt_tokens=self.truncate_prompt_tokens, do_lower_case=encoder_config.get("do_lower_case", False), max_total_tokens_param="max_model_len", ) def to_pooling_params(self, task: PoolingTask = "score"): return PoolingParams( task=task, use_activation=self.use_activation, ) class RerankDocument(BaseModel): text: str | None = None multi_modal: list[ScoreContentPartParam] | None = None class RerankResult(BaseModel): index: int document: RerankDocument relevance_score: float class RerankUsage(BaseModel): prompt_tokens: int total_tokens: int class RerankResponse(OpenAIBaseModel): id: str model: str usage: RerankUsage results: list[RerankResult] class ScoreResponseData(OpenAIBaseModel): index: int object: str = "score" score: float class ScoreResponse(OpenAIBaseModel): id: str = Field(default_factory=lambda: f"embd-{random_uuid()}") object: str = "list" created: int = Field(default_factory=lambda: int(time.time())) model: str data: list[ScoreResponseData] usage: UsageInfo
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/pooling/score/protocol.py", "license": "Apache License 2.0", "lines": 115, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/tokenizers/hf.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import contextlib import copy from pathlib import Path from typing import TypeAlias from transformers import AutoTokenizer, PreTrainedTokenizer, PreTrainedTokenizerFast from vllm.transformers_utils.config import get_sentence_transformer_tokenizer_config from .protocol import TokenizerLike HfTokenizer: TypeAlias = PreTrainedTokenizer | PreTrainedTokenizerFast def get_cached_tokenizer(tokenizer: HfTokenizer) -> HfTokenizer: """ By default, transformers will recompute multiple tokenizer properties each time they are called, leading to a significant slowdown. This proxy caches these properties for faster access. """ cached_tokenizer = copy.copy(tokenizer) tokenizer_all_special_ids = tokenizer.all_special_ids tokenizer_all_special_tokens = tokenizer.all_special_tokens tokenizer_vocab = tokenizer.get_vocab() tokenizer_len = len(tokenizer) max_token_id = max(tokenizer_vocab.values()) max_chars_per_token = max(len(tok) for tok in tokenizer_vocab) # Some tokenizers (e.g., QwenTokenizer) have special tokens that # are added and included in the implementation of the vocab_size # property, but not in get_vocab(); if there is an implementation # of vocab size, we should take the greater value. if hasattr(tokenizer, "vocab_size"): with contextlib.suppress(NotImplementedError): max_token_id = max(max_token_id, tokenizer.vocab_size) class CachedTokenizer(tokenizer.__class__): # type: ignore @property def all_special_ids(self) -> list[int]: return tokenizer_all_special_ids @property def all_special_tokens(self) -> list[str]: return tokenizer_all_special_tokens @property def max_token_id(self) -> int: return max_token_id @property def max_chars_per_token(self) -> int: return max_chars_per_token def get_vocab(self) -> dict[str, int]: return tokenizer_vocab def __len__(self) -> int: return tokenizer_len def __reduce__(self): return get_cached_tokenizer, (tokenizer,) CachedTokenizer.__name__ = f"Cached{tokenizer.__class__.__name__}" cached_tokenizer.__class__ = CachedTokenizer return cached_tokenizer class CachedHfTokenizer(TokenizerLike): @classmethod def from_pretrained( cls, path_or_repo_id: str | Path, *args, trust_remote_code: bool = False, revision: str | None = None, download_dir: str | None = None, **kwargs, ) -> HfTokenizer: try: tokenizer = AutoTokenizer.from_pretrained( path_or_repo_id, *args, trust_remote_code=trust_remote_code, revision=revision, cache_dir=download_dir, **kwargs, ) except ValueError as e: # If the error pertains to the tokenizer class not existing or not # currently being imported, # suggest using the --trust-remote-code flag. if not trust_remote_code and ( "does not exist or is not currently imported." in str(e) or "requires you to execute the tokenizer file" in str(e) ): err_msg = ( "Failed to load the tokenizer. If the tokenizer " "is a custom tokenizer not yet available in the " "HuggingFace transformers library, consider " "setting `trust_remote_code=True` in LLM or using " "the `--trust-remote-code` flag in the CLI." ) raise RuntimeError(err_msg) from e else: raise e # The special_tokens in tokenizer should also be # controlled by do_lower_case in encoder_config encoder_config = get_sentence_transformer_tokenizer_config( path_or_repo_id, revision ) if isinstance(encoder_config, dict) and encoder_config.get( "do_lower_case", False ): special_tokens_map = { k: v.lower() for k, v in tokenizer.special_tokens_map.items() } tokenizer.add_special_tokens(special_tokens_map) return get_cached_tokenizer(tokenizer)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/tokenizers/hf.py", "license": "Apache License 2.0", "lines": 103, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/moe/test_modular_oai_triton_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Test modular OAI Triton MoE """ import pytest import torch from vllm.model_executor.layers.fused_moe.activation import MoEActivation from vllm.utils.import_utils import has_triton_kernels if not has_triton_kernels(): pytest.skip( "triton_kernels not found, skipping all related tests", allow_module_level=True, ) from triton_kernels.matmul_ogs import FlexCtx, PrecisionConfig from triton_kernels.numerics import InFlexData from triton_kernels.numerics_details.mxfp import downcast_to_mxfp, upcast_from_mxfp from triton_kernels.tensor import FP4, convert_layout, wrap_torch_tensor from triton_kernels.tensor_details import layout from triton_kernels.testing import assert_close from vllm.config import VllmConfig, set_current_vllm_config from vllm.model_executor.layers.fused_moe.config import mxfp4_w4a16_moe_quant_config from vllm.model_executor.layers.fused_moe.gpt_oss_triton_kernels_moe import ( OAITritonExperts, UnfusedOAITritonExperts, ) from vllm.model_executor.layers.fused_moe.modular_kernel import FusedMoEModularKernel from vllm.model_executor.layers.fused_moe.prepare_finalize import ( MoEPrepareAndFinalizeNoEP, ) from vllm.model_executor.layers.utils import shuffle_weight from vllm.platforms import current_platform from vllm.utils.torch_utils import set_random_seed from .utils import make_dummy_moe_config MNK = [ (1, 512, 384), (1, 2880, 2880), (2, 512, 384), (2, 2880, 2880), (16, 2880, 2880), ] def unshuffle_weight(w: torch.Tensor): first = w[..., ::2] second = w[..., 1::2] return torch.concat((first, second), dim=-1) def make_weights(dtype, k, n, e): w1 = torch.randn((e, k, 2 * n), dtype=dtype, device="cuda") w1_bias = torch.randn((e, 2 * n), dtype=dtype, device="cuda") w2 = torch.randn((e, n, k), dtype=dtype, device="cuda") w2_bias = torch.randn((e, k), dtype=dtype, device="cuda") w1_tri = w1.clone() w2_tri = w2.clone() w1_bias_tri = w1_bias.clone() w2_bias_tri = w2_bias.clone() w1_bias_tri = w1_bias_tri.to(torch.float32) w2_bias_tri = w2_bias_tri.to(torch.float32) # shuffle weights w1_tri = shuffle_weight(w1_tri) w1_bias_tri = shuffle_weight(w1_bias_tri) # quant triton_weights w1_tri, w1_scale_tri = downcast_to_mxfp(w1_tri, torch.uint8, axis=1) w1 = upcast_from_mxfp(w1_tri, w1_scale_tri, dtype, axis=1) w1 = unshuffle_weight(w1) w2_tri, w2_scale_tri = downcast_to_mxfp(w2_tri, torch.uint8, axis=1) w2 = upcast_from_mxfp(w2_tri, w2_scale_tri, dtype, axis=1) num_warps = 8 w_layout, w_layout_opts = layout.make_default_matmul_mxfp4_w_layout(mx_axis=1) w_scale_layout, w_scale_layout_opts = ( layout.make_default_matmul_mxfp4_w_scale_layout(mx_axis=1, num_warps=num_warps) ) w1_tri = convert_layout(wrap_torch_tensor(w1_tri, FP4), w_layout, **w_layout_opts) w1_scale_tri = convert_layout( wrap_torch_tensor(w1_scale_tri), w_scale_layout, **w_scale_layout_opts, ) w2_tri = convert_layout(wrap_torch_tensor(w2_tri, FP4), w_layout, **w_layout_opts) w2_scale_tri = convert_layout( wrap_torch_tensor(w2_scale_tri), w_scale_layout, **w_scale_layout_opts, ) w1_precision_config = PrecisionConfig( weight_scale=w1_scale_tri, flex_ctx=FlexCtx(rhs_data=InFlexData()) ) w2_precision_config = PrecisionConfig( weight_scale=w2_scale_tri, flex_ctx=FlexCtx(rhs_data=InFlexData()) ) return ( w1, w2, w1_bias, w2_bias, w1_tri, w2_tri, w1_bias_tri, w2_bias_tri, w1_precision_config, w2_precision_config, ) def swiglu(x, alpha: float = 1.702, limit: float = 1.0): # Note we add an extra bias of 1 to the linear layer x_glu, x_linear = torch.chunk(x, 2, dim=-1) if limit is not None: x_glu = x_glu.clamp(max=limit) out_glu = x_glu * torch.sigmoid(alpha * x_glu) if limit is not None: x_linear = x_linear.clamp(min=-limit, max=limit) return out_glu * (x_linear + 1) def torch_moe_impl( hidden_states: torch.Tensor, # (M, K) w1: torch.Tensor, # (E, K, 2N) w2: torch.Tensor, # (E, N, K) w1_bias: torch.Tensor, # (E, 2N) w2_bias: torch.Tensor, # (E, K) topk_weights: torch.Tensor, # (M, topk) topk_ids: torch.Tensor, # (M, topk) ): w1 = w1[topk_ids, ...] w1_bias = w1_bias[topk_ids, ...] hidden_states = torch.einsum("bekc,bk->bec", w1, hidden_states) + w1_bias hidden_states = swiglu(hidden_states, limit=7) w2 = w2[topk_ids, ...] w2_bias = w2_bias[topk_ids, ...] hidden_states = torch.einsum("bekc,bek->bec", w2, hidden_states) + w2_bias # Weighted sum of experts hidden_states = torch.einsum("bec,be->bc", hidden_states, topk_weights) return hidden_states def oai_triton_moe_impl( x: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, w1_scale: "PrecisionConfig", w2_scale: "PrecisionConfig", w1_bias: torch.Tensor | None, w2_bias: torch.Tensor | None, num_experts: int, topk_weights: torch.Tensor, topk_ids: torch.Tensor, unfused: bool = False, ) -> torch.Tensor: quant_config = mxfp4_w4a16_moe_quant_config( w1_bias=w1_bias, w2_bias=w2_bias, w1_scale=w1_scale, w2_scale=w2_scale, ) if unfused: fused_experts = UnfusedOAITritonExperts(make_dummy_moe_config(), quant_config) else: fused_experts = OAITritonExperts(make_dummy_moe_config(), quant_config) mk = FusedMoEModularKernel( MoEPrepareAndFinalizeNoEP(), fused_experts, inplace=False, ) return mk.forward( hidden_states=x, w1=w1, w2=w2, topk_weights=topk_weights, topk_ids=topk_ids, activation=MoEActivation.SWIGLUOAI, global_num_experts=num_experts, expert_map=None, apply_router_weight_on_input=False, ) @pytest.mark.skipif( not current_platform.is_cuda(), reason="This test is skipped on non-CUDA platform." ) @pytest.mark.parametrize("dtype", [torch.bfloat16]) @pytest.mark.parametrize("m,n,k", MNK) @pytest.mark.parametrize("num_experts", [32, 128]) @pytest.mark.parametrize("topk", [4]) @pytest.mark.parametrize("unfused", [True, False]) def test_oai_triton_moe( dtype: torch.dtype, m: int, n: int, k: int, num_experts: int, topk: int, unfused: bool, workspace_init, ): set_random_seed(0) ( w1, w2, w1_bias, w2_bias, w1_tri, w2_tri, w1_bias_tri, w2_bias_tri, w1_precision_config, w2_precision_config, ) = make_weights(dtype, k, n, num_experts) x = torch.randn((m, k), dtype=dtype, device="cuda") router_logits = torch.randn(m, num_experts, device="cuda", dtype=dtype) topk_weights, topk_ids = torch.topk(router_logits, k=topk, dim=-1, sorted=True) topk_weights = torch.nn.functional.softmax(topk_weights, dim=-1) with set_current_vllm_config(VllmConfig()): out_ref = torch_moe_impl(x, w1, w2, w1_bias, w2_bias, topk_weights, topk_ids) out = oai_triton_moe_impl( x, w1_tri, w2_tri, w1_precision_config, w2_precision_config, w1_bias_tri, w2_bias_tri, num_experts, topk_weights, topk_ids, unfused, ) assert_close(ref=out_ref, tri=out, maxtol=0.025, rmstol=0.005)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/moe/test_modular_oai_triton_moe.py", "license": "Apache License 2.0", "lines": 218, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/tokenizers_/test_basic.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import _get_protocol_attrs # type: ignore import pytest from transformers import ( PreTrainedTokenizerBase, PreTrainedTokenizerFast, ) from vllm.tokenizers import TokenizerLike, get_tokenizer from vllm.tokenizers.grok2 import Grok2Tokenizer from vllm.tokenizers.hf import HfTokenizer from vllm.tokenizers.mistral import MistralTokenizer def _get_missing_attrs(obj: object, target: type): return [k for k in _get_protocol_attrs(target) if not hasattr(obj, k)] def _assert_tokenizer_like(tokenizer: object): missing_attrs = _get_missing_attrs(tokenizer, TokenizerLike) assert not missing_attrs, f"Missing attrs: {missing_attrs}" def test_tokenizer_like_protocol(): tokenizer = get_tokenizer("gpt2", use_fast=True) assert isinstance(tokenizer, PreTrainedTokenizerFast) _assert_tokenizer_like(tokenizer) tokenizer = get_tokenizer( "mistralai/Mistral-7B-Instruct-v0.3", tokenizer_mode="mistral" ) assert isinstance(tokenizer, MistralTokenizer) _assert_tokenizer_like(tokenizer) tokenizer = get_tokenizer("xai-org/grok-2", tokenizer_mode="grok2") assert isinstance(tokenizer, Grok2Tokenizer) _assert_tokenizer_like(tokenizer) tokenizer = get_tokenizer("deepseek-ai/DeepSeek-V3", tokenizer_mode="deepseek_v32") assert isinstance(tokenizer, HfTokenizer) # Verify it's a fast tokenizer (required for FastIncrementalDetokenizer) assert isinstance(tokenizer, PreTrainedTokenizerFast) assert "DSV32" in tokenizer.__class__.__name__ _assert_tokenizer_like(tokenizer) @pytest.mark.parametrize("tokenizer_name", ["facebook/opt-125m", "gpt2"]) def test_tokenizer_revision(tokenizer_name: str): # Assume that "main" branch always exists tokenizer = get_tokenizer(tokenizer_name, revision="main") assert isinstance(tokenizer, PreTrainedTokenizerBase) # Assume that "never" branch always does not exist with pytest.raises(OSError, match="not a valid git identifier"): get_tokenizer(tokenizer_name, revision="never") @pytest.mark.parametrize("tokenizer_name", ["BAAI/bge-base-en"]) @pytest.mark.parametrize("n_tokens", [510]) def test_special_tokens(tokenizer_name: str, n_tokens: int): tokenizer = get_tokenizer(tokenizer_name, revision="main") prompts = "[UNK]" * n_tokens prompt_token_ids = tokenizer.encode(prompts) assert len(prompt_token_ids) == n_tokens + 2
{ "repo_id": "vllm-project/vllm", "file_path": "tests/tokenizers_/test_basic.py", "license": "Apache License 2.0", "lines": 50, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/tokenizers_/test_registry.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from pathlib import Path import pytest from vllm.tokenizers import TokenizerLike from vllm.tokenizers.registry import ( TokenizerRegistry, get_tokenizer, resolve_tokenizer_args, ) class TestTokenizer(TokenizerLike): @classmethod def from_pretrained( cls, path_or_repo_id: str | Path, *args, trust_remote_code: bool = False, revision: str | None = None, download_dir: str | None = None, **kwargs, ) -> "TestTokenizer": return TestTokenizer(path_or_repo_id) # type: ignore def __init__(self, path_or_repo_id: str | Path) -> None: super().__init__() self.path_or_repo_id = path_or_repo_id @property def bos_token_id(self) -> int: return 0 @property def eos_token_id(self) -> int: return 1 @property def pad_token_id(self) -> int: return 2 @property def is_fast(self) -> bool: return True @pytest.mark.parametrize("runner_type", ["generate", "pooling"]) def test_resolve_tokenizer_args_idempotent(runner_type): tokenizer_mode, tokenizer_name, args, kwargs = resolve_tokenizer_args( "facebook/opt-125m", runner_type=runner_type, ) assert (tokenizer_mode, tokenizer_name, args, kwargs) == resolve_tokenizer_args( tokenizer_name, *args, **kwargs ) def test_customized_tokenizer(): TokenizerRegistry.register("test_tokenizer", __name__, TestTokenizer.__name__) tokenizer = TokenizerRegistry.load_tokenizer("test_tokenizer", "abc") assert isinstance(tokenizer, TestTokenizer) assert tokenizer.path_or_repo_id == "abc" assert tokenizer.bos_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.pad_token_id == 2 tokenizer = get_tokenizer("abc", tokenizer_mode="test_tokenizer") assert isinstance(tokenizer, TestTokenizer) assert tokenizer.path_or_repo_id == "abc" assert tokenizer.bos_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.pad_token_id == 2
{ "repo_id": "vllm-project/vllm", "file_path": "tests/tokenizers_/test_registry.py", "license": "Apache License 2.0", "lines": 60, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/transformers_utils/test_repo_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import tempfile from pathlib import Path from unittest.mock import MagicMock, call, patch import pytest from vllm.transformers_utils.repo_utils import ( any_pattern_in_repo_files, is_mistral_model_repo, list_filtered_repo_files, ) @pytest.mark.parametrize( "allow_patterns,expected_relative_files", [ ( ["*.json", "correct*.txt"], ["json_file.json", "subfolder/correct.txt", "correct_2.txt"], ), ], ) def test_list_filtered_repo_files( allow_patterns: list[str], expected_relative_files: list[str] ): with tempfile.TemporaryDirectory() as tmp_dir: # Prep folder and files path_tmp_dir = Path(tmp_dir) subfolder = path_tmp_dir / "subfolder" subfolder.mkdir() (path_tmp_dir / "json_file.json").touch() (path_tmp_dir / "correct_2.txt").touch() (path_tmp_dir / "uncorrect.txt").touch() (path_tmp_dir / "uncorrect.jpeg").touch() (subfolder / "correct.txt").touch() (subfolder / "uncorrect_sub.txt").touch() def _glob_path() -> list[str]: return [ str(file.relative_to(path_tmp_dir)) for file in path_tmp_dir.glob("**/*") if file.is_file() ] # Patch list_repo_files called by fn with patch( "vllm.transformers_utils.repo_utils.list_repo_files", MagicMock(return_value=_glob_path()), ) as mock_list_repo_files: out_files = sorted( list_filtered_repo_files( tmp_dir, allow_patterns, "revision", "model", "token" ) ) assert out_files == sorted(expected_relative_files) assert mock_list_repo_files.call_count == 1 assert mock_list_repo_files.call_args_list[0] == call( repo_id=tmp_dir, revision="revision", repo_type="model", token="token", ) @pytest.mark.parametrize( ("allow_patterns", "expected_bool"), [ (["*.json", "correct*.txt"], True), ( ["*.jpeg"], True, ), ( ["not_found.jpeg"], False, ), ], ) def test_one_filtered_repo_files(allow_patterns: list[str], expected_bool: bool): with tempfile.TemporaryDirectory() as tmp_dir: # Prep folder and files path_tmp_dir = Path(tmp_dir) subfolder = path_tmp_dir / "subfolder" subfolder.mkdir() (path_tmp_dir / "uncorrect.jpeg").touch() (subfolder / "correct.txt").touch() def _glob_path() -> list[str]: return [ str(file.relative_to(path_tmp_dir)) for file in path_tmp_dir.glob("**/*") if file.is_file() ] # Patch list_repo_files called by fn with patch( "vllm.transformers_utils.repo_utils.list_repo_files", MagicMock(return_value=_glob_path()), ) as mock_list_repo_files: assert ( any_pattern_in_repo_files( tmp_dir, allow_patterns, "revision", "model", "token" ) ) is expected_bool assert mock_list_repo_files.call_count == 1 assert mock_list_repo_files.call_args_list[0] == call( repo_id=tmp_dir, revision="revision", repo_type="model", token="token", ) @pytest.mark.parametrize( ("files", "expected_bool"), [ (["consolidated.safetensors", "incorrect.txt"], True), (["consolidated-1.safetensors", "incorrect.txt"], True), ( ["consolidated-1.json"], False, ), ], ) def test_is_mistral_model_repo(files: list[str], expected_bool: bool): with tempfile.TemporaryDirectory() as tmp_dir: # Prep folder and files path_tmp_dir = Path(tmp_dir) for file in files: (path_tmp_dir / file).touch() def _glob_path() -> list[str]: return [ str(file.relative_to(path_tmp_dir)) for file in path_tmp_dir.glob("**/*") if file.is_file() ] # Patch list_repo_files called by fn with patch( "vllm.transformers_utils.repo_utils.list_repo_files", MagicMock(return_value=_glob_path()), ) as mock_list_repo_files: assert ( is_mistral_model_repo(tmp_dir, "revision", "model", "token") is expected_bool ) assert mock_list_repo_files.call_count == 1 assert mock_list_repo_files.call_args_list[0] == call( repo_id=tmp_dir, revision="revision", repo_type="model", token="token", )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/transformers_utils/test_repo_utils.py", "license": "Apache License 2.0", "lines": 142, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/tokenizers/protocol.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from pathlib import Path from typing import TYPE_CHECKING, Any, Protocol, overload if TYPE_CHECKING: from transformers import BatchEncoding from vllm.entrypoints.chat_utils import ChatCompletionMessageParam class TokenizerLike(Protocol): @classmethod def from_pretrained( cls, path_or_repo_id: str | Path, *args, trust_remote_code: bool = False, revision: str | None = None, download_dir: str | None = None, **kwargs, ) -> "TokenizerLike": raise NotImplementedError def num_special_tokens_to_add(self) -> int: raise NotImplementedError @property def all_special_tokens(self) -> list[str]: raise NotImplementedError @property def all_special_ids(self) -> list[int]: raise NotImplementedError @property def bos_token_id(self) -> int: raise NotImplementedError @property def eos_token_id(self) -> int: raise NotImplementedError @property def pad_token_id(self) -> int: raise NotImplementedError @property def is_fast(self) -> bool: raise NotImplementedError @property def vocab_size(self) -> int: raise NotImplementedError @property def max_token_id(self) -> int: raise NotImplementedError @property def max_chars_per_token(self) -> int: raise NotImplementedError @property def truncation_side(self) -> str: raise NotImplementedError def __hash__(self) -> int: return hash(id(self)) def __len__(self) -> int: return self.vocab_size def __call__( self, text: str | list[str], text_pair: str | None = None, add_special_tokens: bool = True, truncation: bool = False, max_length: int | None = None, ) -> "BatchEncoding": raise NotImplementedError def get_vocab(self) -> dict[str, int]: raise NotImplementedError def get_added_vocab(self) -> dict[str, int]: raise NotImplementedError def encode( self, text: str, truncation: bool | None = None, max_length: int | None = None, add_special_tokens: bool = True, ) -> list[int]: raise NotImplementedError def apply_chat_template( self, messages: list["ChatCompletionMessageParam"], tools: list[dict[str, Any]] | None = None, **kwargs, ) -> str | list[int]: raise NotImplementedError @overload def convert_tokens_to_ids(self, tokens: str) -> int: ... @overload def convert_tokens_to_ids(self, tokens: list[str]) -> list[int]: ... def convert_tokens_to_ids(self, tokens: str | list[str]) -> int | list[int]: raise NotImplementedError def convert_tokens_to_string(self, tokens: list[str]) -> str: raise NotImplementedError def decode(self, ids: list[int] | int, skip_special_tokens: bool = False) -> str: raise NotImplementedError def convert_ids_to_tokens( self, ids: list[int], skip_special_tokens: bool = False, ) -> list[str]: raise NotImplementedError
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/tokenizers/protocol.py", "license": "Apache License 2.0", "lines": 99, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/tokenizers/registry.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass, field from functools import lru_cache from pathlib import Path from typing import TYPE_CHECKING import huggingface_hub from typing_extensions import TypeVar, assert_never import vllm.envs as envs from vllm.logger import init_logger from vllm.transformers_utils.gguf_utils import ( check_gguf_file, get_gguf_file_path_from_hf, is_gguf, is_remote_gguf, split_remote_gguf, ) from vllm.transformers_utils.repo_utils import ( any_pattern_in_repo_files, is_mistral_model_repo, ) from vllm.utils.import_utils import resolve_obj_by_qualname from .protocol import TokenizerLike if TYPE_CHECKING: from vllm.config.model import ModelConfig, RunnerType logger = init_logger(__name__) _VLLM_TOKENIZERS = { "deepseek_v32": ("deepseek_v32", "DeepseekV32Tokenizer"), "grok2": ("grok2", "Grok2Tokenizer"), "hf": ("hf", "CachedHfTokenizer"), "mistral": ("mistral", "MistralTokenizer"), } @dataclass class _TokenizerRegistry: # Tokenizer mode -> (tokenizer module, tokenizer class) tokenizers: dict[str, tuple[str, str]] = field(default_factory=dict) def register(self, tokenizer_mode: str, module: str, class_name: str) -> None: if tokenizer_mode in self.tokenizers: logger.warning( "%s.%s is already registered for tokenizer_mode=%r. " "It is overwritten by the new one.", module, class_name, tokenizer_mode, ) self.tokenizers[tokenizer_mode] = (module, class_name) return None def load_tokenizer_cls(self, tokenizer_mode: str) -> type[TokenizerLike]: if tokenizer_mode not in self.tokenizers: raise ValueError(f"No tokenizer registered for {tokenizer_mode=!r}.") module, class_name = self.tokenizers[tokenizer_mode] logger.debug_once(f"Loading {class_name} for {tokenizer_mode=!r}") return resolve_obj_by_qualname(f"{module}.{class_name}") def load_tokenizer(self, tokenizer_mode: str, *args, **kwargs) -> TokenizerLike: tokenizer_cls = self.load_tokenizer_cls(tokenizer_mode) return tokenizer_cls.from_pretrained(*args, **kwargs) TokenizerRegistry = _TokenizerRegistry( { mode: (f"vllm.tokenizers.{mod_relname}", cls_name) for mode, (mod_relname, cls_name) in _VLLM_TOKENIZERS.items() } ) def resolve_tokenizer_args( tokenizer_name: str | Path, *args, runner_type: "RunnerType" = "generate", tokenizer_mode: str = "auto", **kwargs, ): revision: str | None = kwargs.get("revision") download_dir: str | None = kwargs.get("download_dir") if envs.VLLM_USE_MODELSCOPE: # download model from ModelScope hub, # lazy import so that modelscope is not required for normal use. from modelscope.hub.snapshot_download import snapshot_download # avoid circular import from vllm.model_executor.model_loader.weight_utils import get_lock # Only set the tokenizer here, model will be downloaded on the workers. if not Path(tokenizer_name).exists(): # Use file lock to prevent multiple processes from # downloading the same file at the same time. with get_lock(tokenizer_name, download_dir): tokenizer_path = snapshot_download( model_id=str(tokenizer_name), cache_dir=download_dir, revision=revision, local_files_only=huggingface_hub.constants.HF_HUB_OFFLINE, # Ignore weights - we only need the tokenizer. ignore_file_pattern=[".*.pt", ".*.safetensors", ".*.bin"], ) tokenizer_name = tokenizer_path # Separate model folder from file path for GGUF models if is_gguf(tokenizer_name): if check_gguf_file(tokenizer_name): kwargs["gguf_file"] = Path(tokenizer_name).name tokenizer_name = Path(tokenizer_name).parent elif is_remote_gguf(tokenizer_name): tokenizer_name, quant_type = split_remote_gguf(tokenizer_name) # Get the HuggingFace Hub path for the GGUF file gguf_file = get_gguf_file_path_from_hf( tokenizer_name, quant_type, revision=revision, ) kwargs["gguf_file"] = gguf_file if "truncation_side" not in kwargs: if runner_type == "generate" or runner_type == "draft": kwargs["truncation_side"] = "left" elif runner_type == "pooling": kwargs["truncation_side"] = "right" else: assert_never(runner_type) if tokenizer_mode == "slow": if kwargs.get("use_fast", False): raise ValueError("Cannot use the fast tokenizer in slow tokenizer mode.") tokenizer_mode = "hf" kwargs["use_fast"] = False # Try to use official Mistral tokenizer if possible if ( tokenizer_mode == "auto" and is_mistral_model_repo( model_name_or_path=str(tokenizer_name), revision=revision ) and any_pattern_in_repo_files( model_name_or_path=str(tokenizer_name), allow_patterns=["tekken.json", "tokenizer.model.v*"], revision=revision, ) ): tokenizer_mode = "mistral" # Try to use Grok2 tiktoken tokenizer if possible if tokenizer_mode == "auto" and any_pattern_in_repo_files( model_name_or_path=str(tokenizer_name), allow_patterns=["tokenizer.tok.json"], revision=revision, ): tokenizer_mode = "grok2" # Fallback to HF tokenizer if tokenizer_mode == "auto": tokenizer_mode = "hf" return tokenizer_mode, tokenizer_name, args, kwargs cached_resolve_tokenizer_args = lru_cache(resolve_tokenizer_args) def tokenizer_args_from_config(config: "ModelConfig", **kwargs): return cached_resolve_tokenizer_args( config.tokenizer, runner_type=config.runner_type, tokenizer_mode=config.tokenizer_mode, revision=config.tokenizer_revision, trust_remote_code=config.trust_remote_code, **kwargs, ) _T = TypeVar("_T", bound=TokenizerLike, default=TokenizerLike) def get_tokenizer( tokenizer_name: str | Path, *args, tokenizer_cls: type[_T] = TokenizerLike, # type: ignore[assignment] trust_remote_code: bool = False, revision: str | None = None, download_dir: str | None = None, **kwargs, ) -> _T: """Gets a tokenizer for the given model name via HuggingFace or ModelScope.""" tokenizer_mode, tokenizer_name, args, kwargs = cached_resolve_tokenizer_args( tokenizer_name, *args, trust_remote_code=trust_remote_code, revision=revision, download_dir=download_dir, **kwargs, ) if tokenizer_cls == TokenizerLike: tokenizer_cls_ = TokenizerRegistry.load_tokenizer_cls(tokenizer_mode) else: tokenizer_cls_ = tokenizer_cls tokenizer = tokenizer_cls_.from_pretrained(tokenizer_name, *args, **kwargs) if not tokenizer.is_fast: logger.warning( "Using a slow tokenizer. This might cause a significant " "slowdown. Consider using a fast tokenizer instead." ) return tokenizer # type: ignore cached_get_tokenizer = lru_cache(get_tokenizer) def cached_tokenizer_from_config(model_config: "ModelConfig", **kwargs): if model_config.skip_tokenizer_init: return None return cached_get_tokenizer( model_config.tokenizer, runner_type=model_config.runner_type, tokenizer_mode=model_config.tokenizer_mode, revision=model_config.tokenizer_revision, trust_remote_code=model_config.trust_remote_code, **kwargs, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/tokenizers/registry.py", "license": "Apache License 2.0", "lines": 193, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/worker/gpu/sample/gumbel.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.triton_utils import tl, triton @triton.jit def _temperature_kernel( logits_ptr, logits_stride, idx_mapping_ptr, temperature_ptr, vocab_size, BLOCK_SIZE: tl.constexpr, ): batch_idx = tl.program_id(0) req_state_idx = tl.load(idx_mapping_ptr + batch_idx) temperature = tl.load(temperature_ptr + req_state_idx).to(tl.float32) if temperature == 0.0 or temperature == 1.0: # Early return to avoid loading logits. return block_idx = tl.program_id(1) block = block_idx * BLOCK_SIZE + tl.arange(0, BLOCK_SIZE) mask = block < vocab_size logits = tl.load(logits_ptr + batch_idx * logits_stride + block, mask=mask) logits = logits.to(tl.float32) logits = logits / temperature tl.store(logits_ptr + batch_idx * logits_stride + block, logits, mask=mask) def apply_temperature( logits: torch.Tensor, idx_mapping: torch.Tensor, temperature: torch.Tensor, ) -> None: num_reqs, vocab_size = logits.shape BLOCK_SIZE = 8192 num_blocks = triton.cdiv(vocab_size, BLOCK_SIZE) _temperature_kernel[(num_reqs, num_blocks)]( logits, logits.stride(0), idx_mapping, temperature, vocab_size, BLOCK_SIZE=BLOCK_SIZE, ) @triton.jit def _gumbel_sample_kernel( local_argmax_ptr, local_argmax_stride, local_max_ptr, local_max_stride, logits_ptr, logits_stride, idx_mapping_ptr, seeds_ptr, pos_ptr, temp_ptr, vocab_size, BLOCK_SIZE: tl.constexpr, APPLY_TEMPERATURE: tl.constexpr, ): batch_idx = tl.program_id(0) req_state_idx = tl.load(idx_mapping_ptr + batch_idx) block_idx = tl.program_id(1) block = block_idx * BLOCK_SIZE + tl.arange(0, BLOCK_SIZE) mask = block < vocab_size logits = tl.load( logits_ptr + batch_idx * logits_stride + block, mask=mask, other=float("-inf"), ) logits = logits.to(tl.float32) temp = tl.load(temp_ptr + req_state_idx).to(tl.float32) if temp != 0.0: # Calculate the seed for gumbel noise. seed = tl.load(seeds_ptr + req_state_idx) pos = tl.load(pos_ptr + batch_idx) gumbel_seed = tl.randint(seed, pos) # Generate gumbel noise in FP32. u = tl.rand(gumbel_seed, block) u = tl.maximum(u, 1e-7) gumbel_noise = -tl.log(-tl.log(u)) # Apply temperature. if APPLY_TEMPERATURE: # NOTE(woosuk): Match the behavior of _temperature_kernel. # E.g., if the kernel uses tl.div_rn, we should use tl.div_rn here too. logits = logits / temp # Apply gumbel noise. logits = tl.where(mask, logits + gumbel_noise, float("-inf")) value, idx = tl.max(logits, axis=0, return_indices=True) token_id = block_idx * BLOCK_SIZE + idx tl.store(local_argmax_ptr + batch_idx * local_argmax_stride + block_idx, token_id) tl.store(local_max_ptr + batch_idx * local_max_stride + block_idx, value) def gumbel_sample( logits: torch.Tensor, # [num_reqs, vocab_size] idx_mapping: torch.Tensor, # [max_num_reqs] temperature: torch.Tensor, # [max_num_reqs] seed: torch.Tensor, # [max_num_reqs] pos: torch.Tensor, # [num_reqs] apply_temperature: bool, ) -> torch.Tensor: num_reqs, vocab_size = logits.shape BLOCK_SIZE = 1024 num_blocks = triton.cdiv(vocab_size, BLOCK_SIZE) local_argmax = torch.empty( num_reqs, num_blocks, dtype=torch.int64, device=logits.device, ) local_max = torch.empty( num_reqs, num_blocks, dtype=torch.float32, device=logits.device, ) _gumbel_sample_kernel[(num_reqs, num_blocks)]( local_argmax, local_argmax.stride(0), local_max, local_max.stride(0), logits, logits.stride(0), idx_mapping, seed, pos, temperature, vocab_size, BLOCK_SIZE=BLOCK_SIZE, APPLY_TEMPERATURE=apply_temperature, ) # NOTE(woosuk): Use int64 for later indexing. max_block_idx = local_max.argmax(dim=-1, keepdim=True) sampled = local_argmax.gather(dim=-1, index=max_block_idx).view(-1) return sampled
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/worker/gpu/sample/gumbel.py", "license": "Apache License 2.0", "lines": 132, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/worker/gpu/sample/logprob.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.triton_utils import tl, triton from vllm.v1.outputs import LogprobsTensors @triton.jit def _topk_log_softmax_kernel( output_ptr, logits_ptr, logits_stride, topk_ids_ptr, topk, vocab_size, BLOCK_SIZE: tl.constexpr, PADDED_TOPK: tl.constexpr, ): req_idx = tl.program_id(0) row_ptr = logits_ptr + req_idx * logits_stride max_val = float("-inf") for i in range(0, vocab_size, BLOCK_SIZE): block = i + tl.arange(0, BLOCK_SIZE) logits = tl.load(row_ptr + block, mask=block < vocab_size, other=float("-inf")) max_val = tl.max(tl.maximum(logits, max_val)) max_val = max_val.to(tl.float32) # type: ignore se = 0.0 for i in range(0, vocab_size, BLOCK_SIZE): block = i + tl.arange(0, BLOCK_SIZE) logits = tl.load(row_ptr + block, mask=block < vocab_size, other=0.0) # NOTE(woosuk): Make sure that logits and all following operations use FP32. logits = logits.to(tl.float32) e = tl.exp(logits - max_val) e = tl.where(block < vocab_size, e, 0.0) se += tl.sum(e) lse = tl.log(se) k_offset = tl.arange(0, PADDED_TOPK) k_mask = k_offset < topk topk_ids = tl.load(topk_ids_ptr + req_idx * topk + k_offset, mask=k_mask, other=0) logits = tl.load(row_ptr + topk_ids, mask=k_mask) logits = logits.to(tl.float32) o = logits - max_val - lse tl.store(output_ptr + req_idx * topk + k_offset, o, mask=k_mask) @triton.jit def _ranks_kernel( output_ptr, logits_ptr, logits_stride, token_ids_ptr, vocab_size, BLOCK_SIZE: tl.constexpr, ): req_idx = tl.program_id(0) row_ptr = logits_ptr + req_idx * logits_stride token_id = tl.load(token_ids_ptr + req_idx) x = tl.load(row_ptr + token_id) n = 0 for i in range(0, vocab_size, BLOCK_SIZE): block = i + tl.arange(0, BLOCK_SIZE) logits = tl.load(row_ptr + block, mask=block < vocab_size, other=float("-inf")) n += tl.sum((logits >= x).to(tl.int32)) tl.store(output_ptr + req_idx, n) def compute_token_logprobs( logits: torch.Tensor, token_ids: torch.Tensor ) -> torch.Tensor: batch_size, vocab_size = logits.shape token_ids = token_ids.to(torch.int64) num_logprobs = token_ids.shape[1] logprobs = logits.new_empty((batch_size, num_logprobs), dtype=torch.float32) _topk_log_softmax_kernel[(batch_size,)]( logprobs, logits, logits.stride(0), token_ids, num_logprobs, vocab_size, BLOCK_SIZE=1024, # type: ignore PADDED_TOPK=triton.next_power_of_2(num_logprobs), ) return logprobs def compute_topk_logprobs( logits: torch.Tensor, num_logprobs: int, sampled_token_ids: torch.Tensor, cu_num_logits: list[int] | None = None, ) -> LogprobsTensors: assert num_logprobs >= 0 batch_size, vocab_size = logits.shape logprob_token_ids = sampled_token_ids.unsqueeze(-1) if num_logprobs > 0: topk_indices = torch.topk(logits, num_logprobs, dim=-1).indices logprob_token_ids = torch.cat((logprob_token_ids, topk_indices), dim=1) # NOTE(woosuk): Here, to save GPU memory, we do not materialize the full # logprobs tensor. Instead, we only compute and return the logprobs of # the topk + 1 tokens. logprobs = compute_token_logprobs(logits, logprob_token_ids) token_ranks = torch.empty(batch_size, dtype=torch.int64, device=logits.device) _ranks_kernel[(batch_size,)]( token_ranks, logits, logits.stride(0), sampled_token_ids, vocab_size, BLOCK_SIZE=8192, # type: ignore ) return LogprobsTensors( logprob_token_ids=logprob_token_ids, logprobs=logprobs, selected_token_ranks=token_ranks, cu_num_generated_tokens=cu_num_logits, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/worker/gpu/sample/logprob.py", "license": "Apache License 2.0", "lines": 109, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/worker/gpu/sample/sampler.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import numpy as np import torch import vllm.envs as envs from vllm.config.model import LogprobsMode from vllm.sampling_params import SamplingParams from vllm.v1.worker.gpu.metrics.logits import get_num_nans from vllm.v1.worker.gpu.sample.bad_words import BadWordsState from vllm.v1.worker.gpu.sample.gumbel import gumbel_sample from vllm.v1.worker.gpu.sample.logit_bias import LogitBiasState from vllm.v1.worker.gpu.sample.logprob import compute_topk_logprobs from vllm.v1.worker.gpu.sample.output import SamplerOutput from vllm.v1.worker.gpu.sample.penalties import PenaltiesState from vllm.v1.worker.gpu.sample.states import NO_LOGPROBS, SamplingStates from vllm.v1.worker.gpu.states import RequestState class Sampler: def __init__( self, max_num_reqs: int, vocab_size: int, device: torch.device, req_states: RequestState, logprobs_mode: LogprobsMode = "raw_logprobs", num_speculative_tokens: int = 1, ): if logprobs_mode not in ("processed_logprobs", "raw_logprobs"): raise NotImplementedError(f"Unsupported logprobs_mode: {logprobs_mode}") self.logprobs_mode = logprobs_mode self.compute_nans = envs.VLLM_COMPUTE_NANS_IN_LOGITS # False by default. self.sampling_states = SamplingStates(max_num_reqs, vocab_size) self.penalties_state = PenaltiesState(req_states) self.logit_bias_state = LogitBiasState(max_num_reqs, device) self.bad_words_state = BadWordsState(req_states) self.num_speculative_tokens = num_speculative_tokens def add_request( self, req_idx: int, prompt_len: int, sampling_params: SamplingParams ) -> None: self.sampling_states.add_request(req_idx, sampling_params) self.penalties_state.add_request(req_idx, sampling_params) self.logit_bias_state.add_request(req_idx, prompt_len, sampling_params) self.bad_words_state.add_request(req_idx, sampling_params) def apply_staged_writes(self) -> None: self.sampling_states.apply_staged_writes() self.penalties_state.apply_staged_writes() self.logit_bias_state.apply_staged_writes() self.bad_words_state.apply_staged_writes() def __call__( self, logits: torch.Tensor, idx_mapping: torch.Tensor, idx_mapping_np: np.ndarray, cu_num_logits_np: np.ndarray, pos: torch.Tensor, input_ids: torch.Tensor, expanded_local_pos: torch.Tensor, ) -> SamplerOutput: # NOTE(woosuk): We intentionally compute num_nans before sampling to make clear # that num_nans is computed before applying penalties and temperature. num_nans = get_num_nans(logits) if self.compute_nans else None sampled, processed_logits = self.sample( logits, idx_mapping, idx_mapping_np, pos, input_ids, expanded_local_pos, ) max_num_logprobs = self.sampling_states.max_num_logprobs(idx_mapping_np) if max_num_logprobs != NO_LOGPROBS: if self.logprobs_mode == "processed_logprobs": logits = processed_logits expanded_logits = logits.shape[0] != idx_mapping_np.shape[0] cu_num_logits = cu_num_logits_np.tolist() if expanded_logits else None logprobs_tensors = compute_topk_logprobs( logits, max_num_logprobs, sampled, cu_num_logits ) else: logprobs_tensors = None # These are GPU tensors. sampler_output = SamplerOutput( # The sampled tokens are expanded to 2D tensor with shape # [num_requests, 1], where each row represents one generated # token per request. sampled_token_ids=sampled.view(-1, 1), logprobs_tensors=logprobs_tensors, num_nans=num_nans, ) return sampler_output def sample( self, logits: torch.Tensor, idx_mapping: torch.Tensor, idx_mapping_np: np.ndarray, pos: torch.Tensor, input_ids: torch.Tensor, expanded_local_pos: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: # Copy logits to a new FP32 tensor. logits = torch.empty_like(logits, dtype=torch.float32).copy_(logits) # Apply logit bias (e.g., allowed_token_ids, min_tokens) in place. self.logit_bias_state.apply_logit_bias(logits, idx_mapping, idx_mapping_np, pos) # Apply penalties in place. self.penalties_state.apply_penalties( logits, idx_mapping, idx_mapping_np, input_ids, expanded_local_pos, self.num_speculative_tokens, ) # Apply bad words masking in place. self.bad_words_state.apply_bad_words( logits, idx_mapping, idx_mapping_np, input_ids, expanded_local_pos, ) # Apply temperature in place. self.sampling_states.apply_temperature(logits, idx_mapping, idx_mapping_np) # Apply min_p in place. self.sampling_states.apply_min_p(logits, idx_mapping, idx_mapping_np) # Apply top_k and/or top_p. This might or might not return a new tensor. logits = self.sampling_states.apply_top_k_top_p( logits, idx_mapping, idx_mapping_np ) # Sample the next token. sampled = gumbel_sample( logits, idx_mapping, self.sampling_states.temperature.gpu, self.sampling_states.seeds.gpu, pos, apply_temperature=False, ) return sampled, logits
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/worker/gpu/sample/sampler.py", "license": "Apache License 2.0", "lines": 137, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/engine/input_processor.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import time from collections.abc import Mapping from typing import Any, Literal import vllm.envs as envs from vllm.config import VllmConfig from vllm.inputs.data import ( ProcessorInputs, PromptType, SingletonInputs, ) from vllm.inputs.parse import split_enc_dec_inputs from vllm.inputs.preprocess import InputPreprocessor from vllm.logger import init_logger from vllm.lora.request import LoRARequest from vllm.multimodal import MULTIMODAL_REGISTRY, MultiModalRegistry from vllm.multimodal.encoder_budget import MultiModalBudget from vllm.multimodal.inputs import ( MultiModalFeatureSpec, ) from vllm.multimodal.utils import argsort_mm_positions from vllm.pooling_params import PoolingParams from vllm.renderers import BaseRenderer, renderer_from_config from vllm.sampling_params import SamplingParams from vllm.tasks import GENERATION_TASKS, POOLING_TASKS, SupportedTask from vllm.tokenizers import TokenizerLike from vllm.utils import length_from_prompt_token_ids_or_embeds, random_uuid from vllm.utils.func_utils import supports_kw from vllm.utils.jsontree import json_iter_leaves from vllm.v1.engine import EngineCoreRequest logger = init_logger(__name__) class InputProcessor: def __init__( self, vllm_config: VllmConfig, renderer: BaseRenderer | None = None, *, mm_registry: MultiModalRegistry = MULTIMODAL_REGISTRY, ) -> None: self.vllm_config = vllm_config self.model_config = model_config = vllm_config.model_config self.cache_config = vllm_config.cache_config self.lora_config = vllm_config.lora_config self.scheduler_config = vllm_config.scheduler_config self.speculative_config = vllm_config.speculative_config self.structured_outputs_config = vllm_config.structured_outputs_config self.observability_config = vllm_config.observability_config self.generation_config_fields = model_config.try_get_generation_config() self.renderer = renderer or renderer_from_config(vllm_config) self.supports_mm_inputs = mm_registry.supports_multimodal_inputs(model_config) self.mm_encoder_cache_size = 0 self.skip_prompt_length_check = False if self.supports_mm_inputs: mm_budget = MultiModalBudget(vllm_config, mm_registry) self.mm_encoder_cache_size = mm_budget.encoder_cache_size self.skip_prompt_length_check = ( mm_budget.processor.info.skip_prompt_length_check ) mm_budget.reset_cache() # Not used anymore self.input_preprocessor = InputPreprocessor( vllm_config, renderer=renderer, mm_registry=mm_registry, ) from vllm.platforms import current_platform platform_validate_request = current_platform.validate_request if supports_kw(platform_validate_request, "prompt"): logger.warning_once( "The signature of Platform.validate_request has changed from " "`(cls, prompt, params, processed_inputs) -> None` to " "`(cls, processed_inputs, params) -> None`. The old signature " "will no longer be supported starting from v0.18." ) orig_validate_request = platform_validate_request def compat_validate_request( processed_inputs: ProcessorInputs, params: SamplingParams | PoolingParams, ): return orig_validate_request( processed_inputs, params, processed_inputs, # type: ignore ) # type: ignore platform_validate_request = compat_validate_request self._platform_validate_request = platform_validate_request @property def tokenizer(self) -> TokenizerLike | None: return self.renderer.tokenizer def get_tokenizer(self) -> TokenizerLike: return self.renderer.get_tokenizer() def _validate_params( self, params: SamplingParams | PoolingParams, supported_tasks: tuple[SupportedTask, ...], ) -> None: """Raise `ValueError` if SamplingParams or PoolingParams is not valid.""" if isinstance(params, SamplingParams): supported_generation_tasks = [ task for task in supported_tasks if task in GENERATION_TASKS ] if not supported_generation_tasks: raise ValueError("This model does not support generation") params.verify( self.model_config, self.speculative_config, self.structured_outputs_config, self.tokenizer, ) elif isinstance(params, PoolingParams): supported_pooling_tasks = [ task for task in supported_tasks if task in POOLING_TASKS ] if not supported_pooling_tasks: raise ValueError("This model does not support pooling") if params.task is None: if "token_embed" in supported_pooling_tasks: params.task = "token_embed" elif "token_classify" in supported_pooling_tasks: params.task = "token_classify" elif "plugin" in supported_pooling_tasks: params.task = "plugin" if params.task not in supported_pooling_tasks: raise ValueError( f"Unsupported task: {params.task!r} " f"Supported tasks: {supported_pooling_tasks}" ) params.verify(self.model_config) else: raise TypeError( f"params must be either SamplingParams or PoolingParams, " f"but got {type(params).__name__}" ) def _validate_lora(self, lora_request: LoRARequest | None) -> None: if lora_request is None: return # LoRA request passed in while LoRA is not enabled if not self.lora_config: raise ValueError( f"Got lora_request {lora_request} but LoRA is not enabled!" ) if self.tokenizer is not None: logger.warning_once( "vLLM has deprecated support for supporting different " "tokenizers for different LoRAs. By default, vLLM uses base " "model's tokenizer. If you are using a LoRA " "with its own tokenizer, consider specifying `--tokenizer " "[lora_path]` to use the LoRA tokenizer." ) def _get_mm_identifier( self, mm_hash: str, lora_request: LoRARequest | None, ) -> str: """ When enable_tower_connector_lora is True, multi-modal embeddings vary depending on the LoRA request. Therefore, the mm_hash must be generated based on the LoRA request to prevent incorrect cache hits. """ if ( lora_request is None or self.lora_config is None or not self.lora_config.enable_tower_connector_lora ): return mm_hash return f"{lora_request.lora_name}:{mm_hash}" @staticmethod def assign_request_id(request: EngineCoreRequest): """Replace the externally supplied request ID with an internal request ID that adds 8 random characters in order to ensure uniquness. """ if request.external_req_id is not None: raise ValueError( "The external_req_id field should not be set on EngineCoreRequests" " passed to vLLM; use the request_id field." ) request.external_req_id = request.request_id if envs.VLLM_DISABLE_REQUEST_ID_RANDOMIZATION: logger.warning_once( "VLLM_DISABLE_REQUEST_ID_RANDOMIZATION is set and will be " "removed in a future release. Duplicate externally-provided " "request IDs may cause failures and/or subtle correctness errors." ) else: request.request_id = f"{request.external_req_id}-{random_uuid():.8}" def process_inputs( self, request_id: str, prompt: PromptType | ProcessorInputs, params: SamplingParams | PoolingParams, supported_tasks: tuple[SupportedTask, ...], arrival_time: float | None = None, lora_request: LoRARequest | None = None, tokenization_kwargs: dict[str, Any] | None = None, trace_headers: Mapping[str, str] | None = None, priority: int = 0, data_parallel_rank: int | None = None, resumable: bool = False, ) -> EngineCoreRequest: self._validate_params(params, supported_tasks) self._validate_lora(lora_request) parallel_config = self.vllm_config.parallel_config dp_size = parallel_config.data_parallel_size dp_local_size = parallel_config.data_parallel_size_local num_ranks = dp_local_size if parallel_config.local_engines_only else dp_size if data_parallel_rank is not None and not (0 <= data_parallel_rank < num_ranks): raise ValueError( f"data_parallel_rank {data_parallel_rank} " f"is out of range [0, {num_ranks})." ) if isinstance(prompt, dict) and "type" in prompt: if tokenization_kwargs: logger.warning_once( "Passing tokenization_kwargs to InputProcessor is deprecated " "and will be removed in v0.18. You should instead pass " "them to Renderer.render_cmpl() or Renderer.render_chat()." ) if arrival_time is None: arrival_time = prompt.get("arrival_time", time.time()) # type: ignore[assignment] processed_inputs: ProcessorInputs = prompt # type: ignore[assignment] else: logger.warning_once( "Passing raw prompts to InputProcessor is deprecated " "and will be removed in v0.18. You should instead pass " "the outputs of Renderer.render_cmpl() or Renderer.render_chat()." ) if arrival_time is None: arrival_time = time.time() processed_inputs = self.input_preprocessor.preprocess( prompt, tokenization_kwargs=tokenization_kwargs, ) self._platform_validate_request(processed_inputs, params) encoder_inputs, decoder_inputs = split_enc_dec_inputs(processed_inputs) self._validate_model_inputs(encoder_inputs, decoder_inputs) # Mypy can be conservative for TypedDict unions; normalize access. if decoder_inputs["type"] == "embeds": prompt_token_ids = None prompt_embeds = decoder_inputs["prompt_embeds"] else: prompt_token_ids = decoder_inputs["prompt_token_ids"] prompt_embeds = None sampling_params = None pooling_params = None if isinstance(params, SamplingParams): # TODO: can we avoid cloning here in multiproc case? sampling_params = params.clone() # If unset max tokens, then generate up to the max_model_len. if sampling_params.max_tokens is None: seq_len = length_from_prompt_token_ids_or_embeds( prompt_token_ids, prompt_embeds ) sampling_params.max_tokens = self.model_config.max_model_len - seq_len sampling_params.update_from_generation_config( self.generation_config_fields, self.renderer.get_eos_token_id(), ) if self.tokenizer is not None: sampling_params.update_from_tokenizer(self.tokenizer) else: pooling_params = params.clone() # Multimodal related. mm_features: list[MultiModalFeatureSpec] | None = None if decoder_inputs["type"] == "multimodal": decoder_mm_inputs = decoder_inputs["mm_kwargs"] decoder_mm_positions = decoder_inputs["mm_placeholders"] decoder_mm_hashes = decoder_inputs["mm_hashes"] if not all( isinstance(leaf, str) for leaf in json_iter_leaves(decoder_mm_hashes) ): raise ValueError( f"mm_hashes must contain only strings, got: {decoder_mm_hashes}. " "This is likely due to an incorrect custom implementation of " "MultiModalProcessor.apply method." ) # Merge and flatten multimodal placeholders, hashes and inputs # from dictionaries to lists, and sort them by each item's position # in the input sequence. sorted_mm_idxs = argsort_mm_positions(decoder_mm_positions) mm_features = [] for modality, idx in sorted_mm_idxs: base_mm_hash = decoder_mm_hashes[modality][idx] mm_features.append( MultiModalFeatureSpec( data=decoder_mm_inputs[modality][idx], modality=modality, identifier=self._get_mm_identifier( base_mm_hash, lora_request, ), mm_position=decoder_mm_positions[modality][idx], mm_hash=base_mm_hash, ) ) return EngineCoreRequest( request_id=request_id, prompt_token_ids=prompt_token_ids, prompt_embeds=prompt_embeds, mm_features=mm_features, sampling_params=sampling_params, pooling_params=pooling_params, arrival_time=arrival_time, lora_request=lora_request, cache_salt=decoder_inputs.get("cache_salt"), priority=priority, data_parallel_rank=data_parallel_rank, trace_headers=trace_headers, resumable=resumable, ) def _validate_prompt_len( self, prompt_len: int, prompt_type: Literal["encoder", "decoder"], ): if self.skip_prompt_length_check: return if prompt_len == 0 and prompt_type == "decoder": raise ValueError(f"The {prompt_type} prompt cannot be empty") model_config = self.model_config max_prompt_len = ( model_config.max_model_len if prompt_type == "decoder" else self.mm_encoder_cache_size ) if prompt_len > max_prompt_len: if self.supports_mm_inputs: suggestion = ( "Make sure that `max_model_len` is no smaller than the " "number of text tokens plus multimodal tokens. For image " "inputs, the number of image tokens depends on the number " "of images, and possibly their aspect ratios as well." ) else: suggestion = ( "Make sure that `max_model_len` is no smaller than the " "number of text tokens." ) raise ValueError( f"The {prompt_type} prompt (length {prompt_len}) is " f"longer than the maximum model length of {max_prompt_len}. " f"{suggestion}" ) elif prompt_len == max_prompt_len and model_config.runner_type == "generate": suggestion = ( "Make sure that `max_model_len` is no smaller than the " "number of text tokens (prompt + requested output tokens)." ) raise ValueError( f"The {prompt_type} prompt (length {prompt_len}) plus the number of " f"requested output tokens (at least 1) is longer than the maximum " f"model length of {max_prompt_len}. {suggestion}" ) def _validate_model_input( self, prompt_inputs: SingletonInputs, prompt_type: Literal["encoder", "decoder"], ) -> None: model_config = self.model_config tokenizer = self.tokenizer prompt_ids = ( None if prompt_inputs["type"] == "embeds" else prompt_inputs["prompt_token_ids"] ) prompt_embeds = ( prompt_inputs["prompt_embeds"] if prompt_inputs["type"] == "embeds" else None ) prompt_len = length_from_prompt_token_ids_or_embeds(prompt_ids, prompt_embeds) self._validate_prompt_len(prompt_len, prompt_type) if prompt_inputs["type"] == "multimodal": decoder_mm_positions = prompt_inputs["mm_placeholders"] for modality, mm_positions in decoder_mm_positions.items(): for mm_position in mm_positions: embed_length = mm_position.get_num_embeds() if embed_length > self.mm_encoder_cache_size: raise ValueError( f"The {prompt_type} prompt contains a(n) {modality} item " f"with length {embed_length}, which exceeds the " f"pre-allocated encoder cache size " f"{self.mm_encoder_cache_size}. Please reduce the input " f"size or increase the encoder cache size " f"by setting --limit-mm-per-prompt at startup." ) if prompt_ids and tokenizer is not None: max_input_id = max(prompt_ids, default=0) # NOTE: tokenizer.max_token_id is the tokenizer’s vocab size while # self.model_config.get_vocab_size() is the model’s vocab size. # For Qwen3 models, the language model has extra tokens that do # not exist in the tokenizer, and vice versa for multimodal # placeholder tokens in some multimodal models. # See https://github.com/QwenLM/Qwen3/issues/29#issuecomment-1933720399 # noqa: E501 # and https://github.com/vllm-project/vllm/pull/22471#discussion_r2312251421 # noqa: E501 # Here we take the max of the two to determine if a token id is # truly out-of-vocabulary. model_vocab_size = model_config.get_vocab_size() if max_input_id > max(tokenizer.max_token_id, model_vocab_size - 1): raise ValueError(f"Token id {max_input_id} is out of vocabulary") def _validate_model_inputs( self, encoder_inputs: SingletonInputs | None, decoder_inputs: SingletonInputs, ): if encoder_inputs is not None: self._validate_model_input(encoder_inputs, prompt_type="encoder") self._validate_model_input(decoder_inputs, prompt_type="decoder")
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/engine/input_processor.py", "license": "Apache License 2.0", "lines": 407, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/transformers_utils/repo_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Utilities for model repo interaction.""" import fnmatch import json import os import time from collections.abc import Callable from functools import cache from pathlib import Path from typing import TypeVar import huggingface_hub from huggingface_hub import hf_hub_download, try_to_load_from_cache from huggingface_hub import list_repo_files as hf_list_repo_files from huggingface_hub.utils import ( EntryNotFoundError, HfHubHTTPError, LocalEntryNotFoundError, RepositoryNotFoundError, RevisionNotFoundError, ) from vllm import envs from vllm.logger import init_logger logger = init_logger(__name__) _R = TypeVar("_R") def with_retry( func: Callable[[], _R], log_msg: str, max_retries: int = 2, retry_delay: int = 2, ) -> _R: for attempt in range(max_retries): try: return func() except Exception as e: if attempt == max_retries - 1: logger.error("%s: %s", log_msg, e) raise logger.error( "%s: %s, retrying %d of %d", log_msg, e, attempt + 1, max_retries ) time.sleep(retry_delay) retry_delay *= 2 raise AssertionError("Should not be reached") # @cache doesn't cache exceptions @cache def list_repo_files( repo_id: str, *, revision: str | None = None, repo_type: str | None = None, token: str | bool | None = None, ) -> list[str]: def lookup_files() -> list[str]: # directly list files if model is local if (local_path := Path(repo_id)).exists(): return [ str(file.relative_to(local_path)) for file in local_path.rglob("*") if file.is_file() ] # if model is remote, use hf_hub api to list files try: if envs.VLLM_USE_MODELSCOPE: from vllm.transformers_utils.utils import modelscope_list_repo_files return modelscope_list_repo_files( repo_id, revision=revision, token=os.getenv("MODELSCOPE_API_TOKEN", None), ) return hf_list_repo_files( repo_id, revision=revision, repo_type=repo_type, token=token ) except huggingface_hub.errors.OfflineModeIsEnabled: # Don't raise in offline mode, # all we know is that we don't have this # file cached. return [] return with_retry(lookup_files, "Error retrieving file list") def list_filtered_repo_files( model_name_or_path: str, allow_patterns: list[str], revision: str | None = None, repo_type: str | None = None, token: str | bool | None = None, ) -> list[str]: try: all_files = list_repo_files( repo_id=model_name_or_path, revision=revision, token=token, repo_type=repo_type, ) except Exception: logger.error( "Error retrieving file list. Please ensure your `model_name_or_path`" "`repo_type`, `token` and `revision` arguments are correctly set. " "Returning an empty list." ) return [] file_list = [] # Filter patterns on filenames for pattern in allow_patterns: file_list.extend( [ file for file in all_files if fnmatch.fnmatch(os.path.basename(file), pattern) ] ) return file_list def any_pattern_in_repo_files( model_name_or_path: str, allow_patterns: list[str], revision: str | None = None, repo_type: str | None = None, token: str | bool | None = None, ): return ( len( list_filtered_repo_files( model_name_or_path=model_name_or_path, allow_patterns=allow_patterns, revision=revision, repo_type=repo_type, token=token, ) ) > 0 ) def is_mistral_model_repo( model_name_or_path: str, revision: str | None = None, repo_type: str | None = None, token: str | bool | None = None, ) -> bool: return any_pattern_in_repo_files( model_name_or_path=model_name_or_path, allow_patterns=["consolidated*.safetensors"], revision=revision, repo_type=repo_type, token=token, ) def file_exists( repo_id: str, file_name: str, *, repo_type: str | None = None, revision: str | None = None, token: str | bool | None = None, ) -> bool: # `list_repo_files` is cached and retried on error, so this is more efficient than # huggingface_hub.file_exists default implementation when looking for multiple files file_list = list_repo_files( repo_id, repo_type=repo_type, revision=revision, token=token ) return file_name in file_list # In offline mode the result can be a false negative def file_or_path_exists( model: str | Path, config_name: str, revision: str | None ) -> bool: if (local_path := Path(model)).exists(): return (local_path / config_name).is_file() # Offline mode support: Check if config file is cached already cached_filepath = try_to_load_from_cache( repo_id=model, filename=config_name, revision=revision ) if isinstance(cached_filepath, str): # The config file exists in cache - we can continue trying to load return True # NB: file_exists will only check for the existence of the config file on # hf_hub. This will fail in offline mode. # Call HF to check if the file exists return file_exists(str(model), config_name, revision=revision) def get_model_path(model: str | Path, revision: str | None = None): if os.path.exists(model): return model assert huggingface_hub.constants.HF_HUB_OFFLINE common_kwargs = { "local_files_only": huggingface_hub.constants.HF_HUB_OFFLINE, "revision": revision, } if envs.VLLM_USE_MODELSCOPE: from modelscope.hub.snapshot_download import snapshot_download return snapshot_download(model_id=model, **common_kwargs) from huggingface_hub import snapshot_download return snapshot_download(repo_id=model, **common_kwargs) def get_hf_file_bytes( file_name: str, model: str | Path, revision: str | None = "main" ) -> bytes | None: """Get file contents from HuggingFace repository as bytes.""" file_path = try_get_local_file(model=model, file_name=file_name, revision=revision) if file_path is None: hf_hub_file = hf_hub_download(model, file_name, revision=revision) file_path = Path(hf_hub_file) if file_path is not None and file_path.is_file(): with open(file_path, "rb") as file: return file.read() return None def try_get_local_file( model: str | Path, file_name: str, revision: str | None = "main" ) -> Path | None: file_path = Path(model) / file_name if file_path.is_file(): return file_path else: try: cached_filepath = try_to_load_from_cache( repo_id=model, filename=file_name, revision=revision ) if isinstance(cached_filepath, str): return Path(cached_filepath) except ValueError: ... return None def get_hf_file_to_dict( file_name: str, model: str | Path, revision: str | None = "main" ): """ Downloads a file from the Hugging Face Hub and returns its contents as a dictionary. Parameters: - file_name (str): The name of the file to download. - model (str): The name of the model on the Hugging Face Hub. - revision (str): The specific version of the model. Returns: - config_dict (dict): A dictionary containing the contents of the downloaded file. """ file_path = try_get_local_file(model=model, file_name=file_name, revision=revision) if file_path is None: try: hf_hub_file = hf_hub_download(model, file_name, revision=revision) except huggingface_hub.errors.OfflineModeIsEnabled: return None except ( RepositoryNotFoundError, RevisionNotFoundError, EntryNotFoundError, LocalEntryNotFoundError, ) as e: logger.debug("File or repository not found in hf_hub_download:", exc_info=e) return None except HfHubHTTPError as e: logger.warning( "Cannot connect to Hugging Face Hub. Skipping file download for '%s':", file_name, exc_info=e, ) return None file_path = Path(hf_hub_file) if file_path is not None and file_path.is_file(): with open(file_path) as file: return json.load(file) return None
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/repo_utils.py", "license": "Apache License 2.0", "lines": 255, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/benchmarks/sweep/plot_pareto.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import math from concurrent.futures import ProcessPoolExecutor from dataclasses import dataclass from functools import partial from pathlib import Path from typing import ClassVar from vllm.utils.collection_utils import full_groupby from vllm.utils.import_utils import PlaceholderModule from .plot import DummyExecutor, _json_load_bytes from .utils import sanitize_filename try: import matplotlib.pyplot as plt except ImportError: plt = PlaceholderModule("matplotlib").placeholder_attr("pyplot") try: import pandas as pd except ImportError: pd = PlaceholderModule("pandas") try: import seaborn as sns except ImportError: seaborn = PlaceholderModule("seaborn") def _first_present(run_data: dict[str, object], keys: list[str]): for key in keys: for candidate in {key, key.replace("_", "-"), key.replace("-", "_")}: if candidate in run_data: return run_data[candidate] return None def _get_numeric( run_data: dict[str, object], keys: list[str], *, allow_zero: bool = True, ) -> float | None: value = _first_present(run_data, keys) if value is None: return None try: numeric = float(value) except (TypeError, ValueError) as exc: raise ValueError( f"Expected numeric value for one of {keys}, " f"but found {value!r} in {run_data=}" ) from exc if not allow_zero and numeric == 0: return None return numeric def _infer_user_count( run_data: dict[str, object], user_count_var: str | None, ) -> float | None: candidates = [user_count_var] if user_count_var else [] candidates.extend(["request_rate"]) user_count = _get_numeric(run_data, candidates, allow_zero=False) if user_count is not None: return user_count # Fallback to the observed peak if configured value is missing. return _get_numeric(run_data, ["max_concurrent_requests"], allow_zero=False) def _infer_gpu_count( run_data: dict[str, object], gpu_count_var: str | None, ) -> float: direct_candidates = [gpu_count_var] if gpu_count_var else [] direct_gpu_count = _get_numeric(run_data, direct_candidates, allow_zero=False) if direct_gpu_count: return direct_gpu_count tp_size = _get_numeric(run_data, ["tensor_parallel_size", "tp"]) pp_size = _get_numeric(run_data, ["pipeline_parallel_size", "pp"]) dp_size = _get_numeric(run_data, ["data_parallel_size", "dp"]) world_size = 1.0 if tp_size: world_size *= tp_size if pp_size: world_size *= pp_size if dp_size: world_size *= dp_size return world_size def _get_throughput( run_data: dict[str, object], throughput_var: str, ) -> float: throughput = _get_numeric(run_data, [throughput_var]) if throughput is None: raise ValueError( f"Cannot find throughput metric {throughput_var!r} in run data. " f"Available keys: {sorted(run_data)}" ) return throughput def _prepare_records( all_data: list[dict[str, object]], *, user_count_var: str | None, gpu_count_var: str | None, ) -> tuple[list[dict[str, object]], int]: prepared = [] skipped_missing_users = 0 for record in all_data: throughput = _get_throughput(record, "output_throughput") user_count = _infer_user_count(record, user_count_var) if user_count is None: skipped_missing_users += 1 continue gpu_count = _infer_gpu_count(record, gpu_count_var) tokens_per_user = throughput / user_count tokens_per_gpu = throughput / gpu_count prepared.append( { **record, "tokens_per_user": tokens_per_user, "tokens_per_gpu": tokens_per_gpu, "user_count_estimate": user_count, "gpu_count": gpu_count, } ) return prepared, skipped_missing_users def _pareto_frontier( df: "pd.DataFrame", x_col: str, y_col: str, *, epsilon: float = 1e-9, ) -> "pd.DataFrame": sorted_df = df.sort_values([x_col, y_col], ascending=[False, False]) frontier_indices = [] best_y = -math.inf for idx, row in sorted_df.iterrows(): y_val = row[y_col] if y_val >= best_y - epsilon: frontier_indices.append(idx) best_y = max(best_y, y_val) return df.loc[frontier_indices] def _get_fig_path( fig_dir: Path, fig_group: tuple[tuple[str, str], ...], ) -> Path: parts = ["PARETO"] if fig_group: parts.extend(f"{k}={v}" for k, v in fig_group) filename = sanitize_filename("-".join(parts) + ".png") return fig_dir / filename def _plot_fig( fig_dir: Path, fig_group_data: tuple[tuple[tuple[str, str], ...], list[dict[str, object]]], label_by: list[str], *, dry_run: bool, ): fig_group, fig_data = fig_group_data fig_path = _get_fig_path(fig_dir, fig_group) print("[BEGIN FIGURE]") print(f"Group: {dict(fig_group)}") print(f"Output file: {fig_path}") if dry_run: print("[END FIGURE]") return df = pd.DataFrame.from_records(fig_data) df = df.dropna(subset=["tokens_per_user", "tokens_per_gpu"]) if df.empty: print("No data points available after filtering; skipping.") print("[END FIGURE]") return frontier = _pareto_frontier(df, "tokens_per_user", "tokens_per_gpu") frontier = frontier.sort_values("tokens_per_user") fig, ax = plt.subplots() sns.scatterplot( data=df, x="tokens_per_user", y="tokens_per_gpu", color="0.5", alpha=0.6, ax=ax, label="All runs", ) sns.lineplot( data=frontier, x="tokens_per_user", y="tokens_per_gpu", marker="o", ax=ax, label="Pareto frontier", ) if label_by: for _, row in frontier.iterrows(): label_parts = [] for key in label_by: if key in row: label_parts.append(f"{key}={row[key]}") if label_parts: ax.text( row["tokens_per_user"], row["tokens_per_gpu"], "\n".join(label_parts), fontsize=8, ) ax.set_xlabel("Tokens/s/user") ax.set_ylabel("Tokens/s/GPU") ax.grid(True, linestyle="--", linewidth=0.5, alpha=0.6) fig.tight_layout() fig.savefig(fig_path) plt.close(fig) print( f"Plotted {len(df)} points; Pareto frontier size: {len(frontier)}.", ) print("[END FIGURE]") def plot_pareto( output_dir: Path, user_count_var: str | None, gpu_count_var: str | None, label_by: list[str], *, dry_run: bool, ): fig_dir = output_dir / "pareto" raw_data = [ run_data for path in output_dir.rglob("**/summary.json") for run_data in _json_load_bytes(path) ] if not raw_data: raise ValueError(f"Did not find any parameter sweep results under {output_dir}") fig_dir.mkdir(parents=True, exist_ok=True) prepared_data, skipped_missing_users = _prepare_records( raw_data, user_count_var=user_count_var, gpu_count_var=gpu_count_var, ) if skipped_missing_users: print( f"Skipped {skipped_missing_users} runs without a user count " "(`max_concurrency` or `max_concurrent_requests`).", ) if not prepared_data: raise ValueError( "No data points with both throughput and user count available " "to plot Pareto frontier.", ) fig_groups = full_groupby( prepared_data, key=lambda item: tuple(), ) with DummyExecutor() if len(fig_groups) <= 1 else ProcessPoolExecutor() as executor: all( executor.map( partial( _plot_fig, fig_dir, label_by=label_by, dry_run=dry_run, ), fig_groups, ) ) @dataclass class SweepPlotParetoArgs: output_dir: Path user_count_var: str | None gpu_count_var: str | None label_by: list[str] dry_run: bool parser_name: ClassVar[str] = "plot_pareto" parser_help: ClassVar[str] = ( "Plot Pareto frontier between tokens/s/user and tokens/s/GPU " "from parameter sweep results." ) @classmethod def from_cli_args(cls, args: argparse.Namespace): output_dir = Path(args.EXPERIMENT_DIR) if not output_dir.exists(): raise ValueError(f"No parameter sweep results under {output_dir}") label_by = [] if not args.label_by else args.label_by.split(",") return cls( output_dir=output_dir, user_count_var=args.user_count_var, gpu_count_var=args.gpu_count_var, label_by=label_by, dry_run=args.dry_run, ) @classmethod def add_cli_args(cls, parser: argparse.ArgumentParser): parser.add_argument( "EXPERIMENT_DIR", type=str, help="The directory containing the sweep results to plot.", ) parser.add_argument( "--user-count-var", type=str, default="max_concurrency", help="Result key that stores concurrent user count. " "Falls back to max_concurrent_requests if missing.", ) parser.add_argument( "--gpu-count-var", type=str, default=None, help="Result key that stores GPU count. " "If not provided, falls back to num_gpus/gpu_count " "or tensor_parallel_size * pipeline_parallel_size.", ) parser.add_argument( "--label-by", type=str, default="max_concurrency,gpu_count", help="Comma-separated list of fields to annotate on Pareto frontier " "points.", ) parser.add_argument( "--dry-run", action="store_true", help="If set, prints the figures to plot without drawing them.", ) return parser def run_main(args: SweepPlotParetoArgs): return plot_pareto( output_dir=args.output_dir, user_count_var=args.user_count_var, gpu_count_var=args.gpu_count_var, label_by=args.label_by, dry_run=args.dry_run, ) def main(args: argparse.Namespace): run_main(SweepPlotParetoArgs.from_cli_args(args)) if __name__ == "__main__": parser = argparse.ArgumentParser(description=SweepPlotParetoArgs.parser_help) SweepPlotParetoArgs.add_cli_args(parser) main(parser.parse_args())
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/benchmarks/sweep/plot_pareto.py", "license": "Apache License 2.0", "lines": 329, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/core/test_output.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.v1.core.sched.output import NewRequestData def _create_new_requests_data(prompt_embeds: torch.Tensor | None) -> NewRequestData: return NewRequestData( req_id="test_req", prompt_token_ids=None, mm_features=[], sampling_params=None, pooling_params=None, block_ids=([],), num_computed_tokens=0, lora_request=None, prompt_embeds=prompt_embeds, ) def test_repr_with_none() -> None: """Test repr when prompt_embeds is None.""" new_requests_data = _create_new_requests_data(None) assert "prompt_embeds_shape=None" in repr(new_requests_data) assert "prompt_embeds_shape=None" in new_requests_data.anon_repr() def test_repr_with_multi_element_tensor() -> None: """Test repr when prompt_embeds is a multi-element tensor.""" prompt_embeds = torch.randn(10, 768) new_requests_data = _create_new_requests_data(prompt_embeds) assert "prompt_embeds_shape=torch.Size([10, 768])" in repr(new_requests_data) assert "prompt_embeds_shape=torch.Size([10, 768])" in new_requests_data.anon_repr()
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/core/test_output.py", "license": "Apache License 2.0", "lines": 27, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/entrypoints/openai/test_transcription_validation_whisper.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # imports for structured outputs tests import asyncio import io import json import librosa import numpy as np import openai import pytest import pytest_asyncio import soundfile as sf from ...utils import RemoteOpenAIServer MODEL_NAME = "openai/whisper-large-v3-turbo" @pytest.fixture(scope="module") def server(): with RemoteOpenAIServer(MODEL_NAME, []) as remote_server: yield remote_server @pytest_asyncio.fixture async def whisper_client(server): async with server.get_async_client() as async_client: yield async_client @pytest.mark.asyncio async def test_basic_audio(whisper_client, mary_had_lamb): # Based on https://github.com/openai/openai-cookbook/blob/main/examples/Whisper_prompting_guide.ipynb. transcription = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=mary_had_lamb, language="en", response_format="text", temperature=0.0, ) out = json.loads(transcription) out_text = out["text"] out_usage = out["usage"] assert "Mary had a little lamb," in out_text assert out_usage["seconds"] == 16, out_usage["seconds"] @pytest.mark.asyncio async def test_basic_audio_batched(mary_had_lamb, winning_call, whisper_client): transcription = whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=mary_had_lamb, language="en", response_format="text", temperature=0.0, ) transcription2 = whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=winning_call, language="en", response_format="text", temperature=0.0, ) # Await both transcriptions by scheduling coroutines together transcription, transcription2 = await asyncio.gather(transcription, transcription2) out = json.loads(transcription) out_text = out["text"] assert "Mary had a little lamb," in out_text out2 = json.loads(transcription2) out_text2 = out2["text"] assert "Edgar Martinez" in out_text2 @pytest.mark.asyncio async def test_bad_requests(mary_had_lamb, whisper_client): # invalid language with pytest.raises(openai.BadRequestError): await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=mary_had_lamb, language="hh", temperature=0.0 ) @pytest.mark.asyncio async def test_long_audio_request(mary_had_lamb, whisper_client): mary_had_lamb.seek(0) audio, sr = librosa.load(mary_had_lamb) # Add small silence after each audio for repeatability in the split process audio = np.pad(audio, (0, 1600)) repeated_audio = np.tile(audio, 10) # Repeated audio to buffer buffer = io.BytesIO() sf.write(buffer, repeated_audio, sr, format="WAV") buffer.seek(0) transcription = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=buffer, language="en", response_format="text", temperature=0.0, ) out = json.loads(transcription) out_text = out["text"] out_usage = out["usage"] counts = out_text.count("Mary had a little lamb") assert counts == 10, counts assert out_usage["seconds"] == 161, out_usage["seconds"] @pytest.mark.asyncio async def test_completion_endpoints(whisper_client): # text to text model with pytest.raises(openai.NotFoundError): await whisper_client.chat.completions.create( model=MODEL_NAME, messages=[{"role": "system", "content": "You are a helpful assistant."}], ) with pytest.raises(openai.NotFoundError): await whisper_client.completions.create(model=MODEL_NAME, prompt="Hello") @pytest.mark.asyncio async def test_streaming_response(winning_call, whisper_client): transcription = "" res_no_stream = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=winning_call, response_format="json", language="en", temperature=0.0, ) res = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=winning_call, language="en", temperature=0.0, stream=True, timeout=30, ) # Reconstruct from chunks and validate async for chunk in res: text = chunk.choices[0]["delta"]["content"] transcription += text assert transcription == res_no_stream.text @pytest.mark.asyncio async def test_stream_options(winning_call, whisper_client): res = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=winning_call, language="en", temperature=0.0, stream=True, extra_body=dict(stream_include_usage=True, stream_continuous_usage_stats=True), timeout=30, ) final = False continuous = True async for chunk in res: if not len(chunk.choices): # final usage sent final = True else: continuous = continuous and hasattr(chunk, "usage") assert final and continuous @pytest.mark.asyncio async def test_sampling_params(mary_had_lamb, whisper_client): """ Compare sampling with params and greedy sampling to assert results are different when extreme sampling parameters values are picked. """ transcription = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=mary_had_lamb, language="en", temperature=0.8, extra_body=dict( seed=42, repetition_penalty=1.9, top_k=12, top_p=0.4, min_p=0.5, frequency_penalty=1.8, presence_penalty=2.0, ), ) greedy_transcription = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=mary_had_lamb, language="en", temperature=0.0, extra_body=dict(seed=42), ) assert greedy_transcription.text != transcription.text @pytest.mark.asyncio async def test_audio_prompt(mary_had_lamb, whisper_client): prompt = "This is a speech, recorded in a phonograph." # Prompts should not omit the part of original prompt while transcribing. prefix = "The first words I spoke in the original phonograph" transcription = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=mary_had_lamb, language="en", response_format="text", temperature=0.0, ) out = json.loads(transcription)["text"] assert prefix in out transcription_wprompt = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=mary_had_lamb, language="en", response_format="text", prompt=prompt, temperature=0.0, ) out_prompt = json.loads(transcription_wprompt)["text"] assert prefix in out_prompt @pytest.mark.asyncio async def test_audio_with_timestamp(mary_had_lamb, whisper_client): transcription = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=mary_had_lamb, language="en", response_format="verbose_json", temperature=0.0, ) assert transcription.segments is not None assert len(transcription.segments) > 0 assert transcription.segments[0].avg_logprob is not None assert transcription.segments[0].compression_ratio is not None @pytest.mark.asyncio async def test_audio_with_max_tokens(whisper_client, mary_had_lamb): transcription = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=mary_had_lamb, language="en", response_format="text", temperature=0.0, extra_body={"max_completion_tokens": 1}, ) out = json.loads(transcription) out_text = out["text"] from transformers import AutoTokenizer tok = AutoTokenizer.from_pretrained(MODEL_NAME) out_tokens = tok(out_text, add_special_tokens=False)["input_ids"] assert len(out_tokens) == 1 # max_completion_tokens > max_model_len transcription = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=mary_had_lamb, language="en", response_format="text", temperature=0.0, extra_body={"max_completion_tokens": int(1e6)}, ) out = json.loads(transcription) out_text = out["text"] out_tokens = tok(out_text, add_special_tokens=False)["input_ids"] assert len(out_tokens) < 450 # ~Whisper max output len @pytest.mark.asyncio @pytest.mark.parametrize( ("fixture_name", "expected_lang", "expected_text"), [ ("mary_had_lamb", "en", ["Mary had a little lamb"]), ("foscolo", "it", ["zacinto", "sacre"]), ], ids=["english", "italian"], ) async def test_language_auto_detect( whisper_client, fixture_name, expected_lang, expected_text, request ): """Auto-detect language when no language param is provided.""" audio_file = request.getfixturevalue(fixture_name) transcription = await whisper_client.audio.transcriptions.create( model=MODEL_NAME, file=audio_file, response_format="verbose_json", temperature=0.0, ) assert transcription.language == expected_lang text_lower = transcription.text.lower() assert any(word.lower() in text_lower for word in expected_text), ( f"Expected {expected_lang} text but got: {transcription.text}" )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/test_transcription_validation_whisper.py", "license": "Apache License 2.0", "lines": 265, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/models/test_gguf_download.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from unittest.mock import MagicMock, patch import pytest from vllm.config import ModelConfig from vllm.config.load import LoadConfig from vllm.model_executor.model_loader.gguf_loader import GGUFModelLoader from vllm.model_executor.model_loader.weight_utils import download_gguf class TestGGUFDownload: """Test GGUF model downloading functionality.""" @patch("vllm.model_executor.model_loader.weight_utils.download_weights_from_hf") def test_download_gguf_single_file(self, mock_download): """Test downloading a single GGUF file.""" # Setup mock mock_folder = "/tmp/mock_cache" mock_download.return_value = mock_folder # Mock glob to return a single file with patch("glob.glob") as mock_glob: mock_glob.side_effect = lambda pattern, **kwargs: ( [f"{mock_folder}/model-IQ1_S.gguf"] if "IQ1_S" in pattern else [] ) result = download_gguf("unsloth/Qwen3-0.6B-GGUF", "IQ1_S") # Verify download_weights_from_hf was called with correct patterns mock_download.assert_called_once_with( model_name_or_path="unsloth/Qwen3-0.6B-GGUF", cache_dir=None, allow_patterns=[ "*-IQ1_S.gguf", "*-IQ1_S-*.gguf", "*/*-IQ1_S.gguf", "*/*-IQ1_S-*.gguf", ], revision=None, ignore_patterns=None, ) # Verify result is the file path, not folder assert result == f"{mock_folder}/model-IQ1_S.gguf" @patch("vllm.model_executor.model_loader.weight_utils.download_weights_from_hf") def test_download_gguf_sharded_files(self, mock_download): """Test downloading sharded GGUF files.""" mock_folder = "/tmp/mock_cache" mock_download.return_value = mock_folder # Mock glob to return sharded files with patch("glob.glob") as mock_glob: mock_glob.side_effect = lambda pattern, **kwargs: ( [ f"{mock_folder}/model-Q2_K-00001-of-00002.gguf", f"{mock_folder}/model-Q2_K-00002-of-00002.gguf", ] if "Q2_K" in pattern else [] ) result = download_gguf("unsloth/gpt-oss-120b-GGUF", "Q2_K") # Should return the first file after sorting assert result == f"{mock_folder}/model-Q2_K-00001-of-00002.gguf" @patch("vllm.model_executor.model_loader.weight_utils.download_weights_from_hf") def test_download_gguf_subdir(self, mock_download): """Test downloading GGUF files from subdirectory.""" mock_folder = "/tmp/mock_cache" mock_download.return_value = mock_folder with patch("glob.glob") as mock_glob: mock_glob.side_effect = lambda pattern, **kwargs: ( [f"{mock_folder}/Q2_K/model-Q2_K.gguf"] if "Q2_K" in pattern or "**/*.gguf" in pattern else [] ) result = download_gguf("unsloth/gpt-oss-120b-GGUF", "Q2_K") assert result == f"{mock_folder}/Q2_K/model-Q2_K.gguf" @patch("vllm.model_executor.model_loader.weight_utils.download_weights_from_hf") @patch("glob.glob", return_value=[]) def test_download_gguf_no_files_found(self, mock_glob, mock_download): """Test error when no GGUF files are found.""" mock_folder = "/tmp/mock_cache" mock_download.return_value = mock_folder with pytest.raises(ValueError, match="Downloaded GGUF files not found"): download_gguf("unsloth/Qwen3-0.6B-GGUF", "IQ1_S") class TestGGUFModelLoader: """Test GGUFModelLoader class methods.""" @patch("os.path.isfile", return_value=True) def test_prepare_weights_local_file(self, mock_isfile): """Test _prepare_weights with local file.""" load_config = LoadConfig(load_format="gguf") loader = GGUFModelLoader(load_config) # Create a simple mock ModelConfig with only the model attribute model_config = MagicMock() model_config.model = "/path/to/model.gguf" result = loader._prepare_weights(model_config) assert result == "/path/to/model.gguf" mock_isfile.assert_called_once_with("/path/to/model.gguf") @patch("vllm.model_executor.model_loader.gguf_loader.hf_hub_download") @patch("os.path.isfile", return_value=False) def test_prepare_weights_repo_filename(self, mock_isfile, mock_hf_download): """Test _prepare_weights with repo_id/filename.gguf format.""" load_config = LoadConfig(load_format="gguf") loader = GGUFModelLoader(load_config) mock_hf_download.return_value = "/downloaded/model.gguf" # Create a simple mock ModelConfig with only the model attribute model_config = MagicMock() model_config.model = "unsloth/Qwen3-0.6B-GGUF/model.gguf" result = loader._prepare_weights(model_config) assert result == "/downloaded/model.gguf" mock_hf_download.assert_called_once_with( repo_id="unsloth/Qwen3-0.6B-GGUF", filename="model.gguf" ) @patch("vllm.config.model.get_hf_image_processor_config", return_value=None) @patch("vllm.transformers_utils.config.file_or_path_exists", return_value=True) @patch("vllm.config.model.get_config") @patch("vllm.config.model.is_gguf", return_value=True) @patch("vllm.model_executor.model_loader.gguf_loader.download_gguf") @patch("os.path.isfile", return_value=False) def test_prepare_weights_repo_quant_type( self, mock_isfile, mock_download_gguf, mock_is_gguf, mock_get_config, mock_file_exists, mock_get_image_config, ): """Test _prepare_weights with repo_id:quant_type format.""" mock_hf_config = MagicMock() mock_hf_config.architectures = ["Qwen3ForCausalLM"] class MockTextConfig: max_position_embeddings = 4096 sliding_window = None model_type = "qwen3" num_attention_heads = 32 mock_text_config = MockTextConfig() mock_hf_config.get_text_config.return_value = mock_text_config mock_hf_config.dtype = "bfloat16" mock_get_config.return_value = mock_hf_config load_config = LoadConfig(load_format="gguf") loader = GGUFModelLoader(load_config) mock_download_gguf.return_value = "/downloaded/model-IQ1_S.gguf" model_config = ModelConfig( model="unsloth/Qwen3-0.6B-GGUF:IQ1_S", tokenizer="Qwen/Qwen3-0.6B" ) result = loader._prepare_weights(model_config) # The actual result will be the downloaded file path from mock assert result == "/downloaded/model-IQ1_S.gguf" mock_download_gguf.assert_called_once_with( "unsloth/Qwen3-0.6B-GGUF", "IQ1_S", cache_dir=None, revision=None, ignore_patterns=["original/**/*"], ) @patch("vllm.config.model.get_hf_image_processor_config", return_value=None) @patch("vllm.config.model.get_config") @patch("vllm.config.model.is_gguf", return_value=False) @patch("vllm.transformers_utils.gguf_utils.check_gguf_file", return_value=False) @patch("os.path.isfile", return_value=False) def test_prepare_weights_invalid_format( self, mock_isfile, mock_check_gguf, mock_is_gguf, mock_get_config, mock_get_image_config, ): """Test _prepare_weights with invalid format.""" mock_hf_config = MagicMock() mock_hf_config.architectures = ["Qwen3ForCausalLM"] class MockTextConfig: max_position_embeddings = 4096 sliding_window = None model_type = "qwen3" num_attention_heads = 32 mock_text_config = MockTextConfig() mock_hf_config.get_text_config.return_value = mock_text_config mock_hf_config.dtype = "bfloat16" mock_get_config.return_value = mock_hf_config load_config = LoadConfig(load_format="gguf") loader = GGUFModelLoader(load_config) # Create ModelConfig with a valid repo_id to avoid validation errors # Then test _prepare_weights with invalid format model_config = ModelConfig(model="unsloth/Qwen3-0.6B") # Manually set model to invalid format after creation model_config.model = "invalid-format" with pytest.raises(ValueError, match="Unrecognised GGUF reference"): loader._prepare_weights(model_config)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/test_gguf_download.py", "license": "Apache License 2.0", "lines": 181, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/entrypoints/openai/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import TypeVar from fastapi import Request from fastapi.exceptions import RequestValidationError from vllm.entrypoints.openai.chat_completion.protocol import ( ChatCompletionRequest, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ) # Used internally _ChatCompletionResponseChoiceT = TypeVar( "_ChatCompletionResponseChoiceT", ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ) def maybe_filter_parallel_tool_calls( choice: _ChatCompletionResponseChoiceT, request: ChatCompletionRequest ) -> _ChatCompletionResponseChoiceT: """Filter to first tool call only when parallel_tool_calls is False.""" if request.parallel_tool_calls: return choice if isinstance(choice, ChatCompletionResponseChoice) and choice.message.tool_calls: choice.message.tool_calls = choice.message.tool_calls[:1] elif ( isinstance(choice, ChatCompletionResponseStreamChoice) and choice.delta.tool_calls ): choice.delta.tool_calls = [ tool_call for tool_call in choice.delta.tool_calls if tool_call.index == 0 ] return choice async def validate_json_request(raw_request: Request): content_type = raw_request.headers.get("content-type", "").lower() media_type = content_type.split(";", maxsplit=1)[0] if media_type != "application/json": raise RequestValidationError( errors=["Unsupported Media Type: Only 'application/json' is allowed"] )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/openai/utils.py", "license": "Apache License 2.0", "lines": 39, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/distributed/test_eagle_dp.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import asyncio import os from contextlib import AsyncExitStack from dataclasses import replace import pytest from vllm import SamplingParams from vllm.engine.arg_utils import AsyncEngineArgs from vllm.platforms import current_platform from vllm.sampling_params import RequestOutputKind from vllm.v1.engine.async_llm import AsyncLLM DP_SIZE = int(os.getenv("DP_SIZE", 2)) if current_platform.is_rocm(): ATTN_BACKENDS = ["ROCM_ATTN", "TRITON_ATTN", "FLEX_ATTENTION"] else: ATTN_BACKENDS = ["FLASH_ATTN"] @pytest.mark.asyncio @pytest.mark.parametrize("attn_backend", ATTN_BACKENDS) @pytest.mark.xfail( current_platform.is_rocm(), reason="Test may fail on ROCm until batch invariance is enabled." "See: https://github.com/vllm-project/vllm/issues/27433", strict=False, ) async def test_run_eagle_dp(monkeypatch: pytest.MonkeyPatch, attn_backend: str): if not current_platform.is_rocm(): # This test checks that running a model with and without eagle # leads to identical tokens. # # NOTE: This is only true in batch invariant mode # (because the target model verifies all draft tokens in one big # forward pass) # # TODO[ROCm]: Test is passing on ROCm CI but may break in future. # Enable batch invariance for ROCm when possible. See: # https://github.com/vllm-project/vllm/issues/27433 monkeypatch.setenv("VLLM_BATCH_INVARIANT", "1") target_model = "meta-llama/Llama-3.1-8B-Instruct" draft_model = "yuhuili/EAGLE-LLaMA3.1-Instruct-8B" engine_args = AsyncEngineArgs( model=target_model, tokenizer_mode="auto", enforce_eager=False, tensor_parallel_size=int(os.getenv("TP_SIZE", 1)), data_parallel_size=DP_SIZE, data_parallel_backend="mp", # ray takes more time trust_remote_code=True, max_model_len=16384, attention_config={"backend": attn_backend}, ) eagle_engine_args = replace( engine_args, speculative_config={ "model": draft_model, "method": "eagle", "num_speculative_tokens": 3, }, ) prompt = "This is a test of data parallel with eagle" # This test might be flaky, see # https://github.com/vllm-project/vllm/issues/31913 num_expected_tokens = 20 sampling_params = SamplingParams( max_tokens=num_expected_tokens, ignore_eos=True, output_kind=RequestOutputKind.FINAL_ONLY, temperature=0, ) async def generate_with_timeout(given_engine: AsyncLLM): async for out in given_engine.generate( request_id="test-eagle-dp", prompt=prompt, sampling_params=sampling_params ): token_ids = out.outputs[0].token_ids assert len(token_ids) == num_expected_tokens return token_ids async def engine_create_and_generate(engine_args: AsyncEngineArgs): async with AsyncExitStack() as after: engine = AsyncLLM.from_engine_args(engine_args) after.callback(engine.shutdown) token_ids = await asyncio.wait_for( generate_with_timeout(engine), timeout=30 ) assert not engine.output_processor.has_unfinished_requests() return token_ids token_ids_with_eagle = await engine_create_and_generate(eagle_engine_args) token_ids_no_eagle = await engine_create_and_generate(engine_args) # Test for correctness assert token_ids_with_eagle == token_ids_no_eagle
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/distributed/test_eagle_dp.py", "license": "Apache License 2.0", "lines": 89, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/xdrope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import numpy as np import torch from .dynamic_ntk_alpha_rope import DynamicNTKAlphaRotaryEmbedding class XDRotaryEmbedding(DynamicNTKAlphaRotaryEmbedding): """DynamicNTKAlphaRotaryEmbedding extended with MultiModal(XD) Sections. Based on the original DynamicNTKAlphaRotaryEmbedding implementation. """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_alpha: float, dtype: torch.dtype, xdrope_section: list[int], ) -> None: self.xdrope_section = xdrope_section super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, scaling_alpha, dtype, ) def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """PyTorch-native implementation equivalent to forward(). Args: positions: [4, num_tokens] (P/W/H/T positions with multimodal inputs) query: [num_tokens, num_heads * head_size] key: [num_tokens, num_kv_heads * head_size] """ assert positions.ndim == 2 assert key is not None num_tokens = positions.shape[-1] cos_sin = self.cos_sin_cache[positions] cos, sin = cos_sin.chunk(2, dim=-1) cos = torch.cat( [m[i] for i, m in enumerate(cos.split(self.xdrope_section, dim=-1))], dim=-1 ) sin = torch.cat( [m[i] for i, m in enumerate(sin.split(self.xdrope_section, dim=-1))], dim=-1 ) query_shape = query.shape query = query.view(num_tokens, -1, self.head_size) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = self.apply_rotary_emb.forward_native( query_rot, cos, sin, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) key_shape = key.shape key = key.view(num_tokens, -1, self.head_size) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = self.apply_rotary_emb.forward_native( key_rot, cos, sin, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key def forward_cuda( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """PyTorch-native implementation equivalent to forward(). Args: positions: [4, num_tokens] (P/W/H/T positions with multimodal inputs) query: [num_tokens, num_heads * head_size] key: [num_tokens, num_kv_heads * head_size] """ assert positions.ndim == 2 assert key is not None num_tokens = positions.shape[-1] cos_sin = self.cos_sin_cache[positions] cos, sin = cos_sin.chunk(2, dim=-1) cos = torch.cat( [m[i] for i, m in enumerate(cos.split(self.xdrope_section, dim=-1))], dim=-1 ) sin = torch.cat( [m[i] for i, m in enumerate(sin.split(self.xdrope_section, dim=-1))], dim=-1 ) query_shape = query.shape query = query.view(num_tokens, -1, self.head_size) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = self.apply_rotary_emb( query_rot, cos, sin, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) key_shape = key.shape key = key.view(num_tokens, -1, self.head_size) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = self.apply_rotary_emb( key_rot, cos, sin, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key @staticmethod def get_next_input_positions( context_len: int, seq_len: int, xd_sections: int = 4, ) -> list[list[int]]: return [list(range(context_len, seq_len)) for _ in range(xd_sections)] @staticmethod def get_next_input_positions_tensor( out: np.ndarray, out_offset: int, context_len: int, num_new_tokens: int, ): values = np.arange( context_len, context_len + num_new_tokens, dtype=out.dtype, ) out[:, out_offset : out_offset + num_new_tokens] = values
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/xdrope.py", "license": "Apache License 2.0", "lines": 142, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/hunyuan_vision.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # coding=utf-8 # Copyright 2025 The HunYuan team. # Copyright 2025 The vLLM team. # Copyright 2025 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Inference-only HunYuan-VL model compatible with HuggingFace weights.""" from collections.abc import Callable, Iterable, Mapping, Sequence from functools import partial from typing import Annotated, Any, Literal, TypeAlias import torch import torch.nn as nn import torch.nn.functional as F from transformers import BatchFeature from vllm.config import VllmConfig from vllm.config.multimodal import BaseDummyOptions from vllm.distributed import parallel_state from vllm.distributed import utils as dist_utils from vllm.logger import init_logger from vllm.model_executor.layers.activation import get_act_fn from vllm.model_executor.layers.attention import MMEncoderAttention from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( ColumnParallelLinear, QKVParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.model_loader.weight_utils import default_weight_loader from vllm.model_executor.models.module_mapping import MultiModelKeys from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import ( ImageItem, ModalityData, MultiModalDataDict, MultiModalFeatureSpec, MultiModalFieldConfig, MultiModalKwargsItems, ) from vllm.multimodal.parse import ( DictEmbeddingItems, ImageSize, ModalityDataItems, MultiModalDataItems, MultiModalDataParser, ) from vllm.multimodal.processing import ( BaseDummyInputsBuilder, BaseMultiModalProcessor, BaseProcessingInfo, PromptReplacement, PromptUpdate, ) from vllm.sequence import IntermediateTensors from vllm.transformers_utils.configs.hunyuan_vl import ( HunYuanVLConfig, HunYuanVLVisionConfig, ) from vllm.transformers_utils.processors.hunyuan_vl import HunYuanVLProcessor from vllm.transformers_utils.processors.hunyuan_vl_image import ( HunYuanVLImageProcessor, smart_resize, ) from vllm.utils.tensor_schema import TensorSchema, TensorShape from .interfaces import ( MultiModalEmbeddings, SupportsEagle3, SupportsLoRA, SupportsMultiModal, SupportsPP, SupportsQuant, SupportsXDRoPE, ) from .utils import ( AutoWeightsLoader, WeightsMapper, init_vllm_registered_model, maybe_prefix, ) from .vision import is_vit_use_data_parallel logger = init_logger(__name__) # === Vision Inputs === # class HunYuanVLImagePixelInputs(TensorSchema): """ Dimensions: - np: Number of patches - ni: Number of images - cps: Number of channels * patch_size * patch_size """ type: Literal["pixel_values"] pixel_values: Annotated[ torch.Tensor, TensorShape("np", "cps"), ] image_grid_thw: Annotated[ torch.Tensor, TensorShape("ni", 3), ] class HunYuanVLImageEmbeddingInputs(TensorSchema): """ Dimensions: - nf: Number of image features - hs: Hidden size - ni: Number of images """ type: Literal["image_embeds"] image_embeds: Annotated[ torch.Tensor, TensorShape("nf", "hs"), ] image_grid_thw: Annotated[ torch.Tensor, TensorShape("ni", 3), ] HunYuanVLImageInputs: TypeAlias = ( HunYuanVLImagePixelInputs | HunYuanVLImageEmbeddingInputs ) # === Vision Encoder === # class HunYuanVisionMLP(nn.Module): def __init__( self, in_features: int, hidden_features: int, bias: bool = True, act_fn: Callable[[torch.Tensor], torch.Tensor] = F.gelu, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() use_data_parallel = is_vit_use_data_parallel() self.dense_h_to_4h = ColumnParallelLinear( in_features, hidden_features, bias=bias, quant_config=quant_config, prefix=f"{prefix}.dense_h_to_4h", disable_tp=use_data_parallel, ) self.dense_4h_to_h = RowParallelLinear( hidden_features, in_features, bias=bias, quant_config=quant_config, prefix=f"{prefix}.dense_4h_to_h", disable_tp=use_data_parallel, ) self.act_fn = act_fn def forward(self, x: torch.Tensor): x_up, _ = self.dense_h_to_4h(x) x_down, _ = self.dense_4h_to_h(self.act_fn(x_up)) return x_down class HunYuanVisionAttention(nn.Module): def __init__( self, embed_dim: int, num_heads: int, projection_size: int, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() # Per attention head and per partition values. use_data_parallel = is_vit_use_data_parallel() self.tp_size = ( 1 if use_data_parallel else parallel_state.get_tensor_model_parallel_world_size() ) self.hidden_size_per_attention_head = dist_utils.divide( projection_size, num_heads ) self.num_attention_heads_per_partition = dist_utils.divide( num_heads, self.tp_size ) self.qkv = QKVParallelLinear( hidden_size=embed_dim, head_size=self.hidden_size_per_attention_head, total_num_heads=num_heads, total_num_kv_heads=num_heads, bias=True, quant_config=quant_config, prefix=f"{prefix}.qkv", disable_tp=use_data_parallel, ) self.o_proj = RowParallelLinear( input_size=projection_size, output_size=embed_dim, quant_config=quant_config, prefix=f"{prefix}.o_proj", disable_tp=use_data_parallel, ) self.scale = self.hidden_size_per_attention_head**-0.5 self.attn = MMEncoderAttention( self.num_attention_heads_per_partition, self.hidden_size_per_attention_head, self.scale, prefix=f"{prefix}.attn", ) def forward( self, x: torch.Tensor, ) -> torch.Tensor: qkv, _ = self.qkv(x) q, k, v = qkv.chunk(3, dim=-1) out = self.attn(q, k, v) output, _ = self.o_proj(out) return output class HunYuanVisionBlock(nn.Module): def __init__( self, dim: int, num_heads: int, mlp_hidden_dim: int, act_fn: Callable[[torch.Tensor], torch.Tensor] = F.gelu, norm_layer: Callable[[int], nn.Module] | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() if norm_layer is None: norm_layer = partial(nn.LayerNorm, eps=1e-6) self.input_layernorm = norm_layer(dim) self.post_attention_layernorm = norm_layer(dim) self.self_attn = HunYuanVisionAttention( embed_dim=dim, num_heads=num_heads, projection_size=dim, quant_config=quant_config, prefix=f"{prefix}.self_attn", ) self.mlp = HunYuanVisionMLP( dim, mlp_hidden_dim, act_fn=act_fn, bias=True, quant_config=quant_config, prefix=f"{prefix}.mlp", ) def forward( self, x: torch.Tensor, ) -> torch.Tensor: x = x + self.self_attn(self.input_layernorm(x)) x = x + self.mlp(self.post_attention_layernorm(x)) return x class HunYuanVisionPatchEmbed(nn.Module): def __init__(self, config: HunYuanVLVisionConfig): super().__init__() self.config = config self.embed_dim = config.hidden_size self.patch_size = config.patch_size self.num_channels = config.num_channels self.spatial_merge_size = config.spatial_merge_size self.interpolate_mode = config.interpolate_mode self.patch_embedding = nn.Conv2d( in_channels=config.num_channels, out_channels=self.embed_dim, kernel_size=self.patch_size, stride=self.patch_size, bias=True, ) self.max_num_patches = (config.max_image_size // self.patch_size) ** 2 self.num_positions = self.max_num_patches + 1 self.position_edge = int(self.num_positions**0.5) # first token is cls token, skip it self.position_embedding = nn.Embedding(self.num_positions, self.embed_dim) self.patch_pos_embed = None def forward( self, pixel_values: torch.Tensor, grid_thw: list[list[int]] ) -> torch.Tensor: num_patches = pixel_values.size(0) pixel_values = pixel_values.reshape( num_patches, self.num_channels, self.patch_size, self.patch_size ) patch_embeds = self.patch_embedding(pixel_values) patch_embeds = patch_embeds.squeeze(-1).squeeze(-1).unsqueeze(0) if self.patch_pos_embed is None: patch_pos_shape = ( 1, self.position_edge, self.position_edge, self.embed_dim, ) self.patch_pos_embed = ( self.position_embedding.weight[1:, :] .reshape(patch_pos_shape) .permute(0, 3, 1, 2) .float() ) patch_pos_embed_list = [] for grid in grid_thw: _, h0, w0 = grid # we add a small number to avoid floating point error in the interpolation # see discussion at https://github.com/facebookresearch/dino/issues/8 h0, w0 = h0 + 0.1, w0 + 0.1 patch_pos_embed = nn.functional.interpolate( self.patch_pos_embed, scale_factor=(h0 / self.position_edge, w0 / self.position_edge), mode=self.interpolate_mode, align_corners=False, ) patch_pos_embed = ( patch_pos_embed.reshape(self.embed_dim, -1) .transpose(0, 1) .unsqueeze(0) .to(patch_embeds.dtype) ) patch_pos_embed_list.append(patch_pos_embed) patch_pos_embed = torch.cat(patch_pos_embed_list, dim=1) embeddings = patch_embeds + patch_pos_embed return embeddings class HunYuanVisionPatchMerger(nn.Module): def __init__( self, in_channels, out_channels, spatial_merge_size=2, rms_norm_eps=1e-5, prefix="", ): super().__init__() self.spatial_merge_size = spatial_merge_size embed_std = out_channels**-0.5 self.proj = nn.Sequential( nn.Conv2d( in_channels, in_channels * 2, kernel_size=spatial_merge_size, stride=spatial_merge_size, ), nn.GELU(), nn.Conv2d(in_channels * 2, in_channels * 4, kernel_size=1), ) self.mlp = nn.Linear(in_channels * 4, out_channels) self.image_newline = nn.Parameter(torch.randn(in_channels * 4) * embed_std) self.image_begin = nn.Parameter(torch.randn(out_channels) * embed_std) self.image_end = nn.Parameter(torch.randn(out_channels) * embed_std) self.image_sep = nn.Parameter(torch.randn(out_channels) * embed_std) self.before_rms = RMSNorm(in_channels, eps=rms_norm_eps) self.after_rms = RMSNorm(out_channels, eps=rms_norm_eps) def forward(self, x, size=(16, 16)): x = self.before_rms(x) h, w = size dtype = x.dtype x = x.permute(0, 2, 1).reshape(x.shape[0], -1, h, w) x = self.proj(x) # b,c,h,w b, c, h, w = x.shape x = torch.cat( [x, self.image_newline.reshape(1, c, 1, 1).expand(b, c, h, 1).to(dtype)], dim=-1, ) x = x.reshape(b, c, -1).permute(0, 2, 1) x = self.mlp(x) begin = self.image_begin.reshape(1, 1, -1).expand(b, 1, x.shape[-1]).to(dtype) end = self.image_end.reshape(1, 1, -1).expand(b, 1, x.shape[-1]).to(dtype) x = torch.cat([begin, x, end], dim=1) return self.after_rms(x) class HunYuanVisionTransformer(nn.Module): def __init__( self, vision_config: HunYuanVLVisionConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() num_hidden_layers = vision_config.num_hidden_layers self.hidden_size = vision_config.hidden_size self.num_heads = vision_config.num_attention_heads self.spatial_merge_size = vision_config.spatial_merge_size from vllm.compilation.backends import set_model_tag with set_model_tag("HunYuanVisionPatchEmbed"): self.embeddings = HunYuanVisionPatchEmbed(vision_config) norm_layer = partial(nn.LayerNorm, eps=vision_config.rms_norm_eps) with set_model_tag("HunYuanVisionBlock"): self.layers = nn.ModuleList( [ HunYuanVisionBlock( dim=vision_config.hidden_size, num_heads=vision_config.num_attention_heads, mlp_hidden_dim=vision_config.intermediate_size, act_fn=get_act_fn(vision_config.hidden_act), norm_layer=norm_layer, quant_config=quant_config, prefix=f"{prefix}.layers.{layer_idx}", ) for layer_idx in range(num_hidden_layers) ] ) with set_model_tag("HunYuanVisionPatchMerger"): self.perceive = HunYuanVisionPatchMerger( vision_config.hidden_size, vision_config.out_hidden_size, spatial_merge_size=vision_config.spatial_merge_size, rms_norm_eps=vision_config.rms_norm_eps, prefix=f"{prefix}.perceive", ) @property def dtype(self) -> torch.dtype: return self.embeddings.patch_embedding.weight.dtype @property def device(self) -> torch.device: return self.embeddings.patch_embedding.weight.device def forward( self, x: torch.Tensor, grid_thw: list[list[int]], ) -> torch.Tensor: # patchify seq_len = x.size(0) cu_seqlens: list = [0] hidden_states = x.to(device=self.device, dtype=self.dtype) # embeddings = patch_embeds + patch_pos_embed hidden_states = self.embeddings(hidden_states, grid_thw) for t, h, w in grid_thw: t, h, w = int(t), int(h), int(w) cu_seqlens.append(h * w) cu_seqlens = torch.tensor(cu_seqlens, dtype=torch.int32) cu_seqlens = torch.cumsum(cu_seqlens, dim=0, dtype=torch.int32) cu_seqlens = cu_seqlens.to(device=self.device, non_blocking=True) hidden_states = hidden_states.reshape(seq_len, -1) hidden_states = hidden_states.unsqueeze(0) # build per-image lengths once split_lengths = [int(h) * int(w) for (_, h, w) in grid_thw] for layer in self.layers: # hidden_states: (1, T_total, D) parts = hidden_states.split(split_lengths, dim=1) # list of (1, L_i, D) parts = [layer(p) for p in parts] hidden_states = torch.cat(parts, dim=1) # adapter split_lengths = (cu_seqlens[1:] - cu_seqlens[:-1]).tolist() split_items = hidden_states.split(split_lengths, dim=1) image_embeds_list = [] for grid, split_item in zip(grid_thw, split_items): image_embeds_list.append( self.perceive(split_item.contiguous(), size=grid[1:]).squeeze(0) ) return image_embeds_list def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ # (param_name, shard_name, shard_id) (".qkv", ".q_proj", "q"), (".qkv", ".k_proj", "k"), (".qkv", ".v_proj", "v"), ] params_dict = dict(self.named_parameters(remove_duplicate=False)) loaded_params: set[str] = set() for name, loaded_weight in weights: for param_name, weight_name, shard_id in stacked_params_mapping: if weight_name not in name: continue name = name.replace(weight_name, param_name) param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params def _hunyuan_vl_field_config(hf_inputs: Mapping[str, torch.Tensor]): image_grid_thw = hf_inputs.get("image_grid_thw", torch.empty((0, 3))) image_grid_sizes = image_grid_thw.prod(-1) return dict( pixel_values=MultiModalFieldConfig.flat_from_sizes("image", image_grid_sizes), image_embeds=MultiModalFieldConfig.flat_from_sizes("image", image_grid_sizes), image_grid_thw=MultiModalFieldConfig.batched("image", keep_on_cpu=True), ) class HunYuanVLMultiModalDataParser(MultiModalDataParser): def _parse_image_data( self, data: dict[str, torch.Tensor] | ModalityData[ImageItem], ) -> ModalityDataItems[Any, Any] | None: if isinstance(data, dict): return DictEmbeddingItems( data, modality="image", required_fields={"image_embeds", "image_grid_thw"}, fields_factory=_hunyuan_vl_field_config, ) return super()._parse_image_data(data) class HunYuanVLProcessingInfo(BaseProcessingInfo): def get_hf_config(self): return self.ctx.get_hf_config(HunYuanVLConfig) def get_hf_processor( self, **kwargs: object, ) -> HunYuanVLProcessor: return self.ctx.get_hf_processor( HunYuanVLProcessor, use_fast=kwargs.pop("use_fast", True), **kwargs, ) def get_image_processor( self, **kwargs: object, ) -> HunYuanVLImageProcessor: return self.get_hf_processor(**kwargs).image_processor def get_data_parser(self): return HunYuanVLMultiModalDataParser( expected_hidden_size=self._get_expected_hidden_size(), ) def get_supported_mm_limits(self) -> Mapping[str, int | None]: return {"image": None} def get_mm_max_tokens_per_item( self, seq_len: int, mm_counts: Mapping[str, int], ) -> Mapping[str, int]: max_image_tokens = self.get_max_image_tokens() # TODO: support video max_video_tokens = 0 return {"image": max_image_tokens, "video": max_video_tokens} def _get_vision_info( self, *, image_width: int, image_height: int, num_frames: int = 1, do_resize: bool = True, image_processor: HunYuanVLImageProcessor, mm_kwargs: Mapping[str, object], ) -> tuple[ImageSize, int]: hf_config = self.get_hf_config() vision_config = hf_config.vision_config patch_size = vision_config.patch_size spatial_merge_size = vision_config.spatial_merge_size mm_kwargs = self.ctx.get_merged_mm_kwargs(mm_kwargs) size = image_processor.size if override_size := mm_kwargs.get("size"): size = size | override_size if (override_min_pixels := mm_kwargs.get("min_pixels")) is not None: size = size | {"shortest_edge": override_min_pixels} if (override_max_pixels := mm_kwargs.get("max_pixels")) is not None: size = size | {"longest_edge": override_max_pixels} if do_resize: resized_height, resized_width = smart_resize( height=image_height, width=image_width, factor=patch_size * spatial_merge_size, min_pixels=size["shortest_edge"], max_pixels=size["longest_edge"], ) preprocessed_size = ImageSize(width=resized_width, height=resized_height) else: preprocessed_size = ImageSize(width=image_width, height=image_height) grid_t = 1 grid_h = preprocessed_size.height // patch_size grid_w = preprocessed_size.width // patch_size num_vision_tokens = ( grid_t * grid_h // spatial_merge_size * (grid_w // spatial_merge_size + 1) + 2 ) return preprocessed_size, num_vision_tokens def get_num_image_tokens( self, *, image_width: int, image_height: int, image_processor: HunYuanVLImageProcessor, mm_kwargs: Mapping[str, object], ) -> int: _, num_image_tokens = self._get_vision_info( image_width=image_width, image_height=image_height, image_processor=image_processor, mm_kwargs=mm_kwargs, ) return num_image_tokens def get_image_size_with_most_features(self) -> ImageSize: image_processor = self.get_image_processor() max_image_size, _ = self._get_vision_info( image_width=512, image_height=8192, image_processor=image_processor, mm_kwargs={}, ) return max_image_size def get_max_image_tokens(self) -> int: image_processor = self.get_image_processor() target_width, target_height = self.get_image_size_with_most_features() return self.get_num_image_tokens( image_width=target_width, image_height=target_height, image_processor=image_processor, mm_kwargs={}, ) class HunYuanVLDummyInputsBuilder(BaseDummyInputsBuilder[HunYuanVLProcessingInfo]): def get_dummy_text(self, mm_counts: Mapping[str, int]) -> str: num_images = mm_counts.get("image", 0) hf_processor = self.info.get_hf_processor(typ=HunYuanVLProcessor) image_token: str = hf_processor.image_token return image_token * num_images def get_dummy_mm_data( self, seq_len: int, mm_counts: Mapping[str, int], mm_options: Mapping[str, BaseDummyOptions], ) -> MultiModalDataDict: num_images = mm_counts.get("image", 1) target_width, target_height = self.info.get_image_size_with_most_features() return { "image": self._get_dummy_images( width=target_width, height=target_height, num_images=num_images ), } class HunYuanVLMultiModalProcessor(BaseMultiModalProcessor[HunYuanVLProcessingInfo]): def _call_hf_processor( self, prompt: str, mm_data: Mapping[str, object], mm_kwargs: Mapping[str, object], tok_kwargs: Mapping[str, object], ) -> BatchFeature: return self.info.ctx.call_hf_processor( self.info.get_hf_processor(**mm_kwargs), dict(text=prompt, **mm_data), dict(**mm_kwargs, **tok_kwargs), ) def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, Any], out_mm_kwargs: MultiModalKwargsItems, ) -> Sequence[PromptUpdate]: hf_processor = self.info.get_hf_processor(**hf_processor_mm_kwargs) image_processor = self.info.get_image_processor(**hf_processor_mm_kwargs) placeholder = { "image": hf_processor.image_token_id, } merge_size = image_processor.merge_size def get_replacement_hunyuan_vl(item_idx: int, modality: str): out_item = out_mm_kwargs[modality][item_idx] grid_thw = out_item[f"{modality}_grid_thw"].data assert isinstance(grid_thw, torch.Tensor) _, grid_h, grid_w = grid_thw num_tokens = (int(grid_h) // merge_size) * ( int(grid_w) // merge_size + 1 ) + 2 return [placeholder[modality]] * num_tokens return [ PromptReplacement( modality=modality, target=[placeholder[modality]], replacement=partial(get_replacement_hunyuan_vl, modality=modality), ) for modality in ("image",) ] def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: return _hunyuan_vl_field_config(hf_inputs) @MULTIMODAL_REGISTRY.register_processor( HunYuanVLMultiModalProcessor, info=HunYuanVLProcessingInfo, dummy_inputs=HunYuanVLDummyInputsBuilder, ) class HunYuanVLForConditionalGeneration( nn.Module, SupportsMultiModal, SupportsLoRA, SupportsPP, SupportsQuant, SupportsXDRoPE, SupportsEagle3, ): # To ensure correct weight loading and mapping. hf_to_vllm_mapper = WeightsMapper( orig_to_new_prefix={ # mapping for new names in checkpoint saved after transformers v4.52 "vit.vit.": "visual.", "vit.": "visual.", "model.": "language_model.model.", } ) supports_encoder_tp_data = True def get_xdrope_input_positions( self, input_tokens: list[int], mm_features: list[MultiModalFeatureSpec], ) -> torch.Tensor: kwargs = MultiModalFeatureSpec.gather_kwargs( mm_features, {"image_grid_thw"}, ) image_grid_thw = [item.tolist() for item in kwargs.get("image_grid_thw", [])] hf_config = self.config image_start_token_id = hf_config.image_start_token_id spatial_merge_size = hf_config.vision_config.spatial_merge_size xd_num = len(hf_config.rope_scaling["xdrope_section"]) input_tokens_tensor = torch.tensor(input_tokens) image_start_indices = torch.argwhere( input_tokens_tensor == image_start_token_id ).squeeze(1) p_index = torch.arange(len(input_tokens_tensor)) w_index = torch.arange(len(input_tokens_tensor)) h_index = torch.arange(len(input_tokens_tensor)) t_index = torch.arange(len(input_tokens_tensor)) for image_index in range(len(image_start_indices)): # +1 : first image_token, +2: for xdrope positions pos = image_start_indices[image_index] + 2 t, h, w = image_grid_thw[image_index] _, llm_grid_h, llm_grid_w = ( t, h // spatial_merge_size, w // spatial_merge_size, ) token_num = (llm_grid_w + 1) * llm_grid_h w_index[pos : pos + token_num].copy_( torch.arange(0, llm_grid_w + 1) .reshape(1, -1) .expand(llm_grid_h, -1) .reshape(-1) ) h_index[pos : pos + token_num].copy_( torch.arange(0, llm_grid_h) .reshape(-1, 1) .expand(-1, llm_grid_w + 1) .reshape(-1) ) t_index[pos : pos + token_num] = image_index if xd_num == 4: llm_positions = torch.stack([p_index, w_index, h_index, t_index]) elif xd_num == 3: llm_positions = torch.stack([w_index, h_index, t_index]) return llm_positions @classmethod def get_placeholder_str(cls, modality: str, i: int) -> str | None: if modality.startswith("image"): return "<|hy_place▁holder▁no▁100|><|hy_place▁holder▁no▁102|><|hy_place▁holder▁no▁101|>" # noqa: E501 raise ValueError("Only image modality is supported") def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config: HunYuanVLConfig = vllm_config.model_config.hf_config self.config = config with self._mark_tower_model(vllm_config, {"image"}): self.visual = HunYuanVisionTransformer( config.vision_config, quant_config=vllm_config.quant_config, prefix=maybe_prefix(prefix, "visual"), ) with self._mark_language_model(vllm_config): self.language_model = init_vllm_registered_model( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "language_model.model"), architectures=[ "HunYuanDenseV1ForCausalLM", "HunYuanMoEV1ForCausalLM", ], ) self.make_empty_intermediate_tensors = ( self.language_model.make_empty_intermediate_tensors ) def _parse_and_validate_image_input( self, **kwargs: object ) -> HunYuanVLImageInputs | None: pixel_values = kwargs.pop("pixel_values", None) image_embeds = kwargs.pop("image_embeds", None) image_grid_thw = kwargs.pop("image_grid_thw", None) if pixel_values is None and image_embeds is None: return None # TODO: refine if isinstance(pixel_values, list): pixel_values = torch.cat(pixel_values, dim=0) if len(pixel_values.shape) == 3: last_dim = pixel_values.shape[-1] pixel_values = pixel_values.reshape(-1, last_dim) image_grid_thw = image_grid_thw.reshape(-1, 3) if pixel_values is not None: return HunYuanVLImagePixelInputs( type="pixel_values", pixel_values=pixel_values, image_grid_thw=image_grid_thw, ) if image_embeds is not None: return HunYuanVLImageEmbeddingInputs( type="image_embeds", image_embeds=image_embeds, image_grid_thw=image_grid_thw, ) def _process_image_input( self, image_input: HunYuanVLImageInputs ) -> tuple[torch.Tensor, ...]: grid_thw = image_input["image_grid_thw"] assert grid_thw.ndim == 2 grid_thw_list = grid_thw.tolist() if image_input["type"] == "image_embeds": image_embeds = image_input["image_embeds"].type(self.visual.dtype) else: pixel_values = image_input["pixel_values"] # TODO: use_data_parallel (split image_embeds in visual) image_embeds = self.visual(pixel_values, grid_thw=grid_thw_list) return image_embeds def _parse_and_validate_multimodal_inputs(self, **kwargs: object) -> dict: mm_input_by_modality = {} # Preserve the order of modalities if there are multiple of them # from the order of kwargs. for input_key in kwargs: if ( input_key in ("pixel_values", "image_embeds") and "image" not in mm_input_by_modality ): mm_input_by_modality["image"] = self._parse_and_validate_image_input( **kwargs ) return mm_input_by_modality def embed_multimodal(self, **kwargs: object) -> MultiModalEmbeddings: mm_input_by_modality = self._parse_and_validate_multimodal_inputs(**kwargs) if not mm_input_by_modality: return [] # The result multimodal_embeddings is tuple of tensors, with each # tensor correspoending to a multimodal data item (image or video). multimodal_embeddings: tuple[torch.Tensor, ...] = () # NOTE: It is important to iterate over the keys in this dictionary # to preserve the order of the modalities. for modality in mm_input_by_modality: multimodal_input = mm_input_by_modality[modality] if modality == "image": image_embeddings = self._process_image_input(multimodal_input) multimodal_embeddings += tuple(image_embeddings) return multimodal_embeddings def set_aux_hidden_state_layers(self, layers: tuple[int, ...]) -> None: self.language_model.model.aux_hidden_state_layers = layers def get_eagle3_aux_hidden_state_layers(self) -> tuple[int, ...]: num_layers = len(self.language_model.model.layers) return (2, num_layers // 2, num_layers - 3) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None, inputs_embeds: torch.Tensor | None, **kwargs: object, ) -> torch.Tensor | IntermediateTensors: if intermediate_tensors is not None: inputs_embeds = None hidden_states = self.language_model( input_ids=input_ids, positions=positions, intermediate_tensors=intermediate_tensors, inputs_embeds=inputs_embeds, ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: return self.language_model.compute_logits(hidden_states) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader( self, skip_prefixes=(["lm_head."] if self.config.tie_word_embeddings else None), ) return loader.load_weights(weights, mapper=self.hf_to_vllm_mapper) def get_mm_mapping(self) -> MultiModelKeys: """ Get the module prefix in multimodal models """ return MultiModelKeys.from_string_field( language_model="language_model.model", connector="visual.perceive", tower_model="visual", )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/hunyuan_vision.py", "license": "Apache License 2.0", "lines": 883, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/transformers_utils/configs/hunyuan_vl.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # adapted from https://github.com/ManaEstras/transformers/blob/v4.57.1.hyvl/src/transformers/models/hunyuan_vl/configuration_hunyuan_vl.py from transformers import PretrainedConfig class HunYuanVLVisionConfig(PretrainedConfig): model_type = "hunyuan_vl" base_config_key = "vision_config" def __init__( self, hidden_act="gelu", hidden_size=1152, intermediate_size=4304, interpolate_mode="bilinear", rms_norm_eps=1e-05, learnable_mlp_pooling_size=0, num_attention_heads=16, num_key_value_heads=None, num_channels=3, num_hidden_layers=27, out_hidden_size=4096, patch_size=16, remove_prenorm=True, spatial_merge_size=2, temporal_patch_size=1, resize_resolution=2048, img_max_token_num=4096, max_image_size=2048, video_max_image_size=768, video_min_image_size=256, min_image_size=512, anyres_vit_max_image_size=2048, max_vit_seq_len=16384, text_hidden_size=3072, **kwargs, ): super().__init__(**kwargs) self.hidden_act = hidden_act self.hidden_size = hidden_size self.intermediate_size = intermediate_size self.interpolate_mode = interpolate_mode self.learnable_mlp_pooling_size = learnable_mlp_pooling_size self.num_attention_heads = num_attention_heads if not num_key_value_heads: self.num_key_value_heads = num_attention_heads else: self.num_key_value_heads = num_key_value_heads self.num_channels = num_channels self.num_hidden_layers = num_hidden_layers self.out_hidden_size = out_hidden_size self.patch_size = patch_size self.remove_prenorm = remove_prenorm self.spatial_merge_size = spatial_merge_size self.temporal_patch_size = temporal_patch_size self.rms_norm_eps = rms_norm_eps self.resize_resolution = resize_resolution self.img_max_token_num = img_max_token_num self.max_image_size = max_image_size self.min_image_size = min_image_size self.video_max_image_size = video_max_image_size self.video_min_image_size = video_min_image_size self.anyres_vit_max_image_size = anyres_vit_max_image_size self.max_vit_seq_len = max_vit_seq_len self.text_hidden_size = text_hidden_size class HunYuanVLTextConfig(PretrainedConfig): r""" This is the configuration class to store the configuration of a [`HunYuanVLTextConfig`]. It is used to instantiate an HunYuan model according to the specified arguments, defining the model architecture. Instantiating a configuration with the defaults will yield a similar configuration to that of the HunYuan-7B. Hunyuan-7B-Instruct [tencent/Hunyuan-7B-Instruct](https://huggingface.co/tencent/Hunyuan-7B-Instruct). Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information. Args: vocab_size (`int`, *optional*, defaults to 290943): Vocabulary size of the HunYuan model. Defines the number of different tokens that can be represented by the `inputs_ids` passed when calling [`HunYuanVLTextConfig`] hidden_size (`int`, *optional*, defaults to 4096): Dimension of the hidden representations. intermediate_size (`int`, *optional*, defaults to 11008): Dimension of the MLP representations or shared MLP representations. num_hidden_layers (`int`, *optional*, defaults to 32): Number of hidden layers in the Transformer decoder. num_attention_heads (`int`, *optional*, defaults to 32): Number of attention heads for each attention layer in the Transformer decoder. num_key_value_heads (`int`, *optional*): This is the number of key_value heads that should be used to implement Grouped Query Attention. If `num_key_value_heads=num_attention_heads`, the model will use Multi Head Attention (MHA), if `num_key_value_heads=1 the model will use Multi Query Attention (MQA) otherwise GQA is used. When converting a multi-head checkpoint to a GQA checkpoint, each group key and value head should be constructed by meanpooling all the original heads within that group. For more details checkout [this paper](https://huggingface.co/papers/2305.13245). If it is not specified, will default to `num_attention_heads`. hidden_act (`str` or `function`, *optional*, defaults to `"silu"`): The non-linear activation function (function or string) in the decoder. max_position_embeddings (`int`, *optional*, defaults to 2048): The maximum sequence length that this model might ever be used with. initializer_range (`float`, *optional*, defaults to 0.02): The standard deviation of the truncated_normal_initializer for initializing all weight matrices. rms_norm_eps (`float`, *optional*, defaults to 1e-05): The epsilon used by the rms normalization layers. use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions (not used by all models). Only relevant if `config.is_decoder=True`. pad_token_id (`int`, *optional*, defaults to 0): Padding token id. bos_token_id (`int`, *optional*, defaults to 1): Beginning of stream token id. eos_token_id (`int`, *optional*, defaults to 2): End of stream token id. eod_token_id (int, *optional*, defaults to 3): Token ID representing the end-of-document marker. Used to indicate the termination of a text sequence. Example: In multi-document processing, this token helps the model distinguish between separate documents. pretraining_tp (`int`, *optional*, defaults to 1): Experimental feature. Tensor parallelism rank used during pretraining. Please refer to [this document](https://huggingface.co/docs/transformers/parallelism) to understand more about it. This value is necessary to ensure exact reproducibility of the pretraining results. Please refer to [this issue](https://github.com/pytorch/pytorch/issues/76232). tie_word_embeddings (`bool`, *optional*, defaults to `False`): Whether to tie weight embeddings rope_theta (`float`, *optional*, defaults to 10000.0): The base period of the RoPE embeddings. rope_scaling (`Dict`, *optional*): Dictionary containing the scaling configuration for the RoPE embeddings. Currently supports two scaling strategies: linear and dynamic. Their scaling factor must be a float greater than 1. The expected format is `{"type": strategy name, "factor": scaling factor}`. When using this flag, don't update `max_position_embeddings` to the expected new maximum. See the following thread for more information on how these scaling strategies behave: https://www.reddit.com/r/LocalLLaMA/comments/14mrgpr/dynamically_scaled_rope_further_increases/. This is an experimental feature, subject to breaking API changes in future versions. attention_bias (`bool`, defaults to `False`, *optional*, defaults to `False`): Whether to use a bias in the query, key, value and output projection layers during self-attention. attention_dropout (`float`, *optional*, defaults to 0.0): The dropout ratio for the attention probabilities. head_dim (`int`, *optional*, defaults to 128): The attention head dimension. """ # noqa: E501 model_type = "hunyuan_vl_text" keys_to_ignore_at_inference = ["past_key_values"] def __init__( self, vocab_size=290943, hidden_size=4096, intermediate_size: int = 11008, num_hidden_layers=32, num_attention_heads=32, num_key_value_heads=None, hidden_act="silu", max_position_embeddings=2048, initializer_range=0.02, rms_norm_eps=1e-5, use_cache=True, pad_token_id=0, bos_token_id=1, eos_token_id=2, eod_token_id=3, pretraining_tp=1, tie_word_embeddings=False, rope_theta=10000.0, rope_scaling=None, attention_bias=False, attention_dropout=0.0, head_dim=None, **kwargs, ): self.vocab_size = vocab_size self.max_position_embeddings = max_position_embeddings self.hidden_size = hidden_size self.intermediate_size = intermediate_size self.num_hidden_layers = num_hidden_layers self.num_attention_heads = num_attention_heads self.head_dim = head_dim # for backward compatibility if num_key_value_heads is None: num_key_value_heads = num_attention_heads self.num_key_value_heads = num_key_value_heads self.hidden_act = hidden_act self.initializer_range = initializer_range self.rms_norm_eps = rms_norm_eps self.pretraining_tp = pretraining_tp self.use_cache = use_cache self.rope_theta = rope_theta self.rope_scaling = rope_scaling # self._rope_scaling_validation() # TODO: Need validation? self.attention_bias = attention_bias self.attention_dropout = attention_dropout super().__init__( pad_token_id=pad_token_id, bos_token_id=bos_token_id, eos_token_id=eos_token_id, tie_word_embeddings=tie_word_embeddings, **kwargs, ) def _rope_scaling_validation(self): """ Validate the `rope_scaling` configuration. """ if self.rope_scaling is None: return if not isinstance(self.rope_scaling, dict) or len(self.rope_scaling) != 2: raise ValueError( "`rope_scaling` must be a dictionary with with two fields, `type` and " f"`factor` or `type` and `alpha`, got {self.rope_scaling}" ) rope_scaling_type = self.rope_scaling.get("type", None) rope_scaling_factor = self.rope_scaling.get("factor", None) rope_scaling_alpha = self.rope_scaling.get("alpha", None) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( "`rope_scaling`'s type field must be one of ['linear', 'dynamic'], " f"got {rope_scaling_type}" ) if rope_scaling_factor is None and rope_scaling_alpha is None: raise ValueError( "`rope_scaling`'s factor or alpha field must be have one, " "got both of none" ) if rope_scaling_factor is not None and ( not isinstance(rope_scaling_factor, float) or rope_scaling_factor <= 1.0 ): raise ValueError( "`rope_scaling`'s factor field must be a float > 1.0, " f"got {rope_scaling_factor}" ) if rope_scaling_alpha is not None and ( not isinstance(rope_scaling_alpha, float) or rope_scaling_alpha <= 1.0 ): raise ValueError( "`rope_scaling`'s alpha field must be a float > 1.0, " f"got {rope_scaling_alpha}" ) class HunYuanVLConfig(PretrainedConfig): model_type = "hunyuan_vl" sub_configs = { "vision_config": HunYuanVLVisionConfig, "text_config": HunYuanVLTextConfig, } keys_to_ignore_at_inference = ["past_key_values"] def __init__( self, text_config=None, vision_config=None, im_start_id=120118, im_end_id=120119, image_token_id=120120, im_newline_id=120121, video_start_id=120122, video_end_id=120123, **kwargs, ): # We need to init super() here so that it does not reset values # that are in text config to the BaseClass defaults. The Base # config has many text related defaults and not all defaults are # same as for `HunYuanVLTextConfig`. super().__init__(**kwargs) if isinstance(vision_config, dict): self.vision_config = self.sub_configs["vision_config"](**vision_config) elif vision_config is None: self.vision_config = self.sub_configs["vision_config"]() if isinstance(text_config, dict): self.text_config = self.sub_configs["text_config"](**text_config) elif text_config is None: # For BC use all kwargs to init `TextConfig` self.text_config = self.sub_configs["text_config"](**kwargs) self.image_token_id = image_token_id self.im_start_id = im_start_id self.im_end_id = im_end_id self.im_newline_id = im_newline_id self.video_start_id = video_start_id self.video_end_id = video_end_id self.vision_config.text_hidden_size = self.text_config.hidden_size # Attention implementation to use. It sets it recursively on sub-configs # so we call it again in the end. self._attn_implementation = kwargs.pop("attn_implementation", None) def __setattr__(self, key, value): if ( (text_config := super().__getattribute__("__dict__").get("text_config")) is not None and key not in ["dtype", "_attn_implementation_internal"] and key in text_config.__dict__ ): setattr(text_config, key, value) else: super().__setattr__(key, value) def __getattribute__(self, key): if "text_config" in super().__getattribute__("__dict__") and key not in [ "_name_or_path", "model_type", "dtype", "_attn_implementation_internal", ]: text_config = super().__getattribute__("text_config") if key in text_config.__dict__: return getattr(text_config, key) return super().__getattribute__(key)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/configs/hunyuan_vl.py", "license": "Apache License 2.0", "lines": 293, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/transformers_utils/processors/hunyuan_vl.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # adapted from https://github.com/ManaEstras/transformers/blob/v4.57.1.hyvl/src/transformers/models/hunyuan_vl/processing_hunyuan_vl.py import numpy as np import torch from transformers import AutoProcessor from transformers.feature_extraction_utils import BatchFeature from transformers.image_utils import ImageInput from transformers.processing_utils import ProcessorMixin from transformers.tokenization_utils_base import PreTokenizedInput, TextInput from transformers.video_utils import VideoInput class HunYuanVLProcessor(ProcessorMixin): attributes = ["image_processor", "tokenizer"] valid_kwargs = ["chat_template"] image_processor_class = "AutoImageProcessor" tokenizer_class = "AutoTokenizer" # ("AutoTokenizer", None) def __init__( self, image_processor=None, tokenizer=None, chat_template=None, **kwargs, ): # TODO Fix the init self.tokenizer = tokenizer self.image_token_id = 120120 # self.tokenizer.image_token_id self.image_token = self.tokenizer.convert_ids_to_tokens(self.image_token_id) self.im_start_token_id = 120118 # self.tokenizer.im_start_id self.im_start_token = self.tokenizer.convert_ids_to_tokens( self.im_start_token_id ) self.im_end_token_id = 120119 # self.tokenizer.im_end_id self.im_end_token = self.tokenizer.convert_ids_to_tokens(self.im_end_token_id) self.placeholder_token = self.tokenizer.convert_ids_to_tokens( self.tokenizer.vocab_size - 1 ) self.pad_id = 120002 # self.tokenizer.pad_token_id super().__init__(image_processor, tokenizer, chat_template=chat_template) def __call__( self, images: ImageInput = None, text: TextInput | PreTokenizedInput | list[TextInput] | list[PreTokenizedInput] = None, videos: VideoInput = None, **kwargs, ) -> BatchFeature: image_inputs = {} if images is not None: image_inputs = self.image_processor(images=images) image_grid_thw = image_inputs["image_grid_thw"] if not isinstance(text, list): text = [text] text = text.copy() # below lines change text in-place image_tokens_cumsum = [0] if images is not None: index = 0 for i in range(len(text)): while self.image_token in text[i]: grid_h, grid_w = image_grid_thw[index][-2:] patch_h = grid_h // self.image_processor.merge_size patch_w = grid_w // self.image_processor.merge_size num_image_tokens = patch_h * (patch_w + 1) + 2 image_tokens_cumsum.append( image_tokens_cumsum[-1] + num_image_tokens ) # text[i] = text[i].replace(self.image_token, self.im_start_token + self.placeholder_token * num_image_tokens + self.im_end_token, 1) # noqa: E501 text[i] = text[i].replace( self.image_token, self.placeholder_token * num_image_tokens, 1 ) index += 1 text[i] = text[i].replace(self.placeholder_token, self.image_token) # text[i] = self.tokenizer.bos_token + text[i] text_inputs = self.tokenizer(text, add_special_tokens=False, **kwargs) self._check_special_mm_tokens(text, text_inputs, modalities=["image"]) input_ids = text_inputs["input_ids"] position_ids = torch.arange(len(input_ids[0])) position_ids_w = torch.arange(len(input_ids[0])) position_ids_h = torch.arange(len(input_ids[0])) position_ids_t = torch.arange(len(input_ids[0])) if images is not None: image_token_pos_indices = torch.where(input_ids[0] == self.image_token_id)[ 0 ] for i in range(len(image_grid_thw)): grid_h, grid_w = image_grid_thw[i][-2:] patch_h = grid_h // self.image_processor.merge_size patch_w = grid_w // self.image_processor.merge_size start_pos = image_token_pos_indices[image_tokens_cumsum[i]].item() + 1 replace_num = (patch_w + 1) * patch_h position_ids_w[start_pos : start_pos + replace_num] = torch.tensor( list(range(patch_w + 1)) * patch_h, dtype=torch.int64 ) patch_h_list = [] for h in range(patch_h): patch_h_list += [h] * (patch_w + 1) position_ids_h[start_pos : start_pos + replace_num] = torch.tensor( patch_h_list, dtype=torch.int64 ) position_ids_t[start_pos : start_pos + replace_num] = 0 position_ids = torch.stack( [position_ids, position_ids_w, position_ids_h, position_ids_t] ).unsqueeze(0) text_inputs["position_ids"] = position_ids attention_mask = input_ids.ne(self.pad_id) text_inputs["attention_mask"] = attention_mask text_inputs["imgs_pos"] = [self.get_imgs_pos(e) for e in input_ids] # image_inputs["imgs"] = [[image_inputs["pixel_values"]]] return_tensors = kwargs.pop("return_tensors", None) return BatchFeature( data={**text_inputs, **image_inputs}, tensor_type=return_tensors, ) def batch_decode(self, *args, **kwargs): return self.tokenizer.batch_decode(*args, **kwargs) def decode(self, *args, **kwargs): return self.tokenizer.decode(*args, **kwargs) def post_process_image_text_to_text( self, generated_outputs, skip_special_tokens=True, clean_up_tokenization_spaces=False, **kwargs, ): assert 0 def apply_chat_template(self, *args, **kwargs): kwargs["return_dict"] = False return self.tokenizer.apply_chat_template(*args, **kwargs) def get_imgs_pos(self, doc_ids): doc_ids = np.array(doc_ids, dtype=np.int64) img_begin_index = np.where(doc_ids == self.im_start_token_id)[0] img_end_index = np.where(doc_ids == self.im_end_token_id)[0] imgs_pos = np.concatenate( ( np.reshape(img_begin_index + 1, (-1, 1)), np.reshape(img_end_index, (-1, 1)), ), axis=-1, ).tolist() return imgs_pos @property def model_input_names(self): tokenizer_input_names = self.tokenizer.model_input_names image_processor_input_names = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names)) def split_image_into_patch_blocks( pixel_values: torch.Tensor, # shape: [batch_size, 3, H, W] patch_size: int = 16, # e.g. 16 adaptor_patch_div: int = 4, # e.g. 4 --> each patch_size is cut into 4x4 small regions, i.e. patch_size // 4 # noqa: E501 ) -> torch.Tensor: """ Split the input image tensor (supporting batch) into large patches of size `patch_size`, and then further divide each large patch into smaller regions of size (patch_size // adaptor_patch_div) x (patch_size // adaptor_patch_div). Each small region is extracted as a tensor of shape [3, patch_size, patch_size]. The final output contains all such small region tensors. Args: pixel_values: Input image tensor of shape [batch_size, 3, H, W]. patch_size: Size of the large patch, e.g., 16. adaptor_patch_div: Each large patch is divided into (patch_size // adaptor_patch_div) x (patch_size // adaptor_patch_div) smaller regions. Returns: patches: A tensor of shape [N, 3, patch_size, patch_size], where N = batch_size * (H // patch_size) * (W // patch_size) * (patch_size // adaptor_patch_div)^2. Each element in the batch corresponds to one small image region. """ # noqa: E501 batch_size, channels, height, width = pixel_values.shape assert channels == 3, "Pixel values must have 3 channels in dim=1" assert height % patch_size == 0 and width % patch_size == 0, ( "H and W must be divisible by patch_size" ) patch_height_num = height // patch_size patch_width_num = width // patch_size # Reshape to [B, 3, ph, ps, pw, ps] img = pixel_values.reshape( batch_size, 3, patch_height_num, patch_size, patch_width_num, patch_size ) # Further split each psxps patch into (ps//aps)x(ps//aps) small regions img = img.reshape( batch_size, 3, patch_height_num, patch_size // adaptor_patch_div, # ps // aps adaptor_patch_div, patch_width_num, patch_size // adaptor_patch_div, # ps // aps adaptor_patch_div, ) # Permute to group the small regions: [B, ph, pw, ps//aps, ps//aps, 3, aps, aps] img = img.permute(0, 2, 5, 3, 6, 1, 4, 7) # Reshape into [B * ph * pw * (ps//aps)^2, 3, patch_size, patch_size] patches = img.reshape(-1, 3, patch_size, patch_size) return patches AutoProcessor.register("HunYuanVLProcessor", HunYuanVLProcessor)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/processors/hunyuan_vl.py", "license": "Apache License 2.0", "lines": 196, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:examples/offline_inference/qwen3_omni/only_thinker.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ This example shows how to use vLLM for running offline inference with the correct prompt format on Qwen3-Omni (thinker only). """ from typing import NamedTuple from vllm import LLM, SamplingParams from vllm.assets.audio import AudioAsset from vllm.assets.image import ImageAsset from vllm.assets.video import VideoAsset from vllm.multimodal.image import convert_image_mode from vllm.utils.argparse_utils import FlexibleArgumentParser class QueryResult(NamedTuple): inputs: dict limit_mm_per_prompt: dict[str, int] # NOTE: The default `max_num_seqs` and `max_model_len` may result in OOM on # lower-end GPUs. # Unless specified, these settings have been tested to work on a single L4. default_system = ( "You are Qwen, a virtual human developed by the Qwen Team, Alibaba " "Group, capable of perceiving auditory and visual inputs, as well as " "generating text and speech." ) def get_mixed_modalities_query() -> QueryResult: question = ( "What is recited in the audio? " "What is the content of this image? Why is this video funny?" ) prompt = ( f"<|im_start|>system\n{default_system}<|im_end|>\n" "<|im_start|>user\n<|audio_start|><|audio_pad|><|audio_end|>" "<|vision_start|><|image_pad|><|vision_end|>" "<|vision_start|><|video_pad|><|vision_end|>" f"{question}<|im_end|>\n" f"<|im_start|>assistant\n" ) return QueryResult( inputs={ "prompt": prompt, "multi_modal_data": { "audio": AudioAsset("mary_had_lamb").audio_and_sample_rate, "image": convert_image_mode( ImageAsset("cherry_blossom").pil_image, "RGB" ), "video": VideoAsset(name="baby_reading", num_frames=16).np_ndarrays, }, }, limit_mm_per_prompt={"audio": 1, "image": 1, "video": 1}, ) def get_use_audio_in_video_query() -> QueryResult: question = ( "Describe the content of the video in details, then convert what the " "baby say into text." ) prompt = ( f"<|im_start|>system\n{default_system}<|im_end|>\n" "<|im_start|>user\n<|vision_start|><|video_pad|><|vision_end|>" f"{question}<|im_end|>\n" f"<|im_start|>assistant\n" ) asset = VideoAsset(name="baby_reading", num_frames=16) audio = asset.get_audio(sampling_rate=16000) return QueryResult( inputs={ "prompt": prompt, "multi_modal_data": { "video": asset.np_ndarrays, "audio": audio, }, "mm_processor_kwargs": { "use_audio_in_video": True, }, }, limit_mm_per_prompt={"audio": 1, "video": 1}, ) def get_multi_audios_query() -> QueryResult: question = "Are these two audio clips the same?" prompt = ( f"<|im_start|>system\n{default_system}<|im_end|>\n" "<|im_start|>user\n<|audio_start|><|audio_pad|><|audio_end|>" "<|audio_start|><|audio_pad|><|audio_end|>" f"{question}<|im_end|>\n" f"<|im_start|>assistant\n" ) return QueryResult( inputs={ "prompt": prompt, "multi_modal_data": { "audio": [ AudioAsset("winning_call").audio_and_sample_rate, AudioAsset("mary_had_lamb").audio_and_sample_rate, ], }, }, limit_mm_per_prompt={ "audio": 2, }, ) def get_multi_images_query() -> QueryResult: question = "What are the differences between these two images?" prompt = ( f"<|im_start|>system\n{default_system}<|im_end|>\n" "<|im_start|>user\n<|vision_start|><|image_pad|><|vision_end|>" "<|vision_start|><|image_pad|><|vision_end|>" f"{question}<|im_end|>\n" f"<|im_start|>assistant\n" ) return QueryResult( inputs={ "prompt": prompt, "multi_modal_data": { "image": [ convert_image_mode(ImageAsset("cherry_blossom").pil_image, "RGB"), convert_image_mode(ImageAsset("stop_sign").pil_image, "RGB"), ], }, }, limit_mm_per_prompt={ "image": 2, }, ) query_map = { "mixed_modalities": get_mixed_modalities_query, "use_audio_in_video": get_use_audio_in_video_query, "multi_audios": get_multi_audios_query, "multi_images": get_multi_images_query, } def main(args): model_name = args.model query_result = query_map[args.query_type]() llm = LLM( model=model_name, max_model_len=args.max_model_len, max_num_seqs=5, limit_mm_per_prompt=query_result.limit_mm_per_prompt, seed=args.seed, tensor_parallel_size=args.tensor_parallel_size, gpu_memory_utilization=args.gpu_memory_utilization, ) # We set temperature to 0.2 so that outputs can be different # even when all prompts are identical when running batch inference. sampling_params = SamplingParams(temperature=0.2, max_tokens=256) outputs = llm.generate(query_result.inputs, sampling_params=sampling_params) for o in outputs: generated_text = o.outputs[0].text print(generated_text) def parse_args(): parser = FlexibleArgumentParser( description="Demo on using vLLM for offline inference with " "audio language models" ) parser.add_argument( "--query-type", "-q", type=str, default="mixed_modalities", choices=query_map.keys(), help="Query type.", ) parser.add_argument( "--seed", type=int, default=0, help="Set the seed when initializing `vllm.LLM`.", ) parser.add_argument( "--model", type=str, default="Qwen/Qwen3-Omni-30B-A3B-Instruct", help="Model name or path.", ) parser.add_argument( "--tensor-parallel-size", "-tp", type=int, default=1, help="Tensor parallel size for distributed inference.", ) parser.add_argument( "--gpu-memory-utilization", type=float, default=0.9, help="GPU memory utilization (0.0 to 1.0).", ) parser.add_argument( "--max-model-len", type=int, default=12800, help="Maximum model context length.", ) return parser.parse_args() if __name__ == "__main__": args = parse_args() main(args)
{ "repo_id": "vllm-project/vllm", "file_path": "examples/offline_inference/qwen3_omni/only_thinker.py", "license": "Apache License 2.0", "lines": 195, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/model_executor/test_qwen3_omni.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from unittest.mock import Mock import pytest from transformers import PretrainedConfig from vllm.multimodal.processing import InputProcessingContext # Helper function to print input IDs with coalesced audio/video tokens. def print_input_ids(input_ids): """ Print input IDs, compressing consecutive special tokens. - 151675: <|audio_pad|> - 151656: <|video_pad|> """ if not input_ids: print("[]") return result = [] i = 0 while i < len(input_ids): current_id = input_ids[i] # Check if it's a special token that should be compressed if current_id in [151675, 151656]: # Count consecutive occurrences count = 1 while i + count < len(input_ids) and input_ids[i + count] == current_id: count += 1 # Add compressed representation token_name = "<|audio_pad|>" if current_id == 151675 else "<|video_pad|>" result.append(f"{token_name} * {count}") i += count else: # Regular token, just add it result.append(str(current_id)) i += 1 print(", ".join(result)) @pytest.fixture def mock_qwen3_omni_config(): """Create a mock Qwen3OmniMoeThinker config.""" config = Mock(spec=PretrainedConfig) # Token IDs from https://huggingface.co/Qwen/Qwen3-Omni-30B-A3B-Instruct/blob/main/tokenizer_config.json config.audio_token_id = 151675 # <|audio_pad|> config.video_token_id = 151656 # <|video_pad|> config.image_token_id = 151655 # <|image_pad|> config.audio_start_token_id = 151669 # <|audio_start|> config.audio_end_token_id = 151670 # <|audio_end|> config.vision_start_token_id = 151652 # <|vision_start|> config.position_id_per_seconds = 12.5 # Vision config vision_config = Mock() vision_config.spatial_merge_size = 2 config.vision_config = vision_config return config @pytest.fixture def mock_processor(): """Create a mock HF processor.""" from transformers.models.whisper import WhisperFeatureExtractor processor = Mock() processor.audio_token = "<|audio_pad|>" processor.image_token = "<|image_pad|>" processor.video_token = "<|video_pad|>" # Create a real WhisperFeatureExtractor instance for the feature_extractor attribute feature_extractor = WhisperFeatureExtractor() processor.feature_extractor = feature_extractor return processor @pytest.fixture def mock_tokenizer(): """Create a mock tokenizer.""" tokenizer = Mock() # Token IDs from https://huggingface.co/Qwen/Qwen3-Omni-30B-A3B-Instruct/blob/main/tokenizer_config.json tokenizer.get_vocab = Mock( return_value={ "<|audio_pad|>": 151675, "<|video_pad|>": 151656, "<|image_pad|>": 151655, "<|audio_start|>": 151669, "<|audio_end|>": 151670, "<|vision_start|>": 151652, "<|vision_end|>": 151653, } ) tokenizer.encode = Mock( side_effect=lambda x: { "<|vision_start|>": [151652], "<|vision_end|>": [151653], "<|audio_start|>": [151669], "<|audio_end|>": [151670], "<|audio_pad|>": [151675], "<|image_pad|>": [151655], "<|video_pad|>": [151656], }.get(x, [0]) ) tokenizer.vision_bos_token = "<|vision_start|>" tokenizer.vision_eos_token = "<|vision_end|>" tokenizer.audio_bos_token = "<|audio_start|>" tokenizer.audio_eos_token = "<|audio_end|>" return tokenizer @pytest.fixture def mock_image_processor(): """Create a mock image processor.""" image_processor = Mock() image_processor.merge_size = 2 return image_processor def test_qwen3_omni_get_updates_use_audio_in_video( mock_qwen3_omni_config, mock_processor, mock_tokenizer, mock_image_processor, ): """Test the get_updates_use_audio_in_video method directly.""" from vllm.model_executor.models.qwen3_omni_moe_thinker import ( Qwen3OmniMoeThinkerMultiModalProcessor, Qwen3OmniMoeThinkerProcessingInfo, ) # Create a mock context mock_ctx = Mock(spec=InputProcessingContext) # Create processing info info = Qwen3OmniMoeThinkerProcessingInfo(mock_ctx) info._get_expected_hidden_size = lambda: 100 info.get_hf_config = Mock(return_value=mock_qwen3_omni_config) info.get_hf_processor = Mock(return_value=mock_processor) info.get_tokenizer = Mock(return_value=mock_tokenizer) info.get_image_processor = Mock(return_value=mock_image_processor) # Create a mock dummy_inputs builder mock_dummy_inputs = Mock() # Create the processor processor = Qwen3OmniMoeThinkerMultiModalProcessor(info, mock_dummy_inputs) # Test parameters from reference video # https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen3-Omni/demo/draw.mp4 audio_len = 85 video_grid_thw = [6, 36, 64] video_second_per_grid_t = 2.0 # Call the method updates = processor.get_updates_use_audio_in_video( thinker_config=mock_qwen3_omni_config, audio_len=audio_len, video_grid_thw=video_grid_thw, video_second_per_grid_t=video_second_per_grid_t, ) # Updated input ids should align with HF implementation. # 151669, # <|video_pad|> * 576, <|audio_pad|> * 25, # <|video_pad|> * 576, <|audio_pad|> * 25, # <|video_pad|> * 576, <|audio_pad|> * 25, # <|video_pad|> * 576, <|audio_pad|> * 10, # <|video_pad|> * 1152, # 151670 print_input_ids(updates) # Verify structure assert isinstance(updates, list) assert len(updates) > 0 # Verify start and end tokens audio_start_token_id = mock_qwen3_omni_config.audio_start_token_id audio_end_token_id = mock_qwen3_omni_config.audio_end_token_id assert updates[0] == audio_start_token_id assert updates[-1] == audio_end_token_id # Verify both audio and video tokens are present audio_token_id = mock_qwen3_omni_config.audio_token_id video_token_id = mock_qwen3_omni_config.video_token_id audio_count = updates.count(audio_token_id) video_count = updates.count(video_token_id) assert audio_count == audio_len, ( f"Expected {audio_len} audio tokens, got {audio_count}" ) # Calculate expected video token count spatial_merge_size = mock_qwen3_omni_config.vision_config.spatial_merge_size height = video_grid_thw[1] // spatial_merge_size width = video_grid_thw[2] // spatial_merge_size expected_video_count = video_grid_thw[0] * height * width assert video_count == expected_video_count, ( f"Expected {expected_video_count} video tokens, got {video_count}" ) # Total tokens should be: 1 (start) + audio_len + video_count + 1 (end) expected_total = 1 + audio_len + expected_video_count + 1 assert len(updates) == expected_total, ( f"Expected {expected_total} total tokens, got {len(updates)}" ) if __name__ == "__main__": pytest.main([__file__, "-v"])
{ "repo_id": "vllm-project/vllm", "file_path": "tests/model_executor/test_qwen3_omni.py", "license": "Apache License 2.0", "lines": 178, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/attention/test_rocm_attention_backends_selection.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests for attention backend selectors.""" from unittest.mock import MagicMock, patch import pytest import torch from vllm.platforms import current_platform from vllm.v1.attention.backends.registry import AttentionBackendEnum from vllm.v1.attention.selector import AttentionSelectorConfig # ROCm-specific attention backend selection tests pytestmark = pytest.mark.skipif( not current_platform.is_rocm(), reason="ROCm-specific tests" ) @pytest.fixture def mock_vllm_config(): """Create a mock VllmConfig for testing.""" config = MagicMock() config.model_config.dtype = torch.float16 config.model_config.hf_config.architectures = ["LlamaForCausalLM"] config.cache_config.block_size = 16 return config @pytest.fixture def mock_on_gfx9(): """Mock the on_gfx9 function to return True.""" with patch("vllm.platforms.rocm.on_gfx9", return_value=True): yield @pytest.mark.parametrize( "env_vars, selected_backend, expected_backend_path", [ # Test Case: Explicit FLEX_ATTENTION backend ( {}, "FLEX_ATTENTION", AttentionBackendEnum.FLEX_ATTENTION.get_path(), ), # Test Case 1: Default (no env vars, no explicit backend) ( {}, None, AttentionBackendEnum.TRITON_ATTN.get_path(), ), # Test Case 2: Explicit TRITON_ATTN backend ( {}, "TRITON_ATTN", AttentionBackendEnum.TRITON_ATTN.get_path(), ), # Test Case 3: Explicit ROCM_ATTN backend ( {}, "ROCM_ATTN", AttentionBackendEnum.ROCM_ATTN.get_path(), ), # Test Case 4: Explicit ROCM_AITER_FA backend ( {}, "ROCM_AITER_FA", AttentionBackendEnum.ROCM_AITER_FA.get_path(), ), # Test Case 5: Explicit ROCM_AITER_UNIFIED_ATTN backend ( {}, "ROCM_AITER_UNIFIED_ATTN", AttentionBackendEnum.ROCM_AITER_UNIFIED_ATTN.get_path(), ), # Test Case 6: VLLM_ROCM_USE_AITER=1 # (defaults to AITER FA when MHA not explicitly disabled) ( {"VLLM_ROCM_USE_AITER": "1"}, None, AttentionBackendEnum.ROCM_AITER_FA.get_path(), ), # Test Case 7: VLLM_ROCM_USE_AITER=1 + VLLM_ROCM_USE_AITER_MHA=1 ( {"VLLM_ROCM_USE_AITER": "1", "VLLM_ROCM_USE_AITER_MHA": "1"}, None, AttentionBackendEnum.ROCM_AITER_FA.get_path(), ), # Test Case 8: VLLM_ROCM_USE_AITER=1 + VLLM_ROCM_USE_AITER_UNIFIED_ATTENTION=1 ( { "VLLM_ROCM_USE_AITER": "1", "VLLM_ROCM_USE_AITER_UNIFIED_ATTENTION": "1", }, None, AttentionBackendEnum.ROCM_AITER_UNIFIED_ATTN.get_path(), ), # Test Case 9: VLLM_ROCM_USE_AITER=1 + explicit TRITON_ATTN ( {"VLLM_ROCM_USE_AITER": "1"}, "TRITON_ATTN", AttentionBackendEnum.TRITON_ATTN.get_path(), ), # Test Case 10: VLLM_ROCM_USE_AITER=1 + VLLM_ROCM_USE_AITER_MHA=0 # (explicitly disabled) ( {"VLLM_ROCM_USE_AITER": "1", "VLLM_ROCM_USE_AITER_MHA": "0"}, None, AttentionBackendEnum.TRITON_ATTN.get_path(), ), # Test Case 11: VLLM_ROCM_USE_AITER=1 + explicit ROCM_ATTN ( {"VLLM_ROCM_USE_AITER": "1"}, "ROCM_ATTN", AttentionBackendEnum.ROCM_ATTN.get_path(), ), ], ) def test_standard_attention_backend_selection( env_vars, selected_backend, expected_backend_path, mock_vllm_config, mock_on_gfx9, monkeypatch, ): """Test standard attention backend selection with various configurations.""" # Set environment variables for key, value in env_vars.items(): monkeypatch.setenv(key, value) # Import after setting env vars to ensure they're picked up # Reload envs to pick up new environment variables import importlib import vllm.envs as envs importlib.reload(envs) # Convert string backend to enum if provided backend_enum = None if selected_backend: backend_enum = getattr(AttentionBackendEnum, selected_backend) # Get the backend class path from vllm.platforms.rocm import RocmPlatform attn_selector_config = AttentionSelectorConfig( head_size=128, dtype=torch.float16, kv_cache_dtype="auto", block_size=16, use_mla=False, has_sink=False, use_sparse=False, ) backend_path = RocmPlatform.get_attn_backend_cls( selected_backend=backend_enum, attn_selector_config=attn_selector_config ) assert backend_path == expected_backend_path @pytest.mark.parametrize( "env_vars, selected_backend, block_size, expected_backend_path, should_raise", [ # Test Case 1: TRITON_MLA with block_size != 1 ( {}, "TRITON_MLA", 16, AttentionBackendEnum.TRITON_MLA.get_path(), False, ), # Test Case 2: TRITON_MLA with block_size == 1 (should raise) ( {}, "TRITON_MLA", 1, None, True, ), # Test Case 3: ROCM_AITER_MLA with block_size == 1 ( {}, "ROCM_AITER_MLA", 1, AttentionBackendEnum.ROCM_AITER_MLA.get_path(), False, ), # Test Case 4: ROCM_AITER_MLA with block_size != 1 (should raise) ( {}, "ROCM_AITER_MLA", 16, AttentionBackendEnum.ROCM_AITER_MLA.get_path(), False, ), # Test Case 5: VLLM_ROCM_USE_AITER=1 with block_size == 1 ( {"VLLM_ROCM_USE_AITER": "1"}, None, 1, AttentionBackendEnum.ROCM_AITER_MLA.get_path(), False, ), # Test Case 6: VLLM_ROCM_USE_AITER=1 with block_size == 16 # (should use ROCM_AITER_MLA now, as it supports block_size 16) ( {"VLLM_ROCM_USE_AITER": "1"}, None, 16, AttentionBackendEnum.ROCM_AITER_MLA.get_path(), False, ), # Test Case 7: VLLM_ROCM_USE_AITER=1 + explicit TRITON_MLA ( {"VLLM_ROCM_USE_AITER": "1"}, "TRITON_MLA", 16, AttentionBackendEnum.TRITON_MLA.get_path(), False, ), # Test Case 8: Explicit ROCM_AITER_TRITON_MLA ( {}, "ROCM_AITER_TRITON_MLA", 16, AttentionBackendEnum.ROCM_AITER_TRITON_MLA.get_path(), False, ), ], ) def test_mla_backend_selection( env_vars, selected_backend, block_size, expected_backend_path, should_raise, mock_vllm_config, monkeypatch, ): """Test MLA backend selection with various configurations.""" # Set environment variables for key, value in env_vars.items(): monkeypatch.setenv(key, value) # Import after setting env vars # Reload envs import importlib import vllm.envs as envs importlib.reload(envs) # Mock is_aiter_mla_enabled based on env vars and block_size aiter_enabled = env_vars.get("VLLM_ROCM_USE_AITER") == "1" mock_rocm_ops = MagicMock() mock_rocm_ops.is_mla_enabled.return_value = aiter_enabled mock_aiter_module = MagicMock() mock_aiter_module.rocm_aiter_ops = mock_rocm_ops with patch.dict("sys.modules", {"vllm._aiter_ops": mock_aiter_module}): # Convert string backend to enum if provided backend_enum = None if selected_backend: backend_enum = getattr(AttentionBackendEnum, selected_backend) from vllm.platforms.rocm import RocmPlatform if should_raise: with pytest.raises(ValueError): attn_selector_config = AttentionSelectorConfig( head_size=128, dtype=torch.float16, kv_cache_dtype="auto", block_size=block_size, use_mla=True, has_sink=False, use_sparse=False, ) attn_selector_config = AttentionSelectorConfig( head_size=128, dtype=torch.float16, kv_cache_dtype="auto", block_size=block_size, use_mla=True, has_sink=False, use_sparse=False, ) backend_path = RocmPlatform.get_attn_backend_cls( selected_backend=backend_enum, attn_selector_config=attn_selector_config, ) else: attn_selector_config = AttentionSelectorConfig( head_size=128, dtype=torch.float16, kv_cache_dtype="auto", block_size=block_size, use_mla=True, has_sink=False, use_sparse=False, ) backend_path = RocmPlatform.get_attn_backend_cls( selected_backend=backend_enum, attn_selector_config=attn_selector_config ) assert backend_path == expected_backend_path def test_aiter_fa_requires_gfx9(mock_vllm_config): """Test that ROCM_AITER_FA requires gfx9 architecture.""" from vllm.platforms.rocm import RocmPlatform # Mock on_gfx9 to return False with ( patch("vllm.platforms.rocm.on_gfx9", return_value=False), pytest.raises( ValueError, match="only supported on gfx9", ), ): attn_selector_config = AttentionSelectorConfig( head_size=128, dtype=torch.float16, kv_cache_dtype="auto", block_size=16, use_mla=False, has_sink=False, use_sparse=False, ) RocmPlatform.get_attn_backend_cls( selected_backend=AttentionBackendEnum.ROCM_AITER_FA, attn_selector_config=attn_selector_config, ) def test_sparse_not_supported(mock_vllm_config): """Test that sparse attention is not supported on ROCm.""" from vllm.platforms.rocm import RocmPlatform with pytest.raises( AssertionError, match="Sparse MLA backend on ROCm only supports block size 1" ): attn_selector_config = AttentionSelectorConfig( head_size=128, dtype=torch.float16, kv_cache_dtype="auto", block_size=16, use_mla=False, has_sink=False, use_sparse=True, ) RocmPlatform.get_attn_backend_cls( selected_backend=None, attn_selector_config=attn_selector_config )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/attention/test_rocm_attention_backends_selection.py", "license": "Apache License 2.0", "lines": 324, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/compile/test_dynamic_shapes_compilation.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import gc import tempfile from contextlib import contextmanager import pytest import torch from vllm import LLM, SamplingParams from vllm.compilation.decorators import support_torch_compile from vllm.config import CompilationConfig, VllmConfig, set_current_vllm_config from vllm.config.compilation import ( CompilationMode, DynamicShapesConfig, DynamicShapesType, ) from vllm.forward_context import set_forward_context from vllm.tokenizers import get_tokenizer from vllm.utils.torch_utils import is_torch_equal_or_newer def get_test_models(): """Get list of models to test based on PyTorch version""" # TODO "Qwen/Qwen3-4B-Instruct-2507" fails Fix issue and support it. return ["gpt2", "Qwen/Qwen2-7B-Instruct", "meta-llama/Llama-3.1-8B"] @pytest.mark.parametrize("model_name", get_test_models()) @pytest.mark.parametrize( "shapes_type", [ DynamicShapesType.BACKED, DynamicShapesType.UNBACKED, DynamicShapesType.BACKED_SIZE_OBLIVIOUS, ], ) @pytest.mark.parametrize("use_aot_compile", ["0", "1"]) @pytest.mark.parametrize("use_bytecode_hook", [True, False]) @pytest.mark.parametrize("evaluate_guards", [False, True]) @pytest.mark.skipif(not is_torch_equal_or_newer("2.10.0"), reason="requires torch 2.10") def test_dynamic_shapes_compilation( monkeypatch, model_name, shapes_type, use_aot_compile, use_bytecode_hook, evaluate_guards, ): """Test that all dynamic shapes types compile successfully""" if use_bytecode_hook and shapes_type == DynamicShapesType.UNBACKED: pytest.skip("UNBACKED dynamic shapes require VLLM_USE_BYTECODE_HOOK=0") if evaluate_guards and shapes_type == DynamicShapesType.UNBACKED: pytest.skip("unbacked dynamic shapes do not add guards") if evaluate_guards and use_aot_compile: pytest.skip("evaluate_guards requires use_aot_compile=0") monkeypatch.setenv("VLLM_USE_AOT_COMPILE", use_aot_compile) monkeypatch.setenv("VLLM_USE_BYTECODE_HOOK", "1" if use_bytecode_hook else "0") prompt = "Hello, my name is" print(f"Testing {shapes_type.name} dynamic shapes...") # Initialize the model with specific dynamic shapes configuration model = LLM( model=model_name, compilation_config={ "mode": CompilationMode.VLLM_COMPILE, "dynamic_shapes_config": { "type": shapes_type.value, "evaluate_guards": evaluate_guards, }, }, max_model_len=1024, ) output = model.generate(prompt) result = output[0].outputs[0].text # Example of setting the sampling parameters tokenizer = get_tokenizer(model_name) yes_tokens = tokenizer.encode("yes", add_special_tokens=False) no_tokens = tokenizer.encode("no", add_special_tokens=False) allowed_ids = list(set(yes_tokens + no_tokens)) sampling_params = SamplingParams( max_tokens=1, temperature=0, allowed_token_ids=allowed_ids ) output = model.generate( "answer with yes or no is " + result + " rubbish for prompt " + prompt + "?", sampling_params=sampling_params, ) result = output[0].outputs[0].text assert result == "yes" # Clean up GPU memory del model gc.collect() torch.cuda.empty_cache() torch.cuda.synchronize() print("GPU memory cleared") @pytest.mark.parametrize("use_aot_compile", ["0", "1"]) @pytest.mark.parametrize( "dynamic_shapes_type", [ DynamicShapesType.BACKED, DynamicShapesType.BACKED_SIZE_OBLIVIOUS, ], ) @pytest.mark.parametrize("evaluate_guards", [False, True]) def test_model_specialization_with_evaluate_guards( monkeypatch, use_aot_compile, dynamic_shapes_type, evaluate_guards ): """Test that evaluate_guards correctly detects shape specialization violations. """ if ( use_aot_compile == "1" and dynamic_shapes_type == DynamicShapesType.BACKED and evaluate_guards ): pytest.skip("evaluate_guards for backed does not work with aot_compile=1") @support_torch_compile class ModelWithSizeCheck(torch.nn.Module): def __init__(self, **kwargs): super().__init__() def forward(self, x: torch.Tensor): # This will cause specialization - torch.compile will guard on # sx.shape[0] if x.shape[0] >= 10: return x * 10 else: return x * 10 @support_torch_compile class ModelWithOneSizeCheck(torch.nn.Module): def __init__(self, **kwargs): super().__init__() def forward(self, x: torch.Tensor): # This will cause 0/1 specializations. if x.shape[0] == 0: return x * 10 if x.shape[0] == 1: return x * 10 else: return x * 10 @contextmanager def use_vllm_config(vllm_config: VllmConfig): with set_forward_context({}, vllm_config), set_current_vllm_config(vllm_config): yield monkeypatch.setenv("TOKENIZERS_PARALLELISM", "true") monkeypatch.setenv("VLLM_USE_AOT_COMPILE", use_aot_compile) monkeypatch.setenv("VLLM_USE_BYTECODE_HOOK", "0") # Create vllm config with the desired settings from vllm.config import CompilationMode vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, dynamic_shapes_config=DynamicShapesConfig( type=dynamic_shapes_type, evaluate_guards=evaluate_guards, ), ) ) def test(model_class, input1, input2, is_01_specialization=False): with ( torch.no_grad(), use_vllm_config(vllm_config), tempfile.TemporaryDirectory() as tmpdirname, ): monkeypatch.setenv("VLLM_CACHE_ROOT", tmpdirname) model = model_class(vllm_config=vllm_config).cuda() model(input1) if evaluate_guards and ( not ( is_01_specialization and dynamic_shapes_type == DynamicShapesType.BACKED ) ): # This should fail because guards were added. with pytest.raises(RuntimeError) as excinfo: model(input2) # Expected failure - guard was violated error_msg = str(excinfo.value) assert ( "GuardManager check failed" in error_msg or "Detected recompile when torch.compile stance" in error_msg ), error_msg else: model(input2) test(ModelWithSizeCheck, torch.randn(20, 10).cuda(), torch.randn(5, 10).cuda()) test(ModelWithSizeCheck, torch.randn(5, 10).cuda(), torch.randn(20, 10).cuda()) test( ModelWithOneSizeCheck, torch.randn(20, 10).cuda(), torch.randn(1, 10).cuda(), is_01_specialization=True, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/compile/test_dynamic_shapes_compilation.py", "license": "Apache License 2.0", "lines": 184, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/distributed/eplb/async_worker.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ The async worker that transfers experts in the background. """ import asyncio import threading from typing import TYPE_CHECKING import torch from torch.distributed import ProcessGroup from vllm.distributed.parallel_state import get_eplb_group from vllm.logger import init_logger from .rebalance_execute import transfer_layer if TYPE_CHECKING: from .eplb_state import EplbModelState, EplbState logger = init_logger(__name__) def start_async_worker( state: "EplbState", is_profile: bool = False, ) -> threading.Thread: eplb_group = get_eplb_group().device_group rank = eplb_group.rank() device_index = state.cuda_device_index assert state.is_async def thread_target() -> None: assert device_index is not None torch.cuda.set_device(device_index) cuda_stream = torch.cuda.Stream(device=device_index) loop = asyncio.new_event_loop() asyncio.set_event_loop(loop) try: loop.run_until_complete( transfer_run_periodically( state=state, eplb_group=eplb_group, cuda_stream=cuda_stream, is_profile=is_profile, ) ) except Exception as exc: # pragma: no cover - diagnostic path logger.exception("async loop error (Rank %d): %s", rank, str(exc)) finally: loop.close() thread = threading.Thread(target=thread_target, daemon=True) thread.start() return thread def run_rebalance_experts( model_state: "EplbModelState", eplb_state: "EplbState", physical_to_logical_map_cpu: torch.Tensor, ) -> None: assert model_state.eplb_stats is not None eplb_stats = model_state.eplb_stats # Wait for the main thread's all-reduce and clone to complete before # accessing the global_expert_load_window tensor. assert model_state.window_ready_event is not None model_state.window_ready_event.wait() model_state.window_ready_event = None # Move the global expert load window to CPU for computation. global_expert_load_window = eplb_stats.global_expert_load_window.cpu() # Compute new expert mappings for the model ( new_physical_to_logical_map, new_logical_to_physical_map, new_logical_replica_count, ) = eplb_state.policy.rebalance_experts( global_expert_load_window, eplb_stats.num_replicas, eplb_stats.num_groups, eplb_stats.num_nodes, eplb_stats.num_gpus, physical_to_logical_map_cpu, ) assert new_physical_to_logical_map.device == torch.device("cpu") model_state.new_physical_to_logical_map = new_physical_to_logical_map max_slots = model_state.logical_to_physical_map.shape[-1] padded_logical = torch.nn.functional.pad( new_logical_to_physical_map, (0, max(0, max_slots - new_logical_to_physical_map.shape[-1])), value=-1, ).to(model_state.logical_to_physical_map.device) new_replica = new_logical_replica_count.to(model_state.logical_replica_count.device) model_state.new_logical_to_physical_map = padded_logical model_state.new_logical_replica_count = new_replica async def transfer_run_periodically( state: "EplbState", eplb_group: ProcessGroup, cuda_stream: torch.cuda.Stream, is_profile: bool = False, ) -> None: while True: await asyncio.to_thread(state.rearrange_event.wait) logger.info("async worker woke up for EPLB transfer") assert state.is_async for model_state in state.model_states.values(): rebalancing_algorithm_executed = False physical_to_logical_map_cpu = None current_num_layers = model_state.model.num_moe_layers while ( model_state.rebalanced and model_state.layer_to_transfer < current_num_layers ): if not model_state.ep_buffer_ready and model_state.rebalanced: # Polling the lock directly in the async thread avoids # the thread switch overhead of asyncio.to_thread. # This is typically faster than offloading to a worker thread. while not model_state.buffer_lock.acquire(blocking=False): await asyncio.sleep(0) try: if model_state.layer_to_transfer >= current_num_layers: break if ( not rebalancing_algorithm_executed or model_state.new_physical_to_logical_map is None ): # Move the physical_to_logical_map to CPU # for rebalancing and transfer_layer. physical_to_logical_map_cpu = ( model_state.physical_to_logical_map.cpu() ) run_rebalance_experts( model_state, state, physical_to_logical_map_cpu ) rebalancing_algorithm_executed = True logger.info( "Async worker computed new indices for model %s", model_state.model_name, ) assert model_state.new_physical_to_logical_map is not None assert physical_to_logical_map_cpu is not None layer_idx = model_state.layer_to_transfer old_layer_indices = physical_to_logical_map_cpu[layer_idx] new_layer_indices = model_state.new_physical_to_logical_map[ layer_idx ] # Wait for the main thread to finish consuming the buffer # before initiating an EPLB transfer on another layer. if model_state.buffer_consumed_event is not None: cuda_stream.wait_event(model_state.buffer_consumed_event) model_state.buffer_consumed_event = None ( model_state.is_unchanged, model_state.is_received_locally, model_state.recv_metadata, ) = await transfer_layer( old_layer_indices=old_layer_indices, new_layer_indices=new_layer_indices, expert_weights=model_state.model.expert_weights[layer_idx], expert_weights_buffer=model_state.expert_buffer, ep_group=eplb_group, is_profile=is_profile, cuda_stream=cuda_stream, ) event = torch.cuda.Event(blocking=False) cuda_stream.record_event(event) model_state.buffer_ready_event = event model_state.ep_buffer_ready = 1 finally: model_state.buffer_lock.release() else: if not model_state.rebalanced: break await asyncio.sleep(0.001) state.rearrange_event.clear()
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/distributed/eplb/async_worker.py", "license": "Apache License 2.0", "lines": 164, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/opencua.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # # Adapted from Qwen2.5-VL implementation # Copyright 2025 The vLLM team. # Copyright 2025 XLANG Lab, The University of Hong Kong """Inference-only OpenCUA-7B model compatible with HuggingFace weights.""" from collections.abc import Mapping, Sequence from typing import Any import torch import torch.nn as nn from transformers import BatchFeature from transformers.models.qwen2_vl import ( Qwen2VLImageProcessor, Qwen2VLProcessor, Qwen2VLVideoProcessor, ) from vllm.config import VllmConfig from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import ( MultiModalFieldConfig, MultiModalKwargsItems, ) from vllm.multimodal.parse import MultiModalDataItems from vllm.multimodal.processing import ( BaseMultiModalProcessor, PromptReplacement, PromptUpdate, ) from vllm.tokenizers import TokenizerLike from .qwen2_5_vl import ( Qwen2_5_VisionTransformer as OpenCUAVisionTransformer, ) from .qwen2_5_vl import ( Qwen2_5_VLForConditionalGeneration, ) from .qwen2_vl import ( Qwen2VLDummyInputsBuilder, Qwen2VLMultiModalDataParser, Qwen2VLProcessingInfo, _create_qwen2vl_field_factory, ) from .utils import ( WeightsMapper, init_vllm_registered_model, maybe_prefix, ) class OpenCUAProcessingInfo(Qwen2VLProcessingInfo): def get_data_parser(self): return Qwen2VLMultiModalDataParser( self.get_hf_config().vision_config.spatial_merge_size, expected_hidden_size=self._get_expected_hidden_size(), ) def get_hf_config(self): return self.ctx.get_hf_config() def get_supported_mm_limits(self) -> Mapping[str, int | None]: return {"image": None} def get_hf_processor(self, **kwargs: object): """Load OpenCUA processor.""" tokenizer = self.get_tokenizer() vision_config = self.ctx.get_hf_image_processor_config() return OpenCUAProcessor( vision_config=vision_config, tokenizer=tokenizer, **kwargs, ) class OpenCUAProcessor(Qwen2VLProcessor): def check_argument_for_proper_class(self, attribute_name: str, arg: object) -> None: if attribute_name == "tokenizer": return return super().check_argument_for_proper_class(attribute_name, arg) def __init__( self, vision_config: dict, tokenizer: TokenizerLike, **kwargs, ): image_processor = Qwen2VLImageProcessor(**vision_config) video_processor = Qwen2VLVideoProcessor(**vision_config) chat_template = kwargs.pop("chat_template", None) super().__init__( image_processor=image_processor, tokenizer=tokenizer, video_processor=video_processor, chat_template=chat_template, **kwargs, ) self.image_token = "<|media_placeholder|>" def __call__( self, text=None, images=None, return_tensors=None, **kwargs, ): if text is not None: if not isinstance(text, list): text = [text] text_inputs = self.tokenizer(text, **kwargs) else: text_inputs = {} image_inputs = {} if images is not None: if not isinstance(images, list): images = [images] if len(images) > 0: image_inputs = self.image_processor( images, return_tensors=return_tensors or "pt" ) combined_inputs = {**text_inputs, **image_inputs} return BatchFeature(combined_inputs, tensor_type=return_tensors) class OpenCUAMultiModalProcessor(BaseMultiModalProcessor[OpenCUAProcessingInfo]): def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: return _create_qwen2vl_field_factory( self.info.get_hf_config().vision_config.spatial_merge_size )(hf_inputs) def _hf_processor_applies_updates( self, prompt_text: str, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, object], tokenization_kwargs: Mapping[str, object], ) -> bool: """vLLM이 prompt 업데이트를 처리하도록 False 반환.""" return False def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, Any], out_mm_kwargs: MultiModalKwargsItems, ) -> Sequence[PromptUpdate]: hf_processor = self.info.get_hf_processor(**hf_processor_mm_kwargs) image_processor = self.info.get_image_processor(**hf_processor_mm_kwargs) tokenizer = self.info.get_tokenizer() vocab = tokenizer.get_vocab() hf_config = self.info.get_hf_config() image_token_str = getattr(hf_processor, "image_token", "<|media_placeholder|>") image_token_id = vocab.get( image_token_str, getattr(hf_config, "media_placeholder_token_id", 151664), ) merge_length = image_processor.merge_size**2 def get_replacement_opencua(item_idx: int): out_item = out_mm_kwargs["image"][item_idx] grid_thw = out_item["image_grid_thw"].data assert isinstance(grid_thw, torch.Tensor) num_tokens = int(grid_thw.prod()) // merge_length return [image_token_id] * num_tokens return [ PromptReplacement( modality="image", target=[image_token_id], replacement=get_replacement_opencua, ) ] class OpenCUADummyInputsBuilder(Qwen2VLDummyInputsBuilder): def get_dummy_text(self, mm_counts: Mapping[str, int]) -> str: num_images = mm_counts.get("image", 0) image_token = "<|media_placeholder|>" return image_token * num_images @MULTIMODAL_REGISTRY.register_processor( OpenCUAMultiModalProcessor, info=OpenCUAProcessingInfo, dummy_inputs=OpenCUADummyInputsBuilder, ) class OpenCUAForConditionalGeneration(Qwen2_5_VLForConditionalGeneration): packed_modules_mapping = { "qkv_proj": ["q_proj", "k_proj", "v_proj"], "gate_up_proj": ["gate_proj", "up_proj"], } hf_to_vllm_mapper = WeightsMapper( orig_to_new_prefix={ "model.language_model.": "language_model.model.", "model.visual.": "visual.", "vision_tower.": "visual.", "lm_head.": "language_model.lm_head.", "model.": "language_model.model.", } ) supports_encoder_tp_data = True @classmethod def get_placeholder_str(cls, modality: str, i: int) -> str | None: if modality.startswith("image"): return "<|media_placeholder|>" raise ValueError("Only image modality is supported") def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): nn.Module.__init__(self) config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config multimodal_config = vllm_config.model_config.multimodal_config self.use_data_parallel = multimodal_config.mm_encoder_tp_mode == "data" self.config = config self.vllm_config = vllm_config self.multimodal_config = multimodal_config self.quant_config = quant_config self.is_multimodal_pruning_enabled = ( multimodal_config.is_multimodal_pruning_enabled() ) with self._mark_tower_model(vllm_config, "image"): self.visual = OpenCUAVisionTransformer( vision_config=config.vision_config, norm_eps=getattr(config, "rms_norm_eps", 1e-6), quant_config=self.quant_config, prefix=maybe_prefix(prefix, "visual"), ) with self._mark_language_model(vllm_config): self.language_model = init_vllm_registered_model( vllm_config=vllm_config, hf_config=config.text_config, prefix=maybe_prefix(prefix, "language_model"), architectures=["Qwen2ForCausalLM"], ) self.make_empty_intermediate_tensors = ( self.language_model.make_empty_intermediate_tensors )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/opencua.py", "license": "Apache License 2.0", "lines": 219, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/worker/gpu/spec_decode/rejection_sample.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.triton_utils import tl, triton @triton.jit def _rejection_sample_kernel( sampled_ptr, # [num_reqs, num_speculative_steps + 1] sampled_stride, num_sampled_ptr, # [num_reqs] target_sampled_ptr, # [num_draft_tokens + num_reqs] input_ids_ptr, # [num_draft_tokens + num_reqs] cu_num_logits_ptr, # [num_reqs + 1] ): req_idx = tl.program_id(0) start_idx = tl.load(cu_num_logits_ptr + req_idx) end_idx = tl.load(cu_num_logits_ptr + req_idx + 1) num_tokens = end_idx - start_idx num_sampled = 0 rejected = False for i in range(num_tokens - 1): if not rejected: target_sampled = tl.load(target_sampled_ptr + start_idx + i) draft_sampled = tl.load(input_ids_ptr + start_idx + i + 1) tl.store(sampled_ptr + req_idx * sampled_stride + i, target_sampled) num_sampled += 1 if target_sampled != draft_sampled: rejected = True if not rejected: target_sampled = tl.load(target_sampled_ptr + start_idx + num_tokens - 1) tl.store( sampled_ptr + req_idx * sampled_stride + num_tokens - 1, target_sampled ) num_sampled += 1 tl.store(num_sampled_ptr + req_idx, num_sampled) def rejection_sample( # [num_draft_tokens + num_reqs] target_sampled: torch.Tensor, # [num_draft_tokens + num_reqs] input_ids: torch.Tensor, # [num_reqs + 1] cu_num_logits: torch.Tensor, num_speculative_steps: int, ) -> tuple[torch.Tensor, torch.Tensor]: num_reqs = cu_num_logits.shape[0] - 1 sampled = torch.empty( num_reqs, num_speculative_steps + 1, dtype=target_sampled.dtype, device=target_sampled.device, ) num_sampled = torch.empty( num_reqs, dtype=torch.int32, device=target_sampled.device, ) _rejection_sample_kernel[(num_reqs,)]( sampled, sampled.stride(0), num_sampled, target_sampled, input_ids, cu_num_logits, num_warps=1, ) return sampled, num_sampled
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/worker/gpu/spec_decode/rejection_sample.py", "license": "Apache License 2.0", "lines": 65, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:examples/online_serving/openai_responses_client.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Set up this example by starting a vLLM OpenAI-compatible server. Reasoning models can be used through the Responses API as seen here https://platform.openai.com/docs/api-reference/responses For example: vllm serve Qwen/Qwen3-8B --reasoning-parser qwen3 """ from openai import OpenAI input_messages = [{"role": "user", "content": "What model are you?"}] def main(): base_url = "http://localhost:8000/v1" client = OpenAI(base_url=base_url, api_key="empty") model = "Qwen/Qwen3-8B" # get_first_model(client) response = client.responses.create( model=model, input=input_messages, ) for message in response.output: if message.type == "reasoning": # append reasoning message input_messages.append(message) response_2 = client.responses.create( model=model, input=input_messages, ) print(response_2.output_text) # I am Qwen, a large language model developed by Alibaba Cloud. # I am designed to assist with a wide range of tasks, including # answering questions, creating content, coding, and engaging in # conversations. I can help with various topics and provide # information or support in multiple languages. How can I assist you today? if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "examples/online_serving/openai_responses_client.py", "license": "Apache License 2.0", "lines": 35, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/distributed/eplb_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os import random import torch import torch.multiprocessing as mp from vllm.config import VllmConfig, set_current_vllm_config from vllm.distributed.parallel_state import ( init_distributed_environment, ) from vllm.utils.system_utils import update_environment_variables mp.set_start_method("spawn", force=True) def distributed_run(fn, world_size, *args): number_of_processes = world_size processes: list[mp.Process] = [] for i in range(number_of_processes): env: dict[str, str] = {} env["RANK"] = str(i) env["LOCAL_RANK"] = str(i) env["WORLD_SIZE"] = str(number_of_processes) env["LOCAL_WORLD_SIZE"] = str(number_of_processes) env["MASTER_ADDR"] = "localhost" env["MASTER_PORT"] = "12345" p = mp.Process(target=fn, args=(env, world_size, *args)) processes.append(p) p.start() for p in processes: p.join() for p in processes: assert p.exitcode == 0 def set_env_vars_and_device(env: dict[str, str]) -> None: update_environment_variables(env) local_rank = os.environ["LOCAL_RANK"] device = torch.device(f"cuda:{local_rank}") torch.cuda.set_device(device) # Create a minimal vllm config for init_distributed_environment vllm_config = VllmConfig() with set_current_vllm_config(vllm_config): init_distributed_environment() # Ensure each worker process has the same random seed random.seed(42) torch.manual_seed(42)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/distributed/eplb_utils.py", "license": "Apache License 2.0", "lines": 42, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test