sample_id
stringlengths
21
196
text
stringlengths
105
936k
metadata
dict
category
stringclasses
6 values
vllm-project/vllm:vllm/config/cache.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math from dataclasses import field from typing import TYPE_CHECKING, Any, Literal from pydantic import Field, SkipValidation, field_validator from vllm.config.utils import config from vllm.logger import init_logger from vllm.utils.mem_constants import GiB_bytes from vllm.utils.mem_utils import format_gib, get_cpu_memory if TYPE_CHECKING: from vllm.config.parallel import ParallelConfig else: ParallelConfig = Any logger = init_logger(__name__) BlockSize = Literal[1, 8, 16, 32, 64, 128, 256] CacheDType = Literal[ "auto", "bfloat16", "fp8", "fp8_e4m3", "fp8_e5m2", "fp8_inc", "fp8_ds_mla", ] MambaDType = Literal["auto", "float32", "float16"] MambaCacheMode = Literal["all", "align", "none"] PrefixCachingHashAlgo = Literal["sha256", "sha256_cbor", "xxhash", "xxhash_cbor"] KVOffloadingBackend = Literal["native", "lmcache"] @config class CacheConfig: """Configuration for the KV cache.""" block_size: SkipValidation[BlockSize] = None # type: ignore[assignment] """Size of a contiguous cache block in number of tokens. On CUDA devices, only block sizes up to 32 are supported. This config has no static default. If left unspecified by the user, it will be set in `Platform.check_and_update_config()` based on the current platform.""" gpu_memory_utilization: float = Field(default=0.9, gt=0, le=1) """The fraction of GPU memory to be used for the model executor, which can range from 0 to 1. For example, a value of 0.5 would imply 50% GPU memory utilization. If unspecified, will use the default value of 0.9. This is a per-instance limit, and only applies to the current vLLM instance. It does not matter if you have another vLLM instance running on the same GPU. For example, if you have two vLLM instances running on the same GPU, you can set the GPU memory utilization to 0.5 for each instance.""" swap_space: float = Field(default=4, ge=0) """Size of the CPU swap space per GPU (in GiB).""" cache_dtype: CacheDType = "auto" """Data type for kv cache storage. If "auto", will use model data type. CUDA 11.8+ supports fp8 (=fp8_e4m3) and fp8_e5m2. ROCm (AMD GPU) supports fp8 (=fp8_e4m3). Intel Gaudi (HPU) supports fp8 (using fp8_inc). Some models (namely DeepSeekV3.2) default to fp8, set to bfloat16 to use bfloat16 instead, this is an invalid option for models that do not default to fp8. """ is_attention_free: bool = False """Whether the model is attention-free. This is primarily set in `ModelConfig` and that value should be manually duplicated here.""" num_gpu_blocks_override: int | None = None """Number of GPU blocks to use. This overrides the profiled `num_gpu_blocks` if specified. Does nothing if `None`. Used for testing preemption.""" sliding_window: int | None = None """Sliding window size for the KV cache. This is primarily set in `ModelConfig` and that value should be manually duplicated here.""" enable_prefix_caching: bool = True """Whether to enable prefix caching.""" prefix_caching_hash_algo: PrefixCachingHashAlgo = "sha256" """Set the hash algorithm for prefix caching:\n - "sha256" uses Pickle for object serialization before hashing. This is the current default, as SHA256 is the most secure choice to avoid potential hash collisions.\n - "sha256_cbor" provides a reproducible, cross-language compatible hash. It serializes objects using canonical CBOR and hashes them with SHA-256.\n - "xxhash" uses Pickle serialization with xxHash (128-bit) for faster, non-cryptographic hashing. Requires the optional ``xxhash`` package. IMPORTANT: Use of a hashing algorithm that is not considered cryptographically secure theoretically increases the risk of hash collisions, which can cause undefined behavior or even leak private information in multi-tenant environments. Even if collisions are still very unlikely, it is important to consider your security risk tolerance against the performance benefits before turning this on.\n - "xxhash_cbor" combines canonical CBOR serialization with xxHash for reproducible hashing. Requires the optional ``xxhash`` package.""" cpu_offload_gb: float = Field(default=0, ge=0) """The space in GiB to offload to CPU, per GPU. Default is 0, which means no offloading. Intuitively, this argument can be seen as a virtual way to increase the GPU memory size. For example, if you have one 24 GB GPU and set this to 10, virtually you can think of it as a 34 GB GPU. Then you can load a 13B model with BF16 weight, which requires at least 26GB GPU memory. Note that this requires fast CPU-GPU interconnect, as part of the model is loaded from CPU memory to GPU memory on the fly in each model forward pass. DEPRECATED: This field is deprecated and will be removed in v0.16. Please use OffloadConfig.uva.cpu_offload_gb instead. """ cpu_offload_params: set[str] = Field(default_factory=set) """The set of parameter name segments to target for CPU offloading. DEPRECATED: This field is deprecated and will be removed in v0.16. Please use OffloadConfig.uva.cpu_offload_params instead. """ calculate_kv_scales: bool = False """This enables dynamic calculation of `k_scale` and `v_scale` when kv_cache_dtype is fp8. If `False`, the scales will be loaded from the model checkpoint if available. Otherwise, the scales will default to 1.0.""" cpu_kvcache_space_bytes: int | None = None """(CPU backend only) CPU key-value cache space.""" mamba_page_size_padded: int | None = None """ Optional override for mamba page size; used by hybrid mamba/attention models to ensure exact alignment with attention page size.""" mamba_block_size: int | None = Field(default=None, gt=0) """Size of a contiguous cache block in number of tokens for mamba cache. Can be set only when prefix caching is enabled. Value must be a multiple of 8 to align with causal_conv1d kernel.""" mamba_cache_dtype: MambaDType = "auto" """The data type to use for the Mamba cache (both the conv as well as the ssm state). If set to 'auto', the data type will be inferred from the model config.""" mamba_ssm_cache_dtype: MambaDType = "auto" """The data type to use for the Mamba cache (ssm state only, conv state will still be controlled by mamba_cache_dtype). If set to 'auto', the data type for the ssm state will be determined by mamba_cache_dtype.""" mamba_cache_mode: MambaCacheMode = "none" """The cache strategy for Mamba layers. - "none": set when prefix caching is disabled. - "all": cache the mamba state of all tokens at position i * block_size. This is the default behavior (for models that support it) when prefix caching is enabled. - "align": only cache the mamba state of the last token of each scheduler step and when the token is at position i * block_size. """ # Will be set after profiling. num_gpu_blocks: int | None = field(default=None, init=False) """The number of blocks to allocate for GPU memory.""" num_cpu_blocks: int | None = field(default=None, init=False) """The number of blocks to allocate for CPU memory.""" kv_sharing_fast_prefill: bool = False """This feature is work in progress and no prefill optimization takes place with this flag enabled currently. In some KV sharing setups, e.g. YOCO (https://arxiv.org/abs/2405.05254), some layers can skip tokens corresponding to prefill. This flag enables attention metadata for eligible layers to be overridden with metadata necessary for implementing this optimization in some models (e.g. Gemma3n) """ kv_cache_memory_bytes: int | None = None """Size of KV Cache per GPU in bytes. By default, this is set to None and vllm can automatically infer the kv cache size based on gpu_memory_utilization. However, users may want to manually specify the kv cache memory size. kv_cache_memory_bytes allows more fine-grain control of how much memory gets used when compared with using gpu_memory_utilization. Note that kv_cache_memory_bytes (when not-None) ignores gpu_memory_utilization""" kv_offloading_size: float | None = None """Size of the KV cache offloading buffer in GiB. When TP > 1, this is the total buffer size summed across all TP ranks. By default, this is set to None, which means no KV offloading is enabled. When set, vLLM will enable KV cache offloading to CPU using the kv_offloading_backend.""" kv_offloading_backend: KVOffloadingBackend = "native" """The backend to use for KV cache offloading. Supported backends include 'native' (vLLM native CPU offloading), 'lmcache'. KV offloading is only activated when kv_offloading_size is set.""" def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ ignored_factors = { # Runtime/derived knobs that don't affect compiled graph shape "gpu_memory_utilization", "swap_space", "is_attention_free", "num_gpu_blocks_override", "enable_prefix_caching", "prefix_caching_hash_algo", "cpu_kvcache_space_bytes", "mamba_page_size_padded", # Post-init/derived counters "num_gpu_blocks", "num_cpu_blocks", # WIP feature toggle not impacting compiled graph shape "kv_sharing_fast_prefill", } from vllm.config.utils import get_hash_factors, hash_factors factors = get_hash_factors(self, ignored_factors) return hash_factors(factors) def metrics_info(self): # convert cache_config to dict(key: str, value: str) for prometheus # metrics info return {key: str(value) for key, value in self.__dict__.items()} @field_validator("cache_dtype", mode="after") @classmethod def _validate_cache_dtype(cls, cache_dtype: CacheDType) -> CacheDType: if cache_dtype.startswith("fp8"): logger.info( "Using fp8 data type to store kv cache. It reduces the GPU " "memory footprint and boosts the performance. " "Meanwhile, it may cause accuracy drop without a proper " "scaling factor." ) return cache_dtype def verify_with_parallel_config( self, parallel_config: ParallelConfig, ) -> None: swap_space_bytes = math.ceil(self.swap_space * GiB_bytes) total_cpu_memory = get_cpu_memory() # FIXME(woosuk): Here, it is assumed that the GPUs in a tensor parallel # group are in the same node. However, the GPUs may span multiple nodes. num_gpus_per_node = parallel_config.tensor_parallel_size cpu_memory_usage = swap_space_bytes * num_gpus_per_node msg = ( f"{format_gib(cpu_memory_usage)} GiB out of the " f"{format_gib(total_cpu_memory)} GiB total CPU memory " "is allocated for the swap space." ) if cpu_memory_usage > 0.7 * total_cpu_memory: raise ValueError("Too large swap space. " + msg) elif cpu_memory_usage > 0.4 * total_cpu_memory: logger.warning("Possibly too large swap space. %s", msg)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/config/cache.py", "license": "Apache License 2.0", "lines": 224, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:benchmarks/kernels/benchmark_mrope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # This script benchmarks the mrope kernel (mainly for Qwen2VL and Qwen2.5VL models). # It generates test data, runs benchmarks, and saves results to a CSV file. # # The CSV file (named with current date/time) contains these columns: # model_name, tp_size, num_tokens, num_heads, num_kv_heads, head_dim, max_position, # is_neox_style, rope_parameters, dtype, torch_mean, torch_median, torch_p99, # torch_min, torch_max, triton_mean, triton_median, triton_p99, triton_min, triton_max, # speedup # # == Usage Examples == # # Single model benchmark: # python3 benchmark_mrope.py --model-name Qwen/Qwen2-VL-7B-Instruct --tp-size 1 \ # --warmup-iter 10 --benchmark-iter 100 --dtype bfloat16 --seed 0 --num-tokens 1024 # # All models benchmark: # python3 benchmark_mrope.py --model-name "" --tp-size 1 --warmup-iter 10 \ # --benchmark-iter 100 --dtype bfloat16 --seed 0 --num-tokens 1024 # # All models with different TP sizes: # python3 benchmark_mrope.py --model-name "" --tp-size 1 2 4 8 --warmup-iter 10 \ # --benchmark-iter 100 --dtype bfloat16 --seed 0 --num-tokens 1024 # # All models with different token counts: # python3 benchmark_mrope.py --model-name "" --tp-size 1 --warmup-iter 10 \ # --benchmark-iter 100 --dtype bfloat16 --seed 0 --num-tokens 1024 4096 16384 import csv import os import time from datetime import datetime from typing import Any import numpy as np import torch from vllm.benchmarks.lib.utils import default_vllm_config from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.transformers_utils.config import get_config from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.torch_utils import set_random_seed device = torch.device("cuda" if torch.cuda.is_available() else "cpu") def generate_test_data( num_tokens: int, num_q_heads: int, num_kv_heads: int, head_size: int, max_position_embeddings: int, dtype: torch.dtype, device: torch.device, ): """Generate test data for given configuration.""" # Create 2D positions (3, num_tokens) for multimodal case positions = torch.randint( 0, max_position_embeddings // 4, (3, num_tokens), device=device ) # Create query and key tensors query = torch.randn(num_tokens, num_q_heads * head_size, dtype=dtype, device=device) key = torch.randn(num_tokens, num_kv_heads * head_size, dtype=dtype, device=device) return positions, query, key def calculate_stats(times: list[float]) -> dict[str, float]: """Calculate statistics from a list of times.""" times_array = np.array(times) return { "mean": np.mean(times_array), "median": np.median(times_array), "p99": np.percentile(times_array, 99), "min": np.min(times_array), "max": np.max(times_array), } @default_vllm_config() def benchmark_mrope( model_name: str, num_tokens: int, head_dim: int, tp_size: int, num_heads: int, num_kv_heads: int, max_position: int = 8192, is_neox_style: bool = True, rope_parameters: dict[str, Any] | None = None, dtype: torch.dtype = torch.bfloat16, seed: int = 0, warmup_iter: int = 10, benchmark_iter: int = 100, csv_writer=None, ): set_random_seed(seed) torch.set_default_device(device) # the parameters to compute the q k v size based on tp_size mrope_helper_class = get_rope( head_size=head_dim, max_position=max_position, is_neox_style=is_neox_style, rope_parameters=rope_parameters, dtype=dtype, ).to(device=device) print(80 * "=") print( f"Evaluating model: {model_name} " f"with tp_size: {tp_size} " f"and num_tokens: {num_tokens}, " f"dtype: {dtype}" ) # create q k v input tensors # create rotary pos emb input tensors positions, query, key = generate_test_data( num_tokens, num_heads, num_kv_heads, head_dim, max_position, dtype, device ) # Warm up for _ in range(warmup_iter): mrope_helper_class.forward_native( positions, query.clone(), key.clone(), ) mrope_helper_class.forward_cuda( positions, query.clone(), key.clone(), ) torch.cuda.synchronize() # Time reference implementation torch_times = [] for _ in range(benchmark_iter): query_clone = query.clone() key_clone = key.clone() torch.cuda.synchronize() start_time = time.time() mrope_helper_class.forward_native( positions, query_clone, key_clone, ) torch.cuda.synchronize() torch_times.append(time.time() - start_time) # Time triton kernel implementation triton_times = [] for _ in range(benchmark_iter): query_clone = query.clone() key_clone = key.clone() torch.cuda.synchronize() start_time = time.time() mrope_helper_class.forward_cuda( positions, query_clone, key_clone, ) torch.cuda.synchronize() triton_times.append(time.time() - start_time) # Calculate statistics torch_stats = calculate_stats(torch_times) triton_stats = calculate_stats(triton_times) print(f"\nPerformance for config ({num_tokens}, {num_heads}, {num_kv_heads}):") print( f"Torch implementation: " f"mean={torch_stats['mean']:.8f}s, " f"median={torch_stats['median']:.8f}s, " f"p99={torch_stats['p99']:.8f}s" ) print( f"Triton implementation: " f"mean={triton_stats['mean']:.8f}s, " f"median={triton_stats['median']:.8f}s, " f"p99={triton_stats['p99']:.8f}s" ) print( f"Triton Speedup over Torch: {torch_stats['mean'] / triton_stats['mean']:.8f}x" ) # Write to CSV if csv_writer: row = [ model_name, tp_size, num_tokens, num_heads, num_kv_heads, head_dim, max_position, is_neox_style, str(rope_parameters), str(dtype).split(".")[-1], torch_stats["mean"], torch_stats["median"], torch_stats["p99"], torch_stats["min"], torch_stats["max"], triton_stats["mean"], triton_stats["median"], triton_stats["p99"], triton_stats["min"], triton_stats["max"], torch_stats["mean"] / triton_stats["mean"], # speedup ] csv_writer.writerow(row) return torch_stats, triton_stats if __name__ == "__main__": parser = FlexibleArgumentParser( description="Benchmark the rotary embedding kernels." ) parser.add_argument("--model-name", type=str, default="") parser.add_argument("--tp-size", type=int, default=1) parser.add_argument("--warmup-iter", type=int, default=10) parser.add_argument("--benchmark-iter", type=int, default=100) parser.add_argument("--dtype", type=str, choices=["bfloat16"], default="bfloat16") parser.add_argument("--seed", type=int, default=0) parser.add_argument("--num-tokens", type=int, nargs="+", required=False) parser.add_argument("--trust-remote-code", action="store_true") parser.add_argument("--output-csv", type=str, default="mrope_benchmark_results.csv") args = parser.parse_args() print(args) # Create CSV file for results timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") csv_filename = f"{os.path.splitext(args.output_csv)[0]}_{timestamp}.csv" with open(csv_filename, "w", newline="") as csvfile: csv_writer = csv.writer(csvfile) # Write header header = [ "model_name", "tp_size", "num_tokens", "num_heads", "num_kv_heads", "head_dim", "max_position", "is_neox_style", "rope_parameters", "dtype", "torch_mean", "torch_median", "torch_p99", "torch_min", "torch_max", "triton_mean", "triton_median", "triton_p99", "triton_min", "triton_max", "speedup", ] csv_writer.writerow(header) model_tp_dict = {} if args.model_name == "": model_tp_dict = { "Qwen/Qwen2-VL-2B-Instruct": [1], "Qwen/Qwen2-VL-7B-Instruct": [1], "Qwen/Qwen2-VL-72B-Instruct": [2, 4, 8], "Qwen/Qwen2.5-VL-3B-Instruct": [1, 2, 4, 8], "Qwen/Qwen2.5-VL-7B-Instruct": [1, 2, 4, 8], "Qwen/Qwen2.5-VL-72B-Instruct": [2, 4, 8], } else: model_tp_dict[args.model_name] = [args.tp_size] if args.num_tokens is None: num_tokens_list = [2**i for i in range(0, 18)] else: num_tokens_list = args.num_tokens for model_name, tp_list in model_tp_dict.items(): config = get_config(model_name, trust_remote_code=args.trust_remote_code) for tp_size in tp_list: # get the model config total_num_kv_heads = config.num_key_value_heads total_num_heads = config.num_attention_heads num_heads = total_num_heads // tp_size num_kv_heads = max(1, total_num_kv_heads // tp_size) head_dim = config.hidden_size // total_num_heads q_size = num_heads * head_dim kv_size = num_kv_heads * head_dim is_neox_style = True rope_parameters = config.rope_parameters max_position = config.max_position_embeddings for num_tokens in num_tokens_list: benchmark_mrope( model_name=model_name, num_tokens=num_tokens, head_dim=head_dim, tp_size=tp_size, num_heads=num_heads, num_kv_heads=num_kv_heads, max_position=max_position, is_neox_style=is_neox_style, rope_parameters=rope_parameters, dtype=getattr(torch, args.dtype), seed=args.seed, warmup_iter=args.warmup_iter, benchmark_iter=args.benchmark_iter, csv_writer=csv_writer, ) print(f"Benchmark results saved to {csv_filename}")
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/kernels/benchmark_mrope.py", "license": "Apache License 2.0", "lines": 288, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/gemma3n_mm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Iterable, Mapping, Sequence from typing import Annotated, Any, Literal import numpy as np import torch from torch import nn from transformers import AutoModel, BatchFeature from transformers.models.gemma3n import ( Gemma3nAudioConfig, Gemma3nAudioFeatureExtractor, Gemma3nConfig, Gemma3nProcessor, Gemma3nTextConfig, Gemma3nVisionConfig, ) from transformers.models.siglip import SiglipImageProcessorFast from vllm.config import ModelConfig, SpeechToTextConfig, VllmConfig from vllm.config.multimodal import BaseDummyOptions from vllm.inputs.data import PromptType, TextPrompt from vllm.logger import init_logger from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import RowParallelLinear from vllm.model_executor.layers.vocab_parallel_embedding import VocabParallelEmbedding from vllm.model_executor.models.gemma3n import Gemma3nForCausalLM from vllm.model_executor.models.gemma3n_audio_utils import ( adjust_audio_features_to_expected_length, ) from vllm.model_executor.models.module_mapping import MultiModelKeys from vllm.model_executor.models.whisper import ISO639_1_SUPPORTED_LANGS from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import ( MultiModalDataDict, MultiModalFieldConfig, MultiModalKwargsItems, ) from vllm.multimodal.parse import ( ImageProcessorItems, MultiModalDataItems, MultiModalDataParser, ) from vllm.multimodal.processing import BaseDummyInputsBuilder from vllm.multimodal.processing.processor import ( BaseMultiModalProcessor, BaseProcessingInfo, MultiModalPromptUpdates, MultiModalPromptUpdatesApplyResult, PlaceholderFeaturesInfo, PromptReplacement, PromptUpdate, PromptUpdateDetails, replace_token_matches, ) from vllm.sequence import IntermediateTensors from vllm.utils.tensor_schema import TensorSchema, TensorShape from .interfaces import MultiModalEmbeddings, SupportsMultiModal, SupportsTranscription from .utils import ( AutoWeightsLoader, WeightsMapper, init_vllm_registered_model, maybe_prefix, ) logger = init_logger(__name__) # This should be based on model config but we hardcode them for now. TOKENS_PER_IMAGE = 256 TOKENS_PER_AUDIO = 188 class Gemma3nImagePixelInputs(TensorSchema): """ Dimensions: - bn: Batch size * number of images - c: Number of channels (3) - h: Height of each patch - w: Width of each patch """ type: Literal["pixel_values"] = "pixel_values" pixel_values: Annotated[torch.Tensor, TensorShape("bn", 3, "h", "w")] class Gemma3nAudioInputs(TensorSchema): """ Dimensions: - bn: Batch size * number of audios - s: seq_length - f: num_features """ type: Literal["audio"] = "audio" input_features_padded: Annotated[torch.Tensor, TensorShape("bn", "s", "f")] input_features_mask: Annotated[torch.Tensor, TensorShape("bn", "s")] Gemma3nImageInputs = Gemma3nImagePixelInputs class Gemma3nProcessingInfo(BaseProcessingInfo): def get_hf_config(self): return self.ctx.get_hf_config(Gemma3nConfig) def get_hf_processor(self, **kwargs: object): return self.ctx.get_hf_processor(Gemma3nProcessor, **kwargs) def get_feature_extractor(self, **kwargs: object) -> Gemma3nAudioFeatureExtractor: return self.get_hf_processor(**kwargs).feature_extractor def get_data_parser(self): feature_extractor = self.get_feature_extractor() return MultiModalDataParser( target_sr=feature_extractor.sampling_rate, expected_hidden_size=self._get_expected_hidden_size(), ) def get_supported_mm_limits(self) -> Mapping[str, int | None]: return {"image": None, "audio": None} def get_max_tokens_per_item( self, seq_len: int, mm_counts: Mapping[str, int] ) -> Mapping[str, int] | None: return {"image": TOKENS_PER_IMAGE, "audio": TOKENS_PER_AUDIO} def get_image_repl( self, *, image_width: int, image_height: int, processor: Gemma3nProcessor, ) -> str: """ Get the replacement text for image tokens. For Gemma3n, this should return the full_image_sequence which includes BOI token, repeated image tokens, and EOI token. """ return PromptUpdateDetails.select_token_id( processor.full_image_sequence, processor.image_token_id ) def get_audio_repl( self, *, processor: Gemma3nProcessor, ) -> str: """ Get the replacement text for audio tokens. For Gemma3n, this should return the full_audio_sequence which includes BOA token, repeated audio tokens, and EOA token. """ # Return the full audio sequence as defined by the processor return PromptUpdateDetails.select_token_id( processor.full_audio_sequence, processor.audio_token_id ) class Gemma3nDummyInputsBuilder(BaseDummyInputsBuilder[Gemma3nProcessingInfo]): def get_dummy_text(self, mm_counts: Mapping[str, int]) -> str: num_images = mm_counts.get("image", 0) num_audios = mm_counts.get("audio", 0) processor = self.info.get_hf_processor() image_token = processor.image_token audio_token = processor.audio_token return image_token * num_images + audio_token * num_audios def get_dummy_mm_data( self, seq_len: int, mm_counts: Mapping[str, int], mm_options: Mapping[str, BaseDummyOptions], ) -> MultiModalDataDict: num_images = mm_counts.get("image", 0) num_audios = mm_counts.get("audio", 0) processor = self.info.get_hf_processor() audio_feature_extractor: Gemma3nAudioFeatureExtractor = ( processor.feature_extractor ) audio_len = audio_feature_extractor.fft_length image_processor: SiglipImageProcessorFast = processor.image_processor img_width = image_processor.size.get("width", 224) img_height = image_processor.size.get("height", 224) image_overrides = mm_options.get("image") audio_overrides = mm_options.get("audio") return { "image": self._get_dummy_images( width=img_width, height=img_height, num_images=num_images, overrides=image_overrides, ), "audio": self._get_dummy_audios( length=audio_len, num_audios=num_audios, overrides=audio_overrides, ), } class Gemma3nMultiModalProcessor(BaseMultiModalProcessor[Gemma3nProcessingInfo]): def _call_hf_processor( self, prompt: str, mm_data: Mapping[str, object], mm_kwargs: Mapping[str, object], tok_kwargs: Mapping[str, object], ) -> BatchFeature: # HF Transformers audio processor no longer accepts `audios` key. # We pop `audios` and replace it with `audio` key to suppress # the warning. if "audios" in mm_data: mm_data["audio"] = mm_data.pop("audios") processed_outputs = super()._call_hf_processor( prompt, mm_data, mm_kwargs, tok_kwargs, ) if "input_features" in processed_outputs: # Padding enables audio_tower to run in batched mode processed_outputs["input_features_padded"] = processed_outputs[ "input_features" ] # Unpad features here since we need the output of each item to be # independent of other items for the cache to work correctly unpadded_features = [ f[mask] for f, mask in zip( processed_outputs["input_features"], processed_outputs["input_features_mask"], ) ] processed_outputs["input_features"] = unpadded_features return processed_outputs def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: return dict( pixel_values=MultiModalFieldConfig.batched("image"), input_features_padded=MultiModalFieldConfig.batched("audio"), input_features_mask=MultiModalFieldConfig.batched("audio"), ) def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, Any], out_mm_kwargs: MultiModalKwargsItems, ) -> Sequence[PromptUpdate]: hf_processor = self.info.get_hf_processor(**hf_processor_mm_kwargs) prompt_updates = [] # Handle image tokens if "image" in mm_items: image_token = hf_processor.image_token def get_replacement_image(item_idx: int): images = mm_items.get_items("image", ImageProcessorItems) image_size = images.get_image_size(item_idx) return self.info.get_image_repl( image_width=image_size.width, image_height=image_size.height, processor=hf_processor, ) prompt_updates.append( PromptReplacement( modality="image", target=image_token, replacement=get_replacement_image, ) ) # Handle audio tokens if "audio" in mm_items: audio_token = hf_processor.audio_token def get_replacement_audio(item_idx: int): return self.info.get_audio_repl( processor=hf_processor, ) prompt_updates.append( PromptReplacement( modality="audio", target=audio_token, replacement=get_replacement_audio, ) ) return prompt_updates def _apply_token_matches( self, prompt: list[int], mm_prompt_updates: MultiModalPromptUpdates, ) -> tuple[list[int], MultiModalPromptUpdatesApplyResult]: token_ids, res = super()._apply_token_matches(prompt, mm_prompt_updates) # "\n\n\n" and "\n\n\n\n" are single tokens # Since our replacement can insert "\n\n" next to "\n" # tokens, we have to combine them to be consistent with # the output of the tokenizer tokenizer = self.info.get_tokenizer() vocab = tokenizer.get_vocab() newline_1 = vocab["\n"] newline_2 = vocab["\n\n"] newline_3 = vocab["\n\n\n"] newline_4 = vocab["\n\n\n\n"] token_ids = replace_token_matches( token_ids, [newline_1, newline_2], [newline_3], ) token_ids = replace_token_matches( token_ids, [newline_2, newline_1], [newline_3], ) token_ids = replace_token_matches( token_ids, [newline_2, newline_2], [newline_4], ) return token_ids, res def _find_mm_placeholders( self, new_token_ids: list[int], mm_prompt_updates: MultiModalPromptUpdates, ) -> Mapping[str, list[PlaceholderFeaturesInfo]]: # We need to detect "\n\n" inside "\n\n\n" and "\n\n\n\n" tokenizer = self.info.get_tokenizer() vocab = tokenizer.get_vocab() newline_1 = vocab["\n"] newline_2 = vocab["\n\n"] newline_3 = vocab["\n\n\n"] newline_4 = vocab["\n\n\n\n"] def get_repl_toks(tok: int) -> list[int]: if tok == newline_3: return [newline_1, newline_2] if tok == newline_4: return [newline_2, newline_2] return [tok] repl_token_ids = list[int]() repl_orig_idxs = list[int]() for orig_idx, orig_tok in enumerate(new_token_ids): repl_toks = get_repl_toks(orig_tok) repl_token_ids.extend(repl_toks) repl_orig_idxs.extend(orig_idx for _ in range(len(repl_toks))) repls = super()._find_mm_placeholders(repl_token_ids, mm_prompt_updates) return { modality: [ PlaceholderFeaturesInfo( modality=p.modality, item_idx=p.item_idx, start_idx=repl_orig_idxs[p.start_idx], tokens=p.tokens, is_embed=p.is_embed, ) for p in placeholders ] for modality, placeholders in repls.items() } class Gemma3nMultimodalEmbedder(nn.Module): """Embeds token ids or soft tokens for multimodal content into language model space.""" def __init__( self, multimodal_config: Gemma3nAudioConfig | Gemma3nVisionConfig, text_config: Gemma3nTextConfig, ): super().__init__() self.multimodal_hidden_size = multimodal_config.hidden_size self.eps = multimodal_config.rms_norm_eps self.vocab_offset = multimodal_config.vocab_offset self.vocab_size = multimodal_config.vocab_size self.text_hidden_size = text_config.hidden_size self.embedding = VocabParallelEmbedding( self.vocab_size, self.multimodal_hidden_size, ) self.hard_embedding_norm = RMSNorm( self.multimodal_hidden_size, eps=self.eps, ) self.soft_embedding_norm = RMSNorm( self.multimodal_hidden_size, eps=self.eps, ) self.embedding_projection = RowParallelLinear( self.multimodal_hidden_size, self.text_hidden_size, bias=False, ) self.embedding_post_projection_norm = RMSNorm( self.text_hidden_size, eps=self.eps, has_weight=False, ) def forward( self, input_ids: torch.LongTensor | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor: """Embeds token ids or soft tokens for multimodal content into language model space. Args: input_ids: A torch.LongTensor containing the token ids to embed. Values should be in the range `[vocab_offset, vocab_offset + vocab_size)`. inputs_embeds: A torch.Tensor containing the soft tokens to embed. Returns: A torch.Tensor of embeddings with shape `[batch_size, seq_len, self.config.text_config.hidden_size]`. """ # noqa: E501 if (input_ids is None) ^ (inputs_embeds is not None): raise ValueError( "You must specify exactly one of input_ids or inputs_embeds" ) if inputs_embeds is not None: emb_norm = self.soft_embedding_norm(inputs_embeds) else: hard_emb = self.embedding(input_ids - self.vocab_offset) emb_norm = self.hard_embedding_norm(hard_emb) emb_norm_proj, _ = self.embedding_projection(emb_norm) return self.embedding_post_projection_norm(emb_norm_proj) @MULTIMODAL_REGISTRY.register_processor( Gemma3nMultiModalProcessor, info=Gemma3nProcessingInfo, dummy_inputs=Gemma3nDummyInputsBuilder, ) class Gemma3nForConditionalGeneration( nn.Module, SupportsMultiModal, SupportsTranscription ): supported_languages = ISO639_1_SUPPORTED_LANGS packed_modules_mapping = { "qkv_proj": [ "q_proj", "k_proj", "v_proj", ], "gate_up_proj": [ "gate_proj", "up_proj", ], } hf_to_vllm_mapper = WeightsMapper( orig_to_new_prefix={ # mapping for new names in checkpoint saved after transformers v4.52 "model.embed_audio.": "embed_audio.", "model.embed_vision.": "embed_vision.", "model.language_model.": "language_model.model.", "model.vision_tower.": "vision_tower.", "model.audio_tower.": "audio_tower.", "model.multi_modal_projector.": "multi_modal_projector.", "lm_head.": "language_model.lm_head.", "model": "language_model.model", } ) def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config multimodal_config = vllm_config.model_config.multimodal_config self.config = config self.quant_config = quant_config self.multimodal_config = multimodal_config self.vocab_size = config.text_config.vocab_size with self._mark_tower_model(vllm_config, "image"): self.vision_tower = AutoModel.from_config(config=config.vision_config) self.embed_vision = Gemma3nMultimodalEmbedder( config.vision_config, config.text_config ) with self._mark_tower_model(vllm_config, "audio"): self.audio_tower = AutoModel.from_config(config=config.audio_config) self.embed_audio = Gemma3nMultimodalEmbedder( config.audio_config, config.text_config ) with self._mark_language_model(vllm_config): self.language_model: Gemma3nForCausalLM = init_vllm_registered_model( vllm_config=vllm_config, hf_config=config.text_config, prefix=maybe_prefix(prefix, "language_model"), architectures=["Gemma3nForCausalLM"], ) # NOTE (NickLucche) In order to be compatible with cudagraph, the # buffer needs to be consistent, so we pre-allocate here. self.per_layer_embeddings = torch.zeros( vllm_config.scheduler_config.max_num_batched_tokens, self.config.text_config.num_hidden_layers, self.config.text_config.hidden_size_per_layer_input, device=self.language_model.model.embed_tokens.weight.device, dtype=self.language_model.model.embed_tokens.weight.dtype, ) def _parse_and_validate_image_input( self, **kwargs: object ) -> Gemma3nImageInputs | None: pixel_values = kwargs.pop("pixel_values", None) image_embeds = kwargs.pop("image_embeds", None) # TODO is this the case? assert image_embeds is None, "Gemma3n does not support image_embeds." if pixel_values is None: return None return Gemma3nImagePixelInputs(pixel_values=pixel_values) def _parse_and_validate_audio_input( self, **kwargs: object ) -> Gemma3nAudioInputs | None: input_features_padded = kwargs.pop("input_features_padded", None) if input_features_padded is None: return None input_features_mask = kwargs.pop("input_features_mask", None) if input_features_mask is None: return None return Gemma3nAudioInputs( input_features_padded=input_features_padded, input_features_mask=input_features_mask, ) def _parse_and_validate_multimodal_inputs(self, **kwargs: object) -> dict: mm_input_by_modality = {} # Preserve the order of modalities if there are multiple of them # from the order of kwargs. for input_key in kwargs: if ( input_key in ("pixel_values", "image_embeds") and "image" not in mm_input_by_modality ): mm_input_by_modality["image"] = self._parse_and_validate_image_input( **kwargs ) if ( input_key == "input_features_padded" and "audio" not in mm_input_by_modality ): mm_input_by_modality["audio"] = self._parse_and_validate_audio_input( **kwargs ) return mm_input_by_modality def _process_image_input( self, image_input: Gemma3nImageInputs, ) -> list[torch.Tensor]: pixel_values = image_input["pixel_values"] vision_outputs = self.vision_tower( pixel_values=pixel_values, do_pooling=False, return_dict=True ).last_hidden_state # TODO try to avoid copy here # (batch, channels, height, width) to (batch, height * width, channels) vision_outputs = ( vision_outputs.reshape( vision_outputs.shape[0], self.config.vision_config.hidden_size, self.config.vision_soft_tokens_per_image, ) .permute(0, 2, 1) .contiguous() ) # Normalize and embed the soft tokens into language model space. vision_outputs *= self.config.vision_config.hidden_size**0.5 # Return a list of embeddings instead of a batched tensor return self.embed_vision(inputs_embeds=vision_outputs).unbind(0) def _process_audio_input( self, audio_input: Gemma3nAudioInputs, ) -> list[torch.Tensor]: # Run on padded features to enable batching input_features = audio_input["input_features_padded"].squeeze(1) input_features_mask = audio_input["input_features_mask"].squeeze(1) audio_outputs = self.audio_tower(input_features, ~input_features_mask) if isinstance(audio_outputs, tuple): # Transformers v4 audio_encodings, audio_mask = audio_outputs else: # Transformers v5 audio_encodings = audio_outputs.last_hidden_state audio_mask = audio_outputs.audio_mel_mask audio_features = self.embed_audio(inputs_embeds=audio_encodings) # The Gemma3nProcessor expects all audio will be 30s in length and # inserts 188 audio soft tokens into the text to account for this. # However, the audio preprocessing and encoder do not guarantee they # will produce exactly 188 soft tokens; they may produce fewer tokens # (for shorter audio) or more tokens (for longer audio or due to # BOA/EOA special tokens in the placeholder sequence). # We handle both cases: # - If fewer tokens: pad with the embedding of the last vocab token # - If more tokens: truncate to the expected count # TODO precompute and cache padding audio_padding_toks = torch.tensor( [[self.vocab_size - 1]], dtype=torch.long, device=audio_features.device ) audio_padding_embs = self.embed_audio(input_ids=audio_padding_toks) audio_features = torch.where( audio_mask.unsqueeze(-1), audio_padding_embs, audio_features ) expected_tokens = self.config.audio_soft_tokens_per_image audio_features, tokens_truncated = adjust_audio_features_to_expected_length( audio_features, expected_tokens, audio_padding_embs ) if tokens_truncated > 0: logger.warning( "Gemma3n audio encoder produced %d extra tokens. " "Truncating to match placeholder count of %d.", tokens_truncated, expected_tokens, ) # Return a list of embeddings instead of a batched tensor return audio_features.unbind(0) def embed_multimodal(self, **kwargs: object) -> MultiModalEmbeddings: mm_input_by_modality = self._parse_and_validate_multimodal_inputs(**kwargs) if mm_input_by_modality is None: return [] multimodal_embeddings: list[torch.Tensor] = [] # NOTE: It is important to iterate over the keys in this dictionary # to preserve the order of the modalities. for modality in mm_input_by_modality: multimodal_input = mm_input_by_modality[modality] if modality == "image": vision_embeddings = self._process_image_input(multimodal_input) multimodal_embeddings.extend(vision_embeddings) if modality == "audio": audio_embeddings = self._process_audio_input(multimodal_input) multimodal_embeddings.extend(audio_embeddings) return multimodal_embeddings def embed_input_ids( self, input_ids: torch.Tensor, multimodal_embeddings: MultiModalEmbeddings | None = None, *, is_multimodal: torch.Tensor | None = None, handle_oov_mm_token: bool = False, ) -> torch.Tensor: # NOTE (NickLucche) Each pass needs tokens to compute PLE so we cache # them here, as the model forward has only access to the input_embeds. if input_ids is not None: per_layer_inputs = self.language_model.model.get_per_layer_input_embeddings( input_ids ) per_layer_inputs = per_layer_inputs.reshape( -1, self.config.text_config.num_hidden_layers, self.config.text_config.hidden_size_per_layer_input, ) self.per_layer_embeddings[: per_layer_inputs.shape[0]].copy_( per_layer_inputs ) # This is to satisfy the type checker for each overload if multimodal_embeddings is None or is_multimodal is None: return super().embed_input_ids(input_ids) return super().embed_input_ids( input_ids, multimodal_embeddings=multimodal_embeddings, is_multimodal=is_multimodal, handle_oov_mm_token=handle_oov_mm_token, ) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs: object, ) -> IntermediateTensors: if intermediate_tensors is not None: inputs_embeds = None # NOTE (NickLucche) During profiling, `embed_input_ids` is not # called, hence we don't have input_ids to compute PLEs. We simply # select a chunk of pre-allocated PLEs. During normal execution, # `embed_input_ids` is called before forward, hence this slice # will contain PLEs computed from the actual input_ids. per_layer_inputs = self.per_layer_embeddings[: inputs_embeds.shape[0]] hidden_states = self.language_model.model( input_ids, positions, per_layer_inputs=per_layer_inputs, intermediate_tensors=intermediate_tensors, inputs_embeds=inputs_embeds, **kwargs, ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: return self.language_model.compute_logits(hidden_states) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader(self) return loader.load_weights(weights, mapper=self.hf_to_vllm_mapper) def get_mm_mapping(self) -> MultiModelKeys: """ Get the module prefix in multimodal models """ return MultiModelKeys.from_string_field( language_model="language_model", connector="multi_modal_projector", tower_model="vision_tower", ) @classmethod def get_placeholder_str(cls, modality: str, i: int) -> str | None: if modality == "image": return "<image_soft_token>" elif modality == "audio": return "<audio_soft_token>" else: raise ValueError(f"Unsupported modality: {modality}") @classmethod def get_generation_prompt( cls, audio: np.ndarray, stt_config: SpeechToTextConfig, model_config: ModelConfig, language: str | None, task_type: Literal["transcribe", "translate"], request_prompt: str, to_language: str | None, ) -> PromptType: """ Gemma3n supports "free-form" transcription. We fix its prompt here to standardize transcriptions/translations requests. """ # Transcribe this audio [into <>] | for transcription # Translate this audio [from <> into <>] | for translation prompt = "<start_of_turn>user\n" prompt += "Transcribe" if task_type == "transcribe" else "Translate" prompt += " this audio" # We assume the language is a valid ISO 639-1 code. full_lang_name = cls.supported_languages.get(language, "") # Translation only for now full_lang_name_to = cls.supported_languages.get(to_language, "") if task_type == "transcribe" and full_lang_name: prompt += f" into {full_lang_name}" elif task_type == "translate": if full_lang_name: prompt += f" from {full_lang_name}" if full_lang_name_to: prompt += f" into {full_lang_name_to}" prompt += ": <audio_soft_token><end_of_turn>\n<start_of_turn>model\n" return TextPrompt( prompt=prompt, multi_modal_data={"audio": (audio, stt_config.sample_rate)}, ) @classmethod def get_speech_to_text_config( cls, model_config: ModelConfig, task_type: str ) -> SpeechToTextConfig: return SpeechToTextConfig( # Let's set this to 30 as suggested in the docs for now, although # the model is only limited by its context length. max_audio_clip_s=30, sample_rate=16000, # TODO enable chunking after more thorough testing. min_energy_split_window_size=None, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/gemma3n_mm.py", "license": "Apache License 2.0", "lines": 716, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/config/parallel.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os from collections.abc import Callable from typing import TYPE_CHECKING, Any, Literal import torch from pydantic import Field, field_validator, model_validator from torch.distributed import ProcessGroup, ReduceOp from typing_extensions import Self import vllm.envs as envs from vllm.config.utils import config from vllm.logger import init_logger from vllm.model_executor.layers.batch_invariant import ( vllm_is_batch_invariant, ) from vllm.platforms import current_platform from vllm.utils.network_utils import get_open_ports_list from vllm.utils.torch_utils import cuda_device_count_stateless if TYPE_CHECKING: from ray.runtime_env import RuntimeEnv from ray.util.placement_group import PlacementGroup from vllm.v1.executor import Executor else: RuntimeEnv = Any PlacementGroup = Any Executor = Any logger = init_logger(__name__) ExpertPlacementStrategy = Literal["linear", "round_robin"] DistributedExecutorBackend = Literal["ray", "mp", "uni", "external_launcher"] DataParallelBackend = Literal["ray", "mp"] EPLBPolicyOption = Literal["default"] All2AllBackend = Literal[ "naive", "pplx", "deepep_high_throughput", "deepep_low_latency", "mori", "allgather_reducescatter", "flashinfer_all2allv", ] @config class EPLBConfig: """Configuration for Expert Parallel Load Balancing (EP).""" window_size: int = 1000 """Window size for expert load recording.""" step_interval: int = 3000 """ Interval for rearranging experts in expert parallelism. Note that if this is greater than the EPLB window size, only the metrics of the last `lb_window_size` steps will be used for rearranging experts. """ num_redundant_experts: int = Field(default=0, ge=0) """Number of redundant experts to use for expert parallelism.""" log_balancedness: bool = False """ Log the balancedness each step of expert parallelism. This is turned off by default since it will cause communication overhead. """ log_balancedness_interval: int = 1 """ Interval for logging the balancedness. """ use_async: bool = False """ Whether to use non-blocking EPLB. """ policy: EPLBPolicyOption = "default" """The policy type for expert parallel load balancing (EPLB).""" @model_validator(mode="after") def _validate_eplb_config(self) -> Self: if self.use_async and self.policy != "default": raise ValueError("Async EPLB is only supported with the default policy.") if self.log_balancedness and self.log_balancedness_interval <= 0: raise ValueError("log_balancedness_interval must be greater than 0.") return self @config class ParallelConfig: """Configuration for the distributed execution.""" pipeline_parallel_size: int = 1 """Number of pipeline parallel groups.""" tensor_parallel_size: int = 1 """Number of tensor parallel groups.""" prefill_context_parallel_size: int = 1 """Number of prefill context parallel groups.""" data_parallel_size: int = 1 """Number of data parallel groups. MoE layers will be sharded according to the product of the tensor parallel size and data parallel size.""" data_parallel_size_local: int = 1 """Number of local data parallel groups.""" data_parallel_rank: int = 0 """Rank of the data parallel group.""" data_parallel_rank_local: int | None = None """Local rank of the data parallel group, set only in SPMD mode.""" data_parallel_master_ip: str = "127.0.0.1" """IP of the data parallel master.""" data_parallel_rpc_port: int = 29550 """Port for data parallel messaging.""" data_parallel_master_port: int = 29500 """Port of the data parallel master.""" data_parallel_backend: DataParallelBackend = "mp" """Backend to use for data parallel, either "mp" or "ray".""" data_parallel_external_lb: bool = False """Whether to use "external" DP LB mode. Applies only to online serving and when data_parallel_size > 0. This is useful for a "one-pod-per-rank" wide-EP setup in Kubernetes. Set implicitly when --data-parallel-rank is provided explicitly to vllm serve.""" data_parallel_hybrid_lb: bool = False """Whether to use "hybrid" DP LB mode. Applies only to online serving and when data_parallel_size > 0. Enables running an AsyncLLM and API server on a "per-node" basis where vLLM load balances between local data parallel ranks, but an external LB balances between vLLM nodes/replicas. Set explicitly in conjunction with --data-parallel-start-rank.""" is_moe_model: bool | None = None """Whether the deployed model is MoE (if known).""" enable_expert_parallel: bool = False """Use expert parallelism instead of tensor parallelism for MoE layers.""" enable_eplb: bool = False """Enable expert parallelism load balancing for MoE layers.""" eplb_config: EPLBConfig = Field(default_factory=EPLBConfig) """Expert parallelism configuration.""" expert_placement_strategy: ExpertPlacementStrategy = "linear" """The expert placement strategy for MoE layers:\n - "linear": Experts are placed in a contiguous manner. For example, with 4 experts and 2 ranks, rank 0 will have experts [0, 1] and rank 1 will have experts [2, 3].\n - "round_robin": Experts are placed in a round-robin manner. For example, with 4 experts and 2 ranks, rank 0 will have experts [0, 2] and rank 1 will have experts [1, 3]. This strategy can help improve load balancing for grouped expert models with no redundant experts.""" all2all_backend: All2AllBackend = "allgather_reducescatter" """All2All backend for MoE expert parallel communication. Available options: - "naive": Naive all2all implementation using broadcasts\n - "allgather_reducescatter": All2all based on allgather and reducescatter\n - "deepep_high_throughput": Use deepep high-throughput kernels\n - "deepep_low_latency": Use deepep low-latency kernels\n - "mori": Use mori kernels\n - "flashinfer_all2allv": Use flashinfer alltoallv kernels for mnnvl""" max_parallel_loading_workers: int | None = None """Maximum number of parallel loading workers when loading model sequentially in multiple batches. To avoid RAM OOM when using tensor parallel and large models.""" disable_custom_all_reduce: bool = False """Disable the custom all-reduce kernel and fall back to NCCL.""" enable_elastic_ep: bool = False """Enable elastic expert parallelism with stateless NCCL groups for DP/EP.""" enable_dbo: bool = False """Enable dual batch overlap for the model executor.""" ubatch_size: int = 0 """Number of ubatch size.""" dbo_decode_token_threshold: int = 32 """The threshold for dual batch overlap for batches only containing decodes. If the number of tokens in the request is greater than this threshold, microbatching will be used. Otherwise, the request will be processed in a single batch.""" dbo_prefill_token_threshold: int = 512 # TODO(lucas): tune """The threshold for dual batch overlap for batches that contain one or more prefills. If the number of tokens in the request is greater than this threshold, microbatching will be used. Otherwise, the request will be processed in a single batch.""" disable_nccl_for_dp_synchronization: bool | None = Field(default=None) """Forces the dp synchronization logic in vllm/v1/worker/dp_utils.py to use Gloo instead of NCCL for its all reduce. Defaults to True when async scheduling is enabled, False otherwise. """ ray_workers_use_nsight: bool = False """Whether to profile Ray workers with nsight, see https://docs.ray.io/en/latest/ray-observability/user-guides/profiling.html#profiling-nsight-profiler.""" ray_runtime_env: RuntimeEnv | None = None """Ray runtime environment to pass to distributed workers.""" placement_group: PlacementGroup | None = None """ray distributed model workers placement group.""" distributed_executor_backend: ( str | DistributedExecutorBackend | type[Executor] | None ) = None """Backend to use for distributed model workers, either "ray" or "mp" (multiprocessing). If the product of pipeline_parallel_size and tensor_parallel_size is less than or equal to the number of GPUs available, "mp" will be used to keep processing on a single host. Otherwise, an error will be raised. To use "mp" you must also set nnodes, and to use "ray" you must manually set distributed_executor_backend to "ray". Note that tpu only support Ray for distributed inference.""" worker_cls: str = "auto" """The full name of the worker class to use. If "auto", the worker class will be determined based on the platform.""" sd_worker_cls: str = "auto" """The full name of the worker class to use for speculative decoding. If "auto", the worker class will be determined based on the platform.""" worker_extension_cls: str = "" """The full name of the worker extension class to use. The worker extension class is dynamically inherited by the worker class. This is used to inject new attributes and methods to the worker class for use in collective_rpc calls.""" master_addr: str = "127.0.0.1" """distributed master address for multi-node distributed inference when distributed_executor_backend is mp.""" master_port: int = 29501 """distributed master port for multi-node distributed inference when distributed_executor_backend is mp.""" node_rank: int = 0 """distributed node rank for multi-node distributed inference when distributed_executor_backend is mp.""" nnodes: int = 1 """num of nodes for multi-node distributed inference when distributed_executor_backend is mp.""" world_size: int = Field(init=False) """world_size is TPxPP, it affects the number of workers we create.""" rank: int = 0 """Global rank in distributed setup.""" _data_parallel_master_port_list: list[int] = Field(default_factory=list) """List of open port auto-queried for data parallel messaging. Set to be private as it's not intended to be configured by users. """ _stateless_dp_group_port_list: list[list[int]] = Field(default_factory=list) """List of open ports for stateless DP groups when enable_elastic_ep is True. Set to be private as it's not intended to be configured by users. It is a list of list[int], with each inner list contains a set of 3 ports to be used for setting up the stateless CPU/device/TCPStore groups in StatelessGroupCoordinator. The number of inner lists is equal to the number of DP groups, i.e., len(self._stateless_dp_group_port_list) == world_size_across_dp // dp_size, and len(self._stateless_dp_group_port_list[i]) == 3 for all i. """ _stateless_ep_group_port_list: list[list[int]] = Field(default_factory=list) """List of open ports for stateless EP groups when enable_elastic_ep is True. Set to be private as it's not intended to be configured by users. len(self._stateless_ep_group_port_list) == world_size_across_dp // ep_size, """ _stateless_eplb_group_port_list: list[list[int]] = Field(default_factory=list) """List of open ports for stateless EPLB groups when enable_elastic_ep is True. Same topology as EP but separate NCCL communicator to avoid deadlocks. """ _stateless_world_group_port_list: list[list[int]] = Field(default_factory=list) """List of open ports for stateless world group when enable_elastic_ep is True. Set to be private as it's not intended to be configured by users. len(self._stateless_world_group_port_list) == 1, """ decode_context_parallel_size: int = 1 """Number of decode context parallel groups, because the world size does not change by dcp, it simply reuse the GPUs of TP group, and tp_size needs to be divisible by dcp_size.""" dcp_kv_cache_interleave_size: int = 1 """ Interleave size of kv_cache storage while using DCP. dcp_kv_cache_interleave_size has been replaced by cp_kv_cache_interleave_size, and will be deprecated when PCP is fully supported. """ cp_kv_cache_interleave_size: int = 1 """Interleave size of kv_cache storage while using DCP or PCP. For `total_cp_rank = pcp_rank * dcp_world_size + dcp_rank`, and `total_cp_world_size = pcp_world_size * dcp_world_size`. store interleave_size tokens on total_cp_rank i, then store next interleave_size tokens on total_cp_rank i+1. Interleave_size=1: token-level alignment, where token `i` is stored on total_cp_rank `i % total_cp_world_size`. Interleave_size=block_size: block-level alignment, where tokens are first populated to the preceding ranks. Tokens are then stored in (rank i+1, block j) only after (rank i, block j) is fully occupied. Block_size should be greater than or equal to cp_kv_cache_interleave_size. Block_size should be divisible by cp_kv_cache_interleave_size. """ data_parallel_index: int = Field(init=False) """Equal to the data parallel rank but not used for torch process groups and not overridden for dense models.""" _api_process_count: int = Field(default=1, gt=0) """ The number of API processes initialized. Note: This is an internal config that is only valid for and should only be set by API server scale-out. """ _api_process_rank: int = Field(default=0, ge=-1) """ The rank of this API process, or `-1` for engine core processes under API server scale-out. Note: This is an internal config that is only valid for and should only be set by API server scale-out. """ @field_validator("disable_nccl_for_dp_synchronization", mode="wrap") @classmethod def _skip_none_validation(cls, value: Any, handler: Callable) -> Any: """Skip validation if the value is `None` when initialisation is delayed.""" return None if value is None else handler(value) @model_validator(mode="after") def _validate_parallel_config(self) -> Self: if self._api_process_rank >= self._api_process_count: raise ValueError( "Invalid value of `_api_process_rank`. " f"Expected to be `-1` or `[0, {self._api_process_count})`, " f"but found: {self._api_process_rank}" ) if self.all2all_backend == "pplx": logger.warning( "The 'pplx' all2all backend has been removed. " "Falling back to 'allgather_reducescatter'." ) self.all2all_backend = "allgather_reducescatter" if self.data_parallel_size_local > self.data_parallel_size: raise ValueError( f"data_parallel_size_local ({self.data_parallel_size_local}) " f"must be <= data_parallel_size ({self.data_parallel_size})" ) if self.data_parallel_size <= 1 and self.data_parallel_external_lb: raise ValueError( "data_parallel_external_lb can only be set when data_parallel_size > 1" ) if self.enable_eplb: if not current_platform.is_cuda_alike(): raise ValueError( "Expert parallelism load balancing is only supported on " "CUDA devices or ROCm devices now." ) if not self.enable_expert_parallel: raise ValueError("enable_expert_parallel must be True to use EPLB.") if self.tensor_parallel_size * self.data_parallel_size <= 1: raise ValueError( "EPLB requires tensor_parallel_size or data_parallel_size " f"to be greater than 1, but got " f"TP={self.tensor_parallel_size},DP={self.data_parallel_size}." ) else: if self.eplb_config.num_redundant_experts != 0: raise ValueError( "num_redundant_experts is set to " f"{self.eplb_config.num_redundant_experts} but EPLB is not " "enabled. Either enable EPLB or unset " "num_redundant_experts." ) # Note(hc): In the current implementation of decode context # parallel(DCP), tp_size needs to be divisible by dcp_size, # because the world size does not change by dcp, it simply # reuses the GPUs of TP group, and split one TP group into # tp_size//dcp_size DCP groups. if self.tensor_parallel_size % self.decode_context_parallel_size != 0: raise ValueError( f"tp_size={self.tensor_parallel_size} must be divisible by" f"dcp_size={self.decode_context_parallel_size}." ) return self @property def world_size_across_dp(self) -> int: """world_size_across_dp is TPxPPxDP, it is the size of the world including data parallelism.""" return self.world_size * self.data_parallel_size @property def use_ubatching(self) -> bool: return self.enable_dbo or self.ubatch_size > 1 @property def num_ubatches(self) -> int: return 2 if self.enable_dbo else self.ubatch_size @property def local_engines_only(self) -> bool: """ Client manages local+remote EngineCores in pure internal LB case. Client manages local EngineCores in hybrid and external LB case. """ return self.data_parallel_external_lb or self.data_parallel_hybrid_lb def get_next_dp_init_port(self) -> int: """ We might need to initialize process groups in multiple processes that is related to data parallelism, e.g. both in the worker and in the engine, which can live in different processes. To avoid port conflicts, we pop a new port from the prepared port list each time we need to initialize a new process group related to data parallelism. """ if self._data_parallel_master_port_list: answer = self._data_parallel_master_port_list.pop() else: answer = self.data_parallel_master_port self.data_parallel_master_port += 1 return answer def allocate_elastic_ep_ports(self) -> None: """Allocate all ports for elastic EP (stateless groups + DP master). Must be called AFTER ray.init() so that ports claimed by Ray's idle worker pool are already in use and won't be returned by get_open_ports_list(). """ if not self.enable_elastic_ep: return if self._stateless_world_group_port_list: return num_world_groups = 1 dp_size = self.data_parallel_size ep_size = self.data_parallel_size * self.world_size_across_dp num_dp_groups = max(1, self.world_size_across_dp // dp_size) num_ep_groups = max(1, self.world_size_across_dp // ep_size) num_eplb_groups = num_ep_groups total_stateless_ports = ( num_world_groups + num_dp_groups + num_ep_groups + num_eplb_groups ) * 3 num_dp_master_ports = 5 all_ports = get_open_ports_list(total_stateless_ports + num_dp_master_ports) self._data_parallel_master_port_list = all_ports[-num_dp_master_ports:] self.data_parallel_master_port = self._data_parallel_master_port_list.pop() all_ports = all_ports[:-num_dp_master_ports] self._stateless_world_group_port_list = [ all_ports[i : i + 3] for i in range(0, num_world_groups * 3, 3) ] start_idx = num_world_groups * 3 self._stateless_dp_group_port_list = [ all_ports[i : i + 3] for i in range(start_idx, start_idx + num_dp_groups * 3, 3) ] start_idx += num_dp_groups * 3 self._stateless_ep_group_port_list = [ all_ports[i : i + 3] for i in range(start_idx, start_idx + num_ep_groups * 3, 3) ] start_idx += num_ep_groups * 3 self._stateless_eplb_group_port_list = [ all_ports[i : i + 3] for i in range(start_idx, start_idx + num_eplb_groups * 3, 3) ] def get_next_stateless_world_group_port(self) -> list[int]: return self._stateless_world_group_port_list.pop() def get_next_stateless_dp_group_port(self) -> list[int]: return self._stateless_dp_group_port_list.pop() def get_next_stateless_ep_group_port(self) -> list[int]: return self._stateless_ep_group_port_list.pop() def get_next_stateless_eplb_group_port(self) -> list[int]: return self._stateless_eplb_group_port_list.pop() def stateless_init_dp_group(self, return_store: bool = False) -> ProcessGroup: # NOTE: In high-concurrency scenarios multiple processes # can pick the same (currently free) port through a race # condition when calling `get_open_port()`. When the first # process binds the port the others will subsequently fail # with `torch.distributed.DistNetworkError: EADDRINUSE`. # To make the initialization more robust we retry a few times # with a fresh port whenever this specific error is observed. from torch.distributed import DistNetworkError from vllm.distributed.utils import ( stateless_init_torch_distributed_process_group, ) max_retries = 5 last_exc: Exception | None = None for _ in range(max_retries): try: # use gloo since the engine process might not have cuda device return stateless_init_torch_distributed_process_group( self.data_parallel_master_ip, self.get_next_dp_init_port(), self.data_parallel_rank, self.data_parallel_size, backend="gloo", return_store=return_store, ) except DistNetworkError as e: # We only want to retry when the root cause is EADDRINUSE. if "EADDRINUSE" in str(e): logger.warning("Address already in use. Retrying with a new port.") last_exc = e continue # try again with a new port raise e # If we get here all retries have failed. assert last_exc is not None raise last_exc # The all_reduce at the end of attention (during o_proj) means that # inputs are replicated across each rank of the tensor parallel group. # If using expert-parallelism with DeepEP All2All ops, replicated # tokens results in useless duplicate computation and communication. # # In this case, ensure the input to the experts is sequence parallel # to avoid the excess work. # @property def use_sequence_parallel_moe(self) -> bool: return ( self.all2all_backend in ( "allgather_reducescatter", "naive", "deepep_high_throughput", "deepep_low_latency", "mori", ) and self.enable_expert_parallel and self.tensor_parallel_size > 1 and self.data_parallel_size > 1 ) @property def node_rank_within_dp(self) -> int: return self.node_rank % self.nnodes_within_dp @property def nnodes_within_dp(self) -> int: if self.nnodes == 1: return 1 data_parallel_node_size = ( self.data_parallel_size // self.data_parallel_size_local ) return self.nnodes // data_parallel_node_size @property def local_world_size(self) -> int: return self.world_size // self.nnodes_within_dp @staticmethod def has_unfinished_dp(dp_group: ProcessGroup, has_unfinished: bool) -> bool: tensor = torch.tensor([has_unfinished], dtype=torch.int32, device="cpu") # dp rank 0: has_unfinished_seqs=True # dp rank 1: has_unfinished_seqs=False # aggregated: has_unfinished_seqs=True # so this is an OR operation, i.e. MAX in integers torch.distributed.all_reduce(tensor, op=ReduceOp.MAX, group=dp_group) aggregated_has_unfinished = bool(tensor.item()) return aggregated_has_unfinished @staticmethod def sync_kv_cache_memory_size(dp_group: ProcessGroup, kv_cache_memory: int) -> int: if kv_cache_memory == -1: kv_cache_memory = torch.iinfo(torch.int64).max tensor = torch.tensor([kv_cache_memory], dtype=torch.int64, device="cpu") # we cannot use broadcast for stateless dp group since it depends # on global rank torch.distributed.all_reduce(tensor, op=ReduceOp.MIN, group=dp_group) return tensor.item() def compute_hash(self): """ Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. This hash is also used for DP worker configuration validation to prevent hangs from mismatched collective communication patterns. """ ignored_factors = { # Derived/runtime topology, networking, or launch details "data_parallel_rank", "data_parallel_rank_local", "data_parallel_size_local", "data_parallel_index", "data_parallel_backend", "data_parallel_external_lb", "data_parallel_hybrid_lb", "data_parallel_master_ip", "data_parallel_master_port", "_data_parallel_master_port_list", "data_parallel_rpc_port", "rank", "master_addr", "master_port", "node_rank", "nnodes", "max_parallel_loading_workers", "disable_custom_all_reduce", "ray_workers_use_nsight", "ray_runtime_env", "placement_group", "distributed_executor_backend", "worker_cls", "sd_worker_cls", "worker_extension_cls", "_api_process_count", "_api_process_rank", } from vllm.config.utils import get_hash_factors, hash_factors factors = get_hash_factors(self, ignored_factors) return hash_factors(factors) def __post_init__(self) -> None: # Continue with the rest of the initialization self.world_size = ( self.pipeline_parallel_size * self.tensor_parallel_size * self.prefill_context_parallel_size ) if self.distributed_executor_backend == "external_launcher": logger.info("Using external launcher for distributed inference.") self.world_size *= self.data_parallel_size if self.enable_elastic_ep: if not self.enable_eplb: raise ValueError("Elastic EP is only supported with enable_eplb=True.") if self.pipeline_parallel_size > 1: raise ValueError( "Elastic EP is not supported with pipeline parallelism " f"(pipeline_parallel_size={self.pipeline_parallel_size})." ) if self.data_parallel_external_lb or self.data_parallel_hybrid_lb: raise NotImplementedError( "Elastic EP is not compatible with data_parallel_external_lb " "or data_parallel_hybrid_lb. Elastic EP relies on a single API " "server and core client to coordinate scale up/down." ) if self.data_parallel_size > 1 or self.data_parallel_size_local == 0: # Data parallel was specified in the engine args. if self.distributed_executor_backend == "external_launcher": # For external launcher, # we need to set the data parallel rank automatically self.data_parallel_rank = int(os.environ["RANK"]) // ( self.world_size // self.data_parallel_size ) logger.info( "Set data_parallel_rank to %d automatically.", self.data_parallel_rank, ) if not self.enable_elastic_ep: if not self._data_parallel_master_port_list: self._data_parallel_master_port_list = get_open_ports_list(5) self.data_parallel_master_port = ( self._data_parallel_master_port_list.pop() ) if not (0 <= self.data_parallel_rank < self.data_parallel_size): raise ValueError( f"data_parallel_rank ({self.data_parallel_rank})" f" must be in the range [0, {self.data_parallel_size})" ) else: # Otherwise fall back to env vars (e.g. for offline SPMD case). self.data_parallel_size = envs.VLLM_DP_SIZE self.data_parallel_rank = envs.VLLM_DP_RANK self.data_parallel_rank_local = envs.VLLM_DP_RANK_LOCAL self.data_parallel_master_ip = envs.VLLM_DP_MASTER_IP self.data_parallel_master_port = envs.VLLM_DP_MASTER_PORT if self.data_parallel_size > 1 and self.is_moe_model is False: raise ValueError( "Offline data parallel mode is not supported/useful" " for dense models." ) self.data_parallel_index = self.data_parallel_rank if self.distributed_executor_backend == "external_launcher": os.environ["VLLM_ENABLE_V1_MULTIPROCESSING"] = "0" logger.info("Disabling V1 multiprocessing for external launcher.") if self.distributed_executor_backend is None and self.world_size_across_dp > 1: # We use multiprocessing by default if world_size fits on the # current node and we aren't in a ray placement group. from vllm.v1.executor import ray_utils backend: DistributedExecutorBackend = "mp" ray_found = ray_utils.ray_is_available() if current_platform.is_tpu() and envs.VLLM_XLA_USE_SPMD: backend = "uni" elif current_platform.is_cuda() and self.nnodes > 1: backend = "mp" elif ( current_platform.is_cuda() and cuda_device_count_stateless() < self.world_size ): gpu_count = cuda_device_count_stateless() raise ValueError( f"World size ({self.world_size}) is larger than the number of " f"available GPUs ({gpu_count}) in this node. If this is " "intentional and you are using:\n" "- ray, set '--distributed-executor-backend ray'.\n" "- multiprocessing, set '--nnodes' appropriately." ) elif self.data_parallel_backend == "ray": logger.info( "Using ray distributed inference because " "data_parallel_backend is ray" ) backend = "ray" elif ray_found: if self.placement_group: backend = "ray" else: from ray import is_initialized as ray_is_initialized if ray_is_initialized(): from ray.util import get_current_placement_group if get_current_placement_group(): backend = "ray" self.distributed_executor_backend = backend logger.debug("Defaulting to use %s for distributed inference", backend) if self.distributed_executor_backend is None and self.world_size == 1: self.distributed_executor_backend = "uni" if self.max_parallel_loading_workers is not None: logger.warning( "max_parallel_loading_workers is currently " "not supported and will be ignored." ) allowed_backends = ("mp", "uni", "external_launcher") if ( self.distributed_executor_backend not in allowed_backends and self.nnodes > 1 ): raise ValueError( "nnodes > 1 can only be set when distributed executor " "backend is mp, uni or external_launcher." ) if ( self.all2all_backend in ("allgather_reducescatter", "naive") and self.eplb_config.use_async ): logger.warning( "Async EPLB causes hangs with the '%s' all2all backend. " "Forcing synchronous EPLB.", self.all2all_backend, ) self.eplb_config.use_async = False @property def use_ray(self) -> bool: return self.distributed_executor_backend == "ray" or ( isinstance(self.distributed_executor_backend, type) and getattr(self.distributed_executor_backend, "uses_ray", False) ) @model_validator(mode="after") def _verify_args(self) -> Self: # Lazy import to avoid circular import from vllm.v1.executor import Executor # Enable batch invariance settings if requested if vllm_is_batch_invariant(): self.disable_custom_all_reduce = True if ( self.distributed_executor_backend is not None and not isinstance(self.distributed_executor_backend, str) and not ( isinstance(self.distributed_executor_backend, type) and issubclass(self.distributed_executor_backend, Executor) ) ): raise ValueError( "Unrecognized distributed executor backend " f"{self.distributed_executor_backend}. Supported " "values are 'ray', 'mp' 'uni', 'external_launcher', " " custom Executor subclass or its import path." ) if self.use_ray: from vllm.v1.executor import ray_utils ray_utils.assert_ray_available() if not current_platform.use_custom_allreduce(): self.disable_custom_all_reduce = True logger.debug( "Disabled the custom all-reduce kernel because it is not " "supported on current platform." ) if self.nnodes > 1: self.disable_custom_all_reduce = True logger.debug( "Disabled the custom all-reduce since we are running on multi-node." ) if self.ray_workers_use_nsight and not self.use_ray: raise ValueError( "Unable to use nsight profiling unless workers run with Ray." ) return self
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/config/parallel.py", "license": "Apache License 2.0", "lines": 731, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/attention/backends/linear_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass import torch from vllm.config import VllmConfig from vllm.v1.attention.backend import ( AttentionBackend, AttentionCGSupport, AttentionMetadataBuilder, CommonAttentionMetadata, ) from vllm.v1.attention.backends.utils import ( mamba_get_block_table_tensor, split_decodes_and_prefills, ) from vllm.v1.kv_cache_interface import AttentionSpec, MambaSpec class LinearAttentionBackend(AttentionBackend): @staticmethod def get_name() -> str: return "LINEAR_ATTN" @staticmethod def get_builder_cls() -> type["LinearAttentionMetadataBuilder"]: return LinearAttentionMetadataBuilder @dataclass class LinearAttentionMetadata: num_prefills: int num_prefill_tokens: int num_decodes: int num_decode_tokens: int query_start_loc: torch.Tensor seq_lens: torch.Tensor state_indices_tensor: torch.Tensor # shape: [batch,] class LinearAttentionMetadataBuilder(AttentionMetadataBuilder[LinearAttentionMetadata]): reorder_batch_threshold: int = 1 _cudagraph_support = AttentionCGSupport.UNIFORM_SINGLE_TOKEN_DECODE def __init__( self, kv_cache_spec: AttentionSpec, layer_names: list[str], vllm_config: VllmConfig, device: torch.device, ): super().__init__(kv_cache_spec, layer_names, vllm_config, device) assert isinstance(kv_cache_spec, MambaSpec) def build( self, common_prefix_len: int, common_attn_metadata: CommonAttentionMetadata, fast_build: bool = False, ) -> LinearAttentionMetadata: query_start_loc = common_attn_metadata.query_start_loc seq_lens = common_attn_metadata.seq_lens state_indices_tensor = mamba_get_block_table_tensor( common_attn_metadata.block_table_tensor, common_attn_metadata.seq_lens, self.kv_cache_spec, self.vllm_config.cache_config.mamba_cache_mode, )[:, 0] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( split_decodes_and_prefills( common_attn_metadata, decode_threshold=self.reorder_batch_threshold ) ) attn_metadata = LinearAttentionMetadata( num_prefills=num_prefills, num_prefill_tokens=num_prefill_tokens, num_decodes=num_decodes, num_decode_tokens=num_decode_tokens, query_start_loc=query_start_loc, seq_lens=seq_lens, state_indices_tensor=state_indices_tensor, ) return attn_metadata
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/attention/backends/linear_attn.py", "license": "Apache License 2.0", "lines": 73, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/multimodal/test_registry.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Unit tests for MultiModalRegistry.supports_multimodal_inputs and Qwen2.5-VL visual component loading behavior. """ import pytest from vllm.multimodal import MULTIMODAL_REGISTRY from ..models.utils import build_model_context pytestmark = pytest.mark.cpu_test @pytest.mark.parametrize( "model_id,limit_mm_per_prompt,expected", [ ("Qwen/Qwen2-0.5B-Instruct", {}, False), ("Qwen/Qwen2.5-VL-3B-Instruct", {}, True), ("Qwen/Qwen2.5-VL-3B-Instruct", {"image": 0, "video": 0}, False), ("Qwen/Qwen2.5-VL-3B-Instruct", {"image": 0}, True), ], ) @pytest.mark.core_model def test_supports_multimodal_inputs(model_id, limit_mm_per_prompt, expected): """Test supports_multimodal_inputs returns correct boolean for various configs.""" ctx = build_model_context( model_id, limit_mm_per_prompt=limit_mm_per_prompt, ) assert MULTIMODAL_REGISTRY.supports_multimodal_inputs(ctx.model_config) is expected
{ "repo_id": "vllm-project/vllm", "file_path": "tests/multimodal/test_registry.py", "license": "Apache License 2.0", "lines": 28, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/config/compilation.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import enum from collections import Counter from collections.abc import Callable from dataclasses import field from pathlib import Path from typing import TYPE_CHECKING, Any, ClassVar, Literal from pydantic import Field, TypeAdapter, field_validator import vllm.envs as envs from vllm.compilation.passes.inductor_pass import CallableInductorPass, InductorPass from vllm.config.utils import ( Range, config, get_hash_factors, hash_factors, ) from vllm.logger import init_logger from vllm.platforms import current_platform from vllm.utils.import_utils import resolve_obj_by_qualname from vllm.utils.math_utils import round_up from vllm.utils.torch_utils import is_torch_equal_or_newer if TYPE_CHECKING: from vllm.config import VllmConfig else: VllmConfig = object logger = init_logger(__name__) class CompilationMode(enum.IntEnum): """The compilation approach used for torch.compile-based compilation of the model.""" NONE = 0 """No torch.compile compilation is applied, model runs in fully eager pytorch mode. The model runs as-is.""" STOCK_TORCH_COMPILE = 1 """The standard `torch.compile` compilation pipeline.""" DYNAMO_TRACE_ONCE = 2 """Single Dynamo trace through the model, avoiding recompilation.""" VLLM_COMPILE = 3 """Custom vLLM Inductor-based backend with caching, piecewise compilation, shape specialization, and custom passes.""" class CUDAGraphMode(enum.Enum): """Constants for the cudagraph mode in CompilationConfig. Meanwhile, the subset enum `NONE`, `PIECEWISE` and `FULL` are also treated as concrete runtime mode for cudagraph runtime dispatching. """ NONE = 0 PIECEWISE = 1 FULL = 2 FULL_DECODE_ONLY = (FULL, NONE) FULL_AND_PIECEWISE = (FULL, PIECEWISE) def decode_mode(self) -> "CUDAGraphMode": return CUDAGraphMode(self.value[0]) if self.separate_routine() else self def mixed_mode(self) -> "CUDAGraphMode": return CUDAGraphMode(self.value[1]) if self.separate_routine() else self def has_mode(self, mode: "CUDAGraphMode") -> bool: assert not mode.separate_routine() if self.separate_routine(): return mode.value in self.value return self == mode def requires_piecewise_compilation(self) -> bool: return self.has_mode(CUDAGraphMode.PIECEWISE) def max_cudagraph_mode(self) -> "CUDAGraphMode": return CUDAGraphMode(max(self.value)) if self.separate_routine() else self def has_full_cudagraphs(self) -> bool: return self.max_cudagraph_mode() == CUDAGraphMode.FULL def has_piecewise_cudagraphs(self) -> bool: return self.requires_piecewise_compilation() def separate_routine(self) -> bool: return isinstance(self.value, tuple) @classmethod def valid_runtime_modes(cls) -> frozenset["CUDAGraphMode"]: return frozenset({cls.NONE, cls.PIECEWISE, cls.FULL}) def is_valid_runtime_mode(self) -> bool: return self in CUDAGraphMode.valid_runtime_modes() def __str__(self) -> str: return self.name @config class PassConfig: """Configuration for custom Inductor passes. This is separate from general `CompilationConfig` so that inductor passes don't all have access to full configuration - that would create a cycle as the `PassManager` is set as a property of config. You must pass PassConfig to VLLMConfig constructor via the CompilationConfig constructor. VLLMConfig's post_init does further initialization. If used outside of the VLLMConfig, some fields may be left in an improper state. """ # New flags fuse_norm_quant: bool = Field(default=None) """Fuse the custom RMSNorm + quant ops.""" fuse_act_quant: bool = Field(default=None) """Fuse the custom SiluMul + quant ops.""" fuse_attn_quant: bool = Field(default=None) """Fuse the custom attention + quant ops.""" eliminate_noops: bool = Field(default=True) """Eliminate no-op ops.""" enable_sp: bool = Field(default=None) """Enable sequence parallelism. Requires TP>1. Automatically disabled if the model's hidden_size is too small for SP to be beneficial (threshold is device-capability dependent).""" fuse_gemm_comms: bool = Field(default=None) """Enable async TP.""" fuse_allreduce_rms: bool = Field(default=None) """Enable flashinfer allreduce fusion.""" enable_qk_norm_rope_fusion: bool = False """Enable fused Q/K RMSNorm + RoPE pass.""" # ROCm/AITER specific fusions fuse_act_padding: bool = Field(default=None) """Fuse the custom RMSNorm + padding ops.""" fuse_rope_kvcache: bool = Field(default=None) """Fuse the QK rope + KV cache ops.""" rope_kvcache_fusion_max_token_num: int = 256 """The threshold for ROCm AITER RoPE+KVCache fusion e.g. for small batch decode. Larger batch sizes e.g. during prefill will use the unfused kernels. """ fi_allreduce_fusion_max_size_mb: float | None = None """The threshold of the communicated tensor sizes under which vllm should use flashinfer fused allreduce. Specified as a float in MB. Unspecified will fallback to default values which are compute capability and world size dependent. FI_ALLREDUCE_FUSION_MAX_SIZE_MB = { 90: { 2: 64, # 64MB 4: 2, # 2MB 8: 1, # 1MB }, 100: { 2: 64, # 64MB 4: 32, # 32MB 8: 1, # 1MB }, }, where key is the device capability""" sp_min_token_num: int | None = None """The minimum number of tokens above which vllm should use sequence parallelism. Specified as an integer token count. Unspecified will fallback to default values which are compute capability and world size dependent.""" # TODO(luka) better pass enabling system. def flashinfer_max_size(self, world_size: int) -> int | None: """ Returns the max communication size in bytes for flashinfer allreduce fusion for the given world size. Returns None if world size is not supported by configs as it's not supported by flashinfer. """ MiB = 1024 * 1024 FI_SUPPORTED_WORLD_SIZES = [2, 4, 8] if world_size not in FI_SUPPORTED_WORLD_SIZES: return None max_size_mb = self.fi_allreduce_fusion_max_size_mb if max_size_mb is None: max_size_mb = self.default_fi_allreduce_fusion_max_size_mb().get(world_size) return int(max_size_mb * MiB) if max_size_mb is not None else None @staticmethod def default_fi_allreduce_fusion_max_size_mb() -> dict[int, float]: from vllm.compilation.passes.fusion.allreduce_rms_fusion import ( FI_ALLREDUCE_FUSION_MAX_SIZE_MB, ) from vllm.platforms import current_platform if not current_platform.is_cuda(): return {} return FI_ALLREDUCE_FUSION_MAX_SIZE_MB.get( current_platform.get_device_capability().to_int(), {} ) def compute_hash(self) -> str: """ Produces a hash unique to the pass configuration. Any new fields that affect compilation should be added to the hash. Any future fields that don't affect compilation should be excluded. """ return hash_factors(get_hash_factors(self, set())) @field_validator( "fuse_norm_quant", "fuse_act_quant", "fuse_attn_quant", "enable_sp", "fuse_gemm_comms", "fuse_allreduce_rms", "fuse_act_padding", "fuse_rope_kvcache", mode="wrap", ) @classmethod def _skip_none_validation(cls, value: Any, handler: Callable) -> Any: """Skip validation if the value is `None` when initialisation is delayed.""" if value is None: return value return handler(value) def __post_init__(self) -> None: # Handle deprecation and defaults if not self.eliminate_noops: if self.fuse_norm_quant or self.fuse_act_quant: logger.warning_once( "Fusion enabled but reshape elimination disabled. " "RMSNorm/SiluMul + quant (fp8) fusion might not work" ) if self.fuse_attn_quant: logger.warning_once( "Fusion enabled but reshape elimination disabled. " "Attention + quant (fp8) fusion might not work" ) if self.fuse_allreduce_rms: logger.warning_once( "Fusion enabled but reshape elimination disabled. " "Allreduce + rms norm + quant (fp8) fusion might not work" ) if self.fuse_act_padding: logger.warning_once( "Fusion enabled but reshape elimination disabled. " "RMSNorm + padding fusion might not work" ) if self.enable_qk_norm_rope_fusion and not current_platform.is_cuda_alike(): logger.warning_once( "QK Norm + RoPE fusion enabled but the current platform is not " "CUDA or ROCm. The fusion will be disabled." ) self.enable_qk_norm_rope_fusion = False if self.fuse_act_padding and not current_platform.is_rocm(): logger.warning_once( "Padding fusion enabled but the current platform is not ROCm. " "The fusion will be disabled." ) self.fuse_act_padding = False if self.fuse_rope_kvcache and not current_platform.is_rocm(): logger.warning_once( "KV cache fusion currently only enabled on ROCm. " "The fusion will be disabled." ) self.fuse_rope_kvcache = False class DynamicShapesType(str, enum.Enum): """Types of dynamic shapes handling in torch.compile(). see Dynamic shapes and vllm guard dropping in torch_compile.md for more details.""" BACKED = "backed" """Use backed dynamic shapes. torch.compile() guards on backed dynamic shapes and may add guards. Symbols are specialized to 0, 1, or >=2 even without encountering branching on those ranges.""" UNBACKED = "unbacked" """Use unbacked dynamic shapes. Guaranteed not to be guarded on and not 0/1 specialized, but may throw data dependent errors when branches require their value without explicit unbacked handling.""" BACKED_SIZE_OBLIVIOUS = "backed_size_oblivious" """Experimental flag that treats backed symbols as unbacked when explicit unbacked handling is defined.""" @config class DynamicShapesConfig: """Configuration to control/debug torch compile dynamic shapes.""" type: DynamicShapesType = DynamicShapesType.BACKED """Controls the type of dynamic shapes handling to use with torch.compile(). - BACKED: Default PyTorch behavior with potential guards ignored. - UNBACKED: No guards guaranteed (most sound) but may throw data dependent errors. - BACKED_SIZE_OBLIVIOUS: Experimental safer alternative to backed/unbacked. """ evaluate_guards: bool = False """ A debug mode to detect and fail if Dynamo ever specializes a dynamic shape by guarding on it. When True, dynamic shape guards are not dropped from dynamo. And a failure will be triggered if a recompilation ever happens due to that. This mode requires VLLM_USE_BYTECODE_HOOK to be 0. Enabling this allow observing the dynamic shapes guards in the tlparse artifacts also. When type is backed, aot_compile must be disabled for this mode to work. until this change picked up https://github.com/pytorch/pytorch/pull/169239. """ assume_32_bit_indexing: bool = False """ whether all tensor sizes can use 32 bit indexing. `True` requires PyTorch 2.10+ """ def compute_hash(self) -> str: """ Provide a hash for DynamicShapesConfig """ from vllm.config.utils import get_hash_factors, hash_factors factors = get_hash_factors(self, {}) return hash_factors(factors) @config class CompilationConfig: """Configuration for compilation. You must pass CompilationConfig to VLLMConfig constructor. VLLMConfig's post_init does further initialization. If used outside of the VLLMConfig, some fields will be left in an improper state. It has three parts: - Top-level Compilation control: - [`mode`][vllm.config.CompilationConfig.mode] - [`debug_dump_path`][vllm.config.CompilationConfig.debug_dump_path] - [`cache_dir`][vllm.config.CompilationConfig.cache_dir] - [`backend`][vllm.config.CompilationConfig.backend] - [`custom_ops`][vllm.config.CompilationConfig.custom_ops] - [`splitting_ops`][vllm.config.CompilationConfig.splitting_ops] - [`compile_mm_encoder`][vllm.config.CompilationConfig.compile_mm_encoder] - CudaGraph capture: - [`cudagraph_mode`][vllm.config.CompilationConfig.cudagraph_mode] - [`cudagraph_capture_sizes`] [vllm.config.CompilationConfig.cudagraph_capture_sizes] - [`max_cudagraph_capture_size`] [vllm.config.CompilationConfig.max_cudagraph_capture_size] - [`cudagraph_num_of_warmups`] [vllm.config.CompilationConfig.cudagraph_num_of_warmups] - [`cudagraph_copy_inputs`] [vllm.config.CompilationConfig.cudagraph_copy_inputs] - Inductor compilation: - [`compile_sizes`][vllm.config.CompilationConfig.compile_sizes] - [`compile_ranges_split_points`] [vllm.config.CompilationConfig.compile_ranges_split_points] - [`inductor_compile_config`] [vllm.config.CompilationConfig.inductor_compile_config] - [`inductor_passes`][vllm.config.CompilationConfig.inductor_passes] - custom inductor passes Why we have different sizes for cudagraph and inductor: - cudagraph: a cudagraph captured for a specific size can only be used for the same size. We need to capture all the sizes we want to use. - inductor: a graph compiled by inductor for a general shape can be used for different sizes. Inductor can also compile for specific sizes, where it can have more information to optimize the graph with fully static shapes. However, we find the general shape compilation is sufficient for most cases. It might be beneficial to compile for certain small batchsizes, where inductor is good at optimizing. """ # Top-level Compilation control level: int = Field(default=None) """ Level is deprecated and will be removed in the next release, either 0.12.0 or 0.11.2 whichever is soonest. Please use mode. Currently all levels are mapped to mode. """ # Top-level Compilation control mode: CompilationMode = Field(default=None) """The compilation approach used for torch.compile-based compilation of the model. - None: If None, we will select the default compilation mode. For V1 engine this is 3. - 0: NONE: No torch.compile compilation is applied, model runs in fully eager pytorch mode. The model runs as-is. - 1: STOCK_TORCH_COMPILE: The standard `torch.compile` compilation pipeline. - 2: DYNAMO_TRACE_ONCE: Single Dynamo trace through the model, avoiding recompilation by removing guards. Requires no dynamic-shape-dependent control-flow. - 3: VLLM_COMPILE: Custom vLLM Inductor-based backend with caching, piecewise compilation, shape specialization, and custom passes.""" debug_dump_path: Path | None = None """The path to dump the debug information.""" cache_dir: str = "" """The directory to store the compiled graph, to accelerate Inductor compilation. By default, it will use model-related information to generate a cache directory.""" compile_cache_save_format: Literal["binary", "unpacked"] = field( default_factory=lambda: envs.VLLM_COMPILE_CACHE_SAVE_FORMAT ) """Format for saving torch compile cache:\n - "binary": saves as binary file (multiprocess safe)\n - "unpacked": saves as directory structure for inspection/debugging (NOT multiprocess safe)\n Defaults to `VLLM_COMPILE_CACHE_SAVE_FORMAT` if not specified. """ backend: str = "" """The backend for compilation. It needs to be a string: - "" (empty string): use the default backend ("inductor" on CUDA-alike platforms). - "eager"/"openxla"/...: use the specified backend registered in PyTorch. - "full.module.name": a qualified name which can be used to import the backend function. We use string to avoid serialization issues when using compilation in a distributed setting. When the compilation mode is 1 or 2, the backend is used for the compilation directly (it sees the whole graph). When the compilation mode is 3, the backend supports both whole graph and piecewise compilation, available backends include eager, inductor, and custom backends, the latter of which can be defined via `get_compile_backend`. Furthermore, compilation is only piecewise if splitting ops is set accordingly and use_inductor_graph_partition is off. Note that the default options for splitting ops are sufficient for piecewise compilation. """ custom_ops: list[str] = field(default_factory=list) """Fine-grained control over which custom ops to enable/disable. Use 'all' to enable all, 'none' to disable all. Also specify a list of custom op names to enable (prefixed with a '+'), or disable (prefixed with a '-'). Examples: - 'all,-op1' to enable all except op1 - 'none,+op1,+op2' to enable only op1 and op2 By default, all custom ops are enabled when running without Inductor and disabled when running with Inductor: mode>CompilationMode.NONE and backend="inductor". Inductor generates (fused) Triton kernels for disabled custom ops.""" splitting_ops: list[str] | None = None """A list of ops to exclude from cudagraphs, used in piecewise compilation. The behavior depends on use_inductor_graph_partition: - When use_inductor_graph_partition=False (default): These ops are used for Dynamo FX-level graph splitting. The graph is split at these ops before Inductor compilation, creating separate subgraphs for cudagraph capture. - When use_inductor_graph_partition=True: These ops are used to register Inductor partition rules. The graph partitioning happens at Inductor codegen time after all passes and fusions are finished, allowing compilation and custom passes to operate on the full graph while still excluding these ops from cudagraphs. If None, defaults to attention ops for piecewise cudagraphs. If empty list [], no ops are excluded (suitable for full cudagraphs).""" compile_mm_encoder: bool = False """Whether or not to compile the multimodal encoder. Currently, this only works for `Qwen2_5_vl` and `mLLaMa4` models on selected platforms. Disabled by default until more models are supported/tested to work.""" # Inductor capture compile_sizes: list[int | str] | None = None """Sizes to compile for inductor. In addition to integers, it also supports "cudagraph_capture_sizes" to specify the sizes for cudagraph capture.""" compile_ranges_split_points: list[int] | None = None """Split points that represent compile ranges for inductor. The compile ranges are [1, split_points[0]], [split_points[0] + 1, split_points[1]], ..., [split_points[-1] + 1, max_num_batched_tokens]. Compile sizes are also used single element ranges, the range is represented as [compile_sizes[i], compile_sizes[i]]. If a range overlaps with the compile size, graph for compile size will be prioritized, i.e. if we have a range [1, 8] and a compile size 4, graph for compile size 4 will be compiled and used instead of the graph for range [1, 8]. """ inductor_compile_config: dict = field(default_factory=dict) """Additional configurations for inductor. - None: use default configurations.""" inductor_passes: dict[str, str] = field(default_factory=dict) """Additional passes for inductor. It is a dictionary from pass name to pass function qualified name. We use function name because the config uses JSON format. If we pass the config from Python, functions can also be passed directly via Python object constructor, e.g. `CompilationConfig(inductor_passes={"a": func})`.""" # CudaGraph compilation cudagraph_mode: CUDAGraphMode = Field(default=None) """ The mode of the cudagraph: - NONE, no cudagraph capture. - PIECEWISE. - FULL. - FULL_DECODE_ONLY. - FULL_AND_PIECEWISE. (v1 default) PIECEWISE mode build piecewise cudagraph only, keeping the cudagraph incompatible ops (i.e. some attention ops) outside the cudagraph for general flexibility. FULL mode: Capture full cudagraph for all batches. Can be good for small models or workloads with small prompts; not supported by many backends. Generally for performance FULL_AND_PIECEWISE is better. FULL_DECODE_ONLY mode: Capture full cudagraph for decode batches only. Mixed prefill-decode batches are run without cudagraphs. Can be good for decode instances in a P/D setup where prefill is not as important so we can save some memory. FULL_AND_PIECEWISE mode: Capture full cudagraph for decode batches and piecewise cudagraph for prefill and mixed prefill-decode batches. This is the most performant mode for most models and is the default. Currently, the cudagraph mode is only used for the v1 engine. Note that the cudagraph logic is generally orthogonal to the compilation logic. While piecewise cudagraphs require piecewise compilation (mode=VLLM_COMPILE and non-empty splitting_ops), full cudagraphs are supported with and without compilation. Warning: This flag is new and subject to change in addition more modes may be added. """ cudagraph_num_of_warmups: int = 0 """Number of warmup runs for cudagraph. It means the first several runs will be treated as warmup runs. Only after that, the execution will be recorded, and the recorded cudagraph will be used for subsequent runs.""" cudagraph_capture_sizes: list[int] | None = None """Sizes to capture cudagraph. - None (default): capture sizes are inferred from vllm config. - list[int]: capture sizes are specified as given.""" cudagraph_copy_inputs: bool = False """Whether to copy input tensors for cudagraph. If the caller can guarantee that the same input buffers are always used, it can set this to False. Otherwise, it should set this to True, and the compiler will copy the input to an internally managed buffer. Default is False. Note that this flag is only effective when cudagraph_mode is PIECEWISE. """ cudagraph_specialize_lora: bool = True """Whether to create separate cuda graphs for cases with and without active LoRA adapters. When set to False, the LoRA-enabled cuda graph will be used for all cases, incurring the overhead of running LoRA ops even when no adapters are active. Setting this to True will remove this overhead at the cost of increased startup time and slightly higher memory usage. When `enable_lora` is False, this option has no effect. """ use_inductor_graph_partition: bool = Field(default=None) """Use inductor graph partition to split the graph at cudagraph_unsafe ops. This partition happens at inductor codegen time after all passes and fusions are finished. It generates a single `call` function which wraps cudagraph-safe ops into partition functions and leave cudagraph-unsafe ops outside the partition functions. For a graph with N cudagraph-unsafe ops (e.g., Attention), there would be N+1 partitions. To mark an op as cudagraph unsafe, we can add `tags=(torch._C.Tag.cudagraph_unsafe)` when register the custom op. This config supports both full cudagraph and piecewise cudagraph without compiling twice. For piecewise cudagraph, it applies vLLM CUDAGraph wrapper to each partition. For N+1 partitions, there would be N+1 CUDAGraph wrapper instances. For full CUDAGraph, we always apply a single CUDAGraph wrapper outside the inductor `call` function in the model runner. The top-level full cudagraph capture ignores all partitioning. """ pass_config: PassConfig = field(default_factory=PassConfig) """Custom inductor passes, see PassConfig for more details""" max_cudagraph_capture_size: int = field(default=None) """The maximum cudagraph capture size. If cudagraph_capture_sizes is specified, this will be set to the largest size in that list (or checked for consistency if specified). If cudagraph_capture_sizes is not specified, the list of sizes is generated automatically following the pattern: [1, 2, 4] + list(range(8, 256, 8)) + list( range(256, max_cudagraph_capture_size + 1, 16)) If not specified, max_cudagraph_capture_size is set to min(max_num_seqs*2, 512) by default. This voids OOM in tight memory scenarios with small max_num_seqs, and prevents capture of many large graphs (>512) that would greatly increase startup time with limited performance benefit. """ dynamic_shapes_config: DynamicShapesConfig = field( default_factory=DynamicShapesConfig ) """Configuration for dynamic shapes options""" local_cache_dir: str = field(default=None, init=False) # type: ignore """local cache dir for each rank""" fast_moe_cold_start: bool | None = None """Optimization for fast MOE cold start. This is a bit of a hack that assumes that: 1. the only decoder forward pass being run is the current model 2. the decoder forward pass runs all of the MOEs in the order in which they are initialized When the above two conditions hold, this option greatly decreases cold start time for MOE models. The options are: - True: optimization is always on - False: optimization is always off - None: optimization is on usually but off for speculative decoding If conditions 1&2 don't hold then this option will lead to silent incorrectness. The only condition in which this doesn't hold is speculative decoding, where there is a draft model that may have MOEs in them. NB: We're working on a longer-term solution that doesn't need these assumptions. """ # keep track of enabled and disabled custom ops enabled_custom_ops: Counter[str] = field(default_factory=Counter, init=False) """custom ops that are enabled""" disabled_custom_ops: Counter[str] = field(default_factory=Counter, init=False) """custom ops that are disabled""" traced_files: set[str] = field(default_factory=set, init=False) """files that are traced for compilation""" compilation_time: float = field(default=0.0, init=False) """time taken for compilation""" static_forward_context: dict[str, Any] = field(default_factory=dict, init=False) """Per-model forward context Map from layer name to layer objects that need to be accessed outside model code, e.g., Attention, FusedMOE when dp_size>1.""" static_all_moe_layers: list[str] = field(default_factory=list, init=False) """The names of all the MOE layers in the model """ # Attention ops; used for piecewise cudagraphs # Use PyTorch operator format: "namespace::name" _attention_ops: ClassVar[list[str]] = [ "vllm::unified_attention", "vllm::unified_attention_with_output", "vllm::unified_mla_attention", "vllm::unified_mla_attention_with_output", "vllm::mamba_mixer2", "vllm::mamba_mixer", "vllm::short_conv", "vllm::linear_attention", "vllm::plamo2_mamba_mixer", "vllm::gdn_attention_core", "vllm::kda_attention", "vllm::sparse_attn_indexer", "vllm::rocm_aiter_sparse_attn_indexer", ] def compute_hash(self) -> str: """ Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # Opt-out: default-include declared fields; keep a tiny exclude set; # normalize types; keep SHA-256. For nested opaque configs, include a # stable identifier (e.g., pass_config.compute_hash()) instead of object id. ignored_factors = { # Paths/dirs and runtime/metrics that don’t affect compiled graph "debug_dump_path", "cache_dir", "local_cache_dir", "traced_files", "compilation_time", "static_forward_context", "pass_config", # handled separately below "dynamic_shapes_config", # handled separately below } from vllm.config.utils import get_hash_factors, hash_factors factors = get_hash_factors(self, ignored_factors) factors["pass_config"] = self.pass_config.compute_hash() factors["dynamic_shapes_config"] = self.dynamic_shapes_config.compute_hash() return hash_factors(factors) def __repr__(self) -> str: exclude = { "static_forward_context": True, "enabled_custom_ops": True, "disabled_custom_ops": True, "compilation_time": True, "traced_files": True, "inductor_compile_config": { "post_grad_custom_post_pass": True, }, } # exclude default attr in pass_config pass_config_exclude = {} for attr, default_val in vars(PassConfig()).items(): if getattr(self.pass_config, attr) == default_val: pass_config_exclude[attr] = True if pass_config_exclude: exclude["pass_config"] = pass_config_exclude config = TypeAdapter(CompilationConfig).dump_python( self, exclude=exclude, exclude_unset=True ) return str(config) __str__ = __repr__ @field_validator("mode", mode="before") @classmethod def validate_mode_before(cls, value: Any) -> Any: """ Enable parsing the `mode` field from string mode names. Accepts both integers (0-3) and string names, like NONE, STOCK_TORCH_COMPILE, DYNAMO_TRACE_ONCE, VLLM_COMPILE. """ if isinstance(value, str): # Convert string mode name to integer value mode_name = value.upper() if mode_name not in CompilationMode.__members__: raise ValueError( f"Invalid compilation mode: {value}. " f"Valid modes are: {', '.join(CompilationMode.__members__.keys())}" ) return CompilationMode[mode_name] return value @field_validator("cudagraph_mode", mode="before") @classmethod def validate_cudagraph_mode_before(cls, value: Any) -> Any: """Enable parsing of the `cudagraph_mode` enum type from string.""" if isinstance(value, str): return CUDAGraphMode[value.upper()] return value @field_validator("pass_config", mode="before") @classmethod def validate_pass_config_before(cls, value: Any) -> Any: """Enable parsing of the `pass_config` field from a dictionary.""" if isinstance(value, dict): return PassConfig(**value) return value @field_validator("compile_cache_save_format") @classmethod def validate_compile_cache_save_format(cls, value: str) -> str: if value not in ("binary", "unpacked"): raise ValueError( f"compile_cache_save_format must be 'binary' or 'unpacked', " f"got: {value}" ) return value @field_validator( "level", "mode", "cudagraph_mode", "max_cudagraph_capture_size", "use_inductor_graph_partition", mode="wrap", ) @classmethod def _skip_none_validation(cls, value: Any, handler: Callable) -> Any: """Skip validation if the value is `None` when initialisation is delayed.""" if value is None: return value return handler(value) def __post_init__(self) -> None: if self.level is not None: logger.warning( "Level is deprecated and will be removed in the next release," "either 0.12.0 or 0.11.2 whichever is soonest." "Use mode instead." "If both level and mode are given," "only mode will be used." ) if self.mode is None: self.mode = self.level count_none = self.custom_ops.count("none") count_all = self.custom_ops.count("all") assert count_none + count_all <= 1, "Can only specify 'none' or 'all'" # TODO(zou3519/luka): There are 2 issues with auto-functionalization V2: # 1. A bug in PyTorch, fixed in 2.7: # https://github.com/pytorch/pytorch/issues/147924 # 2. Custom passes (fusion) rely on auto-functionalization V1 and don't # work with V2. Addressing this will take extra engineering effort # and it is not yet a priority. RFC here: # https://github.com/vllm-project/vllm/issues/14703 KEY = "enable_auto_functionalized_v2" if KEY not in self.inductor_compile_config: self.inductor_compile_config[KEY] = False for k, v in self.inductor_passes.items(): if not isinstance(v, str): assert callable(v), f"pass {k} should be callable or a qualified name" self.inductor_compile_config[k] = ( v if isinstance(v, InductorPass) else CallableInductorPass(v) ) continue # resolve function from qualified name names = v.split(".") module = ".".join(names[:-1]) func_name = names[-1] func = __import__(module).__dict__[func_name] self.inductor_compile_config[k] = ( func if isinstance(func, InductorPass) else CallableInductorPass(func) ) if ( self.pass_config.enable_qk_norm_rope_fusion and "+rotary_embedding" not in self.custom_ops ): # TODO(zhuhaoran): support rope native forward match and remove this. # Linked issue: https://github.com/vllm-project/vllm/issues/28042 self.custom_ops.append("+rotary_embedding") if ( self.pass_config.fuse_rope_kvcache and "+rotary_embedding" not in self.custom_ops ): # TODO(Rohan138): support rope native forward match and remove this. # Linked issue: https://github.com/vllm-project/vllm/issues/28042 self.custom_ops.append("+rotary_embedding") if ( is_torch_equal_or_newer("2.9.0.dev") and "combo_kernels" not in self.inductor_compile_config and "benchmark_combo_kernel" not in self.inductor_compile_config # (fixme @boyuan) combo kernel does not support cpu yet. and not current_platform.is_cpu() ): # use horizontal fusion, which is useful for fusing qk-norm and # qk-rope when query and key have different shapes. self.inductor_compile_config["combo_kernels"] = True self.inductor_compile_config["benchmark_combo_kernel"] = True if self.use_inductor_graph_partition and not is_torch_equal_or_newer( "2.9.0.dev" ): raise ValueError( "use_inductor_graph_partition is only " "supported with torch>=2.9.0.dev. Set " "use_inductor_graph_partition=False instead." ) for op in self.custom_ops: if op[0] not in {"+", "-"} and op not in {"all", "none"}: raise ValueError( f"Invalid syntax '{op}' for custom op, " "must be 'all', 'none', '+op' or '-op' " "(where 'op' is the registered op name)" ) # Currently only eager and inductor backend are supported. # for piecewise compilation. Custom backends are not suppported for # piecewise compilation. Update when more backends are supported. if self.mode == CompilationMode.VLLM_COMPILE and self.backend not in [ "", "eager", "inductor", ]: raise ValueError( f"Invalid backend for piecewise compilation: {self.backend}" ) if self.backend == "": self.backend = current_platform.get_compile_backend() def init_backend(self, vllm_config: "VllmConfig") -> str | Callable: """ Initialize the backend for the compilation config from a vllm config. Arguments: vllm_config: The vllm config to initialize the backend from. Returns: The backend for the compilation config. """ if self.mode is None: raise ValueError( "No compilation mode is set. This method should only be " "called via vllm config where the level is set if none is " "provided." ) if self.mode == CompilationMode.NONE: raise ValueError("No compilation mode is set.") from torch._dynamo.backends.registry import list_backends torch_backends = list_backends(exclude_tags=tuple()) if self.mode in [ CompilationMode.STOCK_TORCH_COMPILE, CompilationMode.DYNAMO_TRACE_ONCE, ]: if self.backend in torch_backends: return self.backend return resolve_obj_by_qualname(self.backend) assert self.mode == CompilationMode.VLLM_COMPILE if self.backend not in ["eager", "inductor"]: logger.info("Using OOT custom backend for compilation.") from vllm.compilation.backends import VllmBackend # TODO[@lucaskabela]: See if we can forward prefix # https://github.com/vllm-project/vllm/issues/27045 return VllmBackend(vllm_config) def post_init_cudagraph_sizes(self) -> None: """To complete the initialization after cudagraph related configs are set. This includes: - initialize compile_sizes """ computed_compile_sizes = [] if self.compile_sizes is not None: # de-duplicate the sizes provided by the config self.compile_sizes = list(set(self.compile_sizes)) for x in self.compile_sizes: if isinstance(x, str): assert x == "cudagraph_capture_sizes", ( "Unrecognized size type in compile_sizes, " f"expect 'cudagraph_capture_sizes', got {x}" ) computed_compile_sizes.extend(self.cudagraph_capture_sizes) else: assert isinstance(x, int) computed_compile_sizes.append(x) self.compile_sizes = computed_compile_sizes # type: ignore # make sure the sizes are in ascending order self.cudagraph_capture_sizes.sort() if self.cudagraph_capture_sizes: assert self.cudagraph_capture_sizes[-1] == self.max_cudagraph_capture_size def set_splitting_ops_for_v1( self, all2all_backend: str, data_parallel_size: int = 1 ): # To compatible with OOT hardware plugin platform (for example vllm-ascend) # which currently only supports sequence parallelism in eager mode. if self.mode != CompilationMode.VLLM_COMPILE: if self.splitting_ops is None: self.splitting_ops = [] return # NOTE: this function needs to be called only when mode is # CompilationMode.VLLM_COMPILE assert self.mode == CompilationMode.VLLM_COMPILE, ( "set_splitting_ops_for_v1 should only be called when " "mode is CompilationMode.VLLM_COMPILE" ) if self.pass_config.fuse_attn_quant and not self.use_inductor_graph_partition: self.set_splitting_ops_for_attn_fusion() else: if self.splitting_ops is None: # NOTE: When using full cudagraph, instead of setting an empty # list and capture the full cudagraph inside the flattened fx # graph, we keep the piecewise fx graph structure but capture # the full cudagraph outside the fx graph. This reduces some # cpu overhead when the runtime batch_size is not cudagraph # captured. see https://github.com/vllm-project/vllm/pull/20059 # for details. Make a copy to avoid mutating the class-level # list via reference. self.splitting_ops = list(self._attention_ops) # unified_kv_cache_update has a string param that prevents Inductor # from reusing piecewise graphs. Remove it from the compiled graph. # This has the side-effect of excluding cache from cudagraphs but # that doesn't seem to affect performance. # https://github.com/vllm-project/vllm/issues/33267 if not self.use_inductor_graph_partition: self.splitting_ops.append("vllm::unified_kv_cache_update") elif len(self.splitting_ops) == 0: if ( self.cudagraph_mode == CUDAGraphMode.PIECEWISE or self.cudagraph_mode == CUDAGraphMode.FULL_AND_PIECEWISE ): logger.warning_once( "Using piecewise cudagraph with empty splitting_ops" ) if self.cudagraph_mode == CUDAGraphMode.PIECEWISE: logger.warning_once( "Piecewise compilation with empty splitting_ops does not " "contain piecewise cudagraph. Setting cudagraph_" "mode to NONE. Hint: If you are using attention " "backends that support cudagraph, consider manually " "setting cudagraph_mode to FULL or FULL_DECODE_ONLY " "to enable full cudagraphs." ) self.cudagraph_mode = CUDAGraphMode.NONE elif self.cudagraph_mode == CUDAGraphMode.FULL_AND_PIECEWISE: logger.warning_once( "Piecewise compilation with empty splitting_ops does " "not contain piecewise cudagraph. Setting " "cudagraph_mode to FULL." ) self.cudagraph_mode = CUDAGraphMode.FULL self.splitting_ops = [] # Disable CUDA graphs for DeepEP high-throughput since its not CG compatible if ( all2all_backend == "deepep_high_throughput" and data_parallel_size > 1 and self.cudagraph_mode != CUDAGraphMode.NONE ): # TODO: Piecewise Cuda graph might be enabled # if torch compile cache key issue fixed # See https://github.com/vllm-project/vllm/pull/25093 logger.info( "DeepEP: Disabling CUDA Graphs since DeepEP high-throughput kernels " "are optimized for prefill and are incompatible with CUDA Graphs. " "In order to use CUDA Graphs for decode-optimized workloads, " "use --all2all-backend with another option, such as " "deepep_low_latency or allgather_reducescatter." ) self.cudagraph_mode = CUDAGraphMode.NONE def set_splitting_ops_for_attn_fusion(self): assert self.pass_config.fuse_attn_quant if self.splitting_ops is None: self.splitting_ops = [] if self.cudagraph_mode.has_piecewise_cudagraphs(): logger.warning_once( "fuse_attn_quant is incompatible with piecewise " "cudagraph when use_inductor_graph_partition is off. " "In this case, splitting_ops will be set to empty " "list, and cudagraph_mode will be set to FULL. " "Please ensure you are using attention backends that " "support cudagraph or set cudagraph_mode to NONE " "explicitly if encountering any problems." ) self.cudagraph_mode = CUDAGraphMode.FULL assert not self.splitting_ops_contain_attention(), ( "attention ops should not be in splitting_ops when fuse_attn_quant is True" ) def splitting_ops_contain_attention(self) -> bool: return self.splitting_ops is not None and all( op in self.splitting_ops for op in self._attention_ops ) def is_attention_compiled_piecewise(self) -> bool: if not self.splitting_ops_contain_attention(): return False if not self.use_inductor_graph_partition: # Dynamo-level FX split case return self.mode == CompilationMode.VLLM_COMPILE # Inductor partition case return self.backend == "inductor" and self.mode != CompilationMode.NONE def custom_op_log_check(self): """ This method logs the enabled/disabled custom ops and checks that the passed custom_ops field only contains relevant ops. It is called at the end of set_current_vllm_config, after the custom ops have been instantiated. """ if len(self.enabled_custom_ops) + len(self.disabled_custom_ops) == 0: logger.debug("No custom ops found in model.") return logger.debug("enabled custom ops: %s", self.enabled_custom_ops) logger.debug("disabled custom ops: %s", self.disabled_custom_ops) all_ops_in_model = self.enabled_custom_ops | self.disabled_custom_ops for op in self.custom_ops: if op in {"all", "none"}: continue assert op[0] in {"+", "-"}, ( "Invalid custom op syntax (should be checked during init)" ) # check if op name exists in model op_name = op[1:] if op_name not in all_ops_in_model: from vllm.model_executor.custom_op import op_registry # Does op exist at all or is it just not present in this model? # Note: Only imported op classes appear in the registry. missing_str = ( "doesn't exist (or wasn't imported/registered)" if op_name not in op_registry else "not present in model" ) enable_str = "enabling" if op[0] == "+" else "disabling" logger.warning_once( "Op '%s' %s, %s with '%s' has no effect", op_name, missing_str, enable_str, op, ) def is_custom_op_enabled(self, op: str) -> bool: if "all" in self.custom_ops: return f"-{op}" not in self.custom_ops assert "none" in self.custom_ops return f"+{op}" in self.custom_ops def adjust_cudagraph_sizes_for_spec_decode( self, uniform_decode_query_len: int, tensor_parallel_size: int ): multiple_of = uniform_decode_query_len if tensor_parallel_size > 1 and self.pass_config.enable_sp: multiple_of = max(uniform_decode_query_len, tensor_parallel_size) if ( multiple_of % uniform_decode_query_len != 0 or multiple_of % tensor_parallel_size != 0 ): raise ValueError( f"Can't determine cudagraph shapes that are both a " f"multiple of {uniform_decode_query_len} " f"(num_speculative_tokens + 1) required by spec-decode " f"and {tensor_parallel_size} (tensor_parallel_size) " f"required by sequence parallelism please adjust " f"num_speculative_tokens or disable sequence parallelism" ) if not self.cudagraph_capture_sizes or multiple_of <= 1: return assert self.max_cudagraph_capture_size is not None rounded_sizes = sorted( set( round_up(size, multiple_of) for size in self.cudagraph_capture_sizes if round_up(size, multiple_of) <= self.max_cudagraph_capture_size ) ) if len(rounded_sizes) == 0 and multiple_of <= self.max_cudagraph_capture_size: # if one valid but would be round_down use that rounded_sizes = [multiple_of] if len(rounded_sizes) == 0: raise ValueError( f"No valid cudagraph sizes after rounding to multiple of {multiple_of} " f"(num_speculative_tokens + 1 or tp if sequence parallelism is enabled)" f" please adjust num_speculative_tokens ({uniform_decode_query_len - 1}" f") or max_cudagraph_capture_size ({self.max_cudagraph_capture_size})" f" or cudagraph_capture_sizes ({self.cudagraph_capture_sizes})" ) self.max_cudagraph_capture_size = rounded_sizes[-1] self.cudagraph_capture_sizes = rounded_sizes def get_compile_ranges(self) -> list[Range]: """Get the compile ranges for the compilation config.""" if self.compile_ranges_split_points is None: return [] split_points = sorted(set(self.compile_ranges_split_points)) return [ Range(start=s + 1, end=e) for s, e in zip([0] + split_points[:-1], split_points) ]
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/config/compilation.py", "license": "Apache License 2.0", "lines": 1029, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/config/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Utility functions for vLLM config dataclasses.""" import ast import enum import hashlib import inspect import json import os import pathlib import textwrap from collections.abc import Callable, Mapping, Sequence, Set from dataclasses import MISSING, field, fields, is_dataclass from itertools import pairwise from typing import TYPE_CHECKING, Any, Protocol, TypeVar, cast import torch from pydantic import ConfigDict from pydantic.dataclasses import dataclass from pydantic.fields import Field as PydanticField from pydantic.fields import FieldInfo from typing_extensions import dataclass_transform, runtime_checkable import vllm.envs as envs from vllm.logger import init_logger logger = init_logger(__name__) if TYPE_CHECKING: from _typeshed import DataclassInstance else: DataclassInstance = Any ConfigType = type[DataclassInstance] ConfigT = TypeVar("ConfigT", bound=DataclassInstance) @dataclass_transform(field_specifiers=(PydanticField,)) def config( cls: type[ConfigT] | None = None, *, config: ConfigDict | None = None, **kwargs: Any, ) -> type[ConfigT] | Callable[[type[ConfigT]], type[ConfigT]]: """Decorator to create a pydantic dataclass with default config. The default config for the dataclass forbids extra fields. All config classes in vLLM should use this decorator. Args: cls: The class to decorate config: The pydantic ConfigDict to use. If provided, it will be merged with the default config. **kwargs: Additional arguments to pass to pydantic.dataclass.""" # Extra fields are forbidden by default merged_config = ConfigDict(extra="forbid") if config is not None: merged_config.update(config) def decorator(cls): return dataclass(cls, config=merged_config, **kwargs) # Called with arguments: @config(config=...) if cls is None: return decorator # Called without arguments: @config return decorator(cls) def get_field(cls: ConfigType, name: str) -> Any: """Get the default factory field of a dataclass by name. Used for getting default factory fields in `EngineArgs`.""" if not is_dataclass(cls): raise TypeError("The given class is not a dataclass.") try: named_field = next(f for f in fields(cls) if f.name == name) except StopIteration as e: raise ValueError(f"Field '{name}' not found in {cls.__name__}.") from e # The arguments to copy to the new field default = named_field.default default_factory = named_field.default_factory init = named_field.init # Handle pydantic.Field if isinstance(default, FieldInfo): if default.init is not None: init = default.init if default.default_factory is not None: default_factory = cast(Callable[[], Any], default.default_factory) default = MISSING else: default = default.default if default is MISSING and default_factory is MISSING: logger.warning_once( "%s.%s has no default or default factory.", cls.__name__, name ) return field(default=default, default_factory=default_factory, init=init) def is_init_field(cls: ConfigType, name: str) -> bool: return get_field(cls, name).init def replace(dataclass_instance: ConfigT, /, **kwargs) -> ConfigT: """Like [`dataclasses.replace`](https://docs.python.org/3/library/dataclasses.html#dataclasses.replace), but compatible with Pydantic dataclasses which use `pydantic.fields.Field` instead of `dataclasses.field`""" cls = type(dataclass_instance) dataclass_dict = dataclass_instance.__dict__ dataclass_dict = {k: v for k, v in dataclass_dict.items() if is_init_field(cls, k)} dataclass_dict.update(kwargs) return cls(**dataclass_dict) def getattr_iter( object: object, names: Sequence[str], default: Any | None = None, default_factory: Callable[[], Any] | None = None, warn: bool = False, ) -> Any: """ A helper function that retrieves an attribute from an object which may have multiple possible names. This is useful when fetching attributes from arbitrary `transformers.PretrainedConfig` instances. In the case where the first name in `names` is the preferred name, and any other names are deprecated aliases, setting `warn=True` will log a warning when a deprecated name is used. """ for i, name in enumerate(names): if hasattr(object, name): if warn and i > 0: logger.warning_once( "%s contains a deprecated attribute name '%s'. " "Please use the preferred attribute name '%s' instead.", type(object).__name__, name, names[0], ) return getattr(object, name) return default_factory() if default_factory is not None else default def get_attr_docs(cls: type[Any]) -> dict[str, str]: """ Get any docstrings placed after attribute assignments in a class body. https://davidism.com/mit-license/ """ cls_node = ast.parse(textwrap.dedent(inspect.getsource(cls))).body[0] if not isinstance(cls_node, ast.ClassDef): raise TypeError("Given object was not a class.") out = {} # Consider each pair of nodes. for a, b in pairwise(cls_node.body): # Must be an assignment then a constant string. if ( not isinstance(a, (ast.Assign, ast.AnnAssign)) or not isinstance(b, ast.Expr) or not isinstance(b.value, ast.Constant) or not isinstance(b.value.value, str) ): continue doc = inspect.cleandoc(b.value.value) # An assignment can have multiple targets (a = b = v), but an # annotated assignment only has one target. targets = a.targets if isinstance(a, ast.Assign) else [a.target] for target in targets: # Must be assigning to a plain name. if not isinstance(target, ast.Name): continue out[target.id] = doc return out @runtime_checkable class SupportsHash(Protocol): def compute_hash(self) -> str: ... class SupportsMetricsInfo(Protocol): def metrics_info(self) -> dict[str, str]: ... def update_config(config: ConfigT, overrides: dict[str, Any]) -> ConfigT: processed_overrides = {} for field_name, value in overrides.items(): assert hasattr(config, field_name), ( f"{type(config)} has no field `{field_name}`" ) current_value = getattr(config, field_name) if is_dataclass(current_value) and not is_dataclass(value): assert isinstance(value, dict), ( f"Overrides to {type(config)}.{field_name} must be a dict" f" or {type(current_value)}, but got {type(value)}" ) value = update_config( current_value, # type: ignore[type-var] value, ) processed_overrides[field_name] = value return replace(config, **processed_overrides) def normalize_value(x): """Return a stable, JSON-serializable canonical form for hashing. Order: primitives, special types (Enum, callable, torch.dtype, Path), then generic containers (Mapping/Set/Sequence) with recursion. """ # Fast path if x is None or isinstance(x, (bool, int, float, str)): return x # Enums: tag with FQN to avoid primitive collisions. # Ex: Enum(1) vs int(1) -> ("module.QualName", value). if isinstance(x, enum.Enum): enum_type = f"{x.__class__.__module__}.{x.__class__.__qualname__}" return (enum_type, normalize_value(x.value)) # Classes (types) are accepted and canonicalized by their fully-qualified # name (module.qualname) for a stable identifier. # Instances are only accepted if they expose uuid(); otherwise they are # rejected to avoid under-hashing object state. # Callables: accept classes only; reject funcs/lambdas/methods. # Used by LogitsProcessor types and ModelConfig.hf_overrides. if isinstance(x, type): module = getattr(x, "__module__", "") qual = getattr(x, "__qualname__", getattr(x, "__name__", "")) return ".".join([p for p in (module, qual) if p]) or repr(x) # Prefer stable uuid identifiers for objects that provide them, even if # they are callable instances (e.g., InductorPass wrappers). if hasattr(x, "uuid") and callable(getattr(x, "uuid", None)): return x.uuid() if callable(x): raise TypeError("normalize_value: function or callable instance unsupported") # Torch dtype: stringify (torch.float64 -> "torch.float64"). # We rely on the string form here; dtype-bearing fields that need additional # disambiguation should encode that at the config layer. if isinstance(x, torch.dtype): return str(x) # Bytes if isinstance(x, (bytes, bytearray)): return x.hex() # Paths (canonicalize) if isinstance(x, pathlib.Path): try: return str(x.expanduser().resolve()) except Exception: return str(x) # Dataclasses: represent as (FQN, sorted(field,value) tuple) for stability. if is_dataclass(x): type_fqn = f"{x.__class__.__module__}.{x.__class__.__qualname__}" items = tuple( (f.name, normalize_value(getattr(x, f.name))) for f in sorted(fields(x), key=lambda f: f.name) ) return (type_fqn, items) # Containers (generic) if isinstance(x, Mapping): return tuple(sorted((str(k), normalize_value(v)) for k, v in x.items())) if isinstance(x, Set): return tuple(sorted(repr(normalize_value(v)) for v in x)) if isinstance(x, Sequence) and not isinstance(x, (str, bytes, bytearray)): return tuple(normalize_value(v) for v in x) # PretrainedConfig if hasattr(x, "to_json_string") and callable(x.to_json_string): return x.to_json_string() # Unsupported type: e.g., modules, generators, open files, or objects # without a stable JSON/UUID representation. Hard-error to avoid # under-hashing. # If you hit this, either reshape your config to use supported primitives # and containers, or extend normalize_value to provide a stable encoding # (e.g., via uuid() or to_json_string()) for this type. raise TypeError( f"normalize_value: unsupported type '{type(x).__name__}'. " "Ensure config values use supported primitives/containers or add a " "stable representation for this type." ) def get_hash_factors(config: ConfigT, ignored_factors: set[str]) -> dict[str, object]: """Gets the factors used for hashing a config class. - Includes all dataclass fields not in `ignored_factors`. - Errors on non-normalizable values. """ factors: dict[str, object] = {} for dc_field in fields(config): factor = dc_field.name if factor in ignored_factors: continue value = getattr(config, factor, None) try: factors[factor] = normalize_value(value) except TypeError as e: raise TypeError( f"get_hash_factors: unsupported type for key '{factor}' " f"({type(value).__name__})" ) from e return factors def hash_factors(items: dict[str, object]) -> str: """Return a SHA-256 hex digest of the canonical items structure.""" return hashlib.sha256(json.dumps(items, sort_keys=True).encode()).hexdigest() @dataclass class Range: """ A range of numbers. Inclusive of start, inclusive of end. """ start: int end: int def is_single_size(self) -> bool: return self.start == self.end def __contains__(self, size: int) -> bool: # Inclusive of start, inclusive of end return self.start <= size <= self.end def __eq__(self, other: object) -> bool: if not isinstance(other, Range): return False return self.start == other.start and self.end == other.end def __hash__(self) -> int: return hash((self.start, self.end)) def __str__(self) -> str: return f"({self.start}, {self.end})" def __repr__(self) -> str: return self.__str__() def handle_deprecated( config: ConfigT, old_name: str, new_name_or_names: str | list[str], removal_version: str, ) -> None: old_val = getattr(config, old_name) if old_val is None: return if isinstance(new_name_or_names, str): new_names = [new_name_or_names] else: new_names = new_name_or_names msg = ( f"{old_name} is deprecated and will be removed in {removal_version}. " f"Use {', '.join(new_names)} instead." ) logger.warning(msg) for new_name in new_names: setattr(config, new_name, old_val) def get_from_deprecated_env_if_set( env_name: str, removal_version: str, field_name: str | None = None, ) -> str | None: """ Get value from deprecated environment variable with warning. Args: env_name: Name of the deprecated environment variable removal_version: Version when it will be removed field_name: Name of the field to suggest as alternative Returns: The environment variable value if set, None otherwise """ if envs.is_set(env_name): value = os.environ.get(env_name) alt_msg = f" Please use {field_name} instead." if field_name else "" logger.warning_once( "Using %s environment variable is deprecated and will be removed in %s.%s", env_name, removal_version, alt_msg, ) return value return None def set_from_deprecated_env_if_set( config: ConfigT, env_name: str, removal_version: str, field_name: str, to_bool: bool = False, to_int: bool = False, ) -> None: """ Set object field from deprecated environment variable with warning. Args: config: Config object to set the field on env_name: Name of the deprecated environment variable removal_version: Version when the env var will be removed field_name: Name of the field to set to_bool: Whether to convert the environment variable value to boolean to_int: Whether to convert the environment variable value to integer Returns: None """ if to_bool and to_int: raise ValueError("Cannot convert to both boolean and integer.") env_value = get_from_deprecated_env_if_set(env_name, removal_version, field_name) if env_value is not None: field_value: str | bool | int = env_value if to_bool: field_value = env_value.lower() in ("1", "true") elif to_int: field_value = int(env_value) setattr(config, field_name, field_value)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/config/utils.py", "license": "Apache License 2.0", "lines": 365, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/entrypoints/openai/test_uds.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from tempfile import TemporaryDirectory import httpx import pytest from vllm.version import __version__ as VLLM_VERSION from ...utils import RemoteOpenAIServer MODEL_NAME = "Qwen/Qwen3-0.6B" @pytest.fixture(scope="module") def server(): with TemporaryDirectory() as tmpdir: args = [ # use half precision for speed and memory savings in CI environment "--dtype", "bfloat16", "--max-model-len", "8192", "--enforce-eager", "--max-num-seqs", "128", "--uds", f"{tmpdir}/vllm.sock", ] with RemoteOpenAIServer(MODEL_NAME, args) as remote_server: yield remote_server @pytest.mark.asyncio async def test_show_version(server: RemoteOpenAIServer): transport = httpx.HTTPTransport(uds=server.uds) client = httpx.Client(transport=transport) response = client.get(server.url_for("version")) response.raise_for_status() assert response.json() == {"version": VLLM_VERSION}
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/test_uds.py", "license": "Apache License 2.0", "lines": 32, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/warmup/deep_gemm_warmup.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Warmup deep_gemm kernels. DeepGEMM JIT's the kernels. The warmup aims to JIT all the kernels that would be used during model execution beforehand. """ import torch from tqdm import tqdm import vllm.envs as envs from vllm.distributed.parallel_state import get_dp_group, is_global_first_rank from vllm.model_executor.layers.fused_moe.deep_gemm_moe import DeepGemmExperts from vllm.model_executor.layers.fused_moe.deep_gemm_utils import compute_aligned_M from vllm.model_executor.layers.fused_moe.layer import FusedMoE, FusedMoEModularMethod from vllm.model_executor.layers.fused_moe.triton_deep_gemm_moe import ( TritonOrDeepGemmExperts, ) from vllm.model_executor.layers.linear import LinearBase from vllm.model_executor.layers.quantization.fp8 import Fp8LinearMethod from vllm.tracing import instrument from vllm.utils.deep_gemm import ( fp8_gemm_nt, get_mk_alignment_for_contiguous_layout, m_grouped_fp8_gemm_nt_contiguous, ) from vllm.utils.math_utils import cdiv from vllm.utils.platform_utils import num_compute_units def _generate_optimal_warmup_m_values( max_tokens: int, n: int, device: torch.device ) -> list[int]: """ Generate M values that cover all possible DeepGEMM kernel configurations. Reference: https://github.com/deepseek-ai/DeepGEMM/blob/79f48ee15a82dd5fad5cd9beaa393c1f755e6b55/csrc/jit_kernels/heuristics/common.hpp Args: max_tokens: Maximum number of tokens to warmup for n: The actual N dimension from the weight tensor device: The torch device to get properties from. """ # DeepGEMM's possible block sizes block_ms = [64, 128, 256] block_ns = list(range(16, min(257, n + 1), 16)) num_sms = num_compute_units(device.index) m_values = set() # Always include small cases m_values.update([1, 2, 4] + [i for i in range(8, 65, 8)]) # Collect M values where different wave patterns occur for block_m in block_ms: for block_n in block_ns: if block_n > n: continue # Add key M boundaries for this block combination for wave in range(1, 11): # Up to 10 waves # M where this block config transitions to next wave target_blocks = wave * num_sms m = target_blocks * block_m // cdiv(n, block_n) if 1 <= m <= max_tokens: m_values.add(m) # Add block_m boundaries for multiple in range(1, max_tokens // block_m + 1): m = multiple * block_m if m <= max_tokens: m_values.add(m) return sorted(m_values) def _extract_data_from_linear_base_module( m: torch.nn.Module, ) -> tuple[torch.Tensor, torch.Tensor, list[int]]: """ Extract weights, weight scales and quantization block sizes from the given LinearBase module. """ assert isinstance(m, LinearBase) assert isinstance(m.quant_method, Fp8LinearMethod) assert m.quant_method.block_quant assert m.quant_method.quant_config is not None w = m.weight ws = m.weight_scale_inv if hasattr(m, "weight_scale_inv") else m.weight_scale quant_block_size = m.quant_method.quant_config.weight_block_size assert isinstance(w, torch.Tensor) assert isinstance(ws, torch.Tensor) assert quant_block_size is not None return (w, ws, quant_block_size) def _extract_data_from_fused_moe_module( m: torch.nn.Module, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, int]: """ Extract weights, weight scales and num_topk from FusedMoE module. """ assert isinstance(m, FusedMoE) w13 = m.w13_weight w13_s = ( m.w13_weight_scale_inv if hasattr(m, "w13_weight_scale_inv") else m.w13_weight_scale ) w2 = m.w2_weight w2_s = ( m.w2_weight_scale_inv if hasattr(m, "w2_weight_scale_inv") else m.w2_weight_scale ) num_topk = m.top_k assert isinstance(w13, torch.Tensor) assert isinstance(w13_s, torch.Tensor) assert isinstance(w2, torch.Tensor) assert isinstance(w2_s, torch.Tensor) return w13, w13_s, w2, w2_s, num_topk def _fp8_linear_may_use_deep_gemm(module: torch.nn.Module) -> bool: """ Return True if the input module/layer could be processed with DeepGEMM. """ # FIXME: this logic is brittle and incorrect - since we # could use DeepGEMM with for than just Fp8LinearMethod block_size = get_mk_alignment_for_contiguous_layout()[0] if not ( isinstance(module, LinearBase) and isinstance(module.quant_method, Fp8LinearMethod) and module.quant_method.block_quant and not module.quant_method.use_marlin ): return False w, _, block_sizes = _extract_data_from_linear_base_module(module) return ( block_sizes == get_mk_alignment_for_contiguous_layout() and w.ndim == 2 and w.shape[0] % block_size == 0 and w.shape[1] % block_size == 0 ) def _fused_moe_grouped_gemm_may_use_deep_gemm(module: torch.nn.Module) -> bool: if not (envs.VLLM_USE_DEEP_GEMM and envs.VLLM_MOE_USE_DEEP_GEMM): return False if not isinstance(module, FusedMoE): return False moe_quant_config = module.quant_method.get_fused_moe_quant_config(module) if ( moe_quant_config is None or moe_quant_config.quant_dtype != torch.float8_e4m3fn or moe_quant_config.block_shape != get_mk_alignment_for_contiguous_layout() ): return False if not isinstance(module.quant_method, FusedMoEModularMethod): # modular kernels could invoke deep_gemm_moe_fp8 return True # Further check if the ModularKernel implementation uses the DeepGemmExperts return isinstance( module.quant_method.moe_mk, (DeepGemmExperts, TritonOrDeepGemmExperts) ) FP8_GEMM_NT_WARMUP_CACHE: set[torch.Size] = set() def _get_fp8_gemm_nt_m_values(w: torch.Tensor, max_tokens: int) -> list[int]: """Get the M values to warmup for a given weight tensor.""" n, _ = w.size() device = w.device # Use optimal M values only if VLLM_DEEP_GEMM_WARMUP is set to "relax". # Otherwise warmup all token sizes to avoid JIT compilation in hotpath if envs.VLLM_DEEP_GEMM_WARMUP == "relax": return _generate_optimal_warmup_m_values(max_tokens, n, device) else: assert envs.VLLM_DEEP_GEMM_WARMUP == "full", ( "Expected " 'VLLM_DEEP_GEMM_WARMUP env to be set to "full" but got ' f"{envs.VLLM_DEEP_GEMM_WARMUP}" ) return list(range(1, max_tokens + 1)) def _deepgemm_fp8_gemm_nt_warmup( w: torch.Tensor, ws: torch.Tensor, max_tokens: int, pbar: tqdm | None = None, ): if w.size() in FP8_GEMM_NT_WARMUP_CACHE: return n, k = w.size() block_m = get_mk_alignment_for_contiguous_layout()[0] device = w.device a1q = torch.empty((max_tokens, k), device=device, dtype=torch.float8_e4m3fn) a1q_scales = torch.empty( (max_tokens, k // block_m), device=device, dtype=torch.float32 ) out = torch.empty((max_tokens, n), device=device, dtype=torch.bfloat16) m_values = _get_fp8_gemm_nt_m_values(w, max_tokens) for num_tokens in m_values: fp8_gemm_nt( (a1q[:num_tokens], a1q_scales[:num_tokens]), (w, ws), out[:num_tokens] ) if pbar is not None: pbar.update(1) FP8_GEMM_NT_WARMUP_CACHE.add(w.size()) GROUPED_FP8_GEMM_NT_CONTIGUOUS_WARMUP_CACHE: set[torch.Size] = set() def _get_grouped_gemm_params( w1: torch.Tensor, w2: torch.Tensor, num_topk: int, max_tokens: int, ) -> tuple[int, int, torch.Tensor]: assert w1.size(0) == w2.size(0), "w1 and w2 must have the same number of experts" block_m = get_mk_alignment_for_contiguous_layout()[0] num_experts = w1.size(0) device = w1.device # Assumes all ranks have the same max_num_batched_tokens max_tokens_across_dp = get_dp_group().world_size * max_tokens max_tokens = min(max_tokens_across_dp, envs.VLLM_FUSED_MOE_CHUNK_SIZE) # This is the maximum GroupedGemm M size that we expect to run # the grouped_gemm with. MAX_M = compute_aligned_M( max_tokens, num_topk, num_experts, block_m, expert_tokens_meta=None ) # Distribute expert-ids evenly. MAX_BLOCKS = MAX_M // block_m expert_ids_block = torch.randint( low=0, high=num_experts, size=(MAX_BLOCKS,), device=device, dtype=torch.int32 ) expert_ids = torch.repeat_interleave(expert_ids_block, block_m, dim=0) return MAX_M, block_m, expert_ids def _deepgemm_grouped_fp8_gemm_nt_contiguous_warmup( w1: torch.Tensor, w2: torch.Tensor, w1_scale: torch.Tensor, w2_scale: torch.Tensor, num_topk: int, max_tokens: int, pbar: tqdm | None = None, ): if ( w1.size() in GROUPED_FP8_GEMM_NT_CONTIGUOUS_WARMUP_CACHE and w2.size() in GROUPED_FP8_GEMM_NT_CONTIGUOUS_WARMUP_CACHE ): return MAX_M, block_m, expert_ids = _get_grouped_gemm_params(w1, w2, num_topk, max_tokens) device = w1.device def _warmup(w: torch.Tensor, w_scale: torch.Tensor): _, n, k = w.size() a1q = torch.empty((MAX_M, k), device=device, dtype=torch.float8_e4m3fn) a1q_scales = torch.empty( (MAX_M, k // block_m), device=device, dtype=torch.float32 ) out = torch.empty((MAX_M, n), device=device, dtype=torch.bfloat16) m_values = list(range(block_m, MAX_M + 1, block_m)) for num_tokens in m_values: m_grouped_fp8_gemm_nt_contiguous( (a1q[:num_tokens], a1q_scales[:num_tokens]), (w, w_scale), out[:num_tokens], expert_ids[:num_tokens], ) if pbar is not None: pbar.update(1) for w, ws in [(w1, w1_scale), (w2, w2_scale)]: if w.size() not in GROUPED_FP8_GEMM_NT_CONTIGUOUS_WARMUP_CACHE: _warmup(w, ws) GROUPED_FP8_GEMM_NT_CONTIGUOUS_WARMUP_CACHE.add(w.size()) def deepgemm_fp8_gemm_nt_warmup( model: torch.nn.Module, max_tokens: int, pbar: tqdm | None = None ): dg_modules = [m for m in model.modules() if _fp8_linear_may_use_deep_gemm(m)] for dgm in dg_modules: w, ws, _ = _extract_data_from_linear_base_module(dgm) _deepgemm_fp8_gemm_nt_warmup(w=w, ws=ws, max_tokens=max_tokens, pbar=pbar) def deepgemm_grouped_fp8_gemm_nt_contiguous_warmup( model: torch.nn.Module, max_tokens: int, pbar: tqdm | None = None ): dg_modules = [ m for m in model.modules() if _fused_moe_grouped_gemm_may_use_deep_gemm(m) ] for dgm in dg_modules: w13, w13_scale, w2, w2_scale, num_topk = _extract_data_from_fused_moe_module( dgm ) _deepgemm_grouped_fp8_gemm_nt_contiguous_warmup( w13, w2, w13_scale, w2_scale, num_topk, max_tokens, pbar=pbar ) def _count_warmup_iterations(model: torch.nn.Module, max_tokens: int) -> int: seen_fp8_sizes: set[torch.Size] = set(FP8_GEMM_NT_WARMUP_CACHE) seen_grouped_sizes: set[torch.Size] = set( GROUPED_FP8_GEMM_NT_CONTIGUOUS_WARMUP_CACHE ) total = 0 for m in model.modules(): if _fp8_linear_may_use_deep_gemm(m): w, _, _ = _extract_data_from_linear_base_module(m) if w.size() not in seen_fp8_sizes: total += len(_get_fp8_gemm_nt_m_values(w, max_tokens)) seen_fp8_sizes.add(w.size()) elif _fused_moe_grouped_gemm_may_use_deep_gemm(m): w13, _, w2, _, num_topk = _extract_data_from_fused_moe_module(m) if w13.size() in seen_grouped_sizes and w2.size() in seen_grouped_sizes: continue MAX_M, block_m, _ = _get_grouped_gemm_params(w13, w2, num_topk, max_tokens) n_values = (MAX_M - block_m) // block_m + 1 if w13.size() not in seen_grouped_sizes: total += n_values seen_grouped_sizes.add(w13.size()) if w2.size() not in seen_grouped_sizes: total += n_values seen_grouped_sizes.add(w2.size()) return total @instrument(span_name="DeepGemm warmup") def deep_gemm_warmup(model: torch.nn.Module, max_tokens: int): total = _count_warmup_iterations(model, max_tokens) if total == 0: return # Only show progress bar on rank 0 to avoid cluttered output if is_global_first_rank(): with tqdm(total=total, desc="DeepGEMM warmup") as pbar: deepgemm_fp8_gemm_nt_warmup(model, max_tokens, pbar) deepgemm_grouped_fp8_gemm_nt_contiguous_warmup(model, max_tokens, pbar) else: deepgemm_fp8_gemm_nt_warmup(model, max_tokens, None) deepgemm_grouped_fp8_gemm_nt_contiguous_warmup(model, max_tokens, None)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/warmup/deep_gemm_warmup.py", "license": "Apache License 2.0", "lines": 307, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/warmup/kernel_warmup.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Warmup kernels used during model execution. This is useful specifically for JIT'ed kernels as we don't want JIT'ing to happen during model execution. """ from typing import TYPE_CHECKING import torch import vllm.envs as envs from vllm.logger import init_logger from vllm.model_executor.warmup.deep_gemm_warmup import deep_gemm_warmup from vllm.platforms import current_platform from vllm.utils.deep_gemm import is_deep_gemm_supported from vllm.utils.flashinfer import has_flashinfer if TYPE_CHECKING: from vllm.v1.worker.gpu_model_runner import GPUModelRunner from vllm.v1.worker.gpu_worker import Worker logger = init_logger(__name__) def kernel_warmup(worker: "Worker"): # Deep GEMM warmup do_deep_gemm_warmup = ( envs.VLLM_USE_DEEP_GEMM and is_deep_gemm_supported() and envs.VLLM_DEEP_GEMM_WARMUP != "skip" ) if do_deep_gemm_warmup: model = worker.get_model() max_tokens = worker.scheduler_config.max_num_batched_tokens deep_gemm_warmup(model, max_tokens) enable_flashinfer_autotune = ( worker.vllm_config.kernel_config.enable_flashinfer_autotune ) # FlashInfer autotune for Hopper (SM 9.0) and Blackwell (SM 10.0) GPUs if enable_flashinfer_autotune is False: logger.info("Skipping FlashInfer autotune because it is disabled.") elif has_flashinfer() and current_platform.has_device_capability(90): flashinfer_autotune(worker.model_runner) # FlashInfer attention warmup # Only warmup if the model has FlashInfer attention groups # and is not a pooling model def _is_flashinfer_backend(backend): try: return backend.get_name() == "FLASHINFER" except NotImplementedError: return False if ( not worker.model_runner.is_pooling_model and worker.model_runner.attn_groups # NOTE: This should be `any` instead of `all` but other hybrid attention # backends don't support this dummy run. Once we remove # `build_for_cudagraph_capture`, we can change it to `any`. and all( _is_flashinfer_backend(group.backend) for groups in worker.model_runner.attn_groups for group in groups ) ): logger.info("Warming up FlashInfer attention.") # Warmup with mixed batch containing both prefill and decode tokens # This is to warm up both prefill and decode attention kernels worker.model_runner._dummy_run( num_tokens=16, skip_eplb=True, is_profile=True, force_attention=True, create_mixed_batch=True, ) def flashinfer_autotune(runner: "GPUModelRunner") -> None: """ Autotune FlashInfer operations. FlashInfer have many implementations for the same operation, autotuning runs benchmarks for each implementation and stores the results. The results are cached transparently and future calls to FlashInfer will use the best implementation. Without autotuning, FlashInfer will rely on heuristics, which may be significantly slower. """ from vllm.utils.flashinfer import autotune with torch.inference_mode(), autotune(): # We skip EPLB here since we don't want to record dummy metrics # When autotuning with number of tokens m, flashinfer will autotune # operations for all number of tokens up to m. # So we only need to run with the max number of tokens. runner._dummy_run( runner.scheduler_config.max_num_batched_tokens, skip_eplb=True, is_profile=True, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/warmup/kernel_warmup.py", "license": "Apache License 2.0", "lines": 89, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:benchmarks/multi_turn/bench_dataset.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import ABC, abstractmethod from statistics import mean from typing import Any, NamedTuple import numpy as np # type: ignore import pandas as pd # type: ignore from bench_utils import ( TEXT_SEPARATOR, Color, logger, ) from tqdm import tqdm from transformers import AutoTokenizer # type: ignore # Conversation ID is a string (e.g: "UzTK34D") ConvId = str # A list of dicts (dicts with keys "id" and "messages") ShareGptConversations = list[dict[str, Any]] # A list of dicts (dicts with keys "role" and "content") MessagesList = list[dict[str, str]] # Map conversation ID to conversation messages ConversationsMap = list[ConvId, MessagesList] class Distribution(ABC): @abstractmethod def sample(self, size: int = 1) -> np.ndarray: pass class UniformDistribution(Distribution): def __init__( self, min_val: int | float, max_val: int | float, is_integer: bool = True, ) -> None: self.min_val = min_val self.max_val = max_val self.is_integer = is_integer def sample(self, size: int = 1) -> np.ndarray: if self.is_integer: return np.random.randint( int(self.min_val), int(self.max_val + 1), size=size ) else: return np.random.uniform(self.min_val, self.max_val, size=size) def __repr__(self) -> str: return f"UniformDistribution[{self.min_val}, {self.max_val}]" class ConstantDistribution(Distribution): def __init__(self, value: int | float) -> None: self.value = value self.max_val = value def sample(self, size: int = 1) -> np.ndarray: return np.full(shape=size, fill_value=self.value) def __repr__(self) -> str: return f"Constant[{self.value}]" class ZipfDistribution(Distribution): def __init__(self, alpha: float, max_val: int | None = None) -> None: self.alpha = alpha self.max_val = max_val def sample(self, size: int = 1) -> np.ndarray: samples = np.random.zipf(self.alpha, size=size) if self.max_val: samples = np.minimum(samples, self.max_val) return samples def __repr__(self) -> str: return f"ZipfDistribution[{self.alpha}]" class PoissonDistribution(Distribution): def __init__(self, alpha: float, max_val: int | None = None) -> None: self.alpha = alpha self.max_val = max_val def sample(self, size: int = 1) -> np.ndarray: samples = np.random.poisson(self.alpha, size=size) if self.max_val: samples = np.minimum(samples, self.max_val) return samples def __repr__(self) -> str: return f"PoissonDistribution[{self.alpha}]" class LognormalDistribution(Distribution): def __init__( self, mean: float | None = None, sigma: float | None = None, average: int | None = None, median_ratio: float | None = None, max_val: int | None = None, ) -> None: self.average = average self.median_ratio = median_ratio self.max_val = max_val if average is not None: if average < 1: raise ValueError("Lognormal average must be positive") if mean or sigma: raise ValueError( "When using lognormal average, you can't provide mean/sigma" ) if self.median_ratio is None: # Default value that provides relatively wide range of values self.median_ratio = 0.85 # Calculate mean/sigma of np.random.lognormal based on the average mean, sigma = self._generate_lognormal_by_median( target_average=self.average, median_ratio=self.median_ratio ) else: if mean is None or sigma is None: raise ValueError( "Must provide both mean and sigma if average is not used" ) if mean <= 0 or sigma < 0: raise ValueError( "Lognormal mean must be positive and sigma must be non-negative" ) # Mean and standard deviation of the underlying normal distribution # Based on numpy.random.lognormal self.mean = mean self.sigma = sigma @staticmethod def _generate_lognormal_by_median( target_average: int, median_ratio: float ) -> tuple[float, float]: """ Compute (mu, sigma) for a lognormal distribution given: - a target average (mean of the distribution) - a ratio of median / mean (controls skewness), assume mean > median Background: If Z ~ Normal(mu, sigma^2), then X = exp(Z) ~ LogNormal(mu, sigma). * mean(X) = exp(mu + sigma^2 / 2) * median(X) = exp(mu) So: median / mean = exp(mu) / exp(mu + sigma^2 / 2) = exp(-sigma^2 / 2) Rearranging: sigma^2 = 2 * ln(mean / median) mu = ln(median) This gives a unique (mu, sigma) for any valid mean and median. """ # Check input validity: median must be smaller than mean if median_ratio <= 0 or median_ratio >= 1: raise ValueError("median_ratio must be in range (0, 1)") target_median = target_average * median_ratio # Solve sigma^2 = 2 * ln(mean / median) sigma = np.sqrt(2 * np.log(target_average / target_median)) mu = np.log(target_median) return mu, sigma def sample(self, size: int = 1) -> np.ndarray: samples = np.random.lognormal(mean=self.mean, sigma=self.sigma, size=size) if self.average is not None: # Scale to average samples *= self.average / samples.mean() if self.max_val: samples = np.minimum(samples, self.max_val) return np.round(samples).astype(int) def __repr__(self) -> str: if self.average: return ( f"LognormalDistribution[{self.average}, " f"{self.median_ratio}, {self.max_val}]" ) return f"LognormalDistribution[{self.mean}, {self.sigma}, {self.max_val}]" class GenConvArgs(NamedTuple): num_conversations: int text_files: list[str] input_num_turns: Distribution input_common_prefix_num_tokens: Distribution input_prefix_num_tokens: Distribution input_num_tokens: Distribution output_num_tokens: Distribution print_stats: bool def verify_field_exists( conf: dict, field_name: str, section: str, subsection: str ) -> None: if field_name not in conf: raise ValueError( f"Missing field '{field_name}' in {section=} and {subsection=}" ) def get_random_distribution( conf: dict, section: str, subsection: str, optional: bool = False ) -> Distribution: # section can be "prompt_input" or "prompt_output" (both required) conf = conf[section] if optional and subsection not in conf: # Optional subsection, if not found assume the value is always 0 return ConstantDistribution(0) # subsection can be "num_turns", "num_tokens" or "prefix_num_tokens" if subsection not in conf: raise ValueError(f"Missing subsection {subsection} in section {section}") conf = conf[subsection] distribution = conf.get("distribution") if distribution is None: raise ValueError( f"Missing field 'distribution' in {section=} and {subsection=}" ) if distribution == "constant": verify_field_exists(conf, "value", section, subsection) return ConstantDistribution(conf["value"]) elif distribution == "zipf": verify_field_exists(conf, "alpha", section, subsection) max_val = conf.get("max", None) return ZipfDistribution(conf["alpha"], max_val=max_val) elif distribution == "poisson": verify_field_exists(conf, "alpha", section, subsection) max_val = conf.get("max", None) return PoissonDistribution(conf["alpha"], max_val=max_val) elif distribution == "lognormal": max_val = conf.get("max", None) if "average" in conf: # Infer lognormal mean/sigma (numpy) from input average median_ratio = conf.get("median_ratio", None) return LognormalDistribution( average=conf["average"], median_ratio=median_ratio, max_val=max_val ) # Use mean/sigma directly (for full control over the distribution) verify_field_exists(conf, "mean", section, subsection) verify_field_exists(conf, "sigma", section, subsection) return LognormalDistribution( mean=conf["mean"], sigma=conf["sigma"], max_val=max_val ) elif distribution == "uniform": verify_field_exists(conf, "min", section, subsection) verify_field_exists(conf, "max", section, subsection) min_value = conf["min"] max_value = conf["max"] assert min_value > 0 assert min_value <= max_value is_integer = isinstance(min_value, int) and isinstance(max_value, int) return UniformDistribution(min_value, max_value, is_integer) else: raise ValueError(f"Unknown distribution: {distribution}") def parse_input_json_file(conf: dict) -> GenConvArgs: # Validate the input file assert isinstance(conf, dict) required_fields = [ "filetype", "num_conversations", "text_files", "prompt_input", "prompt_output", ] for field in required_fields: assert field in conf, f"Missing field {field} in input {conf}" assert conf["filetype"] == "generate_conversations" assert conf["num_conversations"] > 0, "num_conversations should be larger than zero" text_files = conf["text_files"] assert isinstance(text_files, list), "Field 'text_files' should be a list" assert len(text_files) > 0, ( "Field 'text_files' should be a list with at least one file" ) # Parse the parameters for the prompt input/output workload input_num_turns = get_random_distribution(conf, "prompt_input", "num_turns") input_num_tokens = get_random_distribution(conf, "prompt_input", "num_tokens") input_common_prefix_num_tokens = get_random_distribution( conf, "prompt_input", "common_prefix_num_tokens", optional=True ) input_prefix_num_tokens = get_random_distribution( conf, "prompt_input", "prefix_num_tokens" ) output_num_tokens = get_random_distribution(conf, "prompt_output", "num_tokens") print_stats: bool = conf.get("print_stats", False) assert isinstance(print_stats, bool), ( "Field 'print_stats' should be either 'true' or 'false'" ) args = GenConvArgs( num_conversations=conf["num_conversations"], text_files=text_files, input_num_turns=input_num_turns, input_common_prefix_num_tokens=input_common_prefix_num_tokens, input_prefix_num_tokens=input_prefix_num_tokens, input_num_tokens=input_num_tokens, output_num_tokens=output_num_tokens, print_stats=print_stats, ) return args def print_conv_stats(conversations: ConversationsMap, tokenizer: AutoTokenizer) -> None: # Collect statistics conv_stats: list[dict[Any, Any]] = [] req_stats: list[int] = [] print("\nCollecting statistics...") for messages in conversations.values(): # messages is a list of dicts user_tokens: list[int] = [] assistant_tokens: list[int] = [] request_tokens: list[int] = [] req_tokens = 0 for m in messages: content = m["content"] num_tokens = len(tokenizer(content).input_ids) if m["role"] == "user": user_tokens.append(num_tokens) # New user prompt including all chat history req_tokens += num_tokens request_tokens.append(req_tokens) elif m["role"] == "assistant": assistant_tokens.append(num_tokens) # Update assistant answer # (will be part of chat history for the next user prompt) req_tokens += num_tokens item_stats = { "conversation_turns": len(messages), "user_tokens": mean(user_tokens), "assistant_tokens": mean(assistant_tokens), } conv_stats.append(item_stats) req_stats.extend(request_tokens) # Print statistics percentiles = [0.25, 0.5, 0.75, 0.9, 0.99] print(TEXT_SEPARATOR) print(f"{Color.YELLOW}Conversations statistics:{Color.RESET}") print(TEXT_SEPARATOR) df = pd.DataFrame(conv_stats) print(df.describe(percentiles=percentiles).transpose()) print(TEXT_SEPARATOR) print(f"{Color.YELLOW}Request statistics:{Color.RESET}") print(TEXT_SEPARATOR) df = pd.DataFrame(req_stats, columns=["request_tokens"]) print(df.describe(percentiles=percentiles).transpose()) print(TEXT_SEPARATOR) def generate_conversations( args: GenConvArgs, tokenizer: AutoTokenizer ) -> ConversationsMap: # Text for all user prompts # (text from the input text files will be appended to this line) base_prompt_text = "Please rewrite the following text and add more content: " base_prompt_token_count = len( tokenizer.encode(base_prompt_text, add_special_tokens=False) ) logger.info(f"{Color.PURPLE}Generating conversations...{Color.RESET}") logger.info(args) list_of_tokens = [] for filename in args.text_files: # Load text file that will be used to generate prompts with open(filename) as file: data = file.read() tokens_in_file = tokenizer.encode(data, add_special_tokens=False) list_of_tokens.extend(tokens_in_file) logger.info( f"Loaded {len(tokens_in_file)} tokens from file {filename}, " f"total tokens so far: {len(list_of_tokens)}" ) conversations: ConversationsMap = {} conv_id = 0 # Generate number of turns for every conversation turn_count: np.ndarray = args.input_num_turns.sample(args.num_conversations) # Turn count should be at least 2 (one user prompt and one assistant answer) turn_count = np.maximum(turn_count, 2) # Round up to an even number (every user prompt should have an answer) turn_count = turn_count + (turn_count % 2) # Generate number of prefix tokens for every conversation conv_prefix_tokens: np.ndarray = args.input_prefix_num_tokens.sample( args.num_conversations ) # Used to reduce shared text between conversations # (jump/skip over text sections between conversations) base_offset = 0 # Common prefix size for all conversations (only 1 sample required) common_prefix_text = "" common_prefix_tokens: int = args.input_common_prefix_num_tokens.sample(1)[0] if common_prefix_tokens > 0: # Using "." at the end to separate sentences common_prefix_text = ( tokenizer.decode(list_of_tokens[: common_prefix_tokens - 2]) + "." ) base_offset += common_prefix_tokens for conv_id in tqdm( range(args.num_conversations), total=args.num_conversations, desc="Generating conversations", unit="conv", ): # Generate a single conversation messages: MessagesList = [] nturns = turn_count[conv_id] # User prompt token count per turn (with lower limit) input_token_count: np.ndarray = args.input_num_tokens.sample(nturns).astype(int) input_token_count = np.maximum(input_token_count, base_prompt_token_count) # Assistant answer token count per turn (with lower limit) output_token_count: np.ndarray = args.output_num_tokens.sample(nturns).astype( int ) output_token_count = np.maximum(output_token_count, 1) user_turn = True for turn_id in range(nturns): if user_turn: role = "user" num_tokens = input_token_count[turn_id] # Generate the user prompt, # use a unique prefix (the conv_id) for each conversation # (to avoid shared prefix between conversations) content = f"{conv_id} is a nice number... " if len(common_prefix_text) > 0 and turn_id == 0: content = common_prefix_text + content # Update the number of tokens left for the content num_tokens -= len(tokenizer.encode(content, add_special_tokens=False)) if turn_id == 0: prefix_num_tokens = conv_prefix_tokens[conv_id] if prefix_num_tokens > 0: # Add prefix text (context) to the first turn start_offset = base_offset end_offset = start_offset + prefix_num_tokens assert len(list_of_tokens) > end_offset, ( "Not enough input text to generate " f"{prefix_num_tokens} tokens for the " f"prefix text ({start_offset=}, {end_offset=})" ) content += f"{conv_id}, " + tokenizer.decode( list_of_tokens[start_offset:end_offset] ) base_offset += prefix_num_tokens # Add the actual user prompt/question after the prefix text content += base_prompt_text num_tokens -= base_prompt_token_count if num_tokens > 0: # Add text from the input file (to reach the desired token count) start_offset = base_offset + turn_id * input_token_count.max() end_offset = start_offset + num_tokens assert len(list_of_tokens) > end_offset, ( f"Not enough input text to generate {num_tokens} tokens " f"for the prompt ({start_offset=}, {end_offset=})" ) # Convert tokens back to text content += tokenizer.decode(list_of_tokens[start_offset:end_offset]) else: role = "assistant" # This content will not be used as input to the LLM server # (actual answers will be used instead). # Content is only required to determine the min_tokens/max_tokens # (inputs to the LLM server). num_tokens = output_token_count[turn_id] assert len(list_of_tokens) > num_tokens, ( f"Not enough input text to generate {num_tokens} " "tokens for assistant content" ) content = tokenizer.decode(list_of_tokens[:num_tokens]) # Append the user/assistant message to the list of messages messages.append({"role": role, "content": content}) user_turn = not user_turn # Add the new conversation conversations[f"CONV_ID_{conv_id}"] = messages # Increase base offset for the next conversation base_offset += nturns if args.print_stats: print_conv_stats(conversations, tokenizer) return conversations def conversations_list_to_dict(input_list: ShareGptConversations) -> ConversationsMap: conversations: ConversationsMap = {} for item in input_list: conv_id: str = item["id"] assert isinstance(conv_id, str) assert conv_id not in conversations, ( f"Conversation ID {conv_id} found more than once in the input" ) messages: MessagesList = item["messages"] assert isinstance(messages, list), ( f"Conversation messages should be a list (ID: {conv_id})" ) assert len(messages) > 0, f"Conversation with no messages (ID: {conv_id})" conversations[conv_id] = messages logger.info(f"Using {len(conversations)} unique conversations (IDs)") assert len(conversations) == len(input_list) # Print statistics about the selected conversations stats: list[dict[str, Any]] = [] for conv_data in conversations.values(): stats.append({"num_turns": len(conv_data)}) print(TEXT_SEPARATOR) print(f"{Color.YELLOW}Conversations statistics:{Color.RESET}") print(TEXT_SEPARATOR) percentiles = [0.25, 0.5, 0.75, 0.9, 0.99, 0.999, 0.9999] conv_stats = pd.DataFrame(stats).describe(percentiles=percentiles) print(conv_stats.transpose()) print(TEXT_SEPARATOR) return conversations def conversations_dict_to_list(input_dict: ConversationsMap) -> ShareGptConversations: output: ShareGptConversations = [] for conv_id, conv_data in input_dict.items(): new_item = {"id": conv_id, "messages": conv_data} output.append(new_item) return output
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/multi_turn/bench_dataset.py", "license": "Apache License 2.0", "lines": 474, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:benchmarks/multi_turn/bench_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import logging from enum import Enum class Color(Enum): RED = "\033[91m" GREEN = "\033[92m" BLUE = "\033[94m" PURPLE = "\033[95m" CYAN = "\033[96m" YELLOW = "\033[93m" RESET = "\033[0m" def __str__(self): return self.value TEXT_SEPARATOR = "-" * 100 # Configure the logger logging.basicConfig( level=logging.INFO, format="%(asctime)s [%(levelname)s] - %(message)s", datefmt="%d-%m-%Y %H:%M:%S", ) logger = logging.getLogger(__name__)
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/multi_turn/bench_utils.py", "license": "Apache License 2.0", "lines": 22, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:benchmarks/multi_turn/benchmark_serving_multi_turn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import asyncio import json import logging import multiprocessing as mp import os import random import time from collections import Counter, deque from datetime import datetime from enum import Enum from http import HTTPStatus from statistics import mean from typing import NamedTuple import aiohttp # type: ignore import numpy as np # type: ignore import pandas as pd # type: ignore from bench_dataset import ( ConversationsMap, ConvId, GenConvArgs, MessagesList, ShareGptConversations, conversations_dict_to_list, conversations_list_to_dict, generate_conversations, parse_input_json_file, ) from bench_utils import TEXT_SEPARATOR, Color, logger from transformers import AutoTokenizer # type: ignore NUM_TOKENS_FROM_DATASET = 0 TERM_SIGNAL = None class ConversationSampling(str, Enum): ROUND_ROBIN = "round_robin" RANDOM = "random" def __str__(self): return self.value class ClientArgs(NamedTuple): seed: int max_num_requests: int | None skip_first_turn: bool max_turns: int | None max_active_conversations: int verbose: bool print_content: bool verify_output: bool conversation_sampling: ConversationSampling request_rate: float max_retries: int class RequestArgs(NamedTuple): chat_url: str model: str stream: bool limit_min_tokens: int # Use negative value for no limit limit_max_tokens: int # Use negative value for no limit timeout_sec: int class BenchmarkArgs(NamedTuple): url: str num_clients: int early_stop: bool class ServerResponse(NamedTuple): valid: bool ttft_ms: float # time to first chunk tpot_ms: float # time per output chunk (one or more tokens) latency_ms: float start_time_ms: float first_chunk: str # first chunk of the content content: str # includes the first_chunk num_chunks: int def __str__(self) -> str: return f"ttft_ms {self.ttft_ms:.2f}, tpot_ms {self.tpot_ms:.2f}, latency_ms {self.latency_ms:.2f}" # noqa: E501 class RequestStats(NamedTuple): ttft_ms: float tpot_ms: float latency_ms: float start_time_ms: float input_num_turns: int input_num_tokens: int output_num_tokens: int output_num_chunks: int output_num_first_chunk_tokens: int approx_cached_percent: float conversation_id: str client_id: int def __str__(self) -> str: return ( f"ttft_ms {self.ttft_ms:.2f}, tpot_ms {self.tpot_ms:.2f}, latency_ms {self.latency_ms:.2f}, input_num_tokens {self.input_num_tokens}, " # noqa: E501 f"output_num_tokens {self.output_num_tokens} ({self.output_num_chunks} chunks, {self.output_num_first_chunk_tokens} tokens in first chunk), " # noqa: E501 f"approx_cached_percent {self.approx_cached_percent:.2f}%" ) class MetricStats: def __init__(self) -> None: self.min: float | None = None self.max: float | None = None self.avg: float | None = None self.sum = 0.0 self.count = 0 def update(self, value: float) -> None: if self.min is None: self.min = value else: self.min = min(self.min, value) if self.max is None: self.max = value else: self.max = max(self.max, value) self.sum += value self.count += 1 self.avg = self.sum / self.count def __repr__(self) -> str: if self.count == 0: return "no data" return f"avg: {self.avg:>10.3f}, min: {self.min:>10.3f}, max: {self.max:>10.3f}" class MovingAverage: def __init__(self, window_size: int) -> None: self.window_size = window_size self.window = np.zeros(window_size) self.index = 0 self.sum = 0.0 self.count = 0 self.avg: float | None = None def update(self, new_value: float) -> None: if self.count < self.window_size: # Filling up the window self.sum += new_value self.window[self.count] = new_value self.count += 1 else: # Window is full, start replacing old values old_value = self.window[self.index] self.sum = self.sum - old_value + new_value self.window[self.index] = new_value self.index = (self.index + 1) % self.window_size self.avg = self.sum / self.count def __repr__(self) -> str: if self.count == 0: return "no data" return f"avg: {self.avg:>10.3f} ({self.count} samples)" class DebugStats: def __init__(self, logger: logging.Logger, window_size: int) -> None: self.logger = logger self.metrics: dict[str, MovingAverage | MetricStats] = { "moving_avg_ttft_ms": MovingAverage(window_size), "moving_avg_tpot_ms": MovingAverage(window_size), "ttft_ms": MetricStats(), "tpot_ms": MetricStats(), "latency_ms": MetricStats(), "input_num_turns": MetricStats(), "input_num_tokens": MetricStats(), "output_num_tokens": MetricStats(), } def update(self, data: RequestStats) -> None: self.metrics["ttft_ms"].update(data.ttft_ms) self.metrics["moving_avg_ttft_ms"].update(data.ttft_ms) self.metrics["tpot_ms"].update(data.tpot_ms) self.metrics["moving_avg_tpot_ms"].update(data.tpot_ms) self.metrics["latency_ms"].update(data.latency_ms) self.metrics["input_num_turns"].update(data.input_num_turns) self.metrics["input_num_tokens"].update(data.input_num_tokens) self.metrics["output_num_tokens"].update(data.output_num_tokens) def print(self) -> None: self.logger.info("-" * 50) for k, v in self.metrics.items(): kv_info = f"[{k:25}] {v}" self.logger.info(kv_info) self.logger.info("-" * 50) def nanosec_to_millisec(value: float) -> float: return value / 1000000.0 def nanosec_to_sec(value: float) -> float: return value / 1000000000.0 async def send_request( session: aiohttp.ClientSession, messages: list[dict[str, str]], chat_url: str, model: str, stream: bool = True, min_tokens: int | None = None, max_tokens: int | None = None, timeout_sec: int = 120, ) -> ServerResponse: payload = { "model": model, "messages": messages, "seed": 0, "temperature": 0.0, } if stream: payload["stream"] = True payload["stream_options"] = {"include_usage": False} if min_tokens is not None: payload["min_tokens"] = min_tokens if max_tokens is not None: payload["max_tokens"] = max_tokens headers = {"Content-Type": "application/json"} # Calculate the timeout for the request if max_tokens is not None: # Assume TPOT of 200ms and use max_tokens to determine timeout token_based_timeout = int(max_tokens * 0.2) if token_based_timeout > timeout_sec: timeout_sec = token_based_timeout logger.info( "Using timeout of %ds based on max_tokens %d", timeout_sec, max_tokens, ) timeout = aiohttp.ClientTimeout(total=timeout_sec) valid_response = True ttft: float | None = None chunk_delay: list[int] = [] latency: float | None = None first_chunk = "" generated_text = "" start_time: int = time.perf_counter_ns() most_recent_timestamp: int = start_time async with session.post( url=chat_url, json=payload, headers=headers, timeout=timeout ) as response: http_status = HTTPStatus(response.status) if http_status == HTTPStatus.OK: async for chunk_bytes in response.content: chunk_bytes = chunk_bytes.strip() if not chunk_bytes: continue chunk = chunk_bytes.decode("utf-8").removeprefix("data: ") if chunk == "[DONE]": # End of stream latency = time.perf_counter_ns() - start_time elif stream is False: data = json.loads(chunk) message = data["choices"][0]["message"] assert message["role"] == "assistant" generated_text += message["content"] else: timestamp: int = time.perf_counter_ns() data = json.loads(chunk) # Delta is the new content/text/data delta = data["choices"][0]["delta"] if delta.get("content", None): if ttft is None: # First token first_token_time = time.perf_counter_ns() ttft = first_token_time - start_time first_chunk = delta["content"] else: # Decoding phase chunk_delay.append(timestamp - most_recent_timestamp) generated_text += delta["content"] most_recent_timestamp = timestamp else: valid_response = False content = await response.text() logger.warning( f"{Color.YELLOW}Received HTTP status {http_status.value} " f"({http_status.phrase}): {content}{Color.RESET}" ) if latency is None: latency = -1.0 if valid_response: # Streaming is disabled, latency was not set latency = time.perf_counter_ns() - start_time if ttft is None: # The response was a single chunk ttft = latency # Each chunk may include more than one token tpot: float = mean(chunk_delay) if len(chunk_delay) > 0 else 0.0 num_chunks: int = len(chunk_delay) sr = ServerResponse( valid=valid_response, ttft_ms=nanosec_to_millisec(ttft) if ttft > 0.0 else -1.0, tpot_ms=nanosec_to_millisec(tpot), latency_ms=nanosec_to_millisec(latency), start_time_ms=nanosec_to_millisec(start_time), first_chunk=first_chunk, content=generated_text, num_chunks=num_chunks, ) return sr def get_short_string(input: str) -> str: n = 20 if len(input) < 400: return input return f"{input[:n]}...{input[-n:]}" def get_token_count(tokenizer: AutoTokenizer, text: str) -> int: return len(tokenizer(text, add_special_tokens=False).input_ids) def get_messages_token_count( tokenizer: AutoTokenizer, messages: list[dict[str, str]] ) -> int: token_count = 0 for m in messages: token_count += get_token_count(tokenizer, m["content"]) return token_count async def send_turn( session: aiohttp.ClientSession, client_id: int, conv_id: str, conversation_messages: MessagesList, messages_to_use: int, tokenizer: AutoTokenizer, req_args: RequestArgs, verbose: bool, verify_output: bool, ) -> RequestStats | None: assert messages_to_use > 0 assert messages_to_use <= len(conversation_messages) messages = conversation_messages[:messages_to_use] # Index of the next message (the role should be "user") index = messages_to_use - 1 # Verify that the message has only two keys, "role" and "content" assert len(messages[index].keys()) == 2 assert "role" in messages[index] and "content" in messages[index] assert messages[index]["role"] == "user", ( f"Failed on conversation ID {conv_id}, message role should be user" ) if verbose: print( f"{Color.CYAN}Messages (conversation ID {conv_id}," f" {len(messages)} turns):{Color.RESET}", messages, ) # None means that there is no upper/lower limit for the output token count min_tokens = None if req_args.limit_min_tokens < 0 else req_args.limit_min_tokens max_tokens = None if req_args.limit_max_tokens < 0 else req_args.limit_max_tokens if len(conversation_messages) > messages_to_use: # The conversation contains an assistant answer for the next user prompt if ( min_tokens == NUM_TOKENS_FROM_DATASET or max_tokens == NUM_TOKENS_FROM_DATASET ): # Compute number of tokens in the answer (from the input conversation) assistant_answer = conversation_messages[messages_to_use] answer_num_tokens = get_token_count(tokenizer, assistant_answer["content"]) assert assistant_answer["role"] == "assistant" if min_tokens == NUM_TOKENS_FROM_DATASET: min_tokens = max(1, answer_num_tokens) if max_tokens == NUM_TOKENS_FROM_DATASET: max_tokens = max(1, answer_num_tokens) # Send the current conversation to LLM and get a response response: ServerResponse = await send_request( session, messages, req_args.chat_url, req_args.model, req_args.stream, min_tokens, max_tokens, req_args.timeout_sec, ) if response.valid is False: # Request failed return None # Compute number of tokens in input / output input_num_tokens = get_messages_token_count(tokenizer, messages) # Num tokens in the user's last question question_num_tokens = get_token_count(tokenizer, messages[index]["content"]) # Num tokens in the history/context of the question assert input_num_tokens >= question_num_tokens history_num_tokens = input_num_tokens - question_num_tokens # Num tokens in the LLM's answer (first chunk and full answer) first_chunk_tokens = get_token_count(tokenizer, response.first_chunk) output_content = response.content output_num_tokens = get_token_count(tokenizer, output_content) # Prefix caching approximated cached percent approx_cached_percent = ( 100.0 * (history_num_tokens / input_num_tokens) if input_num_tokens > 0 else 0.0 ) # Compute the correct TTFT and TPOT (based on tokens and not chunks). # Required because multiple output tokens may be bundled in a single chunk. if output_num_tokens > 1 and output_num_tokens > first_chunk_tokens: # More than one token and more than one chunk in the output decode_ms = response.latency_ms - response.ttft_ms decode_num_tokens = output_num_tokens - first_chunk_tokens tpot_ms = decode_ms / decode_num_tokens else: # In this case: output_num_tokens == first_chunk_tokens # Output was a single chunk (output_num_tokens > 1) # or even a single token (output_num_tokens == 1) tpot_ms = 0.0 if first_chunk_tokens > 1: # First chunk had multiple tokens, adjust TTFT for a single token delta_ms = (first_chunk_tokens - 1) * tpot_ms ttft_ms = max(0.1, response.ttft_ms - delta_ms) else: # First chunk had only one token ttft_ms = response.ttft_ms rs = RequestStats( ttft_ms=ttft_ms, tpot_ms=tpot_ms, latency_ms=response.latency_ms, start_time_ms=response.start_time_ms, input_num_turns=len(messages), input_num_tokens=input_num_tokens, output_num_tokens=output_num_tokens, output_num_chunks=response.num_chunks, output_num_first_chunk_tokens=first_chunk_tokens, approx_cached_percent=approx_cached_percent, conversation_id=conv_id, client_id=client_id, ) if verbose: print( f"\n{Color.YELLOW}Response ({output_num_tokens} tokens):{Color.RESET}", output_content, ) print(f"{Color.YELLOW}Response metrics: {rs}{Color.RESET}") print("-" * 70) # Save the LLM's answer (will be used as part of the context for the next user turn) answer_index = messages_to_use if len(conversation_messages) > answer_index: assert conversation_messages[answer_index]["role"] == "assistant", ( f"Failed on conversation ID {conv_id}, message role should be assistant" ) orig_content = conversation_messages[answer_index]["content"] if verify_output: # Compare the new answer to the answer from the input file debug_info = ( f"LLM/dataset answers do not match ({conv_id}):" f"\n'{get_short_string(output_content)}' (len: {len(output_content)})," f"\n'{get_short_string(orig_content)}' (len: {len(orig_content)})" ) if orig_content != output_content: raise ValueError(debug_info) # Update the answer conversation_messages[answer_index]["content"] = output_content else: # A user prompt that has no answer, add the answer as a new message new_answer = {"role": "assistant", "content": output_content} conversation_messages.append(new_answer) return rs async def poisson_sleep(request_rate: float, verbose: bool = False) -> None: # Generate a random time interval from the Poisson distribution assert request_rate > 0 interval = np.random.exponential(1.0 / request_rate) if verbose: logger.info(f"Sleeping for {interval:.3f} seconds...") await asyncio.sleep(interval) async def exponential_backoff_sleep( attempt_cnt: int, base_rate: float = 1.0, backoff_factor: float = 2.0, jitter_fraction: float = 0.10, verbose: bool = False, ) -> None: # Sleep with exponential backoff and jitter after a failed request. backoff_delay = base_rate * (backoff_factor**attempt_cnt) jittered_delay = backoff_delay * ( 1 + np.random.uniform(-jitter_fraction, jitter_fraction) ) if verbose: logger.info(f"Backoff for {jittered_delay:.3f} seconds...") await asyncio.sleep(jittered_delay) async def client_main( args: ClientArgs, req_args: RequestArgs, client_id: int, tokenizer: AutoTokenizer, stop_event: mp.Event, # type: ignore task_queue: mp.Queue, result_queue: mp.Queue, conv_queue: mp.Queue, ) -> None: logger.info( f"{Color.CYAN}Started client {client_id}: max_num_requests={args.max_num_requests}, max_active_conversations={args.max_active_conversations}{Color.RESET}" # noqa: E501 ) # Set unique seed per client (each client runs in its own process) # Add 1 to ensure no client uses the same seed as the main process client_seed = args.seed + client_id + 1 random.seed(client_seed) np.random.seed(client_seed) # Active conversations active_convs: ConversationsMap = {} conv_id_queue: deque = deque(maxlen=args.max_active_conversations) # Keep track of how many messages have been used for each conversation turns_count: Counter = Counter() num_successes = 0 num_failures = 0 # Track the timestamp (time.perf_counter()) # of the last turn per conversation (only for debug) time_of_last_turn: dict[ConvId, float] = {} # Flag that indicates that there are no new tasks (conversations) for the client task_queue_empty = False async with aiohttp.ClientSession() as session: # Print progress while task_queue_empty is False: result = None if ( args.max_num_requests and num_successes + num_failures == args.max_num_requests ): logger.info( f"{Color.YELLOW}Client {client_id} reached " f"request limit{Color.RESET}" ) break if stop_event.is_set(): # type: ignore logger.info( f"{Color.YELLOW}Client {client_id} received " f"a termination signal{Color.RESET}" ) break while ( len(active_convs) < args.max_active_conversations and task_queue_empty is False ): # Get a new conversation from the task queue conv_id, messages = task_queue.get() if conv_id is TERM_SIGNAL: task_queue_empty = True break if args.skip_first_turn: # Skip the first turn (both user and assistant), # relevant if warmup was enabled. # Default turns_count[conv_id] will be zero if conv_id # was never inserted/updated in turns_count. turns_count[conv_id] += 2 if turns_count[conv_id] < len(messages): # Add new conversation active_convs[conv_id] = messages conv_id_queue.append(conv_id) if args.verbose: logger.info( f"{Color.GREEN}Client {client_id} will use conversation ID {conv_id} (active conversations {len(active_convs)}){Color.RESET}" # noqa: E501 ) elif args.verbose: # No more messages (conversation finished during the warmup) logger.info( f"{Color.YELLOW}Client {client_id} will not use conversation ID {conv_id} (all {len(messages)} messages already sent){Color.RESET}" # noqa: E501 ) if len(active_convs) == 0 or task_queue_empty: logger.info( f"{Color.YELLOW}Client {client_id} has no more work{Color.RESET}" ) break # Pick an active conversation for the next request if args.conversation_sampling == ConversationSampling.ROUND_ROBIN: conv_id = conv_id_queue.pop() else: # ConversationSampling.RANDOM active_ids = list(active_convs.keys()) conv_id = random.choice(active_ids) messages = active_convs[conv_id] assert isinstance(messages, list) and len(messages) > 0 # Update the amount of messages to use turns_count[conv_id] += 1 current_turn = turns_count[conv_id] assert current_turn < len(messages), ( f"Turn number {current_turn} is invalid for conversation ID {conv_id}" f" that has only {len(messages)} messages" ) if args.verbose: curr_time_sec: float = time.perf_counter() time_since_last_turn: str | float = "N/A" if conv_id in time_of_last_turn: time_since_last_turn = round( curr_time_sec - time_of_last_turn[conv_id], 3 ) logger.info( f"Client {client_id} using conversation ID {conv_id} (turn: {current_turn}, time since last turn [sec]: {time_since_last_turn})" # noqa: E501 ) time_of_last_turn[conv_id] = curr_time_sec success = False for attempt_cnt in range(args.max_retries + 1): try: exception = False result = await send_turn( session, client_id, conv_id, messages, current_turn, tokenizer, req_args, args.print_content, args.verify_output, ) if result is not None: result_queue.put(result) success = True break else: logger.warning( f"{Color.YELLOW}Client {client_id} - Request rejected during conversation ID {conv_id} (turn: {current_turn}){Color.RESET}" # noqa: E501 ) except asyncio.exceptions.TimeoutError: exception = True logger.error( "%sClient %d - Timeout during conversation ID %s (turn: %d). " "Base timeout is %ss (set with --request-timeout-sec), but the " "effective timeout may be longer based on max_tokens. If this " "is unexpected, consider increasing the timeout or checking " "model performance.%s", Color.RED, client_id, conv_id, current_turn, req_args.timeout_sec, Color.RESET, ) except Exception: exception = True logger.exception( f"{Color.RED}Client {client_id} - Exception during conversation ID {conv_id} (turn: {current_turn}){Color.RESET}" # noqa: E501 ) # Sleep before retry if not last attempt if not success and attempt_cnt < args.max_retries: await exponential_backoff_sleep(attempt_cnt, verbose=args.verbose) if not success: num_failures += 1 # Remove the conversation (should not be used again) active_convs.pop(conv_id) if exception: break # Exit gracefully instead of raising an error else: num_successes += 1 # Update the turns counter to include the LLM response # The LLM response will be used as context for the next user turn turns_count[conv_id] += 1 max_turns = len(messages) if args.max_turns is not None: # Limit the number of turns in the conversation max_turns = min(args.max_turns, max_turns) if turns_count[conv_id] >= max_turns: # Conversation has no more turns (no longer active) # save the updated conversation (with the LLM server's answer) conv_queue.put((conv_id, active_convs.pop(conv_id))) if args.verbose: logger.info( f"{Color.GREEN}Client {client_id} finished " f"conversation ID {conv_id}{Color.RESET}" ) else: # Conversation is not finished, insert it at the back of the queue conv_id_queue.appendleft(conv_id) # Sleep between requests (if lambda is positive) if args.request_rate > 0: await poisson_sleep(args.request_rate, args.verbose) # Send indication that the client is done conv_queue.put((TERM_SIGNAL, TERM_SIGNAL)) logger.info( f"{Color.CYAN}Client {client_id} is done " f"({num_successes=}, {num_failures=}){Color.RESET}" ) def worker_function( client_id: int, tokenizer: AutoTokenizer, client_args: ClientArgs, req_args: RequestArgs, stop_event: mp.Event, # type: ignore task_queue: mp.Queue, result_queue: mp.Queue, conv_queue: mp.Queue, ) -> None: asyncio.run( client_main( client_args, req_args, client_id, tokenizer, stop_event, task_queue, result_queue, conv_queue, ) ) def get_client_config( args: argparse.Namespace, input_conv: ConversationsMap ) -> tuple[ClientArgs, RequestArgs]: if args.num_clients < 1: raise ValueError("Number of clients must be a positive number") if len(input_conv) < args.num_clients: raise ValueError( "Number of conversations must be equal or larger than the number of clients" ) max_req_per_client: int | None = None if args.max_num_requests is not None: # Max number of requests per client req_per_client = args.max_num_requests // args.num_clients if req_per_client < 1: raise ValueError("Number of requests should be at least one per client") max_req_per_client = req_per_client max_active_conversations = args.max_active_conversations if max_active_conversations is None: # Each client will have only one active conversation at a time max_active_conversations = args.num_clients if max_active_conversations > len(input_conv): raise ValueError( f"Max active conversations {max_active_conversations} " "must be equal or less than the total number of conversations" ) # Max number of active conversations per client max_active_conv_per_client = max_active_conversations // args.num_clients if max_active_conv_per_client < 1: raise ValueError( f"Max active conversations {max_active_conversations} " "must be equal or greater than the number of clients" ) # Skip the first user turn (as part of the warmup) skip_first_turn = args.warmup_step # Common arguments for all clients client_args = ClientArgs( seed=args.seed, max_num_requests=max_req_per_client, skip_first_turn=skip_first_turn, max_turns=args.max_turns, max_active_conversations=max_active_conv_per_client, verbose=args.verbose, print_content=args.print_content, verify_output=args.verify_output, conversation_sampling=args.conversation_sampling, request_rate=args.request_rate, max_retries=args.max_retries, ) if args.limit_min_tokens > 0 or args.limit_max_tokens > 0: if args.limit_min_tokens < 1 or args.limit_max_tokens < 1: raise ValueError( "Invalid min/max tokens limits (both limits should be provided)" ) if args.limit_min_tokens > args.limit_max_tokens: raise ValueError( "Invalid min/max tokens limits (min should not be larger than max)" ) if args.request_timeout_sec <= 0: raise ValueError("Request timeout must be a positive number") # Arguments for API requests chat_url = f"{args.url}/v1/chat/completions" model_name = args.served_model_name if args.served_model_name else args.model req_args = RequestArgs( chat_url=chat_url, model=model_name, stream=not args.no_stream, limit_min_tokens=args.limit_min_tokens, limit_max_tokens=args.limit_max_tokens, timeout_sec=args.request_timeout_sec, ) return client_args, req_args async def main_mp( client_args: ClientArgs, req_args: RequestArgs, bench_args: BenchmarkArgs, tokenizer: AutoTokenizer, input_conv: ConversationsMap, ) -> tuple[ConversationsMap, list[RequestStats]]: # An event that will trigger graceful termination of all the clients stop_event = mp.Event() # Queue for input conversations (from the input file/dataset) task_queue: mp.Queue = mp.Queue() # Queue for client measurements (TTFT, TPOT, etc. for each request) result_queue: mp.Queue = mp.Queue() # Queue for output conversations (with the LLM answers, sent by the server) conv_queue: mp.Queue = mp.Queue() output_conv: ConversationsMap = {} client_metrics: list[RequestStats] = [] # Start all clients start_time = time.perf_counter_ns() logger.info(f"{Color.GREEN}Starting {bench_args.num_clients} clients{Color.RESET}") clients = [] for client_id in range(bench_args.num_clients): client = mp.Process( name=f"client_{client_id}", target=worker_function, args=( client_id, tokenizer, client_args, req_args, stop_event, task_queue, result_queue, conv_queue, ), ) clients.append(client) client.start() # Submit all the input conversations as tasks for the clients for conv_id, messages in input_conv.items(): task_queue.put((conv_id, messages)) # Add termination signals for clients for _ in range(bench_args.num_clients): task_queue.put((TERM_SIGNAL, TERM_SIGNAL)) # Collect the updated conversations from all clients num_clients_finished = 0 total_convs = len(input_conv) debug_stats = DebugStats(logger, min(15 * bench_args.num_clients, 500)) while num_clients_finished < bench_args.num_clients: # Collect updated conversation conv_id, messages = conv_queue.get() # Collect results (measurements) while not result_queue.empty(): new_data = result_queue.get() client_metrics.append(new_data) debug_stats.update(new_data) if conv_id is TERM_SIGNAL: num_clients_finished += 1 logger.info( f"{Color.CYAN}{num_clients_finished} out of " f"{bench_args.num_clients} clients finished{Color.RESET}" ) if bench_args.early_stop and not stop_event.is_set(): # Once one client finished, stop all other clients. # there is no reason to continue the benchmark with fewer clients. logger.info( f"{Color.YELLOW}Sending termination signal to clients{Color.RESET}" ) stop_event.set() else: output_conv[conv_id] = messages finished_convs = len(output_conv) percent = finished_convs / total_convs # Tuned to control the print rate (can be changed if required) print_cycle = max(3, int(bench_args.num_clients / 4)) if finished_convs % print_cycle == 0: runtime_sec = nanosec_to_sec(time.perf_counter_ns() - start_time) logger.info( f"{Color.CYAN}Finished {finished_convs} out of {total_convs} conversations ({percent:.0%}), " # noqa: E501 f"{num_clients_finished} out of {bench_args.num_clients} clients finished, collected {len(client_metrics)} measurements, runtime {runtime_sec:.3f} sec{Color.RESET}" # noqa: E501 ) rps: str | float = round(len(client_metrics) / runtime_sec, 3) if len(client_metrics) < (5 * bench_args.num_clients): # Do not estimate the RPS if the number of samples is very low # (threshold can be tuned if needed) rps = "N/A" runtime_left_sec: str | float = round( (runtime_sec / finished_convs) * (total_convs - finished_convs), 3 ) if percent < 0.05: # If less than 5% of the conversations were not finished, # the estimation will probably be very inaccurate # (threshold can be tuned if needed). runtime_left_sec = "N/A" logger.info( f"{Color.CYAN}Estimated req/sec {rps}, estimated runtime left {runtime_left_sec} sec{Color.RESET}" # noqa: E501 ) debug_stats.print() logger.info( f"{Color.CYAN}All {bench_args.num_clients} clients finished{Color.RESET}" ) # At this point all the clients finished, # collect results (TTFT, TPOT, etc.) from all the clients. # This needs to happen before calling join on the clients # (result_queue should be emptied). while not result_queue.empty(): client_metrics.append(result_queue.get()) logger.info(f"Collected {len(client_metrics)} samples from all the clients") # Wait for all clients to finish for client in clients: logger.info( f"{Color.CYAN}Waiting for client {client.name} " f"(is alive: {client.is_alive()}){Color.RESET}" ) client.join(timeout=req_args.timeout_sec + 1) if client.is_alive(): logger.warning( f"{Color.YELLOW}Client {client.name} will be terminated{Color.RESET}" ) client.terminate() exitcode = client.exitcode if exitcode != 0: logger.error( f"{Color.RED}Client {client.name} exited " f"with exit code {exitcode}{Color.RESET}" ) logger.info( f"All {bench_args.num_clients} clients exited (successfully " f"finished {len(output_conv)} out of {total_convs} conversations)" ) # Queues should be closed, required to avoid hang at interpreter shutdown unfinished_tasks = 0 while not task_queue.empty(): task_queue.get() unfinished_tasks += 1 if unfinished_tasks > 0: # Can happen if not all tasks (conversations) have finished. # May happen if --max-num-requests was used, # or if an error occurred in one of the clients. logger.debug(f"Discarding {unfinished_tasks} unfinished tasks") task_queue.close() task_queue.join_thread() result_queue.close() result_queue.join_thread() conv_queue.close() conv_queue.join_thread() return output_conv, client_metrics def get_filename_with_timestamp(label: str, extension: str) -> str: time_now = datetime.now() timestamp = time_now.strftime("%d-%m-%Y_%H-%M-%S") filename = f"{label}__{timestamp}.{extension}" return filename def process_statistics( client_metrics: list[RequestStats], warmup_percentages: list[float], test_params: dict, verbose: bool, gen_conv_args: GenConvArgs | None = None, excel_output: bool = False, warmup_runtime_sec: float | None = None, ) -> None: if len(client_metrics) == 0: logger.info("No samples to process") return logger.info(f"Processing {len(client_metrics)} samples...") raw_data = pd.DataFrame(client_metrics) if verbose: # Calculate the time between user turns in each conversation (in a new column) raw_data = raw_data.sort_values(by=["conversation_id", "start_time_ms"]) raw_data["time_between_user_turns_sec"] = raw_data.groupby("conversation_id")[ "start_time_ms" ].diff() # Convert milliseconds to seconds raw_data["time_between_user_turns_sec"] = ( raw_data["time_between_user_turns_sec"] / 1000.0 ) # Final raw data should be sorted by time raw_data = raw_data.sort_values(by=["start_time_ms"]) raw_data["end_time_ms"] = raw_data["start_time_ms"] + raw_data["latency_ms"] percentiles = [0.25, 0.5, 0.75, 0.9] # Add more percentiles if there are enough samples if len(raw_data) >= 100: percentiles.append(0.99) if len(raw_data) >= 1000: percentiles.append(0.999) if len(raw_data) >= 10000: percentiles.append(0.9999) # Set precision for numbers in the output text (the dataframes) pd.set_option("display.precision", 2) # Exclude parameters from RequestStats exclude = [ "start_time_ms", "end_time_ms", "output_num_first_chunk_tokens", "approx_cached_percent", "conversation_id", "client_id", ] print(TEXT_SEPARATOR) print(f"{Color.YELLOW}Parameters:{Color.RESET}") for k, v in test_params.items(): print(f"{k}={v}") # conversations generation parameters if gen_conv_args is not None: gen_params = { "text_files": ", ".join(gen_conv_args.text_files), "input_num_turns": str(gen_conv_args.input_num_turns), "input_common_prefix_num_tokens": str( gen_conv_args.input_common_prefix_num_tokens ), "input_prefix_num_tokens": str(gen_conv_args.input_prefix_num_tokens), "input_num_tokens": str(gen_conv_args.input_num_tokens), "output_num_tokens": str(gen_conv_args.output_num_tokens), } print(f"{Color.YELLOW}Conversations Generation Parameters:{Color.RESET}") for k, v in gen_params.items(): print(f"{k}={v}") print(TEXT_SEPARATOR) params_list = [] df_list = [] for percent in warmup_percentages: # Select samples from the end (tail) of the dataframe warmup_count = int(percent * len(raw_data)) tail_count = len(raw_data) - warmup_count if tail_count == 0: # No reason to process if the count of samples is zero break df = raw_data.tail(tail_count) # Runtime is the diff between the end of the last request # and the start of the first request runtime_sec = df["end_time_ms"].iloc[-1] - df["start_time_ms"].iloc[0] # Convert milliseconds to seconds runtime_sec = runtime_sec / 1000.0 requests_per_sec = float(len(df)) / runtime_sec params = { "runtime_sec": runtime_sec, "requests_per_sec": requests_per_sec, } if warmup_runtime_sec is not None: params["warmup_runtime_sec"] = warmup_runtime_sec params["total_runtime_incl_warmup_sec"] = runtime_sec + warmup_runtime_sec # Generate a summary of relevant metrics (and drop irrelevant data) df = df.drop(columns=exclude).describe(percentiles=percentiles).transpose() # List for Excel file params_list.append(params) df_list.append(df) # Print the statistics summary if percent > 0 or len(warmup_percentages) > 1: print( f"{Color.YELLOW}Statistics summary " f"(assuming {percent:.0%} warmup samples):{Color.RESET}" ) else: print(f"{Color.YELLOW}Statistics summary:{Color.RESET}") for k, v in params.items(): if isinstance(v, float): print(f"{k} = {v:.3f}") else: print(f"{k} = {v}") print(TEXT_SEPARATOR) print(df) print(TEXT_SEPARATOR) if excel_output: prefix = f"statistics_{test_params['num_clients']}_clients" filename = get_filename_with_timestamp(prefix, "xlsx") with pd.ExcelWriter(filename, engine="xlsxwriter") as writer: startrow = 0 test_params_df = pd.DataFrame([test_params]) test_params_df.to_excel( writer, sheet_name="Summary", index=False, startrow=startrow ) startrow += len(test_params_df) + 3 if gen_conv_args is not None: gen_params_df = pd.DataFrame([gen_params]) gen_params_df.to_excel( writer, sheet_name="Summary", index=False, startrow=(startrow - 1) ) startrow += len(gen_params_df) + 3 for params, df_stats in zip(params_list, df_list): df_params = pd.DataFrame([params]) df_params.to_excel( writer, sheet_name="Summary", index=False, startrow=startrow ) startrow += len(df_params) + 2 df_stats.to_excel( writer, sheet_name="Summary", index=True, startrow=startrow ) startrow += len(df_stats) + 3 raw_data.to_excel(writer, sheet_name="Raw data", index=False, startrow=0) logger.info( f"{Color.GREEN}Client metrics exported to file: {filename}{Color.RESET}" ) async def get_server_info(url: str) -> None: logger.info(f"{Color.BLUE}Collecting information from server: {url}{Color.RESET}") async with aiohttp.ClientSession() as session: # Get server version (not mandatory, "version" endpoint may not exist) url_version = f"{url}/version" async with session.get(url_version) as response: if HTTPStatus(response.status) == HTTPStatus.OK: text = await response.text() logger.info(f"{Color.BLUE}Server version: {text}{Color.RESET}") # Get available models url_models = f"{url}/v1/models" async with session.get(url_models) as response: if HTTPStatus(response.status) == HTTPStatus.OK: text = await response.text() logger.info(f"{Color.BLUE}Models:{Color.RESET}") models_data = json.loads(text) models_list = models_data["data"] for model in models_list: model_id = model["id"] max_model_len = model.get("max_model_len", "N/A") logger.info( f"{Color.BLUE}\t{model_id=}, {max_model_len=}{Color.RESET}" ) else: logger.info(f"{Color.RED}Failed to get models{Color.RESET}") async def main() -> None: parser = argparse.ArgumentParser( prog="Benchmark serving with multi-turn conversations", description="Benchmark online inference using REST API", ) parser.add_argument("--version", action="version", version="%(prog)s 1.0") parser.add_argument( "-i", "--input-file", type=str, required=True, help="Input JSON file with ShareGPT conversations or " "configuration file for generation of synthetic conversations", ) parser.add_argument( "-o", "--output-file", type=str, default=None, help="Output JSON file containing conversations with updated assistant answers", ) parser.add_argument( "--seed", type=int, default=0, help="Seed for random number generators (default: 0)", ) parser.add_argument( "-m", "--model", type=str, required=True, help="Path of the LLM model" ) parser.add_argument( "--served-model-name", type=str, default=None, help="The model name used in the API. " "If not specified, the model name will be the " "same as the `--model` argument. ", ) parser.add_argument( "-u", "--url", type=str, default="http://localhost:8000", help="Base URL for the LLM API server", ) parser.add_argument( "-p", "--num-clients", type=int, default=1, help="Number of clients that will send requests in parallel", ) parser.add_argument( "-k", "--max-active-conversations", type=int, default=None, help="Max number of active conversations at a time (for all clients)", ) parser.add_argument( "-n", "--max-num-requests", type=int, default=None, help="Max number of requests to send (total for all clients)", ) parser.add_argument( "--warmup-step", default=False, action="store_true", help="Run a warmup step (using only the first turn of every conversation), " "measurements will not be included in the final benchmark results", ) parser.add_argument( "--max-turns", type=int, default=None, help="Maximum number of turns/messages per conversation, " "includes both user and assistant messages " "(a positive number, e.g: 2, 4, 6, etc.), disabled by default", ) parser.add_argument( "--no-early-stop", default=False, action="store_true", help="By default, the benchmark will stop if at least one client exits." " Use this flag to disable this behavior", ) parser.add_argument( "--limit-max-tokens", type=int, default=NUM_TOKENS_FROM_DATASET, help="Set max_tokens for the output token count of each request " "(must also set --limit-min-tokens). " "Overrides output token count from the input dataset. " "Use a negative value to disable this limit.", ) parser.add_argument( "--limit-min-tokens", type=int, default=NUM_TOKENS_FROM_DATASET, help="Set min_tokens for the output token count of each request " "(must also set --limit-max-tokens). " "Overrides output token count from the input dataset. " "Use a negative value to disable this limit.", ) parser.add_argument( "--request-rate", type=float, default=0, help="Expected request rate (Poisson process) per client in requests/sec." "Set to 0 for no delay between requests.", ) parser.add_argument( "--max-retries", type=int, default=int(os.environ.get("MULTITURN_BENCH_MAX_RETRIES", "0")), help="Maximum number of retry attempts for timed-out requests. " "Default is 0 (no retries). " "Set to higher values to retry failed requests and maintain " "fair workload distribution. " "Can also be set via MULTITURN_BENCH_MAX_RETRIES environment variable.", ) parser.add_argument( "--conversation-sampling", type=ConversationSampling, choices=list(ConversationSampling), default=ConversationSampling.ROUND_ROBIN, help=( "Strategy for selecting which conversation to use for the next request. " "Options: 'round_robin' (cycle through conversations), " "'random' (pick randomly)." ), ) parser.add_argument( "--verify-output", default=False, action="store_true", help="Verify the LLM output (compare to the answers in the input JSON file)", ) parser.add_argument( "--request-timeout-sec", type=int, default=120, help="Timeout in seconds for each API request (default: 120). " "Automatically increased if max tokens imply longer decoding.", ) parser.add_argument( "--no-stream", default=False, action="store_true", help="Disable stream/streaming mode (set 'stream' to False in the API request)", ) parser.add_argument( "-e", "--excel-output", default=False, action="store_true", help="Export summary to Excel file (optional)", ) parser.add_argument( "-v", "--verbose", default=False, action="store_true", help="Enable verbose output", ) parser.add_argument( "--print-content", default=False, action="store_true", help="Print the user prompts and the server's answers", ) parser.add_argument( "--warmup-percentages", type=str, default="0%", help="Ignore the first X samples as warmup (X is a percentage)." " A comma separated list of percentages can be used " "(for example: --warmup-percentages=0%%,50%%)", ) args = parser.parse_args() logger.info(args) logger.info(f"{Color.GREEN}Input parameters:{Color.RESET}") logger.info(f"url={args.url}") logger.info(f"model={args.model}") logger.info(f"num_clients={args.num_clients}") if args.verify_output: logger.info(f"{Color.PURPLE}Verify is enabled{Color.RESET}") # Calculate the amount of samples to filter (as warmup samples/measurements). try: warmup_percentages: list[float] = [0.0] if not args.warmup_step: # Warmup percentage can be used only if the warmup step was used warmup_strings: list[str] = args.warmup_percentages.split(",") warmup_strings = [x.replace("%", "") for x in warmup_strings] warmup_percentages = [float(x) / 100 for x in warmup_strings] # Check for valid range (0 to 1) for p in warmup_percentages: assert p >= 0.0 and p < 1.0 # Sort from high to low warmup percentage warmup_percentages.sort() logger.info( f"Warmup percentages (percentage of samples): {warmup_percentages}" ) except Exception: raise ValueError( f"Invalid --warmup-percentage={args.warmup_percentage}" ) from None # Set global seeds for main process random.seed(args.seed) np.random.seed(args.seed) logger.info("Loading tokenizer") tokenizer = AutoTokenizer.from_pretrained(args.model) await get_server_info(args.url) # Load the input file (either conversations of configuration file) logger.info(f"Reading input file: {args.input_file}") with open(args.input_file) as f: input_data = json.load(f) gen_conv_args = None if isinstance(input_data, list): # The conversations are stored as a list of dicts logger.info(f"Found {len(input_data)} items in the input file") # Convert the list to a ConversationsMap conversations = conversations_list_to_dict(input_data) elif isinstance(input_data, dict): # The input file is a configuration file # (type is determined by the field 'filetype') if "filetype" not in input_data: raise Exception( f"Input file {args.input_file} is invalid (missing 'filetype')" ) logger.info(f"Using input file with filetype: {input_data['filetype']}") gen_conv_args = parse_input_json_file(input_data) # Disable warning from "huggingface/tokenizers" # (when using python multiprocessing and tokenizers) os.environ["TOKENIZERS_PARALLELISM"] = "true" # Generate synthetic conversations conversations = generate_conversations(gen_conv_args, tokenizer) else: raise Exception(f"Input file {args.input_file} is invalid") if args.max_turns is not None: if args.max_turns < 1: raise ValueError("Max turns must be a positive number") logger.info( f"{Color.PURPLE}Max turns per conversation " f"is limited to {args.max_turns}{Color.RESET}" ) # Create benchmark configurations client_args, req_args = get_client_config(args, conversations) bench_args = BenchmarkArgs( url=args.url, num_clients=args.num_clients, early_stop=not args.no_early_stop ) warmup_runtime_sec: float | None = None # Warm-up step if args.warmup_step: # Only send a single user prompt from every conversation. # max_active_conversations must be 1, # otherwise the clients may exit after sending a single request # (because the task queue is empty). warmup_client_args = client_args._replace( skip_first_turn=False, max_turns=1, max_active_conversations=1 ) # Early stop should be disabled, # all clients should finish their work before exiting warmup_bench_args = bench_args._replace(early_stop=False) logger.info("%sWarmup start%s", Color.PURPLE, Color.RESET) warmup_start_ns = time.perf_counter_ns() conversations, _ = await main_mp( warmup_client_args, req_args, warmup_bench_args, tokenizer, conversations ) warmup_runtime_sec = nanosec_to_sec(time.perf_counter_ns() - warmup_start_ns) logger.info( "%sWarmup runtime: %.3f sec (%.3f ms)%s", Color.PURPLE, warmup_runtime_sec, warmup_runtime_sec * 1000, Color.RESET, ) logger.info("%sWarmup done%s", Color.PURPLE, Color.RESET) # Run the benchmark benchmark_start_ns = time.perf_counter_ns() client_convs, client_metrics = await main_mp( client_args, req_args, bench_args, tokenizer, conversations ) benchmark_runtime_sec = nanosec_to_sec(time.perf_counter_ns() - benchmark_start_ns) # Calculate requests per second requests_per_sec = len(client_metrics) / benchmark_runtime_sec benchmark_runtime_ms = benchmark_runtime_sec * 1000.0 logger.info( "%sAll clients finished, benchmark runtime: %.3f sec (%.3f ms), " "requests per second: %.3f%s", Color.GREEN, benchmark_runtime_sec, benchmark_runtime_ms, requests_per_sec, Color.RESET, ) if warmup_runtime_sec is not None: total_runtime_sec = benchmark_runtime_sec + warmup_runtime_sec logger.info( "%sWarmup runtime: %.3f sec (%.3f ms)%s", Color.GREEN, warmup_runtime_sec, warmup_runtime_sec * 1000, Color.RESET, ) logger.info( "%sTotal runtime (including warmup): %.3f sec (%.3f ms)%s", Color.GREEN, total_runtime_sec, total_runtime_sec * 1000, Color.RESET, ) # Benchmark parameters params = { "model": args.model, "num_clients": args.num_clients, "num_conversations": len(conversations), "active_conversations": args.max_active_conversations, "seed": args.seed, } if args.limit_min_tokens > 0: params["min_tokens"] = args.limit_min_tokens if args.limit_max_tokens > 0: params["max_tokens"] = args.limit_max_tokens # Process and print statistics (and save excel file with the statistics) process_statistics( client_metrics, test_params=params, warmup_percentages=warmup_percentages, verbose=args.verbose, gen_conv_args=gen_conv_args, excel_output=args.excel_output, warmup_runtime_sec=warmup_runtime_sec, ) if args.output_file is not None: # Write a JSON file with the updated conversations # The "assistant" content will contain the answers from the tested LLM output_data: ShareGptConversations = conversations_dict_to_list(client_convs) logger.info( f"{Color.GREEN}Writing conversations file: {args.output_file}{Color.RESET}" ) with open(args.output_file, "w") as f: json.dump(output_data, f, indent=4) if __name__ == "__main__": asyncio.run(main())
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/multi_turn/benchmark_serving_multi_turn.py", "license": "Apache License 2.0", "lines": 1400, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:benchmarks/multi_turn/convert_sharegpt_to_openai.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Download dataset from: https://huggingface.co/datasets/philschmid/sharegpt-raw/blob/main/sharegpt_20230401_clean_lang_split.json Convert to OpenAI API: export INPUT_FILE=sharegpt_20230401_clean_lang_split.json python convert_sharegpt_to_openai.py $INPUT_FILE sharegpt_conv_128.json --max-items=128 """ import argparse import json import random from statistics import mean from typing import Any import pandas as pd # type: ignore import tqdm # type: ignore from transformers import AutoTokenizer # type: ignore def has_non_english_chars(text: str) -> bool: return not text.isascii() def content_is_valid( content: str, min_content_len: int | None, max_content_len: int | None ) -> bool: if min_content_len and len(content) < min_content_len: return False if max_content_len and len(content) > max_content_len: return False return has_non_english_chars(content) def print_stats( conversations: "list[dict[Any, Any]]", tokenizer: AutoTokenizer | None = None ) -> None: # Collect statistics stats = [] print("\nCollecting statistics...") for item in tqdm.tqdm(conversations): # item has "id" and "messages" messages = item["messages"] user_turns = 0 assistant_turns = 0 user_words = 0 assistant_words = 0 conv_chars = 0 user_tokens: list[int] = [] assistant_tokens: list[int] = [] for m in messages: content = m["content"] conv_chars += len(content) content_num_words = content.count(" ") + 1 num_tokens = 0 if tokenizer: num_tokens = len(tokenizer(m["content"]).input_ids) if m["role"] == "user": user_turns += 1 user_words += content_num_words if tokenizer: user_tokens.append(num_tokens) elif m["role"] == "assistant": assistant_turns += 1 assistant_words += content_num_words if tokenizer: assistant_tokens.append(num_tokens) # assert user_turns == assistant_turns, \ # f"Invalid conversation ID {item['id']}" conv_words = user_words + assistant_words item_stats = { "user_turns": user_turns, "assistant_turns": assistant_turns, "user_words": user_words, "assistant_words": assistant_words, "conv_turns": len(messages), "conv_words": conv_words, "conv_characters": conv_chars, } if len(user_tokens) > 0: item_stats["user_tokens"] = int(mean(user_tokens)) if len(assistant_tokens) > 0: item_stats["assistant_tokens"] = int(mean(assistant_tokens)) stats.append(item_stats) print("\nStatistics:") percentiles = [0.25, 0.5, 0.75, 0.9, 0.99, 0.999, 0.9999] df = pd.DataFrame(stats) print(df.describe(percentiles=percentiles).transpose()) def convert_sharegpt_to_openai( seed: int, input_file: str, output_file: str, max_items: int | None, min_content_len: int | None = None, max_content_len: int | None = None, min_turns: int | None = None, max_turns: int | None = None, model: str | None = None, ) -> None: if min_turns and max_turns: assert min_turns <= max_turns if min_content_len and max_content_len: # Verify that min is not larger than max if both were given assert min_content_len <= max_content_len print( f"Input parameters:\n{seed=}, {max_items=}, {min_content_len=}," f" {max_content_len=}, {min_turns=}, {max_turns=}\n" ) random.seed(seed) tokenizer = None if model is not None: print(f"Loading tokenizer from: {model}") tokenizer = AutoTokenizer.from_pretrained(model) # Read the ShareGPT JSON file print(f"Reading file: {input_file}") with open(input_file, encoding="utf-8") as f: # Should be a list of dicts # Each dict should have "id" (string) and "conversations" (list of dicts) sharegpt_data = json.load(f) assert isinstance(sharegpt_data, list), "Input file should contain a list of dicts" print(f"Total items in input file: {len(sharegpt_data):,}") print(f"Shuffling dataset with seed {seed}") random.shuffle(sharegpt_data) # Map conversation ID to the all the messages conversation_parts: dict[str, list[Any]] = {} for item in tqdm.tqdm(sharegpt_data): assert "id" in item, "Missing key 'id'" assert "conversations" in item, "Missing key 'conversations'" # Conversation ID (e.g: "hiWPlMD") and part/session (0, 1, 2, etc.) conv_id, _ = item["id"].split("_") new_turns = item["conversations"] if conv_id not in conversation_parts: # Start new conversation conversation_parts[conv_id] = [] elif len(conversation_parts[conv_id]) > 0 and len(new_turns) > 0: prev_turns = conversation_parts[conv_id][-1] if prev_turns[-1]["from"] == new_turns[0]["from"]: new_turns = new_turns[1:] if len(new_turns) > 0: # We assume that parts are in order in the ShareGPT dataset conversation_parts[conv_id].append(new_turns) dataset: list[dict[str, Any]] = [] for conv_id, conv_parts in conversation_parts.items(): new_item = {"id": conv_id} conversations: list[dict[str, str]] = [] # Merge all parts for conv_part in conv_parts: conversations.extend(conv_part) if len(conversations) > 0: new_item["conversations"] = conversations dataset.append(new_item) print(f"Total unique conversations (IDs) in input file: {len(dataset):,}") # Final output data final_openai_dataset: list[dict] = [] # Filter conversations from the ShareGPT dataset and convert to OpenAI format for item in tqdm.tqdm(dataset): messages: list[dict] = [] assert "id" in item, "Missing key 'id'" assert "conversations" in item, "Missing key 'conversations'" conv_id = item["id"] conversations = item["conversations"] if min_turns is not None and len(conversations) < min_turns: # Skip short conversations continue # Convert each message in the conversation, up to max_turns if specified for i, turn in enumerate(conversations): assert "from" in turn and "value" in turn, ( f"Invalid conversation ID {conv_id} - missing 'from' or 'value'" ) role = None turn_from = turn["from"] if turn_from in {"human", "user"}: role = "user" elif turn_from in {"gpt", "bing", "chatgpt", "bard"}: role = "assistant" elif turn_from == "system": role = "system" assert role is not None, ( f"Invalid conversation ID {conv_id} - 'from'='{turn_from}' is invalid" ) if i == 0 and role != "user": # If the first message is from assistant (gpt), skip it. # this happens when the conversation is a follow-up # to a previous conversation (from the same user). continue if max_turns is not None and i >= max_turns: break # Convert message to OpenAI format (with "role" and "content") content = turn["value"] messages.append({"role": role, "content": content}) # Add the converted conversation to the OpenAI format if len(messages) > 0: valid_messages = True # First turn should always be from the user user_turn = True for m in messages: # Make sure that turns alternate between user and assistant if (user_turn and m["role"] != "user") or ( not user_turn and m["role"] != "assistant" ): valid_messages = False break user_turn = not user_turn content = m["content"] valid_messages = content_is_valid( content, min_content_len, max_content_len ) if not valid_messages: break if valid_messages is True: final_openai_dataset.append({"id": conv_id, "messages": messages}) assert len(final_openai_dataset) > 0, "Final number of conversations is zero" print_stats(final_openai_dataset) print_stats_again = False if max_items is not None and len(final_openai_dataset) > max_items: print(f"\n\nSampling {max_items} items from the dataset...") print_stats_again = True final_openai_dataset = random.sample(final_openai_dataset, max_items) if print_stats_again: # Print stats after the dataset changed print_stats(final_openai_dataset, tokenizer) # Write the converted data to a new JSON file final_size = len(final_openai_dataset) print(f"\nTotal conversations converted (after filtering): {final_size:,}") print(f"\nWriting file: {output_file}") with open(output_file, "w", encoding="utf-8") as f: json.dump(final_openai_dataset, f, ensure_ascii=False, indent=2) def main() -> None: parser = argparse.ArgumentParser( description="Convert ShareGPT dataset to OpenAI API format" ) parser.add_argument("input_file", help="Path to the input ShareGPT JSON file") parser.add_argument( "output_file", help="Path to the output OpenAI format JSON file" ) parser.add_argument( "--seed", type=int, default=0, help="Seed for random number generators" ) parser.add_argument( "--max-items", type=int, default=None, help="Maximum number of items in the output file", ) parser.add_argument( "--min-turns", type=int, default=None, help="Minimum number of turns per conversation", ) parser.add_argument( "--max-turns", type=int, default=None, help="Maximum number of turns per conversation", ) parser.add_argument( "--min-content-len", type=int, default=None, help="Min number of characters in the messages' content", ) parser.add_argument( "--max-content-len", type=int, default=None, help="Max number of characters in the messages' content", ) parser.add_argument( "--model", type=str, default=None, help="LLM model, only the tokenizer will be used", ) args = parser.parse_args() convert_sharegpt_to_openai( args.seed, args.input_file, args.output_file, args.max_items, args.min_content_len, args.max_content_len, args.min_turns, args.max_turns, args.model, ) if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/multi_turn/convert_sharegpt_to_openai.py", "license": "Apache License 2.0", "lines": 281, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/moe/test_gpt_oss_triton_kernels.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass, fields import pytest import torch import torch.nn.functional as F from vllm.utils.import_utils import has_triton_kernels if not has_triton_kernels(): pytest.skip( "triton_kernels not found, skipping all related tests", allow_module_level=True, ) import triton_kernels.swiglu from triton_kernels.matmul_ogs import FlexCtx, PrecisionConfig from triton_kernels.numerics import InFlexData from triton_kernels.numerics_details.mxfp import downcast_to_mxfp, upcast_from_mxfp from triton_kernels.tensor import FP4, convert_layout, wrap_torch_tensor from triton_kernels.tensor_details import layout from triton_kernels.testing import assert_close from vllm.model_executor.layers.fused_moe.config import mxfp4_w4a16_moe_quant_config from vllm.model_executor.layers.fused_moe.gpt_oss_triton_kernels_moe import ( triton_kernel_moe_forward, ) from vllm.model_executor.layers.utils import shuffle_weight from vllm.utils.math_utils import round_up def deshuffle(w: torch.Tensor): first = w[..., ::2] second = w[..., 1::2] deshuffled = torch.concat((first, second), dim=-1) return deshuffled def init_compute_data(M, K, N, E, a_dtype: str, w_dtype: str, num_warps: int): randbits = [torch.randperm(E) for _ in range(M)] x_list = [ (-1) ** i * ((16384 + ((i * 512) % 4096) + bits).to(torch.int16).view(torch.bfloat16)) for i, bits in enumerate(randbits) ] exp_data = torch.stack(x_list).to(device="cuda") # simulating gate_output (M, E) # create input tensor x = torch.randn((M, K), dtype=torch.bfloat16, device="cuda") w1 = torch.randn((E, 2 * N, K), dtype=torch.bfloat16, device="cuda") w1_bias = torch.randn((E, 2 * N), dtype=torch.bfloat16, device="cuda") w2 = torch.randn((E, K, N), dtype=torch.bfloat16, device="cuda") w2_bias = torch.randn((E, K), dtype=torch.bfloat16, device="cuda") exp_data_tri = exp_data.clone() x_tri = x.clone() w1_tri = w1.clone() w2_tri = w2.clone() w1_bias_tri = w1_bias.clone() w2_bias_tri = w2_bias.clone() w1_bias_tri = w1_bias_tri.to(torch.float32) w2_bias_tri = w2_bias_tri.to(torch.float32) dtype_dict = { "bf16": torch.bfloat16, "fp8_e4m3": torch.float8_e4m3fn, "fp8_e5m2": torch.float8_e5m2, } x = x.to(dtype_dict[a_dtype]).to(torch.bfloat16) if w_dtype != "mx4": # simulate quantization support on reference impl w1 = w1.to(dtype_dict[w_dtype]).to(torch.bfloat16) w2 = w2.to(dtype_dict[w_dtype]).to(torch.bfloat16) # triton moe kernel use transposed shape for matmul w1_tri = w1_tri.transpose(-2, -1) w2_tri = w2_tri.transpose(-2, -1) # shuffle weights w1_tri = shuffle_weight(w1_tri) w1_bias_tri = shuffle_weight(w1_bias_tri) # quant triton_weights x_tri = x.to(dtype_dict[a_dtype]) if w_dtype != "mx4": pytest.skip("NYI") else: # quantize to mx4 # careful on the padding here, the activation padding need to be # multiple of 64, the actual engine is not implemented w1_bottom_pad = round_up(w1_tri.shape[1], 64) - w1_tri.shape[1] w1_right_pad = round_up(w1_tri.shape[2], 128) - w1_tri.shape[2] w2_bottom_pad = w1_right_pad // 2 w2_right_pad = w1_bottom_pad x_pad = w1_bottom_pad w1_tri = F.pad( w1_tri, (0, w1_right_pad, 0, w1_bottom_pad, 0, 0), mode="constant", value=0, ) w2_tri = F.pad( w2_tri, (0, w2_right_pad, 0, w2_bottom_pad, 0, 0), mode="constant", value=0, ) w1_bias_tri = F.pad( w1_bias_tri, (0, w1_right_pad, 0, 0), mode="constant", value=0 ) w2_bias_tri = F.pad( w2_bias_tri, (0, w2_right_pad, 0, 0), mode="constant", value=0 ) x_tri = F.pad(x_tri, (0, x_pad, 0, 0), mode="constant", value=0) w_layout, w_layout_opts = layout.make_default_matmul_mxfp4_w_layout(mx_axis=1) w_scale_layout, w_scale_layout_opts = ( layout.make_default_matmul_mxfp4_w_scale_layout( mx_axis=1, num_warps=num_warps ) ) w1_tri, w1_scale_tri = downcast_to_mxfp(w1_tri, torch.uint8, axis=1) w1 = upcast_from_mxfp(w1_tri, w1_scale_tri, torch.bfloat16, axis=1) w2_tri, w2_scale_tri = downcast_to_mxfp(w2_tri, torch.uint8, axis=1) w2 = upcast_from_mxfp(w2_tri, w2_scale_tri, torch.bfloat16, axis=1) w1_tri = convert_layout( wrap_torch_tensor(w1_tri, FP4), w_layout, **w_layout_opts ) w1_scale_tri = convert_layout( wrap_torch_tensor(w1_scale_tri), w_scale_layout, **w_scale_layout_opts, ) w2_tri = convert_layout( wrap_torch_tensor(w2_tri, FP4), w_layout, **w_layout_opts ) w2_scale_tri = convert_layout( wrap_torch_tensor(w2_scale_tri), w_scale_layout, **w_scale_layout_opts, ) pc1 = PrecisionConfig( weight_scale=w1_scale_tri, flex_ctx=FlexCtx(rhs_data=InFlexData()) ) pc2 = PrecisionConfig( weight_scale=w2_scale_tri, flex_ctx=FlexCtx(rhs_data=InFlexData()) ) # tucuate so the rest can run properly w1 = w1[..., :K, : 2 * N] w2 = w2[..., :N, :K] w1 = deshuffle(w1) w1 = w1.transpose(-1, -2).contiguous() w2 = w2.transpose(-1, -2).contiguous() return ( x, w1, w1_bias, w2, w2_bias, exp_data, x_tri, w1_tri, w2_tri, exp_data_tri, w1_bias_tri, w2_bias_tri, pc1, pc2, ) @dataclass class ModelConfig: num_hidden_layers: int = 36 num_experts: int = 128 experts_per_token: int = 4 vocab_size: int = 201088 hidden_size: int = 2880 intermediate_size: int = 2880 head_dim: int = 64 num_attention_heads: int = 64 num_key_value_heads: int = 8 sliding_window: int = 128 initial_context_length: int = 4096 rope_theta: float = 150000.0 rope_parameters_factor: float = 32.0 rope_ntk_alpha: float = 1.0 rope_ntk_beta: float = 32.0 def swiglu(x, alpha: float = 1.702, limit: float = 1.0): # Note we add an extra bias of 1 to the linear layer x_glu, x_linear = torch.chunk(x, 2, dim=-1) if limit is not None: x_glu = x_glu.clamp(max=limit) out_glu = x_glu * torch.sigmoid(alpha * x_glu) if limit is not None: x_linear = x_linear.clamp(min=-limit, max=limit) return out_glu * (x_linear + 1) def oai_moe_forward( hidden_states: torch.Tensor, # (M, K) w1: torch.Tensor, # (E, 2N) w1_bias: torch.Tensor, # (E, 2N, K) w2: torch.Tensor, # (E, K, N) w2_bias: torch.Tensor, # (E, N) gating_output: torch.Tensor, # (M, E) topk: int, ): # model.py 309:330, assuming gating and norm t = hidden_states experts = torch.topk(gating_output, k=topk, dim=-1, sorted=True) expert_weights = torch.nn.functional.softmax(experts.values, dim=1) expert_indices = experts.indices # MLP #1 mlp1_weight = w1[expert_indices, ...] mlp1_bias = w1_bias[expert_indices, ...] t = torch.einsum("beck,bk->bec", mlp1_weight, t) + mlp1_bias t = swiglu(t, limit=7) # MLP #2 mlp2_weight = w2[expert_indices, ...] mlp2_bias = w2_bias[expert_indices, ...] t = torch.einsum("beck,bek->bec", mlp2_weight, t) t += mlp2_bias # Weighted sum of experts t = torch.einsum("bec,be->bc", t, expert_weights) return t @dataclass class Case: a_dtype: str w_dtype: str @pytest.mark.parametrize( ", ".join(f.name for f in fields(Case)), [ tuple(getattr(case, f.name) for f in fields(Case)) for case in [ # Case(a_dtype="bf16", w_dtype="bf16"), # Case(a_dtype="fp8_e4m3", w_dtype="fp8_e5m2"), Case(a_dtype="bf16", w_dtype="mx4") ] ], ) @pytest.mark.parametrize("num_token", [2]) @pytest.mark.parametrize("tp", [1, 2, 4, 8]) def test_equiv(num_token, a_dtype, w_dtype, tp, workspace_init): from triton_kernels.tensor_details import layout if not hasattr(layout, "make_default_matmul_mxfp4_w_layout"): pytest.skip("make_default_matmul_mxfp4_w_layout not available") M = num_token E = ModelConfig.num_experts K = ModelConfig.hidden_size N = ModelConfig.intermediate_size // tp topk = ModelConfig.experts_per_token ( x, w1, w1_bias, w2, w2_bias, exp_data, x_tri, w1_tri, w2_tri, exp_data_tri, w1_bias_tri, w2_bias_tri, pc1, pc2, ) = init_compute_data(M, K, N, E, a_dtype, w_dtype, num_warps=8) if a_dtype == "bf16" and w_dtype == "mx4": quant_config = mxfp4_w4a16_moe_quant_config( w1_scale=pc1, w2_scale=pc2, w1_bias=w1_bias_tri, w2_bias=w2_bias_tri, ) else: raise NotImplementedError( f"Quantization configuration for activation={a_dtype} and weight={w_dtype} " f"has not been implemented." ) out_triton_monolithic = triton_kernel_moe_forward( hidden_states=x_tri, w1=w1_tri, w2=w2_tri, gating_output=exp_data_tri, topk=topk, renormalize=True, quant_config=quant_config, ) out_triton_monolithic = out_triton_monolithic[..., :K] out_ref = oai_moe_forward( hidden_states=x, w1=w1, w1_bias=w1_bias, w2=w2, w2_bias=w2_bias, gating_output=exp_data, topk=topk, ) assert_close(ref=out_ref, tri=out_triton_monolithic, maxtol=0.025, rmstol=0.005) def test_unit_shuffle(): N = ModelConfig.intermediate_size K = ModelConfig.hidden_size m = torch.randn((K, 2 * N), dtype=torch.bfloat16, device="cuda") x = torch.randn(K, dtype=torch.bfloat16, device="cuda") m_shuffled = shuffle_weight(m) out_ref = x @ m out_ref = swiglu(out_ref, limit=1.0) out = x @ m_shuffled out = triton_kernels.swiglu.swiglu_torch( out, alpha=1.702, precision_config=triton_kernels.swiglu.PrecisionConfig(limit=1.0), ) assert_close(ref=out_ref, tri=out)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/moe/test_gpt_oss_triton_kernels.py", "license": "Apache License 2.0", "lines": 295, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/layers/fused_moe/gpt_oss_triton_kernels_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm import _custom_ops as ops from vllm._aiter_ops import rocm_aiter_ops from vllm.logger import init_logger from vllm.model_executor.layers.fused_moe.activation import MoEActivation from vllm.model_executor.layers.fused_moe.config import ( FUSED_MOE_UNQUANTIZED_CONFIG, FusedMoEParallelConfig, FusedMoEQuantConfig, ) from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceNoOP, ) from vllm.model_executor.layers.fused_moe.utils import _resize_cache from vllm.model_executor.layers.quantization.utils.quant_utils import ( QuantKey, ) from vllm.platforms import current_platform from vllm.triton_utils import tl, triton from vllm.utils.import_utils import has_triton_kernels logger = init_logger(__name__) use_legacy_triton_kernels = False if has_triton_kernels(): try: import triton_kernels.swiglu from triton_kernels.matmul_ogs import ( FnSpecs, FusedActivation, GatherIndx, RoutingData, ScatterIndx, matmul_ogs, ) from triton_kernels.tensor import ( BIT, Bitmatrix, ) from triton_kernels.topk import topk try: from triton_kernels.tensor import ( SparseMatrix, make_ragged_tensor_metadata, ) except ImportError: if current_platform.is_rocm(): logger.warning_once("Using legacy triton_kernels on ROCm") use_legacy_triton_kernels = True else: raise except (AttributeError, ImportError) as e: logger.error( "Failed to import Triton kernels. Please make sure your triton " "version is compatible. Error: %s", e, ) @triton.jit def pack_bitmatrix( bitmatrix, topk_ids, n_rows, # n_rows in bitmatrix / topk_ids bm_cols: tl.constexpr, # n int32_t bitpacks in bitmatrix n_expts_act, # num_topk BLOCK_SIZE_M: tl.constexpr, BLOCK_SIZE_K: tl.constexpr, ): """ Packs topk_ids into a bitmatrix. code reference: https://github.com/triton-lang/triton/blob/dd1bbc52b34d202dfe5ffea1e04fb16166c5c04e/python/triton_kernels/bench/distributed.py#L264 """ pid_m = tl.program_id(0) offsets_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offsets_k = tl.arange(0, BLOCK_SIZE_K) offsets = offsets_m[:, None] * n_expts_act + offsets_k[None, :] mask = (offsets_m < n_rows)[:, None] & (offsets_k < n_expts_act)[None, :] indices = tl.load(topk_ids + offsets, mask=mask, other=-1) div = indices // 32 rem = indices % 32 one = tl.cast(1, tl.uint32) # Iterate through all the relevant bitmatrix columns. for i in range(bm_cols): # When BLOCK_SIZE_K=32, offs is just the column index. offs = tl.arange(0, BLOCK_SIZE_K // 32) + i * (BLOCK_SIZE_K // 32) # All topks that need to go into this column has the correct bit set. # Other bits are 0. x is a 2D tensor. x = tl.where( div[:, :, None] == offs[None, None, :], (one << rem)[:, :, None], 0 ) # Reduce x to get a single int32_t bitpack. y = tl.reduce_or(x, axis=1) bitmatrix_ptrs = bitmatrix + offsets_m[:, None] * bm_cols + offs[None, :] tl.store(bitmatrix_ptrs, y, mask=offsets_m[:, None] < n_rows) def legacy_routing_from_bitmatrix( bitmatrix: "Bitmatrix", expt_scal: torch.Tensor, expt_indx: torch.Tensor, n_expts_tot: int, n_expts_act: int, ) -> tuple["RoutingData", "GatherIndx", "ScatterIndx"]: """ Replacement for the removed triton_kernels.routing.routing_from_bitmatrix. Creates routing data from a bitmatrix representation. """ if use_legacy_triton_kernels: from triton_kernels.routing import routing_from_bitmatrix return routing_from_bitmatrix( bitmatrix, expt_scal, expt_indx, n_expts_tot, n_expts_act ) sparse_logits = SparseMatrix(indx=expt_indx, vals=expt_scal, mask=bitmatrix) dispatch_indx = sparse_logits.mask_metadata.row_sorted_indx combine_indx = sparse_logits.mask_metadata.col_sorted_indx ragged_batch_metadata = make_ragged_tensor_metadata( sparse_logits.mask_metadata.col_sum, dispatch_indx.shape[0], ) gate_scal = sparse_logits.vals.flatten()[combine_indx] routing_data = RoutingData( gate_scal, ragged_batch_metadata.block_sizes, n_expts_tot, n_expts_act, ragged_batch_metadata, ) gather_idx = GatherIndx(combine_indx, dispatch_indx) scatter_idx = ScatterIndx(dispatch_indx, combine_indx) return routing_data, gather_idx, scatter_idx def legacy_routing( logits: torch.Tensor, n_expts_act: int, sm_first: bool = False, ) -> tuple["RoutingData", "GatherIndx", "ScatterIndx"]: """ Replacement for the removed triton_kernels.routing.routing function. Computes routing data from gating logits. """ if use_legacy_triton_kernels: from triton_kernels.routing import routing return routing(logits, n_expts_act, sm_first=sm_first) if sm_first: logits = torch.softmax(logits, dim=-1) sparse_logits = topk(logits, n_expts_act, apply_softmax=not sm_first) return legacy_routing_from_bitmatrix( sparse_logits.mask, sparse_logits.vals, sparse_logits.indx, logits.shape[-1], n_expts_act, ) def triton_kernel_moe_forward( hidden_states: torch.Tensor, w1, # Tensor or triton_kernels.Tensor w2, # Tensor or triton_kernels.Tensor gating_output: torch.Tensor, topk: int, renormalize: bool, activation: MoEActivation = MoEActivation.SWIGLUOAI, quant_config: FusedMoEQuantConfig | None = None, apply_router_weight_on_input: bool = False, global_num_experts: int = -1, expert_map: torch.Tensor | None = None, unpadded_N_w1=None, unpadded_K_w1=None, unpadded_N_w2=None, unpadded_K_w2=None, ) -> torch.Tensor: if ( quant_config is not None and quant_config.use_mxfp4_w4a8 and rocm_aiter_ops.is_enabled() ): from aiter.ops.triton.moe_routing.routing import routing as aiter_routing routing_data, gather_idx, scatter_idx = aiter_routing( gating_output, topk, sm_first=not renormalize ) return triton_kernel_fused_mxfp4_w4a8_experts( None, hidden_states, w1, w2, routing_data, gather_idx, scatter_idx, activation=activation.value, quant_config=quant_config, apply_router_weight_on_input=apply_router_weight_on_input, global_num_experts=global_num_experts, expert_map=expert_map, unpadded_N_w1=unpadded_N_w1, unpadded_K_w1=unpadded_K_w1, unpadded_N_w2=unpadded_N_w2, unpadded_K_w2=unpadded_K_w2, ) if expert_map is not None: # With expert parallelism, legacy_routing produces routing data # using global expert IDs which don't correspond to local weight # indices. Split the routing into topk selection + expert_map # remapping + local routing data construction (matching the # approach used by OAITritonExperts.apply). from triton_kernels.topk import topk as topk_fn sm_first = not renormalize logits = gating_output if sm_first: logits = torch.softmax(logits, dim=-1) sparse_logits = topk_fn(logits, topk, apply_softmax=not sm_first) # sparse_logits.indx contains global expert IDs – remap to local. topk_ids = expert_map[sparse_logits.indx.to(torch.long)] topk_weights = sparse_logits.vals local_num_experts = w1.size(0) routing_data, gather_idx, scatter_idx = make_routing_data( topk_ids, topk_weights, local_num_experts ) # expert_map already applied; pass None downstream. effective_expert_map = None effective_global_num_experts = local_num_experts else: routing_data, gather_idx, scatter_idx = legacy_routing( gating_output, topk, sm_first=not renormalize ) effective_expert_map = expert_map effective_global_num_experts = global_num_experts output = torch.empty_like(hidden_states) effective_quant_config = ( quant_config if quant_config is not None else FUSED_MOE_UNQUANTIZED_CONFIG ) return triton_kernel_fused_experts( output, hidden_states, w1, w2, routing_data, gather_idx, scatter_idx, topk=topk, activation=activation, quant_config=effective_quant_config, apply_router_weight_on_input=apply_router_weight_on_input, global_num_experts=effective_global_num_experts, expert_map=effective_expert_map, ) # This is a triton implementation of the fused_experts function def triton_kernel_fused_experts( output_tensor: torch.Tensor, hidden_states: torch.Tensor, w1, # Tensor or triton_kernels.Tensor w2, # Tensor or triton_kernels.Tensor routing_data, # RoutingData gather_indx, # GatherIndx scatter_indx, # ScatterIndx topk: int, activation: MoEActivation = MoEActivation.SWIGLUOAI, quant_config: FusedMoEQuantConfig | None = None, swiglu_alpha: float = 1.702, swiglu_limit: float = 7.0, apply_router_weight_on_input: bool = False, global_num_experts: int = -1, expert_map: torch.Tensor | None = None, intermediate_cache: torch.Tensor | None = None, a1q_scale: torch.Tensor | None = None, ) -> torch.Tensor: """Triton implementation of fused expert computation using OAI kernels.""" assert activation == MoEActivation.SWIGLUOAI, ( "Only SWIGLUOAI activation is supported" ) assert quant_config is not None # type check, uint8 means mxfp4 assert hidden_states.dtype == torch.bfloat16 assert quant_config.w1_bias is None or quant_config.w1_bias.dtype == torch.float32 assert quant_config.w2_bias is None or quant_config.w2_bias.dtype == torch.float32 # Shape check, only check non-mxfp4 assert hidden_states.ndim == 2 assert hidden_states.shape[-1] == w1.shape[-2] assert w2.shape[-1] == w1.shape[1] batch_dim = 1 M, K = hidden_states.shape[-2:] E, _, N = w1.shape if global_num_experts == -1: global_num_experts = E if intermediate_cache is None: intermediate_cache = torch.empty( (batch_dim, M * topk, N // 2), device=hidden_states.device, dtype=hidden_states.dtype, ) # Add batch_dim to output buffer because matmul_ogs expects 3D output intermediate_cache = _resize_cache( intermediate_cache, (batch_dim, M * topk, N // 2) ) output_tensor = _resize_cache(output_tensor, (batch_dim, M, K)) act = ( FusedActivation( FnSpecs( "swiglu", triton_kernels.swiglu.swiglu_fn, ("alpha", "limit"), reduction_n=2, ), (swiglu_alpha, swiglu_limit), ) if not use_legacy_triton_kernels else FusedActivation( FnSpecs("swiglu", triton_kernels.swiglu.swiglu_fn, ("alpha", "limit")), (swiglu_alpha, swiglu_limit), 2, ) ) gammas = routing_data.gate_scal if routing_data else None matmul_ogs( hidden_states, w1, quant_config.w1_bias, routing_data, gather_indx=gather_indx, precision_config=quant_config.w1_precision, gammas=gammas if apply_router_weight_on_input else None, fused_activation=act, y=intermediate_cache, ) matmul_ogs( intermediate_cache.view(M * topk, N // 2), w2, quant_config.w2_bias, routing_data, scatter_indx=scatter_indx, precision_config=quant_config.w2_precision, gammas=None if apply_router_weight_on_input else gammas, y=output_tensor, ) output_tensor = output_tensor.view(M, K) return output_tensor # This is a triton implementation of the fused_experts function def triton_kernel_fused_mxfp4_w4a8_experts( output_tensor: torch.Tensor, hidden_states: torch.Tensor, w1, # Tensor or triton_kernels.Tensor w2, # Tensor or triton_kernels.Tensor routing_data, # RoutingData gather_indx, # GatherIndx scatter_indx, # ScatterIndx activation: str = "silu", quant_config: FusedMoEQuantConfig | None = None, swiglu_alpha: float = 1.702, swiglu_limit: float = 7.0, apply_router_weight_on_input: bool = False, global_num_experts: int = -1, expert_map: torch.Tensor | None = None, a1q_scale: torch.Tensor | None = None, unpadded_N_w1=None, unpadded_K_w1=None, unpadded_N_w2=None, unpadded_K_w2=None, ) -> torch.Tensor: assert quant_config is not None # type check, uint8 means mxfp4 assert hidden_states.dtype == torch.bfloat16 assert quant_config.w1_bias is None or quant_config.w1_bias.dtype == torch.float32 assert quant_config.w2_bias is None or quant_config.w2_bias.dtype == torch.float32 # Shape check, only check non-mxfp4 assert hidden_states.shape[-1] == w1.shape[-2] assert w2.shape[-1] == w1.shape[1] E, _, N = w1.shape if global_num_experts == -1: global_num_experts = E gammas = routing_data.gate_scal if routing_data else None from aiter.ops.triton.moe_op_gemm_a8w4 import moe_gemm_a8w4 from aiter.ops.triton.quant_moe import downcast_to_static_fp8 assert quant_config.w1_precision is not None, ( "w1_precision in quant config can't be None" ) assert quant_config.w2_precision is not None, ( "w2_precision in quant config can't be None" ) hidden_states = downcast_to_static_fp8( hidden_states, quant_config.w1_precision.flex_ctx.lhs_data.scale ) intermediate_cache1 = moe_gemm_a8w4( hidden_states, w1.storage.data, None, quant_config.w1_precision.weight_scale.storage.data, quant_config.w1_precision.flex_ctx.lhs_data.scale, quant_config.w2_precision.flex_ctx.lhs_data.scale, quant_config.w1_bias, routing_data, gather_indx=gather_indx, gammas=gammas if apply_router_weight_on_input else None, swizzle_mx_scale="CDNA4_SCALE", out_dtype=torch.float8_e4m3fn, apply_swiglu=True, alpha=swiglu_alpha, limit=swiglu_limit, unpadded_N=unpadded_N_w1, unpadded_K=unpadded_K_w1, ) intermediate_cache3 = moe_gemm_a8w4( intermediate_cache1, w2.storage.data, None, quant_config.w2_precision.weight_scale.storage.data, quant_config.w2_precision.flex_ctx.lhs_data.scale, None, quant_config.w2_bias, routing_data, scatter_indx=scatter_indx, gammas=None if apply_router_weight_on_input else gammas, swizzle_mx_scale="CDNA4_SCALE", unpadded_N=unpadded_N_w2, unpadded_K=unpadded_K_w2, ) return intermediate_cache3 def make_routing_data( topk_ids: torch.Tensor, topk_weights: torch.Tensor, num_local_experts: int, ) -> tuple["RoutingData", torch.Tensor, torch.Tensor]: topk_ids = topk_ids.to(torch.int16) topk_weights = topk_weights.to(torch.bfloat16) n_rows, num_topk = topk_ids.size() BLOCK_SIZE_M = 512 BLOCK_SIZE_K = 32 bm_cols = triton.cdiv(num_local_experts, BLOCK_SIZE_K) # n_bitpacks bitmatrix = torch.zeros( (n_rows, bm_cols), dtype=torch.uint32, device=topk_ids.device ) grid = (triton.cdiv(n_rows, BLOCK_SIZE_M),) pack_bitmatrix[grid]( bitmatrix, topk_ids, n_rows, bm_cols, num_topk, BLOCK_SIZE_M=BLOCK_SIZE_M, BLOCK_SIZE_K=BLOCK_SIZE_K, ) bitmatrix_shape = [n_rows, bm_cols * 32] bitmatrix_shape_max = [n_rows, None] bitmatrix = ( Bitmatrix( bitmatrix, dtype=BIT, shape=bitmatrix_shape, shape_max=bitmatrix_shape_max ) if not use_legacy_triton_kernels else Bitmatrix( bitmatrix, shape=bitmatrix_shape, shape_max=bitmatrix_shape_max, scratchpad=None, ) ) # matmul_ogs expects invalid topk_weights to be -1s topk_weights = torch.where(topk_ids == -1, -1.0, topk_weights) routing_data, gather_indx, scatter_indx = legacy_routing_from_bitmatrix( bitmatrix, topk_weights, topk_ids, num_local_experts, num_topk ) return routing_data, gather_indx, scatter_indx class BaseOAITritonExperts(mk.FusedMoEPermuteExpertsUnpermute): @staticmethod def _supports_current_device() -> bool: raise NotImplementedError( "OAITritonExperts is not yet used by an Oracle. " "This method should not be called." ) @staticmethod def _supports_no_act_and_mul() -> bool: raise NotImplementedError( "OAITritonExperts is not yet used by an Oracle. " "This method should not be called." ) @staticmethod def _supports_quant_scheme( weight_key: QuantKey | None, activation_key: QuantKey | None, ) -> bool: raise NotImplementedError( "OAITritonExperts is not yet used by an Oracle. " "This method should not be called." ) @staticmethod def _supports_activation(activation: MoEActivation) -> bool: raise NotImplementedError( "OAITritonExperts is not yet used by an Oracle. " "This method should not be called." ) @staticmethod def _supports_parallel_config(moe_parallel_config: FusedMoEParallelConfig) -> bool: raise NotImplementedError( "OAITritonExperts is not yet used by an Oracle. " "This method should not be called." ) def supports_expert_map(self) -> bool: return True def moe_problem_size( self, a1: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_ids: torch.Tensor, ) -> tuple[int, int, int, int, int]: """ Extract the MoE problem size from the given tensor arguments: - a: The hidden states, input to the MoE layer. - w1: The first set of expert weights. - w2: The second set of expert weights. - topk_ids: The topk ids. Note: extracting the problem shape from the weight and activation tensors is not obvious. It needs to be done this way specifically due to subtle issues with particular kernels, e.g. the int4 kernels divide the trailing dimension by two, so it's not "correct" to extract N or K from the trailing dimension of w1 or w2. Similarly, some kernels transpose the weights, so this needs to be kept in mind. Note: This implementation covers most cases. However, if experts require a specialized implementation, like MarlinExperts, they are free to override this function. """ assert w1.dim() == 3 and w2.dim() == 3 E, _, N = w1.size() K = a1.size(-1) assert a1.dim() == 2 assert topk_ids.size(0) == a1.size(0), f"{topk_ids.size(0)} != {a1.size(0)}" M = a1.size(0) assert topk_ids.dim() == 2 topk = topk_ids.size(1) return E, M, N, K, topk def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: # Weight application and reduction happens in the fused_experts kernel. return TopKWeightAndReduceNoOP() def _make_routing_data( self, topk_ids: torch.Tensor, topk_weights: torch.Tensor, num_local_experts: int, ) -> tuple["RoutingData", torch.Tensor, torch.Tensor]: return make_routing_data(topk_ids, topk_weights, num_local_experts) class OAITritonExperts(BaseOAITritonExperts): """OAI Triton-based fused MoE expert implementation.""" @staticmethod def activation_format() -> mk.FusedMoEActivationFormat: return mk.FusedMoEActivationFormat.Standard def supports_chunking(self) -> bool: return True def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, activation: MoEActivation, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: # workspace are allocated inside the kernel activation_out_dim = self.adjust_N_for_activation(N, activation) workspace1 = (0, 0) workspace2 = (M * topk, activation_out_dim) output = (M, K) return (workspace1, workspace2, output) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: MoEActivation, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ): if self.quant_config is None: self.quant_config: FusedMoEQuantConfig = FUSED_MOE_UNQUANTIZED_CONFIG if expert_map is not None: topk_ids = expert_map[topk_ids] local_num_experts = w1.size(0) if global_num_experts == -1: global_num_experts = local_num_experts routing_data, gather_indx, scatter_indx = self._make_routing_data( topk_ids, topk_weights, local_num_experts ) topk = topk_ids.size(1) triton_kernel_fused_experts( output, hidden_states, w1, w2, routing_data, gather_indx, scatter_indx, topk=topk, activation=activation, quant_config=self.quant_config, apply_router_weight_on_input=False, global_num_experts=local_num_experts, expert_map=None, # applied already intermediate_cache=workspace2, a1q_scale=a1q_scale, ) class UnfusedOAITritonExperts(BaseOAITritonExperts): """ A Triton based MoE expert class that operates on expert standard format and explicitly keeps the activation and reduction (moe_sum) steps unfused from the matmul_ogs kernel. This exposes injection points for activation and moe_sum. One use case for it is to inject LoRA modules on the activation and moe_sum. """ @staticmethod def activation_format() -> mk.FusedMoEActivationFormat: return mk.FusedMoEActivationFormat.Standard def supports_chunking(self) -> bool: return True def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, activation: MoEActivation, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: # workspace are allocated inside the kernel activation_out_dim = self.adjust_N_for_activation(N, activation) workspace1 = (M * topk, activation_out_dim) workspace2 = (M * topk, max(N, K)) output = (M, K) return (workspace1, workspace2, output) def moe_sum(self, input: torch.Tensor, output: torch.Tensor): ops.moe_sum(input, output) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: MoEActivation, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ): # Use local variable to help mypy narrow the type after None check quant_config = self.quant_config if quant_config is None: quant_config = FUSED_MOE_UNQUANTIZED_CONFIG if expert_map is not None: topk_ids = expert_map[topk_ids] local_num_experts = w1.size(0) if global_num_experts == -1: global_num_experts = local_num_experts routing_data, gather_indx, scatter_indx = self._make_routing_data( topk_ids, topk_weights, local_num_experts ) topk = topk_ids.size(1) # type check, uint8 means mxfp4 assert hidden_states.dtype == torch.bfloat16 assert ( quant_config.w1_bias is None or quant_config.w1_bias.dtype == torch.float32 ) assert ( quant_config.w2_bias is None or quant_config.w2_bias.dtype == torch.float32 ) # Shape check, only check non-mxfp4 assert hidden_states.ndim == 2 assert hidden_states.shape[-1] == w1.shape[-2] assert w2.shape[-1] == w1.shape[1] batch_dim = 1 M, K = hidden_states.shape E, _, N = w1.shape if global_num_experts == -1: global_num_experts = E # Note that the output tensor might be in workspace13 intermediate_cache1 = _resize_cache(workspace2, (batch_dim, M * topk, N)) intermediate_cache3 = _resize_cache(workspace2, (batch_dim, M * topk, K)) activation_out_dim = self.adjust_N_for_activation(N, activation) intermediate_cache2 = _resize_cache(workspace13, (M * topk, activation_out_dim)) gammas = routing_data.gate_scal if routing_data else None matmul_ogs( hidden_states, w1, quant_config.w1_bias, routing_data, gather_indx=gather_indx, precision_config=quant_config.w1_precision, gammas=gammas if apply_router_weight_on_input else None, fused_activation=None, y=intermediate_cache1, ) self.activation( activation, intermediate_cache2, intermediate_cache1.view(-1, N)[gather_indx.dst_indx], ) # matmul_ogs grouped reduction fuse sum across multiple experts: # y[dst_indx // n_expts_act, :] += x # Need to set n_expts_act to 1 to unfuse moe_sum routing_data.n_expts_act = 1 matmul_ogs( intermediate_cache2[gather_indx.src_indx], w2, quant_config.w2_bias, routing_data, scatter_indx=scatter_indx, precision_config=quant_config.w2_precision, gammas=None if apply_router_weight_on_input else gammas, y=intermediate_cache3, ) self.moe_sum(intermediate_cache3.view(-1, topk, K), output)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fused_moe/gpt_oss_triton_kernels_moe.py", "license": "Apache License 2.0", "lines": 721, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/entrypoints/openai/test_completion_with_image_embeds.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json import openai # use the official client for correctness check import pytest import pytest_asyncio import torch from transformers import AutoConfig from tests.conftest import ImageTestAssets from tests.utils import RemoteOpenAIServer from vllm.utils.serial_utils import tensor2base64 # any model with a chat template should work here MODEL_NAME = "llava-hf/llava-1.5-7b-hf" CONFIG = AutoConfig.from_pretrained(MODEL_NAME) MAXIMUM_IMAGES = 2 @pytest.fixture(scope="module") def default_image_embeds_server_args() -> list[str]: return [ "--dtype", "bfloat16", "--max-model-len", "2048", "--max-num-seqs", "4", "--enforce-eager", "--limit-mm-per-prompt", json.dumps({"image": MAXIMUM_IMAGES}), "--enable-mm-embeds", ] @pytest.fixture(scope="module") def server_with_image_embeds(default_image_embeds_server_args): with RemoteOpenAIServer( MODEL_NAME, default_image_embeds_server_args, max_wait_seconds=600 ) as remote_server: yield remote_server @pytest_asyncio.fixture async def client_with_image_embeds(server_with_image_embeds): async with server_with_image_embeds.get_async_client() as async_client: yield async_client @pytest.mark.asyncio @pytest.mark.parametrize("model_name", [MODEL_NAME]) @pytest.mark.parametrize("dtype", [torch.half, torch.float16, torch.float32]) async def test_completions_with_image_embeds( client_with_image_embeds: openai.AsyncOpenAI, model_name: str, image_assets: ImageTestAssets, dtype: torch.dtype, ): # Test case: Single image embeds input image_embeds = image_assets[0].image_embeds.to(dtype=dtype) base64_image_embedding = tensor2base64(image_embeds) chat_completion = await client_with_image_embeds.chat.completions.create( messages=[ {"role": "system", "content": "You are a helpful assistant."}, { "role": "user", "content": [ { "type": "text", "text": "Describe these images separately. For each image," "reply with a short sentence (no more than 10 words).", }, { "type": "image_embeds", "image_embeds": base64_image_embedding, }, ], }, ], model=model_name, ) assert chat_completion.choices[0].message.content is not None assert isinstance(chat_completion.choices[0].message.content, str) assert len(chat_completion.choices[0].message.content) > 0
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/entrypoints/openai/test_completion_with_image_embeds.py", "license": "Apache License 2.0", "lines": 74, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/multimodal/test_cache.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import multiprocessing as mp import numpy as np import pytest import torch from vllm.config import ModelConfig, ParallelConfig, VllmConfig from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.cache import ( BaseMultiModalProcessorCache, BaseMultiModalReceiverCache, MultiModalCache, MultiModalProcessorCacheInItem, MultiModalProcessorCacheItem, MultiModalProcessorCacheItemMetadata, MultiModalProcessorSenderCache, MultiModalReceiverCache, ShmObjectStoreReceiverCache, ShmObjectStoreSenderCache, ) from vllm.multimodal.hasher import MultiModalHasher from vllm.multimodal.inputs import ( MultiModalFeatureSpec, MultiModalFieldElem, MultiModalKwargsItem, MultiModalKwargsItems, MultiModalSharedField, PlaceholderRange, ) from vllm.multimodal.processing import PromptInsertion from vllm.utils.mem_constants import GiB_bytes, MiB_bytes pytestmark = pytest.mark.cpu_test def _dummy_elem( size: int, *, rng: np.random.RandomState | None = None, ): if rng is None: data = torch.empty((size,), dtype=torch.int8) else: data = torch.from_numpy(rng.randint(4, size=(size,), dtype=np.int8)) return MultiModalFieldElem( data=data, field=MultiModalSharedField(batch_size=1), ) def _dummy_item( size_by_key: dict[str, int], *, rng: np.random.RandomState | None = None, ): return MultiModalKwargsItem( {key: _dummy_elem(size, rng=rng) for key, size in size_by_key.items()} ) def _dummy_items( size_by_key_modality: dict[str, dict[str, int]], *, rng: np.random.RandomState | None = None, ): return MultiModalKwargsItems( { modality: [_dummy_item(size_by_key, rng=rng)] for modality, size_by_key in size_by_key_modality.items() } ) @pytest.mark.parametrize( ("item", "expected_size"), [ (_dummy_item({"a1": 100}), 100), (_dummy_item({"a1": 100, "a2": 110}), 210), (_dummy_items({"a": {"a1": 100, "a2": 110}, "b": {"b1": 120, "b2": 130}}), 460), # noqa: E501 ], ) def test_cache_item_size(item, expected_size): cache = MultiModalCache.get_lru_cache(2048, type(item)) cache[""] = item assert cache.currsize == expected_size prompt_update = PromptInsertion("dummy", "target", "insertion").resolve(0) cache[""] = MultiModalProcessorCacheItem(item, [prompt_update]) assert cache.currsize == expected_size cache[""] = MultiModalProcessorCacheItemMetadata(item, [prompt_update]) assert cache.currsize == expected_size cache[""] = item.get_data() assert cache.currsize == expected_size def _create_vllm_config( *, mm_processor_cache_gb: float, enable_ipc: bool, ): return VllmConfig( model_config=ModelConfig( model="llava-hf/llava-onevision-qwen2-0.5b-ov-hf", mm_processor_cache_gb=mm_processor_cache_gb, ), parallel_config=ParallelConfig(data_parallel_size=1 if enable_ipc else 2), ) def _compare_caches( config_0: VllmConfig, config_1: VllmConfig, *, item_capacity: int = 8, hit_rate: float = 0.5, max_items_per_iter: int = 3, is_cached_calls_per_iter: int, n_iter: int = 100, seed: int = 0, ): cache_0_p0 = MULTIMODAL_REGISTRY.processor_cache_from_config(config_0) cache_0_p1 = MULTIMODAL_REGISTRY.engine_receiver_cache_from_config(config_0) cache_1_p0 = MULTIMODAL_REGISTRY.processor_cache_from_config(config_1) cache_1_p1 = MULTIMODAL_REGISTRY.engine_receiver_cache_from_config(config_1) cache_size_gb = max( config_0.model_config.multimodal_config.mm_processor_cache_gb, config_1.model_config.multimodal_config.mm_processor_cache_gb, ) item_size_gb = int(cache_size_gb / item_capacity) rng = np.random.RandomState(seed) all_items = [ _dummy_item({"key": item_size_gb}, rng=rng) for _ in range(int(item_capacity / hit_rate)) ] all_hashes = [ MultiModalHasher.hash_kwargs(item=item.get_data()) for item in all_items ] prompt_update = PromptInsertion("dummy", "target", "insertion").resolve(0) for it in range(n_iter): num_items_to_select = rng.randint(0, max_items_per_iter) item_idxs_to_select = rng.choice(len(all_items), num_items_to_select) selected_items = [all_items[idx] for idx in item_idxs_to_select] selected_hashes = [all_hashes[idx] for idx in item_idxs_to_select] if cache_0_p0 is None: cache_0_p0_out = selected_items else: for _ in range(is_cached_calls_per_iter): cache_0_p0.is_cached(selected_hashes) cache_0_p0_out = [ item for item, _ in cache_0_p0.get_and_update( [(item, [prompt_update]) for item in selected_items], selected_hashes, ) ] if cache_1_p0 is None: cache_1_p0_out = selected_items else: for _ in range(is_cached_calls_per_iter): cache_1_p0.is_cached(selected_hashes) cache_1_p0_out = [ item for item, _ in cache_1_p0.get_and_update( [(item, [prompt_update]) for item in selected_items], selected_hashes, ) ] if cache_0_p1 is None: cache_0_p1_out = cache_0_p0_out else: cache_0_p1_out = cache_0_p1.get_and_update(cache_0_p0_out, selected_hashes) if cache_1_p1 is None: cache_1_p1_out = cache_1_p0_out else: cache_1_p1_out = cache_1_p1.get_and_update(cache_1_p0_out, selected_hashes) assert cache_0_p1_out == cache_1_p1_out, f"Failed at {it=}" @pytest.mark.parametrize("is_cached_calls_per_iter", [1, 2, 3]) def test_ipc_enable_disable_consistency(is_cached_calls_per_iter): cache_size_gb = 1 / (1 << 20) vllm_config_ipc_enabled = _create_vllm_config( mm_processor_cache_gb=cache_size_gb, enable_ipc=True, ) vllm_config_ipc_disabled = _create_vllm_config( mm_processor_cache_gb=0, enable_ipc=False, ) vllm_config_cache_disabled = _create_vllm_config( mm_processor_cache_gb=cache_size_gb, enable_ipc=True, ) _compare_caches( vllm_config_ipc_enabled, vllm_config_ipc_disabled, is_cached_calls_per_iter=is_cached_calls_per_iter, ) _compare_caches( vllm_config_ipc_disabled, vllm_config_cache_disabled, is_cached_calls_per_iter=is_cached_calls_per_iter, ) _compare_caches( vllm_config_cache_disabled, vllm_config_ipc_enabled, is_cached_calls_per_iter=is_cached_calls_per_iter, ) def _run_test_cache_eviction_lru( p0_cache: BaseMultiModalProcessorCache, p1_cache: BaseMultiModalReceiverCache, base_item_size: int, ): request1_hashes = [ "image_A", "image_B", "image_C", ] request1_items = { h: MultiModalKwargsItem.dummy(nbytes=2 * base_item_size) for h in request1_hashes } request2_hashes = ["image_D", "image_E", "image_A", "image_C"] request2_items = { h: MultiModalKwargsItem.dummy(nbytes=1 * base_item_size) for h in request2_hashes } ########################## # STEP 1: Request 1 send ########################## sender_is_cached_item_req1 = p0_cache.is_cached(request1_hashes) # Cache is empty assert sender_is_cached_item_req1 == [False, False, False] # Touch all mm hash for P0 Cache before process for mm_hash in request1_hashes: p0_cache.touch_sender_cache_item(mm_hash) ########################### # Process request 1 for P0 Cache ########################### item_tuple: MultiModalProcessorCacheInItem for i, h in enumerate(request1_hashes): # Use precomputed cache state is_cached = sender_is_cached_item_req1[i] item_tuple = (request1_items[h], []) if not is_cached else None print(f"Request 1: key={h} | cached={is_cached}") p0_cache.get_and_update_item(item_tuple, h) ########################### # Process request 1 for P1 Cache ########################### # Touch all mm hash for P1 Cache before process for mm_hash in request1_hashes: p1_cache.touch_receiver_cache_item(mm_hash) for h in request1_hashes: p1_cache.get_and_update_item(request1_items[h], h) expected_hashes = ["image_A", "image_B", "image_C"] assert list(p0_cache._cache.order) == expected_hashes ########################## # STEP 2: Request 2 send ########################## sender_is_cached_item_req2 = p0_cache.is_cached(request2_hashes) assert sender_is_cached_item_req2 == [False, False, True, True] # Touch all mm hash for P0 Cache before process for mm_hash in request2_hashes: p0_cache.touch_sender_cache_item(mm_hash) ########################### # Process request 2 for P0 Cache ########################### for i, h in enumerate(request2_hashes): # Use precomputed cache state again is_cached = sender_is_cached_item_req2[i] item_tuple = (request2_items[h], []) if not is_cached else None print(f"Request 2: key={h} | cached={is_cached}") p0_cache.get_and_update_item(item_tuple, h) ########################### # Process request 2 for P1 Cache ########################### # Touch all mm hash for P1 Cache before process for mm_hash in request2_hashes: p1_cache.touch_receiver_cache_item(mm_hash) for h in request2_hashes: p1_cache.get_and_update_item(request2_items[h], h) expected_hashes = ["image_D", "image_E", "image_A", "image_C"] assert list(p0_cache._cache.order) == expected_hashes def test_cache_eviction_lru_cache(): model_config = ModelConfig( model="llava-hf/llava-onevision-qwen2-0.5b-ov-hf", mm_processor_cache_gb=6 / GiB_bytes, ) sender_cache = MultiModalProcessorSenderCache(model_config) receiver_cache = MultiModalReceiverCache(model_config) _run_test_cache_eviction_lru(sender_cache, receiver_cache, base_item_size=1) # This test verifies shared-memory cache eviction behavior across processor (p0) # and receiver (p1) caches. # Flow summary: # 1. Request 1 adds images A, B, C — completely filling the cache. # 2. Request 2 tries to add image_G and image_A, but image_G cannot be added because # cache is full and A is protected from eviction — cache remains unchanged. # 3. Request 3 adds image_G, image_H, image_I and image_B # this time, image_A is evicted, freeing 5MB space # and image_G, image_H successfully fits, # image_B is protected from eviction then image_i cannot be added. # This proving normal eviction and reuse behavior. def _run_test_cache_eviction_shm( p0_cache: BaseMultiModalProcessorCache, p1_cache: BaseMultiModalReceiverCache, base_item_size: int, ): request1_hashes = ["image_A", "image_B", "image_C"] request1_items = { h: MultiModalKwargsItem.dummy(5 * base_item_size) for h in request1_hashes } request1_items_p0_result = [] request2_hashes = ["image_G", "image_A"] request2_items = { h: MultiModalKwargsItem.dummy( (5 if h in request1_hashes else 2) * base_item_size ) for h in request2_hashes } request2_items_p0_result = [] request3_hashes = ["image_G", "image_H", "image_I", "image_B"] request3_items = { h: MultiModalKwargsItem.dummy( (5 if h in request1_hashes else 2) * base_item_size ) for h in request3_hashes } request3_items_p0_result = [] ########################## # STEP 1: Request 1 send # This will fill up the cache ########################## sender_is_cached_item_req1 = p0_cache.is_cached(request1_hashes) # Cache is empty assert sender_is_cached_item_req1 == [False, False, False] # Touch all mm hash for P0 Cache before process for mm_hash in request1_hashes: p0_cache.touch_sender_cache_item(mm_hash) ########################### # Process request 1 for P0 Cache ########################### item_tuple: MultiModalProcessorCacheInItem for i, h in enumerate(request1_hashes): # Use precomputed cache state is_cached = sender_is_cached_item_req1[i] item_tuple = (request1_items[h], []) if not is_cached else None print(f"Request 1: key={h} | cached={is_cached}") p0_result = p0_cache.get_and_update_item(item_tuple, h) # Only get mm item, ignore prompt update result request1_items_p0_result.append(p0_result[0]) ########################### # Process request 1 for P1 Cache ########################### # Touch all mm hash for P1 Cache before process for mm_hash, mm_item in zip(request1_hashes, request1_items_p0_result): p1_cache.touch_receiver_cache_item(mm_hash, mm_item) for mm_hash, mm_item in zip(request1_hashes, request1_items_p0_result): p1_cache.get_and_update_item(mm_item, mm_hash) expected_hashes = ["image_A", "image_B", "image_C"] assert list(p0_cache._shm_cache.key_index.keys()) == expected_hashes ########################## # STEP 2: Request 2 send # There is no eviction because image_A is protected # No new item can add to cache ########################## sender_is_cached_item_req2 = p0_cache.is_cached(request2_hashes) assert sender_is_cached_item_req2 == [False, True] # Touch all mm hash for P0 Cache before process for mm_hash in request2_hashes: p0_cache.touch_sender_cache_item(mm_hash) ########################### # Process request 2 for P0 Cache ########################### for i, h in enumerate(request2_hashes): # Use precomputed cache state again is_cached = sender_is_cached_item_req2[i] item_tuple = (request2_items[h], []) if not is_cached else None print(f"Request 2: key={h} | cached={is_cached}") p0_result = p0_cache.get_and_update_item(item_tuple, h) # Only get mm item, ignore prompt update result request2_items_p0_result.append(p0_result[0]) # image_A cannot be evict then # image_G will fail to allocate anyway and image_A still in cache assert p0_cache.is_cached(request2_hashes) == [False, True] ########################### # Process request 2 for P1 Cache ########################### # Touch all mm hash for P1 Cache before process for mm_hash, mm_item in zip(request2_hashes, request2_items_p0_result): p1_cache.touch_receiver_cache_item(mm_hash, mm_item) for mm_hash, mm_item in zip(request2_hashes, request2_items_p0_result): p1_cache.get_and_update_item(mm_item, mm_hash) # Prove that cache state is unchanged expected_hashes = ["image_A", "image_B", "image_C"] assert list(p0_cache._shm_cache.key_index.keys()) == expected_hashes ########################## # STEP 3: Request 3 send ########################## ##### Prove that cache eviction work normally sender_is_cached_item_req3 = p0_cache.is_cached(request3_hashes) assert sender_is_cached_item_req3 == [False, False, False, True] # Touch all mm hash for P0 Cache before process for mm_hash in request3_hashes: p0_cache.touch_sender_cache_item(mm_hash) ########################### # Process request 3 for P0 Cache ########################### for i, h in enumerate(request3_hashes): # Use precomputed cache state again is_cached = sender_is_cached_item_req3[i] item_tuple = (request3_items[h], []) if not is_cached else None print(f"Request 3: key={h} | cached={is_cached}") p0_result = p0_cache.get_and_update_item(item_tuple, h) # Only get mm item, ignore prompt update result request3_items_p0_result.append(p0_result[0]) # image_A got evict and image_G add to cache # image_B is still protected # image_G, image_H fit but image_I cannot fit assert p0_cache.is_cached(request3_hashes) == [True, True, False, True] ########################### # Process request 3 for P1 Cache ########################### # Touch all mm hash for P1 Cache before process for mm_hash, mm_item in zip(request3_hashes, request3_items_p0_result): p1_cache.touch_receiver_cache_item(mm_hash, mm_item) for mm_hash, mm_item in zip(request3_hashes, request3_items_p0_result): p1_cache.get_and_update_item(mm_item, mm_hash) expected_hashes = ["image_B", "image_C", "image_G", "image_H"] assert list(p0_cache._shm_cache.key_index.keys()) == expected_hashes def test_cache_eviction_shm_cache(): vllm_config = VllmConfig( model_config=ModelConfig( model="llava-hf/llava-onevision-qwen2-0.5b-ov-hf", mm_processor_cache_type="shm", mm_shm_cache_max_object_size_mb=6, mm_processor_cache_gb=15.2 * MiB_bytes / GiB_bytes, ), ) sender_cache = ShmObjectStoreSenderCache(vllm_config) receiver_cache = ShmObjectStoreReceiverCache(vllm_config, mp.Lock()) _run_test_cache_eviction_shm(sender_cache, receiver_cache, base_item_size=MiB_bytes) def test_processor_cache_shared_across_loras(): """Test that processor cache uses mm_hash to share data across LoRAs.""" model_config = ModelConfig( model="llava-hf/llava-onevision-qwen2-0.5b-ov-hf", mm_processor_cache_gb=1, ) receiver_cache = MultiModalReceiverCache(model_config) base_mm_hash = "image_hash_abc123" lora_a_identifier = f"12345:{base_mm_hash}" lora_b_identifier = f"67890:{base_mm_hash}" item_data = MultiModalKwargsItem.dummy(1024) feature_lora_a = MultiModalFeatureSpec( data=item_data, modality="image", identifier=lora_a_identifier, mm_position=PlaceholderRange(offset=0, length=100), mm_hash=base_mm_hash, ) receiver_cache.get_and_update_features([feature_lora_a]) assert base_mm_hash in receiver_cache._cache feature_lora_b = MultiModalFeatureSpec( data=None, modality="image", identifier=lora_b_identifier, mm_position=PlaceholderRange(offset=0, length=100), mm_hash=base_mm_hash, ) receiver_cache.get_and_update_features([feature_lora_b]) assert feature_lora_b.data == item_data
{ "repo_id": "vllm-project/vllm", "file_path": "tests/multimodal/test_cache.py", "license": "Apache License 2.0", "lines": 454, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/multimodal/cache.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import operator import sys from abc import ABC, abstractmethod from collections.abc import Mapping, Sequence from multiprocessing.synchronize import Lock as LockType from typing import TYPE_CHECKING, Generic, TypeAlias, TypeVar, cast import torch from typing_extensions import override import vllm.envs as envs from vllm.distributed.device_communicators.shm_object_storage import ( MsgpackSerde, SingleWriterShmObjectStorage, SingleWriterShmRingBuffer, ) from vllm.logger import init_logger from vllm.utils.cache import CacheInfo, LRUCache from vllm.utils.jsontree import json_count_leaves, json_map_leaves, json_reduce_leaves from vllm.utils.mem_constants import GiB_bytes, MiB_bytes from vllm.utils.mem_utils import format_gib from .inputs import ( MultiModalBatchedField, MultiModalFeatureSpec, MultiModalFieldElem, MultiModalKwargsItem, MultiModalKwargsItems, NestedTensors, ) if TYPE_CHECKING: from vllm.config import ModelConfig, VllmConfig from .processing.processor import ResolvedPromptUpdate logger = init_logger(__name__) class MultiModalProcessorCacheItem: """ The data to store inside `MultiModalProcessorOnlyCache`. Args: item: The processed tensor data corresponding to a multi-modal item. prompt_updates: The prompt updates corresponding to `item`. """ def __init__( self, item: MultiModalKwargsItem, prompt_updates: Sequence["ResolvedPromptUpdate"], ) -> None: super().__init__() self.item = item self.prompt_updates = prompt_updates class MultiModalProcessorCacheItemMetadata: """ The metadata to store inside `MultiModalProcessorSenderCache`. Args: item: The processed tensor data corresponding to a multi-modal item. Since P1 already stores the tensor data, we only store its size metadata in P0 to reduce memory usage. The size metadata is still needed to keep the same cache eviction policy as P0. prompt_updates: The prompt updates corresponding to `item`. This needs to stay on P0 because for some models, they are dependent on the processed tensor data (cached on P1). """ def __init__( self, item: MultiModalKwargsItem, prompt_updates: Sequence["ResolvedPromptUpdate"], ) -> None: super().__init__() self.item_size = MultiModalCache.get_item_size(item) self.prompt_updates = prompt_updates MultiModalCacheValue: TypeAlias = ( MultiModalProcessorCacheItem | MultiModalProcessorCacheItemMetadata | MultiModalKwargsItems | MultiModalKwargsItem | Mapping[str, NestedTensors] ) _V = TypeVar("_V", bound=MultiModalCacheValue) class MultiModalCache: @classmethod def get_leaf_size(cls, leaf: object) -> int: if isinstance(leaf, MultiModalProcessorCacheItem): return cls.get_leaf_size(leaf.item) if isinstance(leaf, MultiModalProcessorCacheItemMetadata): return leaf.item_size # These are not subclasses of dict if isinstance( leaf, (MultiModalKwargsItems, MultiModalKwargsItem, MultiModalFieldElem), ): return cls.get_item_size(leaf.data) # type: ignore # sys.getsizeof doesn't work for tensors if isinstance(leaf, torch.Tensor): return leaf.nbytes return sys.getsizeof(leaf) @classmethod def get_item_size( cls, value: MultiModalCacheValue, *, debug: bool = False, ) -> int: size = json_reduce_leaves( operator.add, json_map_leaves(cls.get_leaf_size, value) ) if debug: leaf_count = json_count_leaves(value) logger.debug( "Calculated size of %s to be %s GiB (%d leaves)", type(value), format_gib(size), leaf_count, ) return size @classmethod def get_item_complexity(cls, value: MultiModalCacheValue) -> int: """ Get the number of leaf elements in a multi-modal cache value. This provides a measure of structural complexity that can be useful for debugging cache performance and understanding data patterns. Args: value: The multi-modal cache value to analyze. Returns: The number of leaf elements in the nested structure. """ return json_count_leaves(value) @classmethod def get_lru_cache( cls, capacity_gb: float, value_type: type[_V], *, debug: bool = False, ) -> LRUCache[str, _V]: return LRUCache( GiB_bytes * capacity_gb, getsizeof=lambda x: cls.get_item_size(x, debug=debug), ) _I = TypeVar("_I", contravariant=True) _O = TypeVar("_O", covariant=True) class BaseMultiModalCache(ABC, Generic[_I, _O]): """ Abstract base class to read/write multi-modal items from cache. The idea of multi-modal caching is based on having a client and server where the client executes in the frontend process (=P0) and the server in the core process (=P1). The data flow is as follows: ``` is_cached() x N get_and_update() P0: From API -----------------> -----------------> To P1 get_and_update() P1: From P0 -----------------> To model ``` `is_cached()` can be called any number of times in P0. However, `get_and_update()` must be called in P0 and P1 one after another so that their cache eviction order remains the same. This ensures that the keys in P0 and P1 caches are mirrored, allowing us to determine whether a key is cached in P1 by looking up the P0 cache, without having to communicate with P1. """ @abstractmethod def get_and_update_item( self, mm_item: _I, mm_hash: str, ) -> _O: """ Possibly update a multi-modal item based on whether it is in the underlying cache. This update is done out-of-place and updates the cache eviction order. Args: mm_item: The multi-modal item to update. mm_hash: The hash of `mm_item`. Returns: The update multi-modal item. """ raise NotImplementedError def get_and_update( self, mm_items: Sequence[_I], mm_hashes: list[str], ) -> list[_O]: """ Possibly update a sequence of multi-modal items based on whether they are in the underlying cache. This update is done out-of-place and updates the cache eviction order. Args: mm_items: The multi-modal items to update. mm_hashes: The hash of each item in `mm_items`. Returns: A new list of updated multi-modal items. """ assert len(mm_items) == len(mm_hashes) return [ self.get_and_update_item(mm_item, mm_hash) for mm_item, mm_hash in zip(mm_items, mm_hashes) ] @abstractmethod def clear_cache(self) -> None: """Clear the underlying cache.""" raise NotImplementedError MultiModalProcessorCacheInItem: TypeAlias = ( tuple[MultiModalKwargsItem, Sequence["ResolvedPromptUpdate"]] | None ) MultiModalProcessorCacheOutItem: TypeAlias = tuple[ MultiModalKwargsItem | None, Sequence["ResolvedPromptUpdate"] ] class BaseMultiModalProcessorCache( BaseMultiModalCache[MultiModalProcessorCacheInItem, MultiModalProcessorCacheOutItem] ): """The required interface for caches on P0.""" @abstractmethod def is_cached_item(self, mm_hash: str) -> bool: """ Check whether a multi-modal item is in the underlying cache. This **DOES NOT** update the cache eviction order. Args: mm_hash: The hash of the item to check. Returns: `True` if the item is cached, otherwise `False`. """ raise NotImplementedError def is_cached(self, mm_hashes: list[str]) -> list[bool]: """ Check whether a sequence of multi-modal items are in the underlying cache. This **DOES NOT** update the cache eviction order. Args: mm_hashes: The hash of each item to check. Returns: For each item, `True` if the item is cached, otherwise `False`. """ return [self.is_cached_item(mm_hash) for mm_hash in mm_hashes] def close(self) -> None: """Close the underlying cache, if needed.""" pass @abstractmethod def touch_sender_cache_item(self, mm_hash: str) -> None: """ Update the cache eviction order for a multi-modal item. This is used to touch the item in the cache without changing its value. Args: mm_hash: The hash of the multi-modal item. """ raise NotImplementedError @abstractmethod def make_stats(self, *, delta: bool = False) -> CacheInfo: """ Get (and reset) the multi-modal cache stats. Returns: The current multi-modal caching stats. """ raise NotImplementedError class MultiModalProcessorOnlyCache(BaseMultiModalProcessorCache): """ The cache which is used on P0 when IPC caching is disabled. How to update each item: - If the item is in the cache, replace the input with the cached item. - If the item is not in the cache, store that item (which includes tensor data and metadata) into the cache, and return the input. """ def __init__(self, model_config: "ModelConfig") -> None: super().__init__() mm_config = model_config.get_multimodal_config() self._cache = MultiModalCache.get_lru_cache( mm_config.mm_processor_cache_gb, MultiModalProcessorCacheItem, ) @override def is_cached_item(self, mm_hash: str) -> bool: return mm_hash in self._cache @override def get_and_update_item( self, mm_item: MultiModalProcessorCacheInItem, mm_hash: str, ) -> MultiModalProcessorCacheOutItem: if (cached_item := self._cache.get(mm_hash)) is not None: return cached_item.item, cached_item.prompt_updates assert mm_item is not None, f"Expected a cached item for {mm_hash=}" self._cache[mm_hash] = MultiModalProcessorCacheItem(*mm_item) return mm_item @override def touch_sender_cache_item(self, mm_hash: str) -> None: self._cache.touch(mm_hash) @override def clear_cache(self) -> None: self._cache.clear() @override def make_stats(self, *, delta: bool = False) -> CacheInfo: return self._cache.stat(delta=delta) class MultiModalProcessorSenderCache(BaseMultiModalProcessorCache): """ The cache which is used on P0 when IPC caching is enabled. How to update each item: - If the item is already in the cache, clear the input to avoid unnecessary IPC. - If the item is not in the cache, store the metadata of that item so that the eviction policy remains the same as the cache on P1, and return the input. By only storing the metadata, we avoid keeping the data itself in memory inside P0. """ def __init__(self, model_config: "ModelConfig") -> None: super().__init__() mm_config = model_config.get_multimodal_config() self._cache = MultiModalCache.get_lru_cache( mm_config.mm_processor_cache_gb, MultiModalProcessorCacheItemMetadata, ) @override def is_cached_item(self, mm_hash: str) -> bool: return mm_hash in self._cache @override def get_and_update_item( self, mm_item: MultiModalProcessorCacheInItem, mm_hash: str, ) -> MultiModalProcessorCacheOutItem: if (cached_item := self._cache.get(mm_hash)) is not None: return None, cached_item.prompt_updates assert mm_item is not None, f"Expected a cached item for {mm_hash=}" self._cache[mm_hash] = MultiModalProcessorCacheItemMetadata(*mm_item) return mm_item @override def touch_sender_cache_item(self, mm_hash: str) -> None: self._cache.touch(mm_hash) @override def clear_cache(self) -> None: self._cache.clear() @override def make_stats(self, *, delta: bool = False) -> CacheInfo: return self._cache.stat(delta=delta) class ShmObjectStoreSenderCache(BaseMultiModalProcessorCache): """ The cache which is used on P0 when IPC caching is enabled. How to update each item: - If the item is already in the cache, clear the input to avoid unnecessary IPC. - If the item is not in the cache, store the data in shared memory. """ def __init__(self, vllm_config: "VllmConfig") -> None: super().__init__() self.world_size = vllm_config.parallel_config.world_size mm_config = vllm_config.model_config.get_multimodal_config() ring_buffer = SingleWriterShmRingBuffer( data_buffer_size=int(mm_config.mm_processor_cache_gb * GiB_bytes), name=envs.VLLM_OBJECT_STORAGE_SHM_BUFFER_NAME, create=True, # sender is the writer ) self._shm_cache = SingleWriterShmObjectStorage( max_object_size=mm_config.mm_shm_cache_max_object_size_mb * MiB_bytes, n_readers=self.world_size, ring_buffer=ring_buffer, serde_class=MsgpackSerde, ) # cache prompt_updates for P0 only self._p0_cache: dict[str, Sequence[ResolvedPromptUpdate]] = {} self._hits = 0 self._total = 0 self._last_info = CacheInfo(hits=0, total=0) def _stat(self, *, delta: bool = False) -> CacheInfo: info = CacheInfo(hits=self._hits, total=self._total) if delta: info_delta = info - self._last_info self._last_info = info info = info_delta return info @override def is_cached_item(self, mm_hash: str) -> bool: return self._shm_cache.is_cached(mm_hash) @override def get_and_update_item( self, mm_item: MultiModalProcessorCacheInItem, mm_hash: str, ) -> MultiModalProcessorCacheOutItem: if self._shm_cache.is_cached(mm_hash): self._hits += 1 self._total += 1 address, monotonic_id = self._shm_cache.get_cached(mm_hash) prompt_updates = self._p0_cache[mm_hash] return self.address_as_item(address, monotonic_id), prompt_updates assert mm_item is not None, f"Expected a cached item for {mm_hash=}" item, prompt_updates = mm_item self._total += 1 try: address, monotonic_id = self._shm_cache.put(mm_hash, item) # Try to remove dangling items if p0 cache is too large. if len(self._p0_cache) >= 2 * len(self._shm_cache.key_index): self.remove_dangling_items() self._p0_cache[mm_hash] = prompt_updates return self.address_as_item(address, monotonic_id), prompt_updates except (ValueError, MemoryError) as e: # put may fail if the object is too large or # the cache is full. # In this case we log the error and keep the original mm_input. logger.debug("Failed to cache mm_input with hash %s: %s", mm_hash, e) return mm_item @override def touch_sender_cache_item(self, mm_hash: str) -> None: """Touch the item in shared memory cache to prevent eviction. Increments writer_flag on sender side.""" self._shm_cache.touch(mm_hash) @override def clear_cache(self) -> None: self._shm_cache.clear() self._p0_cache.clear() self._hits = 0 self._total = 0 self._last_info = CacheInfo(hits=0, total=0) @override def make_stats(self, *, delta: bool = False) -> CacheInfo: return self._stat(delta=delta) @override def close(self) -> None: self._shm_cache.close() def remove_dangling_items(self) -> None: """Remove items that are no longer in the shared memory cache.""" cached_hashes = self._shm_cache.key_index.keys() dangling_hashes = set(self._p0_cache.keys()) - cached_hashes for mm_hash in dangling_hashes: del self._p0_cache[mm_hash] def address_as_item( self, address: int, monotonic_id: int, ) -> MultiModalKwargsItem: addr_elem = MultiModalFieldElem( data=address, field=MultiModalBatchedField(), ) id_elem = MultiModalFieldElem( data=monotonic_id, field=MultiModalBatchedField(), ) return MultiModalKwargsItem({"address": addr_elem, "monotonic_id": id_elem}) class BaseMultiModalReceiverCache( BaseMultiModalCache[MultiModalKwargsItem | None, MultiModalKwargsItem] ): """The required interface for caches on P1.""" def get_and_update_features( self, mm_features: list["MultiModalFeatureSpec"], ) -> list["MultiModalFeatureSpec"]: """ Update multimodal features with cached encoder outputs. Touch all identifier at first before update to avoid item in updated list evict during update. Uses mm_hash for cache key to share across LoRAs (falls back to identifier for backward compatibility). """ for feature in mm_features: cache_key = feature.mm_hash or feature.identifier self.touch_receiver_cache_item(cache_key, feature.data) for feature in mm_features: cache_key = feature.mm_hash or feature.identifier feature.data = self.get_and_update_item(feature.data, cache_key) return mm_features @abstractmethod def touch_receiver_cache_item( self, mm_hash: str, mm_item: MultiModalKwargsItem | None = None, ) -> None: """ Update the cache eviction order for a multi-modal item. This is used to touch the item in the cache without changing its value. Args: mm_hash: The hash of the multi-modal item. mm_item: The multi-modal item itself. This is optional and may not be needed by some cache implementations. """ raise NotImplementedError class MultiModalReceiverCache(BaseMultiModalReceiverCache): """ The cache which is used on P1 when IPC caching is enabled. How to update each item: - If the item is in the cache, replace the input with the cached item. - If the item is not in the cache, store that item (which includes tensor data) into the cache, and return the input. """ def __init__(self, model_config: "ModelConfig") -> None: super().__init__() mm_config = model_config.get_multimodal_config() self._cache = MultiModalCache.get_lru_cache( mm_config.mm_processor_cache_gb, MultiModalKwargsItem, ) @override def get_and_update_item( self, mm_item: MultiModalKwargsItem | None, mm_hash: str, ) -> MultiModalKwargsItem: if (cached_item := self._cache.get(mm_hash)) is not None: return cached_item assert mm_item is not None, f"Expected a cached item for {mm_hash=}" self._cache[mm_hash] = mm_item return mm_item @override def touch_receiver_cache_item( self, mm_hash: str, mm_item: MultiModalKwargsItem | None = None, ) -> None: self._cache.touch(mm_hash) @override def clear_cache(self) -> None: self._cache.clear() class ShmObjectStoreReceiverCache(BaseMultiModalReceiverCache): """ The cache which is used on P1 Worker Process when IPC caching is enabled. How to update each item: - If the item has an address, replace the input with the cached item. - If not, return the input. """ def __init__( self, vllm_config: "VllmConfig", shared_worker_lock: LockType, ) -> None: super().__init__() self.world_size = vllm_config.parallel_config.world_size mm_config = vllm_config.model_config.get_multimodal_config() ring_buffer = SingleWriterShmRingBuffer( data_buffer_size=int(mm_config.mm_processor_cache_gb * GiB_bytes), name=envs.VLLM_OBJECT_STORAGE_SHM_BUFFER_NAME, create=False, # Server is a reader ) self._shm_cache = SingleWriterShmObjectStorage( max_object_size=mm_config.mm_shm_cache_max_object_size_mb * MiB_bytes, n_readers=self.world_size, ring_buffer=ring_buffer, serde_class=MsgpackSerde, reader_lock=shared_worker_lock, ) @override def get_and_update_item( self, mm_item: MultiModalKwargsItem | None, mm_hash: str, ) -> MultiModalKwargsItem: assert mm_item is not None, f"Expected an address item for {mm_hash=}" if "address" in mm_item: address = cast(int, mm_item["address"].data) monotonic_id = cast(int, mm_item["monotonic_id"].data) return self._shm_cache.get(address, monotonic_id) return mm_item @override def touch_receiver_cache_item( self, mm_hash: str, mm_item: MultiModalKwargsItem | None = None, ) -> None: """Touch the item in shared memory cache to prevent eviction. Increments reader_count on receiver side.""" assert mm_item is not None if "address" in mm_item: address = cast(int, mm_item["address"].data) monotonic_id = cast(int, mm_item["monotonic_id"].data) self._shm_cache.touch(mm_hash, address=address, monotonic_id=monotonic_id) @override def clear_cache(self) -> None: self._shm_cache.clear()
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/multimodal/cache.py", "license": "Apache License 2.0", "lines": 565, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/attention/backends/mamba1_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass, replace from typing import Any from vllm.v1.attention.backend import AttentionBackend, CommonAttentionMetadata from vllm.v1.attention.backends.mamba_attn import ( BaseMambaAttentionMetadata, BaseMambaAttentionMetadataBuilder, ) class Mamba1AttentionBackend(AttentionBackend): @staticmethod def get_name() -> str: return "MAMBA1_ATTN" @staticmethod def get_builder_cls() -> type["Mamba1AttentionMetadataBuilder"]: return Mamba1AttentionMetadataBuilder @dataclass class Mamba1AttentionMetadata(BaseMambaAttentionMetadata): pass class Mamba1AttentionMetadataBuilder( BaseMambaAttentionMetadataBuilder[Mamba1AttentionMetadata] ): metadata_cls = Mamba1AttentionMetadata def build( self, common_prefix_len: int, common_attn_metadata: CommonAttentionMetadata, fast_build: bool = False, **kwargs: Any, ) -> Mamba1AttentionMetadata: common = self._compute_common_metadata(common_attn_metadata) if ( common.num_prefills > 0 and self.vllm_config.cache_config.mamba_cache_mode == "all" ): cu_chunk_seqlen_p, _, last_chunk_indices_p = ( self._build_chunk_metadata_tensors( self.kv_cache_spec.block_size, common, common_attn_metadata, ) ) return replace( common, cu_chunk_seqlen_p=cu_chunk_seqlen_p, last_chunk_indices_p=last_chunk_indices_p, ) return common
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/attention/backends/mamba1_attn.py", "license": "Apache License 2.0", "lines": 48, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/quantization/mxfp4.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from enum import Enum import torch from torch.nn.parameter import Parameter from vllm import envs from vllm._aiter_ops import rocm_aiter_ops from vllm.config import get_current_vllm_config from vllm.logger import init_logger from vllm.model_executor.layers.attention import Attention from vllm.model_executor.layers.fused_moe import ( FusedMoE, FusedMoEConfig, FusedMoEMethodBase, MoEActivation, ) from vllm.model_executor.layers.fused_moe import modular_kernel as mk from vllm.model_executor.layers.fused_moe.all2all_utils import ( maybe_make_prepare_finalize, ) from vllm.model_executor.layers.fused_moe.config import ( FusedMoEQuantConfig, mxfp4_mxfp8_moe_quant_config, mxfp4_w4a16_moe_quant_config, ocp_mx_moe_quant_config, ) from vllm.model_executor.layers.fused_moe.fused_marlin_moe import ( BatchedMarlinExperts, MarlinExperts, ) from vllm.model_executor.layers.fused_moe.gpt_oss_triton_kernels_moe import ( OAITritonExperts, UnfusedOAITritonExperts, ) from vllm.model_executor.layers.fused_moe.trtllm_moe import TrtLlmGenExperts from vllm.model_executor.layers.linear import LinearBase, UnquantizedLinearMethod from vllm.model_executor.layers.quantization import QuantizationMethods from vllm.model_executor.layers.quantization.base_config import ( QuantizationConfig, QuantizeMethodBase, ) from vllm.model_executor.layers.quantization.utils.marlin_utils import ( get_marlin_input_dtype, ) from vllm.model_executor.layers.quantization.utils.marlin_utils_fp4 import ( prepare_moe_fp4_layer_for_marlin, ) from vllm.model_executor.layers.quantization.utils.mxfp4_utils import ( _can_support_mxfp4, _swizzle_mxfp4, get_padding_alignment, ) from vllm.model_executor.layers.quantization.utils.quant_utils import is_layer_skipped from vllm.model_executor.utils import set_weight_attrs from vllm.platforms import current_platform from vllm.utils.flashinfer import has_flashinfer from vllm.utils.import_utils import has_triton_kernels from vllm.utils.math_utils import round_up logger = init_logger(__name__) # enum for mxfp4 backend class Mxfp4Backend(Enum): NONE = 0 # FlashInfer Backend SM100_FI_MXFP4_MXFP8_TRTLLM = 1 SM100_FI_MXFP4_MXFP8_CUTLASS = 2 SM100_FI_MXFP4_BF16 = 3 SM90_FI_MXFP4_BF16 = 4 # Marlin Backend MARLIN = 5 # Triton Backend TRITON = 6 CK = 7 def get_mxfp4_backend_with_lora() -> Mxfp4Backend: """ Not all MXFP4 backends support LoRA. Select backends that are known to have LoRA support. """ if not current_platform.is_cuda(): return Mxfp4Backend.NONE # If FlashInfer is not available, try either Marlin or Triton triton_kernels_supported = ( has_triton_kernels() # NOTE: triton_kernels are only confirmed to work on SM90 and SM100 # SM110 fails with this error: https://github.com/vllm-project/vllm/issues/29317 # SM120 needs this fix: https://github.com/triton-lang/triton/pull/8498 and (9, 0) <= current_platform.get_device_capability() < (11, 0) ) if envs.VLLM_MXFP4_USE_MARLIN is False and triton_kernels_supported: logger.info_once("[get_mxfp4_backend_with_lora] Using Triton backend") return Mxfp4Backend.TRITON logger.info_once("[get_mxfp4_backend_with_lora] Using Marlin backend") return Mxfp4Backend.MARLIN def get_mxfp4_backend(with_lora_support: bool) -> Mxfp4Backend: # Backend Selection if with_lora_support: return get_mxfp4_backend_with_lora() if current_platform.is_cuda(): if ( current_platform.is_device_capability(90) and has_flashinfer() and envs.VLLM_USE_FLASHINFER_MOE_MXFP4_BF16 ): logger.info_once("Using FlashInfer MXFP4 BF16 backend for SM90") return Mxfp4Backend.SM90_FI_MXFP4_BF16 elif ( current_platform.is_device_capability_family(100) and has_flashinfer() and envs.VLLM_USE_FLASHINFER_MOE_MXFP4_MXFP8_CUTLASS ): logger.info_once("Using FlashInfer MXFP4 MXFP8 CUTLASS backend for SM100") return Mxfp4Backend.SM100_FI_MXFP4_MXFP8_CUTLASS elif ( current_platform.is_device_capability_family(100) and has_flashinfer() and envs.VLLM_USE_FLASHINFER_MOE_MXFP4_MXFP8 ): logger.info_once( "Using FlashInfer MXFP4 MXFP8 TRTLLM backend for SM100", scope="local" ) return Mxfp4Backend.SM100_FI_MXFP4_MXFP8_TRTLLM elif current_platform.is_device_capability_family(100) and has_flashinfer(): logger.info_once( "Using FlashInfer MXFP4 BF16 backend for SM100, " "For faster performance on SM100, consider setting " "VLLM_USE_FLASHINFER_MOE_MXFP4_MXFP8=1, though this may impact " "accuracy." ) return Mxfp4Backend.SM100_FI_MXFP4_BF16 elif ( current_platform.is_device_capability_family(100) or current_platform.is_device_capability(90) ) and not has_flashinfer(): logger.warning_once( "MXFP4 MoE is enabled on Hopper/Blackwell but FlashInfer " "is not available. This may result in degraded performance. " "Please `pip install vllm[flashinfer]` for best results." ) # If FlashInfer is not available, try either Marlin or Triton triton_kernels_supported = ( has_triton_kernels() # NOTE: triton_kernels are only confirmed to work on SM90 and SM100 # SM110 fails with this error: https://github.com/vllm-project/vllm/issues/29317 # SM120 needs this fix: https://github.com/triton-lang/triton/pull/8498 and (9, 0) <= current_platform.get_device_capability() < (11, 0) ) if envs.VLLM_MXFP4_USE_MARLIN or not triton_kernels_supported: logger.info_once("Using Marlin backend") return Mxfp4Backend.MARLIN else: logger.info_once("Using Triton backend") return Mxfp4Backend.TRITON elif current_platform.is_xpu(): logger.info_once("Using xpu backend on XPU") return Mxfp4Backend.MARLIN elif current_platform.is_rocm(): from vllm.platforms.rocm import on_gfx950 if rocm_aiter_ops.is_enabled() and on_gfx950(): logger.info_once("Using CK MXFP4 MoE backend (Aiter ROCm)") return Mxfp4Backend.CK elif has_triton_kernels(): logger.info_once("Using Triton backend") return Mxfp4Backend.TRITON return Mxfp4Backend.NONE class Mxfp4Config(QuantizationConfig): def __init__(self, ignored_layers: list[str] | None = None): super().__init__() self.ignored_layers = ignored_layers @classmethod def from_config(cls, config): return cls() @classmethod def get_min_capability(cls) -> int: return 80 @classmethod def get_name(cls) -> QuantizationMethods: return "mxfp4" @classmethod def get_supported_act_dtypes(cls) -> list[torch.dtype]: return [torch.bfloat16] @classmethod def get_config_filenames(cls) -> list[str]: return [] def get_quant_method( self, layer: torch.nn.Module, prefix: str ) -> "QuantizeMethodBase | None": if isinstance(layer, LinearBase): if self.ignored_layers and is_layer_skipped( prefix=prefix, ignored_layers=self.ignored_layers, fused_mapping=self.packed_modules_mapping, ): return UnquantizedLinearMethod() # TODO: Add support for MXFP4 Linear Method. # MXFP4 LinearMethod is available in AMD-Quark, refer to that implementation # if you are interested in enabling MXFP4 here. logger.debug_once( "MXFP4 linear layer is not implemented - falling back to " "UnquantizedLinearMethod.", scope="local", ) return UnquantizedLinearMethod() elif isinstance(layer, FusedMoE): if current_platform.is_xpu(): return XpuMxfp4MoEMethod(layer.moe_config) else: quant_method = Mxfp4MoEMethod(layer.moe_config) return quant_method elif isinstance(layer, Attention): # TODO: Add support for MXFP4 Attention. logger.debug_once( "MXFP4 attention layer is not implemented. " "Skipping quantization for this layer.", scope="local", ) return None def is_mxfp4_quant(self, prefix: str, layer: torch.nn.Module) -> bool: """MXFP4 config always uses MXFP4 quantization.""" return True class Mxfp4MoEMethod(FusedMoEMethodBase): """MXFP4 MoE quantization method.""" def __init__(self, moe: FusedMoEConfig): super().__init__(moe) self.weight_dtype = "mxfp4" self.mxfp4_backend = get_mxfp4_backend(moe.is_lora_enabled) self.max_capture_size = ( get_current_vllm_config().compilation_config.max_cudagraph_capture_size ) assert self.mxfp4_backend != Mxfp4Backend.NONE, ( f"get_mxfp4_backend(with_lora_support={moe.is_lora_enabled}) found" "no compatible MXFP4 MoE backend (FlashInfer/Marlin/Triton)." "Please check your environment and try again." ) self._cache_permute_indices: dict[torch.Size, torch.Tensor] = {} # Initialized in process_weights_after_loading for CUTLASS/SM90 backends self.moe_mk: mk.FusedMoEModularKernel | None = None def create_weights( self, layer: torch.nn.Module, num_experts: int, hidden_size: int, intermediate_size_per_partition: int, params_dtype: torch.dtype, **extra_weight_attrs, ): self.num_experts = num_experts weight_dtype = torch.uint8 scale_dtype = torch.uint8 # FIXME (zyongye): ship after torch and safetensors support mxfp4 # is_torch_mxfp4_available = ( # hasattr(torch, "float4_e2m1fn_x2") and # hasattr(torch, "float8_e8m0fnu")) # if is_torch_mxfp4_available: # weight_dtype = torch.float4_e2m1fn_x2 # scale_dtype = torch.float8_e8m0fnu mxfp4_block = 32 intermediate_size_per_partition_after_pad = intermediate_size_per_partition if self.mxfp4_backend == Mxfp4Backend.MARLIN: # The moe marlin kernel requires that for each linear # n % 256 == 0 and k % 128 == 0. # In gate_up_proj: # n = 2 * intermediate_size_per_partition_after_pad # k = hidden_size # In down_proj # n = hidden_size # k = intermediate_size_per_partition_after_pad intermediate_size_per_partition_after_pad = round_up( intermediate_size_per_partition, 128 ) if current_platform.is_xpu(): hidden_size = round_up(hidden_size, 128) else: hidden_size = round_up(hidden_size, 256) layer.params_dtype = params_dtype layer.num_experts = num_experts layer.hidden_size = hidden_size layer.intermediate_size_per_partition = ( intermediate_size_per_partition_after_pad ) elif ( self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_MXFP8_TRTLLM or self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_BF16 ): # pad the intermediate size to be a multiple of 2 * mxfp4_block # for to hold non-uniform sharded tensor as well as swizzling # other padding to increase performance intermediate_size_per_partition_after_pad = round_up( intermediate_size_per_partition, 256 ) hidden_size = round_up(hidden_size, 256) elif ( self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_MXFP8_CUTLASS or self.mxfp4_backend == Mxfp4Backend.SM90_FI_MXFP4_BF16 ): intermediate_size_per_partition_after_pad = round_up( intermediate_size_per_partition, 128 ) hidden_size = round_up(hidden_size, 128) elif current_platform.is_rocm(): pad_align = get_padding_alignment() intermediate_size_per_partition_after_pad = round_up( intermediate_size_per_partition, pad_align ) hidden_size = round_up(hidden_size, pad_align) else: intermediate_size_per_partition_after_pad = round_up( intermediate_size_per_partition, 64 ) self.intermediate_size = intermediate_size_per_partition_after_pad self.hidden_size = hidden_size self.hidden_pad = extra_weight_attrs.get("hidden_pad", 0) self.intermediate_pad = ( intermediate_size_per_partition_after_pad - intermediate_size_per_partition ) # Fused gate_up_proj (column parallel) w13_weight = torch.nn.Parameter( torch.zeros( num_experts, 2 * intermediate_size_per_partition_after_pad, hidden_size // 2, dtype=weight_dtype, ), requires_grad=False, ) layer.register_parameter("w13_weight", w13_weight) set_weight_attrs(w13_weight, extra_weight_attrs) w13_weight_scale = torch.nn.Parameter( torch.zeros( num_experts, 2 * intermediate_size_per_partition_after_pad, hidden_size // mxfp4_block, dtype=scale_dtype, ), requires_grad=False, ) layer.register_parameter("w13_weight_scale", w13_weight_scale) set_weight_attrs(w13_weight_scale, extra_weight_attrs) w13_bias = torch.nn.Parameter( torch.zeros( num_experts, 2 * intermediate_size_per_partition_after_pad, dtype=torch.bfloat16, ), requires_grad=False, ) layer.register_parameter("w13_bias", w13_bias) set_weight_attrs(w13_bias, extra_weight_attrs) # down_proj (row parallel) w2_weight = torch.nn.Parameter( torch.zeros( num_experts, hidden_size, intermediate_size_per_partition_after_pad // 2, dtype=weight_dtype, ), requires_grad=False, ) layer.register_parameter("w2_weight", w2_weight) set_weight_attrs(w2_weight, extra_weight_attrs) w2_weight_scale = torch.nn.Parameter( torch.zeros( num_experts, hidden_size, intermediate_size_per_partition_after_pad // mxfp4_block, dtype=scale_dtype, ), requires_grad=False, ) layer.register_parameter("w2_weight_scale", w2_weight_scale) set_weight_attrs(w2_weight_scale, extra_weight_attrs) w2_bias = torch.nn.Parameter( torch.zeros( num_experts, hidden_size, dtype=torch.bfloat16, ), requires_grad=False, ) layer.register_parameter("w2_bias", w2_bias) set_weight_attrs(w2_bias, extra_weight_attrs) def process_weights_after_loading(self, layer): if self.mxfp4_backend == Mxfp4Backend.MARLIN: prepare_moe_fp4_layer_for_marlin( layer, input_dtype=get_marlin_input_dtype() ) self.moe_quant_config = self.get_fused_moe_quant_config(layer) assert self.moe_quant_config is not None prepare_finalize = maybe_make_prepare_finalize( moe=self.moe, quant_config=self.moe_quant_config, routing_tables=layer._maybe_init_expert_routing_tables(), allow_new_interface=True, ) assert prepare_finalize is not None self.moe_mk = mk.FusedMoEModularKernel( prepare_finalize, MarlinExperts( self.moe, self.moe_quant_config, ), inplace=not self.moe.disable_inplace, shared_experts=None, ) elif ( self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_MXFP8_TRTLLM or self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_BF16 ): from flashinfer.fp4_quantization import nvfp4_block_scale_interleave from flashinfer.fused_moe.core import get_w2_permute_indices_with_cache layer.gemm1_alpha = Parameter( torch.tensor([1.702] * self.num_experts, dtype=torch.float32).cuda(), requires_grad=False, ) layer.gemm1_beta = Parameter( torch.tensor([1.0] * self.num_experts, dtype=torch.float32).cuda(), requires_grad=False, ) layer.gemm1_clamp_limit = Parameter( torch.tensor([7.0] * self.num_experts, dtype=torch.float32).cuda(), requires_grad=False, ) sf_block_size = 32 # mxfp4 block size assert ( layer.w13_weight.dim() == 3 and layer.w13_weight.shape[0] == self.num_experts and layer.w13_weight.shape[1] == self.intermediate_size * 2 and layer.w13_weight.shape[2] == self.hidden_size // 2 ) assert ( layer.w13_weight_scale.dim() == 3 and layer.w13_weight_scale.shape[0] == self.num_experts and layer.w13_weight_scale.shape[1] == self.intermediate_size * 2 and layer.w13_weight_scale.shape[2] == self.hidden_size // sf_block_size ) assert ( layer.w2_weight.dim() == 3 and layer.w2_weight.shape[0] == self.num_experts and layer.w2_weight.shape[1] == self.hidden_size and layer.w2_weight.shape[2] == self.intermediate_size // 2 ) assert ( layer.w2_weight_scale.dim() == 3 and layer.w2_weight_scale.shape[1] == self.hidden_size and layer.w2_weight_scale.shape[2] == self.intermediate_size // sf_block_size ) assert ( layer.w13_bias.dim() == 2 and layer.w13_bias.shape[0] == self.num_experts and layer.w13_bias.shape[1] == self.intermediate_size * 2 ) assert ( layer.w2_bias.dim() == 2 and layer.w2_bias.shape[0] == self.num_experts and layer.w2_bias.shape[1] == self.hidden_size ) w13_weight_scale = layer.w13_weight_scale.data w2_weight_scale = layer.w2_weight_scale.data w13_weight = layer.w13_weight.data w2_weight = layer.w2_weight.data w13_bias = layer.w13_bias.data.to(torch.float32) w2_bias = layer.w2_bias.data.to(torch.float32) # Swap w1 and w3 as the definition of # swiglu is different in the trtllm-gen def swap_every_two_rows(x, axis=-1): shape = x.shape if axis < 0: axis = len(shape) + axis # Create a new shape with pairs swapped along specified axis new_shape = list(shape) new_shape[axis] = shape[axis] // 2 new_shape.insert(axis + 1, 2) # Reshape to expose pairs, swap them, and reshape back x = x.reshape(*new_shape) x = x.flip(axis + 1) new_shape = list(shape) return x.reshape(*new_shape) w13_weight_scale = swap_every_two_rows(w13_weight_scale, -2) w13_weight = swap_every_two_rows(w13_weight, -2) w13_bias = swap_every_two_rows(w13_bias, -1) # Do not interleave as the checkpoint is already interleaved # Shuffle weights and scaling factors for transposed mma output gemm1_weights_mxfp4_shuffled = [] gemm1_scales_mxfp4_shuffled = [] gemm2_weights_mxfp4_shuffled = [] gemm2_scales_mxfp4_shuffled = [] gemm1_bias_shuffled = [] gemm2_bias_shuffled = [] epilogue_tile_m = 128 # FIXME: this depends on the kernel internals for i in range(self.num_experts): # w13 weight shuffling permute_indices = get_w2_permute_indices_with_cache( self._cache_permute_indices, w13_weight[i].view(torch.uint8), epilogue_tile_m, ) gemm1_weights_mxfp4_shuffled.append( w13_weight[i] .view(torch.uint8)[permute_indices.to(w13_weight.device)] .contiguous() ) # w13 scale shuffling permute_sf_indices = get_w2_permute_indices_with_cache( self._cache_permute_indices, w13_weight_scale[i].view(torch.uint8), epilogue_tile_m, num_elts_per_sf=16, ) gemm1_scales_mxfp4_shuffled.append( nvfp4_block_scale_interleave( w13_weight_scale[i] .view(torch.uint8)[ permute_sf_indices.to(w13_weight_scale.device) ] .contiguous() ) ) # w13 bias shuffling permute_bias_indices = get_w2_permute_indices_with_cache( self._cache_permute_indices, w13_bias[i].clone().reshape(-1, 1), epilogue_tile_m, ) gemm1_bias_shuffled.append( w13_bias[i] .clone() .reshape(-1, 1)[permute_bias_indices.to(w13_bias.device)] .contiguous() ) # w2 weight shuffling permute_indices = get_w2_permute_indices_with_cache( self._cache_permute_indices, w2_weight[i].view(torch.uint8), epilogue_tile_m, ) gemm2_weights_mxfp4_shuffled.append( w2_weight[i] .view(torch.uint8)[permute_indices.to(w2_weight.device)] .contiguous() ) # w2 scale shuffling permute_sf_indices = get_w2_permute_indices_with_cache( self._cache_permute_indices, w2_weight_scale[i].view(torch.uint8), epilogue_tile_m, num_elts_per_sf=16, ) gemm2_scales_mxfp4_shuffled.append( nvfp4_block_scale_interleave( w2_weight_scale[i] .view(torch.uint8)[ permute_sf_indices.to(w2_weight_scale.device) ] .contiguous() ) ) # w2 bias shuffling permute_indices = get_w2_permute_indices_with_cache( self._cache_permute_indices, w2_bias[i].clone().reshape(-1, 1), epilogue_tile_m, ) gemm2_bias_shuffled.append( w2_bias[i] .clone() .reshape(-1, 1)[permute_indices.to(w2_bias.device)] .contiguous() ) w13_weight = torch.stack(gemm1_weights_mxfp4_shuffled) w13_weight_scale = ( torch.stack(gemm1_scales_mxfp4_shuffled) .reshape( self.num_experts, 2 * self.intermediate_size, self.hidden_size // sf_block_size, ) .view(torch.float8_e4m3fn) ) w2_weight = torch.stack(gemm2_weights_mxfp4_shuffled) w2_weight_scale = ( torch.stack(gemm2_scales_mxfp4_shuffled) .reshape( self.num_experts, self.hidden_size, self.intermediate_size // sf_block_size, ) .view(torch.float8_e4m3fn) ) layer.w13_weight = Parameter(w13_weight, requires_grad=False) layer.w13_weight_scale = Parameter(w13_weight_scale, requires_grad=False) layer.w2_weight = Parameter(w2_weight, requires_grad=False) layer.w2_weight_scale = Parameter(w2_weight_scale, requires_grad=False) layer.w13_bias = Parameter( torch.stack(gemm1_bias_shuffled).reshape(self.num_experts, -1), requires_grad=False, ) layer.w2_bias = Parameter( torch.stack(gemm2_bias_shuffled).reshape(self.num_experts, -1), requires_grad=False, ) elif ( self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_MXFP8_CUTLASS or self.mxfp4_backend == Mxfp4Backend.SM90_FI_MXFP4_BF16 ): sf_block_size = 32 # mxfp4 block size # Common shape assertions assert ( layer.w13_weight.dim() == 3 and layer.w13_weight.shape[0] == self.num_experts and layer.w13_weight.shape[1] == self.intermediate_size * 2 and layer.w13_weight.shape[2] == self.hidden_size // 2 ) assert ( layer.w13_weight_scale.dim() == 3 and layer.w13_weight_scale.shape[0] == self.num_experts and layer.w13_weight_scale.shape[1] == self.intermediate_size * 2 and layer.w13_weight_scale.shape[2] == self.hidden_size // sf_block_size ) assert ( layer.w2_weight.dim() == 3 and layer.w2_weight.shape[0] == self.num_experts and layer.w2_weight.shape[1] == self.hidden_size and layer.w2_weight.shape[2] == self.intermediate_size // 2 ) assert ( layer.w2_weight_scale.dim() == 3 and layer.w2_weight_scale.shape[1] == self.hidden_size and layer.w2_weight_scale.shape[2] == self.intermediate_size // sf_block_size ) assert ( layer.w13_bias.dim() == 2 and layer.w13_bias.shape[0] == self.num_experts and layer.w13_bias.shape[1] == self.intermediate_size * 2 ) assert ( layer.w2_bias.dim() == 2 and layer.w2_bias.shape[0] == self.num_experts and layer.w2_bias.shape[1] == self.hidden_size ) # De-interleave and swap for w13 weight, bias, and scales w13_w = layer.w13_weight.data gate_w, up_w = w13_w[:, ::2, :], w13_w[:, 1::2, :] deinterleaved_w13_w = torch.cat([gate_w, up_w], dim=1) w1_w, w3_w = torch.chunk(deinterleaved_w13_w, 2, dim=1) w13_weight_swapped = torch.cat([w3_w, w1_w], dim=1) w13_b = layer.w13_bias.data.to(torch.float32) gate_b, up_b = w13_b[:, ::2], w13_b[:, 1::2] deinterleaved_w13_b = torch.cat([gate_b, up_b], dim=1) b1, b3 = torch.chunk(deinterleaved_w13_b, 2, dim=-1) w13_bias_swapped = torch.cat([b3, b1], dim=-1).to(torch.bfloat16) w13_s = layer.w13_weight_scale.data gate_s, up_s = w13_s[:, ::2, :], w13_s[:, 1::2, :] deinterleaved_w13_s = torch.cat([gate_s, up_s], dim=1) s1, s3 = torch.chunk(deinterleaved_w13_s, 2, dim=1) w13_scale_swapped = torch.cat([s3, s1], dim=1) if self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_MXFP8_CUTLASS: from flashinfer import block_scale_interleave orig_shape = w13_scale_swapped.shape w13_scale_interleaved = block_scale_interleave( w13_scale_swapped.view(torch.uint8) ).reshape(orig_shape) w2_s = layer.w2_weight_scale.data orig_shape = w2_s.shape w2_scale_interleaved = block_scale_interleave( w2_s.view(torch.uint8) ).reshape(orig_shape) layer.w13_weight = Parameter(w13_weight_swapped, requires_grad=False) layer.w13_weight_scale = Parameter( w13_scale_interleaved, requires_grad=False ) layer.w13_bias = Parameter(w13_bias_swapped, requires_grad=False) layer.w2_weight_scale = Parameter( w2_scale_interleaved, requires_grad=False ) elif self.mxfp4_backend == Mxfp4Backend.SM90_FI_MXFP4_BF16: def _interleave_mxfp4_cutlass_sm90(w): w_shape = w.shape w_interleaved = w.reshape( w_shape[0], w_shape[1], (w_shape[2] // 4), 4 ) w_interleaved = w_interleaved.permute(0, 2, 1, 3) w_interleaved = w_interleaved.reshape( w_shape[0], w_shape[2] // 4, w_shape[1] * 4 ) return w_interleaved w31_scales = w13_scale_swapped.to(torch.uint8).view(torch.uint8) w31_scales_interleaved = _interleave_mxfp4_cutlass_sm90(w31_scales) w2_weight_scale = layer.w2_weight_scale.data w2_scales = w2_weight_scale.to(torch.uint8).view(torch.uint8) w2_scales_interleaved = _interleave_mxfp4_cutlass_sm90(w2_scales) layer.w13_weight = torch.nn.Parameter( torch.cat([w3_w, w1_w], dim=1), requires_grad=False ) layer.w13_bias = torch.nn.Parameter( w13_bias_swapped, requires_grad=False ) layer.w13_weight_scale = torch.nn.Parameter( w31_scales_interleaved, requires_grad=False ) layer.w2_weight_scale = torch.nn.Parameter( w2_scales_interleaved, requires_grad=False ) # theses two kernels go through the `flashinfer_cutlass_fused_moe` path from vllm.model_executor.layers.fused_moe.flashinfer_cutlass_moe import ( FlashInferExperts, ) self.moe_quant_config = self.get_fused_moe_quant_config(layer) assert self.moe_quant_config is not None prepare_finalize = maybe_make_prepare_finalize( moe=self.moe, quant_config=self.moe_quant_config, routing_tables=layer._maybe_init_expert_routing_tables(), allow_new_interface=True, ) assert prepare_finalize is not None self.moe_mk = mk.FusedMoEModularKernel( prepare_finalize, FlashInferExperts( moe_config=self.moe, quant_config=self.moe_quant_config, ), shared_experts=None, ) elif self.mxfp4_backend == Mxfp4Backend.CK: if layer.w13_bias is not None: layer.w13_bias.data = layer.w13_bias.data.to(torch.float32) if layer.w2_bias.data is not None: layer.w2_bias.data = layer.w2_bias.data.to(torch.float32) e, n, k = layer.w13_weight.shape layer.w13_weight.view(torch.uint8).copy_( layer.w13_weight.data.view(torch.uint8) .view(e, n // 2, 2, k) .permute(0, 2, 1, 3) .contiguous() .view(e, n, k) ) layer.w13_weight_scale.data = ( layer.w13_weight_scale.data.view(e, n // 2, 2, -1) .permute(0, 2, 1, 3) .contiguous() .view(e, n, -1) ) layer.w13_weight.data = layer.w13_weight.data.view(torch.float4_e2m1fn_x2) layer.w2_weight.data = layer.w2_weight.data.view(torch.float4_e2m1fn_x2) layer.w13_weight.data = rocm_aiter_ops.shuffle_weight_a16w4( layer.w13_weight, 16, True ) shuffled_w13_scale = rocm_aiter_ops.shuffle_scale_a16w4( layer.w13_weight_scale.view(-1, layer.w13_weight_scale.shape[-1]), self.num_experts, True, ) layer.w2_weight.data = rocm_aiter_ops.shuffle_weight_a16w4( layer.w2_weight, 16, False ) shuffled_w2_scale = rocm_aiter_ops.shuffle_scale_a16w4( layer.w2_weight_scale.view(-1, layer.w2_weight_scale.shape[-1]), self.num_experts, False, ) layer.w13_bias.data = ( layer.w13_bias.data.view(-1, n // 2, 2) .permute(0, 2, 1) .contiguous() .view(-1, n) ) layer.w13_weight_scale = torch.nn.Parameter( shuffled_w13_scale, requires_grad=False ) layer.w2_weight_scale = torch.nn.Parameter( shuffled_w2_scale, requires_grad=False ) # replace_parameter(layer, "w13_bias", w13_bias) # replace_parameter(layer, "w13_weight_scale", w13_weight_scale) # replace_parameter(layer, "w2_weight_scale", w2_weight_scale) # replace_parameter(layer, "w13_weight", w13_weight) # replace_parameter(layer, "w2_weight", w2_weight) elif self.mxfp4_backend == Mxfp4Backend.TRITON: from triton_kernels.matmul_ogs import FlexCtx, PrecisionConfig w13_bias = layer.w13_bias.to(torch.float32) w2_bias = layer.w2_bias.to(torch.float32) layer.w13_bias = Parameter(w13_bias, requires_grad=False) layer.w2_bias = Parameter(w2_bias, requires_grad=False) # Ideally we'd use FusedMoEModularKernel.prepare_finalize object # (stored in self.fused_experts) to determine if the MoE has a # batched activation format. As self.fused_experts is not # initialized at this point, we resort to checking the MoE config # directly. is_batched_moe = self.moe.use_deepep_ll_kernels if is_batched_moe: num_warps = 4 if envs.VLLM_MOE_DP_CHUNK_SIZE <= 512 else 8 else: num_warps = 8 w13_weight, w13_flex, w13_scale = _swizzle_mxfp4( layer.w13_weight, layer.w13_weight_scale, num_warps ) w2_weight, w2_flex, w2_scale = _swizzle_mxfp4( layer.w2_weight, layer.w2_weight_scale, num_warps ) self.w13_precision_config = PrecisionConfig( weight_scale=w13_scale, flex_ctx=FlexCtx(rhs_data=w13_flex) ) self.w2_precision_config = PrecisionConfig( weight_scale=w2_scale, flex_ctx=FlexCtx(rhs_data=w2_flex) ) self.w13_weight = w13_weight self.w2_weight = w2_weight del layer.w13_weight del layer.w2_weight layer.w13_weight = w13_weight layer.w2_weight = w2_weight else: raise ValueError( f"Unsupported mxfp4_backend: {self.mxfp4_backend}: " f"should be one of: {list(Mxfp4Backend)}." ) def get_fused_moe_quant_config( self, layer: torch.nn.Module ) -> FusedMoEQuantConfig | None: if self.mxfp4_backend == Mxfp4Backend.MARLIN: return mxfp4_w4a16_moe_quant_config( w1_bias=layer.w13_bias, w2_bias=layer.w2_bias, w1_scale=layer.w13_weight_scale, w2_scale=layer.w2_weight_scale, ) elif self.mxfp4_backend == Mxfp4Backend.TRITON: w1_scale = self.w13_precision_config w2_scale = self.w2_precision_config return mxfp4_w4a16_moe_quant_config( w1_bias=layer.w13_bias, w2_bias=layer.w2_bias, w1_scale=w1_scale, w2_scale=w2_scale, ) elif self.mxfp4_backend in [ Mxfp4Backend.SM100_FI_MXFP4_MXFP8_TRTLLM, Mxfp4Backend.SM100_FI_MXFP4_MXFP8_CUTLASS, ]: return mxfp4_mxfp8_moe_quant_config( w1_bias=layer.w13_bias, w2_bias=layer.w2_bias, w1_scale=layer.w13_weight_scale, w2_scale=layer.w2_weight_scale, ) elif self.mxfp4_backend in [ Mxfp4Backend.SM100_FI_MXFP4_BF16, Mxfp4Backend.SM90_FI_MXFP4_BF16, Mxfp4Backend.CK, ]: return mxfp4_w4a16_moe_quant_config( w1_bias=layer.w13_bias, w2_bias=layer.w2_bias, w1_scale=layer.w13_weight_scale, w2_scale=layer.w2_weight_scale, ) else: w1_scale = layer.w13_weight_scale w2_scale = layer.w2_weight_scale return ocp_mx_moe_quant_config( quant_dtype="mxfp4", w1_bias=layer.w13_bias, w2_bias=layer.w2_bias, w1_scale=w1_scale, w2_scale=w2_scale, ) def select_gemm_impl( self, prepare_finalize: mk.FusedMoEPrepareAndFinalize, layer: torch.nn.Module, ) -> mk.FusedMoEPermuteExpertsUnpermute: if ( prepare_finalize.activation_format == mk.FusedMoEActivationFormat.BatchedExperts ): if self.mxfp4_backend == Mxfp4Backend.MARLIN: max_num_tokens_per_rank = prepare_finalize.max_num_tokens_per_rank() assert max_num_tokens_per_rank is not None assert self.moe_quant_config is not None return BatchedMarlinExperts( max_num_tokens=max_num_tokens_per_rank, num_dispatchers=prepare_finalize.num_dispatchers(), quant_config=self.moe_quant_config, moe_config=self.moe, ) else: raise NotImplementedError( f"Incompatible Mxfp4 backend ({self.mxfp4_backend}) for " "EP batched experts format" ) else: assert self.moe_quant_config is not None if ( self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_MXFP8_TRTLLM or self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_BF16 ): # B200 code-path kwargs = { # TODO(bnell): part of quant_config "max_capture_size": self.max_capture_size, } return TrtLlmGenExperts(self.moe, self.moe_quant_config, **kwargs) elif self.mxfp4_backend == Mxfp4Backend.MARLIN: return MarlinExperts(self.moe, self.moe_quant_config) elif self.mxfp4_backend == Mxfp4Backend.TRITON: if self.moe.is_lora_enabled: return UnfusedOAITritonExperts(self.moe, self.moe_quant_config) return OAITritonExperts(self.moe, self.moe_quant_config) else: raise NotImplementedError( f"Incompatible Mxfp4 backend ({self.mxfp4_backend}) for EP" ) @property def is_monolithic(self) -> bool: if self.moe.is_lora_enabled: return False return ( self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_MXFP8_TRTLLM or self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_BF16 or self.mxfp4_backend == Mxfp4Backend.TRITON or self.mxfp4_backend == Mxfp4Backend.CK ) def apply( self, layer: FusedMoE, x: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, shared_experts_input: torch.Tensor | None, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: assert not self.is_monolithic if layer.enable_eplb: raise NotImplementedError("EPLB is not supported for mxfp4") assert _can_support_mxfp4( layer.use_grouped_topk, layer.topk_group, layer.num_expert_group, layer.expert_map, layer.custom_routing_function, layer.e_score_correction_bias, layer.apply_router_weight_on_input, layer.scoring_func, layer.activation, layer.eplb_state.expert_load_view, layer.eplb_state.logical_to_physical_map, layer.eplb_state.logical_replica_count, ), "MXFP4 are not supported with this configuration." assert ( self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_MXFP8_CUTLASS or self.mxfp4_backend == Mxfp4Backend.SM90_FI_MXFP4_BF16 or self.mxfp4_backend == Mxfp4Backend.MARLIN ) assert self.moe_mk is not None return self.moe_mk( hidden_states=x, w1=layer.w13_weight, w2=layer.w2_weight, topk_weights=topk_weights, topk_ids=topk_ids, activation=layer.activation, global_num_experts=layer.global_num_experts, apply_router_weight_on_input=layer.apply_router_weight_on_input, expert_map=layer.expert_map, shared_experts_input=shared_experts_input, ) def apply_monolithic( self, layer: FusedMoE, x: torch.Tensor, router_logits: torch.Tensor, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: assert self.is_monolithic if layer.enable_eplb: raise NotImplementedError("EPLB is not supported for mxfp4") assert _can_support_mxfp4( layer.use_grouped_topk, layer.topk_group, layer.num_expert_group, layer.expert_map, layer.custom_routing_function, layer.e_score_correction_bias, layer.apply_router_weight_on_input, layer.scoring_func, layer.activation, layer.eplb_state.expert_load_view, layer.eplb_state.logical_to_physical_map, layer.eplb_state.logical_replica_count, ), "MXFP4 are not supported with this configuration." if ( self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_MXFP8_TRTLLM or self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_BF16 ): from flashinfer import trtllm_fp4_block_scale_moe if self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_BF16: assert x.dtype == torch.bfloat16 x_quant = x x_scale = None elif self.mxfp4_backend == Mxfp4Backend.SM100_FI_MXFP4_MXFP8_TRTLLM: from flashinfer import mxfp8_quantize x_quant, x_scale = mxfp8_quantize(x, False) # to mxfp8 x_scale = x_scale.view(torch.float8_e4m3fn).reshape(*x.shape[:-1], -1) trtllm_gen_output = trtllm_fp4_block_scale_moe( routing_logits=router_logits.to(torch.bfloat16), routing_bias=None, hidden_states=x_quant, hidden_states_scale=x_scale, gemm1_weights=layer.w13_weight, # uint8 (e2m1 x 2) gemm1_weights_scale=layer.w13_weight_scale, # uint8 (e4m3 x 2) gemm1_bias=layer.w13_bias, # fp32 per expert per channel gemm1_alpha=layer.gemm1_alpha, # fp32 per expert gemm1_beta=layer.gemm1_beta, # fp32 per expert gemm1_clamp_limit=layer.gemm1_clamp_limit, # fp32 per expert gemm2_weights=layer.w2_weight, # uint8 (e2m1 x 2) gemm2_weights_scale=layer.w2_weight_scale, # ue8m0 gemm2_bias=layer.w2_bias, # fp32 per expert per channel output1_scale_scalar=None, output1_scale_gate_scalar=None, output2_scale_scalar=None, num_experts=layer.global_num_experts, top_k=layer.top_k, n_group=None, topk_group=None, intermediate_size=self.intermediate_size, # padded to multiple of 256 local_expert_offset=layer.ep_rank * layer.local_num_experts, local_num_experts=self.num_experts, routed_scaling_factor=None, routing_method_type=1 if layer.renormalize else 0, do_finalize=True, tune_max_num_tokens=max(self.max_capture_size, 1), )[0] return trtllm_gen_output elif self.mxfp4_backend == Mxfp4Backend.CK: topk_weights, topk_ids = rocm_aiter_ops.fused_topk( x, router_logits, layer.top_k, True ) output = rocm_aiter_ops.fused_moe( x, layer.w13_weight, layer.w2_weight, topk_weights, topk_ids, activation_method=rocm_aiter_ops.get_aiter_activation_type("swiglu"), quant_method=rocm_aiter_ops.get_aiter_quant_type("per_1x32"), w1_scale=layer.w13_weight_scale, w2_scale=layer.w2_weight_scale, doweight_stage1=False, hidden_pad=self.hidden_pad // 128 * 128, intermediate_pad=self.intermediate_pad // 64 * 64 * 2, bias1=layer.w13_bias, bias2=layer.w2_bias, ) return output elif self.mxfp4_backend == Mxfp4Backend.TRITON: from vllm.model_executor.layers.fused_moe.gpt_oss_triton_kernels_moe import ( # noqa: E501 triton_kernel_moe_forward, ) return triton_kernel_moe_forward( hidden_states=x, w1=layer.w13_weight, w2=layer.w2_weight, gating_output=router_logits, topk=layer.top_k, renormalize=layer.renormalize, global_num_experts=layer.global_num_experts, expert_map=layer.expert_map, quant_config=self.moe_quant_config, apply_router_weight_on_input=layer.apply_router_weight_on_input, ) else: raise ValueError(f"Unsupported backend: {self.mxfp4_backend}") class XpuMxfp4MoEMethod(Mxfp4MoEMethod): def __init__(self, moe_config: FusedMoEConfig): super().__init__(moe_config) self.moe_config = moe_config def create_weights( self, layer: torch.nn.Module, num_experts: int, hidden_size: int, intermediate_size_per_partition: int, params_dtype: torch.dtype, **extra_weight_attrs, ): super().create_weights( layer, num_experts, hidden_size, intermediate_size_per_partition, params_dtype, **extra_weight_attrs, ) self.original_hidden_size = hidden_size def process_weights_after_loading(self, layer: torch.nn.Module) -> None: pass @property def is_monolithic(self) -> bool: return True def apply_monolithic( self, layer: FusedMoE, x: torch.Tensor, router_logits: torch.Tensor, ) -> torch.Tensor: assert layer.activation == MoEActivation.SWIGLUOAI, ( "Only swiglu_oai activation is supported for " f"XPU MXFP4 MoE, not {layer.activation}." ) from vllm_xpu_kernels.fused_moe_interface import xpu_fused_moe M, _ = x.size() routing_weights = torch.empty( M, layer.top_k, dtype=torch.float32, device=x.device ) selected_experts = torch.empty( M, layer.top_k, dtype=torch.int32, device=x.device ) token_expert_indices = torch.empty( M, layer.top_k, dtype=torch.int32, device=x.device ) if layer.use_grouped_topk: routing_weights, selected_experts = torch.ops._moe_C.fused_grouped_topk( x, router_logits, layer.top_k, layer.renormalize, n_expert_group=layer.num_expert_group, n_topk_group=layer.topk_group, scoring_func=layer.scoring_func, routed_scaling_factor=layer.routed_scaling_factor, bias=layer.e_score_correction_bias, ) else: torch.ops._moe_C.topk_softmax( routing_weights, selected_experts, token_expert_indices, router_logits, layer.renormalize, layer.e_score_correction_bias, ) return xpu_fused_moe( hidden_states=x, w13=layer.w13_weight, w13_bias=layer.w13_bias if self.moe.has_bias else None, w13_scales=layer.w13_weight_scale, w2=layer.w2_weight, w2_bias=layer.w2_bias if self.moe.has_bias else None, w2_scales=layer.w2_weight_scale, topk_weights=routing_weights, topk_ids=selected_experts, n_experts_per_token=layer.top_k, activation=layer.activation.value, num_experts=layer.local_num_experts, is_mxfp4=True, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/mxfp4.py", "license": "Apache License 2.0", "lines": 1154, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/reasoning/gptoss_reasoning_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json from collections.abc import Sequence from transformers import PreTrainedTokenizerBase from vllm.entrypoints.mcp.tool_server import ToolServer from vllm.entrypoints.openai.chat_completion.protocol import ( ChatCompletionRequest, ) from vllm.entrypoints.openai.engine.protocol import DeltaMessage from vllm.entrypoints.openai.parser.harmony_utils import parse_chat_output from vllm.logger import init_logger from vllm.reasoning import ReasoningParser logger = init_logger(__name__) no_func_reaonsing_tag = { "type": "structural_tag", "format": { "type": "triggered_tags", "tags": [ { "begin": "<|channel|>analysis<|message|>", "content": {"type": "any_text"}, "end": "<|end|>", } ], "triggers": ["<|channel|>analysis"], "stop_after_first": False, }, } def from_builtin_tool_to_tag(tool: str) -> list[dict]: tag = [ { "begin": f"<|channel|>commentary to={tool}", "content": {"type": "any_text"}, "end": "<|end|>", }, { "begin": f"<|channel|>analysis to={tool}", "content": {"type": "any_text"}, "end": "<|end|>", }, ] return tag def tag_with_builtin_funcs(no_func_reaonsing_tag, builtin_tool_list: list[str]) -> dict: import copy new_tag = copy.deepcopy(no_func_reaonsing_tag) new_tag["format"]["triggers"].append("<|channel|>commentary to=") for tool in builtin_tool_list: new_tag["format"]["tags"].extend(from_builtin_tool_to_tag(tool)) return new_tag class GptOssReasoningParser(ReasoningParser): """ Reasoning parser for GptOss model. The GptOss model uses harmony to extract reasoning content and this parser is only used for detecting the end of the reasoning content. """ def __init__(self, tokenizer: PreTrainedTokenizerBase, *args, **kwargs): super().__init__(tokenizer, *args, **kwargs) # The model can output some special tokens between "final" and "<|message|>" # So we need to look for both sequences to determine the end of reasoning. self.reasoning_end_token_ids_prefix = self.model_tokenizer.encode( "<|channel|>final" ) self.reasoning_end_token_ids_suffix = self.model_tokenizer.encode("<|message|>") # We also need to check for the <|end|> token to avoid false positives from # previous messages in multi-turn conversations. self.eom_token_id = self.model_tokenizer.vocab["<|end|>"] self.reasoning_max_num_between_tokens = 20 def is_reasoning_end(self, input_ids: Sequence[int]) -> bool: end_token_ids_prefix = self.reasoning_end_token_ids_prefix end_token_ids_suffix = self.reasoning_end_token_ids_suffix assert len(end_token_ids_prefix) > 0, "reasoning_end_token_ids_prefix is empty" assert len(end_token_ids_suffix) > 0, "reasoning_end_token_ids_suffix is empty" # Check if the end sequence is present in the input_ids. # We search from the end of input_ids to find the last match. for i in range(len(input_ids) - len(end_token_ids_prefix), -1, -1): if input_ids[i] == self.eom_token_id: # We looped backwards far enough to find the end of a previous message, # which means we have searched the entirety of the current message # and can exit early without searching further back into prior # messages of the conversation. return False if input_ids[i : i + len(end_token_ids_prefix)] == end_token_ids_prefix: # We have found the prefix, now we look for the suffix after the prefix. suffix_start = i + len(end_token_ids_prefix) for j in range( suffix_start, len(input_ids) - len(end_token_ids_suffix) + 1 ): if j - suffix_start >= self.reasoning_max_num_between_tokens: break if ( input_ids[j : j + len(end_token_ids_suffix)] == end_token_ids_suffix ): return True return False def extract_content_ids(self, input_ids: list[int]) -> list[int]: _, content, _ = parse_chat_output(input_ids) if content is None: return [] return self.model_tokenizer.encode(content) def extract_reasoning_streaming( self, previous_text: str, current_text: str, delta_text: str, previous_token_ids: Sequence[int], current_token_ids: Sequence[int], delta_token_ids: Sequence[int], ) -> DeltaMessage | None: prev_reasoning, prev_content, _ = parse_chat_output(list(previous_token_ids)) cur_reasoning, cur_content, _ = parse_chat_output(list(current_token_ids)) reasoning_delta = None content_delta = None if cur_reasoning is not None: prev_r = prev_reasoning or "" if cur_reasoning.startswith(prev_r): reasoning_delta = cur_reasoning[len(prev_r) :] or None else: reasoning_delta = cur_reasoning if cur_content is not None: prev_c = prev_content or "" if cur_content.startswith(prev_c): content_delta = cur_content[len(prev_c) :] or None else: content_delta = cur_content if reasoning_delta is None and content_delta is None: return None return DeltaMessage(reasoning=reasoning_delta, content=content_delta) def extract_reasoning( self, model_output: str, request: ChatCompletionRequest, ) -> tuple[str | None, str | None]: raise NotImplementedError( "gpt-oss has a special branch for parsing reasoning in non-streaming mode. This method shouldn't be used." # noqa: E501 ) # This function prepares the structural tag to format reasoning output def prepare_structured_tag( self, original_tag: str | None, tool_server: ToolServer | None ) -> str | None: if original_tag is None: if tool_server is None: return json.dumps(no_func_reaonsing_tag) else: builtin_tool_list: list[str] = [] if tool_server.has_tool("browser"): builtin_tool_list.append("browser") if tool_server.has_tool("python"): builtin_tool_list.append("python") if tool_server.has_tool("container"): builtin_tool_list.append("container") if len(builtin_tool_list) > 0: logger.info("Builtin_tool_list: %s", builtin_tool_list) func_tag = json.dumps( tag_with_builtin_funcs(no_func_reaonsing_tag, builtin_tool_list) ) else: logger.info("Builtin_tool_list is empty") func_tag = json.dumps(no_func_reaonsing_tag) return func_tag else: # There is potential risk for appending the tag to the original tag return original_tag
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/reasoning/gptoss_reasoning_parser.py", "license": "Apache License 2.0", "lines": 164, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/gpt_oss.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import typing from collections.abc import Callable, Iterable import torch import torch.distributed as dist from torch import nn from transformers import GptOssConfig from vllm.compilation.decorators import support_torch_compile from vllm.config import CacheConfig, VllmConfig from vllm.distributed import ( get_dp_group, get_ep_group, get_pcp_group, get_pp_group, get_tensor_model_parallel_rank, get_tensor_model_parallel_world_size, tensor_model_parallel_all_gather, ) from vllm.model_executor.layers.attention import Attention from vllm.model_executor.layers.fused_moe import FusedMoE from vllm.model_executor.layers.fused_moe.config import FusedMoEParallelConfig from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( QKVParallelLinear, ReplicatedLinear, RowParallelLinear, ) from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.quantization.utils.ocp_mx_utils import OCP_MX_BLOCK_SIZE from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.model_executor.layers.utils import rocm_unquantized_gemm from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import ( default_weight_loader, maybe_remap_kv_scale_name, ) from vllm.model_executor.models.utils import sequence_parallel_chunk from vllm.platforms import current_platform from vllm.sequence import IntermediateTensors from vllm.utils.math_utils import cdiv from vllm.v1.attention.backend import AttentionType from .interfaces import SupportsEagle3, SupportsLoRA, SupportsPP from .utils import ( AutoWeightsLoader, WeightsMapper, extract_layer_index, is_pp_missing_parameter, make_empty_intermediate_tensors_factory, make_layers, maybe_prefix, ) class OAIAttention(nn.Module): def __init__( self, config: GptOssConfig, quant_config: QuantizationConfig | None = None, cache_config: CacheConfig | None = None, prefix: str = "", ): super().__init__() self.layer_idx = extract_layer_index(prefix) self.head_dim = config.head_dim self.num_attention_heads = config.num_attention_heads self.num_key_value_heads = config.num_key_value_heads self.hidden_size = config.hidden_size self.rotary_emb = get_rope( self.head_dim, max_position=config.max_position_embeddings, dtype=torch.float32, rope_parameters={ "rope_theta": config.rope_parameters["rope_theta"], "rope_type": "yarn", "factor": config.rope_parameters["factor"], "original_max_position_embeddings": config.rope_parameters[ "original_max_position_embeddings" ], "beta_fast": config.rope_parameters["beta_fast"], "beta_slow": config.rope_parameters["beta_slow"], "truncate": config.rope_parameters.get("truncate", True), }, is_neox_style=True, ) tp_size = get_tensor_model_parallel_world_size() self.sinks = torch.nn.Parameter( torch.empty(config.num_attention_heads // tp_size, requires_grad=False) ) self.q_size = self.num_attention_heads * self.head_dim // tp_size self.kv_size = self.num_key_value_heads * self.head_dim // tp_size self.scaling = self.head_dim**-0.5 self.qkv_proj = QKVParallelLinear( hidden_size=self.hidden_size, head_size=self.head_dim, total_num_heads=self.num_attention_heads, total_num_kv_heads=self.num_key_value_heads, bias=True, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", ) self.o_proj = RowParallelLinear( input_size=self.num_attention_heads * self.head_dim, output_size=self.hidden_size, bias=True, quant_config=quant_config, prefix=f"{prefix}.o_proj", ) self.num_local_attention_heads = config.num_attention_heads // tp_size self.num_local_key_value_heads = config.num_key_value_heads // tp_size # Only apply sliding window to every other layer sliding_window = config.sliding_window if self.layer_idx % 2 == 0 else None self.attn = Attention( self.num_local_attention_heads, self.head_dim, self.scaling, num_kv_heads=self.num_local_key_value_heads, cache_config=cache_config, quant_config=quant_config, per_layer_sliding_window=sliding_window, attn_type=AttentionType.DECODER, prefix=f"{prefix}.attn", sinks=self.sinks, ) def forward( self, hidden_states: torch.Tensor, positions: torch.Tensor ) -> torch.Tensor: qkv, _ = self.qkv_proj(hidden_states) q, k, v = qkv.split([self.q_size, self.kv_size, self.kv_size], dim=-1) q, k = self.rotary_emb(positions, q, k) attn_output = self.attn(q, k, v) output, _ = self.o_proj(attn_output) return output class MLPBlock(torch.nn.Module): def __init__( self, vllm_config: VllmConfig, layer_idx: int, prefix: str = "", ): super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config parallel_config = vllm_config.parallel_config self.is_sequence_parallel = parallel_config.use_sequence_parallel_moe self.layer_idx = layer_idx self.num_experts = config.num_local_experts self.hidden_size = config.hidden_size self.experts_per_token = config.num_experts_per_tok self.world_size = dist.get_world_size() if dist.is_initialized() else 1 self.router = ReplicatedLinear( config.hidden_size, config.num_local_experts, bias=True, quant_config=None, prefix=f"{prefix}.router", return_bias=False, ) assert config.intermediate_size % self.world_size == 0 self.experts = FusedMoE( num_experts=config.num_local_experts, top_k=config.num_experts_per_tok, hidden_size=config.hidden_size, intermediate_size=config.intermediate_size, reduce_results=True, renormalize=True, quant_config=quant_config, prefix=f"{prefix}.experts", apply_router_weight_on_input=False, has_bias=True, activation="swigluoai", is_sequence_parallel=self.is_sequence_parallel, ) def forward(self, x: torch.Tensor) -> torch.Tensor: num_tokens = x.shape[0] if self.is_sequence_parallel: x = sequence_parallel_chunk(x) if current_platform.is_rocm(): g = rocm_unquantized_gemm( self, x[:, : self.hidden_size], self.router.weight, self.router.bias ) else: g = self.router(x) x = self.experts(hidden_states=x, router_logits=g)[:, : self.hidden_size] if self.is_sequence_parallel: x = tensor_model_parallel_all_gather(x.contiguous(), 0) x = x[:num_tokens] return x class TransformerBlock(torch.nn.Module): def __init__( self, vllm_config: VllmConfig, quant_config: QuantizationConfig, prefix: str = "", ): super().__init__() config = vllm_config.model_config.hf_config cache_config = vllm_config.cache_config self.layer_idx = extract_layer_index(prefix) self.attn = OAIAttention( config, prefix=f"{prefix}.attn", quant_config=quant_config, cache_config=cache_config, ) self.mlp = MLPBlock(vllm_config, self.layer_idx, prefix=f"{prefix}.mlp") self.input_layernorm = RMSNorm(config.hidden_size, eps=1e-5) self.post_attention_layernorm = RMSNorm(config.hidden_size, eps=1e-5) def forward( self, hidden_states: torch.Tensor, positions: torch.Tensor, residual: torch.Tensor | None, ) -> torch.Tensor: # Self Attention if residual is None: residual = hidden_states hidden_states = self.input_layernorm(hidden_states) else: hidden_states, residual = self.input_layernorm(hidden_states, residual) hidden_states = self.attn(hidden_states, positions) # Fully Connected hidden_states, residual = self.post_attention_layernorm(hidden_states, residual) output = self.mlp(hidden_states) return output, residual @support_torch_compile class GptOssModel(nn.Module): def __init__( self, *, vllm_config: VllmConfig, prefix: str = "", ): super().__init__() self.config = vllm_config.model_config.hf_config self.quant_config = vllm_config.quant_config self.parallel_config = vllm_config.parallel_config self.config.hidden_size = self.config.hidden_size self.embedding = VocabParallelEmbedding( self.config.vocab_size, self.config.hidden_size, ) self.start_layer, self.end_layer, self.layers = make_layers( self.config.num_hidden_layers, lambda prefix: TransformerBlock( vllm_config, prefix=prefix, quant_config=self.quant_config, ), prefix=f"{prefix}.layers", ) self.norm = RMSNorm(self.config.hidden_size, eps=1e-5) self.make_empty_intermediate_tensors = make_empty_intermediate_tensors_factory( ["hidden_states", "residual"], self.config.hidden_size ) self.aux_hidden_state_layers = tuple[int, ...]() def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embedding(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor: if get_pp_group().is_first_rank: if inputs_embeds is not None: x = inputs_embeds else: x = self.embed_input_ids(input_ids) residual = None else: assert intermediate_tensors is not None x = intermediate_tensors["hidden_states"] residual = intermediate_tensors["residual"] aux_hidden_states = [] for i in range(self.start_layer, self.end_layer): layer = self.layers[i] if i in self.aux_hidden_state_layers: aux_hidden_states.append(x if residual is None else x + residual) x, residual = layer(x, positions, residual) if not get_pp_group().is_last_rank: return IntermediateTensors({"hidden_states": x, "residual": residual}) x, _ = self.norm(x, residual) if len(aux_hidden_states) > 0: return x, aux_hidden_states return x def get_expert_mapping(self) -> list[tuple[str, str, int, str]]: # Params for weights, weight scales, activation scales # (param_name, weight_name, expert_id, shard_id) # NOTE: this is only used for quark. return FusedMoE.make_expert_params_mapping( self, ckpt_gate_proj_name="w1", ckpt_down_proj_name="w2", ckpt_up_proj_name="w3", num_experts=self.config.num_local_experts, num_redundant_experts=0, ) def _load_weights_mxfp4( self, ep_rank_end: int, ep_rank_start: int, heads_per_rank: int, head_start: int, weights: Iterable[tuple[str, torch.Tensor]], stacked_params_mapping: list[tuple[str, ...]], ) -> set[str]: params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() use_ep = self.parallel_config.enable_expert_parallel num_experts = self.config.num_local_experts # In MoE, we need to flatten the tensor parallel size across the data # parallel size when EP is disabled. tp_size, tp_rank = FusedMoEParallelConfig.flatten_tp_across_dp_and_pcp( tp_size=get_tensor_model_parallel_world_size(), dp_size=get_dp_group().world_size, dp_rank=get_dp_group().rank_in_group, pcp_size=get_pcp_group().world_size, pcp_rank=get_pcp_group().rank_in_group, ) intermediate_size = self.config.intermediate_size intermediate_size_block = intermediate_size // OCP_MX_BLOCK_SIZE per_rank_intermediate_size_block = cdiv(intermediate_size_block, tp_size) per_rank_intermediate_size = ( per_rank_intermediate_size_block * OCP_MX_BLOCK_SIZE ) # Calculate common slicing bounds for current rank tp_rank_start = tp_rank * per_rank_intermediate_size tp_rank_end = min((tp_rank + 1) * per_rank_intermediate_size, intermediate_size) for name, weight in weights: # Skip layers on other devices. if is_pp_missing_parameter(name, self): continue if ".w13_weight_scale" in name: # Handle MLP gate and up projection weights scale if use_ep: narrow_weight = weight[ep_rank_start:ep_rank_end, ...] else: narrow_weight = weight[:, 2 * tp_rank_start : 2 * tp_rank_end, ...] param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader( param, narrow_weight, weight_name=name, shard_id=None, expert_id=None, ) loaded_params.add(name) continue elif ".w2_weight_scale" in name: # Handle MLP down projection weights if use_ep: narrow_weight = weight[ep_rank_start:ep_rank_end, ...] else: narrow_weight = weight[ ..., tp_rank_start // OCP_MX_BLOCK_SIZE : tp_rank_end // OCP_MX_BLOCK_SIZE, ] param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader( param, narrow_weight, weight_name=name, shard_id=None, expert_id=None, ) loaded_params.add(name) continue elif ".w13_weight" in name: # Handle MLP gate and up projection weights # flat weight from (E, 2 * N, block_size, entry_per_block) # to (E, 2 * N, -1), shouldn't trigger copy for contiguous weight = weight.view( num_experts, 2 * intermediate_size, -1 ).contiguous() # Extract gate and up projection parts # since the weight is shuffled, we can slice directly if use_ep: narrow_weight = weight[ep_rank_start:ep_rank_end, ...] else: narrow_weight = weight[:, 2 * tp_rank_start : 2 * tp_rank_end, ...] param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader( param, narrow_weight, weight_name=name, shard_id=None, expert_id=None, ) loaded_params.add(name) continue elif ".w2_weight" in name: # Handle MLP down projection weights # same flatten here, but since 2 mx4 value are packed in 1 # uint8, divide by 2 weight = weight.view( num_experts, -1, intermediate_size // 2 ).contiguous() if use_ep: narrow_weight = weight[ep_rank_start:ep_rank_end, ...] else: narrow_weight = weight[..., tp_rank_start // 2 : tp_rank_end // 2] param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader( param, narrow_weight, weight_name=name, shard_id=None, expert_id=None, ) loaded_params.add(name) continue elif ".w13_bias" in name: # Handle MLP gate and up projection biases # Extract gate and up projection bias parts if use_ep: narrow_weight = weight[ep_rank_start:ep_rank_end, ...] else: narrow_weight = weight[:, 2 * tp_rank_start : 2 * tp_rank_end] param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader( param, narrow_weight, weight_name=name, shard_id=None, expert_id=None, ) loaded_params.add(name) continue elif ".w2_bias" in name: # Handle MLP down projection bias param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) if use_ep: weight = weight[ep_rank_start:ep_rank_end, ...] else: # (only load on rank 0 to avoid duplication) if tp_rank != 0: weight.zero_() weight_loader( param, weight, weight_name=name, shard_id=None, expert_id=None ) loaded_params.add(name) continue elif "sinks" in name: # Handle attention sinks (distributed across ranks) param = params_dict[name] narrow_weight = weight.narrow(0, head_start, heads_per_rank) param.data.copy_(narrow_weight) loaded_params.add(name) continue for param_name, weight_name, shard_id in stacked_params_mapping: if weight_name not in name: continue name = name.replace(weight_name, param_name) param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) if weight_loader == default_weight_loader: weight_loader(param, weight) else: weight_loader(param, weight, shard_id) break else: # Handle all other weights with potential renaming if name not in params_dict: continue param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, weight) loaded_params.add(name) return loaded_params def _load_weights_quark( self, ep_rank_end: int, ep_rank_start: int, heads_per_rank: int, head_start: int, weights: Iterable[tuple[str, torch.Tensor]], stacked_params_mapping: list[tuple[str, ...]], ) -> set[str]: params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() use_ep = self.parallel_config.enable_expert_parallel num_experts = self.config.num_local_experts if use_ep: tp_rank = get_tensor_model_parallel_rank() tp_size = get_tensor_model_parallel_world_size() else: tp_size, tp_rank = FusedMoEParallelConfig.flatten_tp_across_dp_and_pcp( tp_size=get_tensor_model_parallel_world_size(), dp_size=get_dp_group().world_size, dp_rank=get_dp_group().rank_in_group, pcp_size=get_pcp_group().world_size, pcp_rank=get_pcp_group().rank_in_group, ) def _get_moe_weight_dtype(layer_id: int = 0) -> str | None: """Helper function to get MoE quantization weight dtype. Args: layer_id: Layer index to check (default 0, as all layers should have the same quantization method) Returns: Weight dtype string (e.g., "mxfp4", "fp8") or None if not available """ if hasattr(self.layers[layer_id].mlp.experts.quant_method, "weight_dtype"): return self.layers[layer_id].mlp.experts.quant_method.weight_dtype return None intermediate_size = self.config.intermediate_size moe_weight_dtype = _get_moe_weight_dtype(layer_id=0) if moe_weight_dtype == "mxfp4": # MXFP4 requires OCP_MX_BLOCK_SIZE alignment intermediate_size_block = intermediate_size // OCP_MX_BLOCK_SIZE per_rank_intermediate_size_block = cdiv(intermediate_size_block, tp_size) per_rank_intermediate_size = ( per_rank_intermediate_size_block * OCP_MX_BLOCK_SIZE ) else: # FP8 and other formats don't need alignment per_rank_intermediate_size = cdiv(intermediate_size, tp_size) tp_rank_start = tp_rank * per_rank_intermediate_size tp_rank_end = min((tp_rank + 1) * per_rank_intermediate_size, intermediate_size) expert_params_mapping = self.get_expert_mapping() for name, loaded_weight in weights: if is_pp_missing_parameter(name, self): continue layer_id, expert_id, fused_name = None, None, None moe_quant_method = None if "experts" in name: parts = name.split(".") ids = [s for s in parts if s.isdigit()] # for amd-quark format that each expert is separated # need to extract the parameter name with experts fused. # example model: amd/gpt-oss-20b-MoE-Quant-W-MXFP4-A-FP8-KV-FP8 if len(ids) == 2: layer_id, expert_id = int(ids[0]), int(ids[-1]) parts.pop(len(parts) - 1 - parts[::-1].index(str(expert_id))) fused_name = ".".join(parts) # for openai mxfp4 format that all experts are combined # no need to extract the parameter name with experts fused. # models: openai/gpt-oss-20b, openai/gpt-oss-120b elif len(ids) == 1: layer_id, expert_id = int(ids[0]), None fused_name = name else: raise NameError( f"Layer {name} contains more than 2 numeric indices. This is " "an unexpected condition. Please open an issue if encountered." ) moe_quant_method = _get_moe_weight_dtype(layer_id=layer_id) def kv_cache_scale_loader( quant_config: QuantizationConfig, name: str, params_dict: dict[str, typing.Any], weight: torch.Tensor, default_weight_loader: Callable[..., None], loaded_params: set[str], ) -> tuple[bool, set[str]]: """ Load KV cache output scales. Returns: Tuple of (bool, set): - bool: True if KV-cache scale was loaded into loaded_params - set: Updated set of loaded_params if True else the original set """ # load explicit cached KV output scale from quant_config if quant_config is not None and ( scale_name := quant_config.get_cache_scale(name) ): param = params_dict[scale_name] weight_loader = getattr( param, "weight_loader", default_weight_loader ) if weight.numel() != 1: raise ValueError( f"KV cache scale '{scale_name}' is expected to be a " f"scalar, but got a tensor of shape {weight.shape}." ) # Ensure weight is a scalar before passing to loader. weight_loader(param, weight.flatten()[0]) loaded_params.add(scale_name) return True, loaded_params return False, loaded_params load_kv_cache_scale_completed, loaded_params = kv_cache_scale_loader( self.quant_config, name, params_dict, loaded_weight, default_weight_loader, loaded_params, ) if load_kv_cache_scale_completed: continue if ( all(key in name for key in ["input_scale", "mlp.experts"]) and expert_id is not None ): assert loaded_weight.numel() == 1 expert_data = params_dict[fused_name].data[expert_id] expert_data.copy_(loaded_weight) loaded_params.add(fused_name) continue # Unified handler for mxfp4 weights and scales elif moe_quant_method == "mxfp4" and any( name.endswith(suffix) for suffix in [ ".w13_weight_scale", ".w2_weight_scale", ".w13_weight", ".w2_weight", ] ): is_w13 = ".w13_" in name is_scale = "_scale" in name # Reshape weight for mxfp4 if needed (not for scales) if not is_scale and expert_id is None: if is_w13: if loaded_weight.dim() < 3: raise ValueError( f"Expected w13_weight to have at least 3 " f"dimensions, got shape " f"{loaded_weight.shape}" ) if loaded_weight.shape[0] != num_experts: raise ValueError( f"Expected w13_weight first dimension to be " f"{num_experts}, got " f"{loaded_weight.shape[0]}" ) loaded_weight = loaded_weight.view( num_experts, 2 * intermediate_size, -1 ).contiguous() else: if loaded_weight.dim() < 3: raise ValueError( f"Expected w2_weight to have at least 3 " f"dimensions, got shape " f"{loaded_weight.shape}" ) if loaded_weight.shape[0] != num_experts: raise ValueError( f"Expected w2_weight first dimension to be " f"{num_experts}, got " f"{loaded_weight.shape[0]}" ) loaded_weight = loaded_weight.view( num_experts, -1, intermediate_size // 2 ).contiguous() if use_ep: sliced_weight = loaded_weight[ep_rank_start:ep_rank_end, ...] else: if is_w13: if expert_id is None: sliced_weight = loaded_weight[ :, 2 * tp_rank_start : 2 * tp_rank_end, ... ] else: sliced_weight = loaded_weight[ 2 * tp_rank_start : 2 * tp_rank_end, ... ] else: if is_scale: sliced_weight = loaded_weight[ ..., tp_rank_start // OCP_MX_BLOCK_SIZE : tp_rank_end // OCP_MX_BLOCK_SIZE, ] else: sliced_weight = loaded_weight[ ..., tp_rank_start // 2 : tp_rank_end // 2 ] # NOTE(rob): because gpt-oss ckpt has "unique" structure with # fused gate_up_proj fused on disk, we cannot use the existing # weight loaders without added complexity, so just do the # direct load here. param = params_dict[fused_name] expert_data = param.data[expert_id] dim1 = sliced_weight.shape[0] dim2 = sliced_weight.shape[1] expert_data.data[:dim1, :dim2].copy_(sliced_weight) loaded_params.add(fused_name) continue elif name.endswith(".w13_weight") and moe_quant_method == "fp8": if use_ep: narrow_weight = loaded_weight[ep_rank_start:ep_rank_end, ...] else: if expert_id is None: narrow_weight = loaded_weight[ :, 2 * tp_rank_start : 2 * tp_rank_end, : ] else: narrow_weight = loaded_weight[ 2 * tp_rank_start : 2 * tp_rank_end, : ] assert fused_name is not None param = params_dict[fused_name] if expert_id is None: param.data.copy_(narrow_weight) else: param.data[expert_id].copy_(narrow_weight) loaded_params.add(fused_name) continue elif name.endswith(".w13_weight_scale") and moe_quant_method == "fp8": assert fused_name is not None param = params_dict[fused_name] # Check if this is per-channel or per-tensor scale if loaded_weight.numel() > 1 and loaded_weight.dim() == 1: if use_ep: narrow_weight = loaded_weight[ep_rank_start:ep_rank_end, ...] else: narrow_weight = loaded_weight[ 2 * tp_rank_start : 2 * tp_rank_end ] else: narrow_weight = loaded_weight if expert_id is None: param.data.copy_(narrow_weight) else: param.data[expert_id].copy_(narrow_weight) loaded_params.add(fused_name) continue elif name.endswith(".w13_input_scale") and moe_quant_method == "fp8": assert fused_name is not None param = params_dict[fused_name] if expert_id is None: param.data.copy_(loaded_weight) else: param.data[expert_id].copy_(loaded_weight) loaded_params.add(fused_name) continue elif name.endswith(".w2_weight") and moe_quant_method == "fp8": if use_ep: narrow_weight = loaded_weight[ep_rank_start:ep_rank_end, ...] else: if expert_id is None: narrow_weight = loaded_weight[..., tp_rank_start:tp_rank_end] else: narrow_weight = loaded_weight[..., tp_rank_start:tp_rank_end] assert fused_name is not None param = params_dict[fused_name] if expert_id is None: param.data.copy_(narrow_weight) else: param.data[expert_id].copy_(narrow_weight) loaded_params.add(fused_name) continue elif name.endswith(".w2_weight_scale") and moe_quant_method == "fp8": assert fused_name is not None param = params_dict[fused_name] if use_ep: narrow_weight = loaded_weight[ep_rank_start:ep_rank_end, ...] else: narrow_weight = loaded_weight if expert_id is None: param.data.copy_(narrow_weight) else: param.data[expert_id].copy_(narrow_weight) loaded_params.add(fused_name) continue # Unified handler for bias loading (w13_bias and w2_bias) elif name.endswith(".w13_bias") or name.endswith(".w2_bias"): is_w13_bias = name.endswith(".w13_bias") if use_ep: sliced_weight = loaded_weight[ep_rank_start:ep_rank_end, ...] else: if is_w13_bias: if expert_id is None: sliced_weight = loaded_weight[ :, 2 * tp_rank_start : 2 * tp_rank_end ] else: sliced_weight = loaded_weight[ 2 * tp_rank_start : 2 * tp_rank_end ] else: sliced_weight = loaded_weight if tp_rank != 0: sliced_weight = sliced_weight.zero_() # NOTE(rob): because gpt-oss ckpt has "unique" structure with # fused gate_up_proj fused on disk, we cannot use the existing # weight loaders without added complexity, so just do the # direct load here. assert fused_name is not None param = params_dict[fused_name] expert_data = param.data[expert_id] dim1 = sliced_weight.shape[0] expert_data.data[:dim1].copy_(sliced_weight) loaded_params.add(fused_name) continue elif "sinks" in name: # Handle attention sinks (distributed across ranks) param = params_dict[name] narrow_weight = loaded_weight.narrow(0, head_start, heads_per_rank) param.data.copy_(narrow_weight) loaded_params.add(name) continue for param_name, weight_name, shard_id in stacked_params_mapping: # Skip non-stacked layers and experts (experts handled below). if weight_name not in name: continue # We have mlp.experts[0].gate_proj in the checkpoint. # Since we handle the experts below in expert_params_mapping, # we need to skip here BEFORE we update the name, otherwise # name will be updated to mlp.experts[0].gate_up_proj, which # will then be updated below in expert_params_mapping # for mlp.experts[0].gate_gate_up_proj, which breaks load. if ("mlp.experts." in name) and name not in params_dict: continue name = name.replace(weight_name, param_name) if name.endswith("scale"): # Remapping the name of FP8 kv-scale. name = maybe_remap_kv_scale_name(name, params_dict) if name is None: continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) loaded_params.add(name) break else: for mapping in expert_params_mapping: # Anyway, this is an expert weight and should not be # attempted to load as other weights later param_name, weight_name, mapping_expert_id, shard_id = mapping weight_name = ( weight_name[:-1] if weight_name.endswith(".") else weight_name ) if weight_name not in name: continue param = params_dict[fused_name] # We should ask the weight loader to return success or not # here since otherwise we may skip experts with other # available replicas. weight_loader = typing.cast( Callable[..., bool], param.weight_loader ) # Use checkpoint's expert_id for quark format (when expert_id # is extracted from weight name), otherwise use mapping's expert_id actual_expert_id = ( expert_id if expert_id is not None else mapping_expert_id ) success = weight_loader( param, loaded_weight, fused_name, shard_id=shard_id, expert_id=actual_expert_id, return_success=True, ) if success: name = fused_name loaded_params.add(name) break else: if name not in params_dict: continue param = params_dict[name] weight_loader = getattr( param, "weight_loader", default_weight_loader ) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params def _load_weights_other( self, ep_rank_end: int, ep_rank_start: int, heads_per_rank: int, head_start: int, weights: Iterable[tuple[str, torch.Tensor]], stacked_params_mapping: list[tuple[str, ...]], ) -> set[str]: params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() use_ep = self.parallel_config.enable_expert_parallel # In MoE, we need to flatten the tensor parallel size across the data # parallel size when EP is disabled. tp_size, tp_rank = FusedMoEParallelConfig.flatten_tp_across_dp_and_pcp( tp_size=get_tensor_model_parallel_world_size(), dp_size=get_dp_group().world_size, dp_rank=get_dp_group().rank_in_group, pcp_size=get_pcp_group().world_size, pcp_rank=get_pcp_group().rank_in_group, ) intermediate_size = self.config.intermediate_size per_rank_intermediate_size = cdiv(intermediate_size, tp_size) # Calculate common slicing bounds for current rank tp_rank_start = tp_rank * per_rank_intermediate_size tp_rank_end = min((tp_rank + 1) * per_rank_intermediate_size, intermediate_size) for name, weight in weights: # Skip layers on other devices. if is_pp_missing_parameter(name, self): continue if ".w13_weight" in name: # Handle MLP gate and up projection weights # Extract gate and up projection parts if use_ep: narrow_weight = weight[ep_rank_start:ep_rank_end, ...] else: narrow_weight = weight[:, :, 2 * tp_rank_start : 2 * tp_rank_end] narrow_weight = narrow_weight.permute(0, 2, 1).contiguous() param = params_dict[name] param.copy_(narrow_weight) loaded_params.add(name) continue elif ".w2_weight" in name: # Handle MLP down projection weights if use_ep: narrow_weight = weight[ep_rank_start:ep_rank_end, ...] else: narrow_weight = weight[:, tp_rank_start:tp_rank_end, :] narrow_weight = narrow_weight.permute(0, 2, 1).contiguous() param = params_dict[name] param.copy_(narrow_weight) loaded_params.add(name) continue elif ".w13_bias" in name: # Handle MLP gate and up projection biases # Extract gate and up projection bias parts if use_ep: narrow_weight = weight[ep_rank_start:ep_rank_end, ...] else: narrow_weight = weight[:, 2 * tp_rank_start : 2 * tp_rank_end] param = params_dict[name] param.copy_(narrow_weight) loaded_params.add(name) continue elif ".w2_bias" in name: # Handle MLP down projection bias if use_ep: weight = weight[ep_rank_start:ep_rank_end, ...] else: # (only load on rank 0 to avoid duplication) if tp_rank != 0: weight.zero_() param = params_dict[name] param.copy_(weight) loaded_params.add(name) continue elif "sinks" in name: # Handle attention sinks (distributed across ranks) param = params_dict[name] narrow_weight = weight.narrow(0, head_start, heads_per_rank) param.data.copy_(narrow_weight) loaded_params.add(name) continue for param_name, weight_name, shard_id in stacked_params_mapping: if weight_name not in name: continue name = name.replace(weight_name, param_name) param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) if weight_loader == default_weight_loader: weight_loader(param, weight) else: weight_loader(param, weight, shard_id) break else: # Handle all other weights with potential renaming if name not in params_dict: continue param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, weight) loaded_params.add(name) return loaded_params def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ # (param_name, shard_name, shard_id) (".qkv_proj", ".q_proj", "q"), (".qkv_proj", ".k_proj", "k"), (".qkv_proj", ".v_proj", "v"), ] tp_rank = get_tensor_model_parallel_rank() tp_size = get_tensor_model_parallel_world_size() # Attention heads per rank heads_per_rank = self.config.num_attention_heads // tp_size head_start = tp_rank * heads_per_rank ep_size = get_ep_group().world_size ep_rank = get_ep_group().rank num_experts = self.config.num_local_experts experts_per_rank = num_experts // ep_size ep_rank_start = ep_rank * experts_per_rank ep_rank_end = (ep_rank + 1) * experts_per_rank quant_method = ( self.config.quantization_config["quant_method"] if hasattr(self.config, "quantization_config") else None ) if quant_method == "mxfp4": return self._load_weights_mxfp4( ep_rank_end, ep_rank_start, heads_per_rank, head_start, weights, stacked_params_mapping, ) elif quant_method == "quark": return self._load_weights_quark( ep_rank_end, ep_rank_start, heads_per_rank, head_start, weights, stacked_params_mapping, ) else: return self._load_weights_other( ep_rank_end, ep_rank_start, heads_per_rank, head_start, weights, stacked_params_mapping, ) class GptOssForCausalLM(nn.Module, SupportsPP, SupportsEagle3, SupportsLoRA): is_3d_moe_weight: bool = True packed_modules_mapping = {"qkv_proj": ["q_proj", "k_proj", "v_proj"]} hf_to_vllm_mapper = WeightsMapper( orig_to_new_substr={ ".self_attn.": ".attn.", }, orig_to_new_suffix={ ".embed_tokens.weight": ".embedding.weight", # MoE MXFP4 weights ".gate_up_proj_blocks": ".w13_weight", ".down_proj_blocks": ".w2_weight", ".gate_up_proj_scales": ".w13_weight_scale", ".down_proj_scales": ".w2_weight_scale", # MoE other weights ".gate_up_proj": ".w13_weight", ".down_proj": ".w2_weight", # MoE Bias ".gate_up_proj_bias": ".w13_bias", ".down_proj_bias": ".w2_bias", # For quark format ".gate_up_proj.weight": ".w13_weight", ".gate_up_proj.weight_scale": ".w13_weight_scale", ".gate_up_proj.bias": ".w13_bias", ".gate_up_proj.input_scale": ".w13_input_scale", ".down_proj.weight": ".w2_weight", ".down_proj.weight_scale": ".w2_weight_scale", ".down_proj.bias": ".w2_bias", ".down_proj.input_scale": ".w2_input_scale", }, ) def __init__( self, vllm_config: VllmConfig, prefix: str = "", ): super().__init__() self.vllm_config = vllm_config self.config = vllm_config.model_config.hf_config self.model = GptOssModel( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model"), ) self.lm_head = ParallelLMHead( self.config.vocab_size, self.config.hidden_size, prefix=maybe_prefix(prefix, "lm_head"), ) self.logits_processor = LogitsProcessor(self.config.vocab_size) self.make_empty_intermediate_tensors = ( self.model.make_empty_intermediate_tensors ) def set_aux_hidden_state_layers(self, layers: tuple[int, ...]) -> None: self.model.aux_hidden_state_layers = layers def get_eagle3_aux_hidden_state_layers(self) -> tuple[int, ...]: num_layers = len(self.model.layers) return (2, num_layers // 2, num_layers - 3) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor: return self.model(input_ids, positions, intermediate_tensors, inputs_embeds) def compute_logits(self, hidden_states: torch.Tensor) -> torch.Tensor: logits = self.logits_processor(self.lm_head, hidden_states) return logits def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader( self, skip_prefixes=(["lm_head."] if self.config.tie_word_embeddings else None), ) return loader.load_weights(weights, mapper=self.hf_to_vllm_mapper)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/gpt_oss.py", "license": "Apache License 2.0", "lines": 1099, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/transformers_utils/configs/ovis.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # ruff: noqa: E501 # adapted from https://huggingface.co/AIDC-AI/Ovis2-1B/blob/main/configuration_aimv2.py # and https://huggingface.co/AIDC-AI/Ovis2-1B/blob/main/configuration_ovis.py # Ovis Config with AimV2 config registration removed for Transformers compatibility from typing import Any from transformers import AutoConfig, PretrainedConfig class AIMv2Config(PretrainedConfig): """This is the configuration class to store the configuration of an [`AIMv2Model`]. Instantiating a configuration with the defaults will yield a similar configuration to that of the [apple/aimv2-large-patch14-224](https://huggingface.co/apple/aimv2-large-patch14-224). Args: hidden_size: Dimension of the hidden representations. intermediate_size: Dimension of the SwiGLU representations. num_hidden_layers: Number of hidden layers in the Transformer. num_attention_heads: Number of attention heads for each attention layer in the Transformer. num_channels: Number of input channels. image_size: Image size. patch_size: Patch size. rms_norm_eps: Epsilon value used for the RMS normalization layer. attention_dropout: Dropout ratio for attention probabilities. projection_dropout: Dropout ratio for the projection layer after the attention. qkv_bias: Whether to add a bias to the queries, keys and values. use_bias: Whether to add a bias in the feed-forward and projection layers. kwargs: Keyword arguments for the [`PretrainedConfig`]. """ model_type: str = "aimv2" def __init__( self, hidden_size: int = 1024, intermediate_size: int = 2816, num_hidden_layers: int = 24, num_attention_heads: int = 8, num_channels: int = 3, image_size: int = 224, patch_size: int = 14, rms_norm_eps: float = 1e-5, attention_dropout: float = 0.0, projection_dropout: float = 0.0, qkv_bias: bool = False, use_bias: bool = False, **kwargs: Any, ): super().__init__(**kwargs) self.hidden_size = hidden_size self.intermediate_size = intermediate_size self.num_hidden_layers = num_hidden_layers self.num_attention_heads = num_attention_heads self.num_channels = num_channels self.patch_size = patch_size self.image_size = image_size self.attention_dropout = attention_dropout self.rms_norm_eps = rms_norm_eps self.projection_dropout = projection_dropout self.qkv_bias = qkv_bias self.use_bias = use_bias # ---------------------------------------------------------------------- # Visual Tokenizer Configuration # ---------------------------------------------------------------------- class BaseVisualTokenizerConfig(PretrainedConfig): def __init__( self, vocab_size=16384, tokenize_function="softmax", tau=1.0, depths=None, drop_cls_token=False, backbone_config: PretrainedConfig | dict | None = None, hidden_stride: int = 1, **kwargs, ): super().__init__(**kwargs) self.vocab_size = vocab_size self.tokenize_function = tokenize_function self.tau = tau if isinstance(depths, str): depths = [int(x) for x in depths.split("|")] self.depths = depths self.backbone_kwargs = dict[str, Any]() self.drop_cls_token = drop_cls_token if backbone_config is not None: assert isinstance(backbone_config, (PretrainedConfig, dict)), ( f"expect `backbone_config` to be instance of PretrainedConfig or dict, but got {type(backbone_config)} type" ) if not isinstance(backbone_config, PretrainedConfig): model_type = backbone_config["model_type"] if model_type != "aimv2": backbone_config.pop("model_type") backbone_config = AutoConfig.for_model( model_type, **backbone_config ) else: backbone_config = AIMv2Config(**backbone_config) self.backbone_config = backbone_config self.hidden_stride = hidden_stride class Aimv2VisualTokenizerConfig(BaseVisualTokenizerConfig): model_type = "aimv2_visual_tokenizer" def __init__(self, **kwargs): super().__init__(**kwargs) if self.drop_cls_token: self.drop_cls_token = False if self.depths: assert len(self.depths) == 1 self.backbone_kwargs["num_hidden_layers"] = self.depths[0] class SiglipVisualTokenizerConfig(BaseVisualTokenizerConfig): model_type = "siglip_visual_tokenizer" def __init__(self, **kwargs): super().__init__(**kwargs) if self.drop_cls_token: self.drop_cls_token = False if self.depths: assert len(self.depths) == 1 self.backbone_kwargs["num_hidden_layers"] = self.depths[0] AutoConfig.register("siglip_visual_tokenizer", SiglipVisualTokenizerConfig) AutoConfig.register("aimv2_visual_tokenizer", Aimv2VisualTokenizerConfig) # ---------------------------------------------------------------------- # Ovis Configuration # ---------------------------------------------------------------------- class OvisConfig(PretrainedConfig): model_type = "ovis" def __init__( self, llm_config: PretrainedConfig | dict | None = None, visual_tokenizer_config: PretrainedConfig | dict | None = None, multimodal_max_length=8192, hidden_size=None, conversation_formatter_class=None, llm_attn_implementation=None, disable_tie_weight=False, **kwargs, ): super().__init__(**kwargs) if llm_config is not None: assert isinstance(llm_config, (PretrainedConfig, dict)), ( f"expect `llm_config` to be instance of PretrainedConfig or dict, but got {type(llm_config)} type" ) if not isinstance(llm_config, PretrainedConfig): model_type = llm_config["model_type"] llm_config.pop("model_type") llm_config = AutoConfig.for_model(model_type, **llm_config) # map llm_config to text_config self.text_config = llm_config if visual_tokenizer_config is not None: assert isinstance(visual_tokenizer_config, (PretrainedConfig, dict)), ( f"expect `visual_tokenizer_config` to be instance of PretrainedConfig or dict, but got {type(visual_tokenizer_config)} type" ) if not isinstance(visual_tokenizer_config, PretrainedConfig): model_type = visual_tokenizer_config["model_type"] visual_tokenizer_config.pop("model_type") visual_tokenizer_config = AutoConfig.for_model( model_type, **visual_tokenizer_config ) self.visual_tokenizer_config = visual_tokenizer_config self.multimodal_max_length = multimodal_max_length self.hidden_size = hidden_size self.conversation_formatter_class = conversation_formatter_class self.llm_attn_implementation = llm_attn_implementation self.disable_tie_weight = disable_tie_weight
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/configs/ovis.py", "license": "Apache License 2.0", "lines": 160, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:benchmarks/kernels/benchmark_trtllm_prefill_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import csv import os from datetime import datetime import flashinfer import torch from vllm.utils.math_utils import round_up FLOAT32_BYTES = torch.finfo(torch.float).bits // 8 FP8_DTYPE = torch.float8_e4m3fn FP4_DTYPE = torch.uint8 def to_float8(x, dtype=torch.float8_e4m3fn): finfo = torch.finfo(dtype) min_val, max_val = x.aminmax() amax = torch.maximum(min_val.abs(), max_val.abs()).clamp(min=1e-12) scale = finfo.max / amax * 0.1 x_scl_sat = (x * scale).clamp(min=finfo.min, max=finfo.max) return x_scl_sat.to(dtype), scale.float().reciprocal() @torch.no_grad() def benchmark_prefill( dtype: torch.dtype, quant_dtypes: tuple[torch.dtype | None, torch.dtype | None, torch.dtype | None], batch_size: int, max_seq_len: int, num_heads: tuple[int, int] = (64, 8), head_size: int = 128, kv_layout: str = "HND", block_size: int = 16, warmup: int = 10, trials: int = 20, ): torch.set_default_device("cuda") torch.manual_seed(0) q_quant_dtype, kv_quant_dtype, o_quant_dtype = quant_dtypes q_quant_dtype = q_quant_dtype or dtype kv_quant_dtype = kv_quant_dtype or dtype o_quant_dtype = o_quant_dtype or dtype max_q_len = max_kv_len = max_seq_len num_qo_heads, num_kv_heads = num_heads assert num_qo_heads % num_kv_heads == 0 sm_scale = float(1.0 / (head_size**0.5)) # large number to reduce kv_cache reuse NUM_BLOCKS = int(256000 / block_size) kv_cache_shape = None if kv_layout == "NHD": kv_cache_shape = (NUM_BLOCKS, 2, block_size, num_kv_heads, head_size) elif kv_layout == "HND": kv_cache_shape = (NUM_BLOCKS, 2, num_kv_heads, block_size, head_size) else: raise ValueError(f"Invalid kv_layout: {kv_layout}") q_lens = torch.randint(1, max_q_len, (batch_size,), dtype=torch.int32) q_lens[-1] = max_q_len q_indptr = torch.cat( [ torch.tensor([0], dtype=torch.int32), torch.cumsum(q_lens, dim=0, dtype=torch.int32), ] ) # Always using 1.0 scale to reflect the real perf in benchmarking q_scale = 1.0 ref_query = torch.randn( torch.sum(q_lens).item(), num_qo_heads, head_size, dtype=dtype ) if q_quant_dtype == FP8_DTYPE: query, _ = to_float8(ref_query) else: query = ref_query kv_lens = torch.randint(0, max_kv_len, (batch_size,), dtype=torch.int32) kv_lens[-1] = max_kv_len seq_lens = kv_lens + q_lens max_seq_len = torch.max(seq_lens).item() # Always using 1.0 scale to reflect the real perf in benchmarking k_scale = v_scale = 1.0 ref_kv_cache = torch.randn(kv_cache_shape, dtype=dtype) if kv_quant_dtype == FP8_DTYPE: kv_cache, _ = to_float8(ref_kv_cache) else: kv_cache = ref_kv_cache max_num_blocks_per_seq = (max_seq_len + block_size - 1) // block_size block_tables = torch.randint( 0, NUM_BLOCKS, (batch_size, max_num_blocks_per_seq), dtype=torch.int32 ) kv_indptr = [0] kv_indices = [] kv_last_page_lens = [] for i in range(batch_size): seq_len = seq_lens[i] assert seq_len > 0 num_blocks = (seq_len + block_size - 1) // block_size kv_indices.extend(block_tables[i, :num_blocks]) kv_indptr.append(kv_indptr[-1] + num_blocks) kv_last_page_len = seq_len % block_size if kv_last_page_len == 0: kv_last_page_len = block_size kv_last_page_lens.append(kv_last_page_len) kv_indptr = torch.tensor(kv_indptr, dtype=torch.int32) kv_indices = torch.tensor(kv_indices, dtype=torch.int32) kv_last_page_lens = torch.tensor(kv_last_page_lens, dtype=torch.int32) workspace_buffer = torch.zeros(1024 * 1024 * 1024, dtype=torch.int8) wrapper = flashinfer.BatchPrefillWithPagedKVCacheWrapper( workspace_buffer, kv_layout ) wrapper.plan( q_indptr, kv_indptr, kv_indices, kv_last_page_lens, num_qo_heads, num_kv_heads, head_size, block_size, causal=True, sm_scale=sm_scale, q_data_type=dtype, kv_data_type=dtype, ) def time_fn(fn, warmup=10, trials=20): torch.cuda.synchronize() start = torch.Event(enable_timing=True) end = torch.Event(enable_timing=True) times = [] for i in range(warmup): fn() for i in range(trials): start.record() fn() end.record() torch.cuda.synchronize() times.append(start.elapsed_time(end)) # ms return sum(times) / len(times), torch.std(torch.tensor(times)) o_scale = 1.0 o_sf_scale = None output_baseline = torch.empty(ref_query.shape, dtype=dtype) if o_quant_dtype == FP4_DTYPE: o_sf_scale = 500.0 output_trtllm = flashinfer.utils.FP4Tensor( torch.empty(query.shape[:-1] + (query.shape[-1] // 2,), dtype=torch.uint8), torch.empty( ( round_up(query.shape[0], 128), round_up(query.shape[1] * query.shape[2] // 16, 4), ), dtype=torch.float8_e4m3fn, ), ) else: output_trtllm = torch.empty(query.shape, dtype=o_quant_dtype) def baseline_prefill(): return wrapper.run( ref_query, ref_kv_cache, k_scale=k_scale, v_scale=v_scale, out=output_baseline, ) def trtllm_prefill(): return flashinfer.prefill.trtllm_batch_context_with_kv_cache( query=query, kv_cache=kv_cache, workspace_buffer=workspace_buffer, block_tables=block_tables, seq_lens=seq_lens, max_q_len=max_q_len, max_kv_len=max_seq_len, bmm1_scale=q_scale * k_scale * sm_scale, bmm2_scale=v_scale / o_scale, batch_size=batch_size, cum_seq_lens_q=q_indptr, cum_seq_lens_kv=kv_indptr, o_sf_scale=o_sf_scale, out=output_trtllm, ) baseline_mean, baseline_std = time_fn(baseline_prefill) trtllm_mean, trtllm_std = time_fn(trtllm_prefill) # Calculate percentage speedup (positive means TRT is faster) speedup_percent = (baseline_mean - trtllm_mean) / baseline_mean print( f"\t{batch_size}\t{max_seq_len}\t{trtllm_mean:8.3f}\t{trtllm_std.item():8.3f}" f"\t{baseline_mean:8.3f}\t{baseline_std.item():8.3f}\t{speedup_percent:8.3f}" ) # Return results for CSV writing return { "batch_size": batch_size, "trtllm_mean": trtllm_mean, "trtllm_std": trtllm_std.item(), "baseline_mean": baseline_mean, "baseline_std": baseline_std.item(), "speedup_percent": speedup_percent, "q_dtype": str(q_quant_dtype), "kv_cache_dtype": str(kv_quant_dtype), "output_dtype": str(o_quant_dtype), "block_size": block_size, "num_kv_heads": num_kv_heads, "head_size": head_size, "max_seq_len": max_seq_len, } def write_results_to_csv(results, filename=None): """Write benchmark results to CSV file.""" if filename is None: timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") filename = f"flashinfer_trtllm_benchmark_{timestamp}.csv" fieldnames = [ "batch_size", "trtllm_mean", "trtllm_std", "baseline_mean", "baseline_std", "speedup_percent", "q_dtype", "kv_cache_dtype", "output_dtype", "block_size", "num_kv_heads", "head_size", "max_seq_len", ] file_exists = os.path.exists(filename) with open(filename, "a", newline="") as csvfile: writer = csv.DictWriter(csvfile, fieldnames=fieldnames) if not file_exists: writer.writeheader() for result in results: writer.writerow(result) print(f"Results written to {filename}") if __name__ == "__main__": batch_sizes = [1, 4, 8, 16, 32, 64, 128, 256] max_seq_lens = [1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072] all_results = [] dtype = torch.bfloat16 quant_dtypes = [ # (q_quant_dtype, kv_quant_dtype, o_quant_dtype) (None, None, None), (FP8_DTYPE, FP8_DTYPE, None), (FP8_DTYPE, FP8_DTYPE, FP8_DTYPE), (FP8_DTYPE, FP8_DTYPE, FP4_DTYPE), ] for quant_dtype in quant_dtypes: q_quant_dtype, kv_quant_dtype, o_quant_dtype = quant_dtype q_quant_dtype = q_quant_dtype or dtype kv_quant_dtype = kv_quant_dtype or dtype o_quant_dtype = o_quant_dtype or dtype print( f"Running benchmark for q_dtype = {q_quant_dtype}, " f"kv_cache_dtype: {kv_quant_dtype}, " f"output_dtype: {o_quant_dtype}" ) print( "\tbatch_size\tmax_seq_len\ttrtllm_mean\ttrtllm_std\tbaseline_mean\t" "baseline_std\tspeedup_percent" ) for max_seq_len in max_seq_lens: for bs in batch_sizes: result = benchmark_prefill( dtype=dtype, quant_dtypes=quant_dtype, batch_size=bs, max_seq_len=max_seq_len, ) all_results.append(result) # Write all results to CSV write_results_to_csv(all_results)
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/kernels/benchmark_trtllm_prefill_attention.py", "license": "Apache License 2.0", "lines": 261, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/attention/test_flashinfer_trtllm_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from tests.kernels.quantization.nvfp4_utils import ( dequantize_nvfp4_to_dtype, get_nvfp4_global_scale, ) from vllm.platforms import current_platform from vllm.utils.math_utils import round_up from vllm.utils.torch_utils import set_random_seed if not current_platform.is_device_capability_family(100): pytest.skip( "This TRTLLM kernel requires NVIDIA Blackwell.", allow_module_level=True ) else: import flashinfer FLOAT32_BYTES = torch.finfo(torch.float).bits // 8 FP8_DTYPE = current_platform.fp8_dtype() FP4_DTYPE = torch.uint8 def to_float8(x, dtype=torch.float8_e4m3fn): finfo = torch.finfo(dtype) min_val, max_val = x.aminmax() amax = torch.maximum(min_val.abs(), max_val.abs()).clamp(min=1e-12) scale = finfo.max / amax * 0.1 x_scl_sat = (x * scale).clamp(min=finfo.min, max=finfo.max) return x_scl_sat.to(dtype), scale.float().reciprocal() DTYPE = [torch.bfloat16] QUANT_DTYPES = [ # (q_quant_dtype, kv_quant_dtype, o_quant_dtype) (None, None, None), (None, FP8_DTYPE, None), (FP8_DTYPE, FP8_DTYPE, None), (FP8_DTYPE, FP8_DTYPE, FP8_DTYPE), (FP8_DTYPE, FP8_DTYPE, FP4_DTYPE), ] BATCH_SIZE = [4, 12] MAX_SEQ_LENS = [(1024, 4096)] NUM_HEADS = [(64, 8), (40, 8)] HEAD_SIZE = [128] KV_LAYOUT = ["HND"] # currently only HND is supported BLOCK_SIZE = [16] WINDOW_LEFT = [-1, 127] SOFT_CAP = [None, 50.0] HAS_SINKS = [True, False] NUM_BLOCKS = 32768 # Large enough to test overflow in index calculation. @pytest.mark.parametrize("dtype", DTYPE) @pytest.mark.parametrize("quant_dtypes", QUANT_DTYPES) @pytest.mark.parametrize("batch_size", BATCH_SIZE) @pytest.mark.parametrize("max_seq_lens", MAX_SEQ_LENS) @pytest.mark.parametrize("num_heads", NUM_HEADS) @pytest.mark.parametrize("head_size", HEAD_SIZE) @pytest.mark.parametrize("kv_layout", KV_LAYOUT) @pytest.mark.parametrize("block_size", BLOCK_SIZE) @pytest.mark.parametrize("window_left", WINDOW_LEFT) @pytest.mark.parametrize("soft_cap", SOFT_CAP) @pytest.mark.parametrize("has_sinks", HAS_SINKS) @torch.inference_mode def test_flashinfer_trtllm_decode_with_baseline( dtype: torch.dtype, quant_dtypes: tuple[torch.dtype | None, torch.dtype | None, torch.dtype | None], batch_size: int, max_seq_lens: tuple[int, int], num_heads: tuple[int, int], head_size: int, kv_layout: str, block_size: int, window_left: int, soft_cap: float | None, has_sinks: bool, ) -> None: torch.set_default_device("cuda") set_random_seed(42) q_quant_dtype, kv_quant_dtype, o_quant_dtype = quant_dtypes q_quant_dtype = q_quant_dtype or dtype kv_quant_dtype = kv_quant_dtype or dtype o_quant_dtype = o_quant_dtype or dtype _, max_kv_len = max_seq_lens num_qo_heads, num_kv_heads = num_heads assert num_qo_heads % num_kv_heads == 0 sm_scale = float(1.0 / (head_size**0.5)) kv_cache_shape = None if kv_layout == "NHD": kv_cache_shape = (NUM_BLOCKS, 2, block_size, num_kv_heads, head_size) elif kv_layout == "HND": kv_cache_shape = (NUM_BLOCKS, 2, num_kv_heads, block_size, head_size) else: raise ValueError(f"Invalid kv_layout: {kv_layout}") # max_q_len = 1 q_lens = torch.ones((batch_size,), dtype=torch.int32) q_indptr = torch.cat( [ torch.tensor([0], dtype=torch.int32), torch.cumsum(q_lens, dim=0, dtype=torch.int32), ] ) query = torch.randn(torch.sum(q_lens).item(), num_qo_heads, head_size, dtype=dtype) if q_quant_dtype == FP8_DTYPE: query, q_scale = to_float8(query) ref_query = query.to(dtype) * q_scale else: q_scale = 1.0 ref_query = query kv_lens = torch.randint(1, max_kv_len, (batch_size,), dtype=torch.int32) kv_lens[-1] = max_kv_len seq_lens = kv_lens + q_lens max_seq_len = torch.max(seq_lens).item() kv_cache = torch.randn(kv_cache_shape, dtype=dtype) if kv_quant_dtype == FP8_DTYPE: kv_cache, kv_scale = to_float8(kv_cache) ref_kv_cache = kv_cache.to(dtype) * kv_scale else: kv_scale = 1.0 ref_kv_cache = kv_cache k_scale = v_scale = kv_scale max_num_blocks_per_seq = (max_seq_len + block_size - 1) // block_size block_tables = torch.randint( 0, NUM_BLOCKS, (batch_size, max_num_blocks_per_seq), dtype=torch.int32 ) kv_indptr = [0] kv_indices = [] kv_last_page_lens = [] for i in range(batch_size): seq_len = seq_lens[i] assert seq_len > 0 num_blocks = (seq_len + block_size - 1) // block_size kv_indices.extend(block_tables[i, :num_blocks]) kv_indptr.append(kv_indptr[-1] + num_blocks) kv_last_page_len = seq_len % block_size if kv_last_page_len == 0: kv_last_page_len = block_size kv_last_page_lens.append(kv_last_page_len) kv_indptr = torch.tensor(kv_indptr, dtype=torch.int32) kv_indices = torch.tensor(kv_indices, dtype=torch.int32) kv_last_page_lens = torch.tensor(kv_last_page_lens, dtype=torch.int32) workspace_buffer = torch.zeros(128 * 1024 * 1024, dtype=torch.int8) # Baseline Decode if has_sinks: sinks = torch.rand(num_qo_heads, dtype=torch.float32) * 5 wrapper = flashinfer.BatchAttentionWithAttentionSinkWrapper( float_workspace_buffer=workspace_buffer, kv_layout=kv_layout, backend="fa2" ) else: sinks = None wrapper = flashinfer.BatchPrefillWithPagedKVCacheWrapper( float_workspace_buffer=workspace_buffer, kv_layout=kv_layout, backend="fa2" ) wrapper.plan( qo_indptr=q_indptr, paged_kv_indptr=kv_indptr, paged_kv_indices=kv_indices, paged_kv_last_page_len=kv_last_page_lens, num_qo_heads=num_qo_heads, num_kv_heads=num_kv_heads, head_dim_qk=head_size, page_size=block_size, causal=True, sm_scale=sm_scale, window_left=window_left, logits_soft_cap=soft_cap, q_data_type=dtype, kv_data_type=dtype, ) output = torch.empty(ref_query.shape, dtype=dtype) wrapper.run(ref_query, ref_kv_cache, sinks, sm_scale, out=output) o_scale = 1.0 o_sf_scale_float = None if o_quant_dtype == FP8_DTYPE: _, o_scale = to_float8(output) elif o_quant_dtype == FP4_DTYPE: o_sf_scale = get_nvfp4_global_scale(output) o_sf_scale_float = o_sf_scale.item() # TRTLLM Decode if o_quant_dtype == FP4_DTYPE: output_trtllm = flashinfer.utils.FP4Tensor( torch.empty(query.shape[:-1] + (query.shape[-1] // 2,), dtype=torch.uint8), torch.empty( ( round_up(query.shape[0], 128), round_up(query.shape[1] * query.shape[2] // 16, 4), ), dtype=torch.float8_e4m3fn, ), ) else: output_trtllm = torch.empty(query.shape, dtype=o_quant_dtype) flashinfer.decode.trtllm_batch_decode_with_kv_cache( query=query, kv_cache=kv_cache, workspace_buffer=workspace_buffer, block_tables=block_tables, seq_lens=seq_lens, max_seq_len=max_seq_len, bmm1_scale=q_scale * k_scale * sm_scale, bmm2_scale=v_scale / o_scale, window_left=window_left, sinks=sinks, o_sf_scale=o_sf_scale_float, out=output_trtllm, ) if o_quant_dtype == FP8_DTYPE: output_trtllm = output_trtllm.to(dtype) * o_scale elif o_quant_dtype == FP4_DTYPE: output_trtllm.data = output_trtllm.data.reshape( -1, query.shape[1] * query.shape[2] // 2 ) output_trtllm = dequantize_nvfp4_to_dtype( output_trtllm.data, output_trtllm.scale, o_sf_scale, dtype, query.device ) output_trtllm = output_trtllm.reshape(-1, query.shape[1], query.shape[2]) if q_quant_dtype == FP8_DTYPE and o_quant_dtype == FP4_DTYPE: rtol, atol = 7e-2, 9e-2 elif q_quant_dtype == FP8_DTYPE and o_quant_dtype == FP8_DTYPE: rtol, atol = 3e-2, 4e-2 elif q_quant_dtype == FP8_DTYPE and o_quant_dtype == dtype: rtol, atol = 2e-2, 2e-2 elif kv_quant_dtype == FP8_DTYPE: rtol, atol = 4e-2, 6e-2 else: rtol, atol = 1e-2, 1e-2 ( torch.testing.assert_close(output, output_trtllm, atol=atol, rtol=rtol), f"{torch.max(torch.abs(output - output_trtllm))}", ) @pytest.mark.parametrize("dtype", DTYPE) @pytest.mark.parametrize("quant_dtypes", QUANT_DTYPES) @pytest.mark.parametrize("batch_size", BATCH_SIZE) @pytest.mark.parametrize("max_seq_lens", MAX_SEQ_LENS) @pytest.mark.parametrize("num_heads", NUM_HEADS) @pytest.mark.parametrize("head_size", HEAD_SIZE) @pytest.mark.parametrize("kv_layout", KV_LAYOUT) @pytest.mark.parametrize("block_size", BLOCK_SIZE) @pytest.mark.parametrize("window_left", WINDOW_LEFT) @pytest.mark.parametrize("soft_cap", [None]) @pytest.mark.parametrize("has_sinks", HAS_SINKS) @torch.inference_mode def test_flashinfer_trtllm_prefill_with_baseline( dtype: torch.dtype, quant_dtypes: tuple[torch.dtype | None, torch.dtype | None, torch.dtype | None], batch_size: int, max_seq_lens: tuple[int, int], num_heads: tuple[int, int], head_size: int, kv_layout: str, block_size: int, window_left: int, soft_cap: float | None, has_sinks: bool, ) -> None: torch.set_default_device("cuda") set_random_seed(42) q_quant_dtype, kv_quant_dtype, o_quant_dtype = quant_dtypes q_quant_dtype = q_quant_dtype or dtype kv_quant_dtype = kv_quant_dtype or dtype o_quant_dtype = o_quant_dtype or dtype if q_quant_dtype != kv_quant_dtype: pytest.skip("Skipped mixed QKV dtypes for prefill") max_q_len, max_kv_len = max_seq_lens num_qo_heads, num_kv_heads = num_heads assert num_qo_heads % num_kv_heads == 0 sm_scale = float(1.0 / (head_size**0.5)) kv_cache_shape = None if kv_layout == "NHD": kv_cache_shape = (NUM_BLOCKS, 2, block_size, num_kv_heads, head_size) elif kv_layout == "HND": kv_cache_shape = (NUM_BLOCKS, 2, num_kv_heads, block_size, head_size) else: raise ValueError(f"Invalid kv_layout: {kv_layout}") q_lens = torch.randint(1, max_q_len, (batch_size,), dtype=torch.int32) q_lens[-1] = max_q_len q_indptr = torch.cat( [ torch.tensor([0], dtype=torch.int32), torch.cumsum(q_lens, dim=0, dtype=torch.int32), ] ) query = torch.randn(torch.sum(q_lens).item(), num_qo_heads, head_size, dtype=dtype) if q_quant_dtype == FP8_DTYPE: query, q_scale = to_float8(query) ref_query = query.to(dtype) * q_scale else: q_scale = 1.0 ref_query = query kv_lens = torch.randint(1, max_kv_len, (batch_size,), dtype=torch.int32) kv_lens[-1] = max_kv_len seq_lens = kv_lens + q_lens max_seq_len = torch.max(seq_lens).item() kv_cache = torch.randn(kv_cache_shape, dtype=dtype) if kv_quant_dtype == FP8_DTYPE: kv_cache, kv_scale = to_float8(kv_cache) ref_kv_cache = kv_cache.to(dtype) * kv_scale else: kv_scale = 1.0 ref_kv_cache = kv_cache k_scale = v_scale = kv_scale max_num_blocks_per_seq = (max_seq_len + block_size - 1) // block_size block_tables = torch.randint( 0, NUM_BLOCKS, (batch_size, max_num_blocks_per_seq), dtype=torch.int32 ) kv_indptr = [0] kv_indices = [] kv_last_page_lens = [] for i in range(batch_size): seq_len = seq_lens[i] assert seq_len > 0 num_blocks = (seq_len + block_size - 1) // block_size kv_indices.extend(block_tables[i, :num_blocks]) kv_indptr.append(kv_indptr[-1] + num_blocks) kv_last_page_len = seq_len % block_size if kv_last_page_len == 0: kv_last_page_len = block_size kv_last_page_lens.append(kv_last_page_len) kv_indptr = torch.tensor(kv_indptr, dtype=torch.int32) kv_indices = torch.tensor(kv_indices, dtype=torch.int32) kv_last_page_lens = torch.tensor(kv_last_page_lens, dtype=torch.int32) workspace_buffer = torch.zeros(128 * 1024 * 1024, dtype=torch.int8) # Baseline Prefill if has_sinks: sinks = torch.rand(num_qo_heads, dtype=torch.float32) * 5 wrapper = flashinfer.BatchAttentionWithAttentionSinkWrapper( float_workspace_buffer=workspace_buffer, kv_layout=kv_layout, backend="fa2" ) else: sinks = None wrapper = flashinfer.BatchPrefillWithPagedKVCacheWrapper( float_workspace_buffer=workspace_buffer, kv_layout=kv_layout, backend="fa2" ) wrapper.plan( qo_indptr=q_indptr, paged_kv_indptr=kv_indptr, paged_kv_indices=kv_indices, paged_kv_last_page_len=kv_last_page_lens, num_qo_heads=num_qo_heads, num_kv_heads=num_kv_heads, head_dim_qk=head_size, page_size=block_size, causal=True, sm_scale=sm_scale, window_left=window_left, logits_soft_cap=soft_cap, q_data_type=dtype, kv_data_type=dtype, ) output = torch.empty(ref_query.shape, dtype=dtype) wrapper.run(ref_query, ref_kv_cache, sinks, sm_scale, out=output) o_scale = 1.0 o_sf_scale_float = None if o_quant_dtype == FP8_DTYPE: _, o_scale = to_float8(output) elif o_quant_dtype == FP4_DTYPE: o_sf_scale = get_nvfp4_global_scale(output) o_sf_scale_float = o_sf_scale.item() # TRTLLM Prefill if o_quant_dtype == FP4_DTYPE: output_trtllm = flashinfer.utils.FP4Tensor( torch.empty(query.shape[:-1] + (query.shape[-1] // 2,), dtype=torch.uint8), torch.empty( ( round_up(query.shape[0], 128), round_up(query.shape[1] * query.shape[2] // 16, 4), ), dtype=torch.float8_e4m3fn, ), ) else: output_trtllm = torch.empty(query.shape, dtype=o_quant_dtype) flashinfer.prefill.trtllm_batch_context_with_kv_cache( query=query, kv_cache=kv_cache, workspace_buffer=workspace_buffer, block_tables=block_tables, seq_lens=seq_lens, max_q_len=max_q_len, max_kv_len=max_seq_len, bmm1_scale=q_scale * k_scale * sm_scale, bmm2_scale=v_scale / o_scale, batch_size=batch_size, cum_seq_lens_q=q_indptr, cum_seq_lens_kv=kv_indptr, window_left=window_left, sinks=sinks, o_sf_scale=o_sf_scale_float, out=output_trtllm, ) if o_quant_dtype == FP8_DTYPE: output_trtllm = output_trtllm.to(dtype) * o_scale elif o_quant_dtype == FP4_DTYPE: output_trtllm.data = output_trtllm.data.reshape( -1, query.shape[1] * query.shape[2] // 2 ) output_trtllm = dequantize_nvfp4_to_dtype( output_trtllm.data, output_trtllm.scale, o_sf_scale, dtype, query.device ) output_trtllm = output_trtllm.reshape(-1, query.shape[1], query.shape[2]) if q_quant_dtype == FP8_DTYPE and o_quant_dtype == FP4_DTYPE: rtol, atol = 3e-1, 4e-1 elif q_quant_dtype == FP8_DTYPE and o_quant_dtype == FP8_DTYPE: rtol, atol = 4e-2, 6e-2 elif q_quant_dtype == FP8_DTYPE and o_quant_dtype == dtype: rtol, atol = 2e-2, 3e-2 else: rtol, atol = 1e-2, 1e-2 ( torch.testing.assert_close(output, output_trtllm, atol=atol, rtol=rtol), f"{torch.max(torch.abs(output - output_trtllm))}", )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/attention/test_flashinfer_trtllm_attention.py", "license": "Apache License 2.0", "lines": 406, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/test_pooling_params.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass import pytest from tests.models.utils import EmbedModelInfo from vllm import PoolingParams from vllm.config import ModelConfig, PoolerConfig EMBEDDING_MODELS = [ EmbedModelInfo("intfloat/multilingual-e5-small", is_matryoshka=False), EmbedModelInfo( "Snowflake/snowflake-arctic-embed-m-v1.5", is_matryoshka=True, matryoshka_dimensions=[256], ), ] classify_parameters = ["use_activation"] embed_parameters = ["dimensions", "use_activation"] step_pooling_parameters = ["step_tag_id", "returned_token_ids"] @dataclass() class MockModelConfig: pooler_config: PoolerConfig def test_embed(): task = "embed" model_config = MockModelConfig(pooler_config=PoolerConfig(seq_pooling_type="CLS")) pooling_params = PoolingParams(task=task, use_activation=None) pooling_params.verify(model_config) pooling_params = PoolingParams(task=task, use_activation=True) pooling_params.verify(model_config) pooling_params = PoolingParams(task=task, use_activation=False) pooling_params.verify(model_config) invalid_parameters = classify_parameters + step_pooling_parameters for p in set(invalid_parameters) - set(embed_parameters): with pytest.raises(ValueError): pooling_params = PoolingParams(task=task, **{p: True}) pooling_params.verify(model_config) @pytest.mark.parametrize("model_info", EMBEDDING_MODELS) def test_embed_dimensions(model_info: EmbedModelInfo): task = "embed" model_config = ModelConfig( model_info.name, tokenizer=model_info.name, tokenizer_mode="auto", trust_remote_code=False, seed=0, dtype="float16", ) pooling_params = PoolingParams(task=task, dimensions=None) pooling_params.verify(model_config) with pytest.raises(ValueError): pooling_params = PoolingParams(task=task, dimensions=1) pooling_params.verify(model_config) if model_info.is_matryoshka: assert model_info.matryoshka_dimensions is not None pooling_params = PoolingParams( task=task, dimensions=model_info.matryoshka_dimensions[0] ) pooling_params.verify(model_config) @pytest.mark.parametrize("task", ["score", "classify"]) def test_classify(task): model_config = MockModelConfig(pooler_config=PoolerConfig(seq_pooling_type="CLS")) pooling_params = PoolingParams(task=task, use_activation=None) pooling_params.verify(model_config) pooling_params = PoolingParams(task=task, use_activation=True) pooling_params.verify(model_config) pooling_params = PoolingParams(task=task, use_activation=False) pooling_params.verify(model_config) invalid_parameters = embed_parameters + step_pooling_parameters for p in set(invalid_parameters) - set(classify_parameters): with pytest.raises(ValueError): pooling_params = PoolingParams(task=task, **{p: True}) pooling_params.verify(model_config) @pytest.mark.parametrize("pooling_type", ["ALL", "STEP"]) def test_token_embed(pooling_type: str): task = "token_embed" model_config = MockModelConfig( pooler_config=PoolerConfig(tok_pooling_type=pooling_type) ) pooling_params = PoolingParams(task=task, use_activation=None) pooling_params.verify(model_config) pooling_params = PoolingParams(task=task, use_activation=True) pooling_params.verify(model_config) pooling_params = PoolingParams(task=task, use_activation=False) pooling_params.verify(model_config) invalid_parameters = classify_parameters if pooling_type != "STEP": invalid_parameters = classify_parameters + step_pooling_parameters for p in set(invalid_parameters) - set(embed_parameters): with pytest.raises(ValueError): pooling_params = PoolingParams(task=task, **{p: True}) pooling_params.verify(model_config) @pytest.mark.parametrize("pooling_type", ["ALL", "STEP"]) def test_token_classify(pooling_type: str): task = "token_classify" model_config = MockModelConfig( pooler_config=PoolerConfig(tok_pooling_type=pooling_type) ) pooling_params = PoolingParams(task=task, use_activation=None) pooling_params.verify(model_config) pooling_params = PoolingParams(task=task, use_activation=True) pooling_params.verify(model_config) pooling_params = PoolingParams(task=task, use_activation=False) pooling_params.verify(model_config) invalid_parameters = embed_parameters if pooling_type != "STEP": invalid_parameters = embed_parameters + step_pooling_parameters for p in set(invalid_parameters) - set(classify_parameters): with pytest.raises(ValueError): pooling_params = PoolingParams(task=task, **{p: True}) pooling_params.verify(model_config)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/test_pooling_params.py", "license": "Apache License 2.0", "lines": 109, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/base.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Rotary Positional Embeddings Base Class.""" import torch from vllm._aiter_ops import rocm_aiter_ops from vllm.model_executor.custom_op import CustomOp from .common import ApplyRotaryEmb # --8<-- [start:rotary_embedding] @CustomOp.register("rotary_embedding") class RotaryEmbeddingBase(CustomOp): """Original rotary positional embedding.""" # --8<-- [end:rotary_embedding] def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, init_cache: bool = True, ) -> None: super().__init__() self.head_size = head_size self.rotary_dim = rotary_dim self.max_position_embeddings = max_position_embeddings self.base = base self.is_neox_style = is_neox_style self.dtype = dtype # TODO(mgoin): disabled for now due to failures # Flashinfer only supports head_size=64, 128, 256, 512. # https://github.com/flashinfer-ai/flashinfer/blob/ebfd655efe830048dba5d582aaa61d61d1cf9a87/include/flashinfer/utils.cuh#L174-L202 # self.use_flashinfer = (self.enabled() # and dtype in (torch.float16, torch.bfloat16) # and current_platform.is_cuda() # and has_flashinfer() # and self.head_size in [64, 128, 256, 512]) # Check if use_flashinfer is already set if not hasattr(self, "use_flashinfer"): self.use_flashinfer = False self.use_aiter = ( self.enabled() and rocm_aiter_ops.is_triton_rotary_embed_enabled() ) if self.use_aiter: self.rocm_aiter_triton_rotary_embedding = ( rocm_aiter_ops.get_triton_rotary_embedding_op() ) if init_cache: cache = self._compute_cos_sin_cache() if not self.use_flashinfer: cache = cache.to(dtype) self.cos_sin_cache: torch.Tensor self.register_buffer("cos_sin_cache", cache, persistent=False) self.apply_rotary_emb = ApplyRotaryEmb( is_neox_style=self.is_neox_style, ) def _compute_inv_freq(self, base: float) -> torch.Tensor: """Compute the inverse frequency.""" # NOTE(woosuk): To exactly match the HF implementation, we need to # use CPU to compute the cache and then move it to GPU. However, we # create the cache on GPU for faster initialization. This may cause # a slight numerical difference between the HF implementation and ours. inv_freq = 1.0 / ( base ** ( torch.arange(0, self.rotary_dim, 2, dtype=torch.float) / self.rotary_dim ) ) return inv_freq def _compute_cos_sin_cache(self) -> torch.Tensor: """Compute the cos and sin cache.""" inv_freq = self._compute_inv_freq(self.base) t = torch.arange(self.max_position_embeddings, dtype=torch.float) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() sin = freqs.sin() cache = torch.cat((cos, sin), dim=-1) return cache def _match_cos_sin_cache_dtype(self, query: torch.Tensor) -> torch.Tensor: # __setattr__ in nn.Module (called by `self.cos_sin_cache = ...`) # is expensive, so avoid calling it if possible cos_sin_cache = self.cos_sin_cache if ( cos_sin_cache.device == query.device and self.cos_sin_cache.dtype == query.dtype ): return cos_sin_cache cos_sin_cache = cos_sin_cache.to(query.device, dtype=query.dtype) # Avoid mutating buffers during torch.compile (cudagraph) tracing. if torch.compiler.is_compiling(): return cos_sin_cache self.cos_sin_cache = cos_sin_cache return cos_sin_cache def get_cos_sin(self, seqlen: int) -> tuple[torch.Tensor, torch.Tensor]: cos_sin = self.cos_sin_cache[:seqlen] cos, sin = cos_sin.chunk(2, dim=-1) return cos, sin class RotaryEmbedding(RotaryEmbeddingBase): def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, init_cache: bool = True, ) -> None: super().__init__( head_size=head_size, rotary_dim=rotary_dim, max_position_embeddings=max_position_embeddings, base=base, is_neox_style=is_neox_style, dtype=dtype, init_cache=init_cache, ) @staticmethod def forward_static( positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None, head_size: int, rotary_dim: int, cos_sin_cache: torch.Tensor, is_neox_style: bool, ) -> tuple[torch.Tensor, torch.Tensor | None]: """A PyTorch-native implementation of forward().""" positions = positions.flatten() num_tokens = positions.shape[0] cos_sin = cos_sin_cache.index_select(0, positions) cos, sin = cos_sin.chunk(2, dim=-1) query_shape = query.shape query = query.view(num_tokens, -1, head_size) query_rot = query[..., :rotary_dim] query_pass = query[..., rotary_dim:] query_rot = ApplyRotaryEmb.forward_static( query_rot, cos, sin, is_neox_style, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) # key may be None in some cases, e.g. cross-layer KV sharing if key is not None: key_shape = key.shape key = key.view(num_tokens, -1, head_size) key_rot = key[..., :rotary_dim] key_pass = key[..., rotary_dim:] key_rot = ApplyRotaryEmb.forward_static( key_rot, cos, sin, is_neox_style, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """A PyTorch-native implementation of forward().""" cos_sin_cache = self._match_cos_sin_cache_dtype(query) return self.forward_static( positions, query, key, self.head_size, self.rotary_dim, cos_sin_cache, self.is_neox_style, ) def forward_cuda( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: if self.use_flashinfer: torch.ops.vllm.flashinfer_rotary_embedding( positions, query, key, self.head_size, self.cos_sin_cache, self.is_neox_style, ) return query, key from vllm import _custom_ops as ops cos_sin_cache = self._match_cos_sin_cache_dtype(query) # ops.rotary_embedding() is an in-place operation # that updates the query and key tensors. ops.rotary_embedding( positions, query, key, self.head_size, cos_sin_cache, self.is_neox_style, ) return query, key def forward_hip( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: if self.use_aiter: cos_sin_cache = self._match_cos_sin_cache_dtype(query) self.rocm_aiter_triton_rotary_embedding( positions, query, key, self.head_size, cos_sin_cache, self.is_neox_style, ) return query, key return self.forward_cuda(positions, query, key) def forward_xpu( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: self._match_cos_sin_cache_dtype(query) # ops.rotary_embedding() is an in-place operation # that updates the query and key tensors. if key is None: return self.forward_native(positions, query, key) else: from vllm import _custom_ops as ops cos_sin_cache = self._match_cos_sin_cache_dtype(query) ops.rotary_embedding( positions, query, key, self.head_size, cos_sin_cache, self.is_neox_style, ) return query, key def forward_cpu( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: from vllm import _custom_ops as ops cos_sin_cache = self._match_cos_sin_cache_dtype(query) # ops.rotary_embedding() is an in-place operation # that updates the query and key tensors. ops.rotary_embedding( positions, query, key, self.head_size, cos_sin_cache, self.is_neox_style, ) return query, key def extra_repr(self) -> str: s = f"head_size={self.head_size}, rotary_dim={self.rotary_dim}" s += f", max_position_embeddings={self.max_position_embeddings}" s += f", base={self.base}, is_neox_style={self.is_neox_style}" return s
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/base.py", "license": "Apache License 2.0", "lines": 268, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/common.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math from importlib.util import find_spec import torch from vllm.logger import init_logger from vllm.model_executor.custom_op import CustomOp from vllm.utils.torch_utils import direct_register_custom_op logger = init_logger(__name__) # common functions def rotate_neox(x: torch.Tensor) -> torch.Tensor: x1 = x[..., : x.shape[-1] // 2] x2 = x[..., x.shape[-1] // 2 :] return torch.cat((-x2, x1), dim=-1) def rotate_gptj(x: torch.Tensor) -> torch.Tensor: x1 = x[..., ::2] x2 = x[..., 1::2] x = torch.stack((-x2, x1), dim=-1) return x.flatten(-2) # yarn functions # Inverse dim formula to find dim based on number of rotations def yarn_find_correction_dim( num_rotations: int, dim: int, base: float = 10000, max_position_embeddings: int = 2048, ) -> float: return (dim * math.log(max_position_embeddings / (num_rotations * 2 * math.pi))) / ( 2 * math.log(base) ) # Find dim range bounds based on rotations def yarn_find_correction_range( low_rot: int, high_rot: int, dim: int, base: float = 10000, max_position_embeddings: int = 2048, truncate: bool = True, ) -> tuple[float | int, float | int]: low = yarn_find_correction_dim(low_rot, dim, base, max_position_embeddings) high = yarn_find_correction_dim(high_rot, dim, base, max_position_embeddings) if truncate: low = math.floor(low) high = math.ceil(high) return max(low, 0), min(high, dim - 1) # Clamp values just in case def yarn_linear_ramp_mask( low: float, high: float, dim: int, dtype: torch.dtype ) -> torch.Tensor: if low == high: high += 0.001 # Prevent singularity linear_func = (torch.arange(dim, dtype=dtype) - low) / (high - low) ramp_func = torch.clamp(linear_func, 0, 1) return ramp_func def yarn_get_mscale(scale: float = 1) -> float: if scale <= 1: return 1.0 return 0.1 * math.log(scale) + 1.0 def _flashinfer_rotary_embedding( positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor, head_size: int, cos_sin_cache: torch.Tensor, is_neox: bool, ) -> None: """Custom op wrapper for flashinfer's rotary embedding. This is an in-place operation that modifies query and key tensors directly. """ from flashinfer.rope import apply_rope_with_cos_sin_cache_inplace apply_rope_with_cos_sin_cache_inplace( positions=positions, query=query, key=key, head_size=head_size, cos_sin_cache=cos_sin_cache, is_neox=is_neox, ) def _flashinfer_rotary_embedding_fake( positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor, head_size: int, cos_sin_cache: torch.Tensor, is_neox: bool, ) -> None: return # Register flashinfer rotary embedding custom op direct_register_custom_op( op_name="flashinfer_rotary_embedding", op_func=_flashinfer_rotary_embedding, mutates_args=["query", "key"], # These tensors are modified in-place fake_impl=_flashinfer_rotary_embedding_fake, ) # --8<-- [start:apply_rotary_emb] @CustomOp.register("apply_rotary_emb") class ApplyRotaryEmb(CustomOp): # --8<-- [end:apply_rotary_emb] def __init__( self, enforce_enable: bool = False, is_neox_style: bool = True, enable_fp32_compute: bool = False, ) -> None: super().__init__(enforce_enable=enforce_enable) self.is_neox_style = is_neox_style self.enable_fp32_compute = enable_fp32_compute self.apply_rotary_emb_flash_attn = None if find_spec("flash_attn") is not None: from flash_attn.ops.triton.rotary import apply_rotary self.apply_rotary_emb_flash_attn = apply_rotary @staticmethod def forward_static( x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, is_neox_style: bool = True, enable_fp32_compute: bool = False, ) -> torch.Tensor: """ Args: x: [batch_size (optional), seq_len, num_heads, head_size] cos: [seq_len, head_size // 2] sin: [seq_len, head_size // 2] is_neox_style: Whether to use the Neox-style or GPT-J-style. enable_fp32_compute: Temporarily convert x, cos, sin to FP32 dtype for higher accuracy. """ origin_dtype = x.dtype if enable_fp32_compute: x = x.float() cos = cos.unsqueeze(-2).to(x.dtype) sin = sin.unsqueeze(-2).to(x.dtype) if is_neox_style: x1, x2 = torch.chunk(x, 2, dim=-1) else: x1 = x[..., ::2] x2 = x[..., 1::2] o1 = x1 * cos - x2 * sin o2 = x2 * cos + x1 * sin if is_neox_style: output = torch.cat((o1, o2), dim=-1) else: output = torch.stack((o1, o2), dim=-1).flatten(-2) if enable_fp32_compute: output = output.to(origin_dtype) return output def _pre_process( self, x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Size, torch.dtype]: origin_shape = x.shape if len(origin_shape) == 3: # x: [seq_len, num_heads, head_size] x = x.unsqueeze(0) origin_dtype = x.dtype if self.enable_fp32_compute: x = x.float() cos = cos.float() sin = sin.float() return x, cos, sin, origin_shape, origin_dtype def _post_process( self, output: torch.Tensor, origin_shape: torch.Size, origin_dtype: torch.dtype, ) -> torch.Tensor: if len(origin_shape) == 3: output = output.squeeze(0) if self.enable_fp32_compute: output = output.to(origin_dtype) return output def forward_native( self, x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, ) -> torch.Tensor: output = self.forward_static( x, cos, sin, self.is_neox_style, self.enable_fp32_compute ) return output def forward_cuda( self, x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, ) -> torch.Tensor: from vllm.vllm_flash_attn.layers.rotary import apply_rotary_emb x, cos, sin, origin_shape, origin_dtype = self._pre_process(x, cos, sin) """ Arguments of apply_rotary_emb() in vllm_flash_attn: x: [batch_size, seq_len, nheads, headdim] cos, sin: [seqlen_rotary, rotary_dim / 2] interleaved: defalut as False (Neox-style). ... """ interleaved = not self.is_neox_style output = apply_rotary_emb(x, cos, sin, interleaved) output = self._post_process(output, origin_shape, origin_dtype) return output def forward_hip( self, x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, ) -> torch.Tensor: if self.apply_rotary_emb_flash_attn is not None: x, cos, sin, origin_shape, origin_dtype = self._pre_process(x, cos, sin) """ Arguments of apply_rotary() in flash_attn: x: [batch_size, seq_len, nheads, headdim] cos, sin: [seqlen_rotary, rotary_dim / 2] interleaved: defalut as False (Neox-style). ... """ interleaved = not self.is_neox_style output = self.apply_rotary_emb_flash_attn( x, cos, sin, interleaved=interleaved ).type_as(x) output = self._post_process(output, origin_shape, origin_dtype) else: # Falling back to PyTorch native implementation. output = self.forward_native(x, cos, sin) return output def forward_cpu( self, x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, ) -> torch.Tensor: # TODO (bigPYJ1151): need to enable fused CPU ROPE here return self.forward_native(x, cos, sin) def extra_repr(self) -> str: s = f"is_neox_style={self.is_neox_style}" s += f", enable_fp32_compute={self.enable_fp32_compute}" return s
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/common.py", "license": "Apache License 2.0", "lines": 238, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/deepseek_scaling_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import torch from vllm.platforms import current_platform from vllm.utils.flashinfer import has_flashinfer from .base import RotaryEmbeddingBase from .common import ( rotate_gptj, rotate_neox, yarn_find_correction_range, yarn_linear_ramp_mask, ) def yarn_get_mscale(scale: float = 1, mscale: float = 1) -> float: if scale <= 1: return 1.0 return 0.1 * mscale * math.log(scale) + 1.0 class DeepseekScalingRotaryEmbedding(RotaryEmbeddingBase): """RotaryEmbedding extended with YaRN method. Credits to Peng et al. github.com/jquesnelle/yarn """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_factor: float, dtype: torch.dtype, *, extrapolation_factor: float = 1, attn_factor: float = 1, beta_fast: int = 32, beta_slow: int = 1, mscale: float = 1, mscale_all_dim: float = 0, ) -> None: self.scaling_factor = scaling_factor self.extrapolation_factor = extrapolation_factor self.attn_factor = attn_factor self.beta_fast = beta_fast self.beta_slow = beta_slow # Get n-d magnitude scaling corrected for interpolation. self.mscale = float( yarn_get_mscale(self.scaling_factor, float(mscale)) / yarn_get_mscale(self.scaling_factor, float(mscale_all_dim)) * attn_factor ) self.use_flashinfer = ( self.enabled() and dtype in (torch.float16, torch.bfloat16) and current_platform.is_cuda() and has_flashinfer() and head_size in [64, 128, 256, 512] ) super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_inv_freq(self, scaling_factor: float) -> torch.Tensor: pos_freqs = self.base ** ( torch.arange( 0, self.rotary_dim, 2, dtype=torch.float, ) / self.rotary_dim ) inv_freq_extrapolation = 1.0 / pos_freqs inv_freq_interpolation = 1.0 / (scaling_factor * pos_freqs) low, high = yarn_find_correction_range( self.beta_fast, self.beta_slow, self.rotary_dim, self.base, self.max_position_embeddings, ) # Get n-d rotational scaling corrected for extrapolation inv_freq_mask = ( 1 - yarn_linear_ramp_mask(low, high, self.rotary_dim // 2, dtype=torch.float) ) * self.extrapolation_factor inv_freq = ( inv_freq_interpolation * (1 - inv_freq_mask) + inv_freq_extrapolation * inv_freq_mask ) return inv_freq def _compute_cos_sin_cache(self) -> torch.Tensor: inv_freq = self._compute_inv_freq(self.scaling_factor) t = torch.arange( self.max_position_embeddings * self.scaling_factor, dtype=torch.float32, ) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() * self.mscale sin = freqs.sin() * self.mscale cache = torch.cat((cos, sin), dim=-1) return cache def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """PyTorch-native implementation equivalent to forward().""" assert key is not None cos_sin_cache = self._match_cos_sin_cache_dtype(query) query_rot = query[..., : self.rotary_dim] key_rot = key[..., : self.rotary_dim] if self.rotary_dim < self.head_size: query_pass = query[..., self.rotary_dim :] key_pass = key[..., self.rotary_dim :] cos_sin = cos_sin_cache[ torch.add(positions, offsets) if offsets is not None else positions ] cos, sin = cos_sin.chunk(2, dim=-1) if self.is_neox_style: # NOTE(woosuk): Here we assume that the positions tensor has the # shape [batch_size, seq_len]. cos = cos.repeat(1, 1, 2).unsqueeze(-2) sin = sin.repeat(1, 1, 2).unsqueeze(-2) else: cos = cos.repeat_interleave(2, dim=-1).unsqueeze(-2) sin = sin.repeat_interleave(2, dim=-1).unsqueeze(-2) rotate_fn = rotate_neox if self.is_neox_style else rotate_gptj query_rot = query_rot * cos + rotate_fn(query_rot) * sin key_rot = key_rot * cos + rotate_fn(key_rot) * sin if self.rotary_dim < self.head_size: query = torch.cat((query_rot, query_pass), dim=-1) key = torch.cat((key_rot, key_pass), dim=-1) else: query = query_rot key = key_rot return query, key def forward_hip( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: return self.forward_native(positions, query, key, offsets) def forward_cuda( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: if self.use_flashinfer: torch.ops.vllm.flashinfer_rotary_embedding( torch.add(positions, offsets) if offsets is not None else positions, query, key, self.head_size, self.cos_sin_cache, self.is_neox_style, ) return query, key else: return self.forward_native(positions, query, key, offsets)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/deepseek_scaling_rope.py", "license": "Apache License 2.0", "lines": 163, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/dual_chunk_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.model_executor.custom_op import CustomOp from .common import rotate_gptj, rotate_neox # --8<-- [start:dual_chunk_rotary_embedding] @CustomOp.register("dual_chunk_rotary_embedding") class DualChunkRotaryEmbedding(CustomOp): """Rotary positional embedding for Dual Chunk Attention.""" # --8<-- [end:dual_chunk_rotary_embedding] def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, chunk_size: int, local_size: int, ) -> None: super().__init__() self.head_size = head_size self.rotary_dim = rotary_dim self.max_position_embeddings = max_position_embeddings self.base = base self.is_neox_style = is_neox_style self.chunk_size = chunk_size self.local_size = local_size self.dtype = dtype self.device = torch.device(f"cuda:{torch.cuda.current_device()}") (q_cache, qc_cache, k_cache, qc_no_clamp_cache, q_inter_cache) = ( self._compute_cos_sin_cache() ) self.register_buffer("cos_sin_q_cache", q_cache, persistent=False) self.register_buffer("cos_sin_qc_cache", qc_cache, persistent=False) self.register_buffer("cos_sin_k_cache", k_cache, persistent=False) self.register_buffer( "cos_sin_qc_no_clamp_cache", qc_no_clamp_cache, persistent=False ) self.register_buffer("cos_sin_q_inter_cache", q_inter_cache, persistent=False) def _compute_inv_freq(self, base: float) -> torch.Tensor: """Compute the inverse frequency.""" # NOTE(woosuk): The HF implementation uses `torch.arange(...).float()`. # However, we use `torch.arange(..., dtype=torch.float)` instead to # avoid numerical issues with large base values (e.g., 10000000). # This may cause a slight numerical difference between the HF # implementation and ours. # NOTE(woosuk): To exactly match the HF implementation, we need to # use CPU to compute the cache and then move it to GPU. However, we # create the cache on GPU for faster initialization. This may cause # a slight numerical difference between the HF implementation and ours. inv_freq = 1.0 / ( base ** ( torch.arange(0, self.rotary_dim, 2, dtype=torch.float) / self.rotary_dim ) ) return inv_freq def _compute_cos_sin_cache(self) -> torch.Tensor: """Compute the cos and sin cache.""" inv_freq = self._compute_inv_freq(self.base) chunk_len = self.chunk_size - self.local_size q_t = torch.arange(chunk_len, dtype=torch.float) qc_t = (torch.arange(chunk_len, dtype=torch.float) + chunk_len).clamp( max=self.chunk_size ) k_t = torch.arange(self.max_position_embeddings, dtype=torch.float) % chunk_len # count from chunk_len, no clamp(self.chunk_size) restriction qc_no_clamp_t = torch.arange(chunk_len, dtype=torch.float) + chunk_len # count from self.chunk_size for q_inter's rope q_inter_t = torch.arange(chunk_len, dtype=torch.float) + self.chunk_size q_freqs = torch.outer(q_t, inv_freq) qc_freqs = torch.outer(qc_t, inv_freq) k_freqs = torch.outer(k_t, inv_freq) qc_no_clamp_freqs = torch.outer(qc_no_clamp_t, inv_freq) q_inter_freqs = torch.outer(q_inter_t, inv_freq) q_cos = q_freqs.cos() q_sin = q_freqs.sin() qc_cos = qc_freqs.cos() qc_sin = qc_freqs.sin() k_cos = k_freqs.cos() k_sin = k_freqs.sin() qc_no_clamp_cos = qc_no_clamp_freqs.cos() qc_no_clamp_sin = qc_no_clamp_freqs.sin() q_inter_cos = q_inter_freqs.cos() q_inter_sin = q_inter_freqs.sin() q_cache = torch.cat((q_cos, q_sin), dim=-1).to( dtype=self.dtype, device=self.device ) qc_cache = torch.cat((qc_cos, qc_sin), dim=-1).to( dtype=self.dtype, device=self.device ) k_cache = torch.cat((k_cos, k_sin), dim=-1).to( dtype=self.dtype, device=self.device ) qc_no_clamp_cache = torch.cat((qc_no_clamp_cos, qc_no_clamp_sin), dim=-1).to( dtype=self.dtype, device=self.device ) q_inter_cache = torch.cat((q_inter_cos, q_inter_sin), dim=-1).to( dtype=self.dtype, device=self.device ) return q_cache, qc_cache, k_cache, qc_no_clamp_cache, q_inter_cache def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: query = query.view(*query.shape[:-1], -1, self.head_size) key = key.view(*key.shape[:-1], -1, self.head_size) query_rot = query[..., : self.rotary_dim] key_rot = key[..., : self.rotary_dim] if self.rotary_dim < self.head_size: query_pass = query[..., self.rotary_dim :] key_pass = key[..., self.rotary_dim :] else: query_pass = None key_pass = None positions_with_offsets = ( torch.add(positions, offsets) if offsets is not None else positions ) key = self._apply_rotary_embedding( self.cos_sin_k_cache[positions_with_offsets], key_rot, key_pass ) chunk_len = self.chunk_size - self.local_size query = self._apply_rotary_embedding( self.cos_sin_q_cache[positions_with_offsets % chunk_len], query_rot, query_pass, ) query_succ = self._apply_rotary_embedding( self.cos_sin_qc_cache[positions_with_offsets % chunk_len], query_rot, query_pass, ) query_inter = self._apply_rotary_embedding( self.cos_sin_qc_cache[chunk_len - 1].repeat(positions.shape[0], 1), query_rot, query_pass, ) query_succ_critical = self._apply_rotary_embedding( self.cos_sin_qc_no_clamp_cache[positions_with_offsets % chunk_len], query_rot, query_pass, ) query_inter_critical = self._apply_rotary_embedding( self.cos_sin_q_inter_cache[positions_with_offsets % chunk_len], query_rot, query_pass, ) # merge query into one tensor to simplify the interfaces query = torch.cat( ( query, query_succ, query_inter, query_succ_critical, query_inter_critical, ), dim=-1, ) return query, key def forward_cuda( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: return self.forward_native(positions, query, key, offsets) def _apply_rotary_embedding(self, cos_sin, hidden_rot, hidden_pass): cos, sin = cos_sin.chunk(2, dim=-1) if self.is_neox_style: # NOTE(woosuk): Here we assume that the positions tensor has the # shape [batch_size, seq_len]. cos = cos.repeat(1, 1, 2).unsqueeze(-2) sin = sin.repeat(1, 1, 2).unsqueeze(-2) else: cos = cos.repeat_interleave(2, dim=-1).unsqueeze(-2) sin = sin.repeat_interleave(2, dim=-1).unsqueeze(-2) rotate_fn = rotate_neox if self.is_neox_style else rotate_gptj hidden_rot = hidden_rot * cos + rotate_fn(hidden_rot) * sin if self.rotary_dim < self.head_size: hidden = torch.cat((hidden_rot, hidden_pass), dim=-1) else: hidden = hidden_rot return hidden.flatten(-2).squeeze(0) def extra_repr(self) -> str: s = f"head_size={self.head_size}, rotary_dim={self.rotary_dim}" s += f", max_position_embeddings={self.max_position_embeddings}" s += f", base={self.base}, is_neox_style={self.is_neox_style}" s += f", chunk_size={self.chunk_size}, local_size={self.local_size}" return s
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/dual_chunk_rope.py", "license": "Apache License 2.0", "lines": 195, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/dynamic_ntk_alpha_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from .base import RotaryEmbedding class DynamicNTKAlphaRotaryEmbedding(RotaryEmbedding): """RotaryEmbedding extended with Dynamic NTK alpha. Based on the original RotaryEmbedding implementation. """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_alpha: float, dtype: torch.dtype, ) -> None: self.scaling_alpha = scaling_alpha super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_cos_sin_cache(self) -> torch.Tensor: # For Hunyuan DynamicNTKAlphaRotaryEmbedding max_len = self.max_position_embeddings base = self.base * self.scaling_alpha ** ( self.rotary_dim / (self.rotary_dim - 2) ) inv_freq = self._compute_inv_freq(base) t = torch.arange(max_len, dtype=torch.float) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() sin = freqs.sin() cache = torch.cat((cos, sin), dim=-1) return cache
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/dynamic_ntk_alpha_rope.py", "license": "Apache License 2.0", "lines": 35, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/dynamic_ntk_scaling_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Adapted from # https://github.com/huggingface/transformers/blob/v4.33.2/src/transformers/models/llama/modeling_llama.py # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from .base import RotaryEmbedding class DynamicNTKScalingRotaryEmbedding(RotaryEmbedding): """RotaryEmbedding extended with Dynamic NTK scaling. Credits to the Reddit users /u/bloc97 and /u/emozilla """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_factor: float, dtype: torch.dtype, ) -> None: self.scaling_factor = scaling_factor super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_cos_sin_cache(self) -> torch.Tensor: # NOTE(woosuk): self.max_position_embeddings is the original # maximum length before applying the rope scaling. # Thus, the maximum length after applying the rope scaling is # self.max_position_embeddings * self.scaling_factor. max_len = self.max_position_embeddings * self.scaling_factor base = self.base * ( (self.scaling_factor * max_len / self.max_position_embeddings) - (self.scaling_factor - 1) ) ** (self.rotary_dim / (self.rotary_dim - 2)) inv_freq = self._compute_inv_freq(base) t = torch.arange(max_len, dtype=torch.float) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() sin = freqs.sin() cache = torch.cat((cos, sin), dim=-1) return cache
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/dynamic_ntk_scaling_rope.py", "license": "Apache License 2.0", "lines": 60, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/linear_scaling_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Adapted from # https://github.com/huggingface/transformers/blob/v4.33.2/src/transformers/models/llama/modeling_llama.py # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from .base import RotaryEmbedding class LinearScalingRotaryEmbedding(RotaryEmbedding): """RotaryEmbedding extended with linear scaling. It supports multiple scaling factors. Since multiple LoRA adapters may have different scaling factors, we need multiple cos/sin caches. In this way, instead of running rotary embedding kernel per lora, we can run multiple lora in a batched way. In addition to that, we also keep the cos/sin cache for the scaling factor of 1 (default) at all times. Exemplary for two scaling factors x=1, y and z with embeddings [[x11, x12, ... x1m], ..., [xn1, xn2, ..., xnm]] and [[y11, y12, ... y1o], ..., [yn1, yn2, ..., yno]], and [[z11, z12, ... z1p], ..., [zn1, zn2, ..., znp]], we construct the cos/sin cache as follows: [[x11, x12, ... x1m, y11, y12, ... y1o, z11, z12, ... z1p], ... [xn1, xn2, ... xnm, yn1, yn2, ... yno, zn1, zn2, ... znp]] We then use offsets to index into the cos/sin cache for the respective scaling factors. The offset to cache can be accessed via `scaling_factor_to_offset` API. Credits to the Reddit user /u/kaiokendev """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_factors: list[float] | float, dtype: torch.dtype, ) -> None: if isinstance(scaling_factors, float): scaling_factors = [scaling_factors] self.scaling_factors: list[float] = scaling_factors # noqa super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) # Lazy initialized. self._scaling_factor_to_offset: dict[float, int] def _compute_cos_sin_cache(self) -> torch.Tensor: inv_freq = self._compute_inv_freq(self.base) cache_list: list[torch.Tensor] = [] # offsets to the next cache in a tensor. # Each offset corresponds to the same index in scaling_factors. offsets: list[int] = [] for scaling_factor in self.scaling_factors: # NOTE(woosuk): self.max_position_embeddings is the original # maximum length before applying the rope scaling. # Thus, the maximum length after applying the rope scaling is # self.max_position_embeddings * self.scaling_factor. max_len = self.max_position_embeddings * scaling_factor t = torch.arange(max_len, dtype=torch.float) t = t / scaling_factor freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() sin = freqs.sin() cache = torch.cat((cos, sin), dim=-1) if not cache_list: offset = 0 else: last_offset = offsets[-1] next_max_len = cache_list[-1].shape[0] offset = last_offset + next_max_len offsets.append(offset) cache_list.append(cache) self._scaling_factor_to_offset = { float(scaling_factor): offsets[i] for i, scaling_factor in enumerate(self.scaling_factors) } assert len(self.scaling_factors) == len(offsets) return torch.cat(cache_list, dim=0) @property def scaling_factor_to_offset(self) -> dict[float, int]: return self._scaling_factor_to_offset
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/linear_scaling_rope.py", "license": "Apache License 2.0", "lines": 99, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/llama3_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import torch from .base import RotaryEmbedding class Llama3RotaryEmbedding(RotaryEmbedding): def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, scaling_factor: float, low_freq_factor: float, high_freq_factor: float, orig_max_position: int, ) -> None: self.scaling_factor = scaling_factor self.low_freq_factor = low_freq_factor self.high_freq_factor = high_freq_factor self.orig_max_position = orig_max_position super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_inv_freq(self, base: float) -> torch.Tensor: inv_freqs = super()._compute_inv_freq(base) low_freq_wavelen = self.orig_max_position / self.low_freq_factor high_freq_wavelen = self.orig_max_position / self.high_freq_factor wave_len = 2 * math.pi / inv_freqs if self.low_freq_factor != self.high_freq_factor: smooth = (self.orig_max_position / wave_len - self.low_freq_factor) / ( self.high_freq_factor - self.low_freq_factor ) else: smooth = 0 new_freqs = torch.where( wave_len < high_freq_wavelen, inv_freqs, torch.where( wave_len > low_freq_wavelen, inv_freqs / self.scaling_factor, (1 - smooth) * inv_freqs / self.scaling_factor + smooth * inv_freqs, ), ) return new_freqs
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/llama3_rope.py", "license": "Apache License 2.0", "lines": 47, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/llama4_vision_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import torch from .base import RotaryEmbeddingBase class Llama4VisionRotaryEmbedding(RotaryEmbeddingBase): def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, ): super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_inv_freq(self, base: float) -> torch.Tensor: inv_freqs = super()._compute_inv_freq(base) inv_freqs = inv_freqs[: (self.rotary_dim // 2)] return inv_freqs def _compute_cos_sin_cache(self) -> torch.Tensor: inv_freq = self._compute_inv_freq(self.base) # self.max_position_embeddings here is number of image patches # i.e. (image_size // patch_size) ** 2 num_patches = self.max_position_embeddings img_idx = torch.arange(num_patches, dtype=torch.int32).reshape(num_patches, 1) img_idx = torch.cat([img_idx, img_idx[:1]], dim=0) img_idx[-1, -1] = -2 # set to ID_CLS_TOKEN num_patches_single_dim = int(math.sqrt(num_patches)) frequencies_x = img_idx % num_patches_single_dim frequencies_y = img_idx // num_patches_single_dim freqs_x = ( (frequencies_x + 1)[..., None] * inv_freq[None, None, :] ).repeat_interleave(2, dim=-1) freqs_y = ( (frequencies_y + 1)[..., None] * inv_freq[None, None, :] ).repeat_interleave(2, dim=-1) freqs = torch.cat([freqs_x, freqs_y], dim=-1).float().contiguous()[..., ::2] freqs = freqs.masked_fill(img_idx.reshape(-1, 1, 1) < 0, 0) cache = torch.view_as_complex( torch.stack([torch.cos(freqs), torch.sin(freqs)], dim=-1) ) return cache def forward_native( # type: ignore[override] self, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: assert key is not None # self.cos_sin_cache here is complex tensor so we cannot cast into # query's dtype directly with self._match_cos_sin_cache_dtype # NOTE: by not storing cos_sin_cache in self, we can avoid # memory buffer update which is costly to runtime cos_sin_cache: torch.Tensor = self.cos_sin_cache.to(query.device) query_ = torch.view_as_complex(query.float().reshape(*query.shape[:-1], -1, 2)) key_ = torch.view_as_complex(key.float().reshape(*key.shape[:-1], -1, 2)) broadcast_shape = [ d if i == 1 or i == (query_.ndim - 1) else 1 for i, d in enumerate(query_.shape) ] freqs_ci = cos_sin_cache.view(*broadcast_shape) query_out = torch.view_as_real(query_ * freqs_ci).flatten(3) key_out = torch.view_as_real(key_ * freqs_ci).flatten(3) return query_out.type_as(query), key_out.type_as(key) def forward_cuda( # type: ignore[override] self, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: return self.forward_native(query, key)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/llama4_vision_rope.py", "license": "Apache License 2.0", "lines": 72, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/mrope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import numpy as np import torch from vllm.triton_utils import tl, triton from .base import RotaryEmbeddingBase from .yarn_scaling_rope import YaRNScalingRotaryEmbedding, yarn_get_mscale @triton.jit def _triton_mrope_forward( q_ptr, k_ptr, cos, sin, num_tokens, n_qh: tl.constexpr, n_kh: tl.constexpr, hd: tl.constexpr, rd: tl.constexpr, pad_n_qh: tl.constexpr, pad_n_kh: tl.constexpr, pad_hd: tl.constexpr, mrope_section_t: tl.constexpr, mrope_section_h: tl.constexpr, mrope_section_w: tl.constexpr, is_interleaved: tl.constexpr, ): # Adapted from # https://github.com/linkedin/Liger-Kernel/blob/main/src/liger_kernel/ops/qwen2vl_mrope.py # This version supports flatten input tensors from vllm # and supports cos and sin cache with shape (3, num_tokens, head_dim // 2) # instead of (3, bsz, seq_len, head_dim), also supports interleaved rotary pid = tl.program_id(0) # locate start address q_ptr = q_ptr + pid * (n_qh * hd) k_ptr = k_ptr + pid * (n_kh * hd) # #################################################################### # get the cos(mθ_{i...d/2}) and sin(mθ_{i...d/2}) for token position # m of this program instance # #################################################################### # Note: cos and sin now have shape (3, num_tokens, head_dim // 2) # Updated stride calculation for half head_dim half_rd = rd // 2 t_cos = cos + pid * half_rd h_cos = t_cos + num_tokens * half_rd w_cos = h_cos + num_tokens * half_rd t_sin = sin + pid * half_rd h_sin = t_sin + num_tokens * half_rd w_sin = h_sin + num_tokens * half_rd # Updated offsets for half head_dim cos_offsets = tl.arange(0, pad_hd // 2) if is_interleaved: h_mask = ((cos_offsets % 3) == 1) & (cos_offsets <= 3 * mrope_section_h) w_mask = ((cos_offsets % 3) == 2) & (cos_offsets <= 3 * mrope_section_w) t_mask = ~(h_mask | w_mask) else: t_end = mrope_section_t h_end = t_end + mrope_section_h t_mask = cos_offsets < mrope_section_t h_mask = (t_end <= cos_offsets) & (cos_offsets < h_end) w_mask = (h_end <= cos_offsets) & (cos_offsets < half_rd) t_cos_row = tl.load(t_cos + cos_offsets, mask=t_mask, other=0) h_cos_row = tl.load(h_cos + cos_offsets, mask=h_mask, other=0) w_cos_row = tl.load(w_cos + cos_offsets, mask=w_mask, other=0) t_sin_row = tl.load(t_sin + cos_offsets, mask=t_mask, other=0) h_sin_row = tl.load(h_sin + cos_offsets, mask=h_mask, other=0) w_sin_row = tl.load(w_sin + cos_offsets, mask=w_mask, other=0) cos_row = t_cos_row + h_cos_row + w_cos_row sin_row = t_sin_row + h_sin_row + w_sin_row # #################################################################### # Load the left and right half of q and k for the current # program instance (i.e. for the current token) separately # #################################################################### # left half of the head first_half_q_offsets = ( tl.arange(0, pad_n_qh)[:, None] * hd + tl.arange(0, pad_hd // 2)[None, :] ) first_half_k_offsets = ( tl.arange(0, pad_n_kh)[:, None] * hd + tl.arange(0, pad_hd // 2)[None, :] ) first_q_mask = (tl.arange(0, pad_n_qh)[:, None] < n_qh) & ( tl.arange(0, pad_hd // 2)[None, :] < rd // 2 ) first_k_mask = (tl.arange(0, pad_n_kh)[:, None] < n_kh) & ( tl.arange(0, pad_hd // 2)[None, :] < rd // 2 ) q_tile_1 = tl.load(q_ptr + first_half_q_offsets, mask=first_q_mask, other=0).to( sin_row.dtype ) k_tile_1 = tl.load(k_ptr + first_half_k_offsets, mask=first_k_mask, other=0).to( sin_row.dtype ) # right half of the head second_half_q_offsets = first_half_q_offsets + (rd // 2) second_half_k_offsets = first_half_k_offsets + (rd // 2) second_q_mask = first_q_mask second_k_mask = first_k_mask q_tile_2 = tl.load(q_ptr + second_half_q_offsets, mask=second_q_mask, other=0).to( sin_row.dtype ) k_tile_2 = tl.load(k_ptr + second_half_k_offsets, mask=second_k_mask, other=0).to( sin_row.dtype ) # y = [x1, x2] * [cos, cos] + [-x2, x1] * [sin, sin] # Since cos and sin are now half-size, # we use the same cos_row and sin_row for both halves new_q_tile_1 = q_tile_1 * cos_row - q_tile_2 * sin_row tl.store(q_ptr + first_half_q_offsets, new_q_tile_1, mask=first_q_mask) new_q_tile_2 = q_tile_2 * cos_row + q_tile_1 * sin_row tl.store(q_ptr + second_half_q_offsets, new_q_tile_2, mask=second_q_mask) new_k_tile_1 = k_tile_1 * cos_row - k_tile_2 * sin_row tl.store(k_ptr + first_half_k_offsets, new_k_tile_1, mask=first_k_mask) new_k_tile_2 = k_tile_2 * cos_row + k_tile_1 * sin_row tl.store(k_ptr + second_half_k_offsets, new_k_tile_2, mask=second_k_mask) def triton_mrope( q: torch.Tensor, k: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, mrope_section: list[int], head_size: int, rotary_dim: int, mrope_interleaved: bool, ) -> tuple[torch.Tensor, torch.Tensor]: """Qwen2VL mrope kernel. Args: q: [num_tokens, num_heads * head_size] k: [num_tokens, num_kv_heads * head_size] cos: [3, num_tokens, head_size //2 ] (T/H/W positions with multimodal inputs) sin: [3, num_tokens, head_size //2 ] (T/H/W positions with multimodal inputs) mrope_section: [t, h, w] head_size: int """ n_row, n_q_head_head_dim = q.shape n_q_head = n_q_head_head_dim // head_size n_kv_head = k.shape[1] // head_size pad_hd = triton.next_power_of_2(head_size) pad_n_q_head = triton.next_power_of_2(n_q_head) pad_n_kv_head = triton.next_power_of_2(n_kv_head) # ensure tensors passed into the kernel are contiguous. # It will be no-op if they are already contiguous q = q.contiguous() k = k.contiguous() cos = cos.contiguous() sin = sin.contiguous() _triton_mrope_forward[(n_row,)]( q, k, cos, sin, n_row, n_q_head, n_kv_head, head_size, rotary_dim, pad_n_q_head, pad_n_kv_head, pad_hd, mrope_section[0], mrope_section[1], mrope_section[2], mrope_interleaved, ) return q, k def apply_interleaved_rope(x: torch.Tensor, mrope_section: list[int]) -> torch.Tensor: """Apply interleaved MRoPE to 3D rotary embeddings. Reorganizes frequency layout from chunked [TTT...HHH...WWW] to interleaved [THTHWHTHW...TT], preserving frequency continuity. """ x_t = x[0].clone() x_t[..., 1 : mrope_section[1] * 3 : 3] = x[1, ..., 1 : mrope_section[1] * 3 : 3] x_t[..., 2 : mrope_section[2] * 3 : 3] = x[2, ..., 2 : mrope_section[2] * 3 : 3] return x_t class MRotaryEmbedding(RotaryEmbeddingBase): """Rotary Embedding with Multimodal Sections.""" def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, mrope_section: list[int] | None = None, mrope_interleaved: bool = False, # YaRN parameters. *, scaling_factor: float | None = None, extrapolation_factor: float = 1, attn_factor: float = 1, beta_fast: int = 32, beta_slow: int = 1, truncate: bool = True, ) -> None: self.scaling_factor = scaling_factor self.extrapolation_factor = extrapolation_factor self.attn_factor = attn_factor self.beta_fast = beta_fast self.beta_slow = beta_slow self.truncate = truncate if self.scaling_factor is not None: # Get n-d magnitude scaling corrected for interpolation self.mscale = float(yarn_get_mscale(self.scaling_factor) * attn_factor) else: self.mscale = 1.0 # In Qwen2.5-VL, the maximum index value is related to the duration of # the input video. We enlarge max_position_embeddings to 4 times to get # a larger the cos and sin cache. self.cache_max_position_num = max_position_embeddings * 4 super().__init__( head_size, rotary_dim, self.cache_max_position_num, base, is_neox_style, dtype, ) self.mrope_section = mrope_section self.mrope_interleaved = mrope_interleaved if self.mrope_section: assert sum(self.mrope_section) == rotary_dim // 2 def _compute_inv_freq(self, base: float) -> torch.Tensor: if self.scaling_factor is None: return super()._compute_inv_freq(base) return YaRNScalingRotaryEmbedding._compute_inv_freq(self, base) def _compute_cos_sin_cache(self) -> torch.Tensor: if self.scaling_factor is None: return super()._compute_cos_sin_cache() return YaRNScalingRotaryEmbedding._compute_cos_sin_cache(self) def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """PyTorch-native implementation equivalent to forward(). Args: positions: [num_tokens,] (text only) or [3, num_tokens] (T/H/W positions with multimodal inputs) query: [num_tokens, num_heads * head_size] key: [num_tokens, num_kv_heads * head_size] """ assert positions.ndim == 1 or positions.ndim == 2 assert key is not None cos_sin_cache = self._match_cos_sin_cache_dtype(query) num_tokens = positions.shape[-1] cos_sin = cos_sin_cache[positions] cos, sin = cos_sin.chunk(2, dim=-1) if positions.ndim == 2: assert self.mrope_section if self.mrope_interleaved: cos = apply_interleaved_rope(cos, self.mrope_section) sin = apply_interleaved_rope(sin, self.mrope_section) else: cos = torch.cat( [m[i] for i, m in enumerate(cos.split(self.mrope_section, dim=-1))], dim=-1, ) sin = torch.cat( [m[i] for i, m in enumerate(sin.split(self.mrope_section, dim=-1))], dim=-1, ) query_shape = query.shape query = query.view(num_tokens, -1, self.head_size) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = self.apply_rotary_emb.forward_native( query_rot, cos, sin, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) key_shape = key.shape key = key.view(num_tokens, -1, self.head_size) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = self.apply_rotary_emb.forward_native( key_rot, cos, sin, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key def forward_cuda( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: assert positions.ndim == 1 or positions.ndim == 2 assert key is not None cos_sin_cache = self._match_cos_sin_cache_dtype(query) num_tokens = positions.shape[-1] cos_sin = cos_sin_cache[positions] cos, sin = cos_sin.chunk(2, dim=-1) query_shape = query.shape key_shape = key.shape if positions.ndim == 2: assert self.mrope_section q, k = triton_mrope( query, key, cos, sin, self.mrope_section, self.head_size, self.rotary_dim, self.mrope_interleaved, ) return q.reshape(query_shape), k.reshape(key_shape) query = query.view(num_tokens, -1, self.head_size) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = self.apply_rotary_emb( query_rot, cos, sin, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) key = key.view(num_tokens, -1, self.head_size) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = self.apply_rotary_emb( key_rot, cos, sin, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key def forward_cpu( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: return self.forward_native(positions, query, key, offsets) @staticmethod def get_next_input_positions( mrope_position_delta: int, context_len: int, seq_len: int, ) -> list[list[int]]: return [ list( range( context_len + mrope_position_delta, seq_len + mrope_position_delta ) ) for _ in range(3) ] @staticmethod def get_next_input_positions_tensor( out: np.ndarray, out_offset: int, mrope_position_delta: int, context_len: int, num_new_tokens: int, ): values = np.arange( mrope_position_delta + context_len, mrope_position_delta + context_len + num_new_tokens, dtype=out.dtype, ) out[:, out_offset : out_offset + num_new_tokens] = values
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/mrope.py", "license": "Apache License 2.0", "lines": 369, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/ntk_scaling_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from .base import RotaryEmbedding class NTKScalingRotaryEmbedding(RotaryEmbedding): """RotaryEmbedding extended with fixed and mixed NTK scaling. https://kexue.fm/archives/9706""" def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_factor: float, dtype: torch.dtype, mixed_b: float | None = None, ) -> None: self.scaling_factor = scaling_factor self.mixed_b = mixed_b super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_inv_freq(self, base: float) -> torch.Tensor: base = self.base * (self.scaling_factor if self.mixed_b is None else 1) inv_freq = super()._compute_inv_freq(base) if self.mixed_b is None: inv_freq = inv_freq / self.scaling_factor ** (2 / self.rotary_dim) else: a = ( torch.tensor(self.scaling_factor).log() / (self.rotary_dim / 2) ** self.mixed_b ) lambda_1_m = ( a * torch.arange(1, self.rotary_dim // 2 + 1).float() ** self.mixed_b ).exp() inv_freq = inv_freq / lambda_1_m return inv_freq
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/ntk_scaling_rope.py", "license": "Apache License 2.0", "lines": 38, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/phi3_long_rope_scaled_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import torch import torch.nn as nn from vllm.config import get_current_vllm_config from vllm.logger import init_logger from .common import rotate_neox logger = init_logger(__name__) class Phi3LongRoPEScaledRotaryEmbedding(nn.Module): """Phi3 family of models scaled rotary embedding. Based on the original RotaryEmbedding implementation. """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, original_max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, short_factor: list[float], long_factor: list[float], short_mscale: float | None = None, long_mscale: float | None = None, ): super().__init__() if is_neox_style is False: raise ValueError( "`Phi3LongRoPEScaledRotaryEmbedding` only supports neox_style." ) self.rotary_dim = rotary_dim self.head_size = head_size self.max_position_embeddings = max_position_embeddings self.original_max_position_embeddings = original_max_position_embeddings self.base = base self.short_factor = short_factor self.long_factor = long_factor # Force long factors if max_model_len (runtime max length) exceeds # original_max_position_embeddings to prevent KV cache invalidation when # sequences cross this threshold during generation max_model_len = get_current_vllm_config().model_config.max_model_len self.use_long_rope = max_model_len > original_max_position_embeddings if self.use_long_rope: logger.warning_once( "Using LongRoPE scaling factors. This enables longer " "contexts (%d tokens vs original %d tokens) at the cost of " "some performance degradation for shorter sequences. If " "this is not desired, set `max_model_len` to be at most %d.", max_position_embeddings, original_max_position_embeddings, original_max_position_embeddings, ) scale = self.max_position_embeddings / self.original_max_position_embeddings if scale <= 1.0: scaling_factor = 1.0 else: scaling_factor = math.sqrt( 1 + math.log(scale) / math.log(self.original_max_position_embeddings) ) if short_mscale is None: short_mscale = scaling_factor if long_mscale is None: long_mscale = scaling_factor self.short_mscale = short_mscale self.long_mscale = long_mscale short_cache = self._compute_cos_sin_cache( original_max_position_embeddings, short_factor, short_mscale ) short_cache = short_cache.to(dtype) long_cache = self._compute_cos_sin_cache( max_position_embeddings, long_factor, long_mscale ) long_cache = long_cache.to(dtype) long_short_cache = torch.cat([short_cache, long_cache], dim=0) self.register_buffer( "long_short_cos_sin_cache", long_short_cache, persistent=False ) def _compute_inv_freq(self, rescale_factors: list[float]) -> torch.Tensor: rescale_factors = torch.tensor(rescale_factors, dtype=torch.float32) inv_freq = 1.0 / ( rescale_factors * ( self.base ** ( torch.arange(0, self.rotary_dim, 2, dtype=torch.float) / self.rotary_dim ) ) ) return inv_freq def _compute_cos_sin_cache( self, max_position_embeddings: int, rescale_factors: list[float], mscale: float, ) -> torch.Tensor: inv_freq = self._compute_inv_freq(rescale_factors) t = torch.arange(max_position_embeddings, dtype=torch.float) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() * mscale sin = freqs.sin() * mscale cache = torch.cat((cos, sin), dim=-1) return cache def forward( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: assert key is not None query = query.view(*query.shape[:-1], -1, self.head_size) key = key.view(*key.shape[:-1], -1, self.head_size) if self.use_long_rope: k = self.original_max_position_embeddings long_prompt_offset = torch.full_like(positions, k).long() idx = torch.add(positions, long_prompt_offset) else: idx = positions idx = torch.add(idx, offsets) if offsets is not None else idx cos_sin = torch.index_select(self.long_short_cos_sin_cache, 0, idx) cos, sin = cos_sin.chunk(2, dim=-1) cos = cos.repeat(1, 2).unsqueeze(-2) sin = sin.repeat(1, 2).unsqueeze(-2) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = query_rot * cos + rotate_neox(query_rot) * sin query = torch.cat((query_rot, query_pass), dim=-1) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = key_rot * cos + rotate_neox(key_rot) * sin key = torch.cat((key_rot, key_pass), dim=-1) return query.flatten(-2), key.flatten(-2)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/phi3_long_rope_scaled_rope.py", "license": "Apache License 2.0", "lines": 135, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/yarn_scaling_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from .base import RotaryEmbedding from .common import yarn_find_correction_range, yarn_get_mscale, yarn_linear_ramp_mask class YaRNScalingRotaryEmbedding(RotaryEmbedding): """RotaryEmbedding extended with YaRN method. Credits to Peng et al. github.com/jquesnelle/yarn """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_factor: float, dtype: torch.dtype, *, extrapolation_factor: float = 1, attn_factor: float = 1, beta_fast: int = 32, beta_slow: int = 1, apply_yarn_scaling: bool = True, truncate: bool = True, ) -> None: self.scaling_factor = scaling_factor self.extrapolation_factor = extrapolation_factor self.attn_factor = attn_factor self.beta_fast = beta_fast self.beta_slow = beta_slow self.truncate = truncate # Get n-d magnitude scaling corrected for interpolation self.mscale = ( float(yarn_get_mscale(self.scaling_factor) * attn_factor) if apply_yarn_scaling else float(attn_factor) ) super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_inv_freq(self, scaling_factor: float) -> torch.Tensor: pos_freqs = self.base ** ( torch.arange(0, self.rotary_dim, 2, dtype=torch.float) / self.rotary_dim ) inv_freq_extrapolation = 1.0 / pos_freqs inv_freq_interpolation = 1.0 / (scaling_factor * pos_freqs) low, high = yarn_find_correction_range( self.beta_fast, self.beta_slow, self.rotary_dim, self.base, self.max_position_embeddings, self.truncate, ) # Get n-d rotational scaling corrected for extrapolation inv_freq_mask = ( 1 - yarn_linear_ramp_mask(low, high, self.rotary_dim // 2, dtype=torch.float) ) * self.extrapolation_factor inv_freq = ( inv_freq_interpolation * (1 - inv_freq_mask) + inv_freq_extrapolation * inv_freq_mask ) return inv_freq def _compute_cos_sin_cache(self) -> torch.Tensor: inv_freq = self._compute_inv_freq(self.scaling_factor) t = torch.arange( self.max_position_embeddings * self.scaling_factor, dtype=torch.float32 ) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() * self.mscale sin = freqs.sin() * self.mscale cache = torch.cat((cos, sin), dim=-1) return cache
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/yarn_scaling_rope.py", "license": "Apache License 2.0", "lines": 75, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/spec_decode/test_tree_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import pytest import torch from tests.v1.attention.utils import ( create_standard_kv_cache_spec, create_vllm_config, try_backend_includes_kv_cache_update, try_get_attention_backend, ) from vllm.config import ParallelConfig, SpeculativeConfig from vllm.platforms import current_platform from vllm.v1.attention.backend import CommonAttentionMetadata from vllm.v1.attention.backends.fa_utils import is_flash_attn_varlen_func_available from vllm.v1.attention.backends.registry import AttentionBackendEnum if not is_flash_attn_varlen_func_available(): pytest.skip( "This test requires flash_attn_varlen_func, but it's not available.", allow_module_level=True, ) # --------------------------------------------------------------------------- # # KV cache layout adaptation # --------------------------------------------------------------------------- # # Two KV cache layouts exist across backends: # # Flash layout: (2, num_blocks, block_size, num_kv_heads, head_size) # - dim 0 separates key (index 0) and value (index 1) # - Used by: FLASH_ATTN, TREE_ATTN, ROCM_AITER_FA, ROCM_ATTN # # Block layout: (num_blocks, 2, block_size, num_kv_heads, head_size) # - dim 1 separates key (index 0) and value (index 1) # - Used by: TRITON_ATTN # # The test creates KV caches in flash layout (the canonical format used by # tree attention). When a reference backend needs block layout we transpose # dims 0 and 1. # # Note: ROCM_ATTN uses flash layout for storage but its forward path calls # PagedAttention.split_kv_cache which reinterprets the raw memory as paged # layout (num_blocks, num_kv_heads, head_size//x, block_size, x). This is # a view-level incompatibility, not a transpose - see the TODO in # _get_available_reference_backends for details. # # TODO: Replace this mapping with a `KV_CACHE_LAYOUT` class attribute on each # AttentionImpl so the layout is self-documented by the backend itself, e.g.: # class TritonAttentionImpl(AttentionImpl): # KV_CACHE_LAYOUT = "block" # --------------------------------------------------------------------------- # _BLOCK_KV_LAYOUT_BACKENDS = frozenset( { AttentionBackendEnum.TRITON_ATTN, } ) # Backends whose do_kv_cache_update requires engine-level state (e.g. # ForwardContext) that is not available in this test harness, but whose # KV cache is flash layout and can be written with reshape_and_cache_flash. # When a backend is listed here, forward_attention() bypasses # do_kv_cache_update and writes directly to the cache. _NEEDS_DIRECT_CACHE_UPDATE = frozenset( { AttentionBackendEnum.ROCM_AITER_FA, } ) # Backends with known test-harness incompatibilities - see the TODOs # inside _get_available_reference_backends for details. _INCOMPATIBLE_REFERENCE_BACKENDS = frozenset( { AttentionBackendEnum.ROCM_AITER_FA, AttentionBackendEnum.ROCM_ATTN, } ) def _adapt_kv_cache_for_backend( kv_cache: torch.Tensor, backend: AttentionBackendEnum, ) -> torch.Tensor: """Convert kv_cache from flash layout ``(2, num_blocks, ...)`` to block layout ``(num_blocks, 2, ...)`` if the backend requires it. Returns the original tensor unchanged when no conversion is needed.""" if backend in _BLOCK_KV_LAYOUT_BACKENDS: return kv_cache.transpose(0, 1).contiguous() return kv_cache def _get_platform_default_backend() -> AttentionBackendEnum: """Ask the platform what backend it would auto-select at runtime.""" from vllm.v1.attention.selector import AttentionSelectorConfig config = AttentionSelectorConfig( block_size=32, kv_cache_dtype="auto", use_mla=False, use_sparse=False, head_size=128, dtype=torch.bfloat16, ) backend_path = current_platform.get_attn_backend_cls( selected_backend=None, attn_selector_config=config, ) for backend in AttentionBackendEnum: try: if backend.get_path() == backend_path: return backend except ValueError: continue raise RuntimeError( f"Platform returned backend path '{backend_path}' " f"that doesn't match any AttentionBackendEnum member." ) def _get_available_reference_backends() -> list[AttentionBackendEnum]: """Collect all reference backends the current platform can run. On CUDA this is just FLASH_ATTN. On ROCm this includes the platform default plus every backend the hardware supports, so the test validates tree attention against all of them. """ if current_platform.is_rocm(): backends: list[AttentionBackendEnum] = [] # 1. Whatever the platform would auto-select at runtime. default_backend = _get_platform_default_backend() if default_backend not in _INCOMPATIBLE_REFERENCE_BACKENDS: backends.append(default_backend) # 2. TRITON_ATTN - always available on ROCm. if AttentionBackendEnum.TRITON_ATTN not in backends: backends.append(AttentionBackendEnum.TRITON_ATTN) # TODO: Enable ROCM_ATTN. Its forward path uses # PagedAttention.split_kv_cache which reinterprets the raw # cache memory as paged layout: # key: (num_blocks, num_kv_heads, head_size//x, block_size, x) # value: (num_blocks, num_kv_heads, head_size, block_size) # Tree attention writes prefix data in NHD flash layout, so the # same bytes produce completely different values when read in # paged format. Supporting ROCM_ATTN would require writing # prefix data via PagedAttention.write_to_paged_cache into a # separate paged-format KV cache. # TODO: Enable ROCM_AITER_FA. Its metadata builder reads head # counts from the model config at construction time and # allocates extend_workspace with those dimensions. The test # uses independent head count parameters (num_heads=2/4, # num_kv_heads=2) that don't match the model config # (Llama-3-8B: 32 q heads, 8 kv heads), causing a head count # mismatch in flash_attn_varlen_func during extend_forward. # Fixing this requires either matching test head counts to the # model config or decoupling the builder from model config # head geometry. The direct cache update path # (_NEEDS_DIRECT_CACHE_UPDATE) is already in place for when # this is resolved. return backends # CUDA: flash attention. return [AttentionBackendEnum.FLASH_ATTN] class MockAttentionLayer(torch.nn.Module): _q_scale = torch.tensor(1.0, dtype=torch.float32, device="cuda") _k_scale = torch.tensor(1.0, dtype=torch.float32, device="cuda") _v_scale = torch.tensor(1.0, dtype=torch.float32, device="cuda") layer_name = "mock_layer" def __init__(self): super().__init__() def forward(self, x): return x def forward_attention( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, kv_cache: torch.Tensor, block_table: torch.Tensor, slot_mapping: torch.Tensor, seqlen_k: int, backend: AttentionBackendEnum, spec_token_tree: str | None = None, num_spec_tokens: int = 0, ) -> torch.Tensor: """Run a single attention forward pass through the given backend. ``kv_cache`` is expected in **flash layout** ``(2, num_blocks, block_size, num_kv_heads, head_size)``. It is automatically converted when the target backend needs a different layout. """ batch_size, q_len, num_heads, dim_per_head = q.shape num_kv_heads = k.shape[-2] # Initialize the query and KV sequence lengths. query_start_loc = q_len * torch.arange( batch_size + 1, device=q.device, dtype=torch.int32 ) query_lens = torch.diff(query_start_loc) seq_lens = torch.full( (batch_size,), seqlen_k, device=q.device, dtype=torch.int32, ) context_lens = seq_lens - query_lens max_seq_len = int(seq_lens.max()) max_query_len = q_len num_actual_tokens = query_start_loc[-1] softmax_scale = q.shape[-1] ** (-0.5) layer = MockAttentionLayer() # Build common metadata. model_name = "meta-llama/Meta-Llama-3-8B" builder_cls, impl_cls = try_get_attention_backend(backend) vllm_config = create_vllm_config(model_name=model_name, max_model_len=max(seq_lens)) if spec_token_tree is not None: # Create speculative config if token tree is specified. vllm_config.speculative_config = SpeculativeConfig( target_model_config=vllm_config.model_config, target_parallel_config=ParallelConfig(), model=model_name, method="eagle", num_speculative_tokens=num_spec_tokens, speculative_token_tree=spec_token_tree, ) kv_cache_spec = create_standard_kv_cache_spec(vllm_config) builder = builder_cls(kv_cache_spec, [], vllm_config, q.device) common_attn_metadata = CommonAttentionMetadata( query_start_loc=query_start_loc, query_start_loc_cpu=query_start_loc.cpu(), seq_lens=seq_lens, _seq_lens_cpu=seq_lens.cpu(), _num_computed_tokens_cpu=context_lens.cpu(), num_reqs=batch_size, num_actual_tokens=num_actual_tokens, max_query_len=max_query_len, max_seq_len=max_seq_len, block_table_tensor=block_table, slot_mapping=slot_mapping, ) # Build attention metadata. attn_metadata = builder.build( common_prefix_len=0, common_attn_metadata=common_attn_metadata, ) # Initialize the backend implementation. instance = impl_cls( num_heads=num_heads, head_size=dim_per_head, scale=softmax_scale, num_kv_heads=num_kv_heads, alibi_slopes=None, sliding_window=None, kv_cache_dtype="auto", ) # Adapt KV cache layout for this backend. adapted_kv_cache = _adapt_kv_cache_for_backend(kv_cache, backend) # Run forward pass and return output. query = q.view(-1, num_heads, dim_per_head) key = k.view(-1, num_kv_heads, dim_per_head) value = v.view(-1, num_kv_heads, dim_per_head) output = torch.empty_like(query) if not try_backend_includes_kv_cache_update(backend): if backend in _NEEDS_DIRECT_CACHE_UPDATE: # This backend's do_kv_cache_update requires engine-level # ForwardContext that isn't available in this test harness. # Write directly using reshape_and_cache_flash since the # KV cache layout is identical (flash layout, unbind on dim 0). key_cache, value_cache = adapted_kv_cache.unbind(0) torch.ops._C_cache_ops.reshape_and_cache_flash( key, value, key_cache, value_cache, attn_metadata.slot_mapping, "auto", layer._k_scale, layer._v_scale, ) else: instance.do_kv_cache_update( layer=layer, key=key, value=value, kv_cache=adapted_kv_cache, slot_mapping=attn_metadata.slot_mapping, ) return instance.forward( layer=layer, query=query, key=key, value=value, kv_cache=adapted_kv_cache.clone(), attn_metadata=attn_metadata, output=output, ) @pytest.mark.parametrize( "reference_backend", _get_available_reference_backends(), ids=lambda b: b.name, ) def test_tree_attn_correctness( reference_backend: AttentionBackendEnum, ) -> None: torch.manual_seed(42) torch.cuda.manual_seed_all(42) device = "cuda" tree_attn_masks = { # Chain. "[(0,), (0, 0), (0, 0, 0)]": torch.tensor( [ [1, 0, 0, 0], [1, 1, 0, 0], [1, 1, 1, 0], [1, 1, 1, 1], ], device=device, dtype=torch.int32, ), # Tree. "[(0,), (1,), (0, 0), (0, 1), (1, 0), (1, 1)]": torch.tensor( [ [1, 0, 0, 0, 0, 0, 0], [1, 1, 0, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [1, 1, 0, 1, 0, 0, 0], [1, 1, 0, 0, 1, 0, 0], [1, 0, 1, 0, 0, 1, 0], [1, 0, 1, 0, 0, 0, 1], ], device=device, dtype=torch.int32, ), } dim_per_head = 128 num_kv_heads = 2 block_size = 32 max_sequence_length = 8192 randomize_blocks = True for batch_size in [1, 16, 32]: for num_heads in [2, 4]: for sequence_position in [16, 1024, 2048]: for spec_token_tree, tree_attn_mask in tree_attn_masks.items(): # Assert that the number of heads is divisible # by the number of KV heads. assert num_heads % num_kv_heads == 0 # Initialize q, k, and v. tree_size_q = tree_attn_mask.shape[0] seqlen_k = sequence_position + tree_size_q q = torch.randn( (batch_size, tree_size_q, num_heads, dim_per_head), device=device, dtype=torch.bfloat16, ) k = torch.randn( (batch_size, tree_size_q, num_kv_heads, dim_per_head), device=device, dtype=torch.bfloat16, ) v = torch.randn( (batch_size, tree_size_q, num_kv_heads, dim_per_head), device=device, dtype=torch.bfloat16, ) # KV cache in flash layout - the canonical format for # tree attention. forward_attention() handles conversion # when needed. assert max_sequence_length % block_size == 0 max_blocks_per_batch = max_sequence_length // block_size kv_cache = torch.randn( ( 2, batch_size * max_blocks_per_batch, block_size, num_kv_heads, dim_per_head, ), device=q.device, dtype=torch.bfloat16, ) num_alloc_blocks_per_batch = math.ceil(seqlen_k / block_size) block_table = torch.zeros( (batch_size, max_blocks_per_batch), device=q.device, dtype=torch.int32, ) block_ids = torch.arange( 0, batch_size * num_alloc_blocks_per_batch, device=q.device, dtype=torch.int32, ) if randomize_blocks: # Randomize the block ids. block_ids = block_ids[torch.randperm(block_ids.numel())] block_table[:, :num_alloc_blocks_per_batch] = block_ids.view( -1, num_alloc_blocks_per_batch ) # Set up the slot mapping for the input KVs. tree_positions = sequence_position + torch.arange( 0, tree_size_q, device=q.device, dtype=torch.int64, ).repeat(batch_size, 1) tree_slot_mapping = _gen_slot_mapping( tree_positions, block_table, block_size ) # Compute attention for the tree. tree_attn_output = forward_attention( q=q, k=k, v=v, kv_cache=kv_cache, block_table=block_table, slot_mapping=tree_slot_mapping, seqlen_k=seqlen_k, backend=AttentionBackendEnum.TREE_ATTN, spec_token_tree=spec_token_tree, num_spec_tokens=tree_size_q - 1, ).view(batch_size, -1, num_heads, dim_per_head) # Verify each branch against the reference backend. for q_index in range(tree_size_q): # Get the q, k, and v for the branch. branch_mask = tree_attn_mask[q_index, :] branch_indices = torch.nonzero(branch_mask, as_tuple=True)[0] q_len = branch_indices.shape[0] q_branch = q[:, branch_indices] k_branch = k[:, branch_indices] v_branch = v[:, branch_indices] # Setup slot mapping for the branch. branch_positions = sequence_position + torch.arange( 0, q_len, device=q.device, dtype=torch.int64, ).repeat(batch_size, 1) branch_slot_mapping = _gen_slot_mapping( branch_positions, block_table, block_size ) # Reference attention for this branch. ref_output = forward_attention( q=q_branch, k=k_branch, v=v_branch, kv_cache=kv_cache, block_table=block_table, slot_mapping=branch_slot_mapping, seqlen_k=sequence_position + q_len, backend=reference_backend, ).view(batch_size, -1, num_heads, dim_per_head) # Compare the outputs. assert torch.allclose( tree_attn_output[:, branch_indices], ref_output, atol=7.81e-3, ), ( f"outputs are not close for " f"reference_backend: {reference_backend.name}, " f"batch_size: {batch_size}, " f"num_heads: {num_heads}, " f"sequence_position: {sequence_position}, " f"tree_attn_mask: {tree_attn_mask}, " f"q_index: {q_index}." ) def _gen_slot_mapping( positions: torch.Tensor, block_table: torch.Tensor, block_size: int ): block_indices = positions // block_size blocks = block_table.gather(dim=1, index=block_indices) return (blocks * block_size + positions % block_size).view(-1)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/spec_decode/test_tree_attention.py", "license": "Apache License 2.0", "lines": 453, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/v1/attention/backends/tree_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Attention layer with TreeAttention.""" import ast from dataclasses import dataclass from typing import ClassVar import torch from vllm import _custom_ops as ops from vllm.config import VllmConfig from vllm.logger import init_logger from vllm.v1.attention.backend import ( AttentionBackend, AttentionImpl, AttentionMetadataBuilder, AttentionType, CommonAttentionMetadata, MultipleOf, ) from vllm.v1.attention.backends.utils import ( split_decodes_and_prefills, ) from vllm.v1.attention.ops.triton_unified_attention import unified_attention from vllm.v1.kv_cache_interface import AttentionSpec logger = init_logger(__name__) class TreeAttentionBackend(AttentionBackend): accept_output_buffer: bool = True supported_dtypes: ClassVar[list[torch.dtype]] = [torch.float16, torch.bfloat16] @staticmethod def get_supported_kernel_block_sizes() -> list[int | MultipleOf]: return [MultipleOf(16)] @classmethod def get_supported_head_sizes(cls) -> list[int]: return [32, 64, 96, 128, 160, 192, 224, 256] @staticmethod def get_name() -> str: return "TREE_ATTN" @staticmethod def get_impl_cls() -> type["TreeAttentionImpl"]: return TreeAttentionImpl @staticmethod def get_kv_cache_shape( num_blocks: int, block_size: int, num_kv_heads: int, head_size: int, cache_dtype_str: str = "auto", ) -> tuple[int, ...]: if block_size % 16 != 0: raise ValueError("Block size must be a multiple of 16.") return (2, num_blocks, block_size, num_kv_heads, head_size) @staticmethod def get_builder_cls() -> type["TreeAttentionMetadataBuilder"]: return TreeAttentionMetadataBuilder @staticmethod def use_cascade_attention(*args, **kwargs) -> bool: return False @dataclass class TreeAttentionMetadata: num_actual_tokens: int # Number of tokens excluding padding. max_query_len: int query_start_loc: torch.Tensor max_seq_len: int seq_lens: torch.Tensor block_table: torch.Tensor slot_mapping: torch.Tensor num_prefill_tokens: int = 0 num_decode_tokens: int = 0 num_prefills: int = 0 num_decodes: int = 0 tree_attn_bias: torch.Tensor | None = None # Cached Prefill/decode metadata. _cached_prefill_metadata: "TreeAttentionMetadata | None" = None _cached_decode_metadata: "TreeAttentionMetadata | None" = None @property def prefill_metadata(self) -> "TreeAttentionMetadata | None": if self.num_prefills == 0: return None if self._cached_prefill_metadata is not None: # Recover cached prefill-phase attention # metadata structure return self._cached_prefill_metadata q_start_loc = self.query_start_loc[self.num_decodes :] q_seqlens = torch.diff(q_start_loc) kv_seqlens = self.seq_lens[self.num_decodes :] # Construct & cache prefill-phase attention metadata structure self._cached_prefill_metadata = TreeAttentionMetadata( num_actual_tokens=self.num_prefill_tokens, max_query_len=int(q_seqlens.max().item()), query_start_loc=q_start_loc - q_start_loc[0], max_seq_len=int(kv_seqlens.max().item()), seq_lens=kv_seqlens, block_table=self.block_table[self.num_decodes :], slot_mapping=self.slot_mapping[self.num_decode_tokens :], ) return self._cached_prefill_metadata @property def decode_metadata(self) -> "TreeAttentionMetadata | None": if self.num_decode_tokens == 0: return None if self._cached_decode_metadata is not None: # Recover cached decode-phase attention # metadata structure return self._cached_decode_metadata q_start_loc = self.query_start_loc[: self.num_decodes + 1] q_seqlens = torch.diff(q_start_loc) kv_seqlens = self.seq_lens[: self.num_decodes] # Construct & cache decode-phase attention metadata structure self._cached_decode_metadata = TreeAttentionMetadata( num_actual_tokens=self.num_decode_tokens, max_query_len=int(q_seqlens.max().item()), query_start_loc=q_start_loc, max_seq_len=int(kv_seqlens.max().item()), seq_lens=kv_seqlens, block_table=self.block_table[: self.num_decodes], slot_mapping=self.slot_mapping[: self.num_decode_tokens], tree_attn_bias=self.tree_attn_bias, ) return self._cached_decode_metadata class TreeAttentionMetadataBuilder(AttentionMetadataBuilder[TreeAttentionMetadata]): def __init__( self, kv_cache_spec: AttentionSpec, layer_names: list[str], vllm_config: VllmConfig, device: torch.device, ): super().__init__(kv_cache_spec, layer_names, vllm_config, device) self.block_size = kv_cache_spec.block_size spec_config = vllm_config.speculative_config spec_token_tree: str | None = None if spec := spec_config: spec_token_tree = spec.speculative_token_tree tree_choices: list[tuple[int, ...]] = ( ast.literal_eval(spec_token_tree) if spec_token_tree is not None else [(0,)] ) # Construct the tree attention bias. depth_counts = _get_depth_counts(tree_choices) self.tree_attn_bias = _prepare_tree_attn_bias( tree_choices, depth_counts, dtype=torch.float32, device=device, ) self.reorder_batch_threshold = self.tree_attn_bias.shape[0] def build( self, common_prefix_len: int, common_attn_metadata: CommonAttentionMetadata, fast_build: bool = False, ) -> TreeAttentionMetadata: decode_threshold = self.tree_attn_bias.shape[0] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( split_decodes_and_prefills( common_attn_metadata, decode_threshold=decode_threshold ) ) num_actual_tokens = common_attn_metadata.num_actual_tokens q_start_loc = common_attn_metadata.query_start_loc max_query_len = common_attn_metadata.max_query_len kv_seqlens = common_attn_metadata.seq_lens max_seq_len = common_attn_metadata.max_seq_len block_table = common_attn_metadata.block_table_tensor slot_mapping = common_attn_metadata.slot_mapping return TreeAttentionMetadata( num_actual_tokens=num_actual_tokens, num_prefill_tokens=num_prefill_tokens, num_decode_tokens=num_decode_tokens, num_prefills=num_prefills, num_decodes=num_decodes, max_query_len=max_query_len, query_start_loc=q_start_loc, max_seq_len=max_seq_len, seq_lens=kv_seqlens, block_table=block_table, slot_mapping=slot_mapping, tree_attn_bias=self.tree_attn_bias, ) def build_for_drafting( self, common_attn_metadata: CommonAttentionMetadata, draft_index: int, ) -> TreeAttentionMetadata: # Cache the original tree attention bias. orig_tree_attn_bias = self.tree_attn_bias if draft_index == 0: # Use prefill for drafting at the root level. self.tree_attn_bias = torch.empty(0) else: # Slice the tree attention bias for drafting. Exclude # the root level. start, end = 1, 1 + common_attn_metadata.max_query_len self.tree_attn_bias = self.tree_attn_bias[start:end, start:end].contiguous() # Build attention bias. attn_metadata = self.build(0, common_attn_metadata, fast_build=True) # Reset the tree attention bias to the original value. self.tree_attn_bias = orig_tree_attn_bias return attn_metadata def _get_depth_counts(sorted_tree_choices: list[tuple[int, ...]]) -> list[int]: # Count the number of choices at each depth of the tree. depth_counts = [] prev_depth = 0 for path in sorted_tree_choices: depth = len(path) if depth != prev_depth: depth_counts.append(0) depth_counts[depth - 1] += 1 prev_depth = depth return depth_counts def _prepare_tree_attn_bias( sorted_tree_choices: list[tuple[int, ...]], depth_counts: list[int], dtype: torch.dtype | None, device: torch.device | None, ) -> torch.Tensor: # +1 comes from the additional root node. tree_len = len(sorted_tree_choices) + 1 tree_attn_mask = torch.full( (tree_len, tree_len), -torch.inf, device=device, dtype=dtype ) # Set diagonal to all zeros. Each token should # attend to itself. mask_val = 0 for i in range(tree_len): tree_attn_mask[i, i] = mask_val # Set root to all zeros. All tokens attend to it. tree_attn_mask[:, 0] = mask_val # Set all ancestors to zeros. start = 0 for i in range(len(depth_counts)): for j in range(depth_counts[i]): cur_tree_choice = sorted_tree_choices[start + j] # Retrieve ancestor position. if len(cur_tree_choice) == 1: continue ancestor_idx = [] for c in range(len(cur_tree_choice) - 1): ancestor_idx.append( sorted_tree_choices.index(cur_tree_choice[: c + 1]) + 1 ) tree_attn_mask[j + start + 1, ancestor_idx] = mask_val start += depth_counts[i] return tree_attn_mask class TreeAttentionImpl(AttentionImpl): def __init__( self, num_heads: int, head_size: int, scale: float, num_kv_heads: int, alibi_slopes: list[float] | None, sliding_window: int | None, kv_cache_dtype: str, logits_soft_cap: float | None = None, attn_type: AttentionType = AttentionType.DECODER, kv_sharing_target_layer_name: str | None = None, ) -> None: self.num_heads = num_heads self.head_size = head_size self.scale = float(scale) self.num_kv_heads = num_kv_heads self.num_queries_per_kv = self.num_heads // self.num_kv_heads self.kv_cache_dtype = kv_cache_dtype self.kv_sharing_target_layer_name = kv_sharing_target_layer_name if alibi_slopes is not None: alibi_slopes = torch.tensor(alibi_slopes, dtype=torch.float32) self.alibi_slopes = alibi_slopes if logits_soft_cap is None: # Setting logits_soft_cap to 0 means no soft cap. logits_soft_cap = 0 self.logits_soft_cap = logits_soft_cap if sliding_window is None: self.sliding_window = (-1, -1) else: self.sliding_window = (sliding_window - 1, 0) if attn_type != AttentionType.DECODER: raise NotImplementedError( "Encoder self-attention and " "encoder/decoder cross-attention " "are not implemented for " "TreeAttentionImpl." ) def forward( self, layer: torch.nn.Module, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, kv_cache: torch.Tensor, attn_metadata: TreeAttentionMetadata, output: torch.Tensor | None = None, output_scale: torch.Tensor | None = None, output_block_scale: torch.Tensor | None = None, ) -> torch.Tensor: """Forward pass with TreeAttention. Args: query: shape = [num_tokens, num_heads, head_size] key: shape = [num_tokens, num_kv_heads, head_size] value: shape = [num_tokens, num_kv_heads, head_size] kv_cache: shape = [2, num_blocks, block_size, num_kv_heads, head_size] attn_metadata: Metadata for attention. Returns: shape = [num_tokens, num_heads * head_size] """ assert output is not None, "Output tensor must be provided." if output_scale is not None or output_block_scale is not None: raise NotImplementedError( "fused output quantization is not yet supported for TreeAttentionImpl" ) if attn_metadata is None: # Profiling run. return output.fill_(0) # Cache the input KVs. key_cache, value_cache = kv_cache.unbind(0) if self.kv_sharing_target_layer_name is None: # Reshape the input keys and values and store them in the cache. # Skip this if sharing KV cache with an earlier attention layer. # NOTE(woosuk): Here, key and value are padded while slot_mapping is # not padded. However, we don't need to do key[:num_actual_tokens] # and value[:num_actual_tokens] because the reshape_and_cache_flash # op uses the slot_mapping's shape to determine the number of # actual tokens. ops.reshape_and_cache_flash( key, value, key_cache, value_cache, attn_metadata.slot_mapping, self.kv_cache_dtype, layer._k_scale, layer._v_scale, ) num_actual_tokens = attn_metadata.num_actual_tokens num_decode_tokens = attn_metadata.num_decode_tokens descale_shape = (attn_metadata.query_start_loc.shape[0] - 1, key.shape[1]) if prefill_meta := attn_metadata.prefill_metadata: unified_attention( q=query[num_decode_tokens:num_actual_tokens], k=key_cache, v=value_cache, out=output[num_decode_tokens:num_actual_tokens], cu_seqlens_q=prefill_meta.query_start_loc, max_seqlen_q=prefill_meta.max_query_len, seqused_k=prefill_meta.seq_lens, max_seqlen_k=prefill_meta.max_seq_len, softmax_scale=self.scale, causal=True, alibi_slopes=self.alibi_slopes, window_size=self.sliding_window, block_table=prefill_meta.block_table, softcap=self.logits_soft_cap, q_descale=None, # Not supported k_descale=layer._k_scale.expand(descale_shape), v_descale=layer._v_scale.expand(descale_shape), ) if decode_meta := attn_metadata.decode_metadata: unified_attention( q=query[:num_decode_tokens], k=key_cache, v=value_cache, out=output[:num_decode_tokens], cu_seqlens_q=decode_meta.query_start_loc, max_seqlen_q=decode_meta.max_query_len, seqused_k=decode_meta.seq_lens, max_seqlen_k=decode_meta.max_seq_len, softmax_scale=self.scale, causal=True, alibi_slopes=self.alibi_slopes, qq_bias=decode_meta.tree_attn_bias, window_size=self.sliding_window, block_table=decode_meta.block_table, softcap=self.logits_soft_cap, q_descale=None, # Not supported k_descale=layer._k_scale.expand(descale_shape), v_descale=layer._v_scale.expand(descale_shape), ) return output
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/attention/backends/tree_attn.py", "license": "Apache License 2.0", "lines": 377, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/benchmarks/lib/ready_checker.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Utilities for checking endpoint readiness.""" import asyncio import time import aiohttp from tqdm.asyncio import tqdm from vllm.logger import init_logger from .endpoint_request_func import RequestFunc, RequestFuncInput, RequestFuncOutput logger = init_logger(__name__) async def wait_for_endpoint( request_func: RequestFunc, test_input: RequestFuncInput, session: aiohttp.ClientSession, timeout_seconds: int = 600, retry_interval: int = 5, ) -> RequestFuncOutput: """ Wait for an endpoint to become available before starting benchmarks. Args: request_func: The async request function to call test_input: The RequestFuncInput to test with timeout_seconds: Maximum time to wait in seconds (default: 10 minutes) retry_interval: Time between retries in seconds (default: 5 seconds) Returns: RequestFuncOutput: The successful response Raises: ValueError: If the endpoint doesn't become available within the timeout """ deadline = time.perf_counter() + timeout_seconds output = RequestFuncOutput(success=False) print(f"Waiting for endpoint to become up in {timeout_seconds} seconds") with tqdm( total=timeout_seconds, bar_format="{desc} |{bar}| {elapsed} elapsed, {remaining} remaining", unit="s", ) as pbar: while True: # update progress bar remaining = deadline - time.perf_counter() elapsed = timeout_seconds - remaining update_amount = min(elapsed - pbar.n, timeout_seconds - pbar.n) pbar.update(update_amount) pbar.refresh() if remaining <= 0: pbar.close() break # ping the endpoint using request_func try: output = await request_func( request_func_input=test_input, session=session ) if output.success: pbar.close() return output else: err_last_line = str(output.error).rstrip().rsplit("\n", 1)[-1] logger.warning("Endpoint is not ready. Error='%s'", err_last_line) except aiohttp.ClientConnectorError: pass # retry after a delay sleep_duration = min(retry_interval, remaining) if sleep_duration > 0: await asyncio.sleep(sleep_duration) return output
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/benchmarks/lib/ready_checker.py", "license": "Apache License 2.0", "lines": 65, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/ray/lazy_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project def is_ray_initialized(): """Check if Ray is initialized.""" try: import ray return ray.is_initialized() except ImportError: return False except AttributeError: return False def is_in_ray_actor(): """Check if we are in a Ray actor.""" try: import ray return ( ray.is_initialized() and ray.get_runtime_context().get_actor_id() is not None ) except ImportError: return False except AttributeError: return False
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/ray/lazy_utils.py", "license": "Apache License 2.0", "lines": 23, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/attention/test_attention_splitting.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from tests.v1.attention.test_attention_backends import BATCH_SPECS from tests.v1.attention.utils import BatchSpec, create_common_attn_metadata from vllm.v1.attention.backends.utils import ( split_decodes_and_prefills, ) from vllm.v1.worker.ubatch_utils import ( UBatchSlice, _make_metadata_with_slice, maybe_create_ubatch_slices, slice_query_start_locs, split_attn_metadata, ) @pytest.fixture def sample_query_start_loc(): """Sample query_start_loc tensor for testing""" return torch.tensor([0, 5, 12, 20, 35, 50]) def test_basic_slice_middle(sample_query_start_loc): """Test slicing from middle of tensor""" req_slice = slice(1, 3) # slice from index 1 to 3 result = slice_query_start_locs(sample_query_start_loc, req_slice) expected = torch.tensor([0, 7, 15]) assert torch.equal(result, expected) def test_slice_from_beginning(sample_query_start_loc): """Test slicing from the beginning of tensor""" req_slice = slice(0, 2) # slice from index 0 to 2 result = slice_query_start_locs(sample_query_start_loc, req_slice) expected = torch.tensor([0, 5, 12]) assert torch.equal(result, expected) def test_slice_to_end(sample_query_start_loc): """Test slicing to the end of tensor""" req_slice = slice(3, 5) # slice from index 3 to 5 (last index) result = slice_query_start_locs(sample_query_start_loc, req_slice) expected = torch.tensor([0, 15, 30]) assert torch.equal(result, expected) def test_single_element_slice(sample_query_start_loc): """Test slice that results in single element""" req_slice = slice(2, 3) # slice from index 2 to 3 result = slice_query_start_locs(sample_query_start_loc, req_slice) expected = torch.tensor([0, 8]) assert torch.equal(result, expected) def test_full_tensor_slice(sample_query_start_loc): """Test slicing the entire tensor""" req_slice = slice(0, 5) # slice entire tensor result = slice_query_start_locs(sample_query_start_loc, req_slice) expected = torch.tensor([0, 5, 12, 20, 35, 50]) assert torch.equal(result, expected) def test_slice_bounds_edge_cases(sample_query_start_loc): # Test slice that goes exactly to the last element req_slice = slice(4, 5) # Last index result = slice_query_start_locs(sample_query_start_loc, req_slice) expected = torch.tensor([0, 15]) assert torch.equal(result, expected) @pytest.fixture def small_decode_metadata(): """Create metadata for small decode batch""" batch_spec = BATCH_SPECS["small_decode"] device = torch.device("cpu") return create_common_attn_metadata(batch_spec, block_size=16, device=device) @pytest.fixture def large_decode_metadata(): """Create metadata for small decode batch""" batch_spec = BATCH_SPECS["large_decode"] device = torch.device("cpu") return create_common_attn_metadata(batch_spec, block_size=16, device=device) @pytest.fixture def mixed_small_metadata(): """Create metadata for mixed small batch""" batch_spec = BATCH_SPECS["mixed_small"] device = torch.device("cpu") return create_common_attn_metadata(batch_spec, block_size=16, device=device) # Tests for _make_metadata_with_slice def test_make_metadata_with_slice_decode_batch(small_decode_metadata): """Test slicing decode batch metadata""" # Split first request only ubatch_slice = UBatchSlice(slice(0, 1), slice(0, 1)) result = _make_metadata_with_slice(ubatch_slice, small_decode_metadata) # Check sliced results assert result.num_reqs == 1 # slice(0, 1) gives 1 requests assert result.num_actual_tokens == 1 # slice(0, 1) gives 1 token assert result.max_query_len == 1 assert torch.equal(result.query_start_loc, torch.tensor([0, 1])) assert torch.equal(result.seq_lens, torch.tensor([32])) def test_make_metadata_with_slice_mixed_batch(mixed_small_metadata): """Test slicing mixed batch metadata""" ubatch_slice = UBatchSlice(slice(1, 3), slice(1, 7)) # Requests 1-3, tokens 1-7 result = _make_metadata_with_slice(ubatch_slice, mixed_small_metadata) assert result.num_reqs == 2 # slice(1, 3) gives 2 requests assert result.num_actual_tokens == 6 # slice(1, 7) gives 6 tokens assert result.max_query_len == 5 assert torch.equal(result.query_start_loc, torch.tensor([0, 1, 6])) assert torch.equal(result.seq_lens, torch.tensor([40, 48])) def test_split_attn_metadata_decode_batch(large_decode_metadata): """Test splitting decode batch into two equal parts""" num_tokens = large_decode_metadata.num_reqs mid_point = num_tokens // 2 ubatch_slices = [ UBatchSlice(slice(0, mid_point), slice(0, mid_point)), UBatchSlice(slice(mid_point, num_tokens), slice(mid_point, num_tokens)), ] results = split_attn_metadata(ubatch_slices, large_decode_metadata) assert len(results) == 2 # Check first split assert results[0].num_reqs == mid_point assert results[0].num_actual_tokens == mid_point assert torch.equal(results[0].seq_lens, torch.tensor([2048] * mid_point)) # Check second split assert results[1].num_reqs == mid_point assert results[1].num_actual_tokens == mid_point assert torch.equal(results[1].seq_lens, torch.tensor([2048] * mid_point)) def apply_split_decodes_and_prefills( query_lens: list[int], decode_threshold: int, require_uniform: bool, padded_num_tokens: int | None = None, ): """Helper function to apply split_decodes_and_prefills and return the results.""" device = torch.device("cpu") seq_lens = [10 * (i + 1) for i in range(len(query_lens))] common_metadata = create_common_attn_metadata( BatchSpec(seq_lens=seq_lens, query_lens=query_lens), block_size=16, device=device, ) if padded_num_tokens is not None: common_metadata.num_actual_tokens = padded_num_tokens return split_decodes_and_prefills( common_metadata, decode_threshold=decode_threshold, require_uniform=require_uniform, ) def test_split_decodes_and_prefills_nonuniform_all_ones(): query_lens = [1, 1, 1] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( apply_split_decodes_and_prefills(query_lens, 1, False) ) assert num_decodes == 3 assert num_prefills == 0 assert num_decode_tokens == 3 assert num_prefill_tokens == 0 def test_split_decodes_and_prefills_nonuniform_all_short_decodes(): query_lens = [1, 2, 1, 3, 2, 1, 2] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( apply_split_decodes_and_prefills(query_lens, 3, False) ) assert num_decodes == 7 assert num_prefills == 0 assert num_decode_tokens == sum(query_lens) assert num_prefill_tokens == 0 def test_split_decodes_and_prefills_nonuniform_all_prefills(): query_lens = [4, 5, 6, 7] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( apply_split_decodes_and_prefills(query_lens, 3, False) ) assert num_decodes == 0 assert num_prefills == 4 assert num_decode_tokens == 0 assert num_prefill_tokens == sum(query_lens) def test_split_decodes_and_prefills_nonuniform_mixed_batch(): query_lens = [2, 1, 3, 4, 5, 6, 7, 8] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( apply_split_decodes_and_prefills(query_lens, 4, False) ) assert num_decodes == 4 # 2, 1, 3, 4 are all <= 4 assert num_prefills == 4 # 5, 6, 7, 8 are all > 4 assert num_decode_tokens == 10 # 2 + 1 + 3 + 4 assert num_prefill_tokens == 26 # 5 + 6 + 7 + 8 def test_split_decodes_and_prefills_uniform_all_ones(): query_lens = [1, 1, 1] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( apply_split_decodes_and_prefills(query_lens, 1, True) ) assert num_decodes == 3 assert num_prefills == 0 assert num_decode_tokens == 3 assert num_prefill_tokens == 0 def test_split_decodes_and_prefills_uniform_all_short_decodes(): query_lens = [2, 2, 1, 3, 2, 1, 2] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( apply_split_decodes_and_prefills(query_lens, 3, True) ) assert num_decodes == 2 assert num_prefills == 5 assert num_decode_tokens == 4 assert num_prefill_tokens == (1 + 3 + 2 + 1 + 2) def test_split_decodes_and_prefills_uniform_all_prefills(): query_lens = [4, 5, 6, 7] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( apply_split_decodes_and_prefills(query_lens, 3, True) ) assert num_decodes == 0 assert num_prefills == 4 assert num_decode_tokens == 0 assert num_prefill_tokens == sum(query_lens) def test_split_decodes_and_prefills_uniform_mixed_batch_all_uniform_decodes(): query_lens = [2, 2, 2, 4, 5, 6, 7, 8] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( apply_split_decodes_and_prefills(query_lens, 4, True) ) assert num_decodes == 3 # 2, 2, 2 are all <= 4 and uniform assert num_prefills == 5 # 4, 5, 6, 7, 8 are all > 4 assert num_decode_tokens == 6 # 2 + 2 + 2 assert num_prefill_tokens == 30 # 4 + 5 + 6 + 7 + 8 def test_split_decodes_and_prefills_uniform_mixed_batch_non_uniform_decodes(): query_lens = [2, 1, 2, 4, 5, 6, 7, 8] num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( apply_split_decodes_and_prefills(query_lens, 4, True) ) assert num_decodes == 1 # only the first 2 is taken as decode assert num_prefills == 7 # 1, 2, 4, 5, 6, 7, 8 are all > 4 or non-uniform assert num_decode_tokens == 2 # only the first 2 assert num_prefill_tokens == (sum(query_lens) - 2) # rest of the tokens def test_split_decodes_and_prefills_uniform_padded_batch_all_same(): """uniform batch where all query lengths are identical with 0 length padded reqs.""" # All query lengths are 2, with decode_threshold=3 (so 2 <= 3) # This triggers the padded uniform path at line 891 query_lens = [2, 2, 2, 0] padded_num_tokens = 8 num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( apply_split_decodes_and_prefills(query_lens, 3, True, padded_num_tokens) ) # With uniform batch, all requests are treated as decodes assert num_decodes == 4 assert num_prefills == 0 assert num_decode_tokens == padded_num_tokens assert num_prefill_tokens == 0 @pytest.mark.parametrize( "seq_lens,query_lens,split_point,expected_first_reqs,expected_second_reqs", [ # Split in the middle of request 1 ([32, 40], [8, 8], 12, 2, 1), # Split inside the first request ([32, 40], [8, 8], 4, 1, 2), ], ) def test_prefill_split_across_ubatches( seq_lens, query_lens, split_point, expected_first_reqs, expected_second_reqs ): """Test splitting a prefill across ubatches""" import numpy as np device = torch.device("cpu") batch_spec = BatchSpec(seq_lens=seq_lens, query_lens=query_lens) common = create_common_attn_metadata(batch_spec, block_size=16, device=device) num_scheduled_tokens = np.array(query_lens, dtype=np.int32) qsl_np = common.query_start_loc_cpu.numpy() num_tokens = common.num_actual_tokens ubatch_slices, _ = maybe_create_ubatch_slices( True, num_scheduled_tokens, num_tokens, batch_spec.batch_size, split_point=split_point, num_ubatches=2, ) assert ubatch_slices is not None and len(ubatch_slices) == 2 first_meta = _make_metadata_with_slice(ubatch_slices[0], common) second_meta = _make_metadata_with_slice(ubatch_slices[1], common) # Token counts match the split assert first_meta.num_actual_tokens == split_point assert second_meta.num_actual_tokens == num_tokens - split_point # Number of requests per ubatch assert first_meta.num_reqs == expected_first_reqs assert second_meta.num_reqs == expected_second_reqs # Identify which request is split and how many tokens are in the first chunk split_req_idx = int(np.searchsorted(qsl_np, split_point, side="right") - 1) tokens_in_first_chunk = split_point - int(qsl_np[split_req_idx]) orig_q_lens = common.query_start_loc_cpu[1:] - common.query_start_loc_cpu[:-1] # Check query length continuity: first-chunk + second-chunk == original qlen # First ubatch last request query length qlen_first_last = int( first_meta.query_start_loc_cpu[-1] - first_meta.query_start_loc_cpu[-2] ) # Second ubatch first request query length qlen_second_first = int( second_meta.query_start_loc_cpu[1] - second_meta.query_start_loc_cpu[0] ) assert qlen_first_last == tokens_in_first_chunk assert qlen_first_last + qlen_second_first == int(orig_q_lens[split_req_idx]) # Check seq_lens adjustments # Context lengths per original request context_lens = [s - q for s, q in zip(seq_lens, query_lens)] # First ubatch: last request's seq_len should be # context + tokens_in_first_chunk expected_seqlen = context_lens[split_req_idx] + tokens_in_first_chunk assert int(first_meta.seq_lens[-1]) == expected_seqlen # For full preceding requests in first ubatch, seq_lens should match # originals for i in range(first_meta.num_reqs - 1): assert int(first_meta.seq_lens[i]) == seq_lens[i] # Second ubatch: first request (continuation) seq_len should be full # original assert int(second_meta.seq_lens[0]) == seq_lens[split_req_idx] # Any following full requests in second ubatch should match originals for j in range(1, second_meta.num_reqs): # Map to original request index orig_idx = split_req_idx + j assert int(second_meta.seq_lens[j]) == seq_lens[orig_idx]
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/attention/test_attention_splitting.py", "license": "Apache License 2.0", "lines": 301, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/kernels/moe/test_batched_deepgemm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from vllm.model_executor.layers.fused_moe.batched_deep_gemm_moe import ( BatchedDeepGemmExperts, ) from vllm.model_executor.layers.fused_moe.config import fp8_w8a8_moe_quant_config from vllm.model_executor.layers.fused_moe.fused_batched_moe import ( BatchedPrepareAndFinalize, BatchedTritonExperts, ) from vllm.model_executor.layers.fused_moe.modular_kernel import FusedMoEModularKernel from vllm.utils.deep_gemm import calc_diff, is_deep_gemm_supported from .test_deepgemm import make_block_quant_fp8_weights from .utils import make_dummy_moe_config BLOCK_SIZE = [128, 128] @pytest.mark.skipif(not is_deep_gemm_supported(), reason="Requires deep_gemm kernels") @pytest.mark.parametrize("E", [16, 32]) # number of experts @pytest.mark.parametrize("T", [256, 512]) # tokens per expert @pytest.mark.parametrize("K", [128, 256]) # hidden dim @pytest.mark.parametrize("N", [512, 1024]) # intermediate dim per expert @pytest.mark.parametrize("topk", [2, 4]) def test_batched_deepgemm_vs_triton( E: int, T: int, K: int, N: int, topk: int, monkeypatch, workspace_init ): """Compare BatchedDeepGemmExperts to BatchedTritonExperts.""" monkeypatch.setenv("VLLM_USE_DEEP_GEMM", "1") device = "cuda" w1, w2, w1_s, w2_s = make_block_quant_fp8_weights(E, N, K, BLOCK_SIZE) M = E * T # total tokens a = torch.randn(M, K, device=device, dtype=torch.bfloat16) / 10.0 fp8_info = torch.finfo(torch.float8_e4m3fn) a.clamp_(fp8_info.min, fp8_info.max) # random router outputs → top-k indices / weights router_logits = torch.randn(M, E, device=device, dtype=torch.float32) topk_weights, topk_ids = torch.topk(router_logits, k=topk, dim=-1) topk_weights = torch.nn.functional.softmax(topk_weights, dim=-1) # token number for each expert cnt = torch.bincount(topk_ids.flatten(), minlength=E) max_cnt = int(cnt.max().item()) # next power of 2 for max token number max_num_tokens = 1 << (max_cnt - 1).bit_length() prep_finalize = BatchedPrepareAndFinalize( max_num_tokens=max_num_tokens, num_local_experts=E, num_dispatchers=1, rank=0, ) quant_config = fp8_w8a8_moe_quant_config( w1_scale=w1_s, w2_scale=w2_s, per_act_token_quant=False, block_shape=BLOCK_SIZE, ) # triton (reference) triton_experts = BatchedTritonExperts( max_num_tokens=max_num_tokens, num_dispatchers=1, quant_config=quant_config, moe_config=make_dummy_moe_config(), ) mk_triton = FusedMoEModularKernel( prep_finalize, triton_experts, inplace=False, ) out_triton = mk_triton( hidden_states=a, w1=w1, w2=w2, topk_weights=topk_weights, topk_ids=topk_ids, global_num_experts=E, ) # deepgemm deepgemm_experts = BatchedDeepGemmExperts( max_num_tokens=max_num_tokens, num_dispatchers=1, quant_config=quant_config, moe_config=make_dummy_moe_config(), ) mk_deepgemm = FusedMoEModularKernel( prep_finalize, deepgemm_experts, inplace=False, ) out_deepgemm = mk_deepgemm( hidden_states=a, w1=w1, w2=w2, topk_weights=topk_weights, topk_ids=topk_ids, global_num_experts=E, ) diff = calc_diff(out_deepgemm, out_triton) assert diff < 1e-3, f"Output diff too large: {diff}"
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/moe/test_batched_deepgemm.py", "license": "Apache License 2.0", "lines": 97, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:benchmarks/kernels/benchmark_reshape_and_cache_flash.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import random import time import torch from tabulate import tabulate from vllm import _custom_ops as ops from vllm.logger import init_logger from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.torch_utils import ( STR_DTYPE_TO_TORCH_DTYPE, create_kv_caches_with_random_flash, set_random_seed, ) from vllm.v1.attention.ops.triton_reshape_and_cache_flash import ( triton_reshape_and_cache_flash, ) logger = init_logger(__name__) @torch.inference_mode() def run_benchmark( num_tokens: int, num_heads: int, head_size: int, block_size: int, num_blocks: int, dtype: torch.dtype, kv_cache_dtype: str, kv_cache_layout: str, num_iters: int, implementation: str, benchmark_mode: str, device: str = "cuda", ) -> float: """Return latency (seconds) for given num_tokens.""" if kv_cache_dtype == "fp8" and head_size % 16: raise ValueError("fp8 kv-cache requires head_size to be a multiple of 16.") if implementation not in ("cuda", "triton"): raise ValueError( f"Unsupported implementation: {implementation}. " "Only 'cuda' and 'triton' are supported." ) if implementation == "triton" and kv_cache_layout == "HND": return float("nan") # Triton does not support HND layout yet. set_random_seed(42) torch.set_default_device(device) # create random key / value tensors [T, H, D]. key = torch.randn(num_tokens, num_heads, head_size, dtype=dtype, device=device) value = torch.randn_like(key) # prepare the slot mapping. # each token is assigned a unique slot in the KV-cache. num_slots = block_size * num_blocks if num_tokens > num_slots: raise ValueError("num_tokens cannot exceed the total number of cache slots") slot_mapping_lst = random.sample(range(num_slots), num_tokens) slot_mapping = torch.tensor(slot_mapping_lst, dtype=torch.long, device=device) key_caches, value_caches = create_kv_caches_with_random_flash( num_blocks, block_size, 1, # num_layers num_heads, head_size, kv_cache_dtype, dtype, device=device, cache_layout=kv_cache_layout, ) key_cache, value_cache = key_caches[0], value_caches[0] # to free unused memory del key_caches, value_caches # compute per-kernel scaling factors for fp8 conversion (if used). k_scale = (key.amax() / 64.0).to(torch.float32) v_scale = (value.amax() / 64.0).to(torch.float32) if implementation == "cuda": function_under_test = lambda: ops.reshape_and_cache_flash( key, # noqa: F821 value, # noqa: F821 key_cache, # noqa: F821 value_cache, # noqa: F821 slot_mapping, # noqa: F821 kv_cache_dtype, k_scale, v_scale, ) else: function_under_test = lambda: triton_reshape_and_cache_flash( key, # noqa: F821 value, # noqa: F821 key_cache, # noqa: F821 value_cache, # noqa: F821 slot_mapping, # noqa: F821 kv_cache_dtype, k_scale, v_scale, ) if benchmark_mode == "cudagraph": g = torch.cuda.CUDAGraph() with torch.cuda.graph(g): function_under_test() torch.cuda.synchronize() function_under_test = lambda: g.replay() def run_cuda_benchmark(n_iters: int) -> float: nonlocal key, value, key_cache, value_cache, slot_mapping torch.cuda.synchronize() start = time.perf_counter() for _ in range(n_iters): function_under_test() torch.cuda.synchronize() end = time.perf_counter() return (end - start) / n_iters # warm-up run_cuda_benchmark(3) lat = run_cuda_benchmark(num_iters) # free tensors to mitigate OOM when sweeping del key, value, key_cache, value_cache, slot_mapping torch.cuda.empty_cache() return lat def main(args): rows = [] for layout in ["NHD", "HND"]: for exp in range(1, 17): n_tok = 2**exp lat = run_benchmark( num_tokens=n_tok, num_heads=args.num_heads, head_size=args.head_size, block_size=args.block_size, num_blocks=args.num_blocks, dtype=STR_DTYPE_TO_TORCH_DTYPE[args.dtype], kv_cache_dtype=args.kv_cache_dtype, kv_cache_layout=layout, num_iters=args.iters, implementation=args.implementation, benchmark_mode=args.mode, device="cuda", ) rows.append([n_tok, layout, f"{lat * 1e6:.3f}"]) print( f"Benchmark results for implementation {args.implementation}" f" (measuring with {args.mode}):" ) print(tabulate(rows, headers=["num_tokens", "layout", "latency (µs)"])) if __name__ == "__main__": parser = FlexibleArgumentParser() parser.add_argument("--num-heads", type=int, default=128) parser.add_argument( "--head-size", type=int, choices=[64, 80, 96, 112, 120, 128, 192, 256], default=128, ) parser.add_argument("--block-size", type=int, choices=[16, 32], default=16) parser.add_argument("--num-blocks", type=int, default=128 * 512) parser.add_argument( "--dtype", type=str, choices=["half", "bfloat16", "float"], default="bfloat16", ) parser.add_argument( "--kv-cache-dtype", type=str, choices=["auto", "fp8"], default="auto", ) parser.add_argument("--iters", type=int, default=100) parser.add_argument( "--implementation", type=str, choices=["cuda", "triton"], default="cuda", ) parser.add_argument( "--mode", type=str, choices=["cudagraph", "no_graph"], default="cudagraph", ) args = parser.parse_args() main(args)
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/kernels/benchmark_reshape_and_cache_flash.py", "license": "Apache License 2.0", "lines": 179, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/distributed/device_communicators/ray_communicator.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import uuid from typing import Any import ray import torch from ray.exceptions import RayChannelError from ray.experimental.channel.communicator import Communicator, TorchTensorAllocator from torch.distributed import ReduceOp from vllm.distributed.device_communicators.base_device_communicator import ( DeviceCommunicatorBase, ) from vllm.distributed.parallel_state import get_pp_group from vllm.logger import init_logger from vllm.utils.torch_utils import current_stream logger = init_logger(__name__) class RayPPCommunicator(Communicator): """ Communicator to be used for pipeline parallelism in Ray Compiled Graph. This is wraps around the vLLM _PP GroupCoordinator. This class is not thread-safe. """ _comm: DeviceCommunicatorBase | None def __init__( self, world_size: int, comm_id: Any, rank: int | None, actor_handles: list["ray.actor.ActorHandle"], cuda_stream: torch.cuda.Stream | None, use_communication_streams: bool = False, ): """ Initialize a RayPPCommunicator that can be used to communicate with other Ray Compiled Graph actors for pipeline parallelism. Args: world_size: The number of participating actors. comm_id: A unique communicator ID. This is just to conform with the Ray Communicator API and is not used. rank: The rank of this actor. If None, then the caller is not a participant of the RayPPCommunicator group (e.g., the Ray driver). actor_handles: A list of actor handles. cuda_stream: A CUDA stream to dispatch communication ops to. This is not supported. use_communication_streams: Whether to use communication streams. This is not supported. """ self._world_size = world_size self._rank: int | None = None self._actor_handles = actor_handles if use_communication_streams: raise NotImplementedError("use_communication_streams is not supported") if cuda_stream is not None and cuda_stream != current_stream(): raise ValueError( "cuda_stream other than the current stream is not supported" ) if rank is not None: # Rank is not None, this is Ray worker assert ray.get_gpu_ids(), "RayPPCommunicator has no GPUs assigned" self._comm = get_pp_group().device_communicator assert self._comm is not None # Since we wrap around the vLLM _PP communicator, we use # the rank from the vLLM communicator, and ignore the rank # passed in from Ray. # TODO(rui): refactor the Ray Communicator API so that # it also supports no rank passed in. self._rank = self._comm.rank_in_group self._build_actor_rank_mapping() else: # Rank is None, this is Ray driver self._comm = None self._closed = False def _build_actor_rank_mapping(self): """ Use collective communication to build a mapping from actor IDs to ranks. This should be called once during initialization. """ if self._comm is None: return {} current_actor = ray.get_runtime_context().current_actor actor_id_str = current_actor._actor_id.hex() # Ray actor IDs are 32-character hex strings (128 bits) ACTOR_ID_LEN = 32 actor_id_bytes = bytearray(actor_id_str.encode("utf-8")) assert len(actor_id_bytes) == ACTOR_ID_LEN, ( f"Unexpected actor ID length: {len(actor_id_bytes)}" ) actor_id_tensor = torch.frombuffer(actor_id_bytes, dtype=torch.uint8).to( self._comm.device ) # All-gather full actor IDs from all actors gathered_ids = self._comm.all_gather(actor_id_tensor, dim=0) # Build mapping: actor_id -> device_comm_rank self._actor_id_to_rank = {} for rank in range(self._world_size): start_idx = rank * ACTOR_ID_LEN end_idx = (rank + 1) * ACTOR_ID_LEN actor_bytes = gathered_ids[start_idx:end_idx].cpu().numpy().tobytes() actor_id = actor_bytes.decode("utf-8") self._actor_id_to_rank[actor_id] = rank def initialize(self, rank: int) -> None: # No additional initialization is needed. pass def get_actor_handles(self) -> list["ray.actor.ActorHandle"]: return self._actor_handles def get_rank(self, actor: ray.actor.ActorHandle) -> int: """ Return the given actor's rank using device communicator collective ops. """ assert hasattr(self, "_actor_id_to_rank"), ( "Actor rank mapping not built. " "This should have been done during initialization." ) actor_id_str = actor._actor_id.hex() if actor_id_str in self._actor_id_to_rank: return self._actor_id_to_rank[actor_id_str] # type: ignore else: raise ValueError(f"Actor {actor} not found in communicator group") def get_self_rank(self) -> int | None: """ Return this actor's rank. """ return self._rank def get_world_size(self) -> int: """ Return the number of ranks in the RayPPCommunicator group. """ return self._world_size def send(self, buf: "torch.Tensor", peer_rank: int) -> None: """ Send a torch.Tensor to a peer. This returns when the send kernel has been queued, but the kernel may not have completed. Therefore, the caller should ensure that there are no concurrent writes to the sent `buf` until the send has finished. That is, either all writes should be submitted on the current stream (self._cuda_stream) or, if on a different stream, that stream should synchronize with the current stream. Args: buf: The torch.Tensor to send. It should already be on this actor's default device. peer_rank: The rank of the actor to send to. """ if self._closed: raise RayChannelError("RayPPCommunicator has been destroyed.") assert self._comm is not None self._comm.send(buf, peer_rank) def recv( self, shape: tuple[int, ...], dtype: "torch.dtype", peer_rank: int, allocator: TorchTensorAllocator, ) -> "torch.Tensor": """ Receive a torch.Tensor from a peer and synchronize the current stream. After this call returns, the receive buffer is safe to read from any stream. An RayChannelError will be raised if an error occurred (e.g., remote actor died), and the buffer is not safe to read. Args: shape: The shape of the tensor to receive. dtype: The dtype of the tensor to receive. peer_rank: The rank of the actor to receive from. allocator: The allocator to use to create the received tensor. This is ignored for this implementation. """ if self._closed: raise RayChannelError("RayPPCommunicator has been destroyed.") assert self._comm is not None size = torch.Size(shape) buf = self._comm.recv(size, dtype, src=peer_rank) # Buffer values are undefined if NCCL ops are aborted. Therefore, we # need to synchronize here and check that the channel is still # open to ensure that the receive buffer is valid. # TODO(swang): Avoid CUDA synchronization. current_stream().synchronize() if self._closed: raise RayChannelError("RayPPCommunicator has been destroyed.") return buf def allgather( self, send_buf: "torch.Tensor", recv_buf: "torch.Tensor", ): raise NotImplementedError("allgather is not supported") def allreduce( self, send_buf: "torch.Tensor", recv_buf: "torch.Tensor", op: ReduceOp = ReduceOp.SUM, ): raise NotImplementedError("allreduce is not supported") def reducescatter( self, send_buf: "torch.Tensor", recv_buf: "torch.Tensor", op: ReduceOp = ReduceOp.SUM, ): raise NotImplementedError("reducescatter is not supported") @property def recv_stream(self): return torch.cuda.StreamContext(current_stream()) @property def send_stream(self): return torch.cuda.StreamContext(current_stream()) def destroy(self) -> None: # Just sets a flag, vLLM manages the lifecycle of the underlying # _PP GroupCoordinator. self._closed = True def get_transport_name(self) -> str: return "nccl" @classmethod def generate_communicator_id(cls) -> Any: return uuid.uuid4()
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/distributed/device_communicators/ray_communicator.py", "license": "Apache License 2.0", "lines": 214, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/models/multimodal/processing/test_glm4_1v.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from vllm.assets.video import VideoAsset from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import batched_tensors_equal from vllm.multimodal.video import OpenCVDynamicVideoBackend, OpenCVVideoBackend from ...utils import build_model_context @pytest.mark.parametrize("model_id", ["zai-org/GLM-4.1V-9B-Thinking"]) @pytest.mark.parametrize("expected_toks_per_frame", [299]) @pytest.mark.parametrize( "num_frames, fps, expected_grid_t", [ # pre-sampled fixed frames (unexpected behavior, # but we still expect it to work without errors) (32, 1, 16), (32, 2, 16), (128, 1, 64), (128, 2, 64), # post-sampled frames (expected behavior) (-1, 1, 5), (-1, 2, 10), ], ) def test_processor_override( model_id: str, expected_toks_per_frame: int, expected_grid_t: int, fps: int, num_frames: int, ): """Ensure GLM4vMultiModalProcessor can handle video frames properly.""" ctx = build_model_context( model_id, mm_processor_kwargs=None, limit_mm_per_prompt={"video": 1}, ) processor = MULTIMODAL_REGISTRY.create_processor(ctx.model_config) tokenizer = processor.info.get_tokenizer() hf_processor_mm_kwargs = {"fps": fps} # Build the image str / prompt based on the number of images we pass video_assets = VideoAsset(name="baby_reading", num_frames=num_frames) prompt = "<|begin_of_video|><|video|><|end_of_video|>" video, metadata = video_assets.np_ndarrays, video_assets.metadata metadata["fps"] = fps mm_data = {"video": [(video, metadata)]} processed_inputs = processor( prompt, mm_items=processor.info.parse_mm_data(mm_data), hf_processor_mm_kwargs=hf_processor_mm_kwargs, ) # Ensure we have the right number of placeholders per num_crops size hf_processor = processor.info.get_hf_processor(**hf_processor_mm_kwargs) video_token_id = tokenizer.convert_tokens_to_ids(hf_processor.video_token) video_tok_count = processed_inputs["prompt_token_ids"].count(video_token_id) grid_t, _, _ = processed_inputs["mm_kwargs"].get_data()["video_grid_thw"][0] assert grid_t == expected_grid_t assert video_tok_count == expected_toks_per_frame * grid_t @pytest.mark.parametrize("model_id", ["zai-org/GLM-4.1V-9B-Thinking"]) @pytest.mark.parametrize("fps", [2]) def test_video_loader_consistency( model_id: str, fps: int, ): """ Ensure dynamic video loader (pre-sampled by loader) and normal video loader (post-sampled by processor) produce same video processing outputs. """ ctx = build_model_context( model_id, mm_processor_kwargs=None, limit_mm_per_prompt={"video": 1}, ) processor = MULTIMODAL_REGISTRY.create_processor(ctx.model_config) hf_processor_mm_kwargs = {"fps": fps} # Build the image str / prompt based on the number of images we pass prompt = "<|begin_of_video|><|video|><|end_of_video|>" video_path = VideoAsset(name="baby_reading", num_frames=-1).video_path with open(video_path, "rb") as f: video_bytes = f.read() static_video, static_metadata = OpenCVVideoBackend.load_bytes(video_bytes) dynamic_video, dynamic_metadata = OpenCVDynamicVideoBackend.load_bytes( video_bytes, fps=fps ) # pre-sampled loader shouldn't read all frames assert len(dynamic_video) < len(static_video) static_mm_data = {"video": [(static_video, static_metadata)]} dynamic_mm_data = {"video": [(dynamic_video, dynamic_metadata)]} static_outputs = processor( prompt, mm_items=processor.info.parse_mm_data(static_mm_data), hf_processor_mm_kwargs=hf_processor_mm_kwargs, ) dynamic_outputs = processor( prompt, mm_items=processor.info.parse_mm_data(dynamic_mm_data), hf_processor_mm_kwargs=hf_processor_mm_kwargs, ) assert static_outputs["prompt_token_ids"] == dynamic_outputs["prompt_token_ids"] assert batched_tensors_equal( static_outputs["mm_kwargs"].get_data(), dynamic_outputs["mm_kwargs"].get_data(), )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/multimodal/processing/test_glm4_1v.py", "license": "Apache License 2.0", "lines": 103, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/transformers_utils/configs/speculators/algos.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project SUPPORTED_SPECULATORS_TYPES = {} def register_speculator(name): def decorator(fn): SUPPORTED_SPECULATORS_TYPES[name] = fn return fn return decorator @register_speculator("eagle3") def update_eagle3(config_dict: dict, pre_trained_config: dict) -> None: """ Apply Eagle-3 specific configuration transformations to the `dict` used to construct the Transformers PreTrainedConfig. Eagle-3 specific fields: - draft_vocab_size: Size of the draft model's vocabulary - target_hidden_size: Hidden size of the target model - norm_before_residual: Whether to apply norm before residual connection - eagle_aux_hidden_state_layer_ids: List of layer indices from the base model to use as auxiliary inputs for the Eagle3 drafter. These layers provide intermediate hidden states that help the drafter make better predictions. This is the standard field used in Eagle3 checkpoints. """ pre_trained_config["draft_vocab_size"] = config_dict.get("draft_vocab_size") if config_dict.get("target_hidden_size") is not None: pre_trained_config["target_hidden_size"] = config_dict["target_hidden_size"] pre_trained_config["norm_before_residual"] = config_dict.get( "norm_before_residual", True ) pre_trained_config["architectures"] = ["Eagle3LlamaForCausalLM"] if config_dict.get("eagle_aux_hidden_state_layer_ids"): pre_trained_config["eagle_aux_hidden_state_layer_ids"] = config_dict[ "eagle_aux_hidden_state_layer_ids" ]
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/configs/speculators/algos.py", "license": "Apache License 2.0", "lines": 33, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/transformers_utils/configs/speculators/base.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os from typing import Any from transformers import PretrainedConfig from vllm.transformers_utils.configs.speculators.algos import ( SUPPORTED_SPECULATORS_TYPES, ) __all__ = ["SpeculatorsConfig"] class SpeculatorsConfig(PretrainedConfig): model_type = "speculators" @classmethod def from_pretrained( cls, pretrained_model_name_or_path: str | os.PathLike, **kwargs, ) -> "SpeculatorsConfig": """Load speculators Eagle config and convert to vLLM format.""" config_dict, _ = cls.get_config_dict(pretrained_model_name_or_path, **kwargs) vllm_config = cls.extract_transformers_pre_trained_config(config_dict) return cls(**vllm_config) @classmethod def extract_transformers_pre_trained_config( cls, config_dict: dict[str, Any] ) -> dict[str, Any]: """ Extract standard Transformers PreTrainedConfig config from speculators config. """ speculators_model_type = config_dict.get("speculators_model_type") if speculators_model_type not in SUPPORTED_SPECULATORS_TYPES: raise ValueError( f"Expected one of: {SUPPORTED_SPECULATORS_TYPES}. " "Please ensure you're loading a speculators-format model." ) # Start with transformer layer configuration if present pre_trained_config = config_dict.get("transformer_layer_config", {}) # Apply anything specific to the supported algorithm algo_updater = SUPPORTED_SPECULATORS_TYPES[speculators_model_type] algo_updater(config_dict=config_dict, pre_trained_config=pre_trained_config) return pre_trained_config @classmethod def extract_vllm_speculative_config( cls, config_dict: dict[str, Any] ) -> dict[str, Any]: """Extract vLLM speculative config from speculators config.""" # validate fields # TODO: @dsikka - use speculators pydantic model to validate cls.validate_speculators_config(config_dict=config_dict) # Convert from speculators config -> format that can be ingested by vLLM return cls.build_vllm_speculative_config(config_dict=config_dict) @classmethod def validate_speculators_config(cls, config_dict: dict[str, Any]) -> None: try: spec_config = config_dict["speculators_config"] methods = spec_config["proposal_methods"] first_method = methods[0] _ = first_method["speculative_tokens"] _ = spec_config["verifier"]["name_or_path"] _ = config_dict["speculators_model_type"] except (KeyError, IndexError, TypeError) as e: raise ValueError("Invalid speculators config structure") from e if "transformer_layer_config" not in config_dict: raise ValueError("Must provide transformer_layer_config") if not isinstance(config_dict["transformer_layer_config"], dict): raise TypeError( "'transformer_layer_config' must be a dictionary if provided" ) @classmethod def build_vllm_speculative_config( cls, config_dict: dict[str, Any] ) -> dict[str, Any]: """ Build vLLM-compatible speculative configuration from speculators format. This method extracts and transforms speculative configuration from the speculators format into the structure expected by vLLM. Args: config_dict: Configuration dictionary in speculators format Returns: Dictionary with vLLM-compatible speculative configuration """ # Extract speculators configuration spec_config = config_dict["speculators_config"] # Currently we only support one proposal method proposal_methods = spec_config.get("proposal_methods") if not proposal_methods: raise ValueError("No proposal methods found in speculators config") first_method = proposal_methods[0] num_speculative_tokens = first_method.get("speculative_tokens") if num_speculative_tokens is None: raise ValueError( f"Missing 'speculative_tokens' in proposal method. Got: {first_method}" ) # Build base vLLM speculative configuration return { "method": config_dict.get("speculators_model_type"), "num_speculative_tokens": num_speculative_tokens, }
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/configs/speculators/base.py", "license": "Apache License 2.0", "lines": 97, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/quantization/utils/flashinfer_fp4_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Utility helpers for NVFP4 + FlashInfer fused-MoE path""" from typing import TYPE_CHECKING import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm import _custom_ops as ops from vllm.logger import init_logger from vllm.model_executor.layers.fused_moe.activation import MoEActivation from vllm.model_executor.layers.fused_moe.config import ( FusedMoEConfig, FusedMoEParallelConfig, RoutingMethodType, ) from vllm.model_executor.layers.quantization.utils.flashinfer_utils import ( activation_to_flashinfer_int, align_fp4_moe_weights_for_fi, ) from vllm.model_executor.layers.quantization.utils.nvfp4_utils import ( swizzle_blockscale, ) from vllm.model_executor.layers.quantization.utils.quant_utils import ( QuantKey, kNvfp4Dynamic, kNvfp4Static, ) from vllm.platforms import current_platform if TYPE_CHECKING: from vllm.model_executor.layers.fused_moe.layer import FusedMoE from vllm.model_executor.layers.fused_moe.oracle.nvfp4 import ( NvFp4MoeBackend, ) logger = init_logger(__name__) __all__ = [ "reorder_w1w3_to_w3w1", ] # # Methods used by the oracle for kernel selection. # def _supports_current_device() -> bool: """Supports only Blackwell-family GPUs.""" p = current_platform return p.is_cuda() and p.is_device_capability_family(100) def _supports_no_act_and_mul() -> bool: """Supports non-gated MoE.""" return True def _supports_quant_scheme( weight_key: QuantKey | None, activation_key: QuantKey | None, ) -> bool: """Supports Nvfp4 quantization.""" SUPPORTED_W_A = [ (kNvfp4Static, kNvfp4Dynamic), ] return (weight_key, activation_key) in SUPPORTED_W_A def _supports_activation(activation: MoEActivation) -> bool: return activation in [MoEActivation.SILU, MoEActivation.RELU2_NO_MUL] def _supports_routing_method( routing_method: RoutingMethodType, ) -> bool: """Monolithic kernels need to express router support.""" # NOTE(rob): potentially allow others here. This is a conservative list. return routing_method in [ RoutingMethodType.DeepSeekV3, RoutingMethodType.Renormalize, RoutingMethodType.RenormalizeNaive, RoutingMethodType.Llama4, ] def _supports_parallel_config(moe_parallel_config: FusedMoEParallelConfig) -> bool: """ TRTLLM is a monolithic kernel that requires dispatch_router_logits() for the naive dispatch/combine path. DeepEP HT only implements dispatch() for the modular kernel path, so TRTLLM is incompatible with DeepEP HT. """ return not moe_parallel_config.use_deepep_ht_kernels def is_supported_config_trtllm( moe_config: FusedMoEConfig, weight_key: QuantKey | None, activation_key: QuantKey | None, activation_format: mk.FusedMoEActivationFormat, ) -> tuple[bool, str | None]: """ This method mirrors mk.FusedMoEPermuteExpertsUnpermute.is_supported_config """ def _make_reason(reason: str) -> str: return f"kernel does not support {reason}" if not _supports_current_device(): return False, _make_reason(f"current device {current_platform.device_name}") elif not (moe_config.is_act_and_mul or _supports_no_act_and_mul()): return False, _make_reason("no act_and_mul MLP layer") elif not _supports_activation(moe_config.activation): return False, _make_reason(f"{moe_config.activation} activation") elif not _supports_quant_scheme(weight_key, activation_key): return False, _make_reason(f"quantization scheme {weight_key}x{activation_key}") elif not _supports_parallel_config(moe_config.moe_parallel_config): return False, _make_reason(f"parallel config {moe_config.moe_parallel_config}") elif not _supports_routing_method(moe_config.routing_method): return False, _make_reason(f"routing method {moe_config.routing_method}") elif activation_format != mk.FusedMoEActivationFormat.Standard: return False, _make_reason(f"activation format {activation_format}") elif moe_config.hidden_dim % 512 != 0: return False, _make_reason( f"hidden_dim must be divisible by 512, found {moe_config.hidden_dim}" ) return True, None def reorder_w1w3_to_w3w1( weight: torch.Tensor, scale: torch.Tensor, dim: int = -2 ) -> tuple[torch.Tensor, torch.Tensor]: """Re-order the concatenated `[w1, w3]` tensors to `[w3, w1]`""" size = weight.size(dim) assert size % 2 == 0, f"Expected even size in dim {dim}, got {size}" half = size // 2 w1, w3 = weight.split(half, dim=dim) s1, s3 = scale.split(half, dim=dim) return ( torch.cat([w3, w1], dim=dim).contiguous(), torch.cat([s3, s1], dim=dim).contiguous(), ) def prepare_static_weights_for_trtllm_fp4_moe( # args_dequant, # args, gemm1_weights, gemm2_weights, gemm1_scales_linear_fp4_bytes, gemm2_scales_linear_fp4_bytes, hidden_size, intermediate_size, num_experts, is_gated_activation: bool, ): from flashinfer import nvfp4_block_scale_interleave from flashinfer.fused_moe.core import ( _maybe_get_cached_w3_w1_permute_indices, get_w2_permute_indices_with_cache, ) _cache_permute_indices: dict[torch.Size, torch.Tensor] = {} """Prepare quantized weights for kernel (done offline with weights).""" epilogue_tile_m = 128 # FIXME: this depends on the kernel internals gemm1_intermediate_size = ( 2 * intermediate_size if is_gated_activation else intermediate_size ) # Convert quantized weights to proper formats gemm1_weights_fp4 = gemm1_weights.view(torch.float8_e4m3fn).reshape( num_experts, gemm1_intermediate_size, hidden_size // 2 ) # packed fp4 gemm1_scales_linear_fp4 = gemm1_scales_linear_fp4_bytes.view( torch.float8_e4m3fn ).reshape( num_experts, gemm1_intermediate_size, hidden_size // 16 ) # fp8 scaling factors gemm2_weights_fp4 = gemm2_weights.view(torch.float8_e4m3fn).reshape( num_experts, hidden_size, intermediate_size // 2 ) # packed fp4 gemm2_scales_linear_fp4 = gemm2_scales_linear_fp4_bytes.view( torch.float8_e4m3fn ).reshape(num_experts, hidden_size, intermediate_size // 16) # fp8 scaling factors gemm1_weights_fp4_shuffled = [] gemm1_scales_fp4_shuffled = [] gemm2_weights_fp4_shuffled = [] gemm2_scales_fp4_shuffled = [] for i in range(num_experts): # Calculate the permute indices for the following: # 1. Reorder rows of W1 and scales for fused gated activation # 2. Shuffle weights and scaling factors for transposed mma output # for both w3_w1 and w2 weights and scale factors permute_indices = _maybe_get_cached_w3_w1_permute_indices( _cache_permute_indices, gemm1_weights_fp4[i].view(torch.uint8), epilogue_tile_m, is_gated_act_gemm=is_gated_activation, ) gemm1_weights_fp4_shuffled.append( gemm1_weights_fp4[i] .view(torch.uint8)[permute_indices.to(gemm1_weights_fp4.device)] .contiguous() ) permute_sf_indices = _maybe_get_cached_w3_w1_permute_indices( _cache_permute_indices, gemm1_scales_linear_fp4[i].view(torch.uint8), epilogue_tile_m, num_elts_per_sf=16, is_gated_act_gemm=is_gated_activation, ) gemm1_scales_fp4_shuffled.append( nvfp4_block_scale_interleave( gemm1_scales_linear_fp4[i] .view(torch.uint8)[ permute_sf_indices.to(gemm1_scales_linear_fp4.device) ] .contiguous() ) ) permute_indices = get_w2_permute_indices_with_cache( _cache_permute_indices, gemm2_weights_fp4[i].view(torch.uint8), epilogue_tile_m, ) gemm2_weights_fp4_shuffled.append( gemm2_weights_fp4[i] .view(torch.uint8)[permute_indices.to(gemm2_weights_fp4.device)] .contiguous() ) permute_sf_indices = get_w2_permute_indices_with_cache( _cache_permute_indices, gemm2_scales_linear_fp4[i].view(torch.uint8), epilogue_tile_m, num_elts_per_sf=16, ) gemm2_scales_fp4_shuffled.append( nvfp4_block_scale_interleave( gemm2_scales_linear_fp4[i] .view(torch.uint8)[ permute_sf_indices.to(gemm2_scales_linear_fp4.device) ] .contiguous() ) ) # Stack weights for all experts gemm1_weights_fp4_shuffled = torch.stack(gemm1_weights_fp4_shuffled) gemm1_scales_fp4_shuffled = ( torch.stack(gemm1_scales_fp4_shuffled) .view(torch.float8_e4m3fn) .reshape(num_experts, gemm1_intermediate_size, hidden_size // 16) ) gemm2_weights_fp4_shuffled = torch.stack(gemm2_weights_fp4_shuffled) gemm2_scales_fp4_shuffled = ( torch.stack(gemm2_scales_fp4_shuffled) .view(torch.float8_e4m3fn) .reshape(num_experts, hidden_size, intermediate_size // 16) ) return ( gemm1_weights_fp4_shuffled, gemm1_scales_fp4_shuffled, gemm2_weights_fp4_shuffled, gemm2_scales_fp4_shuffled, ) def flashinfer_trtllm_fp4_moe( layer: torch.nn.Module, x: torch.Tensor | tuple[torch.Tensor, torch.Tensor], router_logits: torch.Tensor, top_k: int, activation: MoEActivation, global_num_experts: int, num_expert_group: int | None, topk_group: int | None, custom_routing_function: object | None, e_score_correction_bias: torch.Tensor | None, ) -> torch.Tensor: """ Apply FlashInfer TensorRT-LLM FP4 MoE kernel. Args: layer: The MoE layer with weights and scales x: Input tensor router_logits: Router logits for expert selection top_k: Number of experts to select per token activation: Activation function to use global_num_experts: Total number of experts across all ranks num_expert_group: Number of expert groups (for grouped routing) topk_group: Top-k within each group custom_routing_function: Custom routing function (e.g., Llama4) e_score_correction_bias: Optional routing bias correction Returns: Output tensor from the MoE layer """ import flashinfer from vllm.model_executor.models.llama4 import Llama4MoE SUPPORTED_ACTIVATIONS = [MoEActivation.SILU, MoEActivation.RELU2_NO_MUL] assert activation in SUPPORTED_ACTIVATIONS, ( f"Only {SUPPORTED_ACTIVATIONS} activations are supported for FlashInfer " f"TRTLLM FP4 MoE, {activation} found instead." ) # Quantize input to FP4 if isinstance(x, tuple): hidden_states_fp4, hidden_states_scale_linear_fp4 = x else: # hidden_states is the already quantized (hidden_states_fp4, hidden_states_scale_linear_fp4) = ops.scaled_fp4_quant( x, layer.a1_gscale, is_sf_swizzled_layout=False ) # Determine routing method type use_llama4_routing = custom_routing_function is Llama4MoE.custom_routing_function routing_method_type = layer.routing_method_type if use_llama4_routing: routing_method_type = flashinfer.RoutingMethodType.Llama4 # Cast to Fp32 (required by kernel). router_logits = ( router_logits.to(torch.float32) if routing_method_type == RoutingMethodType.DeepSeekV3 else router_logits ) # Determine activation type activation_type = activation_to_flashinfer_int(layer.activation) # Call TRT-LLM FP4 block-scale MoE kernel out = flashinfer.fused_moe.trtllm_fp4_block_scale_moe( routing_logits=router_logits, routing_bias=e_score_correction_bias, hidden_states=hidden_states_fp4, hidden_states_scale=hidden_states_scale_linear_fp4.view( torch.float8_e4m3fn ).reshape(*hidden_states_fp4.shape[:-1], -1), gemm1_weights=layer.w13_weight.data, gemm1_weights_scale=layer.w13_weight_scale.data.view(torch.float8_e4m3fn), gemm1_bias=None, gemm1_alpha=None, gemm1_beta=None, gemm1_clamp_limit=None, gemm2_weights=layer.w2_weight.data, gemm2_weights_scale=layer.w2_weight_scale.data.view(torch.float8_e4m3fn), gemm2_bias=None, output1_scale_scalar=layer.g1_scale_c.data, output1_scale_gate_scalar=layer.g1_alphas.data, output2_scale_scalar=layer.g2_alphas.data, num_experts=global_num_experts, top_k=top_k, n_group=num_expert_group if num_expert_group is not None else 0, topk_group=topk_group if topk_group is not None else 0, intermediate_size=layer.intermediate_size_per_partition, local_expert_offset=layer.ep_rank * layer.local_num_experts, local_num_experts=layer.local_num_experts, routed_scaling_factor=None, routing_method_type=routing_method_type, do_finalize=True, activation_type=activation_type, )[0] return out def flashinfer_trtllm_fp4_routed_moe( layer: torch.nn.Module, x: torch.Tensor, topk_ids: torch.Tensor, topk_weights: torch.Tensor, top_k: int, activation: MoEActivation, global_num_experts: int, ) -> torch.Tensor: """ Apply FlashInfer TensorRT-LLM FP4 MoE kernel. Uses packed input top k expert indices and scores rather than computing top k expert indices from scores. Args: layer: The MoE layer with weights and scales x: Input tensor topk_ids: Ids of selected experts top_k: Number of experts to select per token activation: Activation function to use global_num_experts: Total number of experts across all ranks Returns: Output tensor from the MoE layer """ import flashinfer # https://github.com/flashinfer-ai/flashinfer/blob/f0277fd1bff90e309e5c19cab36c5dae056d685d/flashinfer/fused_moe/core.py#L2535 assert activation == MoEActivation.SILU, ( "Only SiLU activation is supported for FlashInfer TRTLLM FP4 Routed MoE. " f"{activation} found instead." ) # Pack top k ids and expert weights into a single int32 tensor, as # required by TRT-LLM packed_tensor = (topk_ids.to(torch.int32) << 16) | topk_weights.to( torch.bfloat16 ).view(torch.int16) if isinstance(x, tuple): # Hidden_states is the already quantized hidden_states_fp4, hidden_states_scale_linear_fp4 = x else: # Quantize input to FP4 (hidden_states_fp4, hidden_states_scale_linear_fp4) = ops.scaled_fp4_quant( x, layer.a1_gscale, is_sf_swizzled_layout=False ) # Call TRT-LLM FP4 block-scale MoE kernel out = flashinfer.fused_moe.trtllm_fp4_block_scale_routed_moe( topk_ids=packed_tensor, routing_bias=None, hidden_states=hidden_states_fp4, hidden_states_scale=hidden_states_scale_linear_fp4.view( torch.float8_e4m3fn ).reshape(*hidden_states_fp4.shape[:-1], -1), gemm1_weights=layer.w13_weight.data, gemm1_weights_scale=layer.w13_weight_scale.data.view(torch.float8_e4m3fn), gemm1_bias=None, gemm1_alpha=None, gemm1_beta=None, gemm1_clamp_limit=None, gemm2_weights=layer.w2_weight.data, gemm2_weights_scale=layer.w2_weight_scale.data.view(torch.float8_e4m3fn), gemm2_bias=None, output1_scale_scalar=layer.g1_scale_c.data, output1_scale_gate_scalar=layer.g1_alphas.data, output2_scale_scalar=layer.g2_alphas.data, num_experts=global_num_experts, top_k=top_k, n_group=0, topk_group=0, intermediate_size=layer.intermediate_size_per_partition, local_expert_offset=layer.ep_rank * layer.local_num_experts, local_num_experts=layer.local_num_experts, routed_scaling_factor=None, routing_method_type=1, do_finalize=True, )[0] return out def prepare_nvfp4_moe_layer_for_fi_or_cutlass( backend: "NvFp4MoeBackend", layer: "FusedMoE", w13: torch.Tensor, w13_scale: torch.Tensor, w13_scale_2: torch.Tensor, a13_scale: torch.Tensor, w2: torch.Tensor, w2_scale: torch.Tensor, w2_scale_2: torch.Tensor, a2_scale: torch.Tensor, is_act_and_mul: bool, ) -> tuple[ torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, ]: # Delayed import for circular dependency avoidance. from vllm.model_executor.layers.fused_moe.oracle.nvfp4 import ( NvFp4MoeBackend, is_global_sf_supported_for_nvfp4_backend, ) assert backend in [ NvFp4MoeBackend.VLLM_CUTLASS, NvFp4MoeBackend.FLASHINFER_CUTLASS, NvFp4MoeBackend.FLASHINFER_TRTLLM, NvFp4MoeBackend.FLASHINFER_CUTEDSL, ] # Reorder [w1, w3] to [w3, w1] for FI NVFP4 MoE kernels. is_gated = layer.activation.is_gated if ( is_gated and is_act_and_mul and backend in [ NvFp4MoeBackend.FLASHINFER_CUTLASS, NvFp4MoeBackend.FLASHINFER_TRTLLM, ] ): w13, w13_scale = reorder_w1w3_to_w3w1(w13, w13_scale) # For some FI kernels, the input scales are shared by all experts. if is_global_sf_supported_for_nvfp4_backend(backend): num_experts = w13.shape[0] a13_scale = a13_scale.max().to(torch.float32).expand(num_experts) a2_scale = a2_scale.max().to(torch.float32).expand(num_experts) else: a13_scale = a13_scale.max(dim=1).values.to(torch.float32) # Shuffle weights and scales for FI TRTLLM NVFP4 MoE kernels. if backend == NvFp4MoeBackend.FLASHINFER_TRTLLM: # Align weights for FI NVFP4 MoE kernels. min_alignment = 16 if is_gated else 128 w13, w13_scale, w2, w2_scale, padded_intermediate = ( align_fp4_moe_weights_for_fi( w13, w13_scale, w2, w2_scale, is_act_and_mul, min_alignment ) ) layer.intermediate_size_per_partition = padded_intermediate w13, w13_scale, w2, w2_scale = prepare_static_weights_for_trtllm_fp4_moe( w13, w2, w13_scale, w2_scale, hidden_size=w2.size(-2), intermediate_size=w13.size(-2) // 2 if is_gated else w13.size(-2), num_experts=w13.size(0), is_gated_activation=is_gated, ) # We do not need to make this a parameter, because # it is not used during the weight (re)-loading process. if is_gated: layer.g1_scale_c = a13_scale * w13_scale_2 / a2_scale else: layer.g1_scale_c = torch.ones_like(a13_scale) / a2_scale layer.a1_gscale = 1.0 / a13_scale layer.g1_alphas = a13_scale * w13_scale_2 layer.g2_alphas = a2_scale * w2_scale_2 else: # Swizzle the block scales for other FI NVFP4 MoE kernels. w13_scale = swizzle_blockscale(w13_scale) # Apply padding if needed. pad_size = w13_scale.size(1) - w13.size(1) if pad_size > 0: if is_act_and_mul: raise NotImplementedError( "Intermediate size padding for w1 and w3, for %s " "NvFp4 backend, but this is not currently supported", backend.value, ) w13 = torch.nn.functional.pad(w13, (0, 0, 0, pad_size)) w2 = torch.nn.functional.pad(w2, (0, pad_size // 2, 0, 0)) w2_scale = torch.nn.functional.pad(w2_scale, (0, pad_size // 16)) w2_scale = swizzle_blockscale(w2_scale) return w13, w13_scale, w13_scale_2, a13_scale, w2, w2_scale, w2_scale_2, a2_scale
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/utils/flashinfer_fp4_moe.py", "license": "Apache License 2.0", "lines": 497, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/attention/test_chunked_local_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass import numpy as np import pytest import torch from tests.v1.attention.utils import BatchSpec, create_common_attn_metadata from vllm.v1.attention.backends.utils import make_local_attention_virtual_batches @dataclass class LocalAttentionTestData: # Input parameters batch_spec: BatchSpec attn_chunk_size: int block_size: int # Expected return values expected_q_seqlens: list[int] expected_k_seqlens: list[int] expected_local_block_table: list[list[int]] test_data_list = [ # Same as example in docstring of make_local_attention_virtual_batches # except block table has 9 columns instead of 10 LocalAttentionTestData( batch_spec=BatchSpec( query_lens=[4, 10, 5], seq_lens=[6, 17, 9], ), attn_chunk_size=4, block_size=2, expected_q_seqlens=[2, 2, 1, 4, 4, 1, 4, 1], expected_k_seqlens=[4, 2, 4, 4, 4, 1, 4, 1], # 2 pages per local branch # (chunk size 4 // block size 2) expected_local_block_table=[ [0, 1], # local-batch 0, (batch 0, starting from k[0]) [2, 3], # local-batch 1, (batch 0, starting from k[4]) [11, 12], # local-batch 2, (batch 1, starting from k[4]) [13, 14], # local-batch 3, (batch 1, starting from k[8]) [15, 16], # local-batch 4, (batch 1, starting from k[12]) [17, 17], # local-batch 5, (batch 1, starting from k[16]) [20, 21], # local-batch 6, (batch 2, starting from k[4]) [22, 23], # local-batch 7, (batch 2, starting from k[8]) ], ), # Case where block indices are not clipped to block table ncols-1 # because tokens_in_last_block == attn_chunk_size LocalAttentionTestData( batch_spec=BatchSpec( query_lens=[8], seq_lens=[12], ), attn_chunk_size=4, block_size=2, expected_q_seqlens=[4, 4], expected_k_seqlens=[4, 4], expected_local_block_table=[ [2, 3], [4, 5], ], ), # Case where all kv_seq positions are involved in attn LocalAttentionTestData( batch_spec=BatchSpec( query_lens=[7], # 10 - 7 = 3 previously computed tokens seq_lens=[10], ), attn_chunk_size=4, block_size=2, expected_q_seqlens=[1, 4, 2], expected_k_seqlens=[4, 4, 2], expected_local_block_table=[ [0, 1], [2, 3], [4, 4], ], ), # Case where attn_chunk_size > kv_seq_len # so no extra mini virtual batches are created LocalAttentionTestData( batch_spec=BatchSpec( query_lens=[4], seq_lens=[6], ), # Larger than kv_seq_len attn_chunk_size=10, block_size=2, # No change to q_seqlens and k_seqlens expected_q_seqlens=[4], expected_k_seqlens=[6], # In this case, we only need a block-table like: # block_table = [ [0, 1, 2] ] # 1 batch, 3 pages # But we need to pad it to 5 pages per local batch # because currently the pages_per_local_batch # is calculated as (attn_chunk_size // block_size) expected_local_block_table=[ [0, 1, 2, 2, 2], ], ), # Block size equal to chunk size # Expect single page per batch in local batch table LocalAttentionTestData( batch_spec=BatchSpec( query_lens=[6, 6], seq_lens=[8, 8], ), attn_chunk_size=4, block_size=4, expected_q_seqlens=[2, 4, 2, 4], expected_k_seqlens=[4, 4, 4, 4], # Initial block table = [ # [0, 1], < batch 0 # [2, 3], < batch 1 # ] expected_local_block_table=[ [0], # local-batch 0, (batch 0, starting from k[0]) [1], # local-batch 1, (batch 0, starting from k[4]) [2], # local-batch 1, (batch 0, starting from k[0]) [3], # local-batch 1, (batch 0, starting from k[4]) ], ), # Case where query falls in the second attention chunk # k_toks > 0 1 2 3 4 # q_toks v _____________ # 0 | 1 # 1 | 1 1 # 2 | 1 1 1 # 3 | 1 1 1 1 # 4 | 1 # where tokens 0,1,2,3 have been pre-computed LocalAttentionTestData( batch_spec=BatchSpec( query_lens=[1], seq_lens=[5], ), attn_chunk_size=4, block_size=2, expected_q_seqlens=[1], expected_k_seqlens=[1], expected_local_block_table=[ [2, 2], ], ), ] @pytest.mark.parametrize("test_data", test_data_list) def test_local_attention_virtual_batches(test_data: LocalAttentionTestData): device = torch.device("cuda:0") batch_spec = test_data.batch_spec attn_chunk_size = test_data.attn_chunk_size block_size = test_data.block_size expected_q_seqlens = test_data.expected_q_seqlens expected_k_seqlens = test_data.expected_k_seqlens expected_local_block_table = test_data.expected_local_block_table # Create common attention metadata common_attn_metadata = create_common_attn_metadata( batch_spec, block_size, device, # Use torch.arange instead of torch.randint so we can assert on # block table tensor values. The block table will have shape # (num_batches, cdiv(max_seq_len, block_size)) and the values will be # arranged from 0 to cdiv(max_seq_len, block_size)-1 arange_block_indices=True, ) # Call the function result, _ = make_local_attention_virtual_batches( attn_chunk_size, common_attn_metadata, block_size ) # Convert to numpy for easier comparison actual_q_seqlens = np.diff(result.query_start_loc_cpu.numpy()) actual_k_seqlens = result.seq_lens_cpu.numpy() # Check that all query lengths are less than or equal to attn_chunk_size assert all(q_len <= attn_chunk_size for q_len in actual_q_seqlens) # Check that all key lengths are less than or equal to attn_chunk_size assert all(k_len <= attn_chunk_size for k_len in actual_k_seqlens) # Check that the total number of query tokens is preserved assert sum(actual_q_seqlens) == sum(batch_spec.query_lens) # Verify results np.testing.assert_array_equal(actual_q_seqlens, expected_q_seqlens) np.testing.assert_array_equal(actual_k_seqlens, expected_k_seqlens) expected_block_table_tensor = torch.tensor( expected_local_block_table, dtype=torch.int32, device=device ) print(f"Expected block table:\n{expected_block_table_tensor}") print(f"Actual block table:\n{result.block_table_tensor}") torch.testing.assert_close(result.block_table_tensor, expected_block_table_tensor)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/attention/test_chunked_local_attention.py", "license": "Apache License 2.0", "lines": 185, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/models/step3_text.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Inference-only Jurassic model.""" from collections.abc import Iterable from itertools import islice from typing import Any import torch from torch import nn from vllm.compilation.decorators import support_torch_compile from vllm.config import CacheConfig, ModelConfig, VllmConfig from vllm.distributed import ( get_pp_group, get_tensor_model_parallel_world_size, tensor_model_parallel_all_reduce, ) from vllm.logger import init_logger from vllm.model_executor.layers.activation import SiluAndMul from vllm.model_executor.layers.attention import Attention from vllm.model_executor.layers.fused_moe import FusedMoE from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( ColumnParallelLinear, MergedColumnParallelLinear, ReplicatedLinear, RowParallelLinear, ) from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import default_weight_loader from vllm.sequence import IntermediateTensors from vllm.transformers_utils.configs.step3_vl import Step3TextConfig from .interfaces import SupportsPP from .utils import ( PPMissingLayer, is_pp_missing_parameter, make_empty_intermediate_tensors_factory, make_layers, maybe_prefix, ) logger = init_logger(__name__) class FusedMoEBlock(nn.Module): def __init__( self, config: ModelConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.tp_size = get_tensor_model_parallel_world_size() if self.tp_size > config.moe_num_experts: raise ValueError( f"Tensor parallel size {self.tp_size} is greater than " f"the number of experts {config.moe_num_experts}." ) self.experts = FusedMoE( num_experts=config.moe_num_experts, top_k=config.moe_top_k, hidden_size=config.hidden_size, intermediate_size=config.moe_intermediate_size, reduce_results=False, renormalize=config.norm_expert_weight, quant_config=quant_config, prefix=f"{prefix}.experts", ) self.gate = ReplicatedLinear( config.hidden_size, config.moe_num_experts, bias=False, quant_config=None, prefix=f"{prefix}.gate", ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: orig_shape = hidden_states.shape hidden_dim = hidden_states.shape[-1] hidden_states = hidden_states.view(-1, hidden_dim) router_logits, _ = self.gate(hidden_states) final_hidden_states = self.experts( hidden_states=hidden_states, router_logits=router_logits ) if self.tp_size > 1: final_hidden_states = tensor_model_parallel_all_reduce(final_hidden_states) return final_hidden_states.view(orig_shape) class Step3TextMLP(nn.Module): def __init__( self, hidden_size: int, intermediate_size: int, hidden_act: str, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.gate_up_proj = MergedColumnParallelLinear( hidden_size, [intermediate_size] * 2, bias=False, quant_config=quant_config, prefix=f"{prefix}.gate_up_proj", ) self.down_proj = RowParallelLinear( intermediate_size, hidden_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.down_proj", ) if hidden_act != "silu": raise ValueError( f"Unsupported activation: {hidden_act}. Only silu is supported for now." ) self.act_fn = SiluAndMul() self.hidden_size = hidden_size def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: gate_up, _ = self.gate_up_proj(hidden_states) intermediate_act = self.act_fn(gate_up) output, _ = self.down_proj(intermediate_act) return output class Step3TextAttention(nn.Module): def __init__( self, hidden_size: int, num_heads: int, num_kv_heads: int, norm_eps: float, rope_parameters: dict[str, Any], share_q_dim: int | None = None, max_position_embedding: int = 8192, head_dim: int = 256, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.hidden_size = hidden_size tp_size = get_tensor_model_parallel_world_size() self.total_num_heads = num_heads assert self.total_num_heads % tp_size == 0 self.num_heads = self.total_num_heads // tp_size if num_kv_heads != 1: raise ValueError( f"Step3TextAttention num_kv_heads must be 1, but got {num_kv_heads}." ) self.num_kv_heads = num_kv_heads self.head_dim = head_dim self.kv_size = self.num_kv_heads * self.head_dim self.q_size = share_q_dim if share_q_dim else self.head_dim self.qkv_proj = ReplicatedLinear( hidden_size, self.q_size + self.kv_size * 2, bias=False, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", ) self.o_proj = RowParallelLinear( self.total_num_heads * self.head_dim, hidden_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.o_proj", ) self.inter_norm = RMSNorm(self.q_size, eps=norm_eps) self.wq = ColumnParallelLinear( self.q_size, self.head_dim * self.total_num_heads, bias=False, quant_config=quant_config, prefix=f"{prefix}.wq", ) self.rotary_emb = get_rope( self.head_dim, max_position=max_position_embedding, rope_parameters=rope_parameters, ) scaling = self.head_dim**-0.5 self.attn = Attention( self.num_heads, self.head_dim, scaling, self.num_kv_heads, cache_config=cache_config, prefix=f"{prefix}.attn", ) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor ) -> torch.Tensor: qkv, _ = self.qkv_proj(hidden_states) q, k, v = qkv.split([self.q_size, self.kv_size, self.kv_size], dim=-1) q = self.inter_norm(q) q = self.wq(q)[0] q, k = self.rotary_emb(positions, q, k) attn_output = self.attn(q, k, v) residual, _ = self.o_proj(attn_output) return residual class Step3TextDecoderLayer(nn.Module): def __init__( self, config: Step3TextConfig, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.hidden_size = config.hidden_size self.self_attn = Step3TextAttention( hidden_size=self.hidden_size, num_heads=config.num_attention_heads, num_kv_heads=1, cache_config=cache_config, quant_config=quant_config, norm_eps=config.rms_norm_eps, max_position_embedding=config.max_position_embedding, head_dim=config.head_dim, share_q_dim=config.share_q_dim, rope_parameters=config.rope_parameters, prefix=f"{prefix}.self_attn", ) layer_idx = int(prefix.split("layers.")[1].split(".")[0]) moe_layers_enum = getattr(config, "moe_layers_enum", None) if moe_layers_enum is not None: moe_layers_idx = [int(i) for i in moe_layers_enum.strip().split(",")] else: # Default to 1dense. moe_layers_idx = [i for i in range(1, config.num_hidden_layers)] if layer_idx in moe_layers_idx: self.moe = FusedMoEBlock( config=config, quant_config=quant_config, prefix=f"{prefix}.moe" ) self.share_expert = Step3TextMLP( hidden_size=self.hidden_size, intermediate_size=config.share_expert_dim, hidden_act="silu", quant_config=quant_config, prefix=f"{prefix}.share_expert", ) self.use_moe = True else: self.mlp = Step3TextMLP( hidden_size=config.hidden_size, intermediate_size=config.intermediate_size, hidden_act="silu", quant_config=quant_config, prefix=f"{prefix}.mlp", ) self.use_moe = False self.input_layernorm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.post_attention_layernorm = RMSNorm( config.hidden_size, eps=config.rms_norm_eps ) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, residual: torch.Tensor | None, ) -> tuple[torch.Tensor, torch.Tensor]: if residual is None: residual = hidden_states hidden_states = self.input_layernorm(hidden_states) else: hidden_states, residual = self.input_layernorm(hidden_states, residual) hidden_states = self.self_attn( positions=positions, hidden_states=hidden_states, ) hidden_states, residual = self.post_attention_layernorm(hidden_states, residual) if self.use_moe: share_output = self.share_expert(hidden_states) moe_output = self.moe(hidden_states) hidden_states = share_output + moe_output else: hidden_states = self.mlp(hidden_states) return hidden_states, residual @support_torch_compile class Step3TextModel(nn.Module): def __init__(self, vllm_config: VllmConfig, prefix: str = "") -> None: super().__init__() config = vllm_config.model_config.hf_config cache_config = vllm_config.cache_config quant_config = vllm_config.quant_config self.vocab_size = config.vocab_size self.config = config if get_pp_group().is_first_rank or ( config.tie_word_embeddings and get_pp_group().is_last_rank ): self.embed_tokens = VocabParallelEmbedding( self.vocab_size, config.hidden_size, ) else: self.embed_tokens = PPMissingLayer() self.start_layer, self.end_layer, self.layers = make_layers( config.num_hidden_layers, lambda prefix: Step3TextDecoderLayer( config=config, cache_config=cache_config, quant_config=quant_config, prefix=prefix, ), prefix=f"{prefix}.layers", ) if get_pp_group().is_last_rank: self.norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) else: self.norm = PPMissingLayer() self.make_empty_intermediate_tensors = make_empty_intermediate_tensors_factory( ["hidden_states"], config.hidden_size ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor: if get_pp_group().is_first_rank: if inputs_embeds is not None: hidden_states = inputs_embeds else: hidden_states = self.embed_input_ids(input_ids) residual = None else: assert intermediate_tensors is not None hidden_states = intermediate_tensors["hidden_states"] residual = intermediate_tensors["residual"] for layer in islice(self.layers, self.start_layer, self.end_layer): hidden_states, residual = layer(positions, hidden_states, residual) if not get_pp_group().is_last_rank: return IntermediateTensors( { "hidden_states": hidden_states, "residual": residual, } ) hidden_states, _ = self.norm(hidden_states, residual) return hidden_states class Step3TextForCausalLM(nn.Module, SupportsPP): def __init__( self, *, vllm_config: VllmConfig, prefix: str = "", ): super().__init__() config = vllm_config.model_config.hf_config self.config = config self.vllm_config = vllm_config self.model = Step3TextModel(vllm_config=vllm_config, prefix=prefix) if get_pp_group().is_last_rank: self.lm_head = ParallelLMHead( config.vocab_size, config.hidden_size, prefix=maybe_prefix(prefix, "lm_head"), ) self.logits_processor = LogitsProcessor(config.vocab_size) else: self.lm_head = PPMissingLayer() self.make_empty_intermediate_tensors = ( self.model.make_empty_intermediate_tensors ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ): hidden_states = self.model( input_ids, positions, intermediate_tensors, inputs_embeds ) return hidden_states def compute_logits(self, hidden_states: torch.Tensor) -> torch.Tensor: logits = self.logits_processor(self.lm_head, hidden_states) return logits def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: qkv_params_mapping = [ # (param_name, shard_name, relative_start_idx, relative_end_idx) ( ".qkv_proj", ".q_proj", 0, self.config.share_q_dim / (self.config.share_q_dim + self.config.head_dim * 2), ), ( ".qkv_proj", ".k_proj", self.config.share_q_dim / (self.config.share_q_dim + self.config.head_dim * 2), (self.config.share_q_dim + self.config.head_dim) / (self.config.share_q_dim + self.config.head_dim * 2), ), ( ".qkv_proj", ".v_proj", (self.config.share_q_dim + self.config.head_dim) / (self.config.share_q_dim + self.config.head_dim * 2), (self.config.share_q_dim + self.config.head_dim * 2) / (self.config.share_q_dim + self.config.head_dim * 2), ), ] stacked_params_mapping = [ # (param_name, shard_name, shard_id) (".gate_up_proj", ".gate_proj", 0), (".gate_up_proj", ".up_proj", 1), ] params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() expert_params_mapping = [ (".moe.experts.w13_weight", ".moe.gate_proj.weight", "w1"), (".moe.experts.w13_weight", ".moe.up_proj.weight", "w3"), (".moe.experts.w2_weight", ".moe.down_proj.weight", "w2"), ] disable_moe_stacked_params = [data[1] for data in expert_params_mapping] for name, loaded_weight in weights: for param_name, weight_name, shard_id in stacked_params_mapping: if weight_name not in name: continue if any( disable_moe_stacked_param in name for disable_moe_stacked_param in disable_moe_stacked_params ): continue name = name.replace(weight_name, param_name) if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) loaded_params.add(name) break else: for mapping in expert_params_mapping: param_name, weight_name, shard_id = mapping if weight_name not in name: continue name = name.replace(weight_name, param_name) # Skip layers on other devices. if is_pp_missing_parameter(name, self): continue # Skip loading extra bias for GPTQ models. if ( name.endswith(".bias") or name.endswith("_bias") ) and name not in params_dict: continue param = params_dict[name] weight_loader = param.weight_loader for expert_id in range(loaded_weight.shape[0]): loaded_weight_expert = loaded_weight[expert_id] weight_loader( param, loaded_weight_expert, name, shard_id=shard_id, expert_id=expert_id, ) loaded_params.add(name) break else: for ( param_name, weight_name, start_idx, end_idx, ) in qkv_params_mapping: if weight_name not in name: continue name = name.replace(weight_name, param_name) if is_pp_missing_parameter(name, self): continue param = params_dict[name] dim = param.shape[param.output_dim] begin_idx = int(start_idx * dim) end_idx = int(end_idx * dim) param_slice = param.narrow( param.output_dim, begin_idx, end_idx - begin_idx ) param_slice.copy_(loaded_weight) loaded_params.add(name) break else: if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = getattr( param, "weight_loader", default_weight_loader ) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/step3_text.py", "license": "Apache License 2.0", "lines": 497, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/step3_vl.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math from collections.abc import Iterable, Mapping, Sequence from itertools import product from math import ceil, sqrt from typing import Annotated, Any, Literal, TypeAlias import numpy as np import torch import torch.nn as nn import torch.nn.functional as F from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import BatchFeature, PretrainedConfig, TensorType from vllm.config import VllmConfig from vllm.config.multimodal import BaseDummyOptions from vllm.distributed import get_tensor_model_parallel_world_size from vllm.model_executor.layers.activation import get_act_fn from vllm.model_executor.layers.attention import MMEncoderAttention from vllm.model_executor.layers.conv import Conv2dLayer from vllm.model_executor.layers.linear import ( ColumnParallelLinear, QKVParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import ( MultiModalDataDict, MultiModalFieldConfig, MultiModalKwargsItems, ) from vllm.multimodal.parse import ImageSize, MultiModalDataItems from vllm.multimodal.processing import ( BaseDummyInputsBuilder, BaseMultiModalProcessor, BaseProcessingInfo, PromptReplacement, PromptUpdate, PromptUpdateDetails, ) from vllm.sequence import IntermediateTensors from vllm.tokenizers import TokenizerLike from vllm.transformers_utils.configs import Step3VisionEncoderConfig from vllm.utils.tensor_schema import TensorSchema, TensorShape from .interfaces import MultiModalEmbeddings, SupportsMultiModal, SupportsPP from .utils import ( AutoWeightsLoader, WeightsMapper, init_vllm_registered_model, maybe_prefix, ) from .vision import is_vit_use_data_parallel, run_dp_sharded_vision_model class Step3VLImagePixelInputs(TensorSchema): """ Dimensions: - bn: Batch size * number of images - c: Number of channels (3) - h: Height - w: Width - bnp: Batch size * number of images * number of patches - hp: Height of patch - wp: Width of patch """ type: Literal["pixel_values"] pixel_values: Annotated[torch.Tensor, TensorShape("bn", 3, "h", "w")] patch_pixel_values: Annotated[torch.Tensor, TensorShape("bnp", 3, "hp", "wp")] num_patches: Annotated[torch.Tensor, TensorShape("bn")] class Step3VLImageEmbeddingInputs(TensorSchema): """ Dimensions: - bn: Batch size * number of images - f: Image feature size - h: Hidden size (must match the hidden size of language model backbone) """ type: Literal["image_embeds"] = "image_embeds" data: Annotated[torch.Tensor, TensorShape("bn", "f", "h")] Step3VLImageInputs: TypeAlias = Step3VLImagePixelInputs | Step3VLImageEmbeddingInputs ImageWithPatches = tuple[Image.Image, list[Image.Image], list[bool] | None] MAX_IMAGE_SIZE: int = 3024 class Step3VisionProcessor: def __init__(self, size, interpolation_mode="bicubic", patch_size=None): mean = [0.48145466, 0.4578275, 0.40821073] std = [0.26862954, 0.26130258, 0.27577711] patch_size = patch_size if patch_size is not None else size self.transform = transforms.Compose( [ transforms.ToTensor(), transforms.Normalize(mean, std), transforms.Resize( (size, size), interpolation=InterpolationMode.BICUBIC if interpolation_mode == "bicubic" else InterpolationMode.BILINEAR, antialias=True, ), ] ) self.patch_transform = ( transforms.Compose( [ transforms.ToTensor(), transforms.Normalize(mean, std), transforms.Resize( (patch_size, patch_size), interpolation=InterpolationMode.BICUBIC if interpolation_mode == "bicubic" else InterpolationMode.BILINEAR, antialias=True, ), ] ) if patch_size is not None else None ) def __call__(self, image, is_patch=False): if is_patch: return {"pixel_values": self.patch_transform(image).unsqueeze(0)} else: return {"pixel_values": self.transform(image).unsqueeze(0)} class ImagePatcher: def __init__(self, enable_patch: bool = True) -> None: self.enable_patch = enable_patch def determine_window_size(self, long: int, short: int) -> int: if long < 728: return short if long / short > 1.5 else 0 return min(short, 504) if long / short > 4 else 504 def slide_window( self, width: int, height: int, sizes: list[tuple[int, int]], steps: list[tuple[int, int]], img_rate_thr: float = 0.6, ) -> tuple[list[tuple[int, int, int, int]], tuple[int, int]]: assert 1 >= img_rate_thr >= 0, "The `in_rate_thr` should lie in 0~1" windows = [] # Sliding windows. for size, step in zip(sizes, steps): size_w, size_h = size step_w, step_h = step x_num = 1 if width <= size_w else ceil((width - size_w) / step_w + 1) x_start = [step_w * i for i in range(x_num)] if len(x_start) > 1 and x_start[-1] + size_w > width: x_start[-1] = width - size_w y_num = 1 if height <= size_h else ceil((height - size_h) / step_h + 1) y_start = [step_h * i for i in range(y_num)] if len(y_start) > 1 and y_start[-1] + size_h > height: y_start[-1] = height - size_h start = np.array(list(product(y_start, x_start)), dtype=int) start[:, [0, 1]] = start[:, [1, 0]] windows.append(np.concatenate([start, start + size], axis=1)) windows = np.concatenate(windows, axis=0) return [ (int(box[0]), int(box[1]), int(box[2] - box[0]), int(box[3] - box[1])) for box in windows ], (x_num, y_num) def square_pad(self, img: Image.Image) -> Image.Image: w, h = img.size if w == h: return img size = max(w, h) padded = Image.new(img.mode, (size, size), 0) padded.paste(img, (0, 0)) return padded def get_image_size_for_padding( self, img_width: int, img_height: int ) -> tuple[int, int]: ratio = img_width / img_height if min(img_height, img_width) < 32 and (ratio > 4 or ratio < 1 / 4): new_size = max(img_height, img_width) return new_size, new_size return img_width, img_height def get_image_size_for_preprocess( self, img_width: int, img_height: int ) -> tuple[int, int]: if max(img_height, img_width) > MAX_IMAGE_SIZE: scale_factor = MAX_IMAGE_SIZE / max(img_height, img_width) img_width = int(img_width * scale_factor) img_height = int(img_height * scale_factor) return img_width, img_height def get_image_size_for_crop( self, img_width: int, img_height: int, window_size: int ): w_ratio = img_width / window_size h_ratio = img_height / window_size if w_ratio < 1: width_new = img_width else: decimal_w = w_ratio - img_width // window_size w_ratio = int(w_ratio) + 1 if decimal_w > 0.2 else int(w_ratio) width_new = window_size * w_ratio if h_ratio < 1: height_new = img_height else: decimal_h = h_ratio - img_height // window_size h_ratio = int(h_ratio) + 1 if decimal_h > 0.2 else int(h_ratio) height_new = window_size * h_ratio return int(width_new), int(height_new) def patch_crop(self, img: Image.Image, i: int, j: int, th: int, tw: int): target = img.crop((j, i, j + tw, i + th)) return target def get_num_patches(self, img_width: int, img_height: int) -> tuple[int, int]: img_width, img_height = self.get_image_size_for_padding(img_width, img_height) img_width, img_height = self.get_image_size_for_preprocess( img_width, img_height ) window_size = self.determine_window_size( max(img_height, img_width), min(img_height, img_width) ) if window_size == 0 or not self.enable_patch: return 0, 0 else: img_width, img_height = self.get_image_size_for_crop( img_width, img_height, window_size ) center_list, (x_num, y_num) = self.slide_window( img_width, img_height, [(window_size, window_size)], [(window_size, window_size)], ) full_rows = (len(center_list) - 1) // x_num + 1 if len(center_list) > 0 and len(center_list) % x_num == 0: full_rows -= 1 return len(center_list), full_rows def __call__( self, img: Image.Image ) -> tuple[Image.Image, list[Image.Image], list[bool] | None]: img_width, img_height = img.size new_img_width, new_img_height = self.get_image_size_for_padding( img_width, img_height ) if new_img_width != img_width or new_img_height != img_height: img = self.square_pad(img) img_width, img_height = img.size new_img_width, new_img_height = self.get_image_size_for_preprocess( img_width, img_height ) img = img.resize((new_img_width, new_img_height), Image.Resampling.BILINEAR) window_size = self.determine_window_size( max(new_img_height, new_img_width), min(new_img_height, new_img_width) ) if window_size == 0 or not self.enable_patch: return img, [], None else: new_img_width, new_img_height = self.get_image_size_for_crop( new_img_width, new_img_height, window_size ) if (new_img_width, new_img_height) != (img_width, img_height): img_for_crop = img.resize( (new_img_width, new_img_height), Image.Resampling.BILINEAR ) else: img_for_crop = img patches = [] newlines = [] center_list, (x_num, y_num) = self.slide_window( new_img_width, new_img_height, [(window_size, window_size)], [(window_size, window_size)], ) for patch_id, center_lf_point in enumerate(center_list): x, y, patch_w, patch_h = center_lf_point big_patch = self.patch_crop(img_for_crop, y, x, patch_h, patch_w) patches.append(big_patch) if (patch_id + 1) % x_num == 0: newlines.append(patch_id) if newlines and newlines[-1] == len(patches) - 1: newlines.pop() return ( img, patches, [i in newlines for i in range(len(patches))] if len(patches) > 0 else None, ) class Step3VLProcessor: def __init__( self, config: PretrainedConfig, tokenizer: TokenizerLike, ) -> None: super().__init__() self.config = config self.tokenizer = tokenizer self.image_size = 728 self.patch_size = 504 self.image_preprocessor = Step3VisionProcessor( self.image_size, "bilinear", self.patch_size ) self.num_image_feature_size = 169 self.num_patch_feature_size = 81 self.image_token = "<im_patch>" self.image_feature_placeholder = self.image_token * self.num_image_feature_size self.patch_feature_placeholder = self.image_token * self.num_patch_feature_size # Respect vision config switch to enable/disable patch extraction. # For video understanding, it's preferable to disable patch. enable_patch = getattr(self.config.vision_config, "enable_patch", True) self.patcher = ImagePatcher(enable_patch=enable_patch) @property def image_token_id(self) -> int: return self.tokenizer.get_vocab()[self.image_token] def get_num_image_tokens(self, img_width: int, img_height: int) -> int: num_patches, num_newlines = self.patcher.get_num_patches(img_width, img_height) return ( num_patches * (self.num_patch_feature_size + 2) + self.num_image_feature_size + 2 + num_newlines ) def _split_images(self, images: list[Image.Image]) -> list[ImageWithPatches]: result = [] for img in images: result.append(self.patcher(img)) return result def _convert_images_to_pixel_values( self, images: list[Image.Image], is_patch: bool = False, ) -> list[torch.Tensor]: return [ self.image_preprocessor(img, is_patch=is_patch)["pixel_values"] for img in images ] def _get_patch_repl( self, num_patches: int, patch_newline_mask: list[bool] | None, ) -> tuple[str, list[int]]: text = "" token_ids = [] for i in range(num_patches): assert len(patch_newline_mask) == num_patches text += f"<patch_start>{self.patch_feature_placeholder}<patch_end>" token_ids.extend( [self.tokenizer.convert_tokens_to_ids("<patch_start>")] + [self.image_token_id] * self.num_patch_feature_size + [self.tokenizer.convert_tokens_to_ids("<patch_end>")] ) if patch_newline_mask and patch_newline_mask[i]: text += "<patch_newline>" token_ids.append( self.tokenizer.convert_tokens_to_ids("<patch_newline>") ) return text, token_ids def _get_image_repl( self, num_images: int, ) -> tuple[str, list[int]]: text = f"<im_start>{self.image_feature_placeholder}<im_end>" token_ids = ( [self.tokenizer.convert_tokens_to_ids("<im_start>")] + [self.image_token_id] * self.num_image_feature_size + [self.tokenizer.convert_tokens_to_ids("<im_end>")] ) return text * num_images, token_ids * num_images def _get_image_repl_features( self, num_images: int, num_patches: int, patch_new_line_idx: list[bool] | None, ) -> tuple[str, list[int]]: if num_patches > 0: patch_repl, patch_repl_ids = self._get_patch_repl( num_patches, patch_new_line_idx ) else: patch_repl = "" patch_repl_ids = [] image_repl, image_repl_ids = self._get_image_repl(num_images) return patch_repl + image_repl, patch_repl_ids + image_repl_ids def replace_placeholder(self, text: str, placeholder: str, repls: list[str]) -> str: parts = text.split(placeholder) if len(parts) - 1 != len(repls): raise ValueError( "The number of placeholders does not match the number of replacements." ) result = [parts[0]] for i, repl in enumerate(repls): result.append(repl) result.append(parts[i + 1]) return "".join(result) def __call__( self, text: str | list[str] | None = None, images: Image.Image | list[Image.Image] | None = None, return_tensors: str | TensorType | None = None, ) -> BatchFeature: if text is None: text = [] if not isinstance(text, list): text = [text] if images is None: images = [] if not isinstance(images, list): images = [images] if len(images) == 0: image_inputs = {} text_inputs = self.tokenizer(text) else: splitted_images_data = self._split_images(images) pixel_values_lst = [] patch_pixel_values_lst = [] patch_newline_mask_lst = [] image_repl_str_lst = [] image_repl_ids_lst = [] num_patches = [] for raw_img, img_patches, patch_newline_mask in splitted_images_data: pixel_values_lst.extend(self._convert_images_to_pixel_values([raw_img])) if len(img_patches) > 0: patch_pixel_values_lst.extend( self._convert_images_to_pixel_values(img_patches, is_patch=True) ) num_patches.append(len(img_patches)) image_repl_str, image_repl_ids = self._get_image_repl_features( 1, len(img_patches), patch_newline_mask ) image_repl_str_lst.append(image_repl_str) image_repl_ids_lst.extend(image_repl_ids) if patch_newline_mask is not None: patch_newline_mask_lst.extend(patch_newline_mask) pixel_values = torch.cat(pixel_values_lst) patch_size = self.patch_size image_inputs = { "pixel_values": pixel_values, "num_patches": num_patches, "patch_pixel_values": ( torch.cat(patch_pixel_values_lst) if patch_pixel_values_lst else pixel_values.new_empty((0, 3, patch_size, patch_size)) ), "patch_newline_mask": torch.tensor( patch_newline_mask_lst, dtype=torch.bool ), } text = [ self.replace_placeholder(t, self.image_token, image_repl_str_lst) for t in text ] text_inputs = self.tokenizer(text) return BatchFeature( { **text_inputs, **image_inputs, }, tensor_type=return_tensors, ) class Step3VLProcessingInfo(BaseProcessingInfo): def get_hf_processor(self) -> Step3VLProcessor: return Step3VLProcessor( self.get_hf_config(), self.get_tokenizer(), ) def get_supported_mm_limits(self) -> Mapping[str, int | None]: return {"image": None} def get_max_image_tokens(self) -> int: hf_processor = self.get_hf_processor() return hf_processor.get_num_image_tokens( self.get_image_size_with_most_features().width, self.get_image_size_with_most_features().height, ) def get_mm_max_tokens_per_item( self, seq_len: int, mm_counts: Mapping[str, int], ) -> Mapping[str, int]: return {"image": self.get_max_image_tokens()} def get_image_size_with_most_features(self) -> ImageSize: return ImageSize(3024, 3024) def get_num_mm_tokens(self, mm_data: MultiModalDataDict) -> int: if len(mm_data) != 1 or "image" not in mm_data: raise ValueError("mm_data could only contain one key 'image' for steo1o") image_data = mm_data["image"] if not isinstance(image_data, (list, tuple)): image_data = [image_data] return sum( self.get_hf_processor().get_num_image_tokens(img.width, img.height) for img in image_data ) class Step3VLDummyInputsBuilder(BaseDummyInputsBuilder[Step3VLProcessingInfo]): def get_dummy_text(self, mm_counts: Mapping[str, int]) -> str: num_images = mm_counts.get("image", 0) return "<im_patch>" * num_images def get_dummy_mm_data( self, seq_len: int, mm_counts: Mapping[str, int], mm_options: Mapping[str, BaseDummyOptions], ) -> MultiModalDataDict: target_width, target_height = self.info.get_image_size_with_most_features() num_images = mm_counts.get("image", 0) image_overrides = mm_options.get("image") return { "image": self._get_dummy_images( width=target_width, height=target_height, num_images=num_images, overrides=image_overrides, ) } class Step3VLMultiModalProcessor(BaseMultiModalProcessor[Step3VLProcessingInfo]): def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, Any], out_mm_kwargs: MultiModalKwargsItems, ) -> Sequence[PromptUpdate]: hf_processor = self.info.get_hf_processor(**hf_processor_mm_kwargs) image_placeholder_token_id = hf_processor.image_token_id def get_replacement_step1o(item_idx: int): out_item = out_mm_kwargs["image"][item_idx] num_patches = int(out_item["num_patches"].data) if num_patches > 0: patch_newline_mask = out_item["patch_newline_mask"].data image_repl_ids = hf_processor._get_image_repl_features( 1, num_patches, patch_newline_mask.tolist() )[1] else: image_repl_ids = hf_processor._get_image_repl_features(1, 0, None)[1] return PromptUpdateDetails.select_token_id( seq=image_repl_ids, embed_token_id=image_placeholder_token_id, ) return [ PromptReplacement( modality="image", target=[image_placeholder_token_id], replacement=get_replacement_step1o, ) ] def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: num_patches = hf_inputs.get("num_patches", torch.empty(0)) return dict( pixel_values=MultiModalFieldConfig.batched("image"), patch_pixel_values=MultiModalFieldConfig.flat_from_sizes( "image", num_patches ), num_patches=MultiModalFieldConfig.batched("image"), patch_newline_mask=MultiModalFieldConfig.flat_from_sizes( "image", num_patches ), ) def get_abs_pos(abs_pos, tgt_size): dim = abs_pos.size(-1) abs_pos_new = abs_pos.squeeze(0) cls_token, old_pos_embed = abs_pos_new[:1], abs_pos_new[1:] src_size = int(math.sqrt(abs_pos_new.shape[0] - 1)) tgt_size = int(math.sqrt(tgt_size)) dtype = abs_pos.dtype if src_size != tgt_size: old_pos_embed = ( old_pos_embed.view(1, src_size, src_size, dim) .permute(0, 3, 1, 2) .contiguous() ) old_pos_embed = old_pos_embed.to(torch.float32) new_pos_embed = F.interpolate( old_pos_embed, size=(tgt_size, tgt_size), mode="bicubic", antialias=True, align_corners=False, ).to(dtype) new_pos_embed = new_pos_embed.permute(0, 2, 3, 1) new_pos_embed = new_pos_embed.view(tgt_size * tgt_size, dim) vision_pos_embed = torch.cat([cls_token, new_pos_embed], dim=0) vision_pos_embed = vision_pos_embed.view(1, tgt_size * tgt_size + 1, dim) return vision_pos_embed else: return abs_pos class Step3VisionEmbeddings(nn.Module): def __init__(self, config: Step3VisionEncoderConfig): super().__init__() self.config = config self.embed_dim = config.hidden_size self.image_size = config.image_size self.patch_size = config.patch_size self.class_embedding = nn.Parameter(torch.randn(1, self.embed_dim)) self.patch_embedding = Conv2dLayer( in_channels=config.num_channels, out_channels=self.embed_dim, kernel_size=self.patch_size, stride=self.patch_size, bias=True, ) self.num_patches = (self.image_size // self.patch_size) ** 2 self.pad_tp_size = 4 # hard code for padding # To load the pretrained weights, we still use P+1 as the seqlen self.position_embedding = torch.nn.Embedding( self.num_patches + 1, self.embed_dim ) self.register_buffer( "position_ids", torch.arange(self.num_patches + 1).expand((1, -1)), persistent=False, ) def forward(self, pixel_values: torch.Tensor) -> torch.Tensor: batch_size = pixel_values.shape[0] patch_embeds = self.patch_embedding( pixel_values ) # shape = [*, width, grid, grid] patch_embeds = patch_embeds.flatten(2).transpose(1, 2) # pad class_embeds = self.class_embedding.expand(batch_size, 1, -1) embeddings = torch.cat([class_embeds, patch_embeds], dim=1) embeddings = embeddings + get_abs_pos( self.position_embedding(self.position_ids), patch_embeds.size(1) ) embeddings = torch.cat( [ embeddings[:, 0, :].unsqueeze(1).repeat(1, self.pad_tp_size - 1, 1), embeddings, ], dim=1, ) return embeddings class Step3VisionAttention(nn.Module): """Multi-headed attention from 'Attention Is All You Need' paper""" def __init__( self, config, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.config = config self.embed_dim = config.hidden_size self.total_num_heads = config.num_attention_heads self.head_dim = self.embed_dim // self.total_num_heads self.scale = self.head_dim**-0.5 use_data_parallel = is_vit_use_data_parallel() tp_size = 1 if use_data_parallel else get_tensor_model_parallel_world_size() assert self.total_num_heads % tp_size == 0 self.num_heads = self.total_num_heads // tp_size self.q_size = self.num_heads * self.head_dim self.qkv_proj = QKVParallelLinear( self.embed_dim, self.head_dim, self.total_num_heads, bias=True, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", disable_tp=use_data_parallel, ) self.out_proj = RowParallelLinear( self.embed_dim, self.embed_dim, bias=True, quant_config=quant_config, prefix=f"{prefix}.out_proj", disable_tp=use_data_parallel, ) # Use unified MMEncoderAttention with automatic backend selection self.attn = MMEncoderAttention( self.num_heads, self.head_dim, self.scale, prefix=f"{prefix}.attn", ) def forward( self, hidden_states: torch.Tensor, ): """Input shape: Batch x Time x Channel""" bsz, tgt_len, _ = hidden_states.size() # get query proj qkv, _ = self.qkv_proj(hidden_states) q, k, v = qkv.chunk(chunks=3, dim=-1) # Use unified MMEncoderAttention with automatic backend selection attn_output = self.attn(q, k, v) attn_output, _ = self.out_proj(attn_output) return attn_output class Step3VisionMLP(nn.Module): def __init__( self, config, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.config = config self.activation_fn = get_act_fn(config.hidden_act) use_data_parallel = is_vit_use_data_parallel() self.fc1 = ColumnParallelLinear( config.hidden_size, config.intermediate_size, bias=True, quant_config=quant_config, prefix=f"{prefix}.fc1", disable_tp=use_data_parallel, ) self.fc2 = RowParallelLinear( config.intermediate_size, config.hidden_size, bias=True, quant_config=quant_config, prefix=f"{prefix}.fc2", disable_tp=use_data_parallel, ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states, _ = self.fc1(hidden_states) hidden_states = self.activation_fn(hidden_states) hidden_states, _ = self.fc2(hidden_states) return hidden_states class Step3VisionEncoderLayer(nn.Module): def __init__( self, config: Step3VisionEncoderConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.embed_dim = config.hidden_size self.self_attn = Step3VisionAttention( config, quant_config, prefix=f"{prefix}.self_attn", ) self.layer_norm1 = nn.LayerNorm(self.embed_dim, eps=config.layer_norm_eps) self.mlp = Step3VisionMLP( config, quant_config, prefix=f"{prefix}.mlp", ) self.layer_norm2 = nn.LayerNorm(self.embed_dim, eps=config.layer_norm_eps) def forward( self, hidden_states: torch.Tensor, ) -> torch.FloatTensor: hidden_states = hidden_states + self.layer_norm1(self.self_attn(hidden_states)) hidden_states = hidden_states + self.layer_norm2(self.mlp(hidden_states)) return hidden_states class Step3VisionEncoder(nn.Module): def __init__( self, config: Step3VisionEncoderConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.config = config self.layers = nn.ModuleList( [ Step3VisionEncoderLayer( config, quant_config, prefix=f"{prefix}.layers.{i}", ) for i in range(config.num_hidden_layers) ] ) def forward( self, inputs_embeds, ): hidden_states = inputs_embeds for encoder_layer in self.layers: hidden_states = encoder_layer(hidden_states) return hidden_states class Step3VisionTransformer(nn.Module): def __init__( self, config: Step3VisionEncoderConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.config = config self.use_data_parallel = is_vit_use_data_parallel() self.image_size = config.image_size self.embeddings = Step3VisionEmbeddings(config) self.transformer = Step3VisionEncoder( config, quant_config, prefix=f"{prefix}.transformer", ) def forward( self, pixel_values: torch.Tensor, ): hidden_states = self.embeddings(pixel_values) if self.use_data_parallel: hidden_states = run_dp_sharded_vision_model(hidden_states, self.transformer) else: hidden_states = self.transformer(inputs_embeds=hidden_states) return hidden_states @MULTIMODAL_REGISTRY.register_processor( Step3VLMultiModalProcessor, info=Step3VLProcessingInfo, dummy_inputs=Step3VLDummyInputsBuilder, ) class Step3VLForConditionalGeneration(nn.Module, SupportsMultiModal, SupportsPP): hf_to_vllm_mapper = WeightsMapper( orig_to_new_prefix={ "model.": "language_model.model.", "lm_head.": "language_model.lm_head.", } ) supports_encoder_tp_data = True @classmethod def get_placeholder_str(cls, modality: str, i: int) -> str | None: if modality.startswith("image"): return "<im_patch>" raise ValueError("Only image modality is supported") def __init__(self, *, vllm_config: VllmConfig, prefix: str = "") -> None: super().__init__() config = vllm_config.model_config.hf_config multimodal_config = vllm_config.model_config.multimodal_config self.config = config self.multimodal_config = multimodal_config self.use_data_parallel = multimodal_config.mm_encoder_tp_mode == "data" with self._mark_tower_model(vllm_config, "image"): self.vision_model = Step3VisionTransformer( config.vision_config, None, prefix=maybe_prefix(prefix, "vision_model"), ) self.vit_downsampler = Conv2dLayer( config.vision_config.hidden_size, config.vision_config.output_hidden_size, kernel_size=2, stride=config.understand_projector_stride, ) self.vit_downsampler2 = Conv2dLayer( config.vision_config.output_hidden_size, config.vision_config.output_hidden_size * 2, kernel_size=3, stride=2, padding=1, ) self.vit_large_projector = nn.Linear( config.vision_config.output_hidden_size * 2, config.hidden_size, bias=config.projector_bias, ) with self._mark_language_model(vllm_config): self.language_model = init_vllm_registered_model( vllm_config=vllm_config, hf_config=config.text_config, prefix=maybe_prefix(prefix, "language_model"), ) self.make_empty_intermediate_tensors = ( self.language_model.make_empty_intermediate_tensors ) @property def device(self): return next(self.parameters()).device @property def dtype(self): return next(self.parameters()).dtype def _parse_and_validate_image_input( self, **kwargs: object ) -> Step3VLImageInputs | None: pixel_values = kwargs.pop("pixel_values", None) patch_pixel_values = kwargs.pop("patch_pixel_values", None) num_patches = kwargs.pop("num_patches", None) image_embeds = kwargs.pop("image_embeds", None) if pixel_values is None and image_embeds is None: return None if pixel_values is not None and patch_pixel_values is not None: return Step3VLImagePixelInputs( type="pixel_values", pixel_values=pixel_values.to(self.dtype), patch_pixel_values=patch_pixel_values.to(self.dtype), num_patches=num_patches, ) if image_embeds is not None: return Step3VLImageEmbeddingInputs( type="image_embeds", image_embeds=image_embeds.to(self.dtype), ) raise AssertionError("This line should be unreachable.") def _process_image_features(self, image_features: torch.Tensor) -> torch.Tensor: B, P = image_features.shape[:2] HW = int(sqrt(P)) image_features = image_features.permute(0, 2, 1).view(B, -1, HW, HW) image_features = self.vit_downsampler(image_features) image_features = self.vit_downsampler2(image_features) n_dim = image_features.size(1) image_features = image_features.view(B, n_dim, -1).permute(0, 2, 1) image_features = self.vit_large_projector(image_features) return image_features def _get_vision_model_output(self, input_tensor: torch.Tensor) -> torch.Tensor: return self.vision_model(input_tensor)[:, 4:] def _process_image_input( self, image_input: Step3VLImageInputs ) -> tuple[torch.Tensor, ...]: if image_input["type"] == "image_embeds": image_features = image_input["image_embeds"] else: image_features = self._get_vision_model_output(image_input["pixel_values"]) patch_image_features = ( self._get_vision_model_output(image_input["patch_pixel_values"]) if len(image_input["patch_pixel_values"]) > 0 else None ) num_patches = image_input["num_patches"] image_features = self._process_image_features(image_features) patch_image_features = ( self._process_image_features(patch_image_features) if patch_image_features is not None else None ) merged_image_features = [] cur_patch_idx = 0 for i, num_patch in enumerate(num_patches): cur_feature = [] if num_patch > 0: patch_slice = patch_image_features[ cur_patch_idx : cur_patch_idx + num_patch ] cur_feature.append(patch_slice.view(-1, patch_slice.shape[-1])) cur_feature.append(image_features[i].view(-1, image_features.shape[-1])) cur_patch_idx += num_patch merged_image_features.append( torch.cat(cur_feature) if len(cur_feature) > 1 else cur_feature[0] ) return merged_image_features def embed_multimodal(self, **kwargs) -> MultiModalEmbeddings: image_input = self._parse_and_validate_image_input(**kwargs) if image_input is None: return [] vision_embeddings = self._process_image_input(image_input) return vision_embeddings def embed_input_ids( self, input_ids: torch.Tensor, multimodal_embeddings: MultiModalEmbeddings | None = None, *, is_multimodal: torch.Tensor | None = None, # Multi-modal token ID may exceed vocab size handle_oov_mm_token: bool = True, ) -> torch.Tensor: # This is to satisfy the type checker for each overload if multimodal_embeddings is None or is_multimodal is None: return super().embed_input_ids(input_ids) return super().embed_input_ids( input_ids, multimodal_embeddings=multimodal_embeddings, is_multimodal=is_multimodal, handle_oov_mm_token=handle_oov_mm_token, ) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs: object, ) -> torch.Tensor | IntermediateTensors: if intermediate_tensors is not None: inputs_embeds = None hidden_states = self.language_model( input_ids, positions, intermediate_tensors, inputs_embeds=inputs_embeds ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: return self.language_model.compute_logits(hidden_states) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]): loader = AutoWeightsLoader(self) return loader.load_weights(weights, mapper=self.hf_to_vllm_mapper)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/step3_vl.py", "license": "Apache License 2.0", "lines": 971, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/reasoning/step3_reasoning_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Iterable, Sequence from itertools import islice import regex as re from transformers import PreTrainedTokenizerBase from vllm.entrypoints.openai.chat_completion.protocol import ( ChatCompletionRequest, ) from vllm.entrypoints.openai.engine.protocol import DeltaMessage from vllm.logger import init_logger from vllm.reasoning import ReasoningParser logger = init_logger(__name__) class Step3ReasoningParser(ReasoningParser): """ Reasoning parser for Step3 model. The Step3 model uses </think> token to denote the end of reasoning text. This parser extracts all content before </think> as reasoning content. """ def __init__(self, tokenizer: PreTrainedTokenizerBase, *args, **kwargs): super().__init__(tokenizer, *args, **kwargs) self.think_end_token = "</think>" self.reasoning_regex = re.compile(rf"(.*?){self.think_end_token}", re.DOTALL) if not self.model_tokenizer: raise ValueError( "The model tokenizer must be passed to the ReasoningParser " "constructor during construction." ) self.think_end_token_id = self.vocab.get(self.think_end_token) if self.think_end_token_id is None: raise RuntimeError( "Step3 reasoning parser could not locate think end " "token in the tokenizer!" ) def extract_reasoning_streaming( self, previous_text: str, current_text: str, delta_text: str, previous_token_ids: Sequence[int], current_token_ids: Sequence[int], delta_token_ids: Sequence[int], ) -> DeltaMessage | None: """ Extract reasoning content from a delta message. Handles streaming output where previous + delta = current. Uses token IDs for faster processing. For text "abc</think>xyz": - 'abc' goes to reasoning - 'xyz' goes to content """ # Skip single special token if len(delta_token_ids) == 1 and delta_token_ids[0] == self.think_end_token_id: return None if self.think_end_token_id in delta_token_ids: # </think> in delta, extract reasoning content and remaining content end_index = delta_text.find(self.think_end_token) reasoning = delta_text[:end_index] content = delta_text[end_index + len(self.think_end_token) :] return DeltaMessage( reasoning=reasoning, content=content if content else None, ) elif self.think_end_token_id in previous_token_ids: # </think> already seen in previous text, everything is content return DeltaMessage(content=delta_text) else: # No </think> seen yet, everything is reasoning return DeltaMessage(reasoning=delta_text) def extract_reasoning( self, model_output: str, request: ChatCompletionRequest ) -> tuple[str | None, str | None]: # Check if the model output contains the </think> token if self.think_end_token not in model_output: # If no </think> token, everything is reasoning content return model_output, None else: # Find the first occurrence of </think> end_index = model_output.find(self.think_end_token) reasoning = model_output[:end_index] # Content after </think> token content = model_output[end_index + len(self.think_end_token) :] if len(content) == 0: content = None return reasoning, content def is_reasoning_end(self, input_ids: Sequence[int]) -> bool: return self.think_end_token_id in input_ids def is_reasoning_end_streaming( self, input_ids: Sequence[int], delta_ids: Iterable[int] ) -> bool: end_token_id = self.think_end_token_id return end_token_id in delta_ids def extract_content_ids(self, input_ids: list[int]) -> list[int]: if self.think_end_token_id not in islice( input_ids, 0, max(0, len(input_ids) - 1) ): return [] else: return input_ids[input_ids.index(self.think_end_token_id) + 1 :]
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/reasoning/step3_reasoning_parser.py", "license": "Apache License 2.0", "lines": 99, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/transformers_utils/configs/step3_vl.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any from transformers.configuration_utils import PretrainedConfig class Step3VisionEncoderConfig(PretrainedConfig): model_type = "step3_vision_encoder" def __init__( self, hidden_size=1792, intermediate_size=3072, output_hidden_size=4096, num_hidden_layers=63, num_attention_heads=16, num_channels=3, image_size=728, patch_size=14, hidden_act="quick_gelu", layer_norm_eps=1e-5, **kwargs, ): self.hidden_size = hidden_size self.intermediate_size = intermediate_size self.output_hidden_size = output_hidden_size self.num_hidden_layers = num_hidden_layers self.num_attention_heads = num_attention_heads self.num_channels = num_channels self.patch_size = patch_size self.image_size = image_size self.layer_norm_eps = layer_norm_eps self.hidden_act = hidden_act super().__init__(**kwargs) class Step3TextConfig(PretrainedConfig): model_type = "step3_text" architectures = ["Step3TextForCausalLM"] def __init__( self, hidden_size: int = 7168, intermediate_size: int = 18432, num_attention_heads: int = 64, num_attention_groups: int = 1, num_hidden_layers: int = 61, max_seq_len: int = 65536, vocab_size: int = 128815, rms_norm_eps: float = 1e-5, moe_intermediate_size: int = 5120, moe_num_experts: int = 48, moe_top_k: int = 3, rope_parameters: dict[str, Any] | None = None, max_position_embedding: int = 65536, share_expert_dim: int = 5120, share_q_dim: int = 2048, head_dim: int = 256, norm_expert_weight: bool = False, moe_layers_enum: tuple[int, ...] = ( 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, ), **kwargs, ) -> None: self.hidden_size = hidden_size self.intermediate_size = intermediate_size self.num_attention_heads = num_attention_heads self.num_attention_groups = num_attention_groups self.num_hidden_layers = num_hidden_layers self.max_seq_len = max_seq_len self.vocab_size = vocab_size self.rms_norm_eps = rms_norm_eps self.moe_intermediate_size = moe_intermediate_size self.moe_num_experts = moe_num_experts self.moe_top_k = moe_top_k # Try to set `rope_scaling` if available, otherwise use `rope_parameters` rope_scaling = kwargs.pop("rope_scaling", None) rope_parameters = rope_scaling or rope_parameters or {"rope_type": "default"} rope_theta = kwargs.pop("rope_theta", 500000.0) if "rope_theta" not in rope_parameters: rope_parameters["rope_theta"] = rope_theta self.rope_parameters = rope_parameters self.max_position_embedding = max_position_embedding self.share_expert_dim = share_expert_dim self.share_q_dim = share_q_dim self.head_dim = head_dim self.norm_expert_weight = norm_expert_weight self.moe_layers_enum = moe_layers_enum super().__init__(**kwargs) class Step3VLConfig(PretrainedConfig): model_type = "step3_vl" def __init__( self, vision_config: dict | Step3VisionEncoderConfig | None = None, text_config: dict | Step3TextConfig | None = None, understand_projector_stride: int = 1, projector_bias: bool = True, image_token_id: int = 128001, **kwargs, ) -> None: if vision_config is None: vision_config = Step3VisionEncoderConfig() elif isinstance(vision_config, dict): vision_config = Step3VisionEncoderConfig(**vision_config) self.vision_config = vision_config if text_config is None: text_config = Step3TextConfig() elif isinstance(text_config, dict): text_config = Step3TextConfig(**text_config) self.text_config = text_config self.understand_projector_stride = understand_projector_stride self.projector_bias = projector_bias self.hidden_size = text_config.hidden_size self.image_token_id = image_token_id super().__init__(**kwargs)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/configs/step3_vl.py", "license": "Apache License 2.0", "lines": 164, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/quantization/utils/flashinfer_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from enum import Enum import torch from vllm import envs from vllm.logger import init_logger from vllm.model_executor.layers.fused_moe.activation import MoEActivation from vllm.platforms import current_platform from vllm.utils.math_utils import round_up logger = init_logger(__name__) class FlashinferMoeBackend(Enum): TENSORRT_LLM = "TensorRT-LLM" CUTLASS = "CUTLASS" CUTEDSL = "CUTEDSL" def activation_to_flashinfer_int(activation: MoEActivation) -> int: from flashinfer.fused_moe.core import ActivationType # silu and gelu are mapped to their gated versions SwiGLU and GeGLU respectively ACTIVATION_TO_FI_ACTIVATION = { MoEActivation.SILU_NO_MUL: ActivationType.Silu, MoEActivation.GELU_NO_MUL: ActivationType.Gelu, MoEActivation.SILU: ActivationType.Swiglu, MoEActivation.GELU: ActivationType.Geglu, MoEActivation.RELU2_NO_MUL: ActivationType.Relu2, } return ACTIVATION_TO_FI_ACTIVATION[activation].value def swap_w13_to_w31(x: torch.Tensor) -> torch.Tensor: return ( x.reshape(-1, 2, x.shape[-2] // 2, x.shape[-1]).flip(dims=[1]).reshape(x.shape) ) def rotate_weights_for_fi_trtllm_fp8_per_tensor_moe( gemm1_weights: torch.Tensor, gemm2_weights: torch.Tensor, is_gated_activation: bool ): """Shuffle weights for for FI TRT-LLM Format""" from flashinfer import reorder_rows_for_gated_act_gemm, shuffle_matrix_a epilogue_tile_m = 128 num_experts = gemm1_weights.shape[0] hidden_size = gemm1_weights.shape[-1] intermediate_size = gemm1_weights.shape[1] // 2 # Reorder rows of W1 for fused gated activation gemm1_weights_fp8_interleaved = [] for i in range(num_experts): gemm1_weights_fp8_interleaved.append( reorder_rows_for_gated_act_gemm(gemm1_weights[i]) if is_gated_activation else gemm1_weights[i] ) # Stack weights and scales for all experts gemm1_weights_fp8_interleaved = torch.stack(gemm1_weights_fp8_interleaved).reshape( num_experts, 2 * intermediate_size, hidden_size ) # Shuffle weights and scaling factors for transposed mma output gemm1_weights_fp8_shuffled = [] gemm2_weights_fp8_shuffled = [] for i in range(num_experts): gemm1_weights_fp8_shuffled.append( shuffle_matrix_a( gemm1_weights_fp8_interleaved[i].view(torch.uint8), epilogue_tile_m ) ) gemm2_weights_fp8_shuffled.append( shuffle_matrix_a(gemm2_weights[i].view(torch.uint8), epilogue_tile_m) ) # Stack weights for all experts gemm1_weights.data = torch.stack(gemm1_weights_fp8_shuffled).view( torch.float8_e4m3fn ) gemm2_weights.data = torch.stack(gemm2_weights_fp8_shuffled).view( torch.float8_e4m3fn ) def register_scales_for_trtllm_fp8_per_tensor_moe( layer: torch.nn.Module, w13_scale: torch.Tensor, w13_input_scale: torch.Tensor, w2_scale: torch.Tensor, w2_input_scale: torch.Tensor, ) -> None: """Register necessary scales for FlashInfer TRTLLM FP8 MoE kernel""" g1_alphas, g2_alphas = make_fp8_moe_alpha_scales_for_fi( w13_scale=w13_scale, w13_input_scale=w13_input_scale, w2_scale=w2_scale, w2_input_scale=w2_input_scale, ) layer.w2_input_scale_inv = 1.0 / w2_input_scale layer.output1_scales_gate_scalar = g1_alphas if layer.activation.is_gated: layer.output1_scales_scalar = g1_alphas * layer.w2_input_scale_inv else: layer.output1_scales_scalar = ( torch.ones_like(g1_alphas) * layer.w2_input_scale_inv ) layer.output2_scales_scalar = g2_alphas def apply_fi_trtllm_fp8_per_tensor_moe( layer: torch.nn.Module, hidden_states: torch.Tensor, router_logits: torch.Tensor, routing_bias: torch.Tensor | None, top_k: int, num_expert_group: int | None, topk_group: int | None, global_num_experts: int, apply_router_weight_on_input: bool, ) -> torch.Tensor: from flashinfer.fused_moe import RoutingMethodType import vllm.model_executor.layers.fused_moe.flashinfer_trtllm_moe # noqa: E501, F401 from vllm.model_executor.models.llama4 import Llama4MoE # Added to the layer by: register_scales_for_trtllm_fp8_per_tensor_moe assert ( hasattr(layer, "output1_scales_scalar") and hasattr(layer, "output1_scales_gate_scalar") and hasattr(layer, "output2_scales_scalar") ) if layer.routing_method_type == RoutingMethodType.Llama4: assert ( not layer.renormalize and layer.custom_routing_function == Llama4MoE.custom_routing_function ), ( "FusedMoE flashinfer kernels with Llama4 routing method are only " "supported for Llama4" ) else: assert layer.custom_routing_function is None, ( "Custom routing function is only supported for Llama4" ) activation_type = activation_to_flashinfer_int(layer.activation) return torch.ops.vllm.fi_trtllm_fp8_per_tensor_moe( routing_logits=router_logits, routing_bias=routing_bias, hidden_states=hidden_states, input_scale=layer.w13_input_scale, gemm1_weights=layer.w13_weight, gemm2_weights=layer.w2_weight, output1_scales_scalar=layer.output1_scales_scalar, output1_scales_gate_scalar=layer.output1_scales_gate_scalar, output2_scales_scalar=layer.output2_scales_scalar, num_experts=global_num_experts, top_k=top_k, num_expert_group=num_expert_group, topk_group=topk_group, intermediate_size=layer.intermediate_size_per_partition, local_expert_offset=layer.ep_rank * layer.local_num_experts, local_num_experts=layer.local_num_experts, use_routing_scales_on_input=apply_router_weight_on_input, routing_method_type=layer.routing_method_type, activation_type=activation_type, ) def make_fp8_moe_alpha_scales_for_fi( w13_scale: torch.Tensor, w13_input_scale: torch.Tensor, w2_scale: torch.Tensor, w2_input_scale: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: g1_alphas = (w13_scale * w13_input_scale).squeeze() g2_alphas = (w2_scale * w2_input_scale).squeeze() return g1_alphas, g2_alphas def get_flashinfer_moe_backend() -> FlashinferMoeBackend: backend_map = { "throughput": FlashinferMoeBackend.CUTLASS, "latency": FlashinferMoeBackend.TENSORRT_LLM, "masked_gemm": FlashinferMoeBackend.CUTEDSL, } flashinfer_moe_backend = envs.VLLM_FLASHINFER_MOE_BACKEND if flashinfer_moe_backend in backend_map: if ( flashinfer_moe_backend == "latency" and not current_platform.is_device_capability_family(100) ): logger.info_once( "Flashinfer TRTLLM MOE backend is only supported on " "SM100 and later, using CUTLASS backend instead", scope="local", ) return FlashinferMoeBackend.CUTLASS return backend_map[flashinfer_moe_backend] elif current_platform.is_device_capability(90): return FlashinferMoeBackend.CUTLASS raise ValueError( f"Unknown flashinfer moe backend: {flashinfer_moe_backend!r}. " f"Expected one of {list(backend_map.keys())}." ) def is_flashinfer_supporting_global_sf(backend: FlashinferMoeBackend | None) -> bool: # TODO(shuw@nvidia): Update when new backends are added. backends_supporting_global_sf = ( FlashinferMoeBackend.CUTLASS, FlashinferMoeBackend.TENSORRT_LLM, FlashinferMoeBackend.CUTEDSL, ) return backend in backends_supporting_global_sf def convert_moe_weights_to_flashinfer_trtllm_block_layout( cache_permute_indices: dict[torch.Size, torch.Tensor], w13_weight: torch.Tensor, w2_weight: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: """Convert expert weights to FlashInfer's block layout. This reorders W13 and W2 into the expected epilogue-tiled block layout and returns the shuffled weight tensors. """ if w13_weight.dtype != torch.bfloat16 or w2_weight.dtype != torch.bfloat16: raise ValueError( "Unquantized Moe Backend FlashInfer TRTLLM requires bfloat16 weights" ) from flashinfer.fused_moe.core import ( _maybe_get_cached_w3_w1_permute_indices, convert_to_block_layout, get_w2_permute_indices_with_cache, ) epilogue_tile_m = 128 block_k = 128 # Reorder rows of W13 and W2 for fused gated activation and convert to the # block layout expected by the FlashInfer kernel. num_experts = w13_weight.shape[0] device_w13 = w13_weight.device device_w2 = w2_weight.device w13_weights_shuffled: list[torch.Tensor] = [] w2_weights_shuffled: list[torch.Tensor] = [] for i in range(num_experts): permute_indices = _maybe_get_cached_w3_w1_permute_indices( cache_permute_indices, w13_weight[i].view(torch.uint8), epilogue_tile_m, ) tmp_weights1 = ( w13_weight[i] .clone() .view(torch.uint8)[permute_indices.to(device_w13)] .contiguous() ) permute_indices = get_w2_permute_indices_with_cache( cache_permute_indices, w2_weight[i].view(torch.uint8), epilogue_tile_m, ) tmp_weights2 = ( w2_weight[i] .clone() .view(torch.uint8)[permute_indices.to(device_w2)] .contiguous() ) tmp_weights1 = convert_to_block_layout(tmp_weights1.view(torch.uint8), block_k) tmp_weights2 = convert_to_block_layout(tmp_weights2.view(torch.uint8), block_k) w13_weights_shuffled.append(tmp_weights1.view(torch.bfloat16)) w2_weights_shuffled.append(tmp_weights2.view(torch.bfloat16)) # Stack weights for all experts and return as BF16 tensors. w13_weights_shuffled_tensor = ( torch.stack(w13_weights_shuffled).view(torch.bfloat16).contiguous() ) w2_weights_shuffled_tensor = ( torch.stack(w2_weights_shuffled).view(torch.bfloat16).contiguous() ) return w13_weights_shuffled_tensor, w2_weights_shuffled_tensor def align_fp4_moe_weights_for_fi( w13: torch.Tensor, w13_scale: torch.Tensor, w2: torch.Tensor, w2_scale: torch.Tensor, is_act_and_mul: bool, min_alignment: int = 16, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor, int]: """Pad intermediate size so FlashInfer kernels' alignment constraints hold. Some FlashInfer FP4 MoE kernels require the intermediate size used for GEMM to be divisible by a small alignment value. When this is not satisfied (e.g. with certain tensor-parallel sizes), we pad the gate/up and down projection weights along the intermediate dim. """ # Current local intermediate size (per partition) is the K dimension of # the down projection. num_experts, hidden_size, intermediate = w2.shape intermediate *= 2 # because of packed FP4 padded_intermediate = round_up(intermediate, min_alignment) if padded_intermediate == intermediate: return w13, w13_scale, w2, w2_scale, intermediate logger.info_once( "Padding intermediate size from %d to %d for up/down projection weights.", intermediate, padded_intermediate, scope="local", ) up_mult = 2 if is_act_and_mul else 1 padded_gate_up_dim = up_mult * padded_intermediate # Pad w13 and w2 along its intermediate dimension. padded_w13 = w13.new_zeros((num_experts, padded_gate_up_dim, hidden_size // 2)) padded_w13[:, : w13.shape[1], :] = w13 padded_w2 = w2.new_zeros((num_experts, hidden_size, padded_intermediate // 2)) padded_w2[:, :, : w2.shape[2]] = w2 padded_w13_scale = w13_scale.new_zeros( (num_experts, padded_gate_up_dim, hidden_size // 16) ) padded_w13_scale[:, : w13_scale.shape[1], :] = w13_scale padded_w2_scale = w2_scale.new_zeros( (num_experts, hidden_size, padded_intermediate // 16) ) padded_w2_scale[:, :, : w2_scale.shape[2]] = w2_scale return padded_w13, padded_w13_scale, padded_w2, padded_w2_scale, padded_intermediate def align_fp8_moe_weights_for_fi( w13: torch.Tensor, w2: torch.Tensor, is_act_and_mul: bool, min_alignment: int = 16 ) -> tuple[torch.Tensor, torch.Tensor, int]: """Pad intermediate size so FlashInfer kernels' alignment constraints hold. Some FlashInfer FP8 MoE kernels require the (gated) intermediate size used for GEMM to be divisible by a small alignment value. When this is not satisfied (e.g. with certain tensor-parallel sizes), we pad the gate/up and down projection weights along the intermediate dim. """ # Current local intermediate size (per partition) is the K dimension of # the down projection. num_experts, hidden_size, intermediate = w2.shape padded_intermediate = round_up(intermediate, min_alignment) if padded_intermediate == intermediate: return w13, w2, intermediate logger.info_once( "Padding intermediate size from %d to %d for up/down projection weights.", intermediate, padded_intermediate, scope="local", ) up_mult = 2 if is_act_and_mul else 1 padded_gate_up_dim = up_mult * padded_intermediate # Pad w13 and w2 along its intermediate dimension. padded_w13 = w13.new_zeros((num_experts, padded_gate_up_dim, hidden_size)) padded_w13[:, : w13.shape[1], :] = w13 padded_w2 = w2.new_zeros((num_experts, hidden_size, padded_intermediate)) padded_w2[:, :, :intermediate] = w2 return padded_w13, padded_w2, padded_intermediate def prepare_fp8_moe_layer_for_fi( layer: torch.nn.Module, w13: torch.Tensor, w2: torch.Tensor, w13_scale: torch.Tensor, w13_input_scale: torch.Tensor | None, w2_scale: torch.Tensor, w2_input_scale: torch.Tensor | None, is_trtllm: bool = False, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Convert Fp8 MoE weights to flashinfer kernel format Note that for trtllm we update the model state dict with the scale format needed for these kernels. Note that for per-tensor, we update the layer's intermediate size if the weights needed padding. """ assert hasattr(layer.moe_config, "is_act_and_mul") block_quant = ( hasattr(layer, "weight_block_size") and layer.weight_block_size is not None ) # Some FI MoE kernels require internal alignment of 16 # for the gate-up proj. Pad the weights to respect this. is_gated = layer.activation.is_gated if not block_quant: min_alignment = 16 if is_gated else 128 w13, w2, new_intermediate = align_fp8_moe_weights_for_fi( w13, w2, layer.moe_config.is_act_and_mul, min_alignment, ) layer.intermediate_size_per_partition = new_intermediate # FI kernels require W31 layout rather than W13. if layer.moe_config.is_act_and_mul: w13 = swap_w13_to_w31(w13) if block_quant: w13_scale = swap_w13_to_w31(w13_scale) # FI TRT-LLM FP8 per-tensor MoE kernel requires weight shuffle # and registration of alpha scales. Note that we do not register # as nn.Parameters since they are not needed for weight-reloading. if is_trtllm and not block_quant: assert w13_input_scale is not None assert w2_input_scale is not None rotate_weights_for_fi_trtllm_fp8_per_tensor_moe(w13, w2, is_gated) register_scales_for_trtllm_fp8_per_tensor_moe( layer, w13_scale=w13_scale, w13_input_scale=w13_input_scale, w2_scale=w2_scale, w2_input_scale=w2_input_scale, ) # Clamp block scales to avoid NaN from the FlashInfer CUTLASS kernel. # Some FP8 models have near-zero block scales (~1e-23) for dead/unused # experts. The CUTLASS kernel doesn't handle these correctly on Hopper # (SM 9.0), producing NaN instead of near-zero output. Clamping to a # small minimum prevents this without affecting model accuracy since # these experts' effective weights are already zero. if block_quant: _FI_CUTLASS_MIN_BLOCK_SCALE = 1e-10 w13_scale.clamp_(min=_FI_CUTLASS_MIN_BLOCK_SCALE) w2_scale.clamp_(min=_FI_CUTLASS_MIN_BLOCK_SCALE) return w13, w2, w13_scale
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/utils/flashinfer_utils.py", "license": "Apache License 2.0", "lines": 385, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:examples/offline_inference/async_llm_streaming.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Simple example demonstrating streaming offline inference with AsyncLLM (V1 engine). This script shows the core functionality of vLLM's AsyncLLM engine for streaming token-by-token output in offline inference scenarios. It demonstrates DELTA mode streaming where you receive new tokens as they are generated. Usage: python examples/offline_inference/async_llm_streaming.py """ import asyncio from vllm import SamplingParams from vllm.engine.arg_utils import AsyncEngineArgs from vllm.sampling_params import RequestOutputKind from vllm.v1.engine.async_llm import AsyncLLM async def stream_response(engine: AsyncLLM, prompt: str, request_id: str) -> None: """ Stream response from AsyncLLM and display tokens as they arrive. This function demonstrates the core streaming pattern: 1. Create SamplingParams with DELTA output kind 2. Call engine.generate() and iterate over the async generator 3. Print new tokens as they arrive 4. Handle the finished flag to know when generation is complete """ print(f"\n🚀 Prompt: {prompt!r}") print("💬 Response: ", end="", flush=True) # Configure sampling parameters for streaming sampling_params = SamplingParams( max_tokens=100, temperature=0.8, top_p=0.95, seed=42, # For reproducible results output_kind=RequestOutputKind.DELTA, # Get only new tokens each iteration ) try: # Stream tokens from AsyncLLM async for output in engine.generate( request_id=request_id, prompt=prompt, sampling_params=sampling_params ): # Process each completion in the output for completion in output.outputs: # In DELTA mode, we get only new tokens generated since last iteration new_text = completion.text if new_text: print(new_text, end="", flush=True) # Check if generation is finished if output.finished: print("\n✅ Generation complete!") break except Exception as e: print(f"\n❌ Error during streaming: {e}") raise async def main(): print("🔧 Initializing AsyncLLM...") # Create AsyncLLM engine with simple configuration engine_args = AsyncEngineArgs( model="meta-llama/Llama-3.2-1B-Instruct", enforce_eager=True, # Faster startup for examples ) engine = AsyncLLM.from_engine_args(engine_args) try: # Example prompts to demonstrate streaming prompts = [ "The future of artificial intelligence is", "In a galaxy far, far away", "The key to happiness is", ] print(f"🎯 Running {len(prompts)} streaming examples...") # Process each prompt for i, prompt in enumerate(prompts, 1): print(f"\n{'=' * 60}") print(f"Example {i}/{len(prompts)}") print(f"{'=' * 60}") request_id = f"stream-example-{i}" await stream_response(engine, prompt, request_id) # Brief pause between examples if i < len(prompts): await asyncio.sleep(0.5) print("\n🎉 All streaming examples completed!") finally: # Always clean up the engine print("🔧 Shutting down engine...") engine.shutdown() if __name__ == "__main__": try: asyncio.run(main()) except KeyboardInterrupt: print("\n🛑 Interrupted by user")
{ "repo_id": "vllm-project/vllm", "file_path": "examples/offline_inference/async_llm_streaming.py", "license": "Apache License 2.0", "lines": 88, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/distributed/test_kvlayout.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from vllm.config import ( DeviceConfig, KVTransferConfig, ModelConfig, VllmConfig, set_current_vllm_config, ) from vllm.distributed.kv_transfer.kv_connector.utils import ( get_kv_connector_cache_layout, ) from vllm.logger import init_logger logger = init_logger("test_expert_parallel") def test_get_kv_connector_cache_layout_without_kv_connector(): vllm_config = VllmConfig(device_config=DeviceConfig("cpu")) with set_current_vllm_config(vllm_config): # Test with default settings layout = get_kv_connector_cache_layout() assert layout == "NHD" def test_get_kv_connector_cache_layout_with_lmcache_connector(): kv_transfer_config = KVTransferConfig( kv_connector="LMCacheConnectorV1", kv_role="kv_both", ) vllm_config = VllmConfig( device_config=DeviceConfig("cpu"), kv_transfer_config=kv_transfer_config ) with set_current_vllm_config(vllm_config): # Test with default settings layout = get_kv_connector_cache_layout() assert layout == "NHD" def test_get_kv_connector_cache_layout_with_nixl_connector(): kv_transfer_config = KVTransferConfig( kv_connector="NixlConnector", kv_role="kv_both", ) model_config = ModelConfig() vllm_config = VllmConfig( device_config=DeviceConfig("cpu"), model_config=model_config, kv_transfer_config=kv_transfer_config, ) with set_current_vllm_config(vllm_config): # Test with default settings layout = get_kv_connector_cache_layout() assert layout == "HND" def test_get_kv_connector_cache_layout_with_multi_connector(): kv_transfer_config = KVTransferConfig( kv_connector="MultiConnector", kv_role="kv_both", kv_connector_extra_config={ "connectors": [ {"kv_connector": "ExampleConnector", "kv_role": "kv_both"}, {"kv_connector": "NixlConnector", "kv_role": "kv_both"}, ] }, ) model_config = ModelConfig() vllm_config = VllmConfig( device_config=DeviceConfig("cpu"), model_config=model_config, kv_transfer_config=kv_transfer_config, ) with set_current_vllm_config(vllm_config): # Test with default settings layout = get_kv_connector_cache_layout() assert layout == "HND"
{ "repo_id": "vllm-project/vllm", "file_path": "tests/distributed/test_kvlayout.py", "license": "Apache License 2.0", "lines": 68, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/e2e/test_kv_sharing_fast_prefill.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import random import pytest from vllm import LLM, SamplingParams from vllm.config import CompilationConfig, CompilationMode from vllm.platforms import current_platform from ...utils import check_answers, fork_new_process_for_each_test, prep_prompts # global seed SEED = 42 @pytest.fixture def test_prompts(): """ Adapted from tests/v1/e2e/test_spec_decode.py """ prompt_types = ["repeat", "sentence"] # Setting higher num prompts increases the chance of numerics mismatch # due to matrix multiplication numerics depending on batch dimension num_prompts = 10 prompts = [] random.seed(0) random_prompt_type_choices = random.choices(prompt_types, k=num_prompts) for kind in random_prompt_type_choices: word_choices = ["test", "temp", "hello", "where"] word = random.choice(word_choices) if kind == "repeat": prompt = f"""please repeat the word '{word}' 10 times.""" elif kind == "sentence": prompt = f"""please give a ten-word sentence that uses the word {word} at least once.""" else: raise ValueError(f"Unknown prompt type: {kind}") prompts.append(prompt) return prompts use_fork_for_test = ( fork_new_process_for_each_test if not current_platform.is_rocm() else lambda x: x ) @use_fork_for_test @pytest.mark.parametrize("kv_sharing_fast_prefill", [False, True]) @pytest.mark.parametrize("enforce_eager", [True, False]) def test_kv_sharing_fast_prefill( monkeypatch: pytest.MonkeyPatch, kv_sharing_fast_prefill: bool, enforce_eager: bool, ): if not enforce_eager and current_platform.is_rocm(): # Relevant context: https://github.com/vllm-project/vllm/pull/29244 pytest.skip( "ROCm: torch.compile produces incorrect output for gemma-3n's GELU " "with tanh approximation. Use enforce_eager=True instead." ) sampling_params = SamplingParams(temperature=0.0, max_tokens=100) compilation_config = CompilationConfig( # This allows vLLM compilation backend to handle allocating and # managing buffers for cudagraph cudagraph_copy_inputs=True, mode=CompilationMode.VLLM_COMPILE if not enforce_eager else CompilationMode.NONE, ) batch_size = 10 with monkeypatch.context() as m: # Make scheduling deterministic for reproducibility if current_platform.is_rocm(): # Use spawn to prevent cuda re-initialization error m.setenv("VLLM_WORKER_MULTIPROC_METHOD", "spawn") else: m.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "0") prompts, answer, indices = prep_prompts(batch_size) llm = LLM( model="google/gemma-3n-E2B-it", enforce_eager=enforce_eager, compilation_config=compilation_config, seed=SEED, kv_sharing_fast_prefill=kv_sharing_fast_prefill, ) responses = llm.generate(prompts, sampling_params) check_answers( indices, answer, [response.outputs[0].text for response in responses], accept_rate=1.0, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/e2e/test_kv_sharing_fast_prefill.py", "license": "Apache License 2.0", "lines": 83, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:benchmarks/kernels/benchmark_per_token_group_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import math from collections.abc import Callable from contextlib import contextmanager from unittest.mock import patch import torch from vllm.model_executor.layers.quantization.utils import fp8_utils, int8_utils from vllm.platforms import current_platform @contextmanager def _triton_mode(): """Temporarily force the Triton fallback path""" with patch("vllm.platforms.current_platform.is_cuda", return_value=False): yield def _time_cuda( fn: Callable[[], tuple[torch.Tensor, torch.Tensor]], warmup_iters: int, bench_iters: int, ) -> float: # warmup for _ in range(warmup_iters): fn() torch.cuda.synchronize() start = torch.Event(enable_timing=True) end = torch.Event(enable_timing=True) start.record() for _ in range(bench_iters): fn() end.record() torch.cuda.synchronize() return start.elapsed_time(end) / bench_iters # ms/iter def _run_single( shape: tuple[int, int], group_size: int, dtype: str, *, column_major: bool = False, scale_ue8m0: bool = False, warmup_iters: int, bench_iters: int, ) -> None: num_tokens, hidden_dim = shape device = torch.device("cuda") torch.manual_seed(42) x = torch.randn(num_tokens, hidden_dim, device=device, dtype=torch.bfloat16) * 8 if dtype == "fp8": def cuda_impl(): return fp8_utils.per_token_group_quant_fp8( x, group_size, column_major_scales=column_major, use_ue8m0=scale_ue8m0, ) def triton_impl(): with _triton_mode(): return fp8_utils.per_token_group_quant_fp8( x, group_size, column_major_scales=column_major, use_ue8m0=scale_ue8m0, ) elif dtype == "int8": def cuda_impl(): return int8_utils.per_token_group_quant_int8(x, group_size) def triton_impl(): with _triton_mode(): return int8_utils.per_token_group_quant_int8(x, group_size) else: raise ValueError("dtype must be 'fp8' or 'int8'") cuda_ms = _time_cuda(cuda_impl, warmup_iters, bench_iters) triton_ms = _time_cuda(triton_impl, warmup_iters, bench_iters) speedup = triton_ms / cuda_ms if cuda_ms else math.inf cfg_desc = ( f"shape={shape} gs={group_size:<3} col_major={column_major:<5} " f"ue8m0={scale_ue8m0:<5} dtype={dtype}" ) print( f"{cfg_desc:55} | CUDA {cuda_ms:7.3f} ms | Triton {triton_ms:7.3f} ms | " f"speed-up ×{speedup:5.2f}" ) def parse_args(): parser = argparse.ArgumentParser() parser.add_argument("--warmup-iters", type=int, default=10) parser.add_argument("--bench-iters", type=int, default=100) parser.add_argument("--dtype", choices=["fp8", "int8", "both"], default="both") return parser.parse_args() if __name__ == "__main__": if not current_platform.is_cuda(): raise RuntimeError("CUDA device is required to run this benchmark.") args = parse_args() warmup_iters, bench_iters = args.warmup_iters, args.bench_iters shapes = [(32, 128), (64, 256), (16, 512)] group_sizes = [64, 128] dtypes = ["fp8", "int8"] if args.dtype == "both" else [args.dtype] header = ( "Configuration".ljust(55) + " | " + "CUDA (ms)".center(12) + " | " + "Triton (ms)".center(13) + " | " + "Speed-up" ) print(header) print("-" * len(header)) for dtype in dtypes: for shape in shapes: for gs in group_sizes: if dtype == "fp8": for col_major in (False, True): for ue8m0 in (False, True): _run_single( shape, gs, dtype, column_major=col_major, scale_ue8m0=ue8m0, warmup_iters=warmup_iters, bench_iters=bench_iters, ) else: # INT8 has no col-major / ue8m0 switches _run_single( shape, gs, dtype, warmup_iters=warmup_iters, bench_iters=bench_iters, )
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/kernels/benchmark_per_token_group_quant.py", "license": "Apache License 2.0", "lines": 129, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/models/multimodal/processing/test_mllama4.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests for mllama's multimodal preprocessing and profiling.""" import pytest from torch import prod from transformers import Llama4Config from vllm.multimodal import MULTIMODAL_REGISTRY from ...utils import build_model_context @pytest.mark.parametrize("model_id", ["meta-llama/Llama-Guard-4-12B"]) @pytest.mark.parametrize("max_model_len", [4096, 8192, 25600, 131072]) def test_profiling(model_id: str, max_model_len: int): model_config_kwargs = { "max_model_len": max_model_len, } mm_counts = {"image": 1} ctx = build_model_context( model_id, model_config_kwargs=model_config_kwargs, limit_mm_per_prompt=mm_counts, ) mm_inputs = MULTIMODAL_REGISTRY.get_dummy_mm_inputs( ctx.model_config, mm_counts=mm_counts, ) hf_config = ctx.get_hf_config(Llama4Config) image_size = hf_config.vision_config.image_size patch_size = hf_config.vision_config.patch_size downsample_ratio = int( round(1.0 / (hf_config.vision_config.pixel_shuffle_ratio**2)) ) tokens_per_patch = ((image_size // patch_size) ** 2) // downsample_ratio mm_data = mm_inputs["mm_kwargs"].get_data() chunks_per_image = prod(mm_data["patches_per_image"]) total_num_patches = chunks_per_image * tokens_per_patch num_tiles = ( mm_data["aspect_ratios"][0][0] * mm_data["aspect_ratios"][0][1] ) # x-y separator tokens total_tokens = ( total_num_patches.item() + num_tiles.item() + 3 ) # image start, image, image end assert total_num_patches == sum( item.get_num_embeds() for item in mm_inputs["mm_placeholders"]["image"] ) assert total_tokens == sum( placeholder.length for placeholder in mm_inputs["mm_placeholders"]["image"] )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/multimodal/processing/test_mllama4.py", "license": "Apache License 2.0", "lines": 46, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:examples/offline_inference/basic/reward.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from argparse import Namespace from vllm import LLM, EngineArgs from vllm.utils.argparse_utils import FlexibleArgumentParser def parse_args(): parser = FlexibleArgumentParser() parser = EngineArgs.add_cli_args(parser) # Set example specific arguments parser.set_defaults( model="internlm/internlm2-1_8b-reward", runner="pooling", enforce_eager=True, max_model_len=1024, trust_remote_code=True, ) return parser.parse_args() def main(args: Namespace): # Sample prompts. prompts = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] # Create an LLM. # You should pass runner="pooling" for reward models llm = LLM(**vars(args)) # Generate rewards. The output is a list of PoolingRequestOutput. outputs = llm.reward(prompts) # Print the outputs. print("\nGenerated Outputs:\n" + "-" * 60) for prompt, output in zip(prompts, outputs): rewards = output.outputs.data rewards_trimmed = ( (str(rewards[:16])[:-1] + ", ...]") if len(rewards) > 16 else rewards ) print(f"Prompt: {prompt!r} \nReward: {rewards_trimmed} (size={len(rewards)})") print("-" * 60) if __name__ == "__main__": args = parse_args() main(args)
{ "repo_id": "vllm-project/vllm", "file_path": "examples/offline_inference/basic/reward.py", "license": "Apache License 2.0", "lines": 42, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/quantization/compressed_tensors/schemes/compressed_tensors_w4a8_int.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable import torch from vllm.logger import init_logger from vllm.model_executor.kernels.linear import ( MPLinearLayerConfig, choose_mp_linear_kernel, ) from vllm.model_executor.layers.quantization.compressed_tensors.schemes import ( CompressedTensorsScheme, ) from vllm.model_executor.parameter import ( ChannelQuantScaleParameter, GroupQuantScaleParameter, ModelWeightParameter, ) from vllm.scalar_type import scalar_types logger = init_logger(__name__) __all__ = ["CompressedTensorsW4A8Int"] W4A8_SUPPORTED_TYPES_MAP = { 4: scalar_types.int4, } W4A8_SUPPORTED_BITS = list(W4A8_SUPPORTED_TYPES_MAP.keys()) class CompressedTensorsW4A8Int(CompressedTensorsScheme): _kernel_backends_being_used: set[str] = set() def __init__( self, strategy: str, num_bits: int, group_size: int | None = None, is_static_input_scheme: bool = False, input_symmetric: bool = True, ): self.strategy = strategy self.group_size = -1 if group_size is None else group_size self.is_static_input_scheme = is_static_input_scheme self.input_symmetric = input_symmetric if num_bits not in W4A8_SUPPORTED_TYPES_MAP: raise ValueError( f"Unsupported num_bits = {num_bits}." f"Supported num_bits = {W4A8_SUPPORTED_TYPES_MAP.keys()}" ) self.quant_type = W4A8_SUPPORTED_TYPES_MAP[num_bits] @classmethod def get_min_capability(cls) -> int: return 1 def create_weights( self, layer: torch.nn.Module, output_size: int, input_size: int, output_partition_sizes: list[int], input_size_per_partition: int, params_dtype: torch.dtype, weight_loader: Callable, **kwargs, ): output_size_per_partition = sum(output_partition_sizes) row_parallel = input_size != input_size_per_partition # Compute effective group_size if self.group_size == -1: effective_group_size = ( input_size_per_partition if row_parallel else input_size ) else: effective_group_size = self.group_size # Ensure group_size divides input_size_per_partition assert input_size_per_partition % effective_group_size == 0, ( f"input_size_per_partition {input_size_per_partition}" f" not divisible by group_size {effective_group_size}" ) # Determine scale partitioning is_channelwise = self.group_size == -1 repeat_scales = is_channelwise and row_parallel partition_scales = not repeat_scales mp_linear_kernel_config = MPLinearLayerConfig( full_weight_shape=(input_size, output_size), partition_weight_shape=( input_size_per_partition, output_size_per_partition, ), weight_type=self.quant_type, act_type=params_dtype, group_size=effective_group_size, zero_points=False, has_g_idx=False, ) kernel_type = choose_mp_linear_kernel(mp_linear_kernel_config) if kernel_type.__name__ not in self._kernel_backends_being_used: logger.info("Using %s for CompressedTensorsW4A8Int", kernel_type.__name__) self._kernel_backends_being_used.add(kernel_type.__name__) scales_and_zp_size = input_size_per_partition // effective_group_size weight = ModelWeightParameter( data=torch.empty( output_size_per_partition, input_size_per_partition, dtype=torch.int8 ), input_dim=1, output_dim=0, weight_loader=weight_loader, ) layer.register_parameter("weight", weight) weight_scale_args = { "weight_loader": weight_loader, "data": torch.empty( output_size_per_partition, scales_and_zp_size, dtype=params_dtype ), } if partition_scales: weight_scale = GroupQuantScaleParameter( output_dim=0, input_dim=1, **weight_scale_args ) else: weight_scale = ChannelQuantScaleParameter(output_dim=0, **weight_scale_args) layer.register_parameter("weight_packed", weight) layer.register_parameter("weight_scale", weight_scale) self.kernel = kernel_type( mp_linear_kernel_config, w_q_param_name="weight_packed", w_s_param_name="weight_scale", w_zp_param_name=None, w_gidx_param_name=None, ) def process_weights_after_loading(self, layer: torch.nn.Module) -> None: self.kernel.process_weights_after_loading(layer) def apply_weights( self, layer: torch.nn.Module, x: torch.Tensor, bias: torch.Tensor | None ) -> torch.Tensor: return self.kernel.apply_weights(layer, x, bias)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/compressed_tensors/schemes/compressed_tensors_w4a8_int.py", "license": "Apache License 2.0", "lines": 129, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/lora/punica_wrapper/punica_xpu.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Based on: Chen, L., Ye, Z., Wu, Y., Zhuo, D., Ceze, L., & Krishnamurthy, A. (2023). Punica: Multi-Tenant LoRA Serving. https://arxiv.org/abs/2310.18547 """ from typing import final import torch from vllm import _custom_ops as ops from vllm.lora.layers import LoRAMapping from vllm.lora.ops.xpu_ops import bgmv_expand, bgmv_expand_slice, bgmv_shrink from vllm.triton_utils import HAS_TRITON, triton from vllm.utils.math_utils import round_up if HAS_TRITON: from vllm.lora.ops.triton_ops import ( LoRAKernelMeta, fused_moe_lora, ) from .punica_base import PunicaWrapperBase @final class PunicaWrapperXPU(PunicaWrapperBase): """ PunicaWrapperXPU is designed to manage and provide metadata for the punica kernel. The main function is to maintain the state information for Multi-LoRA, and to provide the interface for the punica ipex kernel. """ def __init__( self, max_num_batched_tokens: int, max_batches: int, device: torch.device | str, **kwargs, ): PunicaWrapperBase.__init__(self, max_num_batched_tokens, max_batches, device) torch._dynamo.mark_dynamic(self._token_lora_indices, 0) torch._dynamo.mark_dynamic(self._embeddings_indices, 1) torch._dynamo.mark_dynamic(self._sampler_indices_padded, 0) self.lora_config = kwargs["lora_config"] self.max_loras = self.lora_config.max_loras self.token_mapping_meta = LoRAKernelMeta.make( self.max_loras, max_num_batched_tokens, device=device ) def update_metadata( self, mapping: LoRAMapping, lora_index_to_id: list[int | None], max_loras: int, vocab_size: int, **kwargs, ): self.is_prefill = mapping.is_prefill self._update_base_metadata(mapping, lora_index_to_id, max_loras, vocab_size) def _get_token_lora_indices(self, x: torch.Tensor) -> torch.IntTensor: return torch.narrow(self._token_lora_indices, 0, 0, x.size(0)) def _apply_shrink( self, y: torch.Tensor, x: torch.Tensor, w_t_all: torch.Tensor, scale: float, ): bgmv_shrink(x, w_t_all, y, self._get_token_lora_indices(x), scale) def _apply_expand( self, y: torch.Tensor, x: torch.Tensor, w_t_all: torch.Tensor, y_offset: int, y_slice_size: int, add_inputs: bool, ): token_lora_indices = self._get_token_lora_indices(x) bgmv_expand_slice( x, w_t_all, y, token_lora_indices, y_offset, y_slice_size, add_inputs ) def add_shrink( self, y: torch.Tensor, x: torch.Tensor, lora_a_stacked: tuple[torch.Tensor, ...], scale: float, **kwargs, ): """ Performs GEMM for multiple slices of lora_a. Semantics: for i in range(len(lora_a_stacked)): y[i] += (x @ lora_a_stacked[i]) * scale Args: y (torch.Tensor): Output tensors x (torch.Tensor): Input tensor lora_a_stacked (tuple[torch.Tensor, ...]): lora_a's weights scale (float): Scaling factor for the operation """ x = x.view(-1, x.shape[-1]) for slice_idx in range(len(lora_a_stacked)): self._apply_shrink(y[slice_idx], x, lora_a_stacked[slice_idx], scale) def add_expand( self, y: torch.Tensor, x: torch.Tensor, lora_b_stacked: tuple[torch.Tensor, ...], output_slices: tuple[int, ...], offset_start: int = 0, add_inputs=True, **kwargs, ) -> None: """ Performs GEMM for multiple slices of lora_b. Semantics: for i in range(len(lora_b_stacked)): slice = output_slices[i] y[:, offset:offset+slice] += x[i] @ lora_b_stacked[i] offset += slice Args: y (torch.Tensor): Output tensor. x (torch.Tensor): Input tensors lora_b_stacked (tuple[torch.Tensor, ...]): lora_b's weight output_slices (tuple[int, ...]): Every slice's size add_inputs (bool): Defaults to True. """ y_org = y y = y.view(-1, y.shape[-1]) assert x.ndim == 3 assert x.size(0) == len(output_slices) # TODO fuse these kernels for slice_idx in range(len(lora_b_stacked)): self._apply_expand( y, x[slice_idx], lora_b_stacked[slice_idx], offset_start, output_slices[slice_idx], add_inputs=add_inputs, ) offset_start += output_slices[slice_idx] y.view_as(y_org) def add_lora_embedding( self, y: torch.Tensor, x: torch.Tensor, lora_b_stacked: torch.Tensor, add_inputs: bool = True, **kwargs, ) -> None: """ Applies lora specifically for VocabParallelEmbeddingWithLoRA. Semantics: y += x @ lora_b_stacked Args: y (torch.Tensor): Output tensor. x (torch.Tensor): Input tensor. lora_b_stacked (torch.Tensor): lora_b's weights. add_inputs (bool): Default to True. """ token_lora_indices = self._get_token_lora_indices(x) bgmv_expand(x, lora_b_stacked, y, token_lora_indices, add_inputs) def add_lora_linear( self, y: torch.Tensor, x: torch.Tensor, lora_a_stacked: tuple[torch.Tensor, ...], lora_b_stacked: tuple[torch.Tensor, ...], scale: float, output_slices: tuple[int, ...], *, buffer: torch.Tensor | None = None, **kwargs, ) -> None: """ Applicable to linear-related lora. Semantics: for i in range(len(lora_a_stacked)): y[i] += ( x[i].unsqueeze(0) @ lora_a_stacked[indices[i], layer_idx, :, :] @ lora_b_stacked[indices[i], layer_idx, :, :] * scale ).squeeze(0) Args: y (torch.Tensor): Output tensor. Will be changed in-place. x (torch.Tensor): Input tensor lora_a_stacked (tuple[torch.Tensor, ...]): lora_a's weight. lora_b_stacked (tuple[torch.Tensor, ...]): lora_b's weight. scale (float): Scaling factor. output_slices (tuple[int, ...]): Every slice's size. buffer (Optional[torch.Tensor]): Defaults to None. """ assert len(lora_a_stacked) == len(lora_b_stacked) == len(output_slices) if buffer is None: r = lora_b_stacked[0].size(-1) buffer = torch.zeros( # type: ignore (len(output_slices), x.size(0), r), dtype=x.dtype, device=x.device, ) self.add_shrink( buffer, # type: ignore x, lora_a_stacked, scale, **kwargs, ) self.add_expand( y, buffer, # type: ignore lora_b_stacked, output_slices, add_inputs=True, **kwargs, ) @property def sampler_indices_padded(self) -> torch.Tensor: """ This property provides access to padded sampler indices. """ return self._sampler_indices_padded[:] def add_lora_logits( self, y: torch.Tensor, x: torch.Tensor, lora_a_stacked: torch.Tensor, lora_b_stacked: torch.Tensor, scale, *, buffer: torch.Tensor | None = None, **kwargs, ) -> None: """ Applies lora specifically for LogitsProcessorWithLoRA. Semantics: buffer = (x @ lora_a_stacked) * scale y += buffer @ lora_b_stacked Args: y (torch.Tensor): Output tensor. x (torch.Tensor): Input tensor. lora_a_stacked (torch.Tensor): lora_a's weights. lora_b_stacked (torch.Tensor): lora_b's weights. scale (float): Scaling factor. buffer (Optional[torch.Tensor]): Default to None. """ y_org = y y = y.view(-1, y.shape[-1]) x = x.view(-1, x.shape[-1]) r = lora_b_stacked.size(-1) if buffer is None: buffer = torch.zeros((x.size(0), r), dtype=x.dtype, device=x.device) sampler_indices = torch.narrow(self._sampler_indices, 0, 0, x.size(0)) bgmv_shrink(x, lora_a_stacked, buffer, sampler_indices, scale) bgmv_expand(buffer, lora_b_stacked, y, sampler_indices, add_inputs=True) return y.view_as(y_org) def moe_lora_align_block_size( self, topk_ids: torch.Tensor, num_tokens: int, block_size: int, num_experts: int, max_loras: int, adapter_enabled: torch.Tensor, expert_map: torch.Tensor | None = None, pad_sorted_ids: bool = False, naive_block_assignment: bool = False, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: """ Aligns tokens and experts into block-sized chunks for LoRA-based mixture-of-experts (MoE) execution. """ (token_lora_mapping, _, _, _, lora_ids, _, _) = ( self.token_mapping_meta.meta_args( num_tokens, self.lora_config.specialize_active_lora ) ) if naive_block_assignment: expert_ids = topk_ids.reshape(-1) sorted_ids = None num_tokens_post_pad = None else: max_num_tokens_padded = topk_ids.numel() + num_experts * (block_size - 1) if pad_sorted_ids: max_num_tokens_padded = round_up(max_num_tokens_padded, block_size) sorted_ids = torch.empty( (max_loras * max_num_tokens_padded,), dtype=torch.int32, device=topk_ids.device, ) max_num_m_blocks = triton.cdiv(max_num_tokens_padded, block_size) # Expert ids must be set default to -1 to prevent a blank block expert_ids = torch.empty( (max_loras * max_num_m_blocks,), dtype=torch.int32, device=topk_ids.device, ) num_tokens_post_pad = torch.empty( (max_loras), dtype=torch.int32, device=topk_ids.device ) ops.moe_lora_align_block_size( topk_ids, token_lora_mapping, num_experts, block_size, max_loras, max_num_tokens_padded, max_num_m_blocks, sorted_ids, expert_ids, num_tokens_post_pad, adapter_enabled, lora_ids, ) if expert_map is not None: expert_ids = expert_map[expert_ids] return None, sorted_ids, expert_ids, num_tokens_post_pad def add_lora_fused_moe( self, y: torch.Tensor, x: torch.Tensor, lora_a_stacked: tuple[torch.Tensor, ...], lora_b_stacked: tuple[torch.Tensor, ...], topk_weights: torch.Tensor, sorted_token_ids: torch.Tensor | None, expert_ids: torch.Tensor, num_tokens_post_padded: torch.Tensor | None, max_lora_rank: int, top_k_num: int, shrink_config, expand_config, adapter_enabled: torch.Tensor, mul_routed_weight=False, fully_sharded: bool = False, offset: int = 0, token_lora_mapping: torch.Tensor | None = None, ): """ Performs a fused forward computation for LoRA of Mixture-of-Experts (MoE) layer. """ ( token_lora_mapping_meta, _, _, _, lora_ids, _, num_active_loras, ) = self.token_mapping_meta.meta_args( x.size(0), self.lora_config.specialize_active_lora ) if token_lora_mapping is None: token_lora_mapping = token_lora_mapping_meta fused_moe_lora( y, x, lora_a_stacked, lora_b_stacked, topk_weights, sorted_token_ids, expert_ids, num_tokens_post_padded, token_lora_mapping, max_lora_rank, top_k_num, lora_ids, num_active_loras, adapter_enabled, shrink_config.get("BLOCK_SIZE_M", 64), shrink_config.get("BLOCK_SIZE_N", 64), shrink_config.get("BLOCK_SIZE_K", 32), shrink_config.get("GROUP_SIZE_M", 8), shrink_config.get("NUM_WARPS", 4), shrink_config.get("NUM_STAGES", 3), shrink_config.get("SPLIT_K", 1), expand_config.get("BLOCK_SIZE_M", 64), expand_config.get("BLOCK_SIZE_N", 64), expand_config.get("BLOCK_SIZE_K", 32), expand_config.get("GROUP_SIZE_M", 8), expand_config.get("NUM_WARPS", 4), expand_config.get("NUM_STAGES", 3), expand_config.get("SPLIT_K", 1), mul_routed_weight, fully_sharded, offset, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/lora/punica_wrapper/punica_xpu.py", "license": "Apache License 2.0", "lines": 383, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/entrypoints/openai/tool_parsers/test_llama3_json_tool_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from unittest.mock import MagicMock, patch import pytest from vllm.entrypoints.openai.engine.protocol import ExtractedToolCallInformation from vllm.tokenizers import TokenizerLike from vllm.tool_parsers.llama_tool_parser import Llama3JsonToolParser @pytest.fixture def parser(default_tokenizer: TokenizerLike): return Llama3JsonToolParser(default_tokenizer) def test_extract_tool_calls_simple(parser): # Test with a simple tool call model_output = ( 'Here is the result: {"name": "getOpenIncidentsTool", ' '"parameters": {}} Would you like to know more?' ) result = parser.extract_tool_calls(model_output, None) assert isinstance(result, ExtractedToolCallInformation) assert result.tools_called is True assert len(result.tool_calls) == 1 assert result.tool_calls[0].type == "function" assert result.tool_calls[0].function.name == "getOpenIncidentsTool" assert result.tool_calls[0].function.arguments == "{}" assert result.content is None def test_extract_tool_calls_with_arguments(parser): # Test with a tool call that has arguments model_output = ( '{"name": "searchTool", "parameters": {"query": "test query", "limit": 10}}' ) result = parser.extract_tool_calls(model_output, None) assert result.tools_called is True assert len(result.tool_calls) == 1 assert result.tool_calls[0].function.name == "searchTool" assert '"query": "test query"' in result.tool_calls[0].function.arguments assert '"limit": 10' in result.tool_calls[0].function.arguments def test_extract_tool_calls_no_json(parser): # Test with text that doesn't contain a JSON object model_output = "This is just some text without any tool calls" result = parser.extract_tool_calls(model_output, None) assert result.tools_called is False assert len(result.tool_calls) == 0 assert result.content == model_output def test_extract_tool_calls_invalid_json(parser): # Test with invalid JSON model_output = '{"name": "invalidTool", "parameters": {invalid json}' result = parser.extract_tool_calls(model_output, None) assert result.tools_called is False assert len(result.tool_calls) == 0 assert result.content == model_output def test_extract_tool_calls_with_arguments_key(parser): # Test with a tool call that uses "arguments" instead of "parameters" model_output = '{"name": "searchTool", "arguments": {"query": "test"}}' result = parser.extract_tool_calls(model_output, None) assert result.tools_called is True assert len(result.tool_calls) == 1 assert result.tool_calls[0].function.name == "searchTool" assert '"query": "test"' in result.tool_calls[0].function.arguments def test_extract_tool_calls_multiple_json(parser): # Test with multiple JSONs separated by semicolons model_output = ( '{"name": "searchTool", "parameters": {"query": "test1"}}; ' '{"name": "getOpenIncidentsTool", "parameters": {}}; ' '{"name": "searchTool", "parameters": {"query": "test2"}}' ) result = parser.extract_tool_calls(model_output, None) assert result.tools_called is True assert len(result.tool_calls) == 3 # Check first tool call assert result.tool_calls[0].function.name == "searchTool" assert '"query": "test1"' in result.tool_calls[0].function.arguments # Check second tool call assert result.tool_calls[1].function.name == "getOpenIncidentsTool" assert result.tool_calls[1].function.arguments == "{}" # Check third tool call assert result.tool_calls[2].function.name == "searchTool" assert '"query": "test2"' in result.tool_calls[2].function.arguments def test_extract_tool_calls_multiple_json_with_whitespace(parser): # Test with multiple JSONs separated by semicolons and extra whitespace model_output = ( '{"name": "searchTool", "parameters": {"query": "test1"}} ; ' '{"name": "getOpenIncidentsTool", "parameters": {}} ; ' '{"name": "searchTool", "parameters": {"query": "test2"}}' ) result = parser.extract_tool_calls(model_output, None) assert result.tools_called is True assert len(result.tool_calls) == 3 assert result.tool_calls[0].function.name == "searchTool" assert result.tool_calls[1].function.name == "getOpenIncidentsTool" assert result.tool_calls[2].function.name == "searchTool" def test_extract_tool_calls_multiple_json_with_surrounding_text(parser): # Test with multiple JSONs and surrounding text model_output = ( "Here are the results: " '{"name": "searchTool", "parameters": {"query": "test1"}}; ' '{"name": "getOpenIncidentsTool", "parameters": {}}; ' '{"name": "searchTool", "parameters": {"query": "test2"}} ' "Would you like to know more?" ) result = parser.extract_tool_calls(model_output, None) assert result.tools_called is True assert len(result.tool_calls) == 3 assert result.tool_calls[0].function.name == "searchTool" assert result.tool_calls[1].function.name == "getOpenIncidentsTool" assert result.tool_calls[2].function.name == "searchTool" def test_extract_tool_calls_deeply_nested_json(parser): # Test with deeply nested JSON parameters (5 levels) model_output = ( '{"name": "complexTool", ' '"parameters": {' '"level1": {' '"level2": {' '"level3": {' '"level4": {' '"value": "deep"' "}}}}}}" ) result = parser.extract_tool_calls(model_output, None) assert result.tools_called is True assert len(result.tool_calls) == 1 assert result.tool_calls[0].function.name == "complexTool" # Verify the nested structure is preserved in the arguments import json args = json.loads(result.tool_calls[0].function.arguments) assert args["level1"]["level2"]["level3"]["level4"]["value"] == "deep" def test_extract_tool_calls_multiple_with_deep_nesting(parser): # Test with multiple tool calls where some have deeply nested parameters model_output = ( '{"name": "simpleTool", "parameters": {"value": "test"}}; ' '{"name": "complexTool", "parameters": ' '{"config": {"database": {"connection": {"pool": {"size": 10}}}}}}' ) result = parser.extract_tool_calls(model_output, None) assert result.tools_called is True assert len(result.tool_calls) == 2 # Check first tool call assert result.tool_calls[0].function.name == "simpleTool" import json args0 = json.loads(result.tool_calls[0].function.arguments) assert args0["value"] == "test" # Check second tool call with deep nesting assert result.tool_calls[1].function.name == "complexTool" args1 = json.loads(result.tool_calls[1].function.arguments) assert args1["config"]["database"]["connection"]["pool"]["size"] == 10 def test_extract_tool_calls_with_quotes_and_brackets_in_string(parser): # Test with quotes and brackets inside quoted string values model_output = ( '{"name": "searchTool", ' '"parameters": {' '"query": "test {value} [complex]",' '"nested": {"inner": "more {brackets}"}' "}}" ) result = parser.extract_tool_calls(model_output, None) assert result.tools_called is True assert len(result.tool_calls) == 1 assert result.tool_calls[0].function.name == "searchTool" # Verify the string values are preserved including brackets and quotes import json args = json.loads(result.tool_calls[0].function.arguments) assert args["query"] == "test {value} [complex]" assert args["nested"]["inner"] == "more {brackets}" def test_extract_tool_calls_with_escaped_quotes_in_nested_json(parser): # Test with escaped quotes in deeply nested JSON model_output = ( '{"name": "parserTool", "parameters": {"text": "He said \\"Hello {world}\\""}}' ) result = parser.extract_tool_calls(model_output, None) assert result.tools_called is True assert len(result.tool_calls) == 1 assert result.tool_calls[0].function.name == "parserTool" # Verify escaped quotes are preserved import json args = json.loads(result.tool_calls[0].function.arguments) assert args["text"] == 'He said "Hello {world}"' def test_extract_tool_calls_missing_name_key(parser): # Test that missing "name" key returns content model_output = '{"parameters": {}}' result = parser.extract_tool_calls(model_output, None) assert result.tools_called is False assert len(result.tool_calls) == 0 assert result.content == model_output def test_extract_tool_calls_missing_parameters_and_arguments_key(parser): # Test that missing both "parameters" and "arguments" keys returns content model_output = '{"name": "toolWithoutParams"}' result = parser.extract_tool_calls(model_output, None) assert result.tools_called is False assert len(result.tool_calls) == 0 assert result.content == model_output def test_regex_timeout_handling(parser): """Test regex timeout is handled gracefully""" fake_problematic_input = "{hello world[A(A=" + "\t)A(A=,\t" * 2 # create a mock regex that raises TimeoutError mock_regex = MagicMock() mock_regex.finditer.side_effect = TimeoutError("Regex timeout") with patch.object(parser, "tool_call_start_regex", mock_regex): result = parser.extract_tool_calls(fake_problematic_input, None) # should treat as regular text when regex times out assert result.content == fake_problematic_input assert result.tools_called is False assert len(result.tool_calls) == 0 mock_regex.finditer.assert_called_once()
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/tool_parsers/test_llama3_json_tool_parser.py", "license": "Apache License 2.0", "lines": 201, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/transformers_utils/dynamic_module.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os from transformers.dynamic_module_utils import ( get_class_from_dynamic_module, resolve_trust_remote_code, ) import vllm.envs as envs from vllm.logger import init_logger logger = init_logger(__name__) def try_get_class_from_dynamic_module( class_reference: str, pretrained_model_name_or_path: str, trust_remote_code: bool, cache_dir: str | os.PathLike | None = None, force_download: bool = False, resume_download: bool | None = None, proxies: dict[str, str] | None = None, token: bool | str | None = None, revision: str | None = None, local_files_only: bool = False, repo_type: str | None = None, code_revision: str | None = None, warn_on_fail: bool = True, **kwargs, ) -> type | None: """ As `transformers.dynamic_module_utils.get_class_from_dynamic_module`, but ignoring any errors. """ try: resolve_trust_remote_code( trust_remote_code, pretrained_model_name_or_path, has_local_code=False, has_remote_code=True, ) return get_class_from_dynamic_module( class_reference, pretrained_model_name_or_path, cache_dir=cache_dir, force_download=force_download, resume_download=resume_download, proxies=proxies, token=token, revision=revision, local_files_only=local_files_only, repo_type=repo_type, code_revision=code_revision, **kwargs, ) except Exception: location = "ModelScope" if envs.VLLM_USE_MODELSCOPE else "HF Hub" if warn_on_fail: logger.warning( "Unable to load %s from %s on %s.", class_reference, pretrained_model_name_or_path, location, exc_info=True, ) return None
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/dynamic_module.py", "license": "Apache License 2.0", "lines": 62, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/interns1.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # -------------------------------------------------------- # InternS1 # Copyright (c) 2025 Shanghai AI Lab # Licensed under The MIT License [see LICENSE for details] # -------------------------------------------------------- from collections.abc import Iterable, Mapping, Sequence from typing import Annotated, Literal, TypeAlias import regex as re import torch import torch.nn as nn from transformers import BatchFeature, InternVLProcessor, PretrainedConfig from transformers.activations import ACT2FN from transformers.models.got_ocr2.image_processing_got_ocr2_fast import ( GotOcr2ImageProcessorFast, ) from transformers.models.internvl.video_processing_internvl import ( InternVLVideoProcessor, ) from vllm.config import VllmConfig from vllm.config.multimodal import BaseDummyOptions from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.models.interns1_vit import InternS1VisionModel from vllm.model_executor.models.module_mapping import MultiModelKeys from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import ( MultiModalDataDict, MultiModalFieldConfig, MultiModalKwargsItems, ) from vllm.multimodal.parse import ( ImageEmbeddingItems, ImageProcessorItems, ImageSize, MultiModalDataItems, ) from vllm.multimodal.processing import ( BaseDummyInputsBuilder, BaseMultiModalProcessor, BaseProcessingInfo, PromptReplacement, PromptUpdate, PromptUpdateDetails, ) from vllm.sequence import IntermediateTensors from vllm.transformers_utils.processor import cached_video_processor_from_config from vllm.utils.tensor_schema import TensorSchema, TensorShape from .interfaces import ( MultiModalEmbeddings, SupportsLoRA, SupportsMultiModal, SupportsPP, ) from .utils import ( AutoWeightsLoader, WeightsMapper, init_vllm_registered_model, maybe_prefix, ) class InternS1MultiModalProjector(nn.Module): def __init__(self, config): super().__init__() self.layer_norm = nn.LayerNorm( config.vision_config.hidden_size * int(1 / config.downsample_ratio) ** 2 ) self.linear_1 = nn.Linear( config.vision_config.hidden_size * int(1 / config.downsample_ratio) ** 2, config.text_config.hidden_size, ) self.act = ACT2FN[config.projector_hidden_act] self.linear_2 = nn.Linear( config.text_config.hidden_size, config.text_config.hidden_size ) def forward(self, image_features): hidden_states = self.layer_norm(image_features) hidden_states = self.linear_1(hidden_states) hidden_states = self.act(hidden_states) hidden_states = self.linear_2(hidden_states) return hidden_states class InternS1ImagePixelInputs(TensorSchema): """ Dimensions: - bnp: Batch size * number of images * (1 + num_patches) - c: Number of channels (3) - h: Height - w: Width - bn: Batch size * number of images """ type: Literal["pixel_values"] = "pixel_values" pixel_values: Annotated[torch.Tensor, TensorShape("bnp", 3, "h", "w")] num_patches: Annotated[torch.Tensor, TensorShape("bn")] class InternS1ImageEmbeddingInputs(TensorSchema): """ Dimensions: - ni: Number of images - tifs: Total image feature size - hs: Hidden size (must match language model backbone) """ type: Literal["image_embeds"] = "image_embeds" data: Annotated[torch.Tensor | list[torch.Tensor], TensorShape("ni", "tifs", "hs")] InternS1ImageInputs: TypeAlias = InternS1ImagePixelInputs | InternS1ImageEmbeddingInputs class InternS1VideoPixelInputs(TensorSchema): """ Dimensions: - bnv: Batch size * number of videos * number of frames - bn: Batch size * number of images - c: Number of channels (3) - h: Height - w: Width """ type: Literal["pixel_values_videos"] = "pixel_values_videos" pixel_values: Annotated[torch.Tensor, TensorShape("bnv", 3, "h", "w")] num_patches: Annotated[torch.Tensor, TensorShape("bn")] class InternS1VideoEmbeddingInputs(TensorSchema): """ Dimensions: - nv: Number of videos - tvfs: Total video feature size - hs: Hidden size (must match language model backbone) """ type: Literal["video_embeds"] = "video_embeds" data: Annotated[torch.Tensor | list[torch.Tensor], TensorShape("nv", "tvfs", "hs")] InternS1VideoInputs: TypeAlias = InternS1VideoPixelInputs | InternS1VideoEmbeddingInputs def resolve_interns1_min_max_num( min_dynamic_patch: int, max_dynamic_patch: int, dynamic_image_size: bool, use_thumbnail: bool, ) -> tuple[int, int]: min_dynamic_patch = min_dynamic_patch if dynamic_image_size else 1 max_dynamic_patch = max_dynamic_patch if dynamic_image_size else 1 if use_thumbnail and max_dynamic_patch != 1: max_dynamic_patch += 1 return min_dynamic_patch, max_dynamic_patch def get_interns1_target_ratios( min_num: int, max_num: int, ) -> list[tuple[int, int]]: target_ratios = { (i, j) for n in range(min_num, max_num + 1) for i in range(1, n + 1) for j in range(1, n + 1) if min_num <= i * j <= max_num } return sorted(target_ratios, key=lambda x: x[0] * x[1]) class InternS1ProcessingInfo(BaseProcessingInfo): """ProcessingInfo for InternS1-style models.""" def get_hf_processor(self, **kwargs: object) -> InternVLProcessor: hf_processor = self.ctx.get_hf_processor(InternVLProcessor, **kwargs) hf_processor.video_processor = cached_video_processor_from_config( self.ctx.model_config, processor_cls=InternVLVideoProcessor, size=hf_processor.image_processor.size, **kwargs, ) return hf_processor def get_supported_mm_limits(self) -> Mapping[str, int | None]: return {"image": None, "video": None} def get_num_image_tokens( self, *, image_width: int, image_height: int, processor: InternVLProcessor, mm_kwargs: Mapping[str, object], ) -> int: image_processor: GotOcr2ImageProcessorFast = processor.image_processor num_image_patches = image_processor.get_number_of_image_patches( image_height, image_width, self.ctx.get_merged_mm_kwargs(mm_kwargs), ) return processor.image_seq_length * num_image_patches def resolve_target_ratios(self, use_thumbnail: bool | None = None): image_processor = self.get_hf_processor().image_processor min_dynamic_patch = image_processor.min_patches max_dynamic_patch = image_processor.max_patches # HF format's InternVL processor uses `crop_to_patches` which is # equivalent to `use_thumbnail` in original format. use_thumbnail = image_processor.crop_to_patches dynamic_image_size = True min_num, max_num = resolve_interns1_min_max_num( min_dynamic_patch, max_dynamic_patch, dynamic_image_size, use_thumbnail=use_thumbnail, ) return get_interns1_target_ratios(min_num, max_num) def get_image_size_with_most_features(self) -> ImageSize: processor = self.get_hf_processor() hf_config = self.ctx.get_hf_config() base_height, base_width = hf_config.vision_config.image_size target_ratios = self.resolve_target_ratios() largest_feature_size, largest_feature_pinpoint = 0, None for wr, hr in target_ratios: width, height = base_width * wr, base_height * hr feat_size = self.get_num_image_tokens( image_width=width, image_height=height, processor=processor, mm_kwargs={}, ) if feat_size > largest_feature_size: largest_feature_size = feat_size largest_feature_pinpoint = ImageSize(width=width, height=height) assert not (largest_feature_size == 0 or largest_feature_pinpoint is None), ( "Cannot have a largest feature size of 0!" ) return largest_feature_pinpoint def get_max_image_tokens(self) -> int: processor = self.get_hf_processor() target_width, target_height = self.get_image_size_with_most_features() return self.get_num_image_tokens( image_width=target_width, image_height=target_height, processor=processor, mm_kwargs={}, ) def get_num_frames_with_most_features( self, seq_len: int, mm_counts: Mapping[str, int], ) -> int: max_images = mm_counts.get("image", 0) max_videos = mm_counts.get("video", 0) processor = self.get_hf_processor() max_image_tokens = self.get_max_image_tokens() * max_images max_total_frames = (seq_len - max_image_tokens) // processor.image_seq_length max_frames_per_video = max_total_frames // max(max_videos, 1) return max(max_frames_per_video, 1) class InternS1DummyInputsBuilder(BaseDummyInputsBuilder[InternS1ProcessingInfo]): """DummyInputsBuilder for InternS1-style models.""" def get_dummy_text(self, mm_counts: Mapping[str, int]) -> str: num_images = mm_counts.get("image", 0) num_videos = mm_counts.get("video", 0) image_token = self.info.get_hf_processor().image_token video_token = self.info.get_hf_processor().video_token return image_token * num_images + video_token * num_videos def get_dummy_mm_data( self, seq_len: int, mm_counts: Mapping[str, int], mm_options: Mapping[str, BaseDummyOptions], ) -> MultiModalDataDict: target_width, target_height = self.info.get_image_size_with_most_features() target_num_frames = self.info.get_num_frames_with_most_features( seq_len, mm_counts ) num_images = mm_counts.get("image", 0) num_videos = mm_counts.get("video", 0) config = self.info.get_hf_config() image_size_h, image_size_w = config.vision_config.image_size image_overrides = mm_options.get("image") video_overrides = mm_options.get("video") return { "image": self._get_dummy_images( width=target_width, height=target_height, num_images=num_images, overrides=image_overrides, ), "video": self._get_dummy_videos( width=image_size_w, height=image_size_h, num_frames=target_num_frames, num_videos=num_videos, overrides=video_overrides, ), } class InternS1MultiModalProcessor(BaseMultiModalProcessor[InternS1ProcessingInfo]): """Basic image-only MultiModalProcessor for InternS1-style models.""" def _call_hf_processor( self, prompt: str, mm_data: Mapping[str, object], mm_kwargs: Mapping[str, object], tok_kwargs: Mapping[str, object], ) -> BatchFeature: mm_data = dict(mm_data) videos = mm_data.pop("videos", []) images = mm_data.pop("images", []) assert isinstance(videos, list) assert isinstance(images, list) hf_processor = self.info.get_hf_processor(**mm_kwargs) tokenizer = hf_processor.tokenizer video_token_id = tokenizer.encode( hf_processor.video_token, add_special_tokens=False ) assert len(video_token_id) == 1 video_token_id = video_token_id[0] prompt = re.sub(hf_processor.image_token, "<image_placeholder>", prompt) prompt = re.sub(hf_processor.video_token, "<video_placeholder>", prompt) image_outputs = {} if images: image_pixel_values = [] for image in images: processed_outputs = super()._call_hf_processor( prompt=hf_processor.image_token, mm_data={"images": image}, mm_kwargs=mm_kwargs, tok_kwargs=tok_kwargs, ) image_pixel_values.append(processed_outputs.pop("pixel_values")) input_ids = processed_outputs.pop("input_ids") image_placeholder = tokenizer.batch_decode(input_ids)[0] prompt = prompt.replace("<image_placeholder>", image_placeholder, 1) num_patches = [len(item) for item in image_pixel_values] image_outputs = { "pixel_values": torch.concat(image_pixel_values), "image_num_patches": torch.tensor(num_patches), "image_token_id": torch.tensor(hf_processor.image_token_id), } video_outputs = {} if videos: video_pixel_values = [] for video in videos: processed_outputs = super()._call_hf_processor( prompt=hf_processor.video_token, mm_data={"videos": video}, mm_kwargs=mm_kwargs, tok_kwargs=tok_kwargs, ) video_pixel_values.append(processed_outputs.pop("pixel_values")) input_ids = processed_outputs.pop("input_ids") input_ids[input_ids == hf_processor.image_token_id] = video_token_id video_placeholder = tokenizer.batch_decode(input_ids)[0] prompt = prompt.replace("<video_placeholder>", video_placeholder, 1) num_frames = [len(item) for item in video_pixel_values] video_outputs = { "pixel_values_videos": torch.concat(video_pixel_values), "video_num_patches": torch.tensor(num_frames), "video_token_id": torch.tensor(video_token_id), } prompt = re.sub("<image_placeholder>", hf_processor.image_token, prompt) prompt = re.sub("<video_placeholder>", hf_processor.video_token, prompt) text_outputs = tokenizer(prompt, **tok_kwargs, return_tensors="pt") return BatchFeature({**text_outputs, **image_outputs, **video_outputs}) def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: image_num_patches = hf_inputs.get("image_num_patches", torch.empty(0)) video_num_patches = hf_inputs.get("video_num_patches", torch.empty(0)) num_images = len(image_num_patches) num_videos = len(video_num_patches) return dict( pixel_values=MultiModalFieldConfig.flat_from_sizes( "image", image_num_patches ), image_num_patches=MultiModalFieldConfig.batched("image"), image_embeds=MultiModalFieldConfig.batched("image"), image_token_id=MultiModalFieldConfig.shared("image", num_images), pixel_values_videos=MultiModalFieldConfig.flat_from_sizes( "video", video_num_patches ), video_num_patches=MultiModalFieldConfig.batched("video"), video_token_id=MultiModalFieldConfig.shared("video", num_videos), ) def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, object], out_mm_kwargs: MultiModalKwargsItems, ) -> Sequence[PromptUpdate]: hf_processor = self.info.get_hf_processor(**hf_processor_mm_kwargs) img_context_token = hf_processor.image_token start_image_token = hf_processor.start_image_token end_image_token = hf_processor.end_image_token video_token = hf_processor.video_token out_mm_data = out_mm_kwargs.get_data() if "video_num_patches" in out_mm_data: video_num_patches = out_mm_data["video_num_patches"] assert isinstance(video_num_patches, torch.Tensor) video_num_patches = video_num_patches.tolist() else: video_num_patches = [] if "image_num_patches" in out_mm_data: image_num_patches = out_mm_data["image_num_patches"] assert isinstance(image_num_patches, torch.Tensor) image_num_patches = image_num_patches.tolist() else: image_num_patches = [] def get_replacement_interns1_image(item_idx: int): images = mm_items.get_items( "image", (ImageEmbeddingItems, ImageProcessorItems) ) if isinstance(images, ImageEmbeddingItems): feature_size = images.get_feature_size(item_idx) else: num_patches = image_num_patches[item_idx] feature_size = num_patches * hf_processor.image_seq_length repl_features = img_context_token * feature_size repl_full = start_image_token + repl_features + end_image_token return PromptUpdateDetails.select_text(repl_full, img_context_token) def get_replacement_interns1_video(item_idx: int): num_patches = video_num_patches[item_idx] repl_features = video_token * hf_processor.image_seq_length repl_features_with_sep = start_image_token + repl_features + end_image_token # num_patches is equal to num_frames repl_full = "\n".join( [f"Frame{i + 1}: {repl_features_with_sep}" for i in range(num_patches)] ) return PromptUpdateDetails.select_text(repl_full, video_token) return [ PromptReplacement( modality="image", target=img_context_token, replacement=get_replacement_interns1_image, ), PromptReplacement( modality="video", target=video_token, replacement=get_replacement_interns1_video, ), ] @MULTIMODAL_REGISTRY.register_processor( InternS1MultiModalProcessor, info=InternS1ProcessingInfo, dummy_inputs=InternS1DummyInputsBuilder, ) class InternS1ForConditionalGeneration( nn.Module, SupportsMultiModal, SupportsPP, SupportsLoRA ): # To ensure correct weight loading and mapping. hf_to_vllm_mapper = WeightsMapper( orig_to_new_prefix={ "lm_head.": "language_model.lm_head.", "model.language_model.": "language_model.model.", "model.vision_tower.": "vision_tower.", "model.multi_modal_projector.": "multi_modal_projector.", } ) @classmethod def get_placeholder_str(cls, modality: str, i: int) -> str | None: # transformers InternVLProcessor uses <IMG_CONTEXT> as the separator # refer to https://github.com/huggingface/transformers/blob/f90de364c2484c7c325bbe05befdcf487bd75b63/src/transformers/models/internvl/processing_internvl.py#L116 if modality.startswith("image"): return "<IMG_CONTEXT>" if modality.startswith("video"): return "<video>" raise ValueError("Only image or video modality is supported") def __init__(self, *, vllm_config: VllmConfig, prefix: str = "") -> None: super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config multimodal_config = vllm_config.model_config.multimodal_config self.config = config self.multimodal_config = multimodal_config image_size = config.vision_config.image_size[0] patch_size = config.vision_config.patch_size[0] self.patch_size = patch_size self.num_image_token = int( (image_size // patch_size) ** 2 * (config.downsample_ratio**2) ) self.downsample_ratio = config.downsample_ratio with self._mark_tower_model(vllm_config, {"image", "video"}): self.vision_tower = self._init_vision_model( config, quant_config=quant_config, prefix=maybe_prefix(prefix, "vision_tower"), ) self.multi_modal_projector = self._init_mlp1(config) with self._mark_language_model(vllm_config): self.language_model = init_vllm_registered_model( vllm_config=vllm_config, hf_config=config.text_config, prefix=maybe_prefix(prefix, "language_model"), ) self.img_context_token_id = None self.video_context_token_id = None self.visual_token_mask = None self.make_empty_intermediate_tensors = ( self.language_model.make_empty_intermediate_tensors ) def _init_vision_model( self, config: PretrainedConfig, quant_config: QuantizationConfig | None, *, prefix: str, ): num_hidden_layers = config.vision_config.num_hidden_layers return InternS1VisionModel( config.vision_config, quant_config=quant_config, num_hidden_layers_override=num_hidden_layers, prefix=prefix, ) def _init_mlp1(self, config: PretrainedConfig) -> nn.Module: return InternS1MultiModalProjector(config) def pixel_shuffle(self, x, scale_factor=0.5): n, w, h, c = x.size() # N, W, H, C --> N, W, H * scale, C // scale x = x.view(n, w, int(h * scale_factor), int(c / scale_factor)) # N, W, H * scale, C // scale --> N, H * scale, W, C // scale x = x.permute(0, 2, 1, 3).contiguous() x = x.view( n, int(h * scale_factor), int(w * scale_factor), int(c / (scale_factor * scale_factor)), ) x = x.permute(0, 2, 1, 3).contiguous() return x def extract_feature(self, pixel_values: torch.Tensor) -> torch.Tensor: vit_embeds = self.vision_tower(pixel_values=pixel_values) vit_embeds = vit_embeds[:, 1:, :] h = w = int(vit_embeds.shape[1] ** 0.5) vit_embeds = vit_embeds.reshape(vit_embeds.shape[0], h, w, -1) vit_embeds = self.pixel_shuffle(vit_embeds, scale_factor=self.downsample_ratio) vit_embeds = vit_embeds.reshape(vit_embeds.shape[0], -1, vit_embeds.shape[-1]) vit_embeds = self.multi_modal_projector(vit_embeds) return vit_embeds def _parse_and_validate_image_input( self, **kwargs: object ) -> InternS1ImageInputs | None: pixel_values = kwargs.pop("pixel_values", None) image_num_patches = kwargs.pop("image_num_patches", None) image_embeds = kwargs.pop("image_embeds", None) if pixel_values is None and image_embeds is None: return None if image_embeds is not None: return InternS1ImageEmbeddingInputs( type="image_embeds", data=image_embeds, ) image_token_id = kwargs["image_token_id"] if isinstance(image_token_id, torch.Tensor): image_token_id = image_token_id.flatten().unique().item() assert isinstance(image_token_id, int) self.img_context_token_id = image_token_id if pixel_values is not None: h, w = self.config.vision_config.image_size return InternS1ImagePixelInputs( type="pixel_values", pixel_values=pixel_values, num_patches=image_num_patches, resolve_bindings={ "h": h, "w": w, }, ) raise AssertionError("This line should be unreachable.") def _parse_and_validate_video_input( self, **kwargs: object ) -> InternS1VideoInputs | None: pixel_values_flat_video = kwargs.pop("pixel_values_videos", None) video_num_patches = kwargs.pop("video_num_patches", None) video_embeds = kwargs.pop("video_embeds", None) if pixel_values_flat_video is None and video_embeds is None: return None if video_embeds is not None: return InternS1VideoEmbeddingInputs( type="video_embeds", data=video_embeds, ) video_token_id = kwargs["video_token_id"] if isinstance(video_token_id, torch.Tensor): video_token_id = video_token_id.flatten().unique().item() assert isinstance(video_token_id, int) self.video_context_token_id = video_token_id if pixel_values_flat_video is not None: h, w = self.config.vision_config.image_size return InternS1VideoPixelInputs( type="pixel_values_videos", num_patches=video_num_patches, pixel_values=pixel_values_flat_video, resolve_bindings={ "h": h, "w": w, }, ) raise AssertionError("This line should be unreachable.") def _process_vision_input( self, image_input: InternS1ImageInputs | InternS1VideoInputs, ) -> tuple[torch.Tensor, ...]: if ( image_input["type"] == "image_embeds" or image_input["type"] == "video_embeds" ): return image_input["data"] image_embeds = self.extract_feature(image_input["pixel_values"]) num_patches = image_input["num_patches"] # Only one image in the current batch if len(num_patches) == 1: return (image_embeds.view(-1, self.config.text_config.hidden_size),) # NOTE: Image embeddings are split into separate tensors for each image # by the size of each embedding. feature_size = image_embeds.shape[1] image_embeds = image_embeds.view(-1, self.config.text_config.hidden_size) image_feature_sizes = [ num_patches * feature_size for num_patches in num_patches ] return image_embeds.split(image_feature_sizes) def _parse_and_validate_multimodal_inputs(self, **kwargs: object) -> dict: modalities = {} # Preserve the order of modalities if there are multiple of them # from the order of kwargs. for input_key in kwargs: if ( input_key in ("pixel_values", "image_embeds") and "images" not in modalities ): modalities["images"] = self._parse_and_validate_image_input(**kwargs) if input_key in ("pixel_values_videos",) and "videos" not in modalities: modalities["videos"] = self._parse_and_validate_video_input(**kwargs) return modalities def _set_visual_token_mask(self, input_ids: torch.Tensor) -> None: self.visual_token_mask = None def embed_multimodal(self, **kwargs: object) -> MultiModalEmbeddings: modalities = self._parse_and_validate_multimodal_inputs(**kwargs) if not modalities: return [] # The result multimodal_embeddings is tuple of tensors, with each # tensor corresponding to a multimodal data item (image or video). multimodal_embeddings: tuple[torch.Tensor, ...] = () # NOTE: It is important to iterate over the keys in this dictionary # to preserve the order of the modalities. for modality in modalities: if modality == "images": image_input = modalities["images"] image_embeddings = self._process_vision_input(image_input) multimodal_embeddings += tuple(image_embeddings) if modality == "videos": video_input = modalities["videos"] video_embeddings = self._process_vision_input(video_input) multimodal_embeddings += tuple(video_embeddings) return multimodal_embeddings def embed_input_ids( self, input_ids: torch.Tensor, multimodal_embeddings: MultiModalEmbeddings | None = None, *, is_multimodal: torch.Tensor | None = None, handle_oov_mm_token: bool = False, ) -> torch.Tensor: if multimodal_embeddings is not None and len(multimodal_embeddings) > 0: self._set_visual_token_mask(input_ids) # This is to satisfy the type checker for each overload if multimodal_embeddings is None or is_multimodal is None: return super().embed_input_ids(input_ids) return super().embed_input_ids( input_ids, multimodal_embeddings=multimodal_embeddings, is_multimodal=is_multimodal, handle_oov_mm_token=handle_oov_mm_token, ) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs: object, ) -> IntermediateTensors: if intermediate_tensors is not None: inputs_embeds = None forward_kwargs = { "input_ids": input_ids, "positions": positions, "intermediate_tensors": intermediate_tensors, "inputs_embeds": inputs_embeds, } hidden_states = self.language_model.model(**forward_kwargs) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: return self.language_model.compute_logits(hidden_states) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader(self) return loader.load_weights(weights, mapper=self.hf_to_vllm_mapper) def get_mm_mapping(self) -> MultiModelKeys: """ Get the module prefix in multimodal models """ return MultiModelKeys.from_string_field( language_model="language_model", connector="multi_modal_projector", tower_model="vision_tower", )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/interns1.py", "license": "Apache License 2.0", "lines": 686, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/interns1_vit.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # adapted from https://huggingface.co/OpenGVLab/InternVL2-4B/blob/main/modeling_intern_vit.py # -------------------------------------------------------- # InternVL # Copyright (c) 2023 OpenGVLab # Licensed under The MIT License [see LICENSE for details] # -------------------------------------------------------- from collections.abc import Iterable import torch import torch.nn as nn from transformers import PretrainedConfig from transformers.utils import torch_int from vllm.model_executor.layers.activation import get_act_fn from vllm.model_executor.layers.attention import MMEncoderAttention from vllm.model_executor.layers.conv import Conv2dLayer from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ColumnParallelLinear, RowParallelLinear from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.model_loader.weight_utils import default_weight_loader NORM2FN = { "rms_norm": RMSNorm, "layer_norm": nn.LayerNorm, } class InternS1VisionPatchEmbeddings(nn.Module): def __init__(self, config): super().__init__() image_size, patch_size = config.image_size, config.patch_size num_channels, hidden_size = config.num_channels, config.hidden_size num_patches = (image_size[1] // patch_size[1]) * ( image_size[0] // patch_size[0] ) patch_shape = (image_size[0] // patch_size[0], image_size[1] // patch_size[1]) self.image_size = image_size self.patch_size = patch_size self.num_channels = num_channels self.num_patches = num_patches self.patch_shape = patch_shape self.projection = Conv2dLayer( num_channels, hidden_size, kernel_size=patch_size, stride=patch_size ) def forward(self, pixel_values: torch.Tensor) -> torch.Tensor: batch_size, num_channels, height, width = pixel_values.shape if num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values " "match with the one set in the configuration." ) embeddings = self.projection(pixel_values.to(self.projection.weight.dtype)) patch_height, patch_width = embeddings.shape[2], embeddings.shape[3] embeddings = embeddings.flatten(2).transpose(1, 2) return embeddings, (patch_height, patch_width) class InternS1VisionEmbeddings(nn.Module): def __init__(self, config: PretrainedConfig): super().__init__() self.config = config self.cls_token = nn.Parameter(torch.zeros(1, 1, config.hidden_size)) if config.use_mask_token: self.mask_token = nn.Parameter(torch.zeros(1, 1, config.hidden_size)) else: self.mask_token = None self.patch_embeddings = InternS1VisionPatchEmbeddings(config) self.patch_size = config.patch_size self.image_size = ( config.image_size if isinstance(config.image_size, Iterable) else (config.image_size, config.image_size) ) num_patches = self.patch_embeddings.num_patches if config.use_absolute_position_embeddings: self.position_embeddings = nn.Parameter( torch.zeros(1, num_patches + 1, config.hidden_size) ) else: self.position_embeddings = None def interpolate_pos_encoding( self, embeddings: torch.Tensor, height: int, width: int ) -> torch.Tensor: """ This method allows to interpolate the pre-trained position encodings, to be able to use the model on higher resolution images. This method is also adapted to support torch.jit tracing. Adapted from: - https://github.com/facebookresearch/dino/blob/de9ee3df6cf39fac952ab558447af1fa1365362a/vision_transformer.py#L174-L194, and - https://github.com/facebookresearch/dinov2/blob/e1277af2ba9496fbadf7aec6eba56e8d882d1e35/dinov2/models/vision_transformer.py#L179-L211 """ # noqa: E501 num_patches = embeddings.shape[1] - 1 num_positions = self.position_embeddings.shape[1] - 1 # always interpolate when tracing to ensure the exported model # works for dynamic input shapes if ( not torch.jit.is_tracing() and num_patches == num_positions and height == width ): return self.position_embeddings class_pos_embed = self.position_embeddings[:, :1] patch_pos_embed = self.position_embeddings[:, 1:] dim = embeddings.shape[-1] new_height = height // self.patch_size[0] new_width = width // self.patch_size[1] sqrt_num_positions = torch_int(num_positions**0.5) patch_pos_embed = patch_pos_embed.reshape( 1, sqrt_num_positions, sqrt_num_positions, dim ) patch_pos_embed = patch_pos_embed.permute(0, 3, 1, 2) patch_pos_embed = nn.functional.interpolate( patch_pos_embed, size=(new_height, new_width), mode="bicubic", align_corners=False, ) patch_pos_embed = patch_pos_embed.permute(0, 2, 3, 1).view(1, -1, dim) return torch.cat((class_pos_embed, patch_pos_embed), dim=1) def forward( self, pixel_values: torch.Tensor, bool_masked_pos: torch.BoolTensor | None = None, ) -> torch.Tensor: _, _, height, width = pixel_values.shape embeddings, (patch_height, patch_width) = self.patch_embeddings(pixel_values) batch_size, seq_len, _ = embeddings.size() if bool_masked_pos is not None: mask_tokens = self.mask_token.expand(batch_size, seq_len, -1) # replace the masked visual tokens by mask_tokens w = bool_masked_pos.unsqueeze(-1).type_as(mask_tokens) embeddings = embeddings * (1 - w) + mask_tokens * w cls_tokens = self.cls_token.expand(batch_size, -1, -1) embeddings = torch.cat((cls_tokens, embeddings), dim=1) if self.position_embeddings is not None: embeddings = embeddings + self.interpolate_pos_encoding( embeddings, height, width ) return embeddings, (patch_height, patch_width) class InternSdpaAttention(nn.Module): """Multi-headed attention from 'Attention Is All You Need' paper""" def __init__( self, config: PretrainedConfig, *, num_dummy_heads: int = 0, prefix: str = "", ) -> None: super().__init__() self.config = config self.embed_dim = config.hidden_size self.num_heads = config.num_attention_heads self.head_dim = self.embed_dim // self.num_heads if self.head_dim * self.num_heads != self.embed_dim: raise ValueError( f"embed_dim must be divisible by num_heads " f"(got `embed_dim`: {self.embed_dim} and `num_heads`:" f" {self.num_heads})." ) # Additional dummy heads are used to enable TP for common GPU counts. self.dummy_dim = (num_dummy_heads + self.num_heads) * self.head_dim self.scale = self.head_dim**-0.5 self.q_proj = nn.Linear( self.embed_dim, self.num_heads * self.head_dim, bias=config.attention_bias ) self.k_proj = nn.Linear( self.embed_dim, self.num_heads * self.head_dim, bias=config.attention_bias ) self.v_proj = nn.Linear( self.embed_dim, self.num_heads * self.head_dim, bias=config.attention_bias ) self.qk_normalization = config.use_qk_norm if self.qk_normalization: self.q_norm = RMSNorm( self.dummy_dim, eps=config.layer_norm_eps, var_hidden_size=self.embed_dim, ) self.k_norm = RMSNorm( self.dummy_dim, eps=config.layer_norm_eps, var_hidden_size=self.embed_dim, ) self.projection_layer = nn.Linear(self.dummy_dim, self.embed_dim) # Use unified MMEncoderAttention with automatic backend selection self.attn = MMEncoderAttention( self.num_heads, self.head_dim, self.scale, prefix=f"{prefix}.attn", ) def forward(self, x: torch.Tensor) -> torch.Tensor: """x shape: (B, N, C)""" q = self.q_proj(x) k = self.k_proj(x) v = self.v_proj(x) if self.qk_normalization: q = self.q_norm(q) k = self.k_norm(k) # Use unified MMEncoderAttention with automatic backend selection x = self.attn(q, k, v) x = self.projection_layer(x) return x class InternS1VisionMLP(nn.Module): def __init__( self, config: PretrainedConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.config = config self.activation_fn = get_act_fn(config.hidden_act) self.fc1 = ColumnParallelLinear( config.hidden_size, config.intermediate_size, bias=True, quant_config=quant_config, prefix=f"{prefix}.fc1", ) self.fc2 = RowParallelLinear( config.intermediate_size, config.hidden_size, bias=True, quant_config=quant_config, prefix=f"{prefix}.fc2", ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states, _ = self.fc1(hidden_states) hidden_states = self.activation_fn(hidden_states) hidden_states, _ = self.fc2(hidden_states) return hidden_states class InternS1VisionLayer(nn.Module): def __init__( self, config: PretrainedConfig, quant_config: QuantizationConfig | None = None, *, num_dummy_heads: int = 0, prefix: str = "", ) -> None: super().__init__() self.attention = self._init_attn( config, quant_config, num_dummy_heads=num_dummy_heads, prefix=f"{prefix}.attention", ) self.mlp = InternS1VisionMLP( config, quant_config=quant_config, prefix=f"{prefix}.mlp" ) self.layernorm_before = NORM2FN[config.norm_type]( config.hidden_size, eps=config.layer_norm_eps ) self.layernorm_after = NORM2FN[config.norm_type]( config.hidden_size, eps=config.layer_norm_eps ) init_values = config.layer_scale_init_value self.lambda_1 = nn.Parameter( init_values * torch.ones(config.hidden_size), requires_grad=True ) self.lambda_2 = nn.Parameter( init_values * torch.ones(config.hidden_size), requires_grad=True ) def _init_attn( self, config: PretrainedConfig, quant_config: QuantizationConfig | None, *, num_dummy_heads: int, prefix: str = "", ): return InternSdpaAttention( config, num_dummy_heads=num_dummy_heads, prefix=prefix, ) def forward( self, hidden_states: torch.Tensor, ): hidden_states = ( hidden_states + self.attention(self.layernorm_before(hidden_states)) * self.lambda_1 ) hidden_states = ( hidden_states + self.mlp(self.layernorm_after(hidden_states)) * self.lambda_2 ) return hidden_states class InternS1VisionEncoder(nn.Module): def __init__( self, config: PretrainedConfig, quant_config: QuantizationConfig | None = None, *, num_hidden_layers_override: int | None = None, num_dummy_heads: int = 0, prefix: str = "", ): super().__init__() self.config = config if num_hidden_layers_override is None: num_hidden_layers = config.num_hidden_layers else: num_hidden_layers = num_hidden_layers_override self.layer = nn.ModuleList( [ InternS1VisionLayer( config, quant_config, num_dummy_heads=num_dummy_heads, prefix=f"{prefix}.layer.{layer_idx}", ) for layer_idx in range(num_hidden_layers) ] ) def forward(self, inputs_embeds: torch.Tensor): hidden_states = inputs_embeds for encoder_layer in self.layer: hidden_states = encoder_layer(hidden_states) return hidden_states class InternS1VisionModel(nn.Module): def __init__( self, config: PretrainedConfig, quant_config: QuantizationConfig | None = None, *, num_hidden_layers_override: int | None = None, num_dummy_heads: int = 0, prefix: str = "", ) -> None: super().__init__() self.config = config self.embeddings = InternS1VisionEmbeddings(config) self.encoder = InternS1VisionEncoder( config=config, num_hidden_layers_override=num_hidden_layers_override, num_dummy_heads=num_dummy_heads, prefix=f"{prefix}.encoder", ) self.layernorm = ( nn.Identity() if config.use_mean_pooling else nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) ) def get_input_embeddings(self): return self.embeddings.patch_embeddings def forward( self, pixel_values: torch.Tensor | None = None, pixel_embeds: torch.Tensor | None = None, ) -> torch.FloatTensor: if pixel_values is None and pixel_embeds is None: raise ValueError("You have to specify pixel_values or pixel_embeds") if pixel_embeds is not None: hidden_states = pixel_embeds elif pixel_values is not None: if pixel_values.ndim == 4: hidden_states, _ = self.embeddings(pixel_values) else: raise ValueError(f"wrong pixel_values size: {pixel_values.shape}") encoder_outputs = self.encoder(inputs_embeds=hidden_states) encoder_outputs = self.layernorm(encoder_outputs) return encoder_outputs def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() for name, loaded_weight in weights: param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/interns1_vit.py", "license": "Apache License 2.0", "lines": 368, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/attention/test_aiter_flash_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from vllm.platforms import current_platform from vllm.utils.torch_utils import set_random_seed # Import AITER backend if on ROCm and aiter is available if current_platform.is_rocm(): from vllm._aiter_ops import is_aiter_found_and_supported if is_aiter_found_and_supported(): import aiter from vllm.v1.attention.backends.rocm_aiter_fa import cp_mha_gather_cache NUM_HEADS = [(4, 4), (8, 2)] HEAD_SIZES = [128, 256] BLOCK_SIZES = [16] DTYPES = [torch.bfloat16] QDTYPES = [None] # one value large enough to test overflow in index calculation. # one value small enough to test the schema op check NUM_BLOCKS = [32768, 2048] def ref_paged_attn( query: torch.Tensor, key_cache: torch.Tensor, value_cache: torch.Tensor, query_lens: list[int], kv_lens: list[int], block_tables: torch.Tensor, scale: float, sliding_window: int | None = None, soft_cap: float | None = None, ) -> torch.Tensor: num_seqs = len(query_lens) block_tables = block_tables.cpu().numpy() _, block_size, num_kv_heads, head_size = key_cache.shape outputs: list[torch.Tensor] = [] start_idx = 0 for i in range(num_seqs): query_len = query_lens[i] kv_len = kv_lens[i] q = query[start_idx : start_idx + query_len] q *= scale num_kv_blocks = (kv_len + block_size - 1) // block_size block_indices = block_tables[i, :num_kv_blocks] k = key_cache[block_indices].view(-1, num_kv_heads, head_size) k = k[:kv_len] v = value_cache[block_indices].view(-1, num_kv_heads, head_size) v = v[:kv_len] if q.shape[1] != k.shape[1]: k = torch.repeat_interleave(k, q.shape[1] // k.shape[1], dim=1) v = torch.repeat_interleave(v, q.shape[1] // v.shape[1], dim=1) attn = torch.einsum("qhd,khd->hqk", q, k).float() empty_mask = torch.ones(query_len, kv_len) mask = torch.triu(empty_mask, diagonal=kv_len - query_len + 1).bool() if sliding_window is not None: sliding_window_mask = ( torch.triu( empty_mask, diagonal=kv_len - (query_len + sliding_window) + 1 ) .bool() .logical_not() ) mask |= sliding_window_mask if soft_cap is not None: attn = soft_cap * torch.tanh(attn / soft_cap) attn.masked_fill_(mask, float("-inf")) attn = torch.softmax(attn, dim=-1).to(v.dtype) out = torch.einsum("hqk,khd->qhd", attn, v) outputs.append(out) start_idx += query_len return torch.cat(outputs, dim=0) @pytest.mark.skipif(not current_platform.is_rocm(), reason="Only ROCm is supported") @pytest.mark.parametrize( "seq_lens", [[(10, 1328), (5, 18), (129, 463)], [(8, 523), (24, 37), (3, 2011)]] ) @pytest.mark.parametrize("num_heads", NUM_HEADS) @pytest.mark.parametrize("head_size", HEAD_SIZES) @pytest.mark.parametrize("block_size", BLOCK_SIZES) @pytest.mark.parametrize("sliding_window", [None, 256]) @pytest.mark.parametrize("dtype", DTYPES) @pytest.mark.parametrize("soft_cap", [None]) @pytest.mark.parametrize("num_blocks", NUM_BLOCKS) @pytest.mark.parametrize("q_dtype", QDTYPES) @torch.inference_mode() def test_varlen_with_paged_kv( seq_lens: list[tuple[int, int]], num_heads: tuple[int, int], head_size: int, sliding_window: int | None, dtype: torch.dtype, block_size: int, soft_cap: float | None, num_blocks: int, q_dtype: torch.dtype | None, ) -> None: from vllm._aiter_ops import is_aiter_found_and_supported if not is_aiter_found_and_supported(): pytest.skip("aiter package required for this test.") torch.set_default_device("cuda") set_random_seed(0) num_seqs = len(seq_lens) query_lens = [x[0] for x in seq_lens] kv_lens = [x[1] for x in seq_lens] num_query_heads = num_heads[0] num_kv_heads = num_heads[1] assert num_query_heads % num_kv_heads == 0 max_query_len = max(query_lens) max_kv_len = max(kv_lens) window_size = (sliding_window - 1, 0) if sliding_window is not None else (-1, -1) scale = head_size**-0.5 query = torch.randn(sum(query_lens), num_query_heads, head_size, dtype=dtype) key_cache = torch.randn( num_blocks, block_size, num_kv_heads, head_size, dtype=dtype ) value_cache = torch.randn_like(key_cache) cu_query_lens = torch.tensor([0] + query_lens, dtype=torch.int32).cumsum( dim=0, dtype=torch.int32 ) cu_seq_lens = torch.tensor([0] + kv_lens, dtype=torch.int32).cumsum( dim=0, dtype=torch.int32 ) # Save kv_lens as list before converting to tensor kv_lens_list = kv_lens kv_lens = torch.tensor(kv_lens, dtype=torch.int32) max_num_blocks_per_seq = (max_kv_len + block_size - 1) // block_size block_tables = torch.randint( 0, num_blocks, (num_seqs, max_num_blocks_per_seq), dtype=torch.int32 ) output = torch.empty_like(query) maybe_quantized_query = query maybe_quantized_key_cache = key_cache maybe_quantized_value_cache = value_cache k_scale_tensor = None v_scale_tensor = None dequant = False if q_dtype is not None: # QKV are drawn from N(0, 1): no need for a fp8 scaling factor maybe_quantized_query = query.to(q_dtype) maybe_quantized_key_cache = key_cache.to(q_dtype) maybe_quantized_value_cache = value_cache.to(q_dtype) dequant = True scale_shape = (num_seqs, num_kv_heads) # For per-seq-per-head scales (matching AITER backend expectation) k_scale_tensor = torch.ones(scale_shape, dtype=torch.float32) v_scale_tensor = torch.ones(scale_shape, dtype=torch.float32) # Prepare metadata for cp_mha_gather_cache # token_to_batch: maps each token to its batch index token_to_batch = torch.zeros(sum(kv_lens_list), dtype=torch.int32) seq_starts = torch.zeros(num_seqs, dtype=torch.int32) token_idx = 0 for batch_idx, kv_len in enumerate(kv_lens_list): token_to_batch[token_idx : token_idx + kv_len] = batch_idx seq_starts[batch_idx] = 0 # Assuming all sequences start at 0 in their blocks token_idx += kv_len # Allocate buffers for gathered KV total_kv_tokens = sum(kv_lens_list) gathered_key = torch.empty( total_kv_tokens, num_kv_heads, head_size, dtype=maybe_quantized_key_cache.dtype ) gathered_value = torch.empty( total_kv_tokens, num_kv_heads, head_size, dtype=maybe_quantized_value_cache.dtype, ) # Gather paged KV cache into contiguous tensors using triton kernel cp_mha_gather_cache( key_cache=maybe_quantized_key_cache, value_cache=maybe_quantized_value_cache, key=gathered_key, value=gathered_value, block_tables=block_tables, k_scales=k_scale_tensor if k_scale_tensor is not None else torch.ones(1, dtype=torch.float32), v_scales=v_scale_tensor if v_scale_tensor is not None else torch.ones(1, dtype=torch.float32), cu_seqlens_kv=cu_seq_lens, token_to_batch=token_to_batch, seq_starts=seq_starts, dequant=dequant, kv_cache_layout="NHD", total_tokens=total_kv_tokens, ) # Call aiter flash attention with gathered KV aiter.flash_attn_varlen_func( q=maybe_quantized_query, k=gathered_key, v=gathered_value, cu_seqlens_q=cu_query_lens, cu_seqlens_k=cu_seq_lens, max_seqlen_q=max_query_len, max_seqlen_k=max_kv_len, min_seqlen_q=1, dropout_p=0.0, softmax_scale=scale, causal=True, window_size=window_size, alibi_slopes=None, return_lse=False, out=output, ) ref_output = ref_paged_attn( query=query, key_cache=key_cache, value_cache=value_cache, query_lens=query_lens, kv_lens=kv_lens_list, block_tables=block_tables, scale=scale, sliding_window=sliding_window, soft_cap=soft_cap, ) atol, rtol = 2e-2, 2e-2 if q_dtype is not None: atol, rtol = 1.5e-1, 1.5e-1 ( torch.testing.assert_close(output, ref_output, atol=atol, rtol=rtol), f"{torch.max(torch.abs(output - ref_output))}", ) # Log diff stats for tracking changes print(f"Max abs diff: {torch.max(torch.abs(output - ref_output))}") print(f"Mean diff: {torch.mean(torch.abs(output - ref_output))}") print(f"Min diff: {torch.std(torch.abs(output - ref_output))}")
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/attention/test_aiter_flash_attn.py", "license": "Apache License 2.0", "lines": 224, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/layers/mamba/ops/layernorm_gated.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) 2024, Tri Dao. # Adapted from https://github.com/state-spaces/mamba/blob/60dadf2e0ee730ac337035d5533de10bc26e4847/mamba_ssm/ops/triton/layernorm_gated.py import torch from vllm.triton_utils import tl, triton @triton.heuristics({"HAS_BIAS": lambda args: args["B"] is not None}) @triton.heuristics({"HAS_Z": lambda args: args["Z"] is not None}) @triton.jit def _layer_norm_fwd_1pass_kernel( X, # pointer to the input Y, # pointer to the output W, # pointer to the weights B, # pointer to the biases Z, # pointer to the other branch Mean, # pointer to the mean Rstd, # pointer to the 1/std stride_x_row: tl.int64, stride_y_row: tl.int64, stride_z_row: tl.int64, M: tl.int64, # number of rows in X N: tl.int64, # number of columns in X eps, # epsilon to avoid division by zero BLOCK_N: tl.constexpr, HAS_BIAS: tl.constexpr, HAS_Z: tl.constexpr, NORM_BEFORE_GATE: tl.constexpr, IS_RMS_NORM: tl.constexpr, ): # Map the program id to the row of X and Y it should compute. row = tl.program_id(0) group = tl.program_id(1) X += row * stride_x_row + group * N Y += row * stride_y_row + group * N if HAS_Z: Z += row * stride_z_row + group * N if not IS_RMS_NORM: Mean += group * M Rstd += group * M W += group * N if HAS_BIAS: B += group * N # Compute mean and variance cols = tl.arange(0, BLOCK_N) x = tl.load(X + cols, mask=cols < N, other=0.0).to(tl.float32) if HAS_Z and not NORM_BEFORE_GATE: z = tl.load(Z + cols, mask=cols < N).to(tl.float32) x *= z * tl.sigmoid(z) if not IS_RMS_NORM: mean = tl.sum(x, axis=0) / N tl.store(Mean + row, mean) xbar = tl.where(cols < N, x - mean, 0.0) var = tl.sum(xbar * xbar, axis=0) / N else: xbar = tl.where(cols < N, x, 0.0) var = tl.sum(xbar * xbar, axis=0) / N rstd = 1 / tl.sqrt(var + eps) tl.store(Rstd + row, rstd) # Normalize and apply linear transformation mask = cols < N w = tl.load(W + cols, mask=mask).to(tl.float32) if HAS_BIAS: b = tl.load(B + cols, mask=mask).to(tl.float32) x_hat = (x - mean) * rstd if not IS_RMS_NORM else x * rstd y = x_hat * w + b if HAS_BIAS else x_hat * w if HAS_Z and NORM_BEFORE_GATE: z = tl.load(Z + cols, mask=mask).to(tl.float32) y *= z * tl.sigmoid(z) # Write output tl.store(Y + cols, y, mask=mask) def _layer_norm_fwd( x, weight, bias, eps, z=None, out=None, group_size=None, norm_before_gate=True, is_rms_norm=False, ): M, N = x.shape if group_size is None: group_size = N assert N % group_size == 0 ngroups = N // group_size assert x.stride(-1) == 1 if z is not None: assert z.stride(-1) == 1 assert z.shape == (M, N) assert weight.shape == (N,) assert weight.stride(-1) == 1 if bias is not None: assert bias.stride(-1) == 1 assert bias.shape == (N,) # allocate output if out is not None: assert out.shape == x.shape else: out = torch.empty_like(x) assert out.stride(-1) == 1 mean = ( torch.empty((ngroups * M,), dtype=torch.float32, device=x.device) if not is_rms_norm else None ) rstd = torch.empty((ngroups * M,), dtype=torch.float32, device=x.device) # Less than 64KB per feature: enqueue fused kernel MAX_FUSED_SIZE = 65536 // x.element_size() BLOCK_N = min(MAX_FUSED_SIZE, triton.next_power_of_2(group_size)) if group_size > BLOCK_N: raise RuntimeError("This layer norm doesn't support feature dim >= 64KB.") # heuristics for number of warps num_warps = min(max(BLOCK_N // 256, 1), 8) grid = (M, ngroups) with torch.cuda.device(x.device.index): _layer_norm_fwd_1pass_kernel[grid]( x, out, weight, bias, z, mean, rstd, x.stride(0), out.stride(0), z.stride(0) if z is not None else 0, M, group_size, eps, BLOCK_N=BLOCK_N, NORM_BEFORE_GATE=norm_before_gate, IS_RMS_NORM=is_rms_norm, num_warps=num_warps, ) return out, mean, rstd def rms_norm_gated( x, weight, bias, z=None, eps=1e-6, group_size=None, norm_before_gate=True ): x_shape_og = x.shape # reshape input data into 2D tensor x = x.reshape(-1, x.shape[-1]) if x.stride(-1) != 1: x = x.contiguous() if z is not None: assert z.shape == x_shape_og z = z.reshape(-1, z.shape[-1]) if z.stride(-1) != 1: z = z.contiguous() weight = weight.contiguous() if bias is not None: bias = bias.contiguous() y, _, _ = _layer_norm_fwd( x, weight, bias, eps, z=z, group_size=group_size, norm_before_gate=norm_before_gate, is_rms_norm=True, ) return y.reshape(x_shape_og)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/mamba/ops/layernorm_gated.py", "license": "Apache License 2.0", "lines": 163, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/hyperclovax_vision.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # copied from : https://github.com/huggingface/transformers import ast from collections import defaultdict from collections.abc import Iterable, Mapping, Sequence from functools import partial from itertools import accumulate from typing import Annotated, Literal import numpy as np import torch import torch.nn as nn from einops import rearrange from timm.layers import LayerNorm, LayerNorm2d from timm.models.regnet import RegStage from transformers import BatchFeature, CLIPVisionConfig, SiglipVisionConfig from vllm.config import VllmConfig from vllm.config.multimodal import BaseDummyOptions from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.cache import BaseMultiModalProcessorCache from vllm.multimodal.inputs import ( MultiModalDataDict, MultiModalFieldConfig, MultiModalKwargsItems, ) from vllm.multimodal.parse import ImageSize, MultiModalDataItems from vllm.multimodal.processing import ( BaseDummyInputsBuilder, BaseMultiModalProcessor, BaseProcessingInfo, InputProcessingContext, PromptReplacement, PromptUpdate, ) from vllm.sequence import IntermediateTensors from vllm.utils.tensor_schema import TensorSchema, TensorShape from .clip import CLIPVisionModel from .interfaces import MultiModalEmbeddings, SupportsMultiModal, SupportsPP from .siglip import SiglipVisionModel from .utils import ( AutoWeightsLoader, flatten_bn, init_vllm_registered_model, maybe_prefix, ) from .vision import get_vision_encoder_info IMAGE_TOKEN: str = "<|dummy3|>" VIDEO_TOKEN: str = "<|_unuse_missing_100270|>" # Based on combine_frames_into_images in # https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B/blob/main/processing_hyperclovax.py def get_num_combined_frames( num_frames: int, max_grid_shape: tuple[int, int] = (3, 3), ) -> int: max_num_grids = max_grid_shape[0] * max_grid_shape[1] # Calculate the number of canvases needed. num_canvases = num_frames // max_num_grids leftover_frames = num_frames % max_num_grids return num_canvases + (leftover_frames > 0) class HCXVisionImagePixelInputs(TensorSchema): """ Dimensions: - n: Number of images - g: Number of grids - c: Number of channels (3) - h: Height - w: Width """ type: Literal["pixel_values"] = "pixel_values" pixel_values_images: Annotated[ list[torch.Tensor], TensorShape("n", "g", 3, "h", "w", dynamic_dims={"g"}) ] image_sizes_images: Annotated[torch.Tensor, TensorShape("n", 2)] HCXVisionImageInputs = HCXVisionImagePixelInputs class HCXVisionVideoPixelInputs(TensorSchema): """ Dimensions: - n: Number of videos - f: Number of frames - g: Number of grids - c: Number of channels (3) - h: Height - w: Width """ type: Literal["pixel_values_videos"] = "pixel_values_videos" pixel_values_videos: Annotated[ list[list[torch.Tensor]], TensorShape("n", "f", "g", 3, "h", "w", dynamic_dims={"f", "g"}), ] HCXVisionVideoInputs = HCXVisionVideoPixelInputs class HCXVisionProcessingInfo(BaseProcessingInfo): def get_vision_encoder_info(self): return get_vision_encoder_info(self.get_hf_config()) def get_supported_mm_limits(self) -> Mapping[str, int | None]: return {"image": None, "video": None} def get_num_image_tokens( self, *, vision_query_length: int | list[int], ) -> int: if isinstance(vision_query_length, int): return vision_query_length else: return sum(vision_query_length) def get_num_video_tokens( self, *, vision_query_length: int | list[int], ) -> int: if isinstance(vision_query_length, int): return vision_query_length else: return sum(vision_query_length) def get_image_size_with_most_features(self) -> ImageSize: vision_encoder_info = self.get_vision_encoder_info() width = height = vision_encoder_info.get_image_size() return ImageSize(width=width, height=height) def get_max_image_tokens(self) -> int: target_width, target_height = self.get_image_size_with_most_features() return self.get_num_image_tokens( image_width=target_width, image_height=target_height, ) class HCXVisionDummyInputsBuilder(BaseDummyInputsBuilder[HCXVisionProcessingInfo]): def get_dummy_text( self, mm_counts: Mapping[str, int], ) -> str: dummy_text = IMAGE_TOKEN * mm_counts.get( "image", 0 ) + VIDEO_TOKEN * mm_counts.get("video", 0) return dummy_text def get_dummy_mm_data( self, seq_len: int, mm_counts: Mapping[str, int], mm_options: Mapping[str, BaseDummyOptions], ) -> MultiModalDataDict: num_images = mm_counts.get("image", 0) num_videos = mm_counts.get("video", 0) target_width, target_height = self.info.get_image_size_with_most_features() target_num_frames = 32 image_overrides = mm_options.get("image") video_overrides = mm_options.get("video") return { "image": self._get_dummy_images( width=target_width, height=target_height, num_images=num_images, overrides=image_overrides, ), "video": self._get_dummy_videos( width=target_width - 1, height=target_height - 1, num_frames=target_num_frames, num_videos=num_videos, overrides=video_overrides, ), } class HCXVisionMultiModalProcessor(BaseMultiModalProcessor[HCXVisionProcessingInfo]): def _call_hf_processor( self, prompt: str, mm_data: Mapping[str, object], mm_kwargs: Mapping[str, object], tok_kwargs: Mapping[str, object], ) -> BatchFeature: for video_idx, video_arr in enumerate(mm_data.get("videos", [])): if video_arr.dtype != np.uint8: mm_data["videos"][video_idx] = video_arr.astype(np.uint8) processed_outputs = self.info.ctx.call_hf_processor( hf_processor=self.info.get_hf_processor(**mm_kwargs), data=dict( text=prompt, images=None, videos=None, ), ) # text-only if len(mm_data) > 0: images = mm_data.get("images") videos = mm_data.get("videos") # batchify input as a single item _processed_outputs = self.info.ctx.call_hf_processor( hf_processor=self.info.get_hf_processor(**mm_kwargs), data=dict( text=None, images=None if images is None else [images], videos=None if videos is None else [videos], ), ) # mm-only for k, v in _processed_outputs.items(): if isinstance(v, list) and len(v) > 0: assert len(v) == 1 _processed_outputs[k] = v[0] if images: _processed_outputs["image_sizes_images"] = torch.tensor( _processed_outputs["image_sizes_images"] ) _processed_outputs["vision_query_lengths_images"] = torch.tensor( _processed_outputs["vision_query_lengths_images"] ) if videos: _idx_per_video = [ 0, *accumulate( get_num_combined_frames(len(video)) for video in videos ), ] _processed_outputs["pixel_values_videos"] = [ _processed_outputs["pixel_values_videos"][ _idx_per_video[i] : _idx_per_video[i + 1] ] for i in range(len(videos)) ] _processed_outputs["vision_query_lengths_videos"] = [ torch.tensor( _processed_outputs["vision_query_lengths_videos"][ _idx_per_video[i] : _idx_per_video[i + 1] ] ) for i in range(len(videos)) ] processed_outputs.update(_processed_outputs) return processed_outputs def _hf_processor_applies_updates( self, prompt_text: str, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, object], tokenization_kwargs: Mapping[str, object], ) -> bool: return False def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, object], out_mm_kwargs: MultiModalKwargsItems, ) -> Sequence[PromptUpdate]: hf_config = self.info.get_hf_config() placeholder = { "image": hf_config.image_token_id, "video": hf_config.video_token_id, } def get_replacement_hyperclovax( item_idx: int, modality: str, out_mm_kwargs: MultiModalKwargsItems, ): out_item = out_mm_kwargs[modality][item_idx] if modality == "image": lens = out_item["vision_query_lengths_images"].data.tolist() num_tokens = self.info.get_num_image_tokens(vision_query_length=lens) elif modality == "video": lens = out_item["vision_query_lengths_videos"].data.tolist() num_tokens = self.info.get_num_video_tokens(vision_query_length=lens) else: raise NotImplementedError(modality) return [placeholder[modality]] * num_tokens return [ PromptReplacement( modality=modality, target=[ placeholder[modality], ], replacement=partial( get_replacement_hyperclovax, modality=modality, out_mm_kwargs=out_mm_kwargs, ), ) for modality in ("image", "video") ] def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: return dict( pixel_values_images=MultiModalFieldConfig.batched("image"), image_sizes_images=MultiModalFieldConfig.batched("image"), vision_query_lengths_images=MultiModalFieldConfig.batched("image"), pixel_values_videos=MultiModalFieldConfig.batched("video"), vision_query_lengths_videos=MultiModalFieldConfig.batched("video"), ) def _build_hcxvision_hf_info( ctx: InputProcessingContext, ) -> HCXVisionProcessingInfo: return HCXVisionProcessingInfo(ctx) def _build_hcxvision_hf_processor( info: HCXVisionProcessingInfo, dummy_inputs: BaseDummyInputsBuilder[HCXVisionProcessingInfo], *, cache: BaseMultiModalProcessorCache | None = None, ) -> BaseMultiModalProcessor: if isinstance(info, HCXVisionProcessingInfo): return HCXVisionMultiModalProcessor( info, dummy_inputs, # type: ignore cache=cache, ) raise NotImplementedError(type(info)) def init_vision_tower_for_hcxvision( vision_config, quant_config: QuantizationConfig | None, *, use_nth_layer: int | None = None, require_post_norm: bool | None = None, prefix: str = "", ) -> CLIPVisionModel | SiglipVisionModel: num_hidden_layers = vision_config.num_hidden_layers if not isinstance(use_nth_layer, int): pass elif use_nth_layer >= 0: num_hidden_layers = use_nth_layer + 1 else: num_hidden_layers = num_hidden_layers + use_nth_layer + 1 if isinstance(vision_config, CLIPVisionConfig): return CLIPVisionModel( vision_config, quant_config=quant_config, num_hidden_layers_override=num_hidden_layers, require_post_norm=require_post_norm, prefix=prefix, ) elif isinstance(vision_config, SiglipVisionConfig): return SiglipVisionModel( vision_config, quant_config=quant_config, num_hidden_layers_override=num_hidden_layers, require_post_norm=require_post_norm, prefix=prefix, ) msg = f"Unsupported vision config: {type(vision_config)}" raise NotImplementedError(msg) class HCXVisionMlp(nn.Module): def __init__( self, mm_projector_type, in_features, hidden_features=None, out_features=None, act_layer=nn.GELU, ): super().__init__() out_features = out_features or in_features hidden_features = hidden_features or in_features self.mm_projector_type = mm_projector_type if self.mm_projector_type == "mlp": self.fc1 = nn.Linear(in_features, hidden_features) self.act = act_layer() self.fc2 = nn.Linear(hidden_features, out_features) elif self.mm_projector_type == "inverted_mlp": self.fc1 = nn.Linear(in_features, 2 * hidden_features) self.act = act_layer() self.fc2 = nn.Linear(2 * hidden_features, out_features) else: raise NotImplementedError( "{} is not implemented".format(self.mm_projector_type) ) def forward(self, x): x = self.fc1(x) x = self.act(x) x = self.fc2(x) return x class HCXVisionCAbstractor(nn.Module): """ This module is based on C-Abstractor, whose license is under apache-2.0. You can check the original code at https://github.com/khanrc/honeybee/blob/main/honeybee/projectors/projectors.py and we made necessary modifications. """ def __init__( self, num_queries: int, num_input_tokens: int, encoder_hidden_size: int, hidden_size: int, output_hidden_size: int, pos_emb: bool = True, prenorm: bool = False, ): super().__init__() self.num_input_tokens = num_input_tokens self.output_hidden_size = output_hidden_size # Positional embedding if pos_emb: self.pos_emb = torch.nn.Parameter( torch.zeros(1, num_input_tokens, encoder_hidden_size) ) self.pos_emb.data.normal_(mean=0.0, std=0.02) else: self.pos_emb = None # (Optional) Pre-normalization layer if prenorm: self.prenorm = LayerNorm(encoder_hidden_size) else: self.prenorm = None self.build_net( num_queries, encoder_hidden_size, hidden_size, output_hidden_size ) self.dtype = next(self.parameters()).dtype def forward( self, x: torch.Tensor, num_queries_vis_abstractors: list[list[int]] | None = None, num_grids: list[int] | None = None, ) -> torch.Tensor: if self.prenorm is not None: x = self.prenorm(x) if self.pos_emb is not None: x = x + self.pos_emb x = self._forward( x, num_queries_vis_abstractors=num_queries_vis_abstractors, num_grids=num_grids, ) # (B, L, output_hidden_size) return x def _forward( self, x: torch.Tensor, num_queries_vis_abstractors: list[list[int]] | None = None, num_grids: list[int] | None = None, ) -> torch.Tensor: # x: [B, L, dim] B, L, dim = x.shape hw = int(L**0.5) x = rearrange(x, "b (h w) d -> b d h w", h=hw, w=hw) if num_queries_vis_abstractors is not None: assert num_grids is not None return self._forward_adaptive_num_query( x, num_queries_vis_abstractors, num_grids ) x = self.net(x) x = rearrange(x, "b d h w -> b (h w) d") x = self.readout(x) return x def _forward_adaptive_num_query( self, x: torch.Tensor, num_queries_vis_abstractors: list[list[int]] | None = None, num_grids: list[int] | None = None, ) -> list[torch.Tensor]: # self.net is consisted by 3 layers (s1, sampler, s2) assert len(self.net) == 3 x = self.net[0](x) # s1 new_x = [] for i, num_queries in enumerate(num_queries_vis_abstractors): hw = int(num_queries**0.5) sampler = nn.AdaptiveAvgPool2d((hw, hw)) out = sampler(x[num_grids[i] : num_grids[i + 1], :]) out = self.net[2](out) # s2 out = rearrange(out, "b d h w -> b (h w) d") out = self.readout(out) new_x.append(out) return new_x def build_net( self, n_queries: int, encoder_hidden_size: int, hidden_size: int, output_hidden_size: int, depth: int = 3, mlp_depth: int = 2, ): assert (n_queries**0.5).is_integer(), ( f"n_queries must be square number. n_queries: {n_queries}" ) hw = int(n_queries**0.5) # RegBlock = ResBlock + SE RegBlock = partial( RegStage, stride=1, dilation=1, act_layer=nn.SiLU, norm_layer=LayerNorm2d, ) s1 = RegBlock( depth, encoder_hidden_size, hidden_size, ) sampler = nn.AdaptiveAvgPool2d((hw, hw)) s2 = RegBlock( depth, hidden_size, hidden_size, ) self.net = nn.Sequential(s1, sampler, s2) self.readout = self.build_mlp(mlp_depth, hidden_size, output_hidden_size) def build_mlp( self, depth: int, hidden_size: int, output_hidden_size: int, ): layers = [nn.Linear(hidden_size, output_hidden_size)] for _ in range(1, depth): layers.append(nn.SiLU()) layers.append(nn.Linear(output_hidden_size, output_hidden_size)) return nn.Sequential(*layers) @MULTIMODAL_REGISTRY.register_processor( _build_hcxvision_hf_processor, info=_build_hcxvision_hf_info, dummy_inputs=HCXVisionDummyInputsBuilder, ) class HCXVisionForCausalLM(nn.Module, SupportsMultiModal, SupportsPP): packed_modules_mapping = { "qkv_proj": ["q_proj", "k_proj", "v_proj"], "gate_up_proj": ["gate_proj", "up_proj"], } def __init__(self, *, vllm_config: VllmConfig, prefix: str = "") -> None: super().__init__() # init configs config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config # text_config text_config = config.text_config if text_config.model_type in ["gpt2", "hyperclovax", "llama"]: text_config._attn_implementation = "sdpa" if text_config.model_type != "hyperclovax": text_config.logits_scaling = 1.0 # vision_config vision_config = config.vision_config vision_config.auto_map = {} vision_config.anyres = config.anyres vision_config.max_num_grids = config.max_num_grids self.dtype = vllm_config.model_config.dtype ## possible_resolution should be matched with preprocessor_config.json config.possible_resolutions = self._init_possible_resolutions( config, vision_config ) with self._mark_tower_model(vllm_config, {"image", "video"}): self.vision_model = init_vision_tower_for_hcxvision( vision_config, quant_config=quant_config, use_nth_layer=getattr(config, "use_nth_layer", -1), require_post_norm=False, prefix=maybe_prefix(prefix, "vision_model"), ) self.mm_projector = self._init_mm_projector( config, text_config, vision_config ) if config.anyres: self.image_newline = nn.Parameter( torch.empty(text_config.hidden_size, dtype=self.dtype) ) with self._mark_language_model(vllm_config): self.language_model = init_vllm_registered_model( vllm_config=vllm_config, hf_config=text_config, prefix=maybe_prefix(prefix, "language_model"), ) self.config = config self.vision_config = vision_config self.text_config = text_config # use_sum_loss = bool(kwargs.pop("use_sum_loss", False)) # self.reduction = self._init_reduction_type(use_sum_loss) @classmethod def get_placeholder_str(cls, modality: str, i: int) -> str | None: if modality.startswith("image"): return IMAGE_TOKEN if modality.startswith("video"): return VIDEO_TOKEN raise ValueError("Only image or video modality is supported") def _parse_and_validate_image_input( self, **kwargs: object, ) -> HCXVisionImageInputs | None: pixel_values_images = kwargs.pop("pixel_values_images", None) if pixel_values_images is None: return None image_sizes_images = kwargs.pop("image_sizes_images") return HCXVisionImagePixelInputs( pixel_values_images=pixel_values_images, image_sizes_images=image_sizes_images, ) def _parse_and_validate_video_input( self, **kwargs: object, ) -> HCXVisionVideoInputs | None: pixel_values_videos = kwargs.pop("pixel_values_videos", None) if pixel_values_videos is None: return None return HCXVisionVideoPixelInputs( pixel_values_videos=pixel_values_videos, ) def _process_image_input( self, image_input: HCXVisionImageInputs, ) -> tuple[torch.Tensor, ...]: return self.forward_images( pixel_values_images=image_input["pixel_values_images"], image_sizes_images=image_input["image_sizes_images"], ) def _process_video_input( self, video_input: HCXVisionVideoInputs, ) -> tuple[torch.Tensor, ...]: return self.forward_videos( pixel_values_videos=video_input["pixel_values_videos"], ) def _parse_and_validate_multimodal_inputs(self, **kwargs: object) -> dict: modalities = {} # Preserve the order of modalities if there are multiple of them # from the order of kwargs. for input_key in kwargs: if input_key == "pixel_values_images" and "images" not in modalities: modalities["images"] = self._parse_and_validate_image_input(**kwargs) if input_key == "pixel_values_videos" and "videos" not in modalities: modalities["videos"] = self._parse_and_validate_video_input(**kwargs) return modalities def embed_multimodal( self, **kwargs: object, ) -> MultiModalEmbeddings: modalities = self._parse_and_validate_multimodal_inputs(**kwargs) if not modalities: return [] # The result multimodal_embeddings is tuple of tensors, with each # tensor correspoending to a multimodal data item (image or video). multimodal_embeddings: tuple[torch.Tensor, ...] = () # NOTE: It is important to iterate over the keys in this dictionary # to preserve the order of the modalities. for modality in modalities: if modality == "images": image_input = modalities["images"] image_embeddings = self._process_image_input(image_input) multimodal_embeddings += tuple(image_embeddings) if modality == "videos": video_input = modalities["videos"] video_embeddings = self._process_video_input(video_input) multimodal_embeddings += tuple(video_embeddings) return multimodal_embeddings def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs: object, ) -> torch.Tensor | IntermediateTensors: if intermediate_tensors is not None: inputs_embeds = None hidden_states = self.language_model.model( input_ids, positions, intermediate_tensors, inputs_embeds=inputs_embeds ) return hidden_states def forward_images( self, pixel_values_images: list[torch.Tensor], image_sizes_images: torch.Tensor, ) -> tuple[torch.Tensor, ...]: pixel_values_image_flat = flatten_bn(pixel_values_images, concat=True) visual_token_idx = 0 if "siglip" in self.vision_config.model_type else 1 image_forward_outs = self.vision_model(pixel_values_image_flat)[ :, visual_token_idx: ] image_forward_outs = image_forward_outs.to(dtype=self.mm_projector.dtype) image_forward_outs = self.mm_projector(image_forward_outs) # b (h w) d split_sizes = [len(item) for item in pixel_values_images] image_forward_outs = torch.split(image_forward_outs, split_sizes, dim=0) # newline for anyres postprocessing image_features = anyres_postprocessing( image_forward_outs=image_forward_outs, image_sizes=image_sizes_images.tolist(), num_queries_vis_abstractor=self.config.num_queries_vis_abstractor_image, unpad=self.config.unpad, patch_size=self.vision_config.patch_size, grid_size=self.vision_config.image_size, image_newline=self.image_newline, possible_resolutions=self.config.possible_resolutions, ) return tuple(image_features) def forward_videos( self, pixel_values_videos: list[list[torch.Tensor]], ) -> tuple[torch.Tensor, ...]: pixel_values_videos_flat = flatten_bn( [frame for frames in pixel_values_videos for frame in frames], concat=True, ) visual_token_idx = 0 if "siglip" in self.vision_config.model_type else 1 video_forward_outs = self.vision_model(pixel_values_videos_flat)[ :, visual_token_idx: ] video_forward_outs = video_forward_outs.to(dtype=self.mm_projector.dtype) # Run MM-Projector # len(num_grids) == len(num_queries_vis_abstractors) + 1 grid_idx = 0 # e.g. [0, 9, 18, 19, 27, 28, 36, 37, 45, 46, 54, 55, 56] num_grids = [grid_idx] # e.g. [81, 81, 81, 9, 81, 9, 81, 9, 81, 9, 81, 9] num_queries_vis_abstractors = [] len_total_frames = video_forward_outs.shape[0] if self.config.first_last_frames_slow: # slowfast (first_last_frames_slow) assert len_total_frames != 0 if len_total_frames <= 2: num_queries_vis_abstractors.append( self.config.num_queries_vis_abstractor_video_slow ) grid_idx += len_total_frames num_grids.append(grid_idx) else: num_queries_vis_abstractors.append( self.config.num_queries_vis_abstractor_video_slow ) grid_idx += 1 num_grids.append(grid_idx) num_queries_vis_abstractors.append( self.config.num_queries_vis_abstractor_video_fast ) grid_idx += len_total_frames - 2 num_grids.append(grid_idx) num_queries_vis_abstractors.append( self.config.num_queries_vis_abstractor_video_slow ) grid_idx += 1 num_grids.append(grid_idx) else: # slowfast for pixel_values_frames in pixel_values_videos: for pixel_values_frame in pixel_values_frames: if len(pixel_values_frame) > 0: num_queries_vis_abstractors.append( self.config.num_queries_vis_abstractor_video_slow ) grid_idx += 1 num_grids.append(grid_idx) num_queries_vis_abstractors.append( self.config.num_queries_vis_abstractor_video_fast ) grid_idx = grid_idx + len(pixel_values_frame) - 1 num_grids.append(grid_idx) video_forward_outs = self.mm_projector( video_forward_outs, num_queries_vis_abstractors, num_grids ) video_features = [] # what we want to return target_features = [] target_group_size = 0 group_counter = 0 video_groups = [ len(frame) for frames in pixel_values_videos for frame in frames ] # for concat video features after projector for forward_out in video_forward_outs: target_group_size += len(forward_out) target_features.append(forward_out.flatten(0, 1)) video_group_size = video_groups[group_counter] if video_group_size == target_group_size: video_features.append(torch.cat(target_features, dim=0)) target_features = [] group_counter += 1 target_group_size = 0 elif video_group_size < target_group_size: raise RuntimeError(f"{video_group_size=} < {target_group_size=}") assert len(target_features) == 0, ( f"target_features is not empty!! {target_features}" ) assert len(video_groups) == len(video_features) feats_per_video = [len(video) for video in pixel_values_videos] idxs_per_video = [0, *accumulate(feats_per_video)] return tuple( torch.cat(video_features[idxs_per_video[i] : idxs_per_video[i + 1]]) for i in range(len(feats_per_video)) ) def _prepare_multimodal_kwargs(self, **kwargs: object): output = defaultdict(list) for k, v in kwargs.items(): if len(v) < 1 or len(v[0]) < 1: continue # if empty batch of empty sample new_k, is_video = k, False if not k.endswith("_images") and not k.endswith("_videos"): pass else: new_k, is_video = k.split("_")[:-1], k.split("_")[-1] new_k = "_".join(new_k) is_video = is_video == "videos" for _sample_idx, _v in enumerate(v): # batch -> sample if new_k not in ["pixel_values"]: if len(output[new_k]) < _sample_idx + 1: output[new_k].append(list()) _v = _v.detach().cpu().numpy().tolist() output[new_k][_sample_idx] += _v elif isinstance(_v, torch.Tensor): if len(output[new_k]) < _sample_idx + 1: output[new_k].append(list()) output["is_videos"].append(list()) _v = list(torch.unbind(_v, dim=0)) output[new_k][_sample_idx] += _v output["is_videos"][_sample_idx] += [ is_video, ] * len(_v) return dict(output) def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: return self.language_model.compute_logits(hidden_states) def load_weights( self, weights: Iterable[tuple[str, torch.Tensor]], ) -> set[str]: loader = AutoWeightsLoader(self) return loader.load_weights(weights) def _init_possible_resolutions( self, config, vision_config, ): if not getattr(config, "possible_resolutions", []): possible_resolutions = [] if config.anyres: assert config.max_num_grids > 0 for i in range(1, config.max_num_grids + 1): for j in range(1, config.max_num_grids + 1): if i == 1 and j == 1 and not config.use_1x1_grid: continue if i * j <= config.max_num_grids: possible_resolutions.append([i, j]) possible_resolutions = [ [ys * vision_config.image_size, xs * vision_config.image_size] for ys, xs in possible_resolutions ] return possible_resolutions else: return config.possible_resolutions def _init_mm_projector( self, config, text_config, vision_config, ): input_hidden_size = vision_config.hidden_size if config.mm_projector_type == "linear": mm_projector = nn.Linear(input_hidden_size, text_config.hidden_size) mm_projector.dtype = next(mm_projector.parameters()).dtype elif config.mm_projector_type == "cabstractor": mm_projector = HCXVisionCAbstractor( num_queries=config.num_queries_vis_abstractor_image, num_input_tokens=(vision_config.image_size // vision_config.patch_size) ** 2, encoder_hidden_size=input_hidden_size, hidden_size=input_hidden_size, output_hidden_size=text_config.hidden_size, pos_emb=config.proj_pos_emb, prenorm=config.proj_prenorm, ) else: mm_projector = HCXVisionMlp( config.mm_projector_type, input_hidden_size, hidden_features=input_hidden_size, out_features=self.text_config.hidden_size, ) return mm_projector def unpad_image(tensor: torch.Tensor, original_size: tuple[int, int]) -> torch.Tensor: original_width, original_height = original_size current_height, current_width = tensor.shape[1:] original_aspect_ratio = original_width / original_height current_aspect_ratio = current_width / current_height if original_aspect_ratio > current_aspect_ratio: scale_factor = current_width / original_width new_height = int(original_height * scale_factor) padding = (current_height - new_height) // 2 unpadded_tensor = tensor[:, padding : current_height - padding, :] else: scale_factor = current_height / original_height new_width = int(original_width * scale_factor) padding = (current_width - new_width) // 2 unpadded_tensor = tensor[:, :, padding : current_width - padding] return unpadded_tensor def select_best_resolution(original_size: tuple, possible_resolutions: list) -> tuple: original_height, original_width = original_size best_fit = None max_effective_resolution = 0 min_wasted_resolution = float("inf") for height, width in possible_resolutions: scale = min(width / original_width, height / original_height) downscaled_width, downscaled_height = ( int(original_width * scale), int(original_height * scale), ) effective_resolution = min( downscaled_width * downscaled_height, original_width * original_height ) wasted_resolution = (width * height) - effective_resolution if effective_resolution > max_effective_resolution or ( effective_resolution == max_effective_resolution and wasted_resolution < min_wasted_resolution ): max_effective_resolution = effective_resolution min_wasted_resolution = wasted_resolution best_fit = (height, width) return best_fit def get_anyres_image_grid_shape( image_size: tuple[int, int], grid_pinpoints: str | list[tuple[int, int]], patch_size: int, ) -> tuple[int, int]: possible_resolutions = ( grid_pinpoints if isinstance(grid_pinpoints, list) else ast.literal_eval(grid_pinpoints) ) original_width, original_height = image_size height, width = select_best_resolution( (original_height, original_width), possible_resolutions ) return width // patch_size, height // patch_size def reshape_and_unpad_image_features( image_feature: torch.Tensor, height: int, width: int, image_size: tuple[int, int], possible_resolutions: list[tuple[int, int]], grid_size: int, unpad: bool, image_newline: torch.Tensor, ) -> torch.Tensor: base_image_feature = image_feature[0] image_feature = image_feature[1:] assert height * width == base_image_feature.shape[0], ( f"{height=} * {width=} != {base_image_feature.shape[0]=}" ) num_patch_width, num_patch_height = get_anyres_image_grid_shape( image_size, possible_resolutions, grid_size ) image_feature = image_feature.view( num_patch_height, num_patch_width, height, width, -1 ) if unpad: image_feature = image_feature.permute(4, 0, 2, 1, 3).contiguous() image_feature = image_feature.flatten(1, 2).flatten(2, 3) image_feature = unpad_image(image_feature, image_size) image_feature = torch.cat( ( image_feature, image_newline[:, None, None] .expand(*image_feature.shape[:-1], 1) .to(image_feature.device), ), dim=-1, ) image_feature = image_feature.flatten(1, 2).transpose(0, 1) else: image_feature = image_feature.permute(0, 2, 1, 3, 4).contiguous() image_feature = image_feature.flatten(0, 3) image_feature = torch.cat((base_image_feature, image_feature), dim=0) return image_feature def anyres_postprocessing( image_forward_outs: list[torch.Tensor], image_sizes: list[list[int]], possible_resolutions: list[tuple[int, int]], patch_size: int, grid_size: int, image_newline: torch.Tensor, num_queries_vis_abstractor: int = -1, unpad: bool = False, ) -> list[torch.Tensor]: height = width = grid_size // patch_size if num_queries_vis_abstractor > 0: assert (num_queries_vis_abstractor**0.5).is_integer(), ( "n_queries must be square number" ) height = width = int(num_queries_vis_abstractor**0.5) # post-processing (unpad, add newline) new_image_features = [] for image_idx, image_feature in enumerate(image_forward_outs): if image_feature.shape[0] > 1: image_feature = reshape_and_unpad_image_features( image_feature=image_feature, height=height, width=width, image_size=image_sizes[image_idx], possible_resolutions=possible_resolutions, grid_size=grid_size, # Pass grid info if needed by helper unpad=unpad, image_newline=image_newline, ) else: image_feature = image_feature[0] image_feature = torch.cat( (image_feature, image_newline[None].to(image_feature.device)), dim=0 ) new_image_features.append(image_feature) return new_image_features
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/hyperclovax_vision.py", "license": "Apache License 2.0", "lines": 987, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/tasks.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Literal, get_args GenerationTask = Literal["generate", "transcription", "realtime"] GENERATION_TASKS: tuple[GenerationTask, ...] = get_args(GenerationTask) PoolingTask = Literal[ "embed", "classify", "score", "token_embed", "token_classify", "plugin" ] POOLING_TASKS: tuple[PoolingTask, ...] = get_args(PoolingTask) SupportedTask = Literal[GenerationTask, PoolingTask]
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/tasks.py", "license": "Apache License 2.0", "lines": 10, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/utils/tensor_schema.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from types import UnionType from typing import Annotated, Any, Union, get_args, get_origin, get_type_hints import torch from vllm.logger import init_logger logger = init_logger(__name__) class TensorShape: def __init__( self, *dims: int | str, dynamic_dims: set[str] | None = None, ) -> None: super().__init__() self.dims = dims self.dynamic_dims = dynamic_dims if dynamic_dims else set() def resolve(self, **bindings: int) -> tuple[int | str, ...]: resolved = list[int | str]() for dim in self.dims: if isinstance(dim, str) and dim in bindings: resolved.append(bindings[dim]) else: resolved.append(dim) return tuple(resolved) def __str__(self) -> str: """Return a string representation of the tensor shape.""" dim_strs = [] for dim in self.dims: if isinstance(dim, str): if dim in self.dynamic_dims: dim_strs.append(f"{dim}*") # Mark dynamic dimensions with * else: dim_strs.append(dim) else: dim_strs.append(str(dim)) return f"({', '.join(dim_strs)})" class TensorSchema: def __init__( self, *, validate: bool = True, resolve_bindings: dict[str, int] | None = None, **kwargs: Any, ) -> None: super().__init__() self._resolve_bindings = resolve_bindings if resolve_bindings else {} for key, value in kwargs.items(): setattr(self, key, value) if validate: self.validate() def __getitem__(self, key: str) -> Any: return getattr(self, key) def get(self, key: str, default: Any = None) -> Any: return getattr(self, key, default) def _match_shape_with_dynamic( self, actual: tuple[int, ...], reference: tuple[int, ...], expected_shape: tuple[int | str, ...], dynamic_dims: set[str], ) -> bool: if len(actual) != len(reference) or len(actual) > len(expected_shape): return False for i, (a, r) in enumerate(zip(actual, reference)): # When validating list inputs, we match shape suffixes only # (e.g. "p", 3, "h", "w"), assuming the list length corresponds # to the leading symbolic dim (e.g. "bn"). This allows comparing # only the trailing dimensions of each element in the list. dim = expected_shape[-len(actual) + i] # Skip this dimension if it's marked dynamic if dim in dynamic_dims: continue if a != r: return False return True def _fmt_indexer(self, idxs: tuple[int, ...]) -> str: if not idxs: return "" return str(list(idxs)) def _validate_field( self, value: object, field_name: str, expected_shape: tuple[int | str, ...], dynamic_dims: set[str], leading_idxs: tuple[int, ...] = (), ) -> tuple[int, ...]: """Validate a field and return the actual shape.""" if isinstance(value, (int, float)): return () # Scalar if isinstance(value, torch.Tensor): return value.shape if not isinstance(value, (list, tuple)): raise TypeError( f"{field_name}{self._fmt_indexer(leading_idxs)} is not " f"one of the expected types: int, float, Tensor, list, tuple. " f"Got: {type(value)}" ) if len(value) == 0: raise ValueError( f"{field_name}{self._fmt_indexer(leading_idxs)} is an empty sequence" ) # Ensure all tensors in the list have the same # shape, besides dynamic dimensions for i, v in enumerate(value): shape = self._validate_field( v, field_name, expected_shape[1:], dynamic_dims, leading_idxs=leading_idxs + (i,), ) if i == 0: first_shape = shape elif not self._match_shape_with_dynamic( shape, first_shape, expected_shape, dynamic_dims, ): raise ValueError( f"{field_name}{self._fmt_indexer(leading_idxs)} " f"contains inconsistent shapes: {first_shape} " f"(index 0) vs {shape} (index {i})" ) # Treat the list as a stacked tensor: # shape = (len(list), *tensor.shape) return (len(value),) + first_shape def _validate_tensor_shape_expected( self, actual_shape: tuple[int, ...], expected_shape: tuple[int | str, ...], field_name: str, shape_env: dict[str, int], dynamic_dims: set[str], ) -> None: """Validate that the actual tensor shape matches the expected shape.""" if len(actual_shape) != len(expected_shape): raise ValueError( f"{field_name} has rank {len(actual_shape)} " f"but expected {len(expected_shape)}. " f"Expected shape: {expected_shape}, " f"but got {actual_shape}" ) for i, dim in enumerate(expected_shape): if dim in dynamic_dims: continue elif isinstance(dim, int): if actual_shape[i] != dim: raise ValueError( f"{field_name} dim[{i}] expected " f"{dim}, got {actual_shape[i]}. " f"Expected shape: {expected_shape}, " f"but got {actual_shape}" ) elif isinstance(dim, str): if dim in shape_env: if actual_shape[i] != shape_env[dim]: raise ValueError( f"{field_name} dim[{i}] expected " f"'{dim}'={shape_env[dim]}, got " f"{actual_shape[i]}" ) else: shape_env[dim] = actual_shape[i] else: raise TypeError( f"{field_name} dim[{i}] has unsupported type: {type(dim)}" ) def validate(self) -> None: type_hints = get_type_hints(self.__class__, include_extras=True) shape_env = dict[str, int]() for field_name, field_type in type_hints.items(): # Check if field is missing if not hasattr(self, field_name) or getattr(self, field_name) is None: # Check if field is marked as optional actual_type = field_type if get_origin(field_type) is Annotated: args = get_args(field_type) actual_type = args[0] # Check arg was provided as Union if get_origin(actual_type) in {Union, UnionType}: # Union for Union[X, Y] and UnionType for X | Y args = get_args(actual_type) # Skip validation when Union contains None if type(None) in args: continue # Otherwise field is required, raise error raise ValueError(f"Required field '{field_name}' is missing") # Field exists, proceed with validation value = getattr(self, field_name) if get_origin(field_type) is not None: args = get_args(field_type) for arg in args: if isinstance(arg, TensorShape): expected_shape = arg.resolve(**self._resolve_bindings) actual_shape = self._validate_field( value, field_name, expected_shape, arg.dynamic_dims, ) self._validate_tensor_shape_expected( actual_shape, expected_shape, field_name, shape_env, arg.dynamic_dims, ) def print_shapes(self) -> None: """Print TensorShape annotations for debugging.""" logger.debug("Shapes in %s:", self.__class__.__name__) type_hints = get_type_hints(self.__class__, include_extras=True) for field_name, field_type in type_hints.items(): if get_origin(field_type) is not None: args = get_args(field_type) for arg in args: if isinstance(arg, TensorShape): logger.debug(" %s: %s", field_name, str(arg))
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/utils/tensor_schema.py", "license": "Apache License 2.0", "lines": 219, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/kv_connector/nixl_integration/test_disagg_accuracy.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import json import os import time import openai import requests MAX_OUTPUT_LEN = 30 SAMPLE_PROMPTS = ( "Red Hat is the best company in the world to work for because it works on " "open source software, which means that all the contributions are " "delivered to the community. As a result, when working on projects like " "vLLM we are able to meet many amazing people from various organizations " "like AMD, Google, NVIDIA, ", "We hold these truths to be self-evident, that all men are created equal, " "that they are endowed by their Creator with certain unalienable Rights, " "that among these are Life, Liberty and the pursuit of Happiness.--That " "to secure these rights, Governments are instituted among Men, deriving " "their just powers from the consent of the governed, ", ) def check_vllm_server(url: str, timeout=5, retries=3) -> bool: """ Checks if the vLLM server is ready by sending a GET request to the /health endpoint. Args: url (str): The base URL of the vLLM server. timeout (int): Timeout in seconds for the request. retries (int): Number of retries if the server is not ready. Returns: bool: True if the server is ready, False otherwise. """ for attempt in range(retries): try: response = requests.get(url, timeout=timeout) if response.status_code == 200: return True else: print( f"Attempt {attempt + 1}: Server returned status code " "{response.status_code}" ) except requests.exceptions.RequestException as e: print(f"Attempt {attempt + 1}: Error connecting to server: {e}") time.sleep(1) # Wait before retrying return False def run_simple_prompt( base_url: str, model_name: str, input_prompt: str, use_chat_endpoint: bool ) -> str: client = openai.OpenAI(api_key="EMPTY", base_url=base_url) if use_chat_endpoint: completion = client.chat.completions.create( model=model_name, messages=[ {"role": "user", "content": [{"type": "text", "text": input_prompt}]} ], max_completion_tokens=MAX_OUTPUT_LEN, temperature=0.0, seed=42, ) return completion.choices[0].message.content else: completion = client.completions.create( model=model_name, prompt=input_prompt, max_tokens=MAX_OUTPUT_LEN, temperature=0.0, seed=42, ) return completion.choices[0].text def main(): """ This script demonstrates how to accept two optional string arguments ("service_url" and "file_name") from the command line, each with a default value of an empty string, using the argparse module. """ parser = argparse.ArgumentParser(description="vLLM client script") parser.add_argument( "--service_url", # Name of the first argument type=str, required=True, help="The vLLM service URL.", ) parser.add_argument( "--model_name", # Name of the first argument type=str, required=True, help="model_name", ) parser.add_argument( "--mode", # Name of the second argument type=str, default="baseline", help="mode: baseline==non-disagg, or disagg", ) parser.add_argument( "--file_name", # Name of the second argument type=str, default=".vllm_output.txt", help="the file that saves the output tokens ", ) args = parser.parse_args() for arg in vars(args): print(f"{arg}: {getattr(args, arg)}") if args.mode == "baseline": # non-disagg health_check_url = f"{args.service_url}/health" else: # disagg proxy health_check_url = f"{args.service_url}/healthcheck" if not os.path.exists(args.file_name): raise ValueError( f"In disagg mode, the output file {args.file_name} from " "non-disagg. baseline does not exist." ) service_url = f"{args.service_url}/v1" if not check_vllm_server(health_check_url): raise RuntimeError(f"vllm server: {args.service_url} is not ready yet!") output_strs = dict() for i, prompt in enumerate(SAMPLE_PROMPTS): use_chat_endpoint = i % 2 == 1 output_str = run_simple_prompt( base_url=service_url, model_name=args.model_name, input_prompt=prompt, use_chat_endpoint=use_chat_endpoint, ) print(f"Prompt: {prompt}, output: {output_str}") output_strs[prompt] = output_str if args.mode == "baseline": # baseline: save outputs try: with open(args.file_name, "w") as json_file: json.dump(output_strs, json_file, indent=4) except OSError as e: print(f"Error writing to file: {e}") raise else: # disagg. verify outputs baseline_outputs = None try: with open(args.file_name) as json_file: baseline_outputs = json.load(json_file) except OSError as e: print(f"Error writing to file: {e}") raise assert isinstance(baseline_outputs, dict) assert len(baseline_outputs) == len(output_strs) for prompt, output in baseline_outputs.items(): assert prompt in output_strs, f"{prompt} not included" assert output == output_strs[prompt], ( f"baseline_output: {output} != PD output: {output_strs[prompt]}" ) if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/kv_connector/nixl_integration/test_disagg_accuracy.py", "license": "Apache License 2.0", "lines": 155, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/v1/worker/kv_connector_model_runner_mixin.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Define KV connector functionality mixin for model runners. """ import copy from collections.abc import Generator from contextlib import AbstractContextManager, contextmanager, nullcontext from typing import TYPE_CHECKING import torch from vllm.config import VllmConfig from vllm.config.cache import CacheDType from vllm.distributed.kv_transfer import ( ensure_kv_transfer_shutdown, get_kv_transfer_group, has_kv_transfer_group, ) from vllm.distributed.kv_transfer.kv_connector.base import KVConnectorBase from vllm.forward_context import get_forward_context, set_forward_context from vllm.logger import init_logger from vllm.v1.attention.backend import AttentionBackend from vllm.v1.kv_cache_interface import AttentionSpec, KVCacheConfig from vllm.v1.outputs import ( EMPTY_MODEL_RUNNER_OUTPUT, KVConnectorOutput, ModelRunnerOutput, ) from vllm.v1.worker.utils import AttentionGroup if TYPE_CHECKING: from vllm.v1.core.sched.output import SchedulerOutput logger = init_logger(__name__) # Defined as a kv connector functionality mixin for ModelRunner (GPU, TPU) class KVConnectorModelRunnerMixin: @staticmethod def ensure_kv_transfer_shutdown() -> None: # has_kv_transfer_group can be None during interpreter shutdown. if has_kv_transfer_group and has_kv_transfer_group(): # type: ignore[truthy-function] ensure_kv_transfer_shutdown() @staticmethod def kv_connector_no_forward( scheduler_output: "SchedulerOutput", vllm_config: VllmConfig ) -> ModelRunnerOutput: # KV send/recv even if no work to do. with ( set_forward_context(None, vllm_config), KVConnectorModelRunnerMixin._get_kv_connector_output( scheduler_output, wait_for_save=False ) as kv_connector_output, ): pass if kv_connector_output.is_empty(): return EMPTY_MODEL_RUNNER_OUTPUT output = copy.copy(EMPTY_MODEL_RUNNER_OUTPUT) output.kv_connector_output = kv_connector_output return output @staticmethod def maybe_get_kv_connector_output( scheduler_output: "SchedulerOutput", clear_metadata: bool = True, ) -> AbstractContextManager[KVConnectorOutput | None]: return ( KVConnectorModelRunnerMixin._get_kv_connector_output( scheduler_output, clear_metadata=clear_metadata ) if has_kv_transfer_group() else nullcontext() ) # This context manager must be used within an active forward context. # It encapsulates the entire KV connector lifecycle within execute_model @staticmethod @contextmanager def _get_kv_connector_output( scheduler_output: "SchedulerOutput", wait_for_save: bool = True, clear_metadata: bool = True, ) -> Generator[KVConnectorOutput, None, None]: output = KVConnectorOutput() # Update KVConnector with the KVConnector metadata forward(). kv_connector = get_kv_transfer_group() assert isinstance(kv_connector, KVConnectorBase) assert scheduler_output.kv_connector_metadata is not None kv_connector.bind_connector_metadata(scheduler_output.kv_connector_metadata) # Background KV cache transfers happen here. # These transfers are designed to be async and the requests # involved may be disjoint from the running requests. # Do this here to save a collective_rpc. kv_connector.start_load_kv(get_forward_context()) try: yield output finally: if wait_for_save: kv_connector.wait_for_save() output.finished_sending, output.finished_recving = ( kv_connector.get_finished(scheduler_output.finished_req_ids) ) output.invalid_block_ids = kv_connector.get_block_ids_with_load_errors() output.kv_connector_stats = kv_connector.get_kv_connector_stats() output.kv_cache_events = kv_connector.get_kv_connector_kv_cache_events() if clear_metadata: kv_connector.clear_connector_metadata() @staticmethod def clear_kv_connector_metadata() -> None: """Clear the KV connector metadata. Call after draft model runs.""" if has_kv_transfer_group(): kv_connector = get_kv_transfer_group() kv_connector.clear_connector_metadata() @staticmethod def use_uniform_kv_cache( attn_groups: list[list[AttentionGroup]], cache_dtype: CacheDType, ) -> bool: """ Determines whether a uniform KV layout should be used. A uniform layout means all layers KV caches will share the same underlying tensor, where for a given block number, the respective KV data for all layers will be contiguous. This will allow efficient KV transfer of per-block KV data for all layers at once. Note this layout will only be applied given 3 conditions: 1. The KV Cache config contains just a single group where all layers have the same page size. 2. A KV connector is configured, and the KV connector instance prefers to use this layout (prefer_cross_layer_blocks() returns True) 2. The flash attention backend supports this layout (get_kv_cache_stride_order(True) includes a placement for a num_layers dimension) Note that the actual placement of the num_layers dimensions in the unified layers tensors will be determined by the attention backend. Thus, the layers KV data may still not be contiguous per block if the attention backend does not support it. Args: attn_groups: The list of attention groups for this model cache_dtype: The KV cache dtype Returns: True if we should use a uniform KV cache layout. """ if not has_kv_transfer_group(): return False if not get_kv_transfer_group().prefer_cross_layer_blocks: return False if len(attn_groups) != 1 or len(attn_groups[0]) != 1: return False attn_group = attn_groups[0][0] kv_cache_spec = attn_group.kv_cache_spec if not isinstance(kv_cache_spec, AttentionSpec): return False attn_backend = attn_group.backend kv_cache_shape = attn_backend.get_kv_cache_shape( 1234, kv_cache_spec.block_size, kv_cache_spec.num_kv_heads, kv_cache_spec.head_size, cache_dtype_str=cache_dtype, ) try: kv_cache_stride_order = attn_backend.get_kv_cache_stride_order( include_num_layers_dimension=True ) except (AttributeError, NotImplementedError): return False # check that attention backend include a layers dimension return len(kv_cache_stride_order) == len(kv_cache_shape) + 1 @staticmethod def allocate_uniform_kv_caches( kv_cache_config: KVCacheConfig, attn_groups: list[list[AttentionGroup]], cache_dtype: CacheDType, device: torch.device, kernel_block_sizes: list[int], ) -> tuple[dict[str, torch.Tensor], torch.Tensor, type[AttentionBackend]]: """ Initializes and reshapes KV caches for the simple case where all layers have the same layout. This function assumes use_uniform_kv_cache() returned True. Args: kv_cache_config: The KV cache config attn_groups: The list of attention groups for this model cache_dtype: The KV cache dtype device: The torch device to allocate on. kernel_block_sizes: The kernel block sizes for each KV cache group. Returns: A tuple (kv_caches, cross_layers_kv_cache, attn_backend) where: kv_caches is a dict mapping between layer names to their corresponding memory buffer for KV cache. cross_layers_kv_cache is the cross layers kv cache tensor attn_backend is the attention backend matching this tensor """ attn_group = attn_groups[0][0] kv_cache_spec = attn_group.kv_cache_spec assert isinstance(kv_cache_spec, AttentionSpec) tensor_sizes = set( kv_cache_tensor.size for kv_cache_tensor in kv_cache_config.kv_cache_tensors ) assert len(tensor_sizes) == 1 tensor_size = tensor_sizes.pop() page_size = kv_cache_spec.page_size_bytes assert tensor_size % page_size == 0 num_blocks = tensor_size // page_size num_layers = len(kv_cache_config.kv_cache_tensors) total_size = tensor_size * num_layers assert len(kernel_block_sizes) == 1 kernel_block_size = kernel_block_sizes[0] num_blocks_per_kv_block = kv_cache_spec.block_size // kernel_block_size kernel_num_blocks = num_blocks * num_blocks_per_kv_block attn_backend = attn_group.backend kv_cache_shape = attn_backend.get_kv_cache_shape( kernel_num_blocks, kernel_block_size, kv_cache_spec.num_kv_heads, kv_cache_spec.head_size, cache_dtype_str=cache_dtype, ) # prepend a num_layers dimension into the shape kv_cache_shape = (num_layers,) + kv_cache_shape try: kv_cache_stride_order = attn_backend.get_kv_cache_stride_order( include_num_layers_dimension=True ) assert len(kv_cache_stride_order) == len(kv_cache_shape) except (AttributeError, NotImplementedError): kv_cache_stride_order = tuple(range(len(kv_cache_shape))) kv_cache_shape = tuple(kv_cache_shape[i] for i in kv_cache_stride_order) logger.info("Allocating a cross layer KV cache of shape %s", kv_cache_shape) # allocate one contiguous buffer for all layers cross_layers_kv_cache = ( torch.zeros(total_size, dtype=torch.int8, device=device) .view(kv_cache_spec.dtype) .view(kv_cache_shape) ) # Maintain original KV shape view. inv_order = [ kv_cache_stride_order.index(i) for i in range(len(kv_cache_stride_order)) ] permuted_kv_cache = cross_layers_kv_cache.permute(*inv_order) kv_caches = {} for i, kv_cache_tensor in enumerate(kv_cache_config.kv_cache_tensors): tensor = permuted_kv_cache[i] for layer_name in kv_cache_tensor.shared_by: kv_caches[layer_name] = tensor return kv_caches, cross_layers_kv_cache, attn_backend
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/worker/kv_connector_model_runner_mixin.py", "license": "Apache License 2.0", "lines": 241, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/test_shuffle_rows.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests for the shuffle_rows function Run `pytest tests/kernels/test_shuffle_rows.py`. """ import pytest import torch from vllm._custom_ops import shuffle_rows from vllm.platforms import current_platform @pytest.mark.parametrize("num_tokens", [1, 16, 64, 128, 256, 512, 1024]) @pytest.mark.parametrize("hidden_size", [128, 256, 512, 1024, 2048, 4096]) @pytest.mark.parametrize("dtype", [torch.float16, torch.bfloat16, torch.float32]) def test_shuffle_rows_basic(num_tokens: int, hidden_size: int, dtype: torch.dtype): """Test basic functionality of shuffle_rows with various tensor sizes and dtypes.""" if not current_platform.is_cuda(): pytest.skip("shuffle_rows requires CUDA") # Create input tensor input_tensor = torch.randn(num_tokens, hidden_size, device="cuda", dtype=dtype) # Create a simple permutation map (identity mapping) dst2src_map = torch.arange(num_tokens, device="cuda", dtype=torch.int32) # Test shuffle_rows output = shuffle_rows(input_tensor, dst2src_map) # With identity mapping, output should be identical to input torch.testing.assert_close(output, input_tensor, atol=0, rtol=0) # Check output shape assert output.shape == (num_tokens, hidden_size) assert output.dtype == dtype assert output.device == input_tensor.device @pytest.mark.parametrize("num_tokens", [16, 64, 128]) @pytest.mark.parametrize("hidden_size", [128, 512, 1024]) @pytest.mark.parametrize("dtype", [torch.float16, torch.bfloat16]) def test_shuffle_rows_permutation( num_tokens: int, hidden_size: int, dtype: torch.dtype ): """Test shuffle_rows with actual permutation.""" if not current_platform.is_cuda(): pytest.skip("shuffle_rows requires CUDA") # Create input tensor input_tensor = torch.randn(num_tokens, hidden_size, device="cuda", dtype=dtype) # Create a reverse permutation map dst2src_map = torch.arange(num_tokens - 1, -1, -1, device="cuda", dtype=torch.int32) # Test shuffle_rows output = shuffle_rows(input_tensor, dst2src_map) # Check that the output is the reverse of the input expected_output = torch.flip(input_tensor, dims=[0]) torch.testing.assert_close(output, expected_output, atol=1e-6, rtol=1e-5) # Check output shape and properties assert output.shape == (num_tokens, hidden_size) assert output.dtype == dtype assert output.device == input_tensor.device @pytest.mark.parametrize("num_tokens", [32, 64]) @pytest.mark.parametrize("hidden_size", [256, 512]) def test_shuffle_rows_expansion(num_tokens: int, hidden_size: int): """Test shuffle_rows with expansion (more output tokens than input tokens).""" if not current_platform.is_cuda(): pytest.skip("shuffle_rows requires CUDA") dtype = torch.float16 # Create input tensor input_tensor = torch.randn(num_tokens, hidden_size, device="cuda", dtype=dtype) # Create a mapping that duplicates some tokens (expansion) expanded_size = num_tokens * 2 dst2src_map = torch.randint( 0, num_tokens, (expanded_size,), device="cuda", dtype=torch.int32 ) # Test shuffle_rows output = shuffle_rows(input_tensor, dst2src_map) # Check output shape assert output.shape == (expanded_size, hidden_size) assert output.dtype == dtype assert output.device == input_tensor.device # Verify that each output row matches the corresponding input row for i in range(expanded_size): src_idx = dst2src_map[i].item() torch.testing.assert_close( output[i], input_tensor[src_idx], atol=1e-6, rtol=1e-5 ) @pytest.mark.parametrize("num_tokens", [16, 64]) @pytest.mark.parametrize("hidden_size", [128, 512]) def test_shuffle_rows_random_permutation(num_tokens: int, hidden_size: int): """Test shuffle_rows with random permutation.""" if not current_platform.is_cuda(): pytest.skip("shuffle_rows requires CUDA") dtype = torch.float16 # Set seed for reproducibility torch.manual_seed(42) # Create input tensor input_tensor = torch.randn(num_tokens, hidden_size, device="cuda", dtype=dtype) # Create a random permutation map dst2src_map = torch.randperm(num_tokens, device="cuda", dtype=torch.int32) # Test shuffle_rows output = shuffle_rows(input_tensor, dst2src_map) # Check output shape and properties assert output.shape == (num_tokens, hidden_size) assert output.dtype == dtype assert output.device == input_tensor.device # Verify that each output row matches the corresponding input row for i in range(num_tokens): src_idx = dst2src_map[i].item() torch.testing.assert_close( output[i], input_tensor[src_idx], atol=1e-6, rtol=1e-5 ) def test_shuffle_rows_edge_cases(): """Test shuffle_rows with edge cases.""" if not current_platform.is_cuda(): pytest.skip("shuffle_rows requires CUDA") dtype = torch.float16 # Test with single token input_tensor = torch.randn(1, 128, device="cuda", dtype=dtype) dst2src_map = torch.tensor([0], device="cuda", dtype=torch.int32) output = shuffle_rows(input_tensor, dst2src_map) torch.testing.assert_close(output, input_tensor, atol=0, rtol=0) # Test with single feature dimension input_tensor = torch.randn(16, 1, device="cuda", dtype=dtype) dst2src_map = torch.arange(16, device="cuda", dtype=torch.int32) output = shuffle_rows(input_tensor, dst2src_map) torch.testing.assert_close(output, input_tensor, atol=0, rtol=0) def test_shuffle_rows_moe_like_scenario(): """Test shuffle_rows in a scenario similar to MoE usage.""" if not current_platform.is_cuda(): pytest.skip("shuffle_rows requires CUDA") dtype = torch.float16 batch_size = 32 hidden_size = 1024 topk = 2 # Simulate input tokens input_tensor = torch.randn(batch_size, hidden_size, device="cuda", dtype=dtype) # Simulate expert assignment (each token goes to topk experts) # This creates a mapping where tokens are duplicated for multiple experts total_tokens = batch_size * topk dst2src_map = torch.zeros(total_tokens, device="cuda", dtype=torch.int32) # Fill the mapping to simulate MoE token distribution for i in range(batch_size): for k in range(topk): dst2src_map[i * topk + k] = i # Test shuffle_rows output = shuffle_rows(input_tensor, dst2src_map) # Check output shape assert output.shape == (total_tokens, hidden_size) assert output.dtype == dtype assert output.device == input_tensor.device # Verify that tokens are correctly duplicated for i in range(batch_size): for k in range(topk): output_idx = i * topk + k torch.testing.assert_close( output[output_idx], input_tensor[i], atol=1e-6, rtol=1e-5 ) @pytest.mark.parametrize("dtype", [torch.float16, torch.bfloat16, torch.float32]) def test_shuffle_rows_dtype_consistency(dtype: torch.dtype): """Test that shuffle_rows preserves dtype correctly.""" if not current_platform.is_cuda(): pytest.skip("shuffle_rows requires CUDA") num_tokens = 64 hidden_size = 512 # Create input tensor with specific dtype input_tensor = torch.randn(num_tokens, hidden_size, device="cuda", dtype=dtype) dst2src_map = torch.arange(num_tokens, device="cuda", dtype=torch.int32) # Test shuffle_rows output = shuffle_rows(input_tensor, dst2src_map) # Verify dtype is preserved assert output.dtype == dtype assert output.device == input_tensor.device torch.testing.assert_close(output, input_tensor, atol=1e-6, rtol=1e-5) def test_shuffle_rows_device_consistency(): """Test that shuffle_rows maintains device consistency.""" if not current_platform.is_cuda(): pytest.skip("shuffle_rows requires CUDA") num_tokens = 32 hidden_size = 256 dtype = torch.float16 # Create input tensor on CUDA input_tensor = torch.randn(num_tokens, hidden_size, device="cuda", dtype=dtype) dst2src_map = torch.arange(num_tokens, device="cuda", dtype=torch.int32) # Test shuffle_rows output = shuffle_rows(input_tensor, dst2src_map) # Verify device is maintained assert output.device == input_tensor.device assert output.device.type == "cuda" def test_shuffle_rows_contiguous_output(): """Test that shuffle_rows produces contiguous output.""" if not current_platform.is_cuda(): pytest.skip("shuffle_rows requires CUDA") num_tokens = 64 hidden_size = 512 dtype = torch.float16 # Create input tensor input_tensor = torch.randn(num_tokens, hidden_size, device="cuda", dtype=dtype) dst2src_map = torch.arange(num_tokens, device="cuda", dtype=torch.int32) # Test shuffle_rows output = shuffle_rows(input_tensor, dst2src_map) # Verify output is contiguous assert output.is_contiguous()
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/test_shuffle_rows.py", "license": "Apache License 2.0", "lines": 194, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/model_executor/model_loader/test_registry.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from torch import nn from vllm.config import ModelConfig from vllm.config.load import LoadConfig from vllm.model_executor.model_loader import get_model_loader, register_model_loader from vllm.model_executor.model_loader.base_loader import BaseModelLoader @register_model_loader("custom_load_format") class CustomModelLoader(BaseModelLoader): def __init__(self, load_config: LoadConfig) -> None: super().__init__(load_config) def download_model(self, model_config: ModelConfig) -> None: pass def load_weights(self, model: nn.Module, model_config: ModelConfig) -> None: pass def test_register_model_loader(): load_config = LoadConfig(load_format="custom_load_format") assert isinstance(get_model_loader(load_config), CustomModelLoader) def test_invalid_model_loader(): with pytest.raises(ValueError): @register_model_loader("invalid_load_format") class InValidModelLoader: pass
{ "repo_id": "vllm-project/vllm", "file_path": "tests/model_executor/model_loader/test_registry.py", "license": "Apache License 2.0", "lines": 24, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/reasoning/test_mistral_reasoning_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from tests.reasoning.utils import run_reasoning_extraction_mistral from vllm.reasoning import ReasoningParser, ReasoningParserManager from vllm.tokenizers.mistral import MistralTokenizer parser_name = "mistral" @pytest.fixture(scope="module") def mistral_tokenizer(): mistral_tokenizer = MistralTokenizer.from_pretrained( "mistralai/Magistral-Small-2509" ) return mistral_tokenizer INVALID_SIMPLE_REASONING = { "output": "This is a reasoning section[/THINK]This is the rest", "reasoning": None, "content": "This is a reasoning sectionThis is the rest", "is_reasoning_end": False, } INVALID_COMPLETE_REASONING = { "output": "This is a reasoning section[/THINK]", "reasoning": None, "content": "This is a reasoning section", "is_reasoning_end": False, } NO_CONTENT = { "output": "[THINK]This is reasoning", "reasoning": "This is reasoning", "content": None, "is_reasoning_end": False, } NO_REASONING = { "output": "This is content", "reasoning": None, "content": "This is content", "is_reasoning_end": False, } NO_REASONING_STREAMING = { "output": "This is a reasoning section", "reasoning": None, "content": "This is a reasoning section", "is_reasoning_end": False, } INVALID_MULTIPLE_LINES = { "output": "This\nThat[/THINK]This is the rest\nThat", "reasoning": None, "content": "This\nThatThis is the rest\nThat", "is_reasoning_end": False, } INVALID_SHORTEST_REASONING_NO_STREAMING = { "output": "[/THINK]This is the rest", "reasoning": None, "content": "This is the rest", "is_reasoning_end": False, } INVALID_SHORTEST_REASONING = { "output": "[/THINK]This is the rest", "reasoning": None, "content": "This is the rest", "is_reasoning_end": False, } REASONING_WITH_THINK = { "output": "[THINK]This is a reasoning section[/THINK]This is the rest", "reasoning": "This is a reasoning section", "content": "This is the rest", "is_reasoning_end": True, } COMPLETE_REASONING_WITH_THINK = { "output": "[THINK]This is a reasoning section[/THINK]", "reasoning": "This is a reasoning section", "content": None, "is_reasoning_end": True, } MULTIPLE_LINES_WITH_THINK = { "output": "[THINK]This\nThat[/THINK]This is the rest\nThat", "reasoning": "This\nThat", "content": "This is the rest\nThat", "is_reasoning_end": True, } INVALID_SHORTEST_REASONING_NO_STREAMING_WITH_THINK = { "output": "[/THINK]This is the rest", "reasoning": None, "content": "This is the rest", "is_reasoning_end": False, } INVALID_SHORTEST_REASONING_WITH_THINK = { "output": "[/THINK]This is the rest", "reasoning": None, "content": "This is the rest", "is_reasoning_end": False, } THINK_NO_END = { "output": "[THINK]This is a reasoning section", "reasoning": "This is a reasoning section", "content": None, "is_reasoning_end": False, } EMPTY = { "output": "", "reasoning": None, "content": "", "is_reasoning_end": False, } EMPTY_STREAMING = { "output": "", "reasoning": None, "content": None, "is_reasoning_end": False, } NEW_LINE = { "output": "Before\n[THINK]This is a reasoning section[/THINK]\nThis is the rest", "reasoning": "This is a reasoning section", "content": "Before\n\nThis is the rest", "is_reasoning_end": True, } NEW_LINE_STREAMING = { "output": "Before\n[THINK]This is a reasoning section[/THINK]\nThis is the rest", "reasoning": "This is a reasoning section", "content": "Before\n\nThis is the rest", "is_reasoning_end": True, } TEST_CASES = [ pytest.param( False, INVALID_SIMPLE_REASONING, id="invalid_simple_reasoning", ), pytest.param( True, INVALID_SIMPLE_REASONING, id="invalid_simple_reasoning_streaming", ), pytest.param( False, INVALID_COMPLETE_REASONING, id="invalid_complete_reasoning", ), pytest.param( True, INVALID_COMPLETE_REASONING, id="invalid_complete_reasoning_streaming", ), pytest.param( False, NO_CONTENT, id="no_content", ), pytest.param( False, NO_REASONING, id="no_reasoning", ), pytest.param( True, NO_REASONING_STREAMING, id="no_reasoning_token_streaming", ), pytest.param( False, INVALID_MULTIPLE_LINES, id="invalid_multiple_lines", ), pytest.param( True, INVALID_MULTIPLE_LINES, id="invalid_multiple_lines_streaming", ), pytest.param( True, INVALID_SHORTEST_REASONING, id="invalid_shortest", ), pytest.param( False, INVALID_SHORTEST_REASONING_NO_STREAMING, id="invalid_shortest_streaming", ), pytest.param( False, REASONING_WITH_THINK, id="reasoning_with_think", ), pytest.param( True, REASONING_WITH_THINK, id="reasoning_with_think_streaming", ), pytest.param( False, COMPLETE_REASONING_WITH_THINK, id="complete_reasoning_with_think", ), pytest.param( True, COMPLETE_REASONING_WITH_THINK, id="complete_reasoning_with_think_streaming", ), pytest.param( False, MULTIPLE_LINES_WITH_THINK, id="multiple_lines_with_think", ), pytest.param( True, MULTIPLE_LINES_WITH_THINK, id="multiple_lines_with_think_streaming", ), pytest.param( False, INVALID_SHORTEST_REASONING_NO_STREAMING_WITH_THINK, id="invalid_shortest_with_think", ), pytest.param( True, INVALID_SHORTEST_REASONING_WITH_THINK, id="invalid_shortest_with_think_streaming", ), pytest.param( False, THINK_NO_END, id="think_no_end", ), pytest.param( True, THINK_NO_END, id="think_no_end_streaming", ), pytest.param( False, EMPTY, id="empty", ), pytest.param( True, EMPTY_STREAMING, id="empty_streaming", ), pytest.param( False, NEW_LINE, id="new_line", ), pytest.param( True, NEW_LINE_STREAMING, id="new_line_streaming", ), ] @pytest.mark.parametrize("streaming, param_dict", TEST_CASES) def test_mistral_reasoning( streaming: bool, param_dict: dict, mistral_tokenizer: MistralTokenizer, ): output = param_dict["output"] index_think = output.find("[THINK]") len_think = len("[THINK]") index_end_think = output.find("[/THINK]") len_end_think = len("[/THINK]") # encode everything to tokens ids output_tokens = [] if index_think != -1: output_before_think = output[:index_think] output_tokens += mistral_tokenizer.tokenizer.encode( output_before_think, False, False ) output_tokens += [mistral_tokenizer.instruct.BEGIN_THINK] if index_end_think != -1: output_middle = output[index_think + len_think : index_end_think] output_after_think = output[index_end_think + len_end_think :] output_tokens += mistral_tokenizer.tokenizer.encode( output_middle, False, False ) output_tokens += [mistral_tokenizer.instruct.END_THINK] output_tokens += mistral_tokenizer.tokenizer.encode( output_after_think, False, False ) else: output_middle = output[index_think + len_think :] output_tokens += mistral_tokenizer.tokenizer.encode( output_middle, False, False ) elif index_end_think != -1: output_before_think = output[:index_end_think] output_after_think = output[index_end_think + len_end_think :] output_tokens += mistral_tokenizer.tokenizer.encode( output_before_think, False, False ) output_tokens += [mistral_tokenizer.instruct.END_THINK] output_tokens += mistral_tokenizer.tokenizer.encode( output_after_think, False, False ) else: output_tokens += mistral_tokenizer.tokenizer.encode(output, False, False) parser: ReasoningParser = ReasoningParserManager.get_reasoning_parser(parser_name)( mistral_tokenizer ) reasoning, content = run_reasoning_extraction_mistral( parser, output_tokens, streaming=streaming ) assert reasoning == param_dict["reasoning"] assert content == param_dict["content"] # Test is_reasoning_end is_reasoning_end = parser.is_reasoning_end(output_tokens) assert is_reasoning_end == param_dict["is_reasoning_end"] # Test extract_content if param_dict["content"] is not None: # Handle the case where there are tokens outputted before Thinking. # This should not occur if the model is well trained and prompted. if "[THINK]" in param_dict["output"] and not param_dict["output"].startswith( "[THINK]" ): before_content = param_dict["output"].split("[THINK]")[0] before_token_ids = mistral_tokenizer.tokenizer.encode( before_content, bos=False, eos=False ) left_to_encode = param_dict["content"][len(before_content) :] # Normal situation. else: before_token_ids = [] left_to_encode = param_dict["content"] content_tokens = parser.extract_content_ids(output_tokens) expected_token_ids = before_token_ids + mistral_tokenizer.tokenizer.encode( left_to_encode, bos=False, eos=False ) assert content_tokens == expected_token_ids else: content = parser.extract_content_ids(output_tokens) assert content == []
{ "repo_id": "vllm-project/vllm", "file_path": "tests/reasoning/test_mistral_reasoning_parser.py", "license": "Apache License 2.0", "lines": 329, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/reasoning/mistral_reasoning_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Sequence from functools import cached_property from vllm.entrypoints.openai.chat_completion.protocol import ( ChatCompletionRequest, ) from vllm.entrypoints.openai.responses.protocol import ( ResponsesRequest, ) from vllm.logger import init_logger from vllm.reasoning import ReasoningParser from vllm.reasoning.basic_parsers import BaseThinkingReasoningParser from vllm.tokenizers.mistral import MistralTokenizer logger = init_logger(__name__) class MistralReasoningParser(BaseThinkingReasoningParser): """ Reasoning parser for Mistral models. The Mistral models uses `[THINK]`...`[/THINK]` tokens to denote reasoning text. This parser extracts the reasoning content from the model output. A valid reasoning trace should always start with a `[THINK]` token and end with a `[/THINK]` token. If `[THINK]` token is not generated, then this parser only returns content. """ def __init__(self, tokenizer: MistralTokenizer, *args, **kwargs): if not isinstance(tokenizer, MistralTokenizer): raise ValueError("The tokenizer must be an instance of MistralTokenizer.") ReasoningParser.__init__(self, tokenizer, *args, **kwargs) if not self.model_tokenizer: raise ValueError( "The model tokenizer must be passed to the ReasoningParser " "constructor during construction." ) self.start_token_id = tokenizer.tokenizer.get_special_token(self.start_token) self.end_token_id = tokenizer.tokenizer.get_special_token(self.end_token) if self.start_token_id is None or self.end_token_id is None: raise RuntimeError( "Mistral reasoning parser could not locate think start/end " "tokens in the tokenizer!" ) @cached_property def start_token(self) -> str: """The token that starts reasoning content.""" from mistral_common.tokens.tokenizers.base import SpecialTokens return SpecialTokens.begin_think @cached_property def end_token(self) -> str: """The token that ends reasoning content.""" from mistral_common.tokens.tokenizers.base import SpecialTokens return SpecialTokens.end_think def is_reasoning_end(self, input_ids: Sequence[int]) -> bool: has_eot_token = False for id in reversed(input_ids): if id == self.start_token_id: # Reasoning ends only if a BOT token is found before a EOT token. return has_eot_token elif id == self.end_token_id: has_eot_token = True return False def extract_content_ids(self, input_ids: list[int]) -> list[int]: """ Extract the content """ has_bot_token = False has_eot_token = False bot_token_index = -1 eot_token_index = -1 # One for loop instead of multiple lookups for i, token_id in enumerate(input_ids): # We filter that we have multiple BOT tokens which should not # happen for a well prompted trained model if token_id == self.start_token_id and not has_bot_token: has_bot_token = True bot_token_index = i elif token_id == self.end_token_id: has_eot_token = True eot_token_index = i break # 1. Only BOT has been outputted if has_bot_token and not has_eot_token: # Should be = [] if model is well prompted and trained. return input_ids[:bot_token_index] # 2. Neither BOT or EOT have been outputted elif not has_bot_token and not has_eot_token: return input_ids # 3. Both BOT and EOT have been outputted. elif has_bot_token and has_eot_token: return input_ids[:bot_token_index] + input_ids[eot_token_index + 1 :] # 4. Only EOT has been outputted => this should not have occurred for a model # well prompted and trained. else: return input_ids[:eot_token_index] + input_ids[eot_token_index + 1 :] def extract_reasoning( self, model_output: str, request: ChatCompletionRequest | ResponsesRequest ) -> tuple[str | None, str | None]: """ Extract reasoning content from the model output. """ if not model_output: return (None, "") # Check if the start token is present in the model output, remove it # if it is present. prev_bot_token, bot_token, post_bot_token = model_output.partition( self.start_token ) has_bot_token = bool(bot_token) # Valid EOT tokens should follow BOT token has_valid_eot_token = has_bot_token and self.end_token in post_bot_token # 1. If there is BOT token followed by EOT token if has_bot_token and has_valid_eot_token: prev_eot_token, _, post_eot_token = post_bot_token.partition(self.end_token) # If model is well prompted and trained prev_bot_token should be "" content = prev_bot_token + post_eot_token return prev_eot_token, content if content else None # 2. Only BOT token elif has_bot_token: # If model is well prompted and trained prev_bot_token should be "" return post_bot_token, prev_bot_token if prev_bot_token else None # 3. EOT token has been outputted without BOT or neither has been outputted else: has_non_valid_eot_token = self.end_token in prev_bot_token # 3.a EOT token has been outputted without BOT # If model is well prompted and trained `has_non_valid_eot_token` should # be `False` and the parser outputs all tokens as 'content' if has_non_valid_eot_token: prev_eot_token, _, post_eot_token = prev_bot_token.partition( self.end_token ) return None, prev_eot_token + post_eot_token # 3.b neither BOT or EOT have been outputted else: return None, prev_bot_token
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/reasoning/mistral_reasoning_parser.py", "license": "Apache License 2.0", "lines": 132, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/models/multimodal/pooling/test_prithvi_mae.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from ....conftest import VllmRunner def _run_test( vllm_runner: type[VllmRunner], model: str, ) -> None: prompt = [ { # This model deals with no text input "prompt_token_ids": [1], "multi_modal_data": { "image": { "pixel_values": torch.ones((6, 512, 512), dtype=torch.float16), "location_coords": torch.ones((1, 2), dtype=torch.float16), } }, } for _ in range(10) ] with vllm_runner( model, runner="pooling", dtype="half", enforce_eager=True, skip_tokenizer_init=True, enable_mm_embeds=True, # Limit the maximum number of sequences to avoid the # test going OOM during the warmup run max_num_seqs=32, default_torch_num_threads=1, ) as vllm_model: vllm_model.llm.encode(prompt, pooling_task="plugin") MODELS = ["ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11"] @pytest.mark.core_model @pytest.mark.parametrize("model", MODELS) def test_models_image( hf_runner, vllm_runner, image_assets, model: str, ) -> None: _run_test( vllm_runner, model, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/multimodal/pooling/test_prithvi_mae.py", "license": "Apache License 2.0", "lines": 48, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/models/multimodal/processing/test_transformers.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from vllm.assets.image import ImageAsset from vllm.config import ModelConfig from vllm.multimodal import MULTIMODAL_REGISTRY @pytest.mark.parametrize("model_id", ["llava-hf/llava-onevision-qwen2-0.5b-ov-hf"]) def test_multimodal_processor(model_id): model_config = ModelConfig( model=model_id, model_impl="transformers", ) mm_processor = MULTIMODAL_REGISTRY.create_processor(model_config) image_pil = ImageAsset("cherry_blossom").pil_image mm_data = {"image": image_pil} str_prompt = "<|im_start|>user <image>\nWhat is the content of this image?<|im_end|><|im_start|>assistant\n" # noqa: E501 str_processed_inputs = mm_processor( prompt=str_prompt, mm_items=mm_processor.info.parse_mm_data(mm_data), hf_processor_mm_kwargs={}, ) ids_prompt = [ 151644, 872, 220, 151646, 198, 3838, 374, 279, 2213, 315, 419, 2168, 30, 151645, 151644, 77091, 198, ] ids_processed_inputs = mm_processor( prompt=ids_prompt, mm_items=mm_processor.info.parse_mm_data(mm_data), hf_processor_mm_kwargs={}, ) assert ( str_processed_inputs["prompt_token_ids"] == ids_processed_inputs["prompt_token_ids"] )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/multimodal/processing/test_transformers.py", "license": "Apache License 2.0", "lines": 49, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/kernels/quantization/test_per_token_group_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from unittest.mock import patch import pytest import torch from vllm.model_executor.layers.quantization.utils import fp8_utils, int8_utils @pytest.mark.parametrize( "shape", [(31, 128), (32, 128), (63, 256), (64, 256), (16, 512)] ) @pytest.mark.parametrize("column_major", [False, True]) @pytest.mark.parametrize("tma_aligned", [False, True]) @pytest.mark.parametrize("scale_ue8m0", [False, True]) @pytest.mark.parametrize("group_size", [64, 128]) @pytest.mark.skipif(not torch.cuda.is_available(), reason="CUDA not available") def test_per_token_group_quant_fp8( shape, column_major: bool, tma_aligned: bool, scale_ue8m0: bool, group_size: int ): device = "cuda" torch.manual_seed(42) num_tokens, hidden_dim = shape x = torch.randn((num_tokens, hidden_dim), device=device, dtype=torch.bfloat16) * 8 # cuda path out_q, scale = fp8_utils.per_token_group_quant_fp8( x, group_size, column_major_scales=column_major, tma_aligned_scales=tma_aligned, use_ue8m0=scale_ue8m0, ) # triton ref with patch("vllm.platforms.current_platform.is_cuda", return_value=False): ref_q, ref_s = fp8_utils.per_token_group_quant_fp8( x, group_size, column_major_scales=column_major, use_ue8m0=scale_ue8m0, ) assert torch.allclose(out_q.float(), ref_q.float(), atol=0.15, rtol=0.15) assert torch.allclose(scale, ref_s, atol=0.01, rtol=0.01) @pytest.mark.parametrize("shape", [(32, 128), (64, 256), (16, 512)]) @pytest.mark.parametrize("group_size", [64, 128]) @pytest.mark.skipif(not torch.cuda.is_available(), reason="CUDA not available") def test_per_token_group_quant_int8(shape, group_size: int): device = "cuda" torch.manual_seed(42) num_tokens, hidden_dim = shape x = torch.randn((num_tokens, hidden_dim), device=device, dtype=torch.bfloat16) * 8 # cuda path out_q, scale = int8_utils.per_token_group_quant_int8( x, group_size, ) # triton ref with patch("vllm.platforms.current_platform.is_cuda", return_value=False): ref_q, ref_s = int8_utils.per_token_group_quant_int8( x, group_size, ) assert torch.allclose(out_q.float(), ref_q.float(), atol=0.15, rtol=0.15) assert torch.allclose(scale, ref_s, atol=0.01, rtol=0.01)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/quantization/test_per_token_group_quant.py", "license": "Apache License 2.0", "lines": 60, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/models/arcee.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright 2023-2025 vLLM Team # Licensed under the Apache License, Version 2.0 (the "License"); # You may not use this file except in compliance with the License. # You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 # # Inference-only Arcee (AFM) model – adds support for ReLU^2 feed-forward # activation. from collections.abc import Iterable from itertools import islice from typing import Any import torch from torch import nn from transformers import LlamaConfig from vllm.compilation.decorators import support_torch_compile from vllm.distributed import get_pp_group from vllm.model_executor.layers.activation import ReLUSquaredActivation from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ColumnParallelLinear, RowParallelLinear from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import ( default_weight_loader, maybe_remap_kv_scale_name, ) from vllm.sequence import IntermediateTensors from .interfaces import SupportsLoRA, SupportsPP from .utils import ( AutoWeightsLoader, PPMissingLayer, is_pp_missing_parameter, make_empty_intermediate_tensors_factory, make_layers, ) class ArceeMLP(nn.Module): """Feed-forward layer for Arcee using ReLU^2 activation (no gating as in LLaMA).""" def __init__( self, hidden_size: int, intermediate_size: int, hidden_act: str, quant_config: Any | None = None, bias: bool = False, prefix: str = "", reduce_results: bool = True, ) -> None: super().__init__() # Single linear projection up to intermediate size # (no separate gate projection) self.up_proj = ColumnParallelLinear( input_size=hidden_size, output_size=intermediate_size, bias=bias, quant_config=quant_config, prefix=f"{prefix}.up_proj", ) # Down projection back to hidden size self.down_proj = RowParallelLinear( input_size=intermediate_size, output_size=hidden_size, bias=bias, quant_config=quant_config, reduce_results=reduce_results, prefix=f"{prefix}.down_proj", ) if hidden_act != "relu2": raise ValueError( f"Unsupported activation: {hidden_act}. " "Only 'relu2' is supported for AFM." ) # Define ReLU^2 activation: (ReLU(x))^2 elementwise self.act_fn = ReLUSquaredActivation() def forward(self, x: torch.Tensor) -> torch.Tensor: x, _ = self.up_proj(x) # Project to intermediate size x = self.act_fn(x) # Apply ReLU^2 activation elementwise x, _ = self.down_proj(x) # Project back down to hidden size return x class ArceeDecoderLayer(nn.Module): """Transformer decoder block for Arcee, with self-attention and ReLU^2 MLP.""" def __init__( self, config: LlamaConfig, cache_config: Any | None = None, quant_config: Any | None = None, prefix: str = "", ) -> None: super().__init__() self.hidden_size = config.hidden_size max_position_embeddings = getattr(config, "max_position_embeddings", 8192) # Determine if attention bias is needed (some variants use bias terms) attention_bias = getattr(config, "attention_bias", False) or getattr( config, "bias", False ) bias_o_proj = attention_bias if hasattr(config, "qkv_bias"): attention_bias = config.qkv_bias # Self-Attention (using LLaMA's attention structure) from vllm.model_executor.models.llama import ( LlamaAttention, # import here to avoid circular import ) self.self_attn = LlamaAttention( config=config, hidden_size=self.hidden_size, num_heads=config.num_attention_heads, num_kv_heads=getattr( config, "num_key_value_heads", config.num_attention_heads ), max_position_embeddings=max_position_embeddings, quant_config=quant_config, bias=attention_bias, bias_o_proj=bias_o_proj, cache_config=cache_config, prefix=f"{prefix}.self_attn", attn_type=getattr( config, "attn_type", "decoder" ), # assume decoder (causal) unless specified ) # MLP with ReLU^2 activation self.mlp = ArceeMLP( hidden_size=self.hidden_size, intermediate_size=config.intermediate_size, hidden_act=config.hidden_act, quant_config=quant_config, bias=getattr(config, "mlp_bias", False), prefix=f"{prefix}.mlp", ) # Layer normalization layers (RMSNorm as in LLaMA) self.input_layernorm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.post_attention_layernorm = RMSNorm( config.hidden_size, eps=config.rms_norm_eps ) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, residual: torch.Tensor | None, ) -> tuple[torch.Tensor, torch.Tensor]: # Self-Attention block if residual is None: residual = hidden_states hidden_states = self.input_layernorm(hidden_states) else: # Fused residual add + layernorm if supported hidden_states, residual = self.input_layernorm(hidden_states, residual) hidden_states = self.self_attn(positions=positions, hidden_states=hidden_states) # Feed-forward block hidden_states, residual = self.post_attention_layernorm(hidden_states, residual) hidden_states = self.mlp(hidden_states) return hidden_states, residual @support_torch_compile class ArceeModel(nn.Module): """The transformer model backbone for Arcee (embedding layer + stacked decoder blocks + final norm).""" def __init__( self, *, vllm_config, prefix: str = "", layer_type: type[nn.Module] = ArceeDecoderLayer, ) -> None: super().__init__() config: LlamaConfig = vllm_config.model_config.hf_config cache_config = vllm_config.cache_config quant_config = vllm_config.quant_config self.quant_config = quant_config self.config = config self.vocab_size = config.vocab_size # Word embeddings (parallelized if using pipeline parallel) if get_pp_group().is_first_rank or ( config.tie_word_embeddings and get_pp_group().is_last_rank ): self.embed_tokens = VocabParallelEmbedding( self.vocab_size, config.hidden_size, quant_config=quant_config, ) else: self.embed_tokens = PPMissingLayer() # placeholder on non-embedding ranks # Build decoder layers across pipeline ranks self.start_layer, self.end_layer, self.layers = make_layers( config.num_hidden_layers, lambda prefix: layer_type( config=config, cache_config=cache_config, quant_config=quant_config, prefix=prefix, ), prefix=f"{prefix}.layers", ) # Final RMSNorm on the last pipeline stage if get_pp_group().is_last_rank: self.norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) else: self.norm = PPMissingLayer() # For optional capturing of intermediate hidden states # (not used by default) self.aux_hidden_state_layers: tuple[int, ...] = tuple() # Prepare factory for empty intermediate tensors # (for pipeline scheduling) self.make_empty_intermediate_tensors = make_empty_intermediate_tensors_factory( ["hidden_states", "residual"], config.hidden_size ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor | IntermediateTensors | tuple[torch.Tensor, list[torch.Tensor]]: # Embedding lookup (on first pipeline rank) if get_pp_group().is_first_rank: hidden_states = ( inputs_embeds if inputs_embeds is not None else self.embed_input_ids(input_ids) ) residual = None else: assert intermediate_tensors is not None, ( "IntermediateTensors must be provided for non-first pipeline ranks" ) hidden_states = intermediate_tensors["hidden_states"] residual = intermediate_tensors["residual"] aux_hidden_states: list[torch.Tensor] = [] for idx, layer in enumerate( islice(self.layers, self.start_layer, self.end_layer) ): if idx in self.aux_hidden_state_layers: aux_hidden_states.append( hidden_states + residual ) # capture pre-layer hidden state if needed hidden_states, residual = layer(positions, hidden_states, residual) if not get_pp_group().is_last_rank: # Send intermediate results to the next pipeline stage return IntermediateTensors( {"hidden_states": hidden_states, "residual": residual} ) # On last rank: apply final layer norm hidden_states, _ = self.norm(hidden_states, residual) if len(aux_hidden_states) > 0: return hidden_states, aux_hidden_states return hidden_states def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: """Load weights, mapping q/k/v projections to fused qkv_proj.""" stacked_params_mapping = [ (".qkv_proj", ".q_proj", "q"), (".qkv_proj", ".k_proj", "k"), (".qkv_proj", ".v_proj", "v"), ] params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() for name, loaded_weight in weights: if "rotary_emb.inv_freq" in name: continue if "rotary_emb.cos_cached" in name or "rotary_emb.sin_cached" in name: continue if self.quant_config is not None and ( scale_name := self.quant_config.get_cache_scale(name) ): param = params_dict[scale_name] weight_loader = getattr(param, "weight_loader", default_weight_loader) loaded_weight = ( loaded_weight if loaded_weight.dim() == 0 else loaded_weight[0] ) weight_loader(param, loaded_weight) loaded_params.add(scale_name) continue if "scale" in name or "zero_point" in name: remapped_name = maybe_remap_kv_scale_name(name, params_dict) if remapped_name is None: continue name = remapped_name mapped = False for param_name, weight_name, shard_id in stacked_params_mapping: if weight_name not in name: continue name = name.replace(weight_name, param_name) if name.endswith(".bias") and name not in params_dict: mapped = True break if is_pp_missing_parameter(name, self): mapped = True break param = params_dict[name] weight_loader = param.weight_loader # type: ignore[attr-defined] weight_loader(param, loaded_weight, shard_id) loaded_params.add(name) mapped = True break if mapped: continue if name.endswith(".bias") and name not in params_dict: continue if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params class ArceeForCausalLM(nn.Module, SupportsLoRA, SupportsPP): """Arcee Model for causal language modeling, integrated with vLLM runtime.""" # Map fused module names to their submodule components # (for quantization and LoRA) packed_modules_mapping = { "qkv_proj": ["q_proj", "k_proj", "v_proj"], } def __init__(self, *, vllm_config, prefix: str = "") -> None: super().__init__() config = vllm_config.model_config.hf_config self.config = config # Initialize the inner Transformer model (ArceeModel) self.model = ArceeModel(vllm_config=vllm_config, prefix=f"{prefix}.model") # On the last pipeline stage, set up the LM head and logits processor if get_pp_group().is_last_rank: # Determine vocabulary size (including any LoRA extra tokens # for padded LM head) self.lm_head = ParallelLMHead( config.vocab_size, config.hidden_size, quant_config=vllm_config.quant_config, bias=getattr(config, "lm_head_bias", False), prefix=f"{prefix}.lm_head", ) if config.tie_word_embeddings: # Tie output weights with input embedding matrix self.lm_head = self.lm_head.tie_weights(self.model.embed_tokens) logit_scale = getattr(config, "logit_scale", 1.0) self.logits_processor = LogitsProcessor( config.vocab_size, scale=logit_scale ) else: # Placeholder for lm_head on non-last ranks self.lm_head = PPMissingLayer() self.make_empty_intermediate_tensors = ( self.model.make_empty_intermediate_tensors ) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor | IntermediateTensors: model_output = self.model( input_ids=input_ids, positions=positions, intermediate_tensors=intermediate_tensors, inputs_embeds=inputs_embeds, ) return model_output def compute_logits(self, hidden_states: torch.Tensor) -> torch.Tensor | None: # Compute final logits from hidden states (last pipeline rank only) logits = self.logits_processor(self.lm_head, hidden_states) return logits def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: """Load weights into the model (delegates to inner model and handles tied embeddings).""" loader = AutoWeightsLoader( self, skip_prefixes=(["lm_head."] if self.config.tie_word_embeddings else None), skip_substrs=["gate_proj"], ) # AutoWeightLoader handles weight name remapping, including fusing # separate q_proj, k_proj, v_proj into qkv_proj return loader.load_weights(weights)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/arcee.py", "license": "Apache License 2.0", "lines": 377, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/sample/ops/logprobs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Some utilities for logprobs, including logits.""" import torch from vllm.platforms import current_platform @torch.compile(dynamic=True, backend=current_platform.simple_compile_backend) def batched_count_greater_than(x: torch.Tensor, values: torch.Tensor) -> torch.Tensor: """ Counts elements in each row of x that are greater than the corresponding value in values. Use torch.compile to generate an optimized kernel for this function. otherwise, it will create additional copies of the input tensors and cause memory issues. Args: x (torch.Tensor): A 2D tensor of shape (batch_size, n_elements). values (torch.Tensor): A 2D tensor of shape (batch_size, 1). Returns: torch.Tensor: A 1D tensor of shape (batch_size,) with the counts. """ return (x >= values).sum(-1)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/sample/ops/logprobs.py", "license": "Apache License 2.0", "lines": 19, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/models/multimodal/generation/test_maverick.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Create a reduced-layer version of the Maverick model for testing purposes. This script creates a new model with fewer layers by: 1. Loading the original Maverick model configuration 2. Creating a reduced configuration 3. Generating compatible safetensors files with appropriate weights 4. Creating the necessary index files for vLLM compatibility """ import json import shutil from pathlib import Path from typing import Any import pytest import torch from safetensors.torch import save_file from transformers import AutoConfig, AutoProcessor, AutoTokenizer, GenerationConfig from vllm import LLM, SamplingParams from vllm.v1.executor.abstract import Executor from vllm.v1.kv_cache_interface import ChunkedLocalAttentionSpec, FullAttentionSpec from ....utils import multi_gpu_test # Sample prompts for testing PROMPTS: list[str] = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] def run_maverick_serving(model: str): """Test Llama-4-Maverick model with vLLM LLM class using CLI equivalent options with reduced layers. """ try: sampling_params = SamplingParams(temperature=0.8, top_p=0.95) llm = LLM( model=model, max_model_len=2048, enforce_eager=True, tensor_parallel_size=8, enable_expert_parallel=True, trust_remote_code=True, gpu_memory_utilization=0.4, kv_cache_dtype="fp8", ) outputs = llm.generate(PROMPTS, sampling_params) # Print the outputs print("\nGenerated Outputs:\n" + "-" * 60) for output in outputs: prompt = output.prompt generated_text = output.outputs[0].text print(f"Prompt: {prompt!r}") print(f"Output: {generated_text!r}") print("-" * 60) except Exception as e: print(f"Error initializing or running model: {e}") raise def get_rope_layers_config(model_path: str) -> list[int]: """ Get the interleaved RoPE configuration from HuggingFace config Args: model_path: Path to the local directory containing the reduced Maverick model checkpoint Returns: List of 0 or 1 indicating whether each layer uses RoPE and local attn 0 indicates that RoPE is not used while 1 indicates that RoPE is used. """ config_path = Path(model_path) / "config.json" model_config = json.loads(config_path.read_text()) text_config = model_config["text_config"] no_rope_layers = text_config["no_rope_layers"] print(f"Found no_rope_layers: {no_rope_layers}") return no_rope_layers def create_reduced_maverick_model( original_model_name: str = "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8", output_dir: str = "/tmp/reduced_maverick", text_layers: int = 4, num_experts: int = 4, vision_layers: int = 2, force_recreate: bool = False, ) -> str: """ Create a reduced-layer version of the Maverick model. Args: original_model_name: Name of the original Maverick model output_dir: Directory to save the reduced model text_layers: Number of text transformer layers num_experts: Number of experts per layer vision_layers: Number of vision transformer layers force_recreate: Whether to recreate if output_dir already exists Returns: Path to the created reduced model directory """ print( f"Creating reduced Maverick model with {text_layers} text layers and " f"{vision_layers} vision layers..." ) # Create output directory output_path = Path(output_dir) if output_path.exists(): if force_recreate: shutil.rmtree(output_path) else: print( f"Output directory {output_dir} already exists. " "Use --force-recreate to overwrite." ) return str(output_path) output_path.mkdir(parents=True, exist_ok=True) try: print("Loading original model configuration...") original_config = AutoConfig.from_pretrained( original_model_name, trust_remote_code=True ) print("Creating reduced configuration...") reduced_config = create_reduced_config( original_config, text_layers, num_experts, vision_layers ) config_path = output_path / "config.json" with open(config_path, "w") as f: json.dump(reduced_config, f, indent=2) print(f"Saved reduced config to {config_path}") print("Copying tokenizer files...") copy_tokenizer_files(original_model_name, output_path) print("Creating reduced safetensors files...") create_reduced_safetensors(original_config, reduced_config, output_path) print("Creating preprocessor config...") create_preprocessor_config(original_config, output_path) try: gen_config = GenerationConfig.from_pretrained(original_model_name) gen_config.save_pretrained(output_path) print("Copied generation config") except Exception as e: print(f"Could not copy generation config: {e}") print(f"Successfully created reduced Maverick model at {output_path}") return str(output_path) except Exception as e: print(f"Error creating reduced model: {e}") # Clean up on failure if output_path.exists(): shutil.rmtree(output_path) raise def create_reduced_config( original_config: Any, text_layers: int, num_experts: int, vision_layers: int ) -> dict[str, Any]: """Create a reduced configuration based on the original.""" # Convert config to dictionary config_dict = original_config.to_dict() # Reduce text layers if "text_config" in config_dict: original_text_layers = config_dict["text_config"]["num_hidden_layers"] config_dict["text_config"]["num_hidden_layers"] = text_layers original_layer_types = config_dict["text_config"]["layer_types"] config_dict["text_config"]["layer_types"] = original_layer_types[:text_layers] print(f"Reduced text layers from {original_text_layers} to {text_layers}") original_num_experts = config_dict["text_config"]["num_local_experts"] config_dict["text_config"]["num_local_experts"] = num_experts print(f"Reduced num experts from {original_num_experts} to {num_experts}") hidden_dim_divisor = 4 original_hidden_size = config_dict["text_config"]["hidden_size"] new_hidden_size = original_hidden_size // hidden_dim_divisor config_dict["text_config"]["hidden_size"] = new_hidden_size print(f"Reduced hidden size from {original_hidden_size} to {new_hidden_size}") original_head_dim = config_dict["text_config"]["head_dim"] new_head_dim = original_head_dim // hidden_dim_divisor config_dict["text_config"]["head_dim"] = new_head_dim print(f"Reduced head dim from {original_head_dim} to {new_head_dim}") # Reduce vision layers if "vision_config" in config_dict: original_vision_layers = config_dict["vision_config"]["num_hidden_layers"] config_dict["vision_config"]["num_hidden_layers"] = vision_layers print(f"Reduced vision layers from {original_vision_layers} to {vision_layers}") # Update model name to indicate it's a reduced version config_dict["_name_or_path"] = f"reduced_maverick_{text_layers}t_{vision_layers}v" return config_dict def copy_tokenizer_files(original_model_name: str, output_path: Path) -> None: """Copy tokenizer files from the original model.""" try: tokenizer = AutoTokenizer.from_pretrained( original_model_name, trust_remote_code=True ) tokenizer.save_pretrained(output_path) print("Tokenizer files copied successfully") except Exception as e: print(f"Warning: Could not copy tokenizer files: {e}") def create_preprocessor_config(original_config: Any, output_path: Path) -> None: """Create preprocessor_config.json for multimodal model.""" # Try to load the original preprocessor config try: processor = AutoProcessor.from_pretrained( original_config._name_or_path or "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8", trust_remote_code=True, ) processor.save_pretrained(output_path) print("Copied original preprocessor config") return except Exception as e: print(f"Could not copy original preprocessor config: {e}") raise def create_reduced_safetensors( original_config: Any, reduced_config: dict[str, Any], output_path: Path ) -> None: """Create safetensors files with weights for the reduced model.""" print("Generating synthetic weights for reduced model...") text_config = reduced_config["text_config"] vision_config = reduced_config["vision_config"] weights = {} print("Creating text model weights...") weights.update(create_text_model_weights(text_config)) print("Creating vision model weights...") weights.update(create_vision_model_weights(vision_config)) print("Creating shared model weights...") weights.update(create_shared_weights(text_config, vision_config)) print("Saving weights to safetensors files...") save_weights_to_safetensors(weights, output_path) def create_text_model_weights(text_config: dict[str, Any]) -> dict[str, torch.Tensor]: """Create synthetic weights for the text model with MoE structure.""" weights = {} vocab_size = text_config["vocab_size"] hidden_size = text_config["hidden_size"] intermediate_size = text_config["intermediate_size"] intermediate_size_mlp = text_config["intermediate_size_mlp"] num_layers = text_config["num_hidden_layers"] num_attention_heads = text_config["num_attention_heads"] num_key_value_heads = text_config.get("num_key_value_heads", num_attention_heads) # MoE specific parameters num_experts = text_config.get("num_local_experts") assert num_experts is not None, "num_local_experts must be specified for MoE" head_dim = hidden_size // num_attention_heads # Embedding layers weights["language_model.model.embed_tokens.weight"] = torch.randn( vocab_size, hidden_size, dtype=torch.float16 ) # Transformer layers for layer_idx in range(num_layers): layer_prefix = f"language_model.model.layers.{layer_idx}" print(f"Creating weights for layer {layer_prefix}...") # Self-attention weights (separate q, k, v projections) weights[f"{layer_prefix}.self_attn.q_proj.weight"] = torch.randn( num_attention_heads * head_dim, hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.self_attn.k_proj.weight"] = torch.randn( num_key_value_heads * head_dim, hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.self_attn.v_proj.weight"] = torch.randn( num_key_value_heads * head_dim, hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.self_attn.o_proj.weight"] = torch.randn( hidden_size, num_attention_heads * head_dim, dtype=torch.bfloat16 ) print("Self-attention weights created.") # Feed-forward weights - MoE pattern based on interleave_moe_layer_step # For interleave_moe_layer_step=2: layers 1,3,5,... are MoE, layers # 0,2,4,... are dense interleave_step = text_config.get("interleave_moe_layer_step", 1) is_moe_layer = interleave_step > 0 and (layer_idx + 1) % interleave_step == 0 if is_moe_layer: # MoE layer structure # 1. Router weights weights[f"{layer_prefix}.feed_forward.router.weight"] = torch.randn( num_experts, hidden_size, dtype=torch.float16 ) # 2. Individual expert weights (not fused) for expert_idx in range(num_experts): expert_prefix = f"{layer_prefix}.feed_forward.experts.{expert_idx}" weights[f"{expert_prefix}.gate_proj.weight"] = torch.randn( intermediate_size, hidden_size, dtype=torch.bfloat16 ) weights[f"{expert_prefix}.up_proj.weight"] = torch.randn( intermediate_size, hidden_size, dtype=torch.bfloat16 ) weights[f"{expert_prefix}.down_proj.weight"] = torch.randn( hidden_size, intermediate_size, dtype=torch.bfloat16 ) # Expert weight scales (FP8 quantization) weights[f"{expert_prefix}.gate_proj.weight_scale"] = torch.ones( intermediate_size, 1, dtype=torch.bfloat16 ) weights[f"{expert_prefix}.up_proj.weight_scale"] = torch.ones( intermediate_size, 1, dtype=torch.bfloat16 ) weights[f"{expert_prefix}.down_proj.weight_scale"] = torch.ones( hidden_size, 1, dtype=torch.bfloat16 ) # 3. Shared expert weights shared_expert_prefix = f"{layer_prefix}.feed_forward.shared_expert" weights[f"{shared_expert_prefix}.gate_proj.weight"] = torch.randn( intermediate_size, hidden_size, dtype=torch.bfloat16 ) weights[f"{shared_expert_prefix}.up_proj.weight"] = torch.randn( intermediate_size, hidden_size, dtype=torch.bfloat16 ) weights[f"{shared_expert_prefix}.down_proj.weight"] = torch.randn( hidden_size, intermediate_size, dtype=torch.bfloat16 ) print(f"MoE feed-forward weights created for layer {layer_idx}.") else: # Dense layer structure weights[f"{layer_prefix}.feed_forward.gate_proj.weight"] = torch.randn( intermediate_size_mlp, hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.feed_forward.up_proj.weight"] = torch.randn( intermediate_size_mlp, hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.feed_forward.down_proj.weight"] = torch.randn( hidden_size, intermediate_size_mlp, dtype=torch.bfloat16 ) print(f"Dense feed-forward weights created for layer {layer_idx}.") # Layer norms weights[f"{layer_prefix}.input_layernorm.weight"] = torch.ones( hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.post_attention_layernorm.weight"] = torch.ones( hidden_size, dtype=torch.bfloat16 ) print("Layer norms created.") # Final layer norm and output projection weights["language_model.model.norm.weight"] = torch.ones( hidden_size, dtype=torch.bfloat16 ) weights["language_model.lm_head.weight"] = torch.randn( vocab_size, hidden_size, dtype=torch.bfloat16 ) return weights def create_vision_model_weights( vision_config: dict[str, Any], ) -> dict[str, torch.Tensor]: """Create synthetic weights for the vision model.""" weights = {} hidden_size = vision_config["hidden_size"] intermediate_size = vision_config["intermediate_size"] num_layers = vision_config["num_hidden_layers"] # Vision transformer layers for layer_idx in range(num_layers): layer_prefix = f"vision_model.model.layers.{layer_idx}" weights[f"{layer_prefix}.self_attn.q_proj.weight"] = torch.randn( hidden_size, hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.self_attn.q_proj.bias"] = torch.zeros( hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.self_attn.k_proj.weight"] = torch.randn( hidden_size, hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.self_attn.k_proj.bias"] = torch.zeros( hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.self_attn.v_proj.weight"] = torch.randn( hidden_size, hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.self_attn.v_proj.bias"] = torch.zeros( hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.self_attn.o_proj.weight"] = torch.randn( hidden_size, hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.self_attn.o_proj.bias"] = torch.zeros( hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.mlp.fc1.weight"] = torch.randn( intermediate_size, hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.mlp.fc1.bias"] = torch.zeros( intermediate_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.mlp.fc2.weight"] = torch.randn( hidden_size, intermediate_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.mlp.fc2.bias"] = torch.zeros( hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.input_layernorm.weight"] = torch.ones( hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.input_layernorm.bias"] = torch.zeros( hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.post_attention_layernorm.weight"] = torch.ones( hidden_size, dtype=torch.bfloat16 ) weights[f"{layer_prefix}.post_attention_layernorm.bias"] = torch.zeros( hidden_size, dtype=torch.bfloat16 ) return weights def create_shared_weights( text_config: dict[str, Any], vision_config: dict[str, Any] ) -> dict[str, torch.Tensor]: """Create weights for shared components (vision-language connector)""" weights = {} text_hidden_size = text_config["hidden_size"] projector_input_dim = vision_config["projector_input_dim"] # Vision-language connector (projects vision features to text space) weights["multi_modal_projector.linear_1.weight"] = torch.randn( text_hidden_size, projector_input_dim, dtype=torch.bfloat16 ) return weights def save_weights_to_safetensors( weights: dict[str, torch.Tensor], output_path: Path ) -> None: """Save weights to safetensors files and create index.""" # Determine how to shard the weights max_shard_size = 5 * 1024 * 1024 * 1024 # 5GB per shard # Calculate sizes and create shards shards = [] current_shard: dict[str, torch.Tensor] = {} current_size = 0 for name, tensor in weights.items(): tensor_size = tensor.numel() * tensor.element_size() if current_size + tensor_size > max_shard_size and current_shard: shards.append(current_shard) current_shard = {} current_size = 0 current_shard[name] = tensor current_size += tensor_size if current_shard: shards.append(current_shard) # Save shards and create index weight_map = {} if len(shards) == 1: # Single file filename = "model.safetensors" save_file(shards[0], output_path / filename) weight_map = {name: filename for name in shards[0]} print(f"Saved weights to single file: {filename}") else: # Multiple shards for i, shard in enumerate(shards): filename = f"model-{i + 1:05d}-of-{len(shards):05d}.safetensors" save_file(shard, output_path / filename) for name in shard: weight_map[name] = filename print(f"Saved shard {i + 1}/{len(shards)}: {filename}") # Create index file index_data = { "metadata": { "total_size": sum( tensor.numel() * tensor.element_size() for tensor in weights.values() ) }, "weight_map": weight_map, } index_path = output_path / "model.safetensors.index.json" with open(index_path, "w") as f: json.dump(index_data, f, indent=2) print(f"Created index file: {index_path}") print( f"Total model size: {index_data['metadata']['total_size'] / (1024**3):.2f} GB" ) def check_attention_spec_interleaved_rope( llm: LLM, num_attention_layers: int, num_ranks: int, rope_layers: list[int], ): """Check that the attention spec is correct.""" assert isinstance(llm.llm_engine.model_executor, Executor) kv_cache_specs_per_rank = llm.llm_engine.model_executor.get_kv_cache_specs() for rank in range(num_ranks): kv_cache_specs = kv_cache_specs_per_rank[rank] assert len(kv_cache_specs.keys()) == num_attention_layers for i in range(num_attention_layers): if rope_layers[i] == 0: expected_spec = FullAttentionSpec else: expected_spec = ChunkedLocalAttentionSpec assert isinstance( kv_cache_specs[f"language_model.model.layers.{i}.self_attn.attn"], expected_spec, ) def run_reduced_model(llm: LLM, should_profile: bool = False) -> None: """Test the created reduced model with vLLM.""" sampling_params = SamplingParams(temperature=0.8, top_p=0.95, max_tokens=50) if should_profile: llm.start_profile() outputs = llm.generate(PROMPTS, sampling_params) if should_profile: llm.stop_profile() print("Test generation successful!") for output in outputs: print(f"Prompt: {output.prompt}") print(f"Output: {output.outputs[0].text}") print("-" * 40) @multi_gpu_test(num_gpus=2) @pytest.mark.parametrize( "original_model_name,text_layers,num_experts,vision_layers,", [("meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8", 4, 4, 2)], ) @pytest.mark.parametrize("enforce_eager", [True, False]) @pytest.mark.parametrize("tp,ep", [(2, True)]) @pytest.mark.skipif(not torch.cuda.is_available(), reason="CUDA not available") def test_dummy_maverick( monkeypatch, original_model_name: str, text_layers: int, num_experts: int, vision_layers: int, enforce_eager: bool, tp: int, ep: bool, output_dir: str = "/tmp/reduced_maverick", force_recreate: bool = True, profile: bool = False, ) -> None: # Disable multiprocessing allows us to access model executor from LLM engine monkeypatch.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "0") model_path = create_reduced_maverick_model( original_model_name=original_model_name, output_dir=output_dir, text_layers=text_layers, num_experts=num_experts, vision_layers=vision_layers, force_recreate=force_recreate, ) print(f"\nReduced model created successfully at: {model_path}") rope_layers = get_rope_layers_config(model_path) llm = LLM( model=model_path, trust_remote_code=True, max_model_len=512, # Small context for testing gpu_memory_utilization=0.3, # Conservative memory usage enforce_eager=enforce_eager, tensor_parallel_size=tp, enable_expert_parallel=ep, ) check_attention_spec_interleaved_rope( llm, text_layers, tp, rope_layers, ) print(f"\nTesting reduced model at {model_path}...") run_reduced_model(llm=llm, should_profile=profile) def main(): """Main function to create and test the reduced model.""" import argparse parser = argparse.ArgumentParser( description="Create a reduced-layer Maverick model" ) parser.add_argument( "--output-dir", default="/tmp/reduced_maverick", help="Output directory for the reduced model", ) parser.add_argument( "--text-layers", type=int, default=4, help="Number of text transformer layers", ) parser.add_argument("--num-experts", type=int, default=4, help="Number of experts") parser.add_argument( "--vision-layers", type=int, default=2, help="Number of vision transformer layers", ) parser.add_argument( "--force-recreate", action="store_true", help="Force recreation if output directory exists", ) parser.add_argument( "--test", action="store_true", help="Test the created model with vLLM" ) parser.add_argument( "--profile", action="store_true", help="Profile the created model with vLLM" ) parser.add_argument( "--test-original", action="store_true", help="Test the original model with vLLM", ) parser.add_argument( "--original-model", default="meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8", help="Original model name to base the reduction on", ) args = parser.parse_args() if args.test: test_dummy_maverick( original_model_name=args.original_model, output_dir=args.output_dir, text_layers=args.text_layers, num_experts=args.num_experts, vision_layers=args.vision_layers, force_recreate=args.force_recreate, tp=2, ep=True, enforce_eager=True, profile=args.profile, ) if args.test_original: run_maverick_serving(args.original_model) if __name__ == "__main__": exit(main())
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/multimodal/generation/test_maverick.py", "license": "Apache License 2.0", "lines": 596, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/quantization/test_modelopt.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Test ModelOpt quantization method setup and weight loading. Run `pytest tests/quantization/test_modelopt.py`. """ import os from typing import NoReturn import pytest import torch from tests.quantization.utils import is_quant_method_supported @pytest.fixture(scope="function", autouse=True) def enable_pickle(monkeypatch): """`LLM.apply_model` requires pickling a function.""" monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") def _skip(msg: str) -> NoReturn: pytest.skip(msg) raise RuntimeError(msg) def _snapshot_download_or_skip(model_id: str) -> str: try: from huggingface_hub import snapshot_download except Exception as e: # pragma: no cover _skip(f"huggingface_hub is required to download {model_id}: {e}") try: return snapshot_download( repo_id=model_id, repo_type="model", # These checkpoints are already small; download full repo for simplicity. allow_patterns=["*"], ) except Exception as e: _skip(f"Failed to download {model_id} from the HF Hub: {e}") @pytest.mark.skipif( not is_quant_method_supported("modelopt"), reason="ModelOpt FP8 is not supported on this GPU type.", ) def test_modelopt_fp8_checkpoint_setup(vllm_runner): """Test ModelOpt FP8 checkpoint loading and structure validation.""" # TODO: provide a small publicly available test checkpoint model_path = ( "/home/scratch.omniml_data_1/zhiyu/ckpts/test_ckpts/" "TinyLlama-1.1B-Chat-v1.0-fp8-0710" ) # Skip test if checkpoint doesn't exist if not os.path.exists(model_path): pytest.skip( f"Test checkpoint not found at {model_path}. " "This test requires a local ModelOpt FP8 checkpoint." ) with vllm_runner(model_path, quantization="modelopt", enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj o_proj = layer.self_attn.o_proj gate_up_proj = layer.mlp.gate_up_proj down_proj = layer.mlp.down_proj # Check that ModelOpt quantization method is properly applied from vllm.model_executor.layers.quantization.modelopt import ( ModelOptFp8LinearMethod, ) assert isinstance(qkv_proj.quant_method, ModelOptFp8LinearMethod) assert isinstance(o_proj.quant_method, ModelOptFp8LinearMethod) assert isinstance(gate_up_proj.quant_method, ModelOptFp8LinearMethod) assert isinstance(down_proj.quant_method, ModelOptFp8LinearMethod) # Check weight dtype is FP8 assert qkv_proj.weight.dtype == torch.float8_e4m3fn assert o_proj.weight.dtype == torch.float8_e4m3fn assert gate_up_proj.weight.dtype == torch.float8_e4m3fn assert down_proj.weight.dtype == torch.float8_e4m3fn # Check scales are present and have correct dtype assert hasattr(qkv_proj, "weight_scale") assert hasattr(qkv_proj, "input_scale") assert qkv_proj.weight_scale.dtype == torch.float32 assert qkv_proj.input_scale.dtype == torch.float32 assert hasattr(o_proj, "weight_scale") assert hasattr(o_proj, "input_scale") assert o_proj.weight_scale.dtype == torch.float32 assert o_proj.input_scale.dtype == torch.float32 assert hasattr(gate_up_proj, "weight_scale") assert hasattr(gate_up_proj, "input_scale") assert gate_up_proj.weight_scale.dtype == torch.float32 assert gate_up_proj.input_scale.dtype == torch.float32 assert hasattr(down_proj, "weight_scale") assert hasattr(down_proj, "input_scale") assert down_proj.weight_scale.dtype == torch.float32 assert down_proj.input_scale.dtype == torch.float32 llm.apply_model(check_model) # Run a simple generation test to ensure the model works output = llm.generate_greedy(["Hello my name is"], max_tokens=4) assert output print(f"ModelOpt FP8 output: {output}") @pytest.mark.skipif( not is_quant_method_supported("modelopt"), reason="ModelOpt FP8 is not supported on this GPU type.", ) def test_modelopt_fp8_pc_pt_checkpoint_setup(vllm_runner): """Test ModelOpt FP8_PER_CHANNEL_PER_TOKEN checkpoint setup.""" model_id = "CedricHwang/qwen2.5-0.5b-modelopt-fp8-pc-pt" model_path = _snapshot_download_or_skip(model_id) with vllm_runner(model_path, quantization="modelopt", enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj o_proj = layer.self_attn.o_proj gate_up_proj = layer.mlp.gate_up_proj down_proj = layer.mlp.down_proj from vllm.model_executor.layers.quantization.modelopt import ( ModelOptFp8PcPtLinearMethod, ) assert isinstance(qkv_proj.quant_method, ModelOptFp8PcPtLinearMethod) assert isinstance(o_proj.quant_method, ModelOptFp8PcPtLinearMethod) assert isinstance(gate_up_proj.quant_method, ModelOptFp8PcPtLinearMethod) assert isinstance(down_proj.quant_method, ModelOptFp8PcPtLinearMethod) assert qkv_proj.weight.dtype == torch.float8_e4m3fn assert o_proj.weight.dtype == torch.float8_e4m3fn assert gate_up_proj.weight.dtype == torch.float8_e4m3fn assert down_proj.weight.dtype == torch.float8_e4m3fn # Per-channel scales; activations are dynamically scaled per token. assert hasattr(qkv_proj, "weight_scale") assert qkv_proj.weight_scale.dtype == torch.float32 assert qkv_proj.weight_scale.dim() == 1 assert not hasattr(qkv_proj, "input_scale") assert hasattr(o_proj, "weight_scale") assert o_proj.weight_scale.dtype == torch.float32 assert o_proj.weight_scale.dim() == 1 assert not hasattr(o_proj, "input_scale") assert hasattr(gate_up_proj, "weight_scale") assert gate_up_proj.weight_scale.dtype == torch.float32 assert gate_up_proj.weight_scale.dim() == 1 assert not hasattr(gate_up_proj, "input_scale") assert hasattr(down_proj, "weight_scale") assert down_proj.weight_scale.dtype == torch.float32 assert down_proj.weight_scale.dim() == 1 assert not hasattr(down_proj, "input_scale") llm.apply_model(check_model) output = llm.generate_greedy(["Hello my name is"], max_tokens=4) assert output print(f"ModelOpt FP8_PER_CHANNEL_PER_TOKEN output: {output}") @pytest.mark.skipif( not is_quant_method_supported("modelopt"), reason="ModelOpt FP8 is not supported on this GPU type.", ) def test_modelopt_fp8_pb_wo_checkpoint_setup(vllm_runner): """Test ModelOpt FP8_PB_WO checkpoint setup.""" model_id = "CedricHwang/qwen2.5-0.5b-modelopt-fp8-pb-wo" model_path = _snapshot_download_or_skip(model_id) with vllm_runner(model_path, quantization="modelopt", enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj o_proj = layer.self_attn.o_proj gate_up_proj = layer.mlp.gate_up_proj down_proj = layer.mlp.down_proj from vllm.model_executor.layers.quantization.modelopt import ( ModelOptFp8PbWoLinearMethod, ) assert isinstance(qkv_proj.quant_method, ModelOptFp8PbWoLinearMethod) assert isinstance(o_proj.quant_method, ModelOptFp8PbWoLinearMethod) assert isinstance(gate_up_proj.quant_method, ModelOptFp8PbWoLinearMethod) assert isinstance(down_proj.quant_method, ModelOptFp8PbWoLinearMethod) assert qkv_proj.weight.dtype == torch.float8_e4m3fn assert o_proj.weight.dtype == torch.float8_e4m3fn assert gate_up_proj.weight.dtype == torch.float8_e4m3fn assert down_proj.weight.dtype == torch.float8_e4m3fn # Block scales; should be materialized as a 2D [out_blk, in_blk] tensor. assert hasattr(qkv_proj, "weight_scale") assert qkv_proj.weight_scale.dtype == torch.float32 assert qkv_proj.weight_scale.dim() == 2 assert hasattr(o_proj, "weight_scale") assert o_proj.weight_scale.dtype == torch.float32 assert o_proj.weight_scale.dim() == 2 assert hasattr(gate_up_proj, "weight_scale") assert gate_up_proj.weight_scale.dtype == torch.float32 assert gate_up_proj.weight_scale.dim() == 2 assert hasattr(down_proj, "weight_scale") assert down_proj.weight_scale.dtype == torch.float32 assert down_proj.weight_scale.dim() == 2 llm.apply_model(check_model) output = llm.generate_greedy(["Hello my name is"], max_tokens=4) assert output print(f"ModelOpt FP8_PB_WO output: {output}")
{ "repo_id": "vllm-project/vllm", "file_path": "tests/quantization/test_modelopt.py", "license": "Apache License 2.0", "lines": 180, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/models/glm4_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright 2025 The ZhipuAI Team. # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Inference-only GLM-4.5, GLM-4.6, GLM-4.7 model compatible with HuggingFace weights.""" import typing from collections.abc import Callable, Iterable from itertools import islice import torch from torch import nn from transformers.models.glm4_moe import Glm4MoeConfig from vllm.compilation.decorators import support_torch_compile from vllm.config import CacheConfig, VllmConfig, get_current_vllm_config from vllm.distributed import ( get_ep_group, get_pp_group, get_tensor_model_parallel_world_size, ) from vllm.logger import init_logger from vllm.model_executor.layers.activation import SiluAndMul from vllm.model_executor.layers.attention import Attention from vllm.model_executor.layers.fused_moe import SharedFusedMoE from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( MergedColumnParallelLinear, QKVParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import ( default_weight_loader, maybe_remap_kv_scale_name, ) from vllm.sequence import IntermediateTensors from .interfaces import MixtureOfExperts, SupportsLoRA, SupportsPP from .utils import ( AutoWeightsLoader, PPMissingLayer, is_pp_missing_parameter, make_empty_intermediate_tensors_factory, make_layers, maybe_prefix, ) logger = init_logger(__name__) class Glm4MoeMLP(nn.Module): def __init__( self, hidden_size: int, intermediate_size: int, hidden_act: str, quant_config: QuantizationConfig | None = None, reduce_results: bool = True, prefix: str = "", ) -> None: super().__init__() self.gate_up_proj = MergedColumnParallelLinear( hidden_size, [intermediate_size] * 2, bias=False, quant_config=quant_config, prefix=f"{prefix}.gate_up_proj", ) self.down_proj = RowParallelLinear( intermediate_size, hidden_size, bias=False, quant_config=quant_config, reduce_results=reduce_results, prefix=f"{prefix}.down_proj", ) if hidden_act != "silu": raise ValueError( f"Unsupported activation: {hidden_act}. Only silu is supported for now." ) self.act_fn = SiluAndMul() def forward(self, x): gate_up, _ = self.gate_up_proj(x) x = self.act_fn(gate_up) x, _ = self.down_proj(x) return x class Glm4MoE(nn.Module): def __init__( self, config: Glm4MoeConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", enable_eplb: bool = False, ): super().__init__() self.tp_size = get_tensor_model_parallel_world_size() self.routed_scaling_factor = config.routed_scaling_factor self.ep_group = get_ep_group().device_group self.ep_rank = get_ep_group().rank_in_group self.ep_size = self.ep_group.size() self.n_routed_experts: int = config.n_routed_experts self.n_shared_experts: int = config.n_shared_experts if config.hidden_act != "silu": raise ValueError( f"Unsupported activation: {config.hidden_act}. " "Only silu is supported for now." ) # NOTE In the transformers implementation, the gate isn't an nn.Linear, # so we cannot use ReplicatedLinear here. # See: https://github.com/huggingface/transformers/blob/v4.55.1/src/transformers/models/glm4_moe/modeling_glm4_moe.py#L260 self.gate = nn.Linear( config.hidden_size, config.n_routed_experts, bias=False, dtype=torch.float32, ) self.gate.e_score_correction_bias = nn.Parameter( torch.empty(config.n_routed_experts, dtype=torch.float32) ) # Load balancing settings. vllm_config = get_current_vllm_config() eplb_config = vllm_config.parallel_config.eplb_config self.enable_eplb = enable_eplb self.n_redundant_experts = eplb_config.num_redundant_experts self.n_logical_experts = self.n_routed_experts self.n_physical_experts = self.n_logical_experts + self.n_redundant_experts self.n_local_physical_experts = self.n_physical_experts // self.ep_size self.physical_expert_start = self.ep_rank * self.n_local_physical_experts self.physical_expert_end = ( self.physical_expert_start + self.n_local_physical_experts ) if config.n_shared_experts is not None: intermediate_size = config.moe_intermediate_size * config.n_shared_experts self.shared_experts = Glm4MoeMLP( hidden_size=config.hidden_size, intermediate_size=intermediate_size, hidden_act=config.hidden_act, quant_config=quant_config, reduce_results=False, prefix=f"{prefix}.shared_experts", ) else: self.shared_experts = None self.experts = SharedFusedMoE( shared_experts=self.shared_experts, num_experts=config.n_routed_experts, top_k=config.num_experts_per_tok, hidden_size=config.hidden_size, intermediate_size=config.moe_intermediate_size, reduce_results=False, renormalize=config.norm_topk_prob, quant_config=quant_config, use_grouped_topk=True, num_expert_group=config.n_group, topk_group=config.topk_group, prefix=f"{prefix}.experts", scoring_func="sigmoid", # we do scaling outside, set factor to 1.0 to avoid double mul routed_scaling_factor=1.0, e_score_correction_bias=self.gate.e_score_correction_bias, enable_eplb=self.enable_eplb, num_redundant_experts=self.n_redundant_experts, router_logits_dtype=torch.float32, ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: num_tokens, hidden_dim = hidden_states.shape hidden_states = hidden_states.view(-1, hidden_dim) # router_logits: (num_tokens, n_experts) router_logits = self.gate(hidden_states.to(dtype=torch.float32)) fused_moe_out = self.experts( hidden_states=hidden_states, router_logits=router_logits ) if self.shared_experts is not None: shared_output, final_hidden_states = fused_moe_out assert shared_output is not None final_hidden_states = ( final_hidden_states * self.routed_scaling_factor + shared_output ) else: final_hidden_states = fused_moe_out * self.routed_scaling_factor if self.tp_size > 1: final_hidden_states = self.experts.maybe_all_reduce_tensor_model_parallel( final_hidden_states ) return final_hidden_states.view(num_tokens, hidden_dim) class Glm4MoeAttention(nn.Module): def __init__( self, config: Glm4MoeConfig, hidden_size: int, num_heads: int, num_kv_heads: int, max_position_embeddings: int = 131072, head_dim: int | None = None, rms_norm_eps: float = 1e-05, qkv_bias: bool = False, use_qk_norm: bool = False, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.hidden_size = hidden_size tp_size = get_tensor_model_parallel_world_size() self.total_num_heads = num_heads assert self.total_num_heads % tp_size == 0 self.num_heads = self.total_num_heads // tp_size self.total_num_kv_heads = num_kv_heads if self.total_num_kv_heads >= tp_size: # Number of KV heads is greater than TP size, so we partition # the KV heads across multiple tensor parallel GPUs. assert self.total_num_kv_heads % tp_size == 0 else: # Number of KV heads is less than TP size, so we replicate # the KV heads across multiple tensor parallel GPUs. assert tp_size % self.total_num_kv_heads == 0 self.num_kv_heads = max(1, self.total_num_kv_heads // tp_size) self.head_dim = head_dim or (hidden_size // self.total_num_heads) self.q_size = self.num_heads * self.head_dim self.kv_size = self.num_kv_heads * self.head_dim self.scaling = self.head_dim**-0.5 self.max_position_embeddings = max_position_embeddings self.use_qk_norm = use_qk_norm self.qkv_proj = QKVParallelLinear( hidden_size, self.head_dim, self.total_num_heads, self.total_num_kv_heads, bias=qkv_bias, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", ) self.o_proj = RowParallelLinear( self.total_num_heads * self.head_dim, hidden_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.o_proj", ) config.rope_parameters.setdefault("partial_rotary_factor", 0.5) self.rotary_emb = get_rope( self.head_dim, max_position=max_position_embeddings, rope_parameters=config.rope_parameters, ) self.attn = Attention( self.num_heads, self.head_dim, self.scaling, num_kv_heads=self.num_kv_heads, cache_config=cache_config, quant_config=quant_config, prefix=f"{prefix}.attn", ) if self.use_qk_norm: self.q_norm = RMSNorm(self.head_dim, eps=rms_norm_eps) self.k_norm = RMSNorm(self.head_dim, eps=rms_norm_eps) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, ) -> torch.Tensor: qkv, _ = self.qkv_proj(hidden_states) q, k, v = qkv.split([self.q_size, self.kv_size, self.kv_size], dim=-1) if self.use_qk_norm: q = self.q_norm(q.reshape(-1, self.num_heads, self.head_dim)).reshape( q.shape ) k = self.k_norm(k.reshape(-1, self.num_kv_heads, self.head_dim)).reshape( k.shape ) q, k = self.rotary_emb(positions, q, k) attn_output = self.attn(q, k, v) output, _ = self.o_proj(attn_output) return output class Glm4MoeDecoderLayer(nn.Module): def __init__( self, config: Glm4MoeConfig, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", enable_eplb: bool = False, ) -> None: super().__init__() self.hidden_size = config.hidden_size max_position_embeddings = getattr(config, "max_position_embeddings", 131072) # DecoderLayers are created with `make_layers` which passes the prefix # with the layer's index. layer_idx = int(prefix.split(sep=".")[-1]) self.layer_idx = layer_idx self.self_attn = Glm4MoeAttention( config=config, hidden_size=self.hidden_size, num_heads=config.num_attention_heads, num_kv_heads=config.num_key_value_heads, max_position_embeddings=max_position_embeddings, head_dim=config.head_dim, rms_norm_eps=config.rms_norm_eps, qkv_bias=config.attention_bias, cache_config=cache_config, quant_config=quant_config, prefix=f"{prefix}.self_attn", use_qk_norm=config.use_qk_norm, ) if ( config.n_routed_experts is not None and layer_idx >= config.first_k_dense_replace ): self.mlp = Glm4MoE( config=config, quant_config=quant_config, prefix=f"{prefix}.mlp", enable_eplb=enable_eplb, ) else: self.mlp = Glm4MoeMLP( hidden_size=config.hidden_size, intermediate_size=config.intermediate_size, hidden_act=config.hidden_act, quant_config=quant_config, prefix=f"{prefix}.mlp", ) self.input_layernorm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.post_attention_layernorm = RMSNorm( config.hidden_size, eps=config.rms_norm_eps ) self.routed_scaling_factor = config.routed_scaling_factor def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, residual: torch.Tensor | None, ) -> tuple[torch.Tensor, torch.Tensor]: if residual is None: residual = hidden_states hidden_states = self.input_layernorm(hidden_states) else: hidden_states, residual = self.input_layernorm(hidden_states, residual) hidden_states = self.self_attn(positions=positions, hidden_states=hidden_states) hidden_states, residual = self.post_attention_layernorm(hidden_states, residual) hidden_states = self.mlp(hidden_states) return hidden_states, residual @support_torch_compile( dynamic_arg_dims={ "input_ids": 0, "positions": -1, "intermediate_tensors": 0, "inputs_embeds": 0, } ) class Glm4MoeModel(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config cache_config = vllm_config.cache_config quant_config = vllm_config.quant_config enable_eplb = vllm_config.parallel_config.enable_eplb self.config = config self.vocab_size = config.vocab_size if get_pp_group().is_first_rank: self.embed_tokens = VocabParallelEmbedding( config.vocab_size, config.hidden_size, prefix=f"{prefix}.embed_tokens" ) else: self.embed_tokens = PPMissingLayer() self.start_layer, self.end_layer, self.layers = make_layers( config.num_hidden_layers, lambda prefix: Glm4MoeDecoderLayer( config=config, cache_config=cache_config, quant_config=quant_config, prefix=prefix, enable_eplb=enable_eplb, ), prefix=f"{prefix}.layers", ) if get_pp_group().is_last_rank: self.norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) else: self.norm = PPMissingLayer() self.make_empty_intermediate_tensors = make_empty_intermediate_tensors_factory( ["hidden_states", "residual"], config.hidden_size ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor | IntermediateTensors: if get_pp_group().is_first_rank: if inputs_embeds is not None: hidden_states = inputs_embeds else: hidden_states = self.embed_input_ids(input_ids) residual = None else: assert intermediate_tensors is not None hidden_states = intermediate_tensors["hidden_states"] residual = intermediate_tensors["residual"] for layer in islice(self.layers, self.start_layer, self.end_layer): hidden_states, residual = layer(positions, hidden_states, residual) if not get_pp_group().is_last_rank: return IntermediateTensors( {"hidden_states": hidden_states, "residual": residual} ) hidden_states, _ = self.norm(hidden_states, residual) return hidden_states def get_expert_mapping(self) -> list[tuple[str, str, int, str]]: # Params for weights, fp8 weight scales, fp8 activation scales # (param_name, weight_name, expert_id, shard_id) return SharedFusedMoE.make_expert_params_mapping( self, ckpt_gate_proj_name="gate_proj", ckpt_down_proj_name="down_proj", ckpt_up_proj_name="up_proj", num_experts=self.config.n_routed_experts, ) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ # (param_name, shard_name, shard_id) ("qkv_proj", "q_proj", "q"), ("qkv_proj", "k_proj", "k"), ("qkv_proj", "v_proj", "v"), ("gate_up_proj", "gate_proj", 0), ("gate_up_proj", "up_proj", 1), ] params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() expert_params_mapping = self.get_expert_mapping() for name, loaded_weight in weights: spec_layer = get_spec_layer_idx_from_weight_name(self.config, name) if spec_layer is not None: continue for param_name, weight_name, shard_id in stacked_params_mapping: # Skip non-stacked layers and experts (experts handled below). if weight_name not in name: continue # We have mlp.experts[0].gate_proj in the checkpoint. # Since we handle the experts below in expert_params_mapping, # we need to skip here BEFORE we update the name, otherwise # name will be updated to mlp.experts[0].gate_up_proj, which # will then be updated below in expert_params_mapping # for mlp.experts[0].gate_gate_up_proj, which breaks load. if ("mlp.experts." in name) and name not in params_dict: continue name = name.replace(weight_name, param_name) # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: is_expert_weight = False for mapping in expert_params_mapping: param_name, weight_name, expert_id, shard_id = mapping if weight_name not in name: continue # Anyway, this is an expert weight and should not be # attempted to load as other weights later is_expert_weight = True # Do not modify `name` since the loop may continue here # Instead, create a new variable name_mapped = name.replace(weight_name, param_name) if is_pp_missing_parameter(name_mapped, self): continue param = params_dict[name_mapped] # We should ask the weight loader to return success or not # here since otherwise we may skip experts with other # available replicas. weight_loader = typing.cast( Callable[..., bool], param.weight_loader ) success = weight_loader( param, loaded_weight, name_mapped, shard_id=shard_id, expert_id=expert_id, return_success=True, ) if success: name = name_mapped break else: if is_expert_weight: # We've checked that this is an expert weight # However it's not mapped locally to this rank # So we simply skip it continue # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue # Remapping the name of FP8 kv-scale. name = maybe_remap_kv_scale_name(name, params_dict) if name is None: continue if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = getattr( param, "weight_loader", default_weight_loader ) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params class Glm4MixtureOfExperts(MixtureOfExperts): def extract_moe_parameters(self, example_moe: Glm4MoE | None) -> None: if example_moe is None: raise RuntimeError("No Glm4MoE layer found in model.layers.") else: self.num_logical_experts = example_moe.n_logical_experts self.num_physical_experts = example_moe.n_physical_experts self.num_local_physical_experts = example_moe.n_local_physical_experts self.num_routed_experts = example_moe.n_routed_experts self.num_shared_experts = example_moe.n_shared_experts self.num_redundant_experts = example_moe.n_redundant_experts def update_physical_experts_metadata( self, num_physical_experts: int, num_local_physical_experts: int, ) -> None: assert self.num_local_physical_experts == num_local_physical_experts self.num_physical_experts = num_physical_experts self.num_local_physical_experts = num_local_physical_experts self.num_redundant_experts = num_physical_experts - self.num_logical_experts for moe in self.moe_mlp_layers: moe.n_local_physical_experts = num_local_physical_experts moe.n_physical_experts = num_physical_experts moe.n_redundant_experts = self.num_redundant_experts moe.experts.update_expert_map() class Glm4MoeForCausalLM(nn.Module, SupportsPP, SupportsLoRA, Glm4MixtureOfExperts): packed_modules_mapping = { "qkv_proj": [ "q_proj", "k_proj", "v_proj", ], "gate_up_proj": [ "gate_proj", "up_proj", ], } fall_back_to_pt_during_load = False def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config self.config = config self.quant_config = quant_config self.model = Glm4MoeModel( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model") ) if get_pp_group().is_last_rank: self.lm_head = ParallelLMHead( config.vocab_size, config.hidden_size, quant_config=quant_config, prefix=maybe_prefix(prefix, "lm_head"), ) else: self.lm_head = PPMissingLayer() self.logits_processor = LogitsProcessor(config.vocab_size) self.make_empty_intermediate_tensors = ( self.model.make_empty_intermediate_tensors ) self.expert_weights = [] # Set MoE hyperparameters self.num_moe_layers = config.num_hidden_layers - config.first_k_dense_replace self.num_expert_groups = config.n_group self.moe_layers = [] self.moe_mlp_layers: list[Glm4MoE] = [] example_moe = None for layer in self.model.layers: if isinstance(layer, PPMissingLayer): continue assert isinstance(layer, Glm4MoeDecoderLayer) if isinstance(layer.mlp, Glm4MoE): # Pick last one layer since the first ones may be dense layers. example_moe = layer.mlp self.moe_mlp_layers.append(layer.mlp) self.moe_layers.append(layer.mlp.experts) self.extract_moe_parameters(example_moe) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor | IntermediateTensors: hidden_states = self.model( input_ids, positions, intermediate_tensors, inputs_embeds ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: logits = self.logits_processor(self.lm_head, hidden_states) return logits def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader(self) return loader.load_weights(weights) def get_expert_mapping(self) -> list[tuple[str, str, int, str]]: return self.model.get_expert_mapping() def get_spec_layer_idx_from_weight_name( config: Glm4MoeConfig, weight_name: str ) -> int | None: if hasattr(config, "num_nextn_predict_layers") and ( config.num_nextn_predict_layers > 0 ): layer_idx = config.num_hidden_layers for i in range(config.num_nextn_predict_layers): if f"layers.{layer_idx + i}." in weight_name: return layer_idx + i return None
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/glm4_moe.py", "license": "Apache License 2.0", "lines": 644, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/glm4_moe_mtp.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright 2025 The ZhipuAI Team. # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Inference-only GLM-4.5, GLM-4.6, GLM-4.7 MTP model compatible with HuggingFace weights.""" from collections.abc import Iterable import torch import torch.nn as nn from transformers import PretrainedConfig from vllm.config import CacheConfig, ParallelConfig, VllmConfig from vllm.model_executor.layers.fused_moe import FusedMoE from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import default_weight_loader from vllm.sequence import IntermediateTensors from .glm4_moe import ( Glm4MixtureOfExperts, Glm4MoE, Glm4MoeDecoderLayer, get_spec_layer_idx_from_weight_name, ) from .utils import maybe_prefix class SharedHead(nn.Module): def __init__( self, config: PretrainedConfig, prefix: str, quant_config: QuantizationConfig | None = None, ) -> None: super().__init__() self.norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.head = ParallelLMHead( config.vocab_size, config.hidden_size, quant_config=quant_config, prefix=maybe_prefix(prefix, "head"), ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: return self.norm(hidden_states) class Glm4MoeMultiTokenPredictorLayer(nn.Module): def __init__( self, config: PretrainedConfig, prefix: str, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, parallel_config: ParallelConfig | None = None, ) -> None: super().__init__() self.enorm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.hnorm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.eh_proj = nn.Linear(config.hidden_size * 2, config.hidden_size, bias=False) self.shared_head = SharedHead( config=config, prefix=prefix, quant_config=quant_config ) self.enable_eplb = parallel_config.enable_eplb self.mtp_block = Glm4MoeDecoderLayer( config=config, cache_config=cache_config, quant_config=quant_config, prefix=prefix, enable_eplb=self.enable_eplb, ) def forward( self, input_ids: torch.Tensor, positions: torch.Tensor, previous_hidden_states: torch.Tensor, inputs_embeds: torch.Tensor | None = None, spec_step_index: int = 0, ) -> torch.Tensor: assert inputs_embeds is not None # masking inputs at position 0, as not needed by MTP inputs_embeds = torch.where(positions.unsqueeze(-1) == 0, 0, inputs_embeds) inputs_embeds = self.enorm(inputs_embeds) previous_hidden_states = self.hnorm(previous_hidden_states) hidden_states = self.eh_proj( torch.cat([inputs_embeds, previous_hidden_states], dim=-1) ) hidden_states, residual = self.mtp_block( positions=positions, hidden_states=hidden_states, residual=None ) hidden_states = residual + hidden_states return hidden_states class Glm4MoeMultiTokenPredictor(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config self.mtp_start_layer_idx = config.num_hidden_layers self.num_mtp_layers = config.num_nextn_predict_layers # to map the exact layer index from weights self.layers = torch.nn.ModuleDict( { str(idx): Glm4MoeMultiTokenPredictorLayer( config, f"{prefix}.layers.{idx}", cache_config=vllm_config.cache_config, quant_config=vllm_config.quant_config, parallel_config=vllm_config.parallel_config, ) for idx in range( self.mtp_start_layer_idx, self.mtp_start_layer_idx + self.num_mtp_layers, ) } ) self.embed_tokens = VocabParallelEmbedding( config.vocab_size, config.hidden_size, ) self.logits_processor = LogitsProcessor(config.vocab_size) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor, positions: torch.Tensor, previous_hidden_states: torch.Tensor, inputs_embeds: torch.Tensor | None = None, spec_step_idx: int = 0, ) -> torch.Tensor: if inputs_embeds is None: inputs_embeds = self.embed_tokens(input_ids) current_step_idx = spec_step_idx % self.num_mtp_layers return self.layers[str(self.mtp_start_layer_idx + current_step_idx)]( input_ids, positions, previous_hidden_states, inputs_embeds, current_step_idx, ) def compute_logits( self, hidden_states: torch.Tensor, spec_step_idx: int = 0, ) -> torch.Tensor: current_step_idx = spec_step_idx % self.num_mtp_layers mtp_layer = self.layers[str(self.mtp_start_layer_idx + current_step_idx)] logits = self.logits_processor( mtp_layer.shared_head.head, mtp_layer.shared_head(hidden_states) ) return logits class Glm4MoeMTP(nn.Module, Glm4MixtureOfExperts): def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() self.config = vllm_config.model_config.hf_config self.model = Glm4MoeMultiTokenPredictor( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model") ) self.expert_weights = [] # Set MoE hyperparameters self.num_moe_layers = self.config.num_nextn_predict_layers self.num_expert_groups = self.config.n_group self.moe_layers: list[FusedMoE] = [] self.moe_mlp_layers: list[Glm4MoE] = [] example_moe = None for layer in self.model.layers.values(): assert isinstance(layer, Glm4MoeMultiTokenPredictorLayer) layer = layer.mtp_block assert isinstance(layer, Glm4MoeDecoderLayer) if isinstance(layer.mlp, Glm4MoE): example_moe = layer.mlp self.moe_mlp_layers.append(layer.mlp) self.moe_layers.append(layer.mlp.experts) self.extract_moe_parameters(example_moe) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, hidden_states: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, spec_step_idx: int = 0, ) -> torch.Tensor: hidden_states = self.model( input_ids, positions, hidden_states, inputs_embeds, spec_step_idx ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, spec_step_idx: int = 0, ) -> torch.Tensor | None: return self.model.compute_logits(hidden_states, spec_step_idx) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ # (param_name, shard_name, shard_id) ("qkv_proj", "q_proj", "q"), ("qkv_proj", "k_proj", "k"), ("qkv_proj", "v_proj", "v"), ("gate_up_proj", "gate_proj", 0), ("gate_up_proj", "up_proj", 1), ] # Params for weights, fp8 weight scales, fp8 activation scales # (param_name, weight_name, expert_id, shard_id) expert_params_mapping = FusedMoE.make_expert_params_mapping( self, ckpt_gate_proj_name="gate_proj", ckpt_down_proj_name="down_proj", ckpt_up_proj_name="up_proj", num_experts=self.config.n_routed_experts, ) params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() for name, loaded_weight in weights: if name == "lm_head.weight": spec_layer = self.model.mtp_start_layer_idx name = f"model.layers.{spec_layer}.shared_head.head.weight" elif name == "model.embed_tokens.weight": spec_layer = self.model.mtp_start_layer_idx else: spec_layer = get_spec_layer_idx_from_weight_name(self.config, name) if spec_layer is None: continue name = self._rewrite_spec_layer_name(spec_layer, name) for param_name, weight_name, shard_id in stacked_params_mapping: # Skip non-stacked layers and experts (experts handled below). if weight_name not in name: continue # We have mlp.experts[0].gate_proj in the checkpoint. # Since we handle the experts below in expert_params_mapping, # we need to skip here BEFORE we update the name, otherwise # name will be updated to mlp.experts[0].gate_up_proj, which # will then be updated below in expert_params_mapping # for mlp.experts[0].gate_gate_up_proj, which breaks load. if ("mlp.experts." in name) and name not in params_dict: continue name = name.replace(weight_name, param_name) # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: for mapping in expert_params_mapping: param_name, weight_name, expert_id, shard_id = mapping if weight_name not in name: continue name = name.replace(weight_name, param_name) param = params_dict[name] weight_loader = param.weight_loader weight_loader( param, loaded_weight, name, shard_id=shard_id, expert_id=expert_id, ) break else: # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue # Some checkpoints include weight scale tensors for the # LM head even when the quantized head isn't built. Skip # them if the model does not expose a matching parameter # to avoid KeyError during load. if name.endswith(".weight_scale") and name not in params_dict: continue # According to DeepSeek-V3 Technical Report, MTP modules # shares embedding layer. We only load the first weights. if ( spec_layer != self.model.mtp_start_layer_idx and ".layers" not in name ): continue param = params_dict[name] weight_loader = getattr( param, "weight_loader", default_weight_loader ) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params def _rewrite_spec_layer_name(self, spec_layer: int, name: str) -> str: """ Rewrite the weight name to match the format of the original model. Add .mtp_block for modules in transformer layer block for spec layer and rename shared layer weights to be top level. """ spec_layer_weight_names = [ "embed_tokens", "enorm", "hnorm", "eh_proj", "shared_head", ] shared_weight_names = ["embed_tokens"] spec_layer_weight = False shared_weight = False for weight_name in spec_layer_weight_names: if weight_name in name: spec_layer_weight = True if weight_name in shared_weight_names: shared_weight = True break if not spec_layer_weight: # treat rest weights as weights for transformer layer block name = name.replace( f"model.layers.{spec_layer}.", f"model.layers.{spec_layer}.mtp_block." ) elif shared_weight: # treat shared weights as top level weights name = name.replace(f"model.layers.{spec_layer}.", "model.") return name
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/glm4_moe_mtp.py", "license": "Apache License 2.0", "lines": 332, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/exaone4.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # ruff: noqa: E501 # Adapted from # https://github.com/lgai-exaone/transformers/blob/add-exaone4/src/transformers/models/exaone4/modeling_exaone4.py # Copyright 2025 The LG CNS Gen AI Solution Delivery Team. # Copyright 2025 The LG AI Research and HuggingFace Inc. team. All rights reserved. # # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Inference-only Exaone model compatible with HuggingFace weights.""" from collections.abc import Iterable from itertools import islice import torch from torch import nn from transformers import Exaone4Config from vllm.compilation.decorators import support_torch_compile from vllm.config import CacheConfig, VllmConfig from vllm.distributed import get_pp_group, get_tensor_model_parallel_world_size from vllm.model_executor.layers.activation import SiluAndMul from vllm.model_executor.layers.attention import Attention from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( MergedColumnParallelLinear, QKVParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import ( default_weight_loader, maybe_remap_kv_scale_name, ) from vllm.sequence import IntermediateTensors from vllm.transformers_utils.config import set_default_rope_theta from .interfaces import SupportsLoRA, SupportsPP from .utils import ( AutoWeightsLoader, PPMissingLayer, extract_layer_index, is_pp_missing_parameter, make_empty_intermediate_tensors_factory, make_layers, maybe_prefix, ) class Exaone4GatedMLP(nn.Module): def __init__( self, hidden_size: int, intermediate_size: int, hidden_act: str, quant_config: QuantizationConfig | None = None, reduce_results: bool = True, bias: bool = False, prefix: str = "", ) -> None: super().__init__() self.gate_up_proj = MergedColumnParallelLinear( input_size=hidden_size, output_sizes=[intermediate_size] * 2, bias=bias, quant_config=quant_config, prefix=f"{prefix}.gate_up_proj", ) self.down_proj = RowParallelLinear( input_size=intermediate_size, output_size=hidden_size, bias=bias, quant_config=quant_config, reduce_results=reduce_results, prefix=f"{prefix}.down_proj", ) if hidden_act != "silu": raise ValueError( f"Unsupported activation: {hidden_act}. Only silu is supported for now." ) self.act_fn = SiluAndMul() def forward(self, x): gate_up, _ = self.gate_up_proj(x) x = self.act_fn(gate_up) x, _ = self.down_proj(x) return x class Exaone4Attention(nn.Module): def __init__( self, config: Exaone4Config, hidden_size: int, num_heads: int, num_kv_heads: int, max_position_embeddings: int = 8192, quant_config: QuantizationConfig | None = None, bias: bool = False, cache_config: CacheConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.hidden_size = hidden_size tp_size = get_tensor_model_parallel_world_size() self.total_num_heads = num_heads assert self.total_num_heads % tp_size == 0 self.num_heads = self.total_num_heads // tp_size self.total_num_kv_heads = num_kv_heads if self.total_num_kv_heads >= tp_size: # Number of KV heads is greater than TP size, so we partition # the KV heads across multiple tensor parallel GPUs. assert self.total_num_kv_heads % tp_size == 0 else: # Number of KV heads is less than TP size, so we replicate # the KV heads across multiple tensor parallel GPUs. assert tp_size % self.total_num_kv_heads == 0 self.num_kv_heads = max(1, self.total_num_kv_heads // tp_size) # MistralConfig has an optional head_dim introduced by Mistral-Nemo self.head_dim = getattr(config, "head_dim", None) if self.head_dim is None: self.head_dim = self.hidden_size // self.total_num_heads self.q_size = self.num_heads * self.head_dim self.kv_size = self.num_kv_heads * self.head_dim self.scaling = self.head_dim**-0.5 self.max_position_embeddings = max_position_embeddings self.qkv_proj = QKVParallelLinear( hidden_size=hidden_size, head_size=self.head_dim, total_num_heads=self.total_num_heads, total_num_kv_heads=self.total_num_kv_heads, bias=bias, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", ) self.o_proj = RowParallelLinear( input_size=self.total_num_heads * self.head_dim, output_size=hidden_size, bias=bias, quant_config=quant_config, prefix=f"{prefix}.o_proj", ) self.q_norm = RMSNorm(self.head_dim, eps=config.rms_norm_eps) self.k_norm = RMSNorm(self.head_dim, eps=config.rms_norm_eps) is_neox_style = True if quant_config is not None and quant_config.get_name() == "gguf": is_neox_style = False layer_idx = extract_layer_index(prefix) is_sliding = config.layer_types[layer_idx] == "sliding_attention" self.sliding_window = config.sliding_window if is_sliding else None # apply rotary embeddings to every layer in full attention models self.apply_rope_all_layers = "sliding_attention" not in config.layer_types set_default_rope_theta(config, default_theta=1000000) self.rotary_emb = get_rope( self.head_dim, max_position=max_position_embeddings, rope_parameters=config.rope_parameters, is_neox_style=is_neox_style, ) self.attn = Attention( self.num_heads, self.head_dim, self.scaling, num_kv_heads=self.num_kv_heads, cache_config=cache_config, quant_config=quant_config, per_layer_sliding_window=self.sliding_window, prefix=f"{prefix}.attn", ) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, ) -> torch.Tensor: qkv, _ = self.qkv_proj(hidden_states) q, k, v = qkv.split([self.q_size, self.kv_size, self.kv_size], dim=-1) q = q.unflatten(-1, (self.num_heads, self.head_dim)) q = self.q_norm(q) q = q.flatten(-2, -1) k = k.unflatten(-1, (self.num_kv_heads, self.head_dim)) k = self.k_norm(k) k = k.flatten(-2, -1) if self.sliding_window or self.apply_rope_all_layers: q, k = self.rotary_emb(positions, q, k) attn_output = self.attn(q, k, v) output, _ = self.o_proj(attn_output) return output class Exaone4DecoderLayer(nn.Module): def __init__( self, config: Exaone4Config, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.hidden_size = config.hidden_size max_position_embeddings = getattr(config, "max_position_embeddings", 8192) # Support abacusai/Smaug-72B-v0.1 with attention_bias # Support internlm/internlm-7b with bias attention_bias = getattr(config, "attention_bias", False) or getattr( config, "bias", False ) self.self_attn = Exaone4Attention( config=config, hidden_size=self.hidden_size, num_heads=config.num_attention_heads, num_kv_heads=getattr( config, "num_key_value_heads", config.num_attention_heads ), max_position_embeddings=max_position_embeddings, quant_config=quant_config, bias=attention_bias, cache_config=cache_config, prefix=f"{prefix}.self_attn", ) self.mlp = Exaone4GatedMLP( hidden_size=self.hidden_size, intermediate_size=config.intermediate_size, hidden_act=config.hidden_act, quant_config=quant_config, bias=getattr(config, "mlp_bias", False), prefix=f"{prefix}.mlp", ) self.post_attention_layernorm = RMSNorm( config.hidden_size, eps=config.rms_norm_eps ) self.post_feedforward_layernorm = RMSNorm( config.hidden_size, eps=config.rms_norm_eps ) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, residual: torch.Tensor | None, ) -> tuple[torch.Tensor, torch.Tensor]: residual = hidden_states # Self Attention hidden_states = self.self_attn( positions=positions, hidden_states=hidden_states, ) # Use post-LN hidden_states = self.post_attention_layernorm(hidden_states) hidden_states = residual + hidden_states residual = hidden_states # Fully Connected hidden_states = self.mlp(hidden_states) # Use post-LN hidden_states = self.post_feedforward_layernorm(hidden_states) hidden_states = residual + hidden_states return hidden_states, residual @support_torch_compile class Exaone4Model(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config cache_config = vllm_config.cache_config quant_config = vllm_config.quant_config self.config = config self.quant_config = quant_config self.vocab_size = config.vocab_size if get_pp_group().is_first_rank or ( config.tie_word_embeddings and get_pp_group().is_last_rank ): self.embed_tokens = VocabParallelEmbedding( self.vocab_size, config.hidden_size, quant_config=quant_config, ) else: self.embed_tokens = PPMissingLayer() self.start_layer, self.end_layer, self.layers = make_layers( config.num_hidden_layers, lambda prefix: Exaone4DecoderLayer( config=config, cache_config=cache_config, quant_config=quant_config, prefix=prefix, ), prefix=f"{prefix}.layers", ) if get_pp_group().is_last_rank: self.norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) else: self.norm = PPMissingLayer() self.make_empty_intermediate_tensors = make_empty_intermediate_tensors_factory( ["hidden_states", "residual"], config.hidden_size ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor | IntermediateTensors: if get_pp_group().is_first_rank: if inputs_embeds is not None: hidden_states = inputs_embeds else: hidden_states = self.embed_input_ids(input_ids) residual = None else: assert intermediate_tensors is not None hidden_states = intermediate_tensors["hidden_states"] residual = intermediate_tensors["residual"] for layer in islice(self.layers, self.start_layer, self.end_layer): hidden_states, residual = layer( positions, hidden_states, residual, ) if not get_pp_group().is_last_rank: return IntermediateTensors( {"hidden_states": hidden_states, "residual": residual} ) hidden_states = self.norm(hidden_states) return hidden_states def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ # (param_name, shard_name, shard_id) (".qkv_proj", ".q_proj", "q"), (".qkv_proj", ".k_proj", "k"), (".qkv_proj", ".v_proj", "v"), (".gate_up_proj", ".gate_proj", 0), (".gate_up_proj", ".up_proj", 1), ] params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() for name, loaded_weight in weights: if "rotary_emb.inv_freq" in name: continue if "rotary_emb.cos_cached" in name or "rotary_emb.sin_cached" in name: # Models trained using ColossalAI may include these tensors in # the checkpoint. Skip them. continue if self.quant_config is not None and ( scale_name := self.quant_config.get_cache_scale(name) ): # Loading kv cache quantization scales param = params_dict[scale_name] weight_loader = getattr(param, "weight_loader", default_weight_loader) loaded_weight = ( loaded_weight if loaded_weight.dim() == 0 else loaded_weight[0] ) weight_loader(param, loaded_weight) loaded_params.add(scale_name) continue for param_name, weight_name, shard_id in stacked_params_mapping: if weight_name not in name: continue name = name.replace(weight_name, param_name) # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue # Remapping the name of FP8 kv-scale. name = maybe_remap_kv_scale_name(name, params_dict) if name is None: continue if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params class Exaone4ForCausalLM(nn.Module, SupportsLoRA, SupportsPP): packed_modules_mapping = { "qkv_proj": [ "q_proj", "k_proj", "v_proj", ], "gate_up_proj": [ "gate_proj", "up_proj", ], } # LoRA specific attributes embedding_modules = { "embed_tokens": "input_embeddings", "lm_head": "output_embeddings", } def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config self.config = config self.quant_config = quant_config self.model = Exaone4Model( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model"), ) if get_pp_group().is_last_rank: self.lm_head = ParallelLMHead( config.vocab_size, config.hidden_size, quant_config=quant_config, prefix=maybe_prefix(prefix, "lm_head"), ) if config.tie_word_embeddings: self.lm_head.weight = self.model.embed_tokens.weight logit_scale = getattr(config, "logit_scale", 1.0) self.logits_processor = LogitsProcessor( config.vocab_size, scale=logit_scale ) else: self.lm_head = PPMissingLayer() self.make_empty_intermediate_tensors = ( self.model.make_empty_intermediate_tensors ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor | IntermediateTensors: model_output = self.model( input_ids, positions, intermediate_tensors, inputs_embeds ) return model_output def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: logits = self.logits_processor(self.lm_head, hidden_states) return logits def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader( self, # With tie_word_embeddings, we can skip lm_head.weight # The weight might appear unnecessarily in the files if the model is # processed with quantization, LoRA, fine-tuning, etc. skip_prefixes=(["lm_head."] if self.config.tie_word_embeddings else None), ) return loader.load_weights(weights)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/exaone4.py", "license": "Apache License 2.0", "lines": 459, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:examples/online_serving/elastic_ep/scale.py
#!/usr/bin/env python3 # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import json import sys import requests def scale(host, port, new_dp_size): url = f"http://{host}:{port}/scale_elastic_ep" payload = {"new_data_parallel_size": new_dp_size} headers = {"Content-Type": "application/json"} print(f"Sending scale request to {url}") print(f"Payload: {json.dumps(payload, indent=2)}") try: response = requests.post(url, json=payload, headers=headers, timeout=300) print(f"Status Code: {response.status_code}") print(f"Response: {response.text}") if response.status_code == 200: print("Scale up/down request successful!") return True else: print("Scale up/down request failed!") return False except requests.exceptions.RequestException as e: print(f"Request failed: {e}") return False def main(): parser = argparse.ArgumentParser(description="Test scale up/down functionality") parser.add_argument("--host", default="localhost", help="API server host") parser.add_argument("--port", type=int, default=8006, help="API server port") parser.add_argument( "--new-dp-size", type=int, default=2, help="New data parallel size" ) args = parser.parse_args() success = scale(args.host, args.port, args.new_dp_size) sys.exit(0 if success else 1) if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "examples/online_serving/elastic_ep/scale.py", "license": "Apache License 2.0", "lines": 38, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fused_moe/flashinfer_cutlass_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm.config import get_current_vllm_config from vllm.logger import init_logger from vllm.model_executor.layers.fused_moe.activation import MoEActivation from vllm.model_executor.layers.fused_moe.config import ( FusedMoEParallelConfig, FusedMoEQuantConfig, ) from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceNoOP, ) from vllm.model_executor.layers.quantization.utils.quant_utils import ( QuantKey, kFp8Dynamic128Sym, kFp8Static128BlockSym, kFp8StaticTensorSym, kMxfp4Static, kMxfp8Dynamic, kNvfp4Dynamic, kNvfp4Static, ) from vllm.platforms import current_platform from vllm.utils.flashinfer import ( flashinfer_cutlass_fused_moe, has_flashinfer_cutlass_fused_moe, ) logger = init_logger(__name__) def is_valid_flashinfer_cutlass_fused_moe( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor ) -> bool: """ Check if the given problem size is supported by the FlashInfer CUTLASS MoE kernel. """ if not has_flashinfer_cutlass_fused_moe(): logger.debug_once( "FlashInferExperts disabled: flashinfer_cutlass_fused_moe not available." ) return False # Data type checks if ( w1.dtype != torch.uint8 or w2.dtype != torch.uint8 or hidden_states.dtype not in [torch.float32, torch.float16, torch.bfloat16] ): logger.debug_once( "FlashInferExperts disabled: w1/w2 must be torch.uint8 " f"(got w1={w1.dtype}, w2={w2.dtype}), hidden_states must be " f"float32, float16, or bfloat16 (got {hidden_states.dtype})." ) return False return True class FlashInferExperts(mk.FusedMoEPermuteExpertsUnpermute): def __init__( self, moe_config: mk.FusedMoEConfig, quant_config: FusedMoEQuantConfig, ): super().__init__(moe_config, quant_config) assert quant_config.weight_quant_dtype in ( "mxfp4", "nvfp4", torch.float8_e4m3fn, None, ), ( "Only mxfp4, nvfp4, fp8, bfloat16 and" " float16 quantization are currently supported." ) self.device = moe_config.device self.num_experts = moe_config.num_local_experts self.ep_rank = moe_config.moe_parallel_config.ep_rank self.ep_size = moe_config.moe_parallel_config.ep_size self.tp_rank = moe_config.moe_parallel_config.tp_rank self.tp_size = moe_config.moe_parallel_config.tp_size self.out_dtype = moe_config.in_dtype self.use_dp = moe_config.moe_parallel_config.dp_size > 1 # Enables DeepSeek-style FP8 block-scale path: # - pass per-block weight scales to the kernel # - skip input activation quantization (kernel applies scaling) self.use_deepseek_fp8_block_scale = quant_config.is_block_quantized self.max_capture_size = ( get_current_vllm_config().compilation_config.max_cudagraph_capture_size ) if quant_config.weight_quant_dtype == "mxfp4": # This value is used specifically for gpt-oss, # Need to revisit this for other models self.gemm1_alpha = torch.tensor( [1.702] * self.num_experts, dtype=torch.float32, device=self.device ) self.gemm1_beta = torch.tensor( [1.0] * self.num_experts, dtype=torch.float32, device=self.device ) self.gemm1_clamp_limit = torch.tensor( [7.0] * self.num_experts, dtype=torch.float32, device=self.device ) if quant_config.quant_dtype == "mxfp8": self.fake_input_scale = torch.ones( self.num_experts, device=self.device, dtype=torch.float32, ) @property def expects_unquantized_inputs(self) -> bool: return self.quant_config.use_fp8_w8a8 and self.quant_config.is_block_quantized @staticmethod def _supports_current_device() -> bool: p = current_platform return ( p.is_cuda() and ( p.is_device_capability(90) or p.is_device_capability_family(100) or p.is_device_capability_family(110) or p.is_device_capability_family(120) ) and has_flashinfer_cutlass_fused_moe() ) @staticmethod def _supports_no_act_and_mul() -> bool: return True @staticmethod def _supports_quant_scheme( weight_key: QuantKey | None, activation_key: QuantKey | None, ) -> bool: p = current_platform scheme = (weight_key, activation_key) # The following are supported by FlashInferExperts: return ( # unquantized and fp8 static per-tensor on 9.0+ ( scheme in [ (None, None), (kFp8StaticTensorSym, kFp8StaticTensorSym), ] and p.has_device_capability(90) ) # fp8 block-scale, wmxfp4a16 on 9.0 or ( scheme in [ (kMxfp4Static, None), (kFp8Static128BlockSym, kFp8Dynamic128Sym), ] and p.is_device_capability(90) ) # nvfp4, wmxfp4amxfp8 on 10.0+ or ( scheme in [ (kMxfp4Static, kMxfp8Dynamic), (kNvfp4Static, kNvfp4Dynamic), ] and p.has_device_capability(100) ) ) @staticmethod def _supports_activation(activation: MoEActivation) -> bool: return activation in [ MoEActivation.SILU, MoEActivation.RELU2_NO_MUL, MoEActivation.SWIGLUOAI, ] @staticmethod def _supports_parallel_config(moe_parallel_config: FusedMoEParallelConfig) -> bool: # FLASHINFER_CUTLASS currently uses its down P/F, which does not # work with SP. This will be removed in follow up after we get # rid of the FlashInfer specific P/F function. # TODO: the per-tensor fp8 kernels don't work with MNNVL FI A2As. return True @staticmethod def activation_format() -> mk.FusedMoEActivationFormat: return mk.FusedMoEActivationFormat.Standard def supports_expert_map(self) -> bool: return False def supports_chunking(self) -> bool: # This refers to TP chunking; DP chunking is handled separately. return True def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: return TopKWeightAndReduceNoOP() def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, activation: MoEActivation, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: # We use global_num_experts due to how moe_align_block_size handles # expert_maps. """ Compute the shapes for the temporary and final outputs of the two gemms and activation in the fused expert function. Since the gemms are independent, the workspace for the first gemm can be shared with the workspace for the last gemm. Returns a tuple of: - workspace13 shape tuple: must be large enough to hold the result of either expert gemm. - workspace2 shape tuple: must be large enough to hold the result of the activation function. - output shape tuple: must be exact size of the final gemm output. - Workspace type: The dtype to use for the workspace tensors. - Note: in order for activation chunking to work, the first dimension of each tuple must be the number of tokens. """ workspace1 = (M, K) workspace2 = (0,) # For NVFP4, the output is stored in a packed int8 format, # so the actual hidden dim is 2x the size of K here. output_shape = (M, K * 2 if self.quant_dtype == "nvfp4" else K) # The workspace is determined by `aq`, since it comes after any # potential communication op and is involved in the expert computation. return (workspace1, workspace2, output_shape) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: MoEActivation, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor | None, workspace2: torch.Tensor | None, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool | None, ): from flashinfer.fused_moe.core import ActivationType activation_str_to_value_map = { MoEActivation.SILU: ActivationType.Swiglu, # This is the default MoEActivation.SWIGLUOAI: ActivationType.Swiglu, # gpt-oss alias MoEActivation.RELU2_NO_MUL: ActivationType.Relu2, } assert activation in activation_str_to_value_map, ( f"{activation=} missing from {activation_str_to_value_map.keys()=}" ) quant_scales = None fc1_expert_weights = None fc2_expert_weights = None fc1_expert_biases = None fc2_expert_biases = None swiglu_alpha = None swiglu_beta = None swiglu_limit = None use_mxfp8_act_scaling = False use_w4_group_scaling = False # Select quantization metadata based on FP8 format/path if ( self.quant_dtype == torch.float8_e4m3fn and not self.use_deepseek_fp8_block_scale ): # FP8 per-tensor path: use global alphas/scales; do not pass input_sf quant_scales = [ self.g1_alphas, # w13_weight_scale * w13_input_scale self.a2_gscale, # 1.0 / w2_input_scale self.g2_alphas, # w2_weight_scale * w2_input_scale self.a1_scale, ] a1q_scale = None # not passing input_sf in fp8 fc1_expert_weights = w1 fc2_expert_weights = w2 elif self.quant_dtype == "nvfp4": # Ensure w1_scale and w2_scale are not None before calling view assert self.w1_scale is not None and self.w2_scale is not None, ( "w1_scale and w2_scale must not be None for FlashInferExperts" ) # Flashinfer CUTLASS kernel takes scalar global scales, # min because inv_scale. quant_scales = [ self.a1_gscale, self.w1_scale.view(torch.int32), self.g1_alphas, self.a2_gscale, self.w2_scale.view(torch.int32), self.g2_alphas, ] # FlashInfer API requires weight to be long for nvfp4 fc1_expert_weights = w1.view(torch.long) fc2_expert_weights = w2.view(torch.long) elif self.weight_quant_dtype == "mxfp4": assert self.w1_scale is not None and self.w2_scale is not None assert w1.is_contiguous() and w2.is_contiguous() assert self.gemm1_alpha is not None assert self.gemm1_beta is not None assert self.gemm1_clamp_limit is not None assert topk_ids.is_contiguous() fc1_expert_biases = self.w1_bias fc2_expert_biases = self.w2_bias swiglu_alpha = self.gemm1_alpha swiglu_beta = self.gemm1_beta swiglu_limit = self.gemm1_clamp_limit if self.quant_dtype == "mxfp8": assert self.fake_input_scale is not None fc1_expert_weights = w1.view(torch.long) fc2_expert_weights = w2.view(torch.long) quant_scales = [ self.w1_scale.view(torch.int32), self.fake_input_scale, self.w2_scale.view(torch.int32), self.fake_input_scale, ] use_mxfp8_act_scaling = True else: assert hidden_states.dtype == torch.bfloat16 fc1_expert_weights = w1 fc2_expert_weights = w2 quant_scales = [ self.w1_scale, self.w2_scale, ] a1q_scale = None use_w4_group_scaling = True elif self.use_deepseek_fp8_block_scale: # FP8 block-scale path: provide block-scale weights, omit a1q_scale quant_scales = [ self.w1_scale, self.w2_scale, ] a1q_scale = None fc1_expert_weights = w1 fc2_expert_weights = w2 else: quant_scales = None a1q_scale = None fc1_expert_weights = w1 fc2_expert_weights = w2 _ = flashinfer_cutlass_fused_moe( input=hidden_states, token_selected_experts=topk_ids.to(torch.int), token_final_scales=topk_weights, fc1_expert_weights=fc1_expert_weights, fc2_expert_weights=fc2_expert_weights, fc1_expert_biases=fc1_expert_biases, fc2_expert_biases=fc2_expert_biases, swiglu_alpha=swiglu_alpha, swiglu_beta=swiglu_beta, swiglu_limit=swiglu_limit, output=output, output_dtype=self.out_dtype, quant_scales=quant_scales, input_sf=a1q_scale, tp_size=self.tp_size, tp_rank=self.tp_rank, ep_size=self.ep_size, ep_rank=self.ep_rank, activation_type=activation_str_to_value_map[activation], # Informs FlashInfer to use the block-scale decoding path when True use_deepseek_fp8_block_scale=self.use_deepseek_fp8_block_scale, use_mxfp8_act_scaling=use_mxfp8_act_scaling, use_w4_group_scaling=use_w4_group_scaling, tune_max_num_tokens=max(self.max_capture_size, 1), ) def moe_sum(self, input: torch.Tensor, output: torch.Tensor) -> None: # No support for LoRA in flashinfer_cutlass_fused_moe. # See TODOs in flashinfer functions runMoe and runMoeMinLantency. raise NotImplementedError("LoRA is not supported for flashinfer_cutlass_moe")
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fused_moe/flashinfer_cutlass_moe.py", "license": "Apache License 2.0", "lines": 367, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/utils/flashinfer.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Compatibility wrapper for FlashInfer API changes. Users of vLLM should always import **only** these wrappers. """ import contextlib import functools import importlib import importlib.util import os import shutil from collections.abc import Callable from typing import Any, NoReturn import requests import torch import vllm.envs as envs from vllm.logger import init_logger from vllm.model_executor.layers.batch_invariant import ( vllm_is_batch_invariant, ) from vllm.platforms import current_platform logger = init_logger(__name__) # This is the storage path for the cubins, it can be replaced # with a local path for testing. # Referenced from https://github.com/flashinfer-ai/flashinfer/blob/0c9a92c3d9a7e043ab6f3f7b2273269caf6ab044/flashinfer/jit/cubin_loader.py#L35 # noqa: E501 FLASHINFER_CUBINS_REPOSITORY = os.environ.get( "FLASHINFER_CUBINS_REPOSITORY", "https://edge.urm.nvidia.com/artifactory/sw-kernelinferencelibrary-public-generic-local/", # noqa: E501 ) @functools.cache def has_flashinfer_cubin() -> bool: """Return `True` if flashinfer-cubin package is available.""" if envs.VLLM_HAS_FLASHINFER_CUBIN: return True if importlib.util.find_spec("flashinfer_cubin") is not None: return True logger.debug_once("flashinfer-cubin package was not found") return False @functools.cache def has_flashinfer() -> bool: """Return `True` if flashinfer-python package is available.""" # Use find_spec to check if the module exists without importing it # This avoids potential CUDA initialization side effects if importlib.util.find_spec("flashinfer") is None: logger.debug_once("FlashInfer unavailable since package was not found") return False # When not using flashinfer cubin, # Also check if nvcc is available since it's required to JIT compile flashinfer if not has_flashinfer_cubin() and shutil.which("nvcc") is None: logger.debug_once( "FlashInfer unavailable since nvcc was not found " "and not using pre-downloaded cubins" ) return False return True def _missing(*_: Any, **__: Any) -> NoReturn: """Placeholder for unavailable FlashInfer backend.""" raise RuntimeError( "FlashInfer backend is not available. Please install the package " "to enable FlashInfer kernels: " "https://github.com/flashinfer-ai/flashinfer" ) def _get_submodule(module_name: str) -> Any | None: """Safely import a submodule and return it, or None if not available.""" try: return importlib.import_module(module_name) except (ImportError, ModuleNotFoundError): return None # General lazy import wrapper def _lazy_import_wrapper( module_name: str, attr_name: str, fallback_fn: Callable[..., Any] = _missing ): """Create a lazy import wrapper for a specific function.""" @functools.cache def _get_impl(): if not has_flashinfer(): return None mod = _get_submodule(module_name) return getattr(mod, attr_name, None) if mod else None def wrapper(*args, **kwargs): impl = _get_impl() if impl is None: return fallback_fn(*args, **kwargs) return impl(*args, **kwargs) return wrapper # Create lazy wrappers for each function flashinfer_trtllm_bf16_moe = _lazy_import_wrapper( "flashinfer.fused_moe", "trtllm_bf16_moe" ) flashinfer_trtllm_fp8_block_scale_moe = _lazy_import_wrapper( "flashinfer.fused_moe", "trtllm_fp8_block_scale_moe" ) flashinfer_trtllm_fp8_per_tensor_scale_moe = _lazy_import_wrapper( "flashinfer.fused_moe", "trtllm_fp8_per_tensor_scale_moe" ) flashinfer_cutlass_fused_moe = _lazy_import_wrapper( "flashinfer.fused_moe", "cutlass_fused_moe" ) flashinfer_cutedsl_grouped_gemm_nt_masked = _lazy_import_wrapper( "flashinfer.cute_dsl.blockscaled_gemm", "grouped_gemm_nt_masked" ) flashinfer_fp4_quantize = _lazy_import_wrapper("flashinfer", "fp4_quantize") nvfp4_batched_quantize = _lazy_import_wrapper("flashinfer", "nvfp4_batched_quantize") silu_and_mul_scaled_nvfp4_experts_quantize = _lazy_import_wrapper( "flashinfer", "silu_and_mul_scaled_nvfp4_experts_quantize" ) scaled_fp4_grouped_quantize = _lazy_import_wrapper( "flashinfer", "scaled_fp4_grouped_quantize" ) nvfp4_block_scale_interleave = _lazy_import_wrapper( "flashinfer.fp4_quantization", "block_scale_interleave" ) trtllm_fp4_block_scale_moe = _lazy_import_wrapper( "flashinfer", "trtllm_fp4_block_scale_moe" ) # Special case for autotune since it returns a context manager autotune = _lazy_import_wrapper( "flashinfer.autotuner", "autotune", fallback_fn=lambda *args, **kwargs: contextlib.nullcontext(), ) @functools.cache def has_flashinfer_comm() -> bool: """Return `True` if FlashInfer comm module is available.""" return has_flashinfer() and importlib.util.find_spec("flashinfer.comm") is not None @functools.cache def has_flashinfer_all2all() -> bool: """Return `True` if FlashInfer mnnvl all2all is available.""" if not has_flashinfer_comm(): return False # Check if all required functions are available required_functions = [ ("flashinfer.comm", "Mapping"), ("flashinfer.comm.mnnvl", "MnnvlMemory"), ("flashinfer.comm.trtllm_alltoall", "MnnvlMoe"), ("flashinfer.comm.trtllm_alltoall", "MoEAlltoallInfo"), ] for module_name, attr_name in required_functions: mod = _get_submodule(module_name) if not mod or not hasattr(mod, attr_name): return False return True @functools.cache def has_flashinfer_moe() -> bool: """Return `True` if FlashInfer MoE module is available.""" return ( has_flashinfer() and importlib.util.find_spec("flashinfer.fused_moe") is not None ) @functools.cache def has_flashinfer_cutedsl() -> bool: """Return ``True`` if FlashInfer cutedsl module is available.""" return ( has_flashinfer() and importlib.util.find_spec("flashinfer.cute_dsl") is not None ) @functools.cache def has_flashinfer_trtllm_fused_moe() -> bool: """Return `True` if FlashInfer TRTLLM fused MoE is available.""" if not has_flashinfer_moe(): return False required_functions = [ ("flashinfer.fused_moe", "trtllm_fp8_block_scale_moe"), ("flashinfer.fused_moe", "trtllm_fp8_per_tensor_scale_moe"), ("flashinfer.fused_moe", "trtllm_fp4_block_scale_moe"), ("flashinfer.fused_moe", "trtllm_mxint4_block_scale_moe"), ] for module_name, attr_name in required_functions: mod = _get_submodule(module_name) if not mod or not hasattr(mod, attr_name): return False return True @functools.cache def has_flashinfer_cutlass_fused_moe() -> bool: """Return `True` if FlashInfer CUTLASS fused MoE is available.""" if not has_flashinfer_moe(): return False # Check if all required functions are available required_functions = [ ("flashinfer.fused_moe", "cutlass_fused_moe"), ("flashinfer", "fp4_quantize"), ("flashinfer", "nvfp4_block_scale_interleave"), ("flashinfer.fused_moe", "trtllm_fp4_block_scale_moe"), ] for module_name, attr_name in required_functions: mod = _get_submodule(module_name) if not mod or not hasattr(mod, attr_name): return False return True @functools.cache def has_flashinfer_cutedsl_grouped_gemm_nt_masked() -> bool: """Return ``True`` if FlashInfer CUTLASS fused MoE is available.""" if not has_flashinfer_cutedsl(): return False # Check if all required functions are available required_functions = [ ("flashinfer.cute_dsl.blockscaled_gemm", "grouped_gemm_nt_masked"), ("flashinfer", "scaled_fp4_grouped_quantize"), ("flashinfer", "silu_and_scaled_nvfp4_experts_quantize"), ] for module_name, attr_name in required_functions: mod = _get_submodule(module_name) if not mod or not hasattr(mod, attr_name): return False return True @functools.cache def has_nvidia_artifactory() -> bool: """Return `True` if NVIDIA's artifactory is accessible. This checks connectivity to the kernel inference library artifactory which is required for downloading certain cubin kernels like TRTLLM FHMA. """ # If we have pre-downloaded cubins, we can assume the cubins are available. if has_flashinfer_cubin(): return True try: # Use a short timeout to avoid blocking for too long response = requests.get(FLASHINFER_CUBINS_REPOSITORY, timeout=5) accessible = response.status_code == 200 if accessible: logger.debug_once("NVIDIA artifactory is accessible") else: logger.warning_once( "NVIDIA artifactory returned failed status code: %d", response.status_code, ) return accessible except Exception as e: logger.warning_once("Failed to connect to NVIDIA artifactory: %s", e) return False @functools.cache def supports_trtllm_attention() -> bool: """ TRTLLM attention is supported if the platform is SM100, NVIDIA artifactory is accessible, and batch-invariant mode is not enabled. """ # Batch-invariant mode disables TRTLLM attention if vllm_is_batch_invariant(): return False # Requires SM100 and NVIDIA artifactory to be accessible to download cubins return ( current_platform.is_device_capability_family(100) and has_nvidia_artifactory() ) def force_use_trtllm_attention() -> bool | None: """ This function should only be called during initialization stage when vllm config is set. Return `None` if --attention-config.use_trtllm_attention is not set, return `True` if TRTLLM attention is forced to be used, return `False` if TRTLLM attention is forced to be not used. """ from vllm.config import get_current_vllm_config vllm_config = get_current_vllm_config() return vllm_config.attention_config.use_trtllm_attention def can_use_trtllm_attention(num_qo_heads: int, num_kv_heads: int) -> bool: """Check if the current configuration supports TRTLLM attention.""" if force_use_trtllm_attention() is False: return False has_trtllm = supports_trtllm_attention() return has_trtllm and (num_qo_heads % num_kv_heads == 0) def use_trtllm_attention( num_qo_heads: int, num_kv_heads: int, num_tokens: int, max_seq_len: int, dcp_world_size: int, kv_cache_dtype: str, q_dtype: torch.dtype, is_prefill: bool, # None means auto-detection, True means force on, False means force off force_use_trtllm: bool | None = None, has_sinks: bool = False, has_spec: bool = False, ) -> bool: """Return `True` if TRTLLM attention is used.""" # CLI argument is set to 0 - respect it if force_use_trtllm is not None and not force_use_trtllm: return False # Decode context parallel is not supported if dcp_world_size > 1: logger.warning_once( "Trtllm does not support returning LSE and as a result " "does not support DCP, reverting to FlashInfer" ) return False # The platform is not supported if not supports_trtllm_attention(): if force_use_trtllm: logger.warning_once( "TRTLLM attention is not supported on this platform, " "but --attention-config.use_trtllm_attention is set to 1" ) return False # The combination of query and key heads is not supported if num_qo_heads % num_kv_heads != 0: if force_use_trtllm: logger.warning_once( "TRTLLM attention is not supported for this combination of " "query and key heads, but --attention-config.use_trtllm_attention is " "set to 1" ) return False if has_spec and not is_prefill: # Speculative decoding requires TRTLLM attention for decodes logger.info_once("Using TRTLLM attention (enabled for speculative decoding).") return True # Must use TRTLLM attention if query is FP8 quantized if q_dtype == current_platform.fp8_dtype(): logger.info_once("Using TRTLLM attention (query is quantized).") return True # If sinks are being used, we must use TRTLLM attention as it's # the only backend that supports them if has_sinks: logger.info_once("Using TRTLLM attention (required for attention sinks).") return True if force_use_trtllm is None: # CLI argument not set - use auto-detection if is_prefill: # Prefill auto-detection use_trtllm = kv_cache_dtype == "auto" if use_trtllm: logger.warning_once("Using TRTLLM prefill attention (auto-detected).") else: # Decode auto-detection use_trtllm = num_tokens <= 256 and kv_cache_dtype == "auto" if use_trtllm: logger.warning_once("Using TRTLLM decode attention (auto-detected).") return use_trtllm # CLI argument is set to 1 - respect it logger.info_once( "Using TRTLLM attention (--attention-config.use_trtllm_attention is set to 1)" ) return True if has_flashinfer(): from vllm.utils.torch_utils import direct_register_custom_op def _flashinfer_concat_mla_k( k: torch.Tensor, k_nope: torch.Tensor, k_pe: torch.Tensor, ) -> None: """Custom op wrapper for flashinfer's concat_mla_k. This is an in-place operation that concatenates k_nope and k_pe into k. The kernel is optimized for DeepSeek V3 dimensions: - num_heads=128 - nope_dim=128 - rope_dim=64 Key optimizations: - Warp-based processing with software pipelining - Vectorized memory access (int2 for nope, int for rope) - L2 prefetching for next row while processing current - Register reuse for rope values across all heads Args: k: Output tensor, shape [num_tokens, num_heads, nope_dim + rope_dim]. Modified in-place. k_nope: The nope part of k, shape [num_tokens, num_heads, nope_dim]. k_pe: The rope part of k (shared), shape [num_tokens, 1, rope_dim]. This is broadcast to all heads. """ from flashinfer.concat_ops import concat_mla_k concat_mla_k(k, k_nope, k_pe) def _flashinfer_concat_mla_k_fake( k: torch.Tensor, k_nope: torch.Tensor, k_pe: torch.Tensor, ) -> None: return # Register flashinfer concat_mla_k custom op direct_register_custom_op( op_name="flashinfer_concat_mla_k", op_func=_flashinfer_concat_mla_k, mutates_args=["k"], # k tensor is modified in-place fake_impl=_flashinfer_concat_mla_k_fake, ) @torch.library.custom_op( "vllm::flashinfer_mm_fp4", mutates_args=[], device_types="cuda", ) def flashinfer_mm_fp4( A: torch.Tensor, B: torch.Tensor, A_scale: torch.Tensor, B_scale: torch.Tensor, g_scale: torch.Tensor, dtype: torch.dtype, use_8x4_sf_layout: bool, backend: str, ) -> torch.Tensor: from flashinfer import mm_fp4 as flashinfer_mm_fp4_ return flashinfer_mm_fp4_( A, B, A_scale, B_scale, g_scale, dtype, block_size=16, use_8x4_sf_layout=use_8x4_sf_layout, backend=backend, ) @torch.library.register_fake( "vllm::flashinfer_mm_fp4", ) def flashinfer_mm_fp4_fake( A: torch.Tensor, B: torch.Tensor, A_scale: torch.Tensor, B_scale: torch.Tensor, g_scale: torch.Tensor, dtype: torch.dtype, use_8x4_sf_layout: bool, backend: str, ) -> torch.Tensor: return torch.empty(A.shape[0], B.shape[1], dtype=dtype, device=A.device) @torch.library.custom_op( "vllm::bmm_fp8", mutates_args=[], device_types="cuda", ) def bmm_fp8( A: torch.Tensor, B: torch.Tensor, A_scale: torch.Tensor, B_scale: torch.Tensor, dtype: torch.dtype, backend: str, ) -> torch.Tensor: from flashinfer import bmm_fp8 as bmm_fp8_ return bmm_fp8_(A, B, A_scale, B_scale, dtype, None, backend) @torch.library.register_fake( "vllm::bmm_fp8", ) def bmm_fp8_fake( A: torch.Tensor, B: torch.Tensor, A_scale: torch.Tensor, B_scale: torch.Tensor, dtype: torch.dtype, backend: str, ) -> torch.Tensor: return torch.empty( A.shape[0], A.shape[1], B.shape[2], dtype=dtype, device=A.device ) @torch.library.custom_op( "vllm::flashinfer_nvfp4_quantize", mutates_args=[], device_types="cuda", ) def flashinfer_nvfp4_quantize( a: torch.Tensor, a_global_sf: torch.Tensor ) -> tuple[torch.Tensor, torch.Tensor]: from flashinfer import SfLayout from flashinfer import nvfp4_quantize as nvfp4_quantize_ return nvfp4_quantize_( a, a_global_sf, sfLayout=SfLayout.layout_8x4, do_shuffle=False ) @torch.library.register_fake( "vllm::flashinfer_nvfp4_quantize", ) def flashinfer_nvfp4_quantize_fake( a: torch.Tensor, a_global_sf: torch.Tensor ) -> tuple[torch.Tensor, torch.Tensor]: m, n = a.shape round_up = lambda x, y: (x + y - 1) // y * y rounded_m = round_up(m, 8) scale_n = n // 16 rounded_n = round_up(scale_n, 4) return torch.empty(m, n // 2, dtype=torch.uint8, device=a.device), torch.empty( rounded_m, rounded_n, dtype=torch.uint8, device=a.device ) @torch.library.custom_op( "vllm::mm_mxfp8", mutates_args=[], device_types="cuda", ) def mm_mxfp8( A: torch.Tensor, B: torch.Tensor, A_scale: torch.Tensor, B_scale: torch.Tensor, out_dtype: torch.dtype, backend: str = "cutlass", ) -> torch.Tensor: from flashinfer import mm_mxfp8 as mm_mxfp8_ return mm_mxfp8_( A, B, A_scale, B_scale, out=None, out_dtype=out_dtype, backend=backend, ) @torch.library.register_fake( "vllm::mm_mxfp8", ) def mm_mxfp8_fake( A: torch.Tensor, B: torch.Tensor, A_scale: torch.Tensor, B_scale: torch.Tensor, out_dtype: torch.dtype, backend: str = "cutlass", ) -> torch.Tensor: # A is [m, k], B is [k, n] -> output [m, n] return torch.empty(A.shape[0], B.shape[1], dtype=out_dtype, device=A.device) def flashinfer_mm_mxfp8( a: torch.Tensor, b: torch.Tensor, block_scale_a: torch.Tensor, block_scale_b: torch.Tensor, out_dtype: torch.dtype, backend: str = "cutlass", ) -> torch.Tensor: """MXFP8 MM helper - mirrors flashinfer_scaled_fp4_mm API. Takes non-transposed weights and handles transpose internally. CRITICAL: mm_mxfp8 CUTLASS kernel requires SWIZZLED 1D scales for optimal performance and accuracy. Both input and weight scales should be in swizzled format from FlashInfer's mxfp8_quantize(is_sf_swizzled_layout=True). """ # a shape [M, K] # b shape [K, N] assert a.ndim == 2 and b.ndim == 2 assert a.shape[1] == b.shape[1] # K dimension must match if block_scale_b.ndim != 1: raise ValueError( "mm_mxfp8 expects 1D swizzled weight scales for CUTLASS; " f"got shape={tuple(block_scale_b.shape)}" ) # Output tensor [M, N] return mm_mxfp8( a, b.t(), # Transpose weight: [N, K] -> [K, N] block_scale_a, block_scale_b, out_dtype, backend=backend, ) def flashinfer_scaled_fp4_mm( a: torch.Tensor, b: torch.Tensor, block_scale_a: torch.Tensor, block_scale_b: torch.Tensor, alpha: torch.Tensor, out_dtype: torch.dtype, backend: str, ) -> torch.Tensor: assert a.ndim == 2 and b.ndim == 2 assert block_scale_a.ndim == 2 and block_scale_b.ndim == 2 assert a.stride(-1) == 1 and b.stride(-1) == 1 assert a.shape[1] == b.shape[1] if backend in ("cutlass", "cudnn"): block_scale_a = block_scale_a.view(torch.uint8) block_scale_b = block_scale_b.view(torch.uint8) use_8x4_sf_layout = True if backend == "trtllm" and a.shape[0] <= 32 else False # noqa: SIM210 return flashinfer_mm_fp4( a, b.t(), block_scale_a, block_scale_b.t(), alpha, out_dtype, use_8x4_sf_layout=use_8x4_sf_layout, backend=backend, ) def flashinfer_scaled_fp8_mm( a: torch.Tensor, b: torch.Tensor, scale_a: torch.Tensor, scale_b: torch.Tensor, out_dtype: torch.dtype, bias: torch.Tensor | None = None, ) -> torch.Tensor: assert a.ndim == 2 and b.ndim == 2 assert a.shape[1] == b.shape[0] assert scale_a.numel() == 1 and scale_b.numel() == 1 assert a.dtype == torch.float8_e4m3fn and b.dtype == torch.float8_e4m3fn assert a.device.type == "cuda" and b.device.type == "cuda" assert scale_a.dtype == torch.float32 and scale_b.dtype == torch.float32 assert scale_a.device.type == "cuda" and scale_b.device.type == "cuda" output = bmm_fp8( a.unsqueeze(0), b.unsqueeze(0), scale_a, scale_b, out_dtype, "auto", ).view(a.shape[0], b.shape[1]) if bias is not None: output = output + bias return output def flashinfer_quant_nvfp4_8x4_sf_layout( a: torch.Tensor, a_global_sf: torch.Tensor ) -> tuple[torch.Tensor, torch.Tensor]: return flashinfer_nvfp4_quantize(a, a_global_sf) flashinfer_fp8_blockscale_gemm = _lazy_import_wrapper( "flashinfer.gemm", "fp8_blockscale_gemm_sm90" ) @functools.cache def has_flashinfer_fp8_blockscale_gemm() -> bool: """Return `True` if FlashInfer block-scale FP8 GEMM is available.""" return ( has_flashinfer() and current_platform.is_device_capability(90) and hasattr(_get_submodule("flashinfer.gemm"), "fp8_blockscale_gemm_sm90") ) @functools.cache def is_flashinfer_fp8_blockscale_gemm_supported() -> bool: """Return `True` if FlashInfer block-scale FP8 GEMM is supported.""" return ( envs.VLLM_BLOCKSCALE_FP8_GEMM_FLASHINFER and has_flashinfer_fp8_blockscale_gemm() ) def should_use_flashinfer_for_blockscale_fp8_gemm( is_flashinfer_supported: bool, output_dtype: torch.dtype, input: torch.Tensor, weight: torch.Tensor, ): if not is_flashinfer_supported: return False # Verify DeepGEMM N/K dims requirements # NOTE: Also synchronized with test_w8a8_block_fp8_deep_gemm_matmul # test inside kernels/quantization/test_block_fp8.py N_MULTIPLE = 64 K_MULTIPLE = 128 weight_dtype = weight.dtype input_dtype = input.dtype should_use_flashinfer = ( output_dtype == torch.bfloat16 and input_dtype == torch.bfloat16 and weight_dtype == torch.float8_e4m3fn and weight.shape[0] % N_MULTIPLE == 0 and weight.shape[1] % K_MULTIPLE == 0 ) return should_use_flashinfer __all__ = [ "has_flashinfer", "flashinfer_trtllm_fp8_block_scale_moe", "flashinfer_cutlass_fused_moe", "flashinfer_cutedsl_grouped_gemm_nt_masked", "flashinfer_fp4_quantize", "silu_and_mul_scaled_nvfp4_experts_quantize", "scaled_fp4_grouped_quantize", "nvfp4_block_scale_interleave", "trtllm_fp4_block_scale_moe", "autotune", "has_flashinfer_moe", "has_flashinfer_comm", "has_flashinfer_all2all", "has_flashinfer_cutlass_fused_moe", "has_flashinfer_cutedsl_grouped_gemm_nt_masked", "has_flashinfer_fp8_blockscale_gemm", "has_nvidia_artifactory", "supports_trtllm_attention", "can_use_trtllm_attention", "use_trtllm_attention", "flashinfer_scaled_fp4_mm", "flashinfer_scaled_fp8_mm", "flashinfer_quant_nvfp4_8x4_sf_layout", "flashinfer_fp8_blockscale_gemm", "should_use_flashinfer_for_blockscale_fp8_gemm", "is_flashinfer_fp8_blockscale_gemm_supported", ]
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/utils/flashinfer.py", "license": "Apache License 2.0", "lines": 662, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:examples/offline_inference/skip_loading_weights_in_engine_init.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from vllm import LLM, RequestOutput, SamplingParams # Sample prompts. prompts = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] # Create a sampling params object. sampling_params = SamplingParams(temperature=0.8, top_p=0.95) def print_prompts_and_outputs(outputs: list[RequestOutput]) -> None: print("-" * 60) for output in outputs: prompt = output.prompt generated_text = output.outputs[0].text print(f"Prompt: {prompt!r}") print(f"Output: {generated_text!r}") print("-" * 60) def main(): # Create an LLM without loading real weights llm = LLM( model="Qwen/Qwen3-0.6B", load_format="dummy", enforce_eager=True, tensor_parallel_size=4, ) outputs = llm.generate(prompts, sampling_params) print("\nOutputs do not make sense:") print_prompts_and_outputs(outputs) # Update load format from `dummy` to `auto` llm.collective_rpc( "update_config", args=({"load_config": {"load_format": "auto"}},) ) # Now reload real weights inplace llm.collective_rpc("reload_weights") # Check outputs make sense outputs = llm.generate(prompts, sampling_params) print("\nOutputs make sense after loading real weights:") print_prompts_and_outputs(outputs) if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "examples/offline_inference/skip_loading_weights_in_engine_init.py", "license": "Apache License 2.0", "lines": 43, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/models/multimodal/processing/test_nemotron_vl.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests for Nemotron-Nano-VL's multimodal preprocessing kwargs.""" from collections.abc import Mapping import pytest from PIL import Image from transformers import PretrainedConfig from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.image import rescale_image_size from vllm.multimodal.processing import BaseMultiModalProcessor from ....conftest import ImageTestAssets from ...utils import build_model_context def _get_expected_num_patches( config: PretrainedConfig, image: Image.Image, num_imgs: int, min_num: int, max_num: int, ): from vllm.model_executor.models.nemotron_vl import ( calculate_nemotron_vl_targets, get_nemotron_vl_target_ratios, ) width, height = image.size blocks, _, _ = calculate_nemotron_vl_targets( orig_width=width, orig_height=height, target_ratios=get_nemotron_vl_target_ratios( min_num, max_num, ), image_size=config.force_image_size, use_thumbnail=False, ) expected_num_patches = blocks if config.use_thumbnail and expected_num_patches > 1: expected_num_patches += 1 return expected_num_patches def _run_check( processor: BaseMultiModalProcessor, images: list[Image.Image], min_num: int, max_num: int, mm_processor_kwargs: Mapping[str, object], ): tokenizer = processor.info.get_tokenizer() config = processor.info.get_hf_config() image_processor = processor.info.get_image_processor() config.use_thumbnail = image_processor.use_thumbnail prompt = "<image>" * len(images) mm_data = {"image": images} total_expected_num_patches = sum( _get_expected_num_patches(config, image, len(images), min_num, max_num) for image in images ) print(total_expected_num_patches) processed_inputs = processor( prompt, mm_items=processor.info.parse_mm_data(mm_data), hf_processor_mm_kwargs=mm_processor_kwargs, ) # Ensure we have the right number of placeholders per num_crops size image_token_id = tokenizer.convert_tokens_to_ids("<image>") img_tok_count = processed_inputs["prompt_token_ids"].count(image_token_id) pixel_shape = processed_inputs["mm_kwargs"].get_data()["pixel_values_flat"].shape print("Image token count:", img_tok_count, "Pixel shape:", pixel_shape) assert img_tok_count == 256 * total_expected_num_patches assert pixel_shape[0] == total_expected_num_patches @pytest.mark.parametrize("model_id", ["nvidia/Llama-3.1-Nemotron-Nano-VL-8B-V1"]) @pytest.mark.parametrize( "size_factors", [ # Single-scale [1.0], # Single-scale, batched [1.0, 1.0, 1.0], # Multi-scale [0.25, 0.5, 1.0], [4.0, 2.0, 1.0], ], ) @pytest.mark.parametrize( ("min_dynamic_patch", "max_dynamic_patch"), [(1, 1), (1, 2), (1, 4), (1, 8), (2, 4), (4, 8)], ) @pytest.mark.parametrize("dynamic_image_size", [True, False]) @pytest.mark.parametrize("kwargs_on_init", [True, False]) def test_processor_override( model_id: str, image_assets: ImageTestAssets, size_factors: list[int], min_dynamic_patch: int, max_dynamic_patch: int, dynamic_image_size: bool | None, kwargs_on_init: bool, ): mm_processor_kwargs = { "min_dynamic_patch": min_dynamic_patch, "max_dynamic_patch": max_dynamic_patch, "dynamic_image_size": dynamic_image_size, } ctx = build_model_context( model_id, mm_processor_kwargs=mm_processor_kwargs if kwargs_on_init else None, limit_mm_per_prompt={"image": len(size_factors)}, ) processor = MULTIMODAL_REGISTRY.create_processor(ctx.model_config) hf_processor_mm_kwargs = {} if kwargs_on_init else mm_processor_kwargs min_num = min_dynamic_patch if dynamic_image_size else 1 max_num = max_dynamic_patch if dynamic_image_size else 1 _run_check( processor, [rescale_image_size(image_assets[0].pil_image, f) for f in size_factors], min_num, max_num, hf_processor_mm_kwargs, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/multimodal/processing/test_nemotron_vl.py", "license": "Apache License 2.0", "lines": 117, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/models/nemotron_vl.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # adapted from https://huggingface.co/OpenGVLab/InternVL2-4B/blob/main/modeling_internvl_chat.py # -------------------------------------------------------- # InternVL # Copyright (c) 2023 OpenGVLab # Licensed under The MIT License [see LICENSE for details] # -------------------------------------------------------- from abc import ABC from collections.abc import Iterable import torch import torch.nn as nn import torchvision.transforms as T from PIL import Image from transformers import AutoModel, PretrainedConfig from transformers.image_processing_utils_fast import BaseImageProcessorFast from vllm.config import VllmConfig from vllm.model_executor.layers.pooler import DispatchPooler from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.quantization.awq import AWQConfig from vllm.model_executor.models.internvl import ( BaseInternVLDummyInputsBuilder, BaseInternVLMultiModalProcessor, BaseInternVLProcessingInfo, InternVLImageEmbeddingInputs, InternVLImageInputs, InternVLImagePixelInputs, InternVLProcessor, ) from vllm.model_executor.models.module_mapping import MultiModelKeys from vllm.model_executor.models.siglip import SiglipVisionModel from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.image import convert_image_mode from vllm.multimodal.processing import PromptUpdateDetails from vllm.sequence import IntermediateTensors from vllm.tokenizers import TokenizerLike from vllm.transformers_utils.processor import cached_image_processor_from_config from vllm.transformers_utils.repo_utils import get_hf_file_to_dict from .interfaces import ( MultiModalEmbeddings, SupportsLoRA, SupportsMultiModal, SupportsPP, ) from .interfaces_base import VllmModelForPooling from .utils import ( AutoWeightsLoader, WeightsMapper, init_vllm_registered_model, maybe_prefix, ) def build_transform(input_size: int): return T.Compose( [ T.Lambda(lambda img: convert_image_mode(img, "RGB")), T.Resize( (input_size, input_size), interpolation=T.InterpolationMode.BICUBIC ), T.ToTensor(), ] ) # adapted from https://huggingface.co/nvidia/Llama-3.1-Nemotron-Nano-VL-8B-V1 def find_closest_aspect_ratio( aspect_ratio: float, target_ratios: list[tuple[int, int]], *, width: int, height: int, image_size: int, ) -> tuple[int, int]: best_factor = float("-inf") best_ratio = (1, 1) area = width * height for rw, rh in target_ratios: target_aspect_ratio = rw / rh size_factor = min((rw * rh * image_size * image_size) / area, 0.6) ratio_closeness = min( target_aspect_ratio / aspect_ratio, aspect_ratio / target_aspect_ratio ) factor = size_factor * ratio_closeness if factor > best_factor: best_factor = factor best_ratio = (rw, rh) return best_ratio def calculate_nemotron_vl_targets( *, orig_width: int, orig_height: int, target_ratios: list[tuple[int, int]], image_size: int, use_thumbnail: bool, ) -> tuple[int, int, int]: aspect_ratio = orig_width / orig_height # find the closest aspect ratio to the target target_aspect_ratio = find_closest_aspect_ratio( aspect_ratio, target_ratios, width=orig_width, height=orig_height, image_size=image_size, ) # calculate the target width and height target_width = image_size * target_aspect_ratio[0] target_height = image_size * target_aspect_ratio[1] blocks = target_aspect_ratio[0] * target_aspect_ratio[1] # add thumbnail image if num_blocks != 1 if use_thumbnail and blocks != 1: blocks += 1 return blocks, target_width, target_height def dynamic_preprocess_nemotron_vl( image: Image.Image, *, target_ratios: list[tuple[int, int]], image_size: int, use_thumbnail: bool, ) -> list[Image.Image]: orig_width, orig_height = image.size # calculate the number of blocks without thumbnail blocks, target_width, target_height = calculate_nemotron_vl_targets( orig_width=orig_width, orig_height=orig_height, target_ratios=target_ratios, image_size=image_size, use_thumbnail=False, ) # resize the image resized_img = image.resize((target_width, target_height)) processed_images = [] for i in range(blocks): box = ( (i % (target_width // image_size)) * image_size, (i // (target_width // image_size)) * image_size, ((i % (target_width // image_size)) + 1) * image_size, ((i // (target_width // image_size)) + 1) * image_size, ) # split the image split_img = resized_img.crop(box) processed_images.append(split_img) assert len(processed_images) == blocks if use_thumbnail and len(processed_images) != 1: thumbnail_img = image.resize((image_size, image_size)) processed_images.append(thumbnail_img) return processed_images def get_nemotron_vl_target_ratios( min_num: int, max_num: int, ) -> list[tuple[int, int]]: target_ratios = { (i, j) for n in range(min_num, max_num + 1) for i in range(1, n + 1) for j in range(1, n + 1) if min_num <= i * j <= max_num } return sorted(target_ratios, key=lambda x: x[0] * x[1]) def image_to_pixel_values_nemotron_vl( image: Image.Image, *, input_size: int, min_num: int, max_num: int, use_thumbnail: bool, transform: T.Compose | None = None, ) -> torch.Tensor: target_ratios = get_nemotron_vl_target_ratios(min_num, max_num) if transform is None: transform = build_transform(input_size=input_size) images = dynamic_preprocess_nemotron_vl( image, target_ratios=target_ratios, image_size=input_size, use_thumbnail=use_thumbnail, ) pixel_values = torch.stack([transform(image) for image in images]) return pixel_values class NemotronVLProcessor(InternVLProcessor): IMG_START = "<img>" IMG_END = "</img>" IMG_CONTEXT = "<image>" def __init__( self, config: PretrainedConfig, tokenizer: TokenizerLike, image_processor: BaseImageProcessorFast | None = None, *, min_dynamic_patch: int | None = None, max_dynamic_patch: int | None = None, dynamic_image_size: bool | None = None, ) -> None: ABC.__init__(self) self.config = config self.tokenizer = tokenizer self.image_processor = image_processor image_size: int = config.force_image_size patch_size: int = config.patch_size if min_dynamic_patch is None: min_dynamic_patch = 1 assert isinstance(min_dynamic_patch, int) if max_dynamic_patch is None: max_dynamic_patch = self.image_processor.max_num_tiles assert isinstance(max_dynamic_patch, int) if dynamic_image_size is None: dynamic_image_size = True assert isinstance(dynamic_image_size, bool) self.num_image_token = int( (image_size // patch_size) ** 2 * (config.downsample_ratio**2) ) self.image_size = image_size self.min_dynamic_patch = min_dynamic_patch self.max_dynamic_patch = max_dynamic_patch self.dynamic_image_size = dynamic_image_size if image_processor is not None: self.use_thumbnail = image_processor.use_thumbnail else: self.use_thumbnail = getattr(config, "use_thumbnail", True) @property def image_token_id(self) -> int: return self.tokenizer.get_vocab()[self.IMG_CONTEXT] def _get_transform(self) -> T.Compose: return build_transform(input_size=self.image_size) def get_num_image_tokens( self, *, image_width: int, image_height: int, ) -> int: target_ratios = self.resolve_target_ratios( use_thumbnail=False, # Applied in calculate_targets ) num_patches, _, _ = calculate_nemotron_vl_targets( orig_width=image_width, orig_height=image_height, image_size=self.image_size, target_ratios=target_ratios, use_thumbnail=self.use_thumbnail, ) return num_patches * self.num_image_token def _images_to_pixel_values_lst( self, images: list[Image.Image], min_dynamic_patch: int | None = None, max_dynamic_patch: int | None = None, dynamic_image_size: bool | None = None, ) -> list[torch.Tensor]: min_num, max_num = self.resolve_min_max_num( min_dynamic_patch=min_dynamic_patch, max_dynamic_patch=max_dynamic_patch, dynamic_image_size=dynamic_image_size, use_thumbnail=False, # Applied in image_to_pixel_values ) return [ image_to_pixel_values_nemotron_vl( image, input_size=self.image_size, min_num=min_num, max_num=max_num, use_thumbnail=self.use_thumbnail, transform=self._get_transform(), ) for image in images ] def _replace_image_tokens( self, text: list[str], pixel_values_lst: list[torch.Tensor], ) -> list[str]: """Replace <image> placeholders with image tokens.""" for pixel_values in pixel_values_lst: num_patches = pixel_values.shape[0] feature_size = num_patches * self.num_image_token image_repl = self.get_image_repl(feature_size, num_patches) # Use temporary placeholder to avoid replacing tokens we just inserted NVL_IMAGE_CONTEXT = image_repl.full.replace("<image>", "<NVL_IMG_CONTEXT>") text = [t.replace("<image>", NVL_IMAGE_CONTEXT, 1) for t in text] return [t.replace("<NVL_IMG_CONTEXT>", self.IMG_CONTEXT) for t in text] def _preprocess_image( self, text: list[str], images: list[Image.Image], min_dynamic_patch: int | None = None, max_dynamic_patch: int | None = None, dynamic_image_size: bool | None = None, ) -> tuple[list[str], dict[str, torch.Tensor]]: if len(images) == 0: image_inputs = {} else: pixel_values_lst = self._images_to_pixel_values_lst( images, min_dynamic_patch=min_dynamic_patch, max_dynamic_patch=max_dynamic_patch, dynamic_image_size=dynamic_image_size, ) image_inputs = { "pixel_values_flat": torch.cat(pixel_values_lst), "image_num_patches": torch.tensor( [len(item) for item in pixel_values_lst] ), } text = self._replace_image_tokens(text, pixel_values_lst) return text, image_inputs def get_image_repl( self, feature_size: int, num_patches: int | None, ) -> PromptUpdateDetails[str]: repl_features = self.IMG_CONTEXT * feature_size repl_full = self.IMG_START + repl_features + self.IMG_END return PromptUpdateDetails.select_text(repl_full, self.IMG_CONTEXT) class NemotronVLProcessingInfo(BaseInternVLProcessingInfo): """Processing info for Nemotron VL models.""" def get_hf_processor(self, **kwargs: object) -> NemotronVLProcessor: return self.ctx.init_processor( NemotronVLProcessor, config=self.get_hf_config(), tokenizer=self.get_tokenizer(), image_processor=self.get_image_processor(), **kwargs, ) def get_image_processor(self, **kwargs: object): return cached_image_processor_from_config( self.ctx.model_config, **kwargs, ) @MULTIMODAL_REGISTRY.register_processor( BaseInternVLMultiModalProcessor[NemotronVLProcessingInfo], info=NemotronVLProcessingInfo, dummy_inputs=BaseInternVLDummyInputsBuilder[NemotronVLProcessingInfo], ) class LlamaNemotronVLChatModel(nn.Module, SupportsMultiModal, SupportsPP, SupportsLoRA): @classmethod def get_placeholder_str(cls, modality: str, i: int) -> str | None: if modality.startswith("image"): return "<image>" raise ValueError("Only image modality is supported") def __init__(self, *, vllm_config: VllmConfig, prefix: str = "") -> None: super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config multimodal_config = vllm_config.model_config.multimodal_config self.config = config self.multimodal_config = multimodal_config self._patch_quant_config(config, quant_config) image_size = config.force_image_size or config.vision_config.image_size patch_size = config.vision_config.patch_size self.patch_size = patch_size self.num_image_token = int( (image_size // patch_size) ** 2 * (config.downsample_ratio**2) ) self.downsample_ratio = config.downsample_ratio self.ps_version = config.ps_version with self._mark_tower_model(vllm_config, "image"): self.vision_model = self._init_vision_model( config, quant_config=quant_config, prefix=maybe_prefix(prefix, "vision_model"), ) self.mlp1 = self._init_mlp1(config) with self._mark_language_model(vllm_config): self.language_model = init_vllm_registered_model( vllm_config=vllm_config, hf_config=config.get_text_config(), prefix=maybe_prefix(prefix, "language_model"), ) self.img_context_token_id = None self.visual_token_mask = None self.make_empty_intermediate_tensors = ( self.language_model.make_empty_intermediate_tensors ) def _patch_quant_config( self, config: PretrainedConfig, quant_config: QuantizationConfig ): # the awq models from OpenGVLab missing `modules_to_not_convert` # patch the quant_config to add `modules_to_not_convert` back if isinstance(quant_config, AWQConfig): text_config = config.get_text_config() llm_quant_config = getattr(text_config, "quantization_config", None) if (not quant_config.modules_to_not_convert) and ( llm_quant_config is not None ): quant_config.modules_to_not_convert.append("vision_model") def _init_vision_model( self, config: PretrainedConfig, quant_config: QuantizationConfig | None, *, prefix: str, ): return AutoModel.from_config(config.vision_config, trust_remote_code=True) def _init_mlp1( self, config: PretrainedConfig, vit_hidden_size: int | None = None, vision_projection_hidden_size: int | None = None, ) -> nn.Module: if vit_hidden_size is None: vit_hidden_size = config.vit_hidden_size if vision_projection_hidden_size is None: vision_projection_hidden_size = config.projector_hidden_size llm_hidden_size = config.get_text_config().hidden_size return nn.Sequential( nn.LayerNorm( vit_hidden_size * int(1 / self.downsample_ratio) ** 2, bias=True ), nn.Linear( vit_hidden_size * int(1 / self.downsample_ratio) ** 2, vision_projection_hidden_size, bias=True, ), nn.GELU(), nn.Linear(vision_projection_hidden_size, llm_hidden_size), ) def pixel_shuffle(self, x, scale_factor=0.5): n, w, h, c = x.size() # N, W, H, C --> N, W, H * scale, C // scale x = x.view(n, w, int(h * scale_factor), int(c / scale_factor)) # N, W, H * scale, C // scale --> N, H * scale, W, C // scale x = x.permute(0, 2, 1, 3).contiguous() x = x.view( n, int(h * scale_factor), int(w * scale_factor), int(c / (scale_factor * scale_factor)), ) if self.ps_version == "v1": pass else: x = x.permute(0, 2, 1, 3).contiguous() return x def _call_vision_model(self, pixel_values: torch.Tensor) -> torch.Tensor: """Call vision model and return embeddings. Override this method in subclasses to handle different vision model interfaces (e.g., SigLIP vs C-RADIO). """ vit_embeds = self.vision_model(x=pixel_values).features return vit_embeds.to(dtype=torch.bfloat16) def extract_feature(self, pixel_values: torch.Tensor) -> torch.Tensor: # https://huggingface.co/nvidia/Llama-3.1-Nemotron-Nano-VL-8B-V1/blob/main/modeling.py#L177 vit_embeds = self._call_vision_model(pixel_values) h = w = int(vit_embeds.shape[1] ** 0.5) vit_embeds = vit_embeds.reshape(vit_embeds.shape[0], h, w, -1) vit_embeds = self.pixel_shuffle(vit_embeds, scale_factor=self.downsample_ratio) vit_embeds = vit_embeds.reshape(vit_embeds.shape[0], -1, vit_embeds.shape[-1]) vit_embeds = self.mlp1(vit_embeds) return vit_embeds def _parse_and_validate_image_input( self, **kwargs: object ) -> InternVLImageInputs | None: pixel_values_flat = kwargs.pop("pixel_values_flat", None) image_num_patches = kwargs.pop("image_num_patches", None) image_embeds = kwargs.pop("image_embeds", None) if pixel_values_flat is None and image_embeds is None: return None if image_embeds is not None: return InternVLImageEmbeddingInputs( type="image_embeds", data=image_embeds, ) image_token_id = kwargs["image_token_id"] if isinstance(image_token_id, torch.Tensor): image_token_id = image_token_id.flatten().unique().item() assert isinstance(image_token_id, int) self.img_context_token_id = image_token_id if pixel_values_flat is not None: return InternVLImagePixelInputs( type="pixel_values", pixel_values_flat=pixel_values_flat, num_patches=image_num_patches, resolve_bindings={ "h": self.config.force_image_size, "w": self.config.force_image_size, }, ) raise AssertionError("This line should be unreachable.") def _process_image_input( self, image_input: InternVLImageInputs, ) -> tuple[torch.Tensor, ...]: if image_input["type"] == "image_embeds": return image_input["data"] image_embeds = self.extract_feature(image_input["pixel_values_flat"]) num_patches = image_input["num_patches"] hidden_size = self.config.get_text_config().hidden_size # Only one image in the current batch if len(num_patches) == 1: return (image_embeds.view(-1, hidden_size),) # NOTE: Image embeddings are split into separate tensors for each image # by the size of each embedding. feature_size = image_embeds.shape[1] image_embeds = image_embeds.view(-1, hidden_size) image_feature_sizes = [ num_patches * feature_size for num_patches in num_patches ] return image_embeds.split(image_feature_sizes) def _parse_and_validate_multimodal_inputs(self, **kwargs: object) -> dict: modalities = {} # Preserve the order of modalities if there are multiple of them # from the order of kwargs. for input_key in kwargs: if ( input_key in ("pixel_values_flat", "image_embeds") and "images" not in modalities ): modalities["images"] = self._parse_and_validate_image_input(**kwargs) return modalities def _set_visual_token_mask(self, input_ids: torch.Tensor) -> None: self.visual_token_mask = None def embed_multimodal(self, **kwargs: object) -> MultiModalEmbeddings: modalities = self._parse_and_validate_multimodal_inputs(**kwargs) if not modalities: return [] # The result multimodal_embeddings is tuple of tensors, with each # tensor corresponding to a multimodal data item (image). multimodal_embeddings: tuple[torch.Tensor, ...] = () # NOTE: It is important to iterate over the keys in this dictionary # to preserve the order of the modalities. for modality in modalities: if modality == "images": image_input = modalities["images"] image_embeddings = self._process_image_input(image_input) multimodal_embeddings += tuple(image_embeddings) return multimodal_embeddings def embed_input_ids( self, input_ids: torch.Tensor, multimodal_embeddings: MultiModalEmbeddings | None = None, *, is_multimodal: torch.Tensor | None = None, handle_oov_mm_token: bool = False, ) -> torch.Tensor: if multimodal_embeddings is not None and len(multimodal_embeddings) > 0: self._set_visual_token_mask(input_ids) # This is to satisfy the type checker for each overload if multimodal_embeddings is None or is_multimodal is None: return super().embed_input_ids(input_ids) return super().embed_input_ids( input_ids, multimodal_embeddings=multimodal_embeddings, is_multimodal=is_multimodal, handle_oov_mm_token=handle_oov_mm_token, ) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs: object, ) -> IntermediateTensors: if intermediate_tensors is not None: inputs_embeds = None forward_kwargs = { "input_ids": input_ids, "positions": positions, "intermediate_tensors": intermediate_tensors, "inputs_embeds": inputs_embeds, } # Only required if the model is mono-architecture if self.visual_token_mask is not None: forward_kwargs.update({"visual_token_mask": self.visual_token_mask}) self.visual_token_mask = None hidden_states = self.language_model.model(**forward_kwargs) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: return self.language_model.compute_logits(hidden_states) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: ## Ignore registered_buffers ## see https://huggingface.co/nvidia/C-RADIOv2-H/blob/main/input_conditioner.py#L28 # noqa: E501 skip_substrs = ["norm_mean", "norm_std"] loader = AutoWeightsLoader(self, skip_substrs=skip_substrs) return loader.load_weights(weights) def get_mm_mapping(self) -> MultiModelKeys: """ Get the module prefix in multimodal models """ return MultiModelKeys.from_string_field( language_model="language_model", connector="mlp1", tower_model="vision_model", ) # -------------------------------------------------------- # LlamaNemotronVL Embedding Model (nvidia/llama-nemotron-embed-vl-1b-v2) # Extends LlamaNemotronVLChatModel for embedding/pooling tasks: # - SigLIP vision encoder (instead of C-RADIO) # - Bidirectional (non-causal) LLaMA language model # - Pooler output instead of generative logits # -------------------------------------------------------- # SigLIP normalization constants SIGLIP_MEAN = (0.5, 0.5, 0.5) SIGLIP_STD = (0.5, 0.5, 0.5) def build_siglip_transform(input_size: int): """Build transform for SigLIP vision encoder with normalization. Extends the base transform from nemotron_vl with SigLIP-specific normalization. """ base_transform = build_transform(input_size=input_size) return T.Compose( [ base_transform, T.Normalize(mean=SIGLIP_MEAN, std=SIGLIP_STD), ] ) class LlamaNemotronVLEmbedProcessor(NemotronVLProcessor): """ Processor for LlamaNemotronVL embedding model. Inherits from NemotronVLProcessor and specializes it for embedding tasks: - Uses SigLIP transform with normalization instead of base transform - Uses different image context token (<IMG_CONTEXT> vs <image>) """ IMG_CONTEXT = "<IMG_CONTEXT>" def __init__( self, config: PretrainedConfig, tokenizer: TokenizerLike, processor_config: dict, *, min_dynamic_patch: int | None = None, max_dynamic_patch: int | None = None, dynamic_image_size: bool | None = None, ) -> None: if min_dynamic_patch is None: min_dynamic_patch = processor_config.get( "min_input_tiles", getattr(config, "min_dynamic_patch", 1), ) if max_dynamic_patch is None: max_dynamic_patch = processor_config.get( "max_input_tiles", getattr(config, "max_dynamic_patch", 1), ) if dynamic_image_size is None: dynamic_image_size = processor_config.get( "dynamic_image_size", getattr(config, "dynamic_image_size", True), ) super().__init__( config=config, tokenizer=tokenizer, image_processor=None, min_dynamic_patch=min_dynamic_patch, max_dynamic_patch=max_dynamic_patch, dynamic_image_size=dynamic_image_size, ) def _get_transform(self) -> T.Compose: """Override to add SigLIP normalization.""" return build_siglip_transform(input_size=self.image_size) def _replace_image_tokens( self, text: list[str], pixel_values_lst: list[torch.Tensor], ) -> list[str]: """Override with simpler token replacement for embedding model. No temporary placeholder needed because IMG_CONTEXT is <IMG_CONTEXT>, not <image>, so there's no collision risk. """ for pixel_values in pixel_values_lst: num_patches = pixel_values.shape[0] feature_size = num_patches * self.num_image_token image_repl = self.get_image_repl(feature_size, num_patches) text = [t.replace("<image>", image_repl.full, 1) for t in text] return text class LlamaNemotronVLEmbedProcessingInfo(NemotronVLProcessingInfo): """Processing info for LlamaNemotronVL embedding model.""" def get_hf_processor(self, **kwargs: object) -> LlamaNemotronVLEmbedProcessor: """Override to create embedding-specific processor without image_processor.""" model_config = self.ctx.model_config processor_config = {} if model_config.model is not None: processor_config = ( get_hf_file_to_dict( "processor_config.json", model_config.model, model_config.revision, ) or {} ) return self.ctx.init_processor( LlamaNemotronVLEmbedProcessor, config=self.get_hf_config(), tokenizer=self.get_tokenizer(), processor_config=processor_config, **kwargs, ) @MULTIMODAL_REGISTRY.register_processor( BaseInternVLMultiModalProcessor[LlamaNemotronVLEmbedProcessingInfo], info=LlamaNemotronVLEmbedProcessingInfo, dummy_inputs=BaseInternVLDummyInputsBuilder[LlamaNemotronVLEmbedProcessingInfo], ) class LlamaNemotronVLForEmbedding(LlamaNemotronVLChatModel, VllmModelForPooling): """ LlamaNemotronVL model for embeddings. Inherits from LlamaNemotronVLChatModel and specializes it for embedding tasks: - Uses SigLIP vision encoder instead of C-RADIO - Uses bidirectional LLaMA (via llm_config) instead of causal LLaMA - Adds pooler for embedding output instead of generating logits """ is_pooling_model = True # Weight mapping from checkpoint format to vLLM format # Different from parent class due to different vision model structure weight_mapper = WeightsMapper( orig_to_new_prefix={ # Language model mapping "language_model.layers.": "language_model.model.layers.", "language_model.embed_tokens.": "language_model.model.embed_tokens.", "language_model.norm.": "language_model.model.norm.", # Vision model mapping (SiglipVisionModel has nested vision_model) "vision_model.encoder.": "vision_model.vision_model.encoder.", "vision_model.embeddings.": "vision_model.vision_model.embeddings.", "vision_model.post_layernorm.": "vision_model.vision_model.post_layernorm.", } ) def __init__(self, *, vllm_config: VllmConfig, prefix: str = "") -> None: super().__init__(vllm_config=vllm_config, prefix=prefix) config = vllm_config.model_config.hf_config # Override: get img_context_token_id from config (parent sets None) self.img_context_token_id = getattr(config, "img_context_token_id", None) # Initialize pooler for embedding output pooler_config = vllm_config.model_config.pooler_config assert pooler_config is not None self.pooler = DispatchPooler.for_embedding(pooler_config) def _init_vision_model( self, config: PretrainedConfig, quant_config, *, prefix: str, ) -> nn.Module: """Override to use SigLIP instead of C-RADIO.""" return SiglipVisionModel( config.vision_config, quant_config=quant_config, prefix=prefix, use_head=False, ) def _init_mlp1(self, config: PretrainedConfig) -> nn.Module: """Override to use different MLP structure for embedding model.""" return super()._init_mlp1( config, vit_hidden_size=config.vision_config.hidden_size, vision_projection_hidden_size=config.get_text_config().hidden_size, ) def _call_vision_model(self, pixel_values: torch.Tensor) -> torch.Tensor: """Override to handle SigLIP interface.""" return self.vision_model(pixel_values) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: """Override to use different weight mapping for SigLIP.""" loader = AutoWeightsLoader(self) return loader.load_weights(weights, mapper=self.weight_mapper)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/nemotron_vl.py", "license": "Apache License 2.0", "lines": 749, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/entrypoints/openai/tool_parsers/test_hunyuan_a13b_tool_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # ruff: noqa: E501 import json from unittest.mock import MagicMock import pytest from tests.entrypoints.openai.tool_parsers.utils import ( run_tool_extraction, run_tool_extraction_streaming, ) from vllm.entrypoints.openai.engine.protocol import FunctionCall, ToolCall from vllm.tool_parsers import ToolParser, ToolParserManager def make_tool_call(name, arguments): return ToolCall( type="function", function=FunctionCall(name=name, arguments=json.dumps(arguments)), ) # TODO: add reason prefix and suffix. @pytest.mark.parametrize( "model_output,expected_tool_calls,expected_content", [ # No tool call ("How can I help you today?", [], "How can I help you today?"), # Single tool call, no content ( '<tool_calls>[{"name": "get_weather", "arguments": {"city": "San Francisco", "metric": "celsius"}}]</tool_calls>', # noqa: E501 [ make_tool_call( "get_weather", {"city": "San Francisco", "metric": "celsius"} ) ], None, ), # Multiple tool calls ( '<tool_calls>[{"name": "get_weather", "arguments": {"city": "San Francisco", "metric": "celsius"}}, {"name": "register_user", "arguments": {"name": "John Doe", "age": 37, "address": {"city": "San Francisco", "state": "CA"}, "role": null, "passed_test": true, "aliases": ["John", "Johnny"]}}]</tool_calls>', # noqa: E501 [ make_tool_call( "get_weather", {"city": "San Francisco", "metric": "celsius"} ), make_tool_call( "register_user", { "name": "John Doe", "age": 37, "address": {"city": "San Francisco", "state": "CA"}, "role": None, "passed_test": True, "aliases": ["John", "Johnny"], }, ), ], None, ), # Content before tool call ( 'I will call the tool now. <tool_calls>[{"name": "get_weather", "arguments": {"city": "Boston"}}]</tool_calls>', # noqa: E501 [make_tool_call("get_weather", {"city": "Boston"})], "I will call the tool now. ", ), # Content after tool call (should be stripped) ( '<tool_calls>[{"name": "get_weather", "arguments": {"city": "Seattle"}}]</tool_calls>\nThank you!', # noqa: E501 [make_tool_call("get_weather", {"city": "Seattle"})], None, ), ( '<tool_calls>[{"name": "complex_tool", "arguments": {"level1": {"level2": {"level3": {"value": 123}}}}}]</tool_calls>', [ make_tool_call( "complex_tool", {"level1": {"level2": {"level3": {"value": 123}}}} ) ], None, ), ], ) def test_hunyuan_a13b_tool_parser_extract( model_output, expected_tool_calls, expected_content ): mock_tokenizer = MagicMock() tool_parser: ToolParser = ToolParserManager.get_tool_parser("hunyuan_a13b")( mock_tokenizer ) content, tool_calls = run_tool_extraction( tool_parser, model_output, streaming=False ) # align the random id. for idx in range(len(tool_calls)): tool_calls[idx].id = expected_tool_calls[idx].id assert tool_calls == expected_tool_calls assert content == expected_content # Streaming test: simulate incremental output @pytest.mark.parametrize( "model_deltas,expected_tool_calls", [ ( [ '<tool_calls>[{"name": "get_weather", ', '"arguments": {"city": "San Francisco", ', '"metric": "celsius"}}]', "</tool_calls>", ], [ make_tool_call( "get_weather", {"city": "San Francisco", "metric": "celsius"} ) ], ), ( [ '<tool_calls>[{"name":', ' "get_weather",', ' "arguments":', ' {"city": "Boston"}', "}]", "</tool_calls>", ], [make_tool_call("get_weather", {"city": "Boston"})], ), ( [ "", '<tool_calls>[{"name":', ' "get_weather",', ' "arguments":', ' {"city": "Boston"}', "}]", "</tool_calls>", "\n</answer>", ], [make_tool_call("get_weather", {"city": "Boston"})], ), pytest.param( [ '<tool_calls>[{"name": "complex_tool",', ' "arguments": ', ' {"level1": {"level2": ', '{"level3": {"value": 123}}}}}', "]</tool_calls>", ], [ make_tool_call( "complex_tool", {"level1": {"level2": {"level3": {"value": 123}}}} ) ], marks=pytest.mark.xfail( reason="stream parsing not support nested json yet." ), ), ], ) def test_hunyuan_a13b_tool_parser_streaming(model_deltas, expected_tool_calls): mock_tokenizer = MagicMock() tool_parser: ToolParser = ToolParserManager.get_tool_parser("hunyuan_a13b")( mock_tokenizer ) reconstructor = run_tool_extraction_streaming( tool_parser, model_deltas, assert_one_tool_per_delta=False ) # align the random id. for idx in range(len(reconstructor.tool_calls)): reconstructor.tool_calls[idx].id = expected_tool_calls[idx].id assert reconstructor.tool_calls == expected_tool_calls
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/tool_parsers/test_hunyuan_a13b_tool_parser.py", "license": "Apache License 2.0", "lines": 164, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/layers/fused_moe/deep_gemm_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Taken from https://github.com/ModelTC/LightLLM/blob/8ed97c74c18f11505b048b1ba00ba5c0cef8bff6/lightllm/common/fused_moe/deepep_scatter_gather.py and updated to fit vllm needs and terminology. """ import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm.model_executor.layers.fused_moe.utils import count_expert_num_tokens from vllm.triton_utils import tl, triton from vllm.utils.deep_gemm import get_mk_alignment_for_contiguous_layout from vllm.utils.math_utils import round_up def expert_num_tokens_round_up_and_sum( expert_num_tokens: torch.Tensor, alignment: int ) -> int: # Round up each element in expert_num_tokens to the nearest multiple of # alignment. ent = (expert_num_tokens.to(torch.int64) + (alignment - 1)) // alignment * alignment return torch.sum(ent).item() def compute_aligned_M( M: int, num_topk: int, local_num_experts: int, alignment: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, ): if (expert_tokens_meta is not None) and ( expert_tokens_meta.expert_num_tokens_cpu is not None ): return expert_num_tokens_round_up_and_sum( expert_tokens_meta.expert_num_tokens_cpu, alignment=alignment ) # expert_num_tokens information is not available on the cpu. # compute the max required size. M_sum = (M * num_topk) + local_num_experts * (alignment - 1) M_sum = round_up(M_sum, alignment) return M_sum @triton.jit def apply_expert_map(expert_id, expert_map): if expert_id != -1: expert_id = tl.load(expert_map + expert_id).to(expert_id.dtype) return expert_id @triton.jit def round_up_128(x: int) -> int: y = 128 return ((x + y - 1) // y) * y @triton.jit def _fwd_kernel_ep_scatter_1( num_recv_tokens_per_expert, expert_start_loc, m_indices, num_experts: tl.constexpr, BLOCK_E: tl.constexpr, BLOCK_EXPERT_NUM: tl.constexpr, ): cur_expert = tl.program_id(0) offset_cumsum = tl.arange(0, BLOCK_EXPERT_NUM) tokens_per_expert = tl.load( num_recv_tokens_per_expert + offset_cumsum, mask=offset_cumsum < num_experts, other=0, ) tokens_per_expert = round_up_128(tokens_per_expert) cumsum = tl.cumsum(tokens_per_expert) - tokens_per_expert tl.store(expert_start_loc + offset_cumsum, cumsum, mask=offset_cumsum < num_experts) cur_expert_start = tl.load(expert_start_loc + cur_expert) cur_expert_token_num = tl.load(num_recv_tokens_per_expert + cur_expert) m_indices_start_ptr = m_indices + cur_expert_start off_expert = tl.arange(0, BLOCK_E) # any rows in the per-expert aligned region that do not correspond to # real tokens are left untouched here and should remain initialized to # -1 so DeepGEMM can skip them for start_m in tl.range(0, cur_expert_token_num, BLOCK_E, num_stages=4): offs = start_m + off_expert mask = offs < cur_expert_token_num tl.store( m_indices_start_ptr + offs, cur_expert, mask=mask, ) @triton.jit def _fwd_kernel_ep_scatter_2( total_token_num, expert_start_loc, recv_x, recv_x_stride0, recv_x_stride1, recv_x_scale, recv_x_scale_stride0, recv_x_scale_stride1, recv_topk, recv_topk_stride0, recv_topk_stride1, output_tensor, output_tensor_stride0, output_tensor_stride1, output_tensor_scale, output_tensor_scale_stride0, output_tensor_scale_stride1, output_index, output_index_stride0, output_index_stride1, topk_num: tl.constexpr, expert_map, HAS_EXPERT_MAP: tl.constexpr, HIDDEN_SIZE: tl.constexpr, HIDDEN_SIZE_PAD: tl.constexpr, SCALE_HIDDEN_SIZE: tl.constexpr, SCALE_HIDDEN_SIZE_PAD: tl.constexpr, ): start_token_id = tl.program_id(0) grid_num = tl.num_programs(0) offset_in = tl.arange(0, HIDDEN_SIZE_PAD) mask = offset_in < HIDDEN_SIZE offset_in_s = tl.arange(0, SCALE_HIDDEN_SIZE_PAD) mask_s = offset_in_s < SCALE_HIDDEN_SIZE for token_id in range(start_token_id, total_token_num, grid_num): to_copy = tl.load(recv_x + token_id * recv_x_stride0 + offset_in, mask=mask) to_copy_s = tl.load( recv_x_scale + token_id * recv_x_scale_stride0 + offset_in_s, mask=mask_s ) for topk_index in tl.range(0, topk_num, 1, num_stages=4): expert_id = tl.load(recv_topk + token_id * recv_topk_stride0 + topk_index) if HAS_EXPERT_MAP: expert_id = apply_expert_map(expert_id, expert_map) if expert_id >= 0: dest_token_index = tl.atomic_add(expert_start_loc + expert_id, 1) tl.store( output_index + token_id * output_index_stride0 + topk_index, dest_token_index, ) output_tensor_ptr = ( output_tensor + dest_token_index * output_tensor_stride0 ) output_tensor_scale_ptr = ( output_tensor_scale + dest_token_index * output_tensor_scale_stride0 ) tl.store(output_tensor_ptr + offset_in, to_copy, mask=mask) tl.store(output_tensor_scale_ptr + offset_in_s, to_copy_s, mask=mask_s) @torch.no_grad() def ep_scatter( recv_x: torch.Tensor, recv_x_scale: torch.Tensor, recv_topk: torch.Tensor, num_recv_tokens_per_expert: torch.Tensor, expert_map: torch.Tensor | None, expert_start_loc: torch.Tensor, output_tensor: torch.Tensor, output_tensor_scale: torch.Tensor, m_indices: torch.Tensor, output_index: torch.Tensor, ): BLOCK_E = 128 # token num of per expert is aligned to 128 BLOCK_D = 128 # block size of quantization num_warps = 8 num_experts = num_recv_tokens_per_expert.shape[0] hidden_size = recv_x.shape[1] # grid = (triton.cdiv(hidden_size, BLOCK_D), num_experts) grid = num_experts assert m_indices.shape[0] % BLOCK_E == 0 _fwd_kernel_ep_scatter_1[(grid,)]( num_recv_tokens_per_expert, expert_start_loc, m_indices, num_experts=num_experts, num_warps=num_warps, BLOCK_E=BLOCK_E, BLOCK_EXPERT_NUM=triton.next_power_of_2(num_experts), ) grid = min(recv_topk.shape[0], 1024 * 8) _fwd_kernel_ep_scatter_2[(grid,)]( recv_topk.shape[0], expert_start_loc, recv_x, recv_x.stride(0), recv_x.stride(1), recv_x_scale, recv_x_scale.stride(0), recv_x_scale.stride(1), recv_topk, recv_topk.stride(0), recv_topk.stride(1), output_tensor, output_tensor.stride(0), output_tensor.stride(1), output_tensor_scale, output_tensor_scale.stride(0), output_tensor_scale.stride(1), output_index, output_index.stride(0), output_index.stride(1), topk_num=recv_topk.shape[1], expert_map=expert_map, HAS_EXPERT_MAP=expert_map is not None, num_warps=num_warps, HIDDEN_SIZE=hidden_size, HIDDEN_SIZE_PAD=triton.next_power_of_2(hidden_size), SCALE_HIDDEN_SIZE=hidden_size // BLOCK_D, SCALE_HIDDEN_SIZE_PAD=triton.next_power_of_2(hidden_size // BLOCK_D), ) return @triton.jit def _fwd_kernel_ep_gather( total_token_num, input_tensor, input_tensor_stride0, input_tensor_stride1, recv_topk_ids, recv_topk_ids_stride0, recv_topk_ids_stride1, recv_topk_weight, recv_topk_weight_stride0, recv_topk_weight_stride1, input_index, input_index_stride0, input_index_stride1, output_tensor, output_tensor_stride0, output_tensor_stride1, topk_num: tl.constexpr, expert_map, HAS_EXPERT_MAP: tl.constexpr, BLOCK_D: tl.constexpr, ): cur_block = tl.program_id(0) start_cur_token = tl.program_id(1) grid_num = tl.num_programs(1) for cur_token in range(start_cur_token, total_token_num, grid_num): off_d = tl.arange(0, BLOCK_D) accumulator = tl.zeros([BLOCK_D], dtype=tl.float32) for topk_index in range(0, topk_num): expert_id = tl.load( recv_topk_ids + cur_token * recv_topk_ids_stride0 + topk_index ) if HAS_EXPERT_MAP: expert_id = apply_expert_map(expert_id, expert_map) if expert_id >= 0: source_token_index = tl.load( input_index + cur_token * input_index_stride0 + topk_index ) acc_weight = tl.load( recv_topk_weight + cur_token * recv_topk_weight_stride0 + topk_index ) tmp = tl.load( input_tensor + source_token_index * input_tensor_stride0 + cur_block * BLOCK_D + off_d ) accumulator += tmp.to(tl.float32) * acc_weight tl.store( output_tensor + cur_token * output_tensor_stride0 + cur_block * BLOCK_D + off_d, accumulator.to(output_tensor.dtype.element_ty), ) @torch.no_grad() def ep_gather( input_tensor: torch.Tensor, recv_topk_ids: torch.Tensor, recv_topk_weight: torch.Tensor, input_index: torch.Tensor, expert_map: torch.Tensor | None, output_tensor: torch.Tensor, ): num_warps = 2 num_tokens = output_tensor.shape[0] hidden_size = input_tensor.shape[1] BLOCK_D = min(hidden_size, 1024) assert hidden_size % BLOCK_D == 0 grid = (triton.cdiv(hidden_size, BLOCK_D), min(num_tokens, 1024)) _fwd_kernel_ep_gather[grid]( num_tokens, input_tensor, input_tensor.stride(0), input_tensor.stride(1), recv_topk_ids, recv_topk_ids.stride(0), recv_topk_ids.stride(1), recv_topk_weight, recv_topk_weight.stride(0), recv_topk_weight.stride(1), input_index, input_index.stride(0), input_index.stride(1), output_tensor, output_tensor.stride(0), output_tensor.stride(1), topk_num=recv_topk_ids.shape[1], expert_map=expert_map, HAS_EXPERT_MAP=expert_map is not None, num_warps=num_warps, BLOCK_D=BLOCK_D, ) return def deepgemm_moe_permute( aq: torch.Tensor, aq_scale: torch.Tensor, topk_ids: torch.Tensor, local_num_experts: int, expert_map: torch.Tensor | None, expert_tokens_meta: mk.ExpertTokensMetadata | None, aq_out: torch.Tensor | None = None, ): assert aq.ndim == 2 assert topk_ids.dtype.is_signed, "The kernel uses -1 to represent invalid topk_ids" H = aq.size(1) device = aq.device block_m, block_k = get_mk_alignment_for_contiguous_layout() M_sum = compute_aligned_M( M=topk_ids.size(0), num_topk=topk_ids.size(1), local_num_experts=local_num_experts, alignment=block_m, expert_tokens_meta=expert_tokens_meta, ) expert_start_loc = torch.empty( (local_num_experts), device=device, dtype=torch.int32 ) assert aq_out is None or aq_out.shape == (M_sum, H) if aq_out is None: aq_out = torch.empty((M_sum, H), device=device, dtype=aq.dtype) aq_scale_out = torch.empty( (M_sum, H // block_k), device=device, dtype=torch.float32 ) # DeepGEMM uses negative values in m_indices (here expert_ids) to mark # completely invalid / padded blocks that should be skipped. We always # initialize expert_ids to -1 so any row that is not explicitly written # by the scatter kernel will be treated as invalid and skipped by # DeepGEMM's scheduler. expert_ids = torch.full( (M_sum,), fill_value=-1, device=device, dtype=torch.int32, ) inv_perm = torch.empty(topk_ids.shape, device=device, dtype=torch.int32) expert_num_tokens = None if expert_tokens_meta is not None: expert_num_tokens = expert_tokens_meta.expert_num_tokens else: expert_num_tokens = count_expert_num_tokens( topk_ids, local_num_experts, expert_map ) ep_scatter( recv_x=aq, recv_x_scale=aq_scale, recv_topk=topk_ids, num_recv_tokens_per_expert=expert_num_tokens, expert_start_loc=expert_start_loc, expert_map=expert_map, output_tensor=aq_out, output_tensor_scale=aq_scale_out, m_indices=expert_ids, output_index=inv_perm, ) return aq_out, aq_scale_out, expert_ids, inv_perm def deepgemm_unpermute_and_reduce( a: torch.Tensor, # Grouped gemm output topk_ids: torch.Tensor, topk_weights: torch.Tensor, inv_perm: torch.Tensor, expert_map: torch.Tensor | None, output: torch.Tensor, ): return ep_gather( input_tensor=a, recv_topk_ids=topk_ids, recv_topk_weight=topk_weights, input_index=inv_perm, expert_map=expert_map, output_tensor=output, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fused_moe/deep_gemm_utils.py", "license": "Apache License 2.0", "lines": 374, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license