repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/shared_fused_moe.py
vllm/model_executor/layers/fused_moe/shared_fused_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.distributed import ( get_tensor_model_parallel_world_size, tensor_model_parallel_all_reduce, ) from vllm.model_executor.layers.fused_moe.layer import FusedMoE # TODO(bnell): Add shared + fused combo function? e.g. + class SharedFusedMoE(FusedMoE): """ A FusedMoE operation that also computes the results of shared experts. If an all2all communicator is being used the shared expert computation can be interleaved with the fused all2all dispatch communication step. """ def __init__( self, shared_experts: torch.nn.Module | None, gate: torch.nn.Module | None = None, use_overlapped: bool = True, **kwargs, ): super().__init__(**kwargs) self._shared_experts = shared_experts # Disable shared expert overlap if: # - we are using eplb with non-default backend, because of correctness issues # - we are using flashinfer with DP, since there nothint to gain # - we are using marlin kernels backend = self.moe_parallel_config.all2all_backend self.use_overlapped = ( use_overlapped and not ( (self.enable_eplb and backend != "allgather_reducescatter") or (self.moe_config.use_flashinfer_cutlass_kernels and self.dp_size > 1) ) and self._shared_experts is not None ) self._gate = gate @property def shared_experts(self) -> torch.nn.Module | None: return self._shared_experts if self.use_overlapped else None @property def gate(self) -> torch.nn.Module | None: return self._gate if self.use_overlapped else None @property def is_internal_router(self) -> bool: return self.gate is not None def forward( self, hidden_states: torch.Tensor, router_logits: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: if not self.use_overlapped: if self._shared_experts is not None: shared_out = self._shared_experts(hidden_states) # Reduce shared expert outputs if necessary, since the MLP # should have been created with reduce_results=False. if ( self.reduce_results and get_tensor_model_parallel_world_size() > 1 and self.must_reduce_shared_expert_outputs() ): shared_out = tensor_model_parallel_all_reduce(shared_out) else: shared_out = None fused_out = super().forward( hidden_states=hidden_states, router_logits=router_logits, ) else: shared_out, fused_out = super().forward( hidden_states=hidden_states, router_logits=router_logits, ) # ensure early TP reduction of shared expert outputs when required if ( shared_out is not None and self.reduce_results and get_tensor_model_parallel_world_size() > 1 and self.must_reduce_shared_expert_outputs() ): shared_out = tensor_model_parallel_all_reduce(shared_out) return shared_out, fused_out
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/cutlass_moe.py
vllm/model_executor/layers/fused_moe/cutlass_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """CUTLASS based Fused MoE kernels.""" from collections.abc import Callable import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm import _custom_ops as ops from vllm.logger import init_logger from vllm.model_executor.layers.fused_moe.config import FusedMoEQuantConfig from vllm.model_executor.layers.fused_moe.moe_permute_unpermute import ( moe_permute, moe_unpermute, ) from vllm.model_executor.layers.fused_moe.prepare_finalize import ( MoEPrepareAndFinalizeNoEP, ) from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceDelegate, TopKWeightAndReduceNoOP, ) from vllm.model_executor.layers.fused_moe.utils import _resize_cache from vllm.scalar_type import scalar_types logger = init_logger(__name__) def run_cutlass_moe_fp8( output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_ids: torch.Tensor, activation_callable: Callable, global_num_experts: int, expert_map: torch.Tensor | None, w1_scale: torch.Tensor | None, w2_scale: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, ab_strides1: torch.Tensor, ab_strides2: torch.Tensor, c_strides1: torch.Tensor, c_strides2: torch.Tensor, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_num_tokens: torch.Tensor | None, out_dtype: torch.dtype, per_act_token: bool, per_out_ch: bool, use_batched_format: bool, topk_weights: torch.Tensor | None, ): a1q = hidden_states assert w1_scale is not None assert w2_scale is not None assert w1.dtype == torch.float8_e4m3fn assert w2.dtype == torch.float8_e4m3fn assert a1q.size(-1) == w1.size(2), "Hidden size mismatch w1" assert w1.size(1) == w2.size(2) * 2, "Hidden size mismatch w2" assert ( w1_scale.dim() == 1 or w1_scale.size(1) == 1 or w1_scale.shape[1] == w1.size(1) ), "W1 scale shape mismatch" assert ( w2_scale.dim() == 1 or w2_scale.size(1) == 1 or w2_scale.shape[1] == w2.size(1) ), "W2 scale shape mismatch" assert w1.size(0) == w2.size(0), "Expert number mismatch" assert ( a1q_scale is None or a1q_scale.dim() == 0 or a1q_scale.size(0) == 1 or a1q_scale.size(0) == a1q.shape[0] ), "Input scale shape mismatch" assert w1.size(0) == w2.size(0), "Weights expert number mismatch" assert w1.size(0) == w1_scale.size(0), "w1 scales expert number mismatch" assert w1.size(0) == w2_scale.size(0), "w2 scales expert number mismatch" assert ( a2_scale is None or a2_scale.dim() == 0 or a2_scale.size(0) == 1 or a2_scale.size(0) == a1q.shape[0] ), "Intermediate scale shape mismatch" assert out_dtype in [torch.half, torch.bfloat16], "Invalid output dtype" if expert_map is not None: assert expert_num_tokens is None # We have two modes: batched experts and non-batched experts. # In the non-batched mode, the input tokens are not padded: thus, the shape # of the input is [total_num_tokens, hidden_size]. The input and output # require shuffling by a_map and c_map such that the tokens assigned to # each expert are contiguous. # In the batched mode, the input tokens are padded per expert to ensure that # the batched dispatch and combine functions work correctly: thus, the shape # of the input is [num_experts, max_num_tokens_per_expert, hidden_size]. # The batched input and output require no shuffling by a_map and c_map since # their tokens are already contiguous for each expert as a result of # the dispatch function. M = a1q.size(0) # non batched expert M padded_M = a1q.size(1) # batched expert M _, K, N = w2.shape device = a1q.device assert w1.size(2) == K assert global_num_experts != -1 assert a1q_scale is not None if expert_map is not None: "Translate info from expert_map to topk_ids" local_topk_ids = torch.where( expert_map[topk_ids] != -1, expert_map[topk_ids], -1 ) else: local_topk_ids = topk_ids topk = local_topk_ids.size(1) local_E = w1.size(0) if use_batched_format: mm1_out = _resize_cache(workspace13, (local_E * padded_M, N * 2)) act_out = _resize_cache(workspace2, (local_E * padded_M, N)) quant_out = _resize_cache( workspace13.view(dtype=torch.float8_e4m3fn), (local_E * padded_M, N) ) mm2_out = _resize_cache(workspace2, (local_E * padded_M, K)) else: a1q_perm = _resize_cache( workspace2.view(dtype=torch.float8_e4m3fn), (M * topk, K) ) mm1_out = _resize_cache(workspace13, (M * topk, N * 2)) act_out = _resize_cache(workspace2, (M * topk, N)) # original workspace are based on input hidden_states dtype (bf16) quant_out = _resize_cache( workspace13.view(dtype=torch.float8_e4m3fn), (M * topk, N) ) mm2_out = _resize_cache(workspace2, (M * topk, K)) if use_batched_format: assert expert_num_tokens is not None expert_offsets = torch.empty((local_E), dtype=torch.int32, device=device) problem_sizes1 = torch.empty((local_E, 3), dtype=torch.int32, device=device) problem_sizes2 = torch.empty((local_E, 3), dtype=torch.int32, device=device) ops.get_cutlass_pplx_moe_mm_data( expert_offsets, problem_sizes1, problem_sizes2, expert_num_tokens, local_E, padded_M, N, K, ) w1_scale = w1_scale.reshape(w1_scale.size(0), -1) w2_scale = w2_scale.reshape(w2_scale.size(0), -1) a1q = a1q.reshape(-1, a1q.size(2)) a1q_scale = a1q_scale.reshape(-1, a1q_scale.size(2)).contiguous() # c3x get_group_gemm_starts expects int64 to avoid overflow # during offset calculations expert_offsets = expert_offsets.to(torch.int64) else: problem_sizes1 = torch.empty( (global_num_experts, 3), dtype=torch.int32, device=device ) problem_sizes2 = torch.empty( (global_num_experts, 3), dtype=torch.int32, device=device ) num_expert = global_num_experts if expert_map is None else expert_map.size(0) # permuted a1q reuses workspace2 a1q, a1q_scale, expert_offsets, inv_perm, _ = moe_permute( a1q, a1q_scale, topk_ids, num_expert, local_E, expert_map, permuted_hidden_states=a1q_perm, ) expert_offsets = expert_offsets[:-1] ops.get_cutlass_moe_mm_problem_sizes( local_topk_ids, problem_sizes1, problem_sizes2, global_num_experts, N, K ) if not per_act_token and (expert_map is not None or use_batched_format): # this is necessary to avoid imprecise scale calculation caused by # random data in the unused workspace. The workspace is unused when # this rank handles only partial tokens, or when it is batched . mm1_out.fill_(0) ops.cutlass_moe_mm( mm1_out, a1q, w1, a1q_scale, w1_scale, expert_offsets, problem_sizes1, ab_strides1, ab_strides1, c_strides1, per_act_token, per_out_ch, ) activation_callable(act_out, mm1_out) a2q, a2q_scale = ops.scaled_fp8_quant( act_out, a2_scale, use_per_token_if_dynamic=per_act_token, output=quant_out ) if expert_map is not None: mm2_out.fill_(0) ops.cutlass_moe_mm( mm2_out, a2q, w2, a2q_scale, w2_scale, expert_offsets, problem_sizes2, ab_strides2, ab_strides2, c_strides2, per_act_token, per_out_ch, ) if use_batched_format: output.copy_(mm2_out.reshape(local_E, padded_M, K), non_blocking=True) else: # for non-chunking mode the output is resized from workspace13 # so we need to make sure mm2_out uses workspace2. moe_unpermute( out=output, permuted_hidden_states=mm2_out, topk_weights=topk_weights, inv_permuted_idx=inv_perm, ) class CutlassExpertsFp8Base(mk.FusedMoEPermuteExpertsUnpermute): def __init__( self, out_dtype: torch.dtype | None, ab_strides1: torch.Tensor, ab_strides2: torch.Tensor, c_strides1: torch.Tensor, c_strides2: torch.Tensor, quant_config: FusedMoEQuantConfig, ): assert quant_config.use_fp8_w8a8 super().__init__(quant_config) self.out_dtype = out_dtype self.ab_strides1 = ab_strides1 self.ab_strides2 = ab_strides2 self.c_strides1 = c_strides1 self.c_strides2 = c_strides2 def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: # Let PrepareAndFinalize::finalize() decide the impl. return TopKWeightAndReduceDelegate() def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: str, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ): assert self.w1_zp is None, "w1_zp is not supported in CUTLASS MoE" assert self.w2_zp is None, "w2_zp is not supported in CUTLASS MoE" expert_num_tokens = None if expert_tokens_meta is not None: expert_num_tokens = expert_tokens_meta.expert_num_tokens activation_callable = lambda o, i: self.activation(activation, o, i) use_batched_format = ( self.activation_formats[0] == mk.FusedMoEActivationFormat.BatchedExperts ) in_dtype = hidden_states.dtype run_cutlass_moe_fp8( output, hidden_states, w1, w2, topk_ids, activation_callable, global_num_experts, expert_map, self.w1_scale, self.w2_scale, a1q_scale, a2_scale, self.ab_strides1, self.ab_strides2, self.c_strides1, self.c_strides2, workspace13, workspace2, expert_num_tokens, self.out_dtype if self.out_dtype is not None else in_dtype, self.per_act_token_quant, self.per_out_ch_quant, use_batched_format, topk_weights, ) class CutlassExpertsFp8(CutlassExpertsFp8Base): def __init__( self, out_dtype: torch.dtype | None, ab_strides1: torch.Tensor, ab_strides2: torch.Tensor, c_strides1: torch.Tensor, c_strides2: torch.Tensor, quant_config: FusedMoEQuantConfig, ): super().__init__( out_dtype, ab_strides1, ab_strides2, c_strides1, c_strides2, quant_config, ) @property def activation_formats( self, ) -> tuple[mk.FusedMoEActivationFormat, mk.FusedMoEActivationFormat]: return ( mk.FusedMoEActivationFormat.Standard, mk.FusedMoEActivationFormat.Standard, ) def supports_chunking(self) -> bool: return True def supports_expert_map(self) -> bool: return True def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: # topk weights and reduction are fused in moe_unpermute cuda kernel return TopKWeightAndReduceNoOP() def workspace_dtype(self, act_dtype: torch.dtype) -> torch.dtype: return self.out_dtype if self.out_dtype is not None else act_dtype def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: workspace1 = (M * topk, max(N, K)) workspace2 = (M * topk, max(N // 2, K)) output = (M, K) return (workspace1, workspace2, output) class CutlassBatchedExpertsFp8(CutlassExpertsFp8Base): def __init__( self, max_experts_per_worker: int, num_dispatchers: int, out_dtype: torch.dtype | None, ab_strides1: torch.Tensor, ab_strides2: torch.Tensor, c_strides1: torch.Tensor, c_strides2: torch.Tensor, quant_config: FusedMoEQuantConfig, ): super().__init__( out_dtype, ab_strides1, ab_strides2, c_strides1, c_strides2, quant_config, ) assert max_experts_per_worker > 0 self.max_experts_per_worker = max_experts_per_worker self.num_dispatchers = num_dispatchers @property def activation_formats( self, ) -> tuple[mk.FusedMoEActivationFormat, mk.FusedMoEActivationFormat]: return ( mk.FusedMoEActivationFormat.BatchedExperts, mk.FusedMoEActivationFormat.BatchedExperts, ) def supports_chunking(self) -> bool: return False def supports_expert_map(self) -> bool: return False def workspace_dtype(self, act_dtype: torch.dtype) -> torch.dtype: return self.out_dtype if self.out_dtype is not None else act_dtype def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: num_dp = self.num_dispatchers assert num_dp is not None workspace1 = (self.max_experts_per_worker, M * num_dp, max(N, K)) workspace2 = (self.max_experts_per_worker, M * num_dp, max(N // 2, K)) output = (self.max_experts_per_worker, M, K) return (workspace1, workspace2, output) def cutlass_moe_fp8( a: torch.Tensor, w1_q: torch.Tensor, w2_q: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, ab_strides1: torch.Tensor, ab_strides2: torch.Tensor, c_strides1: torch.Tensor, c_strides2: torch.Tensor, quant_config: FusedMoEQuantConfig, activation: str = "silu", expert_map: torch.Tensor | None = None, apply_router_weight_on_input: bool = False, global_num_experts: int = -1, ) -> torch.Tensor: """ This function computes a a8w8-quantized Mixture of Experts (MoE) layer using two sets of quantized weights, w1_q and w2_q, and top-k gating mechanism. The matrix multiplications are implemented with CUTLASS grouped gemm. Parameters: - a (torch.Tensor): The input tensor to the MoE layer. Shape: [M, K] - w1_q (torch.Tensor): The first set of fp8-quantized expert weights. Shape: [num_experts, K, 2N] (the weights are passed transposed) - w2_q (torch.Tensor): The second set of fp8-quantized expert weights. Shape: [num_experts, N, K] (the weights are passed transposed) - topk_weights (torch.Tensor): The weights of each token->expert mapping. - topk_ids (torch.Tensor): The token->expert mappings. - w1_scale (torch.Tensor): The fp32 scale to dequantize w1_q. Shape: [num_experts] or [num_experts, 2N] - w2_scale (torch.Tensor): The fp32 scale to dequantize w2_q. Shape: [num_experts] or [num_experts, K] - ab_strides1 (torch.Tensor): The input/weight strides for the first gemm. Shape: [num_experts] - ab_strides2 (torch.Tensor): The input/weight strides for the second gemm. Shape: [num_experts] - c_strides1 (torch.Tensor): The output strides for the first gemm. Shape: [num_experts] - c_strides2 (torch.Tensor): The output strides for the second gemm. Shape: [num_experts] - per_act_token (Optional[bool]): Whether the scale is per-token or per-tensor. - activation (str): The activation function to use. - a1_scale (Optional[torch.Tensor]): The optional fp32 scale to quantize a. Shape: scalar or [M] - a2_scale (Optional[torch.Tensor]): The optional fp32 scale to quantize the intermediate result between the gemms. Shape: scalar or [M] - expert_map (Optional[torch.Tensor]): In the case of Expert parallel, every Rank is responsible for a subset of experts. expert_map is a mapping from global expert-id to local expert-id. When expert_map[i] is -1, it means that this Rank is not responsible for global expert-id i. - apply_router_weight_on_input (bool): When true, the topk weights are applied directly on the inputs. This is only applicable when topk is 1. - global_num_experts (int): The total number of experts. Returns: - torch.Tensor: The fp16 output tensor after applying the MoE layer. """ assert quant_config is not None if quant_config.a1_scale is not None: assert quant_config.per_act_token_quant == (quant_config.a1_scale.numel() != 1) if quant_config.a2_scale is not None: assert quant_config.per_act_token_quant == (quant_config.a2_scale.numel() != 1) if quant_config.w1_scale is not None: if quant_config.per_out_ch_quant: assert quant_config.w1_scale.dim() > 1 and quant_config.w1_scale.size( 1 ) == w1_q.size(1) else: assert ( quant_config.w1_scale.dim() == 1 or quant_config.w1_scale.size(1) == 1 ) num_experts = global_num_experts if global_num_experts != -1 else w1_q.size(0) fn = mk.FusedMoEModularKernel( MoEPrepareAndFinalizeNoEP(), CutlassExpertsFp8( out_dtype=a.dtype, ab_strides1=ab_strides1, ab_strides2=ab_strides2, c_strides1=c_strides1, c_strides2=c_strides2, quant_config=quant_config, ), ) return fn( a, w1_q, w2_q, topk_weights, topk_ids, activation=activation, global_num_experts=num_experts, expert_map=expert_map, apply_router_weight_on_input=apply_router_weight_on_input, ) FLOAT4_E2M1_MAX = scalar_types.float4_e2m1f.max() FLOAT8_E4M3_MAX = torch.finfo(torch.float8_e4m3fn).max def run_cutlass_moe_fp4( output: torch.Tensor, a: torch.Tensor, a1_gscale: torch.Tensor, w1_fp4: torch.Tensor, w1_blockscale: torch.Tensor, w1_alphas: torch.Tensor, a2_gscale: torch.Tensor, w2_fp4: torch.Tensor, w2_blockscale: torch.Tensor, w2_alphas: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, workspace13: torch.Tensor, workspace2: torch.Tensor, m: int, n: int, k: int, e: int, device: torch.device, apply_router_weight_on_input: bool = False, ) -> None: """ MoE implementation for FP4 Inputs # Gemm 1 a: Input tensor: [m, k] (half/bfloat16) a1_gscale: Activation scale per expert: [e] (float32) w1(gate up) (not an argument to cutlass_moe_fp4): [e, 2 * n, k] w1_fp4: [e, 2 * n, k // 2], dtype: torch.uint8 (stacked fp4: E2M1) (Note: `n` is the up projection output dim, `k` is the input dim in full precision) w1_blockscale: [e, 2 * n, k // block_size] (float8_e4m3) (Block size = 16 for NVFP4) # Gemm 2 a2_gscale: Activation scale per expert: [e] w2(down projection) (not an argument to cutlass_moe_fp4): [e, k, n] w2_fp4: [e, k, n // 2], dtype: torch.uint8 (stacked E2M1) w2_blockscale: [e, k, n // block_size], dtype: float8_e4m3 topk_weights: [m, topk] dtype: float8 topk_ids: [m, topk] dtype: float8 m, n, k: Unquantized weight shapes, dtype: int e: number of experts, dtype: int assumes that topk < k < n to satisfy - up/down projection expectations. """ assert topk_weights.shape == topk_ids.shape, "topk shape mismatch" assert w1_fp4.dtype == torch.uint8, "weight 1 must be uint8" assert w2_fp4.dtype == torch.uint8, "weight 2 must be uint8" assert ( w1_fp4.ndim == 3 and w2_fp4.ndim == 3 and w1_blockscale.ndim == 3 and w2_blockscale.ndim == 3 ), "All Weights must be of rank 3 for cutlass_moe_fp4" m_a, k_a = a.shape e_w1, nx2_w1, half_k_w1 = w1_fp4.shape e_w2, k_w2, half_n_w2 = w2_fp4.shape assert e_w1 == e_w2 and e_w1 == e, ( "Number of experts must match", f" between weights. {e_w1}, {e_w2}, {e}", ) assert k_a == half_k_w1 * 2 and k == k_w2, ( "Hidden size mismatch between a, w1 and w2" ) assert nx2_w1 == n * 2 and half_n_w2 * 2 == n, "mismatch in expected `n`" assert m == m_a, "input shape mismatch" assert 2 * half_k_w1 == k_w2, "Hidden size mismatch w2 and w1" assert a.dtype in [torch.half, torch.bfloat16], "Invalid input dtype" assert topk_weights.size(0) == m and topk_ids.size(0) == m, ( "topk must be provided for each row of a" ) topk = topk_ids.size(1) out_dtype = a.dtype num_topk = topk_ids.size(1) expert_offsets = torch.empty((e + 1), dtype=torch.int32, device=device) blockscale_offsets = torch.empty((e + 1), dtype=torch.int32, device=device) # Problem size: (num_experts, (m,2n,k)) problem_sizes1 = torch.empty((e, 3), dtype=torch.int32, device=device) # Problem size: (num_experts, (m,n,k)) problem_sizes2 = torch.empty((e, 3), dtype=torch.int32, device=device) a_map = torch.empty((topk_ids.numel()), dtype=torch.int32, device=device) c_map = torch.empty((topk_ids.numel()), dtype=torch.int32, device=device) if apply_router_weight_on_input: # TODO: this only works for topK=1, will need to update for topK>1 assert num_topk == 1, ( "apply_router_weight_on_input is only implemented for topk=1" ) a.mul_(topk_weights.to(out_dtype)) # problem shapes should have [m, n, k] # Note that problem sizes are based on logical number of elements. ops.get_cutlass_moe_mm_data( topk_ids, expert_offsets, problem_sizes1, problem_sizes2, a_map, c_map, e, n, k, blockscale_offsets, ) a = ops.shuffle_rows(a, a_map) rep_a_fp4, rep_a_blockscale = ops.scaled_fp4_experts_quant( a, a1_gscale, expert_offsets, blockscale_offsets, num_topk, ) c1 = _resize_cache(workspace13, (m * topk, n * 2)) c2 = _resize_cache(workspace2, (m * topk, n)) c3 = _resize_cache(workspace13, (m * topk, k)) ops.cutlass_fp4_moe_mm( c1, rep_a_fp4, w1_fp4, rep_a_blockscale, w1_blockscale, w1_alphas, problem_sizes1, expert_offsets[:-1], blockscale_offsets[:-1], ) del rep_a_fp4, rep_a_blockscale torch.ops._C.silu_and_mul(c2, c1) int_fp4, int_blockscale = ops.scaled_fp4_experts_quant( c2, a2_gscale, expert_offsets, blockscale_offsets, num_topk ) ops.cutlass_fp4_moe_mm( c3, int_fp4, w2_fp4, int_blockscale, w2_blockscale, w2_alphas, problem_sizes2, expert_offsets[:-1], blockscale_offsets[:-1], ) del int_fp4, int_blockscale c3 = ops.shuffle_rows(c3, c_map) assert output.dtype == out_dtype if not apply_router_weight_on_input: output.copy_( ( c3.view(m, num_topk, k) * topk_weights.view(m, num_topk, 1).to(out_dtype) ).sum(dim=1), non_blocking=True, ) else: output.copy_(c3.view(m, num_topk, k).sum(dim=1), non_blocking=True) return # Split into batched and non-batched class CutlassExpertsFp4(mk.FusedMoEPermuteExpertsUnpermute): def __init__( self, max_experts_per_worker: int, out_dtype: torch.dtype, quant_config: FusedMoEQuantConfig, use_batched_format: bool = False, ): super().__init__(quant_config) self.max_experts_per_worker = max_experts_per_worker self.out_dtype = out_dtype self.use_batched_format = use_batched_format @property def activation_formats( self, ) -> tuple[mk.FusedMoEActivationFormat, mk.FusedMoEActivationFormat]: if self.use_batched_format: return ( mk.FusedMoEActivationFormat.BatchedExperts, mk.FusedMoEActivationFormat.BatchedExperts, ) else: return ( mk.FusedMoEActivationFormat.Standard, mk.FusedMoEActivationFormat.Standard, ) def supports_expert_map(self) -> bool: return False def supports_chunking(self) -> bool: return True def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: return TopKWeightAndReduceNoOP() def workspace_dtype(self, act_dtype: torch.dtype) -> torch.dtype: return self.out_dtype if self.out_dtype is not None else act_dtype def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: workspace1: tuple[int, ...] = () workspace2: tuple[int, ...] = () output: tuple[int, ...] = () if self.use_batched_format: workspace1 = (self.max_experts_per_worker, M, max(N, K)) workspace2 = (self.max_experts_per_worker, M, (N // 2)) output = (self.max_experts_per_worker, M, K) else: workspace1 = (M * topk, max(2 * N, K)) workspace2 = (M * topk, N) output = (M, K) return (workspace1, workspace2, output) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: str, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, # unused a2_scale: torch.Tensor | None, # unused workspace13: torch.Tensor | None, workspace2: torch.Tensor | None, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ): e, m, n, k, _ = self.moe_problem_size(hidden_states, w1, w2, topk_ids) n = w2.shape[2] * 2 run_cutlass_moe_fp4( output=output, a=hidden_states, a1_gscale=self.a1_gscale, w1_fp4=w1, w1_blockscale=self.w1_scale, w1_alphas=self.g1_alphas, a2_gscale=self.a2_gscale, w2_fp4=w2, w2_blockscale=self.w2_scale, w2_alphas=self.g2_alphas, topk_weights=topk_weights, topk_ids=topk_ids, workspace13=workspace13, workspace2=workspace2, m=m, n=n, k=k, e=e, device=hidden_states.device, apply_router_weight_on_input=apply_router_weight_on_input, ) def cutlass_moe_fp4( a: torch.Tensor, w1_fp4: torch.Tensor, w2_fp4: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, quant_config: FusedMoEQuantConfig, m: int, n: int, k: int, e: int, expert_map: torch.Tensor | None = None, apply_router_weight_on_input: bool = False, ) -> torch.Tensor: assert expert_map is None, ( "Expert Parallelism / expert_map " "is currently not supported for " "ModelOptNvFp4FusedMoE's cutlass_moe_fp4." ) # TODO(bnell): this feels a bit hacky # NVFP4 requires two levels of quantization, which involves # computing some scaling factors dynamically. This makes it # incompatible with the typical prepare -> MoE -> finalize # pipeline. Move the quantization logic into the MoE body. quant_config = FusedMoEQuantConfig.make( quant_dtype=None, # skip quantization in prepare/finalize per_act_token_quant=quant_config.per_act_token_quant, per_out_ch_quant=quant_config.per_out_ch_quant, block_shape=quant_config.block_shape, g1_alphas=quant_config.g1_alphas, g2_alphas=quant_config.g2_alphas, a1_gscale=quant_config.a1_gscale, a2_gscale=quant_config.a2_gscale, w1_scale=quant_config.w1_scale, w2_scale=quant_config.w2_scale, ) fn = mk.FusedMoEModularKernel( MoEPrepareAndFinalizeNoEP(), CutlassExpertsFp4( max_experts_per_worker=e, out_dtype=a.dtype, quant_config=quant_config, use_batched_format=False, ), ) return fn( hidden_states=a, w1=w1_fp4, w2=w2_fp4, topk_weights=topk_weights, topk_ids=topk_ids, inplace=False, activation="silu", global_num_experts=e, expert_map=None, apply_router_weight_on_input=apply_router_weight_on_input, ) # W4A8 def run_cutlass_moe_w4a8_fp8( output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_ids: torch.Tensor, activation_callable: Callable, global_num_experts: int, expert_map: torch.Tensor | None, w1_scale: torch.Tensor | None, w2_scale: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, w1_chan_scale: torch.Tensor, w2_chan_scale: torch.Tensor, a_strides1: torch.Tensor, a_strides2: torch.Tensor, b_strides1: torch.Tensor, b_strides2: torch.Tensor, c_strides1: torch.Tensor, c_strides2: torch.Tensor, s_strides1: torch.Tensor, s_strides2: torch.Tensor, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_num_tokens: torch.Tensor | None, out_dtype: torch.dtype, per_act_token: bool, per_out_ch: bool, use_batched_format: bool, topk_weights: torch.Tensor | None, group_size: int, ): a1q = hidden_states M = a1q.size(0) local_E = w1.size(0) device = a1q.device _, K, N_packed = w2.shape N = N_packed * 8 # logical N, pack 8 int4 into 1 int32 assert per_act_token, "W4A8 must use per-token scales" assert per_out_ch, "W4A8 must use per-channel scales" assert w1_scale is not None assert w2_scale is not None assert w1_scale.dtype == torch.float8_e4m3fn assert w2_scale.dtype == torch.float8_e4m3fn assert w1.dtype == torch.int32 assert w2.dtype == torch.int32 assert w1_chan_scale.dtype == torch.float32 assert w2_chan_scale.dtype == torch.float32 assert w1.size(0) == w2.size(0), "Weights expert number mismatch" assert a1q_scale is not None assert a2_scale is None assert out_dtype in [torch.bfloat16], f"Invalid output dtype: {out_dtype}" if expert_map is not None: assert expert_num_tokens is None assert not use_batched_format, "batched format not supported yet" assert group_size == 128, f"Only group size 128 supported but got {group_size=}" assert global_num_experts != -1 assert w1.size(2) * 8 == K, ( f"w1 hidden size mismatch: got {w1.size(2) * 8}, expected {K=}" ) # Translate info from expert_map to topk_ids if expert_map is not None: local_topk_ids = torch.where( expert_map[topk_ids] != -1, expert_map[topk_ids], -1 ) else: local_topk_ids = topk_ids topk = local_topk_ids.size(1) a1q_perm = _resize_cache(workspace2.view(dtype=torch.float8_e4m3fn), (M * topk, K)) mm1_out = _resize_cache(workspace13, (M * topk, N * 2))
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/utils.py
vllm/model_executor/layers/fused_moe/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools from math import prod import torch from vllm import _custom_ops as ops from vllm.model_executor.layers.quantization.utils.fp8_utils import ( per_token_group_quant_fp8, ) from vllm.model_executor.layers.quantization.utils.int8_utils import ( per_token_group_quant_int8, per_token_quant_int8, ) from vllm.model_executor.layers.quantization.utils.mxfp4_utils import ( quant_dequant_mxfp4, ) from vllm.model_executor.layers.quantization.utils.mxfp6_utils import ( quant_dequant_mxfp6, ) from vllm.model_executor.layers.quantization.utils.mxfp8_utils import ( mxfp8_e4m3_quantize, ) from vllm.triton_utils import tl, triton from vllm.utils.flashinfer import flashinfer_fp4_quantize from vllm.utils.math_utils import cdiv from vllm.utils.torch_utils import is_torch_equal_or_newer @triton.jit def _count_expert_num_tokens( topk_ids_ptr, expert_num_tokens_ptr, num_experts, topk_numel, expert_map, HAS_EXPERT_MAP: tl.constexpr, BLOCK_SIZE: tl.constexpr, ): curr_expert = tl.program_id(0) offsets = tl.arange(0, BLOCK_SIZE) topk_ids_ptrs = topk_ids_ptr + offsets acc = tl.zeros((BLOCK_SIZE,), dtype=tl.int32) for x in range(tl.cdiv(topk_numel, BLOCK_SIZE)): mask = offsets < (topk_numel - x * BLOCK_SIZE) expert_ids = tl.load(topk_ids_ptrs, mask=mask, other=-1) if HAS_EXPERT_MAP: expert_map_ptrs = expert_map + expert_ids expert_map_mask = expert_ids >= 0 expert_ids = tl.load(expert_map_ptrs, mask=expert_map_mask, other=-1) has_curr_expert = tl.where(expert_ids == curr_expert, 1, 0) acc = acc + has_curr_expert topk_ids_ptrs += BLOCK_SIZE if curr_expert < num_experts: tl.store(expert_num_tokens_ptr + curr_expert, tl.sum(acc)) def count_expert_num_tokens( topk_ids: torch.Tensor, num_local_experts: int, expert_map: torch.Tensor | None ) -> torch.Tensor: """ Count the number to tokens assigned to each expert. Parameters: - topk_ids (torch.Tensor): Tensor mapping each token to its list of experts. - num_local_experts (int): Number of experts in this rank. - expert_map (Optional[torch.Tensor]): A tensor mapping expert indices from the global expert space to the local expert space of the expert parallel shard. Returns: A tensor of size num_local_experts, where tensor[i] holds the number of tokens assigned to the ith expert. """ assert topk_ids.dtype.is_signed, "The kernel uses -1 to represent invalid topk_ids" expert_num_tokens = torch.empty( (num_local_experts), device=topk_ids.device, dtype=torch.int32 ) grid = num_local_experts BLOCK_SIZE = min(topk_ids.numel(), 1024) BLOCK_SIZE = triton.next_power_of_2(BLOCK_SIZE) _count_expert_num_tokens[(grid,)]( topk_ids, expert_num_tokens, num_local_experts, topk_ids.numel(), expert_map, HAS_EXPERT_MAP=expert_map is not None, BLOCK_SIZE=BLOCK_SIZE, ) return expert_num_tokens def _resize_cache(x: torch.Tensor, v: tuple[int, ...]) -> torch.Tensor: """ Shrink the given tensor and apply the given view to it. This is used to resize the intermediate fused_moe caches. """ assert prod(v) <= x.numel(), ( f"{v} ({prod(v)}) <= {x.shape} ({x.numel()})" ) # CUDAGRAPH unfriendly? return x.flatten()[: prod(v)].view(*v) def _nvfp4_quantize( A: torch.Tensor, A_scale: torch.Tensor | None, is_sf_swizzled_layout: bool, ) -> tuple[torch.Tensor, torch.Tensor]: return flashinfer_fp4_quantize( A, A_scale, is_sf_swizzled_layout=is_sf_swizzled_layout ) def _fp8_quantize( A: torch.Tensor, A_scale: torch.Tensor | None, per_act_token: bool, block_shape: list[int] | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: """ Perform fp8 quantization on the inputs. If a block_shape is provided, the output will be blocked. """ if block_shape is None: # TODO(luka): use QuantFP8 custom op # https://github.com/vllm-project/vllm/issues/20711 A, A_scale = ops.scaled_fp8_quant( A, A_scale, use_per_token_if_dynamic=per_act_token ) else: assert not per_act_token assert len(block_shape) == 2 _, block_k = block_shape[0], block_shape[1] A, A_scale = per_token_group_quant_fp8(A, block_k) assert cdiv(A.size(-1), block_k) == A_scale.size(-1) return A, A_scale def _int8_quantize( A: torch.Tensor, A_scale: torch.Tensor | None, per_act_token: bool, block_shape: list[int] | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: """ Perform int8 quantization on the inputs. If a block_shape is provided, the output will be blocked. """ # If weights are per-channel (per_channel_quant=True), then # activations apply per-token quantization. Otherwise, assume # activation tensor-wise fp8/int8 quantization, dynamic or static if block_shape is None: assert per_act_token, "int8 quantization only supports block or channel-wise" A, A_scale = per_token_quant_int8(A) else: assert not per_act_token assert len(block_shape) == 2 _, block_k = block_shape[0], block_shape[1] A, A_scale = per_token_group_quant_int8(A, block_k) assert cdiv(A.size(-1), block_k) == A_scale.size(-1) return A, A_scale def _mxfp4_quantize( A: torch.Tensor, A_scale: torch.Tensor | None, per_act_token_quant: bool, block_shape: list[int] | None = None, ) -> tuple[torch.Tensor, None]: assert block_shape is None # TODO: native mxfp4 is currently not integrated in vllm, # so simulating even on devices supporting this data type natively. # Once integrated, `current_platform.supports_mx()` should be used to # control quantize+dequantize, or simply quantize here down to mxfp4. A = quant_dequant_mxfp4(A) return A, None def _mxfp8_e4m3_quantize( A: torch.Tensor, A_scale: torch.Tensor | None, per_act_token_quant: bool, block_shape: list[int] | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: assert A_scale is None assert not per_act_token_quant assert block_shape is None return mxfp8_e4m3_quantize(A) def _mxfp6_e3m2_quantize( A: torch.Tensor, A_scale: torch.Tensor | None, per_act_token_quant: bool, block_shape: list[int] | None = None, ) -> tuple[torch.Tensor, None]: assert block_shape is None # TODO: native mxfp6 is currently not integrated in vllm, # so simulating even on devices supporting this data type natively. # Eventually, there should be a check based on # `current_platform.supports_mx()` here. A = quant_dequant_mxfp6(A, quant_dtype="fp6_e3m2") return A, None def _mxfp6_e2m3_quantize( A: torch.Tensor, A_scale: torch.Tensor | None, per_act_token_quant: bool, block_shape: list[int] | None = None, ) -> tuple[torch.Tensor, None]: assert block_shape is None # TODO: native mxfp6 is currently not integrated in vllm, # so simulating even on devices supporting this data type natively. # Eventually, there should be a check based on # `current_platform.supports_mx()` here. A = quant_dequant_mxfp6(A, quant_dtype="fp6_e2m3") return A, None def moe_kernel_quantize_input( A: torch.Tensor, A_scale: torch.Tensor | None, quant_dtype: None | torch.dtype | str, per_act_token_quant: bool, block_shape: list[int] | None = None, is_fp4_scale_swizzled: bool = True, ) -> tuple[torch.Tensor, torch.Tensor | None]: if quant_dtype == torch.float8_e4m3fn: return _fp8_quantize(A, A_scale, per_act_token_quant, block_shape) elif quant_dtype == torch.int8: return _int8_quantize(A, A_scale, per_act_token_quant, block_shape) elif quant_dtype == "nvfp4": return _nvfp4_quantize(A, A_scale, is_sf_swizzled_layout=is_fp4_scale_swizzled) elif quant_dtype == "mxfp4": return _mxfp4_quantize(A, A_scale, per_act_token_quant, block_shape) elif quant_dtype == "mxfp8": # TODO: `quant_dtype == "mxfp8"` is ambiguous, # should be fp8_e4m3. OCP MX also defines `fp8_e5m2`. return _mxfp8_e4m3_quantize(A, A_scale, per_act_token_quant, block_shape) elif quant_dtype == "mxfp6_e3m2": return _mxfp6_e3m2_quantize(A, A_scale, per_act_token_quant, block_shape) elif quant_dtype == "mxfp6_e2m3": return _mxfp6_e2m3_quantize(A, A_scale, per_act_token_quant, block_shape) else: return A, A_scale def _fp8_perm(m: torch.Tensor, idx: torch.Tensor) -> torch.Tensor: """ A permutation routine that works on fp8 types. """ if torch.is_floating_point(m) and m.dtype.itemsize == 1: return m.view(dtype=torch.uint8)[idx, ...].view(dtype=m.dtype) else: return m[idx, ...] def normalize_scales_shape(scales: torch.Tensor | None) -> torch.Tensor | None: if scales is not None: if scales.numel() == 1: scales = scales.view(1, 1) else: scales = scales.view(-1, scales.size(-1)) return scales def normalize_batched_scales_shape( scales: torch.Tensor | None, num_experts: int, ) -> torch.Tensor | None: if scales is not None and scales.ndim < 3: if scales.numel() == 1: scales = scales.view(1) scales = torch.repeat_interleave(scales, num_experts, dim=0).view( num_experts, 1, 1 ) else: scales = scales.view(num_experts, -1, scales.size(-1)) return scales def _validate_scale_shape( a: torch.Tensor, a_scale: torch.Tensor | None, per_act_token_quant: bool, block_shape: list[int] | None, ) -> None: if a_scale is None: return if not per_act_token_quant and block_shape is None: assert a_scale.numel() == 1, f"{a_scale.shape}" elif per_act_token_quant: assert a_scale.shape[0] == a.shape[0] and a_scale.shape[1] == 1, ( f"{a_scale.shape[0]} == {a.shape[0]} and {a_scale.shape[1]} == 1" ) else: assert block_shape is not None expected = (a.shape[0], cdiv(a.shape[1], block_shape[1])) assert a_scale.shape == expected, f"{a_scale.shape} == {expected}" def activation_without_mul(activation: str) -> str: return activation + "_no_mul" # Torch custom ops can't deal with outputs aliasing inputs so we need to # disable inplace for torch >= 2.9. # See https://github.com/vllm-project/vllm/issues/26378 @functools.cache def disable_inplace() -> bool: return is_torch_equal_or_newer("2.9")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/gpt_oss_triton_kernels_moe.py
vllm/model_executor/layers/fused_moe/gpt_oss_triton_kernels_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm import _custom_ops as ops from vllm.logger import init_logger from vllm.model_executor.layers.fused_moe.config import ( FUSED_MOE_UNQUANTIZED_CONFIG, FusedMoEQuantConfig, ) from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceNoOP, ) from vllm.model_executor.layers.fused_moe.utils import _resize_cache from vllm.triton_utils import tl, triton from vllm.utils.import_utils import has_triton_kernels logger = init_logger(__name__) if has_triton_kernels(): try: import triton_kernels.swiglu from triton_kernels.matmul_ogs import FnSpecs, FusedActivation, matmul_ogs from triton_kernels.routing import RoutingData, routing, routing_from_bitmatrix from triton_kernels.tensor import Bitmatrix except (AttributeError, ImportError) as e: logger.error( "Failed to import Triton kernels. Please make sure your triton " "version is compatible. Error: %s", e, ) @triton.jit def pack_bitmatrix( bitmatrix, topk_ids, n_rows, # n_rows in bitmatrix / topk_ids bm_cols: tl.constexpr, # n int32_t bitpacks in bitmatrix n_expts_act, # num_topk BLOCK_SIZE_M: tl.constexpr, BLOCK_SIZE_K: tl.constexpr, ): """ Packs topk_ids into a bitmatrix. code reference: https://github.com/triton-lang/triton/blob/dd1bbc52b34d202dfe5ffea1e04fb16166c5c04e/python/triton_kernels/bench/distributed.py#L264 """ pid_m = tl.program_id(0) offsets_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offsets_k = tl.arange(0, BLOCK_SIZE_K) offsets = offsets_m[:, None] * n_expts_act + offsets_k[None, :] mask = (offsets_m < n_rows)[:, None] & (offsets_k < n_expts_act)[None, :] indices = tl.load(topk_ids + offsets, mask=mask, other=-1) div = indices // 32 rem = indices % 32 one = tl.cast(1, tl.uint32) # Iterate through all the relevant bitmatrix columns. for i in range(bm_cols): # When BLOCK_SIZE_K=32, offs is just the column index. offs = tl.arange(0, BLOCK_SIZE_K // 32) + i * (BLOCK_SIZE_K // 32) # All topks that need to go into this column has the correct bit set. # Other bits are 0. x is a 2D tensor. x = tl.where( div[:, :, None] == offs[None, None, :], (one << rem)[:, :, None], 0 ) # Reduce x to get a single int32_t bitpack. y = tl.reduce_or(x, axis=1) bitmatrix_ptrs = bitmatrix + offsets_m[:, None] * bm_cols + offs[None, :] tl.store(bitmatrix_ptrs, y, mask=offsets_m[:, None] < n_rows) def triton_kernel_moe_forward( hidden_states: torch.Tensor, w1, # Tensor or triton_kernels.Tensor w2, # Tensor or triton_kernels.Tensor gating_output: torch.Tensor, topk: int, renormalize: bool, activation: str = "silu", quant_config: FusedMoEQuantConfig | None = None, apply_router_weight_on_input: bool = False, global_num_experts: int = -1, expert_map: torch.Tensor | None = None, ) -> torch.Tensor: routing_data, gather_idx, scatter_idx = routing( gating_output, topk, sm_first=not renormalize ) output = torch.empty_like(hidden_states) return triton_kernel_fused_experts( output, hidden_states, w1, w2, routing_data, gather_idx, scatter_idx, topk=topk, activation=activation, quant_config=quant_config, apply_router_weight_on_input=apply_router_weight_on_input, global_num_experts=global_num_experts, expert_map=expert_map, ) # This is a triton implementation of the fused_experts function def triton_kernel_fused_experts( output_tensor: torch.Tensor, hidden_states: torch.Tensor, w1, # Tensor or triton_kernels.Tensor w2, # Tensor or triton_kernels.Tensor routing_data, # RoutingData gather_indx, # GatherIndx scatter_indx, # ScatterIndx topk: int, activation: str = "silu", quant_config: FusedMoEQuantConfig | None = None, swiglu_alpha: float = 1.702, swiglu_limit: float = 7.0, apply_router_weight_on_input: bool = False, global_num_experts: int = -1, expert_map: torch.Tensor | None = None, intermediate_cache: torch.Tensor | None = None, a1q_scale: torch.Tensor | None = None, ) -> torch.Tensor: if quant_config is None: quant_config = FUSED_MOE_UNQUANTIZED_CONFIG # type check, uint8 means mxfp4 assert hidden_states.dtype == torch.bfloat16 assert quant_config.w1_bias is None or quant_config.w1_bias.dtype == torch.float32 assert quant_config.w2_bias is None or quant_config.w2_bias.dtype == torch.float32 # Shape check, only check non-mxfp4 assert hidden_states.ndim == 2 assert hidden_states.shape[-1] == w1.shape[-2] assert w2.shape[-1] == w1.shape[1] batch_dim = 1 M, K = hidden_states.shape[-2:] E, _, N = w1.shape if global_num_experts == -1: global_num_experts = E if intermediate_cache is None: intermediate_cache = torch.empty( (batch_dim, M * topk, N // 2), device=hidden_states.device, dtype=hidden_states.dtype, ) # Add batch_dim to output buffer because matmul_ogs expects 3D output intermediate_cache = _resize_cache( intermediate_cache, (batch_dim, M * topk, N // 2) ) output_tensor = _resize_cache(output_tensor, (batch_dim, M, K)) act = FusedActivation( FnSpecs("swiglu", triton_kernels.swiglu.swiglu_fn, ("alpha", "limit")), (swiglu_alpha, swiglu_limit), 2, ) gammas = routing_data.gate_scal if routing_data else None matmul_ogs( hidden_states, w1, quant_config.w1_bias, routing_data, gather_indx=gather_indx, precision_config=quant_config.w1_precision, gammas=gammas if apply_router_weight_on_input else None, fused_activation=act, y=intermediate_cache, ) matmul_ogs( intermediate_cache.view(M * topk, N // 2), w2, quant_config.w2_bias, routing_data, scatter_indx=scatter_indx, precision_config=quant_config.w2_precision, gammas=None if apply_router_weight_on_input else gammas, y=output_tensor, ) output_tensor = output_tensor.view(M, K) return output_tensor def make_routing_data( topk_ids: torch.Tensor, topk_weights: torch.Tensor, num_local_experts: int, ) -> tuple["RoutingData", torch.Tensor, torch.Tensor]: topk_ids = topk_ids.to(torch.int16) topk_weights = topk_weights.to(torch.bfloat16) n_rows, num_topk = topk_ids.size() BLOCK_SIZE_M = 512 BLOCK_SIZE_K = 32 bm_cols = triton.cdiv(num_local_experts, BLOCK_SIZE_K) # n_bitpacks bitmatrix = torch.zeros( (n_rows, bm_cols), dtype=torch.uint32, device=topk_ids.device ) grid = (triton.cdiv(n_rows, BLOCK_SIZE_M),) pack_bitmatrix[grid]( bitmatrix, topk_ids, n_rows, bm_cols, num_topk, BLOCK_SIZE_M=BLOCK_SIZE_M, BLOCK_SIZE_K=BLOCK_SIZE_K, ) bitmatrix_shape = [n_rows, bm_cols * 32] bitmatrix_shape_max = [n_rows, None] bitmatrix = Bitmatrix( bitmatrix, shape=bitmatrix_shape, shape_max=bitmatrix_shape_max, scratchpad=None ) # matmul_ogs expects invalid topk_weights to be -1s topk_weights = torch.where(topk_ids == -1, -1.0, topk_weights) routing_data, gather_indx, scatter_indx = routing_from_bitmatrix( bitmatrix, topk_weights, topk_ids, num_local_experts, num_topk ) return routing_data, gather_indx, scatter_indx class BaseOAITritonExperts(mk.FusedMoEPermuteExpertsUnpermute): def __init__(self, quant_config: FusedMoEQuantConfig): super().__init__(quant_config) def supports_expert_map(self) -> bool: return True def moe_problem_size( self, a1: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_ids: torch.Tensor, ) -> tuple[int, int, int, int, int]: """ Extract the MoE problem size from the given tensor arguments: - a: The hidden states, input to the MoE layer. - w1: The first set of expert weights. - w2: The second set of expert weights. - topk_ids: The topk ids. Note: extracting the problem shape from the weight and activation tensors is not obvious. It needs to be done this way specifically due to subtle issues with particular kernels, e.g. the int4 kernels divide the trailing dimension by two, so it's not "correct" to extract N or K from the trailing dimension of w1 or w2. Similarly, some kernels transpose the weights, so this needs to be kept in mind. Note: This implementation covers most cases. However, if experts require a specialized implementation, like MarlinExperts, they are free to override this function. """ assert w1.dim() == 3 and w2.dim() == 3 E, _, N = w1.size() K = a1.size(-1) assert a1.dim() == 2 assert topk_ids.size(0) == a1.size(0), f"{topk_ids.size(0)} != {a1.size(0)}" M = a1.size(0) assert topk_ids.dim() == 2 topk = topk_ids.size(1) return E, M, N, K, topk def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: # Weight application and reduction happens in the fused_experts kernel. return TopKWeightAndReduceNoOP() def _make_routing_data( self, topk_ids: torch.Tensor, topk_weights: torch.Tensor, num_local_experts: int, ) -> tuple["RoutingData", torch.Tensor, torch.Tensor]: return make_routing_data(topk_ids, topk_weights, num_local_experts) class OAITritonExperts(BaseOAITritonExperts): def __init__(self, quant_config: FusedMoEQuantConfig): # TODO (varun) : Enable activation quantization assert quant_config.use_mxfp4_w4a16, "Supports only mxfp4_w4a16" super().__init__(quant_config) @property def activation_formats( self, ) -> tuple[mk.FusedMoEActivationFormat, mk.FusedMoEActivationFormat]: return ( mk.FusedMoEActivationFormat.Standard, mk.FusedMoEActivationFormat.Standard, ) def supports_chunking(self) -> bool: return True def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: # workspace are allocated inside the kernel workspace1 = (0, 0) workspace2 = (M * topk, N // 2) output = (M, K) return (workspace1, workspace2, output) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: str, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ): if expert_map is not None: topk_ids = expert_map[topk_ids] local_num_experts = w1.size(0) if global_num_experts == -1: global_num_experts = local_num_experts routing_data, gather_indx, scatter_indx = self._make_routing_data( topk_ids, topk_weights, local_num_experts ) topk = topk_ids.size(1) triton_kernel_fused_experts( output, hidden_states, w1, w2, routing_data, gather_indx, scatter_indx, topk=topk, activation=activation, quant_config=self.quant_config, apply_router_weight_on_input=False, global_num_experts=local_num_experts, expert_map=None, # applied already intermediate_cache=workspace2, a1q_scale=a1q_scale, ) class UnfusedOAITritonExperts(BaseOAITritonExperts): """ A Triton based MoE expert class that operates on expert standard format and explicitly keeps the activation and reduction (moe_sum) steps unfused from the matmul_ogs kernel. This exposes injection points for activation and moe_sum. One use case for it is to inject LoRA modules on the activation and moe_sum. """ def __init__(self, quant_config: FusedMoEQuantConfig): # TODO (varun) : Enable activation quantization assert quant_config.use_mxfp4_w4a16, "Supports only mxfp4_w4a16" super().__init__(quant_config) @property def activation_formats( self, ) -> tuple[mk.FusedMoEActivationFormat, mk.FusedMoEActivationFormat]: return ( mk.FusedMoEActivationFormat.Standard, mk.FusedMoEActivationFormat.Standard, ) def supports_chunking(self) -> bool: return True def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: # workspace are allocated inside the kernel workspace1 = (M * topk, N // 2) workspace2 = (M * topk, max(N, K)) output = (M, K) return (workspace1, workspace2, output) def moe_sum(self, input: torch.Tensor, output: torch.Tensor): ops.moe_sum(input, output) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: str, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ): if self.quant_config is None: self.quant_config = FUSED_MOE_UNQUANTIZED_CONFIG if expert_map is not None: topk_ids = expert_map[topk_ids] local_num_experts = w1.size(0) if global_num_experts == -1: global_num_experts = local_num_experts routing_data, gather_indx, scatter_indx = self._make_routing_data( topk_ids, topk_weights, local_num_experts ) topk = topk_ids.size(1) # type check, uint8 means mxfp4 assert hidden_states.dtype == torch.bfloat16 assert ( self.quant_config.w1_bias is None or self.quant_config.w1_bias.dtype == torch.float32 ) assert ( self.quant_config.w2_bias is None or self.quant_config.w2_bias.dtype == torch.float32 ) # Shape check, only check non-mxfp4 assert hidden_states.ndim == 2 assert hidden_states.shape[-1] == w1.shape[-2] assert w2.shape[-1] == w1.shape[1] batch_dim = 1 M, K = hidden_states.shape E, _, N = w1.shape if global_num_experts == -1: global_num_experts = E # Note that the output tensor might be in workspace13 intermediate_cache1 = _resize_cache(workspace2, (batch_dim, M * topk, N)) intermediate_cache3 = _resize_cache(workspace2, (batch_dim, M * topk, K)) intermediate_cache2 = _resize_cache(workspace13, (M * topk, N // 2)) gammas = routing_data.gate_scal if routing_data else None matmul_ogs( hidden_states, w1, self.quant_config.w1_bias, routing_data, gather_indx=gather_indx, precision_config=self.quant_config.w1_precision, gammas=gammas if apply_router_weight_on_input else None, fused_activation=None, y=intermediate_cache1, ) self.activation( activation, intermediate_cache2, intermediate_cache1.view(-1, N) ) # matmul_ogs grouped reduction fuse sum across multiple experts: # y[dst_ind // n_expts_act, :] += x[src_ind, :] # Need to set n_expts_act to 1 to unfuse moe_sum routing_data.n_expts_act = 1 matmul_ogs( intermediate_cache2, w2, self.quant_config.w2_bias, routing_data, scatter_indx=scatter_indx, precision_config=self.quant_config.w2_precision, gammas=None if apply_router_weight_on_input else gammas, y=intermediate_cache3, ) self.moe_sum(intermediate_cache3.view(-1, topk, K), output)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/all2all_utils.py
vllm/model_executor/layers/fused_moe/all2all_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.distributed import ( get_ep_group, ) from vllm.model_executor.layers.fused_moe.config import ( FusedMoEConfig, FusedMoEParallelConfig, FusedMoEQuantConfig, ) from vllm.model_executor.layers.fused_moe.modular_kernel import ( FusedMoEPrepareAndFinalize, ) from vllm.model_executor.layers.quantization.utils.flashinfer_utils import ( build_flashinfer_fp8_cutlass_moe_prepare_finalize, ) from vllm.platforms import current_platform from vllm.utils.import_utils import has_deep_ep, has_pplx if current_platform.is_cuda_alike(): if has_pplx(): from .pplx_prepare_finalize import ( PplxPrepareAndFinalize, pplx_hidden_dim_scale_bytes, ) if has_deep_ep(): from .deepep_ht_prepare_finalize import DeepEPHTPrepareAndFinalize from .deepep_ll_prepare_finalize import ( DEEPEP_QUANT_BLOCK_SHAPE, DeepEPLLPrepareAndFinalize, ) def maybe_roundup_layer_hidden_size( hidden_size: int, act_dtype: torch.dtype, moe_parallel_config: FusedMoEParallelConfig, ) -> int: """ Given layer hidden size and MoE configurations, round up hidden_size if necessary. Args: hidden_size: Layer hidden-size act_dtype: Data type of the layer activations. moe_parallel_config: Fused MoE parallelization strategy configuration. Return: Rounded up hidden_size if rounding up is required based on the configs and all2all backend. Original hidden size otherwise. """ if moe_parallel_config.use_deepep_ht_kernels: hidden_size = DeepEPHTPrepareAndFinalize.maybe_roundup_layer_hidden_size( hidden_size, act_dtype ) if moe_parallel_config.use_deepep_ll_kernels: hidden_size = DeepEPLLPrepareAndFinalize.maybe_roundup_layer_hidden_size( hidden_size ) return hidden_size def maybe_make_prepare_finalize( moe: FusedMoEConfig, quant_config: FusedMoEQuantConfig | None, routing_tables: tuple[torch.Tensor, torch.Tensor, torch.Tensor] | None = None, ) -> FusedMoEPrepareAndFinalize | None: if not moe.moe_parallel_config.use_all2all_kernels: return None all2all_manager = get_ep_group().device_communicator.all2all_manager assert all2all_manager is not None prepare_finalize: FusedMoEPrepareAndFinalize | None = None if moe.use_flashinfer_cutlass_kernels: assert quant_config is not None use_deepseek_fp8_block_scale = ( quant_config is not None and quant_config.is_block_quantized ) prepare_finalize = build_flashinfer_fp8_cutlass_moe_prepare_finalize( moe=moe, use_deepseek_fp8_block_scale=use_deepseek_fp8_block_scale, ) elif moe.use_pplx_kernels: assert quant_config is not None hidden_dim_bytes, hidden_scale_bytes = pplx_hidden_dim_scale_bytes( moe.max_num_tokens, moe.hidden_dim, moe.in_dtype, quant_config.quant_dtype, per_act_token_quant=quant_config.per_act_token_quant, block_shape=quant_config.block_shape, ) all_to_all_args = dict( max_num_tokens=moe.max_num_tokens, num_experts=moe.num_experts, experts_per_token=moe.experts_per_token, # topk rank=all2all_manager.rank, world_size=all2all_manager.world_size, # dp_size actually means tp_size, bug in pplx kernels dp_size=all2all_manager.tp_group.world_size, hidden_dim=moe.hidden_dim, hidden_dim_bytes=hidden_dim_bytes, hidden_dim_scale_bytes=hidden_scale_bytes, ) num_dispatchers = ( all2all_manager.world_size // all2all_manager.tp_group.world_size ) # Intranode pplx a2a takes a group name while internode does not. if not all2all_manager.internode: all_to_all_args["group_name"] = all2all_manager.cpu_group.group_name handle = all2all_manager.get_handle(all_to_all_args) prepare_finalize = PplxPrepareAndFinalize( handle, max_num_tokens=moe.max_num_tokens, num_local_experts=moe.num_local_experts, num_dispatchers=num_dispatchers, ) elif moe.use_deepep_ht_kernels: assert moe.dp_size == all2all_manager.dp_world_size all_to_all_args = dict() handle = all2all_manager.get_handle(all_to_all_args) prepare_finalize = DeepEPHTPrepareAndFinalize( handle, num_dispatchers=all2all_manager.world_size, dp_size=all2all_manager.dp_world_size, rank_expert_offset=all2all_manager.rank * moe.num_local_experts, ) elif moe.use_deepep_ll_kernels: assert quant_config is not None global_to_physical = physical_to_global = local_expert_global_ids = None if routing_tables is not None: ( global_to_physical, physical_to_global, local_expert_global_ids, ) = routing_tables all_to_all_args = dict( max_num_tokens_per_dp_rank=moe.max_num_tokens, token_hidden_size=moe.hidden_dim, num_ep_ranks=all2all_manager.world_size, num_global_experts=moe.num_experts, num_local_experts=moe.num_experts // all2all_manager.world_size, ) handle = all2all_manager.get_handle(all_to_all_args) # Note: We may want to use FP8 dispatch just to reduce # data movement. use_fp8_dispatch = ( quant_config.quant_dtype == current_platform.fp8_dtype() and quant_config.block_shape == DEEPEP_QUANT_BLOCK_SHAPE ) prepare_finalize = DeepEPLLPrepareAndFinalize( handle, max_tokens_per_rank=moe.max_num_tokens, num_dispatchers=all2all_manager.world_size, use_fp8_dispatch=use_fp8_dispatch, global_to_physical=global_to_physical, physical_to_global=physical_to_global, local_expert_global_ids=local_expert_global_ids, ) return prepare_finalize
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/config.py
vllm/model_executor/layers/fused_moe/config.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass from enum import IntEnum from typing import Optional, Union import torch import vllm.envs as envs from vllm.config import ParallelConfig from vllm.distributed import ( get_dp_group, get_pcp_group, get_tensor_model_parallel_rank, ) from vllm.logger import init_logger from vllm.model_executor.layers.quantization.utils.ocp_mx_utils import ( OCP_MX_DTYPES, OCP_MX_Scheme, ) from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape from vllm.platforms import current_platform from vllm.utils.flashinfer import has_flashinfer_cutlass_fused_moe from vllm.utils.import_utils import has_triton_kernels from vllm.utils.math_utils import cdiv logger = init_logger(__name__) if has_triton_kernels(): try: from triton_kernels.matmul_ogs import PrecisionConfig except (ImportError, AttributeError) as e: logger.error( "Failed to import Triton kernels. Please make sure your triton " "version is compatible. Error: %s", e, ) def _get_config_dtype_str( dtype: torch.dtype, use_fp8_w8a8: bool = False, use_fp8_w8a16: bool = False, use_int8_w8a16: bool = False, use_int4_w4a16: bool = False, ocp_mx_scheme: str | None = None, ) -> str | None: """ Return a string used to construct the filename that contains the tuning info for a particular quantization scheme. See try_get_optimal_moe_config in fused_moe.py. """ if use_fp8_w8a8: return "fp8_w8a8" elif use_fp8_w8a16: return "fp8_w8a16" elif use_int8_w8a16: return "int8_w8a16" elif use_int4_w4a16: return "int4_w4a16" elif ocp_mx_scheme is not None: # The output of this function is passed to `try_get_optimal_moe_config`, # and as we only simulate OCP MX execution in fused_moe for now, # we will NOT look for `*,dtype=w_mxfp4_a_mxfp4.json` for now. return None elif dtype == torch.float: # avoiding cases where kernel fails when float32 MoE # use fp16/bfloat16 configs return "float32" return None def _quant_flags_to_group_shape( quant_dtype: torch.dtype | str | None, per_act_token_quant: bool, per_out_ch_quant: bool, block_shape: list[int] | None, ) -> tuple[GroupShape | None, GroupShape | None]: """ Convert MoE quantization flags into more generic GroupShapes. """ a_shape: GroupShape | None w_shape: GroupShape | None if block_shape is not None: assert not per_act_token_quant assert not per_out_ch_quant # TODO(bnell): this is not quite right for activations since first # dim should be 1. a_shape = GroupShape(row=block_shape[0], col=block_shape[1]) w_shape = GroupShape(row=block_shape[0], col=block_shape[1]) else: w_shape = None a_shape = None if quant_dtype is None else GroupShape.PER_TENSOR if per_act_token_quant: a_shape = GroupShape.PER_TOKEN if per_out_ch_quant: w_shape = GroupShape.PER_TOKEN return a_shape, w_shape # The type of method in top-K routing # Please keep this in sync with the counterpart defined in https://github.com/flashinfer-ai/flashinfer/blob/main/include/flashinfer/trtllm/fused_moe/runner.h class RoutingMethodType(IntEnum): # Default: Softmax -> TopK Default = (0,) # Renormalize: TopK -> Softmax Renormalize = (1,) # DeepSeekV3: Sigmoid -> RoutingBiasAdd -> Top2 in group -> Top4 groups # -> Top8 experts from the Top4 groups DeepSeekV3 = (2,) # Llama4: Top1 -> Sigmoid Llama4 = (3,) # RenormalizeNaive: Softmax -> TopK -> Renormalize RenormalizeNaive = (4,) # TopK: TopK (no softmax) TopK = (5,) # Unspecified Unspecified = 6.0 @dataclass class FusedMoEQuantDesc: """ A quantization descriptor for fused MoE ops. This class can describe either activations or weights. """ # The quantized type of this parameters. None means unquantized or # already quantized. # TODO (bnell): use scalar_type instead of Union. dtype: torch.dtype | str | None = None # A field that describes the quantization group shape, from quant_utils.py. # * (-1, -1) for per-tensor quantization # * (1, -1) for per-row quantization # * (-1, 1) for per-column quantization # * (128, 128) for 128x128 deepseek style block quantization # * (1, 128) for deepseek style activation quantization # (i.e. per-token-per-group) shape: GroupShape | None = None # Quantization scales. # TODO(bnell): maybe put PrecisionConfigs in subclass of QuantDesc? scale: Union[torch.Tensor, "PrecisionConfig", None] = None # Quantization alphas or gscales, used for nvfp4 types. # W4A8 FP8: used for per-channel scales # TODO(bnell): put some of these in subclasses alpha_or_gscale: torch.Tensor | None = None # Zero points for int4/int8 types zp: torch.Tensor | None = None # Biases for GPT triton MoE bias: torch.Tensor | None = None # TODO(bnell): have subclasses for specific moe methods? # e.g. for specific arguments bias, precision, etc. @dataclass class FusedMoEQuantConfig: """ The FusedMoEQuantConfig contains all the quantization parameters for a single FusedMoEMethodBase operation. It consists of four FusedMoEQuantDescs, one for each activation and set of weights. Each FusedMoEMethodBase must implement a get_fused_moe_quant_config method to construct a FusedMoEQuantConfig for use with that class. FusedMoEQuant configs are only used for modular kernels, fused_experts (from fused_moe.py), cutlass_moe_fp[48], rocm_aiter_fused_experts and triton_kernel_moe_forward. Other MoE methods can ignore the FusedMoEQuantConfig (for now) and hardcode it to None. There are currently some restrictions on what can be expressed: - Most MoE ops only support similar quantization strategies for each parameter, e.g. both weights must have the same GroupShape and both activations must share the same GroupShape. One exception to this is the cutlass moe which allows per channel quantization on the outputs. Note: this restrictions are not always rigorously checked. - Not all fused MoE functions support all the parameters, e.g. zero points, global scales, alphas and biases are not universally supported. - Fully general GroupShapes are not allowed. Activations only support per token, per tensor or K-blocked. - Weights are not required to have a GroupShape since they have already been quantized. Other notes: - PrecisionConfigs are specific to GPT OSS Triton. - As a follow up it would probably make sense to subclass FusedMoEQuantDesc or FusedMoEQuantConfig for particular FusedMoEMethodBase subclasses so that only the required quantization parameters are used/stored. """ # TODO(bnell) make sure a1_scales/a2_scales don't interfere with chunking _a1: FusedMoEQuantDesc _a2: FusedMoEQuantDesc _w1: FusedMoEQuantDesc _w2: FusedMoEQuantDesc def __post_init__(self): assert not self.per_act_token_quant or self.block_shape is None, ( "illegal quantization" ) # # Convenience accessors for various properties. # @property def quant_dtype(self) -> torch.dtype | str | None: return self._a1.dtype @property def is_quantized(self) -> bool: return self.quant_dtype is not None @property def is_per_act_token(self) -> bool: return self._a1.shape == GroupShape.PER_TOKEN @property def per_act_token_quant(self) -> bool: return self._a1.shape == GroupShape.PER_TOKEN @property def per_out_ch_quant(self) -> bool: return self._w1.shape == GroupShape.PER_TOKEN @property def is_per_tensor(self) -> bool: return self._a1.shape == GroupShape.PER_TENSOR @property def block_shape(self) -> list[int] | None: if ( self._a1.shape is not None and self._a1.shape != GroupShape.PER_TENSOR and self._a1.shape != GroupShape.PER_TOKEN ): return [self._a1.shape.row, self._a1.shape.col] else: return None @property def is_block_quantized(self) -> bool: return self.block_shape is not None @property def a1_scale(self) -> torch.Tensor | None: assert self._a1.scale is None or isinstance(self._a1.scale, torch.Tensor) return self._a1.scale @property def a1_gscale(self) -> torch.Tensor | None: return self._a1.alpha_or_gscale @property def a2_scale(self) -> torch.Tensor | None: assert self._a2.scale is None or isinstance(self._a2.scale, torch.Tensor) return self._a2.scale @property def a2_gscale(self) -> torch.Tensor | None: return self._a2.alpha_or_gscale @property def w1_scale(self) -> torch.Tensor | None: assert self._w1.scale is None or isinstance(self._w1.scale, torch.Tensor) return self._w1.scale @property def w1_zp(self) -> torch.Tensor | None: return self._w1.zp @property def w1_bias(self) -> torch.Tensor | None: return self._w1.bias @property def w1_precision(self) -> Optional["PrecisionConfig"]: assert self._w1.scale is None or isinstance(self._w1.scale, PrecisionConfig) return self._w1.scale @property def g1_alphas(self) -> torch.Tensor | None: return self._w1.alpha_or_gscale @property def w2_scale(self) -> torch.Tensor | None: assert self._w2.scale is None or isinstance(self._w2.scale, torch.Tensor) return self._w2.scale @property def w2_zp(self) -> torch.Tensor | None: return self._w2.zp @property def w2_bias(self) -> torch.Tensor | None: return self._w2.bias @property def w2_precision(self) -> Optional["PrecisionConfig"]: assert self._w2.scale is None or isinstance(self._w2.scale, PrecisionConfig) return self._w2.scale @property def g2_alphas(self) -> torch.Tensor | None: return self._w2.alpha_or_gscale @property def use_fp8_w8a8(self) -> bool: return self.quant_dtype == torch.float8_e4m3fn @property def use_int8_w8a8(self) -> bool: return self.quant_dtype == torch.int8 @property def use_int8_w8a16(self) -> bool: return self._a1.dtype is None and self._w1.dtype == torch.int8 @property def use_fp8_w8a16(self) -> bool: return self._a1.dtype is None and self._w1.dtype == current_platform.fp8_dtype() @property def use_int4_w4a16(self) -> bool: return self._a1.dtype is None and self._w1.dtype == "int4" @property def ocp_mx_scheme(self) -> str | None: if not hasattr(self, "_ocp_mx_scheme"): if (self._a1.dtype is not None and not isinstance(self._a1.dtype, str)) or ( self._w1.dtype is not None and not isinstance(self._w1.dtype, str) ): self._ocp_mx_scheme = None else: ocp_mx_scheme = OCP_MX_Scheme.from_quant_dtype( self._a1.dtype, self._w1.dtype ) if ocp_mx_scheme is not None: ocp_mx_scheme = ocp_mx_scheme.value self._ocp_mx_scheme = ocp_mx_scheme return self._ocp_mx_scheme @property def use_mxfp4_w4a16(self) -> bool: return self._a1.dtype is None and self._w1.dtype == "mxfp4" @property def use_mxfp4_w4a4(self) -> bool: return self._a1.dtype == "mxfp4" and self._w1.dtype == "mxfp4" @property def use_nvfp4_w4a4(self) -> bool: return self.quant_dtype == "nvfp4" def config_name(self, dtype: torch.dtype) -> str | None: """ Return a string used to construct the filename that contains the tuning info for a particular quantization scheme. See try_get_optimal_moe_config in fused_moe.py. """ return _get_config_dtype_str( use_fp8_w8a8=self.use_fp8_w8a8, use_fp8_w8a16=self.use_fp8_w8a16, use_int8_w8a16=self.use_int8_w8a16, use_int4_w4a16=self.use_int4_w4a16, ocp_mx_scheme=self.ocp_mx_scheme, dtype=dtype, ) def scale_shape( self, max_tokens: int, hidden_dim: int, ) -> tuple[int, int] | None: """ Construct the proper activation scale shape for this config. """ if self.is_quantized: if self.is_block_quantized: assert self.block_shape is not None _, block_k = self.block_shape k_tiles = cdiv(hidden_dim, block_k) return (max_tokens, k_tiles) elif self.is_per_act_token: return (max_tokens, 1) else: return (1, 1) else: return None def batched_scale_shape( self, num_experts: int, max_tokens: int, hidden_dim: int, ) -> tuple[int, int, int] | None: """ Construct the proper activation batched scale shape for this config, e.g. (num experts, *scale_shape). """ if self.is_quantized: scale_shape = self.scale_shape(max_tokens, hidden_dim) assert scale_shape is not None return (num_experts, *scale_shape) else: return None @staticmethod def make( quant_dtype: torch.dtype | str | None = None, per_act_token_quant: bool = False, per_out_ch_quant: bool = False, block_shape: list[int] | None = None, w1_scale: Union[torch.Tensor, "PrecisionConfig", None] = None, w2_scale: Union[torch.Tensor, "PrecisionConfig", None] = None, a1_scale: torch.Tensor | None = None, a2_scale: torch.Tensor | None = None, g1_alphas: torch.Tensor | None = None, g2_alphas: torch.Tensor | None = None, a1_gscale: torch.Tensor | None = None, a2_gscale: torch.Tensor | None = None, w1_bias: torch.Tensor | None = None, w2_bias: torch.Tensor | None = None, w1_zp: torch.Tensor | None = None, w2_zp: torch.Tensor | None = None, weight_dtype: torch.dtype | str | None = None, ) -> "FusedMoEQuantConfig": """ General builder function for a FusedMoEQuantConfig. - quant_dtype: Optional quantization type. None if activations are unquantized or quantized prior to calling. Note: "nvfp4", "mxfp4", "mxfp6_e3m2", "mxfp6_e2m3" are the only valid string values for quant_dtype. - per_act_token_quant: Activations have per token quantization. - per_out_ch_quant: Outputs have per channel quantization. (only for cutlass). - block_shape: Optional block size for block-wise quantization. Incompatible with per_act_token and per_out_ch quant. - w1_scale: Optional scale to be used for w1. - w2_scale: Optional scale to be used for w2. - a1_scale: Optional scale to be used for a1. - a2_scale: Optional scale to be used for a2. - g1_alphas: Optional global quantization scales for w1 (for nvfp4). per-channel scales for w1 (for W4A8 FP8). - g2_alphas: Optional global quantization scales for w2 (for nvfp4). per-channel scales for w2 (for W4A8 FP8). - a1_gscale: Optional global quantization scales for a1 (for nvfp4). - a2_gscale: Optional global quantization scales for a2 (for nvfp4). - w1_bias: Optional biases for w1 (GPT OSS Triton). - w2_bias: Optional biases for w1 (GPT OSS Triton). - w1_zp: Optional w1 zero points for int4/int8 quantization. - w2_zp: Optional w2 zero points for int4/int8 quantization. """ assert not isinstance(quant_dtype, str) or quant_dtype in { "nvfp4", "mxfp4", "mxfp6_e3m2", "mxfp6_e2m3", } assert not isinstance(weight_dtype, str) or weight_dtype in { "nvfp4", "mxfp4", "mxfp6_e3m2", "mxfp6_e2m3", "int4", } if weight_dtype is None: weight_dtype = quant_dtype a_shape, w_shape = _quant_flags_to_group_shape( quant_dtype, per_act_token_quant, per_out_ch_quant, block_shape ) quant_config = FusedMoEQuantConfig( _a1=FusedMoEQuantDesc(quant_dtype, a_shape, a1_scale, a1_gscale), _a2=FusedMoEQuantDesc(quant_dtype, a_shape, a2_scale, a2_gscale), _w1=FusedMoEQuantDesc( weight_dtype, w_shape, w1_scale, g1_alphas, w1_zp, w1_bias ), _w2=FusedMoEQuantDesc( weight_dtype, w_shape, w2_scale, g2_alphas, w2_zp, w2_bias ), ) assert quant_config.per_act_token_quant == per_act_token_quant assert quant_config.per_out_ch_quant == per_out_ch_quant assert quant_config.block_shape == block_shape return quant_config def fp8_w8a8_moe_quant_config( w1_scale: torch.Tensor, w2_scale: torch.Tensor, a1_scale: torch.Tensor | None = None, a2_scale: torch.Tensor | None = None, per_act_token_quant: bool = False, per_out_ch_quant: bool = False, block_shape: list[int] | None = None, a1_gscale: torch.Tensor | None = None, a2_gscale: torch.Tensor | None = None, g1_alphas: torch.Tensor | None = None, g2_alphas: torch.Tensor | None = None, ) -> FusedMoEQuantConfig: """ Construct a quant config for fp8 activations and fp8 weights. """ return FusedMoEQuantConfig.make( torch.float8_e4m3fn, w1_scale=w1_scale, g1_alphas=g1_alphas, w2_scale=w2_scale, g2_alphas=g2_alphas, a1_scale=a1_scale, a1_gscale=a1_gscale, a2_scale=a2_scale, a2_gscale=a2_gscale, per_act_token_quant=per_act_token_quant, per_out_ch_quant=per_out_ch_quant, block_shape=block_shape, ) def int8_w8a8_moe_quant_config( w1_scale: torch.Tensor, w2_scale: torch.Tensor, a1_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, per_act_token_quant: bool = False, ) -> FusedMoEQuantConfig: """ Construct a quant config for int8 activations and int8 weights. """ return FusedMoEQuantConfig.make( torch.int8, w1_scale=w1_scale, w2_scale=w2_scale, a1_scale=a1_scale, a2_scale=a2_scale, per_act_token_quant=per_act_token_quant, per_out_ch_quant=False, block_shape=None, ) def gptq_marlin_moe_quant_config( w1_scale: torch.Tensor, w2_scale: torch.Tensor, weight_bits: int, group_size: int, w1_zp: torch.Tensor | None = None, w2_zp: torch.Tensor | None = None, w1_bias: torch.Tensor | None = None, w2_bias: torch.Tensor | None = None, ): """ Construct a quant config for gptq marlin quantization. """ from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape w_shape = None if group_size == -1 else GroupShape(row=1, col=group_size) # Activations are NOT quantized for GPTQ (fp16/bf16) a_shape = w_shape # Same as weight shape for alignment # Determine weight dtype if weight_bits == 4: weight_dtype = "int4" elif weight_bits == 8: weight_dtype = torch.int8 else: raise ValueError(f"Unsupported weight_bits: {weight_bits}") return FusedMoEQuantConfig( _a1=FusedMoEQuantDesc(dtype=None, shape=a_shape), _a2=FusedMoEQuantDesc(dtype=None, shape=a_shape), _w1=FusedMoEQuantDesc(weight_dtype, w_shape, w1_scale, None, w1_zp, w1_bias), _w2=FusedMoEQuantDesc(weight_dtype, w_shape, w2_scale, None, w2_zp, w2_bias), ) def mxfp4_w4a16_moe_quant_config( w1_scale: Union[torch.Tensor, "PrecisionConfig"], w2_scale: Union[torch.Tensor, "PrecisionConfig"], w1_bias: torch.Tensor | None = None, w2_bias: torch.Tensor | None = None, ) -> FusedMoEQuantConfig: """ Construct a quant config for unquantized activations and mxfp4 weights. """ return FusedMoEQuantConfig( _a1=FusedMoEQuantDesc(), _a2=FusedMoEQuantDesc(), _w1=FusedMoEQuantDesc("mxfp4", None, w1_scale, None, None, w1_bias), _w2=FusedMoEQuantDesc("mxfp4", None, w2_scale, None, None, w2_bias), ) def mxfp4_mxfp8_moe_quant_config( w1_scale: Union[torch.Tensor, "PrecisionConfig"], w2_scale: Union[torch.Tensor, "PrecisionConfig"], a1_scale: torch.Tensor | None = None, a2_scale: torch.Tensor | None = None, w1_bias: torch.Tensor | None = None, w2_bias: torch.Tensor | None = None, block_shape: list[int] | None = None, ) -> FusedMoEQuantConfig: """ Construct a quant config for mxfp4 activations and mxfp4 weights. """ return FusedMoEQuantConfig( _a1=FusedMoEQuantDesc("mxfp8"), _a2=FusedMoEQuantDesc("mxfp8"), _w1=FusedMoEQuantDesc("mxfp4", None, w1_scale, None, None, w1_bias), _w2=FusedMoEQuantDesc("mxfp4", None, w2_scale, None, None, w2_bias), ) def ocp_mx_moe_quant_config( quant_dtype: str, w1_scale: Union[torch.Tensor, "PrecisionConfig"], w2_scale: Union[torch.Tensor, "PrecisionConfig"], weight_dtype: str | None = None, a1_scale: torch.Tensor | None = None, a2_scale: torch.Tensor | None = None, w1_bias: torch.Tensor | None = None, w2_bias: torch.Tensor | None = None, block_shape: list[int] | None = None, ) -> FusedMoEQuantConfig: """ Construct a quant config for mxfp4 activations and mxfp4 weights. """ assert quant_dtype in OCP_MX_DTYPES return FusedMoEQuantConfig.make( quant_dtype=quant_dtype, weight_dtype=weight_dtype, w1_scale=w1_scale, w2_scale=w2_scale, a1_scale=a1_scale, a2_scale=a2_scale, w1_bias=w1_bias, w2_bias=w2_bias, per_act_token_quant=False, per_out_ch_quant=False, block_shape=block_shape, ) def nvfp4_moe_quant_config( g1_alphas: torch.Tensor, g2_alphas: torch.Tensor, a1_gscale: torch.Tensor, a2_gscale: torch.Tensor, w1_scale: torch.Tensor, w2_scale: torch.Tensor, ) -> FusedMoEQuantConfig: """ Construct a quant config for mxfp4 activations and nvp4 weights. """ return FusedMoEQuantConfig.make( "nvfp4", w1_scale=w1_scale, w2_scale=w2_scale, a1_gscale=a1_gscale, a2_gscale=a2_gscale, g1_alphas=g1_alphas, g2_alphas=g2_alphas, per_act_token_quant=False, per_out_ch_quant=False, block_shape=None, ) def int4_w4a16_moe_quant_config( w1_scale: torch.Tensor, w2_scale: torch.Tensor, w1_zp: torch.Tensor | None, w2_zp: torch.Tensor | None, block_shape: list[int] | None = None, ) -> FusedMoEQuantConfig: """ Construct a quant config for 16-bit float activations and int4 weights. """ group_shape = GroupShape(*block_shape) if block_shape is not None else None return FusedMoEQuantConfig( _a1=FusedMoEQuantDesc(shape=group_shape), _a2=FusedMoEQuantDesc(shape=group_shape), _w1=FusedMoEQuantDesc("int4", group_shape, w1_scale, None, w1_zp), _w2=FusedMoEQuantDesc("int4", group_shape, w2_scale, None, w2_zp), ) def fp8_w8a16_moe_quant_config( w1_scale: torch.Tensor, w2_scale: torch.Tensor, block_shape: list[int] | None = None, ) -> FusedMoEQuantConfig: """ Construct a quant config for 16-bit float activations and fp8 weights. """ group_shape = GroupShape(*block_shape) if block_shape is not None else None return FusedMoEQuantConfig( _a1=FusedMoEQuantDesc(), _a2=FusedMoEQuantDesc(), _w1=FusedMoEQuantDesc( current_platform.fp8_dtype(), group_shape, w1_scale, None, None ), _w2=FusedMoEQuantDesc( current_platform.fp8_dtype(), group_shape, w2_scale, None, None ), ) def int8_w8a16_moe_quant_config( w1_scale: torch.Tensor, w2_scale: torch.Tensor, w1_zp: torch.Tensor | None, w2_zp: torch.Tensor | None, block_shape: list[int] | None = None, ) -> FusedMoEQuantConfig: """ Construct a quant config for 16-bit float activations and int8 weights. """ group_shape = GroupShape(*block_shape) if block_shape is not None else None return FusedMoEQuantConfig( _a1=FusedMoEQuantDesc(shape=group_shape), _a2=FusedMoEQuantDesc(shape=group_shape), _w1=FusedMoEQuantDesc(torch.int8, group_shape, w1_scale, None, w1_zp), _w2=FusedMoEQuantDesc(torch.int8, group_shape, w2_scale, None, w2_zp), ) def int4_w4afp8_moe_quant_config( w1_scale: torch.Tensor, w2_scale: torch.Tensor, g1_alphas: torch.Tensor, g2_alphas: torch.Tensor, per_act_token_quant: bool = False, per_out_ch_quant: bool = False, block_shape: list[int] | None = None, ) -> FusedMoEQuantConfig: """ Construct a quant config for fp8 activations and int4 weights. """ return FusedMoEQuantConfig.make( torch.float8_e4m3fn, # quant dtype for activations w1_scale=w1_scale, w2_scale=w2_scale, g1_alphas=g1_alphas, g2_alphas=g2_alphas, per_act_token_quant=per_act_token_quant, per_out_ch_quant=per_out_ch_quant, block_shape=block_shape, weight_dtype="int4", # weight dtype for weights ) def awq_marlin_moe_quant_config( w1_scale: torch.Tensor, w2_scale: torch.Tensor, w1_zp: torch.Tensor | None, w2_zp: torch.Tensor | None, weight_bits: int, group_size: int, w1_bias: torch.Tensor | None = None, w2_bias: torch.Tensor | None = None, ) -> FusedMoEQuantConfig: """ Construct a quant config for awq marlin quantization. """ from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape w_shape = None if group_size == -1 else GroupShape(row=1, col=group_size) # Activations are NOT quantized for AWQ (fp16/bf16) a_shape = w_shape # Same as weight shape for alignment # Determine weight dtype if weight_bits == 4: weight_dtype = "int4" elif weight_bits == 8: weight_dtype = torch.int8 else: raise ValueError(f"Unsupported weight_bits: {weight_bits}") return FusedMoEQuantConfig( _a1=FusedMoEQuantDesc(dtype=None, shape=a_shape), _a2=FusedMoEQuantDesc(dtype=None, shape=a_shape), _w1=FusedMoEQuantDesc(weight_dtype, w_shape, w1_scale, None, w1_zp, w1_bias), _w2=FusedMoEQuantDesc(weight_dtype, w_shape, w2_scale, None, w2_zp, w2_bias), ) def biased_moe_quant_config( w1_bias: torch.Tensor | None, w2_bias: torch.Tensor | None, ) -> FusedMoEQuantConfig: """ Construct a quant config for unquantized activations with biases. """ return FusedMoEQuantConfig( _a1=FusedMoEQuantDesc(), _a2=FusedMoEQuantDesc(), _w1=FusedMoEQuantDesc(bias=w1_bias), _w2=FusedMoEQuantDesc(bias=w2_bias), ) # A FusedMoEQuantConfig constant for an unquantized MoE op. FUSED_MOE_UNQUANTIZED_CONFIG: FusedMoEQuantConfig = FusedMoEQuantConfig.make() @dataclass class FusedMoEParallelConfig: tp_size: int pcp_size: int dp_size: int ep_size: int tp_rank: int pcp_rank: int dp_rank: int ep_rank: int use_ep: bool # whether to use EP or not all2all_backend: str # all2all backend for MoE communication @property def use_all2all_kernels(self): return self.dp_size > 1 and self.use_ep @property def use_pplx_kernels(self): return self.use_all2all_kernels and self.all2all_backend == "pplx" @property def use_deepep_ht_kernels(self): return ( self.use_all2all_kernels and self.all2all_backend == "deepep_high_throughput" ) @property def use_deepep_ll_kernels(self): return self.use_all2all_kernels and self.all2all_backend == "deepep_low_latency" @staticmethod def flatten_tp_across_dp_and_pcp( tp_size: int, dp_size: int, dp_rank: int, pcp_size: int, pcp_rank: int ) -> tuple[int, int]: tp_rank = 0 if tp_size == 1 else get_tensor_model_parallel_rank() # There are actually dp_size * pcp_size * tp_size devices. # Update tp_size and tp_rank so we shard across all devices. flatten_tp_size = dp_size * pcp_size * tp_size flatten_tp_rank = dp_rank * pcp_size * tp_size + pcp_rank * tp_size + tp_rank return flatten_tp_size, flatten_tp_rank @staticmethod def make( tp_size_: int, pcp_size_: int, dp_size_: int, vllm_parallel_config: ParallelConfig, ) -> "FusedMoEParallelConfig": """ Determine MoE parallel configuration. Based on the input `tp_size_`, `dp_size_` and vllm's parallel config, determine what level's of parallelism to use in the fused moe layer. Args: tp_size_ (int): `tp_size` passed into the FusedMoE constructor. pcp_size_ (int): `pcp_size` passed into the FusedMoE constructor. dp_size_ (int): `dp_size` passed into the FusedMoE constructor. vllm_parallel_config (ParallelConfig): vLLM's parallel config object which contains the `enable_expert_parallel` flag. Examples: When there is no parallelism requested, i.e. `tp_size_` = `pcp_size_` = `dp_size_` = 1, we simply return the sizes unaltered and the ranks set to 0. Expert Parallelism is considered only when either `dp_size_`, `pcp_size_` or `tp_size_` is non trivial. Note that PCP serves the same function as DP here. When TP = 2, DP(PCP) = 1 and EP = False, the configuration on different devices: - device 0 : TP = {2, 0} DP = {1, 0} EP = {1, 0} // legend : {size, rank} - device 1 : TP = {2, 1} DP = {1, 0} EP = {1, 0} - Comment : Tensors are sharded across 2 devices. When TP = 1, DP(PCP) = 2 and EP = False, the configuration on different devices: - device 0 : TP = {2, 0} DP = {2, 0} EP = {1, 0} - device 1 : TP = {2, 1} DP = {2, 1} EP = {1, 0} - Comment: There are 2 engine instances and the tensors are sharded across 2 decvices. When TP = 2, DP(PCP) = 2 and EP = False, the configuration on different devices: - device 0: TP = {4, 0} DP = {2, 0} EP = {1, 0} - device 1: TP = {4, 1} DP = {2, 0} EP = {1, 0} - device 2: TP = {4, 2} DP = {2, 1} EP = {1, 0} - device 3: TP = {4, 3} DP = {2, 1} EP = {1, 0} - Comment: There are 2 engine instances and the tensors are sharded across 4 devices. When, TP = 2, DP(PCP) = 1 and EP = True, the configuration on different devices: - device 0: TP = {1, 0} DP = {1, 0} EP = {2, 0} - device 1: TP = {1, 0} DP = {1, 0} EP = {2, 1} - Comment: The experts are split between the 2 devices. When, TP = 1, DP(PCP) = 2 and EP = True, the configuration on different devices: - device 0: TP = {1, 0} DP = {2, 0} EP = {2, 0} - device 1: TP = {1, 0} DP = {2, 1} EP = {2, 1} - Comment: There are 2 engine instances and the experts are split between the 2 devices. When TP = 2, DP(PCP) = 2 and EP = True, the configuration on different devices: - device 0: TP = {1, 0} DP = {2, 0} EP = {4, 0} - device 1: TP = {1, 0} DP = {2, 0} EP = {4, 1} - device 2: TP = {1, 0} DP = {2, 1} EP = {4, 2} - device 3: TP = {1, 0} DP = {2, 1} EP = {4, 3} - Comment: There are 2 engine instances and the experts are split between the 4 devices. """ use_ep = ( dp_size_ * pcp_size_ * tp_size_ > 1 and vllm_parallel_config.enable_expert_parallel )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/flashinfer_cutlass_moe.py
vllm/model_executor/layers/fused_moe/flashinfer_cutlass_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm.logger import init_logger from vllm.model_executor.layers.fused_moe.config import FusedMoEQuantConfig from vllm.model_executor.layers.fused_moe.flashinfer_cutlass_prepare_finalize import ( # noqa: E501 create_flashinfer_prepare_finalize, ) from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceNoOP, ) from vllm.utils.flashinfer import ( flashinfer_cutlass_fused_moe, has_flashinfer_cutlass_fused_moe, ) logger = init_logger(__name__) def is_valid_flashinfer_cutlass_fused_moe( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor ) -> bool: """ Check if the given problem size is supported by the FlashInfer CUTLASS MoE kernel. """ if not has_flashinfer_cutlass_fused_moe(): logger.debug_once( "FlashInferExperts disabled: flashinfer_cutlass_fused_moe not available." ) return False # Data type checks if ( w1.dtype != torch.uint8 or w2.dtype != torch.uint8 or hidden_states.dtype not in [torch.float32, torch.float16, torch.bfloat16] ): logger.debug_once( "FlashInferExperts disabled: w1/w2 must be torch.uint8 " f"(got w1={w1.dtype}, w2={w2.dtype}), hidden_states must be " f"float32, float16, or bfloat16 (got {hidden_states.dtype})." ) return False return True class FlashInferExperts(mk.FusedMoEPermuteExpertsUnpermute): def __init__( self, out_dtype: torch.dtype, quant_config: FusedMoEQuantConfig, ep_rank: int = 0, ep_size: int = 1, tp_rank: int = 0, tp_size: int = 1, use_dp: bool = False, use_deepseek_fp8_block_scale: bool = False, ): super().__init__(quant_config) assert quant_config.quant_dtype in ("nvfp4", torch.float8_e4m3fn, None), ( "Only nvfp4, fp8, bfloat16 and" " float16 quantization are currently supported." ) self.ep_rank = ep_rank self.ep_size = ep_size self.tp_rank = tp_rank self.tp_size = tp_size self.out_dtype = out_dtype self.use_dp = use_dp # Enables DeepSeek-style FP8 block-scale path: # - pass per-block weight scales to the kernel # - skip input activation quantization (kernel applies scaling) self.use_deepseek_fp8_block_scale = use_deepseek_fp8_block_scale @property def activation_formats( self, ) -> tuple[mk.FusedMoEActivationFormat, mk.FusedMoEActivationFormat]: return ( mk.FusedMoEActivationFormat.Standard, mk.FusedMoEActivationFormat.Standard, ) def supports_expert_map(self) -> bool: return False def supports_chunking(self) -> bool: # This refers to TP chunking; DP chunking is handled separately. return True def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: return TopKWeightAndReduceNoOP() def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: # We use global_num_experts due to how moe_align_block_size handles # expert_maps. """ Compute the shapes for the temporary and final outputs of the two gemms and activation in the fused expert function. Since the gemms are independent, the workspace for the first gemm can be shared with the workspace for the last gemm. Returns a tuple of: - workspace13 shape tuple: must be large enough to hold the result of either expert gemm. - workspace2 shape tuple: must be large enough to hold the result of the activation function. - output shape tuple: must be exact size of the final gemm output. - Workspace type: The dtype to use for the workspace tensors. - Note: in order for activation chunking to work, the first dimension of each tuple must be the number of tokens. """ workspace1 = (M, K) workspace2 = (0,) # For TP, the quantization is fused with fused_moe call. output_shape = (M, K * 2 if self.quant_dtype == "nvfp4" and self.use_dp else K) # The workspace is determined by `aq`, since it comes after any # potential communication op and is involved in the expert computation. return (workspace1, workspace2, output_shape) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: str, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor | None, workspace2: torch.Tensor | None, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool | None, ): from flashinfer.fused_moe.core import ActivationType activation_str_to_value_map = { "silu": ActivationType.Swiglu, # This is the default "relu2_no_mul": ActivationType.Relu2, } assert activation in activation_str_to_value_map, ( f"{activation=} missing from {activation_str_to_value_map.keys()=}" ) # Select quantization metadata based on FP8 format/path if ( self.quant_dtype == torch.float8_e4m3fn and not self.use_deepseek_fp8_block_scale ): # FP8 per-tensor path: use global alphas/scales; do not pass input_sf quant_scales = [ self.g1_alphas, self.a2_gscale, self.g2_alphas, self.a1_gscale, ] a1q_scale = None # not passing input_sf in fp8 fc1_expert_weights = w1 fc2_expert_weights = w2 elif self.quant_dtype == "nvfp4": # Ensure w1_scale and w2_scale are not None before calling view assert self.w1_scale is not None and self.w2_scale is not None, ( "w1_scale and w2_scale must not be None for FlashInferExperts" ) # Flashinfer CUTLASS kernel takes scalar global scales, # min because inv_scale. quant_scales = [ self.a1_gscale, self.w1_scale.view(torch.int32), self.g1_alphas, self.a2_gscale, self.w2_scale.view(torch.int32), self.g2_alphas, ] # FlashInfer API requires weight to be long for nvfp4 fc1_expert_weights = w1.view(torch.long) fc2_expert_weights = w2.view(torch.long) elif self.use_deepseek_fp8_block_scale: # FP8 block-scale path: provide block-scale weights, omit a1q_scale quant_scales = [ self.w1_scale, self.w2_scale, ] a1q_scale = None fc1_expert_weights = w1 fc2_expert_weights = w2 else: quant_scales = None a1q_scale = None fc1_expert_weights = w1 fc2_expert_weights = w2 _ = flashinfer_cutlass_fused_moe( input=hidden_states, token_selected_experts=topk_ids.to(torch.int), token_final_scales=topk_weights, fc1_expert_weights=fc1_expert_weights, fc2_expert_weights=fc2_expert_weights, output_dtype=self.out_dtype, quant_scales=quant_scales, input_sf=a1q_scale, tp_size=self.tp_size, tp_rank=self.tp_rank, ep_size=self.ep_size, ep_rank=self.ep_rank, output=output, activation_type=activation_str_to_value_map[activation], # Informs FlashInfer to use the block-scale decoding path when True use_deepseek_fp8_block_scale=self.use_deepseek_fp8_block_scale, ) def flashinfer_cutlass_moe_fp4( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, quant_config: FusedMoEQuantConfig, inplace: bool = False, activation: str = "silu", global_num_experts: int = -1, expert_map: torch.Tensor | None = None, apply_router_weight_on_input: bool = False, ) -> torch.Tensor: fused_experts = mk.FusedMoEModularKernel( create_flashinfer_prepare_finalize(use_dp=False), FlashInferExperts( out_dtype=hidden_states.dtype, quant_config=quant_config, use_dp=False, ), ) return fused_experts( hidden_states=hidden_states, w1=w1, w2=w2, topk_weights=topk_weights, topk_ids=topk_ids, inplace=inplace, activation=activation, global_num_experts=global_num_experts, expert_map=expert_map, apply_router_weight_on_input=apply_router_weight_on_input, ) def flashinfer_cutlass_moe( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, quant_config: FusedMoEQuantConfig, inplace: bool = False, activation: str = "silu", global_num_experts: int = -1, expert_map: torch.Tensor | None = None, apply_router_weight_on_input: bool = False, tp_rank: int = 0, tp_size: int = 1, ep_rank: int = 0, ep_size: int = 1, use_dp: bool = False, ) -> torch.Tensor: fused_experts = mk.FusedMoEModularKernel( create_flashinfer_prepare_finalize(use_dp=use_dp), FlashInferExperts( out_dtype=hidden_states.dtype, quant_config=quant_config, tp_rank=tp_rank, tp_size=tp_size, ep_rank=ep_rank, ep_size=ep_size, ), ) return fused_experts( hidden_states=hidden_states, w1=w1, w2=w2, topk_weights=topk_weights, topk_ids=topk_ids, inplace=inplace, activation=activation, global_num_experts=global_num_experts, expert_map=expert_map, apply_router_weight_on_input=apply_router_weight_on_input, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/prepare_finalize.py
vllm/model_executor/layers/fused_moe/prepare_finalize.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm.model_executor.layers.fused_moe.config import FusedMoEQuantConfig from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceContiguous, TopKWeightAndReduceDelegate, ) from vllm.model_executor.layers.fused_moe.utils import moe_kernel_quantize_input class MoEPrepareAndFinalizeNoEP(mk.FusedMoEPrepareAndFinalize): def __init__(self, defer_input_quant: bool = False) -> None: super().__init__() self.defer_input_quant = defer_input_quant @property def activation_format(self) -> mk.FusedMoEActivationFormat: return mk.FusedMoEActivationFormat.Standard def max_num_tokens_per_rank(self) -> int | None: return None def topk_indices_dtype(self) -> torch.dtype | None: return None def num_dispatchers(self) -> int: return 1 def output_is_reduced(self) -> bool: return False def prepare( self, a1: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_experts: int, expert_map: torch.Tensor | None, apply_router_weight_on_input: bool, quant_config: FusedMoEQuantConfig, ) -> mk.PrepareResultType: if apply_router_weight_on_input: topk = topk_ids.size(1) # TODO: this only works for topK=1, will need to update for topK>1 assert topk == 1, ( "apply_router_weight_on_input is only implemented for topk=1" ) # Note: do not use inplace for shared experts overlap a1 = a1 * topk_weights.to(a1.dtype) # Defer input quant to moe kernel for backends (e.g. AITER, FI) # which use a single kernel call for quant + experts. if self.defer_input_quant: return a1, None, None, None, None a1q, a1q_scale = moe_kernel_quantize_input( a1, quant_config.a1_scale, quant_config.quant_dtype, quant_config.per_act_token_quant, quant_config.block_shape, ) return a1q, a1q_scale, None, None, None def finalize( self, output: torch.Tensor, fused_expert_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, apply_router_weight_on_input: bool, weight_and_reduce_impl: mk.TopKWeightAndReduce, ) -> None: if isinstance(weight_and_reduce_impl, TopKWeightAndReduceDelegate): weight_and_reduce_impl = TopKWeightAndReduceContiguous() weight_and_reduce_impl.apply( output=output, fused_expert_output=fused_expert_output, topk_weights=topk_weights, topk_ids=topk_ids, apply_router_weight_on_input=apply_router_weight_on_input, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/moe_permute_unpermute.py
vllm/model_executor/layers/fused_moe/moe_permute_unpermute.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm import _custom_ops as ops from vllm.model_executor.layers.fused_moe.moe_align_block_size import ( moe_align_block_size, ) from vllm.model_executor.layers.fused_moe.utils import _fp8_perm def _moe_permute( curr_hidden_states: torch.Tensor, a1q_scale: torch.Tensor | None, curr_topk_ids: torch.Tensor, global_num_experts: int, expert_map: torch.Tensor | None, block_m: int, ) -> tuple[torch.Tensor, torch.Tensor | None, torch.Tensor, torch.Tensor, torch.Tensor]: """ Determine the sorted_token_ids, expert_ids for the given problem size. Permute the hidden states and scales according to `sorted_token_ids`. """ top_k_num = curr_topk_ids.size(1) tokens_in_chunk = curr_hidden_states.size(0) sorted_token_ids, expert_ids, num_tokens_post_padded = moe_align_block_size( curr_topk_ids, block_m, global_num_experts, expert_map, pad_sorted_ids=True ) inv_perm: torch.Tensor | None = None num_tokens = top_k_num * tokens_in_chunk expert_ids = torch.repeat_interleave(expert_ids, block_m, dim=0) inv_perm = torch.argsort(sorted_token_ids)[:num_tokens] # Permute according to sorted token ids. sorted_token_ids = sorted_token_ids.clamp(max=num_tokens - 1) curr_hidden_states = _fp8_perm(curr_hidden_states, sorted_token_ids // top_k_num) if a1q_scale is not None: a1q_scale = a1q_scale[sorted_token_ids // top_k_num] return (curr_hidden_states, a1q_scale, sorted_token_ids, expert_ids, inv_perm) def _moe_unpermute_and_reduce( out: torch.Tensor, curr_hidden: torch.Tensor, inv_perm: torch.Tensor | None, topk_weight: torch.Tensor, apply_router_weight_on_input: bool, ) -> None: """ Unpermute the final result and apply topk_weights, then perform the final reduction on the hidden states. """ M, topk = topk_weight.size() K = curr_hidden.size(-1) if inv_perm is not None: curr_hidden = curr_hidden[inv_perm, ...] curr_hidden = curr_hidden.view(-1, topk, K) if not apply_router_weight_on_input: curr_hidden.mul_(topk_weight.view(M, -1, 1)) ops.moe_sum(curr_hidden, out) def moe_permute( hidden_states: torch.Tensor, a1q_scale: torch.Tensor | None, topk_ids: torch.Tensor, n_expert: int, n_local_expert: int = -1, expert_map: torch.Tensor | None = None, align_block_size: int | None = None, fill_invalid_expert: int = -1, permuted_hidden_states: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None, torch.Tensor, torch.Tensor, torch.Tensor]: """ This function expands and permutes activation to gather uncontinuous tokens for each expert. Parameters: - hidden_states (torch.Tensor): The input tensor to the MoE layer. - a1q_scale (Optional[torch.Tensor]): quant scale for hidden_states - topk_ids (torch.Tensor): topk expert route id for each token. - n_expert (int): The number of expert. - n_local_expert (int): The number of expert in current EP rank. - expert_map (Optional[torch.Tensor]): A tensor mapping expert indices from the global expert space to the local expert space of the expert parallel shard. - align_block_size (Optional[int]): align group gemm block size for deepgemm - fill_invalid_expert(int): fill expert id in m_indices for invalid expert to workaround DeepGemm unsupported -1 in m_indices - permuted_hidden_states (Optional[torch.Tensor]): Optional output tensor. If None, the output tensor will be created in this function. Returns: - permuted_hidden_states (torch.Tensor): permuted activation. - a1q_scale (Optional[torch.Tensor]): permuted quant scale for hidden_states if original scale not per-tensor scaling - expert_first_token_offset (torch.Tensor): offset of the first token of each expert for standard grouped gemm. if enable 'align_block_size' expert_first_token_offset will align up to 'align_block_size'. - inv_permuted_idx (torch.Tensor): idx map for moe_unpermute. - permuted_idx (torch.Tensor): idx map from hidden to permuted_hidden. - m_indices: m_indices for grouped gemm in deepgemm,`m_indices[i]` records the group which the j-th row of the LHS belong to.` """ n_token, n_hidden = hidden_states.size() topk = topk_ids.size(1) assert (n_hidden * hidden_states.element_size()) % 16 == 0, ( "permue kernel need hidden dim align to 16B" ) permuted_row_size = n_token * topk if align_block_size is not None: permuted_row_size = ( ( permuted_row_size + n_expert * (align_block_size - 1) + align_block_size - 1 ) // align_block_size * align_block_size ) if n_local_expert == -1: n_local_expert = n_expert if permuted_hidden_states is None: permuted_hidden_states = torch.empty( (permuted_row_size, n_hidden), dtype=hidden_states.dtype, device=hidden_states.device, ) assert permuted_hidden_states.size() == (permuted_row_size, n_hidden), ( f"Expected permuted hidden states to be {(permuted_row_size, n_hidden)}" f" but got {permuted_hidden_states.size()}" ) token_expert_indices = torch.arange( 0, n_token * topk, dtype=torch.int32, device=hidden_states.device ).reshape((n_token, topk)) m_indices = torch.full( (permuted_row_size,), fill_invalid_expert, dtype=torch.int32, device=hidden_states.device, ) expert_first_token_offset = torch.empty( n_local_expert + 1, dtype=torch.int64, device=hidden_states.device ) permuted_idx = torch.full( (permuted_row_size,), n_token * topk, dtype=torch.int32, device=hidden_states.device, ) inv_permuted_idx = torch.empty( (n_token, topk), dtype=torch.int32, device=hidden_states.device ) topk_ids = topk_ids.to(torch.int32) torch.ops._moe_C.moe_permute( hidden_states, topk_ids, token_expert_indices, expert_map, n_expert, n_local_expert, topk, align_block_size, permuted_hidden_states, expert_first_token_offset, inv_permuted_idx, permuted_idx, m_indices, ) if a1q_scale is not None and a1q_scale.dim() > 1: a1q_scale = a1q_scale[permuted_idx.clamp(max=n_token * topk - 1) // topk] return ( permuted_hidden_states, a1q_scale, expert_first_token_offset, inv_permuted_idx.flatten(), m_indices, ) def moe_unpermute( out: torch.Tensor, permuted_hidden_states: torch.Tensor, topk_weights: torch.Tensor, inv_permuted_idx: torch.Tensor, expert_first_token_offset: torch.Tensor | None = None, ) -> None: """ This function expands and permutes activation to gathering uncontinuous tokens for each expert. Parameters: - out (torch.Tensor): output tensor - permuted_hidden_states (torch.Tensor): permuted activation. - topk_weights (torch.Tensor): topk expert route weight for each token. - inv_permuted_idx (torch.Tensor): row idx map for moe_unpermute. - expert_first_token_offset (Optional[torch.Tensor]): offset of the first token of each expert for grouped gemm. Returns: - hidden_states (torch.Tensor): The reduced and unpermuted activation tensor. """ topk = topk_weights.size(1) n_hidden = permuted_hidden_states.size(-1) assert (n_hidden * permuted_hidden_states.element_size()) % 16 == 0, ( "unpermue kernel need hidden dim align to 16B" ) torch.ops._moe_C.moe_unpermute( permuted_hidden_states, topk_weights, inv_permuted_idx, expert_first_token_offset, topk, out, ) def moe_permute_unpermute_supported(): return torch.ops._moe_C.moe_permute_unpermute_supported()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/deepep_ht_prepare_finalize.py
vllm/model_executor/layers/fused_moe/deepep_ht_prepare_finalize.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable import deep_ep import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm.model_executor.layers.fused_moe.config import FusedMoEQuantConfig from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceContiguous, TopKWeightAndReduceDelegate, ) from vllm.model_executor.layers.fused_moe.utils import moe_kernel_quantize_input from vllm.utils.math_utils import round_up from vllm.v1.worker.ubatching import ( dbo_current_ubatch_id, dbo_enabled, dbo_get_previous_event, dbo_switch_to_comm, dbo_switch_to_compute, dbo_switch_to_compute_sync, dbo_yield_and_switch_from_comm_to_compute, dbo_yield_and_switch_from_compute_to_comm, ) class DeepEPHTPrepareAndFinalize(mk.FusedMoEPrepareAndFinalize): """ Prepare/Finalize using DeepEP High-Throughput kernels. """ @staticmethod def maybe_roundup_layer_hidden_size(hidden_size: int, dtype: torch.dtype) -> int: # Round up hidden size so it is compatible with DeepEP High Throughput # kernels. # DeepEP intranode kernels make copies in units of, # 32(warp-size) int4 elements. Round up hidden size to respect this. # For example, an input hidden size of 2880 with dtype torch.bfloat16 # will be rounded up to 3072. hidden_size_bytes = hidden_size * dtype.itemsize xfer_atom_size = 512 # 32 * 16 (size(int4)) if hidden_size_bytes % xfer_atom_size == 0: return hidden_size hidden_size_bytes = round_up(hidden_size_bytes, xfer_atom_size) return hidden_size_bytes // dtype.itemsize def __init__( self, buffer: deep_ep.Buffer, num_dispatchers: int, dp_size: int, rank_expert_offset: int, ): super().__init__() self.buffer = buffer self.num_dispatchers_ = num_dispatchers self.dp_size = dp_size self.rank_expert_offset = rank_expert_offset self.async_prepare = True # The dispatch function returns a handle that the combine function # requires. Under DBO microbatching we must track one handle per # micro-batch to avoid races between threads. self.handles = [None, None] # From https://github.com/deepseek-ai/DeepEP/blob/9fe9021f29c9083cd1808ab36b740208524d9f63/deep_ep/buffer.py#L164 self.available_rank_configs = [2, 4, 8, 16, 24, 32, 64, 128, 144, 160] def num_dispatchers(self) -> int: return self.num_dispatchers_ def output_is_reduced(self) -> bool: return True @property def activation_format(self) -> mk.FusedMoEActivationFormat: return mk.FusedMoEActivationFormat.Standard def max_num_tokens_per_rank(self) -> int | None: return None def topk_indices_dtype(self) -> torch.dtype | None: return torch.int64 def _get_dispatch_config(self) -> deep_ep.Config | None: if self.num_dispatchers_ not in self.available_rank_configs: return None return deep_ep.Buffer.get_dispatch_config(self.num_dispatchers_) def _get_combine_config(self) -> deep_ep.Config | None: if self.num_dispatchers_ not in self.available_rank_configs: return None return deep_ep.Buffer.get_combine_config(self.num_dispatchers_) def _do_dispatch( self, tokens: torch.Tensor, token_scales: torch.Tensor | None, rank_topk_ids: torch.Tensor, rank_topk_weights: torch.Tensor, num_experts: int, a1_scale: torch.Tensor | None, quant_config: FusedMoEQuantConfig, ) -> Callable: has_scales = token_scales is not None # We yield before launching the dispatch kernel since the dispatch # kernel will block the CPU so we want to queue up all the compute # for the other ubatch before the dispatch kernel starts. dbo_yield_and_switch_from_compute_to_comm() # capture a DeepEP event and pass it as previous_event so # DeepEP honors the dependency internally. previous_event = dbo_get_previous_event(self.buffer.capture) ( num_tokens_per_rank, num_tokens_per_rdma_rank, dispatch_expert_num_tokens, is_token_in_rank, event, ) = self.buffer.get_dispatch_layout( topk_idx=rank_topk_ids, num_experts=num_experts, previous_event=previous_event, async_finish=False, allocate_on_comm_stream=False, ) token_data = tokens if has_scales: token_data = (tokens, token_scales) ( token_data, expert_topk_ids, expert_topk_weights, expert_num_tokens_per_expert_list, handle, event, ) = self.buffer.dispatch( x=token_data, handle=None, num_tokens_per_rank=num_tokens_per_rank, num_tokens_per_rdma_rank=num_tokens_per_rdma_rank, is_token_in_rank=is_token_in_rank, num_tokens_per_expert=dispatch_expert_num_tokens, topk_idx=rank_topk_ids, topk_weights=rank_topk_weights, # expert_alignment rounds the number of tokens per expert # to this value. expert_alignment=1, config=self._get_dispatch_config(), previous_event=previous_event, async_finish=self.async_prepare and not dbo_enabled(), allocate_on_comm_stream=False, ) # record the handle for this ubatch a2a_idx = dbo_current_ubatch_id() self.handles[a2a_idx] = handle dbo_switch_to_compute_sync() return lambda: self._receiver( event, has_scales, token_data, expert_topk_ids, num_experts, expert_num_tokens_per_expert_list, expert_topk_weights, a1_scale, quant_config, ) def _receiver( self, event: deep_ep.EventOverlap, has_scales: bool, token_data: tuple[torch.Tensor, torch.Tensor] | torch.Tensor, expert_topk_ids: torch.Tensor | None, num_experts: int, expert_num_tokens_per_expert_list: list[int], expert_topk_weights: torch.Tensor | None, a1_scale: torch.Tensor | None, quant_config: FusedMoEQuantConfig, ) -> mk.PrepareResultType: if event.event is not None: event.current_stream_wait() if has_scales: expert_x, expert_x_scale = token_data else: expert_x, expert_x_scale = token_data, None # The existing MOE kernels assume that all entries of topk_ids are # valid. To that effect, set the -1s in expert_topk_ids to some expert # outside this rank so the expert_map can remap it to -1 when safe. # With Expert Parallel, the experts are divided amongst the rank # sequentially. For rank 0, set it to num_experts - 1 and for all other # ranks set it to 0 as we know that expert_map will have a -1 in those # regions for those ranks. # # DeepEP's topk_ids output refers to the local experts directly. Offset # the topk_ids to move it back to the global experts space so it aligns # with existing vLLM interfaces. assert expert_topk_ids is not None expert_topk_ids = torch.where( expert_topk_ids == -1, num_experts - 1 if self.rank_expert_offset == 0 else 0, expert_topk_ids + self.rank_expert_offset, ) # Makes a GPU-CPU copy. # TODO (varun): Maybe it is better to re-compute the expert_num_tokens # on GPU. expert_tokens_meta = mk.ExpertTokensMetadata.make_from_list( expert_num_tokens_per_expert_list, device=expert_x.device ) # Dispatch and Quant # DeepEP kernels only support dispatching block-quantized # activation scales. # Dispatch in bfloat16 and quantize afterwards if not quant_config.is_block_quantized: # Quantize after dispatch. expert_x_scale = None if expert_x.numel() != 0: expert_x, expert_x_scale = moe_kernel_quantize_input( expert_x, a1_scale, quant_dtype=quant_config.quant_dtype, per_act_token_quant=False, block_shape=quant_config.block_shape, ) return ( expert_x, expert_x_scale, expert_tokens_meta, expert_topk_ids, expert_topk_weights, ) def supports_async(self) -> bool: return True def prepare_async( self, a1: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_experts: int, expert_map: torch.Tensor | None, apply_router_weight_on_input: bool, quant_config: FusedMoEQuantConfig, ) -> mk.ReceiverType: if apply_router_weight_on_input: topk = topk_ids.size(1) # TODO: this only works for topK=1, will need to update for topK>1 assert topk == 1, ( "apply_router_weight_on_input is only implemented for topk=1" ) a1 = a1 * topk_weights.to(a1.dtype) if quant_config.is_block_quantized: # Quant and Dispatch a1q, a1q_scale = moe_kernel_quantize_input( a1, quant_config.a1_scale, quant_dtype=quant_config.quant_dtype, per_act_token_quant=quant_config.per_act_token_quant, block_shape=quant_config.block_shape, ) if a1q_scale is not None and a1q_scale.numel() == 1: a1q_scale = a1q_scale.view(1, 1) a1_post_scale = None else: a1q = a1 a1q_scale = None a1_post_scale = quant_config.a1_scale return self._do_dispatch( tokens=a1q, token_scales=a1q_scale, rank_topk_ids=topk_ids, rank_topk_weights=topk_weights, num_experts=num_experts, a1_scale=a1_post_scale, quant_config=quant_config, ) def prepare( self, a1: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_experts: int, expert_map: torch.Tensor | None, apply_router_weight_on_input: bool, quant_config: FusedMoEQuantConfig, ) -> mk.PrepareResultType: receiver = self.prepare_async( a1, topk_weights, topk_ids, num_experts, expert_map, apply_router_weight_on_input, quant_config, ) return receiver() def _finalize( self, output: torch.Tensor, fused_expert_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, apply_router_weight_on_input: bool, weight_and_reduce_impl: mk.TopKWeightAndReduce, do_async: bool, ) -> Callable | None: a2a_idx = dbo_current_ubatch_id() handle = self.handles[a2a_idx] assert handle is not None # fused_expert_output can have 0 tokens - This happens when none of the # tokens from the all2all reach this EP rank. if fused_expert_output.numel() != 0: if isinstance(weight_and_reduce_impl, TopKWeightAndReduceDelegate): weight_and_reduce_impl = TopKWeightAndReduceContiguous() fused_expert_output = weight_and_reduce_impl.apply( output=None, fused_expert_output=fused_expert_output, topk_weights=topk_weights, topk_ids=topk_ids, apply_router_weight_on_input=apply_router_weight_on_input, ) dbo_yield_and_switch_from_compute_to_comm() assert fused_expert_output.dtype == torch.bfloat16, ( f"Expected fused_expert_output bfloat16, got {fused_expert_output.dtype}" ) previous_event = dbo_get_previous_event(self.buffer.capture) combined_x, _, event = self.buffer.combine( # HT combine only supports BF16 x=fused_expert_output, handle=handle, topk_weights=None, config=self._get_combine_config(), previous_event=previous_event, async_finish=do_async and not dbo_enabled(), allocate_on_comm_stream=False, ) dbo_switch_to_compute() if do_async: def _receiver(): if event.event is not None: event.current_stream_wait() dbo_switch_to_comm() # Respect inplace outputs. output.copy_(combined_x, non_blocking=True) # TODO(lucas): refactor the modular kernel so this will be # handled there dbo_yield_and_switch_from_comm_to_compute() return _receiver else: # TODO(lucas): support this case with the refactored modular kernel assert not dbo_enabled() # Respect inplace outputs. output.copy_(combined_x, non_blocking=True) return None def finalize_async( self, output: torch.Tensor, fused_expert_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, apply_router_weight_on_input: bool, weight_and_reduce_impl: mk.TopKWeightAndReduce, ) -> Callable: receiver = self._finalize( output, fused_expert_output, topk_weights, topk_ids, apply_router_weight_on_input, weight_and_reduce_impl, True, ) assert receiver is not None return receiver def finalize( self, output: torch.Tensor, fused_expert_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, apply_router_weight_on_input: bool, weight_and_reduce_impl: mk.TopKWeightAndReduce, ) -> None: self._finalize( output, fused_expert_output, topk_weights, topk_ids, apply_router_weight_on_input, weight_and_reduce_impl, False, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/__init__.py
vllm/model_executor/layers/fused_moe/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from contextlib import contextmanager from typing import Any from vllm.model_executor.layers.fused_moe.config import ( FusedMoEConfig, RoutingMethodType, ) from vllm.model_executor.layers.fused_moe.fused_moe_method_base import ( FusedMoEMethodBase, ) from vllm.model_executor.layers.fused_moe.layer import ( FusedMoE, FusedMoeWeightScaleSupported, ) from vllm.model_executor.layers.fused_moe.modular_kernel import ( FusedMoEActivationFormat, FusedMoEPermuteExpertsUnpermute, FusedMoEPrepareAndFinalize, ) from vllm.model_executor.layers.fused_moe.shared_fused_moe import SharedFusedMoE from vllm.model_executor.layers.fused_moe.unquantized_fused_moe_method import ( UnquantizedFusedMoEMethod, ) from vllm.model_executor.layers.fused_moe.utils import activation_without_mul from vllm.model_executor.layers.fused_moe.zero_expert_fused_moe import ( ZeroExpertFusedMoE, ) from vllm.triton_utils import HAS_TRITON _config: dict[str, Any] | None = None @contextmanager def override_config(config): global _config old_config = _config _config = config yield _config = old_config def get_config() -> dict[str, Any] | None: return _config __all__ = [ "FusedMoE", "FusedMoEConfig", "FusedMoEMethodBase", "UnquantizedFusedMoEMethod", "FusedMoeWeightScaleSupported", "FusedMoEPermuteExpertsUnpermute", "FusedMoEActivationFormat", "FusedMoEPrepareAndFinalize", "RoutingMethodType", "SharedFusedMoE", "ZeroExpertFusedMoE", "activation_without_mul", "override_config", "get_config", ] if HAS_TRITON: # import to register the custom ops from vllm.model_executor.layers.fused_moe.batched_deep_gemm_moe import ( BatchedDeepGemmExperts, ) from vllm.model_executor.layers.fused_moe.cutlass_moe import ( CutlassBatchedExpertsFp8, CutlassExpertsFp8, CutlassExpertsW4A8Fp8, cutlass_moe_fp4, cutlass_moe_fp8, cutlass_moe_w4a8_fp8, ) from vllm.model_executor.layers.fused_moe.deep_gemm_moe import DeepGemmExperts from vllm.model_executor.layers.fused_moe.fused_batched_moe import ( BatchedTritonExperts, ) from vllm.model_executor.layers.fused_moe.fused_moe import ( GroupedTopk, TritonExperts, fused_experts, fused_topk, get_config_file_name, ) from vllm.model_executor.layers.fused_moe.triton_deep_gemm_moe import ( TritonOrDeepGemmExperts, ) __all__ += [ "fused_topk", "fused_experts", "get_config_file_name", "GroupedTopk", "cutlass_moe_fp8", "cutlass_moe_fp4", "cutlass_moe_w4a8_fp8", "CutlassExpertsFp8", "CutlassBatchedExpertsFp8", "CutlassExpertsW4A8Fp8", "TritonExperts", "BatchedTritonExperts", "DeepGemmExperts", "BatchedDeepGemmExperts", "TritonOrDeepGemmExperts", ] else: # Some model classes directly use the custom ops. Add placeholders # to avoid import errors. def _raise_exception(method: str): raise NotImplementedError(f"{method} is not implemented as lack of triton.") fused_topk = lambda *args, **kwargs: _raise_exception("fused_topk") fused_experts = lambda *args, **kwargs: _raise_exception("fused_experts")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/routing_simulator.py
vllm/model_executor/layers/fused_moe/routing_simulator.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Token-to-Expert Routing Simulator This module provides a framework for simulating and testing different token-to-expert routing strategies for Mixture of Experts (MoE) models. It supports routing logic customization and includes example implementations like uniform random routing. """ from abc import ABC, abstractmethod from typing import Any import torch from vllm.logger import init_logger logger = init_logger(__name__) class RoutingStrategy(ABC): """Base class for token-to-expert routing strategies.""" @abstractmethod def route_tokens( self, hidden_states: torch.Tensor, router_logits: torch.Tensor, top_k: int, indices_type: torch.dtype | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: """ Route tokens to experts. Args: hidden_states: Input hidden states [num_tokens, hidden_size] router_logits: Router logits [num_tokens, num_experts] top_k: Number of experts to select per token indices_type: Data type for expert indices Returns: tuple of (topk_weights, topk_ids) """ pass class DistributionBasedRouting(RoutingStrategy): """ Distribution-based random routing strategy with configurable distributions. This routing strategy randomly selects experts for each token based on different probability distributions. Currently supports uniform and normal distributions for testing different routing patterns. """ def __init__(self, distribution: str = "uniform", **distribution_params: Any): """ Initialize distribution-based routing. Args: distribution: Type of distribution to use for sampling - "uniform": Uniform distribution (default) - "normal": Normal/Gaussian distribution **distribution_params: Parameters specific to the chosen distribution For "uniform": No additional parameters needed For "normal": mean (default: 0.0), std (default: 1.0) """ self.distribution = distribution.lower() self.distribution_params = distribution_params # Validate distribution and parameters self._validate_distribution_params() def _validate_distribution_params(self): """Validate distribution type and parameters.""" valid_distributions = ["uniform", "normal"] if self.distribution not in valid_distributions: raise ValueError( f"Unsupported distribution: {self.distribution}. " f"Supported distributions: {valid_distributions}" ) # Set default parameters if not provided if self.distribution == "normal": self.distribution_params.setdefault("mean", 0.0) self.distribution_params.setdefault("std", 1.0) def route_tokens( self, hidden_states: torch.Tensor, router_logits: torch.Tensor, top_k: int, indices_type: torch.dtype | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: """ Randomly select experts for each token using the specified distribution. Args: hidden_states: Input hidden states [num_tokens, hidden_size] router_logits: Router logits [num_tokens, num_experts] top_k: Number of experts to select per token indices_type: Data type for expert indices Returns: tuple of (topk_weights, topk_ids) where: - topk_weights: Weights based on distribution sampling - topk_ids: Expert indices sampled from the distribution """ num_tokens = hidden_states.shape[0] num_experts = router_logits.shape[-1] if indices_type is None: indices_type = torch.long # Generate expert IDs based on the specified distribution topk_ids = self._sample_expert_ids( num_tokens, num_experts, top_k, hidden_states.device, indices_type ) # Generate weights based on the distribution topk_weights = self._generate_weights(num_tokens, top_k, hidden_states.device) return topk_weights, topk_ids def _sample_expert_ids( self, num_tokens: int, num_experts: int, top_k: int, device: torch.device, indices_type: torch.dtype, ) -> torch.Tensor: """Sample expert IDs based on the specified distribution.""" if self.distribution == "uniform": # Uniform random sampling return torch.randint( low=0, high=num_experts, size=(num_tokens, top_k), dtype=indices_type, device=device, ) elif self.distribution == "normal": # For normal distribution, sample continuous values and map to # expert IDs continuous_samples = self._sample_continuous_distribution( num_tokens, top_k, device ) # Map continuous samples to expert indices # Normalize to [0, 1] range and scale to [0, num_experts) normalized_samples = self._normalize_samples(continuous_samples) expert_ids = (normalized_samples * num_experts).long() expert_ids = torch.clamp(expert_ids, 0, num_experts - 1) return expert_ids.to(dtype=indices_type) else: raise ValueError(f"Unsupported distribution: {self.distribution}") def _sample_continuous_distribution( self, num_tokens: int, top_k: int, device: torch.device ) -> torch.Tensor: """Sample from continuous distributions.""" shape = (num_tokens, top_k) if self.distribution == "normal": mean = self.distribution_params["mean"] std = self.distribution_params["std"] return torch.normal(mean, std, size=shape, device=device) else: raise ValueError( f"Unsupported continuous distribution: {self.distribution}" ) def _normalize_samples(self, samples: torch.Tensor) -> torch.Tensor: """Normalize samples to [0, 1] range.""" if self.distribution == "normal": # Use sigmoid to map normal distribution to [0, 1] return torch.sigmoid(samples) else: raise ValueError( f"Unsupported distribution for normalization: {self.distribution}" ) def _generate_weights( self, num_tokens: int, top_k: int, device: torch.device ) -> torch.Tensor: """Generate weights based on the distribution.""" if self.distribution == "uniform": # All-ones weights for uniform distribution return torch.ones( (num_tokens, top_k), dtype=torch.float32, device=device, ) elif self.distribution == "normal": # For normal distribution, generate weights from the same # distribution continuous_weights = self._sample_continuous_distribution( num_tokens, top_k, device ) # Normalize to positive values and sum to 1 weights = torch.abs(continuous_weights) weights = weights / weights.sum(dim=-1, keepdim=True) return weights else: raise ValueError( f"Unsupported distribution for weight generation: {self.distribution}" ) def get_distribution_info(self) -> dict: """Get information about the current distribution configuration.""" return { "distribution": self.distribution, "parameters": self.distribution_params.copy(), } class RoutingSimulator: """ Token-to-Expert Routing Simulator. This class provides a framework for testing and comparing different routing strategies for MoE models. It can simulate routing behavior and collect statistics for analysis. """ # Class-level registry of routing strategies _routing_strategies: dict[str, RoutingStrategy] = { # Basic routing strategies "uniform_random": DistributionBasedRouting( distribution="uniform", mean=0.0, std=1.0 ), "normal_routing": DistributionBasedRouting( distribution="normal", mean=0.0, std=1.0 ), } @classmethod def register_strategy(cls, name: str, strategy: RoutingStrategy): """ Register a custom routing strategy. Args: name: Name of the strategy strategy: RoutingStrategy instance """ cls._routing_strategies[name] = strategy @classmethod def get_available_strategies(cls) -> list[str]: """ Get list of available routing strategy names. Returns: List of available strategy names """ return list(cls._routing_strategies.keys()) @staticmethod def simulate_routing( hidden_states: torch.Tensor, router_logits: torch.Tensor, strategy_name: str, top_k: int, indices_type: torch.dtype | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: """ Simulate token-to-expert routing using the specified strategy. Args: hidden_states: Input hidden states [num_tokens, hidden_size] router_logits: Router logits [num_tokens, num_experts] strategy_name: Name of the routing strategy to use top_k: Number of experts to select per token indices_type: Data type for expert indices Returns: tuple of (topk_weights, topk_ids) """ if strategy_name not in RoutingSimulator._routing_strategies: raise ValueError( f"Unknown routing strategy: {strategy_name}. " f"Available strategies: " f"{list(RoutingSimulator._routing_strategies.keys())}" ) logger.warning_once( "Simulating MoE routing using a %s strategy. " "This should only be used for performance testing. " "Model outputs will not be valid.", strategy_name, ) strategy = RoutingSimulator._routing_strategies[strategy_name] return strategy.route_tokens( hidden_states=hidden_states, router_logits=router_logits, top_k=top_k, indices_type=indices_type, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/triton_deep_gemm_moe.py
vllm/model_executor/layers/fused_moe/triton_deep_gemm_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm.model_executor.layers.fused_moe.config import FusedMoEQuantConfig from vllm.model_executor.layers.fused_moe.deep_gemm_moe import ( DeepGemmExperts, _valid_deep_gemm, _valid_deep_gemm_shape, ) from vllm.model_executor.layers.fused_moe.fused_moe import TritonExperts from vllm.utils.deep_gemm import ( get_mk_alignment_for_contiguous_layout, is_deep_gemm_e8m0_used, ) class TritonOrDeepGemmExperts(mk.FusedMoEPermuteExpertsUnpermute): def __init__( self, quant_config: FusedMoEQuantConfig, allow_deep_gemm: bool = False, ): super().__init__(quant_config) self.triton_expert = TritonExperts(quant_config) self.allow_deep_gemm = ( allow_deep_gemm and self.quant_config.use_fp8_w8a8 and self.block_shape == get_mk_alignment_for_contiguous_layout() ) self.deep_gemm_expert = ( DeepGemmExperts(self.quant_config) if self.allow_deep_gemm else None ) @property def activation_formats( self, ) -> tuple[mk.FusedMoEActivationFormat, mk.FusedMoEActivationFormat]: assert ( self.deep_gemm_expert is None or self.triton_expert.activation_formats == self.deep_gemm_expert.activation_formats ) return self.triton_expert.activation_formats def supports_chunking(self) -> bool: dge = self.deep_gemm_expert te = self.triton_expert return (dge is None or dge.supports_chunking()) and ( te is None or te.supports_chunking() ) def supports_expert_map(self) -> bool: dge = self.deep_gemm_expert te = self.triton_expert return (dge is None or dge.supports_expert_map()) and ( te is None or te.supports_expert_map() ) def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: dge = self.deep_gemm_expert te = self.triton_expert dge_war = dge.finalize_weight_and_reduce_impl() if dge else None te_war = te.finalize_weight_and_reduce_impl() if te else None is_dge_war = dge_war is not None is_te_war = te_war is not None if is_dge_war and is_te_war: assert dge_war == te_war, ( "Both implementations should agree on WeightAndReduce impls. " f"Got dge_war: {dge_war}, and te_war: {te_war}" ) if dge_war is not None: return dge_war assert te_war is not None return te_war def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: # Note: the deep gemm workspaces are strictly larger than the triton # workspaces so we can be pessimistic here and allocate for DeepGemm # even if we fall back to triton later, e.g. if expert maps are set. if self.allow_deep_gemm and ( is_deep_gemm_e8m0_used() or _valid_deep_gemm_shape(M, N, K) ): assert self.deep_gemm_expert is not None return self.deep_gemm_expert.workspace_shapes( M, N, K, topk, global_num_experts, local_num_experts, expert_tokens_meta, ) else: return self.triton_expert.workspace_shapes( M, N, K, topk, global_num_experts, local_num_experts, expert_tokens_meta, ) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: str, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ): use_deep_gemm = self.allow_deep_gemm and ( is_deep_gemm_e8m0_used() or _valid_deep_gemm(hidden_states, w1, w2) ) experts = self.deep_gemm_expert if use_deep_gemm else self.triton_expert assert experts is not None experts.apply( output, hidden_states, w1, w2, topk_weights, topk_ids, activation, global_num_experts, expert_map, a1q_scale, a2_scale, workspace13, workspace2, expert_tokens_meta, apply_router_weight_on_input, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/fused_marlin_moe.py
vllm/model_executor/layers/fused_moe/fused_marlin_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Fused MoE utilities for GPTQ.""" from collections.abc import Callable import torch import vllm._custom_ops as ops import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm.model_executor.layers.fused_moe.config import FusedMoEQuantConfig from vllm.model_executor.layers.fused_moe.moe_align_block_size import ( batched_moe_align_block_size, moe_align_block_size, ) from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceDelegate, TopKWeightAndReduceNoOP, ) from vllm.model_executor.layers.fused_moe.utils import _resize_cache, disable_inplace from vllm.model_executor.layers.quantization.utils.marlin_utils import ( marlin_make_workspace_new, marlin_moe_intermediate_size, marlin_quant_input, ) from vllm.platforms import current_platform from vllm.scalar_type import ScalarType, scalar_types def default_activation_func( activation: str, output: torch.Tensor, input: torch.Tensor ) -> None: if activation == "silu": torch.ops._C.silu_and_mul(output, input) elif activation == "swigluoai": # alpha = 1.702, limit = 7.0 torch.ops._C.swigluoai_and_mul(output, input) else: raise ValueError( f"Unsupported activation: {activation}. " "Only silu and swigluoai activations are supported." ) def _fused_marlin_moe( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, bias1: torch.Tensor | None, bias2: torch.Tensor | None, w1_scale: torch.Tensor, w2_scale: torch.Tensor, topk_weights: torch.Tensor, num_topk: int, quant_type: ScalarType, apply_router_weight_on_input: bool, expert_map: torch.Tensor | None, block_size_m: int, sorted_token_ids: torch.Tensor, expert_ids: torch.Tensor, num_tokens_post_padded: torch.Tensor, activation: str = "silu", activation_func: Callable[ [str, torch.Tensor, torch.Tensor], None ] = default_activation_func, input_global_scale1: torch.Tensor | None = None, input_global_scale2: torch.Tensor | None = None, global_scale1: torch.Tensor | None = None, global_scale2: torch.Tensor | None = None, g_idx1: torch.Tensor | None = None, g_idx2: torch.Tensor | None = None, sort_indices1: torch.Tensor | None = None, sort_indices2: torch.Tensor | None = None, w1_zeros: torch.Tensor | None = None, w2_zeros: torch.Tensor | None = None, workspace: torch.Tensor | None = None, intermediate_cache13: torch.Tensor | None = None, intermediate_cache2: torch.Tensor | None = None, output: torch.Tensor | None = None, input_dtype: torch.dtype | None = None, is_k_full: bool = True, ) -> torch.Tensor: assert hidden_states.ndim == 2 M, K = hidden_states.size() N = marlin_moe_intermediate_size(w1, w2) if workspace is None: workspace = marlin_make_workspace_new(hidden_states.device, 4) if intermediate_cache13 is None: intermediate_cache13 = torch.empty( (M * num_topk * max(2 * N, K),), device=hidden_states.device, dtype=hidden_states.dtype, ) if intermediate_cache2 is None: intermediate_cache2 = torch.empty( (M * num_topk, N), device=hidden_states.device, dtype=hidden_states.dtype, ) intermediate_cache1 = _resize_cache(intermediate_cache13, (M * num_topk, 2 * N)) intermediate_cache3 = _resize_cache(intermediate_cache13, (M * num_topk, K)) intermediate_cache2 = _resize_cache(intermediate_cache2, (M * num_topk, N)) a_scales1 = None gate_up_input = hidden_states if input_dtype == torch.int8: gate_up_input, a_scales1 = marlin_quant_input(hidden_states, input_dtype) if input_global_scale1 is not None: a_scales1 = a_scales1 * input_global_scale1 elif input_dtype == torch.float8_e4m3fn: gate_up_input, a_scales1 = marlin_quant_input(hidden_states, input_dtype) intermediate_cache1 = ops.moe_wna16_marlin_gemm( gate_up_input, intermediate_cache1, w1, bias1, w1_scale, a_scales1, global_scale1, w1_zeros, g_idx1, sort_indices1, workspace, sorted_token_ids, expert_ids, num_tokens_post_padded, topk_weights, moe_block_size=block_size_m, top_k=num_topk, mul_topk_weights=apply_router_weight_on_input, is_ep=expert_map is not None, b_q_type=quant_type, size_m=M, size_n=2 * N, size_k=K, is_k_full=is_k_full, use_atomic_add=False, use_fp32_reduce=True, is_zp_float=False, ) activation_func( activation, intermediate_cache2, intermediate_cache1.view(-1, 2 * N) ) if output is None: output = intermediate_cache3 if expert_map is not None: output.zero_() a_scales2 = None if input_dtype == torch.int8: intermediate_cache2, a_scales2 = marlin_quant_input( intermediate_cache2, input_dtype ) if input_global_scale2 is not None: a_scales2 = a_scales2 * input_global_scale2 elif input_dtype == torch.float8_e4m3fn: intermediate_cache2, a_scales2 = marlin_quant_input( intermediate_cache2, input_dtype ) output = ops.moe_wna16_marlin_gemm( intermediate_cache2, output, w2, bias2, w2_scale, a_scales2, global_scale2, w2_zeros, g_idx2, sort_indices2, workspace, sorted_token_ids, expert_ids, num_tokens_post_padded, topk_weights, moe_block_size=block_size_m, top_k=1, mul_topk_weights=not apply_router_weight_on_input, is_ep=expert_map is not None, b_q_type=quant_type, size_m=M * num_topk, size_n=K, size_k=N, is_k_full=is_k_full, use_atomic_add=False, use_fp32_reduce=True, is_zp_float=False, ) return output def fused_marlin_moe( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, bias1: torch.Tensor | None, bias2: torch.Tensor | None, w1_scale: torch.Tensor, w2_scale: torch.Tensor, gating_output: torch.Tensor | None, topk_weights: torch.Tensor, topk_ids: torch.Tensor, quant_type_id: int, apply_router_weight_on_input: bool = False, global_num_experts: int = -1, activation: str = "silu", activation_func: Callable[ [str, torch.Tensor, torch.Tensor], None ] = default_activation_func, moe_sum: Callable[[torch.Tensor, torch.Tensor], None] | None = None, expert_map: torch.Tensor | None = None, input_global_scale1: torch.Tensor | None = None, input_global_scale2: torch.Tensor | None = None, global_scale1: torch.Tensor | None = None, global_scale2: torch.Tensor | None = None, g_idx1: torch.Tensor | None = None, g_idx2: torch.Tensor | None = None, sort_indices1: torch.Tensor | None = None, sort_indices2: torch.Tensor | None = None, w1_zeros: torch.Tensor | None = None, w2_zeros: torch.Tensor | None = None, workspace: torch.Tensor | None = None, intermediate_cache13: torch.Tensor | None = None, intermediate_cache2: torch.Tensor | None = None, is_k_full: bool = True, output: torch.Tensor | None = None, input_dtype: torch.dtype | None = None, inplace: bool = False, ) -> torch.Tensor: """ This function computes a Mixture of Experts (MoE) layer using two sets of weights, w1 and w2, and top-k gating mechanism. Parameters: - hidden_states (torch.Tensor): The input tensor to the MoE layer. - w1 (torch.Tensor): The first set of expert weights. - w2 (torch.Tensor): The second set of expert weights. - w1_scale (torch.Tensor): Scale to be used for w1. - w2_scale (torch.Tensor): Scale to be used for w2. - gating_output (torch.Tensor|None): The output of the gating operation (before softmax). - g_idx1 (torch.Tensor|None): The first set of act_order indices. - g_idx2 (torch.Tensor|None): The second set of act_order indices. - sort_indices1 (torch.Tensor|None): The first act_order input permutation. - sort_indices2 (torch.Tensor|None): The second act_order input permutation. - topk_weights (torch.Tensor): Top-k weights. - topk_ids (torch.Tensor): Indices of topk-k elements. - w1_zeros (torch.Tensor|None): Optional zero points to be used for w1. - w2_zeros (torch.Tensor|None): Optional zero points to be used for w2. - num_bits (bool): The number of bits in expert weights quantization. Returns: - torch.Tensor: The output tensor after applying the MoE layer. """ if inplace: assert output is None, "Conflicting request" quant_type = ScalarType.from_id(quant_type_id) assert quant_type in [ scalar_types.uint4, scalar_types.uint8b128, scalar_types.uint4b8, scalar_types.float8_e4m3fn, scalar_types.float4_e2m1f, ] bit4_scalar_types = [ scalar_types.uint4, scalar_types.uint4b8, scalar_types.float4_e2m1f, ] num_bits = 4 if quant_type in bit4_scalar_types else 8 M, K = hidden_states.size() E = w1.size(0) topk = topk_ids.size(1) # Check constraints. if gating_output is not None: assert gating_output.size(0) == M, "Number of tokens mismatch" assert w1.size(1) * 16 == K, "Hidden size mismatch w1" assert w2.size(2) // (num_bits // 2) == K, "Hidden size mismatch w2" assert hidden_states.is_contiguous(), "Hidden_states must be contiguous" assert w1.is_contiguous(), "Expert weights1 must be contiguous" assert w2.is_contiguous(), "Expert weights2 must be contiguous" assert hidden_states.dtype in [torch.float16, torch.bfloat16] assert num_bits in [4, 8] assert topk_weights.dtype == torch.float32 # M block size selection logic # TODO: tune this further for specific models for block_size_m in [8, 16, 32, 48, 64]: if M * topk / E / block_size_m < 0.9: break if input_dtype is not None and input_dtype.itemsize == 1: block_size_m = max(block_size_m, 16) if global_num_experts == -1: global_num_experts = E sorted_token_ids, expert_ids, num_tokens_post_padded = moe_align_block_size( topk_ids, block_size_m, global_num_experts, expert_map, ignore_invalid_experts=True, ) assert activation is not None moe_output = _fused_marlin_moe( hidden_states=hidden_states, w1=w1, w2=w2, bias1=bias1, bias2=bias2, w1_scale=w1_scale, w2_scale=w2_scale, topk_weights=topk_weights, num_topk=topk, quant_type=quant_type, apply_router_weight_on_input=apply_router_weight_on_input, expert_map=expert_map, block_size_m=block_size_m, sorted_token_ids=sorted_token_ids, expert_ids=expert_ids, num_tokens_post_padded=num_tokens_post_padded, activation=activation, activation_func=activation_func, input_global_scale1=input_global_scale1, input_global_scale2=input_global_scale2, global_scale1=global_scale1, global_scale2=global_scale2, g_idx1=g_idx1, g_idx2=g_idx2, sort_indices1=sort_indices1, sort_indices2=sort_indices2, w1_zeros=w1_zeros, w2_zeros=w2_zeros, workspace=workspace, intermediate_cache13=intermediate_cache13, intermediate_cache2=intermediate_cache2, output=None, input_dtype=input_dtype, is_k_full=is_k_full, ).view(-1, topk, K) if output is None: if inplace and not disable_inplace(): output = hidden_states else: output = torch.empty_like(hidden_states) if moe_sum is None: return torch.sum(moe_output.view(-1, topk, K), dim=1, out=output) else: return moe_sum(moe_output, output) def batched_fused_marlin_moe( hidden_states: torch.Tensor, expert_num_tokens: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, bias1: torch.Tensor | None, bias2: torch.Tensor | None, w1_scale: torch.Tensor, w2_scale: torch.Tensor, gating_output: torch.Tensor | None, quant_type_id: int, apply_router_weight_on_input: bool = False, global_num_experts: int = -1, activation: str | None = "silu", expert_map: torch.Tensor | None = None, global_scale1: torch.Tensor | None = None, global_scale2: torch.Tensor | None = None, g_idx1: torch.Tensor | None = None, g_idx2: torch.Tensor | None = None, sort_indices1: torch.Tensor | None = None, sort_indices2: torch.Tensor | None = None, w1_zeros: torch.Tensor | None = None, w2_zeros: torch.Tensor | None = None, workspace: torch.Tensor | None = None, intermediate_cache13: torch.Tensor | None = None, intermediate_cache2: torch.Tensor | None = None, is_k_full: bool = True, output: torch.Tensor | None = None, inplace: bool = False, ) -> torch.Tensor: """ This function massages the inputs so the batched hidden_states can be presented as a 2D contiguous tensor that could be used with _fused_marlin_moe. Note that both batched_fused_marlin_moe and fused_marlin_moe ultimately use `ops.moe_wna16_marlin_gemm` for the gemm operation and `ops.moe_mna16_marlin_gemm` supports only 2D contiguous hidden_states. Note that the moe_align_block_size function indicates, - What rows of the A matrix (hidden_states) to access during the matmul, via sorted_ids output. - What expert_id to use for each block matmul, via expert_ids ouptut. In the batched version, the tokens are already grouped/batched by experts they subscribe to. Due to this, we can represent the batched hidden_states tensor of shape [B, MAX_TOKENS_PER_BATCH, K] as a 2D tensor of shape, [B * MAX_TOKENS_PER_BATCH, K]. We may treat this a 2D contiguous tensor with topk=1 as each token (row in the tensor) subscribes to exactly one expert_id (which is the batch_id). With the expert_num_tokens tensor, that indicates how many tokens are actually valid in each batch, the batched_moe_align_block_size function constructs the sorted_ids and expert_ids tensors, so only relevant/valid rows of A (hidden_states) are accessed and are processed with the correct expert_ids. """ assert hidden_states.ndim == 3, ( f"hidden states must be batched. e.g. [B, MAX_TOKENS, K]." f"But got {hidden_states.size()}" ) if inplace: assert output is None, "Conflicting request." quant_type = ScalarType.from_id(quant_type_id) assert quant_type in [ scalar_types.uint4, scalar_types.uint8b128, scalar_types.uint4b8, scalar_types.float8_e4m3fn, scalar_types.float4_e2m1f, ] bit4_scalar_types = [ scalar_types.uint4, scalar_types.uint4b8, scalar_types.float4_e2m1f, ] num_bits = 4 if quant_type in bit4_scalar_types else 8 B, BATCH_TOKENS_MAX, K = hidden_states.size() M = hidden_states.view(-1, K).size(0) E = w1.size(0) # Check constraints. assert hidden_states.is_contiguous(), "Hidden_states must be contiguous" assert hidden_states.dtype in [torch.float16, torch.bfloat16] assert expert_num_tokens.size(0) == E assert B == E, ( "Batch must be as big as number of experts as the tokens" "are sorted into the batch/expert they belong to" ) assert w1.size(1) * 16 == K, "Hidden size mismatch w1" assert w2.size(2) // (num_bits // 2) == K, "Hidden size mismatch w2" assert w1.is_contiguous(), "Expert weights1 must be contiguous" assert w2.is_contiguous(), "Expert weights2 must be contiguous" assert num_bits in [4, 8] # Technically, the tokens are already separated by their expert ids. # Hidden-States can just be squeezed to have just 2 dimensions, # [B * MAX_TOKENS, K] and top_k can be interpreted as just 1. topk = 1 # TODO(varun) : Choose a decent block size like in fused_marlin_moe block_size_m = 64 sorted_token_ids, expert_ids, num_tokens_post_padded = batched_moe_align_block_size( max_tokens_per_batch=BATCH_TOKENS_MAX, block_size=block_size_m, expert_num_tokens=expert_num_tokens, ) if output is None and inplace: output = hidden_states # TODO (varun): This can be avoided by plumbing the marlin kernel to # ignore topk_weights when topk_weights_ptr is a nullptr. topk_weights = torch.ones( (M, topk), device=hidden_states.device, dtype=torch.float32 ) assert activation is not None output = _fused_marlin_moe( hidden_states=hidden_states.view(-1, K), w1=w1, w2=w2, bias1=bias1, bias2=bias2, w1_scale=w1_scale, w2_scale=w2_scale, topk_weights=topk_weights, num_topk=topk, quant_type=quant_type, apply_router_weight_on_input=apply_router_weight_on_input, activation=activation, expert_map=expert_map, block_size_m=block_size_m, sorted_token_ids=sorted_token_ids, expert_ids=expert_ids, num_tokens_post_padded=num_tokens_post_padded, global_scale1=global_scale1, global_scale2=global_scale2, g_idx1=g_idx1, g_idx2=g_idx2, sort_indices1=sort_indices1, sort_indices2=sort_indices2, w1_zeros=w1_zeros, w2_zeros=w2_zeros, workspace=workspace, intermediate_cache13=intermediate_cache13, intermediate_cache2=intermediate_cache2, output=output.view(-1, K) if output is not None else output, is_k_full=is_k_full, ) output = output.view(B, BATCH_TOKENS_MAX, K) return output class MarlinExpertsBase(mk.FusedMoEPermuteExpertsUnpermute): def __init__( self, quant_config: FusedMoEQuantConfig, w13_g_idx: torch.Tensor | None = None, w2_g_idx: torch.Tensor | None = None, w13_g_idx_sort_indices: torch.Tensor | None = None, w2_g_idx_sort_indices: torch.Tensor | None = None, is_k_full: bool = True, ): # TODO (varun) : Enable activation quantization assert ( quant_config.use_mxfp4_w4a16 or quant_config.use_int4_w4a16 or quant_config.use_fp8_w8a16 ), "Supports only mxfp4_w4a16, int4_w4a16 or fp8_w8a16" self.w13_g_idx = w13_g_idx self.w2_g_idx = w2_g_idx self.w13_g_idx_sort_indices = w13_g_idx_sort_indices self.w2_g_idx_sort_indices = w2_g_idx_sort_indices self.is_k_full = is_k_full super().__init__(quant_config) @property def quant_type_id(self) -> int: # uint4b8 will be set for int4 weight and float4_e2m1f will be used for mxfp4 if self.quant_config.use_int4_w4a16: return scalar_types.uint4b8.id elif self.quant_config.use_mxfp4_w4a16: return scalar_types.float4_e2m1f.id elif ( self.quant_config.use_fp8_w8a16 and current_platform.fp8_dtype() == torch.float8_e4m3fn ): return scalar_types.float8_e4m3fn.id else: raise NotImplementedError("Unsupported quantization type.") def moe_problem_size( self, a1: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_ids: torch.Tensor, ) -> tuple[int, int, int, int, int]: assert w1.dim() == 3 and w2.dim() == 3 E = w1.size(0) K = a1.size(-1) N = marlin_moe_intermediate_size(w1, w2) if a1.dim() == 2: # Make sure we are using the correct a1 (pre-permute). assert topk_ids.size(0) == a1.size(0), f"{topk_ids.size(0)} != {a1.size(0)}" M = a1.size(0) else: assert a1.dim() == 3 assert a1.size(0) == E, f"{a1.size(0)} == {E}" M = a1.size(1) # This is max_num_tokens assert topk_ids.dim() == 2 topk = topk_ids.size(1) return E, M, N, K, topk class MarlinExperts(MarlinExpertsBase): def __init__( self, quant_config: FusedMoEQuantConfig, w13_g_idx: torch.Tensor | None = None, w2_g_idx: torch.Tensor | None = None, w13_g_idx_sort_indices: torch.Tensor | None = None, w2_g_idx_sort_indices: torch.Tensor | None = None, is_k_full: bool = True, ): super().__init__( quant_config, w13_g_idx, w2_g_idx, w13_g_idx_sort_indices, w2_g_idx_sort_indices, is_k_full, ) def supports_expert_map(self) -> bool: return True def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: return TopKWeightAndReduceNoOP() @property def activation_formats( self, ) -> tuple[mk.FusedMoEActivationFormat, mk.FusedMoEActivationFormat]: return ( mk.FusedMoEActivationFormat.Standard, mk.FusedMoEActivationFormat.Standard, ) def supports_chunking(self) -> bool: return True def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: # Modular Kernel provisions output buffer from workspace1. However in # the fused_marlin_moe() function, the final torch.sum(), is defined # essentially as, # `torch.sum(workspace1, dim=1, out=output)` # Having overlapping input and output tensors for torch.sum seems # error prone and depends on how the torch.sum is implemented. # For this reason we swap let the output buffer provision from # workspace2. # Workspace/IntermediateCache allocation matching fused_marlin_moe() # workspace1 = (M * topk * max(2 * N, K),) # workspace2 = (M * topk, N) # Workspace/IntermediateCache allocation accounting for output buffer # provisioning workspace1 = (M * topk, max(N, K)) workspace2 = (M * topk * max(2 * N, K),) output = (M, K) return (workspace1, workspace2, output) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: str, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ): assert self.w1_scale is not None assert self.w2_scale is not None return fused_marlin_moe( hidden_states=hidden_states, w1=w1, w2=w2, bias1=self.w1_bias, bias2=self.w2_bias, w1_scale=self.w1_scale, w2_scale=self.w2_scale, gating_output=None, topk_weights=topk_weights, topk_ids=topk_ids, quant_type_id=self.quant_type_id, apply_router_weight_on_input=apply_router_weight_on_input, global_num_experts=global_num_experts, activation=activation, activation_func=self.activation, moe_sum=self.moe_sum, expert_map=expert_map, output=output, # Workspaces are swapped in workspace_shapes() to account for proper # output buffer allocation. Please refer to workspace_shapes(). intermediate_cache13=workspace2, intermediate_cache2=workspace13, g_idx1=self.w13_g_idx, g_idx2=self.w2_g_idx, sort_indices1=self.w13_g_idx_sort_indices, sort_indices2=self.w2_g_idx_sort_indices, is_k_full=self.is_k_full, ) def moe_sum(self, input: torch.Tensor, output: torch.Tensor) -> None: ops.moe_sum(input, output) class BatchedMarlinExperts(MarlinExpertsBase): def __init__( self, max_num_tokens: int, num_dispatchers: int, quant_config: FusedMoEQuantConfig, w13_g_idx: torch.Tensor | None = None, w2_g_idx: torch.Tensor | None = None, w13_g_idx_sort_indices: torch.Tensor | None = None, w2_g_idx_sort_indices: torch.Tensor | None = None, is_k_full: bool = True, ): super().__init__( quant_config, w13_g_idx, w2_g_idx, w13_g_idx_sort_indices, w2_g_idx_sort_indices, is_k_full, ) self.max_num_tokens = max_num_tokens self.num_dispatchers = num_dispatchers def supports_expert_map(self) -> bool: return True def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: return TopKWeightAndReduceDelegate() @property def activation_formats( self, ) -> tuple[mk.FusedMoEActivationFormat, mk.FusedMoEActivationFormat]: return ( mk.FusedMoEActivationFormat.BatchedExperts, mk.FusedMoEActivationFormat.BatchedExperts, ) def supports_chunking(self) -> bool: return False def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: num_dispatchers = self.num_dispatchers num_experts = local_num_experts max_num_tokens = M if self.max_num_tokens is None else self.max_num_tokens workspace13 = (num_experts * max_num_tokens * num_dispatchers, max(K, N * 2)) workspace2 = (num_experts * max_num_tokens * num_dispatchers, N) output = (num_experts, max_num_tokens * num_dispatchers, K) return (workspace13, workspace2, output) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: str, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ): assert expert_tokens_meta is not None, "Num valid tokens per batch is required" return batched_fused_marlin_moe( hidden_states=hidden_states, expert_num_tokens=expert_tokens_meta.expert_num_tokens, w1=w1, w2=w2, bias1=self.w1_bias, bias2=self.w2_bias, w1_scale=self.w1_scale, w2_scale=self.w2_scale, gating_output=None, quant_type_id=self.quant_type_id, apply_router_weight_on_input=apply_router_weight_on_input, global_num_experts=global_num_experts, activation=activation, expert_map=expert_map, output=output, intermediate_cache13=workspace13, intermediate_cache2=workspace2, g_idx1=self.w13_g_idx, g_idx2=self.w2_g_idx, sort_indices1=self.w13_g_idx_sort_indices, sort_indices2=self.w2_g_idx_sort_indices, is_k_full=self.is_k_full, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/deepep_ll_prepare_finalize.py
vllm/model_executor/layers/fused_moe/deepep_ll_prepare_finalize.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable import deep_ep import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm import envs from vllm.logger import init_logger from vllm.model_executor.layers.fused_moe.config import FusedMoEQuantConfig from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceDelegate, ) from vllm.model_executor.layers.fused_moe.utils import ( moe_kernel_quantize_input, normalize_batched_scales_shape, ) from vllm.v1.worker.ubatching import ( dbo_current_ubatch_id, dbo_enabled, dbo_maybe_run_recv_hook, ) logger = init_logger(__name__) # DeepEP kernels quantize dispatch inputs in 128 element chunks. DEEPEP_QUANT_BLOCK_SIZE = 128 DEEPEP_QUANT_BLOCK_SHAPE = [DEEPEP_QUANT_BLOCK_SIZE, DEEPEP_QUANT_BLOCK_SIZE] logger = init_logger(__name__) def dequant_fp8( expert_x_fp8: torch.Tensor, expert_x_scales: torch.Tensor ) -> torch.Tensor: """ Return dequantized tensor in fp32 """ # TODO (varun) : Optimize leverage num_tokens_per_expert counts assert expert_x_fp8.is_contiguous() expert_x_scales = expert_x_scales.contiguous() num_experts = expert_x_fp8.size(0) expert_x_fp32 = expert_x_fp8.to(torch.float32).view( num_experts, -1, DEEPEP_QUANT_BLOCK_SIZE ) expert_x_scales = expert_x_scales.view(num_experts, -1, 1) return (expert_x_fp32 * expert_x_scales).view(expert_x_fp8.size()) class DeepEPLLPrepareAndFinalize(mk.FusedMoEPrepareAndFinalize): """ Prepare/Finalize using DeepEP low-latency kernels. """ # DeepEP low-latency kernels are compiled only for certain # specific hidden sizes. # NOTE: Keep this list sorted, maybe_roundup_layer_hidden_size depends # on it. SUPPORTED_HIDDEN_SIZES = [2048, 2560, 3072, 4096, 5120, 6144, 7168, 8192] @staticmethod def maybe_roundup_layer_hidden_size(hidden_size: int) -> int: # Round up hidden size to the closest supported hidden size. _supported_hs = DeepEPLLPrepareAndFinalize.SUPPORTED_HIDDEN_SIZES # Check sorted num_supported_hs = len(_supported_hs) assert all( [ _supported_hs[i] < _supported_hs[i + 1] for i in range(num_supported_hs - 1) ] ) for x in _supported_hs: if x >= hidden_size: return x raise ValueError( f"Hidden Size {hidden_size} is greater than the " f"maximum supported hidden size {_supported_hs[-1]}" ) def __init__( self, buffer: deep_ep.Buffer, max_tokens_per_rank: int, num_dispatchers: int, use_fp8_dispatch: bool = False, global_to_physical: torch.Tensor | None = None, physical_to_global: torch.Tensor | None = None, local_expert_global_ids: torch.Tensor | None = None, ): super().__init__() self.buffer = buffer self.max_tokens_per_rank = max_tokens_per_rank self.use_fp8_dispatch = use_fp8_dispatch # The dispatch function returns a handle that the combine function # requires. We store the handle here so it is available to the # combine function. self.handles: list[tuple | None] = [None, None] self.num_dispatchers_ = num_dispatchers topk_indices_dtype = self.topk_indices_dtype() def _maybe_cast(tensor: torch.Tensor | None) -> torch.Tensor | None: if tensor is None or topk_indices_dtype is None: return tensor return tensor.to(dtype=topk_indices_dtype) self.global_to_physical = _maybe_cast(global_to_physical) self.physical_to_global = _maybe_cast(physical_to_global) self.local_expert_global_ids = _maybe_cast(local_expert_global_ids) # We don't have enough information to determine if we should dispatch # activation scales in a packed ue8m0 format during object construction # time. This setting is handled by post_init_setup. self.use_ue8m0_dispatch = False def post_init_setup(self, fused_experts: mk.FusedMoEPermuteExpertsUnpermute): if not fused_experts.supports_packed_ue8m0_act_scales(): # Early exit. return if self.use_fp8_dispatch: logger.debug_once( "Update DeepEPLLPrepareFinalize to do packed ue8m0 scales dispatch." ) self.use_ue8m0_dispatch = True else: logger.warning_once( "DeepEPLLPrepareAndFinalize is setup to dispatch raw/unquantized " f"activations despite ({fused_experts.__class__.__name__}) being able " "to support quantized activations.", scope="local", ) def num_dispatchers(self) -> int: return self.num_dispatchers_ def output_is_reduced(self) -> bool: return True @property def activation_format(self) -> mk.FusedMoEActivationFormat: return mk.FusedMoEActivationFormat.BatchedExperts def max_num_tokens_per_rank(self) -> int | None: return self.max_tokens_per_rank def topk_indices_dtype(self) -> torch.dtype | None: return torch.int64 def _map_global_to_physical_ids(self, topk_ids: torch.Tensor) -> torch.Tensor: if self.global_to_physical is None: return topk_ids return self.global_to_physical[topk_ids] def _map_local_to_global_ids(self, expert_topk_ids: torch.Tensor) -> torch.Tensor: if self.local_expert_global_ids is None: return expert_topk_ids return self.local_expert_global_ids[expert_topk_ids] def _do_quant( self, x: torch.Tensor | tuple[torch.Tensor, torch.Tensor], a1_dtype: torch.dtype, quant_config: FusedMoEQuantConfig, ) -> tuple[torch.Tensor, torch.Tensor | None]: if self.use_fp8_dispatch: block_k = ( quant_config.block_shape[1] if quant_config.block_shape is not None else None ) if block_k == DEEPEP_QUANT_BLOCK_SIZE: # DeepEP kernels did the quantization for us. x, x_scales = x return x, x_scales # Dequant to get back the tokens in the datatype we dispatched in. x_fp8, x_scales = x x = dequant_fp8(x_fp8, x_scales).to(dtype=a1_dtype) assert isinstance(x, (torch.Tensor, tuple)) q_dtype = quant_config.quant_dtype if q_dtype == "nvfp4" and envs.VLLM_DEEPEPLL_NVFP4_DISPATCH: logger.info_once( "Since VLLM_DEEPEPLL_NVFP4_DISPATCH==1, make sure " "using the hybrid-ep branch of DeepEP" "(https://github.com/deepseek-ai/DeepEP/tree/hybrid-ep)" ) assert isinstance(x, tuple) x_scales = x[1] x = x[0].permute(2, 0, 1) num_experts, max_tokens, hidden_dim_by_2 = x.shape hidden_dim = hidden_dim_by_2 * 2 assert envs.VLLM_FLASHINFER_MOE_BACKEND == "masked_gemm" logger.info_once( "Quantization is fused with DeepEP nvfp4 dispatch for " "FlashInfer CUTEDSL as VLLM_DEEPEPLL_NVFP4_DISPATCH==1" ) else: if q_dtype == "nvfp4": q_dtype = None logger.info_once( "Using DeepEP bfloat16 dispatch for FlashInfer CUTEDSL as " "VLLM_DEEPEPLL_NVFP4_DISPATCH==0" ) assert isinstance(x, torch.Tensor) num_experts, max_tokens, hidden_dim = x.size() # TODO (varun): Optimization - Use a batched version of quant x = x.view((-1, hidden_dim)) x, x_scales = moe_kernel_quantize_input( x, quant_config.a1_scale, q_dtype, quant_config.per_act_token_quant, quant_config.block_shape, ) x = x.view((num_experts, -1, hidden_dim)) if q_dtype is not None and q_dtype != "nvfp4": assert x_scales is not None x_scales = normalize_batched_scales_shape(x_scales, num_experts) return x, x_scales def supports_async(self) -> bool: return True def prepare_async( self, a1: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_experts: int, expert_map: torch.Tensor | None, apply_router_weight_on_input: bool, quant_config: FusedMoEQuantConfig, ) -> tuple[Callable, mk.ReceiverType]: hidden_size = a1.size(1) assert hidden_size in self.SUPPORTED_HIDDEN_SIZES, ( f"Hidden Size {hidden_size} not in supported list of hidden sizes" f"{self.SUPPORTED_HIDDEN_SIZES}" ) a2a_idx = dbo_current_ubatch_id() if self.use_fp8_dispatch: assert hidden_size % 128 == 0, ( "DeepEP kernels quantize the inputs in blocks of shape 128" ) use_nvfp4 = False nvfp4_dispatch = ( quant_config.quant_dtype == "nvfp4" and envs.VLLM_DEEPEPLL_NVFP4_DISPATCH ) if nvfp4_dispatch: use_nvfp4 = True qc_a1_gscale_or_scale = ( quant_config.a1_gscale if nvfp4_dispatch else quant_config.a1_scale ) has_per_token_scales = ( qc_a1_gscale_or_scale.numel() != 1 if qc_a1_gscale_or_scale is not None else ( quant_config.a2_scale.numel() != 1 if quant_config.a2_scale is not None else False ) ) if not use_nvfp4: assert not has_per_token_scales, ( "low_latency kernels doesn't support dispatching per-token scales" ) if apply_router_weight_on_input: topk = topk_ids.size(1) # TODO: this only works for topK=1, will need to update for topK>1 assert topk == 1, ( "apply_router_weight_on_input is only implemented for topk=1" ) a1 = a1 * topk_weights.to(a1.dtype) # Dispatch dispatch_topk_ids = self._map_global_to_physical_ids(topk_ids) expert_x, expert_num_tokens, handle, _, hook = self.buffer.low_latency_dispatch( a1, dispatch_topk_ids, self.max_tokens_per_rank, num_experts, use_fp8=self.use_fp8_dispatch, **(dict(use_nvfp4=True) if use_nvfp4 else dict()), **( dict(x_global_scale=qc_a1_gscale_or_scale) if qc_a1_gscale_or_scale is not None else dict() ), async_finish=False, return_recv_hook=True, ) self.handles[a2a_idx] = handle return ( hook, lambda: self._receiver( expert_x, expert_num_tokens, quant_config.a1_scale, a1.dtype, quant_config, ), ) def _receiver( self, expert_x: torch.Tensor | tuple[torch.Tensor, torch.Tensor], expert_num_tokens: torch.Tensor, a1_scale: torch.Tensor | None, a1_dtype: torch.dtype, quant_config: FusedMoEQuantConfig, ) -> mk.PrepareResultType: expert_x, expert_x_scale = self._do_quant(expert_x, a1_dtype, quant_config) expert_tokens_meta = mk.ExpertTokensMetadata( expert_num_tokens=expert_num_tokens, expert_num_tokens_cpu=None ) return expert_x, expert_x_scale, expert_tokens_meta, None, None def prepare( self, a1: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_experts: int, expert_map: torch.Tensor | None, apply_router_weight_on_input: bool, quant_config: FusedMoEQuantConfig, ) -> mk.PrepareResultType: hook, receiver = self.prepare_async( a1, topk_weights, topk_ids, num_experts, expert_map, apply_router_weight_on_input, quant_config, ) hook() return receiver() def _finalize( self, output: torch.Tensor, fused_expert_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, apply_router_weight_on_input: bool, weight_and_reduce_impl: mk.TopKWeightAndReduce, do_async: bool, ) -> tuple[Callable, Callable]: assert isinstance(weight_and_reduce_impl, TopKWeightAndReduceDelegate), ( "Weight application and reduction happens in the combine kernel." ) a2a_idx = dbo_current_ubatch_id() do_recv_hook = dbo_enabled() or do_async handle = self.handles[a2a_idx] assert handle is not None combine_topk_weights = topk_weights if apply_router_weight_on_input: # weights have already been applied. combine_topk_weights = torch.ones_like(topk_weights) combine_topk_ids = self._map_global_to_physical_ids(topk_ids) # TODO (varun) : Enable zero copy mode dbo_maybe_run_recv_hook() _, _, recv_hook = self.buffer.low_latency_combine( fused_expert_output, combine_topk_ids, combine_topk_weights, handle, async_finish=False, zero_copy=False, return_recv_hook=do_recv_hook, out=output, ) return recv_hook, lambda: None def finalize_async( self, output: torch.Tensor, fused_expert_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, apply_router_weight_on_input: bool, weight_and_reduce_impl: mk.TopKWeightAndReduce, ) -> tuple[Callable, Callable]: return self._finalize( output, fused_expert_output, topk_weights, topk_ids, apply_router_weight_on_input, weight_and_reduce_impl, do_async=True, ) def finalize( self, output: torch.Tensor, fused_expert_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, apply_router_weight_on_input: bool, weight_and_reduce_impl: mk.TopKWeightAndReduce, ) -> None: self._finalize( output, fused_expert_output, topk_weights, topk_ids, apply_router_weight_on_input, weight_and_reduce_impl, do_async=False, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/moe_align_block_size.py
vllm/model_executor/layers/fused_moe/moe_align_block_size.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm import _custom_ops as ops from vllm.triton_utils import triton from vllm.utils.math_utils import round_up def moe_align_block_size( topk_ids: torch.Tensor, block_size: int, num_experts: int, expert_map: torch.Tensor | None = None, pad_sorted_ids: bool = False, ignore_invalid_experts: bool = False, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Aligns the token distribution across experts to be compatible with block size for matrix multiplication. Note: In the case of expert_parallel, moe_align_block_size initially considers all experts as valid and aligns all tokens appropriately. Before the function returns it marks the experts_ids that are not in the current GPU rank as -1 so the MoE matmuls could skip those blocks. This requires the num_experts input arg to be the num global experts. Parameters: - topk_ids: A tensor of shape [total_tokens, top_k] representing the top-k expert indices for each token. - block_size: The block size used in block matrix multiplication. - num_experts: The total number of experts. - expert_map: A tensor of shape [num_experts] that maps the expert index from the global space to the local index space of the current expert parallel shard. If the expert is not in the current expert parallel shard, the mapping is set to -1. - pad_sorted_ids: A flag indicating whether the sorted_token_ids length should be padded to a multiple of block_size, - ignore_invalid_experts: A flag indicating whether to ignore invalid experts. When False, all expert_ids in topk_ids will participate in counting and ranking, but invalid experts in expert_ids will be marked as -1. When True, all invalid expert_ids in topk_ids will be ignored and will not participate in counting or ranking, and there will be no -1 in expert_ids. Returns: - sorted_token_ids: A tensor containing the sorted token indices according to their allocated expert. - expert_ids: A tensor indicating the assigned expert index for each block. - num_tokens_post_padded: The total number of tokens after padding, ensuring divisibility by block_size. This function pads the number of tokens that each expert needs to process so that it is divisible by block_size. Padding ensures that during block matrix multiplication, the dimensions align correctly. Example: Given topk_ids = [[2, 3, 4], [1, 2, 4], [1, 3, 4], [1, 2, 3]], block_size = 4, and num_experts = 4: - We initially have 12 tokens (after repeating 'top_k' times) and 4 experts, with each expert needing to process 3 tokens. - As block_size is 4, we pad 1 token for each expert. - First, flatten topk_ids to [2, 3, 4, 1, 2, 4, 1, 3, 4, 1, 2, 3]. - Then append padding tokens [12, 12, 12, 12] for each block. - After sorting by expert index, we obtain token_ids [3, 6, 9, 12, 0, 4, 10, 12, 1, 7, 11, 12, 2, 5, 8, 12]. Tokens 12 are non-existent (padding) and are ignored in the subsequent matrix multiplication. - The padding ensures that the total number of tokens is now divisible by block_size for proper block matrix operations. """ max_num_tokens_padded = topk_ids.numel() + num_experts * (block_size - 1) if pad_sorted_ids: max_num_tokens_padded = round_up(max_num_tokens_padded, block_size) if topk_ids.numel() < num_experts: max_num_tokens_padded = min( topk_ids.numel() * block_size, max_num_tokens_padded ) sorted_ids = torch.empty( (max_num_tokens_padded,), dtype=torch.int32, device=topk_ids.device ) max_num_m_blocks = triton.cdiv(max_num_tokens_padded, block_size) expert_ids = torch.empty( (max_num_m_blocks,), dtype=torch.int32, device=topk_ids.device ) num_tokens_post_pad = torch.empty((1), dtype=torch.int32, device=topk_ids.device) ops.moe_align_block_size( topk_ids, num_experts, block_size, sorted_ids, expert_ids, num_tokens_post_pad, expert_map if ignore_invalid_experts else None, ) if expert_map is not None and not ignore_invalid_experts: expert_ids = expert_map[expert_ids] return sorted_ids, expert_ids, num_tokens_post_pad def batched_moe_align_block_size( max_tokens_per_batch: int, block_size: int, expert_num_tokens: torch.Tensor ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Given num_batches, max_tokens_per_batch, block_size and the number of valid-tokens in each batch, prepare sorted_token_ids, expert_ids and num_tokens_post_pad. sorted_token_ids, expert_ids and num_tokens_post_pad have the same semantics as in moe_align_block_size. This function is intended to be a drop in replacement for moe_align_batch_size for the batched case. Parameters: - max_tokens_per_batch (int): Number of tokens in each batch (both valid and invalid). - block_size (int): block_size to align the data to. - expert_num_tokens (torch.Tensor): expert_num_tokens[i], indicates the number of valid tokens in batch i. Returns: - sorted_token_ids (torch.Tensor): Torch tensor of size (num_batches * max_tokens_per_batch) indicating the token indices for that block. - expert_ids (torch.Tensor): Torch tensor of size ceil((num_batches * max_tokens_per_batch) / block_size) indicating what expert to use for each block. - num_tokens_post_pad (torch.Tensor): Torch tensor of size 1 indicating the number of valid blocks with actual data to process. This is represented in terms of num tokens. Example: Let num_batches=5, max_tokens_per_batch=8, block_size=4, and expert_num_tokens=[2, 3, 0, 6, 8]. This expert_num_tokens tensor indicates that, - The first 2 tokens in the 0th batch are valid and the rest 6 are invalid (i.e. in the 2D hidden_states tensor of shape, [num_batches * max_tokens_per_batch, K], indices 0, 1 are valid) - The first 3 tokens in the 1st batch are valid. i.e. indices 8, 9, 10 - 0 tokens in the 2nd batch are valid - first 6 tokens in the 3rd batch are valid. i.e. indices, 24, 25, 26, 27, 28, 29 - so on ... In this case, sorted_token_ids will be [0, 1, 40, 40, 8, 9, 10, 40, 24, 25, 26, 27, 28, 29, 40, 40, 32, 33, 34, 35, 36, 37, 38, 39, 40, 40, 40, 40, (rest all 40, 40, 40, 40) ...] Here, 40 represents an invalid index. as there is no token index 40. The gemm kernel using this sorted_token_ids is expected to skip the gemm computation when it encounters this invalid index. expert_ids will be [0, 1, 3, 3, 4, 5, 5, -1, -1, (rest all -1) ...] Here, -1 represents an invalid expert. The gemm kernel using this expert_ids is expected to skip the gemm computation when it encounters an expert of id -1. num_tokens_post_pad will be 24 as sorted_token_ids has valid entries until 24. """ B = expert_num_tokens.size(0) device = expert_num_tokens.device # Round up so each batch can be split to blocks evenly. max_num_tokens_padded = B * round_up(max_tokens_per_batch, block_size) sorted_ids = torch.empty((max_num_tokens_padded,), dtype=torch.int32, device=device) assert max_num_tokens_padded % block_size == 0 max_num_m_blocks = max_num_tokens_padded // block_size expert_ids = torch.empty((max_num_m_blocks,), dtype=torch.int32, device=device) num_tokens_post_pad = torch.empty((1), dtype=torch.int32, device=device) ops.batched_moe_align_block_size( max_tokens_per_batch, block_size, expert_num_tokens, sorted_ids, expert_ids, num_tokens_post_pad, ) return sorted_ids, expert_ids, num_tokens_post_pad
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/fused_moe.py
vllm/model_executor/layers/fused_moe/fused_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Fused MoE Triton kernels.""" import functools import json import os from collections.abc import Callable from typing import Any import torch import torch.nn.functional as F import vllm.envs as envs import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm import _custom_ops as ops from vllm._aiter_ops import rocm_aiter_ops from vllm.logger import init_logger from vllm.model_executor.custom_op import CustomOp from vllm.model_executor.layers.batch_invariant import ( vllm_is_batch_invariant, ) from vllm.model_executor.layers.fused_moe.config import ( FUSED_MOE_UNQUANTIZED_CONFIG, FusedMoEQuantConfig, _get_config_dtype_str, ) from vllm.model_executor.layers.fused_moe.deep_gemm_moe import ( _valid_deep_gemm, deep_gemm_moe_fp8, ) from vllm.model_executor.layers.fused_moe.moe_align_block_size import ( moe_align_block_size, ) from vllm.model_executor.layers.fused_moe.prepare_finalize import ( MoEPrepareAndFinalizeNoEP, ) from vllm.model_executor.layers.fused_moe.rocm_aiter_fused_moe import ( # noqa: E501 rocm_aiter_grouped_topk, ) from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceNoOP, ) from vllm.model_executor.layers.fused_moe.utils import ( _resize_cache, activation_without_mul, disable_inplace, moe_kernel_quantize_input, ) from vllm.model_executor.layers.quantization.utils.mxfp4_utils import dequant_mxfp4 from vllm.model_executor.layers.quantization.utils.mxfp6_utils import dequant_mxfp6 from vllm.model_executor.layers.quantization.utils.ocp_mx_utils import OCP_MX_Scheme from vllm.model_executor.utils import maybe_disable_graph_partition from vllm.platforms import current_platform from vllm.triton_utils import tl, triton from vllm.utils.deep_gemm import is_deep_gemm_e8m0_used from vllm.utils.torch_utils import direct_register_custom_op, is_torch_equal_or_newer logger = init_logger(__name__) @triton.jit def write_zeros_to_output( c_ptr, stride_cm, stride_cn, pid_n, N, offs_token, token_mask, BLOCK_SIZE_M, BLOCK_SIZE_N, compute_type, ): accumulator = tl.zeros((BLOCK_SIZE_M, BLOCK_SIZE_N), dtype=compute_type) offs_cn = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) c_ptrs = c_ptr + stride_cm * offs_token[:, None] + stride_cn * offs_cn[None, :] c_mask = token_mask[:, None] & (offs_cn[None, :] < N) tl.store(c_ptrs, accumulator, mask=c_mask) @triton.jit def fused_moe_kernel_gptq_awq( # Pointers to matrices a_ptr, b_ptr, c_ptr, b_scale_ptr, b_zp_ptr, topk_weights_ptr, sorted_token_ids_ptr, expert_ids_ptr, num_tokens_post_padded_ptr, # Matrix dimensions N: tl.constexpr, K: tl.constexpr, EM, num_valid_tokens, # The stride variables represent how much to increase the ptr by when # moving by 1 element in a particular dimension. E.g. `stride_am` is # how much to increase `a_ptr` by to get the element one row down # (A has M rows). stride_am, stride_ak, stride_be, stride_bk, stride_bn, stride_cm, stride_cn, stride_bse, stride_bsk, stride_bsn, stride_bze, stride_bzk, stride_bzn, block_k_diviable: tl.constexpr, group_size: tl.constexpr, # Meta-parameters BLOCK_SIZE_M: tl.constexpr, BLOCK_SIZE_N: tl.constexpr, BLOCK_SIZE_K: tl.constexpr, GROUP_SIZE_M: tl.constexpr, SPLIT_K: tl.constexpr, MUL_ROUTED_WEIGHT: tl.constexpr, top_k: tl.constexpr, compute_type: tl.constexpr, has_zp: tl.constexpr, use_int4_w4a16: tl.constexpr, use_int8_w8a16: tl.constexpr, ): """ Implements the fused computation for a Mixture of Experts (MOE) using token and expert matrices. Key Parameters: - A: The input tensor representing tokens with shape (*, K), where '*' can be any shape representing batches and K is the feature dimension of each token. - B: The stacked MOE weight tensor with shape (E, N, K), where E is the number of experts, K is the input feature dimension, and N is the output feature dimension. - C: The output cache tensor with shape (M, topk, N), where M is the total number of tokens post padding, topk is the number of times each token is repeated, and N is the output feature dimension. - sorted_token_ids: A tensor containing the sorted indices of tokens, repeated topk times and arranged by the expert index they are assigned to. - expert_ids: A tensor containing the indices of the expert for each block. It determines which expert matrix from B should be used for each block in A. This kernel performs the multiplication of a token by its corresponding expert matrix as determined by `expert_ids`. The sorting of `sorted_token_ids` by expert index and padding ensures divisibility by BLOCK_SIZE_M, which is necessary to maintain consistency in block matrix multiplication across different blocks processed by the same expert. """ # ----------------------------------------------------------- # Map program ids `pid` to the block of C it should compute. # This is done in a grouped ordering to promote L2 data reuse. pid = tl.program_id(axis=0) num_pid_m = tl.cdiv(EM, BLOCK_SIZE_M) num_pid_n = tl.cdiv(N, BLOCK_SIZE_N) num_pid_in_group = GROUP_SIZE_M * num_pid_n group_id = pid // num_pid_in_group first_pid_m = group_id * GROUP_SIZE_M group_size_m = min(num_pid_m - first_pid_m, GROUP_SIZE_M) pid_m = first_pid_m + ((pid % num_pid_in_group) % group_size_m) pid_n = (pid % num_pid_in_group) // group_size_m # ---------------------------------------------------------- # Create pointers for the first blocks of A and B. # We will advance this pointer as we move in the K direction # and accumulate # `a_ptrs` is a block of [BLOCK_SIZE_M, BLOCK_SIZE_K] pointers # `b_ptrs` is a block of [BLOCK_SIZE_K, BLOCK_SIZE_N] pointers num_tokens_post_padded = tl.load(num_tokens_post_padded_ptr) if pid_m * BLOCK_SIZE_M >= num_tokens_post_padded: return offs_token_id = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M).to(tl.int64) offs_token = tl.load(sorted_token_ids_ptr + offs_token_id) token_mask = offs_token < num_valid_tokens off_experts = tl.load(expert_ids_ptr + pid_m).to(tl.int64) if off_experts == -1: # ----------------------------------------------------------- # Write back zeros to the output when the expert is not # in the current expert parallel rank. write_zeros_to_output( c_ptr, stride_cm, stride_cn, pid_n, N, offs_token, token_mask, BLOCK_SIZE_M, BLOCK_SIZE_N, compute_type, ) return offs_bn = (pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N).to(tl.int64)) % N offs_k = tl.arange(0, BLOCK_SIZE_K) a_ptrs = a_ptr + ( offs_token[:, None] // top_k * stride_am + offs_k[None, :] * stride_ak ) if use_int4_w4a16: b_ptrs = ( b_ptr + off_experts * stride_be + (offs_k[:, None] // 2) * stride_bk + offs_bn[None, :] * stride_bn ) b_shifter = (offs_k[:, None] % 2) * 4 elif use_int8_w8a16: b_ptrs = ( b_ptr + off_experts * stride_be + offs_k[:, None] * stride_bk + offs_bn[None, :] * stride_bn ) if not has_zp and use_int4_w4a16: b_zp_num = 8 if not has_zp and use_int8_w8a16: b_zp_num = 128 elif has_zp and use_int4_w4a16: b_zp_shifter = (offs_bn[None, :] % 2) * 4 # ----------------------------------------------------------- # Iterate to compute a block of the C matrix. # We accumulate into a `[BLOCK_SIZE_M, BLOCK_SIZE_N]` block # of fp32 values for higher accuracy. # `accumulator` will be converted back to fp16 after the loop. accumulator = tl.zeros((BLOCK_SIZE_M, BLOCK_SIZE_N), dtype=tl.float32) for k in range(0, tl.cdiv(K, BLOCK_SIZE_K)): # Load the next block of A and B, generate a mask by checking the # K dimension. if not block_k_diviable: k_mask = offs_k[:, None] < K - k * BLOCK_SIZE_K k_other = 0.0 else: k_mask = None k_other = None a = tl.load( a_ptrs, mask=token_mask[:, None] & (offs_k[None, :] < K - k * BLOCK_SIZE_K), other=0.0, ) b = tl.load(b_ptrs) if use_int4_w4a16: b = (b >> b_shifter) & 0xF b_scale_ptrs = ( b_scale_ptr + off_experts * stride_bse + offs_bn[None, :] * stride_bsn + ((offs_k[:, None] + BLOCK_SIZE_K * k) // group_size) * stride_bsk ) b_scale = tl.load(b_scale_ptrs, mask=k_mask, other=k_other) b_scale = b_scale.to(tl.float32) if has_zp and use_int4_w4a16: offs_k_true = (offs_k[:, None] + BLOCK_SIZE_K * k) // group_size b_zp_ptrs = ( b_zp_ptr + off_experts * stride_bze + (offs_bn[None, :] // 2) * stride_bzn + offs_k_true * stride_bzk ) b_zp = tl.load(b_zp_ptrs, mask=k_mask, other=k_other) b_zp = (b_zp >> b_zp_shifter) & 0xF b_zp = b_zp.to(tl.float32) elif has_zp and use_int8_w8a16: offs_k_true = (offs_k[:, None] + BLOCK_SIZE_K * k) // group_size b_zp_ptrs = ( b_zp_ptr + off_experts * stride_bze + offs_bn[None, :] * stride_bzn + offs_k_true * stride_bzk ) b_zp = tl.load(b_zp_ptrs, mask=k_mask, other=k_other) b_zp = b_zp.to(tl.float32) # We accumulate along the K dimension. if has_zp: b = ((b.to(tl.float32) - b_zp) * b_scale).to(compute_type) else: b = ((b.to(tl.float32) - b_zp_num) * b_scale).to(compute_type) accumulator = tl.dot(a, b, acc=accumulator) # Advance the ptrs to the next K block. a_ptrs += BLOCK_SIZE_K * stride_ak if use_int4_w4a16: b_ptrs += (BLOCK_SIZE_K // 2) * stride_bk else: b_ptrs += BLOCK_SIZE_K * stride_bk if MUL_ROUTED_WEIGHT: moe_weight = tl.load(topk_weights_ptr + offs_token, mask=token_mask, other=0) accumulator = accumulator * moe_weight[:, None] accumulator = accumulator.to(compute_type) # ----------------------------------------------------------- # Write back the block of the output offs_cn = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) c_ptrs = c_ptr + stride_cm * offs_token[:, None] + stride_cn * offs_cn[None, :] c_mask = token_mask[:, None] & (offs_cn[None, :] < N) tl.store(c_ptrs, accumulator, mask=c_mask) @triton.jit def fused_moe_kernel( # Pointers to matrices a_ptr, b_ptr, c_ptr, b_bias_ptr, a_scale_ptr, b_scale_ptr, topk_weights_ptr, sorted_token_ids_ptr, expert_ids_ptr, num_tokens_post_padded_ptr, # Matrix dimensions N, K, EM, num_valid_tokens, # The stride variables represent how much to increase the ptr by when # moving by 1 element in a particular dimension. E.g. `stride_am` is # how much to increase `a_ptr` by to get the element one row down # (A has M rows). stride_am, stride_ak, stride_be, stride_bk, stride_bn, stride_cm, stride_cn, stride_asm, stride_ask, stride_bse, stride_bsk, stride_bsn, stride_bbe, # bias expert stride stride_bbn, # bias N stride # Block size for block-wise quantization group_n: tl.constexpr, group_k: tl.constexpr, # Meta-parameters BLOCK_SIZE_M: tl.constexpr, BLOCK_SIZE_N: tl.constexpr, BLOCK_SIZE_K: tl.constexpr, GROUP_SIZE_M: tl.constexpr, SPLIT_K: tl.constexpr, MUL_ROUTED_WEIGHT: tl.constexpr, top_k: tl.constexpr, compute_type: tl.constexpr, use_fp8_w8a8: tl.constexpr, use_int8_w8a8: tl.constexpr, use_int8_w8a16: tl.constexpr, per_channel_quant: tl.constexpr, HAS_BIAS: tl.constexpr, ): """ Implements the fused computation for a Mixture of Experts (MOE) using token and expert matrices. Key Parameters: - A: The input tensor representing tokens with shape (*, K), where '*' can be any shape representing batches and K is the feature dimension of each token. - B: The stacked MOE weight tensor with shape (E, N, K), where E is the number of experts, K is the input feature dimension, and N is the output feature dimension. - C: The output cache tensor with shape (M, topk, N), where M is the total number of tokens post padding, topk is the number of times each token is repeated, and N is the output feature dimension. - sorted_token_ids: A tensor containing the sorted indices of tokens, repeated topk times and arranged by the expert index they are assigned to. - expert_ids: A tensor containing the indices of the expert for each block. It determines which expert matrix from B should be used for each block in A. This kernel performs the multiplication of a token by its corresponding expert matrix as determined by `expert_ids`. The sorting of `sorted_token_ids` by expert index and padding ensures divisibility by BLOCK_SIZE_M, which is necessary to maintain consistency in block matrix multiplication across different blocks processed by the same expert. """ # ----------------------------------------------------------- # Map program ids `pid` to the block of C it should compute. # This is done in a grouped ordering to promote L2 data reuse. pid = tl.program_id(axis=0) num_pid_m = tl.cdiv(EM, BLOCK_SIZE_M) num_pid_n = tl.cdiv(N, BLOCK_SIZE_N) num_pid_in_group = GROUP_SIZE_M * num_pid_n group_id = pid // num_pid_in_group first_pid_m = group_id * GROUP_SIZE_M group_size_m = min(num_pid_m - first_pid_m, GROUP_SIZE_M) pid_m = first_pid_m + ((pid % num_pid_in_group) % group_size_m) pid_n = (pid % num_pid_in_group) // group_size_m # ---------------------------------------------------------- # Create pointers for the first blocks of A and B. # We will advance this pointer as we move in the K direction # and accumulate # `a_ptrs` is a block of [BLOCK_SIZE_M, BLOCK_SIZE_K] pointers # `b_ptrs` is a block of [BLOCK_SIZE_K, BLOCK_SIZE_N] pointers num_tokens_post_padded = tl.load(num_tokens_post_padded_ptr) if pid_m * BLOCK_SIZE_M >= num_tokens_post_padded: return offs_token_id = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M).to(tl.int64) offs_token = tl.load(sorted_token_ids_ptr + offs_token_id) token_mask = offs_token < num_valid_tokens off_experts = tl.load(expert_ids_ptr + pid_m).to(tl.int64) if off_experts == -1: # ----------------------------------------------------------- # Write back zeros to the output when the expert is not # in the current expert parallel rank. write_zeros_to_output( c_ptr, stride_cm, stride_cn, pid_n, N, offs_token, token_mask, BLOCK_SIZE_M, BLOCK_SIZE_N, compute_type, ) return offs_bn = (pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N).to(tl.int64)) % N offs_k = tl.arange(0, BLOCK_SIZE_K) a_ptrs = a_ptr + ( offs_token[:, None] // top_k * stride_am + offs_k[None, :] * stride_ak ) b_ptrs = ( b_ptr + off_experts * stride_be + (offs_k[:, None] * stride_bk + offs_bn[None, :] * stride_bn) ) if use_int8_w8a16: b_scale_ptrs = ( b_scale_ptr + off_experts * stride_bse + offs_bn[None, :] * stride_bsn ) b_scale = tl.load(b_scale_ptrs) if use_fp8_w8a8 or use_int8_w8a8: # block-wise if group_k > 0 and group_n > 0: a_scale_ptrs = a_scale_ptr + (offs_token // top_k) * stride_asm offs_bsn = offs_bn // group_n b_scale_ptrs = ( b_scale_ptr + off_experts * stride_bse + offs_bsn * stride_bsn ) # channel-wise elif per_channel_quant: b_scale_ptrs = ( b_scale_ptr + off_experts * stride_bse + offs_bn[None, :] * stride_bsn ) b_scale = tl.load(b_scale_ptrs) # Load per-token scale for activations a_scale_ptrs = a_scale_ptr + (offs_token // top_k) * stride_asm a_scale = tl.load(a_scale_ptrs, mask=token_mask, other=0.0)[:, None] # tensor-wise else: a_scale = tl.load(a_scale_ptr) b_scale = tl.load(b_scale_ptr + off_experts) if HAS_BIAS: # bias shape: [num_experts, N] bias_ptrs = b_bias_ptr + off_experts * stride_bbe + offs_bn * stride_bbn bias = tl.load(bias_ptrs, mask=(offs_bn < N), other=0.0) # ----------------------------------------------------------- # Iterate to compute a block of the C matrix. # We accumulate into a `[BLOCK_SIZE_M, BLOCK_SIZE_N]` block # of fp32 values for higher accuracy. # `accumulator` will be converted back to fp16 after the loop. accumulator = tl.zeros((BLOCK_SIZE_M, BLOCK_SIZE_N), dtype=tl.float32) for k in range(0, tl.cdiv(K, BLOCK_SIZE_K)): # Load the next block of A and B, generate a mask by checking the # K dimension. a = tl.load( a_ptrs, mask=token_mask[:, None] & (offs_k[None, :] < K - k * BLOCK_SIZE_K), other=0.0, ) b = tl.load(b_ptrs, mask=offs_k[:, None] < K - k * BLOCK_SIZE_K, other=0.0) # We accumulate along the K dimension. if use_int8_w8a16: accumulator = tl.dot(a, b.to(compute_type), acc=accumulator) elif use_fp8_w8a8 or use_int8_w8a8: if group_k > 0 and group_n > 0: k_start = k * BLOCK_SIZE_K offs_ks = k_start // group_k a_scale = tl.load( a_scale_ptrs + offs_ks * stride_ask, mask=token_mask, other=0.0 ) b_scale = tl.load(b_scale_ptrs + offs_ks * stride_bsk) accumulator += tl.dot(a, b) * a_scale[:, None] * b_scale[None, :] else: if use_fp8_w8a8: # acc used to enable fp8_fast_accum accumulator = tl.dot(a, b, acc=accumulator) else: accumulator += tl.dot(a, b) else: accumulator += tl.dot(a, b) # Advance the ptrs to the next K block. a_ptrs += BLOCK_SIZE_K * stride_ak b_ptrs += BLOCK_SIZE_K * stride_bk if HAS_BIAS: accumulator = accumulator + bias[None, :] if MUL_ROUTED_WEIGHT: moe_weight = tl.load(topk_weights_ptr + offs_token, mask=token_mask, other=0) accumulator = accumulator * moe_weight[:, None] if use_int8_w8a16: accumulator = (accumulator * b_scale).to(compute_type) elif use_fp8_w8a8 or use_int8_w8a8: if group_k > 0 and group_n > 0: accumulator = accumulator.to(compute_type) else: accumulator = (accumulator * a_scale * b_scale).to(compute_type) else: accumulator = accumulator.to(compute_type) # ----------------------------------------------------------- # Write back the block of the output offs_cn = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) c_ptrs = c_ptr + stride_cm * offs_token[:, None] + stride_cn * offs_cn[None, :] c_mask = token_mask[:, None] & (offs_cn[None, :] < N) tl.store(c_ptrs, accumulator, mask=c_mask) # NOTE(zyongye): we can remove all the wna16 kernel # once we drop off sm75 support def invoke_fused_moe_wna16_cuda_kernel( A: torch.Tensor, B: torch.Tensor, C: torch.Tensor, B_scale: torch.Tensor | None, B_zp: torch.Tensor | None, topk_weights: torch.Tensor | None, sorted_token_ids: torch.Tensor, expert_ids: torch.Tensor, num_tokens_post_padded: torch.Tensor, mul_routed_weight: bool, top_k: int, config: dict[str, Any], block_shape: list[int], ): assert B_scale is not None and B_scale.ndim == 3 assert B_zp is None or B_zp.ndim == 3 assert block_shape is None or block_shape[0] == 0 M = A.size(0) num_tokens = M * top_k bit = 4 config = config.copy() config.update( get_moe_wna16_block_config( config=config, use_moe_wna16_cuda=True, num_valid_tokens=num_tokens, size_k=A.size(1), size_n=B.size(1), num_experts=B.size(1), group_size=block_shape[1], real_top_k=top_k, block_size_m=config["BLOCK_SIZE_M"], ) ) ops.moe_wna16_gemm( A, C, B, B_scale, B_zp, topk_weights if mul_routed_weight else None, sorted_token_ids, expert_ids, num_tokens_post_padded, top_k, config["BLOCK_SIZE_M"], config["BLOCK_SIZE_N"], config["BLOCK_SIZE_K"], bit, ) # NOTE(zyongye): we can remove all the wna16 kernel # once we drop off sm75 support def invoke_fused_moe_wna16_triton_kernel( A: torch.Tensor, B: torch.Tensor, C: torch.Tensor, B_scale: torch.Tensor | None, B_zp: torch.Tensor | None, topk_weights: torch.Tensor | None, sorted_token_ids: torch.Tensor, expert_ids: torch.Tensor, num_tokens_post_padded: torch.Tensor, mul_routed_weight: bool, top_k: int, config: dict[str, Any], compute_type: tl.dtype, use_int8_w8a16: bool, use_int4_w4a16: bool, block_shape: list[int], ): assert B_scale is not None and B_scale.ndim == 3 assert B_zp is None or B_zp.ndim == 3 assert block_shape is None or block_shape[0] == 0 M = A.size(0) num_tokens = M * top_k EM = sorted_token_ids.size(0) if A.size(0) < config["BLOCK_SIZE_M"]: # optimize for small batch_size. # We assume that top_ids of each token is unique, # so num_valid_experts <= batch_size <= BLOCK_SIZE_M, # and we can skip some invalid blocks. EM = min(sorted_token_ids.size(0), A.size(0) * top_k * config["BLOCK_SIZE_M"]) grid = lambda META: ( triton.cdiv(EM, META["BLOCK_SIZE_M"]) * triton.cdiv(B.size(1), META["BLOCK_SIZE_N"]), ) config = config.copy() config.update( get_moe_wna16_block_config( config=config, use_moe_wna16_cuda=False, num_valid_tokens=num_tokens, size_k=A.size(1), size_n=B.size(1), num_experts=B.size(1), group_size=block_shape[1], real_top_k=top_k, block_size_m=config["BLOCK_SIZE_M"], ) ) fused_moe_kernel_gptq_awq[grid]( A, B, C, B_scale, B_zp, topk_weights, sorted_token_ids, expert_ids, num_tokens_post_padded, B.size(1), A.size(1), EM, num_tokens, A.stride(0), A.stride(1), B.stride(0), B.stride(2), B.stride(1), C.stride(1), C.stride(2), B_scale.stride(0), B_scale.stride(2), B_scale.stride(1), B_zp.stride(0) if B_zp is not None else 0, B_zp.stride(2) if B_zp is not None else 0, B_zp.stride(1) if B_zp is not None else 0, block_k_diviable=A.size(1) % config["BLOCK_SIZE_K"] == 0, group_size=block_shape[1], MUL_ROUTED_WEIGHT=mul_routed_weight, top_k=top_k, compute_type=compute_type, has_zp=B_zp is not None, use_int4_w4a16=use_int4_w4a16, use_int8_w8a16=use_int8_w8a16, **config, ) def invoke_fused_moe_triton_kernel( A: torch.Tensor, B: torch.Tensor, C: torch.Tensor, A_scale: torch.Tensor | None, B_scale: torch.Tensor | None, topk_weights: torch.Tensor | None, sorted_token_ids: torch.Tensor, expert_ids: torch.Tensor, num_tokens_post_padded: torch.Tensor, mul_routed_weight: bool, top_k: int, config: dict[str, Any], compute_type: tl.dtype, use_fp8_w8a8: bool, use_int8_w8a8: bool, use_int8_w8a16: bool, use_int4_w4a16: bool, per_channel_quant: bool, block_shape: list[int] | None = None, B_bias: torch.Tensor | None = None, ): assert topk_weights is not None or not mul_routed_weight assert topk_weights is None or topk_weights.stride(1) == 1 assert sorted_token_ids.stride(0) == 1 if use_fp8_w8a8 or use_int8_w8a8: assert B_scale is not None assert block_shape is None or triton.cdiv( B.size(-2), block_shape[0] ) == B_scale.size(-2) assert block_shape is None or triton.cdiv( B.size(-1), block_shape[1] ) == B_scale.size(-1) elif use_int8_w8a16 or use_int4_w4a16: assert B_scale is not None assert block_shape is None or block_shape[0] == 0 else: assert A_scale is None assert B_scale is None M = A.size(0) num_tokens = M * top_k EM = sorted_token_ids.size(0) if A.size(0) < config["BLOCK_SIZE_M"]: # optimize for small batch_size. # We assume that top_ids of each token is unique, # so num_valid_experts <= batch_size <= BLOCK_SIZE_M, # and we can skip some invalid blocks. EM = min(sorted_token_ids.size(0), A.size(0) * top_k * config["BLOCK_SIZE_M"]) grid = lambda META: ( triton.cdiv(EM, META["BLOCK_SIZE_M"]) * triton.cdiv(B.size(1), META["BLOCK_SIZE_N"]), ) HAS_BIAS = B_bias is not None config = config.copy() config["SPLIT_K"] = 1 BLOCK_SIZE_K = config.pop("BLOCK_SIZE_K") if block_shape is not None: BLOCK_SIZE_K = min(BLOCK_SIZE_K, min(block_shape[0], block_shape[1])) fused_moe_kernel[grid]( A, B, C, B_bias, A_scale, B_scale, topk_weights, sorted_token_ids, expert_ids, num_tokens_post_padded, B.size(1), B.size(2), EM, num_tokens, A.stride(0), A.stride(1), B.stride(0), B.stride(2), B.stride(1), C.stride(1), C.stride(2), A_scale.stride(0) if A_scale is not None and A_scale.ndim == 2 else 0, A_scale.stride(1) if A_scale is not None and A_scale.ndim == 2 else 0, B_scale.stride(0) if B_scale is not None and B_scale.ndim >= 2 else 0, B_scale.stride(2) if B_scale is not None and B_scale.ndim == 3 else 0, B_scale.stride(1) if B_scale is not None and B_scale.ndim >= 2 else 0, B_bias.stride(0) if B_bias is not None else 0, B_bias.stride(1) if B_bias is not None else 0, 0 if block_shape is None else block_shape[0], 0 if block_shape is None else block_shape[1], MUL_ROUTED_WEIGHT=mul_routed_weight, top_k=top_k, compute_type=compute_type, use_fp8_w8a8=use_fp8_w8a8, use_int8_w8a8=use_int8_w8a8, use_int8_w8a16=use_int8_w8a16, per_channel_quant=per_channel_quant, HAS_BIAS=HAS_BIAS, BLOCK_SIZE_K=BLOCK_SIZE_K, **config, ) def dispatch_fused_moe_kernel( A: torch.Tensor, B: torch.Tensor, C: torch.Tensor, A_scale: torch.Tensor | None, B_scale: torch.Tensor | None, B_zp: torch.Tensor | None, topk_weights: torch.Tensor | None, sorted_token_ids: torch.Tensor, expert_ids: torch.Tensor, num_tokens_post_padded: torch.Tensor, mul_routed_weight: bool, top_k: int, config: dict[str, Any], compute_type: tl.dtype, use_fp8_w8a8: bool, use_int8_w8a8: bool, use_int8_w8a16: bool, use_int4_w4a16: bool, per_channel_quant: bool, block_shape: list[int] | None = None, B_bias: torch.Tensor | None = None, ) -> None: assert topk_weights is not None or not mul_routed_weight assert topk_weights is None or topk_weights.stride(1) == 1 assert sorted_token_ids.stride(0) == 1 M = A.size(0) num_tokens = M * top_k if (use_int8_w8a16 or use_int4_w4a16) and ( block_shape is not None and block_shape[1] > 0 ): assert B_bias is None use_moe_wna16_cuda = should_moe_wna16_use_cuda( num_valid_tokens=num_tokens, group_size=block_shape[1], num_experts=B.size(0), bit=4 if use_int4_w4a16 else 8, ) if use_moe_wna16_cuda: invoke_fused_moe_wna16_cuda_kernel( A, B, C, B_scale, B_zp, topk_weights, sorted_token_ids, expert_ids, num_tokens_post_padded, mul_routed_weight, top_k, config, block_shape, ) return invoke_fused_moe_wna16_triton_kernel( A, B, C, B_scale, B_zp, topk_weights, sorted_token_ids, expert_ids, num_tokens_post_padded, mul_routed_weight, top_k, config, compute_type, use_int8_w8a16, use_int4_w4a16, block_shape, ) else: invoke_fused_moe_triton_kernel( A, B, C, A_scale, B_scale, topk_weights, sorted_token_ids, expert_ids, num_tokens_post_padded, mul_routed_weight, top_k, config, compute_type, use_fp8_w8a8, use_int8_w8a8, use_int8_w8a16, use_int4_w4a16, per_channel_quant, block_shape, B_bias, ) @triton.jit def compute_identity_kernel( top_k: int, hidden_states_ptr: tl.tensor, expert_scales_ptr: tl.tensor, num_tokens: int, output_ptr: tl.tensor, hidden_dim: int, scales_stride: int, BLOCK_SIZE: tl.constexpr, ) -> None: pid = tl.program_id(0) batch_id = pid // (hidden_dim // BLOCK_SIZE) dim_offset = pid % (hidden_dim // BLOCK_SIZE) * BLOCK_SIZE if batch_id >= num_tokens or dim_offset >= hidden_dim: return h = tl.load( hidden_states_ptr + batch_id * hidden_dim + dim_offset + tl.arange(0, BLOCK_SIZE), mask=(dim_offset + tl.arange(0, BLOCK_SIZE)) < hidden_dim, ) result = tl.zeros([BLOCK_SIZE], dtype=tl.float32) for i in range(top_k): scale = tl.load(expert_scales_ptr + batch_id * scales_stride + i) result += h * scale tl.store( output_ptr + batch_id * hidden_dim + dim_offset + tl.arange(0, BLOCK_SIZE), result, mask=(dim_offset + tl.arange(0, BLOCK_SIZE)) < hidden_dim, ) def zero_experts_compute_triton( expert_indices: torch.Tensor, expert_scales: torch.Tensor, num_experts: int, zero_expert_type: str, hidden_states: torch.Tensor, ) -> torch.Tensor: N = expert_indices.numel() top_k = expert_indices.size(-1) grid = lambda meta: (triton.cdiv(N, meta["BLOCK_SIZE"]),) if zero_expert_type == "identity": zero_expert_mask = expert_indices < num_experts zero_expert_scales = expert_scales.clone() zero_expert_scales[zero_expert_mask] = 0.0 normal_expert_mask = expert_indices >= num_experts expert_indices[normal_expert_mask] = 0 expert_scales[normal_expert_mask] = 0.0 output = torch.zeros_like(hidden_states).to(hidden_states.device) hidden_dim = hidden_states.size(-1) num_tokens = hidden_states.size(0) grid = lambda meta: (num_tokens * (hidden_dim // meta["BLOCK_SIZE"]),) compute_identity_kernel[grid]( top_k, hidden_states, zero_expert_scales, num_tokens, output, hidden_dim, zero_expert_scales.stride(0), BLOCK_SIZE=256,
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/unquantized_fused_moe_method.py
vllm/model_executor/layers/fused_moe/unquantized_fused_moe_method.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import torch.nn.functional as F import vllm.envs as envs import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm._aiter_ops import rocm_aiter_ops from vllm.logger import init_logger from vllm.model_executor.custom_op import CustomOp from vllm.model_executor.layers.fused_moe.config import ( FUSED_MOE_UNQUANTIZED_CONFIG, FusedMoEConfig, FusedMoEQuantConfig, biased_moe_quant_config, ) from vllm.model_executor.layers.fused_moe.flashinfer_cutlass_moe import ( FlashInferExperts, ) from vllm.model_executor.layers.fused_moe.fused_moe_method_base import ( FusedMoEMethodBase, ) from vllm.model_executor.layers.fused_moe.modular_kernel import ( FusedMoEActivationFormat, FusedMoEPermuteExpertsUnpermute, FusedMoEPrepareAndFinalize, ) from vllm.model_executor.layers.fused_moe.prepare_finalize import ( MoEPrepareAndFinalizeNoEP, ) from vllm.model_executor.layers.quantization.utils.flashinfer_utils import ( swap_w13_to_w31, ) from vllm.model_executor.utils import replace_parameter, set_weight_attrs from vllm.platforms import current_platform from vllm.platforms.interface import CpuArchEnum from vllm.utils.flashinfer import has_flashinfer_cutlass_fused_moe if current_platform.is_cuda_alike(): from .fused_batched_moe import BatchedTritonExperts from .fused_moe import TritonExperts else: TritonExperts = None # type: ignore if current_platform.is_tpu(): from .moe_pallas import fused_moe as fused_moe_pallas else: fused_moe_pallas = None # type: ignore logger = init_logger(__name__) @CustomOp.register("unquantized_fused_moe") class UnquantizedFusedMoEMethod(FusedMoEMethodBase, CustomOp): """MoE method without quantization.""" def __init__(self, moe: FusedMoEConfig): super().__init__(moe) self.rocm_aiter_moe_enabled = rocm_aiter_ops.is_fused_moe_enabled() if self.rocm_aiter_moe_enabled: from .rocm_aiter_fused_moe import rocm_aiter_fused_experts self.rocm_aiter_fused_experts = rocm_aiter_fused_experts else: self.rocm_aiter_fused_experts = None # type: ignore # FlashInfer CUTLASS MoE is only supported on Hopper and later GPUS self.flashinfer_cutlass_moe_enabled = ( has_flashinfer_cutlass_fused_moe() and envs.VLLM_USE_FLASHINFER_MOE_FP16 and self.moe.moe_parallel_config.use_ep and self.moe.moe_parallel_config.dp_size == 1 and current_platform.get_device_capability()[0] >= 9 ) if self.flashinfer_cutlass_moe_enabled: logger.info_once( "Enabling FlashInfer CUTLASS MoE for UnquantizedFusedMoEMethod" ) else: if ( self.moe.moe_parallel_config.use_ep and self.moe.moe_parallel_config.dp_size == 1 ): logger.info_once( "FlashInfer CUTLASS MoE is available for EP" " but not enabled, consider setting" " VLLM_USE_FLASHINFER_MOE_FP16=1 to enable it.", scope="local", ) elif self.moe.moe_parallel_config.dp_size > 1: logger.info_once( "FlashInfer CUTLASS MoE is currently not available for DP.", scope="local", ) @property def supports_eplb(self) -> bool: return True @property def allow_inplace(self) -> bool: return True def maybe_make_prepare_finalize( self, routing_tables: tuple[torch.Tensor, torch.Tensor, torch.Tensor] | None = None, ) -> FusedMoEPrepareAndFinalize | None: if self.rocm_aiter_moe_enabled: return None else: return super().maybe_make_prepare_finalize(routing_tables) def select_gemm_impl( self, prepare_finalize: FusedMoEPrepareAndFinalize, layer: torch.nn.Module, ) -> FusedMoEPermuteExpertsUnpermute: assert self.moe_quant_config is not None if ( prepare_finalize.activation_format == FusedMoEActivationFormat.BatchedExperts ): logger.debug("BatchedTritonExperts %s", self.moe) return BatchedTritonExperts( max_num_tokens=self.moe.max_num_tokens, num_dispatchers=prepare_finalize.num_dispatchers(), quant_config=self.moe_quant_config, ) else: logger.debug("TritonExperts %s", self.moe) return TritonExperts(self.moe_quant_config) def create_weights( self, layer: torch.nn.Module, num_experts: int, hidden_size: int, intermediate_size_per_partition: int, params_dtype: torch.dtype, **extra_weight_attrs, ): if self.moe.is_act_and_mul: w13_up_dim = 2 * intermediate_size_per_partition else: w13_up_dim = intermediate_size_per_partition # Fused gate_up_proj (column parallel) w13_weight = torch.nn.Parameter( torch.empty( num_experts, w13_up_dim, hidden_size, dtype=params_dtype, ), requires_grad=False, ) layer.register_parameter("w13_weight", w13_weight) set_weight_attrs(w13_weight, extra_weight_attrs) if self.moe.has_bias: w13_bias = torch.nn.Parameter( torch.zeros(num_experts, w13_up_dim, dtype=params_dtype), requires_grad=False, ) layer.register_parameter("w13_bias", w13_bias) set_weight_attrs(w13_bias, extra_weight_attrs) # down_proj (row parallel) w2_weight = torch.nn.Parameter( torch.empty( num_experts, hidden_size, intermediate_size_per_partition, dtype=params_dtype, ), requires_grad=False, ) layer.register_parameter("w2_weight", w2_weight) set_weight_attrs(w2_weight, extra_weight_attrs) if self.moe.has_bias: w2_bias = torch.nn.Parameter( torch.zeros(num_experts, hidden_size, dtype=params_dtype), requires_grad=False, ) layer.register_parameter("w2_bias", w2_bias) set_weight_attrs(w2_bias, extra_weight_attrs) def _maybe_pad_weight(self, weight: torch.Tensor) -> torch.Tensor: # Pad the weight tensor. This is an optimization on ROCm platform, which # can benefit from tensors located far enough from one another in memory if ( envs.VLLM_ROCM_MOE_PADDING and current_platform.is_rocm() and weight.stride(-1) == 1 and (weight.stride(-2) * weight.element_size()) % 512 == 0 ): num_pad = 256 // weight.element_size() weight = F.pad(weight, (0, num_pad), "constant", 0)[..., :-num_pad] torch.cuda.empty_cache() return weight def process_weights_after_loading(self, layer: torch.nn.Module) -> None: super().process_weights_after_loading(layer) # Padding the weight for better performance on ROCm layer.w13_weight.data = self._maybe_pad_weight(layer.w13_weight.data) layer.w2_weight.data = self._maybe_pad_weight(layer.w2_weight.data) if self.rocm_aiter_moe_enabled: shuffled_w13, shuffled_w2 = rocm_aiter_ops.shuffle_weights( layer.w13_weight.data, layer.w2_weight.data ) layer.w13_weight.data = shuffled_w13 layer.w2_weight.data = shuffled_w2 if current_platform.is_xpu(): import intel_extension_for_pytorch as ipex ep_rank_start = self.moe.ep_rank * self.moe.num_local_experts layer.ipex_fusion = ipex.llm.modules.GatedMLPMOE( layer.w13_weight, layer.w2_weight, use_prepack=True, experts_start_id=ep_rank_start, ) elif current_platform.is_cpu(): from vllm.model_executor.layers.fused_moe import cpu_fused_moe if current_platform.get_cpu_architecture() == CpuArchEnum.X86: from vllm.model_executor.layers.utils import check_cpu_sgl_kernel dtype_w13 = layer.w13_weight.dtype _, n_w13, k_w13 = layer.w13_weight.size() dtype_w2 = layer.w2_weight.dtype _, n_w2, k_w2 = layer.w2_weight.size() if ( envs.VLLM_CPU_SGL_KERNEL and check_cpu_sgl_kernel(n_w13, k_w13, dtype_w13) and check_cpu_sgl_kernel(n_w2, k_w2, dtype_w2) ): packed_w13_weight = torch.ops._C.convert_weight_packed( layer.w13_weight ) assert packed_w13_weight.size() == layer.w13_weight.size() layer.w13_weight.copy_(packed_w13_weight) del packed_w13_weight packed_w2_weight = torch.ops._C.convert_weight_packed( layer.w2_weight ) assert packed_w2_weight.size() == layer.w2_weight.size() layer.w2_weight.copy_(packed_w2_weight) layer.cpu_fused_moe = cpu_fused_moe.SGLFusedMOE(layer) else: layer.cpu_fused_moe = cpu_fused_moe.CPUFusedMOE(layer) else: layer.cpu_fused_moe = cpu_fused_moe.CPUFusedMOE(layer) elif current_platform.is_cuda_alike(): self.moe_quant_config = self.get_fused_moe_quant_config(layer) if self.flashinfer_cutlass_moe_enabled: self.use_inplace = False # Swap halves to arrange as [w3; w1] (kernel expectation) w13_weight = swap_w13_to_w31(layer.w13_weight.data) replace_parameter(layer, "w13_weight", w13_weight) self.kernel = mk.FusedMoEModularKernel( MoEPrepareAndFinalizeNoEP(), FlashInferExperts( out_dtype=layer.params_dtype, quant_config=self.moe_quant_config, tp_rank=self.moe.moe_parallel_config.tp_rank, tp_size=self.moe.moe_parallel_config.tp_size, ep_rank=self.moe.moe_parallel_config.ep_rank, ep_size=self.moe.moe_parallel_config.ep_size, ), ) else: self.use_inplace = True self.kernel = mk.FusedMoEModularKernel( MoEPrepareAndFinalizeNoEP(), TritonExperts(self.moe_quant_config), shared_experts=None, ) def apply( self, layer: "FusedMoE", # type: ignore[name-defined] # noqa: F821 x: torch.Tensor, router_logits: torch.Tensor, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: return self.forward( layer=layer, x=x, router_logits=router_logits, ) def get_fused_moe_quant_config(self, layer: torch.nn.Module) -> FusedMoEQuantConfig: if self.moe.has_bias: return biased_moe_quant_config( layer.w13_bias, layer.w2_bias, ) else: return FUSED_MOE_UNQUANTIZED_CONFIG def forward_cuda( self, layer: "FusedMoE", # type: ignore[name-defined] # noqa: F821 x: torch.Tensor, router_logits: torch.Tensor, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: topk_weights, topk_ids = layer.select_experts( hidden_states=x, router_logits=router_logits, ) if self.rocm_aiter_moe_enabled: result = self.rocm_aiter_fused_experts( hidden_states=x, w1=layer.w13_weight, w2=layer.w2_weight, topk_weights=topk_weights, topk_ids=topk_ids, expert_map=layer.expert_map, activation=layer.activation, apply_router_weight_on_input=layer.apply_router_weight_on_input, ) else: result = self.kernel( hidden_states=x, w1=layer.w13_weight, w2=layer.w2_weight, topk_weights=topk_weights, topk_ids=topk_ids, inplace=self.use_inplace, activation=layer.activation, apply_router_weight_on_input=layer.apply_router_weight_on_input, global_num_experts=layer.global_num_experts, expert_map=layer.expert_map, ) return result def forward_cpu( self, layer: "FusedMoE", # type: ignore[name-defined] # noqa: F821 x: torch.Tensor, router_logits: torch.Tensor, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: if ( layer.enable_eplb is not False or layer.expert_load_view is not None or layer.logical_to_physical_map is not None or layer.logical_replica_count is not None ): raise NotImplementedError("Expert load balancing is not supported for CPU.") return layer.cpu_fused_moe( layer, x, layer.use_grouped_topk, layer.top_k, router_logits, layer.renormalize, layer.topk_group, layer.num_expert_group, layer.global_num_experts, layer.expert_map, layer.custom_routing_function, layer.scoring_func, layer.routed_scaling_factor, layer.e_score_correction_bias, layer.apply_router_weight_on_input, layer.activation, ) def forward_xpu( self, layer: "FusedMoE", # type: ignore[name-defined] # noqa: F821 x: torch.Tensor, router_logits: torch.Tensor, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: if ( layer.enable_eplb is not False or layer.expert_load_view is not None or layer.logical_to_physical_map is not None or layer.logical_replica_count is not None ): raise NotImplementedError("Expert load balancing is not supported for XPU.") return layer.ipex_fusion( x, layer.use_grouped_topk, layer.top_k, router_logits, layer.renormalize, layer.topk_group, layer.num_expert_group, custom_routing_function=layer.custom_routing_function, ) def forward_tpu( self, layer: "FusedMoE", # type: ignore[name-defined] # noqa: F821 x: torch.Tensor, router_logits: torch.Tensor, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: assert not layer.use_grouped_topk assert layer.num_expert_group is None assert layer.topk_group is None assert layer.custom_routing_function is None assert layer.apply_router_weight_on_input is False if layer.scoring_func != "softmax": raise NotImplementedError( "Only softmax scoring function is supported for TPU." ) if layer.e_score_correction_bias is not None: raise NotImplementedError( "Expert score correction bias is not supported for TPU." ) assert layer.activation == "silu", ( f"{layer.activation} is not supported for TPU." ) assert layer.routed_scaling_factor == 1.0, ( f"routed_scaling_factor {layer.routed_scaling_factor} is " "not supported for TPU." ) if ( layer.enable_eplb is not False or layer.expert_load_view is not None or layer.logical_to_physical_map is not None or layer.logical_replica_count is not None ): raise NotImplementedError("Expert load balancing is not supported for TPU.") return fused_moe_pallas( hidden_states=x, w1=layer.w13_weight, w2=layer.w2_weight, topk=layer.top_k, gating_output=router_logits, global_num_experts=layer.global_num_experts, expert_map=layer.expert_map, renormalize=layer.renormalize, ) if current_platform.is_tpu(): forward_native = forward_tpu elif current_platform.is_cpu(): forward_native = forward_cpu elif current_platform.is_xpu(): forward_native = forward_xpu else: forward_native = forward_cuda
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/modular_kernel.py
vllm/model_executor/layers/fused_moe/modular_kernel.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import ABC, abstractmethod from collections.abc import Callable from dataclasses import dataclass from enum import Enum from math import prod from typing import final import torch import vllm.envs as envs from vllm.forward_context import get_forward_context, is_forward_context_available from vllm.logger import init_logger from vllm.model_executor.layers.fused_moe.config import ( FusedMoEParallelConfig, FusedMoEQuantConfig, ) from vllm.model_executor.layers.fused_moe.utils import ( _resize_cache, count_expert_num_tokens, disable_inplace, ) from vllm.utils.math_utils import cdiv from vllm.v1.worker.ubatching import ( dbo_enabled, dbo_maybe_run_recv_hook, dbo_register_recv_hook, dbo_yield, ) from vllm.v1.worker.workspace import current_workspace_manager logger = init_logger(__name__) # # This file defines a set of base classes used to make MoE kernels more modular. # The goal is to be able to utilize different communication mechanisms with # any fused MoE kernel without needing to have combinatoric implementations. # # The fused moe kernels are broken down into the following components: # # [Router] → [Quantize-Dispatch] → [Permute-Experts-Unpermute] → [Combine] # # Each component will be independent of (but may inform) the others except for # [Quantize-Dispatch] and `[Combine] (see below). The components can then be # mixed and matched with so that DP+EP can be supported easily for multiple # MoE kernel implementations. # # The following main classes are defined: # * FusedMoEPrepareAndFinalize - an abstract base class for preparation of MoE # inputs (e.g. quantization, distribution) and finalization of Moe outputs. # The prepare method must take care of any needed quantization and the # finalize method, informed by the FusedMoEPermuteExpertsUnpermute method, # may apply weights and/or do the final reduction of the output. # * FusedMoEPermuteExpertsUnpermute - an abstract base class for the main fused # MoE operation, i.e matmul + act_mul + optionally quant + matmul. # Some FusedMoEPermuteExpertsUnpermute implementations may choose to do # the weight application and/or reduction. The class communicates this # to [Finalize] via a TopKWeightAndReduce object. # * FusedMoEModularKernel - an interface class that combines a # FusedMoEPrepareAndFinalize and a FusedMoEPermuteExpertsUnpermute to # provide the standard fused MoE kernel interface. # * TopKWeightAndReduce - A TopKWeightAndReduce implementation chosen # by the FusedMoEPermuteExpertsUnpermute implementation that is passed # on to [Finalize]. # # [Quantize-Prepare] and [Finalize] functionality are bundled into a single # class `FusedMoEPrepareAndFinalize` since they could use collective # communication mechanisms that need to be consistent. # class FusedMoEActivationFormat(Enum): """ The standard activation format (num_tokens, hidden dim). """ Standard = ("standard",) """ The batched experts format (num experts, max tokens per expert, hidden dim) """ BatchedExperts = ("batched_experts",) @dataclass class ExpertTokensMetadata: """ Metadata regarding expert-token routing. """ expert_num_tokens: torch.Tensor expert_num_tokens_cpu: torch.Tensor | None @staticmethod def make_from_list( expert_num_tokens_list: list[int], device: str ) -> "ExpertTokensMetadata": expert_num_tokens_cpu = torch.tensor( expert_num_tokens_list, device="cpu", dtype=torch.int32 ) return ExpertTokensMetadata( expert_num_tokens=expert_num_tokens_cpu.to(device, non_blocking=True), expert_num_tokens_cpu=expert_num_tokens_cpu, ) class TopKWeightAndReduce(ABC): """ An abstract base class for weight application and reduction implementations. """ @abstractmethod def apply( self, output: torch.Tensor | None, fused_expert_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, apply_router_weight_on_input: bool, ) -> torch.Tensor: """ Apply topk_weights to the fused_experts_outputs and/or reduce. If an output tensor is not passed, it will be created in the function. """ raise NotImplementedError # # PrepareResultType is a tuple of: # - quantized + dispatched a. # - quantized + dispatched a1_scales. # - Optional ExpertTokensMetadata containing gpu/cpu tensors # as big as the number of local experts with the information about the # number of tokens assigned to each local expert. # - Optional dispatched expert topk IDs # - Optional dispatched expert topk weight # # See `prepare` method below. # PrepareResultType = tuple[ torch.Tensor, torch.Tensor | None, ExpertTokensMetadata | None, torch.Tensor | None, torch.Tensor | None, ] ReceiverType = Callable[[], PrepareResultType] # TODO: pass FusedMoEParallelConfig in as ctor parameter? class FusedMoEPrepareAndFinalize(ABC): """ An abstract base class for the [Quantize-Prepare] and [Finalize] steps described above. """ def post_init_setup(self, fused_experts: "FusedMoEPermuteExpertsUnpermute"): """ Initialize FusedMoEPrepareAndFinalize settings that depend on FusedMoEPermuteExpertsUnpermute experts object. The FusedMoEPrepareAndFinalize implementations that have such dependencies may choose to override this function. """ return @abstractmethod def prepare( self, a1: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_experts: int, expert_map: torch.Tensor | None, apply_router_weight_on_input: bool, quant_config: FusedMoEQuantConfig, ) -> PrepareResultType: """ Perform any quantization (and/or) dispatching needed for this kernel. - a1: The (unquantized) input to the MoE layer. - topk_ids: The topk ids. - topk_weights: The topk weights. - num_experts: The total number of experts in the global expert space. - expert_map: A tensor mapping expert indices from the global expert space to the local expert space of the expert parallel shard. - apply_router_weight_on_input: When True, apply the weights to the activations, before quantization + dispatching. - quant_config: Quantization info provided by the fused experts. Returns a tuple of: - quantized + dispatched a. - Optional quantized + dispatched a1_scales. - Optional ExpertTokensMetadata containing gpu/cpu tensors as big as the number of local experts with the information about the number of tokens assigned to each local expert. - Optional dispatched expert topk IDs - Optional dispatched expert topk weight """ raise NotImplementedError def supports_async(self) -> bool: """ Indicates whether or not this class implements prepare_async and finalize_async. """ return False def prepare_async( self, a1: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_experts: int, expert_map: torch.Tensor | None, apply_router_weight_on_input: bool, quant_config: FusedMoEQuantConfig, ) -> tuple[Callable, ReceiverType] | ReceiverType: """ Perform any quantization (and/or) dispatching needed for this kernel but do not wait for results from other workers. - a1: The (unquantized) input to the MoE layer. - a1_scale: Optional scales for a1 - a2_scale: Optional scales for the second MoE gemm. Required to make sure the quantization is consistent for both gemms. - topk_ids: The topk ids. - topk_weights: The topk weights. - num_experts: The total number of experts in the global expert space. - expert_map: A tensor mapping expert indices from the global expert space to the local expert space of the expert parallel shard. - apply_router_weight_on_input: When True, apply the weights to the activations, before quantization + dispatching. Returns a callback or a hook callback pair that when invoked waits for results from other workers and has the same return signature as `prepare`, if a hook is returned this is more lightweight check that the recv is complete without doing extra work (used by DBO, will be refactored in the very near future) e.g. ret = obj.prepare_async(...) if isinstance(ret, tuple): hook, receiver = ret hook() if hook is not None: a, a_scales, expert_meta, topk_ids, topk_weights = receiver() is equivalent to: a, a_scales, expert_meta, topk_ids, topk_weights = obj.prepare(...) """ raise NotImplementedError @abstractmethod def finalize( self, output: torch.Tensor, fused_expert_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, apply_router_weight_on_input: bool, weight_and_reduce_impl: TopKWeightAndReduce, ) -> None: """ Perform any combine plus apply weights and perform a reduction on the fused experts output. - output: The output tensor, written in place. Must be (M, K) shape. - fused_expert_output: The unweighted, unreduced output of the fused experts, it will have (M, topk, K) shape. - topk_weights: The weights to be applied to the fused_experts_output. - topk_ids: The topk_ids. - apply_router_weight_on_input: When False, apply the weights to fused_expert_output. - weight_and_reduce_impl: An optional TopKWeightAndReduce implementation. """ raise NotImplementedError def finalize_async( self, output: torch.Tensor, fused_expert_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, apply_router_weight_on_input: bool, weight_and_reduce_impl: TopKWeightAndReduce, ) -> tuple[Callable, Callable] | Callable: """ Perform any combine plus apply weights and perform a reduction on the fused experts output but do not wait for results from other workers. - output: The output tensor, written in place. Must be (M, K) shape. - fused_expert_output: The unweighted, unreduced output of the fused experts, it will have (M, topk, K) shape. - topk_weights: The weights to be applied to the fused_experts_output. - topk_ids: The topk_ids. - apply_router_weight_on_input: When False, apply the weights to fused_expert_output. - weight_and_reduce_impl: An optional TopKWeightAndReduce implementation. Returns a callback or a hook callback pair that when invoked waits for results from other workers and has the same return signature as `finalize`, if a hook is returned this is more lightweight check that the recv is complete without doing extra work (used by DBO, will be refactored in the very near future) ret = obj.finalize_async(output, ...) ... output not valid yet ... if isinstance(ret, tuple): hook, receiver = ret hook() receiver() ... output valid here ... is equivalent to: obj.finalize(output, ...) """ raise NotImplementedError @property @abstractmethod def activation_format(self) -> FusedMoEActivationFormat: """ A property indicating the output format of the activations for the 'prepare' method. """ raise NotImplementedError @abstractmethod def topk_indices_dtype(self) -> torch.dtype | None: """ The PrepareFinalize All2All implementations generally constrain the dtype of the topk_ids they support. This function returns the required topk indices dtype so it can be respected. Return None if there are no such restrictions. """ raise NotImplementedError @abstractmethod def max_num_tokens_per_rank(self) -> int | None: """ Some PrepareFinalize All2All implementations are batched. Meaning, they can process only as set of tokens at a time. This function returns the batch size i.e the maximum number of tokens the implementation can process at a time. Return None if there are no such restrictions. """ raise NotImplementedError @abstractmethod def num_dispatchers(self) -> int: raise NotImplementedError @abstractmethod def output_is_reduced(self) -> bool: """ Indicates whether or not the output of finalize is reduced across all ranks. """ raise NotImplementedError # TODO: add supported activations method (return string) class FusedMoEPermuteExpertsUnpermute(ABC): """ An abstract base class for the [Permute-Experts-Unpermute] step described above. """ def __init__( self, quant_config: FusedMoEQuantConfig, ): """ quant_config: Quantization parameters for this experts instance. """ self.quant_config = quant_config @property @abstractmethod def activation_formats( self, ) -> tuple[FusedMoEActivationFormat, FusedMoEActivationFormat]: """ A property which is a tuple of the input and output activation formats for the 'apply' method. """ raise NotImplementedError def moe_problem_size( self, a1: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_ids: torch.Tensor, ) -> tuple[int, int, int, int, int]: """ Extract the MoE problem size from the given tensor arguments: - a: The hidden states, input to the MoE layer. - w1: The first set of expert weights. - w2: The second set of expert weights. - topk_ids: The topk ids. Note: extracting the problem shape from the weight and activation tensors is not obvious. It needs to be done this way specifically due to subtle issues with particular kernels, e.g. the int4 kernels divide the trailing dimension by two, so it's not "correct" to extract N or K from the trailing dimension of w1 or w2. Similarly, some kernels transpose the weights, so this needs to be kept in mind. Note: This implementation covers most cases. However, if experts require a specialized implementation, like MarlinExperts, they are free to override this function. """ assert w1.dim() == 3 and w2.dim() == 3 E, N, _ = w1.size() K = a1.size(-1) if a1.dim() == 2: # Make sure we are using the correct a1 (pre-permute). assert topk_ids.size(0) == a1.size(0), f"{topk_ids.size(0)} != {a1.size(0)}" M = a1.size(0) else: assert a1.dim() == 3 assert a1.size(0) == E, f"{a1.size(0)} == {E}" M = a1.size(1) # This is max_num_tokens assert topk_ids.dim() == 2 topk = topk_ids.size(1) return E, M, N, K, topk # # Various helpers for accessing quantization parameters from the # quant_config. # @property def quant_dtype(self) -> torch.dtype | None: return self.quant_config.quant_dtype @property def block_shape(self) -> list[int] | None: return self.quant_config.block_shape @property def per_act_token_quant(self) -> bool: return self.quant_config.per_act_token_quant @property def per_out_ch_quant(self) -> bool: return self.quant_config.per_out_ch_quant @property def a1_scale(self) -> torch.Tensor | None: return self.quant_config.a1_scale @property def a2_scale(self) -> torch.Tensor | None: return self.quant_config.a2_scale @property def a1_gscale(self) -> torch.Tensor | None: return self.quant_config.a1_gscale @property def a2_gscale(self) -> torch.Tensor | None: return self.quant_config.a2_gscale @property def w1_scale(self) -> torch.Tensor | None: return self.quant_config.w1_scale @property def w2_scale(self) -> torch.Tensor | None: return self.quant_config.w2_scale @property def w1_zp(self) -> torch.Tensor | None: return self.quant_config.w1_zp @property def w2_zp(self) -> torch.Tensor | None: return self.quant_config.w2_zp @property def w1_bias(self) -> torch.Tensor | None: return self.quant_config.w1_bias @property def w2_bias(self) -> torch.Tensor | None: return self.quant_config.w2_bias @property def g1_alphas(self) -> torch.Tensor | None: return self.quant_config.g1_alphas @property def g2_alphas(self) -> torch.Tensor | None: return self.quant_config.g2_alphas # TODO (bnell): make this return a CHUNK_SIZE or None instead? @abstractmethod def supports_chunking(self) -> bool: """ A flag indicating whether or not this class supports activation chunking. """ raise NotImplementedError @abstractmethod def supports_expert_map(self) -> bool: """ A flag indicating whether or not this class supports expert maps """ raise NotImplementedError def supports_packed_ue8m0_act_scales(self) -> bool: """ A flag indicating whether or not this class can process packed ue8m0 activation scales. """ return False def workspace_dtype(self, act_dtype: torch.dtype) -> torch.dtype: """ Workspace type: The dtype to use for the workspace tensors. """ return act_dtype @abstractmethod def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: ExpertTokensMetadata | None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: """ Compute the shapes for the temporary and final outputs of the two gemms and activation in the fused expert function. Since the gemms are independent, the workspace for the first gemm can be shared with the workspace for the last gemm. Inputs: - M: number of tokens. - N: Row (or column) dimension of expert weights. - K: hidden dimension - topk: The number of top-k experts to select. - global_num_experts: global number of experts. - local_num_experts: local number of experts due to DP/EP. - expert_tokens_meta: number of tokens per expert metadata for batched format. Returns a tuple of: - workspace13 shape tuple: must be large enough to hold the result of either expert gemm. - workspace2 shape tuple: must be large enough to hold the result of the activation function. - output shape tuple: must be exact size of the final gemm output. - Note: workspace shapes can be 0 if the workspace is not needed. But in order for activation chunking to work, the first dimension of each tuple must be the number of tokens when the shape is not 0. """ raise NotImplementedError def activation( self, activation: str, output: torch.Tensor, input: torch.Tensor ) -> None: assert output.size(-1) * 2 == input.size(-1) if activation == "silu": torch.ops._C.silu_and_mul(output, input) elif activation == "gelu": torch.ops._C.gelu_and_mul(output, input) elif activation == "swigluoai": # alpha = 1.702, limit = 7.0 torch.ops._C.swigluoai_and_mul(output, input) else: raise ValueError(f"Unsupported FusedMoe activation: {activation}") def enable_chunking(self): return ( envs.VLLM_ENABLE_FUSED_MOE_ACTIVATION_CHUNKING and self.supports_chunking() ) def finalize_weight_and_reduce_impl(self) -> TopKWeightAndReduce: raise NotImplementedError @abstractmethod def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: str, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_tokens_meta: ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ) -> None: """ This function computes the intermediate result of a Mixture of Experts (MoE) layer using two sets of weights, w1 and w2. Parameters: - output: (torch.Tensor): The unweighted, unreduced output tensor. - hidden_states: (torch.Tensor): The (quantized) input tensor to the MoE layer. - w1 (torch.Tensor): The first set of expert weights. - w2 (torch.Tensor): The second set of expert weights. - topk_weights: A map of row to expert weights. Some implementations choose to do weight application. - topk_ids (torch.Tensor): A map of row to expert id. - activation (str): The activation function to apply after the first MoE layer. - global_num_experts (int): The total number of experts in the global expert space. - expert_map (Optional[torch.Tensor]): A tensor mapping expert indices from the global expert space to the local expert space of the expert parallel shard. - a1q_scale (Optional[torch.Tensor]): Optional quantized scale to be used for a1. Result of quantization from prepare/finalize and not from the FusedMoEQuantConfig. - workspace13 (torch.Tensor): A scratch tensor used for gemm outputs must be large enough to hold output of either MoE gemm. - workspace2 (torch.Tensor): A scratch tensor used for the activation function. - expert_tokens_meta (Optional[ExpertTokensMetadata]) - An optional ExpertTokensMetadata object containing gpu/cpu tensors as big as the number of local experts with the information about the number of tokens assigned to each local expert. - apply_router_weight_on_input: True if router weights are already applied on the input. This is relevant if the implementation chooses to do weight application. """ raise NotImplementedError def _slice_scales( scales: torch.Tensor | None, start: int, end: int ) -> torch.Tensor | None: if scales is not None: if scales.numel() == 1: return scales else: return scales[start:end] return None @final class FusedMoEModularKernel(torch.nn.Module): """ This class combines a FusedMoEPrepareAndFinalize instance and a FusedMoEPermuteExpertsUnpermute to provide an interface that is compatible with the `fused_experts` function in fused_moe.py. It takes care of managing any required scratch space. Note: Instances of this class should only be used for a single model layer due to any layer specific state that may be used by the component objects. """ def __init__( self, prepare_finalize: FusedMoEPrepareAndFinalize, fused_experts: FusedMoEPermuteExpertsUnpermute, shared_experts: torch.nn.Module | None = None, moe_parallel_config: FusedMoEParallelConfig | None = None, ): super().__init__() self.prepare_finalize = prepare_finalize self.fused_experts = fused_experts self.shared_experts = shared_experts # prefer an explicit FusedMoEParallelConfig when available (from # FusedMoE layers / tests). # if not provided, assume this kernel is # running in a non-DP+EP context self.moe_parallel_config: FusedMoEParallelConfig | None = moe_parallel_config self.is_dp_ep = ( moe_parallel_config is not None and moe_parallel_config.dp_size > 1 and moe_parallel_config.use_ep ) self._post_init_setup() assert ( prepare_finalize.activation_format == fused_experts.activation_formats[0] ), ( f"{prepare_finalize.__class__.__name__}." f"{prepare_finalize.activation_format} == " f"{fused_experts.__class__.__name__}." f"{fused_experts.activation_formats[0]}" ) def _post_init_setup(self): """ Resolve any leftover setup dependencies between self.prepare_finalize and self.fused_experts here. """ self.prepare_finalize.post_init_setup(self.fused_experts) def supports_expert_map(self) -> bool: """ A flag indicating whether or not this class supports expert maps. """ return self.fused_experts.supports_expert_map() def output_is_reduced(self) -> bool: """ Indicates whether or not the output of fused MoE kernel is reduced across all ranks. """ return self.prepare_finalize.output_is_reduced() def _chunk_info(self, M: int) -> tuple[int, int]: """ Compute number of chunks and chunk size for given M. If chunking is not supported, set the CHUNK_SIZE to M so we get num_chunks == 1. Take max(M, 1) to avoid divide by zero. If there are no tokens to process, the number of chunks will be zero. """ CHUNK_SIZE = max( 1, ( M if not self.fused_experts.enable_chunking() else min(M, envs.VLLM_FUSED_MOE_CHUNK_SIZE) ), ) num_chunks = cdiv(M, CHUNK_SIZE) # If there are no tokens, then there should be no loop iterations. assert M > 0 or num_chunks == 0 return num_chunks, CHUNK_SIZE def _allocate_buffers( self, out_dtype: torch.dtype, device: torch.device, M_chunk: int, M_full: int, N: int, K: int, top_k: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: ExpertTokensMetadata | None, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Allocate temporary and output buffers for the fused experts op. Inputs: - out_dtype: output type of workspace and output tensors. - device: the device of the workspace and output tensors. See `workspace_shapes` for a description of the remainder of arguments. Returns a tuple of (workspace13, workspace2, output) tensors. """ assert M_full > 0 and M_chunk > 0 num_chunks, _ = self._chunk_info(M_full) workspace_dtype = self.fused_experts.workspace_dtype(out_dtype) # Force worst-case allocation in profiling run for # "mk.FusedMoEModularKernel.Standard" formats where this is only bounded # by `VLLM_FUSED_MOE_CHUNK_SIZE` and may not be seen during profiling with # DP+EP due to the random token routing. is_profile_run = ( is_forward_context_available() and get_forward_context().attn_metadata is None ) if is_profile_run and self.fused_experts.enable_chunking() and self.is_dp_ep: max_workspace_13, max_workspace_2, max_fused_out_shape = ( self.fused_experts.workspace_shapes( envs.VLLM_FUSED_MOE_CHUNK_SIZE, N, K, top_k, global_num_experts, local_num_experts, # expert_tokens_meta help in allocating optimal/minimal # amount of workspace. Mark it None, so we allocate for # the worst-case scenario. expert_tokens_meta=None, ) ) current_workspace_manager().get_simultaneous( (max_workspace_13, workspace_dtype), (max_workspace_2, workspace_dtype), (max_fused_out_shape, out_dtype), ) # Get intermediate workspace shapes based off the chunked M size. workspace13_shape, workspace2_shape, _ = self.fused_experts.workspace_shapes( M_chunk, N, K, top_k, global_num_experts, local_num_experts, expert_tokens_meta, ) # Get final output shape based on the full M size. _, _, fused_out_shape = self.fused_experts.workspace_shapes( M_full, N, K, top_k, global_num_experts, local_num_experts, expert_tokens_meta, ) # We can reuse the memory between cache1 and cache3 because by the # time we need cache3, we're done with cache1. # Construct the entire output that can then be processed in chunks. # Reuse workspace13 for the output in the non-chunked case as long # as it is large enough. This will not always be the case for standard # format experts and with experts that have empty workspaces. if num_chunks == 1 and prod(workspace13_shape) >= prod(fused_out_shape): workspace13, workspace2 = current_workspace_manager().get_simultaneous( (workspace13_shape, workspace_dtype), (workspace2_shape, workspace_dtype), ) fused_out = _resize_cache(workspace13, fused_out_shape) else: workspace13, workspace2, fused_out = ( current_workspace_manager().get_simultaneous( (workspace13_shape, workspace_dtype), (workspace2_shape, workspace_dtype), (fused_out_shape, out_dtype), ) ) return workspace13, workspace2, fused_out @staticmethod def _slice_output_tensor( fused_out: torch.Tensor, chunk_idx: int, num_chunks: int, CHUNK_SIZE: int, M: int, ) -> torch.Tensor: if num_chunks == 1: return fused_out assert fused_out.size(0) % M == 0, f"fused_out shape {fused_out.shape} vs M {M}" factor = fused_out.size(0) // M out_chunk_size = CHUNK_SIZE * factor s = chunk_idx * out_chunk_size e = min(s + out_chunk_size, fused_out.size(0)) return fused_out[s:e] @staticmethod def _slice_expert_tokens_metadata( num_chunks: int, full_expert_tokens_meta: ExpertTokensMetadata | None, chunk_topk_ids: torch.Tensor, local_num_experts: int, expert_map: torch.Tensor | None, ) -> ExpertTokensMetadata | None: if num_chunks == 1 or full_expert_tokens_meta is None: return full_expert_tokens_meta # The existing expert_num_tokens is for the entire a1q # input. Chunking forces recomputation of the number # of tokens assigned to each expert. c_expert_num_tokens = count_expert_num_tokens( chunk_topk_ids, local_num_experts, expert_map ) c_expert_num_tokens_cpu = None need_expert_num_tokens_cpu = ( full_expert_tokens_meta.expert_num_tokens_cpu is not None ) if need_expert_num_tokens_cpu: # This is blocking as some implementations need the count # on the CPU to determine appropriate input/out fused-moe # buffers c_expert_num_tokens_cpu = c_expert_num_tokens.to("cpu", non_blocking=False) return ExpertTokensMetadata( expert_num_tokens=c_expert_num_tokens, expert_num_tokens_cpu=c_expert_num_tokens_cpu, ) def _prepare( self,
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/fused_moe_method_base.py
vllm/model_executor/layers/fused_moe/fused_moe_method_base.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import abstractmethod import torch from vllm.logger import init_logger from vllm.model_executor.layers.fused_moe.config import ( FusedMoEConfig, FusedMoEQuantConfig, ) from vllm.model_executor.layers.fused_moe.modular_kernel import ( FusedMoEPermuteExpertsUnpermute, FusedMoEPrepareAndFinalize, ) from vllm.model_executor.layers.quantization.base_config import ( QuantizeMethodBase, ) logger = init_logger(__name__) class FusedMoEMethodBase(QuantizeMethodBase): def __init__(self, moe: FusedMoEConfig): super().__init__() self.moe: FusedMoEConfig = moe self.moe_quant_config: FusedMoEQuantConfig | None = None @abstractmethod def create_weights( self, layer: torch.nn.Module, num_experts: int, hidden_size: int, intermediate_size_per_partition: int, params_dtype: torch.dtype, **extra_weight_attrs, ): raise NotImplementedError def uses_weight_scale_2_pattern(self) -> bool: """ Returns True if this quantization method uses 'weight_scale_2' pattern for per-tensor weight scales (e.g., FP4 variants), False otherwise. This method should be overridden by subclasses that use the 'weight_scale_2' pattern instead of the standard 'weight_scale' pattern. """ return False def maybe_make_prepare_finalize( self, routing_tables: tuple[torch.Tensor, torch.Tensor, torch.Tensor] | None = None, ) -> FusedMoEPrepareAndFinalize | None: from .all2all_utils import maybe_make_prepare_finalize return maybe_make_prepare_finalize( self.moe, self.moe_quant_config, routing_tables ) def select_gemm_impl( self, prepare_finalize: FusedMoEPrepareAndFinalize, layer: torch.nn.Module, ) -> FusedMoEPermuteExpertsUnpermute: # based on the all2all implementation, select the appropriate # gemm implementation raise NotImplementedError( f"{self.__class__.__name__} must select appropriate gemm " "implementation based on the prepare_finalize" ) def prepare_dp_allgather_tensor( self, layer: "FusedMoE", # type: ignore[name-defined] # noqa: F821 hidden_states: torch.Tensor, router_logits: torch.Tensor, ) -> tuple[torch.Tensor, list[torch.Tensor]]: """Hook to prepare tensors and extra tensors for DP allgather + EP dispatch.""" raise NotImplementedError( "Method 'prepare_dp_allgather_tensor' is not implemented in " f"{self.__class__.__name__}." ) @abstractmethod def get_fused_moe_quant_config( self, layer: torch.nn.Module ) -> FusedMoEQuantConfig | None: raise NotImplementedError @property def topk_indices_dtype(self) -> torch.dtype | None: return None @property def supports_eplb(self) -> bool: return False @property def allow_inplace(self) -> bool: return False @property def method_name(self) -> str: return self.__class__.__name__ @abstractmethod def apply( self, layer: "FusedMoE", # type: ignore[name-defined] # noqa: F821 x: torch.Tensor, router_logits: torch.Tensor, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: raise NotImplementedError
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/moe_pallas.py
vllm/model_executor/layers/fused_moe/moe_pallas.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import torch.nn.functional as F def _histogram(input: torch.Tensor, min: int, max: int) -> torch.Tensor: """ Compute the histogram of an int32 tensor. The bin edges are defined by the min and max values, with step = 1. """ assert input.dtype == torch.int32, "input must be of torch.int32 dtype." assert min <= max, "min must be less than or equal to max." def searchsorted( sorted_sequence: torch.Tensor, values_to_search: torch.Tensor ) -> torch.Tensor: return (sorted_sequence.unsqueeze(1) == values_to_search).sum(dim=1) bin_edges = torch.linspace(min, max, max - min + 1, dtype=input.dtype).to( input.device ) return searchsorted(bin_edges, input).to(torch.int32) def fused_moe( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, gating_output: torch.Tensor, topk: int, global_num_experts: int, expert_map: torch.Tensor = None, renormalize: bool = False, ) -> torch.Tensor: """ Args: hidden_states: [*, hidden_size] w1: [num_experts, intermediate_size * 2, hidden_size] w2: [num_experts, hidden_size, intermediate_size] gating_output: [*, num_experts] """ assert expert_map is None, "expert_map is not supported for pallas MoE." import torch_xla.experimental.custom_kernel # noqa: F401 orig_shape = hidden_states.shape hidden_size = hidden_states.shape[-1] num_tokens = hidden_states.shape[:-1].numel() num_experts = w1.shape[0] intermediate_size = w2.shape[-1] device = hidden_states.device dtype = hidden_states.dtype assert (num_tokens * topk) % 16 == 0, ( "The Pallas GMM kernel requires num_tokens * topk to be a multiple of " f"16 but got {num_tokens * topk}" ) hidden_states = hidden_states.view(num_tokens, hidden_size) gating_output = gating_output.view(num_tokens, num_experts) topk_weights = gating_output.softmax(dim=-1, dtype=torch.float) topk_weights, topk_indices = topk_weights.topk(topk, dim=-1) if renormalize: topk_weights = topk_weights / topk_weights.sum(dim=-1, keepdim=True) topk_weights = topk_weights.to(dtype) topk_indices = topk_indices.flatten() topk_argsort_indices = topk_indices.argsort() topk_argsort_revert_indices = topk_argsort_indices.argsort() token_indices = torch.arange(num_tokens, device=device).repeat_interleave(topk) token_indices = token_indices[topk_argsort_indices] group_sizes = _histogram(topk_indices.to(torch.int32), 0, num_experts - 1) x = hidden_states[token_indices] x = torch.ops.xla.gmm(x, w1, group_sizes, transpose_rhs=True) x = F.silu(x[..., :intermediate_size]) * x[..., intermediate_size:] x = torch.ops.xla.gmm(x, w2, group_sizes, transpose_rhs=True) x = x[topk_argsort_revert_indices].reshape(-1, topk, hidden_size) x = x * topk_weights.unsqueeze(dim=-1) x = x.sum(dim=-2) x = x.reshape(orig_shape) return x
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fused_moe/flashinfer_trtllm_moe.py
vllm/model_executor/layers/fused_moe/flashinfer_trtllm_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.model_executor.layers.fused_moe.config import RoutingMethodType from vllm.model_executor.layers.fused_moe.utils import moe_kernel_quantize_input from vllm.model_executor.layers.quantization.utils.flashinfer_utils import ( calculate_tile_tokens_dim, ) from vllm.model_executor.layers.quantization.utils.fp8_utils import ( per_token_group_quant_fp8, ) from vllm.utils.torch_utils import direct_register_custom_op def flashinfer_fused_moe_blockscale_fp8( routing_logits: torch.Tensor, routing_bias: torch.Tensor, x: torch.Tensor, w13_weight: torch.Tensor, w13_weight_scale_inv: torch.Tensor, w2_weight: torch.Tensor, w2_weight_scale_inv: torch.Tensor, global_num_experts: int, top_k: int, num_expert_group: int | None, topk_group: int | None, intermediate_size: int, expert_offset: int, local_num_experts: int, block_shape: list[int], routing_method_type: int = RoutingMethodType.DeepSeekV3, routed_scaling: float | None = 1.0, ) -> torch.Tensor: from vllm.utils.flashinfer import flashinfer_trtllm_fp8_block_scale_moe topk_group = topk_group if topk_group is not None else 0 assert top_k <= global_num_experts assert top_k <= 10 assert global_num_experts % 4 == 0 assert block_shape == [128, 128] # Routing kernel expects #experts <= #threads 512 assert global_num_experts <= 512 a_q, a_sf = per_token_group_quant_fp8(x, block_shape[1]) # NOTE: scales of hidden states have to be transposed! a_sf_t = a_sf.t().contiguous() return flashinfer_trtllm_fp8_block_scale_moe( routing_logits=routing_logits, routing_bias=routing_bias, hidden_states=a_q, hidden_states_scale=a_sf_t, gemm1_weights=w13_weight, gemm1_weights_scale=w13_weight_scale_inv, gemm2_weights=w2_weight, gemm2_weights_scale=w2_weight_scale_inv, num_experts=global_num_experts, top_k=top_k, n_group=num_expert_group, topk_group=topk_group, intermediate_size=intermediate_size, local_expert_offset=expert_offset, local_num_experts=local_num_experts, routed_scaling_factor=routed_scaling, tile_tokens_dim=None, routing_method_type=routing_method_type, use_shuffled_weight=False, ) def flashinfer_fused_moe_blockscale_fp8_fake( routing_logits: torch.Tensor, routing_bias: torch.Tensor, x: torch.Tensor, w13_weight: torch.Tensor, w13_weight_scale_inv: torch.Tensor, w2_weight: torch.Tensor, w2_weight_scale_inv: torch.Tensor, global_num_experts: int, top_k: int, num_expert_group: int, topk_group: int, intermediate_size: int, expert_offset: int, local_num_experts: int, block_shape: list[int], routing_method_type: int, routed_scaling: float = 1.0, ) -> torch.Tensor: return torch.empty_like(x) # TODO(bnell): Does this really need to be a torch.op? direct_register_custom_op( op_name="flashinfer_fused_moe_blockscale_fp8", op_func=flashinfer_fused_moe_blockscale_fp8, fake_impl=flashinfer_fused_moe_blockscale_fp8_fake, tags=(torch.Tag.needs_fixed_stride_order,), ) def flashinfer_fused_moe_per_tensor_scale_fp8( routing_logits: torch.Tensor, routing_bias: torch.Tensor | None, hidden_states: torch.Tensor, input_scale: torch.Tensor, gemm1_weights: torch.Tensor, gemm2_weights: torch.Tensor, output1_scales_scalar: torch.Tensor, output1_scales_gate_scalar: torch.Tensor, output2_scales_scalar: torch.Tensor, num_experts: int, top_k: int, num_expert_group: int | None, topk_group: int | None, intermediate_size: int, local_expert_offset: int, local_num_experts: int, use_routing_scales_on_input: bool, routing_method_type: int, routed_scaling_factor: float = 1.0, ) -> torch.Tensor: num_expert_group = num_expert_group if num_expert_group is not None else 0 topk_group = topk_group if topk_group is not None else 0 quant_hidden_states, _ = moe_kernel_quantize_input( hidden_states, input_scale, quant_dtype=torch.float8_e4m3fn, per_act_token_quant=False, ) from vllm.utils.flashinfer import flashinfer_trtllm_fp8_per_tensor_scale_moe return flashinfer_trtllm_fp8_per_tensor_scale_moe( routing_logits=routing_logits, routing_bias=routing_bias, hidden_states=quant_hidden_states, gemm1_weights=gemm1_weights, output1_scales_scalar=output1_scales_scalar, output1_scales_gate_scalar=output1_scales_gate_scalar, gemm2_weights=gemm2_weights, output2_scales_scalar=output2_scales_scalar, num_experts=num_experts, top_k=top_k, n_group=num_expert_group, topk_group=topk_group, intermediate_size=intermediate_size, local_expert_offset=local_expert_offset, local_num_experts=local_num_experts, routed_scaling_factor=routed_scaling_factor, use_routing_scales_on_input=use_routing_scales_on_input, tile_tokens_dim=calculate_tile_tokens_dim( hidden_states.shape[0], top_k, num_experts ), routing_method_type=routing_method_type, ) def flashinfer_fused_moe_per_tensor_scale_fp8_fake( routing_logits: torch.Tensor, routing_bias: torch.Tensor | None, hidden_states: torch.Tensor, input_scale: torch.Tensor, gemm1_weights: torch.Tensor, gemm2_weights: torch.Tensor, output1_scales_scalar: torch.Tensor, output1_scales_gate_scalar: torch.Tensor, output2_scales_scalar: torch.Tensor, num_experts: int, top_k: int, num_expert_group: int | None, topk_group: int | None, intermediate_size: int, local_expert_offset: int, local_num_experts: int, use_routing_scales_on_input: bool, routing_method_type: int, routed_scaling_factor: float = 1.0, ) -> torch.Tensor: return torch.empty_like(hidden_states) # TODO(bnell): Does this really need to be a torch.op? direct_register_custom_op( op_name="flashinfer_fused_moe_per_tensor_scale_fp8", op_func=flashinfer_fused_moe_per_tensor_scale_fp8, mutates_args=["hidden_states"], fake_impl=flashinfer_fused_moe_per_tensor_scale_fp8_fake, tags=(torch.Tag.needs_fixed_stride_order,), )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/dynamic_ntk_alpha_rope.py
vllm/model_executor/layers/rotary_embedding/dynamic_ntk_alpha_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from .base import RotaryEmbedding class DynamicNTKAlphaRotaryEmbedding(RotaryEmbedding): """RotaryEmbedding extended with Dynamic NTK alpha. Based on the original RotaryEmbedding implementation. """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_alpha: float, dtype: torch.dtype, ) -> None: self.scaling_alpha = scaling_alpha super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_cos_sin_cache(self) -> torch.Tensor: # For Hunyuan DynamicNTKAlphaRotaryEmbedding max_len = self.max_position_embeddings base = self.base * self.scaling_alpha ** ( self.rotary_dim / (self.rotary_dim - 2) ) inv_freq = self._compute_inv_freq(base) t = torch.arange(max_len, dtype=torch.float) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() sin = freqs.sin() cache = torch.cat((cos, sin), dim=-1) return cache
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/dynamic_ntk_scaling_rope.py
vllm/model_executor/layers/rotary_embedding/dynamic_ntk_scaling_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Adapted from # https://github.com/huggingface/transformers/blob/v4.33.2/src/transformers/models/llama/modeling_llama.py # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from .base import RotaryEmbedding class DynamicNTKScalingRotaryEmbedding(RotaryEmbedding): """RotaryEmbedding extended with Dynamic NTK scaling. Credits to the Reddit users /u/bloc97 and /u/emozilla """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_factor: float, dtype: torch.dtype, ) -> None: self.scaling_factor = scaling_factor super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_cos_sin_cache(self) -> torch.Tensor: # NOTE(woosuk): self.max_position_embeddings is the original # maximum length before applying the rope scaling. # Thus, the maximum length after applying the rope scaling is # self.max_position_embeddings * self.scaling_factor. max_len = self.max_position_embeddings * self.scaling_factor base = self.base * ( (self.scaling_factor * max_len / self.max_position_embeddings) - (self.scaling_factor - 1) ) ** (self.rotary_dim / (self.rotary_dim - 2)) inv_freq = self._compute_inv_freq(base) t = torch.arange(max_len, dtype=torch.float) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() sin = freqs.sin() cache = torch.cat((cos, sin), dim=-1) return cache
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/llama4_vision_rope.py
vllm/model_executor/layers/rotary_embedding/llama4_vision_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import torch from .base import RotaryEmbeddingBase class Llama4VisionRotaryEmbedding(RotaryEmbeddingBase): def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, ): super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_inv_freq(self, base: float) -> torch.Tensor: inv_freqs = super()._compute_inv_freq(base) inv_freqs = inv_freqs[: (self.rotary_dim // 2)] return inv_freqs def _compute_cos_sin_cache(self) -> torch.Tensor: inv_freq = self._compute_inv_freq(self.base) # self.max_position_embeddings here is number of image patches # i.e. (image_size // patch_size) ** 2 num_patches = self.max_position_embeddings img_idx = torch.arange(num_patches, dtype=torch.int32).reshape(num_patches, 1) img_idx = torch.cat([img_idx, img_idx[:1]], dim=0) img_idx[-1, -1] = -2 # set to ID_CLS_TOKEN num_patches_single_dim = int(math.sqrt(num_patches)) frequencies_x = img_idx % num_patches_single_dim frequencies_y = img_idx // num_patches_single_dim freqs_x = ( (frequencies_x + 1)[..., None] * inv_freq[None, None, :] ).repeat_interleave(2, dim=-1) freqs_y = ( (frequencies_y + 1)[..., None] * inv_freq[None, None, :] ).repeat_interleave(2, dim=-1) freqs = torch.cat([freqs_x, freqs_y], dim=-1).float().contiguous()[..., ::2] freqs = freqs.masked_fill(img_idx.reshape(-1, 1, 1) < 0, 0) cache = torch.view_as_complex( torch.stack([torch.cos(freqs), torch.sin(freqs)], dim=-1) ) return cache def forward_native( # type: ignore[override] self, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: assert key is not None # self.cos_sin_cache here is complex tensor so we cannot cast into # query's dtype directly with self._match_cos_sin_cache_dtype self.cos_sin_cache: torch.Tensor = self.cos_sin_cache.to(query.device) query_ = torch.view_as_complex(query.float().reshape(*query.shape[:-1], -1, 2)) key_ = torch.view_as_complex(key.float().reshape(*key.shape[:-1], -1, 2)) broadcast_shape = [ d if i == 1 or i == (query_.ndim - 1) else 1 for i, d in enumerate(query_.shape) ] freqs_ci = self.cos_sin_cache.view(*broadcast_shape) query_out = torch.view_as_real(query_ * freqs_ci).flatten(3) key_out = torch.view_as_real(key_ * freqs_ci).flatten(3) return query_out.type_as(query), key_out.type_as(key) def forward_cuda( # type: ignore[override] self, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: return self.forward_native(query, key)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/phi3_long_rope_scaled_rope.py
vllm/model_executor/layers/rotary_embedding/phi3_long_rope_scaled_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import torch import torch.nn as nn from vllm.config import get_current_vllm_config from vllm.logger import init_logger from .common import rotate_neox logger = init_logger(__name__) class Phi3LongRoPEScaledRotaryEmbedding(nn.Module): """Phi3 family of models scaled rotary embedding. Based on the original RotaryEmbedding implementation. """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, original_max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, short_factor: list[float], long_factor: list[float], short_mscale: float | None = None, long_mscale: float | None = None, ): super().__init__() if is_neox_style is False: raise ValueError( "`Phi3LongRoPEScaledRotaryEmbedding` only supports neox_style." ) self.rotary_dim = rotary_dim self.head_size = head_size self.max_position_embeddings = max_position_embeddings self.original_max_position_embeddings = original_max_position_embeddings self.base = base self.short_factor = short_factor self.long_factor = long_factor # Force long factors if max_model_len (runtime max length) exceeds # original_max_position_embeddings to prevent KV cache invalidation when # sequences cross this threshold during generation max_model_len = get_current_vllm_config().model_config.max_model_len self.use_long_rope = max_model_len > original_max_position_embeddings if self.use_long_rope: logger.warning_once( "Using LongRoPE scaling factors. This enables longer " "contexts (%d tokens vs original %d tokens) at the cost of " "some performance degradation for shorter sequences. If " "this is not desired, set `max_model_len` to be at most %d.", max_position_embeddings, original_max_position_embeddings, original_max_position_embeddings, ) scale = self.max_position_embeddings / self.original_max_position_embeddings if scale <= 1.0: scaling_factor = 1.0 else: scaling_factor = math.sqrt( 1 + math.log(scale) / math.log(self.original_max_position_embeddings) ) if short_mscale is None: short_mscale = scaling_factor if long_mscale is None: long_mscale = scaling_factor self.short_mscale = short_mscale self.long_mscale = long_mscale short_cache = self._compute_cos_sin_cache( original_max_position_embeddings, short_factor, short_mscale ) short_cache = short_cache.to(dtype) long_cache = self._compute_cos_sin_cache( max_position_embeddings, long_factor, long_mscale ) long_cache = long_cache.to(dtype) long_short_cache = torch.cat([short_cache, long_cache], dim=0) self.register_buffer( "long_short_cos_sin_cache", long_short_cache, persistent=False ) def _compute_inv_freq(self, rescale_factors: list[float]) -> torch.Tensor: rescale_factors = torch.tensor(rescale_factors, dtype=torch.float32) inv_freq = 1.0 / ( rescale_factors * ( self.base ** ( torch.arange(0, self.rotary_dim, 2, dtype=torch.float) / self.rotary_dim ) ) ) return inv_freq def _compute_cos_sin_cache( self, max_position_embeddings: int, rescale_factors: list[float], mscale: float, ) -> torch.Tensor: inv_freq = self._compute_inv_freq(rescale_factors) t = torch.arange(max_position_embeddings, dtype=torch.float) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() * mscale sin = freqs.sin() * mscale cache = torch.cat((cos, sin), dim=-1) return cache def forward( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: assert key is not None query = query.view(*query.shape[:-1], -1, self.head_size) key = key.view(*key.shape[:-1], -1, self.head_size) if self.use_long_rope: k = self.original_max_position_embeddings long_prompt_offset = torch.full_like(positions, k).long() idx = torch.add(positions, long_prompt_offset) else: idx = positions idx = torch.add(idx, offsets) if offsets is not None else idx cos_sin = torch.index_select(self.long_short_cos_sin_cache, 0, idx) cos, sin = cos_sin.chunk(2, dim=-1) cos = cos.repeat(1, 2).unsqueeze(-2) sin = sin.repeat(1, 2).unsqueeze(-2) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = query_rot * cos + rotate_neox(query_rot) * sin query = torch.cat((query_rot, query_pass), dim=-1) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = key_rot * cos + rotate_neox(key_rot) * sin key = torch.cat((key_rot, key_pass), dim=-1) return query.flatten(-2), key.flatten(-2)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/dual_chunk_rope.py
vllm/model_executor/layers/rotary_embedding/dual_chunk_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.model_executor.custom_op import CustomOp from .common import rotate_gptj, rotate_neox @CustomOp.register("dual_chunk_rotary_embedding") class DualChunkRotaryEmbedding(CustomOp): """Rotary positional embedding for Dual Chunk Attention.""" def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, chunk_size: int, local_size: int, ) -> None: super().__init__() self.head_size = head_size self.rotary_dim = rotary_dim self.max_position_embeddings = max_position_embeddings self.base = base self.is_neox_style = is_neox_style self.chunk_size = chunk_size self.local_size = local_size self.dtype = dtype self.device = torch.device(f"cuda:{torch.cuda.current_device()}") (q_cache, qc_cache, k_cache, qc_no_clamp_cache, q_inter_cache) = ( self._compute_cos_sin_cache() ) self.register_buffer("cos_sin_q_cache", q_cache, persistent=False) self.register_buffer("cos_sin_qc_cache", qc_cache, persistent=False) self.register_buffer("cos_sin_k_cache", k_cache, persistent=False) self.register_buffer( "cos_sin_qc_no_clamp_cache", qc_no_clamp_cache, persistent=False ) self.register_buffer("cos_sin_q_inter_cache", q_inter_cache, persistent=False) def _compute_inv_freq(self, base: float) -> torch.Tensor: """Compute the inverse frequency.""" # NOTE(woosuk): The HF implementation uses `torch.arange(...).float()`. # However, we use `torch.arange(..., dtype=torch.float)` instead to # avoid numerical issues with large base values (e.g., 10000000). # This may cause a slight numerical difference between the HF # implementation and ours. # NOTE(woosuk): To exactly match the HF implementation, we need to # use CPU to compute the cache and then move it to GPU. However, we # create the cache on GPU for faster initialization. This may cause # a slight numerical difference between the HF implementation and ours. inv_freq = 1.0 / ( base ** ( torch.arange(0, self.rotary_dim, 2, dtype=torch.float) / self.rotary_dim ) ) return inv_freq def _compute_cos_sin_cache(self) -> torch.Tensor: """Compute the cos and sin cache.""" inv_freq = self._compute_inv_freq(self.base) chunk_len = self.chunk_size - self.local_size q_t = torch.arange(chunk_len, dtype=torch.float) qc_t = (torch.arange(chunk_len, dtype=torch.float) + chunk_len).clamp( max=self.chunk_size ) k_t = torch.arange(self.max_position_embeddings, dtype=torch.float) % chunk_len # count from chunk_len, no clamp(self.chunk_size) restriction qc_no_clamp_t = torch.arange(chunk_len, dtype=torch.float) + chunk_len # count from self.chunk_size for q_inter's rope q_inter_t = torch.arange(chunk_len, dtype=torch.float) + self.chunk_size q_freqs = torch.outer(q_t, inv_freq) qc_freqs = torch.outer(qc_t, inv_freq) k_freqs = torch.outer(k_t, inv_freq) qc_no_clamp_freqs = torch.outer(qc_no_clamp_t, inv_freq) q_inter_freqs = torch.outer(q_inter_t, inv_freq) q_cos = q_freqs.cos() q_sin = q_freqs.sin() qc_cos = qc_freqs.cos() qc_sin = qc_freqs.sin() k_cos = k_freqs.cos() k_sin = k_freqs.sin() qc_no_clamp_cos = qc_no_clamp_freqs.cos() qc_no_clamp_sin = qc_no_clamp_freqs.sin() q_inter_cos = q_inter_freqs.cos() q_inter_sin = q_inter_freqs.sin() q_cache = torch.cat((q_cos, q_sin), dim=-1).to( dtype=self.dtype, device=self.device ) qc_cache = torch.cat((qc_cos, qc_sin), dim=-1).to( dtype=self.dtype, device=self.device ) k_cache = torch.cat((k_cos, k_sin), dim=-1).to( dtype=self.dtype, device=self.device ) qc_no_clamp_cache = torch.cat((qc_no_clamp_cos, qc_no_clamp_sin), dim=-1).to( dtype=self.dtype, device=self.device ) q_inter_cache = torch.cat((q_inter_cos, q_inter_sin), dim=-1).to( dtype=self.dtype, device=self.device ) return q_cache, qc_cache, k_cache, qc_no_clamp_cache, q_inter_cache def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: query = query.view(*query.shape[:-1], -1, self.head_size) key = key.view(*key.shape[:-1], -1, self.head_size) query_rot = query[..., : self.rotary_dim] key_rot = key[..., : self.rotary_dim] if self.rotary_dim < self.head_size: query_pass = query[..., self.rotary_dim :] key_pass = key[..., self.rotary_dim :] else: query_pass = None key_pass = None positions_with_offsets = ( torch.add(positions, offsets) if offsets is not None else positions ) key = self._apply_rotary_embedding( self.cos_sin_k_cache[positions_with_offsets], key_rot, key_pass ) chunk_len = self.chunk_size - self.local_size query = self._apply_rotary_embedding( self.cos_sin_q_cache[positions_with_offsets % chunk_len], query_rot, query_pass, ) query_succ = self._apply_rotary_embedding( self.cos_sin_qc_cache[positions_with_offsets % chunk_len], query_rot, query_pass, ) query_inter = self._apply_rotary_embedding( self.cos_sin_qc_cache[chunk_len - 1].repeat(positions.shape[0], 1), query_rot, query_pass, ) query_succ_critical = self._apply_rotary_embedding( self.cos_sin_qc_no_clamp_cache[positions_with_offsets % chunk_len], query_rot, query_pass, ) query_inter_critical = self._apply_rotary_embedding( self.cos_sin_q_inter_cache[positions_with_offsets % chunk_len], query_rot, query_pass, ) # merge query into one tensor to simplify the interfaces query = torch.cat( ( query, query_succ, query_inter, query_succ_critical, query_inter_critical, ), dim=-1, ) return query, key def forward_cuda( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: return self.forward_native(positions, query, key, offsets) def _apply_rotary_embedding(self, cos_sin, hidden_rot, hidden_pass): cos, sin = cos_sin.chunk(2, dim=-1) if self.is_neox_style: # NOTE(woosuk): Here we assume that the positions tensor has the # shape [batch_size, seq_len]. cos = cos.repeat(1, 1, 2).unsqueeze(-2) sin = sin.repeat(1, 1, 2).unsqueeze(-2) else: cos = cos.repeat_interleave(2, dim=-1).unsqueeze(-2) sin = sin.repeat_interleave(2, dim=-1).unsqueeze(-2) rotate_fn = rotate_neox if self.is_neox_style else rotate_gptj hidden_rot = hidden_rot * cos + rotate_fn(hidden_rot) * sin if self.rotary_dim < self.head_size: hidden = torch.cat((hidden_rot, hidden_pass), dim=-1) else: hidden = hidden_rot return hidden.flatten(-2).squeeze(0) def extra_repr(self) -> str: s = f"head_size={self.head_size}, rotary_dim={self.rotary_dim}" s += f", max_position_embeddings={self.max_position_embeddings}" s += f", base={self.base}, is_neox_style={self.is_neox_style}" s += f", chunk_size={self.chunk_size}, local_size={self.local_size}" return s
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/deepseek_scaling_rope.py
vllm/model_executor/layers/rotary_embedding/deepseek_scaling_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import torch from vllm.platforms import current_platform from vllm.utils.flashinfer import has_flashinfer from .base import RotaryEmbeddingBase from .common import ( rotate_gptj, rotate_neox, yarn_find_correction_range, yarn_linear_ramp_mask, ) def yarn_get_mscale(scale: float = 1, mscale: float = 1) -> float: if scale <= 1: return 1.0 return 0.1 * mscale * math.log(scale) + 1.0 class DeepseekScalingRotaryEmbedding(RotaryEmbeddingBase): """RotaryEmbedding extended with YaRN method. Credits to Peng et al. github.com/jquesnelle/yarn """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_factor: float, dtype: torch.dtype, *, extrapolation_factor: float = 1, attn_factor: float = 1, beta_fast: int = 32, beta_slow: int = 1, mscale: float = 1, mscale_all_dim: float = 0, ) -> None: self.scaling_factor = scaling_factor self.extrapolation_factor = extrapolation_factor self.attn_factor = attn_factor self.beta_fast = beta_fast self.beta_slow = beta_slow # Get n-d magnitude scaling corrected for interpolation. self.mscale = float( yarn_get_mscale(self.scaling_factor, float(mscale)) / yarn_get_mscale(self.scaling_factor, float(mscale_all_dim)) * attn_factor ) self.use_flashinfer = ( self.enabled() and dtype in (torch.float16, torch.bfloat16) and current_platform.is_cuda() and has_flashinfer() and head_size in [64, 128, 256, 512] ) super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_inv_freq(self, scaling_factor: float) -> torch.Tensor: pos_freqs = self.base ** ( torch.arange( 0, self.rotary_dim, 2, dtype=torch.float, device=current_platform.device_type, ) / self.rotary_dim ) inv_freq_extrapolation = 1.0 / pos_freqs inv_freq_interpolation = 1.0 / (scaling_factor * pos_freqs) low, high = yarn_find_correction_range( self.beta_fast, self.beta_slow, self.rotary_dim, self.base, self.max_position_embeddings, ) # Get n-d rotational scaling corrected for extrapolation inv_freq_mask = ( 1 - yarn_linear_ramp_mask(low, high, self.rotary_dim // 2, dtype=torch.float) ) * self.extrapolation_factor inv_freq = ( inv_freq_interpolation * (1 - inv_freq_mask) + inv_freq_extrapolation * inv_freq_mask ) return inv_freq def _compute_cos_sin_cache(self) -> torch.Tensor: inv_freq = self._compute_inv_freq(self.scaling_factor) t = torch.arange( self.max_position_embeddings * self.scaling_factor, device=current_platform.device_type, dtype=torch.float32, ) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() * self.mscale sin = freqs.sin() * self.mscale cache = torch.cat((cos, sin), dim=-1) return cache def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """PyTorch-native implementation equivalent to forward().""" assert key is not None self._match_cos_sin_cache_dtype(query) query_rot = query[..., : self.rotary_dim] key_rot = key[..., : self.rotary_dim] if self.rotary_dim < self.head_size: query_pass = query[..., self.rotary_dim :] key_pass = key[..., self.rotary_dim :] cos_sin = self.cos_sin_cache[ torch.add(positions, offsets) if offsets is not None else positions ] cos, sin = cos_sin.chunk(2, dim=-1) if self.is_neox_style: # NOTE(woosuk): Here we assume that the positions tensor has the # shape [batch_size, seq_len]. cos = cos.repeat(1, 1, 2).unsqueeze(-2) sin = sin.repeat(1, 1, 2).unsqueeze(-2) else: cos = cos.repeat_interleave(2, dim=-1).unsqueeze(-2) sin = sin.repeat_interleave(2, dim=-1).unsqueeze(-2) rotate_fn = rotate_neox if self.is_neox_style else rotate_gptj query_rot = query_rot * cos + rotate_fn(query_rot) * sin key_rot = key_rot * cos + rotate_fn(key_rot) * sin if self.rotary_dim < self.head_size: query = torch.cat((query_rot, query_pass), dim=-1) key = torch.cat((key_rot, key_pass), dim=-1) else: query = query_rot key = key_rot return query, key def forward_hip( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: return self.forward_native(positions, query, key, offsets) def forward_cuda( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: if self.use_flashinfer: torch.ops.vllm.flashinfer_rotary_embedding( torch.add(positions, offsets) if offsets is not None else positions, query, key, self.head_size, self.cos_sin_cache, self.is_neox_style, ) return query, key else: return self.forward_native(positions, query, key, offsets)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/common.py
vllm/model_executor/layers/rotary_embedding/common.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math from importlib.util import find_spec import torch from vllm.logger import init_logger from vllm.model_executor.custom_op import CustomOp from vllm.utils.torch_utils import direct_register_custom_op logger = init_logger(__name__) # common functions def rotate_neox(x: torch.Tensor) -> torch.Tensor: x1 = x[..., : x.shape[-1] // 2] x2 = x[..., x.shape[-1] // 2 :] return torch.cat((-x2, x1), dim=-1) def rotate_gptj(x: torch.Tensor) -> torch.Tensor: x1 = x[..., ::2] x2 = x[..., 1::2] x = torch.stack((-x2, x1), dim=-1) return x.flatten(-2) # yarn functions # Inverse dim formula to find dim based on number of rotations def yarn_find_correction_dim( num_rotations: int, dim: int, base: float = 10000, max_position_embeddings: int = 2048, ) -> float: return (dim * math.log(max_position_embeddings / (num_rotations * 2 * math.pi))) / ( 2 * math.log(base) ) # Find dim range bounds based on rotations def yarn_find_correction_range( low_rot: int, high_rot: int, dim: int, base: float = 10000, max_position_embeddings: int = 2048, truncate: bool = True, ) -> tuple[float | int, float | int]: low = yarn_find_correction_dim(low_rot, dim, base, max_position_embeddings) high = yarn_find_correction_dim(high_rot, dim, base, max_position_embeddings) if truncate: low = math.floor(low) high = math.ceil(high) return max(low, 0), min(high, dim - 1) # Clamp values just in case def yarn_linear_ramp_mask( low: float, high: float, dim: int, dtype: torch.dtype ) -> torch.Tensor: if low == high: high += 0.001 # Prevent singularity linear_func = (torch.arange(dim, dtype=dtype) - low) / (high - low) ramp_func = torch.clamp(linear_func, 0, 1) return ramp_func def yarn_get_mscale(scale: float = 1) -> float: if scale <= 1: return 1.0 return 0.1 * math.log(scale) + 1.0 def _flashinfer_rotary_embedding( positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor, head_size: int, cos_sin_cache: torch.Tensor, is_neox: bool, ) -> None: """Custom op wrapper for flashinfer's rotary embedding. This is an in-place operation that modifies query and key tensors directly. """ from flashinfer.rope import apply_rope_with_cos_sin_cache_inplace apply_rope_with_cos_sin_cache_inplace( positions=positions, query=query, key=key, head_size=head_size, cos_sin_cache=cos_sin_cache, is_neox=is_neox, ) def _flashinfer_rotary_embedding_fake( positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor, head_size: int, cos_sin_cache: torch.Tensor, is_neox: bool, ) -> None: return # Register flashinfer rotary embedding custom op direct_register_custom_op( op_name="flashinfer_rotary_embedding", op_func=_flashinfer_rotary_embedding, mutates_args=["query", "key"], # These tensors are modified in-place fake_impl=_flashinfer_rotary_embedding_fake, ) @CustomOp.register("apply_rotary_emb") class ApplyRotaryEmb(CustomOp): def __init__( self, enforce_enable: bool = False, is_neox_style: bool = True, enable_fp32_compute: bool = False, ) -> None: super().__init__(enforce_enable) self.is_neox_style = is_neox_style self.enable_fp32_compute = enable_fp32_compute self.apply_rotary_emb_flash_attn = None if find_spec("flash_attn") is not None: from flash_attn.ops.triton.rotary import apply_rotary self.apply_rotary_emb_flash_attn = apply_rotary @staticmethod def forward_static( x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, is_neox_style: bool = True, enable_fp32_compute: bool = False, ) -> torch.Tensor: """ Args: x: [batch_size (optional), seq_len, num_heads, head_size] cos: [seq_len, head_size // 2] sin: [seq_len, head_size // 2] is_neox_style: Whether to use the Neox-style or GPT-J-style. enable_fp32_compute: Temporarily convert x, cos, sin to FP32 dtype for higher accuracy. """ origin_dtype = x.dtype if enable_fp32_compute: x = x.float() cos = cos.unsqueeze(-2).to(x.dtype) sin = sin.unsqueeze(-2).to(x.dtype) if is_neox_style: x1, x2 = torch.chunk(x, 2, dim=-1) else: x1 = x[..., ::2] x2 = x[..., 1::2] o1 = x1 * cos - x2 * sin o2 = x2 * cos + x1 * sin if is_neox_style: output = torch.cat((o1, o2), dim=-1) else: output = torch.stack((o1, o2), dim=-1).flatten(-2) if enable_fp32_compute: output = output.to(origin_dtype) return output def _pre_process( self, x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Size, torch.dtype]: origin_shape = x.shape if len(origin_shape) == 3: # x: [seq_len, num_heads, head_size] x = x.unsqueeze(0) origin_dtype = x.dtype if self.enable_fp32_compute: x = x.float() cos = cos.float() sin = sin.float() return x, cos, sin, origin_shape, origin_dtype def _post_process( self, output: torch.Tensor, origin_shape: torch.Size, origin_dtype: torch.dtype, ) -> torch.Tensor: if len(origin_shape) == 3: output = output.squeeze(0) if self.enable_fp32_compute: output = output.to(origin_dtype) return output def forward_native( self, x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, ) -> torch.Tensor: output = self.forward_static( x, cos, sin, self.is_neox_style, self.enable_fp32_compute ) return output def forward_cuda( self, x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, ) -> torch.Tensor: from vllm.vllm_flash_attn.layers.rotary import apply_rotary_emb x, cos, sin, origin_shape, origin_dtype = self._pre_process(x, cos, sin) """ Arguments of apply_rotary_emb() in vllm_flash_attn: x: [batch_size, seq_len, nheads, headdim] cos, sin: [seqlen_rotary, rotary_dim / 2] interleaved: defalut as False (Neox-style). ... """ interleaved = not self.is_neox_style output = apply_rotary_emb(x, cos, sin, interleaved) output = self._post_process(output, origin_shape, origin_dtype) return output def forward_hip( self, x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, ) -> torch.Tensor: if self.apply_rotary_emb_flash_attn is not None: x, cos, sin, origin_shape, origin_dtype = self._pre_process(x, cos, sin) """ Arguments of apply_rotary() in flash_attn: x: [batch_size, seq_len, nheads, headdim] cos, sin: [seqlen_rotary, rotary_dim / 2] interleaved: defalut as False (Neox-style). ... """ interleaved = not self.is_neox_style output = self.apply_rotary_emb_flash_attn( x, cos, sin, interleaved=interleaved ).type_as(x) output = self._post_process(output, origin_shape, origin_dtype) else: # Falling back to PyTorch native implementation. output = self.forward_native(x, cos, sin) return output def forward_cpu( self, x: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, ) -> torch.Tensor: # TODO (bigPYJ1151): need to enable fused CPU ROPE here return self.forward_native(x, cos, sin) def extra_repr(self) -> str: s = f"is_neox_style={self.is_neox_style}" s += f"enable_fp32_compute={self.enable_fp32_compute}" return s
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/mrope.py
vllm/model_executor/layers/rotary_embedding/mrope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import numpy as np import torch from vllm.triton_utils import tl, triton from .base import RotaryEmbeddingBase from .yarn_scaling_rope import YaRNScalingRotaryEmbedding, yarn_get_mscale @triton.jit def _triton_mrope_forward( q_ptr, k_ptr, cos, sin, num_tokens, n_qh: tl.constexpr, n_kh: tl.constexpr, hd: tl.constexpr, rd: tl.constexpr, pad_n_qh: tl.constexpr, pad_n_kh: tl.constexpr, pad_hd: tl.constexpr, mrope_section_t: tl.constexpr, mrope_section_h: tl.constexpr, mrope_section_w: tl.constexpr, is_interleaved: tl.constexpr, ): # Adapted from # https://github.com/linkedin/Liger-Kernel/blob/main/src/liger_kernel/ops/qwen2vl_mrope.py # This version supports flatten input tensors from vllm # and supports cos and sin cache with shape (3, num_tokens, head_dim // 2) # instead of (3, bsz, seq_len, head_dim), also supports interleaved rotary pid = tl.program_id(0) # locate start address q_ptr = q_ptr + pid * (n_qh * hd) k_ptr = k_ptr + pid * (n_kh * hd) # #################################################################### # get the cos(mθ_{i...d/2}) and sin(mθ_{i...d/2}) for token position # m of this program instance # #################################################################### # Note: cos and sin now have shape (3, num_tokens, head_dim // 2) # Updated stride calculation for half head_dim half_rd = rd // 2 t_cos = cos + pid * half_rd h_cos = t_cos + num_tokens * half_rd w_cos = h_cos + num_tokens * half_rd t_sin = sin + pid * half_rd h_sin = t_sin + num_tokens * half_rd w_sin = h_sin + num_tokens * half_rd # Updated offsets for half head_dim cos_offsets = tl.arange(0, pad_hd // 2) if is_interleaved: h_mask = ((cos_offsets % 3) == 1) & (cos_offsets <= 3 * mrope_section_h) w_mask = ((cos_offsets % 3) == 2) & (cos_offsets <= 3 * mrope_section_w) t_mask = ~(h_mask | w_mask) else: t_end = mrope_section_t h_end = t_end + mrope_section_h t_mask = cos_offsets < mrope_section_t h_mask = (t_end <= cos_offsets) & (cos_offsets < h_end) w_mask = (h_end <= cos_offsets) & (cos_offsets < half_rd) t_cos_row = tl.load(t_cos + cos_offsets, mask=t_mask, other=0) h_cos_row = tl.load(h_cos + cos_offsets, mask=h_mask, other=0) w_cos_row = tl.load(w_cos + cos_offsets, mask=w_mask, other=0) t_sin_row = tl.load(t_sin + cos_offsets, mask=t_mask, other=0) h_sin_row = tl.load(h_sin + cos_offsets, mask=h_mask, other=0) w_sin_row = tl.load(w_sin + cos_offsets, mask=w_mask, other=0) cos_row = t_cos_row + h_cos_row + w_cos_row sin_row = t_sin_row + h_sin_row + w_sin_row # #################################################################### # Load the left and right half of q and k for the current # program instance (i.e. for the current token) separately # #################################################################### # left half of the head first_half_q_offsets = ( tl.arange(0, pad_n_qh)[:, None] * hd + tl.arange(0, pad_hd // 2)[None, :] ) first_half_k_offsets = ( tl.arange(0, pad_n_kh)[:, None] * hd + tl.arange(0, pad_hd // 2)[None, :] ) first_q_mask = (tl.arange(0, pad_n_qh)[:, None] < n_qh) & ( tl.arange(0, pad_hd // 2)[None, :] < rd // 2 ) first_k_mask = (tl.arange(0, pad_n_kh)[:, None] < n_kh) & ( tl.arange(0, pad_hd // 2)[None, :] < rd // 2 ) q_tile_1 = tl.load(q_ptr + first_half_q_offsets, mask=first_q_mask, other=0).to( sin_row.dtype ) k_tile_1 = tl.load(k_ptr + first_half_k_offsets, mask=first_k_mask, other=0).to( sin_row.dtype ) # right half of the head second_half_q_offsets = first_half_q_offsets + (rd // 2) second_half_k_offsets = first_half_k_offsets + (rd // 2) second_q_mask = first_q_mask second_k_mask = first_k_mask q_tile_2 = tl.load(q_ptr + second_half_q_offsets, mask=second_q_mask, other=0).to( sin_row.dtype ) k_tile_2 = tl.load(k_ptr + second_half_k_offsets, mask=second_k_mask, other=0).to( sin_row.dtype ) # y = [x1, x2] * [cos, cos] + [-x2, x1] * [sin, sin] # Since cos and sin are now half-size, # we use the same cos_row and sin_row for both halves new_q_tile_1 = q_tile_1 * cos_row - q_tile_2 * sin_row tl.store(q_ptr + first_half_q_offsets, new_q_tile_1, mask=first_q_mask) new_q_tile_2 = q_tile_2 * cos_row + q_tile_1 * sin_row tl.store(q_ptr + second_half_q_offsets, new_q_tile_2, mask=second_q_mask) new_k_tile_1 = k_tile_1 * cos_row - k_tile_2 * sin_row tl.store(k_ptr + first_half_k_offsets, new_k_tile_1, mask=first_k_mask) new_k_tile_2 = k_tile_2 * cos_row + k_tile_1 * sin_row tl.store(k_ptr + second_half_k_offsets, new_k_tile_2, mask=second_k_mask) def triton_mrope( q: torch.Tensor, k: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, mrope_section: list[int], head_size: int, rotary_dim: int, mrope_interleaved: bool, ) -> tuple[torch.Tensor, torch.Tensor]: """Qwen2VL mrope kernel. Args: q: [num_tokens, num_heads * head_size] k: [num_tokens, num_kv_heads * head_size] cos: [3, num_tokens, head_size //2 ] (T/H/W positions with multimodal inputs) sin: [3, num_tokens, head_size //2 ] (T/H/W positions with multimodal inputs) mrope_section: [t, h, w] head_size: int """ n_row, n_q_head_head_dim = q.shape n_q_head = n_q_head_head_dim // head_size n_kv_head = k.shape[1] // head_size pad_hd = triton.next_power_of_2(head_size) pad_n_q_head = triton.next_power_of_2(n_q_head) pad_n_kv_head = triton.next_power_of_2(n_kv_head) # ensure tensors passed into the kernel are contiguous. # It will be no-op if they are already contiguous q = q.contiguous() k = k.contiguous() cos = cos.contiguous() sin = sin.contiguous() _triton_mrope_forward[(n_row,)]( q, k, cos, sin, n_row, n_q_head, n_kv_head, head_size, rotary_dim, pad_n_q_head, pad_n_kv_head, pad_hd, mrope_section[0], mrope_section[1], mrope_section[2], mrope_interleaved, ) return q, k def apply_interleaved_rope(x: torch.Tensor, mrope_section: list[int]) -> torch.Tensor: """Apply interleaved MRoPE to 3D rotary embeddings. Reorganizes frequency layout from chunked [TTT...HHH...WWW] to interleaved [THTHWHTHW...TT], preserving frequency continuity. """ x_t = x[0].clone() x_t[..., 1 : mrope_section[1] * 3 : 3] = x[1, ..., 1 : mrope_section[1] * 3 : 3] x_t[..., 2 : mrope_section[2] * 3 : 3] = x[2, ..., 2 : mrope_section[2] * 3 : 3] return x_t class MRotaryEmbedding(RotaryEmbeddingBase): """Rotary Embedding with Multimodal Sections.""" def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, mrope_section: list[int] | None = None, mrope_interleaved: bool = False, # YaRN parameters. *, scaling_factor: float | None = None, extrapolation_factor: float = 1, attn_factor: float = 1, beta_fast: int = 32, beta_slow: int = 1, ) -> None: self.scaling_factor = scaling_factor self.extrapolation_factor = extrapolation_factor self.attn_factor = attn_factor self.beta_fast = beta_fast self.beta_slow = beta_slow if self.scaling_factor is not None: # Get n-d magnitude scaling corrected for interpolation self.mscale = float(yarn_get_mscale(self.scaling_factor) * attn_factor) else: self.mscale = 1.0 # In Qwen2.5-VL, the maximum index value is related to the duration of # the input video. We enlarge max_position_embeddings to 4 times to get # a larger the cos and sin cache. self.cache_max_position_num = max_position_embeddings * 4 super().__init__( head_size, rotary_dim, self.cache_max_position_num, base, is_neox_style, dtype, ) self.mrope_section = mrope_section self.mrope_interleaved = mrope_interleaved if self.mrope_section: assert sum(self.mrope_section) == rotary_dim // 2 def _compute_inv_freq(self, base: float) -> torch.Tensor: if self.scaling_factor is None: return super()._compute_inv_freq(base) return YaRNScalingRotaryEmbedding._compute_inv_freq(self, base) def _compute_cos_sin_cache(self) -> torch.Tensor: if self.scaling_factor is None: return super()._compute_cos_sin_cache() return YaRNScalingRotaryEmbedding._compute_cos_sin_cache(self) def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """PyTorch-native implementation equivalent to forward(). Args: positions: [num_tokens,] (text only) or [3, num_tokens] (T/H/W positions with multimodal inputs) query: [num_tokens, num_heads * head_size] key: [num_tokens, num_kv_heads * head_size] """ assert positions.ndim == 1 or positions.ndim == 2 assert key is not None self._match_cos_sin_cache_dtype(query) num_tokens = positions.shape[-1] cos_sin = self.cos_sin_cache[positions] cos, sin = cos_sin.chunk(2, dim=-1) if positions.ndim == 2: assert self.mrope_section if self.mrope_interleaved: cos = apply_interleaved_rope(cos, self.mrope_section) sin = apply_interleaved_rope(sin, self.mrope_section) else: cos = torch.cat( [m[i] for i, m in enumerate(cos.split(self.mrope_section, dim=-1))], dim=-1, ) sin = torch.cat( [m[i] for i, m in enumerate(sin.split(self.mrope_section, dim=-1))], dim=-1, ) query_shape = query.shape query = query.view(num_tokens, -1, self.head_size) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = self.apply_rotary_emb.forward_native( query_rot, cos, sin, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) key_shape = key.shape key = key.view(num_tokens, -1, self.head_size) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = self.apply_rotary_emb.forward_native( key_rot, cos, sin, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key def forward_cuda( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: assert positions.ndim == 1 or positions.ndim == 2 assert key is not None self._match_cos_sin_cache_dtype(query) num_tokens = positions.shape[-1] cos_sin = self.cos_sin_cache[positions] cos, sin = cos_sin.chunk(2, dim=-1) query_shape = query.shape key_shape = key.shape if positions.ndim == 2: assert self.mrope_section q, k = triton_mrope( query, key, cos, sin, self.mrope_section, self.head_size, self.rotary_dim, self.mrope_interleaved, ) return q.reshape(query_shape), k.reshape(key_shape) query = query.view(num_tokens, -1, self.head_size) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = self.apply_rotary_emb( query_rot, cos, sin, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) key = key.view(num_tokens, -1, self.head_size) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = self.apply_rotary_emb( key_rot, cos, sin, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key def forward_cpu( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: return self.forward_native(positions, query, key, offsets) @staticmethod def get_next_input_positions( mrope_position_delta: int, context_len: int, seq_len: int, ) -> list[list[int]]: return [ list( range( context_len + mrope_position_delta, seq_len + mrope_position_delta ) ) for _ in range(3) ] @staticmethod def get_next_input_positions_tensor( out: np.ndarray, out_offset: int, mrope_position_delta: int, context_len: int, num_new_tokens: int, ): values = np.arange( mrope_position_delta + context_len, mrope_position_delta + context_len + num_new_tokens, dtype=out.dtype, ) out[:, out_offset : out_offset + num_new_tokens] = values
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/yarn_scaling_rope.py
vllm/model_executor/layers/rotary_embedding/yarn_scaling_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from .base import RotaryEmbedding from .common import yarn_find_correction_range, yarn_get_mscale, yarn_linear_ramp_mask class YaRNScalingRotaryEmbedding(RotaryEmbedding): """RotaryEmbedding extended with YaRN method. Credits to Peng et al. github.com/jquesnelle/yarn """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_factor: float, dtype: torch.dtype, *, extrapolation_factor: float = 1, attn_factor: float = 1, beta_fast: int = 32, beta_slow: int = 1, apply_yarn_scaling: bool = True, truncate: bool = True, ) -> None: self.scaling_factor = scaling_factor self.extrapolation_factor = extrapolation_factor self.attn_factor = attn_factor self.beta_fast = beta_fast self.beta_slow = beta_slow self.truncate = truncate # Get n-d magnitude scaling corrected for interpolation self.mscale = ( float(yarn_get_mscale(self.scaling_factor) * attn_factor) if apply_yarn_scaling else float(attn_factor) ) super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_inv_freq(self, scaling_factor: float) -> torch.Tensor: pos_freqs = self.base ** ( torch.arange(0, self.rotary_dim, 2, dtype=torch.float) / self.rotary_dim ) inv_freq_extrapolation = 1.0 / pos_freqs inv_freq_interpolation = 1.0 / (scaling_factor * pos_freqs) low, high = yarn_find_correction_range( self.beta_fast, self.beta_slow, self.rotary_dim, self.base, self.max_position_embeddings, self.truncate, ) # Get n-d rotational scaling corrected for extrapolation inv_freq_mask = ( 1 - yarn_linear_ramp_mask(low, high, self.rotary_dim // 2, dtype=torch.float) ) * self.extrapolation_factor inv_freq = ( inv_freq_interpolation * (1 - inv_freq_mask) + inv_freq_extrapolation * inv_freq_mask ) return inv_freq def _compute_cos_sin_cache(self) -> torch.Tensor: inv_freq = self._compute_inv_freq(self.scaling_factor) t = torch.arange( self.max_position_embeddings * self.scaling_factor, dtype=torch.float32 ) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() * self.mscale sin = freqs.sin() * self.mscale cache = torch.cat((cos, sin), dim=-1) return cache
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/xdrope.py
vllm/model_executor/layers/rotary_embedding/xdrope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import numpy as np import torch from .dynamic_ntk_alpha_rope import DynamicNTKAlphaRotaryEmbedding class XDRotaryEmbedding(DynamicNTKAlphaRotaryEmbedding): """DynamicNTKAlphaRotaryEmbedding extended with MultiModal(XD) Sections. Based on the original DynamicNTKAlphaRotaryEmbedding implementation. """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_alpha: float, dtype: torch.dtype, xdrope_section: list[int], ) -> None: self.xdrope_section = xdrope_section super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, scaling_alpha, dtype, ) def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """PyTorch-native implementation equivalent to forward(). Args: positions: [4, num_tokens] (P/W/H/T positions with multimodal inputs) query: [num_tokens, num_heads * head_size] key: [num_tokens, num_kv_heads * head_size] """ assert positions.ndim == 2 assert key is not None num_tokens = positions.shape[-1] cos_sin = self.cos_sin_cache[positions] cos, sin = cos_sin.chunk(2, dim=-1) cos = torch.cat( [m[i] for i, m in enumerate(cos.split(self.xdrope_section, dim=-1))], dim=-1 ) sin = torch.cat( [m[i] for i, m in enumerate(sin.split(self.xdrope_section, dim=-1))], dim=-1 ) query_shape = query.shape query = query.view(num_tokens, -1, self.head_size) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = self.apply_rotary_emb.forward_native( query_rot, cos, sin, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) key_shape = key.shape key = key.view(num_tokens, -1, self.head_size) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = self.apply_rotary_emb.forward_native( key_rot, cos, sin, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key def forward_cuda( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, offsets: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """PyTorch-native implementation equivalent to forward(). Args: positions: [4, num_tokens] (P/W/H/T positions with multimodal inputs) query: [num_tokens, num_heads * head_size] key: [num_tokens, num_kv_heads * head_size] """ assert positions.ndim == 2 assert key is not None num_tokens = positions.shape[-1] cos_sin = self.cos_sin_cache[positions] cos, sin = cos_sin.chunk(2, dim=-1) cos = torch.cat( [m[i] for i, m in enumerate(cos.split(self.xdrope_section, dim=-1))], dim=-1 ) sin = torch.cat( [m[i] for i, m in enumerate(sin.split(self.xdrope_section, dim=-1))], dim=-1 ) query_shape = query.shape query = query.view(num_tokens, -1, self.head_size) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = self.apply_rotary_emb( query_rot, cos, sin, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) key_shape = key.shape key = key.view(num_tokens, -1, self.head_size) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = self.apply_rotary_emb( key_rot, cos, sin, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key @staticmethod def get_next_input_positions( context_len: int, seq_len: int, xd_sections: int = 4, ) -> list[list[int]]: return [list(range(context_len, seq_len)) for _ in range(xd_sections)] @staticmethod def get_next_input_positions_tensor( out: np.ndarray, out_offset: int, context_len: int, num_new_tokens: int, ): values = np.arange( context_len, context_len + num_new_tokens, dtype=out.dtype, ) out[:, out_offset : out_offset + num_new_tokens] = values
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/linear_scaling_rope.py
vllm/model_executor/layers/rotary_embedding/linear_scaling_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Adapted from # https://github.com/huggingface/transformers/blob/v4.33.2/src/transformers/models/llama/modeling_llama.py # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from .base import RotaryEmbedding class LinearScalingRotaryEmbedding(RotaryEmbedding): """RotaryEmbedding extended with linear scaling. It supports multiple scaling factors. Since multiple LoRA adapters may have different scaling factors, we need multiple cos/sin caches. In this way, instead of running rotary embedding kernel per lora, we can run multiple lora in a batched way. In addition to that, we also keep the cos/sin cache for the scaling factor of 1 (default) at all times. Exemplary for two scaling factors x=1, y and z with embeddings [[x11, x12, ... x1m], ..., [xn1, xn2, ..., xnm]] and [[y11, y12, ... y1o], ..., [yn1, yn2, ..., yno]], and [[z11, z12, ... z1p], ..., [zn1, zn2, ..., znp]], we construct the cos/sin cache as follows: [[x11, x12, ... x1m, y11, y12, ... y1o, z11, z12, ... z1p], ... [xn1, xn2, ... xnm, yn1, yn2, ... yno, zn1, zn2, ... znp]] We then use offsets to index into the cos/sin cache for the respective scaling factors. The offset to cache can be accessed via `scaling_factor_to_offset` API. Credits to the Reddit user /u/kaiokendev """ def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_factors: list[float] | float, dtype: torch.dtype, ) -> None: if isinstance(scaling_factors, float): scaling_factors = [scaling_factors] self.scaling_factors: list[float] = scaling_factors # noqa super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) # Lazy initialized. self._scaling_factor_to_offset: dict[float, int] def _compute_cos_sin_cache(self) -> torch.Tensor: inv_freq = self._compute_inv_freq(self.base) cache_list: list[torch.Tensor] = [] # offsets to the next cache in a tensor. # Each offset corresponds to the same index in scaling_factors. offsets: list[int] = [] for scaling_factor in self.scaling_factors: # NOTE(woosuk): self.max_position_embeddings is the original # maximum length before applying the rope scaling. # Thus, the maximum length after applying the rope scaling is # self.max_position_embeddings * self.scaling_factor. max_len = self.max_position_embeddings * scaling_factor t = torch.arange(max_len, dtype=torch.float) t = t / scaling_factor freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() sin = freqs.sin() cache = torch.cat((cos, sin), dim=-1) if not cache_list: offset = 0 else: last_offset = offsets[-1] next_max_len = cache_list[-1].shape[0] offset = last_offset + next_max_len offsets.append(offset) cache_list.append(cache) self._scaling_factor_to_offset = { float(scaling_factor): offsets[i] for i, scaling_factor in enumerate(self.scaling_factors) } assert len(self.scaling_factors) == len(offsets) return torch.cat(cache_list, dim=0) @property def scaling_factor_to_offset(self) -> dict[float, int]: return self._scaling_factor_to_offset
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/ntk_scaling_rope.py
vllm/model_executor/layers/rotary_embedding/ntk_scaling_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from .base import RotaryEmbedding class NTKScalingRotaryEmbedding(RotaryEmbedding): """RotaryEmbedding extended with fixed and mixed NTK scaling. https://kexue.fm/archives/9706""" def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, scaling_factor: float, dtype: torch.dtype, mixed_b: float | None = None, ) -> None: self.scaling_factor = scaling_factor self.mixed_b = mixed_b super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_inv_freq(self, base: float) -> torch.Tensor: base = self.base * (self.scaling_factor if self.mixed_b is None else 1) inv_freq = super()._compute_inv_freq(base) if self.mixed_b is None: inv_freq = inv_freq / self.scaling_factor ** (2 / self.rotary_dim) else: a = ( torch.tensor(self.scaling_factor).log() / (self.rotary_dim / 2) ** self.mixed_b ) lambda_1_m = ( a * torch.arange(1, self.rotary_dim // 2 + 1).float() ** self.mixed_b ).exp() inv_freq = inv_freq / lambda_1_m return inv_freq
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/ernie45_vl_rope.py
vllm/model_executor/layers/rotary_embedding/ernie45_vl_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from .mrope import MRotaryEmbedding class Ernie4_5_VLRotaryEmbedding(MRotaryEmbedding): """3D rotary positional embedding. 3D is t:time h:height w:width""" def forward_native( # type: ignore[override] self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: assert positions.ndim == 1 or positions.ndim == 2 assert key is not None num_tokens = positions.shape[-1] cos_sin = self.cos_sin_cache[positions] cos, sin = cos_sin.chunk(2, dim=-1) if positions.ndim == 2: assert self.mrope_section section_h = self.mrope_section[0] # 22 section_w = self.mrope_section[1] # 22 section_t = self.mrope_section[2] # 20 assert section_h == section_w # Split according to [h w h w h w h w... t t t...] section_cos_t = cos[..., -section_t:] section_cos_h = cos[..., : section_h + section_w : 2] section_cos_w = cos[..., 1 : section_h + section_w : 2] cos_t, cos_h, cos_w = section_cos_t[0], section_cos_h[1], section_cos_w[2] cos_hw = torch.stack([cos_h, cos_w], dim=-1).reshape( cos_h.shape[:-1] + (cos_h.shape[-1] * 2,) ) cos = torch.cat([cos_hw, cos_t], dim=-1) section_sin_t = sin[..., -section_t:] section_sin_h = sin[..., : section_h + section_w : 2] section_sin_w = sin[..., 1 : section_h + section_w : 2] sin_t, sin_h, sin_w = section_sin_t[0], section_sin_h[1], section_sin_w[2] sin_hw = torch.stack([sin_h, sin_w], dim=-1).reshape( sin_h.shape[:-1] + (sin_h.shape[-1] * 2,) ) sin = torch.cat([sin_hw, sin_t], dim=-1) query_shape = query.shape query = query.view(num_tokens, -1, self.head_size) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = self.apply_rotary_emb.forward_native( query_rot, cos, sin, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) key_shape = key.shape key = key.view(num_tokens, -1, self.head_size) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = self.apply_rotary_emb.forward_native( key_rot, cos, sin, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key def forward_cuda( # type: ignore[override] self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: return self.forward_native(positions, query, key)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/__init__.py
vllm/model_executor/layers/rotary_embedding/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Rotary Positional Embeddings.""" from typing import Any import torch from .base import RotaryEmbedding from .deepseek_scaling_rope import DeepseekScalingRotaryEmbedding from .dual_chunk_rope import DualChunkRotaryEmbedding from .dynamic_ntk_alpha_rope import DynamicNTKAlphaRotaryEmbedding from .dynamic_ntk_scaling_rope import DynamicNTKScalingRotaryEmbedding from .linear_scaling_rope import LinearScalingRotaryEmbedding from .llama3_rope import Llama3RotaryEmbedding from .llama4_vision_rope import Llama4VisionRotaryEmbedding from .mrope import MRotaryEmbedding from .ntk_scaling_rope import NTKScalingRotaryEmbedding from .phi3_long_rope_scaled_rope import Phi3LongRoPEScaledRotaryEmbedding from .xdrope import XDRotaryEmbedding from .yarn_scaling_rope import YaRNScalingRotaryEmbedding _ROPE_DICT: dict[tuple, RotaryEmbedding] = {} def get_rope( head_size: int, max_position: int, is_neox_style: bool = True, rope_parameters: dict[str, Any] | None = None, dtype: torch.dtype | None = None, dual_chunk_attention_config: dict[str, Any] | None = None, ) -> RotaryEmbedding: if dtype is None: dtype = torch.get_default_dtype() if rope_parameters is not None: # Transforms every value that is a list into a tuple for caching calls rope_parameters_tuple = { k: tuple(v) if isinstance(v, list) else v for k, v in rope_parameters.items() } rope_parameters_args = tuple(rope_parameters_tuple.items()) else: rope_parameters_args = None if dual_chunk_attention_config is not None: dual_chunk_attention_tuple = { k: tuple(v) if isinstance(v, list) else v for k, v in dual_chunk_attention_config.items() if k != "sparse_attention_config" } dual_chunk_attention_args = tuple(dual_chunk_attention_tuple.items()) else: dual_chunk_attention_args = None rope_parameters = rope_parameters or {} base = rope_parameters.get("rope_theta", 10000) scaling_type = rope_parameters.get("rope_type", "default") partial_rotary_factor = rope_parameters.get("partial_rotary_factor", 1.0) if partial_rotary_factor <= 0.0 or partial_rotary_factor > 1.0: raise ValueError(f"{partial_rotary_factor=} must be between 0.0 and 1.0") rotary_dim = int(head_size * partial_rotary_factor) key = ( head_size, rotary_dim, max_position, is_neox_style, rope_parameters_args, dual_chunk_attention_args, dtype, ) if key in _ROPE_DICT: return _ROPE_DICT[key] if dual_chunk_attention_config is not None: extra_kwargs = { k: v for k, v in dual_chunk_attention_config.items() if k in ("chunk_size", "local_size") } rotary_emb = DualChunkRotaryEmbedding( head_size, rotary_dim, max_position, base, is_neox_style, dtype, **extra_kwargs, ) elif scaling_type == "default": if "mrope_section" in rope_parameters: rotary_emb = MRotaryEmbedding( head_size, rotary_dim, max_position, base, is_neox_style, dtype, mrope_section=rope_parameters["mrope_section"], mrope_interleaved=rope_parameters.get("mrope_interleaved", False), ) else: rotary_emb = RotaryEmbedding( head_size, rotary_dim, max_position, base, is_neox_style, dtype, ) elif scaling_type == "llama3": scaling_factor = rope_parameters["factor"] low_freq_factor = rope_parameters["low_freq_factor"] high_freq_factor = rope_parameters["high_freq_factor"] original_max_position = rope_parameters["original_max_position_embeddings"] rotary_emb = Llama3RotaryEmbedding( head_size, rotary_dim, max_position, base, is_neox_style, dtype, scaling_factor, low_freq_factor, high_freq_factor, original_max_position, ) elif scaling_type == "mllama4": rotary_emb = Llama4VisionRotaryEmbedding( head_size, rotary_dim, max_position, base, is_neox_style, dtype ) elif scaling_type == "linear": scaling_factor = rope_parameters["factor"] rotary_emb = LinearScalingRotaryEmbedding( head_size, rotary_dim, max_position, base, is_neox_style, scaling_factor, dtype, ) elif scaling_type == "ntk": scaling_factor = rope_parameters["factor"] mixed_b = rope_parameters.get("mixed_b") rotary_emb = NTKScalingRotaryEmbedding( head_size, rotary_dim, max_position, base, is_neox_style, scaling_factor, dtype, mixed_b, ) elif scaling_type == "dynamic": if "alpha" in rope_parameters: scaling_alpha = rope_parameters["alpha"] rotary_emb = DynamicNTKAlphaRotaryEmbedding( head_size, rotary_dim, max_position, base, is_neox_style, scaling_alpha, dtype, ) elif "factor" in rope_parameters: scaling_factor = rope_parameters["factor"] rotary_emb = DynamicNTKScalingRotaryEmbedding( head_size, rotary_dim, max_position, base, is_neox_style, scaling_factor, dtype, ) else: raise ValueError( "Dynamic rope scaling must contain either 'alpha' or 'factor' field" ) elif scaling_type == "xdrope": scaling_alpha = rope_parameters["alpha"] rotary_emb = XDRotaryEmbedding( head_size, rotary_dim, max_position, base, is_neox_style, scaling_alpha, dtype, xdrope_section=rope_parameters["xdrope_section"], ) elif scaling_type == "yarn": scaling_factor = rope_parameters["factor"] original_max_position = rope_parameters["original_max_position_embeddings"] extra_kwargs = { k: v for k, v in rope_parameters.items() if k in ( "extrapolation_factor", "attn_factor", "beta_fast", "beta_slow", "apply_yarn_scaling", "truncate", ) } if "mrope_section" in rope_parameters: extra_kwargs.pop("apply_yarn_scaling", None) rotary_emb = MRotaryEmbedding( head_size, rotary_dim, original_max_position, base, is_neox_style, dtype, mrope_section=rope_parameters["mrope_section"], mrope_interleaved=rope_parameters.get("mrope_interleaved", False), scaling_factor=scaling_factor, **extra_kwargs, ) else: rotary_emb = YaRNScalingRotaryEmbedding( head_size, rotary_dim, original_max_position, base, is_neox_style, scaling_factor, dtype, **extra_kwargs, ) elif scaling_type in ["deepseek_yarn", "deepseek_llama_scaling"]: scaling_factor = rope_parameters["factor"] original_max_position = rope_parameters["original_max_position_embeddings"] # assert max_position == original_max_position * scaling_factor extra_kwargs = { k: v for k, v in rope_parameters.items() if k in ( "extrapolation_factor", "attn_factor", "beta_fast", "beta_slow", "mscale", "mscale_all_dim", ) } rotary_emb = DeepseekScalingRotaryEmbedding( head_size, rotary_dim, original_max_position, base, is_neox_style, scaling_factor, dtype, **extra_kwargs, ) elif scaling_type == "longrope": short_factor = rope_parameters["short_factor"] long_factor = rope_parameters["long_factor"] original_max_position = rope_parameters["original_max_position_embeddings"] extra_kwargs = { k: v for k, v in rope_parameters.items() if k in ("short_mscale", "long_mscale") } rotary_emb = Phi3LongRoPEScaledRotaryEmbedding( head_size, rotary_dim, max_position, original_max_position, base, is_neox_style, dtype, short_factor, long_factor, **extra_kwargs, ) else: raise ValueError(f"Unknown RoPE scaling type {scaling_type}") _ROPE_DICT[key] = rotary_emb return rotary_emb
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/base.py
vllm/model_executor/layers/rotary_embedding/base.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Rotary Positional Embeddings Base Class.""" import torch from vllm._aiter_ops import rocm_aiter_ops from vllm.model_executor.custom_op import CustomOp from .common import ApplyRotaryEmb @CustomOp.register("rotary_embedding") class RotaryEmbeddingBase(CustomOp): """Original rotary positional embedding.""" def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, ) -> None: super().__init__() self.head_size = head_size self.rotary_dim = rotary_dim self.max_position_embeddings = max_position_embeddings self.base = base self.is_neox_style = is_neox_style self.dtype = dtype # TODO(mgoin): disabled for now due to failures # Flashinfer only supports head_size=64, 128, 256, 512. # https://github.com/flashinfer-ai/flashinfer/blob/ebfd655efe830048dba5d582aaa61d61d1cf9a87/include/flashinfer/utils.cuh#L174-L202 # self.use_flashinfer = (self.enabled() # and dtype in (torch.float16, torch.bfloat16) # and current_platform.is_cuda() # and has_flashinfer() # and self.head_size in [64, 128, 256, 512]) # Check if use_flashinfer is already set if not hasattr(self, "use_flashinfer"): self.use_flashinfer = False cache = self._compute_cos_sin_cache() if not self.use_flashinfer: cache = cache.to(dtype) self.cos_sin_cache: torch.Tensor self.register_buffer("cos_sin_cache", cache, persistent=False) self.is_rocm_triton_rotary_embed_enabled = ( rocm_aiter_ops.is_triton_rotary_embed_enabled() ) self.apply_rotary_emb = ApplyRotaryEmb( is_neox_style=self.is_neox_style, ) def _compute_inv_freq(self, base: float) -> torch.Tensor: """Compute the inverse frequency.""" # NOTE(woosuk): To exactly match the HF implementation, we need to # use CPU to compute the cache and then move it to GPU. However, we # create the cache on GPU for faster initialization. This may cause # a slight numerical difference between the HF implementation and ours. inv_freq = 1.0 / ( base ** ( torch.arange(0, self.rotary_dim, 2, dtype=torch.float) / self.rotary_dim ) ) return inv_freq def _compute_cos_sin_cache(self) -> torch.Tensor: """Compute the cos and sin cache.""" inv_freq = self._compute_inv_freq(self.base) t = torch.arange(self.max_position_embeddings, dtype=torch.float) freqs = torch.einsum("i,j -> ij", t, inv_freq) cos = freqs.cos() sin = freqs.sin() cache = torch.cat((cos, sin), dim=-1) return cache def _match_cos_sin_cache_dtype(self, query: torch.Tensor) -> None: # __setattr__ in nn.Module (called by `self.cos_sin_cache = ...`) # is expensive, so avoid calling it if possible if ( self.cos_sin_cache.device != query.device or self.cos_sin_cache.dtype != query.dtype ): self.cos_sin_cache = self.cos_sin_cache.to(query.device, dtype=query.dtype) def get_cos_sin(self, seqlen: int) -> tuple[torch.Tensor, torch.Tensor]: cos_sin = self.cos_sin_cache[:seqlen] cos, sin = cos_sin.chunk(2, dim=-1) return cos, sin class RotaryEmbedding(RotaryEmbeddingBase): def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, ) -> None: super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) @staticmethod def forward_static( positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None, head_size: int, rotary_dim: int, cos_sin_cache: torch.Tensor, is_neox_style: bool, ) -> tuple[torch.Tensor, torch.Tensor | None]: """A PyTorch-native implementation of forward().""" positions = positions.flatten() num_tokens = positions.shape[0] cos_sin = cos_sin_cache.index_select(0, positions) cos, sin = cos_sin.chunk(2, dim=-1) query_shape = query.shape query = query.view(num_tokens, -1, head_size) query_rot = query[..., :rotary_dim] query_pass = query[..., rotary_dim:] query_rot = ApplyRotaryEmb.forward_static( query_rot, cos, sin, is_neox_style, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) # key may be None in some cases, e.g. cross-layer KV sharing if key is not None: key_shape = key.shape key = key.view(num_tokens, -1, head_size) key_rot = key[..., :rotary_dim] key_pass = key[..., rotary_dim:] key_rot = ApplyRotaryEmb.forward_static( key_rot, cos, sin, is_neox_style, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """A PyTorch-native implementation of forward().""" return self.forward_static( positions, query, key, self.head_size, self.rotary_dim, self.cos_sin_cache, self.is_neox_style, ) def forward_cuda( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: if self.use_flashinfer: torch.ops.vllm.flashinfer_rotary_embedding( positions, query, key, self.head_size, self.cos_sin_cache, self.is_neox_style, ) return query, key from vllm import _custom_ops as ops self._match_cos_sin_cache_dtype(query) # ops.rotary_embedding() is an in-place operation # that updates the query and key tensors. ops.rotary_embedding( positions, query, key, self.head_size, self.cos_sin_cache, self.is_neox_style, ) return query, key def forward_hip( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: if self.is_rocm_triton_rotary_embed_enabled: self._match_cos_sin_cache_dtype(query) rocm_aiter_ops.triton_rotary_embed( positions, query, key, self.cos_sin_cache, self.head_size, self.rotary_dim, self.is_neox_style, ) return query, key return self.forward_cuda(positions, query, key) def forward_xpu( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: from vllm._ipex_ops import ipex_ops as ops self._match_cos_sin_cache_dtype(query) # ops.rotary_embedding() is an in-place operation # that updates the query and key tensors. if key is None: # XPU kernel doesn't support key=None so fall back to native impl # TODO(sarckk): add support for optional key in # ipex.llm.functional.rotary_embedding_batched return self.forward_native(positions, query, key) else: ops.rotary_embedding( positions, query, key, self.head_size, self.cos_sin_cache, self.is_neox_style, ) return query, key def extra_repr(self) -> str: s = f"head_size={self.head_size}, rotary_dim={self.rotary_dim}" s += f", max_position_embeddings={self.max_position_embeddings}" s += f", base={self.base}, is_neox_style={self.is_neox_style}" return s
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/rotary_embedding/llama3_rope.py
vllm/model_executor/layers/rotary_embedding/llama3_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import torch from .base import RotaryEmbedding class Llama3RotaryEmbedding(RotaryEmbedding): def __init__( self, head_size: int, rotary_dim: int, max_position_embeddings: int, base: float, is_neox_style: bool, dtype: torch.dtype, scaling_factor: float, low_freq_factor: float, high_freq_factor: float, orig_max_position: int, ) -> None: self.scaling_factor = scaling_factor self.low_freq_factor = low_freq_factor self.high_freq_factor = high_freq_factor self.orig_max_position = orig_max_position super().__init__( head_size, rotary_dim, max_position_embeddings, base, is_neox_style, dtype ) def _compute_inv_freq(self, base: float) -> torch.Tensor: inv_freqs = super()._compute_inv_freq(base) low_freq_wavelen = self.orig_max_position / self.low_freq_factor high_freq_wavelen = self.orig_max_position / self.high_freq_factor wave_len = 2 * math.pi / inv_freqs if self.low_freq_factor != self.high_freq_factor: smooth = (self.orig_max_position / wave_len - self.low_freq_factor) / ( self.high_freq_factor - self.low_freq_factor ) else: smooth = 0 new_freqs = torch.where( wave_len < high_freq_wavelen, inv_freqs, torch.where( wave_len > low_freq_wavelen, inv_freqs / self.scaling_factor, (1 - smooth) * inv_freqs / self.scaling_factor + smooth * inv_freqs, ), ) return new_freqs
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/abstract.py
vllm/model_executor/layers/mamba/abstract.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import abstractmethod from collections.abc import Iterable import torch from vllm.attention.backends.abstract import AttentionBackend from vllm.attention.selector import get_mamba_attn_backend from vllm.config import VllmConfig from vllm.model_executor.layers.attention_layer_base import AttentionLayerBase from vllm.v1.kv_cache_interface import KVCacheSpec, MambaSpec class MambaBase(AttentionLayerBase): """ Base class for Mamba-like layers which support the v1 engine. Inherit from this class if you implement a custom layer. """ # Contains the KV cache (mamba state) for the layer # in the shape specified by `self.get_state_shape`. kv_cache: tuple[torch.Tensor, ...] @abstractmethod def get_state_shape(self) -> Iterable[tuple[int, ...]]: """ Defines the shape of the state. For mamba layers this is usually a (conv_state, ssm_state) tuple. In this case, returns (conv_state_shape, ssm_state_shape). """ pass @property @abstractmethod def mamba_type(self) -> str: pass @abstractmethod def get_state_dtype(self) -> tuple[torch.dtype, ...]: pass def get_kv_cache_spec(self, vllm_config: VllmConfig) -> KVCacheSpec | None: if ( vllm_config.speculative_config is not None and vllm_config.model_config.hf_config.model_type not in ["qwen3_next"] ): raise NotImplementedError( "Mamba with speculative decoding is not supported yet." ) mamba_block_size = vllm_config.cache_config.mamba_block_size page_size_padded = vllm_config.cache_config.mamba_page_size_padded return MambaSpec( shapes=self.get_state_shape(), dtypes=self.get_state_dtype(), block_size=mamba_block_size, page_size_padded=page_size_padded, mamba_type=self.mamba_type, num_speculative_blocks=( vllm_config.speculative_config.num_speculative_tokens if vllm_config.speculative_config else 0 ), ) def get_attn_backend(self) -> type[AttentionBackend]: """Get the attention backend class for this Mamba layer.""" return get_mamba_attn_backend(self.mamba_type)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/mamba_utils.py
vllm/model_executor/layers/mamba/mamba_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.config.cache import MambaDType from vllm.config.model import ModelDType from vllm.distributed import divide from vllm.utils.torch_utils import ( STR_DTYPE_TO_TORCH_DTYPE, get_kv_cache_torch_dtype, ) class MambaStateDtypeCalculator: @classmethod def linear_attention_state_dtype( cls, model_dtype: ModelDType | torch.dtype, mamba_cache_dtype: MambaDType, ) -> tuple[torch.dtype, ...]: # TODO (tdoublep) requires testing if mamba_cache_dtype == "float32": raise ValueError("fp32 state for minimax is not yet supported") state_dtype = get_kv_cache_torch_dtype(mamba_cache_dtype, model_dtype) return (state_dtype,) @classmethod def mamba1_state_dtype( cls, model_dtype: ModelDType | torch.dtype, mamba_cache_dtype: MambaDType, mamba_ssm_cache_dtype: MambaDType, ) -> tuple[torch.dtype, ...]: return cls._mamba_state_dtype( model_dtype, mamba_cache_dtype, mamba_ssm_cache_dtype ) @classmethod def mamba2_state_dtype( cls, model_dtype: ModelDType | torch.dtype, mamba_cache_dtype: MambaDType, mamba_ssm_cache_dtype: MambaDType, ) -> tuple[torch.dtype, ...]: return cls._mamba_state_dtype( model_dtype, mamba_cache_dtype, mamba_ssm_cache_dtype ) @classmethod def _mamba_state_dtype( cls, model_dtype: ModelDType | torch.dtype, mamba_cache_dtype: MambaDType, mamba_ssm_cache_dtype: MambaDType, ) -> tuple[torch.dtype, ...]: conv_state_dtype = get_kv_cache_torch_dtype(mamba_cache_dtype, model_dtype) if mamba_ssm_cache_dtype == "auto": temporal_state_dtype = conv_state_dtype else: temporal_state_dtype = STR_DTYPE_TO_TORCH_DTYPE[mamba_ssm_cache_dtype] return (conv_state_dtype, temporal_state_dtype) @classmethod def short_conv_state_dtype( cls, model_dtype: ModelDType | torch.dtype, mamba_cache_dtype: MambaDType, ) -> tuple[torch.dtype, ...]: conv_state_dtype = get_kv_cache_torch_dtype(mamba_cache_dtype, model_dtype) return (conv_state_dtype,) @classmethod def gated_delta_net_state_dtype( cls, model_dtype: ModelDType | torch.dtype, mamba_cache_dtype: MambaDType, ) -> tuple[torch.dtype, torch.dtype]: state_dtype = get_kv_cache_torch_dtype(mamba_cache_dtype, model_dtype) return (state_dtype, state_dtype) @classmethod def kda_state_dtype( cls, model_dtype: ModelDType | torch.dtype, mamba_cache_dtype: MambaDType, ): state_dtype = get_kv_cache_torch_dtype(mamba_cache_dtype, model_dtype) return (state_dtype, state_dtype, state_dtype, torch.float32) class MambaStateShapeCalculator: @classmethod def linear_attention_state_shape( cls, num_heads: int, tp_size: int, head_dim: int, ) -> tuple[tuple[int, int, int], ...]: state_shape = (num_heads // tp_size, head_dim, head_dim) return (state_shape,) @classmethod def mamba1_state_shape( cls, tp_world_size: int, intermediate_size: int, state_size: int, conv_kernel: int, ) -> tuple[tuple[int, int], tuple[int, int]]: conv_state_shape = (divide(intermediate_size, tp_world_size), conv_kernel - 1) temporal_state_shape = (divide(intermediate_size, tp_world_size), state_size) conv_state_shape = conv_state_shape[1], conv_state_shape[0] return conv_state_shape, temporal_state_shape @classmethod def mamba2_state_shape( cls, tp_world_size: int, intermediate_size: int, n_groups: int, num_heads: int, head_dim: int, state_size: int, conv_kernel: int, ) -> tuple[tuple[int, int], tuple[int, int, int]]: # if n_groups is not divisible by world_size, need to extend the shards # to ensure all groups needed by a head is sharded along with it n_groups = n_groups + cls.extra_groups_for_head_shards(n_groups, tp_world_size) # heads and n_groups are TP-ed conv_dim = intermediate_size + 2 * n_groups * state_size # contiguous along 'dim' axis conv_state_shape = (conv_kernel - 1, divide(conv_dim, tp_world_size)) # These are not TP-ed as they depend on A, dt_bias, D # - they are typically small # e.g., (h_heads, head_dim, state_size) = (128, 64, 128) temporal_state_shape = (divide(num_heads, tp_world_size), head_dim, state_size) return conv_state_shape, temporal_state_shape @classmethod def short_conv_state_shape( cls, tp_world_size: int, intermediate_size: int, conv_kernel: int, ) -> tuple[tuple[int, int]]: conv_dim = divide(intermediate_size, tp_world_size) conv_state_shape = (conv_kernel - 1, conv_dim) return (conv_state_shape,) @classmethod def extra_groups_for_head_shards(cls, ngroups: int, tp_size: int): """Compute the increase in group numbers to account for replication in order to accompany the head shards.""" # in the case ngoups % tp_size == 0, this will be zero if ngroups % tp_size == 0: return 0 # for n_groups == 1, this is exactly tp_size - n_groups return tp_size - ngroups @classmethod def gated_delta_net_state_shape( cls, tp_world_size: int, num_k_heads: int, num_v_heads: int, head_k_dim: int, head_v_dim: int, conv_kernel_size: int, num_spec: int = 0, ): conv_dim = head_k_dim * num_k_heads * 2 + head_v_dim * num_v_heads conv_state_shape = ( divide(conv_dim, tp_world_size), conv_kernel_size - 1 + num_spec, ) conv_state_shape = conv_state_shape[1], conv_state_shape[0] temporal_state_shape = ( divide(num_v_heads, tp_world_size), head_k_dim, head_v_dim, ) return conv_state_shape, temporal_state_shape @classmethod def kda_state_shape( cls, tp_world_size: int, num_heads: int, head_dim: int, num_k_heads: int | None = None, head_k_dim: int | None = None, conv_kernel_size: int = 4, num_spec: int = 0, ) -> tuple[tuple[int, int], tuple[int, int], tuple[int, int], tuple[int, int, int]]: if num_k_heads is None: num_k_heads = num_heads if head_k_dim is None: head_k_dim = head_dim proj_size = num_heads * head_dim proj_k_size = num_k_heads * head_k_dim conv_state_shape = (divide(proj_size, tp_world_size), conv_kernel_size - 1) conv_state_k_shape = (divide(proj_k_size, tp_world_size), conv_kernel_size - 1) recurrent_state_shape = (divide(num_heads, tp_world_size), head_dim, head_dim) conv_state_shape = conv_state_shape[1], conv_state_shape[0] conv_state_k_shape = conv_state_k_shape[1], conv_state_k_shape[0] return ( conv_state_shape, conv_state_k_shape, conv_state_k_shape, recurrent_state_shape, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/mamba_mixer.py
vllm/model_executor/layers/mamba/mamba_mixer.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import NamedTuple import torch from torch import nn from torch.nn.parameter import Parameter from vllm.config import CacheConfig, ModelConfig, get_current_vllm_config from vllm.distributed.parallel_state import ( get_tensor_model_parallel_rank, get_tensor_model_parallel_world_size, ) from vllm.forward_context import ForwardContext, get_forward_context from vllm.model_executor.custom_op import CustomOp from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( ColumnParallelLinear, MergedColumnParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.mamba.abstract import MambaBase from vllm.model_executor.layers.mamba.mamba_utils import ( MambaStateDtypeCalculator, MambaStateShapeCalculator, ) from vllm.model_executor.layers.mamba.ops.causal_conv1d import ( causal_conv1d_fn, causal_conv1d_update, ) from vllm.model_executor.layers.mamba.ops.mamba_ssm import ( selective_scan_fn, selective_state_update, ) from vllm.model_executor.utils import set_weight_attrs from vllm.utils.torch_utils import direct_register_custom_op from vllm.v1.attention.backends.mamba1_attn import Mamba1AttentionMetadata # Adapted from transformers.models.mamba.modeling_mamba.MambaMixer @CustomOp.register("mamba_mixer") class MambaMixer(MambaBase, CustomOp): """ Compute ∆, A, B, C, and D the state space parameters and compute the `contextualized_states`. A, D are input independent (see Mamba paper [1] Section 3.5.2 "Interpretation of A" for why A isn't selective) ∆, B, C are input-dependent (this is a key difference between Mamba and the linear time invariant S4, and is why Mamba is called **selective** state spaces) """ def __init__( self, hidden_size: int, ssm_state_size: int, conv_kernel_size: int, intermediate_size: int, time_step_rank: int, use_conv_bias: bool, use_bias: bool, use_rms_norm: bool, rms_norm_has_weight: bool = True, rms_norm_eps: float = 1e-5, activation="silu", is_lora_enabled: bool = False, model_config: ModelConfig | None = None, cache_config: CacheConfig | None = None, prefix: str = "", ): super().__init__() self.time_step_rank = time_step_rank self.ssm_state_size = ssm_state_size self.use_rms_norm = use_rms_norm self.activation = activation self.is_lora_enabled = is_lora_enabled self.conv_kernel_size = conv_kernel_size self.intermediate_size = intermediate_size self.conv1d = ColumnParallelLinear( input_size=conv_kernel_size, output_size=intermediate_size, bias=use_conv_bias, ) # unsqueeze to fit conv1d weights shape into the linear weights shape. # Can't do this in `weight_loader` since it already exists in # `ColumnParallelLinear` and `set_weight_attrs` # doesn't allow to override it self.conv1d.weight.data = self.conv1d.weight.data.unsqueeze(1) self.in_proj = MergedColumnParallelLinear( hidden_size, [intermediate_size] * 2, bias=use_bias ) # selective projection used to make dt, B and C input dependent self.x_proj = RowParallelLinear( intermediate_size, time_step_rank + ssm_state_size * 2, bias=False, ) # time step projection (discretization) - # In the forward we need to apply dt_proj without the bias, # as the bias is added in the selective scan kernel. self.dt_proj = ColumnParallelLinear( time_step_rank, intermediate_size, bias=True, skip_bias_add=True ) def weight_loader(param: Parameter, loaded_weight: torch.Tensor): tp_rank = get_tensor_model_parallel_rank() tp_size = get_tensor_model_parallel_world_size() param.data.copy_( loaded_weight.data.split(loaded_weight.shape[0] // tp_size, dim=0)[ tp_rank ] ) def A_weight_loader(param: Parameter, loaded_weight: torch.Tensor): weight_loader(param, -torch.exp(loaded_weight.float())) tp_size = get_tensor_model_parallel_world_size() self.A = nn.Parameter( torch.empty( intermediate_size // tp_size, ssm_state_size, dtype=torch.float32, ) ) self.D = nn.Parameter(torch.ones(intermediate_size // tp_size)) set_weight_attrs(self.D, {"weight_loader": weight_loader}) set_weight_attrs(self.A, {"weight_loader": A_weight_loader}) self.out_proj = RowParallelLinear( intermediate_size, hidden_size, bias=use_bias, input_is_parallel=True, ) self.dt_layernorm = ( RMSNorm( time_step_rank, eps=rms_norm_eps, has_weight=rms_norm_has_weight, ) if use_rms_norm else None ) self.b_layernorm = ( RMSNorm( ssm_state_size, eps=rms_norm_eps, has_weight=rms_norm_has_weight, ) if use_rms_norm else None ) self.c_layernorm = ( RMSNorm( ssm_state_size, eps=rms_norm_eps, has_weight=rms_norm_has_weight, ) if use_rms_norm else None ) compilation_config = get_current_vllm_config().compilation_config if prefix in compilation_config.static_forward_context: raise ValueError(f"Duplicate layer name: {prefix}") compilation_config.static_forward_context[prefix] = self # The inner tuple is (conv_state, ssm_state) self.kv_cache = (torch.tensor([]), torch.tensor([])) self.model_config = model_config self.cache_config = cache_config self.prefix = prefix def _ssm_transform( self, x: torch.Tensor ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: if self.is_lora_enabled: # Lora kernel requires contiguous tensor. ssm_params = self.x_proj(x.contiguous())[0] else: ssm_params = self.x_proj(x)[0] time_step, B, C = torch.split( ssm_params, [self.time_step_rank, self.ssm_state_size, self.ssm_state_size], dim=-1, ) if self.use_rms_norm: assert self.dt_layernorm is not None assert self.b_layernorm is not None assert self.c_layernorm is not None time_step = self.dt_layernorm(time_step.contiguous()) B = self.b_layernorm(B.contiguous()) C = self.c_layernorm(C.contiguous()) discrete_time_step = self.dt_proj(time_step)[0].transpose(-2, -1) return discrete_time_step, B, C def forward(self, hidden_states: torch.Tensor, output: torch.Tensor): torch.ops.vllm.mamba_mixer( hidden_states, output, self.prefix, ) def forward_native(self, hidden_states: torch.Tensor, output: torch.Tensor): pass def forward_cuda(self, hidden_states: torch.Tensor, output: torch.Tensor): """ Run the Mamba-1 SSM pipeline. Steps ----- 1. Apply the gated-MLP linear projection to the raw input. 2. Pass the projected sequence through the convolutional mixing layer. 3. Feed the result into the State-Space Model (SSM) blocks. 4. Perform the recurrence y ← SSM(A, B, C, Δ)(x) to produce contextual representations. 5. Project the contextualised sequence back to the output embedding dimension. Batch handling -------------- Prefill and decode tokens are processed by dedicated CUDA kernels for both the convolutional (conv1d) and SSM stages. In the case of a mixed batch (containing both prefill and decode tokens), both sets of kernels are executed independently and their outputs are concatenated before the final output projection. """ forward_context: ForwardContext = get_forward_context() attn_metadata = forward_context.attn_metadata assert self.cache_config is not None mamba_block_size = self.cache_config.mamba_block_size prefix_caching_enabled = self.cache_config.enable_prefix_caching if attn_metadata is not None: assert isinstance(attn_metadata, dict) attn_metadata = attn_metadata[self.prefix] assert isinstance(attn_metadata, Mamba1AttentionMetadata) query_start_loc_p = attn_metadata.query_start_loc_p state_indices_tensor = attn_metadata.state_indices_tensor self_kv_cache = self.kv_cache[forward_context.virtual_engine] conv_state = self_kv_cache[0].transpose(-1, -2) ssm_state = self_kv_cache[1] has_initial_states_p = attn_metadata.has_initial_states_p # 1. Gated MLP's linear projection projected_states = self.in_proj(hidden_states)[0].transpose(-2, -1) hidden_states_BC, gate = projected_states.chunk(2, dim=-2) conv_weights = self.conv1d.weight.view( self.conv1d.weight.size(0), self.conv1d.weight.size(2) ) if attn_metadata is None: # V1 profile run hidden_states_BC = hidden_states_BC.contiguous() return self.out_proj(hidden_states_BC.transpose(-2, -1))[0] num_prefill_tokens = attn_metadata.num_prefill_tokens # token count num_decode_tokens = attn_metadata.num_decode_tokens num_prefills = attn_metadata.num_prefills # request count num_decodes = attn_metadata.num_decode_tokens # token count (=request) has_prefill = num_prefill_tokens > 0 has_decode = num_decode_tokens > 0 num_actual_tokens = num_prefill_tokens + num_decode_tokens prefill_decode_split = split_batch_to_prefill_and_decode( hidden_states_BC, gate, state_indices_tensor, num_prefill_tokens, num_prefills, num_decode_tokens, ) hidden_states_BC_p = prefill_decode_split.hidden_states_BC_p hidden_states_BC_d = prefill_decode_split.hidden_states_BC_d gate_p = prefill_decode_split.gate_p gate_d = prefill_decode_split.gate_d state_indices_tensor_p = prefill_decode_split.state_indices_tensor_p state_indices_tensor_d = prefill_decode_split.state_indices_tensor_d if prefix_caching_enabled: block_idx_last_computed_token_d, block_idx_last_computed_token_p = ( torch.split( attn_metadata.block_idx_last_computed_token, [num_decodes, num_prefills], dim=0, ) ) block_idx_last_scheduled_token_d, block_idx_last_scheduled_token_p = ( torch.split( attn_metadata.block_idx_last_scheduled_token, [num_decodes, num_prefills], dim=0, ) ) block_idx_first_scheduled_token_p = ( attn_metadata.block_idx_first_scheduled_token_p ) num_computed_tokens_p = attn_metadata.num_computed_tokens_p else: block_idx_last_computed_token_d = None block_idx_last_computed_token_p = None block_idx_last_scheduled_token_d = None block_idx_last_scheduled_token_p = None block_idx_first_scheduled_token_p = None num_computed_tokens_p = None ssm_outputs = [] if has_prefill: # 2. Convolution sequence transformation conv_out_p = causal_conv1d_fn( hidden_states_BC_p, conv_weights, self.conv1d.bias, activation=self.activation, conv_states=conv_state, has_initial_state=has_initial_states_p, cache_indices=state_indices_tensor_p, query_start_loc=query_start_loc_p, block_idx_first_scheduled_token=block_idx_first_scheduled_token_p, block_idx_last_scheduled_token=block_idx_last_scheduled_token_p, initial_state_idx=block_idx_last_computed_token_p, num_computed_tokens=num_computed_tokens_p, block_size_to_align=mamba_block_size, ) # 3. State Space Model sequence transformations. discrete_time_step_p, B_p, C_p = self._ssm_transform( conv_out_p.transpose(-2, -1) ) time_proj_bias = self._time_proj_bias() # 4. Perform the recurrence y ← SSM(A, B, C, Δ)(x) scan_out_p = selective_scan_fn( conv_out_p, ssm_state, discrete_time_step_p, self.A, B_p.transpose(-2, -1), C_p.transpose(-2, -1), self.D.float(), gate_p, time_proj_bias, delta_softplus=True, cache_indices=state_indices_tensor_p, has_initial_state=has_initial_states_p, query_start_loc=query_start_loc_p, block_size=mamba_block_size, block_idx_first_scheduled_token=block_idx_first_scheduled_token_p, block_idx_last_scheduled_token=block_idx_last_scheduled_token_p, initial_state_idx=block_idx_last_computed_token_p, ) ssm_outputs.append(scan_out_p) if has_decode: if prefix_caching_enabled: state_indices_tensor_d_input = state_indices_tensor_d.gather( 1, block_idx_last_computed_token_d.unsqueeze(1) ).squeeze(1) state_indices_tensor_d_output = state_indices_tensor_d.gather( 1, block_idx_last_scheduled_token_d.unsqueeze(1) ).squeeze(1) else: state_indices_tensor_d_input = state_indices_tensor_d state_indices_tensor_d_output = state_indices_tensor_d # 2. Convolution sequence transformation conv_out_d = causal_conv1d_update( hidden_states_BC_d.transpose(0, 1), conv_state, conv_weights, self.conv1d.bias, self.activation, conv_state_indices=state_indices_tensor_d, block_idx_last_scheduled_token=block_idx_last_scheduled_token_d, initial_state_idx=block_idx_last_computed_token_d, ).transpose(0, 1) # 3. State Space Model sequence transformation. discrete_time_step_d, B_d, C_d = self._ssm_transform( conv_out_d.transpose(-2, -1) ) time_proj_bias = self._time_proj_bias() # 4. Perform the recurrence y ← SSM(A, B, C, Δ)(x) scan_outputs_d = torch.empty_like(hidden_states_BC_d.transpose(0, 1)) selective_state_update( ssm_state, conv_out_d.transpose(0, 1), discrete_time_step_d.transpose(0, 1), self.A, B_d, C_d, self.D, gate_d.transpose(0, 1), time_proj_bias, dt_softplus=True, state_batch_indices=state_indices_tensor_d_input, dst_state_batch_indices=state_indices_tensor_d_output, out=scan_outputs_d, ) scan_outputs_d = scan_outputs_d.transpose(0, 1) ssm_outputs.insert(0, scan_outputs_d) scan_outputs_combined = ( ssm_outputs[0] if len(ssm_outputs) == 1 else torch.cat(ssm_outputs, dim=-1) ) # 5. Final output projection if self.is_lora_enabled: # Lora kernel requires contiguous tensor. scan_outputs_combined = scan_outputs_combined.transpose(-2, -1).contiguous() out = self.out_proj(scan_outputs_combined)[0] else: out = self.out_proj(scan_outputs_combined.transpose(-2, -1))[0] output[:num_actual_tokens] = out def get_state_dtype(self) -> tuple[torch.dtype]: assert self.model_config is not None assert self.cache_config is not None return MambaStateDtypeCalculator.mamba1_state_dtype( self.model_config.dtype, self.cache_config.mamba_cache_dtype, self.cache_config.mamba_ssm_cache_dtype, ) def get_state_shape(self) -> tuple[tuple[int, ...], tuple[int, ...]]: return MambaStateShapeCalculator.mamba1_state_shape( tp_world_size=get_tensor_model_parallel_world_size(), intermediate_size=self.intermediate_size, state_size=self.ssm_state_size, conv_kernel=self.conv_kernel_size, ) @property def mamba_type(self) -> str: return "mamba1" def _time_proj_bias(self) -> torch.Tensor | None: if hasattr(self.dt_proj, "bias") and self.dt_proj.bias is not None: return self.dt_proj.bias.float() return None class PrefillDecodeSplit(NamedTuple): hidden_states_BC_p: torch.Tensor hidden_states_BC_d: torch.Tensor gate_p: torch.Tensor gate_d: torch.Tensor state_indices_tensor_p: torch.Tensor state_indices_tensor_d: torch.Tensor def split_batch_to_prefill_and_decode( hidden_states_BC: torch.Tensor, gate: torch.Tensor, state_indices_tensor: torch.Tensor, num_prefill_tokens: int, num_prefills: int, num_decode_tokens: int, ) -> PrefillDecodeSplit: num_actual_tokens = num_prefill_tokens + num_decode_tokens # In v1, decode tokens come first, then prefill tokens. hidden_states_BC_d, hidden_states_BC_p = torch.split( hidden_states_BC[..., :num_actual_tokens], [num_decode_tokens, num_prefill_tokens], dim=-1, ) gate_d, gate_p = torch.split( gate[..., :num_actual_tokens], [num_decode_tokens, num_prefill_tokens], dim=-1 ) # num_decode_tokens accounts for CUDA graph padding when applicable state_indices_tensor_d, state_indices_tensor_p = torch.split( state_indices_tensor[: num_decode_tokens + num_prefills], [num_decode_tokens, num_prefills], dim=0, ) return PrefillDecodeSplit( hidden_states_BC_p=hidden_states_BC_p, hidden_states_BC_d=hidden_states_BC_d, gate_p=gate_p, gate_d=gate_d, state_indices_tensor_p=state_indices_tensor_p, state_indices_tensor_d=state_indices_tensor_d, ) def mamba_mixer( hidden_states: torch.Tensor, output: torch.Tensor, layer_name: str, ) -> None: forward_context: ForwardContext = get_forward_context() self = forward_context.no_compile_layers[layer_name] self.forward_cuda(hidden_states=hidden_states, output=output) def mamba_mixer_fake( hidden_states: torch.Tensor, output: torch.Tensor, layer_name: str, ) -> None: return direct_register_custom_op( op_name="mamba_mixer", op_func=mamba_mixer, mutates_args=["output"], fake_impl=mamba_mixer_fake, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/short_conv.py
vllm/model_executor/layers/mamba/short_conv.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.attention.backends.abstract import AttentionMetadata from vllm.config import CacheConfig, ModelConfig, get_current_vllm_config from vllm.distributed import get_tensor_model_parallel_world_size from vllm.forward_context import ForwardContext, get_forward_context from vllm.model_executor.custom_op import CustomOp from vllm.model_executor.layers.linear import ( ColumnParallelLinear, MergedColumnParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.mamba.abstract import MambaBase from vllm.model_executor.layers.mamba.mamba_utils import ( MambaStateDtypeCalculator, MambaStateShapeCalculator, ) from vllm.model_executor.layers.mamba.ops.causal_conv1d import ( causal_conv1d_fn, causal_conv1d_update, ) from vllm.utils.torch_utils import direct_register_custom_op from vllm.v1.attention.backends.short_conv_attn import ShortConvAttentionMetadata @CustomOp.register("short_conv") class ShortConv(MambaBase, CustomOp): def __init__( self, config, dim: int, layer_idx: int, model_config: ModelConfig | None = None, cache_config: CacheConfig | None = None, prefix: str = "", ): super().__init__() self.config = config self.layer_idx = layer_idx self.conv_dim = dim self.L_cache = config.conv_L_cache self.bias = config.conv_bias self.conv = ColumnParallelLinear( input_size=self.L_cache, output_size=dim, bias=self.bias, prefix=f"{prefix}.conv1d", ) # unsqueeze to fit conv1d weights shape into the linear weights shape. # Can't do this in `weight_loader` since it already exists in # `ColumnParallelLinear` and `set_weight_attrs` # doesn't allow to override it self.conv.weight.data = self.conv.weight.data.unsqueeze(1) self.in_proj = MergedColumnParallelLinear( input_size=dim, output_sizes=[dim] * 3, bias=self.bias, prefix=f"{prefix}.in_proj", ) self.out_proj = RowParallelLinear( input_size=dim, output_size=dim, bias=self.bias, prefix=f"{prefix}.out_proj", ) compilation_config = get_current_vllm_config().compilation_config if prefix in compilation_config.static_forward_context: raise ValueError(f"Duplicate layer name: {prefix}") compilation_config.static_forward_context[prefix] = self self.kv_cache = (torch.tensor([]),) self.model_config = model_config self.cache_config = cache_config self.prefix = prefix def forward_native( self, hidden_states: torch.Tensor, output: torch.Tensor, ): return def forward( self, hidden_states: torch.Tensor, output: torch.Tensor, ): torch.ops.vllm.short_conv( hidden_states, output, self.prefix, ) def forward_cuda( self, hidden_states: torch.Tensor, output: torch.Tensor, ): forward_context = get_forward_context() # ShortConvAttentionMetadata contains metadata necessary for the # short_conv triton kernels to operate in continuous batching and in # chunked prefill modes; they are computed at top-level model forward # since they stay the same and reused for all mamba layers in the same # iteration. attn_metadata: AttentionMetadata = forward_context.attn_metadata if attn_metadata is not None: assert isinstance(attn_metadata, dict) attn_metadata = attn_metadata[self.prefix] assert isinstance(attn_metadata, ShortConvAttentionMetadata) self_kv_cache = self.kv_cache[forward_context.virtual_engine] conv_state = self_kv_cache[0].transpose(-1, -2) state_indices_tensor = attn_metadata.state_indices_tensor has_initial_states_p = attn_metadata.has_initial_states_p query_start_loc_p = attn_metadata.query_start_loc_p BCx, _ = self.in_proj(hidden_states) B, C, x = BCx.chunk(3, dim=-1) conv_weights = self.conv.weight.view( self.conv.weight.size(0), self.conv.weight.size(2) ) if attn_metadata is None: # V1 profile run Bx = (B * x).contiguous() hidden_states = C * Bx contextualized_states, _ = self.out_proj(hidden_states) return contextualized_states num_prefills = attn_metadata.num_prefills # request count num_decodes = attn_metadata.num_decode_tokens # token count (=request) num_prefill_tokens = attn_metadata.num_prefill_tokens # token count has_prefill = num_prefills > 0 has_decode = num_decodes > 0 num_actual_tokens = num_decodes + num_prefill_tokens # NOTE: V1 puts decode before prefill # Separate prefill and decode by splitting varlen input # Split along token dimension B_d, B_p = torch.split( B[:num_actual_tokens], [num_decodes, num_prefill_tokens], dim=0, ) C_d, C_p = torch.split( C[:num_actual_tokens], [num_decodes, num_prefill_tokens], dim=0, ) x_d, x_p = torch.split( x[:num_actual_tokens], [num_decodes, num_prefill_tokens], dim=0, ) # Split along batch dimension state_indices_tensor_d, state_indices_tensor_p = torch.split( state_indices_tensor, [num_decodes, num_prefills], dim=0, ) conv_output_list = [] if has_prefill: Bx_p = (B_p * x_p).transpose(0, 1) Bx = causal_conv1d_fn( Bx_p, conv_weights, self.conv.bias, activation=None, conv_states=conv_state, has_initial_state=has_initial_states_p, cache_indices=state_indices_tensor_p, metadata=attn_metadata, query_start_loc=query_start_loc_p, ).transpose(0, 1)[:num_prefill_tokens] y = C_p * Bx conv_output_list.append(y) if has_decode: Bx_d = (B_d * x_d).contiguous() Bx = causal_conv1d_update( Bx_d, conv_state, conv_weights, self.conv.bias, activation=None, conv_state_indices=state_indices_tensor_d, ) y = C_d * Bx conv_output_list.insert(0, y) # Merge prefill and decode outputs before passing to gated MLP hidden_states = torch.vstack(conv_output_list) # Final linear projection output[:num_actual_tokens], _ = self.out_proj(hidden_states) def get_state_dtype(self) -> tuple[torch.dtype, ...]: assert self.model_config is not None assert self.cache_config is not None return MambaStateDtypeCalculator.short_conv_state_dtype( self.model_config.dtype, self.cache_config.mamba_cache_dtype, ) def get_state_shape(self) -> tuple[tuple[int, ...]]: return MambaStateShapeCalculator.short_conv_state_shape( tp_world_size=get_tensor_model_parallel_world_size(), intermediate_size=self.conv_dim, conv_kernel=self.L_cache, ) @property def mamba_type(self) -> str: return "short_conv" def short_conv( hidden_states: torch.Tensor, output: torch.Tensor, layer_name: str, ) -> None: forward_context: ForwardContext = get_forward_context() self = forward_context.no_compile_layers[layer_name] self.forward_cuda(hidden_states=hidden_states, output=output) def short_conv_fake( hidden_states: torch.Tensor, output: torch.Tensor, layer_name: str, ) -> None: return direct_register_custom_op( op_name="short_conv", op_func=short_conv, mutates_args=["output"], fake_impl=short_conv_fake, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/linear_attn.py
vllm/model_executor/layers/mamba/linear_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import torch import torch.nn.functional as F from einops import rearrange from torch import nn from vllm.attention.backends.abstract import AttentionMetadata from vllm.config import CacheConfig, ModelConfig, get_current_vllm_config from vllm.distributed.communication_op import tensor_model_parallel_all_reduce from vllm.distributed.parallel_state import ( get_tensor_model_parallel_rank, get_tensor_model_parallel_world_size, ) from vllm.forward_context import ForwardContext, get_forward_context from vllm.model_executor.custom_op import CustomOp from vllm.model_executor.layers.lightning_attn import ( lightning_attention, linear_decode_forward_triton, ) from vllm.model_executor.layers.linear import ColumnParallelLinear, RowParallelLinear from vllm.model_executor.layers.mamba.abstract import MambaBase from vllm.model_executor.layers.mamba.mamba_utils import ( MambaStateDtypeCalculator, MambaStateShapeCalculator, ) from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.utils.torch_utils import direct_register_custom_op from vllm.v1.attention.backends.linear_attn import LinearAttentionMetadata class MiniMaxText01RMSNormTP(CustomOp): name = "MiniMaxText01RMSNormTP" def __init__(self, hidden_size: int, eps: float = 1e-6) -> None: super().__init__() self.tp_world = get_tensor_model_parallel_world_size() self.tp_rank = get_tensor_model_parallel_rank() self.weight = nn.Parameter(torch.ones(int(hidden_size / self.tp_world))) self.weight.weight_loader = self.weight_loader self.variance_epsilon = eps return @staticmethod def weight_loader( param: nn.Parameter, loaded_weight: torch.Tensor, ) -> None: tp_world = get_tensor_model_parallel_world_size() tp_rank = get_tensor_model_parallel_rank() shard_size = loaded_weight.shape[0] // tp_world shard = slice(tp_rank * shard_size, (tp_rank + 1) * shard_size) param.data.copy_(loaded_weight[shard]) return def _forward( self, x: torch.Tensor, ) -> torch.Tensor: orig_dtype = x.dtype x = x.to(torch.float32) variance = x.pow(2).mean(dim=-1, keepdim=True, dtype=torch.float32) if self.tp_world > 1: variance = tensor_model_parallel_all_reduce(variance) / self.tp_world x = x * torch.rsqrt(variance + self.variance_epsilon) x = (x * self.weight).to(orig_dtype) return x def forward( self, x: torch.Tensor, residual: torch.Tensor | None = None, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: assert residual is None, "RMSNorm does not support residual connection." return self._forward(x) @staticmethod def forward_qk( q_norm: "MiniMaxText01RMSNormTP", k_norm: "MiniMaxText01RMSNormTP", q: torch.Tensor, k: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: orig_dtype = q.dtype q = q.to(torch.float32) k = k.to(torch.float32) q_var = q.pow(2).mean(dim=-1, keepdim=True) k_var = k.pow(2).mean(dim=-1, keepdim=True) if q_norm.tp_world > 1: qk_var = torch.cat([q_var, k_var], dim=-1) qk_var = tensor_model_parallel_all_reduce(qk_var) / q_norm.tp_world q_var, k_var = qk_var.chunk(2, dim=-1) q = q * torch.rsqrt(q_var + q_norm.variance_epsilon) * q_norm.weight k = k * torch.rsqrt(k_var + k_norm.variance_epsilon) * k_norm.weight q = q.to(orig_dtype) k = k.to(orig_dtype) return q, k class MiniMaxText01LinearKernel: @staticmethod def jit_linear_forward_prefix( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, kv_caches: torch.Tensor, slope_rate: torch.Tensor, block_size: int, layer_idx: int | None = None, **kwargs, ) -> torch.Tensor: slope_rate = slope_rate.to(torch.float32) should_pad_dim = q.dim() == 3 if should_pad_dim: q = q.unsqueeze(0) k = k.unsqueeze(0) v = v.unsqueeze(0) b, h, n, d = q.shape e = d kv_history = kv_caches.reshape(1, h, d, e).contiguous() output, kv_history = lightning_attention( q, k, v, slope_rate, block_size=block_size, kv_history=kv_history ) kv_caches.copy_(kv_history[:, :, -1, :, :].reshape(h, d, e)) assert output.shape[0] == 1, "batch size must be 1" return rearrange(output.squeeze(0), "h n d -> n (h d)") class MiniMaxText01LinearAttention(nn.Module, MambaBase): @property def mamba_type(self) -> str: return "linear_attention" def get_state_dtype(self) -> tuple[torch.dtype]: assert self.model_config is not None assert self.cache_config is not None return MambaStateDtypeCalculator.linear_attention_state_dtype( self.model_config.dtype, self.cache_config.mamba_cache_dtype, ) def get_state_shape(self) -> tuple[tuple[int, int, int], ...]: return MambaStateShapeCalculator.linear_attention_state_shape( num_heads=self.num_heads, tp_size=self.tp_size, head_dim=self.head_dim ) def __init__( self, hidden_size: int, hidden_inner_size: int, num_heads: int, head_dim: int, max_position: int, block_size: int, num_hidden_layer: int, model_config: ModelConfig | None = None, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, layer_idx: int = 0, linear_layer_idx: int = 0, prefix: str = "linear_attn", ) -> None: super().__init__() self.layer_idx = layer_idx self.BLOCK = block_size self.hidden_size = hidden_size self.num_heads = num_heads self.head_dim = head_dim self.total_num_heads = num_heads self.hidden_inner_size = hidden_inner_size self.tp_size = get_tensor_model_parallel_world_size() self.tp_rank = get_tensor_model_parallel_rank() assert self.total_num_heads % self.tp_size == 0 self.tp_heads = self.total_num_heads // self.tp_size self.qkv_size = self.num_heads * self.head_dim self.tp_hidden = self.head_dim * self.tp_heads self.model_config = model_config self.cache_config = cache_config self.prefix = prefix self.qkv_proj = ColumnParallelLinear( hidden_size, self.hidden_inner_size * 3, bias=False, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", ) self.output_gate = ColumnParallelLinear( hidden_size, self.hidden_inner_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.output_gate", ) self.out_proj = RowParallelLinear( self.hidden_inner_size, hidden_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.out_proj", ) self.norm = MiniMaxText01RMSNormTP( self.hidden_inner_size, eps=1e-5, ) slope_rate = MiniMaxText01LinearAttention._build_slope_tensor(self.num_heads) if num_hidden_layer <= 1: self.slope_rate = slope_rate * (1 + 1e-5) else: self.slope_rate = slope_rate * ( 1 - layer_idx / (num_hidden_layer - 1) + 1e-5 ) self.tp_slope = self.slope_rate[ self.tp_rank * self.tp_heads : (self.tp_rank + 1) * self.tp_heads ].contiguous() compilation_config = get_current_vllm_config().compilation_config if prefix in compilation_config.static_forward_context: raise ValueError(f"Duplicate layer name: {prefix}") compilation_config.static_forward_context[prefix] = self @staticmethod def weight_direct_load(param: torch.Tensor, loaded_weight: torch.Tensor) -> None: assert param.size() == loaded_weight.size() param.data.copy_(loaded_weight) return @staticmethod def _build_slope_tensor(n_attention_heads: int): def get_slopes(n): def get_slopes_power_of_2(n): start = 2 ** (-(2 ** -(math.log2(n) - 3))) ratio = start return [start * ratio**i for i in range(n)] if math.log2(n).is_integer(): return get_slopes_power_of_2(n) else: closest_power_of_2 = 2 ** math.floor(math.log2(n)) return ( get_slopes_power_of_2(closest_power_of_2) + get_slopes(2 * closest_power_of_2)[0::2][: n - closest_power_of_2] ) slopes = torch.tensor( get_slopes(n_attention_heads), dtype=torch.float32 ).reshape(n_attention_heads, 1, 1) return slopes def _prefill_and_mix_infer( self, q, k, v, kv_cache, state_indices_tensor, attn_metadata ): hidden = [] for _prefill_idx in range(getattr(attn_metadata, "num_prefills", 0)): if _prefill_idx >= len(attn_metadata.query_start_loc): break if _prefill_idx >= len(state_indices_tensor): break offset = attn_metadata.num_decode_tokens _start = attn_metadata.query_start_loc[offset + _prefill_idx] _end = attn_metadata.query_start_loc[offset + _prefill_idx + 1] slot_id = state_indices_tensor[offset + _prefill_idx] qs = q[_start:_end].transpose(0, 1).contiguous() ks = k[_start:_end].transpose(0, 1).contiguous() vs = v[_start:_end].transpose(0, 1).contiguous() slice_layer_cache = kv_cache[slot_id, ...] out_slice = MiniMaxText01LinearKernel.jit_linear_forward_prefix( qs, ks, vs, slice_layer_cache, self.tp_slope, self.BLOCK, layer_idx=self.layer_idx, ) hidden.append(out_slice.contiguous()) if attn_metadata.num_decode_tokens > 0: hidden_decode = self._decode_infer( q, k, v, kv_cache, state_indices_tensor, attn_metadata ) hidden.insert(0, hidden_decode) if not hidden: return torch.empty((0, q.size(-1)), device=q.device, dtype=q.dtype) hidden = torch.concat(hidden, dim=0).contiguous() return hidden def _decode_infer(self, q, k, v, kv_cache, state_indices_tensor, attn_metadata): q = q[: attn_metadata.num_decode_tokens].unsqueeze(2).contiguous() k = k[: attn_metadata.num_decode_tokens].unsqueeze(2).contiguous() v = v[: attn_metadata.num_decode_tokens].unsqueeze(2).contiguous() slot_id = state_indices_tensor[: attn_metadata.num_decodes] hidden = linear_decode_forward_triton( q, k, v, kv_cache, self.tp_slope, slot_id, 32 ) return hidden def forward( self, hidden_states: torch.Tensor, output: torch.Tensor, positions: torch.Tensor ) -> None: torch.ops.vllm.linear_attention( hidden_states, output, positions, self.prefix, ) def _forward( self, hidden_states: torch.Tensor, output: torch.Tensor, positions: torch.Tensor ) -> None: forward_context = get_forward_context() attn_metadata: AttentionMetadata = forward_context.attn_metadata if attn_metadata is not None: assert isinstance(attn_metadata, dict) attn_metadata = attn_metadata[self.prefix] assert isinstance(attn_metadata, LinearAttentionMetadata) num_actual_tokens = ( attn_metadata.num_prefill_tokens + attn_metadata.num_decode_tokens ) else: num_actual_tokens = hidden_states.shape[0] qkv, _ = self.qkv_proj(hidden_states[:num_actual_tokens]) qkv32 = qkv.to(torch.float32) qkvact = torch.nn.functional.silu(qkv32) qkvact = qkvact.view((qkv.shape[0], self.tp_heads, -1)) q, k, v = torch.split(qkvact, [self.head_dim] * 3, dim=-1) if attn_metadata is not None: kv_cache = self.kv_cache[forward_context.virtual_engine][0] state_indices_tensor = attn_metadata.state_indices_tensor num_prefills = getattr(attn_metadata, "num_prefills", 0) if num_prefills > 0: num_decode_tokens = getattr(attn_metadata, "num_decode_tokens", 0) for prefill_idx in range(num_prefills): q_start = attn_metadata.query_start_loc[ num_decode_tokens + prefill_idx ] q_end = attn_metadata.query_start_loc[ num_decode_tokens + prefill_idx + 1 ] query_len = q_end - q_start context_len = ( attn_metadata.seq_lens[num_decode_tokens + prefill_idx] - query_len ) if context_len == 0: block_to_clear = state_indices_tensor[ num_decode_tokens + prefill_idx ] kv_cache[block_to_clear, ...] = 0 decode_only = getattr(attn_metadata, "num_prefills", 0) == 0 if attn_metadata is None: hidden = torch.empty( (q.shape[0], q.shape[1] * q.shape[2]), device=q.device, dtype=q.dtype ) else: if not decode_only: hidden = self._prefill_and_mix_infer( q, k, v, kv_cache, state_indices_tensor, attn_metadata ) else: hidden = self._decode_infer( q, k, v, kv_cache, state_indices_tensor, attn_metadata ) hidden = self.norm._forward(hidden) gate, _ = self.output_gate(hidden_states[:num_actual_tokens]) hidden = F.sigmoid(gate) * hidden hidden = hidden.to(hidden_states.dtype) output[:num_actual_tokens], _ = self.out_proj(hidden) def linear_attention( hidden_states: torch.Tensor, output: torch.Tensor, positions: torch.Tensor, layer_name: str, ) -> None: forward_context: ForwardContext = get_forward_context() self = forward_context.no_compile_layers[layer_name] self._forward(hidden_states=hidden_states, output=output, positions=positions) def linear_attention_fake( hidden_states: torch.Tensor, output: torch.Tensor, positions: torch.Tensor, layer_name: str, ) -> None: return direct_register_custom_op( op_name="linear_attention", op_func=linear_attention, mutates_args=["output"], fake_impl=linear_attention_fake, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/mamba_mixer2.py
vllm/model_executor/layers/mamba/mamba_mixer2.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from torch import nn from vllm.attention.backends.abstract import AttentionMetadata from vllm.config import CacheConfig, ModelConfig, get_current_vllm_config from vllm.distributed import ( divide, get_tensor_model_parallel_rank, get_tensor_model_parallel_world_size, tensor_model_parallel_all_gather, tensor_model_parallel_all_reduce, ) from vllm.forward_context import ForwardContext, get_forward_context from vllm.model_executor.custom_op import CustomOp from vllm.model_executor.layers.linear import ( ColumnParallelLinear, MergedColumnParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.mamba.abstract import MambaBase from vllm.model_executor.layers.mamba.mamba_utils import ( MambaStateDtypeCalculator, MambaStateShapeCalculator, ) from vllm.model_executor.layers.mamba.ops.causal_conv1d import ( causal_conv1d_fn, causal_conv1d_update, ) from vllm.model_executor.layers.mamba.ops.layernorm_gated import rms_norm_gated from vllm.model_executor.layers.mamba.ops.mamba_ssm import selective_state_update from vllm.model_executor.layers.mamba.ops.ssd_combined import ( mamba_chunk_scan_combined_varlen, ) from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.model_loader.weight_utils import ( LoaderFunction, composed_weight_loader, sharded_weight_loader, ) from vllm.model_executor.utils import set_weight_attrs from vllm.utils.torch_utils import direct_register_custom_op from vllm.v1.attention.backends.mamba2_attn import Mamba2AttentionMetadata # Added by the IBM Team, 2024 # Adapted from transformers.models.mamba2.modeling_mamba2.MambaRMSNormGated @CustomOp.register("mixer2_gated_rms_norm") class Mixer2RMSNormGated(CustomOp): def __init__( self, full_hidden_size: int, full_n_groups: int, use_rms_norm: bool = True, eps: float = 1e-6, ): super().__init__() self.tp_size = get_tensor_model_parallel_world_size() self.tp_rank = get_tensor_model_parallel_rank() self.full_hidden_size = full_hidden_size self.group_size = full_hidden_size // full_n_groups self.per_rank_hidden_size = full_hidden_size // self.tp_size self.n_groups = full_hidden_size // self.group_size self.variance_epsilon = eps self.use_rms_norm = use_rms_norm if self.use_rms_norm: # Register norm weight only if we're actually applying RMSNorm self.weight = nn.Parameter(torch.ones(self.per_rank_hidden_size)) set_weight_attrs(self.weight, {"weight_loader": sharded_weight_loader(0)}) else: # Avoid checkpoint mismatch by skipping unused parameter self.register_parameter("weight", None) assert self.full_hidden_size % self.tp_size == 0, ( "Tensor parallel world size must divide hidden size." ) def forward_native( self, x: torch.Tensor, gate: torch.Tensor, ): # Three tensor-parallel cases: # 1. n_groups is 1 # In this case we parallelize along the reduction dim. # Each rank computes a local sum of squares followed by AllReduce # 2. tp_size divides n_groups # Each rank only reduces within its local group(s). # No collective ops necessary. # 3. The general case can be pretty complicated so we AllGather # the input and then redundantly compute the RMSNorm. input_dtype = x.dtype x = x * nn.functional.silu(gate.to(torch.float32)) if not self.use_rms_norm: return x.to(input_dtype) if self.n_groups == 1: if self.tp_size > 1: # Compute local sum and then reduce to obtain global sum local_sums = x.pow(2).sum(dim=-1, keepdim=True) global_sums = tensor_model_parallel_all_reduce(local_sums) # Calculate the variance count = self.tp_size * x.shape[-1] variance = global_sums / count else: variance = x.pow(2).mean(-1, keepdim=True) x = x * torch.rsqrt(variance + self.variance_epsilon) else: redundant_tp: bool = self.n_groups % self.tp_size != 0 if redundant_tp: # To handle the general case, redundantly apply the variance x = tensor_model_parallel_all_gather(x, -1) *prefix_dims, hidden_dim = x.shape group_count = hidden_dim // self.group_size x_grouped = x.view(*prefix_dims, group_count, self.group_size) variance = x_grouped.pow(2).mean(-1, keepdim=True) x_grouped = x_grouped * torch.rsqrt(variance + self.variance_epsilon) x = x_grouped.view(*prefix_dims, hidden_dim) if redundant_tp: start = self.per_rank_hidden_size * self.tp_rank end = start + self.per_rank_hidden_size x = x[..., start:end] return self.weight * x.to(input_dtype) def forward_cuda( self, x: torch.Tensor, gate: torch.Tensor, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: input_dtype = x.dtype if not self.use_rms_norm: # Keep gate in float32 for numerical stability during silu return x * nn.functional.silu(gate.to(torch.float32)).to(input_dtype) if ((self.n_groups % self.tp_size) != 0) or self.n_groups != 1: return self.forward_native(x, gate) return rms_norm_gated( x, self.weight.data, bias=None, z=gate, eps=self.variance_epsilon, norm_before_gate=False, ) def mamba_v2_sharded_weight_loader( shard_spec: list[tuple[int, int, float]], tp_size: int, tp_rank: int, ) -> LoaderFunction: """Create a weight loader for mamba v2. This ensures that the projections are correctly sharded so that they can be split into x, B, C. It also ensures that all the groups corresponding to a head shard is placed together with it. """ def loader(param: torch.Tensor, loaded_weight: torch.Tensor) -> None: # - track boundary of (sharded) param, and loaded_weight, respectively boundary, loaded_boundary = 0, 0 # - iterate over the shard specs for full_dim, extra, duplicate_groups in shard_spec: # - full dim is the model dim (before TP). # - extra > 0, means there is expected overall increase # of dimensions. This is so because of replication. # - ratio is used map the tp_rank to the actual shard # rank. This is useful when there is replication of # groups to accompany head shards. # - size of the loaded shard shard_size = full_dim // tp_size # - compute the rank into the loaded shard. # - if there is replication, different TP shards will # take from the same rank. # NOTE: currently we only support duplication # in the case where num_groups == 1 rank = 0 if duplicate_groups else tp_rank # - leftmost boundary index into loaded weight. loaded_skip = rank * shard_size loaded_start_idx = loaded_boundary + loaded_skip # - take these many dims from the loaded weight. take = min(shard_size, full_dim - extra - loaded_skip) # - always shard on dim 0 # - the ignore is for a mundane mypy error as it does not # seem to handle slices well. # https://github.com/python/mypy/issues/2410 param.data[ boundary : (boundary + take), ... # type: ignore[misc] ] = loaded_weight[ loaded_start_idx : ( loaded_start_idx + take ) # type: ignore[misc] ] # type: ignore[misc] # move indexing boundaries boundary += shard_size loaded_boundary += full_dim - extra return loader # Adapted from transformers.models.mamba.modeling_mamba.MambaMixer @CustomOp.register("mamba_mixer2") class MambaMixer2(MambaBase, CustomOp): """ Compute ∆, A, B, C, and D the state space parameters and compute the `contextualized_states`. A, D are input independent (see Mamba paper [1] Section 3.5.2 "Interpretation of A" for why A isn't selective) ∆, B, C are input-dependent (this is a key difference between Mamba and the linear time invariant S4, and is why Mamba is called **selective** state spaces) """ def __init__( self, hidden_size: int, ssm_state_size: int, conv_kernel_size: int, intermediate_size: int, use_conv_bias: bool, use_bias: bool, n_groups: int = 1, num_heads: int = 128, head_dim: int = 64, rms_norm_eps: float = 1e-5, activation: str = "silu", use_rms_norm: bool = True, model_config: ModelConfig | None = None, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() # For TP, the sharding plan is as follows: # - for the conv modules, since # conv_dim = intermediate_size * 2 * n_groups * ssm_state_size, # we shard intermediate_size and n_groups # - since intermediate_size = n_heads * head_dim, sharding on # intermediate_size is achieved by sharding on n_heads. # - IF, world_size divides groups, then sharding # (n_groups / world_size, n_heads / world_size) # also maintains the invariant n_heads % n_groups == 0 # - HOWEVER IF, world_size DOES NOT divide groups, then we need # to allocate extra space in the shard, such that groups # may be replicated to follow the head shard. # - NOTE: currently for the world size DOES NOT divide groups # case, we only support the case when n_groups == 1 self.tp_size = get_tensor_model_parallel_world_size() tp_rank = get_tensor_model_parallel_rank() assert num_heads % self.tp_size == 0, ( "Tensor parallel world size must divide num heads." ) assert (n_groups % self.tp_size) == 0 or n_groups == 1, ( "If tensor parallel world size does not divide num_groups, " "then num_groups must equal 1." ) assert ( (n_groups % self.tp_size == 0) or self.tp_size == 1 or quant_config is None ), ( "Tensor parallel currently supported for quantized models only " "if tensor parallel world size divides num groups." ) self.ssm_state_size = ssm_state_size self.conv_kernel_size = conv_kernel_size self.activation = activation self.intermediate_size = intermediate_size self.head_dim = head_dim self.num_heads = num_heads self.n_groups = n_groups if n_groups % self.tp_size != 0: # - for TP we shard conv_dim by sharding on n_groups, # - but if n_groups cannot divide tp_size, we need to # extend some extra groups groups = MambaStateShapeCalculator.extra_groups_for_head_shards( n_groups, self.tp_size ) self.n_groups = n_groups + groups self.groups_ssm_state_size = self.n_groups * self.ssm_state_size self.conv_dim = intermediate_size + 2 * self.groups_ssm_state_size if n_groups % self.tp_size == 0: self.conv1d = MergedColumnParallelLinear( input_size=conv_kernel_size, output_sizes=[ intermediate_size, self.groups_ssm_state_size, self.groups_ssm_state_size, ], bias=use_conv_bias, quant_config=None, prefix=f"{prefix}.conv1d", ) self.in_proj = MergedColumnParallelLinear( input_size=hidden_size, output_sizes=[ intermediate_size, intermediate_size, self.groups_ssm_state_size, self.groups_ssm_state_size, self.num_heads, ], bias=use_bias, quant_config=quant_config, prefix=f"{prefix}.in_proj", ) else: # This is the n_groups == 1 case, # where we need to duplicate groups if TP>1. self.conv1d = ColumnParallelLinear( input_size=conv_kernel_size, output_size=self.conv_dim, bias=use_conv_bias, quant_config=None, prefix=f"{prefix}.conv1d", ) self.in_proj = ColumnParallelLinear( input_size=hidden_size, output_size=intermediate_size + self.conv_dim + self.num_heads, bias=use_bias, quant_config=quant_config, prefix=f"{prefix}.in_proj", ) # - because in_proj is a concatenation of 3 weights, we # need to interleave them before sharding # - use the custom weight loader mamba_v2_sharded_weight_loader # for conv1d.bias, covn1d.weight and in_proj.weight # - need to set these settings, to assign the groups # to the head shards group_shard_settings = ( self.groups_ssm_state_size, # expected model size (self.n_groups - n_groups) * self.ssm_state_size, # extra dims assigned n_groups == 1, # if there was only one group ) intermediate_settings = (intermediate_size, 0, False) head_settings = (self.num_heads, 0, False) # - the weight already has a "weight_loader" attribute # which set_weight_attrs will raise if we do not # delete before trying to override it # - ditto for the other two weights below delattr(self.conv1d.bias, "weight_loader") set_weight_attrs( self.conv1d.bias, { "weight_loader": mamba_v2_sharded_weight_loader( [ intermediate_settings, group_shard_settings, group_shard_settings, ], self.tp_size, tp_rank, ) }, ) delattr(self.conv1d.weight, "weight_loader") set_weight_attrs( self.conv1d.weight, { "weight_loader": mamba_v2_sharded_weight_loader( [ intermediate_settings, group_shard_settings, group_shard_settings, ], self.tp_size, tp_rank, ) }, ) if quant_config is None: # - quant layers do not have a weight loader delattr(self.in_proj.weight, "weight_loader") set_weight_attrs( self.in_proj.weight, { "weight_loader": mamba_v2_sharded_weight_loader( [ intermediate_settings, # for gate intermediate_settings, group_shard_settings, group_shard_settings, head_settings, # for dt ], self.tp_size, tp_rank, ) }, ) # unsqueeze to fit conv1d weights shape into the linear weights shape. # Can't do this in `weight_loader` since it already exists in # `ColumnParallelLinear` and `MergedColumnParallelLinear`, # and `set_weight_attrs` doesn't allow to override it self.conv1d.weight.data = self.conv1d.weight.data.unsqueeze(1) conv_weights = self.conv1d.weight.view( self.conv1d.weight.size(0), self.conv1d.weight.size(2) ) self.register_buffer("conv_weights", conv_weights, persistent=False) # - these are TPed by heads to reduce the size of the # temporal shape self.A = nn.Parameter( torch.empty( divide(num_heads, self.tp_size), dtype=torch.float32, ) ) self.D = nn.Parameter(torch.ones(num_heads // self.tp_size)) self.dt_bias = nn.Parameter(torch.ones(num_heads // self.tp_size)) self.use_rms_norm = use_rms_norm set_weight_attrs(self.D, {"weight_loader": sharded_weight_loader(0)}) a_weight_loader = composed_weight_loader( sharded_weight_loader(0), lambda x: -torch.exp(x.float()) ) set_weight_attrs(self.A, {"weight_loader": a_weight_loader}) set_weight_attrs(self.dt_bias, {"weight_loader": sharded_weight_loader(0)}) self.out_proj = RowParallelLinear( intermediate_size, hidden_size, bias=use_bias, input_is_parallel=True, quant_config=quant_config, prefix=f"{prefix}.out_proj", ) self.norm = Mixer2RMSNormGated( intermediate_size, n_groups, self.use_rms_norm, eps=rms_norm_eps ) # - get hidden_states, B and C after depthwise convolution. self.split_hidden_states_B_C_fn = lambda hidden_states_B_C: torch.split( hidden_states_B_C, [ self.intermediate_size // self.tp_size, self.groups_ssm_state_size // self.tp_size, self.groups_ssm_state_size // self.tp_size, ], dim=-1, ) compilation_config = get_current_vllm_config().compilation_config if prefix in compilation_config.static_forward_context: raise ValueError(f"Duplicate layer name: {prefix}") compilation_config.static_forward_context[prefix] = self # The tuple is (conv_state, ssm_state) self.kv_cache = (torch.tensor([]), torch.tensor([])) self.model_config = model_config self.cache_config = cache_config self.prefix = prefix # Pre-compute sizes for forward pass self.tped_intermediate_size = self.intermediate_size // self.tp_size self.tped_conv_size = self.conv_dim // self.tp_size self.tped_dt_size = self.num_heads // self.tp_size self.split_hidden_states_B_C_fn = lambda hidden_states_B_C: torch.split( hidden_states_B_C, [ self.tped_intermediate_size, self.groups_ssm_state_size // self.tp_size, self.groups_ssm_state_size // self.tp_size, ], dim=-1, ) def forward_native( self, hidden_states: torch.Tensor, mup_vector: torch.Tensor | None = None, ): pass def forward( self, hidden_states: torch.Tensor, mup_vector: torch.Tensor | None = None, ): # 1. Gated MLP's linear projection projected_states, _ = self.in_proj(hidden_states) if mup_vector is not None: projected_states = projected_states * mup_vector # 2. Prepare inputs for conv + SSM ssm_output = torch.empty( [ hidden_states.shape[0], (self.num_heads // self.tp_size) * self.head_dim, ], dtype=hidden_states.dtype, device=hidden_states.device, ) # 3. conv + SSM # (split `projected_states` into hidden_states_B_C, dt in the custom op to # ensure it is not treated as an intermediate tensor by torch compile) torch.ops.vllm.mamba_mixer2( projected_states, ssm_output, self.prefix, ) # 4. gated MLP # GatedRMSNorm internally applying SiLU to the gate # SiLU is applied internally before normalization, unlike standard # norm usage gate = projected_states[..., : self.tped_intermediate_size] hidden_states = self.norm(ssm_output, gate) # 5. Final linear projection output, _ = self.out_proj(hidden_states) return output def conv_ssm_forward( self, projected_states: torch.Tensor, output: torch.Tensor, ): hidden_states_B_C, dt = torch.split( projected_states[..., self.tped_intermediate_size :], [self.tped_conv_size, self.tped_dt_size], dim=-1, ) forward_context = get_forward_context() # attn_metadata contains metadata necessary for the mamba2 triton # kernels to operate in continuous batching and in chunked prefill # modes; they are computed at top-level model forward since they # stay the same and reused for all mamba layers in the same iteration attn_metadata: AttentionMetadata = forward_context.attn_metadata assert self.cache_config is not None mamba_block_size = self.cache_config.mamba_block_size prefix_caching_enabled = self.cache_config.enable_prefix_caching if attn_metadata is not None: assert isinstance(attn_metadata, dict) attn_metadata = attn_metadata[self.prefix] assert isinstance(attn_metadata, Mamba2AttentionMetadata) self_kv_cache = self.kv_cache[forward_context.virtual_engine] # conv_state = (..., dim, width-1) yet contiguous along 'dim' conv_state = self_kv_cache[0].transpose(-1, -2) ssm_state = self_kv_cache[1] state_indices_tensor = attn_metadata.state_indices_tensor has_initial_states_p = attn_metadata.has_initial_states_p prep_initial_states = attn_metadata.prep_initial_states chunk_size = attn_metadata.chunk_size seq_idx_p = attn_metadata.seq_idx_p query_start_loc_p = attn_metadata.query_start_loc_p cu_chunk_seqlen_p = attn_metadata.cu_chunk_seqlen_p last_chunk_indices_p = attn_metadata.last_chunk_indices_p if attn_metadata is None: # profile run hidden_states_B_C = ( hidden_states_B_C.transpose(0, 1).clone().transpose(0, 1) ).contiguous() hidden_states, _B, _C = self.split_hidden_states_B_C_fn(hidden_states_B_C) return hidden_states num_prefills = attn_metadata.num_prefills # request count num_decodes = attn_metadata.num_decode_tokens # token count (=request) num_prefill_tokens = attn_metadata.num_prefill_tokens # token count has_prefill = num_prefills > 0 has_decode = num_decodes > 0 num_actual_tokens = num_prefill_tokens + num_decodes # Separate prefill and decode by splitting varlen input # Split along token dimension hidden_states_B_C_d, hidden_states_B_C_p = torch.split( hidden_states_B_C[:num_actual_tokens], [num_decodes, num_prefill_tokens], dim=0, ) dt_d, dt_p = torch.split( dt[:num_actual_tokens], [num_decodes, num_prefill_tokens], dim=0, ) # Split along batch dimension state_indices_tensor_d, state_indices_tensor_p = torch.split( state_indices_tensor[:num_actual_tokens], [num_decodes, num_prefills], dim=0, ) if prefix_caching_enabled: # If prefix caching is enabled, retrieve the relevant variables # for prefill and decode block_idx_last_computed_token_d, block_idx_last_computed_token_p = ( torch.split( attn_metadata.block_idx_last_computed_token, [num_decodes, num_prefills], dim=0, ) ) block_idx_last_scheduled_token_d, block_idx_last_scheduled_token_p = ( torch.split( attn_metadata.block_idx_last_scheduled_token, [num_decodes, num_prefills], dim=0, ) ) # Prefill-only variables: block_idx_first_scheduled_token_p = ( attn_metadata.block_idx_first_scheduled_token_p ) num_computed_tokens_p = attn_metadata.num_computed_tokens_p else: block_idx_last_computed_token_d = None block_idx_last_computed_token_p = None block_idx_last_scheduled_token_d = None block_idx_last_scheduled_token_p = None block_idx_first_scheduled_token_p = None num_computed_tokens_p = None preallocated_ssm_out_d, preallocated_ssm_out_p = torch.split( output[:num_actual_tokens], [num_decodes, num_prefill_tokens], dim=0, ) # Process prefill requests if has_prefill: # 2. Convolution sequence transformation # - It will read the initial states for every sequence, # that has "has_initial_states_p" == True, # from "cache_indices", using "state_indices_tensor_p". # - It updates the "conv_state" cache in positions pointed # to by "state_indices_tensor_p". # In particular, it will always write the state at the # sequence end. # In addition, "block_idx_first_scheduled_token_p" and # "block_idx_last_scheduled_token_p" # are provided (which are pointers into # "state_indices_tensor_p"), it will write additional cache # states aligned at "block_size_to_align". x = hidden_states_B_C_p.transpose( 0, 1 ) # this is the form that causal-conv see hidden_states_B_C_p = causal_conv1d_fn( x, self.conv_weights, self.conv1d.bias, activation=self.activation, conv_states=conv_state, has_initial_state=has_initial_states_p, cache_indices=state_indices_tensor_p, block_idx_first_scheduled_token=block_idx_first_scheduled_token_p, block_idx_last_scheduled_token=block_idx_last_scheduled_token_p, initial_state_idx=block_idx_last_computed_token_p, num_computed_tokens=num_computed_tokens_p, block_size_to_align=mamba_block_size, metadata=attn_metadata, query_start_loc=query_start_loc_p, ).transpose(0, 1)[:num_prefill_tokens] hidden_states_p, B_p, C_p = self.split_hidden_states_B_C_fn( hidden_states_B_C_p ) # 3. State Space Model sequence transformation initial_states = None if has_initial_states_p is not None and prep_initial_states: kernel_ssm_indices = state_indices_tensor_p if prefix_caching_enabled: kernel_ssm_indices = state_indices_tensor_p.gather( 1, block_idx_last_computed_token_p.unsqueeze(1) ).squeeze(1) initial_states = torch.where( has_initial_states_p[:, None, None, None], ssm_state[kernel_ssm_indices], 0, ) # NOTE: final output is an in-place update of out tensor varlen_states = mamba_chunk_scan_combined_varlen( hidden_states_p.view( num_prefill_tokens, self.num_heads // self.tp_size, self.head_dim ), dt_p, self.A, B_p.view(num_prefill_tokens, self.n_groups // self.tp_size, -1), C_p.view(num_prefill_tokens, self.n_groups // self.tp_size, -1), chunk_size=chunk_size, D=self.D, z=None, dt_bias=self.dt_bias, seq_idx=seq_idx_p, cu_seqlens=query_start_loc_p, cu_chunk_seqlens=cu_chunk_seqlen_p, last_chunk_indices=last_chunk_indices_p, initial_states=initial_states, return_intermediate_states=prefix_caching_enabled, dt_softplus=True, dt_limit=(0.0, float("inf")), out=preallocated_ssm_out_p.view(num_prefill_tokens, -1, self.head_dim), state_dtype=ssm_state.dtype, ) if prefix_caching_enabled: # The chunk_stride is the number of chunks per mamba block # e.g., if mamba_block_size = 512 and chunk_size = 256, # then chunk_stride = 2 chunk_stride = mamba_block_size // chunk_size # Save state for sequences with more than just final state for seq_idx in range(num_prefills): # Block index for the first scheduled token block_idx_first_scheduled_token = block_idx_first_scheduled_token_p[ seq_idx ] # Block index for the last scheduled token block_idx_last_scheduled_token = block_idx_last_scheduled_token_p[ seq_idx ] # Number of blocks that need to be written n_blocks_to_fill = ( block_idx_last_scheduled_token - block_idx_first_scheduled_token ) # Skip sequences that don't have any blocks to fill if n_blocks_to_fill == 0: continue # Look up the state indices cache_blocks_to_fill = state_indices_tensor_p[ seq_idx, block_idx_first_scheduled_token:block_idx_last_scheduled_token, ] # First chunk index for this sequence if seq_idx == 0: first_chunk = 0 else: first_chunk = 1 + last_chunk_indices_p[seq_idx - 1] # First chunk that is aligned on the mamba block boundary first_aligned_chunk = first_chunk + chunk_stride - 1 # Calculate the number of computed tokens that were not # already cached num_unaligned_computed_tokens = ( num_computed_tokens_p[seq_idx] % mamba_block_size ) if num_unaligned_computed_tokens > 0: # If the number of computed tokens is not block aligned, # then we need to shift the index accordingly first_aligned_chunk -= ( num_unaligned_computed_tokens // chunk_size ) # Get states to write from_where = varlen_states[ first_aligned_chunk : first_aligned_chunk + n_blocks_to_fill * chunk_stride : chunk_stride ] # Write the states ssm_state[cache_blocks_to_fill] = from_where # For all seqs, store the last state (note: might be partial): ssm_state[ state_indices_tensor_p.gather( 1, block_idx_last_scheduled_token_p.unsqueeze(1) ).squeeze(1) ] = varlen_states[last_chunk_indices_p] else: # update ssm states # - varlen state is a (num_prefills, nheads, headdim, dstate)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/__init__.py
vllm/model_executor/layers/mamba/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/ops/ssd_chunk_scan.py
vllm/model_executor/layers/mamba/ops/ssd_chunk_scan.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) 2024, Tri Dao, Albert Gu. # Adapted from https://github.com/state-spaces/mamba/blob/v2.2.4/mamba_ssm/ops/triton/ssd_chunk_scan.py # ruff: noqa: E501,SIM102 from packaging import version from vllm.triton_utils import tl, triton TRITON_22 = version.parse(triton.__version__) >= version.parse("2.2.0") @triton.autotune( configs=[ triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 256, "BLOCK_SIZE_K": 64}, num_stages=3, num_warps=8, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 256, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 128, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 128, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 64}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 128, "BLOCK_SIZE_K": 64}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 32, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 32, "BLOCK_SIZE_K": 32}, num_stages=5, num_warps=2, ), triton.Config( {"BLOCK_SIZE_M": 32, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=5, num_warps=2, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=2, ), ], key=["chunk_size", "hdim", "dstate", "IS_CAUSAL"], ) @triton.jit def _chunk_scan_fwd_kernel( # Pointers to matrices cb_ptr, x_ptr, z_ptr, out_ptr, dt_ptr, dA_cumsum_ptr, seq_idx_ptr, C_ptr, states_ptr, D_ptr, initstates_ptr, cu_chunk_seqlens_ptr, # Matrix dimensions chunk_size: tl.constexpr, hdim: tl.constexpr, dstate: tl.constexpr, seqlen, nheads_ngroups_ratio: tl.constexpr, # Strides stride_cb_chunk: tl.int64, stride_cb_head: tl.int64, stride_cb_csize_m: tl.int64, stride_cb_csize_k: tl.constexpr, stride_x_seqlen: tl.int64, stride_x_head: tl.int64, stride_x_hdim: tl.constexpr, stride_z_seqlen: tl.int64, stride_z_head: tl.int64, stride_z_hdim: tl.constexpr, stride_out_seqlen: tl.int64, stride_out_head: tl.int64, stride_out_hdim: tl.constexpr, stride_dt_chunk: tl.int64, stride_dt_head: tl.int64, stride_dt_csize: tl.constexpr, stride_dA_cs_chunk: tl.int64, stride_dA_cs_head: tl.int64, stride_dA_cs_csize: tl.constexpr, stride_seq_idx_chunk: tl.constexpr, stride_C_seqlen: tl.int64, stride_C_head: tl.int64, stride_C_dstate: tl.constexpr, stride_states_chunk: tl.int64, stride_states_head: tl.int64, stride_states_hdim: tl.int64, stride_states_dstate: tl.constexpr, stride_init_states_batch: tl.int64, stride_init_states_head: tl.int64, stride_init_states_hdim: tl.int64, stride_init_states_dstate: tl.constexpr, stride_D_head: tl.constexpr, # Meta-parameters IS_CAUSAL: tl.constexpr, HAS_D: tl.constexpr, D_HAS_HDIM: tl.constexpr, HAS_Z: tl.constexpr, BLOCK_SIZE_M: tl.constexpr, BLOCK_SIZE_N: tl.constexpr, BLOCK_SIZE_K: tl.constexpr, BLOCK_SIZE_DSTATE: tl.constexpr, IS_TRITON_22: tl.constexpr, HAS_INITSTATES: tl.constexpr, ): pid_c = tl.program_id(axis=1).to(tl.int64) pid_h = tl.program_id(axis=2) num_pid_n = tl.cdiv(hdim, BLOCK_SIZE_N) pid_m = tl.program_id(axis=0) // num_pid_n pid_n = tl.program_id(axis=0) % num_pid_n cb_ptr += pid_c * stride_cb_chunk + (pid_h // nheads_ngroups_ratio) * stride_cb_head chunk_seqlen_start = tl.load(cu_chunk_seqlens_ptr + pid_c) chunk_seqlen_end = tl.load(cu_chunk_seqlens_ptr + pid_c + 1) x_ptr += chunk_seqlen_start * stride_x_seqlen + pid_h * stride_x_head dt_ptr += pid_c * stride_dt_chunk + pid_h * stride_dt_head dA_cumsum_ptr += pid_c * stride_dA_cs_chunk + pid_h * stride_dA_cs_head C_ptr += ( chunk_seqlen_start * stride_C_seqlen + (pid_h // nheads_ngroups_ratio) * stride_C_head ) # M-block offsets and prev states # - logic in next block may override these if there is an active offset offs_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) seq_idx_ptr += pid_c * stride_seq_idx_chunk seq_idx = tl.load(seq_idx_ptr) seq_idx_prev = tl.load( seq_idx_ptr - stride_seq_idx_chunk, mask=pid_c >= 1, other=-1 ) if HAS_INITSTATES and (seq_idx != seq_idx_prev): prev_states_ptr = ( initstates_ptr + seq_idx * stride_init_states_batch + pid_h * stride_init_states_head ) prev_states_hdim = stride_init_states_hdim prev_states_dstate = stride_init_states_dstate else: prev_states_ptr = ( states_ptr + (pid_c - 1) * stride_states_chunk + pid_h * stride_states_head ) prev_states_hdim = stride_states_hdim prev_states_dstate = stride_states_dstate chunk_size_limit = chunk_seqlen_end - chunk_seqlen_start offs_n = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) dA_cs_m = tl.load( dA_cumsum_ptr + offs_m * stride_dA_cs_csize, mask=offs_m < chunk_size, other=0.0 ).to(tl.float32) acc = tl.zeros((BLOCK_SIZE_M, BLOCK_SIZE_N), dtype=tl.float32) offs_out_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offs_out_n = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) # Faster to just do 1 iteration with larger BLOCK_SIZE_K, up to block size 128 offs_k_dstate = tl.arange( 0, BLOCK_SIZE_DSTATE if BLOCK_SIZE_DSTATE <= 128 else BLOCK_SIZE_K ) C_ptrs = C_ptr + ( offs_m[:, None] * stride_C_seqlen + offs_k_dstate[None, :] * stride_C_dstate ) scale_m = tl.exp(dA_cs_m) if BLOCK_SIZE_DSTATE <= 128: C = tl.load( C_ptrs, mask=(offs_m[:, None] < chunk_size_limit) & (offs_k_dstate[None, :] < dstate), other=0.0, ) if not HAS_INITSTATES and (seq_idx != seq_idx_prev): # if no init states AND starting a new sequence, we need zeros prev_states = tl.zeros( (BLOCK_SIZE_DSTATE, BLOCK_SIZE_N), dtype=C_ptr.dtype.element_ty ) else: # otherwise read the previous state prev_states_ptrs = ( prev_states_ptr + offs_n[None, :] * prev_states_hdim + offs_k_dstate[:, None] * prev_states_dstate ) prev_states = tl.load( prev_states_ptrs, mask=(offs_k_dstate[:, None] < dstate) & (offs_n[None, :] < hdim), other=0.0, ) prev_states = prev_states.to(C_ptr.dtype.element_ty) acc = tl.dot(C, prev_states) * scale_m[:, None] else: prev_states_ptrs = ( prev_states_ptr + offs_n[None, :] * prev_states_hdim + offs_k_dstate[:, None] * prev_states_dstate ) for k in range(0, dstate, BLOCK_SIZE_K): C = tl.load( C_ptrs, mask=(offs_m[:, None] < chunk_size_limit) & (offs_k_dstate[None, :] < dstate - k), other=0.0, ) if not HAS_INITSTATES and (seq_idx != seq_idx_prev): prev_states = tl.zeros( (BLOCK_SIZE_K, BLOCK_SIZE_N), dtype=C_ptr.dtype.element_ty ) else: prev_states = tl.load( prev_states_ptrs, mask=(offs_k_dstate[:, None] < dstate - k) & (offs_n[None, :] < hdim), other=0.0, ) prev_states = prev_states.to(C_ptr.dtype.element_ty) acc += tl.dot(C, prev_states) C_ptrs += BLOCK_SIZE_K prev_states_ptrs += BLOCK_SIZE_K acc *= scale_m[:, None] offs_k = tl.arange(0, BLOCK_SIZE_K) cb_ptrs = cb_ptr + ( offs_m[:, None] * stride_cb_csize_m + offs_k[None, :] * stride_cb_csize_k ) x_ptrs = x_ptr + ( offs_k[:, None] * stride_x_seqlen + offs_n[None, :] * stride_x_hdim ) dt_ptrs = dt_ptr + offs_k * stride_dt_csize dA_cumsum_ptrs = dA_cumsum_ptr + offs_k * stride_dA_cs_csize K_MAX = ( chunk_size_limit if not IS_CAUSAL else min((pid_m + 1) * BLOCK_SIZE_M, chunk_size_limit) ) for k in range(0, K_MAX, BLOCK_SIZE_K): cb = tl.load( cb_ptrs, mask=(offs_m[:, None] < chunk_size) & (offs_k[None, :] < chunk_size - k), other=0.0, ).to(tl.float32) dA_cs_k = tl.load(dA_cumsum_ptrs, mask=offs_k < chunk_size - k, other=0.0).to( tl.float32 ) # If there's seq_idx, we already set cb[i, j] = 0 for seq_idx[i] != seq_idx[j]. # So we don't need masking wrt seq_idx here. cb *= tl.exp(dA_cs_m[:, None] - dA_cs_k[None, :]) dt_k = tl.load(dt_ptrs, mask=offs_k < chunk_size - k, other=0.0).to(tl.float32) cb *= dt_k if IS_CAUSAL: mask = offs_m[:, None] >= k + offs_k[None, :] cb = tl.where(mask, cb, 0.0) cb = cb.to(x_ptr.dtype.element_ty) x = tl.load( x_ptrs, mask=(offs_k[:, None] < chunk_size_limit - k) & (offs_n[None, :] < hdim), other=0.0, ) acc += tl.dot(cb, x) cb_ptrs += BLOCK_SIZE_K * stride_cb_csize_k x_ptrs += BLOCK_SIZE_K * stride_x_seqlen dt_ptrs += BLOCK_SIZE_K * stride_dt_csize dA_cumsum_ptrs += BLOCK_SIZE_K * stride_dA_cs_csize offs_out_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offs_out_n = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) if HAS_D: if D_HAS_HDIM: D = tl.load( D_ptr + pid_h * stride_D_head + offs_n, mask=offs_n < hdim, other=0.0 ).to(tl.float32) else: D = tl.load(D_ptr + pid_h * stride_D_head).to(tl.float32) x_residual = tl.load( x_ptr + (offs_m[:, None] * stride_x_seqlen + offs_n[None, :] * stride_x_hdim), mask=(offs_m[:, None] < chunk_size_limit) & (offs_n[None, :] < hdim), other=0.0, ).to(tl.float32) acc += x_residual * D if HAS_Z: z_ptr += chunk_seqlen_start * stride_z_seqlen + pid_h * stride_z_head z_ptrs = z_ptr + ( stride_z_seqlen * offs_out_m[:, None] + stride_z_hdim * offs_out_n[None, :] ) z = tl.load( z_ptrs, mask=(offs_out_m[:, None] < chunk_size_limit) & (offs_out_n[None, :] < hdim), other=0.0, ).to(tl.float32) acc *= z * tl.sigmoid(z) out_ptr += chunk_seqlen_start * stride_out_seqlen + pid_h * stride_out_head out_ptrs = out_ptr + ( stride_out_seqlen * offs_out_m[:, None] + offs_out_n[None, :] * stride_out_hdim ) tl.store( out_ptrs, acc, mask=(offs_out_m[:, None] < chunk_size_limit) & (offs_out_n[None, :] < hdim), ) def _chunk_scan_fwd( cb, x, dt, dA_cumsum, C, states, cu_chunk_seqlens, out, seq_idx, D=None, z=None, initial_states=None, ): assert seq_idx is not None, "this implementation requires seq_idx" seqlen, nheads, headdim = x.shape _, nchunks, chunk_size = dt.shape _, ngroups, dstate = C.shape assert nheads % ngroups == 0 assert C.shape == (seqlen, ngroups, dstate) assert cb.shape == (nchunks, ngroups, chunk_size, chunk_size) if D is not None: assert D.shape == (nheads, headdim) or D.shape == (nheads,) if z is not None: assert z.shape == x.shape assert dt.shape == (nheads, nchunks, chunk_size) assert dA_cumsum.shape == (nheads, nchunks, chunk_size) assert states.shape == (nchunks, nheads, headdim, dstate) assert seq_idx.shape == (nchunks,) grid = lambda META: ( triton.cdiv(chunk_size, META["BLOCK_SIZE_M"]) * triton.cdiv(headdim, META["BLOCK_SIZE_N"]), nchunks, nheads, ) z_strides = (z.stride(0), z.stride(1), z.stride(2)) if z is not None else (0, 0, 0) initial_states_strides = ( ( initial_states.stride(0), initial_states.stride(1), initial_states.stride(2), initial_states.stride(3), ) if initial_states is not None else (0, 0, 0, 0) ) _chunk_scan_fwd_kernel[grid]( cb_ptr=cb, x_ptr=x, z_ptr=z, out_ptr=out, dt_ptr=dt, dA_cumsum_ptr=dA_cumsum, seq_idx_ptr=seq_idx, C_ptr=C, states_ptr=states, D_ptr=D, initstates_ptr=initial_states, cu_chunk_seqlens_ptr=cu_chunk_seqlens, chunk_size=chunk_size, hdim=headdim, dstate=dstate, seqlen=seqlen, nheads_ngroups_ratio=nheads // ngroups, stride_cb_chunk=cb.stride(0), stride_cb_head=cb.stride(1), stride_cb_csize_m=cb.stride(2), stride_cb_csize_k=cb.stride(3), stride_x_seqlen=x.stride(0), stride_x_head=x.stride(1), stride_x_hdim=x.stride(2), stride_z_seqlen=z_strides[0], stride_z_head=z_strides[1], stride_z_hdim=z_strides[2], stride_out_seqlen=out.stride(0), stride_out_head=out.stride(1), stride_out_hdim=out.stride(2), stride_dt_chunk=dt.stride(1), stride_dt_head=dt.stride(0), stride_dt_csize=dt.stride(2), stride_dA_cs_chunk=dA_cumsum.stride(1), stride_dA_cs_head=dA_cumsum.stride(0), stride_dA_cs_csize=dA_cumsum.stride(2), stride_seq_idx_chunk=seq_idx.stride(0), stride_C_seqlen=C.stride(0), stride_C_head=C.stride(1), stride_C_dstate=C.stride(2), stride_states_chunk=states.stride(0), stride_states_head=states.stride(1), stride_states_hdim=states.stride(2), stride_states_dstate=states.stride(3), stride_init_states_batch=initial_states_strides[0], stride_init_states_head=initial_states_strides[1], stride_init_states_hdim=initial_states_strides[2], stride_init_states_dstate=initial_states_strides[3], stride_D_head=D.stride(0) if D is not None else 0, IS_CAUSAL=True, HAS_D=D is not None, D_HAS_HDIM=D.dim() == 2 if D is not None else True, HAS_Z=z is not None, BLOCK_SIZE_DSTATE=max(triton.next_power_of_2(dstate), 16), IS_TRITON_22=TRITON_22, HAS_INITSTATES=initial_states is not None, ) return
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/ops/ssd_combined.py
vllm/model_executor/layers/mamba/ops/ssd_combined.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) 2024, Tri Dao, Albert Gu. # Adapted from https://github.com/state-spaces/mamba/blob/v2.2.4/mamba_ssm/ops/triton/ssd_combined.py # ruff: noqa: E501 import torch from einops import rearrange from packaging import version from vllm.triton_utils import triton from .ssd_bmm import _bmm_chunk_fwd from .ssd_chunk_scan import _chunk_scan_fwd from .ssd_chunk_state import _chunk_cumsum_fwd, _chunk_state_fwd from .ssd_state_passing import _state_passing_fwd TRITON_22 = version.parse(triton.__version__) >= version.parse("2.2.0") def is_int_pow_2(n): return isinstance(n, int) and n > 0 and (n & (n - 1)) == 0 def _mamba_chunk_scan_combined_fwd( x, dt, A, B, C, chunk_size, out, D=None, z=None, dt_bias=None, initial_states=None, return_intermediate_states=False, seq_idx=None, cu_seqlens=None, cu_chunk_seqlens=None, last_chunk_indices=None, dt_softplus=False, dt_limit=(0.0, float("inf")), state_dtype=None, ): assert is_int_pow_2(chunk_size), "chunk_size must be integer power of 2" seqlen, nheads, headdim = x.shape _, ngroups, dstate = B.shape assert nheads % ngroups == 0 assert B.shape == (seqlen, ngroups, dstate) assert dt.shape == (seqlen, nheads) assert A.shape == (nheads,) assert C.shape == B.shape if z is not None: assert z.shape == x.shape if D is not None: assert D.shape == (nheads, headdim) or D.shape == (nheads,) if seq_idx is not None: assert seq_idx.shape == (cu_chunk_seqlens.shape[0] - 1,) if B.stride(-1) != 1: B = B.contiguous() if C.stride(-1) != 1: C = C.contiguous() if ( x.stride(-1) != 1 and x.stride(0) != 1 ): # Either M or K dimension should be contiguous x = x.contiguous() if ( z is not None and z.stride(-1) != 1 and z.stride(0) != 1 ): # Either M or K dimension should be contiguous z = z.contiguous() if D is not None and D.stride(-1) != 1: D = D.contiguous() assert cu_seqlens is not None, "Assuming varlen input - must supply cu_seqlens" if initial_states is not None: assert initial_states.shape == (len(cu_seqlens) - 1, nheads, headdim, dstate) # This function executes 5 sub-functions for computing mamba # - a good resource is the blog https://goombalab.github.io/blog/2024/mamba2-part3-algorithm/ # which has a minimal implementation to understand the below operations # - as explained by the blog, mamba is a special case of causal attention # - the idea is to chunk the attention matrix and compute each # submatrix separately using different optimizations. # - see the blog and paper for a visualization of the submatrices # which we refer to in the comments below # 1. Compute chunked cumsum of A * dt # - here dt may go through a softplus activation dA_cumsum, dt = _chunk_cumsum_fwd( dt, A, chunk_size, cu_chunk_seqlens, dt_bias=dt_bias, dt_softplus=dt_softplus, dt_limit=dt_limit, ) # 2. Compute the state for each intra-chunk # (right term of low-rank factorization of off-diagonal blocks; B terms) states = _chunk_state_fwd( B, x, dt, dA_cumsum, cu_chunk_seqlens, states_in_fp32=True ) # 3. Compute the inter-chunk SSM recurrence; produces correct SSM states at chunk boundaries # (middle term of factorization of off-diag blocks; A terms) # - for handling chunked prefill, this requires i) initial_states and # ii) seq_idx to be all specified. # - When a new seq_idx is detected, we will stop passing the prev_state # and switch accordingly to the init_state corresponding to the new seq_idx. states = _state_passing_fwd( rearrange(states, "... p n -> ... (p n)"), dA_cumsum, # (nheads, nchunks, chunk_size) cu_chunk_seqlens, initial_states=rearrange(initial_states, "... p n -> ... (p n)") if initial_states is not None else None, # (batch, nheads, headdim*dstate) seq_idx=seq_idx, out_dtype=state_dtype if state_dtype is not None else C.dtype, ) states = rearrange(states, "... (p n) -> ... p n", n=dstate) # 4. Compute batched matrix multiply for C_j^T B_i terms CB = _bmm_chunk_fwd(C, B, chunk_size, cu_chunk_seqlens, output_dtype=torch.float32) # 5. Scan and compute the diagonal blocks, taking into # account past causal states. # - if initial states are provided, then states information will be # augmented with initial_states. # - to do this properly, we need to account for example changes in # the continuous batch, therefore we introduce pseudo chunks, which is # a chunk that is split up each time an example changes. # - in each (pseudo) chunk, we detect if the previous (pseudo) chunk had # a seq_idx change, in which case we take states information from # init_states. _chunk_scan_fwd( CB, x, dt, dA_cumsum, C, states, cu_chunk_seqlens, out, # in-place update seq_idx, D=D, z=z, initial_states=initial_states, ) if return_intermediate_states: return states else: return states[last_chunk_indices] def mamba_chunk_scan_combined_varlen( x, dt, A, B, C, chunk_size, cu_seqlens, cu_chunk_seqlens, last_chunk_indices, seq_idx, out, D=None, z=None, dt_bias=None, initial_states=None, dt_softplus=False, dt_limit=(0.0, float("inf")), return_intermediate_states=False, state_dtype=None, ): """ Argument: x: (seqlen, nheads, headdim) dt: (seqlen, nheads) A: (nheads) B: (seqlen, ngroups, dstate) C: (seqlen, ngroups, dstate) chunk_size: int cu_seqlens: (batch + 1,) cu_chunk_seqlens: (nchunks + 1,) last_chunk_indices: (batch,) seq_idx: (nchunks,) out: (seqlen, nheads, headdim) preallocated output tensor D: (nheads, headdim) or (nheads,) z: (seqlen, nheads, headdim) dt_bias: (nheads,) initial_states: (batch, nheads, headdim, dstate) dt_softplus: Whether to apply softplus to dt out: (seqlen, nheads, headdim) preallocated output tensor state_dtype: The data type of the ssm state Return: varlen_states: (batch, nheads, headdim, dstate) """ assert cu_seqlens is not None, "cu_seqlens must be provided assuming varlen input" assert seq_idx is not None varlen_states = _mamba_chunk_scan_combined_fwd( x, dt, A, B, C, chunk_size, out, D=D, z=z, dt_bias=dt_bias, initial_states=initial_states, return_intermediate_states=return_intermediate_states, seq_idx=seq_idx, cu_seqlens=cu_seqlens, cu_chunk_seqlens=cu_chunk_seqlens, last_chunk_indices=last_chunk_indices, dt_softplus=dt_softplus, dt_limit=dt_limit, state_dtype=state_dtype, ) return varlen_states
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/ops/causal_conv1d.py
vllm/model_executor/layers/mamba/ops/causal_conv1d.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) 2024, Tri Dao. # Adapted from https://github.com/Dao-AILab/causal-conv1d/blob/main/causal_conv1d/causal_conv1d_interface.py import numpy as np import torch from vllm.attention.backends.utils import PAD_SLOT_ID from vllm.triton_utils import tl, triton @triton.jit() def _causal_conv1d_fwd_kernel( # continuous batching # Pointers to matrices x_ptr, # (dim, cu_seqlen) holding `batch` of actual sequences + padded sequences w_ptr, # (dim, width) bias_ptr, initial_states_ptr, # conv_states_ptr cache_indices_ptr, # (batch, n_blocks + padding) The second dimension contains # the block indices relevant for each sequence # plus potential 0-padding at the beginning and at the end has_initial_states_ptr, query_start_loc_ptr, batch_ptr, token_chunk_offset_ptr, block_idx_first_scheduled_token, # (batch,) block_idx_last_scheduled_token, # (batch,) initial_state_idx, # (batch,) num_computed_tokens, # (batch,) o_ptr, # (dim, seqlen) - actually pointing to x_ptr # Matrix dimensions dim: tl.constexpr, seqlen: tl.int32, # cu_seqlen num_cache_lines: tl.constexpr, # added to support vLLM larger cache lines # Strides stride_x_dim: tl.constexpr, # stride to get to next feature-value, stride_x_token: tl.constexpr, # stride to get to next token (same feature-index, same sequence-index) stride_w_dim: tl.constexpr, # stride to get to next dim-axis value stride_w_width: tl.constexpr, # stride to get to next width-axis value stride_istate_seq: tl.constexpr, stride_istate_dim: tl.constexpr, stride_istate_token: tl.constexpr, stride_cache_indices: tl.constexpr, stride_o_dim: tl.constexpr, stride_o_token: tl.constexpr, stride_block_m: tl.constexpr, # Stride block to align divided by BLOCK_M # others pad_slot_id: tl.constexpr, # Meta-parameters HAS_BIAS: tl.constexpr, KERNEL_WIDTH: tl.constexpr, SILU_ACTIVATION: tl.constexpr, IS_APC_ENABLED: tl.constexpr, USE_PAD_SLOT: tl.constexpr, NP2_STATELEN: tl.constexpr, BLOCK_M: tl.constexpr, BLOCK_N: tl.constexpr, ): conv_states_ptr = initial_states_ptr conv_state_indices_ptr = cache_indices_ptr stride_conv_state_seq = stride_istate_seq stride_conv_state_dim = stride_istate_dim stride_conv_state_tok = stride_istate_token state_len = ( KERNEL_WIDTH - 1 ) # can be passed via argument if it's not the same as this value # one program handles one chunk in a single sequence # rather than mixing sequences - to make updating initial_states across sequences efficiently # single-sequence id idx_seq = tl.load(batch_ptr + tl.program_id(0)).to(tl.int64) chunk_offset = tl.load(token_chunk_offset_ptr + tl.program_id(0)) # BLOCK_N elements along the feature-dimension (channel) idx_feats = tl.program_id(1) * BLOCK_N + tl.arange(0, BLOCK_N) if idx_seq == pad_slot_id: return sequence_start_index = tl.load(query_start_loc_ptr + idx_seq) sequence_end_index = tl.load(query_start_loc_ptr + idx_seq + 1) # find the actual sequence length seqlen = sequence_end_index - sequence_start_index B_size: tl.constexpr = stride_block_m * BLOCK_M if IS_APC_ENABLED: # Handle the case if prefix caching is enabled. # In particular, if prefix caching is enabled, the program write additional cache states to "cache_indices_ptr" # Get the length of the completed sequence so far and compute the offset. current_first_index = tl.load(block_idx_first_scheduled_token + idx_seq) current_last_index = tl.load(block_idx_last_scheduled_token + idx_seq) sequence_completed_index = tl.load(num_computed_tokens + idx_seq) # Compute the offset where the first stride_block_m-aligned first full block is # Value in "token-space" sequence_completed_offset_token = sequence_completed_index % B_size seq_completed_offset = B_size - sequence_completed_offset_token seq_end_offset = (seqlen - seq_completed_offset) % B_size last_full_block_token_index = sequence_end_index - seq_end_offset # If the sequence without the sequence_offset_index is stride_cache_chunk-aligned, then the last full chunk is the second-to-last one if seq_end_offset == 0: last_full_block_token_index = last_full_block_token_index - B_size # Get the number of blocks to be filled for the current sequence # If n_block_to_fill = 0, then only the state at the sequence end is stored n_block_to_fill = current_last_index - current_first_index # Get the index of the init block conv_state_init_index = tl.load(initial_state_idx + idx_seq) else: n_block_to_fill = 0 current_last_index = 0 conv_state_init_index = 0 current_first_index = 0 last_full_block_token_index = 0 token_offset = BLOCK_M * chunk_offset segment_len = min(BLOCK_M, seqlen - token_offset) # base of the sequence x_base = ( x_ptr + sequence_start_index * stride_x_token + idx_feats * stride_x_dim ) # [BLOCK_N,] # cache_idx conv_states_input_coord = tl.load( conv_state_indices_ptr + idx_seq * stride_cache_indices + conv_state_init_index ).to(tl.int64) if USE_PAD_SLOT: # noqa if conv_states_input_coord == pad_slot_id: # not processing as this is not the actual sequence return conv_states_base = ( conv_states_ptr + (conv_states_input_coord * stride_conv_state_seq) + (idx_feats * stride_conv_state_dim) ) # [BLOCK_N,] w_base = w_ptr + (idx_feats * stride_w_dim) # [BLOCK_N,] # Does 2 things: # 1. READ prior-block init-state data - [done by every Triton programs] # 2. update conv_state with new data [only by the Triton program handles chunk_offset=0] if chunk_offset == 0: # read from conv_states load_init_state = tl.load(has_initial_states_ptr + idx_seq).to(tl.int1) if load_init_state: # load from conv_states prior_tokens = conv_states_base + (state_len - 1) * stride_conv_state_tok mask_w = idx_feats < dim if KERNEL_WIDTH == 2: conv_states_ptrs = prior_tokens # [BLOCK_N] col0 = tl.load(conv_states_ptrs, mask_w, 0.0) if KERNEL_WIDTH == 3: conv_states_ptrs = prior_tokens # [BLOCK_N] col1 = tl.load(conv_states_ptrs, mask_w, 0.0) conv_states_ptrs = prior_tokens - 1 * stride_conv_state_tok # [BLOCK_N] col0 = tl.load(conv_states_ptrs, mask_w, 0.0) if KERNEL_WIDTH == 4: conv_states_ptrs = prior_tokens # [BLOCK_N] col2 = tl.load(conv_states_ptrs, mask_w, 0.0) conv_states_ptrs = prior_tokens - 1 * stride_conv_state_tok # [BLOCK_N] col1 = tl.load(conv_states_ptrs, mask_w, 0.0) conv_states_ptrs = prior_tokens - 2 * stride_conv_state_tok # [BLOCK_N] col0 = tl.load(conv_states_ptrs, mask_w, 0.0) if KERNEL_WIDTH == 5: conv_states_ptrs = prior_tokens # [BLOCK_N] col3 = tl.load(conv_states_ptrs, mask_w, 0.0) conv_states_ptrs = prior_tokens - 1 * stride_conv_state_tok # [BLOCK_N] col2 = tl.load(conv_states_ptrs, mask_w, 0.0) conv_states_ptrs = prior_tokens - 2 * stride_conv_state_tok # [BLOCK_N] col1 = tl.load(conv_states_ptrs, mask_w, 0.0) conv_states_ptrs = prior_tokens - 3 * stride_conv_state_tok # [BLOCK_N] col0 = tl.load(conv_states_ptrs, mask_w, 0.0) else: # prior-tokens are zeros if KERNEL_WIDTH >= 2: # STRATEGY1 # first chunk and does not have prior-token, so just set to 0 col0 = tl.zeros((BLOCK_N,), dtype=x_ptr.dtype.element_ty) if KERNEL_WIDTH >= 3: # STRATEGY1 col1 = tl.zeros((BLOCK_N,), dtype=x_ptr.dtype.element_ty) if KERNEL_WIDTH >= 4: # STRATEGY1 col2 = tl.zeros((BLOCK_N,), dtype=x_ptr.dtype.element_ty) if KERNEL_WIDTH >= 5: # STRATEGY1 col3 = tl.zeros((BLOCK_N,), dtype=x_ptr.dtype.element_ty) # STEP 2: # here prepare data for updating conv_state if ( state_len <= seqlen ): # SMALL_CACHE=True (only move part of 'x' into conv_state cache) # just read from 'x' # copy 'x' data to conv_state # load only 'x' data (and set 0 before 'x' if seqlen < state_len) idx_tokens_last = (seqlen - state_len) + tl.arange( 0, NP2_STATELEN ) # [BLOCK_M] x_ptrs = ( x_ptr + ((sequence_start_index + idx_tokens_last) * stride_x_token)[:, None] + (idx_feats * stride_x_dim)[None, :] ) # [BLOCK_M,BLOCK_N,] mask_x = ( (idx_tokens_last >= 0)[:, None] & (idx_tokens_last < seqlen)[:, None] & (idx_feats < dim)[None, :] ) # token-index # token-index # feature-index loaded_x = tl.load(x_ptrs, mask_x, 0.0) idx_tokens_conv = tl.arange(0, NP2_STATELEN) # [BLOCK_M] # Compute the offset where the last block should be written in the conv_states conv_states_output_coord = tl.load( conv_state_indices_ptr + idx_seq * stride_cache_indices + current_last_index ).to(tl.int64) conv_states_ptrs_target = ( conv_states_ptr + (conv_states_output_coord * stride_conv_state_seq) # Offset from seq + (idx_feats * stride_conv_state_dim) )[None, :] + ( # [BLOCK_N,] idx_tokens_conv * stride_conv_state_tok )[:, None] mask = (idx_tokens_conv < state_len)[:, None] & (idx_feats < dim)[None, :] tl.debug_barrier() # NOTE: use this due to bug in Triton compiler tl.store(conv_states_ptrs_target, loaded_x, mask) else: if load_init_state: # update conv_state by shifting left, i.e. take last few cols from conv_state + cols from 'x' idx_tokens_conv = tl.arange(0, NP2_STATELEN) # [BLOCK_M] conv_states_ptrs_source = ( conv_states_ptr + (conv_states_input_coord * stride_conv_state_seq) + (idx_feats * stride_conv_state_dim)[None, :] + ((idx_tokens_conv + seqlen) * stride_conv_state_tok)[:, None] ) # [BLOCK_M, BLOCK_N] mask = ( (conv_states_input_coord < num_cache_lines) & ((idx_tokens_conv + seqlen) < state_len)[:, None] & (idx_feats < dim)[None, :] ) conv_state = tl.load(conv_states_ptrs_source, mask, other=0.0) VAL = state_len - seqlen x_ptrs = ( x_base[None, :] + ((idx_tokens_conv - VAL) * stride_x_token)[:, None] ) # [BLOCK_M, BLOCK_N] mask_x = ( (idx_tokens_conv - VAL >= 0)[:, None] & (idx_tokens_conv - VAL < seqlen)[:, None] & (idx_feats < dim)[None, :] ) # token-index # token-index # feature-index loaded_x = tl.load(x_ptrs, mask_x, 0.0) tl.debug_barrier() # need this due to the bug in tl.where not enforcing this when data is the result of another tl.load new_conv_state = tl.where( mask, conv_state, loaded_x ) # BUG in 'tl.where' which requires a barrier before this conv_states_ptrs_target = ( conv_states_base + (idx_tokens_conv * stride_conv_state_tok)[:, None] ) # [BLOCK_M, BLOCK_N] mask = (idx_tokens_conv < state_len)[:, None] & (idx_feats < dim)[ None, : ] tl.store(conv_states_ptrs_target, new_conv_state, mask) else: # load_init_state == False # update conv_state by shifting left, BUT # set cols prior to 'x' as zeros + cols from 'x' idx_tokens_conv = tl.arange(0, NP2_STATELEN) # [BLOCK_M] VAL = state_len - seqlen x_ptrs = ( x_base[None, :] + ((idx_tokens_conv - VAL) * stride_x_token)[:, None] ) # [BLOCK_M, BLOCK_N] mask_x = ( (idx_tokens_conv - VAL >= 0)[:, None] & (idx_tokens_conv - VAL < seqlen)[:, None] & (idx_feats < dim)[None, :] ) # token-index # token-index # feature-index new_conv_state = tl.load(x_ptrs, mask_x, 0.0) conv_states_ptrs_target = ( conv_states_base + (idx_tokens_conv * stride_conv_state_tok)[:, None] ) # [BLOCK_M, BLOCK_N] mask = (idx_tokens_conv < state_len)[:, None] & (idx_feats < dim)[ None, : ] tl.store(conv_states_ptrs_target, new_conv_state, mask) else: # chunk_offset > 0 # read prior-token data from `x` load_init_state = True prior_tokens = x_base + (token_offset - 1) * stride_x_token mask_w = idx_feats < dim if KERNEL_WIDTH == 2: conv_states_ptrs = prior_tokens # [BLOCK_N] col0 = tl.load(conv_states_ptrs, mask_w, 0.0, cache_modifier=".ca") if KERNEL_WIDTH == 3: conv_states_ptrs = prior_tokens # [BLOCK_N] col1 = tl.load(conv_states_ptrs, mask_w, 0.0, cache_modifier=".ca") conv_states_ptrs = prior_tokens - 1 * stride_x_token # [BLOCK_N] col0 = tl.load(conv_states_ptrs, mask_w, 0.0, cache_modifier=".ca") if KERNEL_WIDTH == 4: conv_states_ptrs = prior_tokens # [BLOCK_N] col2 = tl.load(conv_states_ptrs, mask_w, 0.0, cache_modifier=".ca") conv_states_ptrs = prior_tokens - 1 * stride_x_token # [BLOCK_N] col1 = tl.load(conv_states_ptrs, mask_w, 0.0, cache_modifier=".ca") conv_states_ptrs = prior_tokens - 2 * stride_x_token # [BLOCK_N] col0 = tl.load(conv_states_ptrs, mask_w, 0.0, cache_modifier=".ca") if KERNEL_WIDTH == 5: # ruff: noqa: F841 conv_states_ptrs = prior_tokens # [BLOCK_N] col3 = tl.load(conv_states_ptrs, mask_w, 0.0, cache_modifier=".ca") conv_states_ptrs = prior_tokens - 1 * stride_x_token # [BLOCK_N] col2 = tl.load(conv_states_ptrs, mask_w, 0.0, cache_modifier=".ca") conv_states_ptrs = prior_tokens - 2 * stride_x_token # [BLOCK_N] col1 = tl.load(conv_states_ptrs, mask_w, 0.0, cache_modifier=".ca") conv_states_ptrs = prior_tokens - 3 * stride_x_token # [BLOCK_N] col0 = tl.load(conv_states_ptrs, mask_w, 0.0, cache_modifier=".ca") # Store intermediate states aligned with stride_block_m # The additional states are cached starting from the last stride_block_m. # For example: # If n_block_to_fill = 0, then only the state at the sequence end is cached and the process below is not involved. # If n_block_to_fill > 0, then the states at the sequence end and at the n_block_to_fill-last # stride_block_m are cached. # For example chunk_offset = n_block_to_fill stores the state at last_full_block if (chunk_offset - 1) < n_block_to_fill: # Store the states at the chunk boundaries from the start of the sequence idx_tokens_last = ( last_full_block_token_index - (n_block_to_fill - chunk_offset) * B_size - state_len ) + tl.arange(0, NP2_STATELEN) # [BLOCK_M] x_ptrs = ( x_ptr + (idx_tokens_last * stride_x_token)[:, None] + (idx_feats * stride_x_dim)[None, :] ) # [BLOCK_M,BLOCK_N,] mask_x = (idx_tokens_last >= 0)[:, None] & (idx_feats < dim)[ None, : ] # token-index # token-index # feature-index loaded_x = tl.load(x_ptrs, mask_x, 0.0) idx_tokens_conv = tl.arange(0, NP2_STATELEN) # [BLOCK_M] # cache_idx conv_states_output_coord = tl.load( conv_state_indices_ptr + idx_seq * stride_cache_indices + current_first_index + (chunk_offset - 1) ).to(tl.int64) conv_states_ptrs_target = ( conv_states_ptr + (conv_states_output_coord * stride_conv_state_seq) # Offset from seq + (idx_feats * stride_conv_state_dim) )[None, :] + ( # [BLOCK_N,] idx_tokens_conv * stride_conv_state_tok )[:, None] mask = (idx_tokens_conv < state_len)[:, None] & (idx_feats < dim)[None, :] tl.debug_barrier() # NOTE: use this due to bug in Triton compiler tl.store(conv_states_ptrs_target, loaded_x, mask) if HAS_BIAS: bias = bias_ptr + idx_feats mask_bias = idx_feats < dim acc_preload = tl.load(bias, mask=mask_bias, other=0.0).to( tl.float32 ) # [BLOCK_N] else: acc_preload = tl.zeros((BLOCK_N,), dtype=tl.float32) x_base_1d = x_base + token_offset * stride_x_token # starting of chunk # PRE-LOAD WEIGHTS mask_w = idx_feats < dim if KERNEL_WIDTH >= 2: w_ptrs = w_base + (0 * stride_w_width) # [BLOCK_N] tensor w_col0 = tl.load(w_ptrs, mask_w, other=0.0) w_ptrs = w_base + (1 * stride_w_width) # [BLOCK_N] tensor w_col1 = tl.load(w_ptrs, mask_w, other=0.0) if KERNEL_WIDTH >= 3: w_ptrs = w_base + (2 * stride_w_width) # [BLOCK_N] tensor w_col2 = tl.load(w_ptrs, mask_w, other=0.0) if KERNEL_WIDTH >= 4: w_ptrs = w_base + (3 * stride_w_width) # [BLOCK_N] tensor w_col3 = tl.load(w_ptrs, mask_w, other=0.0) mask_x_1d = idx_feats < dim for idx_token in range(segment_len): acc = acc_preload matrix_w = w_col0 matrix_x = col0 for j in tl.static_range(KERNEL_WIDTH): if KERNEL_WIDTH == 2: if j == 1: # KERNEL_WIDTH-1: matrix_w = w_col1 x_ptrs_1d = x_base_1d + idx_token * stride_x_token # [BLOCK_N] matrix_x = tl.load(x_ptrs_1d, mask=mask_x_1d) elif KERNEL_WIDTH == 3: if j == 1: matrix_w = w_col1 matrix_x = col1 elif j == 2: matrix_w = w_col2 x_ptrs_1d = x_base_1d + idx_token * stride_x_token # [BLOCK_N] matrix_x = tl.load(x_ptrs_1d, mask=mask_x_1d) elif KERNEL_WIDTH == 4: if j == 1: matrix_w = w_col1 matrix_x = col1 elif j == 2: matrix_w = w_col2 matrix_x = col2 elif j == 3: matrix_w = w_col3 x_ptrs_1d = x_base_1d + idx_token * stride_x_token # [BLOCK_N] matrix_x = tl.load(x_ptrs_1d, mask=mask_x_1d) acc += matrix_x * matrix_w # [BLOCK_N] if KERNEL_WIDTH == 2: col0 = matrix_x elif KERNEL_WIDTH == 3: col0 = col1 col1 = matrix_x elif KERNEL_WIDTH == 4: col0 = col1 col1 = col2 col2 = matrix_x if SILU_ACTIVATION: acc = acc / (1 + tl.exp(-acc)) mask_1d = (idx_token < segment_len) & ( idx_feats < dim ) # token-index # feature-index o_ptrs = ( o_ptr + (sequence_start_index + token_offset + idx_token) * stride_o_token + (idx_feats * stride_o_dim) ) tl.store(o_ptrs, acc, mask=mask_1d) def causal_conv1d_fn( x: torch.Tensor, weight: torch.Tensor, bias: torch.Tensor | None, conv_states: torch.Tensor, query_start_loc: torch.Tensor, cache_indices: torch.Tensor | None = None, has_initial_state: torch.Tensor | None = None, activation: str | None = "silu", pad_slot_id: int = PAD_SLOT_ID, block_idx_first_scheduled_token: torch.Tensor | None = None, block_idx_last_scheduled_token: torch.Tensor | None = None, initial_state_idx: torch.Tensor | None = None, num_computed_tokens: torch.Tensor | None = None, block_size_to_align=0, metadata=None, validate_data=False, ): """support varlen + continuous batching when x is 2D tensor x: (dim,cu_seq_len) cu_seq_len = total tokens of all seqs in that batch sequences are concatenated from left to right for varlen weight: (dim, width) conv_states: (...,dim,width - 1) itype updated inplace if cache_indices are not provided [it use `cache_indices` to get the index to the cache of conv_state for that sequence conv_state[cache_indices[i]] for seq-i - to be used as initial_state when has_initial_state[i] = True and after that conv_state[cache_indices[i]] need to be shift-left and updated with values from 'x' ] query_start_loc: (batch + 1) int32 The cumulative sequence lengths of the sequences in the batch, used to index into sequence. prepended by 0. if x = [5, 1, 1, 1] <- continuous batching (batch=4) then query_start_loc = [0, 5, 6, 7, 8] <- the starting index of the next sequence; while the last value is the ending index of the last sequence [length(query_start_loc)-1 == batch] for example: query_start_loc = torch.Tensor([0,10,16,17]), x.shape=(dim,17) cache_indices: (batch) int32 indicates the corresponding state index, like so: conv_state = conv_states[cache_indices[batch_id]] has_initial_state: (batch) bool indicates whether should the kernel take the current state as initial state for the calculations [single boolean for each sequence in the batch: True or False] bias: (dim,) activation: either None or "silu" or "swish" or True pad_slot_id: int if cache_indices is passed, lets the kernel identify padded entries that will not be processed, for example: cache_indices = [pad_slot_id, 1, 20, pad_slot_id] in this case, the kernel will not process entries at indices 0 and 3 block_idx_first_scheduled_token: (batch,), dtype int32 The pointer into cache_indices, where the first cache block to be filled is located. block_idx_last_scheduled_token: (batch,), dtype int32 The pointer into cache_indices, where the last cache block to be filled is located. initial_state_idx: (batch,), dtype int32 The pointer into cache_indices, where the cache block containing the initial state is located. num_computed_tokens: (batch,), dtype int32 The number of tokens already completed for each sequence block_size_to_align: int The block size to align the cached states to out: same shape as `x` """ if isinstance(activation, bool) and activation: activation = "silu" args = None # Store original dtype to cast back at the end original_x_dtype = x.dtype x = x.to(conv_states.dtype) out = torch.empty_like(x) if metadata is not None: nums_dict = metadata.nums_dict args = nums_dict batch_ptr = metadata.batch_ptr token_chunk_offset_ptr = metadata.token_chunk_offset_ptr else: seqlens = query_start_loc.diff().to("cpu") args = seqlens MAX_NUM_PROGRAMS = 1024 batch_ptr = torch.full( (MAX_NUM_PROGRAMS,), PAD_SLOT_ID, dtype=torch.int32, device=x.device ) # tracking which seq-idx the Triton program is handling token_chunk_offset_ptr = torch.full( (MAX_NUM_PROGRAMS,), PAD_SLOT_ID, dtype=torch.int32, device=x.device ) # tracking BLOCK_M-based index in the sequence the Triton program is handling is_channel_last = (x.stride(0) == 1) & (x.stride(1) > 1) dim, cu_seqlen = x.shape _, width = weight.shape state_len = width - 1 np2_statelen = triton.next_power_of_2(state_len) padded_batch = query_start_loc.size(0) - 1 stride_x_dim = x.stride(0) stride_x_token = x.stride(1) stride_w_dim = weight.stride(0) stride_w_width = weight.stride(1) stride_istate_seq = 0 stride_istate_dim = 0 stride_istate_token = 0 num_cache_lines = 0 BLOCK_M = 8 if conv_states is not None: # extensions to support vLLM: # 1. conv_states is used to replaced initial_states # 2. conv_states serve as a cache with num cache lines can be larger than batch size # 3. mapping from sequence x[idx] to a cache line at index as specified via cache_indices[idx] # 4. computation can be skipped if cache_indices[idx] == pad_slot_id num_cache_lines = conv_states.size(0) assert ( num_cache_lines == conv_states.shape[0] and dim == conv_states.shape[1] and width - 1 <= conv_states.shape[2] ) stride_istate_seq = conv_states.stride(0) stride_istate_dim = conv_states.stride(1) stride_istate_token = conv_states.stride(2) assert stride_istate_dim == 1 if out.dim() == 2: stride_o_dim = out.stride(0) stride_o_token = out.stride(1) else: stride_o_dim = out.stride(1) stride_o_token = out.stride(2) stride_cache_indices = cache_indices.stride(0) if cache_indices is not None else 0 if validate_data: assert x.dim() == 2 assert query_start_loc is not None assert query_start_loc.dim() == 1 assert x.stride(0) == 1 or x.stride(1) == 1 if bias is not None: assert bias.dim() == 1 assert dim == bias.size(0) if cache_indices is not None: assert cache_indices.dim() == 1 assert padded_batch == cache_indices.size(0) if has_initial_state is not None: assert has_initial_state.size() == (padded_batch,) assert conv_states is not None, ( "ERROR: `has_initial_state` is used, which needs also `conv_states`" ) assert weight.stride(1) == 1 assert (dim, width) == weight.shape assert is_channel_last, "Need to run in channel-last layout" if block_size_to_align is not None and block_size_to_align > 0: assert (block_size_to_align % BLOCK_M) == 0, ( "The mamba block size needs to be divisible by the BLOCK_M" ) else: block_size_to_align = BLOCK_M if metadata is None: def num_program(META, seqlens): tot = 0 mlist = [] offsetlist = [] # type: ignore nums = -(-seqlens // META["BLOCK_M"]) tot = nums.sum().item() mlist = np.repeat(np.arange(len(nums)), nums) for idx, num in enumerate(nums): offsetlist.extend( range(num) ) # chunk-idx if a sequence is split into multiple chunks if META["batch_ptr"].nelement() < len(mlist): newlen = len(mlist) + 1 META["batch_ptr"].resize_(newlen).fill_(PAD_SLOT_ID) META["token_chunk_offset_ptr"].resize_(newlen).fill_(PAD_SLOT_ID) if META["batch_ptr"].nelement() >= len(mlist): META["batch_ptr"][0 : len(mlist)].copy_( torch.from_numpy(np.array(mlist)) ) META["token_chunk_offset_ptr"][0 : len(mlist)].copy_( torch.from_numpy(np.array(offsetlist)) ) META["batch_ptr"] = META["batch_ptr"].to(META["x_ptr"].device) META["token_chunk_offset_ptr"] = META["token_chunk_offset_ptr"].to( META["x_ptr"].device ) return tot else: def num_program(META, nums_dict): tot = nums_dict[META["BLOCK_M"]]["tot"] mlist = nums_dict[META["BLOCK_M"]]["mlist"] mlist_len = nums_dict[META["BLOCK_M"]]["mlist_len"] offsetlist = nums_dict[META["BLOCK_M"]]["offsetlist"] if nums_dict[META["BLOCK_M"]]["batch_ptr"] is not None: META["batch_ptr"] = nums_dict[META["BLOCK_M"]]["batch_ptr"] META["token_chunk_offset_ptr"] = nums_dict[META["BLOCK_M"]][ "token_chunk_offset_ptr" ] else: if META["batch_ptr"].nelement() < mlist_len: newlen = mlist_len + 1 META["batch_ptr"].resize_(newlen).fill_(PAD_SLOT_ID) META["token_chunk_offset_ptr"].resize_(newlen).fill_(PAD_SLOT_ID) if META["batch_ptr"].nelement() >= mlist_len: META["batch_ptr"][0:mlist_len].copy_(mlist) META["token_chunk_offset_ptr"][0:mlist_len].copy_(offsetlist) return tot def grid(META): return ( num_program(META, args), triton.cdiv(dim, META["BLOCK_N"]), ) if batch_ptr.device != x.device: batch_ptr = batch_ptr.to(x.device) token_chunk_offset_ptr = token_chunk_offset_ptr.to(x.device) _causal_conv1d_fwd_kernel[grid]( # Pointers to matrices x, weight, bias, conv_states, cache_indices, has_initial_state, query_start_loc, batch_ptr, token_chunk_offset_ptr, block_idx_first_scheduled_token, block_idx_last_scheduled_token, initial_state_idx, num_computed_tokens, out, # Matrix dimensions dim, cu_seqlen, num_cache_lines, # stride stride_x_dim, stride_x_token, stride_w_dim, stride_w_width, stride_istate_seq, stride_istate_dim, stride_istate_token, stride_cache_indices, stride_o_dim, stride_o_token, block_size_to_align // BLOCK_M, # others pad_slot_id, # META HAS_BIAS=bias is not None, KERNEL_WIDTH=width, SILU_ACTIVATION=activation in ["silu", "swish"], IS_APC_ENABLED=block_idx_last_scheduled_token is not None, USE_PAD_SLOT=pad_slot_id is not None, NP2_STATELEN=np2_statelen, # launch_cooperative_grid=True BLOCK_M=BLOCK_M, BLOCK_N=256, num_stages=2, ) return out.to(original_x_dtype) @triton.jit() def _causal_conv1d_update_kernel( # Pointers to matrices x_ptr, # (batch, dim, seqlen) w_ptr, # (dim, width) bias_ptr, conv_state_ptr, conv_state_indices_ptr, num_accepted_tokens_ptr, query_start_loc_ptr, # (batch + 1) block_idx_last_scheduled_token, # (batch,) initial_state_idx, # (batch,) o_ptr, # (batch, dim, seqlen) # Matrix dimensions batch: int, dim: tl.constexpr, seqlen: tl.constexpr, state_len: tl.constexpr, num_cache_lines: tl.constexpr, # added to support vLLM larger cache lines # Strides stride_x_seq: tl.constexpr, stride_x_dim: tl.constexpr, stride_x_token: tl.constexpr, stride_w_dim: tl.constexpr, stride_w_width: tl.constexpr, stride_conv_state_seq: tl.constexpr, stride_conv_state_dim: tl.constexpr, stride_conv_state_tok: tl.constexpr, stride_state_indices: tl.constexpr, stride_o_seq: tl.constexpr, stride_o_dim: tl.constexpr, stride_o_token: tl.constexpr, # others pad_slot_id: tl.constexpr, # Meta-parameters HAS_BIAS: tl.constexpr, KERNEL_WIDTH: tl.constexpr,
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/ops/ssd_state_passing.py
vllm/model_executor/layers/mamba/ops/ssd_state_passing.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) 2024, Tri Dao, Albert Gu. # Adapted from https://github.com/state-spaces/mamba/blob/v2.2.4/mamba_ssm/ops/triton/ssd_state_passing.py # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton @triton.autotune( configs=[ triton.Config({"BLOCK_SIZE": 64}), triton.Config({"BLOCK_SIZE": 128}), triton.Config({"BLOCK_SIZE": 256}), triton.Config({"BLOCK_SIZE": 512}), triton.Config({"BLOCK_SIZE": 1024}), triton.Config({"BLOCK_SIZE": 2048}), ], key=["dim"], ) @triton.jit def _state_passing_fwd_kernel( # Pointers to matrices states_ptr, out_ptr, dA_cs_ptr, initstates_ptr, seq_idx_ptr, cu_chunk_seqlens_ptr, # Matrix dimensions dim: tl.constexpr, nchunks, seqlen, chunk_size: tl.constexpr, # Strides stride_states_chunk: tl.int64, stride_states_head: tl.int64, stride_states_dim: tl.constexpr, stride_out_chunk: tl.int64, stride_out_head: tl.int64, stride_out_dim: tl.constexpr, stride_dA_cs_head: tl.int64, stride_dA_cs_chunk: tl.int64, stride_dA_cs_csize: tl.constexpr, stride_initstates_batch: tl.int64, stride_initstates_head: tl.int64, stride_initstates_dim: tl.constexpr, stride_seq_idx_chunk: tl.constexpr, # Meta-parameters HAS_INITSTATES: tl.constexpr, BLOCK_SIZE: tl.constexpr, ): pid_h = tl.program_id(axis=1) pid_m = tl.program_id(axis=0) states_ptr += pid_h * stride_states_head dA_cs_ptr += pid_h * stride_dA_cs_head + (chunk_size - 1) * stride_dA_cs_csize out_ptr += pid_h * stride_out_head offs_m = pid_m * BLOCK_SIZE + tl.arange(0, BLOCK_SIZE) states_ptrs = states_ptr + offs_m * stride_states_dim out_ptrs = out_ptr + offs_m * stride_out_dim if HAS_INITSTATES: initstates_ptrs = ( initstates_ptr + pid_h * stride_initstates_head + offs_m * stride_initstates_dim ) states = tl.load(initstates_ptrs, mask=offs_m < dim, other=0.0).to(tl.float32) else: states = tl.zeros((BLOCK_SIZE,), dtype=tl.float32) prev_seq_idx = 0 for c in range(nchunks): new_states = tl.load(states_ptrs, mask=offs_m < dim, other=0.0).to(tl.float32) dA_cs = tl.load(dA_cs_ptr).to(tl.float32) seq_idx = tl.load(seq_idx_ptr + c * stride_seq_idx_chunk) # we have started a new sequence if prev_seq_idx != seq_idx: if HAS_INITSTATES: initstates_ptrs = ( initstates_ptr + seq_idx * stride_initstates_batch + pid_h * stride_initstates_head + offs_m * stride_initstates_dim ) states = tl.load(initstates_ptrs, mask=offs_m < dim, other=0.0).to( tl.float32 ) else: states = tl.zeros((BLOCK_SIZE,), dtype=tl.float32) prev_seq_idx = seq_idx states = tl.exp(dA_cs) * states + new_states tl.store(out_ptrs, states, mask=offs_m < dim) states_ptrs += stride_states_chunk dA_cs_ptr += stride_dA_cs_chunk out_ptrs += stride_out_chunk def _state_passing_fwd( states, dA_cumsum, cu_chunk_seqlens, seq_idx, initial_states=None, out_dtype=None, ): nchunks, nheads, dim = states.shape chunk_size = dA_cumsum.shape[-1] assert dA_cumsum.shape == (nheads, nchunks, chunk_size) seqlen = seq_idx.shape[-1] out_dtype = states.dtype if out_dtype is None else out_dtype out = torch.empty((nchunks, nheads, dim), device=states.device, dtype=out_dtype) initial_states_strides = ( (initial_states.stride(0), initial_states.stride(1), initial_states.stride(2)) if initial_states is not None else (0, 0, 0) ) grid = lambda META: (triton.cdiv(dim, META["BLOCK_SIZE"]), nheads) with torch.cuda.device(states.device.index): _state_passing_fwd_kernel[grid]( states_ptr=states, out_ptr=out, dA_cs_ptr=dA_cumsum, initstates_ptr=initial_states, seq_idx_ptr=seq_idx, cu_chunk_seqlens_ptr=cu_chunk_seqlens, dim=dim, nchunks=nchunks, seqlen=seqlen if seq_idx is not None else 0, chunk_size=chunk_size if seq_idx is not None else 0, stride_states_chunk=states.stride(0), stride_states_head=states.stride(1), stride_states_dim=states.stride(2), stride_out_chunk=out.stride(0), stride_out_head=out.stride(1), stride_out_dim=out.stride(2), stride_dA_cs_head=dA_cumsum.stride(0), stride_dA_cs_chunk=dA_cumsum.stride(1), stride_dA_cs_csize=dA_cumsum.stride(2), stride_initstates_batch=initial_states_strides[0], stride_initstates_head=initial_states_strides[1], stride_initstates_dim=initial_states_strides[2], stride_seq_idx_chunk=seq_idx.stride(0), HAS_INITSTATES=initial_states is not None, ) return out
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/ops/ssd_bmm.py
vllm/model_executor/layers/mamba/ops/ssd_bmm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) 2024, Tri Dao, Albert Gu. # Adapted from https://github.com/state-spaces/mamba/blob/v2.2.4/mamba_ssm/ops/triton/ssd_bmm.py # ruff: noqa: E501,SIM102 import torch from vllm.triton_utils import tl, triton @triton.autotune( configs=[ triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 256, "BLOCK_SIZE_K": 64}, num_stages=3, num_warps=8, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 256, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 128, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 128, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 32, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 32, "BLOCK_SIZE_K": 32}, num_stages=5, num_warps=2, ), triton.Config( {"BLOCK_SIZE_M": 32, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=5, num_warps=2, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=2, ), ], key=["chunk_size", "K", "IS_CAUSAL"], ) @triton.jit def _bmm_chunk_fwd_kernel( # Pointers to matrices a_ptr, b_ptr, out_ptr, cu_chunk_seqlens_ptr, # Matrix dimensions seqlen, chunk_size: tl.constexpr, K: tl.constexpr, ngroups: tl.constexpr, stride_a_seqlen: tl.int64, stride_a_head: tl.int64, stride_ak: tl.constexpr, stride_b_seqlen: tl.int64, stride_b_head: tl.int64, stride_bk: tl.constexpr, stride_out_chunk: tl.int64, stride_out_head: tl.int64, stride_outm: tl.int64, stride_outn: tl.constexpr, # Meta-parameters IS_CAUSAL: tl.constexpr, dot_dtype: tl.constexpr, BLOCK_SIZE_M: tl.constexpr, BLOCK_SIZE_N: tl.constexpr, BLOCK_SIZE_K: tl.constexpr, ): pid_ch = tl.program_id(axis=1).to(tl.int64) pid_c = pid_ch // ngroups pid_h = pid_ch - pid_c * ngroups num_pid_n = tl.cdiv(chunk_size, BLOCK_SIZE_N) pid_m = tl.program_id(axis=0) // num_pid_n pid_n = tl.program_id(axis=0) % num_pid_n if IS_CAUSAL: if pid_n * BLOCK_SIZE_N >= (pid_m + 1) * BLOCK_SIZE_M: return chunk_seqlen_start = tl.load(cu_chunk_seqlens_ptr + pid_c) chunk_seqlen_end = tl.load(cu_chunk_seqlens_ptr + pid_c + 1) a_ptr += chunk_seqlen_start * stride_a_seqlen + pid_h * stride_a_head b_ptr += chunk_seqlen_start * stride_b_seqlen + pid_h * stride_b_head offs_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offs_n = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) offs_k = tl.arange(0, BLOCK_SIZE_K) a_ptrs = a_ptr + (offs_m[:, None] * stride_a_seqlen + offs_k[None, :] * stride_ak) b_ptrs = b_ptr + (offs_k[:, None] * stride_bk + offs_n[None, :] * stride_b_seqlen) chunk_size_limit = chunk_seqlen_end - chunk_seqlen_start acc = tl.zeros((BLOCK_SIZE_M, BLOCK_SIZE_N), dtype=tl.float32) # compute a * b.T for k in range(0, tl.cdiv(K, BLOCK_SIZE_K)): a = tl.load( a_ptrs, mask=(offs_m[:, None] < chunk_size_limit) & (offs_k[None, :] < K - k * BLOCK_SIZE_K), other=0.0, ).to(dot_dtype) b = tl.load( b_ptrs, mask=(offs_k[:, None] < K - k * BLOCK_SIZE_K) & (offs_n[None, :] < chunk_size_limit), other=0.0, ).to(dot_dtype) acc += tl.dot(a, b) a_ptrs += BLOCK_SIZE_K * stride_ak b_ptrs += BLOCK_SIZE_K * stride_bk offs_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offs_n = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) out = acc.to(out_ptr.dtype.element_ty) out_ptr += pid_c * stride_out_chunk + pid_h * stride_out_head out_ptrs = out_ptr + (stride_outm * offs_m[:, None] + offs_n[None, :] * stride_outn) tl.store( out_ptrs, out, mask=(offs_m[:, None] < chunk_size) & (offs_n[None, :] < chunk_size), ) def _bmm_chunk_fwd(a, b, chunk_size, cu_chunk_seqlens, causal=False, output_dtype=None): """ Argument: a: (seqlen, ngroups, k) b: (seqlen, ngroups, k) chunk_size: int cu_chunk_seq_lens: (nchunks+1,) causal: if True, then out[i, j] for i > j will be arbitrary, only out[i, j] for i <= j are guaranteed to be correct. Return: out: (nchunks, ngroups, chunk_size, chunk_size) """ seqlen, ngroups, k = a.shape assert b.shape == a.shape if a.stride(-1) != 1 and a.stride(0) != 1: a = a.contiguous() if b.stride(-1) != 1 and b.stride(0) != 1: b = b.contiguous() nchunks = len(cu_chunk_seqlens) - 1 # Allocates output. out_dtype = a.dtype if output_dtype is None else output_dtype out = torch.empty( (nchunks, ngroups, chunk_size, chunk_size), device=a.device, dtype=out_dtype ) dot_dtype = ( tl.bfloat16 if a.dtype == torch.bfloat16 or b.dtype == torch.bfloat16 else ( tl.float16 if a.dtype == torch.float16 or b.dtype == torch.float16 else tl.float32 ) ) grid = lambda META: ( triton.cdiv(chunk_size, META["BLOCK_SIZE_M"]) * triton.cdiv(chunk_size, META["BLOCK_SIZE_N"]), nchunks * ngroups, ) with torch.cuda.device(a.device.index): _bmm_chunk_fwd_kernel[grid]( a_ptr=a, b_ptr=b, out_ptr=out, cu_chunk_seqlens_ptr=cu_chunk_seqlens, seqlen=seqlen, chunk_size=chunk_size, K=k, ngroups=ngroups, stride_a_seqlen=a.stride(0), stride_a_head=a.stride(1), stride_ak=a.stride(2), stride_b_seqlen=b.stride(0), stride_b_head=b.stride(1), stride_bk=b.stride(2), stride_out_chunk=out.stride(0), stride_out_head=out.stride(1), stride_outm=out.stride(-2), stride_outn=out.stride(-1), IS_CAUSAL=causal, dot_dtype=dot_dtype, ) return out
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/ops/layernorm_gated.py
vllm/model_executor/layers/mamba/ops/layernorm_gated.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) 2024, Tri Dao. # Adapted from https://github.com/state-spaces/mamba/blob/60dadf2e0ee730ac337035d5533de10bc26e4847/mamba_ssm/ops/triton/layernorm_gated.py import torch from vllm.triton_utils import tl, triton @triton.heuristics({"HAS_BIAS": lambda args: args["B"] is not None}) @triton.heuristics({"HAS_Z": lambda args: args["Z"] is not None}) @triton.jit def _layer_norm_fwd_1pass_kernel( X, # pointer to the input Y, # pointer to the output W, # pointer to the weights B, # pointer to the biases Z, # pointer to the other branch Mean, # pointer to the mean Rstd, # pointer to the 1/std stride_x_row: tl.int64, stride_y_row: tl.int64, stride_z_row: tl.int64, M: tl.int64, # number of rows in X N: tl.int64, # number of columns in X eps, # epsilon to avoid division by zero BLOCK_N: tl.constexpr, HAS_BIAS: tl.constexpr, HAS_Z: tl.constexpr, NORM_BEFORE_GATE: tl.constexpr, IS_RMS_NORM: tl.constexpr, ): # Map the program id to the row of X and Y it should compute. row = tl.program_id(0) group = tl.program_id(1) X += row * stride_x_row + group * N Y += row * stride_y_row + group * N if HAS_Z: Z += row * stride_z_row + group * N if not IS_RMS_NORM: Mean += group * M Rstd += group * M W += group * N if HAS_BIAS: B += group * N # Compute mean and variance cols = tl.arange(0, BLOCK_N) x = tl.load(X + cols, mask=cols < N, other=0.0).to(tl.float32) if HAS_Z and not NORM_BEFORE_GATE: z = tl.load(Z + cols, mask=cols < N).to(tl.float32) x *= z * tl.sigmoid(z) if not IS_RMS_NORM: mean = tl.sum(x, axis=0) / N tl.store(Mean + row, mean) xbar = tl.where(cols < N, x - mean, 0.0) var = tl.sum(xbar * xbar, axis=0) / N else: xbar = tl.where(cols < N, x, 0.0) var = tl.sum(xbar * xbar, axis=0) / N rstd = 1 / tl.sqrt(var + eps) tl.store(Rstd + row, rstd) # Normalize and apply linear transformation mask = cols < N w = tl.load(W + cols, mask=mask).to(tl.float32) if HAS_BIAS: b = tl.load(B + cols, mask=mask).to(tl.float32) x_hat = (x - mean) * rstd if not IS_RMS_NORM else x * rstd y = x_hat * w + b if HAS_BIAS else x_hat * w if HAS_Z and NORM_BEFORE_GATE: z = tl.load(Z + cols, mask=mask).to(tl.float32) y *= z * tl.sigmoid(z) # Write output tl.store(Y + cols, y, mask=mask) def _layer_norm_fwd( x, weight, bias, eps, z=None, out=None, group_size=None, norm_before_gate=True, is_rms_norm=False, ): M, N = x.shape if group_size is None: group_size = N assert N % group_size == 0 ngroups = N // group_size assert x.stride(-1) == 1 if z is not None: assert z.stride(-1) == 1 assert z.shape == (M, N) assert weight.shape == (N,) assert weight.stride(-1) == 1 if bias is not None: assert bias.stride(-1) == 1 assert bias.shape == (N,) # allocate output if out is not None: assert out.shape == x.shape else: out = torch.empty_like(x) assert out.stride(-1) == 1 mean = ( torch.empty((ngroups * M,), dtype=torch.float32, device=x.device) if not is_rms_norm else None ) rstd = torch.empty((ngroups * M,), dtype=torch.float32, device=x.device) # Less than 64KB per feature: enqueue fused kernel MAX_FUSED_SIZE = 65536 // x.element_size() BLOCK_N = min(MAX_FUSED_SIZE, triton.next_power_of_2(group_size)) if group_size > BLOCK_N: raise RuntimeError("This layer norm doesn't support feature dim >= 64KB.") # heuristics for number of warps num_warps = min(max(BLOCK_N // 256, 1), 8) grid = (M, ngroups) with torch.cuda.device(x.device.index): _layer_norm_fwd_1pass_kernel[grid]( x, out, weight, bias, z, mean, rstd, x.stride(0), out.stride(0), z.stride(0) if z is not None else 0, M, group_size, eps, BLOCK_N=BLOCK_N, NORM_BEFORE_GATE=norm_before_gate, IS_RMS_NORM=is_rms_norm, num_warps=num_warps, ) return out, mean, rstd def rms_norm_gated( x, weight, bias, z=None, eps=1e-6, group_size=None, norm_before_gate=True ): x_shape_og = x.shape # reshape input data into 2D tensor x = x.reshape(-1, x.shape[-1]) if x.stride(-1) != 1: x = x.contiguous() if z is not None: assert z.shape == x_shape_og z = z.reshape(-1, z.shape[-1]) if z.stride(-1) != 1: z = z.contiguous() weight = weight.contiguous() if bias is not None: bias = bias.contiguous() y, _, _ = _layer_norm_fwd( x, weight, bias, eps, z=z, group_size=group_size, norm_before_gate=norm_before_gate, is_rms_norm=True, ) return y.reshape(x_shape_og)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/ops/__init__.py
vllm/model_executor/layers/mamba/ops/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/ops/ssd_chunk_state.py
vllm/model_executor/layers/mamba/ops/ssd_chunk_state.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) 2024, Tri Dao, Albert Gu. # Adapted from https://github.com/state-spaces/mamba/blob/v2.2.4/mamba_ssm/ops/triton/ssd_chunk_state.py # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .mamba_ssm import softplus @triton.autotune( configs=[ triton.Config({"BLOCK_SIZE_H": 2}), triton.Config({"BLOCK_SIZE_H": 4}), triton.Config({"BLOCK_SIZE_H": 8}), triton.Config({"BLOCK_SIZE_H": 16}), triton.Config({"BLOCK_SIZE_H": 32}), triton.Config({"BLOCK_SIZE_H": 64}), ], key=["chunk_size", "nheads"], ) @triton.jit def _chunk_cumsum_fwd_kernel( # Pointers to matrices dt_ptr, A_ptr, dt_bias_ptr, dt_out_ptr, dA_cumsum_ptr, cu_chunk_seqlens_ptr, # Matrix dimension seqlen, nheads: tl.constexpr, chunk_size: tl.constexpr, dt_min: tl.constexpr, dt_max: tl.constexpr, # Strides stride_dt_seqlen: tl.int64, stride_dt_head: tl.constexpr, stride_A_head: tl.constexpr, stride_dt_bias_head: tl.constexpr, stride_dt_out_head: tl.int64, stride_dt_out_chunk: tl.int64, stride_dt_out_csize: tl.constexpr, stride_dA_cs_head: tl.int64, stride_dA_cs_chunk: tl.int64, stride_dA_cs_csize: tl.constexpr, # Meta-parameters DT_SOFTPLUS: tl.constexpr, HAS_DT_BIAS: tl.constexpr, BLOCK_SIZE_H: tl.constexpr, BLOCK_SIZE_CHUNK: tl.constexpr, ): # if dt is long, may cause problems, so use 64 bit # https://github.com/triton-lang/triton/issues/1058 pid_c = tl.program_id(axis=0).to(tl.int64) pid_h = tl.program_id(axis=1) chunk_seqlen_start = tl.load(cu_chunk_seqlens_ptr + pid_c) chunk_seqlen_end = tl.load(cu_chunk_seqlens_ptr + pid_c + 1) dt_ptr += chunk_seqlen_start * stride_dt_seqlen dt_out_ptr += pid_c * stride_dt_out_chunk dA_cumsum_ptr += pid_c * stride_dA_cs_chunk offs_h = pid_h * BLOCK_SIZE_H + tl.arange(0, BLOCK_SIZE_H) offs_c = tl.arange(0, BLOCK_SIZE_CHUNK) dt_ptrs = dt_ptr + ( offs_h[:, None] * stride_dt_head + offs_c[None, :] * stride_dt_seqlen ) A_ptrs = A_ptr + offs_h * stride_A_head dt_out_ptrs = dt_out_ptr + ( offs_h[:, None] * stride_dt_out_head + offs_c[None, :] * stride_dt_out_csize ) dA_cs_ptrs = dA_cumsum_ptr + ( offs_h[:, None] * stride_dA_cs_head + offs_c[None, :] * stride_dA_cs_csize ) chunk_size_limit = chunk_seqlen_end - chunk_seqlen_start dt = tl.load( dt_ptrs, mask=(offs_h[:, None] < nheads) & (offs_c[None, :] < chunk_size_limit), other=0.0, ).to(tl.float32) if HAS_DT_BIAS: dt_bias = tl.load( dt_bias_ptr + offs_h * stride_dt_bias_head, mask=offs_h < nheads, other=0.0 ).to(tl.float32) dt += dt_bias[:, None] if DT_SOFTPLUS: dt = tl.where(dt <= 20.0, softplus(dt), dt) dt = tl.clamp(dt, dt_min, dt_max) dt = tl.where( (offs_h[:, None] < nheads) & (offs_c[None, :] < chunk_size_limit), dt, 0.0 ) tl.store( dt_out_ptrs, dt, mask=(offs_h[:, None] < nheads) & (offs_c[None, :] < chunk_size), ) A = tl.load(A_ptrs, mask=offs_h < nheads, other=0.0).to(tl.float32) dA = dt * A[:, None] dA_cs = tl.cumsum(dA, axis=1) tl.store( dA_cs_ptrs, dA_cs, mask=(offs_h[:, None] < nheads) & (offs_c[None, :] < chunk_size), ) @triton.autotune( configs=[ triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 256, "BLOCK_SIZE_K": 64}, num_stages=3, num_warps=8, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 256, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 128, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 128, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 32, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 32, "BLOCK_SIZE_K": 32}, num_stages=5, num_warps=2, ), triton.Config( {"BLOCK_SIZE_M": 32, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=5, num_warps=2, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=2, ), ], key=["hdim", "dstate", "chunk_size"], ) @triton.jit def _chunk_state_fwd_kernel( # Pointers to matrices x_ptr, b_ptr, states_ptr, dt_ptr, dA_cumsum_ptr, cu_chunk_seqlens_ptr, # Matrix dimensions hdim: tl.constexpr, dstate: tl.constexpr, chunk_size: tl.constexpr, seqlen, nheads_ngroups_ratio: tl.constexpr, # Strides stride_x_seqlen: tl.int64, stride_x_head: tl.int64, stride_x_hdim: tl.constexpr, stride_b_seqlen: tl.int64, stride_b_head: tl.int64, stride_b_dstate: tl.constexpr, stride_states_chunk: tl.int64, stride_states_head: tl.int64, stride_states_hdim: tl.int64, stride_states_dstate: tl.constexpr, stride_dt_head: tl.int64, stride_dt_chunk: tl.int64, stride_dt_csize: tl.constexpr, stride_dA_cs_head: tl.int64, stride_dA_cs_chunk: tl.int64, stride_dA_cs_csize: tl.constexpr, # Meta-parameters BLOCK_SIZE_M: tl.constexpr, BLOCK_SIZE_N: tl.constexpr, BLOCK_SIZE_K: tl.constexpr, ): pid_c = tl.program_id(axis=1).to(tl.int64) pid_h = tl.program_id(axis=2) num_pid_n = tl.cdiv(dstate, BLOCK_SIZE_N) pid_m = tl.program_id(axis=0) // num_pid_n pid_n = tl.program_id(axis=0) % num_pid_n chunk_seqlen_start = tl.load(cu_chunk_seqlens_ptr + pid_c) chunk_seqlen_end = tl.load(cu_chunk_seqlens_ptr + pid_c + 1) b_ptr += ( chunk_seqlen_start * stride_b_seqlen + (pid_h // nheads_ngroups_ratio) * stride_b_head ) x_ptr += chunk_seqlen_start * stride_x_seqlen + pid_h * stride_x_head dt_ptr += pid_c * stride_dt_chunk + pid_h * stride_dt_head dA_cumsum_ptr += pid_c * stride_dA_cs_chunk + pid_h * stride_dA_cs_head offs_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offs_n = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) offs_k = tl.arange(0, BLOCK_SIZE_K) x_ptrs = x_ptr + ( offs_m[:, None] * stride_x_hdim + offs_k[None, :] * stride_x_seqlen ) b_ptrs = b_ptr + ( offs_n[None, :] * stride_b_dstate + offs_k[:, None] * stride_b_seqlen ) dt_ptrs = dt_ptr + offs_k * stride_dt_csize dA_cs_last = tl.load(dA_cumsum_ptr + (chunk_size - 1) * stride_dA_cs_csize).to( tl.float32 ) dA_cumsum_ptrs = dA_cumsum_ptr + offs_k * stride_dA_cs_csize chunk_size_limit = chunk_seqlen_end - chunk_seqlen_start acc = tl.zeros((BLOCK_SIZE_M, BLOCK_SIZE_N), dtype=tl.float32) for k in range(0, chunk_size_limit, BLOCK_SIZE_K): x = tl.load( x_ptrs, mask=(offs_m[:, None] < hdim) & (offs_k[None, :] < chunk_size_limit - k), other=0.0, ) b = tl.load( b_ptrs, mask=(offs_k[:, None] < chunk_size_limit - k) & (offs_n[None, :] < dstate), other=0.0, ).to(tl.float32) dA_cs_k = tl.load( dA_cumsum_ptrs, mask=offs_k < chunk_size_limit - k, other=0.0 ).to(tl.float32) dt_k = tl.load(dt_ptrs, mask=offs_k < chunk_size_limit - k, other=0.0).to( tl.float32 ) scale = tl.exp(dA_cs_last - dA_cs_k) * dt_k b *= scale[:, None] b = b.to(x_ptr.dtype.element_ty) acc += tl.dot(x, b) x_ptrs += BLOCK_SIZE_K * stride_x_seqlen b_ptrs += BLOCK_SIZE_K * stride_b_seqlen dt_ptrs += BLOCK_SIZE_K * stride_dt_csize dA_cumsum_ptrs += BLOCK_SIZE_K * stride_dA_cs_csize states = acc.to(states_ptr.dtype.element_ty) states_ptr += pid_c * stride_states_chunk + pid_h * stride_states_head offs_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offs_n = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) states_ptrs = states_ptr + ( offs_m[:, None] * stride_states_hdim + offs_n[None, :] * stride_states_dstate ) c_mask = (offs_m[:, None] < hdim) & (offs_n[None, :] < dstate) tl.store(states_ptrs, states, mask=c_mask) @triton.autotune( configs=[ triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 256, "BLOCK_SIZE_K": 64}, num_stages=3, num_warps=8, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 256, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 128, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 128, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 128, "BLOCK_SIZE_N": 32, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=4, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 32, "BLOCK_SIZE_K": 32}, num_stages=5, num_warps=2, ), triton.Config( {"BLOCK_SIZE_M": 32, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=5, num_warps=2, ), triton.Config( {"BLOCK_SIZE_M": 64, "BLOCK_SIZE_N": 64, "BLOCK_SIZE_K": 32}, num_stages=4, num_warps=2, ), ], key=["hdim", "dstate", "chunk_size"], ) @triton.jit def _chunk_state_varlen_kernel( # Pointers to matrices x_ptr, b_ptr, dt_ptr, dA_cumsum_ptr, chunk_states_ptr, cu_seqlens_ptr, states_ptr, initstates_ptr, # Matrix dimensions hdim: tl.constexpr, dstate: tl.constexpr, chunk_size: tl.constexpr, nheads_ngroups_ratio: tl.constexpr, # Strides stride_x_seqlen: tl.int64, stride_x_head: tl.int64, stride_x_hdim: tl.constexpr, stride_b_seqlen: tl.int64, stride_b_head: tl.int64, stride_b_dstate: tl.constexpr, stride_dt_head: tl.int64, stride_dt_chunk: tl.int64, stride_dt_csize: tl.constexpr, stride_dA_cs_head: tl.int64, stride_dA_cs_chunk: tl.int64, stride_dA_cs_csize: tl.constexpr, stride_chunk_states_chunk: tl.int64, stride_chunk_states_head: tl.int64, stride_chunk_states_hdim: tl.int64, stride_chunk_states_dstate: tl.constexpr, stride_states_batch: tl.int64, stride_states_head: tl.int64, stride_states_hdim: tl.int64, stride_states_dstate: tl.constexpr, stride_init_states_batch: tl.int64, stride_init_states_head: tl.int64, stride_init_states_hdim: tl.int64, stride_init_states_dstate: tl.constexpr, # Meta-parameters BLOCK_SIZE_M: tl.constexpr, BLOCK_SIZE_N: tl.constexpr, BLOCK_SIZE_K: tl.constexpr, HAS_INITSTATES: tl.constexpr, ): pid_b = tl.program_id(axis=1) pid_h = tl.program_id(axis=2) num_pid_n = tl.cdiv(dstate, BLOCK_SIZE_N) pid_m = tl.program_id(axis=0) // num_pid_n pid_n = tl.program_id(axis=0) % num_pid_n end_idx = tl.load(cu_seqlens_ptr + pid_b + 1) pid_c = (end_idx - 1) // chunk_size b_ptr += ( pid_c * chunk_size * stride_b_seqlen + (pid_h // nheads_ngroups_ratio) * stride_b_head ) x_ptr += pid_c * chunk_size * stride_x_seqlen + pid_h * stride_x_head dt_ptr += pid_c * stride_dt_chunk + pid_h * stride_dt_head dA_cumsum_ptr += pid_c * stride_dA_cs_chunk + pid_h * stride_dA_cs_head chunk_states_ptr += ( pid_c * stride_chunk_states_chunk + pid_h * stride_chunk_states_head ) if HAS_INITSTATES: # if there are init states provided, we differentiate between states (which # are boundary conditions at a chunk boundary) and initstates (which are boundary # conditions when a new example in a cont batch starts) initstates_ptr += pid_h * stride_init_states_head offs_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offs_n = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) offs_k = tl.arange(0, BLOCK_SIZE_K) x_ptrs = x_ptr + ( offs_m[:, None] * stride_x_hdim + offs_k[None, :] * stride_x_seqlen ) b_ptrs = b_ptr + ( offs_n[None, :] * stride_b_dstate + offs_k[:, None] * stride_b_seqlen ) dt_ptrs = dt_ptr + offs_k * stride_dt_csize dA_cs_last = tl.load( dA_cumsum_ptr + (end_idx - pid_c * chunk_size - 1) * stride_dA_cs_csize ).to(tl.float32) dA_cumsum_ptrs = dA_cumsum_ptr + offs_k * stride_dA_cs_csize chunk_size_limit = end_idx - pid_c * chunk_size start_idx = tl.load(cu_seqlens_ptr + pid_b) start_idx_cur = tl.maximum(start_idx - pid_c * chunk_size, 0) acc = tl.zeros((BLOCK_SIZE_M, BLOCK_SIZE_N), dtype=tl.float32) for k in range(0, chunk_size_limit, BLOCK_SIZE_K): x = tl.load( x_ptrs, mask=(offs_m[:, None] < hdim) & (offs_k[None, :] < chunk_size_limit - k) & (offs_k[None, :] >= start_idx_cur - k), other=0.0, ) b = tl.load( b_ptrs, mask=(offs_k[:, None] < chunk_size_limit - k) & (offs_n[None, :] < dstate) & (offs_k[:, None] >= start_idx_cur - k), other=0.0, ).to(tl.float32) dA_cs_k = tl.load( dA_cumsum_ptrs, mask=offs_k < chunk_size_limit - k, other=0.0 ).to(tl.float32) dt_k = tl.load(dt_ptrs, mask=offs_k < chunk_size_limit - k, other=0.0).to( tl.float32 ) scale = tl.where( (offs_k >= start_idx_cur - k) & (offs_k < chunk_size_limit - k), tl.exp(dA_cs_last - dA_cs_k) * dt_k, 0.0, ) b *= scale[:, None] b = b.to(x_ptr.dtype.element_ty) acc += tl.dot(x, b) x_ptrs += BLOCK_SIZE_K * stride_x_seqlen b_ptrs += BLOCK_SIZE_K * stride_b_seqlen dt_ptrs += BLOCK_SIZE_K * stride_dt_csize dA_cumsum_ptrs += BLOCK_SIZE_K * stride_dA_cs_csize # If the sequence starts after the last chunk idx, we don't need to add the contribution from the last chunk # If HAS_INITSTATES==True need to consider two possibilities # - if start_idx < pid_c * chunk_size, then we need to take the past_states_ptrs # - if state_idx >= pid * chunk_size, then we need to insert initstates if ( (start_idx < pid_c * chunk_size) # first chunk or (HAS_INITSTATES) ): dA_cs_boundary = 0.0 # default if not HAS_INITSTATES: past_states_ptrs = chunk_states_ptr + ( offs_m[:, None] * stride_chunk_states_hdim + offs_n[None, :] * stride_chunk_states_dstate ) else: # - this seems repetitive, buts its to help the compiler if start_idx < pid_c * chunk_size: past_states_ptrs = chunk_states_ptr + ( offs_m[:, None] * stride_chunk_states_hdim + offs_n[None, :] * stride_chunk_states_dstate ) else: past_states_ptrs = initstates_ptr + ( pid_b * stride_init_states_batch + offs_m[:, None] * stride_init_states_hdim + offs_n[None, :] * stride_init_states_dstate ) # need to adjust the boundary if start_idx > pid_c * chunk_size: dA_cs_boundary = tl.load( dA_cumsum_ptr + (start_idx - pid_c * chunk_size - 1) * stride_dA_cs_csize ).to(tl.float32) past_states = tl.load( past_states_ptrs, mask=(offs_m[:, None] < hdim) & (offs_n[None, :] < dstate), other=0.0, ).to(tl.float32) scale = tl.exp(dA_cs_last - dA_cs_boundary) acc += past_states * scale states = acc.to(states_ptr.dtype.element_ty) states_ptr += pid_b * stride_states_batch + pid_h * stride_states_head offs_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offs_n = pid_n * BLOCK_SIZE_N + tl.arange(0, BLOCK_SIZE_N) states_ptrs = states_ptr + ( offs_m[:, None] * stride_states_hdim + offs_n[None, :] * stride_states_dstate ) c_mask = (offs_m[:, None] < hdim) & (offs_n[None, :] < dstate) tl.store(states_ptrs, states, mask=c_mask) def _chunk_cumsum_fwd( dt, A, chunk_size, cu_chunk_seqlens, dt_bias=None, dt_softplus=False, dt_limit=(0.0, float("inf")), ): seqlen, nheads = dt.shape assert A.shape == (nheads,) if dt_bias is not None: assert dt_bias.shape == (nheads,) nchunks = cu_chunk_seqlens.shape[0] - 1 dt_out = torch.empty( nheads, nchunks, chunk_size, device=dt.device, dtype=torch.float32 ) dA_cumsum = torch.empty( nheads, nchunks, chunk_size, device=dt.device, dtype=torch.float32 ) grid_chunk_cs = lambda META: (nchunks, triton.cdiv(nheads, META["BLOCK_SIZE_H"])) with torch.cuda.device(dt.device.index): _chunk_cumsum_fwd_kernel[grid_chunk_cs]( dt_ptr=dt, A_ptr=A, dt_bias_ptr=dt_bias, dt_out_ptr=dt_out, dA_cumsum_ptr=dA_cumsum, cu_chunk_seqlens_ptr=cu_chunk_seqlens, seqlen=seqlen, nheads=nheads, chunk_size=chunk_size, dt_min=dt_limit[0], dt_max=dt_limit[1], stride_dt_seqlen=dt.stride(0), stride_dt_head=dt.stride(1), stride_A_head=A.stride(0), stride_dt_bias_head=dt_bias.stride(0) if dt_bias is not None else 0, stride_dt_out_head=dt_out.stride(0), stride_dt_out_chunk=dt_out.stride(1), stride_dt_out_csize=dt_out.stride(2), stride_dA_cs_head=dA_cumsum.stride(0), stride_dA_cs_chunk=dA_cumsum.stride(1), stride_dA_cs_csize=dA_cumsum.stride(2), DT_SOFTPLUS=dt_softplus, HAS_DT_BIAS=dt_bias is not None, BLOCK_SIZE_CHUNK=triton.next_power_of_2(chunk_size), ) return dA_cumsum, dt_out def _chunk_state_fwd( B, x, dt, dA_cumsum, cu_chunk_seqlens, states=None, states_in_fp32=True ): seqlen, nheads, headdim = x.shape _, nchunks, chunk_size = dt.shape _, ngroups, dstate = B.shape assert nheads % ngroups == 0 assert B.shape == (seqlen, ngroups, dstate) assert dt.shape == (nheads, nchunks, chunk_size) assert dA_cumsum.shape == dt.shape if states is not None: assert states.shape == (nchunks, nheads, headdim, dstate) else: states_dtype = torch.float32 if states_in_fp32 else B.dtype states = torch.empty( (nchunks, nheads, headdim, dstate), device=x.device, dtype=states_dtype ) grid = lambda META: ( triton.cdiv(headdim, META["BLOCK_SIZE_M"]) * triton.cdiv(dstate, META["BLOCK_SIZE_N"]), nchunks, nheads, ) with torch.cuda.device(x.device.index): _chunk_state_fwd_kernel[grid]( x_ptr=x, b_ptr=B, states_ptr=states, dt_ptr=dt, dA_cumsum_ptr=dA_cumsum, cu_chunk_seqlens_ptr=cu_chunk_seqlens, hdim=headdim, dstate=dstate, chunk_size=chunk_size, seqlen=seqlen, nheads_ngroups_ratio=nheads // ngroups, stride_x_seqlen=x.stride(0), stride_x_head=x.stride(1), stride_x_hdim=x.stride(2), stride_b_seqlen=B.stride(0), stride_b_head=B.stride(1), stride_b_dstate=B.stride(2), stride_states_chunk=states.stride(0), stride_states_head=states.stride(1), stride_states_hdim=states.stride(2), stride_states_dstate=states.stride(3), stride_dt_head=dt.stride(0), stride_dt_chunk=dt.stride(1), stride_dt_csize=dt.stride(2), stride_dA_cs_head=dA_cumsum.stride(0), stride_dA_cs_chunk=dA_cumsum.stride(1), stride_dA_cs_csize=dA_cumsum.stride(2), ) return states def chunk_state_varlen( B, x, dt, dA_cumsum, cu_seqlens, chunk_states, initial_states=None ): total_seqlen, nheads, headdim = x.shape _, nchunks, chunk_size = dt.shape _, ngroups, dstate = B.shape batch = cu_seqlens.shape[0] - 1 cu_seqlens = cu_seqlens.contiguous() assert nheads % ngroups == 0 assert B.shape == (total_seqlen, ngroups, dstate) assert dt.shape == (nheads, nchunks, chunk_size) assert dA_cumsum.shape == dt.shape assert chunk_states.shape == (nchunks, nheads, headdim, dstate) if initial_states is not None: assert initial_states.shape == (batch, nheads, headdim, dstate) states = torch.empty( batch, nheads, headdim, dstate, dtype=chunk_states.dtype, device=chunk_states.device, ) initial_states_strides = ( ( initial_states.stride(0), initial_states.stride(1), initial_states.stride(2), initial_states.stride(3), ) if initial_states is not None else (0, 0, 0, 0) ) grid = lambda META: ( triton.cdiv(headdim, META["BLOCK_SIZE_M"]) * triton.cdiv(dstate, META["BLOCK_SIZE_N"]), batch, nheads, ) with torch.cuda.device(x.device.index): _chunk_state_varlen_kernel[grid]( x_ptr=x, b_ptr=B, dt_ptr=dt, dA_cumsum_ptr=dA_cumsum, chunk_states_ptr=chunk_states, cu_seqlens_ptr=cu_seqlens, states_ptr=states, initstates_ptr=initial_states, hdim=headdim, dstate=dstate, chunk_size=chunk_size, nheads_ngroups_ratio=nheads // ngroups, stride_x_seqlen=x.stride(0), stride_x_head=x.stride(1), stride_x_hdim=x.stride(2), stride_b_seqlen=B.stride(0), stride_b_head=B.stride(1), stride_b_dstate=B.stride(2), stride_dt_head=dt.stride(0), stride_dt_chunk=dt.stride(1), stride_dt_csize=dt.stride(2), stride_dA_cs_head=dA_cumsum.stride(0), stride_dA_cs_chunk=dA_cumsum.stride(1), stride_dA_cs_csize=dA_cumsum.stride(2), stride_chunk_states_chunk=chunk_states.stride(0), stride_chunk_states_head=chunk_states.stride(1), stride_chunk_states_hdim=chunk_states.stride(2), stride_chunk_states_dstate=chunk_states.stride(3), stride_states_batch=states.stride(0), stride_states_head=states.stride(1), stride_states_hdim=states.stride(2), stride_states_dstate=states.stride(3), stride_init_states_batch=initial_states_strides[0], stride_init_states_head=initial_states_strides[1], stride_init_states_hdim=initial_states_strides[2], stride_init_states_dstate=initial_states_strides[3], HAS_INITSTATES=initial_states is not None, ) return states
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/mamba/ops/mamba_ssm.py
vllm/model_executor/layers/mamba/ops/mamba_ssm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) 2024, Tri Dao, Albert Gu. # Adapted from https://github.com/state-spaces/mamba/blob/v2.2.4/mamba_ssm/ops/triton/selective_state_update.py import torch from packaging import version from vllm import _custom_ops as ops from vllm.attention.backends.utils import PAD_SLOT_ID from vllm.triton_utils import HAS_TRITON, tl, triton TRITON3 = HAS_TRITON and (version.parse(triton.__version__) >= version.parse("3.0.0")) if TRITON3: @triton.jit def softplus(dt): dt = tl.where(dt <= 20.0, tl.math.log(tl.math.exp(dt) + 1), dt) return dt else: @triton.jit def softplus(dt): dt = tl.where(dt <= 20.0, tl.math.log1p(tl.exp(dt)), dt) return dt @triton.heuristics({"HAS_DT_BIAS": lambda args: args["dt_bias_ptr"] is not None}) @triton.heuristics({"HAS_D": lambda args: args["D_ptr"] is not None}) @triton.heuristics({"HAS_Z": lambda args: args["z_ptr"] is not None}) @triton.heuristics( { "HAS_STATE_BATCH_INDICES": lambda args: args["state_batch_indices_ptr"] is not None } ) @triton.heuristics( {"IS_SPEC_DECODING": lambda args: args["num_accepted_tokens_ptr"] is not None} ) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens_ptr"] is not None}) @triton.heuristics( {"BLOCK_SIZE_DSTATE": lambda args: triton.next_power_of_2(args["dstate"])} ) @triton.jit(do_not_specialize=["N"]) def _selective_scan_update_kernel( # Pointers to matrices state_ptr, x_ptr, dt_ptr, dt_bias_ptr, A_ptr, B_ptr, C_ptr, D_ptr, z_ptr, out_ptr, state_batch_indices_ptr, dst_state_batch_indices_ptr, pad_slot_id, num_accepted_tokens_ptr, cu_seqlens_ptr, # Matrix dimensions N, nheads, dim, dstate, nheads_ngroups_ratio, # Strides stride_state_batch, stride_state_head, stride_state_dim, stride_state_dstate, stride_x_batch, stride_x_head, stride_x_dim, stride_dt_batch, stride_dt_head, stride_dt_dim, stride_dt_bias_head, stride_dt_bias_dim, stride_A_head, stride_A_dim, stride_A_dstate, stride_B_batch, stride_B_group, stride_B_dstate, stride_C_batch, stride_C_group, stride_C_dstate, stride_D_head, stride_D_dim, stride_z_batch, stride_z_head, stride_z_dim, stride_out_batch, stride_out_head, stride_out_dim, stride_state_indices_batch, stride_state_indices_T, stride_dst_state_indices_batch, stride_dst_state_indices_T, # Meta-parameters DT_SOFTPLUS: tl.constexpr, TIE_HDIM: tl.constexpr, BLOCK_SIZE_M: tl.constexpr, HAS_DT_BIAS: tl.constexpr, HAS_D: tl.constexpr, HAS_Z: tl.constexpr, HAS_STATE_BATCH_INDICES: tl.constexpr, IS_SPEC_DECODING: tl.constexpr, IS_VARLEN: tl.constexpr, BLOCK_SIZE_DSTATE: tl.constexpr, ): pid_m = tl.program_id(axis=0) pid_b = tl.program_id(axis=1) pid_h = tl.program_id(axis=2) if IS_VARLEN: bos = tl.load(cu_seqlens_ptr + pid_b).to(tl.int64) eos = tl.load(cu_seqlens_ptr + pid_b + 1).to(tl.int64) seq_len = eos - bos if seq_len == 0: return else: bos = pid_b seq_len = 1 state_ptr_base = state_ptr # If HAS_STATE_BATCH_INDICES is true, then the ssm state's batch coordinate # is taken from the state_batch_indices_ptr Otherwise, the state coordinate # is the same as the batch id. if HAS_STATE_BATCH_INDICES: if IS_SPEC_DECODING: num_accepted = tl.load(num_accepted_tokens_ptr + pid_b).to(tl.int64) init_token_idx = tl.maximum(num_accepted - 1, 0) else: init_token_idx = 0 dst_state_batch_indices_ptr += pid_b * stride_dst_state_indices_batch if not IS_SPEC_DECODING: dst_state_batch_idx = tl.load( dst_state_batch_indices_ptr + init_token_idx * stride_dst_state_indices_T ).to(tl.int64) dst_state_ptr = state_ptr + ( dst_state_batch_idx * stride_state_batch + pid_h * stride_state_head ) state_batch_indices_ptr += ( pid_b * stride_state_indices_batch + init_token_idx * stride_state_indices_T ) state_batch_idx = tl.load(state_batch_indices_ptr).to(tl.int64) state_ptr += state_batch_idx * stride_state_batch + pid_h * stride_state_head else: dst_state_ptr = ( state_ptr + pid_b * stride_state_batch + pid_h * stride_state_head ) state_ptr += pid_b * stride_state_batch + pid_h * stride_state_head x_ptr += bos * stride_x_batch + pid_h * stride_x_head dt_ptr += bos * stride_dt_batch + pid_h * stride_dt_head if HAS_DT_BIAS: dt_bias_ptr += pid_h * stride_dt_bias_head A_ptr += pid_h * stride_A_head B_ptr += bos * stride_B_batch + (pid_h // nheads_ngroups_ratio) * stride_B_group C_ptr += bos * stride_C_batch + (pid_h // nheads_ngroups_ratio) * stride_C_group if HAS_Z: z_ptr += bos * stride_z_batch + pid_h * stride_z_head out_ptr += bos * stride_out_batch + pid_h * stride_out_head offs_m = pid_m * BLOCK_SIZE_M + tl.arange(0, BLOCK_SIZE_M) offs_n = tl.arange(0, BLOCK_SIZE_DSTATE) state_ptrs = state_ptr + ( offs_m[:, None] * stride_state_dim + offs_n[None, :] * stride_state_dstate ) if not IS_SPEC_DECODING: dst_state_ptrs = dst_state_ptr + ( offs_m[:, None] * stride_state_dim + offs_n[None, :] * stride_state_dstate ) mask = (offs_m[:, None] < dim) & (offs_n[None, :] < dstate) if HAS_STATE_BATCH_INDICES: mask &= state_batch_idx != pad_slot_id state = tl.load(state_ptrs, mask=mask, other=0.0).to(tl.float32) if HAS_DT_BIAS: dt_bias_ptrs = dt_bias_ptr + offs_m * stride_dt_bias_dim if HAS_D: D_ptr += pid_h * stride_D_head D_ptrs = D_ptr + offs_m * stride_D_dim A_ptrs = A_ptr + offs_m[:, None] * stride_A_dim + offs_n[None, :] * stride_A_dstate for i_t in range(seq_len): x_ptrs = x_ptr + offs_m * stride_x_dim dt_ptrs = dt_ptr + offs_m * stride_dt_dim B_ptrs = B_ptr + offs_n * stride_B_dstate C_ptrs = C_ptr + offs_n * stride_C_dstate if HAS_Z: z_ptrs = z_ptr + offs_m * stride_z_dim out_ptrs = out_ptr + offs_m * stride_out_dim x = tl.load(x_ptrs, mask=offs_m < dim, other=0.0).to(tl.float32) if not TIE_HDIM: dt = tl.load(dt_ptrs, mask=offs_m < dim, other=0.0).to(tl.float32) if HAS_DT_BIAS: dt += tl.load(dt_bias_ptrs, mask=offs_m < dim, other=0.0).to(tl.float32) if DT_SOFTPLUS: dt = softplus(dt) A = tl.load( A_ptrs, mask=(offs_m[:, None] < dim) & (offs_n[None, :] < dstate), other=0.0, ).to(tl.float32) dA = tl.exp(A * dt[:, None]) else: dt = tl.load(dt_ptr).to(tl.float32) if HAS_DT_BIAS: dt += tl.load(dt_bias_ptr).to(tl.float32) if DT_SOFTPLUS: dt = softplus(dt) A = tl.load(A_ptr).to(tl.float32) dA = tl.exp(A * dt) # scalar, not a matrix B = tl.load(B_ptrs, mask=offs_n < dstate, other=0.0).to(tl.float32) C = tl.load(C_ptrs, mask=offs_n < dstate, other=0.0).to(tl.float32) if HAS_D: D = tl.load(D_ptrs, mask=offs_m < dim, other=0.0).to(tl.float32) if HAS_Z: z = tl.load(z_ptrs, mask=offs_m < dim, other=0.0).to(tl.float32) dB = B[None, :] * dt[:, None] if not TIE_HDIM else B * dt state = state * dA + dB * x[:, None] if IS_SPEC_DECODING: dst_idx_ptr = dst_state_batch_indices_ptr + i_t * stride_dst_state_indices_T token_dst_idx = tl.load(dst_idx_ptr).to(tl.int64) if token_dst_idx != pad_slot_id: token_dst_ptrs = ( state_ptr_base + token_dst_idx * stride_state_batch + pid_h * stride_state_head + offs_m[:, None] * stride_state_dim + offs_n[None, :] * stride_state_dstate ) tl.store( token_dst_ptrs, state.to(token_dst_ptrs.dtype.element_ty), mask=mask ) out = tl.sum(state * C[None, :], axis=1) if HAS_D: out += x * D if HAS_Z: out *= z * tl.sigmoid(z) tl.store(out_ptrs, out, mask=offs_m < dim) x_ptr += stride_x_batch dt_ptr += stride_dt_batch B_ptr += stride_B_batch C_ptr += stride_C_batch out_ptr += stride_out_batch if HAS_Z: z_ptr += stride_z_batch if not IS_SPEC_DECODING: tl.store(dst_state_ptrs, state.to(dst_state_ptrs.dtype.element_ty), mask=mask) def selective_state_update( state, x, dt, A, B, C, D=None, z=None, dt_bias=None, dt_softplus=False, state_batch_indices=None, dst_state_batch_indices=None, pad_slot_id=PAD_SLOT_ID, out=None, num_accepted_tokens=None, cu_seqlens=None, ): """ Argument: state: (batch, dim, dstate) or (batch, nheads, dim, dstate) x: (batch, dim) or (batch, nheads, dim) dt: (batch, dim) or (batch, nheads, dim) A: (dim, dstate) or (nheads, dim, dstate) B: (batch, dstate) or (batch, ngroups, dstate) C: (batch, dstate) or (batch, ngroups, dstate) D: (dim,) or (nheads, dim) z: (batch, dim) or (batch, nheads, dim) dt_bias: (dim,) or (nheads, dim) pad_slot_id: int if cache_indices is passed, lets the kernel identify padded entries that will not be processed, for example: cache_indices = [pad_slot_id, 1, 20, pad_slot_id] in this case, the kernel will not process entries at indices 0 and 3 out: Preallocated ssm output tensor. Assume same shape as x. In-place updated. num_accepted_tokens: (batch,) number of accepted tokens from previous verification step, tells the kernel which initial state to use cu_seqlens: (batch,) length per sequence, for variable length in speculative decoding cases """ if state.dim() == 3: state = state.unsqueeze(1) if x.dim() == 2: x = x.unsqueeze(1) if dt.dim() == 2: dt = dt.unsqueeze(1) if A.dim() == 2: A = A.unsqueeze(0) if B.dim() == 2: B = B.unsqueeze(1) if C.dim() == 2: C = C.unsqueeze(1) if D is not None and D.dim() == 1: D = D.unsqueeze(0) if z is not None and z.dim() == 2: z = z.unsqueeze(1) if dt_bias is not None and dt_bias.dim() == 1: dt_bias = dt_bias.unsqueeze(0) if out.dim() == 2: out = out.unsqueeze(1) if num_accepted_tokens is not None: assert state_batch_indices is not None and state_batch_indices.dim() == 2 assert dst_state_batch_indices is None or dst_state_batch_indices.dim() == 2 if state_batch_indices is not None and state_batch_indices.dim() == 1: state_batch_indices = state_batch_indices.unsqueeze(1) if dst_state_batch_indices is not None and dst_state_batch_indices.dim() == 1: dst_state_batch_indices = dst_state_batch_indices.unsqueeze(1) _, nheads, dim, dstate = state.shape batch = x.shape[0] if cu_seqlens is not None: N = len(cu_seqlens) - 1 # Only used to verify the shape of # state_batch_indices and dst_state_batch_indices max_seqlen = ( state_batch_indices.size(-1) if state_batch_indices is not None else 1 ) else: N = batch max_seqlen = 1 assert x.shape == (batch, nheads, dim) assert dt.shape == x.shape assert A.shape == (nheads, dim, dstate) ngroups = B.shape[1] assert nheads % ngroups == 0, "nheads must be divisible by ngroups" assert B.shape == (batch, ngroups, dstate) assert C.shape == B.shape if D is not None: assert D.shape == (nheads, dim) if z is not None: assert z.shape == x.shape if dt_bias is not None: assert dt_bias.shape == (nheads, dim) if state_batch_indices is not None: assert state_batch_indices.shape[0] >= N assert state_batch_indices.shape[1] >= max_seqlen if dst_state_batch_indices is not None: assert dst_state_batch_indices.shape[0] >= N assert dst_state_batch_indices.shape[1] >= max_seqlen else: # revert to the default behavior of in-place state updates dst_state_batch_indices = state_batch_indices assert out.shape == x.shape if num_accepted_tokens is not None: assert num_accepted_tokens.shape == (N,) grid = lambda META: (triton.cdiv(dim, META["BLOCK_SIZE_M"]), N, nheads) z_strides = (z.stride(0), z.stride(1), z.stride(2)) if z is not None else (0, 0, 0) state_batch_indices_strides = ( (state_batch_indices.stride(0), state_batch_indices.stride(1)) if state_batch_indices is not None else (0, 0) ) dst_state_batch_indices_strides = ( (dst_state_batch_indices.stride(0), dst_state_batch_indices.stride(1)) if dst_state_batch_indices is not None else (0, 0) ) # We don't want autotune since it will overwrite the state # We instead tune by hand. BLOCK_SIZE_M, num_warps = ( (32, 4) if dstate <= 16 else ( (16, 4) if dstate <= 32 else ((8, 4) if dstate <= 64 else ((4, 4) if dstate <= 128 else ((4, 8)))) ) ) tie_hdim = ( A.stride(-1) == 0 and A.stride(-2) == 0 and dt.stride(-1) == 0 and dt_bias.stride(-1) == 0 ) with torch.cuda.device(x.device.index): _selective_scan_update_kernel[grid]( state, x, dt, dt_bias, A, B, C, D, z, out, state_batch_indices, dst_state_batch_indices, pad_slot_id, num_accepted_tokens, cu_seqlens, N, nheads, dim, dstate, nheads // ngroups, state.stride(0), state.stride(1), state.stride(2), state.stride(3), x.stride(0), x.stride(1), x.stride(2), dt.stride(0), dt.stride(1), dt.stride(2), *(dt_bias.stride(0), dt_bias.stride(1)) if dt_bias is not None else 0, A.stride(0), A.stride(1), A.stride(2), B.stride(0), B.stride(1), B.stride(2), C.stride(0), C.stride(1), C.stride(2), *(D.stride(0), D.stride(1)) if D is not None else 0, z_strides[0], z_strides[1], z_strides[2], out.stride(0), out.stride(1), out.stride(2), state_batch_indices_strides[0], state_batch_indices_strides[1], dst_state_batch_indices_strides[0], dst_state_batch_indices_strides[1], dt_softplus, tie_hdim, BLOCK_SIZE_M, num_warps=num_warps, ) def selective_scan_fn( u, ssm_states, delta, A, B, C, D=None, z=None, delta_bias=None, delta_softplus=False, query_start_loc=None, cache_indices=None, has_initial_state=None, pad_slot_id=PAD_SLOT_ID, block_size=1024, block_idx_first_scheduled_token=None, block_idx_last_scheduled_token=None, initial_state_idx=None, ) -> torch.Tensor: """ u: (dim, total_length) for varlen or (batch, dim, seqlen) applies changes in place. ssm_states: (batch, dim, dstate) or (batch, nheads, dim, dstate) applies changes in place. delta: (dim, total_length) for varlen or (batch, dim, seqlen) A: (dim, dstate) B: (ngroups, dstate, total_length) for varlen or (batch,ngroups,dstate,seqlen) C: (ngroups, dstate, total_length) for varlen or (batch,ngroups,dstate,seqlen) D: (dim,) z: (dim, total_length) for varlen or (batch, dim, seqlen) dt_bias: (dim,) or (dim) query_start_loc: (batch + 1) int32 The cumulative sequence lengths of the sequences in the batch, used to index into sequence. prepended with 0. for example: query_start_loc = torch.Tensor([0,10,16,17]), x.shape=(dim,17) cache_indices: (batch) int32 A tensor with each cell is a correspondent input and output ssm_state indices - Without APC: (batch,) - single state index per batch item - With APC: (batch, max_positions) - cache block indices for read/write Each non-zero value indicates a cache block to load from and/or write to. has_initial_state: (batch) bool A tensor populated with ones and zeros, indicate if the ssm_state at the corresponding index should be used as initial state. Not providing argument assumes there's no initial state pad_slot_id: int if cache_indices is passed, lets the kernel identify padding entries that will not be processed, for example: cache_indices = [pad_slot_id, 1 ,20 ,pad_slot_id] in this case, the kernel will not process entries at indices 0 and 3 block_size: int The block size to align the cached states to block_idx_first_scheduled_token: (batch,), dtype int32 The pointer into cache_indices, where the first cache block to be filled is located. block_idx_last_scheduled_token: (batch,), dtype int32 The pointer into cache_indices, where the last cache block to be filled is located. initial_state_idx: (batch,), dtype int32 The pointer into cache_indices, where the cache block containing the initial state is located. returns output: (dim, total_length) for varlen or (batch, dim, seqlen) supports inplace replacement """ if u.stride(-1) != 1: u = u.contiguous() if delta.stride(-1) != 1: delta = delta.contiguous() if D is not None: D = D.contiguous() if B.stride(-1) != 1: B = B.contiguous() if C.stride(-1) != 1: C = C.contiguous() if z is not None and z.stride(-1) != 1: z = z.contiguous() if B.dim() == 3 and query_start_loc is None: B = B.unsqueeze(1) if B.dim() == 2 and query_start_loc is not None: B = B.unsqueeze(0) if C.dim() == 3 and query_start_loc is None: C = C.unsqueeze(1) if C.dim() == 2 and query_start_loc is not None: C = C.unsqueeze(0) ops.selective_scan_fwd( u, delta, A, B, C, D, z, delta_bias, delta_softplus, query_start_loc, cache_indices, has_initial_state, ssm_states, pad_slot_id, block_size, block_idx_first_scheduled_token, block_idx_last_scheduled_token, initial_state_idx, ) if z is None: return delta # output written inplace to delta else: return z # output written inplace to z
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/__init__.py
vllm/model_executor/layers/fla/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/kda.py
vllm/model_executor/layers/fla/ops/kda.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch import torch.nn as nn from vllm.triton_utils import tl, triton from vllm.utils.math_utils import cdiv, next_power_of_2 from .chunk_delta_h import chunk_gated_delta_rule_fwd_h from .cumsum import chunk_local_cumsum from .fused_recurrent import fused_recurrent_gated_delta_rule_fwd_kernel from .index import prepare_chunk_indices from .l2norm import l2norm_fwd from .op import exp, log from .solve_tril import solve_tril from .utils import is_amd BT_LIST_AUTOTUNE = [32, 64, 128] NUM_WARPS_AUTOTUNE = [2, 4, 8, 16] if is_amd else [4, 8, 16, 32] def fused_recurrent_kda_fwd( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float, initial_state: torch.Tensor, inplace_final_state: bool = True, cu_seqlens: torch.LongTensor | None = None, ssm_state_indices: torch.Tensor | None = None, num_accepted_tokens: torch.Tensor | None = None, use_qk_l2norm_in_kernel: bool = False, ) -> tuple[torch.Tensor, torch.Tensor]: B, T, H, K, V = *k.shape, v.shape[-1] HV = v.shape[2] N = B if cu_seqlens is None else len(cu_seqlens) - 1 BK, BV = next_power_of_2(K), min(next_power_of_2(V), 8) NK, NV = cdiv(K, BK), cdiv(V, BV) assert NK == 1, "NK > 1 is not supported yet" num_stages = 3 num_warps = 1 o = torch.empty_like(k) if inplace_final_state: final_state = initial_state else: final_state = q.new_empty(T, HV, K, V, dtype=initial_state.dtype) stride_init_state_token = initial_state.stride(0) stride_final_state_token = final_state.stride(0) if ssm_state_indices is None: stride_indices_seq, stride_indices_tok = 1, 1 elif ssm_state_indices.ndim == 1: stride_indices_seq, stride_indices_tok = ssm_state_indices.stride(0), 1 else: stride_indices_seq, stride_indices_tok = ssm_state_indices.stride() grid = (NK, NV, N * HV) fused_recurrent_gated_delta_rule_fwd_kernel[grid]( q=q, k=k, v=v, g=g, beta=beta, o=o, h0=initial_state, ht=final_state, cu_seqlens=cu_seqlens, ssm_state_indices=ssm_state_indices, num_accepted_tokens=num_accepted_tokens, scale=scale, N=N, T=T, B=B, H=H, HV=HV, K=K, V=V, BK=BK, BV=BV, stride_init_state_token=stride_init_state_token, stride_final_state_token=stride_final_state_token, stride_indices_seq=stride_indices_seq, stride_indices_tok=stride_indices_tok, IS_BETA_HEADWISE=beta.ndim == v.ndim, USE_QK_L2NORM_IN_KERNEL=use_qk_l2norm_in_kernel, INPLACE_FINAL_STATE=inplace_final_state, IS_KDA=True, num_warps=num_warps, num_stages=num_stages, ) return o, final_state def fused_recurrent_kda( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor = None, scale: float = None, initial_state: torch.Tensor = None, inplace_final_state: bool = True, use_qk_l2norm_in_kernel: bool = True, cu_seqlens: torch.LongTensor | None = None, ssm_state_indices: torch.LongTensor | None = None, **kwargs, ) -> tuple[torch.Tensor, torch.Tensor]: if cu_seqlens is not None and q.shape[0] != 1: raise ValueError( f"The batch size is expected to be 1 rather than {q.shape[0]} when using `cu_seqlens`." f"Please flatten variable-length inputs before processing." ) if scale is None: scale = k.shape[-1] ** -0.5 o, final_state = fused_recurrent_kda_fwd( q=q.contiguous(), k=k.contiguous(), v=v.contiguous(), g=g.contiguous(), beta=beta.contiguous(), scale=scale, initial_state=initial_state, inplace_final_state=inplace_final_state, cu_seqlens=cu_seqlens, ssm_state_indices=ssm_state_indices, num_accepted_tokens=None, use_qk_l2norm_in_kernel=use_qk_l2norm_in_kernel, ) return o, final_state @triton.heuristics( { "STORE_RESIDUAL_OUT": lambda args: args["residual_out"] is not None, "HAS_RESIDUAL": lambda args: args["residual"] is not None, "HAS_WEIGHT": lambda args: args["w"] is not None, "HAS_BIAS": lambda args: args["b"] is not None, } ) @triton.jit def layer_norm_gated_fwd_kernel( x, # pointer to the input g, # pointer to the gate y, # pointer to the output w, # pointer to the weights b, # pointer to the biases residual, # pointer to the residual residual_out, # pointer to the residual mean, # pointer to the mean rstd, # pointer to the 1/std eps, # epsilon to avoid division by zero T, # number of rows in x D: tl.constexpr, # number of columns in x BT: tl.constexpr, BD: tl.constexpr, ACTIVATION: tl.constexpr, IS_RMS_NORM: tl.constexpr, STORE_RESIDUAL_OUT: tl.constexpr, HAS_RESIDUAL: tl.constexpr, HAS_WEIGHT: tl.constexpr, HAS_BIAS: tl.constexpr, ): i_t = tl.program_id(0) o_d = tl.arange(0, BD) m_d = o_d < D p_x = tl.make_block_ptr(x, (T, D), (D, 1), (i_t * BT, 0), (BT, BD), (1, 0)) b_x = tl.load(p_x, boundary_check=(0, 1)).to(tl.float32) if HAS_RESIDUAL: p_res = tl.make_block_ptr( residual, (T, D), (D, 1), (i_t * BT, 0), (BT, BD), (1, 0) ) b_x += tl.load(p_res, boundary_check=(0, 1)).to(tl.float32) if STORE_RESIDUAL_OUT: p_res_out = tl.make_block_ptr( residual_out, (T, D), (D, 1), (i_t * BT, 0), (BT, BD), (1, 0) ) tl.store(p_res_out, b_x.to(p_res_out.dtype.element_ty), boundary_check=(0, 1)) if not IS_RMS_NORM: b_mean = tl.sum(b_x, axis=1) / D p_mean = tl.make_block_ptr(mean, (T,), (1,), (i_t * BT,), (BT,), (0,)) tl.store(p_mean, b_mean.to(p_mean.dtype.element_ty), boundary_check=(0,)) b_xbar = tl.where(m_d[None, :], b_x - b_mean[:, None], 0.0) b_var = tl.sum(b_xbar * b_xbar, axis=1) / D else: b_xbar = tl.where(m_d[None, :], b_x, 0.0) b_var = tl.sum(b_xbar * b_xbar, axis=1) / D b_rstd = 1 / tl.sqrt(b_var + eps) p_rstd = tl.make_block_ptr(rstd, (T,), (1,), (i_t * BT,), (BT,), (0,)) tl.store(p_rstd, b_rstd.to(p_rstd.dtype.element_ty), boundary_check=(0,)) if HAS_WEIGHT: b_w = tl.load(w + o_d, mask=m_d).to(tl.float32) if HAS_BIAS: b_b = tl.load(b + o_d, mask=m_d).to(tl.float32) b_x_hat = ( (b_x - b_mean[:, None]) * b_rstd[:, None] if not IS_RMS_NORM else b_x * b_rstd[:, None] ) b_y = b_x_hat * b_w[None, :] if HAS_WEIGHT else b_x_hat if HAS_BIAS: b_y = b_y + b_b[None, :] # swish/sigmoid output gate p_g = tl.make_block_ptr(g, (T, D), (D, 1), (i_t * BT, 0), (BT, BD), (1, 0)) b_g = tl.load(p_g, boundary_check=(0, 1)).to(tl.float32) if ACTIVATION == "swish" or ACTIVATION == "silu": b_y = b_y * b_g * tl.sigmoid(b_g) elif ACTIVATION == "sigmoid": b_y = b_y * tl.sigmoid(b_g) # Write output p_y = tl.make_block_ptr(y, (T, D), (D, 1), (i_t * BT, 0), (BT, BD), (1, 0)) tl.store(p_y, b_y.to(p_y.dtype.element_ty), boundary_check=(0, 1)) @triton.heuristics( { "STORE_RESIDUAL_OUT": lambda args: args["residual_out"] is not None, "HAS_RESIDUAL": lambda args: args["residual"] is not None, "HAS_WEIGHT": lambda args: args["w"] is not None, "HAS_BIAS": lambda args: args["b"] is not None, } ) @triton.jit def layer_norm_gated_fwd_kernel1( x, # pointer to the input g, # pointer to the gate y, # pointer to the output w, # pointer to the weights b, # pointer to the biases residual, # pointer to the residual residual_out, # pointer to the residual mean, # pointer to the mean rstd, # pointer to the 1/std eps, # epsilon to avoid division by zero D: tl.constexpr, # number of columns in x BD: tl.constexpr, ACTIVATION: tl.constexpr, IS_RMS_NORM: tl.constexpr, STORE_RESIDUAL_OUT: tl.constexpr, HAS_RESIDUAL: tl.constexpr, HAS_WEIGHT: tl.constexpr, HAS_BIAS: tl.constexpr, ): i_t = tl.program_id(0) x += i_t * D y += i_t * D g += i_t * D if HAS_RESIDUAL: residual += i_t * D if STORE_RESIDUAL_OUT: residual_out += i_t * D o_d = tl.arange(0, BD) m_d = o_d < D b_x = tl.load(x + o_d, mask=m_d, other=0.0).to(tl.float32) if HAS_RESIDUAL: b_x += tl.load(residual + o_d, mask=m_d, other=0.0).to(tl.float32) if STORE_RESIDUAL_OUT: tl.store(residual_out + o_d, b_x, mask=m_d) if not IS_RMS_NORM: b_mean = tl.sum(b_x, axis=0) / D tl.store(mean + i_t, b_mean) b_xbar = tl.where(m_d, b_x - b_mean, 0.0) b_var = tl.sum(b_xbar * b_xbar, axis=0) / D else: b_xbar = tl.where(m_d, b_x, 0.0) b_var = tl.sum(b_xbar * b_xbar, axis=0) / D b_rstd = 1 / tl.sqrt(b_var + eps) tl.store(rstd + i_t, b_rstd) if HAS_WEIGHT: b_w = tl.load(w + o_d, mask=m_d).to(tl.float32) if HAS_BIAS: b_b = tl.load(b + o_d, mask=m_d).to(tl.float32) b_x_hat = (b_x - b_mean) * b_rstd if not IS_RMS_NORM else b_x * b_rstd b_y = b_x_hat * b_w if HAS_WEIGHT else b_x_hat if HAS_BIAS: b_y = b_y + b_b # swish/sigmoid output gate b_g = tl.load(g + o_d, mask=m_d, other=0.0).to(tl.float32) if ACTIVATION == "swish" or ACTIVATION == "silu": b_y = b_y * b_g * tl.sigmoid(b_g) elif ACTIVATION == "sigmoid": b_y = b_y * tl.sigmoid(b_g) # Write output tl.store(y + o_d, b_y, mask=m_d) def layer_norm_gated_fwd( x: torch.Tensor, g: torch.Tensor, weight: torch.Tensor, bias: torch.Tensor, activation: str = "swish", eps: float = 1e-5, residual: torch.Tensor = None, out_dtype: torch.dtype = None, residual_dtype: torch.dtype = None, is_rms_norm: bool = False, ): if residual is not None: residual_dtype = residual.dtype T, D = x.shape if residual is not None: assert residual.shape == (T, D) if weight is not None: assert weight.shape == (D,) if bias is not None: assert bias.shape == (D,) # allocate output y = x if out_dtype is None else torch.empty_like(x, dtype=out_dtype) if residual is not None or ( residual_dtype is not None and residual_dtype != x.dtype ): residual_out = torch.empty(T, D, device=x.device, dtype=residual_dtype) else: residual_out = None mean = ( torch.empty((T,), dtype=torch.float, device=x.device) if not is_rms_norm else None ) rstd = torch.empty((T,), dtype=torch.float, device=x.device) # Less than 64KB per feature: enqueue fused kernel MAX_FUSED_SIZE = 65536 // x.element_size() BD = min(MAX_FUSED_SIZE, next_power_of_2(D)) if D > BD: raise RuntimeError("This layer norm doesn't support feature dim >= 64KB.") # heuristics for number of warps if D <= 512: BT = 32 layer_norm_gated_fwd_kernel[(cdiv(T, BT),)]( x=x, g=g, y=y, w=weight, b=bias, residual=residual, residual_out=residual_out, mean=mean, rstd=rstd, eps=eps, T=T, D=D, BD=BD, BT=BT, ACTIVATION=activation, IS_RMS_NORM=is_rms_norm, num_warps=4, ) else: layer_norm_gated_fwd_kernel1[(T,)]( x=x, g=g, y=y, w=weight, b=bias, residual=residual, residual_out=residual_out, mean=mean, rstd=rstd, eps=eps, D=D, BD=BD, ACTIVATION=activation, IS_RMS_NORM=is_rms_norm, num_warps=4, ) # residual_out is None if residual is None and residual_dtype == input_dtype return y, mean, rstd, residual_out if residual_out is not None else x def rms_norm_gated( x: torch.Tensor, g: torch.Tensor, weight: torch.Tensor, bias: torch.Tensor, activation: str = "swish", residual: torch.Tensor | None = None, prenorm: bool = False, residual_in_fp32: bool = False, eps: float = 1e-6, ): x_shape_og = x.shape # reshape input data into 2D tensor x = x.contiguous().reshape(-1, x.shape[-1]) g = g.contiguous().reshape(-1, g.shape[-1]) if residual is not None: assert residual.shape == x_shape_og residual = residual.contiguous().reshape(-1, residual.shape[-1]) residual_dtype = ( residual.dtype if residual is not None else (torch.float if residual_in_fp32 else None) ) y, _, _, residual_out = layer_norm_gated_fwd( x=x, g=g, weight=weight, bias=bias, activation=activation, eps=eps, residual=residual, residual_dtype=residual_dtype, is_rms_norm=True, ) y = y.reshape(x_shape_og) return y if not prenorm else (y, residual_out.reshape(x_shape_og)) class FusedRMSNormGated(nn.Module): def __init__( self, hidden_size: int, elementwise_affine: bool = True, eps: float = 1e-5, activation: str = "swish", device: torch.device | None = None, dtype: torch.dtype | None = None, ) -> None: factory_kwargs = {"device": device, "dtype": dtype} super().__init__() self.hidden_size = hidden_size self.elementwise_affine = elementwise_affine self.eps = eps self.activation = activation if self.activation not in ["swish", "silu", "sigmoid"]: raise ValueError(f"Unsupported activation: {self.activation}") if elementwise_affine: self.weight = nn.Parameter(torch.empty(hidden_size, **factory_kwargs)) else: self.register_parameter("weight", None) self.register_parameter("bias", None) def forward( self, x: torch.Tensor, g: torch.Tensor, residual: torch.Tensor | None = None, prenorm: bool = False, residual_in_fp32: bool = False, ) -> torch.Tensor: return rms_norm_gated( x, g, self.weight, self.bias, self.activation, residual=residual, eps=self.eps, prenorm=prenorm, residual_in_fp32=residual_in_fp32, ) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({"BK": BK}, num_warps=num_warps, num_stages=num_stages) for BK in [32, 64] for num_warps in [1, 2, 4, 8] for num_stages in [2, 3, 4] ], key=["BC"], ) @triton.jit(do_not_specialize=["T"]) def chunk_kda_scaled_dot_kkt_fwd_kernel_intra_sub_inter( q, k, g, beta, A, Aqk, scale, cu_seqlens, chunk_indices, T, H: tl.constexpr, K: tl.constexpr, BT: tl.constexpr, BC: tl.constexpr, BK: tl.constexpr, NC: tl.constexpr, IS_VARLEN: tl.constexpr, ): i_t, i_c, i_bh = tl.program_id(0), tl.program_id(1), tl.program_id(2) i_b, i_h = i_bh // H, i_bh % H i_i, i_j = i_c // NC, i_c % NC if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T if i_t * BT + i_i * BC >= T: return if i_i <= i_j: return q += (bos * H + i_h) * K k += (bos * H + i_h) * K g += (bos * H + i_h) * K A += (bos * H + i_h) * BT Aqk += (bos * H + i_h) * BT p_b = tl.make_block_ptr( beta + bos * H + i_h, (T,), (H,), (i_t * BT + i_i * BC,), (BC,), (0,) ) b_b = tl.load(p_b, boundary_check=(0,)) b_A = tl.zeros([BC, BC], dtype=tl.float32) b_Aqk = tl.zeros([BC, BC], dtype=tl.float32) for i_k in range(tl.cdiv(K, BK)): p_q = tl.make_block_ptr( q, (T, K), (H * K, 1), (i_t * BT + i_i * BC, i_k * BK), (BC, BK), (1, 0) ) p_k = tl.make_block_ptr( k, (T, K), (H * K, 1), (i_t * BT + i_i * BC, i_k * BK), (BC, BK), (1, 0) ) p_g = tl.make_block_ptr( g, (T, K), (H * K, 1), (i_t * BT + i_i * BC, i_k * BK), (BC, BK), (1, 0) ) b_kt = tl.make_block_ptr( k, (K, T), (1, H * K), (i_k * BK, i_t * BT + i_j * BC), (BK, BC), (0, 1) ) p_gk = tl.make_block_ptr( g, (K, T), (1, H * K), (i_k * BK, i_t * BT + i_j * BC), (BK, BC), (0, 1) ) o_k = i_k * BK + tl.arange(0, BK) m_k = o_k < K # [BK,] b_gn = tl.load(g + (i_t * BT + i_i * BC) * H * K + o_k, mask=m_k, other=0) # [BC, BK] b_g = tl.load(p_g, boundary_check=(0, 1)) b_k = tl.load(p_k, boundary_check=(0, 1)) * exp(b_g - b_gn[None, :]) # [BK, BC] b_gk = tl.load(p_gk, boundary_check=(0, 1)) b_kt = tl.load(b_kt, boundary_check=(0, 1)) # [BC, BC] b_ktg = b_kt * exp(b_gn[:, None] - b_gk) b_A += tl.dot(b_k, b_ktg) b_q = tl.load(p_q, boundary_check=(0, 1)) b_qg = b_q * exp(b_g - b_gn[None, :]) * scale b_Aqk += tl.dot(b_qg, b_ktg) b_A *= b_b[:, None] p_A = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + i_i * BC, i_j * BC), (BC, BC), (1, 0) ) tl.store(p_A, b_A.to(A.dtype.element_ty), boundary_check=(0, 1)) p_Aqk = tl.make_block_ptr( Aqk, (T, BT), (H * BT, 1), (i_t * BT + i_i * BC, i_j * BC), (BC, BC), (1, 0) ) tl.store(p_Aqk, b_Aqk.to(Aqk.dtype.element_ty), boundary_check=(0, 1)) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[triton.Config({}, num_warps=num_warps) for num_warps in [1, 2, 4, 8]], key=["BK", "BT"], ) @triton.jit(do_not_specialize=["T"]) def chunk_kda_scaled_dot_kkt_fwd_kernel_intra_sub_intra( q, k, g, beta, A, Aqk, scale, cu_seqlens, chunk_indices, T, H: tl.constexpr, K: tl.constexpr, BT: tl.constexpr, BC: tl.constexpr, BK: tl.constexpr, IS_VARLEN: tl.constexpr, ): i_t, i_i, i_bh = tl.program_id(0), tl.program_id(1), tl.program_id(2) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T if i_t * BT + i_i * BC >= T: return o_i = tl.arange(0, BC) o_k = tl.arange(0, BK) m_k = o_k < K m_A = (i_t * BT + i_i * BC + o_i) < T o_A = (bos + i_t * BT + i_i * BC + o_i) * H * BT + i_h * BT + i_i * BC p_q = tl.make_block_ptr( q + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT + i_i * BC, 0), (BC, BK), (1, 0), ) p_k = tl.make_block_ptr( k + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT + i_i * BC, 0), (BC, BK), (1, 0), ) p_g = tl.make_block_ptr( g + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT + i_i * BC, 0), (BC, BK), (1, 0), ) b_q = tl.load(p_q, boundary_check=(0, 1)) b_k = tl.load(p_k, boundary_check=(0, 1)) b_g = tl.load(p_g, boundary_check=(0, 1)) p_b = beta + (bos + i_t * BT + i_i * BC + o_i) * H + i_h b_k = b_k * tl.load(p_b, mask=m_A, other=0)[:, None] p_kt = k + (bos + i_t * BT + i_i * BC) * H * K + i_h * K + o_k p_gk = g + (bos + i_t * BT + i_i * BC) * H * K + i_h * K + o_k for j in range(0, min(BC, T - i_t * BT - i_i * BC)): b_kt = tl.load(p_kt, mask=m_k, other=0).to(tl.float32) b_gk = tl.load(p_gk, mask=m_k, other=0).to(tl.float32) b_ktg = b_kt[None, :] * exp(b_g - b_gk[None, :]) b_A = tl.sum(b_k * b_ktg, 1) b_A = tl.where(o_i > j, b_A, 0.0) b_Aqk = tl.sum(b_q * b_ktg, 1) b_Aqk = tl.where(o_i >= j, b_Aqk * scale, 0.0) tl.store(A + o_A + j, b_A, mask=m_A) tl.store(Aqk + o_A + j, b_Aqk, mask=m_A) p_kt += H * K p_gk += H * K def chunk_kda_scaled_dot_kkt_fwd( q: torch.Tensor, k: torch.Tensor, gk: torch.Tensor | None = None, beta: torch.Tensor | None = None, scale: float | None = None, cu_seqlens: torch.LongTensor | None = None, chunk_size: int = 64, output_dtype: torch.dtype = torch.float32, ) -> tuple[torch.Tensor, torch.Tensor]: r""" Compute beta * K * K^T. Args: k (torch.Tensor): The key tensor of shape `[B, T, H, K]`. beta (torch.Tensor): The beta tensor of shape `[B, T, H]`. gk (torch.Tensor): The cumulative sum of the gate tensor of shape `[B, T, H, K]` applied to the key tensor. Default: `None`. cu_seqlens (torch.LongTensor): The cumulative sequence lengths of the input tensor. Default: None chunk_size (int): The chunk size. Default: 64. output_dtype (torch.dtype): The dtype of the output tensor. Default: `torch.float32` Returns: beta * K * K^T of shape `[B, T, H, BT]` where `BT` is the chunk size. """ B, T, H, K = k.shape assert K <= 256 BT = chunk_size chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) BC = min(16, BT) NC = cdiv(BT, BC) BK = max(next_power_of_2(K), 16) A = torch.zeros(B, T, H, BT, device=k.device, dtype=output_dtype) Aqk = torch.zeros(B, T, H, BT, device=k.device, dtype=output_dtype) grid = (NT, NC * NC, B * H) chunk_kda_scaled_dot_kkt_fwd_kernel_intra_sub_inter[grid]( q=q, k=k, g=gk, beta=beta, A=A, Aqk=Aqk, scale=scale, cu_seqlens=cu_seqlens, chunk_indices=chunk_indices, T=T, H=H, K=K, BT=BT, BC=BC, NC=NC, ) grid = (NT, NC, B * H) chunk_kda_scaled_dot_kkt_fwd_kernel_intra_sub_intra[grid]( q=q, k=k, g=gk, beta=beta, A=A, Aqk=Aqk, scale=scale, cu_seqlens=cu_seqlens, chunk_indices=chunk_indices, T=T, H=H, K=K, BT=BT, BC=BC, BK=BK, ) return A, Aqk @triton.heuristics( { "STORE_QG": lambda args: args["qg"] is not None, "STORE_KG": lambda args: args["kg"] is not None, "IS_VARLEN": lambda args: args["cu_seqlens"] is not None, } ) @triton.autotune( configs=[ triton.Config({}, num_warps=num_warps, num_stages=num_stages) for num_warps in [2, 4, 8] for num_stages in [2, 3, 4] ], key=["H", "K", "V", "BT", "BK", "BV", "IS_VARLEN"], ) @triton.jit(do_not_specialize=["T"]) def recompute_w_u_fwd_kernel( q, k, qg, kg, v, beta, w, u, A, gk, cu_seqlens, chunk_indices, T, H: tl.constexpr, K: tl.constexpr, V: tl.constexpr, BT: tl.constexpr, BK: tl.constexpr, BV: tl.constexpr, STORE_QG: tl.constexpr, STORE_KG: tl.constexpr, IS_VARLEN: tl.constexpr, DOT_PRECISION: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T p_b = tl.make_block_ptr(beta + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,)) b_b = tl.load(p_b, boundary_check=(0,)) p_A = tl.make_block_ptr( A + (bos * H + i_h) * BT, (T, BT), (H * BT, 1), (i_t * BT, 0), (BT, BT), (1, 0) ) b_A = tl.load(p_A, boundary_check=(0, 1)) for i_v in range(tl.cdiv(V, BV)): p_v = tl.make_block_ptr( v + (bos * H + i_h) * V, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0), ) p_u = tl.make_block_ptr( u + (bos * H + i_h) * V, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0), ) b_v = tl.load(p_v, boundary_check=(0, 1)) b_vb = (b_v * b_b[:, None]).to(b_v.dtype) b_u = tl.dot(b_A, b_vb, input_precision=DOT_PRECISION) tl.store(p_u, b_u.to(p_u.dtype.element_ty), boundary_check=(0, 1)) for i_k in range(tl.cdiv(K, BK)): p_w = tl.make_block_ptr( w + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) p_k = tl.make_block_ptr( k + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_kb = b_k * b_b[:, None] p_gk = tl.make_block_ptr( gk + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) b_gk = tl.load(p_gk, boundary_check=(0, 1)) b_kb *= exp(b_gk) if STORE_QG: p_q = tl.make_block_ptr( q + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) p_qg = tl.make_block_ptr( qg + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) b_q = tl.load(p_q, boundary_check=(0, 1)) b_qg = b_q * exp(b_gk) tl.store(p_qg, b_qg.to(p_qg.dtype.element_ty), boundary_check=(0, 1)) if STORE_KG: last_idx = min(i_t * BT + BT, T) - 1 o_k = i_k * BK + tl.arange(0, BK) m_k = o_k < K b_gn = tl.load( gk + ((bos + last_idx) * H + i_h) * K + o_k, mask=m_k, other=0.0 ) b_kg = b_k * exp(b_gn - b_gk) p_kg = tl.make_block_ptr( kg + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) tl.store(p_kg, b_kg.to(p_kg.dtype.element_ty), boundary_check=(0, 1)) b_w = tl.dot(b_A, b_kb.to(b_k.dtype)) tl.store(p_w, b_w.to(p_w.dtype.element_ty), boundary_check=(0, 1)) def recompute_w_u_fwd( k: torch.Tensor, v: torch.Tensor, beta: torch.Tensor, A: torch.Tensor, q: torch.Tensor | None = None, gk: torch.Tensor | None = None, cu_seqlens: torch.LongTensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: B, T, H, K, V = *k.shape, v.shape[-1] BT = A.shape[-1] BK = 64 BV = 64 chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) w = torch.empty_like(k) u = torch.empty_like(v) kg = torch.empty_like(k) if gk is not None else None recompute_w_u_fwd_kernel[(NT, B * H)]( q=q, k=k, qg=None, kg=kg, v=v, beta=beta, w=w, u=u, A=A, gk=gk, cu_seqlens=cu_seqlens, chunk_indices=chunk_indices, T=T, H=H, K=K, V=V, BT=BT, BK=BK, BV=BV, DOT_PRECISION="ieee", ) return w, u, None, kg @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({"BK": BK, "BV": BV}, num_warps=num_warps, num_stages=num_stages) for BK in [32, 64] for BV in [64, 128] for num_warps in [2, 4, 8] for num_stages in [2, 3, 4] ], key=["BT"], ) @triton.jit(do_not_specialize=["T"]) def chunk_gla_fwd_kernel_o( q, v, g, h, o, A, cu_seqlens, chunk_indices, scale, T, H: tl.constexpr, K: tl.constexpr, V: tl.constexpr, BT: tl.constexpr, BK: tl.constexpr, BV: tl.constexpr, IS_VARLEN: tl.constexpr, ): i_v, i_t, i_bh = tl.program_id(0), tl.program_id(1), tl.program_id(2) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_tg = i_t i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos NT = tl.cdiv(T, BT) else: NT = tl.cdiv(T, BT) i_tg = i_b * NT + i_t bos, eos = i_b * T, i_b * T + T m_s = tl.arange(0, BT)[:, None] >= tl.arange(0, BT)[None, :] b_o = tl.zeros([BT, BV], dtype=tl.float32) for i_k in range(tl.cdiv(K, BK)): p_q = tl.make_block_ptr( q + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) p_g = tl.make_block_ptr( g + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) p_h = tl.make_block_ptr( h + (i_tg * H + i_h) * K * V, (K, V), (V, 1), (i_k * BK, i_v * BV), (BK, BV), (1, 0), ) # [BT, BK] b_q = tl.load(p_q, boundary_check=(0, 1)) b_q = (b_q * scale).to(b_q.dtype) # [BT, BK] b_g = tl.load(p_g, boundary_check=(0, 1)) # [BT, BK] b_qg = (b_q * exp(b_g)).to(b_q.dtype) # [BK, BV] b_h = tl.load(p_h, boundary_check=(0, 1)) # works but dkw, owing to divine benevolence # [BT, BV] if i_k >= 0: b_o += tl.dot(b_qg, b_h.to(b_qg.dtype)) p_v = tl.make_block_ptr( v + (bos * H + i_h) * V, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0), ) p_o = tl.make_block_ptr( o + (bos * H + i_h) * V, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0), ) p_A = tl.make_block_ptr( A + (bos * H + i_h) * BT, (T, BT), (H * BT, 1), (i_t * BT, 0), (BT, BT), (1, 0) ) # [BT, BV] b_v = tl.load(p_v, boundary_check=(0, 1)) # [BT, BT] b_A = tl.load(p_A, boundary_check=(0, 1)) b_A = tl.where(m_s, b_A, 0.0).to(b_v.dtype) b_o += tl.dot(b_A, b_v, allow_tf32=False) tl.store(p_o, b_o.to(p_o.dtype.element_ty), boundary_check=(0, 1)) def chunk_gla_fwd_o_gk( q: torch.Tensor, v: torch.Tensor, g: torch.Tensor, A: torch.Tensor, h: torch.Tensor, o: torch.Tensor, scale: float, cu_seqlens: torch.LongTensor | None = None, chunk_size: int = 64, ):
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/chunk_scaled_dot_kkt.py
vllm/model_executor/layers/fla/ops/chunk_scaled_dot_kkt.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices from .op import exp @triton.heuristics( { "USE_G": lambda args: args["g"] is not None, "IS_VARLEN": lambda args: args["cu_seqlens"] is not None, } ) @triton.autotune( configs=[ triton.Config({"BK": BK}, num_warps=num_warps, num_stages=num_stages) for BK in [32, 64, 128] for num_warps in [2, 4, 8] for num_stages in [2, 3, 4] ], key=["H", "K", "BT", "IS_VARLEN"], ) @triton.jit(do_not_specialize=["T"]) def chunk_scaled_dot_kkt_fwd_kernel( k, beta, g, A, cu_seqlens, chunk_indices, T, H: tl.constexpr, Hg: tl.constexpr, K: tl.constexpr, BT: tl.constexpr, BK: tl.constexpr, IS_VARLEN: tl.constexpr, USE_G: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T o_t = i_t * BT + tl.arange(0, BT) m_t = o_t < T p_beta = tl.make_block_ptr( beta + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,) ) b_beta = tl.load(p_beta, boundary_check=(0,)) b_A = tl.zeros([BT, BT], dtype=tl.float32) for i_k in range(tl.cdiv(K, BK)): p_k = tl.make_block_ptr( k + (bos * Hg + i_h // (H // Hg)) * K, (T, K), (Hg * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_kb = b_k * b_beta[:, None] b_A += tl.dot(b_kb.to(b_k.dtype), tl.trans(b_k)) if USE_G: p_g = tl.make_block_ptr(g + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,)) b_g = tl.load(p_g, boundary_check=(0,)) b_g_diff = b_g[:, None] - b_g[None, :] b_A = b_A * exp(b_g_diff) m_A = (o_t[:, None] > o_t[None, :]) & (m_t[:, None] & m_t) b_A = tl.where(m_A, b_A, 0) p_A = tl.make_block_ptr( A + (bos * H + i_h) * BT, (T, BT), (BT * H, 1), (i_t * BT, 0), (BT, BT), (1, 0) ) tl.store(p_A, b_A.to(p_A.dtype.element_ty), boundary_check=(0, 1)) def chunk_scaled_dot_kkt_fwd( k: torch.Tensor, g: torch.Tensor | None = None, beta: torch.Tensor | None = None, cu_seqlens: torch.LongTensor | None = None, chunk_size: int = 64, output_dtype: torch.dtype = torch.float32, ) -> torch.Tensor: r""" Compute beta * K * K^T. Args: k (torch.Tensor): The key tensor of shape `[B, T, H, K]`. beta (torch.Tensor): The beta tensor of shape `[B, T, H]`. g (torch.Tensor): The cumulative sum of the gate tensor of shape `[B, T, H]`. Default: `None`. cu_seqlens (torch.LongTensor): The cumulative sequence lengths of the input tensor. Default: None chunk_size (int): The chunk size. Default: 64. output_dtype (torch.dtype): The dtype of the output tensor. Default: `torch.float32` Returns: beta * K * K^T of shape `[B, T, H, BT]` where `BT` is the chunk size. """ # This kernel is slightly different from fla to support Q/K with different head numbers. # In fla, Q/K always have the same head number, so Hg is always equal to H. B, T, Hg, K = k.shape H = beta.shape[-1] BT = chunk_size chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = triton.cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) A = torch.empty(B, T, H, BT, device=k.device, dtype=output_dtype) chunk_scaled_dot_kkt_fwd_kernel[(NT, B * H)]( k=k, g=g, beta=beta, A=A, cu_seqlens=cu_seqlens, chunk_indices=chunk_indices, T=T, H=H, Hg=Hg, K=K, BT=BT, ) return A
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/wy_fast.py
vllm/model_executor/layers/fla/ops/wy_fast.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({}, num_warps=num_warps, num_stages=num_stages) for num_warps in [2, 4, 8] for num_stages in [2, 3, 4] ], key=["H", "K", "V", "BT", "BK", "BV", "IS_VARLEN"], ) @triton.jit(do_not_specialize=["T"]) def recompute_w_u_fwd_kernel( k, v, beta, w, u, A, g, cu_seqlens, chunk_indices, T, H: tl.constexpr, Hg: tl.constexpr, K: tl.constexpr, V: tl.constexpr, BT: tl.constexpr, BK: tl.constexpr, BV: tl.constexpr, IS_VARLEN: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T p_beta = tl.make_block_ptr( beta + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,) ) p_g = tl.make_block_ptr(g + (bos * H + i_h), (T,), (H,), (i_t * BT,), (BT,), (0,)) p_A = tl.make_block_ptr( A + (bos * H + i_h) * BT, (T, BT), (H * BT, 1), (i_t * BT, 0), (BT, BT), (1, 0) ) b_beta = tl.load(p_beta, boundary_check=(0,)) b_A = tl.load(p_A, boundary_check=(0, 1)) b_g = tl.exp(tl.load(p_g, boundary_check=(0,))) for i_v in range(tl.cdiv(V, BV)): p_v = tl.make_block_ptr( v + (bos * H + i_h) * V, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0), ) p_u = tl.make_block_ptr( u + (bos * H + i_h) * V, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0), ) b_v = tl.load(p_v, boundary_check=(0, 1)) b_vb = (b_v * b_beta[:, None]).to(b_v.dtype) b_u = tl.dot(b_A, b_vb, allow_tf32=False) tl.store(p_u, b_u.to(p_u.dtype.element_ty), boundary_check=(0, 1)) for i_k in range(tl.cdiv(K, BK)): p_k = tl.make_block_ptr( k + (bos * Hg + i_h // (H // Hg)) * K, (T, K), (Hg * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) p_w = tl.make_block_ptr( w + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_kb = (b_k * b_beta[:, None] * b_g[:, None]).to(b_k.dtype) b_w = tl.dot(b_A, b_kb) tl.store(p_w, b_w.to(p_w.dtype.element_ty), boundary_check=(0, 1)) def recompute_w_u_fwd( k: torch.Tensor, v: torch.Tensor, beta: torch.Tensor, g_cumsum: torch.Tensor, A: torch.Tensor, cu_seqlens: torch.LongTensor | None, ) -> tuple[torch.Tensor, torch.Tensor]: B, T, Hg, K, V = *k.shape, v.shape[-1] H = v.shape[-2] BT = A.shape[-1] chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = triton.cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) BK = 64 BV = 64 u = torch.empty_like(v) w = k.new_empty(B, T, H, K) recompute_w_u_fwd_kernel[(NT, B * H)]( k=k, v=v, beta=beta, w=w, u=u, A=A, g=g_cumsum, cu_seqlens=cu_seqlens, chunk_indices=chunk_indices, T=T, H=H, Hg=Hg, K=K, V=V, BT=BT, BK=BK, BV=BV, ) return w, u
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/cumsum.py
vllm/model_executor/layers/fla/ops/cumsum.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import warnings import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices from .utils import check_shared_mem, input_guard BS_LIST = [32, 64] if check_shared_mem() else [16, 32] @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[triton.Config({}, num_warps=num_warps) for num_warps in [1, 2, 4, 8]], key=["B", "H", "BT", "IS_VARLEN", "REVERSE"], ) @triton.jit(do_not_specialize=["T"]) def chunk_local_cumsum_scalar_kernel( s, o, cu_seqlens, chunk_indices, T, B: tl.constexpr, H: tl.constexpr, BT: tl.constexpr, REVERSE: tl.constexpr, IS_VARLEN: tl.constexpr, HEAD_FIRST: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T if HEAD_FIRST: p_s = tl.make_block_ptr( s + bos * H + i_h * T, (T,), (1,), (i_t * BT,), (BT,), (0,) ) p_o = tl.make_block_ptr( o + bos * H + i_h * T, (T,), (1,), (i_t * BT,), (BT,), (0,) ) else: p_s = tl.make_block_ptr(s + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,)) p_o = tl.make_block_ptr(o + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,)) # [BT] b_s = tl.load(p_s, boundary_check=(0,)).to(tl.float32) b_o = tl.cumsum(b_s, axis=0) if REVERSE: b_z = tl.sum(b_s, axis=0) b_o = -b_o + b_z[None] + b_s tl.store(p_o, b_o.to(p_o.dtype.element_ty), boundary_check=(0,)) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({"BS": BS}, num_warps=num_warps) for BS in BS_LIST for num_warps in [2, 4, 8] ], key=["B", "H", "S", "BT", "IS_VARLEN", "REVERSE"], ) @triton.jit(do_not_specialize=["T"]) def chunk_local_cumsum_vector_kernel( s, o, cu_seqlens, chunk_indices, T, B: tl.constexpr, H: tl.constexpr, S: tl.constexpr, BT: tl.constexpr, BS: tl.constexpr, REVERSE: tl.constexpr, IS_VARLEN: tl.constexpr, HEAD_FIRST: tl.constexpr, ): i_s, i_t, i_bh = tl.program_id(0), tl.program_id(1), tl.program_id(2) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T o_i = tl.arange(0, BT) if REVERSE: m_s = tl.where(o_i[:, None] <= o_i[None, :], 1.0, 0.0) else: m_s = tl.where(o_i[:, None] >= o_i[None, :], 1.0, 0.0) if HEAD_FIRST: p_s = tl.make_block_ptr( s + (bos * H + i_h * T) * S, (T, S), (S, 1), (i_t * BT, i_s * BS), (BT, BS), (1, 0), ) p_o = tl.make_block_ptr( o + (bos * H + i_h * T) * S, (T, S), (S, 1), (i_t * BT, i_s * BS), (BT, BS), (1, 0), ) else: p_s = tl.make_block_ptr( s + (bos * H + i_h) * S, (T, S), (H * S, 1), (i_t * BT, i_s * BS), (BT, BS), (1, 0), ) p_o = tl.make_block_ptr( o + (bos * H + i_h) * S, (T, S), (H * S, 1), (i_t * BT, i_s * BS), (BT, BS), (1, 0), ) # [BT, BS] b_s = tl.load(p_s, boundary_check=(0, 1)).to(tl.float32) b_o = tl.dot(m_s, b_s, allow_tf32=False) tl.store(p_o, b_o.to(p_o.dtype.element_ty), boundary_check=(0, 1)) def chunk_local_cumsum_scalar( g: torch.Tensor, chunk_size: int, reverse: bool = False, cu_seqlens: torch.Tensor | None = None, head_first: bool = False, output_dtype: torch.dtype | None = torch.float, ) -> torch.Tensor: if head_first: B, H, T = g.shape else: B, T, H = g.shape assert chunk_size == 2 ** (chunk_size.bit_length() - 1), ( "chunk_size must be a power of 2" ) BT = chunk_size chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = triton.cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) g_org, g = g, torch.empty_like(g, dtype=output_dtype or g.dtype) grid = (NT, B * H) chunk_local_cumsum_scalar_kernel[grid]( g_org, g, cu_seqlens, chunk_indices, T=T, B=B, H=H, BT=BT, HEAD_FIRST=head_first, REVERSE=reverse, ) return g def chunk_local_cumsum_vector( g: torch.Tensor, chunk_size: int, reverse: bool = False, cu_seqlens: torch.Tensor | None = None, head_first: bool = False, output_dtype: torch.dtype | None = torch.float, ) -> torch.Tensor: if head_first: B, H, T, S = g.shape else: B, T, H, S = g.shape BT = chunk_size chunk_indices = ( prepare_chunk_indices(cu_seqlens, chunk_size) if cu_seqlens is not None else None ) NT = triton.cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) assert chunk_size == 2 ** (chunk_size.bit_length() - 1), ( "chunk_size must be a power of 2" ) g_org, g = g, torch.empty_like(g, dtype=output_dtype or g.dtype) def grid(meta): return (triton.cdiv(meta["S"], meta["BS"]), NT, B * H) # keep cumulative normalizer in fp32 # this kernel is equivalent to # g = g.view(B, H, NT, BT, -1).cumsum(-2).view(B, H, T, -1) chunk_local_cumsum_vector_kernel[grid]( g_org, g, cu_seqlens, chunk_indices, T=T, B=B, H=H, S=S, BT=BT, HEAD_FIRST=head_first, REVERSE=reverse, ) return g @input_guard def chunk_local_cumsum( g: torch.Tensor, chunk_size: int, reverse: bool = False, cu_seqlens: torch.Tensor | None = None, head_first: bool = False, output_dtype: torch.dtype | None = torch.float, **kwargs, ) -> torch.Tensor: if not head_first and g.shape[1] < g.shape[2]: warnings.warn( f"Input tensor shape suggests potential format mismatch: seq_len ({g.shape[1]}) < num_heads ({g.shape[2]}). " "This may indicate the inputs were passed in head-first format [B, H, T, ...] " "when head_first=False was specified. " "Please verify your input tensor format matches the expected shape [B, T, H, ...].", stacklevel=2, ) if cu_seqlens is not None: assert g.shape[0] == 1, ( "Only batch size 1 is supported when cu_seqlens are provided" ) if len(g.shape) == 3: return chunk_local_cumsum_scalar( g, chunk_size, reverse, cu_seqlens, head_first, output_dtype ) elif len(g.shape) == 4: return chunk_local_cumsum_vector( g, chunk_size, reverse, cu_seqlens, head_first, output_dtype ) else: raise ValueError( f"Unsupported input shape {g.shape}. " f"which should be (B, T, H, D) if `head_first=False` " f"or (B, H, T, D) otherwise" )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/chunk.py
vllm/model_executor/layers/fla/ops/chunk.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import warnings import torch from einops import rearrange from .chunk_delta_h import chunk_gated_delta_rule_fwd_h from .chunk_o import chunk_fwd_o from .chunk_scaled_dot_kkt import chunk_scaled_dot_kkt_fwd from .cumsum import chunk_local_cumsum from .l2norm import l2norm_fwd from .solve_tril import solve_tril from .utils import SUPPRESS_LEVEL, input_guard from .wy_fast import recompute_w_u_fwd def chunk_gated_delta_rule_fwd( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float, initial_state: torch.Tensor, output_final_state: bool, cu_seqlens: torch.LongTensor | None = None, ): g = chunk_local_cumsum(g, chunk_size=64, cu_seqlens=cu_seqlens) # obtain WY representation. u is actually the new v. A = chunk_scaled_dot_kkt_fwd( k=k, beta=beta, g=g, cu_seqlens=cu_seqlens, output_dtype=torch.float32 ) A = solve_tril(A=A, cu_seqlens=cu_seqlens, output_dtype=k.dtype) w, u = recompute_w_u_fwd( k=k, v=v, beta=beta, A=A, g_cumsum=g, cu_seqlens=cu_seqlens, ) h, v_new, final_state = chunk_gated_delta_rule_fwd_h( k=k, w=w, u=u, g=g, initial_state=initial_state, output_final_state=output_final_state, cu_seqlens=cu_seqlens, ) o = chunk_fwd_o( q=q, k=k, v=v_new, h=h, g=g, scale=scale, cu_seqlens=cu_seqlens, ) if SUPPRESS_LEVEL < 3: return g, o, A, final_state, None, None, None elif SUPPRESS_LEVEL >= 3: return g, o, A, final_state, w, h, v_new class ChunkGatedDeltaRuleFunction(torch.autograd.Function): @staticmethod @input_guard @torch.amp.custom_fwd(device_type="cuda") def forward( ctx, q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float, initial_state: torch.Tensor, output_final_state: bool, cu_seqlens: torch.LongTensor | None = None, use_qk_l2norm_in_kernel: bool = False, ): if use_qk_l2norm_in_kernel: q = l2norm_fwd(q) k = l2norm_fwd(k) g, o, A, final_state, w, h, v_new = chunk_gated_delta_rule_fwd( q=q, k=k, v=v, g=g, beta=beta, scale=scale, initial_state=initial_state, output_final_state=output_final_state, cu_seqlens=cu_seqlens, ) ctx.scale = scale ctx.use_qk_l2norm_in_kernel = use_qk_l2norm_in_kernel return o.to(q.dtype), final_state @torch.compiler.disable def chunk_gated_delta_rule( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float = None, initial_state: torch.Tensor = None, output_final_state: bool = False, cu_seqlens: torch.LongTensor | None = None, head_first: bool = False, use_qk_l2norm_in_kernel: bool = False, ): r""" Args: q (torch.Tensor): queries of shape `[B, T, H, K]` if `head_first=False` else `[B, H, T, K]`. k (torch.Tensor): keys of shape `[B, T, H, K]` if `head_first=False` else `[B, H, T, K]`. v (torch.Tensor): values of shape `[B, T, H, V]` if `head_first=False` else `[B, H, T, V]`. g (torch.Tensor): (forget) gating tensor (in log space!) of shape `[B, T, H]` if `head_first=False` else `[B, H, T]`. beta (torch.Tensor): betas of shape `[B, T, H]` if `head_first=False` else `[B, H, T]`. scale (Optional[int]): Scale factor for the RetNet attention scores. If not provided, it will default to `1 / sqrt(K)`. Default: `None`. initial_state (Optional[torch.Tensor]): Initial state of shape `[N, H, K, V]` for `N` input sequences. For equal-length input sequences, `N` equals the batch size `B`. Default: `None`. output_final_state (Optional[bool]): Whether to output the final state of shape `[N, H, K, V]`. Default: `False`. cu_seqlens (torch.LongTensor): Cumulative sequence lengths of shape `[N+1]` used for variable-length training, consistent with the FlashAttention API. head_first (Optional[bool]): Whether the inputs are in the head-first format, which is not supported for variable-length inputs. Default: `False`. Returns: o (torch.Tensor): Outputs of shape `[B, T, H, V]` if `head_first=False` else `[B, H, T, V]`. final_state (torch.Tensor): Final state of shape `[N, H, K, V]` if `output_final_state=True` else `None`. Examples:: >>> import torch >>> import torch.nn.functional as F >>> from einops import rearrange >>> from fla.ops.gated_delta_rule import chunk_gated_delta_rule # inputs with equal lengths >>> B, T, H, K, V = 4, 2048, 4, 512, 512 >>> q = torch.randn(B, T, H, K, dtype=torch.bfloat16, device='cuda') >>> k = F.normalize(torch.randn(B, T, H, K, dtype=torch.bfloat16, device='cuda'), p=2, dim=-1) >>> v = torch.randn(B, T, H, V, dtype=torch.bfloat16, device='cuda') >>> beta = torch.rand(B, T, H, dtype=torch.bfloat16, device='cuda').sigmoid() >>> g = F.logsigmoid(torch.rand(B, T, H, dtype=torch.bfloat16, device='cuda')) >>> h0 = torch.randn(B, H, K, V, dtype=torch.bfloat16, device='cuda') >>> o, ht = chunk_gated_delta_rule( q, k, v, g, beta, initial_state=h0, output_final_state=True ) # for variable-length inputs, the batch size `B` is expected to be 1 and `cu_seqlens` is required >>> q, k, v, beta, g = map(lambda x: rearrange(x, 'b t ... -> 1 (b t) ...'), (q, k, v, beta, g)) # for a batch with 4 sequences, `cu_seqlens` with 5 start/end positions are expected >>> cu_seqlens = q.new_tensor([0, 2048, 4096, 6144, 8192], dtype=torch.long) >>> o_var, ht_var = chunk_gated_delta_rule( q, k, v, g, beta, initial_state=h0, output_final_state=True, cu_seqlens=cu_seqlens ) """ assert q.dtype == k.dtype == v.dtype assert q.dtype != torch.float32, ( "ChunkGatedDeltaRuleFunction does not support float32. Please use bfloat16." ) assert len(beta.shape) == 3, ( "beta must be of shape [B, T, H] if head_first=False, or [B, H, T] otherwise." ) if head_first: raise DeprecationWarning( "head_first is deprecated and will be removed in a future version. " "Please use head_first=False for now instead.", stacklevel=2, ) q, k, v, beta, g = map( lambda x: rearrange(x, "b h t ... -> b t h ..."), (q, k, v, beta, g) ) if not head_first and q.shape[1] < q.shape[2]: warnings.warn( f"Input tensor shape suggests potential format mismatch: seq_len ({q.shape[1]}) < num_heads ({q.shape[2]}). " "This may indicate the inputs were passed in head-first format [B, H, T, ...] " "when head_first=False was specified. " "Please verify your input tensor format matches the expected shape [B, T, H, ...].", stacklevel=2, ) if cu_seqlens is not None: if q.shape[0] != 1: raise ValueError( f"The batch size is expected to be 1 rather than {q.shape[0]} when using `cu_seqlens`." f"Please flatten variable-length inputs before processing." ) if initial_state is not None and initial_state.shape[0] != len(cu_seqlens) - 1: raise ValueError( f"The number of initial states is expected to be equal to the number of input sequences, " f"i.e., {len(cu_seqlens) - 1} rather than {initial_state.shape[0]}." ) if scale is None: scale = k.shape[-1] ** -0.5 o, final_state = ChunkGatedDeltaRuleFunction.apply( q, k, v, g, beta, scale, initial_state, output_final_state, cu_seqlens, use_qk_l2norm_in_kernel, ) if head_first: o = rearrange(o, "b t h ... -> b h t ...") return o, final_state
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/l2norm.py
vllm/model_executor/layers/fla/ops/l2norm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang import os import torch from vllm.triton_utils import tl, triton BT_LIST = [8, 16, 32, 64, 128] USE_DEFAULT_FLA_NORM = int(os.getenv("USE_DEFAULT_FLA_NORM", "0")) @triton.autotune( configs=[ triton.Config({}, num_warps=num_warps) for num_warps in [1, 2, 4, 8, 16, 32] ], key=["D"], ) @triton.jit def l2norm_fwd_kernel1( x, y, D, BD: tl.constexpr, eps, ): i_t = tl.program_id(0) x += i_t * D y += i_t * D # Compute mean and variance cols = tl.arange(0, BD) mask = cols < D b_x = tl.load(x + cols, mask=mask, other=0.0).to(tl.float32) b_var = tl.sum(b_x * b_x, axis=0) b_rstd = 1 / tl.sqrt(b_var + eps) # tl.store(Rstd + i_t, rstd) # Normalize and apply linear transformation b_y = b_x * b_rstd tl.store(y + cols, b_y, mask=mask) @triton.autotune( configs=[ triton.Config({"BT": BT}, num_warps=num_warps) for num_warps in [1, 2, 4, 8, 16] for BT in BT_LIST ], key=["D"], ) @triton.jit(do_not_specialize=["NB"]) def l2norm_fwd_kernel( x, y, eps, NB, T, D: tl.constexpr, BT: tl.constexpr, BD: tl.constexpr, ): i_t = tl.program_id(0) p_x = tl.make_block_ptr(x, (T, D), (D, 1), (i_t * BT, 0), (BT, BD), (1, 0)) b_x = tl.load(p_x, boundary_check=(0, 1)).to(tl.float32) b_var = tl.sum(b_x * b_x, axis=1) b_y = b_x / tl.sqrt(b_var + eps)[:, None] p_y = tl.make_block_ptr(y, (T, D), (D, 1), (i_t * BT, 0), (BT, BD), (1, 0)) tl.store(p_y, b_y.to(p_y.dtype.element_ty), boundary_check=(0, 1)) @triton.jit def l2norm_fwd_kernel2(X, Y, eps, M, N: tl.constexpr, MBLOCK: tl.constexpr): xoffset = tl.program_id(0) * MBLOCK row_idx = xoffset + tl.arange(0, MBLOCK)[:, None] xmask = row_idx < M rindex = tl.arange(0, N)[None, :] xs = tl.load(X + (rindex + N * row_idx), xmask).to(tl.float32) square = tl.broadcast_to(xs * xs, [MBLOCK, N]) square_sum = tl.sum(tl.where(xmask, square, 0), 1)[:, None] rsqrt = tl.rsqrt(square_sum + eps) tl.store(Y + (rindex + N * row_idx), xs * rsqrt, xmask) def l2norm_fwd( x: torch.Tensor, eps: float = 1e-6, output_dtype: torch.dtype | None = None ): x_shape_og = x.shape x = x.view(-1, x.shape[-1]) # allocate output if output_dtype is None: y = torch.empty_like(x) else: y = torch.empty_like(x, dtype=output_dtype) assert y.stride(-1) == 1 T, D = x.shape[0], x.shape[-1] # rstd = torch.empty((T,), dtype=torch.float32, device=x.device) # Less than 64KB per feature: enqueue fused kernel MAX_FUSED_SIZE = 65536 // x.element_size() BD = min(MAX_FUSED_SIZE, triton.next_power_of_2(D)) if D > BD: raise RuntimeError("This layer doesn't support feature dim >= 64KB.") if not USE_DEFAULT_FLA_NORM: MBLOCK = 32 # M, N = x.shape l2norm_fwd_kernel2[(triton.cdiv(T, MBLOCK),)]( x, y, eps, T, D, MBLOCK, ) else: if D <= 512: NB = triton.cdiv(T, 2048) def grid(meta): return (triton.cdiv(T, meta["BT"]),) l2norm_fwd_kernel[grid]( x, y, eps, NB=NB, T=T, D=D, BD=BD, ) else: l2norm_fwd_kernel1[(T,)]( x, y, eps=eps, D=D, BD=BD, ) return y.view(x_shape_og)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/chunk_o.py
vllm/model_executor/layers/fla/ops/chunk_o.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices from .op import exp from .utils import FLA_GDN_FIX_BT, check_shared_mem, is_nvidia_hopper BKV_LIST = [64, 128] if check_shared_mem() else [32, 64] NUM_WARPS = [2, 4] if is_nvidia_hopper else [2, 4, 8] @triton.heuristics( { "USE_G": lambda args: args["g"] is not None, "IS_VARLEN": lambda args: args["cu_seqlens"] is not None, } ) @triton.autotune( configs=[ triton.Config({"BK": BK, "BV": BV}, num_warps=num_warps, num_stages=num_stages) for BK in BKV_LIST for BV in BKV_LIST for num_warps in NUM_WARPS for num_stages in [2, 3, 4] ], key=["H", "K", "V", "BT"], ) @triton.jit(do_not_specialize=["T"]) def chunk_fwd_kernel_o( q, k, v, h, g, o, cu_seqlens, chunk_indices, scale, T, H: tl.constexpr, Hg: tl.constexpr, K: tl.constexpr, V: tl.constexpr, BT: tl.constexpr, BK: tl.constexpr, BV: tl.constexpr, USE_G: tl.constexpr, IS_VARLEN: tl.constexpr, ): i_v, i_t, i_bh = tl.program_id(0), tl.program_id(1), tl.program_id(2) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_tg = i_t i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos NT = tl.cdiv(T, BT) else: NT = tl.cdiv(T, BT) i_tg = i_b * NT + i_t bos, eos = i_b * T, i_b * T + T # offset calculation q += (bos * Hg + i_h // (H // Hg)) * K k += (bos * Hg + i_h // (H // Hg)) * K v += (bos * H + i_h) * V o += (bos * H + i_h) * V h += (i_tg * H + i_h).to(tl.int64) * K * V b_o = tl.zeros([BT, BV], dtype=tl.float32) b_A = tl.zeros([BT, BT], dtype=tl.float32) for i_k in range(tl.cdiv(K, BK)): p_q = tl.make_block_ptr( q, (T, K), (Hg * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0) ) p_k = tl.make_block_ptr( k, (K, T), (1, Hg * K), (i_k * BK, i_t * BT), (BK, BT), (0, 1) ) p_h = tl.make_block_ptr( h, (K, V), (V, 1), (i_k * BK, i_v * BV), (BK, BV), (1, 0) ) # [BT, BK] b_q = tl.load(p_q, boundary_check=(0, 1)) # [BK, BT] b_k = tl.load(p_k, boundary_check=(0, 1)) # [BK, BV] b_h = tl.load(p_h, boundary_check=(0, 1)) # [BT, BK] @ [BK, BV] -> [BT, BV] b_o += tl.dot(b_q, b_h) # [BT, BK] @ [BK, BT] -> [BT, BT] b_A += tl.dot(b_q, b_k) if USE_G: g += bos * H + i_h p_g = tl.make_block_ptr(g, (T,), (H,), (i_t * BT,), (BT,), (0,)) b_g = tl.load(p_g, boundary_check=(0,)) b_o = b_o * exp(b_g)[:, None] b_A = b_A * exp(b_g[:, None] - b_g[None, :]) o_t = i_t * BT + tl.arange(0, BT) m_t = o_t < T m_A = (o_t[:, None] >= o_t[None, :]) & (m_t[:, None] & m_t) b_A = tl.where(m_A, b_A, 0) p_v = tl.make_block_ptr( v, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0) ) p_o = tl.make_block_ptr( o, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0) ) b_v = tl.load(p_v, boundary_check=(0, 1)) # to fix mma -> mma layout conversion # already solved by triton v3.2 or higher b_o = b_o * scale + tl.dot(b_A.to(b_v.dtype), b_v) * scale tl.store(p_o, b_o.to(p_o.dtype.element_ty), boundary_check=(0, 1)) def chunk_fwd_o( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, h: torch.Tensor, g: torch.Tensor | None = None, # cumsum of log decay scale: float | None = None, cu_seqlens: torch.LongTensor | None = None, chunk_size: int = 64, ) -> torch.Tensor: B, T, Hg, K, V = *q.shape, v.shape[-1] H = v.shape[-2] BT = 64 if FLA_GDN_FIX_BT else min(chunk_size, max(16, triton.next_power_of_2(T))) chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = triton.cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) if scale is None: scale = k.shape[-1] ** -0.5 o = torch.empty_like(v) def grid(meta): return (triton.cdiv(V, meta["BV"]), NT, B * H) chunk_fwd_kernel_o[grid]( q, k, v, h, g, o, cu_seqlens, chunk_indices, scale, T=T, H=H, Hg=Hg, K=K, V=V, BT=BT, ) return o
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/chunk_delta_h.py
vllm/model_executor/layers/fla/ops/chunk_delta_h.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices, prepare_chunk_offsets from .op import exp from .utils import use_cuda_graph NUM_WARPS = [2, 4, 8, 16] @triton.heuristics( { "USE_G": lambda args: args["g"] is not None, "USE_GK": lambda args: args["gk"] is not None, "USE_INITIAL_STATE": lambda args: args["h0"] is not None, "STORE_FINAL_STATE": lambda args: args["ht"] is not None, "SAVE_NEW_VALUE": lambda args: args["v_new"] is not None, "IS_VARLEN": lambda args: args["cu_seqlens"] is not None, } ) @triton.autotune( configs=[ triton.Config({"BV": BV}, num_warps=num_warps, num_stages=num_stages) for num_warps in [2, 4] for num_stages in [2, 3, 4] for BV in [32, 64] ], key=["H", "K", "V", "BT"], use_cuda_graph=use_cuda_graph, ) @triton.jit(do_not_specialize=["T"]) def chunk_gated_delta_rule_fwd_kernel_h_blockdim64( k, v, w, v_new, g, gk, h, h0, ht, cu_seqlens, chunk_offsets, T, H: tl.constexpr, Hg: tl.constexpr, K: tl.constexpr, V: tl.constexpr, BT: tl.constexpr, BV: tl.constexpr, USE_G: tl.constexpr, USE_GK: tl.constexpr, USE_INITIAL_STATE: tl.constexpr, STORE_FINAL_STATE: tl.constexpr, SAVE_NEW_VALUE: tl.constexpr, IS_VARLEN: tl.constexpr, ): i_v, i_nh = tl.program_id(0), tl.program_id(1) i_n, i_h = i_nh // H, i_nh % H if IS_VARLEN: bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos NT = tl.cdiv(T, BT) boh = tl.load(chunk_offsets + i_n).to(tl.int32) else: bos, eos = i_n * T, i_n * T + T NT = tl.cdiv(T, BT) boh = i_n * NT # [BK, BV] b_h1 = tl.zeros([64, BV], dtype=tl.float32) if K > 64: b_h2 = tl.zeros([64, BV], dtype=tl.float32) if K > 128: b_h3 = tl.zeros([64, BV], dtype=tl.float32) if K > 192: b_h4 = tl.zeros([64, BV], dtype=tl.float32) # calculate offset h += ((boh * H + i_h) * K * V).to(tl.int64) v += ((bos * H + i_h) * V).to(tl.int64) k += ((bos * Hg + i_h // (H // Hg)) * K).to(tl.int64) w += ((bos * H + i_h) * K).to(tl.int64) if SAVE_NEW_VALUE: v_new += ((bos * H + i_h) * V).to(tl.int64) stride_v = H * V stride_h = H * K * V stride_k = Hg * K stride_w = H * K if USE_INITIAL_STATE: h0 = h0 + i_nh * K * V if STORE_FINAL_STATE: ht = ht + i_nh * K * V # load initial state if USE_INITIAL_STATE: p_h0_1 = tl.make_block_ptr(h0, (K, V), (V, 1), (0, i_v * BV), (64, BV), (1, 0)) b_h1 += tl.load(p_h0_1, boundary_check=(0, 1)).to(tl.float32) if K > 64: p_h0_2 = tl.make_block_ptr( h0, (K, V), (V, 1), (64, i_v * BV), (64, BV), (1, 0) ) b_h2 += tl.load(p_h0_2, boundary_check=(0, 1)).to(tl.float32) if K > 128: p_h0_3 = tl.make_block_ptr( h0, (K, V), (V, 1), (128, i_v * BV), (64, BV), (1, 0) ) b_h3 += tl.load(p_h0_3, boundary_check=(0, 1)).to(tl.float32) if K > 192: p_h0_4 = tl.make_block_ptr( h0, (K, V), (V, 1), (192, i_v * BV), (64, BV), (1, 0) ) b_h4 += tl.load(p_h0_4, boundary_check=(0, 1)).to(tl.float32) # main recurrence for i_t in range(NT): p_h1 = tl.make_block_ptr( h + i_t * stride_h, (K, V), (V, 1), (0, i_v * BV), (64, BV), (1, 0) ) tl.store(p_h1, b_h1.to(p_h1.dtype.element_ty), boundary_check=(0, 1)) if K > 64: p_h2 = tl.make_block_ptr( h + i_t * stride_h, (K, V), (V, 1), (64, i_v * BV), (64, BV), (1, 0) ) tl.store(p_h2, b_h2.to(p_h2.dtype.element_ty), boundary_check=(0, 1)) if K > 128: p_h3 = tl.make_block_ptr( h + i_t * stride_h, (K, V), (V, 1), (128, i_v * BV), (64, BV), (1, 0) ) tl.store(p_h3, b_h3.to(p_h3.dtype.element_ty), boundary_check=(0, 1)) if K > 192: p_h4 = tl.make_block_ptr( h + i_t * stride_h, (K, V), (V, 1), (192, i_v * BV), (64, BV), (1, 0) ) tl.store(p_h4, b_h4.to(p_h4.dtype.element_ty), boundary_check=(0, 1)) p_w = tl.make_block_ptr( w, (T, K), (stride_w, 1), (i_t * BT, 0), (BT, 64), (1, 0) ) b_w = tl.load(p_w, boundary_check=(0, 1)) b_v = tl.dot(b_w, b_h1.to(b_w.dtype)) if K > 64: p_w = tl.make_block_ptr( w, (T, K), (stride_w, 1), (i_t * BT, 64), (BT, 64), (1, 0) ) b_w = tl.load(p_w, boundary_check=(0, 1)) b_v += tl.dot(b_w, b_h2.to(b_w.dtype)) if K > 128: p_w = tl.make_block_ptr( w, (T, K), (stride_w, 1), (i_t * BT, 128), (BT, 64), (1, 0) ) b_w = tl.load(p_w, boundary_check=(0, 1)) b_v += tl.dot(b_w, b_h3.to(b_w.dtype)) if K > 192: p_w = tl.make_block_ptr( w, (T, K), (stride_w, 1), (i_t * BT, 192), (BT, 64), (1, 0) ) b_w = tl.load(p_w, boundary_check=(0, 1)) b_v += tl.dot(b_w, b_h4.to(b_w.dtype)) p_v = tl.make_block_ptr( v, (T, V), (stride_v, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0) ) b_v = tl.load(p_v, boundary_check=(0, 1)) - b_v if SAVE_NEW_VALUE: p_v = tl.make_block_ptr( v_new, (T, V), (stride_v, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0) ) tl.store(p_v, b_v.to(p_v.dtype.element_ty), boundary_check=(0, 1)) last_idx = min((i_t + 1) * BT, T) - 1 if USE_G: m_t = (i_t * BT + tl.arange(0, BT)) < T b_g_last = tl.load(g + bos * H + last_idx * H + i_h) p_g = tl.make_block_ptr( g + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,) ) b_g = tl.load(p_g, boundary_check=(0,)) b_v = b_v * tl.where(m_t, exp(b_g_last - b_g), 0)[:, None] b_g_last = exp(b_g_last) b_h1 *= b_g_last if K > 64: b_h2 *= b_g_last if K > 128: b_h3 *= b_g_last if K > 192: b_h4 *= b_g_last if USE_GK: o_k1 = tl.arange(0, 64) b_gk_last1 = tl.load( gk + (bos + last_idx) * H * K + i_h * K + o_k1, mask=(o_k1 < K), other=0.0, ) b_h1 *= exp(b_gk_last1)[:, None] if K > 64: o_k2 = 64 + o_k1 b_gk_last2 = tl.load( gk + (bos + last_idx) * H * K + i_h * K + o_k2, mask=(o_k2 < K), other=0.0, ) b_h2 *= exp(b_gk_last2)[:, None] if K > 128: o_k3 = 128 + o_k1 b_gk_last3 = tl.load( gk + (bos + last_idx) * H * K + i_h * K + o_k3, mask=(o_k3 < K), other=0.0, ) b_h3 *= exp(b_gk_last3)[:, None] if K > 192: o_k4 = 192 + o_k1 b_gk_last4 = tl.load( gk + (bos + last_idx) * H * K + i_h * K + o_k4, mask=(o_k4 < K), other=0.0, ) b_h4 *= exp(b_gk_last4)[:, None] b_v = b_v.to(k.dtype.element_ty) p_k = tl.make_block_ptr( k, (K, T), (1, stride_k), (0, i_t * BT), (64, BT), (0, 1) ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_h1 += tl.dot(b_k, b_v) if K > 64: p_k = tl.make_block_ptr( k, (K, T), (1, stride_k), (64, i_t * BT), (64, BT), (0, 1) ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_h2 += tl.dot(b_k, b_v) if K > 128: p_k = tl.make_block_ptr( k, (K, T), (1, stride_k), (128, i_t * BT), (64, BT), (0, 1) ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_h3 += tl.dot(b_k, b_v) if K > 192: p_k = tl.make_block_ptr( k, (K, T), (1, stride_k), (192, i_t * BT), (64, BT), (0, 1) ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_h4 += tl.dot(b_k, b_v) # epilogue if STORE_FINAL_STATE: p_ht = tl.make_block_ptr(ht, (K, V), (V, 1), (0, i_v * BV), (64, BV), (1, 0)) tl.store(p_ht, b_h1.to(p_ht.dtype.element_ty), boundary_check=(0, 1)) if K > 64: p_ht = tl.make_block_ptr( ht, (K, V), (V, 1), (64, i_v * BV), (64, BV), (1, 0) ) tl.store(p_ht, b_h2.to(p_ht.dtype.element_ty), boundary_check=(0, 1)) if K > 128: p_ht = tl.make_block_ptr( ht, (K, V), (V, 1), (128, i_v * BV), (64, BV), (1, 0) ) tl.store(p_ht, b_h3.to(p_ht.dtype.element_ty), boundary_check=(0, 1)) if K > 192: p_ht = tl.make_block_ptr( ht, (K, V), (V, 1), (192, i_v * BV), (64, BV), (1, 0) ) tl.store(p_ht, b_h4.to(p_ht.dtype.element_ty), boundary_check=(0, 1)) def chunk_gated_delta_rule_fwd_h( k: torch.Tensor, w: torch.Tensor, u: torch.Tensor, g: torch.Tensor | None = None, gk: torch.Tensor | None = None, initial_state: torch.Tensor | None = None, output_final_state: bool = False, chunk_size: int = 64, # SY: remove this argument and force chunk size 64? save_new_value: bool = True, cu_seqlens: torch.LongTensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: # This kernel is slightly different from fla to support Q/K with different head numbers. # In fla, Q/K always have the same head number, so Hg is always equal to H. B, T, Hg, K, V = *k.shape, u.shape[-1] H = u.shape[-2] BT = chunk_size chunk_indices = ( prepare_chunk_indices(cu_seqlens, chunk_size) if cu_seqlens is not None else None ) # N: the actual number of sequences in the batch with either equal or variable lengths if cu_seqlens is None: N, NT, chunk_offsets = B, triton.cdiv(T, BT), None else: N, NT, chunk_offsets = ( len(cu_seqlens) - 1, len(chunk_indices), prepare_chunk_offsets(cu_seqlens, BT), ) assert K <= 256, "current kernel does not support head dimension larger than 256." h = k.new_empty(B, NT, H, K, V) final_state = ( k.new_empty(N, H, K, V, dtype=torch.float32) if output_final_state else None ) v_new = torch.empty_like(u) if save_new_value else None def grid(meta): return (triton.cdiv(V, meta["BV"]), N * H) chunk_gated_delta_rule_fwd_kernel_h_blockdim64[grid]( k=k, v=u, w=w, v_new=v_new, g=g, gk=gk, h=h, h0=initial_state, ht=final_state, cu_seqlens=cu_seqlens, chunk_offsets=chunk_offsets, T=T, H=H, Hg=Hg, K=K, V=V, BT=BT, ) return h, v_new, final_state
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/utils.py
vllm/model_executor/layers/fla/ops/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import contextlib import functools import logging import os from collections.abc import Callable from enum import Enum from typing import Any, Literal import torch from vllm.platforms import current_platform from vllm.triton_utils import triton logger = logging.getLogger(__name__) COMPILER_MODE = os.getenv("FLA_COMPILER_MODE") == "1" FLA_CI_ENV = os.getenv("FLA_CI_ENV") == "1" FLA_GDN_FIX_BT = os.getenv("FLA_GDN_FIX_BT", "0") == "1" SUPPRESS_LEVEL = int(os.getenv("GDN_RECOMPUTE_SUPPRESS_LEVEL", "0")) def tensor_cache(fn: Callable[..., torch.Tensor]) -> Callable[..., torch.Tensor]: """ A decorator that caches the most recent results of a function with tensor inputs. This decorator will store the output of the decorated function for the most recent set of input tensors. The cache is limited to a fixed size (default is 4). When the cache is full, the oldest entry will be removed. Args: fn (Callable[..., torch.Tensor]): The function to be decorated. It should take tensor inputs and return tensor outputs. Returns: Callable[..., torch.Tensor]: A wrapped version of the input function with single-entry caching. """ cache_entries: tuple[tuple | None, dict | None, Any] = [] cache_size = 8 @functools.wraps(fn) def wrapper(*args: Any, **kwargs: Any) -> Any: nonlocal cache_entries, cache_size for i, entry in enumerate(cache_entries): last_args, last_kwargs, last_result = entry if ( len(args) == len(last_args) and len(kwargs) == len(last_kwargs) and all(a is b for a, b in zip(args, last_args)) and all( k in last_kwargs and v is last_kwargs[k] for k, v in kwargs.items() ) ): cache_entries = ( cache_entries[:i] + cache_entries[i + 1 :] + [(args, kwargs, last_result)] ) return last_result result = fn(*args, **kwargs) if len(cache_entries) >= cache_size: cache_entries = cache_entries[1:] cache_entries.append((args, kwargs, result)) return result return wrapper def input_guard(fn: Callable[..., torch.Tensor]) -> Callable[..., torch.Tensor]: """ A decorator to make sure all input tensors are contiguous and set the device based on input tensors. """ @functools.wraps(fn) def wrapper(*args, **kwargs): contiguous_args = ( i if not isinstance(i, torch.Tensor) else i.contiguous() for i in args ) contiguous_kwargs = { k: (v if not isinstance(v, torch.Tensor) else v.contiguous()) for k, v in kwargs.items() } tensor = None for arg in args: if isinstance(arg, torch.Tensor): tensor = arg break if tensor is None: for value in kwargs.values(): if isinstance(value, torch.Tensor): tensor = value break if tensor is not None: ctx = torch.cuda.device(tensor.device.index) else: ctx = contextlib.nullcontext() with ctx: return fn(*contiguous_args, **contiguous_kwargs) return wrapper @functools.cache def get_available_device() -> str: try: return triton.runtime.driver.active.get_current_target().backend except (RuntimeError, AttributeError): return "cpu" @functools.cache def _check_platform() -> Literal["nvidia", "amd", "intel", "musa"]: device = get_available_device() mapping = { "cuda": "nvidia", "hip": "amd", "xpu": "intel", } # return the mapped value, or the original if not found return mapping.get(device, device) # For AMD GPUs, the triton backend is 'hip', while for Nvidia GPUs, the triton backend is 'cuda'. # However, the torch backend is 'cuda' for both Nvidia and AMD GPUs. # Therefore, we need to check the triton backend to determine the actual GPU vendor. device = "cuda" if current_platform.is_cuda_alike() else get_available_device() device_torch_lib = getattr(torch, device, None) device_platform = _check_platform() is_amd = device_platform == "amd" is_intel = device_platform == "intel" is_nvidia = device_platform == "nvidia" is_intel_alchemist = is_intel and "Intel(R) Arc(TM) A" in torch.xpu.get_device_name(0) is_nvidia_hopper = is_nvidia and ( "NVIDIA H" in torch.cuda.get_device_name(0) or torch.cuda.get_device_capability()[0] >= 9 ) use_cuda_graph = is_nvidia and os.environ.get("FLA_USE_CUDA_GRAPH", "0") == "1" is_gather_supported = hasattr(triton.language, "gather") is_tma_supported = (is_nvidia and torch.cuda.get_device_capability(0)[0] >= 9) and ( hasattr(triton.language, "_experimental_make_tensor_descriptor") or hasattr(triton.language, "make_tensor_descriptor") ) def get_all_max_shared_mem(): try: return [ triton.runtime.driver.active.utils.get_device_properties(i)[ "max_shared_mem" ] for i in range(device_torch_lib.device_count()) ] except BaseException: return [-1] class Backend(Enum): ADA = 101376 # RTX 4090 AMPERE = 166912 # A100 HOPPER = 232448 # H100 DEFAULT = 102400 # Default @classmethod def get_shared_memory(cls, arch: str) -> int: try: return cls[arch.upper()].value except KeyError: return cls.DEFAULT.value @functools.cache def check_shared_mem(arch: str = "none", tensor_idx: int = 0) -> bool: try: device_shared_mem_list = get_all_max_shared_mem() max_shared_memory = device_shared_mem_list[tensor_idx] return max_shared_memory >= Backend.get_shared_memory(arch) except Exception: return False
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/layernorm_guard.py
vllm/model_executor/layers/fla/ops/layernorm_guard.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Tri Dao # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2024, Tri Dao. # ruff: noqa: E501 # Based on the Triton LayerNorm tutorial: https://triton-lang.org/main/getting-started/tutorials/05-layer-norm.html # For the backward pass, we keep weight_grad and bias_grad in registers and accumulate. # This backward pass is faster for dimensions up to 8k, but after that it's much slower due to register spilling. # The models we train have hidden dim up to 8k anyway (e.g. Llama 70B), so this is fine. from functools import lru_cache import torch import torch.nn as nn import torch.nn.functional as F from einops import rearrange from vllm.triton_utils import tl, triton from vllm.utils.math_utils import cdiv, next_power_of_2 from .utils import input_guard def rms_norm_ref( x, weight, bias, z=None, eps=1e-6, group_size=None, norm_before_gate=True, upcast=True, ): dtype = x.dtype weight = weight.float() bias = bias.float() if bias is not None else None if upcast: x = x.float() z = z.float() if z is not None else z if z is not None and not norm_before_gate: x = x * F.silu(z) if group_size is None: rstd = 1 / torch.sqrt((x.square()).mean(dim=-1, keepdim=True) + eps) out = (x * rstd * weight) + bias if bias is not None else (x * rstd * weight) else: x_group = rearrange(x, "... (g d) -> ... g d", d=group_size) rstd = 1 / torch.sqrt((x_group.square()).mean(dim=-1, keepdim=True) + eps) out = rearrange(x_group * rstd, "... g d -> ... (g d)") * weight if bias is not None: out = out + bias if z is not None and norm_before_gate: out *= F.silu(z) return out.to(dtype) @triton.heuristics( { "HAS_BIAS": lambda args: args["B"] is not None, "HAS_Z": lambda args: args["Z"] is not None, } ) @triton.jit def layer_norm_fwd_kernel( X, # pointer to the input Y, # pointer to the output W, # pointer to the weights B, # pointer to the biases Z, # pointer to the other branch Mean, # pointer to the mean Rstd, # pointer to the 1/std stride_x_row, # how much to increase the pointer when moving by 1 row stride_y_row, stride_z_row, M, # number of rows in X N: tl.constexpr, # number of columns in X eps, # epsilon to avoid division by zero BLOCK_N: tl.constexpr, ROWS_PER_BLOCK: tl.constexpr, HAS_BIAS: tl.constexpr, HAS_Z: tl.constexpr, NORM_BEFORE_GATE: tl.constexpr, IS_RMS_NORM: tl.constexpr, ): # Map the program id to the starting row of X and Y it should compute. row_start = tl.program_id(0) * ROWS_PER_BLOCK group = tl.program_id(1) # Create 2D tile: [ROWS_PER_BLOCK, BLOCK_N] rows = row_start + tl.arange(0, ROWS_PER_BLOCK) cols = tl.arange(0, BLOCK_N) # Compute offsets for 2D tile row_offsets = rows[:, None] * stride_x_row col_offsets = cols[None, :] + group * N # Base pointers X_base = X + row_offsets + col_offsets Y_base = Y + rows[:, None] * stride_y_row + col_offsets # Create mask for valid rows and columns row_mask = rows[:, None] < M col_mask = cols[None, :] < N mask = row_mask & col_mask # Load input data with 2D tile x = tl.load(X_base, mask=mask, other=0.0).to(tl.float32) if HAS_Z and not NORM_BEFORE_GATE: Z_base = Z + rows[:, None] * stride_z_row + col_offsets z = tl.load(Z_base, mask=mask, other=0.0).to(tl.float32) x *= z * tl.sigmoid(z) # Compute mean and variance per row (reduce along axis 1) if not IS_RMS_NORM: mean = tl.sum(x, axis=1) / N # Shape: [ROWS_PER_BLOCK] # Store mean for each row mean_offsets = group * M + rows mean_mask = rows < M tl.store(Mean + mean_offsets, mean, mask=mean_mask) # Broadcast mean back to 2D for subtraction xbar = tl.where(mask, x - mean[:, None], 0.0) var = tl.sum(xbar * xbar, axis=1) / N # Shape: [ROWS_PER_BLOCK] else: xbar = tl.where(mask, x, 0.0) var = tl.sum(xbar * xbar, axis=1) / N # Shape: [ROWS_PER_BLOCK] mean = 0.0 # Placeholder for RMS norm rstd = tl.rsqrt(var + eps) # Shape: [ROWS_PER_BLOCK] # Store rstd for each row rstd_offsets = group * M + rows rstd_mask = rows < M tl.store(Rstd + rstd_offsets, rstd, mask=rstd_mask) # Load weights and biases (broadcast across rows) w_offsets = cols + group * N w_mask = cols < N w = tl.load(W + w_offsets, mask=w_mask, other=0.0).to(tl.float32) if HAS_BIAS: b = tl.load(B + w_offsets, mask=w_mask, other=0.0).to(tl.float32) # Normalize and apply linear transformation if not IS_RMS_NORM: x_hat = (x - mean[:, None]) * rstd[:, None] else: x_hat = x * rstd[:, None] y = x_hat * w[None, :] + b[None, :] if HAS_BIAS else x_hat * w[None, :] if HAS_Z and NORM_BEFORE_GATE: Z_base = Z + rows[:, None] * stride_z_row + col_offsets z = tl.load(Z_base, mask=mask, other=0.0).to(tl.float32) y *= z * tl.sigmoid(z) # Write output tl.store(Y_base, y, mask=mask) @lru_cache def _get_sm_count(device: torch.device) -> int: """Get and cache the SM count for a given device.""" props = torch.cuda.get_device_properties(device) return props.multi_processor_count def calc_rows_per_block(M: int, device: torch.device) -> int: sm_count = _get_sm_count(device) rows_per_block = next_power_of_2(cdiv(M, 2 * sm_count)) rows_per_block = min(rows_per_block, 4) return rows_per_block def layer_norm_fwd( x: torch.Tensor, weight: torch.Tensor, bias: torch.Tensor, eps: float, z: torch.Tensor = None, out: torch.Tensor = None, group_size: int = None, norm_before_gate: bool = True, is_rms_norm: bool = False, ): M, N = x.shape if group_size is None: group_size = N assert N % group_size == 0 ngroups = N // group_size assert x.stride(-1) == 1 if z is not None: assert z.stride(-1) == 1 assert z.shape == (M, N) assert weight.shape == (N,) assert weight.stride(-1) == 1 if bias is not None: assert bias.stride(-1) == 1 assert bias.shape == (N,) # allocate output if out is not None: assert out.shape == x.shape else: out = torch.empty_like(x) assert out.stride(-1) == 1 mean = ( torch.empty((ngroups * M,), dtype=torch.float32, device=x.device) if not is_rms_norm else None ) rstd = torch.empty((ngroups * M,), dtype=torch.float32, device=x.device) # Less than 64KB per feature: enqueue fused kernel MAX_FUSED_SIZE = 65536 // x.element_size() BLOCK_N = min(MAX_FUSED_SIZE, triton.next_power_of_2(group_size)) if group_size > BLOCK_N: raise RuntimeError("This layer norm doesn't support feature dim >= 64KB.") # heuristics for number of warps num_warps = min(max(BLOCK_N // 256, 1), 8) # Calculate rows per block based on SM count rows_per_block = calc_rows_per_block(M, x.device) # Update grid to use rows_per_block grid = (cdiv(M, rows_per_block), ngroups) layer_norm_fwd_kernel[grid]( x, out, weight, bias, z, mean, rstd, x.stride(0), out.stride(0), z.stride(0) if z is not None else 0, M, group_size, eps, BLOCK_N=BLOCK_N, ROWS_PER_BLOCK=rows_per_block, NORM_BEFORE_GATE=norm_before_gate, IS_RMS_NORM=is_rms_norm, num_warps=num_warps, ) return out, mean, rstd class LayerNormFn(torch.autograd.Function): @input_guard @staticmethod def forward( ctx, x, weight, bias, z=None, eps=1e-6, group_size=None, norm_before_gate=True, is_rms_norm=False, ): """If z is not None, we do norm(x) * silu(z) if norm_before_gate, else norm(x * silu(z))""" x_shape_og = x.shape # reshape input data into 2D tensor x = x.reshape(-1, x.shape[-1]) if x.stride(-1) != 1: x = x.contiguous() if z is not None: assert z.shape == x_shape_og z = z.reshape(-1, z.shape[-1]) if z.stride(-1) != 1: z = z.contiguous() weight = weight.contiguous() if bias is not None: bias = bias.contiguous() y, mean, rstd = layer_norm_fwd( x, weight, bias, eps, z=z, group_size=group_size, norm_before_gate=norm_before_gate, is_rms_norm=is_rms_norm, ) ctx.save_for_backward(x, weight, bias, mean, rstd, z) ctx.x_shape_og = x_shape_og ctx.eps = eps ctx.group_size = group_size ctx.norm_before_gate = norm_before_gate ctx.is_rms_norm = is_rms_norm return y.reshape(x_shape_og) def layernorm_fn( x, weight, bias, z=None, eps=1e-6, group_size=None, norm_before_gate=True, is_rms_norm=False, ): return LayerNormFn.apply( x, weight, bias, z, eps, group_size, norm_before_gate, is_rms_norm ) def rmsnorm_fn( x, weight, bias, z=None, eps=1e-6, group_size=None, norm_before_gate=True ): return LayerNormFn.apply( x, weight, bias, z, eps, group_size, norm_before_gate, True ) class LayerNormGated(nn.Module): def __init__( self, hidden_size, eps: float = 1e-5, group_size: int | None = None, norm_before_gate: bool = True, device: torch.device | None = None, dtype: torch.dtype | None = None, ): """If group_size is not None, we do GroupNorm with each group having group_size elements. group_size=None is equivalent to group_size=hidden_size (i.e. there's only 1 group). """ factory_kwargs = {"device": device, "dtype": dtype} super().__init__() self.eps = eps self.weight = nn.Parameter(torch.empty(hidden_size, **factory_kwargs)) self.bias = nn.Parameter(torch.empty(hidden_size, **factory_kwargs)) self.group_size = group_size self.norm_before_gate = norm_before_gate self.reset_parameters() def reset_parameters(self): torch.nn.init.ones_(self.weight) torch.nn.init.zeros_(self.bias) def forward(self, x, z=None): """If z is not None, we do norm(x) * silu(z) if norm_before_gate, else norm(x * silu(z))""" return layernorm_fn( x, self.weight, self.bias, z=z, group_size=self.group_size, eps=self.eps, norm_before_gate=self.norm_before_gate, ) class RMSNormGated(nn.Module): def __init__( self, hidden_size, eps: float = 1e-5, group_size: int | None = None, norm_before_gate: bool = False, device: torch.device | None = None, dtype: torch.dtype | None = None, ): """If group_size is not None, we do GroupNorm with each group having group_size elements. group_size=None is equivalent to group_size=hidden_size (i.e. there's only 1 group). """ factory_kwargs = {"device": device, "dtype": dtype} super().__init__() self.eps = eps self.weight = nn.Parameter(torch.empty(hidden_size, **factory_kwargs)) self.register_parameter("bias", None) self.group_size = group_size self.norm_before_gate = norm_before_gate self.reset_parameters() def reset_parameters(self): torch.nn.init.ones_(self.weight) def forward(self, x, z=None): """If z is not None, we do norm(x) * silu(z) if norm_before_gate, else norm(x * silu(z))""" return rmsnorm_fn( x, self.weight, self.bias, z=z, eps=self.eps, group_size=self.group_size, norm_before_gate=self.norm_before_gate, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/fused_recurrent.py
vllm/model_executor/layers/fla/ops/fused_recurrent.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .op import exp @triton.heuristics( { "USE_INITIAL_STATE": lambda args: args["h0"] is not None, "IS_VARLEN": lambda args: args["cu_seqlens"] is not None, "IS_CONTINUOUS_BATCHING": lambda args: args["ssm_state_indices"] is not None, "IS_SPEC_DECODING": lambda args: args["num_accepted_tokens"] is not None, } ) @triton.jit(do_not_specialize=["N", "T"]) def fused_recurrent_gated_delta_rule_fwd_kernel( q, k, v, g, beta, o, h0, ht, cu_seqlens, ssm_state_indices, num_accepted_tokens, scale, N: tl.int64, # num of sequences T: tl.int64, # num of tokens B: tl.constexpr, H: tl.constexpr, HV: tl.constexpr, K: tl.constexpr, V: tl.constexpr, BK: tl.constexpr, BV: tl.constexpr, stride_init_state_token: tl.constexpr, stride_final_state_token: tl.constexpr, stride_indices_seq: tl.constexpr, stride_indices_tok: tl.constexpr, USE_INITIAL_STATE: tl.constexpr, # whether to use initial state INPLACE_FINAL_STATE: tl.constexpr, # whether to store final state inplace IS_BETA_HEADWISE: tl.constexpr, # whether beta is headwise vector or scalar, USE_QK_L2NORM_IN_KERNEL: tl.constexpr, IS_VARLEN: tl.constexpr, IS_CONTINUOUS_BATCHING: tl.constexpr, IS_SPEC_DECODING: tl.constexpr, IS_KDA: tl.constexpr, ): i_k, i_v, i_nh = tl.program_id(0), tl.program_id(1), tl.program_id(2) i_n, i_hv = i_nh // HV, i_nh % HV i_h = i_hv // (HV // H) if IS_VARLEN: bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int64), tl.load(cu_seqlens + i_n + 1).to(tl.int64), ) all = T T = eos - bos else: bos, eos = i_n * T, i_n * T + T all = B * T if T == 0: # no tokens to process for this sequence return o_k = i_k * BK + tl.arange(0, BK) o_v = i_v * BV + tl.arange(0, BV) p_q = q + (bos * H + i_h) * K + o_k p_k = k + (bos * H + i_h) * K + o_k p_v = v + (bos * HV + i_hv) * V + o_v if IS_BETA_HEADWISE: p_beta = beta + (bos * HV + i_hv) * V + o_v else: p_beta = beta + bos * HV + i_hv if not IS_KDA: p_g = g + bos * HV + i_hv else: p_gk = g + (bos * HV + i_hv) * K + o_k p_o = o + ((i_k * all + bos) * HV + i_hv) * V + o_v mask_k = o_k < K mask_v = o_v < V mask_h = mask_k[:, None] & mask_v[None, :] b_h = tl.zeros([BK, BV], dtype=tl.float32) if USE_INITIAL_STATE: if IS_CONTINUOUS_BATCHING: if IS_SPEC_DECODING: i_t = tl.load(num_accepted_tokens + i_n).to(tl.int64) - 1 else: i_t = 0 p_h0 = ( h0 + tl.load(ssm_state_indices + i_n * stride_indices_seq + i_t).to( tl.int64 ) * stride_init_state_token ) else: p_h0 = h0 + bos * HV * K * V p_h0 = p_h0 + i_hv * K * V + o_k[:, None] * V + o_v[None, :] b_h += tl.load(p_h0, mask=mask_h, other=0).to(tl.float32) for i_t in range(0, T): b_q = tl.load(p_q, mask=mask_k, other=0).to(tl.float32) b_k = tl.load(p_k, mask=mask_k, other=0).to(tl.float32) b_v = tl.load(p_v, mask=mask_v, other=0).to(tl.float32) if USE_QK_L2NORM_IN_KERNEL: b_q = b_q / tl.sqrt(tl.sum(b_q * b_q) + 1e-6) b_k = b_k / tl.sqrt(tl.sum(b_k * b_k) + 1e-6) b_q = b_q * scale # [BK, BV] if not IS_KDA: b_g = tl.load(p_g).to(tl.float32) b_h *= exp(b_g) else: b_gk = tl.load(p_gk).to(tl.float32) b_h *= exp(b_gk[:, None]) # [BV] b_v -= tl.sum(b_h * b_k[:, None], 0) if IS_BETA_HEADWISE: b_beta = tl.load(p_beta, mask=mask_v, other=0).to(tl.float32) else: b_beta = tl.load(p_beta).to(tl.float32) b_v *= b_beta # [BK, BV] b_h += b_k[:, None] * b_v[None, :] # [BV] b_o = tl.sum(b_h * b_q[:, None], 0) tl.store(p_o, b_o.to(p_o.dtype.element_ty), mask=mask_v) # keep the states for multi-query tokens if INPLACE_FINAL_STATE: p_ht = ( ht + tl.load(ssm_state_indices + i_n * stride_indices_seq + i_t).to( tl.int64 ) * stride_final_state_token ) else: p_ht = ht + (bos + i_t) * stride_final_state_token p_ht = p_ht + i_hv * K * V + o_k[:, None] * V + o_v[None, :] tl.store(p_ht, b_h.to(p_ht.dtype.element_ty), mask=mask_h) p_q += H * K p_k += H * K p_o += HV * V p_v += HV * V if not IS_KDA: p_g += HV else: p_gk += HV * K p_beta += HV * (V if IS_BETA_HEADWISE else 1) def fused_recurrent_gated_delta_rule_fwd( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float, initial_state: torch.Tensor, inplace_final_state: bool = True, cu_seqlens: torch.LongTensor | None = None, ssm_state_indices: torch.Tensor | None = None, num_accepted_tokens: torch.Tensor | None = None, use_qk_l2norm_in_kernel: bool = False, ) -> tuple[torch.Tensor, torch.Tensor]: B, T, H, K, V = *k.shape, v.shape[-1] HV = v.shape[2] N = B if cu_seqlens is None else len(cu_seqlens) - 1 BK, BV = triton.next_power_of_2(K), min(triton.next_power_of_2(V), 8) NK, NV = triton.cdiv(K, BK), triton.cdiv(V, BV) assert NK == 1, "NK > 1 is not supported yet" num_stages = 3 num_warps = 1 o = q.new_empty(NK, *v.shape) if inplace_final_state: final_state = initial_state else: final_state = q.new_empty(T, HV, K, V, dtype=initial_state.dtype) stride_init_state_token = initial_state.stride(0) stride_final_state_token = final_state.stride(0) if ssm_state_indices is None: stride_indices_seq, stride_indices_tok = 1, 1 elif ssm_state_indices.ndim == 1: stride_indices_seq, stride_indices_tok = ssm_state_indices.stride(0), 1 else: stride_indices_seq, stride_indices_tok = ssm_state_indices.stride() grid = (NK, NV, N * HV) fused_recurrent_gated_delta_rule_fwd_kernel[grid]( q=q, k=k, v=v, g=g, beta=beta, o=o, h0=initial_state, ht=final_state, cu_seqlens=cu_seqlens, ssm_state_indices=ssm_state_indices, num_accepted_tokens=num_accepted_tokens, scale=scale, N=N, T=T, B=B, H=H, HV=HV, K=K, V=V, BK=BK, BV=BV, stride_init_state_token=stride_init_state_token, stride_final_state_token=stride_final_state_token, stride_indices_seq=stride_indices_seq, stride_indices_tok=stride_indices_tok, IS_BETA_HEADWISE=beta.ndim == v.ndim, USE_QK_L2NORM_IN_KERNEL=use_qk_l2norm_in_kernel, INPLACE_FINAL_STATE=inplace_final_state, IS_KDA=False, num_warps=num_warps, num_stages=num_stages, ) o = o.squeeze(0) return o, final_state class FusedRecurrentFunction(torch.autograd.Function): @staticmethod def forward( ctx, q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float, initial_state: torch.Tensor, inplace_final_state: bool = True, cu_seqlens: torch.LongTensor | None = None, ssm_state_indices: torch.Tensor | None = None, num_accepted_tokens: torch.Tensor | None = None, use_qk_l2norm_in_kernel: bool = False, ): o, final_state = fused_recurrent_gated_delta_rule_fwd( q=q.contiguous(), k=k.contiguous(), v=v.contiguous(), g=g.contiguous(), beta=beta.contiguous(), scale=scale, initial_state=initial_state, inplace_final_state=inplace_final_state, cu_seqlens=cu_seqlens, ssm_state_indices=ssm_state_indices, num_accepted_tokens=num_accepted_tokens, use_qk_l2norm_in_kernel=use_qk_l2norm_in_kernel, ) return o, final_state def fused_recurrent_gated_delta_rule( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor = None, scale: float = None, initial_state: torch.Tensor = None, inplace_final_state: bool = True, cu_seqlens: torch.LongTensor | None = None, ssm_state_indices: torch.Tensor | None = None, num_accepted_tokens: torch.Tensor | None = None, use_qk_l2norm_in_kernel: bool = False, ) -> tuple[torch.Tensor, torch.Tensor]: r""" Args: q (torch.Tensor): queries of shape `[B, T, H, K]`. k (torch.Tensor): keys of shape `[B, T, H, K]`. v (torch.Tensor): values of shape `[B, T, HV, V]`. GVA is applied if `HV > H`. g (torch.Tensor): g (decays) of shape `[B, T, HV]`. beta (torch.Tensor): betas of shape `[B, T, HV]`. scale (Optional[int]): Scale factor for the RetNet attention scores. If not provided, it will default to `1 / sqrt(K)`. Default: `None`. initial_state (Optional[torch.Tensor]): Initial state of shape `[N, HV, K, V]` for `N` input sequences. For equal-length input sequences, `N` equals the batch size `B`. Default: `None`. inplace_final_state: bool: Whether to store the final state in-place to save memory. Default: `True`. cu_seqlens (torch.LongTensor): Cumulative sequence lengths of shape `[N+1]` used for variable-length training, consistent with the FlashAttention API. ssm_state_indices (Optional[torch.Tensor]): Indices to map the input sequences to the initial/final states. num_accepted_tokens (Optional[torch.Tensor]): Number of accepted tokens for each sequence during decoding. Returns: o (torch.Tensor): Outputs of shape `[B, T, HV, V]`. final_state (torch.Tensor): Final state of shape `[N, HV, K, V]`. Examples:: >>> import torch >>> import torch.nn.functional as F >>> from einops import rearrange >>> from fla.ops.gated_delta_rule import fused_recurrent_gated_delta_rule # inputs with equal lengths >>> B, T, H, HV, K, V = 4, 2048, 4, 8, 512, 512 >>> q = torch.randn(B, T, H, K, device='cuda') >>> k = F.normalize(torch.randn(B, T, H, K, device='cuda'), p=2, dim=-1) >>> v = torch.randn(B, T, HV, V, device='cuda') >>> g = F.logsigmoid(torch.rand(B, T, HV, device='cuda')) >>> beta = torch.rand(B, T, HV, device='cuda').sigmoid() >>> h0 = torch.randn(B, HV, K, V, device='cuda') >>> o, ht = fused_gated_recurrent_delta_rule( q, k, v, g, beta, initial_state=h0, ) # for variable-length inputs, the batch size `B` is expected to be 1 and `cu_seqlens` is required >>> q, k, v, g, beta = map(lambda x: rearrange(x, 'b t ... -> 1 (b t) ...'), (q, k, v, g, beta)) # for a batch with 4 sequences, `cu_seqlens` with 5 start/end positions are expected >>> cu_seqlens = q.new_tensor([0, 2048, 4096, 6144, 8192], dtype=torch.long) >>> o_var, ht_var = fused_gated_recurrent_delta_rule( q, k, v, g, beta, initial_state=h0, cu_seqlens=cu_seqlens ) """ if cu_seqlens is not None and q.shape[0] != 1: raise ValueError( f"The batch size is expected to be 1 rather than {q.shape[0]} when using `cu_seqlens`." f"Please flatten variable-length inputs before processing." ) if scale is None: scale = k.shape[-1] ** -0.5 else: assert scale > 0, "scale must be positive" if beta is None: beta = torch.ones_like(q[..., 0]) o, final_state = FusedRecurrentFunction.apply( q, k, v, g, beta, scale, initial_state, inplace_final_state, cu_seqlens, ssm_state_indices, num_accepted_tokens, use_qk_l2norm_in_kernel, ) return o, final_state
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/solve_tril.py
vllm/model_executor/layers/fla/ops/solve_tril.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import os import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices from .op import make_tensor_descriptor from .utils import input_guard, is_amd, is_tma_supported FLA_TRIL_PRECISION = os.environ.get("FLA_TRIL_PRECISION", "ieee") ALLOWED_TRIL_PRECISIONS = ["ieee", "tf32"] if is_amd else ["ieee", "tf32", "tf32x3"] assert FLA_TRIL_PRECISION in ALLOWED_TRIL_PRECISIONS, ( f"FLA_TRIL_PRECISION must be one of {ALLOWED_TRIL_PRECISIONS}, but got {FLA_TRIL_PRECISION}" ) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({}, num_warps=num_warps, num_stages=num_stages) for num_warps in [1, 2, 4, 8] for num_stages in [2, 3, 4, 5] ], key=["BT"], ) @triton.jit(do_not_specialize=["T"]) def solve_tril_16x16_kernel( A, Ai, cu_seqlens, chunk_indices, T, H: tl.constexpr, BT: tl.constexpr, USE_TMA: tl.constexpr, IS_VARLEN: tl.constexpr, DOT_PRECISION: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T o_i = tl.arange(0, 16) m_A = o_i[:, None] > o_i[None, :] m_I = o_i[:, None] == o_i[None, :] A = A + (bos * H + i_h) * BT Ai = Ai + (bos * H + i_h) * 16 offset = (i_t * 16) % BT if not USE_TMA: p_A = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * 16, offset), (16, 16), (1, 0) ) # [16, 16] b_A = tl.load(p_A, boundary_check=(0, 1)).to(tl.float32) else: desc = make_tensor_descriptor(A, [T, BT], [H * BT, 1], [16, 16]) desc_o = make_tensor_descriptor(Ai, [T, 16], [H * 16, 1], [16, 16]) b_A = desc.load([i_t * 16, offset]).to(tl.float32) b_A = -tl.where(m_A, b_A, 0) for i in range(2, min(16, T - i_t * 16)): # [16] b_a = -tl.load(A + (i_t * 16 + i) * H * BT + o_i + offset) b_a = b_a + tl.sum(b_a[:, None] * b_A, 0) b_A = tl.where((o_i == i)[:, None], b_a, b_A) b_A += m_I if not USE_TMA: p_Ai = tl.make_block_ptr( Ai, (T, 16), (H * 16, 1), (i_t * 16, 0), (16, 16), (1, 0) ) tl.store( p_Ai, b_A.to(p_Ai.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) else: desc_o.store([i_t * 16, 0], b_A.to(desc_o.dtype, fp_downcast_rounding="rtne")) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({}, num_warps=num_warps, num_stages=num_stages) for num_warps in [1, 2, 4, 8] for num_stages in [2, 3, 4, 5] ], key=["H", "BT", "IS_VARLEN"], ) @triton.jit(do_not_specialize=["T"]) def merge_16x16_to_32x32_inverse_kernel( A, Ai, cu_seqlens, chunk_indices, T, H: tl.constexpr, BT: tl.constexpr, USE_TMA: tl.constexpr, IS_VARLEN: tl.constexpr, DOT_PRECISION: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T o_i = tl.arange(0, 16) m_A = o_i[:, None] > o_i[None, :] m_I = o_i[:, None] == o_i[None, :] A += (bos * H + i_h) * BT Ai += (bos * H + i_h) * BT if not USE_TMA: p_A_11 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT, 0), (16, 16), (1, 0) ) p_A_22 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 16, 16), (16, 16), (1, 0) ) b_Ai_11 = tl.load(p_A_11, boundary_check=(0, 1)).to(tl.float32) b_Ai_22 = tl.load(p_A_22, boundary_check=(0, 1)).to(tl.float32) else: desc = make_tensor_descriptor(A, [T, BT], [H * BT, 1], [16, 16]) desc_o = make_tensor_descriptor(Ai, [T, BT], [H * BT, 1], [16, 16]) b_Ai_11 = desc.load([i_t * BT + 0, 0]).to(tl.float32) b_Ai_22 = desc.load([i_t * BT + 16, 16]).to(tl.float32) # [16, 16] b_Ai_11 = -tl.where(m_A, b_Ai_11, 0) b_Ai_22 = -tl.where(m_A, b_Ai_22, 0) for i in range(2, min(16, T - i_t * BT)): b_a_11 = -tl.load(A + (i_t * BT + i) * H * BT + o_i) b_a_11 += tl.sum(b_a_11[:, None] * b_Ai_11, 0) b_Ai_11 = tl.where((o_i == i)[:, None], b_a_11, b_Ai_11) for i in range(16 + 2, min(32, T - i_t * BT)): b_a_22 = -tl.load(A + (i_t * BT + i) * H * BT + o_i + 16) b_a_22 += tl.sum(b_a_22[:, None] * b_Ai_22, 0) b_Ai_22 = tl.where((o_i == i - 16)[:, None], b_a_22, b_Ai_22) b_Ai_11 += m_I b_Ai_22 += m_I if not USE_TMA: p_A_21 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 16, 0), (16, 16), (1, 0) ) b_A_21 = tl.load(p_A_21, boundary_check=(0, 1)).to(tl.float32) else: b_A_21 = desc.load([i_t * BT + 16, 0]).to(tl.float32) b_Ai_21 = -tl.dot( tl.dot(b_Ai_22, b_A_21, input_precision=DOT_PRECISION), b_Ai_11, input_precision=DOT_PRECISION, ) if not USE_TMA: p_Ai_11 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT, 0), (16, 16), (1, 0) ) p_Ai_21 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 16, 0), (16, 16), (1, 0) ) p_Ai_22 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 16, 16), (16, 16), (1, 0) ) tl.store( p_Ai_11, b_Ai_11.to(p_Ai_11.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_22, b_Ai_22.to(p_Ai_22.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_21, b_Ai_21.to(p_Ai_21.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) else: desc_o.store( [i_t * BT + 0, 0], b_Ai_11.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 16, 0], b_Ai_21.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 16, 16], b_Ai_22.to(desc_o.dtype, fp_downcast_rounding="rtne") ) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({}, num_warps=num_warps, num_stages=num_stages) for num_warps in [2, 4, 8] for num_stages in [2, 3, 4, 5] ], key=["H", "BT", "IS_VARLEN"], ) @triton.jit(do_not_specialize=["T"]) def merge_16x16_to_64x64_inverse_kernel( A, Ai, cu_seqlens, chunk_indices, T, H: tl.constexpr, BT: tl.constexpr, USE_TMA: tl.constexpr, IS_VARLEN: tl.constexpr, DOT_PRECISION: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T o_i = tl.arange(0, 16) m_A = o_i[:, None] > o_i[None, :] m_I = o_i[:, None] == o_i[None, :] A += (bos * H + i_h) * BT Ai += (bos * H + i_h) * BT if not USE_TMA: p_A_11 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT, 0), (16, 16), (1, 0) ) p_A_22 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 16, 16), (16, 16), (1, 0) ) p_A_33 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 32, 32), (16, 16), (1, 0) ) p_A_44 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 48, 48), (16, 16), (1, 0) ) b_Ai_11 = tl.load(p_A_11, boundary_check=(0, 1)).to(tl.float32) b_Ai_22 = tl.load(p_A_22, boundary_check=(0, 1)).to(tl.float32) b_Ai_33 = tl.load(p_A_33, boundary_check=(0, 1)).to(tl.float32) b_Ai_44 = tl.load(p_A_44, boundary_check=(0, 1)).to(tl.float32) else: desc = make_tensor_descriptor(A, [T, BT], [H * BT, 1], [16, 16]) desc_o = make_tensor_descriptor(Ai, [T, BT], [H * BT, 1], [16, 16]) b_Ai_11 = desc.load([i_t * BT + 0, 0]).to(tl.float32) b_Ai_22 = desc.load([i_t * BT + 16, 16]).to(tl.float32) b_Ai_33 = desc.load([i_t * BT + 32, 32]).to(tl.float32) b_Ai_44 = desc.load([i_t * BT + 48, 48]).to(tl.float32) # [16, 16] b_Ai_11 = -tl.where(m_A, b_Ai_11, 0) b_Ai_22 = -tl.where(m_A, b_Ai_22, 0) b_Ai_33 = -tl.where(m_A, b_Ai_33, 0) b_Ai_44 = -tl.where(m_A, b_Ai_44, 0) for i in range(2, min(16, T - i_t * BT)): b_a_11 = -tl.load(A + (i_t * BT + i) * H * BT + o_i) b_a_11 += tl.sum(b_a_11[:, None] * b_Ai_11, 0) b_Ai_11 = tl.where((o_i == i)[:, None], b_a_11, b_Ai_11) for i in range(16 + 2, min(32, T - i_t * BT)): b_a_22 = -tl.load(A + (i_t * BT + i) * H * BT + o_i + 16) b_a_22 += tl.sum(b_a_22[:, None] * b_Ai_22, 0) b_Ai_22 = tl.where((o_i == i - 16)[:, None], b_a_22, b_Ai_22) for i in range(32 + 2, min(48, T - i_t * BT)): b_a_33 = -tl.load(A + (i_t * BT + i) * H * BT + o_i + 32) b_a_33 += tl.sum(b_a_33[:, None] * b_Ai_33, 0) b_Ai_33 = tl.where((o_i == i - 32)[:, None], b_a_33, b_Ai_33) for i in range(48 + 2, min(64, T - i_t * BT)): b_a_44 = -tl.load(A + (i_t * BT + i) * H * BT + o_i + 48) b_a_44 += tl.sum(b_a_44[:, None] * b_Ai_44, 0) b_Ai_44 = tl.where((o_i == i - 48)[:, None], b_a_44, b_Ai_44) b_Ai_11 += m_I b_Ai_22 += m_I b_Ai_33 += m_I b_Ai_44 += m_I if not USE_TMA: p_A_21 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 16, 0), (16, 16), (1, 0) ) p_A_31 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 32, 0), (16, 16), (1, 0) ) p_A_32 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 32, 16), (16, 16), (1, 0) ) p_A_41 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 48, 0), (16, 16), (1, 0) ) p_A_42 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 48, 16), (16, 16), (1, 0) ) p_A_43 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 48, 32), (16, 16), (1, 0) ) b_A_21 = tl.load(p_A_21, boundary_check=(0, 1)).to(tl.float32) b_A_31 = tl.load(p_A_31, boundary_check=(0, 1)).to(tl.float32) b_A_32 = tl.load(p_A_32, boundary_check=(0, 1)).to(tl.float32) b_A_41 = tl.load(p_A_41, boundary_check=(0, 1)).to(tl.float32) b_A_42 = tl.load(p_A_42, boundary_check=(0, 1)).to(tl.float32) b_A_43 = tl.load(p_A_43, boundary_check=(0, 1)).to(tl.float32) else: b_A_21 = desc.load([i_t * BT + 16, 0]).to(tl.float32) b_A_31 = desc.load([i_t * BT + 32, 0]).to(tl.float32) b_A_32 = desc.load([i_t * BT + 32, 16]).to(tl.float32) b_A_41 = desc.load([i_t * BT + 48, 0]).to(tl.float32) b_A_42 = desc.load([i_t * BT + 48, 16]).to(tl.float32) b_A_43 = desc.load([i_t * BT + 48, 32]).to(tl.float32) b_Ai_21 = -tl.dot( tl.dot(b_Ai_22, b_A_21, input_precision=DOT_PRECISION), b_Ai_11, input_precision=DOT_PRECISION, ) b_Ai_32 = -tl.dot( tl.dot(b_Ai_33, b_A_32, input_precision=DOT_PRECISION), b_Ai_22, input_precision=DOT_PRECISION, ) b_Ai_43 = -tl.dot( tl.dot(b_Ai_44, b_A_43, input_precision=DOT_PRECISION), b_Ai_33, input_precision=DOT_PRECISION, ) b_Ai_31 = -tl.dot( b_Ai_33, tl.dot(b_A_31, b_Ai_11, input_precision=DOT_PRECISION) + tl.dot(b_A_32, b_Ai_21, input_precision=DOT_PRECISION), input_precision=DOT_PRECISION, ) b_Ai_42 = -tl.dot( b_Ai_44, tl.dot(b_A_42, b_Ai_22, input_precision=DOT_PRECISION) + tl.dot(b_A_43, b_Ai_32, input_precision=DOT_PRECISION), input_precision=DOT_PRECISION, ) b_Ai_41 = -tl.dot( b_Ai_44, tl.dot(b_A_41, b_Ai_11, input_precision=DOT_PRECISION) + tl.dot(b_A_42, b_Ai_21, input_precision=DOT_PRECISION) + tl.dot(b_A_43, b_Ai_31, input_precision=DOT_PRECISION), input_precision=DOT_PRECISION, ) if not USE_TMA: p_Ai_11 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT, 0), (16, 16), (1, 0) ) p_Ai_22 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 16, 16), (16, 16), (1, 0) ) p_Ai_33 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 32, 32), (16, 16), (1, 0) ) p_Ai_44 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 48, 48), (16, 16), (1, 0) ) p_Ai_21 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 16, 0), (16, 16), (1, 0) ) p_Ai_31 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 32, 0), (16, 16), (1, 0) ) p_Ai_32 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 32, 16), (16, 16), (1, 0) ) p_Ai_41 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 48, 0), (16, 16), (1, 0) ) p_Ai_42 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 48, 16), (16, 16), (1, 0) ) p_Ai_43 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 48, 32), (16, 16), (1, 0) ) tl.store( p_Ai_11, b_Ai_11.to(p_Ai_11.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_22, b_Ai_22.to(p_Ai_22.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_33, b_Ai_33.to(p_Ai_33.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_44, b_Ai_44.to(p_Ai_44.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_21, b_Ai_21.to(p_Ai_21.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_31, b_Ai_31.to(p_Ai_31.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_32, b_Ai_32.to(p_Ai_32.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_41, b_Ai_41.to(p_Ai_41.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_42, b_Ai_42.to(p_Ai_42.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_43, b_Ai_43.to(p_Ai_43.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) else: desc_o.store( [i_t * BT + 0, 0], b_Ai_11.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 16, 16], b_Ai_22.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 32, 32], b_Ai_33.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 48, 48], b_Ai_44.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 16, 0], b_Ai_21.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 32, 0], b_Ai_31.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 32, 16], b_Ai_32.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 48, 0], b_Ai_41.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 48, 16], b_Ai_42.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 48, 32], b_Ai_43.to(desc_o.dtype, fp_downcast_rounding="rtne") ) @input_guard def solve_tril( A: torch.Tensor, cu_seqlens: torch.Tensor | None = None, output_dtype: torch.dtype = torch.float, ) -> torch.Tensor: """ Compute the inverse of the matrix I + A A should be strictly lower triangular, i.e., A.triu() == 0. Args: A (torch.Tensor): [B, T, H, BT], where BT should only be 16, 32, or 64. cu_seqlens (torch.Tensor): The cumulative sequence lengths of the input tensor. Default: `None`. output_dtype (torch.dtype): The dtype of the output tensor. Default: `torch.float`. If `None`, the output dtype will be the same as the input dtype. Returns: (I + A)^-1 with the same shape as A """ assert A.shape[-1] in [16, 32, 64] output_dtype = A.dtype if output_dtype is None else output_dtype B, T, H, BT = A.shape chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = len(chunk_indices) if cu_seqlens is not None else triton.cdiv(T, BT) Ai = torch.zeros_like(A, dtype=output_dtype) if BT == 16: merge_fn = solve_tril_16x16_kernel elif BT == 32: merge_fn = merge_16x16_to_32x32_inverse_kernel elif BT == 64: merge_fn = merge_16x16_to_64x64_inverse_kernel merge_fn[NT, B * H]( A=A, Ai=Ai, cu_seqlens=cu_seqlens, chunk_indices=chunk_indices, T=T, H=H, BT=BT, USE_TMA=is_tma_supported, DOT_PRECISION=FLA_TRIL_PRECISION, ) return Ai
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/__init__.py
vllm/model_executor/layers/fla/ops/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang from .chunk import chunk_gated_delta_rule from .fused_recurrent import fused_recurrent_gated_delta_rule from .layernorm_guard import RMSNormGated __all__ = [ "RMSNormGated", "chunk_gated_delta_rule", "fused_recurrent_gated_delta_rule", ]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/op.py
vllm/model_executor/layers/fla/ops/op.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang import os from vllm.triton_utils import tl, tldevice, triton from .utils import is_gather_supported if os.environ.get("FLA_USE_FAST_OPS", "0") == "1": exp = tldevice.fast_expf log = tldevice.fast_logf log2 = tldevice.fast_log2f else: exp = tl.exp log = tl.log log2 = tl.log2 if not is_gather_supported: @triton.jit def gather(src, index, axis, _builder=None): """ Gather operation that works when tl.gather is not supported. This is a fallback implementation that returns None. Just to make triton compiler happy. """ return None else: gather = tl.gather if hasattr(triton.language, "_experimental_make_tensor_descriptor"): # For Triton 3.3.x make_tensor_descriptor = triton.language._experimental_make_tensor_descriptor elif hasattr(triton.language, "make_tensor_descriptor"): # For Triton 3.4.x and later make_tensor_descriptor = triton.language.make_tensor_descriptor else: """ Fallback implementation when TMA is not supported. Returns None to indicate TMA descriptors are unavailable. Just make triton compiler happy. """ @triton.jit def make_tensor_descriptor( base, shape, strides, block_shape, _builder=None, ): return None
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/model_executor/layers/fla/ops/index.py
vllm/model_executor/layers/fla/ops/index.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import triton from .utils import tensor_cache @tensor_cache def prepare_lens(cu_seqlens: torch.LongTensor) -> torch.LongTensor: return cu_seqlens[1:] - cu_seqlens[:-1] @tensor_cache def prepare_chunk_indices( cu_seqlens: torch.LongTensor, chunk_size: int ) -> torch.LongTensor: indices = torch.cat( [ torch.arange(n) for n in triton.cdiv(prepare_lens(cu_seqlens), chunk_size).tolist() ] ) return torch.stack([indices.eq(0).cumsum(0) - 1, indices], 1).to(cu_seqlens) @tensor_cache def prepare_chunk_offsets( cu_seqlens: torch.LongTensor, chunk_size: int ) -> torch.LongTensor: return torch.cat( [cu_seqlens.new_tensor([0]), triton.cdiv(prepare_lens(cu_seqlens), chunk_size)] ).cumsum(-1)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/load.py
vllm/config/load.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import TYPE_CHECKING, Any from pydantic import Field, field_validator from pydantic.dataclasses import dataclass from vllm.config.utils import config from vllm.logger import init_logger from vllm.utils.hashing import safe_hash if TYPE_CHECKING: from vllm.model_executor.model_loader import LoadFormats from vllm.model_executor.model_loader.tensorizer import TensorizerConfig else: LoadFormats = Any TensorizerConfig = Any logger = init_logger(__name__) @config @dataclass class LoadConfig: """Configuration for loading the model weights.""" load_format: str | LoadFormats = "auto" """The format of the model weights to load:\n - "auto" will try to load the weights in the safetensors format and fall back to the pytorch bin format if safetensors format is not available.\n - "pt" will load the weights in the pytorch bin format.\n - "safetensors" will load the weights in the safetensors format.\n - "npcache" will load the weights in pytorch format and store a numpy cache to speed up the loading.\n - "dummy" will initialize the weights with random values, which is mainly for profiling.\n - "tensorizer" will use CoreWeave's tensorizer library for fast weight loading. See the Tensorize vLLM Model script in the Examples section for more information.\n - "runai_streamer" will load the Safetensors weights using Run:ai Model Streamer.\n - "runai_streamer_sharded" will load weights from pre-sharded checkpoint files using Run:ai Model Streamer.\n - "bitsandbytes" will load the weights using bitsandbytes quantization.\n - "sharded_state" will load weights from pre-sharded checkpoint files, supporting efficient loading of tensor-parallel models.\n - "gguf" will load weights from GGUF format files (details specified in https://github.com/ggml-org/ggml/blob/master/docs/gguf.md).\n - "mistral" will load weights from consolidated safetensors files used by Mistral models. - Other custom values can be supported via plugins.""" download_dir: str | None = None """Directory to download and load the weights, default to the default cache directory of Hugging Face.""" safetensors_load_strategy: str = "lazy" """Specifies the loading strategy for safetensors weights. - "lazy" (default): Weights are memory-mapped from the file. This enables on-demand loading and is highly efficient for models on local storage. - "eager": The entire file is read into CPU memory upfront before loading. This is recommended for models on network filesystems (e.g., Lustre, NFS) as it avoids inefficient random reads, significantly speeding up model initialization. However, it uses more CPU RAM. - "torchao": Weights are loaded in upfront and then reconstructed into torchao tensor subclasses. This is used when the checkpoint was quantized using torchao and saved using safetensors. Needs torchao >= 0.14.0 """ model_loader_extra_config: dict | TensorizerConfig = Field(default_factory=dict) """Extra config for model loader. This will be passed to the model loader corresponding to the chosen load_format.""" device: str | None = None """Device to which model weights will be loaded, default to device_config.device""" ignore_patterns: list[str] | str = Field(default_factory=lambda: ["original/**/*"]) """The list of patterns to ignore when loading the model. Default to "original/**/*" to avoid repeated loading of llama's checkpoints.""" use_tqdm_on_load: bool = True """Whether to enable tqdm for showing progress bar when loading model weights.""" pt_load_map_location: str | dict[str, str] = "cpu" """ pt_load_map_location: the map location for loading pytorch checkpoint, to support loading checkpoints can only be loaded on certain devices like "cuda", this is equivalent to {"": "cuda"}. Another supported format is mapping from different devices like from GPU 1 to GPU 0: {"cuda:1": "cuda:0"}. Note that when passed from command line, the strings in dictionary needs to be double quoted for json parsing. For more details, see original doc for `map_location` in https://pytorch.org/docs/stable/generated/torch.load.html """ def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # no factors to consider. # this config will not affect the computation graph. factors: list[Any] = [] hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str @field_validator("load_format", mode="after") def _lowercase_load_format(cls, load_format: str) -> str: return load_format.lower() @field_validator("ignore_patterns", mode="after") def _validate_ignore_patterns( cls, ignore_patterns: list[str] | str ) -> list[str] | str: if ignore_patterns != ["original/**/*"] and len(ignore_patterns) > 0: logger.info( "Ignoring the following patterns when downloading weights: %s", ignore_patterns, ) return ignore_patterns
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/pooler.py
vllm/config/pooler.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any, Literal from pydantic.dataclasses import dataclass from vllm.config.utils import config from vllm.logger import init_logger from vllm.utils.hashing import safe_hash logger = init_logger(__name__) PoolingTypeStr = Literal["LAST", "ALL", "CLS", "STEP", "MEAN"] @config @dataclass class PoolerConfig: """Controls the behavior of output pooling in pooling models.""" pooling_type: PoolingTypeStr | None = None """ The pooling method of the pooling model. This should be a key in [`vllm.model_executor.layers.pooler.PoolingType`][]. """ ## for embeddings models normalize: bool | None = None """ Whether to normalize the embeddings outputs. Defaults to True. """ dimensions: int | None = None """ Reduce the dimensions of embeddings if model support matryoshka representation. Defaults to None. """ enable_chunked_processing: bool | None = None """ Whether to enable chunked processing for long inputs that exceed the model's maximum position embeddings. When enabled, long inputs will be split into chunks, processed separately, and then aggregated using weighted averaging. This allows embedding models to handle arbitrarily long text without CUDA errors. Defaults to False. """ max_embed_len: int | None = None """ Maximum input length allowed for embedding generation. When set, allows inputs longer than max_embed_len to be accepted for embedding models. When an input exceeds max_embed_len, it will be handled according to the original max_model_len validation logic. Defaults to None (i.e. set to max_model_len). """ ## for classification models softmax: float | None = None """ softmax will be deprecated, please use use_activation instead. """ activation: float | None = None """ activation will be deprecated, please use use_activation instead. """ use_activation: bool | None = None """ Whether to apply activation function to the classification outputs. Defaults to True. """ logit_bias: float | None = None """ If provided, apply classification logit biases. Defaults to None. """ ## for reward models step_tag_id: int | None = None """ If set, only the score corresponding to the `step_tag_id` in the generated sentence should be returned. Otherwise, the scores for all tokens are returned. """ returned_token_ids: list[int] | None = None """ A list of indices for the vocabulary dimensions to be extracted, such as the token IDs of `good_token` and `bad_token` in the `math-shepherd-mistral-7b-prm` model. """ def __post_init__(self): # raise deprecated warning for softmax and activation self.use_activation = get_use_activation(self) def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # no factors to consider. # this config will not affect the computation graph. factors: list[Any] = [] hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str def get_use_activation(o: object): if softmax := getattr(o, "softmax", None) is not None: logger.warning_once( "softmax will be deprecated and will be removed in v0.15. " "Please use use_activation instead." ) return softmax if activation := getattr(o, "activation", None) is not None: logger.warning_once( "activation will be deprecated and will be removed in v0.15. " "Please use use_activation instead." ) return activation return getattr(o, "use_activation", None)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/profiler.py
vllm/config/profiler.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os from typing import Any, Literal from pydantic import Field, model_validator from pydantic.dataclasses import dataclass from typing_extensions import Self import vllm.envs as envs from vllm.config.utils import config from vllm.logger import init_logger from vllm.utils.hashing import safe_hash logger = init_logger(__name__) ProfilerKind = Literal["torch", "cuda"] @config @dataclass class ProfilerConfig: """Dataclass which contains profiler config for the engine.""" profiler: ProfilerKind | None = None """Which profiler to use. Defaults to None. Options are: - 'torch': Use PyTorch profiler.\n - 'cuda': Use CUDA profiler.""" torch_profiler_dir: str = "" """Directory to save torch profiler traces. Both AsyncLLM's CPU traces and worker's traces (CPU & GPU) will be saved under this directory. Note that it must be an absolute path.""" torch_profiler_with_stack: bool = True """If `True`, enables stack tracing in the torch profiler. Enabled by default.""" torch_profiler_with_flops: bool = False """If `True`, enables FLOPS counting in the torch profiler. Disabled by default.""" torch_profiler_use_gzip: bool = True """If `True`, saves torch profiler traces in gzip format. Enabled by default""" torch_profiler_dump_cuda_time_total: bool = True """If `True`, dumps total CUDA time in torch profiler traces. Enabled by default.""" torch_profiler_record_shapes: bool = False """If `True`, records tensor shapes in the torch profiler. Disabled by default.""" torch_profiler_with_memory: bool = False """If `True`, enables memory profiling in the torch profiler. Disabled by default.""" ignore_frontend: bool = False """If `True`, disables the front-end profiling of AsyncLLM when using the 'torch' profiler. This is needed to reduce overhead when using delay/limit options, since the front-end profiling does not track iterations and will capture the entire range. """ delay_iterations: int = Field(default=0, ge=0) """Number of engine iterations to skip before starting profiling. Defaults to 0, meaning profiling starts immediately after receiving /start_profile. """ max_iterations: int = Field(default=0, ge=0) """Maximum number of engine iterations to profile after starting profiling. Defaults to 0, meaning no limit. """ def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # no factors to consider. # this config will not affect the computation graph. factors: list[Any] = [] hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str def _get_from_env_if_set(self, field_name: str, env_var_name: str) -> None: """Get field from env var if set, with deprecation warning.""" if envs.is_set(env_var_name): value = getattr(envs, env_var_name) logger.warning_once( "Using %s environment variable is deprecated and will be removed in " "v0.14.0 or v1.0.0, whichever is soonest. Please use " "--profiler-config.%s command line argument or " "ProfilerConfig(%s=...) config field instead.", env_var_name, field_name, field_name, ) return value return None def _set_from_env_if_set( self, field_name: str, env_var_name: str, to_bool: bool = True, to_int: bool = False, ) -> None: """Set field from env var if set, with deprecation warning.""" value = self._get_from_env_if_set(field_name, env_var_name) if value is not None: if to_bool: value = value == "1" if to_int: value = int(value) setattr(self, field_name, value) @model_validator(mode="after") def _validate_profiler_config(self) -> Self: maybe_use_cuda_profiler = self._get_from_env_if_set( "profiler", "VLLM_TORCH_CUDA_PROFILE" ) if maybe_use_cuda_profiler is not None: self.profiler = "cuda" if maybe_use_cuda_profiler == "1" else None else: self._set_from_env_if_set( "torch_profiler_dir", "VLLM_TORCH_PROFILER_DIR", to_bool=False ) if self.torch_profiler_dir: self.profiler = "torch" self._set_from_env_if_set( "torch_profiler_record_shapes", "VLLM_TORCH_PROFILER_RECORD_SHAPES", ) self._set_from_env_if_set( "torch_profiler_with_memory", "VLLM_TORCH_PROFILER_WITH_PROFILE_MEMORY", ) self._set_from_env_if_set( "torch_profiler_with_stack", "VLLM_TORCH_PROFILER_WITH_STACK", ) self._set_from_env_if_set( "torch_profiler_with_flops", "VLLM_TORCH_PROFILER_WITH_FLOPS", ) self._set_from_env_if_set( "ignore_frontend", "VLLM_TORCH_PROFILER_DISABLE_ASYNC_LLM", ) self._set_from_env_if_set( "torch_profiler_use_gzip", "VLLM_TORCH_PROFILER_USE_GZIP", ) self._set_from_env_if_set( "torch_profiler_dump_cuda_time_total", "VLLM_TORCH_PROFILER_DUMP_CUDA_TIME_TOTAL", ) self._set_from_env_if_set( "delay_iterations", "VLLM_PROFILER_DELAY_ITERS", to_bool=False, to_int=True ) self._set_from_env_if_set( "max_iterations", "VLLM_PROFILER_MAX_ITERS", to_bool=False, to_int=True ) has_delay_or_limit = self.delay_iterations > 0 or self.max_iterations > 0 if self.profiler == "torch" and has_delay_or_limit and not self.ignore_frontend: logger.warning_once( "Using 'torch' profiler with delay_iterations or max_iterations " "while ignore_frontend is False may result in high overhead." ) profiler_dir = self.torch_profiler_dir if profiler_dir and self.profiler != "torch": raise ValueError( "torch_profiler_dir is only applicable when profiler is set to 'torch'" ) if self.profiler == "torch" and not profiler_dir: raise ValueError("torch_profiler_dir must be set when profiler is 'torch'") if profiler_dir: is_gs_path = ( profiler_dir.startswith("gs://") and profiler_dir[5:] and profiler_dir[5] != "/" ) if not is_gs_path: self.torch_profiler_dir = os.path.abspath( os.path.expanduser(profiler_dir) ) return self
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/model.py
vllm/config/model.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import warnings from collections.abc import Callable from dataclasses import InitVar, field from functools import cached_property from typing import TYPE_CHECKING, Any, Literal, cast, get_args import torch from pydantic import ConfigDict, Field, field_validator, model_validator from pydantic.dataclasses import dataclass import vllm.envs as envs from vllm.attention.backends.registry import AttentionBackendEnum from vllm.config.model_arch import ( ModelArchitectureConfig, ) from vllm.config.multimodal import MMCacheType, MMEncoderTPMode, MultiModalConfig from vllm.config.pooler import PoolerConfig from vllm.config.scheduler import RunnerType from vllm.config.utils import config, getattr_iter from vllm.logger import init_logger from vllm.platforms import current_platform from vllm.transformers_utils.config import ( ConfigFormat, get_config, get_hf_image_processor_config, get_hf_text_config, get_pooling_config, get_sentence_transformer_tokenizer_config, is_encoder_decoder, is_rope_parameters_nested, try_get_dense_modules, try_get_generation_config, try_get_tokenizer_config, uses_mrope, uses_xdrope_dim, ) from vllm.transformers_utils.gguf_utils import ( is_gguf, is_remote_gguf, maybe_patch_hf_config_from_gguf, split_remote_gguf, ) from vllm.transformers_utils.model_arch_config_convertor import ( MODEL_ARCH_CONFIG_CONVERTORS, ModelArchConfigConvertorBase, ) from vllm.transformers_utils.runai_utils import ObjectStorageModel, is_runai_obj_uri from vllm.transformers_utils.utils import maybe_model_redirect from vllm.utils.import_utils import LazyLoader if TYPE_CHECKING: from transformers import PretrainedConfig import vllm.model_executor.layers.quantization as me_quant import vllm.model_executor.models as me_models from vllm.config.load import LoadConfig from vllm.config.parallel import ParallelConfig from vllm.model_executor.layers.quantization import QuantizationMethods from vllm.v1.sample.logits_processor import LogitsProcessor else: PretrainedConfig = Any me_quant = LazyLoader( "model_executor", globals(), "vllm.model_executor.layers.quantization" ) me_models = LazyLoader("model_executor", globals(), "vllm.model_executor.models") LoadConfig = Any ParallelConfig = Any QuantizationMethods = Any LogitsProcessor = Any logger = init_logger(__name__) RunnerOption = Literal["auto", RunnerType] ConvertType = Literal["none", "embed", "classify", "reward", "mm_encoder_only"] ConvertOption = Literal["auto", ConvertType] TokenizerMode = Literal["auto", "hf", "slow", "mistral", "deepseek_v32"] ModelDType = Literal["auto", "half", "float16", "bfloat16", "float", "float32"] LogprobsMode = Literal[ "raw_logits", "raw_logprobs", "processed_logits", "processed_logprobs" ] HfOverrides = dict[str, Any] | Callable[[PretrainedConfig], PretrainedConfig] ModelImpl = Literal["auto", "vllm", "transformers", "terratorch"] LayerBlockType = Literal["attention", "linear_attention", "mamba"] _RUNNER_CONVERTS: dict[RunnerType, list[ConvertType]] = { "generate": [], "pooling": ["embed", "classify", "reward"], "draft": [], } AttnTypeStr = Literal[ "decoder", "encoder", "encoder_only", "encoder_decoder", "attention_free", "hybrid" ] @config @dataclass(config=ConfigDict(arbitrary_types_allowed=True)) class ModelConfig: """Configuration for the model.""" model: str = "Qwen/Qwen3-0.6B" """Name or path of the Hugging Face model to use. It is also used as the content for `model_name` tag in metrics output when `served_model_name` is not specified.""" runner: RunnerOption = "auto" """The type of model runner to use. Each vLLM instance only supports one model runner, even if the same model can be used for multiple types.""" convert: ConvertOption = "auto" """Convert the model using adapters defined in [vllm.model_executor.models.adapters][]. The most common use case is to adapt a text generation model to be used for pooling tasks.""" tokenizer: str = Field(default=None) """Name or path of the Hugging Face tokenizer to use. If unspecified, model name or path will be used.""" tokenizer_mode: TokenizerMode | str = "auto" """Tokenizer mode:\n - "auto" will use the tokenizer from `mistral_common` for Mistral models if available, otherwise it will use the "hf" tokenizer.\n - "hf" will use the fast tokenizer if available.\n - "slow" will always use the slow tokenizer.\n - "mistral" will always use the tokenizer from `mistral_common`.\n - "deepseek_v32" will always use the tokenizer from `deepseek_v32`.\n - Other custom values can be supported via plugins.""" trust_remote_code: bool = False """Trust remote code (e.g., from HuggingFace) when downloading the model and tokenizer.""" dtype: ModelDType | torch.dtype = "auto" """Data type for model weights and activations:\n - "auto" will use FP16 precision for FP32 and FP16 models, and BF16 precision for BF16 models.\n - "half" for FP16. Recommended for AWQ quantization.\n - "float16" is the same as "half".\n - "bfloat16" for a balance between precision and range.\n - "float" is shorthand for FP32 precision.\n - "float32" for FP32 precision.""" seed: int = 0 """Random seed for reproducibility. We must set the global seed because otherwise, different tensor parallel workers would sample different tokens, leading to inconsistent results.""" hf_config: PretrainedConfig = field(init=False) """The Hugging Face config of the model.""" hf_text_config: PretrainedConfig = field(init=False) """The Hugging Face config of the text model (same as hf_config for text models).""" hf_config_path: str | None = None """Name or path of the Hugging Face config to use. If unspecified, model name or path will be used.""" allowed_local_media_path: str = "" """Allowing API requests to read local images or videos from directories specified by the server file system. This is a security risk. Should only be enabled in trusted environments.""" allowed_media_domains: list[str] | None = None """If set, only media URLs that belong to this domain can be used for multi-modal inputs. """ revision: str | None = None """The specific model version to use. It can be a branch name, a tag name, or a commit id. If unspecified, will use the default version.""" code_revision: str | None = None """The specific revision to use for the model code on the Hugging Face Hub. It can be a branch name, a tag name, or a commit id. If unspecified, will use the default version.""" tokenizer_revision: str | None = None """The specific revision to use for the tokenizer on the Hugging Face Hub. It can be a branch name, a tag name, or a commit id. If unspecified, will use the default version.""" max_model_len: int = Field(default=None, ge=-1) """Model context length (prompt and output). If unspecified, will be automatically derived from the model config. When passing via `--max-model-len`, supports k/m/g/K/M/G in human-readable format. Examples:\n - 1k -> 1000\n - 1K -> 1024\n - 25.6k -> 25,600\n - -1 or 'auto' -> Automatically choose the maximum model length that fits in GPU memory. This will use the model's maximum context length if it fits, otherwise it will find the largest length that can be accommodated.""" spec_target_max_model_len: int | None = None """Specify the maximum length for spec decoding draft models.""" quantization: QuantizationMethods | str | None = None """Method used to quantize the weights. If `None`, we first check the `quantization_config` attribute in the model config file. If that is `None`, we assume the model weights are not quantized and use `dtype` to determine the data type of the weights.""" enforce_eager: bool = False """Whether to always use eager-mode PyTorch. If True, we will disable CUDA graph and always execute the model in eager mode. If False, we will use CUDA graph and eager execution in hybrid for maximal performance and flexibility.""" max_logprobs: int = 20 """Maximum number of log probabilities to return when `logprobs` is specified in `SamplingParams`. The default value comes the default for the OpenAI Chat Completions API. -1 means no cap, i.e. all (output_length * vocab_size) logprobs are allowed to be returned and it may cause OOM.""" logprobs_mode: LogprobsMode = "raw_logprobs" """Indicates the content returned in the logprobs and prompt_logprobs. Supported mode: 1) raw_logprobs, 2) processed_logprobs, 3) raw_logits, 4) processed_logits. Raw means the values before applying any logit processors, like bad words. Processed means the values after applying all processors, including temperature and top_k/top_p. """ disable_sliding_window: bool = False """Whether to disable sliding window. If True, we will disable the sliding window functionality of the model, capping to sliding window size. If the model does not support sliding window, this argument is ignored.""" disable_cascade_attn: bool = False """Disable cascade attention for V1. While cascade attention does not change the mathematical correctness, disabling it could be useful for preventing potential numerical issues. Note that even if this is set to False, cascade attention will be only used when the heuristic tells that it's beneficial.""" skip_tokenizer_init: bool = False """Skip initialization of tokenizer and detokenizer. Expects valid `prompt_token_ids` and `None` for prompt from the input. The generated output will contain token ids.""" enable_prompt_embeds: bool = False """If `True`, enables passing text embeddings as inputs via the `prompt_embeds` key. WARNING: The vLLM engine may crash if incorrect shape of embeddings is passed. Only enable this flag for trusted users!""" served_model_name: str | list[str] | None = None """The model name(s) used in the API. If multiple names are provided, the server will respond to any of the provided names. The model name in the model field of a response will be the first name in this list. If not specified, the model name will be the same as the `--model` argument. Noted that this name(s) will also be used in `model_name` tag content of prometheus metrics, if multiple names provided, metrics tag will take the first one.""" config_format: str | ConfigFormat = "auto" """The format of the model config to load:\n - "auto" will try to load the config in hf format if available after trying to load in mistral format.\n - "hf" will load the config in hf format.\n - "mistral" will load the config in mistral format.""" hf_token: bool | str | None = None """The token to use as HTTP bearer authorization for remote files . If `True`, will use the token generated when running `huggingface-cli login` (stored in `~/.huggingface`).""" hf_overrides: HfOverrides = field(default_factory=dict) """If a dictionary, contains arguments to be forwarded to the Hugging Face config. If a callable, it is called to update the HuggingFace config.""" logits_processor_pattern: str | None = None """Optional regex pattern specifying valid logits processor qualified names that can be passed with the `logits_processors` extra completion argument. Defaults to `None`, which allows no processors.""" generation_config: str = "auto" """The folder path to the generation config. Defaults to `"auto"`, the generation config will be loaded from model path. If set to `"vllm"`, no generation config is loaded, vLLM defaults will be used. If set to a folder path, the generation config will be loaded from the specified folder path. If `max_new_tokens` is specified in generation config, then it sets a server-wide limit on the number of output tokens for all requests.""" override_generation_config: dict[str, Any] = field(default_factory=dict) """Overrides or sets generation config. e.g. `{"temperature": 0.5}`. If used with `--generation-config auto`, the override parameters will be merged with the default config from the model. If used with `--generation-config vllm`, only the override parameters are used.""" enable_sleep_mode: bool = False """Enable sleep mode for the engine (only cuda and hip platforms are supported).""" model_impl: str | ModelImpl = "auto" """Which implementation of the model to use:\n - "auto" will try to use the vLLM implementation, if it exists, and fall back to the Transformers implementation if no vLLM implementation is available.\n - "vllm" will use the vLLM model implementation.\n - "transformers" will use the Transformers model implementation.\n - "terratorch" will use the TerraTorch model implementation. """ override_attention_dtype: str | None = None """Override dtype for attention""" logits_processors: list[str | type[LogitsProcessor]] | None = None """One or more logits processors' fully-qualified class names or class definitions""" io_processor_plugin: str | None = None """IOProcessor plugin name to load at model startup""" # Pooler config pooler_config: PoolerConfig | None = None """Pooler config which controls the behaviour of output pooling in pooling models.""" # Multimodal config and init vars multimodal_config: MultiModalConfig | None = None """Configuration for multimodal model. If `None`, this will be inferred from the architecture of `self.model`.""" limit_mm_per_prompt: InitVar[dict[str, int | dict[str, int]] | None] = None enable_mm_embeds: InitVar[bool | None] = None media_io_kwargs: InitVar[dict[str, dict[str, Any]] | None] = None mm_processor_kwargs: InitVar[dict[str, Any] | None] = None mm_processor_cache_gb: InitVar[float | None] = None mm_processor_cache_type: InitVar[MMCacheType | None] = None mm_shm_cache_max_object_size_mb: InitVar[int | None] = None mm_encoder_tp_mode: InitVar[MMEncoderTPMode | None] = None mm_encoder_attn_backend: InitVar[AttentionBackendEnum | str | None] = None interleave_mm_strings: InitVar[bool | None] = None skip_mm_profiling: InitVar[bool | None] = None video_pruning_rate: InitVar[float | None] = None def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ ignored_factors = { "runner", "convert", "tokenizer", "tokenizer_mode", "seed", "hf_config_path", "allowed_local_media_path", "allowed_media_domains", "tokenizer_revision", "spec_target_max_model_len", "enforce_eager", "logprobs_mode", "disable_cascade_attn", "skip_tokenizer_init", "served_model_name", "config_format", "hf_token", "hf_overrides", "logits_processor_pattern", "override_attention_dtype", "logits_processors", "io_processor_plugin", "pooler_config", "multimodal_config", "limit_mm_per_prompt", "media_io_kwargs", "mm_processor_kwargs", "mm_processor_cache_gb", "mm_processor_cache_type", "mm_shm_cache_max_object_size_mb", "mm_encoder_tp_mode", "interleave_mm_strings", "skip_mm_profiling", } from vllm.config.utils import get_hash_factors, hash_factors factors = get_hash_factors(self, ignored_factors) return hash_factors(factors) def _update_nested( self, target: PretrainedConfig | dict[str, Any], updates: dict[str, Any], ) -> None: """Recursively updates a config or dict with nested updates.""" for key, value in updates.items(): if isinstance(value, dict): # Get the nested target if isinstance(target, dict): nested_target = target.get(key) else: nested_target = getattr(target, key, None) # If nested target exists and can be updated recursively if nested_target is not None and ( isinstance(nested_target, dict) or hasattr(nested_target, "__dict__") ): self._update_nested(nested_target, value) continue # Set the value (base case) if isinstance(target, dict): target[key] = value else: setattr(target, key, value) def _apply_dict_overrides( self, config: PretrainedConfig, overrides: dict[str, Any], ) -> None: """Apply dict overrides, handling both nested configs and dict values.""" from transformers import PretrainedConfig for key, value in overrides.items(): attr = getattr(config, key, None) if attr is not None and isinstance(attr, PretrainedConfig): # It's a nested config - recursively update it self._update_nested(attr, value) else: # It's a dict-valued parameter - set it directly setattr(config, key, value) def __post_init__( self, # Multimodal config init vars limit_mm_per_prompt: dict[str, int | dict[str, int]] | None, enable_mm_embeds: bool | None, media_io_kwargs: dict[str, dict[str, Any]] | None, mm_processor_kwargs: dict[str, Any] | None, mm_processor_cache_gb: float | None, mm_processor_cache_type: MMCacheType | None, mm_shm_cache_max_object_size_mb: int | None, mm_encoder_tp_mode: MMEncoderTPMode | None, mm_encoder_attn_backend: AttentionBackendEnum | str | None, interleave_mm_strings: bool | None, skip_mm_profiling: bool | None, video_pruning_rate: float | None, ) -> None: # Keep set served_model_name before maybe_model_redirect(self.model) self.served_model_name = get_served_model_name( self.model, self.served_model_name ) self.model = maybe_model_redirect(self.model) # The tokenizer is consistent with the model by default. if self.tokenizer is None: self.tokenizer = self.model if self.tokenizer_revision is None: self.tokenizer_revision = self.revision self.tokenizer = maybe_model_redirect(self.tokenizer) if isinstance(self.hf_config_path, str): self.hf_config_path = maybe_model_redirect(self.hf_config_path) if callable(self.hf_overrides): hf_overrides_kw = {} hf_overrides_fn = self.hf_overrides dict_overrides: dict[str, Any] = {} else: # Separate dict overrides from flat ones # We'll determine how to apply dict overrides after loading the config hf_overrides_kw = {} dict_overrides = {} for key, value in self.hf_overrides.items(): if isinstance(value, dict): dict_overrides[key] = value else: hf_overrides_kw[key] = value hf_overrides_fn = None self.maybe_pull_model_tokenizer_for_runai(self.model, self.tokenizer) from vllm.platforms import current_platform if self.override_attention_dtype is not None and not current_platform.is_rocm(): warnings.warn( "override-attention-dtype is set but not using ROCm platform", stacklevel=2, ) if self.enable_sleep_mode and not current_platform.is_sleep_mode_available(): raise ValueError("Sleep mode is not supported on current platform.") hf_config = get_config( self.hf_config_path or self.model, self.trust_remote_code, self.revision, self.code_revision, self.config_format, hf_overrides_kw=hf_overrides_kw, hf_overrides_fn=hf_overrides_fn, ) hf_config = maybe_patch_hf_config_from_gguf( self.model, hf_config, ) self.hf_config = hf_config if dict_overrides: self._apply_dict_overrides(hf_config, dict_overrides) self.hf_text_config = get_hf_text_config(self.hf_config) self.attention_chunk_size = getattr( self.hf_text_config, "attention_chunk_size", None ) self.encoder_config = self._get_encoder_config() self.hf_image_processor_config = get_hf_image_processor_config( self.model, hf_token=self.hf_token, revision=self.revision ) self.model_arch_config = self.get_model_arch_config() architectures = self.architectures registry = self.registry is_generative_model = registry.is_text_generation_model(architectures, self) is_pooling_model = registry.is_pooling_model(architectures, self) self.runner_type = self._get_runner_type(architectures, self.runner) self.convert_type = self._get_convert_type( architectures, self.runner_type, self.convert ) if self.runner_type == "generate" and not is_generative_model: generate_converts = _RUNNER_CONVERTS["generate"] if self.convert_type not in generate_converts: # Currently we don't have any converters for generative models raise ValueError("This model does not support `--runner generate`.") if self.runner_type == "pooling" and not is_pooling_model: pooling_converts = _RUNNER_CONVERTS["pooling"] if self.convert_type not in pooling_converts: convert_option = "<" + "|".join(pooling_converts) + ">" raise ValueError( "This model does not support `--runner pooling`. " f"You can pass `--convert {convert_option} to adapt " "it into a pooling model." ) # Note: Initialize these attributes early because transformers fallback # may fail to load dynamic modules in child processes model_info, arch = registry.inspect_model_cls(architectures, self) self._model_info = model_info self._architecture = arch logger.info("Resolved architecture: %s", arch) # Init pooler config if needed if self.runner_type == "pooling": if self.pooler_config is None: self.pooler_config = PoolerConfig() base_config = get_pooling_config(self.model, self.revision) if base_config is not None: # Only set values that are not overridden by the user for k, v in base_config.items(): if getattr(self.pooler_config, k) is None: setattr(self.pooler_config, k, v) default_pooling_type = self._model_info.default_pooling_type if self.pooler_config.pooling_type is None: self.pooler_config.pooling_type = default_pooling_type self.dtype: torch.dtype = _get_and_verify_dtype( self.model, self.hf_config, self.dtype, is_pooling_model=self.runner_type == "pooling", revision=self.revision, ) self.original_max_model_len = self.max_model_len self.max_model_len = self.get_and_verify_max_len(self.max_model_len) if self.is_encoder_decoder: self.mm_processor_cache_gb = 0 logger.info("Encoder-decoder model detected, disabling mm processor cache.") # Init multimodal config if needed if self._model_info.supports_multimodal: if ( mm_encoder_tp_mode == "data" and not self._model_info.supports_multimodal_encoder_tp_data ): logger.warning_once( "This model does not support `--mm-encoder-tp-mode data`. " "Falling back to `--mm-encoder-tp-mode weights`." ) mm_encoder_tp_mode = "weights" mm_config_kwargs = dict( limit_per_prompt=limit_mm_per_prompt, enable_mm_embeds=enable_mm_embeds, media_io_kwargs=media_io_kwargs, mm_processor_kwargs=mm_processor_kwargs, mm_processor_cache_gb=mm_processor_cache_gb, mm_processor_cache_type=mm_processor_cache_type, mm_shm_cache_max_object_size_mb=mm_shm_cache_max_object_size_mb, mm_encoder_tp_mode=mm_encoder_tp_mode, mm_encoder_attn_backend=mm_encoder_attn_backend, interleave_mm_strings=interleave_mm_strings, skip_mm_profiling=skip_mm_profiling, video_pruning_rate=video_pruning_rate, ) mm_config_kwargs = { k: v for k, v in mm_config_kwargs.items() if v is not None } self.multimodal_config = MultiModalConfig(**mm_config_kwargs) # Multimodal GGUF models must use original repo for mm processing if is_gguf(self.tokenizer) and self.is_multimodal_model: raise ValueError( "Loading a multimodal GGUF model needs to use original " "tokenizer. Please specify the unquantized hf model's " "repo name or path using the --tokenizer argument." ) if self.disable_sliding_window: # Set after get_and_verify_max_len to ensure that max_model_len # can be correctly capped to sliding window size self.hf_text_config.sliding_window = None # Avoid running try_verify_and_update_config multiple times self.config_updated = False self._try_verify_and_update_model_config() self._verify_quantization() self._verify_cuda_graph() self._verify_bnb_config() def get_model_arch_config( self, ) -> ModelArchitectureConfig: convertor_cls = MODEL_ARCH_CONFIG_CONVERTORS.get( self.hf_config.model_type, ModelArchConfigConvertorBase ) convertor = convertor_cls(self.hf_config, self.hf_text_config) return convertor.convert() @field_validator("tokenizer", "max_model_len", mode="wrap") @classmethod def _skip_none_validation(cls, value: Any, handler: Callable) -> Any: """Skip validation if the value is `None` when initialisation is delayed.""" if value is None: return value return handler(value) @field_validator("tokenizer_mode", mode="after") def _lowercase_tokenizer_mode(cls, tokenizer_mode: str) -> str: return tokenizer_mode.lower() @field_validator("quantization", mode="before") @classmethod def validate_quantization_before(cls, value: Any) -> Any: if isinstance(value, str): return value.lower() return value @model_validator(mode="after") def validate_model_config_after(self: "ModelConfig") -> "ModelConfig": """Called after __post_init__""" if not isinstance(self.tokenizer, str): raise ValueError( f"tokenizer must be a string, got " f"{type(self.tokenizer).__name__}: {self.tokenizer!r}. " "Please provide a valid tokenizer path or HuggingFace model ID." ) if not isinstance(self.max_model_len, int): raise ValueError( f"max_model_len must be a positive integer, " f"got {type(self.max_model_len).__name__}: {self.max_model_len!r}. " "Example: max_model_len=2048" ) return self def _get_transformers_backend_cls(self) -> str: """Determine which Transformers modeling backend class will be used if `model_impl` is set to `transformers` or `auto`.""" cls = "Transformers" # If 'hf_config != hf_text_config' it's a nested config, i.e. multimodal cls += "MultiModal" if self.hf_config != self.hf_text_config else "" cls += "MoE" if self.is_moe else "" # Check if the architecture we're wrapping has defaults runner = None task = None if defaults := try_match_architecture_defaults(self.architectures[0]): _, (runner, task) = defaults # User specified value take precedence if self.runner != "auto": runner = self.runner # Only consider Transformers modeling backend pooling classes if we're wrapping # an architecture that defaults to pooling. Otherwise, we return the LM class # and use adapters. if runner == "pooling" and task in {"embed", "classify"}: if task == "embed": cls += "EmbeddingModel" elif task == "classify": cls += "ForSequenceClassification" else: cls += "ForCausalLM" return cls def using_transformers_backend(self) -> bool: """Check if the model is using the Transformers modeling backend class.""" used_cls = self._model_info.architecture transformers_backend_cls = self._get_transformers_backend_cls() return used_cls == transformers_backend_cls @property def registry(self): return me_models.ModelRegistry @property def architectures(self) -> list[str]: return self.model_arch_config.architectures @property def architecture(self) -> str: """The architecture vllm actually used.""" return self._architecture def maybe_pull_model_tokenizer_for_runai(self, model: str, tokenizer: str) -> None: """Pull model/tokenizer from Object Storage to temporary directory when needed. Args: model: Model name or path tokenizer: Tokenizer name or path """ if not (is_runai_obj_uri(model) or is_runai_obj_uri(tokenizer)): return if is_runai_obj_uri(model): object_storage_model = ObjectStorageModel(url=model) object_storage_model.pull_files( model, allow_pattern=["*.model", "*.py", "*.json"] ) self.model_weights = model self.model = object_storage_model.dir # If tokenizer is same as model, download to same directory if model == tokenizer: object_storage_model.pull_files( model, ignore_pattern=[ "*.pt", "*.safetensors", "*.bin", "*.tensors", "*.pth", ], ) self.tokenizer = object_storage_model.dir return # Only download tokenizer if needed and not already handled if is_runai_obj_uri(tokenizer): object_storage_tokenizer = ObjectStorageModel(url=tokenizer) object_storage_tokenizer.pull_files( model, ignore_pattern=["*.pt", "*.safetensors", "*.bin", "*.tensors", "*.pth"], ) self.tokenizer = object_storage_tokenizer.dir def _get_encoder_config(self): model = self.model if is_remote_gguf(model): model, _ = split_remote_gguf(model) return get_sentence_transformer_tokenizer_config(model, self.revision) def _get_default_runner_type( self, architectures: list[str], ) -> RunnerType: registry = self.registry
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/observability.py
vllm/config/observability.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from functools import cached_property from typing import Any, Literal, cast from packaging.version import parse from pydantic import Field, field_validator, model_validator from pydantic.dataclasses import dataclass from vllm import version from vllm.config.utils import config from vllm.utils.hashing import safe_hash DetailedTraceModules = Literal["model", "worker", "all"] @config @dataclass class ObservabilityConfig: """Configuration for observability - metrics and tracing.""" show_hidden_metrics_for_version: str | None = None """Enable deprecated Prometheus metrics that have been hidden since the specified version. For example, if a previously deprecated metric has been hidden since the v0.7.0 release, you use `--show-hidden-metrics-for-version=0.7` as a temporary escape hatch while you migrate to new metrics. The metric is likely to be removed completely in an upcoming release.""" @cached_property def show_hidden_metrics(self) -> bool: """Check if the hidden metrics should be shown.""" if self.show_hidden_metrics_for_version is None: return False return version._prev_minor_version_was(self.show_hidden_metrics_for_version) otlp_traces_endpoint: str | None = None """Target URL to which OpenTelemetry traces will be sent.""" collect_detailed_traces: list[DetailedTraceModules] | None = None """It makes sense to set this only if `--otlp-traces-endpoint` is set. If set, it will collect detailed traces for the specified modules. This involves use of possibly costly and or blocking operations and hence might have a performance impact. Note that collecting detailed timing information for each request can be expensive.""" kv_cache_metrics: bool = False """Enable KV cache residency metrics (lifetime, idle time, reuse gaps). Uses sampling to minimize overhead. Requires log stats to be enabled (i.e., --disable-log-stats not set).""" kv_cache_metrics_sample: float = Field(default=0.01, gt=0, le=1) """Sampling rate for KV cache metrics (0.0, 1.0]. Default 0.01 = 1% of blocks.""" cudagraph_metrics: bool = False """Enable CUDA graph metrics (number of padded/unpadded tokens, runtime cudagraph dispatch modes, and their observed frequencies at every logging interval).""" enable_layerwise_nvtx_tracing: bool = False """Enable layerwise NVTX tracing. This traces the execution of each layer or module in the model and attach informations such as input/output shapes to nvtx range markers. Noted that this doesn't work with CUDA graphs enabled.""" enable_mfu_metrics: bool = False """Enable Model FLOPs Utilization (MFU) metrics.""" enable_mm_processor_stats: bool = False """Enable collection of timing statistics for multimodal processor operations. This is for internal use only (e.g., benchmarks) and is not exposed as a CLI argument.""" enable_mfu_metrics: bool = False """Enable Model FLOPs Utilization (MFU) metrics.""" @cached_property def collect_model_forward_time(self) -> bool: """Whether to collect model forward time for the request.""" return self.collect_detailed_traces is not None and ( "model" in self.collect_detailed_traces or "all" in self.collect_detailed_traces ) @cached_property def collect_model_execute_time(self) -> bool: """Whether to collect model execute time for the request.""" return self.collect_detailed_traces is not None and ( "worker" in self.collect_detailed_traces or "all" in self.collect_detailed_traces ) def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # no factors to consider. # this config will not affect the computation graph. factors: list[Any] = [] hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str @field_validator("show_hidden_metrics_for_version") @classmethod def _validate_show_hidden_metrics_for_version(cls, value: str | None) -> str | None: if value is not None: # Raises an exception if the string is not a valid version. parse(value) return value @field_validator("otlp_traces_endpoint") @classmethod def _validate_otlp_traces_endpoint(cls, value: str | None) -> str | None: if value is not None: from vllm.tracing import is_otel_available, otel_import_error_traceback if not is_otel_available(): raise ValueError( "OpenTelemetry is not available. Unable to configure " "'otlp_traces_endpoint'. Ensure OpenTelemetry packages are " f"installed. Original error:\n{otel_import_error_traceback}" ) return value @field_validator("collect_detailed_traces") @classmethod def _validate_collect_detailed_traces( cls, value: list[DetailedTraceModules] | None ) -> list[DetailedTraceModules] | None: """Handle the legacy case where users might provide a comma-separated string instead of a list of strings.""" if value is not None and len(value) == 1 and "," in value[0]: value = cast(list[DetailedTraceModules], value[0].split(",")) return value @model_validator(mode="after") def _validate_tracing_config(self): if self.collect_detailed_traces and not self.otlp_traces_endpoint: raise ValueError( "collect_detailed_traces requires `--otlp-traces-endpoint` to be set." ) return self
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/speech_to_text.py
vllm/config/speech_to_text.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from pydantic.dataclasses import dataclass from vllm.config.utils import config @config @dataclass class SpeechToTextConfig: """Configuration for speech-to-text models.""" sample_rate: float = 16_000 """Sample rate (Hz) to resample input audio to. Most speech models expect 16kHz audio input. The input audio will be automatically resampled to this rate before processing.""" max_audio_clip_s: int = 30 """Maximum duration in seconds for a single audio clip without chunking. Audio longer than this will be split into smaller chunks if `allow_audio_chunking` evaluates to True, otherwise it will be rejected.""" overlap_chunk_second: int = 1 """Overlap duration in seconds between consecutive audio chunks when splitting long audio. This helps maintain context across chunk boundaries and improves transcription quality at split points.""" min_energy_split_window_size: int | None = 1600 """Window size in samples for finding low-energy (quiet) regions to split audio chunks. The algorithm looks for the quietest moment within this window to minimize cutting through speech. Default 1600 samples ≈ 100ms at 16kHz. If None, no chunking will be done.""" @property def allow_audio_chunking(self) -> bool: return self.min_energy_split_window_size is not None
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/kv_events.py
vllm/config/kv_events.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Literal from pydantic import Field from pydantic.dataclasses import dataclass from vllm.config.utils import config @config @dataclass class KVEventsConfig: """Configuration for KV event publishing.""" enable_kv_cache_events: bool = False """If True, enable KV cache events for tracking block storage and removal. Events can be published externally by zmq using the event publisher config. """ publisher: Literal["null", "zmq"] = Field(default=None) """The publisher to use for publishing kv events. Can be "null", "zmq". """ endpoint: str = "tcp://*:5557" """The zmq endpoint to use for publishing kv events. """ replay_endpoint: str | None = None """The zmq endpoint to use for replaying kv events. """ buffer_steps: int = 10_000 """The number of steps to cache for replay endpoint. Will only save events from the last N steps for the replay endpoint. """ hwm: int = 100_000 """The zmq high water mark for the event publisher. After queueing N events, events will start dropping if the consumer is not keeping up. """ max_queue_size: int = 100_000 """The maximum number of events to queue while waiting for publishing. """ topic: str = "" """The topic to use for the event publisher. Consumers can subscribe to this topic to receive events. """ def __post_init__(self): if self.publisher is None: self.publisher = "zmq" if self.enable_kv_cache_events else "null"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/vllm.py
vllm/config/vllm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import copy import getpass import json import os import tempfile import threading import time from contextlib import contextmanager from dataclasses import is_dataclass, replace from datetime import datetime from enum import IntEnum from functools import lru_cache from pathlib import Path from typing import TYPE_CHECKING, Any, TypeVar, get_args import torch from pydantic import ConfigDict, Field, model_validator from pydantic.dataclasses import dataclass import vllm.envs as envs from vllm.config.speculative import EagleModelTypes from vllm.logger import enable_trace_function_call, init_logger from vllm.transformers_utils.runai_utils import is_runai_obj_uri from vllm.utils import random_uuid from vllm.utils.hashing import safe_hash from .attention import AttentionConfig from .cache import CacheConfig from .compilation import CompilationConfig, CompilationMode, CUDAGraphMode from .device import DeviceConfig from .ec_transfer import ECTransferConfig from .kv_events import KVEventsConfig from .kv_transfer import KVTransferConfig from .load import LoadConfig from .lora import LoRAConfig from .model import ModelConfig from .observability import ObservabilityConfig from .parallel import ParallelConfig from .profiler import ProfilerConfig from .scheduler import SchedulerConfig from .speculative import SpeculativeConfig from .structured_outputs import StructuredOutputsConfig from .utils import SupportsHash, config if TYPE_CHECKING: from transformers import PretrainedConfig from vllm.model_executor.layers.quantization.base_config import QuantizationConfig from vllm.v1.kv_cache_interface import KVCacheConfig else: PretrainedConfig = Any QuantizationConfig = Any KVCacheConfig = Any logger = init_logger(__name__) class OptimizationLevel(IntEnum): """Optimization level enum.""" O0 = 0 """O0 : No optimization. no compilation, no cudagraphs, no other optimization, just starting up immediately""" O1 = 1 """O1: Quick optimizations. Dynamo+Inductor compilation and Piecewise cudagraphs""" O2 = 2 """O2: Full optimizations. -O1 as well as Full and Piecewise cudagraphs.""" O3 = 3 """O3: Currently the same as -O2s.""" IS_QUANTIZED = False IS_DENSE = False # The optimizations that depend on these properties currently set to False # in all cases. # if model_config is not None: # IS_QUANTIZED = lambda c: c.model_config.is_quantized() # IS_DENSE = lambda c: not c.model_config.is_model_moe() # See https://github.com/vllm-project/vllm/issues/25689. def enable_norm_fusion(cfg: "VllmConfig") -> bool: """Enable if either RMS norm or quant FP8 custom op is active; otherwise Inductor handles fusion.""" return cfg.compilation_config.is_custom_op_enabled( "rms_norm" ) or cfg.compilation_config.is_custom_op_enabled("quant_fp8") def enable_act_fusion(cfg: "VllmConfig") -> bool: """Enable if either SiLU+Mul or quant FP8 custom op is active; otherwise Inductor handles fusion.""" return cfg.compilation_config.is_custom_op_enabled( "silu_and_mul" ) or cfg.compilation_config.is_custom_op_enabled("quant_fp8") OPTIMIZATION_LEVEL_00 = { "compilation_config": { "pass_config": { "eliminate_noops": False, "fuse_norm_quant": False, "fuse_act_quant": False, "fuse_allreduce_rms": False, "fuse_attn_quant": False, "enable_sp": False, "fuse_gemm_comms": False, }, "cudagraph_mode": CUDAGraphMode.NONE, "use_inductor_graph_partition": False, }, } OPTIMIZATION_LEVEL_01 = { "compilation_config": { "pass_config": { "eliminate_noops": True, "fuse_norm_quant": enable_norm_fusion, "fuse_act_quant": enable_act_fusion, "fuse_allreduce_rms": False, "fuse_attn_quant": False, "enable_sp": False, "fuse_gemm_comms": False, }, "cudagraph_mode": CUDAGraphMode.PIECEWISE, "use_inductor_graph_partition": False, }, } OPTIMIZATION_LEVEL_02 = { "compilation_config": { "pass_config": { "eliminate_noops": True, "fuse_norm_quant": enable_norm_fusion, "fuse_act_quant": enable_act_fusion, "fuse_allreduce_rms": False, "fuse_attn_quant": IS_QUANTIZED, "enable_sp": IS_DENSE, "fuse_gemm_comms": IS_DENSE, }, "cudagraph_mode": CUDAGraphMode.FULL_AND_PIECEWISE, "use_inductor_graph_partition": False, }, } OPTIMIZATION_LEVEL_03 = { "compilation_config": { "pass_config": { "eliminate_noops": True, "fuse_norm_quant": enable_norm_fusion, "fuse_act_quant": enable_act_fusion, "fuse_allreduce_rms": False, "fuse_attn_quant": IS_QUANTIZED, "enable_sp": IS_DENSE, "fuse_gemm_comms": IS_DENSE, }, "cudagraph_mode": CUDAGraphMode.FULL_AND_PIECEWISE, "use_inductor_graph_partition": False, }, } OPTIMIZATION_LEVEL_TO_CONFIG = { OptimizationLevel.O0: OPTIMIZATION_LEVEL_00, OptimizationLevel.O1: OPTIMIZATION_LEVEL_01, OptimizationLevel.O2: OPTIMIZATION_LEVEL_02, OptimizationLevel.O3: OPTIMIZATION_LEVEL_03, } @config @dataclass(config=ConfigDict(arbitrary_types_allowed=True)) class VllmConfig: """Dataclass which contains all vllm-related configuration. This simplifies passing around the distinct configurations in the codebase. """ # TODO: use default_factory once default constructing ModelConfig doesn't # try to download a model model_config: ModelConfig = Field(default=None) """Model configuration.""" cache_config: CacheConfig = Field(default_factory=CacheConfig) """Cache configuration.""" parallel_config: ParallelConfig = Field(default_factory=ParallelConfig) """Parallel configuration.""" scheduler_config: SchedulerConfig = Field( default_factory=SchedulerConfig.default_factory, ) """Scheduler configuration.""" device_config: DeviceConfig = Field(default_factory=DeviceConfig) """Device configuration.""" load_config: LoadConfig = Field(default_factory=LoadConfig) """Load configuration.""" attention_config: AttentionConfig = Field(default_factory=AttentionConfig) """Attention configuration.""" lora_config: LoRAConfig | None = None """LoRA configuration.""" speculative_config: SpeculativeConfig | None = None """Speculative decoding configuration.""" structured_outputs_config: StructuredOutputsConfig = Field( default_factory=StructuredOutputsConfig ) """Structured outputs configuration.""" observability_config: ObservabilityConfig = Field( default_factory=ObservabilityConfig ) """Observability configuration.""" quant_config: QuantizationConfig | None = None """Quantization configuration.""" compilation_config: CompilationConfig = Field(default_factory=CompilationConfig) """`torch.compile` and cudagraph capture configuration for the model. As a shorthand, one can append compilation arguments via -cc.parameter=argument such as `-cc.mode=3` (same as `-cc='{"mode":3}'`). You can specify the full compilation config like so: `{"mode": 3, "cudagraph_capture_sizes": [1, 2, 4, 8]}` """ profiler_config: ProfilerConfig = Field(default_factory=ProfilerConfig) """Profiling configuration.""" kv_transfer_config: KVTransferConfig | None = None """The configurations for distributed KV cache transfer.""" kv_events_config: KVEventsConfig | None = None """The configurations for event publishing.""" ec_transfer_config: ECTransferConfig | None = None """The configurations for distributed EC cache transfer.""" # some opaque config, only used to provide additional information # for the hash computation, mainly used for testing, debugging or out of # tree config registration. additional_config: dict | SupportsHash = Field(default_factory=dict) """Additional config for specified platform. Different platforms may support different configs. Make sure the configs are valid for the platform you are using. Contents must be hashable.""" instance_id: str = "" """The ID of the vLLM instance.""" optimization_level: OptimizationLevel = OptimizationLevel.O2 """The optimization level. These levels trade startup time cost for performance, with -O0 having the best startup time and -O3 having the best performance. -02 is used by defult. See OptimizationLevel for full description.""" def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ factors: list[Any] = [] # summarize vllm config vllm_factors: list[Any] = [] from vllm import __version__ vllm_factors.append(__version__) if self.model_config: vllm_factors.append(self.model_config.compute_hash()) else: vllm_factors.append("None") if self.cache_config: vllm_factors.append(self.cache_config.compute_hash()) else: vllm_factors.append("None") if self.parallel_config: vllm_factors.append(self.parallel_config.compute_hash()) else: vllm_factors.append("None") if self.scheduler_config: vllm_factors.append(self.scheduler_config.compute_hash()) else: vllm_factors.append("None") if self.device_config: vllm_factors.append(self.device_config.compute_hash()) else: vllm_factors.append("None") if self.load_config: vllm_factors.append(self.load_config.compute_hash()) else: vllm_factors.append("None") if self.attention_config: vllm_factors.append(self.attention_config.compute_hash()) else: vllm_factors.append("None") if self.lora_config: vllm_factors.append(self.lora_config.compute_hash()) else: vllm_factors.append("None") if self.speculative_config: vllm_factors.append(self.speculative_config.compute_hash()) else: vllm_factors.append("None") if self.structured_outputs_config: vllm_factors.append(self.structured_outputs_config.compute_hash()) if self.profiler_config: vllm_factors.append(self.profiler_config.compute_hash()) else: vllm_factors.append("None") vllm_factors.append(self.observability_config.compute_hash()) if self.quant_config: pass # should be captured by model_config.quantization if self.compilation_config: vllm_factors.append(self.compilation_config.compute_hash()) else: vllm_factors.append("None") if self.kv_transfer_config: vllm_factors.append(self.kv_transfer_config.compute_hash()) else: vllm_factors.append("None") if self.ec_transfer_config: vllm_factors.append(self.ec_transfer_config.compute_hash()) else: vllm_factors.append("None") if self.additional_config: if isinstance(additional_config := self.additional_config, dict): additional_config_hash = safe_hash( json.dumps(additional_config, sort_keys=True).encode(), usedforsecurity=False, ).hexdigest() else: additional_config_hash = additional_config.compute_hash() vllm_factors.append(additional_config_hash) else: vllm_factors.append("None") factors.append(vllm_factors) hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest()[ :10 ] return hash_str def pad_for_cudagraph(self, batch_size: int) -> int: # if batch_size > self.compilation_config.max_cudagraph_capture_size, # it should raise an IndexError. # the caller should make sure the batch_size is within the range, # i.e., batch_size <= self.compilation_config.max_cudagraph_capture_size return self.compilation_config.bs_to_padded_graph_size[batch_size] @property def needs_dp_coordinator(self) -> bool: """ Determine if the DPCoordinator process is needed. The DPCoordinator is needed in two cases: 1. For MoE models with DP > 1: to handle wave coordination (even in external LB mode, since wave coordination runs in the coordinator) 2. For non-MoE models in internal/hybrid LB mode: to collect and publish queue stats for load balancing across DP ranks Returns: True if DPCoordinator process is needed, False otherwise. """ # For non-MoE models, only need coordinator in internal/hybrid LB mode # (for stats collection). return self.parallel_config.data_parallel_size > 1 and ( self.model_config is None or self.model_config.is_moe or not self.parallel_config.data_parallel_external_lb ) def enable_trace_function_call_for_thread(self) -> None: """ Set up function tracing for the current thread, if enabled via the `VLLM_TRACE_FUNCTION` environment variable. """ if envs.VLLM_TRACE_FUNCTION: tmp_dir = tempfile.gettempdir() # add username to tmp_dir to avoid permission issues tmp_dir = os.path.join(tmp_dir, getpass.getuser()) filename = ( f"VLLM_TRACE_FUNCTION_for_process_{os.getpid()}" f"_thread_{threading.get_ident()}_at_{datetime.now()}.log" ).replace(" ", "_") log_path = os.path.join( tmp_dir, "vllm", f"vllm-instance-{self.instance_id}", filename, ) os.makedirs(os.path.dirname(log_path), exist_ok=True) enable_trace_function_call(log_path) @staticmethod def _get_quantization_config( model_config: ModelConfig, load_config: LoadConfig ) -> QuantizationConfig | None: """Get the quantization config.""" from vllm.platforms import current_platform if model_config.quantization is not None: from vllm.model_executor.model_loader.weight_utils import get_quant_config quant_config = get_quant_config(model_config, load_config) capability_tuple = current_platform.get_device_capability() if capability_tuple is not None: capability = capability_tuple.to_int() if capability < quant_config.get_min_capability(): raise ValueError( f"The quantization method {model_config.quantization} " "is not supported for the current GPU. Minimum " f"capability: {quant_config.get_min_capability()}. " f"Current capability: {capability}." ) supported_dtypes = quant_config.get_supported_act_dtypes() if model_config.dtype not in supported_dtypes: raise ValueError( f"{model_config.dtype} is not supported for quantization " f"method {model_config.quantization}. Supported dtypes: " f"{supported_dtypes}" ) quant_config.maybe_update_config(model_config.model) return quant_config return None @staticmethod def get_quantization_config( model_config: ModelConfig, load_config: LoadConfig ) -> QuantizationConfig | None: import copy # For some reason, the _ version of this modifies the model_config # object, so using deepcopy to avoid this problem. return VllmConfig._get_quantization_config( copy.deepcopy(model_config), load_config ) def with_hf_config( self, hf_config: PretrainedConfig, architectures: list[str] | None = None, ) -> "VllmConfig": if architectures is not None: hf_config = copy.deepcopy(hf_config) hf_config.architectures = architectures model_config = copy.deepcopy(self.model_config) model_config.hf_config = hf_config model_config.model_arch_config = model_config.get_model_arch_config() return replace(self, model_config=model_config) def _set_config_default(self, config_obj: Any, key: str, value: Any) -> None: """Set config attribute to default if not already set by user. Args: config_obj: Configuration object to update. key: Attribute name. value: Default value (static or callable). """ if getattr(config_obj, key) is None: # Some config values are known before initialization and are # hard coded. # Other values depend on the user given configuration, so they are # implemented with lambda functions and decided at run time. setattr(config_obj, key, value(self) if callable(value) else value) def _apply_optimization_level_defaults(self, defaults: dict[str, Any]) -> None: """Apply optimization level defaults using self as root. Recursively applies values from defaults into nested config objects. Only fields present in defaults are overwritten. If the user configuration does not specify a value for a default field and if the default field is still None after all user selections are applied, then default values will be applied to the field. User speciied fields will not be overridden by the default. Args: defaults: Dictionary of default values to apply. """ def apply_recursive(config_obj: Any, config_defaults: dict[str, Any]) -> None: """Recursively apply defaults to config_obj, using self as root.""" for key, value in config_defaults.items(): if not hasattr(config_obj, key): continue current = getattr(config_obj, key) if isinstance(value, dict) and is_dataclass(current): apply_recursive(current, value) else: self._set_config_default(config_obj, key, value) apply_recursive(self, defaults) def _post_init_kv_transfer_config(self) -> None: """Update KVTransferConfig based on top-level configs in VllmConfig. Right now, this function reads the offloading settings from CacheConfig and configures the KVTransferConfig accordingly. """ if (kv_offloading_backend := self.cache_config.kv_offloading_backend) is None: return # If no KVTransferConfig is provided, create a default one. if self.kv_transfer_config is None: self.kv_transfer_config = KVTransferConfig() if (kv_offloading_size := self.cache_config.kv_offloading_size) is None: raise ValueError( "You must set kv_offloading_size when kv_offloading_backend is set." ) num_kv_ranks = ( self.parallel_config.tensor_parallel_size * self.parallel_config.pipeline_parallel_size ) if kv_offloading_backend == "native": self.kv_transfer_config.kv_connector = "OffloadingConnector" kv_bytes_per_rank = kv_offloading_size * (1 << 30) / num_kv_ranks # NOTE(ApostaC): the actual calculation for num_cpu_blocks should be # done after the model's KV cache is initialized self.kv_transfer_config.kv_connector_extra_config.update( {"kv_bytes_per_rank": kv_bytes_per_rank, "num_cpu_blocks": 0} ) elif kv_offloading_backend == "lmcache": self.kv_transfer_config.kv_connector = "LMCacheConnectorV1" kv_gb_per_rank = kv_offloading_size / num_kv_ranks self.kv_transfer_config.kv_connector_extra_config = { "lmcache.local_cpu": True, "lmcache.max_local_cpu_size": kv_gb_per_rank, } # This is the same for all backends self.kv_transfer_config.kv_role = "kv_both" def __post_init__(self): """Verify configs are valid & consistent with each other.""" # To give each torch profile run a unique instance name. self.instance_id = f"{time.time_ns()}" self.try_verify_and_update_config() if self.model_config is not None: self.model_config.verify_with_parallel_config(self.parallel_config) self.model_config.verify_dual_chunk_attention_config(self.load_config) self.parallel_config.is_moe_model = self.model_config.is_moe self.cache_config.verify_with_parallel_config(self.parallel_config) if self.lora_config is not None: self.lora_config.verify_with_model_config(self.model_config) if self.quant_config is None and self.model_config is not None: self.quant_config = VllmConfig._get_quantization_config( self.model_config, self.load_config ) executor_backend = self.parallel_config.distributed_executor_backend executor_supports_async_sched = executor_backend in ( "mp", "uni", "external_launcher", ) if self.scheduler_config.async_scheduling: # Async scheduling explicitly enabled, hard fail any incompatibilities. if self.parallel_config.pipeline_parallel_size > 1: raise ValueError( "Async scheduling is not yet compatible with " "pipeline_parallel_size > 1." ) # Currently, async scheduling only support eagle speculative # decoding. if self.speculative_config is not None: if self.speculative_config.method not in get_args(EagleModelTypes): raise ValueError( "Currently, async scheduling is only supported " "with EAGLE/MTP kind of speculative decoding." ) if self.speculative_config.disable_padded_drafter_batch: raise ValueError( "async scheduling for EAGLE/MTP kind of speculative " "decoding is enabled, but disable_padded_drafter_batch=True " "disable_padded_drafter_batch=True is not supported for " "this situation now. please set " "disable_padded_drafter_batch=Fasle" ) if not executor_supports_async_sched: raise ValueError( "Currently, async scheduling only supports `mp`, `uni`, or " "`external_launcher` distributed executor backend, but you chose " f"`{executor_backend}`." ) elif self.scheduler_config.async_scheduling is None: # Enable async scheduling unless there is an incompatible option. if self.parallel_config.pipeline_parallel_size > 1: logger.warning( "Async scheduling is not yet supported with " "pipeline_parallel_size > 1 and will be disabled." ) self.scheduler_config.async_scheduling = False elif self.speculative_config is not None: if self.speculative_config.method not in get_args(EagleModelTypes): logger.warning( "Async scheduling not supported with %s-based " "speculative decoding and will be disabled.", self.speculative_config.method, ) else: logger.warning( "Async scheduling will be disabled because some features do " "not currently work in conjunction with speculative decoding. " "To use async scheduling with spec decoding anyway, " "enable it explicitly via async_scheduling=True." ) self.scheduler_config.async_scheduling = False elif not executor_supports_async_sched: logger.warning( "Async scheduling will be disabled because it is not supported " "with the `%s` distributed executor backend (only `mp`, `uni`, and " "`external_launcher` are supported).", executor_backend, ) self.scheduler_config.async_scheduling = False else: self.scheduler_config.async_scheduling = True if ( self.scheduler_config.async_scheduling and not self.parallel_config.disable_nccl_for_dp_synchronization ): logger.info_once( "Disabling NCCL for DP synchronization when using async scheduling." ) self.parallel_config.disable_nccl_for_dp_synchronization = True logger.info_once( "Asynchronous scheduling is %s.", "enabled" if self.scheduler_config.async_scheduling else "disabled", ) from vllm.platforms import current_platform if ( self.model_config is not None and self.scheduler_config.enable_chunked_prefill and self.model_config.dtype == torch.float32 and current_platform.get_device_capability() == (7, 5) ): logger.warning_once( "Turing devices tensor cores do not support float32 matmul. " "To workaround this limitation, vLLM will set 'ieee' input " "precision for chunked prefill triton kernels." ) if ( self.optimization_level > OptimizationLevel.O0 and self.model_config is not None and self.model_config.enforce_eager ): logger.warning("Enforce eager set, overriding optimization level to -O0") self.optimization_level = OptimizationLevel.O0 if self.compilation_config.backend == "eager" or ( self.compilation_config.mode is not None and self.compilation_config.mode != CompilationMode.VLLM_COMPILE ): logger.warning( "Inductor compilation was disabled by user settings," "Optimizations settings that are only active during" "Inductor compilation will be ignored." ) def has_blocked_weights(): if self.quant_config is not None: if hasattr(self.quant_config, "weight_block_size"): return self.quant_config.weight_block_size is not None elif hasattr(self.quant_config, "has_blocked_weights"): return self.quant_config.has_blocked_weights() return False # Enable quant_fp8 CUDA ops (TODO disable in follow up) # On H100 the CUDA kernel is faster than # native implementation # https://github.com/vllm-project/vllm/issues/25094 if has_blocked_weights(): custom_ops = self.compilation_config.custom_ops if "-quant_fp8" not in custom_ops: custom_ops.append("+quant_fp8") if self.compilation_config.mode is None: if self.optimization_level > OptimizationLevel.O0: self.compilation_config.mode = CompilationMode.VLLM_COMPILE else: self.compilation_config.mode = CompilationMode.NONE if all(s not in self.compilation_config.custom_ops for s in ("all", "none")): if ( self.compilation_config.backend == "inductor" and self.compilation_config.mode != CompilationMode.NONE ): self.compilation_config.custom_ops.append("none") else: self.compilation_config.custom_ops.append("all") default_config = OPTIMIZATION_LEVEL_TO_CONFIG[self.optimization_level] self._apply_optimization_level_defaults(default_config) if ( self.compilation_config.cudagraph_mode.requires_piecewise_compilation() and self.compilation_config.mode != CompilationMode.VLLM_COMPILE ): logger.info( "Cudagraph mode %s is not compatible with compilation mode %s." "Overriding to NONE.", self.compilation_config.cudagraph_mode, self.compilation_config.mode, ) self.compilation_config.cudagraph_mode = CUDAGraphMode.NONE # async tp is built on top of sequence parallelism # and requires it to be enabled. if self.compilation_config.pass_config.fuse_gemm_comms: self.compilation_config.pass_config.enable_sp = True if self.compilation_config.pass_config.enable_sp: if "-rms_norm" in self.compilation_config.custom_ops: logger.warning( "RMS norm force disabled, sequence parallelism might break" ) else: self.compilation_config.custom_ops.append("+rms_norm") if current_platform.support_static_graph_mode(): # if cudagraph_mode has full cudagraphs, we need to check support if model_config := self.model_config: if ( self.compilation_config.cudagraph_mode.has_full_cudagraphs() and model_config.pooler_config is not None ): logger.warning_once( "Pooling models do not support full cudagraphs. " "Overriding cudagraph_mode to PIECEWISE." ) self.compilation_config.cudagraph_mode = CUDAGraphMode.PIECEWISE elif ( model_config.is_encoder_decoder and self.compilation_config.cudagraph_mode not in (CUDAGraphMode.NONE, CUDAGraphMode.FULL_DECODE_ONLY) ): logger.info_once( "Encoder-decoder models do not support %s. " "Overriding cudagraph_mode to FULL_DECODE_ONLY.", self.compilation_config.cudagraph_mode.name, ) self.compilation_config.cudagraph_mode = ( CUDAGraphMode.FULL_DECODE_ONLY ) # disable cudagraph when enforce eager execution if self.model_config is not None and self.model_config.enforce_eager: logger.info("Cudagraph is disabled under eager mode") self.compilation_config.cudagraph_mode = CUDAGraphMode.NONE # override related settings when enforce eager self.compilation_config.max_cudagraph_capture_size = 0 self.compilation_config.cudagraph_capture_sizes = [] else: self.compilation_config.cudagraph_num_of_warmups = 1 self._set_cudagraph_sizes() else: self.compilation_config.cudagraph_mode = CUDAGraphMode.NONE if self.cache_config.kv_sharing_fast_prefill: if ( self.speculative_config is not None and self.speculative_config.use_eagle() ): raise ValueError( "Fast prefill optimization for KV sharing is not " "compatible with EAGLE as EAGLE requires correct logits " "for all tokens while fast prefill gives incorrect logits " "for prompt tokens." )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/lora.py
vllm/config/lora.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import TYPE_CHECKING, Any, Literal import torch from pydantic import ConfigDict, Field, model_validator from pydantic.dataclasses import dataclass from typing_extensions import Self from vllm.config.utils import config from vllm.logger import init_logger from vllm.utils.hashing import safe_hash if TYPE_CHECKING: from vllm.config import ModelConfig from vllm.config.cache import CacheConfig else: ModelConfig = Any CacheConfig = Any logger = init_logger(__name__) LoRADType = Literal["auto", "float16", "bfloat16"] MaxLoRARanks = Literal[1, 8, 16, 32, 64, 128, 256, 320, 512] LoRAExtraVocabSize = Literal[256, 512] @config @dataclass(config=ConfigDict(arbitrary_types_allowed=True)) class LoRAConfig: """Configuration for LoRA.""" max_lora_rank: MaxLoRARanks = 16 """Max LoRA rank.""" max_loras: int = Field(default=1, ge=1) """Max number of LoRAs in a single batch.""" fully_sharded_loras: bool = False """By default, only half of the LoRA computation is sharded with tensor parallelism. Enabling this will use the fully sharded layers. At high sequence length, max rank or tensor parallel size, this is likely faster. """ max_cpu_loras: int | None = None """Maximum number of LoRAs to store in CPU memory. Must be >= than `max_loras`.""" lora_dtype: torch.dtype | LoRADType = "auto" """Data type for LoRA. If auto, will default to base model dtype.""" default_mm_loras: dict[str, str] | None = None """Dictionary mapping specific modalities to LoRA model paths; this field is only applicable to multimodal models and should be leveraged when a model always expects a LoRA to be active when a given modality is present. Note that currently, if a request provides multiple additional modalities, each of which have their own LoRA, we do NOT apply default_mm_loras because we currently only support one lora adapter per prompt. When run in offline mode, the lora IDs for n modalities will be automatically assigned to 1-n with the names of the modalities in alphabetic order.""" enable_tower_connector_lora: bool = False """If `True`, LoRA support for the tower (vision encoder) and connector of multimodal models will be enabled. This is an experimental feature and currently only supports some MM models such as the Qwen VL series. The default is False.""" def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ factors: list[Any] = [] factors.append(self.max_lora_rank) factors.append(self.max_loras) factors.append(self.fully_sharded_loras) factors.append(self.lora_dtype) factors.append(self.enable_tower_connector_lora) hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str @model_validator(mode="after") def _validate_lora_config(self) -> Self: if self.max_cpu_loras is None: self.max_cpu_loras = self.max_loras elif self.max_cpu_loras < self.max_loras: raise ValueError( f"max_cpu_loras ({self.max_cpu_loras}) must be >= " f"max_loras ({self.max_loras})" ) return self def verify_with_model_config(self, model_config: ModelConfig): if self.lora_dtype in (None, "auto"): self.lora_dtype = model_config.dtype elif isinstance(self.lora_dtype, str): self.lora_dtype = getattr(torch, self.lora_dtype)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/compilation.py
vllm/config/compilation.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import enum from collections import Counter from collections.abc import Callable from dataclasses import field from pathlib import Path from typing import TYPE_CHECKING, Any, ClassVar, Literal from pydantic import ConfigDict, Field, TypeAdapter, field_validator from pydantic.dataclasses import dataclass import vllm.envs as envs from vllm.compilation.inductor_pass import CallableInductorPass, InductorPass from vllm.config.utils import ( Range, config, get_hash_factors, hash_factors, ) from vllm.logger import init_logger from vllm.platforms import current_platform from vllm.utils.import_utils import resolve_obj_by_qualname from vllm.utils.math_utils import round_up from vllm.utils.torch_utils import is_torch_equal_or_newer if TYPE_CHECKING: from vllm.config import VllmConfig else: VllmConfig = object logger = init_logger(__name__) class CompilationMode(enum.IntEnum): """The compilation approach used for torch.compile-based compilation of the model.""" NONE = 0 """No torch.compile compilation is applied, model runs in fully eager pytorch mode. The model runs as-is.""" STOCK_TORCH_COMPILE = 1 """The standard `torch.compile` compilation pipeline.""" DYNAMO_TRACE_ONCE = 2 """Single Dynamo trace through the model, avoiding recompilation.""" VLLM_COMPILE = 3 """Custom vLLM Inductor-based backend with caching, piecewise compilation, shape specialization, and custom passes.""" class CUDAGraphMode(enum.Enum): """Constants for the cudagraph mode in CompilationConfig. Meanwhile, the subset enum `NONE`, `PIECEWISE` and `FULL` are also treated as concrete runtime mode for cudagraph runtime dispatching. """ NONE = 0 PIECEWISE = 1 FULL = 2 FULL_DECODE_ONLY = (FULL, NONE) FULL_AND_PIECEWISE = (FULL, PIECEWISE) def decode_mode(self) -> "CUDAGraphMode": return CUDAGraphMode(self.value[0]) if self.separate_routine() else self def mixed_mode(self) -> "CUDAGraphMode": return CUDAGraphMode(self.value[1]) if self.separate_routine() else self def has_mode(self, mode: "CUDAGraphMode") -> bool: assert not mode.separate_routine() if self.separate_routine(): return mode.value in self.value return self == mode def requires_piecewise_compilation(self) -> bool: return self.has_mode(CUDAGraphMode.PIECEWISE) def max_cudagraph_mode(self) -> "CUDAGraphMode": return CUDAGraphMode(max(self.value)) if self.separate_routine() else self def has_full_cudagraphs(self) -> bool: return self.max_cudagraph_mode() == CUDAGraphMode.FULL def has_piecewise_cudagraphs(self) -> bool: return self.requires_piecewise_compilation() def separate_routine(self) -> bool: return isinstance(self.value, tuple) def valid_runtime_modes(self) -> bool: return self in [CUDAGraphMode.NONE, CUDAGraphMode.PIECEWISE, CUDAGraphMode.FULL] def __str__(self) -> str: return self.name @config @dataclass(config=ConfigDict(extra="forbid")) class PassConfig: """Configuration for custom Inductor passes. This is separate from general `CompilationConfig` so that inductor passes don't all have access to full configuration - that would create a cycle as the `PassManager` is set as a property of config. You must pass PassConfig to VLLMConfig constructor via the CompilationConfig constructor. VLLMConfig's post_init does further initialization. If used outside of the VLLMConfig, some fields may be left in an improper state. """ # New flags fuse_norm_quant: bool = Field(default=None) """Fuse the custom RMSNorm + quant ops.""" fuse_act_quant: bool = Field(default=None) """Fuse the custom SiluMul + quant ops.""" fuse_attn_quant: bool = Field(default=None) """Fuse the custom attention + quant ops.""" eliminate_noops: bool = Field(default=None) """Eliminate no-op ops.""" enable_sp: bool = Field(default=None) """Enable sequence parallelism.""" fuse_gemm_comms: bool = Field(default=None) """Enable async TP.""" fuse_allreduce_rms: bool = Field(default=None) """Enable flashinfer allreduce fusion.""" fi_allreduce_fusion_max_size_mb: float | None = None """The threshold of the communicated tensor sizes under which vllm should use flashinfer fused allreduce. Specified as a float in MB. Unspecified will fallback to default values which are compute capability and world size dependent. FI_ALLREDUCE_FUSION_MAX_SIZE_MB = { 90: { 2: 64, # 64MB 4: 2, # 2MB 8: 1, # 1MB }, 100: { 2: 64, # 64MB 4: 32, # 32MB 8: 1, # 1MB }, }, where key is the device capability""" enable_qk_norm_rope_fusion: bool = False """Enable fused Q/K RMSNorm + RoPE pass.""" # TODO(luka) better pass enabling system. def flashinfer_max_size(self, world_size: int) -> int | None: """ Returns the max communication size in bytes for flashinfer allreduce fusion for the given world size. Returns None if world size is not supported by configs as it's not supported by flashinfer. """ MiB = 1024 * 1024 FI_SUPPORTED_WORLD_SIZES = [2, 4, 8] if world_size not in FI_SUPPORTED_WORLD_SIZES: return None max_size_mb = self.fi_allreduce_fusion_max_size_mb if max_size_mb is None: max_size_mb = self.default_fi_allreduce_fusion_max_size_mb().get(world_size) return int(max_size_mb * MiB) if max_size_mb is not None else None @staticmethod def default_fi_allreduce_fusion_max_size_mb() -> dict[int, float]: from vllm.compilation.collective_fusion import FI_ALLREDUCE_FUSION_MAX_SIZE_MB from vllm.platforms import current_platform if not current_platform.is_cuda(): return {} return FI_ALLREDUCE_FUSION_MAX_SIZE_MB.get( current_platform.get_device_capability().to_int(), {} ) def compute_hash(self) -> str: """ Produces a hash unique to the pass configuration. Any new fields that affect compilation should be added to the hash. Any future fields that don't affect compilation should be excluded. """ return hash_factors(get_hash_factors(self, set())) @field_validator( "fuse_norm_quant", "fuse_act_quant", "fuse_attn_quant", "eliminate_noops", "enable_sp", "fuse_gemm_comms", "fuse_allreduce_rms", mode="wrap", ) @classmethod def _skip_none_validation(cls, value: Any, handler: Callable) -> Any: """Skip validation if the value is `None` when initialisation is delayed.""" if value is None: return value return handler(value) def __post_init__(self) -> None: # Handle deprecation and defaults if not self.eliminate_noops: if self.fuse_norm_quant or self.fuse_act_quant: logger.warning_once( "Fusion enabled but reshape elimination disabled. " "RMSNorm/SiluMul + quant (fp8) fusion might not work" ) if self.fuse_attn_quant: logger.warning_once( "Fusion enabled but reshape elimination disabled. " "Attention + quant (fp8) fusion might not work" ) if self.fuse_allreduce_rms: logger.warning_once( "Fusion enabled but reshape elimination disabled. " "Allreduce + rms norm + quant (fp8) fusion might not work" ) if self.enable_qk_norm_rope_fusion and not current_platform.is_cuda_alike(): logger.warning_once( "QK Norm + RoPE fusion enabled but the current platform is not " "CUDA or ROCm. The fusion will be disabled." ) self.enable_qk_norm_rope_fusion = False class DynamicShapesType(str, enum.Enum): """Types of dynamic shapes handling in torch.compile(). see Dynamic shapes and vllm guard dropping in torch_compile.md for more details.""" BACKED = "backed" """Use backed dynamic shapes. torch.compile() guards on backed dynamic shapes and may add guards. Symbols are specialized to 0, 1, or >=2 even without encountering branching on those ranges.""" UNBACKED = "unbacked" """Use unbacked dynamic shapes. Guaranteed not to be guarded on and not 0/1 specialized, but may throw data dependent errors when branches require their value without explicit unbacked handling.""" BACKED_SIZE_OBLIVIOUS = "backed_size_oblivious" """Experimental flag that treats backed symbols as unbacked when explicit unbacked handling is defined.""" @config @dataclass(config=ConfigDict(extra="forbid")) class DynamicShapesConfig: """Configuration to control/debug torch compile dynamic shapes.""" type: DynamicShapesType = DynamicShapesType.BACKED """Controls the type of dynamic shapes handling to use with torch.compile(). - BACKED: Default PyTorch behavior with potential guards ignored. - UNBACKED: No guards guaranteed (most sound) but may throw data dependent errors. - BACKED_SIZE_OBLIVIOUS: Experimental safer alternative to backed/unbacked. """ evaluate_guards: bool = False """ A debug mode to detect and fail if Dynamo ever specializes a dynamic shape by guarding on it. When True, dynamic shape guards are not dropped from dynamo. And a failure will be triggered if a recompilation ever happens due to that. This mode requires VLLM_USE_BYTECODE_HOOK to be 0. Enabling this allow observing the dynamic shapes guards in the tlparse artifacts also. When type is backed, aot_compile must be disabled for this mode to work. until this change picked up https://github.com/pytorch/pytorch/pull/169239. """ def compute_hash(self) -> str: """ Provide a hash for DynamicShapesConfig """ from vllm.config.utils import get_hash_factors, hash_factors factors = get_hash_factors(self, {}) return hash_factors(factors) @config @dataclass(config=ConfigDict(extra="forbid")) class CompilationConfig: """Configuration for compilation. You must pass CompilationConfig to VLLMConfig constructor. VLLMConfig's post_init does further initialization. If used outside of the VLLMConfig, some fields will be left in an improper state. It has three parts: - Top-level Compilation control: - [`mode`][vllm.config.CompilationConfig.mode] - [`debug_dump_path`][vllm.config.CompilationConfig.debug_dump_path] - [`cache_dir`][vllm.config.CompilationConfig.cache_dir] - [`backend`][vllm.config.CompilationConfig.backend] - [`custom_ops`][vllm.config.CompilationConfig.custom_ops] - [`splitting_ops`][vllm.config.CompilationConfig.splitting_ops] - [`compile_mm_encoder`][vllm.config.CompilationConfig.compile_mm_encoder] - CudaGraph capture: - [`cudagraph_mode`][vllm.config.CompilationConfig.cudagraph_mode] - [`cudagraph_capture_sizes`] [vllm.config.CompilationConfig.cudagraph_capture_sizes] - [`max_cudagraph_capture_size`] [vllm.config.CompilationConfig.max_cudagraph_capture_size] - [`cudagraph_num_of_warmups`] [vllm.config.CompilationConfig.cudagraph_num_of_warmups] - [`cudagraph_copy_inputs`] [vllm.config.CompilationConfig.cudagraph_copy_inputs] - Inductor compilation: - [`compile_sizes`][vllm.config.CompilationConfig.compile_sizes] - [`compile_ranges_split_points`] [vllm.config.CompilationConfig.compile_ranges_split_points] - [`inductor_compile_config`] [vllm.config.CompilationConfig.inductor_compile_config] - [`inductor_passes`][vllm.config.CompilationConfig.inductor_passes] - custom inductor passes Why we have different sizes for cudagraph and inductor: - cudagraph: a cudagraph captured for a specific size can only be used for the same size. We need to capture all the sizes we want to use. - inductor: a graph compiled by inductor for a general shape can be used for different sizes. Inductor can also compile for specific sizes, where it can have more information to optimize the graph with fully static shapes. However, we find the general shape compilation is sufficient for most cases. It might be beneficial to compile for certain small batchsizes, where inductor is good at optimizing. """ # Top-level Compilation control level: int = Field(default=None) """ Level is deprecated and will be removed in the next release, either 0.12.0 or 0.11.2 whichever is soonest. Please use mode. Currently all levels are mapped to mode. """ # Top-level Compilation control mode: CompilationMode = Field(default=None) """The compilation approach used for torch.compile-based compilation of the model. - None: If None, we will select the default compilation mode. For V1 engine this is 3. - 0: NONE: No torch.compile compilation is applied, model runs in fully eager pytorch mode. The model runs as-is. - 1: STOCK_TORCH_COMPILE: The standard `torch.compile` compilation pipeline. - 2: DYNAMO_TRACE_ONCE: Single Dynamo trace through the model, avoiding recompilation by removing guards. Requires no dynamic-shape-dependent control-flow. - 3: VLLM_COMPILE: Custom vLLM Inductor-based backend with caching, piecewise compilation, shape specialization, and custom passes.""" debug_dump_path: Path | None = None """The path to dump the debug information.""" cache_dir: str = "" """The directory to store the compiled graph, to accelerate Inductor compilation. By default, it will use model-related information to generate a cache directory.""" compile_cache_save_format: Literal["binary", "unpacked"] = field( default_factory=lambda: envs.VLLM_COMPILE_CACHE_SAVE_FORMAT ) """Format for saving torch compile cache:\n - "binary": saves as binary file (multiprocess safe)\n - "unpacked": saves as directory structure for inspection/debugging (NOT multiprocess safe)\n Defaults to `VLLM_COMPILE_CACHE_SAVE_FORMAT` if not specified. """ backend: str = "" """The backend for compilation. It needs to be a string: - "" (empty string): use the default backend ("inductor" on CUDA-alike platforms). - "eager"/"openxla"/...: use the specified backend registered in PyTorch. - "full.module.name": a qualified name which can be used to import the backend function. We use string to avoid serialization issues when using compilation in a distributed setting. When the compilation mode is 1 or 2, the backend is used for the compilation directly (it sees the whole graph). When the compilation mode is 3, the backend supports both whole graph and piecewise compilation, available backends include eager, inductor, and custom backends, the latter of which can be defined via `get_compile_backend`. Furthermore, compilation is only piecewise if splitting ops is set accordingly and use_inductor_graph_partition is off. Note that the default options for splitting ops are sufficient for piecewise compilation. """ custom_ops: list[str] = field(default_factory=list) """Fine-grained control over which custom ops to enable/disable. Use 'all' to enable all, 'none' to disable all. Also specify a list of custom op names to enable (prefixed with a '+'), or disable (prefixed with a '-'). Examples: - 'all,-op1' to enable all except op1 - 'none,+op1,+op2' to enable only op1 and op2 By default, all custom ops are enabled when running without Inductor and disabled when running with Inductor: mode>=VLLM_COMPILE and backend="inductor". Inductor generates (fused) Triton kernels for disabled custom ops.""" splitting_ops: list[str] | None = None """A list of ops to exclude from cudagraphs, used in piecewise compilation. The behavior depends on use_inductor_graph_partition: - When use_inductor_graph_partition=False (default): These ops are used for Dynamo FX-level graph splitting. The graph is split at these ops before Inductor compilation, creating separate subgraphs for cudagraph capture. - When use_inductor_graph_partition=True: These ops are used to register Inductor partition rules. The graph partitioning happens at Inductor codegen time after all passes and fusions are finished, allowing compilation and custom passes to operate on the full graph while still excluding these ops from cudagraphs. If None, defaults to attention ops for piecewise cudagraphs. If empty list [], no ops are excluded (suitable for full cudagraphs).""" compile_mm_encoder: bool = False """Whether or not to compile the multimodal encoder. Currently, this only works for `Qwen2_5_vl` on selected platforms. Disabled by default until more models are supported/tested to work.""" # Inductor capture compile_sizes: list[int | str] | None = None """Sizes to compile for inductor. In addition to integers, it also supports "cudagraph_capture_sizes" to specify the sizes for cudagraph capture.""" compile_ranges_split_points: list[int] | None = None """Split points that represent compile ranges for inductor. The compile ranges are [1, split_points[0]], [split_points[0] + 1, split_points[1]], ..., [split_points[-1] + 1, max_num_batched_tokens]. Compile sizes are also used single element ranges, the range is represented as [compile_sizes[i], compile_sizes[i]]. If a range overlaps with the compile size, graph for compile size will be prioritized, i.e. if we have a range [1, 8] and a compile size 4, graph for compile size 4 will be compiled and used instead of the graph for range [1, 8]. """ inductor_compile_config: dict = field(default_factory=dict) """Additional configurations for inductor. - None: use default configurations.""" inductor_passes: dict[str, str] = field(default_factory=dict) """Additional passes for inductor. It is a dictionary from pass name to pass function qualified name. We use function name because the config uses JSON format. If we pass the config from Python, functions can also be passed directly via Python object constructor, e.g. `CompilationConfig(inductor_passes={"a": func})`.""" # CudaGraph compilation cudagraph_mode: CUDAGraphMode = Field(default=None) """ The mode of the cudagraph: - NONE, no cudagraph capture. - PIECEWISE. - FULL. - FULL_DECODE_ONLY. - FULL_AND_PIECEWISE. (v1 default) PIECEWISE mode build piecewise cudagraph only, keeping the cudagraph incompatible ops (i.e. some attention ops) outside the cudagraph for general flexibility. FULL mode: Capture full cudagraph for all batches. Can be good for small models or workloads with small prompts; not supported by many backends. Generally for performance FULL_AND_PIECEWISE is better. FULL_DECODE_ONLY mode: Capture full cudagraph for decode batches only. Mixed prefill-decode batches are run without cudagraphs. Can be good for decode instances in a P/D setup where prefill is not as important so we can save some memory. FULL_AND_PIECEWISE mode: Capture full cudagraph for decode batches and piecewise cudagraph for prefill and mixed prefill-decode batches. This is the most performant mode for most models and is the default. Currently, the cudagraph mode is only used for the v1 engine. Note that the cudagraph logic is generally orthogonal to the compilation logic. While piecewise cudagraphs require piecewise compilation (mode=VLLM_COMPILE and non-empty splitting_ops), full cudagraphs are supported with and without compilation. Warning: This flag is new and subject to change in addition more modes may be added. """ cudagraph_num_of_warmups: int = 0 """Number of warmup runs for cudagraph. It means the first several runs will be treated as warmup runs. Only after that, the execution will be recorded, and the recorded cudagraph will be used for subsequent runs.""" cudagraph_capture_sizes: list[int] | None = None """Sizes to capture cudagraph. - None (default): capture sizes are inferred from vllm config. - list[int]: capture sizes are specified as given.""" cudagraph_copy_inputs: bool = False """Whether to copy input tensors for cudagraph. If the caller can guarantee that the same input buffers are always used, it can set this to False. Otherwise, it should set this to True, and the compiler will copy the input to an internally managed buffer. Default is False. Note that this flag is only effective when cudagraph_mode is PIECEWISE. """ cudagraph_specialize_lora: bool = True """Whether to create separate cuda graphs for cases with and without active LoRA adapters. When set to False, the LoRA-enabled cuda graph will be used for all cases, incurring the overhead of running LoRA ops even when no adapters are active. Setting this to True will remove this overhead at the cost of increased startup time and slightly higher memory usage. When `enable_lora` is False, this option has no effect. """ use_inductor_graph_partition: bool = Field(default=None) """Use inductor graph partition to split the graph at cudagraph_unsafe ops. This partition happens at inductor codegen time after all passes and fusions are finished. It generates a single `call` function which wraps cudagraph-safe ops into partition functions and leave cudagraph-unsafe ops outside the partition functions. For a graph with N cudagraph-unsafe ops (e.g., Attention), there would be N+1 partitions. To mark an op as cudagraph unsafe, we can add `tags=(torch._C.Tag.cudagraph_unsafe)` when register the custom op. This config supports both full cudagraph and piecewise cudagraph without compiling twice. For piecewise cudagraph, it applies vLLM CUDAGraph wrapper to each partition. For N+1 partitions, there would be N+1 CUDAGraph wrapper instances. For full CUDAGraph, we always apply a single CUDAGraph wrapper outside the inductor `call` function in the model runner. The top-level full cudagraph capture ignores all partitioning. """ pass_config: PassConfig = field(default_factory=PassConfig) """Custom inductor passes, see PassConfig for more details""" max_cudagraph_capture_size: int | None = field(default=None) """The maximum cudagraph capture size. If cudagraph_capture_sizes is specified, this will be set to the largest size in that list (or checked for consistency if specified). If cudagraph_capture_sizes is not specified, the list of sizes is generated automatically following the pattern: [1, 2, 4] + list(range(8, 256, 8)) + list( range(256, max_cudagraph_capture_size + 1, 16)) If not specified, max_cudagraph_capture_size is set to min(max_num_seqs*2, 512) by default. This voids OOM in tight memory scenarios with small max_num_seqs, and prevents capture of many large graphs (>512) that would greatly increase startup time with limited performance benefit. """ dynamic_shapes_config: DynamicShapesConfig = field( default_factory=DynamicShapesConfig ) """Configuration for dynamic shapes options""" local_cache_dir: str = field(default=None, init=False) # type: ignore """local cache dir for each rank""" bs_to_padded_graph_size: list[int] = field( default=None, # type: ignore init=False, ) """optimization: Intuitively, bs_to_padded_graph_size should be dict[int, int]. since we know all keys are in a range [0, max_cudagraph_capture_size], we can optimize it to list[int] for better lookup performance.""" # keep track of enabled and disabled custom ops enabled_custom_ops: Counter[str] = field(default_factory=Counter, init=False) """custom ops that are enabled""" disabled_custom_ops: Counter[str] = field(default_factory=Counter, init=False) """custom ops that are disabled""" traced_files: set[str] = field(default_factory=set, init=False) """files that are traced for compilation""" compilation_time: float = field(default=0.0, init=False) """time taken for compilation""" static_forward_context: dict[str, Any] = field(default_factory=dict, init=False) """Per-model forward context Map from layer name to layer objects that need to be accessed outside model code, e.g., Attention, FusedMOE when dp_size>1.""" # Attention ops; used for piecewise cudagraphs # Use PyTorch operator format: "namespace::name" _attention_ops: ClassVar[list[str]] = [ "vllm::unified_attention", "vllm::unified_attention_with_output", "vllm::unified_mla_attention", "vllm::unified_mla_attention_with_output", "vllm::mamba_mixer2", "vllm::mamba_mixer", "vllm::short_conv", "vllm::linear_attention", "vllm::plamo2_mamba_mixer", "vllm::gdn_attention_core", "vllm::kda_attention", "vllm::sparse_attn_indexer", ] def compute_hash(self) -> str: """ Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # Opt-out: default-include declared fields; keep a tiny exclude set; # normalize types; keep SHA-256. For nested opaque configs, include a # stable identifier (e.g., pass_config.compute_hash()) instead of object id. ignored_factors = { # Paths/dirs and runtime/metrics that don’t affect compiled graph "debug_dump_path", "cache_dir", "local_cache_dir", "bs_to_padded_graph_size", "traced_files", "compilation_time", "static_forward_context", "pass_config", # handled separately below } from vllm.config.utils import get_hash_factors, hash_factors factors = get_hash_factors(self, ignored_factors) factors["pass_config"] = self.pass_config.compute_hash() return hash_factors(factors) def __repr__(self) -> str: exclude = { "static_forward_context": True, "enabled_custom_ops": True, "disabled_custom_ops": True, "compilation_time": True, "bs_to_padded_graph_size": True, "traced_files": True, "inductor_compile_config": { "post_grad_custom_post_pass": True, }, } # exclude default attr in pass_config pass_config_exclude = {} for attr, default_val in vars(PassConfig()).items(): if getattr(self.pass_config, attr) == default_val: pass_config_exclude[attr] = True if pass_config_exclude: exclude["pass_config"] = pass_config_exclude config = TypeAdapter(CompilationConfig).dump_python( self, exclude=exclude, exclude_unset=True ) return str(config) __str__ = __repr__ @field_validator("mode", mode="before") @classmethod def validate_mode_before(cls, value: Any) -> Any: """ Enable parsing the `mode` field from string mode names. Accepts both integers (0-3) and string names, like NONE, STOCK_TORCH_COMPILE, DYNAMO_TRACE_ONCE, VLLM_COMPILE. """ if isinstance(value, str): # Convert string mode name to integer value mode_name = value.upper() if mode_name not in CompilationMode.__members__: raise ValueError( f"Invalid compilation mode: {value}. " f"Valid modes are: {', '.join(CompilationMode.__members__.keys())}" ) return CompilationMode[mode_name] return value @field_validator("cudagraph_mode", mode="before") @classmethod def validate_cudagraph_mode_before(cls, value: Any) -> Any: """Enable parsing of the `cudagraph_mode` enum type from string.""" if isinstance(value, str): return CUDAGraphMode[value.upper()] return value @field_validator("pass_config", mode="before") @classmethod def validate_pass_config_before(cls, value: Any) -> Any: """Enable parsing of the `pass_config` field from a dictionary.""" if isinstance(value, dict): return PassConfig(**value) return value @field_validator("compile_cache_save_format") @classmethod def validate_compile_cache_save_format(cls, value: str) -> str: if value not in ("binary", "unpacked"): raise ValueError( f"compile_cache_save_format must be 'binary' or 'unpacked', " f"got: {value}" ) return value @field_validator( "level", "mode", "cudagraph_mode", "use_inductor_graph_partition", mode="wrap", ) @classmethod def _skip_none_validation(cls, value: Any, handler: Callable) -> Any: """Skip validation if the value is `None` when initialisation is delayed.""" if value is None: return value return handler(value) def __post_init__(self) -> None: if self.level is not None: logger.warning( "Level is deprecated and will be removed in the next release," "either 0.12.0 or 0.11.2 whichever is soonest." "Use mode instead." "If both level and mode are given," "only mode will be used." ) if self.mode is None: self.mode = self.level count_none = self.custom_ops.count("none") count_all = self.custom_ops.count("all") assert count_none + count_all <= 1, "Can only specify 'none' or 'all'" # TODO(zou3519/luka): There are 2 issues with auto-functionalization V2: # 1. A bug in PyTorch, fixed in 2.7: # https://github.com/pytorch/pytorch/issues/147924 # 2. Custom passes (fusion) rely on auto-functionalization V1 and don't # work with V2. Addressing this will take extra engineering effort # and it is not yet a priority. RFC here: # https://github.com/vllm-project/vllm/issues/14703 if is_torch_equal_or_newer("2.6"): KEY = "enable_auto_functionalized_v2" if KEY not in self.inductor_compile_config: self.inductor_compile_config[KEY] = False for k, v in self.inductor_passes.items(): if not isinstance(v, str): assert callable(v), f"pass {k} should be callable or a qualified name" self.inductor_compile_config[k] = ( v if isinstance(v, InductorPass) else CallableInductorPass(v) ) continue # resolve function from qualified name names = v.split(".") module = ".".join(names[:-1]) func_name = names[-1] func = __import__(module).__dict__[func_name] self.inductor_compile_config[k] = ( func if isinstance(func, InductorPass) else CallableInductorPass(func) ) if self.pass_config.enable_qk_norm_rope_fusion: # TODO(zhuhaoran): support rope native forward match and remove this. # Linked issue: https://github.com/vllm-project/vllm/issues/28042 self.custom_ops.append("+rotary_embedding") if ( is_torch_equal_or_newer("2.9.0.dev") and "combo_kernels" not in self.inductor_compile_config
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/ec_transfer.py
vllm/config/ec_transfer.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import hashlib import uuid from dataclasses import field from typing import Any, Literal, get_args from pydantic.dataclasses import dataclass from vllm.config.utils import config ECProducer = Literal["ec_producer"] ECConsumer = Literal["ec_consumer"] ECRole = Literal[ECProducer, ECConsumer] @config @dataclass class ECTransferConfig: """Configuration for distributed EC cache transfer.""" ec_connector: str | None = None """The EC connector for vLLM to transmit EC caches between vLLM instances. """ engine_id: str | None = None """The engine id for EC transfers.""" ec_buffer_device: str | None = "cuda" """The device used by ec connector to buffer the EC cache. Currently only support 'cuda'.""" ec_buffer_size: float = 1e9 """The buffer size for TorchDistributedConnector. Measured in number of bytes. Recommended value: 1e9 (about 1GB).""" ec_role: ECRole | None = None """Whether this vLLM instance produces, consumes EC cache, or both. Choices are 'ec_producer', 'ec_consumer'.""" ec_rank: int | None = None """The rank of this vLLM instance in the EC cache transfer. Typical value: 0 for encoder, 1 for pd instance. Currently only 1P1D is supported.""" ec_parallel_size: int = 1 """The number of parallel instances for EC cache transfer. For PyNcclConnector, this should be 2.""" ec_ip: str = "127.0.0.1" """The EC connector ip, used to build distributed connection.""" ec_port: int = 14579 """The EC connector port, used to build distributed connection.""" ec_connector_extra_config: dict[str, Any] = field(default_factory=dict) """any extra config that the connector may need.""" ec_connector_module_path: str | None = None """The Python module path to dynamically load the EC connector from. Only supported in V1.""" def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # no factors to consider. # this config will not affect the computation graph. factors: list[Any] = [] hash_str = hashlib.md5(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str def __post_init__(self) -> None: if self.engine_id is None: self.engine_id = str(uuid.uuid4()) if self.ec_role is not None and self.ec_role not in get_args(ECRole): raise ValueError( f"Unsupported ec_role: {self.ec_role}. " f"Supported roles are {get_args(ECRole)}" ) if self.ec_connector is not None and self.ec_role is None: raise ValueError( "Please specify ec_role when ec_connector " f"is set, supported roles are {get_args(ECRole)}" ) @property def is_ec_transfer_instance(self) -> bool: return self.ec_connector is not None and self.ec_role in get_args(ECRole) @property def is_ec_producer(self) -> bool: return self.ec_connector is not None and self.ec_role in get_args(ECProducer) @property def is_ec_consumer(self) -> bool: return self.ec_connector is not None and self.ec_role in get_args(ECConsumer) def get_from_extra_config(self, key, default) -> Any: return self.ec_connector_extra_config.get(key, default)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/utils.py
vllm/config/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Utility functions for vLLM config dataclasses.""" import ast import enum import hashlib import inspect import json import pathlib import textwrap from collections.abc import Callable, Iterable, Mapping, Sequence, Set from dataclasses import MISSING, Field, dataclass, field, fields, is_dataclass, replace from itertools import pairwise from typing import TYPE_CHECKING, Any, Protocol, TypeVar import regex as re import torch from pydantic.fields import FieldInfo from typing_extensions import runtime_checkable from vllm.logger import init_logger logger = init_logger(__name__) if TYPE_CHECKING: from _typeshed import DataclassInstance else: DataclassInstance = Any ConfigType = type[DataclassInstance] ConfigT = TypeVar("ConfigT", bound=ConfigType) def config(cls: ConfigT) -> ConfigT: """ A decorator that ensures all fields in a dataclass have default values and that each field has a docstring. If a `ConfigT` is used as a CLI argument itself, the `type` keyword argument provided by `get_kwargs` will be `pydantic.TypeAdapter(ConfigT).validate_json(cli_arg)` which treats the `cli_arg` as a JSON string which gets validated by `pydantic`. Config validation is performed by the tools/pre_commit/validate_config.py script, which is invoked during the pre-commit checks. """ return cls def get_field(cls: ConfigType, name: str) -> Field: """Get the default factory field of a dataclass by name. Used for getting default factory fields in `EngineArgs`.""" if not is_dataclass(cls): raise TypeError("The given class is not a dataclass.") cls_fields = {f.name: f for f in fields(cls)} if name not in cls_fields: raise ValueError(f"Field '{name}' not found in {cls.__name__}.") named_field: Field = cls_fields[name] if (default_factory := named_field.default_factory) is not MISSING: return field(default_factory=default_factory) if (default := named_field.default) is not MISSING: if isinstance(default, FieldInfo): # Handle pydantic.Field defaults if default.default_factory is not None: return field(default_factory=default.default_factory) else: default = default.default return field(default=default) raise ValueError( f"{cls.__name__}.{name} must have a default value or default factory." ) def getattr_iter( object: object, names: Iterable[str], default: Any | None = None, default_factory: Callable[[], Any] | None = None, warn: bool = False, ) -> Any: """ A helper function that retrieves an attribute from an object which may have multiple possible names. This is useful when fetching attributes from arbitrary `transformers.PretrainedConfig` instances. In the case where the first name in `names` is the preferred name, and any other names are deprecated aliases, setting `warn=True` will log a warning when a deprecated name is used. """ for i, name in enumerate(names): if hasattr(object, name): if warn and i > 0: logger.warning_once( "%s contains a deprecated attribute name '%s'. " "Please use the preferred attribute name '%s' instead.", type(object).__name__, name, names[0], ) return getattr(object, name) return default_factory() if default_factory is not None else default def contains_object_print(text: str) -> bool: """ Check if the text looks like a printed Python object, e.g. contains any substring matching the pattern: "at 0xFFFFFFF>" We match against 0x followed by 2-16 hex chars (there's a max of 16 on a 64-bit system). Args: text (str): The text to check Returns: result (bool): `True` if a match is found, `False` otherwise. """ pattern = r"at 0x[a-fA-F0-9]{2,16}>" match = re.search(pattern, text) return match is not None def assert_hashable(text: str) -> bool: if not contains_object_print(text): return True raise AssertionError( f"vLLM tried to hash some configs that may have Python objects ids " f"in them. This is a bug, please file an issue. " f"Text being hashed: {text}" ) def get_attr_docs(cls: type[Any]) -> dict[str, str]: """ Get any docstrings placed after attribute assignments in a class body. https://davidism.com/mit-license/ """ cls_node = ast.parse(textwrap.dedent(inspect.getsource(cls))).body[0] if not isinstance(cls_node, ast.ClassDef): raise TypeError("Given object was not a class.") out = {} # Consider each pair of nodes. for a, b in pairwise(cls_node.body): # Must be an assignment then a constant string. if ( not isinstance(a, (ast.Assign, ast.AnnAssign)) or not isinstance(b, ast.Expr) or not isinstance(b.value, ast.Constant) or not isinstance(b.value.value, str) ): continue doc = inspect.cleandoc(b.value.value) # An assignment can have multiple targets (a = b = v), but an # annotated assignment only has one target. targets = a.targets if isinstance(a, ast.Assign) else [a.target] for target in targets: # Must be assigning to a plain name. if not isinstance(target, ast.Name): continue out[target.id] = doc return out def is_init_field(cls: ConfigType, name: str) -> bool: return next(f for f in fields(cls) if f.name == name).init @runtime_checkable class SupportsHash(Protocol): def compute_hash(self) -> str: ... class SupportsMetricsInfo(Protocol): def metrics_info(self) -> dict[str, str]: ... def update_config(config: ConfigT, overrides: dict[str, Any]) -> ConfigT: processed_overrides = {} for field_name, value in overrides.items(): assert hasattr(config, field_name), ( f"{type(config)} has no field `{field_name}`" ) current_value = getattr(config, field_name) if is_dataclass(current_value) and not is_dataclass(value): assert isinstance(value, dict), ( f"Overrides to {type(config)}.{field_name} must be a dict" f" or {type(current_value)}, but got {type(value)}" ) value = update_config( current_value, # type: ignore[type-var] value, ) processed_overrides[field_name] = value return replace(config, **processed_overrides) def normalize_value(x): """Return a stable, JSON-serializable canonical form for hashing. Order: primitives, special types (Enum, callable, torch.dtype, Path), then generic containers (Mapping/Set/Sequence) with recursion. """ # Fast path if x is None or isinstance(x, (bool, int, float, str)): return x # Enums: tag with FQN to avoid primitive collisions. # Ex: Enum(1) vs int(1) -> ("module.QualName", value). if isinstance(x, enum.Enum): enum_type = f"{x.__class__.__module__}.{x.__class__.__qualname__}" return (enum_type, normalize_value(x.value)) # Classes (types) are accepted and canonicalized by their fully-qualified # name (module.qualname) for a stable identifier. # Instances are only accepted if they expose uuid(); otherwise they are # rejected to avoid under-hashing object state. # Callables: accept classes only; reject funcs/lambdas/methods. # Used by LogitsProcessor types and ModelConfig.hf_overrides. if isinstance(x, type): module = getattr(x, "__module__", "") qual = getattr(x, "__qualname__", getattr(x, "__name__", "")) return ".".join([p for p in (module, qual) if p]) or repr(x) # Prefer stable uuid identifiers for objects that provide them, even if # they are callable instances (e.g., InductorPass wrappers). if hasattr(x, "uuid") and callable(getattr(x, "uuid", None)): return x.uuid() if callable(x): raise TypeError("normalize_value: function or callable instance unsupported") # Torch dtype: stringify (torch.float64 -> "torch.float64"). # We rely on the string form here; dtype-bearing fields that need additional # disambiguation should encode that at the config layer. if isinstance(x, torch.dtype): return str(x) # Bytes if isinstance(x, (bytes, bytearray)): return x.hex() # Paths (canonicalize) if isinstance(x, pathlib.Path): try: return str(x.expanduser().resolve()) except Exception: return str(x) # Dataclasses: represent as (FQN, sorted(field,value) tuple) for stability. if is_dataclass(x): type_fqn = f"{x.__class__.__module__}.{x.__class__.__qualname__}" items = tuple( (f.name, normalize_value(getattr(x, f.name))) for f in sorted(fields(x), key=lambda f: f.name) ) return (type_fqn, items) # Containers (generic) if isinstance(x, Mapping): return tuple(sorted((str(k), normalize_value(v)) for k, v in x.items())) if isinstance(x, Set): return tuple(sorted(repr(normalize_value(v)) for v in x)) if isinstance(x, Sequence) and not isinstance(x, (str, bytes, bytearray)): return tuple(normalize_value(v) for v in x) # PretrainedConfig if hasattr(x, "to_json_string") and callable(x.to_json_string): return x.to_json_string() # Unsupported type: e.g., modules, generators, open files, or objects # without a stable JSON/UUID representation. Hard-error to avoid # under-hashing. # If you hit this, either reshape your config to use supported primitives # and containers, or extend normalize_value to provide a stable encoding # (e.g., via uuid() or to_json_string()) for this type. raise TypeError( f"normalize_value: unsupported type '{type(x).__name__}'. " "Ensure config values use supported primitives/containers or add a " "stable representation for this type." ) def get_hash_factors(config: ConfigT, ignored_factors: set[str]) -> dict[str, object]: """Gets the factors used for hashing a config class. - Includes all dataclass fields not in `ignored_factors`. - Errors on non-normalizable values. """ factors: dict[str, object] = {} for dc_field in fields(config): factor = dc_field.name if factor in ignored_factors: continue value = getattr(config, factor, None) try: factors[factor] = normalize_value(value) except TypeError as e: raise TypeError( f"get_hash_factors: unsupported type for key '{factor}' " f"({type(value).__name__})" ) from e return factors def hash_factors(items: dict[str, object]) -> str: """Return a SHA-256 hex digest of the canonical items structure.""" return hashlib.sha256(json.dumps(items, sort_keys=True).encode()).hexdigest() def handle_deprecated( config: ConfigT, old_name: str, new_name_or_names: str | list[str], removal_version: str, ) -> None: old_val = getattr(config, old_name) if old_val is None: return if isinstance(new_name_or_names, str): new_names = [new_name_or_names] else: new_names = new_name_or_names msg = ( f"{old_name} is deprecated and will be removed in {removal_version}. " f"Use {', '.join(new_names)} instead." ) logger.warning(msg) for new_name in new_names: setattr(config, new_name, old_val) @dataclass class Range: """ A range of numbers. Inclusive of start, inclusive of end. """ start: int end: int def is_single_size(self) -> bool: return self.start == self.end def __contains__(self, size: int) -> bool: # Inclusive of start, inclusive of end return self.start <= size <= self.end def __eq__(self, other: object) -> bool: if not isinstance(other, Range): return False return self.start == other.start and self.end == other.end def __hash__(self) -> int: return hash((self.start, self.end)) def __str__(self) -> str: return f"({self.start}, {self.end})" def __repr__(self) -> str: return self.__str__()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/speculative.py
vllm/config/speculative.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import ast from typing import TYPE_CHECKING, Any, Literal, get_args from pydantic import Field, SkipValidation, model_validator from pydantic.dataclasses import dataclass from typing_extensions import Self from vllm.config.model import ModelConfig from vllm.config.parallel import ParallelConfig from vllm.config.utils import config from vllm.logger import init_logger from vllm.utils.hashing import safe_hash from vllm.utils.import_utils import LazyLoader, has_arctic_inference if TYPE_CHECKING: from transformers import PretrainedConfig import vllm.model_executor.layers.quantization as me_quant else: PretrainedConfig = Any me_quant = LazyLoader( "model_executor", globals(), "vllm.model_executor.layers.quantization" ) logger = init_logger(__name__) MTPModelTypes = Literal[ "deepseek_mtp", "mimo_mtp", "glm4_moe_mtp", "ernie_mtp", "qwen3_next_mtp", "longcat_flash_mtp", "mtp", "pangu_ultra_moe_mtp", ] EagleModelTypes = Literal["eagle", "eagle3", MTPModelTypes] SpeculativeMethod = Literal[ "ngram", "medusa", "mlp_speculator", "draft_model", "suffix", EagleModelTypes, ] @config @dataclass class SpeculativeConfig: """Configuration for speculative decoding.""" enforce_eager: bool | None = None """Override the default enforce_eager from model_config""" # General speculative decoding control num_speculative_tokens: int = Field(default=None, gt=0) """The number of speculative tokens, if provided. It will default to the number in the draft model config if present, otherwise, it is required.""" model: str | None = None """The name of the draft model, eagle head, or additional weights, if provided.""" method: SpeculativeMethod | None = None """The name of the speculative method to use. If users provide and set the `model` param, the speculative method type will be detected automatically if possible, if `model` param is not provided, the method name must be provided. If using `ngram` method, the related configuration `prompt_lookup_max` and `prompt_lookup_min` should be considered.""" draft_tensor_parallel_size: int | None = Field(default=None, ge=1) """The degree of the tensor parallelism for the draft model. Can only be 1 or the same as the target model's tensor parallel size.""" # Draft model configuration quantization: me_quant.QuantizationMethods | None = None """Quantization method that was used to quantize the draft model weights. If `None`, we assume the model weights are not quantized. Note that it only takes effect when using the draft model-based speculative method.""" max_model_len: int | None = Field(default=None, ge=1) """The maximum model length of the draft model. Used when testing the ability to skip speculation for some sequences.""" revision: str | None = None """The specific model version to use for the draft model. It can be a branch name, a tag name, or a commit id. If unspecified, will use the default version.""" code_revision: str | None = None """The specific revision to use for the draft model code on Hugging Face Hub. It can be a branch name, a tag name, or a commit id. If unspecified, will use the default version.""" # Advanced control disable_by_batch_size: int | None = Field(default=None, ge=2) """Disable speculative decoding for new incoming requests when the number of enqueued requests is larger than this value, if provided.""" disable_padded_drafter_batch: bool = False """Disable input padding for speculative decoding. If set to True, speculative input batches can contain sequences of different lengths, which may only be supported by certain attention backends. This currently only affects the EAGLE method of speculation.""" # Ngram proposer configuration prompt_lookup_max: int | None = Field(default=None, ge=1) """Maximum size of ngram token window when using Ngram proposer, required when method is set to ngram.""" prompt_lookup_min: int | None = Field(default=None, ge=1) """Minimum size of ngram token window when using Ngram proposer, if provided. Defaults to 1.""" speculative_token_tree: str | None = None """Specifies the tree structure for speculative token generation. """ # required configuration params passed from engine target_model_config: SkipValidation[ModelConfig] = None # type: ignore """The configuration of the target model.""" target_parallel_config: SkipValidation[ParallelConfig] = None # type: ignore """The parallel configuration for the target model.""" # params generated in the post-init stage draft_model_config: SkipValidation[ModelConfig] = None # type: ignore """The configuration of the draft model initialized internal.""" draft_parallel_config: SkipValidation[ParallelConfig] = None # type: ignore """The parallel configuration for the draft model initialized internal.""" # Suffix decoding configuration suffix_decoding_max_tree_depth: int = 24 """The maximum depth of the suffix decoding global and prompt trees. The tree depth limits the sum of the prefix match and speculation lengths.""" suffix_decoding_max_cached_requests: int = 10000 """The maximum number of requests to cache in the global suffix tree. If exceeded, will trigger eviction in FIFO order. If set to 0, the global suffix tree is disabled and past responses are not cached (prompt trees are still used).""" suffix_decoding_max_spec_factor: float = 1.0 """The maximum spec factor for suffix decoding. The spec factor controls speculation lengths based on the prefix match length: max_spec_tokens = max_spec_factor * prefix_match_length.""" suffix_decoding_min_token_prob: float = 0.1 """The minimum token probability for suffix decoding. Will only speculate tokens with estimated probability (based on frequency counts) greater than or equal to this value.""" def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ factors: list[Any] = [] # Eagle3 affects the computation graph because it returns intermediate # hidden states in addition to the final hidden state. factors.append(self.method == "eagle3") hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str @staticmethod def hf_config_override(hf_config: PretrainedConfig) -> PretrainedConfig: initial_architecture = hf_config.architectures[0] if hf_config.model_type in ("deepseek_v3", "deepseek_v32"): hf_config.model_type = "deepseek_mtp" if hf_config.model_type == "deepseek_mtp": n_predict = getattr(hf_config, "num_nextn_predict_layers", None) hf_config.update( {"n_predict": n_predict, "architectures": ["DeepSeekMTPModel"]} ) if hf_config.model_type in ("pangu_ultra_moe"): hf_config.model_type = "pangu_ultra_moe_mtp" if hf_config.model_type == "pangu_ultra_moe_mtp": n_predict = getattr(hf_config, "num_nextn_predict_layers", None) hf_config.update( {"n_predict": n_predict, "architectures": ["OpenPanguMTPModel"]} ) if hf_config.architectures[0] == "MiMoForCausalLM": hf_config.model_type = "mimo_mtp" n_predict = getattr(hf_config, "num_nextn_predict_layers", None) hf_config.update( { "num_hidden_layers": 0, "n_predict": n_predict, "architectures": ["MiMoMTPModel"], } ) if hf_config.architectures[0] == "Glm4MoeForCausalLM": hf_config.model_type = "glm4_moe_mtp" n_predict = getattr(hf_config, "num_nextn_predict_layers", None) hf_config.update( { "num_hidden_layers": 0, "n_predict": n_predict, "architectures": ["Glm4MoeMTPModel"], } ) if hf_config.model_type == "ernie4_5_moe": hf_config.model_type = "ernie_mtp" if hf_config.model_type == "ernie_mtp": n_predict = getattr(hf_config, "num_nextn_predict_layers", None) hf_config.update( {"n_predict": n_predict, "architectures": ["ErnieMTPModel"]} ) if hf_config.model_type == "qwen3_next": hf_config.model_type = "qwen3_next_mtp" if hf_config.model_type == "qwen3_next_mtp": n_predict = getattr(hf_config, "num_nextn_predict_layers", None) hf_config.update( {"n_predict": n_predict, "architectures": ["Qwen3NextMTP"]} ) if hf_config.model_type == "longcat_flash": hf_config.model_type = "longcat_flash_mtp" n_predict = getattr(hf_config, "num_nextn_predict_layers", 1) hf_config.update( {"n_predict": n_predict, "architectures": ["LongCatFlashMTPModel"]} ) if initial_architecture == "MistralLarge3ForCausalLM": hf_config.update({"architectures": ["EagleMistralLarge3ForCausalLM"]}) return hf_config def __post_init__(self): # Note: "method" is a new parameter that helps to extend the # configuration of non-model-based proposers, and the "model" parameter # will be used to set the draft model, eagle head, or additional weight # when needed. If users do not specify "method", the speculative method # will be detected automatically if possible. If the speculative method # can not be detected, it will be considered as the "draft_model" by # default. if self.method in get_args(MTPModelTypes) and self.method != "mtp": logger.warning( "method `%s` is deprecated and replaced with mtp.", self.method ) self.method = "mtp" if self.model is None and self.num_speculative_tokens is not None: if self.method == "mtp": if self.target_model_config is None: raise ValueError("target_model_config must be present for mtp") if self.target_model_config.hf_text_config.model_type == "deepseek_v32": # FIXME(luccafong): cudgraph with v32 MTP is not supported, # remove this when the issue is fixed. self.enforce_eager = True # use the draft model from the same model: self.model = self.target_model_config.model # Align the quantization of draft model for cases such as # --quantization fp8 with a bf16 checkpoint. if not self.quantization: self.quantization = self.target_model_config.quantization elif self.method in ("ngram", "[ngram]"): self.model = "ngram" elif self.method == "suffix": self.model = "suffix" else: raise ValueError( "num_speculative_tokens was provided but without speculative model." ) # Automatically configure the method for ngram when "model" is used # instead of "method" if self.method is None and ( self.model is not None and self.model in ("ngram", "[ngram]") ): self.method = "ngram" if self.method in ("ngram", "[ngram]"): # Unified to "ngram" internally self.method = "ngram" # Set default values if not provided if self.prompt_lookup_min is None and self.prompt_lookup_max is None: # TODO(woosuk): Tune these values. They are arbitrarily chosen. self.prompt_lookup_min = 5 self.prompt_lookup_max = 5 elif self.prompt_lookup_min is None: if self.prompt_lookup_max is None: raise ValueError( "Either prompt_lookup_max or prompt_lookup_min must be " "provided when using the ngram method." ) self.prompt_lookup_min = self.prompt_lookup_max elif self.prompt_lookup_max is None: if self.prompt_lookup_min is None: raise ValueError( "Either prompt_lookup_max or prompt_lookup_min must be " "provided when using the ngram method." ) self.prompt_lookup_max = self.prompt_lookup_min # Validate values if self.prompt_lookup_min > self.prompt_lookup_max: raise ValueError( f"prompt_lookup_min={self.prompt_lookup_min} must " f"be <= prompt_lookup_max={self.prompt_lookup_max}" ) # TODO: current we still need extract vocab_size from target model # config, in future, we may try refactor it out, and set # draft related config as None here. self.draft_model_config = self.target_model_config self.draft_parallel_config = self.target_parallel_config elif self.method == "suffix": self._validate_suffix_decoding() else: self.prompt_lookup_max = 0 self.prompt_lookup_min = 0 if self.model is not None: self.draft_model_config = ModelConfig( model=self.model, runner="draft", tokenizer=self.target_model_config.tokenizer, tokenizer_mode=self.target_model_config.tokenizer_mode, trust_remote_code=self.target_model_config.trust_remote_code, allowed_local_media_path=self.target_model_config.allowed_local_media_path, allowed_media_domains=self.target_model_config.allowed_media_domains, dtype=self.target_model_config.dtype, seed=self.target_model_config.seed, revision=self.revision, code_revision=self.code_revision, tokenizer_revision=self.target_model_config.tokenizer_revision, spec_target_max_model_len=self.target_model_config.max_model_len, quantization=self.quantization, enforce_eager=self.target_model_config.enforce_eager, max_logprobs=self.target_model_config.max_logprobs, hf_overrides=SpeculativeConfig.hf_config_override, config_format=self.target_model_config.config_format, ) # Automatically detect the method if self.method in ("eagle", "eagle3"): pass # examples: # yuhuili/EAGLE-LLaMA3-Instruct-8B # yuhuili/EAGLE3-LLaMA3.1-Instruct-8B # AngelSlim/Qwen3-8B_eagle3 elif "eagle-" in self.draft_model_config.model.lower(): self.method = "eagle" elif "eagle3" in self.draft_model_config.model.lower(): self.method = "eagle3" elif self.draft_model_config.hf_config.model_type == "medusa": self.method = "medusa" elif self.draft_model_config.hf_config.model_type == "mlp_speculator": self.method = "mlp_speculator" elif self.draft_model_config.hf_config.model_type in get_args( MTPModelTypes ): self.method = "mtp" if self.num_speculative_tokens > 1: logger.warning( "Enabling num_speculative_tokens > 1 will run" "multiple times of forward on same MTP layer" ",which may result in lower acceptance rate" ) elif self.draft_model_config.hf_config.model_type in ( "longcat_flash_mtp" ): self.method = "longcat_flash_mtp" if self.num_speculative_tokens > 1: logger.warning( "LongCat MTP models only have " "one layer. Might need some code changes " "to support multiple layers." ) else: self.method = "draft_model" raise NotImplementedError( "Speculative decoding with draft model is not " "supported yet. Please consider using other " "speculative decoding methods such as ngram, medusa, " "eagle, or mtp." ) # Replace hf_config for EAGLE draft_model if self.method in ("eagle", "eagle3"): from vllm.transformers_utils.configs import SpeculatorsConfig from vllm.transformers_utils.configs.eagle import EAGLEConfig if isinstance( self.draft_model_config.hf_config, (EAGLEConfig, SpeculatorsConfig), ): pass else: eagle_config = EAGLEConfig( self.draft_model_config.hf_config, method=self.method, model_type="eagle", ) self.draft_model_config.hf_config = eagle_config self.draft_model_config.model_arch_config = ( self.draft_model_config.get_model_arch_config() ) if self.num_speculative_tokens is not None and hasattr( self.draft_model_config.hf_config, "num_lookahead_tokens" ): self.draft_model_config.hf_config.num_lookahead_tokens = ( self.num_speculative_tokens ) n_predict = getattr( self.draft_model_config.hf_config, "n_predict", None ) if n_predict is not None: if self.num_speculative_tokens is None: # Default to max value defined in draft model config. self.num_speculative_tokens = n_predict elif ( self.num_speculative_tokens > n_predict and self.num_speculative_tokens % n_predict != 0 ): # Ensure divisibility for MTP module reuse. raise ValueError( f"num_speculative_tokens:{self.num_speculative_tokens}" f" must be divisible by {n_predict=}" ) if self.speculative_token_tree is None: # Generate chain of tokens. self.speculative_token_tree = str( [(i + 1) * (0,) for i in range(self.num_speculative_tokens)] ) else: # Sort the token tree breadth-first. tree_choices = ast.literal_eval(self.speculative_token_tree) self.speculative_token_tree = str( sorted(tree_choices, key=lambda t: (len(t), t)) ) self.draft_tensor_parallel_size = ( SpeculativeConfig._verify_and_get_draft_tp( self.target_parallel_config, self.draft_tensor_parallel_size, self.draft_model_config.hf_config, ) ) self.draft_model_config.max_model_len = ( SpeculativeConfig._maybe_override_draft_max_model_len( self.max_model_len, self.draft_model_config.max_model_len, self.target_model_config.max_model_len, ) ) self.draft_parallel_config = ( SpeculativeConfig.create_draft_parallel_config( self.target_parallel_config, self.draft_tensor_parallel_size ) ) return self def _validate_suffix_decoding(self): if not has_arctic_inference(): raise ImportError( "Arctic Inference is required for suffix decoding. " "Install via `pip install arctic-inference==0.1.1`." ) if self.num_speculative_tokens is None: # Suffix decoding decides the actual number of speculative tokens # dynamically and treats num_speculative_tokens as a maximum limit. self.num_speculative_tokens = self.suffix_decoding_max_tree_depth logger.warning( "Defaulted num_speculative_tokens to %s for suffix decoding.", self.num_speculative_tokens, ) # Validate values if self.suffix_decoding_max_tree_depth < 1: raise ValueError( f"suffix_decoding_max_tree_depth=" f"{self.suffix_decoding_max_tree_depth} must be >= 1" ) if self.suffix_decoding_max_cached_requests < 0: raise ValueError( f"suffix_decoding_max_cached_requests=" f"{self.suffix_decoding_max_cached_requests} must be >= 0" ) if self.suffix_decoding_max_spec_factor < 0: raise ValueError( f"suffix_decoding_max_spec_factor=" f"{self.suffix_decoding_max_spec_factor} must be >= 0" ) if not 0 <= self.suffix_decoding_min_token_prob <= 1: raise ValueError( f"suffix_decoding_min_token_prob=" f"{self.suffix_decoding_min_token_prob} must be in [0, 1]" ) @staticmethod def _maybe_override_draft_max_model_len( speculative_max_model_len: int | None, draft_max_model_len: int, target_max_model_len: int, ) -> int: """Determine the max sequence len for the draft model. This is usually the draft_max_model_len, but may be the target_max_model_len if it is less than the draft_max_model_len, or may be speculative_max_model_len if it is specified. This is necessary so that sequences do not exceed the capacity of the draft model or the target model. speculative_max_model_len is mainly used for testing that sequences can skip speculation. """ if speculative_max_model_len is not None: if speculative_max_model_len > draft_max_model_len: raise ValueError( f"{speculative_max_model_len=} cannot be " f"larger than {draft_max_model_len=}" ) if speculative_max_model_len > target_max_model_len: raise ValueError( f"{speculative_max_model_len=} cannot be " f"larger than {target_max_model_len=}" ) return speculative_max_model_len return min( draft_max_model_len, target_max_model_len, ) @staticmethod def _verify_and_get_draft_tp( target_parallel_config: ParallelConfig, speculative_draft_tensor_parallel_size: int | None, draft_hf_config: PretrainedConfig, ) -> int: """ Verifies and adjusts the tensor parallel size for a draft model specified using speculative_draft_tensor_parallel_size. """ # If speculative_draft_tensor_parallel_size is unset then set it # appropriately else verify that it is set correctly. if speculative_draft_tensor_parallel_size is None: if draft_hf_config.model_type == "mlp_speculator": speculative_draft_tensor_parallel_size = 1 if target_parallel_config.tensor_parallel_size > 1: logger.warning( "%s cannot currently be run with tp>1; " "setting speculative_draft_tensor_parallel_size=1", draft_hf_config.model_type, ) else: speculative_draft_tensor_parallel_size = ( target_parallel_config.tensor_parallel_size ) elif speculative_draft_tensor_parallel_size not in ( 1, target_parallel_config.tensor_parallel_size, ): raise ValueError( f"{speculative_draft_tensor_parallel_size=} cannot be " f"other value than 1 or target model tensor_parallel_size" ) return speculative_draft_tensor_parallel_size @staticmethod def create_draft_parallel_config( target_parallel_config: ParallelConfig, speculative_draft_tensor_parallel_size: int, ) -> ParallelConfig: """Create a parallel config for use by the draft worker. This is mostly a copy of the target parallel config, except the tp_size. """ draft_parallel_config = ParallelConfig( pipeline_parallel_size=target_parallel_config.pipeline_parallel_size, tensor_parallel_size=speculative_draft_tensor_parallel_size, distributed_executor_backend=target_parallel_config.distributed_executor_backend, max_parallel_loading_workers=target_parallel_config.max_parallel_loading_workers, disable_custom_all_reduce=target_parallel_config.disable_custom_all_reduce, ray_workers_use_nsight=target_parallel_config.ray_workers_use_nsight, placement_group=target_parallel_config.placement_group, ) return draft_parallel_config @model_validator(mode="after") def _verify_args(self) -> Self: if self.num_speculative_tokens is None: raise ValueError( "num_speculative_tokens must be provided with " "speculative model unless the draft model config contains an " "n_predict parameter." ) if self.num_speculative_tokens <= 0: raise ValueError( "Expected num_speculative_tokens to be greater " f"than zero ({self.num_speculative_tokens})." ) if self.draft_model_config: self.draft_model_config.verify_with_parallel_config( self.draft_parallel_config ) if self.disable_by_batch_size is not None and self.disable_by_batch_size < 2: raise ValueError( "Expect the batch size threshold of disabling " "speculative decoding is > 1, but got " f"{self.disable_by_batch_size=}" ) eagle3_target_supported = ["llama", "qwen", "minicpm", "gpt_oss"] if ( self.method == "eagle3" and self.target_model_config and not any( supported_model in self.target_model_config.hf_text_config.model_type for supported_model in eagle3_target_supported ) ): raise ValueError( f"Eagle3 is only supported for {eagle3_target_supported} models. " # noqa: E501 f"Got {self.target_model_config.hf_text_config.model_type=}" ) return self def use_eagle(self) -> bool: return self.method in ("eagle", "eagle3", "mtp") def __repr__(self) -> str: method = self.method model = None if method in ("ngram", "suffix") else self.draft_model_config.model num_spec_tokens = self.num_speculative_tokens return f"SpeculativeConfig({method=}, {model=}, {num_spec_tokens=})"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/structured_outputs.py
vllm/config/structured_outputs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any, Literal from pydantic import model_validator from pydantic.dataclasses import dataclass from typing_extensions import Self from vllm.config.utils import config from vllm.utils.hashing import safe_hash StructuredOutputsBackend = Literal[ "auto", "xgrammar", "guidance", "outlines", "lm-format-enforcer" ] @config @dataclass class StructuredOutputsConfig: """Dataclass which contains structured outputs config for the engine.""" backend: StructuredOutputsBackend = "auto" """Which engine will be used for structured outputs (e.g. JSON schema, regex, etc) by default. With "auto", we will make opinionated choices based on request contents and what the backend libraries currently support, so the behavior is subject to change in each release.""" disable_fallback: bool = False """If `True`, vLLM will not fallback to a different backend on error.""" disable_any_whitespace: bool = False """If `True`, json output will always be compact without any whitespace. If `False`, the model may generate whitespace between JSON fields, which is still valid JSON. This is only supported for xgrammar and guidance backends.""" disable_additional_properties: bool = False """If `True`, the `guidance` backend will not use `additionalProperties` in the JSON schema. This is only supported for the `guidance` backend and is used to better align its behaviour with `outlines` and `xgrammar`.""" reasoning_parser: str = "" """Select the reasoning parser depending on the model that you're using. This is used to parse the reasoning content into OpenAI API format.""" reasoning_parser_plugin: str = "" """Path to a dynamically reasoning parser plugin that can be dynamically loaded and registered.""" enable_in_reasoning: bool = False """Whether to use structured input for reasoning.""" def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # no factors to consider. # this config will not affect the computation graph. factors: list[Any] = [] hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str @model_validator(mode="after") def _validate_structured_output_config(self) -> Self: if self.disable_any_whitespace and self.backend not in ("xgrammar", "guidance"): raise ValueError( "disable_any_whitespace is only supported for " "xgrammar and guidance backends." ) if self.disable_additional_properties and self.backend != "guidance": raise ValueError( "disable_additional_properties is only supported " "for the guidance backend." ) return self
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/__init__.py
vllm/config/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from vllm.config.attention import AttentionConfig from vllm.config.cache import CacheConfig from vllm.config.compilation import ( CompilationConfig, CompilationMode, CUDAGraphMode, PassConfig, ) from vllm.config.device import DeviceConfig from vllm.config.ec_transfer import ECTransferConfig from vllm.config.kv_events import KVEventsConfig from vllm.config.kv_transfer import KVTransferConfig from vllm.config.load import LoadConfig from vllm.config.lora import LoRAConfig from vllm.config.model import ( ModelConfig, iter_architecture_defaults, str_dtype_to_torch_dtype, try_match_architecture_defaults, ) from vllm.config.multimodal import MultiModalConfig from vllm.config.observability import ObservabilityConfig from vllm.config.parallel import EPLBConfig, ParallelConfig from vllm.config.pooler import PoolerConfig from vllm.config.profiler import ProfilerConfig from vllm.config.scheduler import SchedulerConfig from vllm.config.speculative import SpeculativeConfig from vllm.config.speech_to_text import SpeechToTextConfig from vllm.config.structured_outputs import StructuredOutputsConfig from vllm.config.utils import ( ConfigType, SupportsMetricsInfo, config, get_attr_docs, is_init_field, update_config, ) from vllm.config.vllm import ( VllmConfig, get_cached_compilation_config, get_current_vllm_config, get_layers_from_vllm_config, set_current_vllm_config, ) # __all__ should only contain classes and functions. # Types and globals should be imported from their respective modules. __all__ = [ # From vllm.config.attention "AttentionConfig", # From vllm.config.cache "CacheConfig", # From vllm.config.compilation "CompilationConfig", "CompilationMode", "CUDAGraphMode", "PassConfig", # From vllm.config.device "DeviceConfig", # From vllm.config.ec_transfer "ECTransferConfig", # From vllm.config.kv_events "KVEventsConfig", # From vllm.config.kv_transfer "KVTransferConfig", # From vllm.config.load "LoadConfig", # From vllm.config.lora "LoRAConfig", # From vllm.config.model "ModelConfig", "iter_architecture_defaults", "str_dtype_to_torch_dtype", "try_match_architecture_defaults", # From vllm.config.multimodal "MultiModalConfig", # From vllm.config.observability "ObservabilityConfig", # From vllm.config.parallel "EPLBConfig", "ParallelConfig", # From vllm.config.pooler "PoolerConfig", # From vllm.config.scheduler "SchedulerConfig", # From vllm.config.speculative "SpeculativeConfig", # From vllm.config.speech_to_text "SpeechToTextConfig", # From vllm.config.structured_outputs "StructuredOutputsConfig", # From vllm.config.profiler "ProfilerConfig", # From vllm.config.utils "ConfigType", "SupportsMetricsInfo", "config", "get_attr_docs", "is_init_field", "update_config", # From vllm.config.vllm "VllmConfig", "get_cached_compilation_config", "get_current_vllm_config", "set_current_vllm_config", "get_layers_from_vllm_config", ]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/multimodal.py
vllm/config/multimodal.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Mapping from typing import Any, Literal, TypeAlias from pydantic import ConfigDict, Field, field_validator, model_validator from pydantic.dataclasses import dataclass from vllm.attention.backends.registry import AttentionBackendEnum from vllm.config.utils import config from vllm.utils.hashing import safe_hash @dataclass class BaseDummyOptions: """Base options for generating dummy data during profiling.""" count: int = Field(999, ge=0) @dataclass(config=ConfigDict(extra="forbid")) class VideoDummyOptions(BaseDummyOptions): """Options for generating dummy video data during profiling.""" num_frames: int | None = Field(None, gt=0) width: int | None = Field(None, gt=0) height: int | None = Field(None, gt=0) @dataclass(config=ConfigDict(extra="forbid")) class ImageDummyOptions(BaseDummyOptions): """Options for generating dummy image data during profiling.""" width: int | None = Field(None, gt=0) height: int | None = Field(None, gt=0) @dataclass(config=ConfigDict(extra="forbid")) class AudioDummyOptions(BaseDummyOptions): """Options for generating dummy audio data during profiling.""" length: int | None = Field(None, gt=0) MMEncoderTPMode = Literal["weights", "data"] MMCacheType = Literal["shm", "lru"] DummyOptions: TypeAlias = ( BaseDummyOptions | VideoDummyOptions | ImageDummyOptions | AudioDummyOptions ) @config @dataclass class MultiModalConfig: """Controls the behavior of multimodal models.""" limit_per_prompt: dict[str, DummyOptions] = Field(default_factory=dict) """The maximum number of input items and options allowed per prompt for each modality. Defaults to 999 for each modality. Legacy format (count only): {"image": 16, "video": 2} Configurable format (with options): {"video": {"count": 1, "num_frames": 32, "width": 512, "height": 512}, "image": {"count": 5, "width": 512, "height": 512}} Mixed format (combining both): {"image": 16, "video": {"count": 1, "num_frames": 32, "width": 512, "height": 512}} """ enable_mm_embeds: bool = False """If `True`, enables passing multimodal embeddings: for `LLM` class, this refers to tensor inputs under `multi_modal_data`; for the OpenAI-compatible server, this refers to chat messages with content `"type": "*_embeds"`. WARNING: The vLLM engine may crash if incorrect shape of embeddings is passed. Only enable this flag for trusted users!""" media_io_kwargs: dict[str, dict[str, Any]] = Field(default_factory=dict) """Additional args passed to process media inputs, keyed by modalities. For example, to set num_frames for video, set `--media-io-kwargs '{"video": {"num_frames": 40} }'`""" mm_processor_kwargs: dict[str, object] | None = None """Arguments to be forwarded to the model's processor for multi-modal data, e.g., image processor. Overrides for the multi-modal processor obtained from `transformers.AutoProcessor.from_pretrained`. The available overrides depend on the model that is being run. For example, for Phi-3-Vision: `{"num_crops": 4}`.""" mm_processor_cache_gb: float = Field(default=4, ge=0) """The size (in GiB) of the multi-modal processor cache, which is used to avoid re-processing past multi-modal inputs. This cache is duplicated for each API process and engine core process, resulting in a total memory usage of `mm_processor_cache_gb * (api_server_count + data_parallel_size)`. Set to `0` to disable this cache completely (not recommended).""" mm_processor_cache_type: MMCacheType = "lru" """Type of cache to use for the multi-modal preprocessor/mapper. If `shm`, use shared memory FIFO cache. If `lru`, use mirrored LRU cache.""" mm_shm_cache_max_object_size_mb: int = Field(default=128, ge=0) """Size limit (in MiB) for each object stored in the multi-modal processor shared memory cache. Only effective when `mm_processor_cache_type` is `"shm"`.""" mm_encoder_tp_mode: MMEncoderTPMode = "weights" """Indicates how to optimize multi-modal encoder inference using tensor parallelism (TP). - `"weights"`: Within the same vLLM engine, split the weights of each layer across TP ranks. (default TP behavior)\n - `"data"`: Within the same vLLM engine, split the batched input data across TP ranks to process the data in parallel, while hosting the full weights on each TP rank. This batch-level DP is not to be confused with API request-level DP (which is controlled by `--data-parallel-size`). This is only supported on a per-model basis and falls back to `"weights"` if the encoder does not support DP.""" mm_encoder_attn_backend: AttentionBackendEnum | None = None """Optional override for the multi-modal encoder attention backend when using vision transformers. Accepts any value from `vllm.attention.backends.registry.AttentionBackendEnum` (e.g. `FLASH_ATTN`).""" interleave_mm_strings: bool = False """Enable fully interleaved support for multimodal prompts, while using --chat-template-content-format=string.""" skip_mm_profiling: bool = False """When enabled, skips multimodal memory profiling and only profiles with language backbone model during engine initialization. This reduces engine startup time but shifts the responsibility to users for estimating the peak memory usage of the activation of multimodal encoder and embedding cache.""" video_pruning_rate: float | None = Field(default=None, ge=0.0, lt=1.0) """Sets pruning rate for video pruning via Efficient Video Sampling. Value sits in range [0;1) and determines fraction of media tokens from each video to be pruned. """ @field_validator("limit_per_prompt", mode="before") @classmethod def _validate_limit_per_prompt( cls, value: dict[str, int | dict[str, int]] ) -> dict[str, DummyOptions]: for k, v in value.items(): # Handle legacy format where only count is specified if isinstance(v, int): v = {"count": v} # Convert to the appropriate DummyOptions subclass if k == "video": value[k] = VideoDummyOptions(**v) elif k == "image": value[k] = ImageDummyOptions(**v) elif k == "audio": value[k] = AudioDummyOptions(**v) else: value[k] = BaseDummyOptions(**v) return value @field_validator("mm_encoder_attn_backend", mode="before") @classmethod def _validate_mm_encoder_attn_backend( cls, value: str | AttentionBackendEnum | None ) -> AttentionBackendEnum | None: if isinstance(value, str) and value.upper() == "XFORMERS": raise ValueError( "Attention backend 'XFORMERS' has been removed (See PR #29262 for " "details). Please select a supported attention backend." ) if value is None or isinstance(value, AttentionBackendEnum): return value assert isinstance(value, str), ( "mm_encoder_attn_backend must be a string or an AttentionBackendEnum." ) return AttentionBackendEnum[value.upper()] @model_validator(mode="after") def _validate_multimodal_config(self): if self.mm_processor_cache_type != "shm" and ( self.mm_shm_cache_max_object_size_mb != MultiModalConfig.mm_shm_cache_max_object_size_mb ): raise ValueError( "'mm_shm_cache_max_object_size_mb' should only be set when " "'mm_processor_cache_type' is 'shm'." ) return self def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ factors: list[Any] = [ self.mm_encoder_attn_backend.name if self.mm_encoder_attn_backend is not None else None ] hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str def get_limit_per_prompt(self, modality: str) -> int: """ Get the maximum number of input items allowed per prompt for the given modality (backward compatible). """ limit_data = self.limit_per_prompt.get(modality) if limit_data is None: # Unspecified modality is set to 999 by default return 999 return limit_data.count def get_dummy_options(self, modality: str) -> BaseDummyOptions | None: """ Get the configurable dummy data options for a modality. Returns None if no options are configured for this modality. """ # All values are now DummyOptions after normalization return self.limit_per_prompt.get(modality) def merge_mm_processor_kwargs( self, inference_kwargs: Mapping[str, object], ) -> dict[str, object]: """ Get the keyword arguments to pass to the multi-modal processor according to the extra arguments passed during inference. """ kwargs = self.mm_processor_kwargs or {} return kwargs | dict(inference_kwargs) def is_multimodal_pruning_enabled(self): return self.video_pruning_rate is not None and self.video_pruning_rate > 0
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/attention.py
vllm/config/attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any, Literal from pydantic import field_validator from pydantic.dataclasses import dataclass from vllm.attention.backends.registry import AttentionBackendEnum from vllm.config.utils import config from vllm.logger import init_logger logger = init_logger(__name__) @config @dataclass class AttentionConfig: """Configuration for attention mechanisms in vLLM.""" backend: AttentionBackendEnum | None = None """Attention backend to use. If None, will be selected automatically.""" flash_attn_version: Literal[2, 3] | None = None """Force vllm to use a specific flash-attention version (2 or 3). Only valid when using the flash-attention backend.""" use_prefill_decode_attention: bool = False """Use separate prefill and decode kernels for attention instead of the unified triton kernel.""" flash_attn_max_num_splits_for_cuda_graph: int = 32 """Flash Attention max number splits for cuda graph decode.""" use_cudnn_prefill: bool = False """Whether to use cudnn prefill.""" use_trtllm_ragged_deepseek_prefill: bool = False """Whether to use TRTLLM ragged deepseek prefill.""" use_trtllm_attention: bool | None = None """If set to True/False, use or don't use the TRTLLM attention backend in flashinfer. If None, auto-detect the attention backend in flashinfer.""" disable_flashinfer_prefill: bool = False """Whether to disable flashinfer prefill.""" disable_flashinfer_q_quantization: bool = False """If set, when using fp8 kv, do not quantize Q to fp8.""" def compute_hash(self) -> str: """ Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ from vllm.config.utils import get_hash_factors, hash_factors ignored_factors: list[str] = [] factors = get_hash_factors(self, ignored_factors) return hash_factors(factors) @field_validator("backend", mode="before") @classmethod def validate_backend_before(cls, value: Any) -> Any: """Enable parsing of the `backend` enum type from string.""" if isinstance(value, str): return AttentionBackendEnum[value.upper()] return value def _set_from_env_if_set(self, field_name: str, env_var_name: str) -> None: """Set field from env var if set, with deprecation warning.""" from vllm import envs if envs.is_set(env_var_name): value = getattr(envs, env_var_name) if field_name == "backend": value = self.validate_backend_before(value) setattr(self, field_name, value) logger.warning_once( "Using %s environment variable is deprecated and will be removed in " "v0.14.0 or v1.0.0, whichever is soonest. Please use " "--attention-config.%s command line argument or " "AttentionConfig(%s=...) config field instead.", env_var_name, field_name, field_name, ) def __post_init__(self) -> None: self._set_from_env_if_set("backend", "VLLM_ATTENTION_BACKEND") self._set_from_env_if_set("flash_attn_version", "VLLM_FLASH_ATTN_VERSION") self._set_from_env_if_set( "use_prefill_decode_attention", "VLLM_V1_USE_PREFILL_DECODE_ATTENTION" ) self._set_from_env_if_set( "flash_attn_max_num_splits_for_cuda_graph", "VLLM_FLASH_ATTN_MAX_NUM_SPLITS_FOR_CUDA_GRAPH", ) self._set_from_env_if_set("use_cudnn_prefill", "VLLM_USE_CUDNN_PREFILL") self._set_from_env_if_set( "use_trtllm_ragged_deepseek_prefill", "VLLM_USE_TRTLLM_RAGGED_DEEPSEEK_PREFILL", ) self._set_from_env_if_set("use_trtllm_attention", "VLLM_USE_TRTLLM_ATTENTION") self._set_from_env_if_set( "disable_flashinfer_prefill", "VLLM_DISABLE_FLASHINFER_PREFILL" ) self._set_from_env_if_set( "disable_flashinfer_q_quantization", "VLLM_FLASHINFER_DISABLE_Q_QUANTIZATION", )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/kv_transfer.py
vllm/config/kv_transfer.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import uuid from dataclasses import field from typing import Any, Literal, get_args from pydantic.dataclasses import dataclass from vllm.config.utils import config from vllm.utils.hashing import safe_hash KVProducer = Literal["kv_producer", "kv_both"] KVConsumer = Literal["kv_consumer", "kv_both"] KVRole = Literal[KVProducer, KVConsumer] @config @dataclass class KVTransferConfig: """Configuration for distributed KV cache transfer.""" kv_connector: str | None = None """The KV connector for vLLM to transmit KV caches between vLLM instances. """ engine_id: str | None = None """The engine id for KV transfers.""" kv_buffer_device: str = "cuda" """The device used by kv connector to buffer the KV cache. Choices are 'cuda' and 'cpu'.""" kv_buffer_size: float = 1e9 """The buffer size for TorchDistributedConnector. Measured in number of bytes. Recommended value: 1e9 (about 1GB).""" kv_role: KVRole | None = None """Whether this vLLM instance produces, consumes KV cache, or both. Choices are 'kv_producer', 'kv_consumer', and 'kv_both'.""" kv_rank: int | None = None """The rank of this vLLM instance in the KV cache transfer. Typical value: 0 for prefill instance, 1 for decode instance. Currently only 1P1D is supported.""" kv_parallel_size: int = 1 """The number of parallel instances for KV cache transfer. For P2pNcclConnector, this should be 2.""" kv_ip: str = "127.0.0.1" """The KV connector ip, used to build distributed connection.""" kv_port: int = 14579 """The KV connector port, used to build distributed connection.""" kv_connector_extra_config: dict[str, Any] = field(default_factory=dict) """any extra config that the connector may need.""" kv_connector_module_path: str | None = None """The Python module path to dynamically load the KV connector from. Only supported in V1.""" enable_permute_local_kv: bool = False """Experiment feature flag to enable HND to NHD KV Transfer""" kv_load_failure_policy: Literal["recompute", "fail"] = "recompute" """Policy for handling KV cache load failures. 'recompute': reschedule the request to recompute failed blocks (default) 'fail': immediately fail the request with an error finish reason""" def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # no factors to consider. # this config will not affect the computation graph. factors: list[Any] = [] hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str def __post_init__(self) -> None: if self.engine_id is None: self.engine_id = str(uuid.uuid4()) if self.kv_role is not None and self.kv_role not in get_args(KVRole): raise ValueError( f"Unsupported kv_role: {self.kv_role}. " f"Supported roles are {get_args(KVRole)}" ) if self.kv_connector is not None and self.kv_role is None: raise ValueError( "Please specify kv_role when kv_connector " f"is set, supported roles are {get_args(KVRole)}" ) @property def is_kv_transfer_instance(self) -> bool: return self.kv_connector is not None and self.kv_role in get_args(KVRole) @property def is_kv_producer(self) -> bool: return self.kv_connector is not None and self.kv_role in get_args(KVProducer) @property def is_kv_consumer(self) -> bool: return self.kv_connector is not None and self.kv_role in get_args(KVConsumer) def get_from_extra_config(self, key, default) -> Any: return self.kv_connector_extra_config.get(key, default)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/cache.py
vllm/config/cache.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import field from typing import TYPE_CHECKING, Any, Literal from pydantic import Field, SkipValidation, field_validator from pydantic.dataclasses import dataclass from vllm.config.utils import config from vllm.logger import init_logger from vllm.utils.mem_constants import GiB_bytes from vllm.utils.mem_utils import get_cpu_memory if TYPE_CHECKING: from vllm.config.parallel import ParallelConfig else: ParallelConfig = Any logger = init_logger(__name__) BlockSize = Literal[1, 8, 16, 32, 64, 128, 256] CacheDType = Literal[ "auto", "bfloat16", "fp8", "fp8_e4m3", "fp8_e5m2", "fp8_inc", "fp8_ds_mla", ] MambaDType = Literal["auto", "float32", "float16"] PrefixCachingHashAlgo = Literal["sha256", "sha256_cbor", "xxhash", "xxhash_cbor"] KVOffloadingBackend = Literal["native", "lmcache"] @config @dataclass class CacheConfig: """Configuration for the KV cache.""" block_size: SkipValidation[BlockSize] = None # type: ignore """Size of a contiguous cache block in number of tokens. On CUDA devices, only block sizes up to 32 are supported. This config has no static default. If left unspecified by the user, it will be set in `Platform.check_and_update_config()` based on the current platform.""" gpu_memory_utilization: float = Field(default=0.9, gt=0, le=1) """The fraction of GPU memory to be used for the model executor, which can range from 0 to 1. For example, a value of 0.5 would imply 50% GPU memory utilization. If unspecified, will use the default value of 0.9. This is a per-instance limit, and only applies to the current vLLM instance. It does not matter if you have another vLLM instance running on the same GPU. For example, if you have two vLLM instances running on the same GPU, you can set the GPU memory utilization to 0.5 for each instance.""" swap_space: float = Field(default=4, ge=0) """Size of the CPU swap space per GPU (in GiB).""" cache_dtype: CacheDType = "auto" """Data type for kv cache storage. If "auto", will use model data type. CUDA 11.8+ supports fp8 (=fp8_e4m3) and fp8_e5m2. ROCm (AMD GPU) supports fp8 (=fp8_e4m3). Intel Gaudi (HPU) supports fp8 (using fp8_inc). Some models (namely DeepSeekV3.2) default to fp8, set to bfloat16 to use bfloat16 instead, this is an invalid option for models that do not default to fp8. """ is_attention_free: bool = False """Whether the model is attention-free. This is primarily set in `ModelConfig` and that value should be manually duplicated here.""" num_gpu_blocks_override: int | None = None """Number of GPU blocks to use. This overrides the profiled `num_gpu_blocks` if specified. Does nothing if `None`. Used for testing preemption.""" sliding_window: int | None = None """Sliding window size for the KV cache. This is primarily set in `ModelConfig` and that value should be manually duplicated here.""" enable_prefix_caching: bool = True """Whether to enable prefix caching.""" prefix_caching_hash_algo: PrefixCachingHashAlgo = "sha256" """Set the hash algorithm for prefix caching:\n - "sha256" uses Pickle for object serialization before hashing. This is the current default, as SHA256 is the most secure choice to avoid potential hash collisions.\n - "sha256_cbor" provides a reproducible, cross-language compatible hash. It serializes objects using canonical CBOR and hashes them with SHA-256.\n - "xxhash" uses Pickle serialization with xxHash (128-bit) for faster, non-cryptographic hashing. Requires the optional ``xxhash`` package. IMPORTANT: Use of a hashing algorithm that is not considered cryptographically secure theoretically increases the risk of hash collisions, which can cause undefined behavior or even leak private information in multi-tenant environments. Even if collisions are still very unlikely, it is important to consider your security risk tolerance against the performance benefits before turning this on.\n - "xxhash_cbor" combines canonical CBOR serialization with xxHash for reproducible hashing. Requires the optional ``xxhash`` package.""" cpu_offload_gb: float = Field(default=0, ge=0) """The space in GiB to offload to CPU, per GPU. Default is 0, which means no offloading. Intuitively, this argument can be seen as a virtual way to increase the GPU memory size. For example, if you have one 24 GB GPU and set this to 10, virtually you can think of it as a 34 GB GPU. Then you can load a 13B model with BF16 weight, which requires at least 26GB GPU memory. Note that this requires fast CPU-GPU interconnect, as part of the model is loaded from CPU memory to GPU memory on the fly in each model forward pass. """ calculate_kv_scales: bool = False """This enables dynamic calculation of `k_scale` and `v_scale` when kv_cache_dtype is fp8. If `False`, the scales will be loaded from the model checkpoint if available. Otherwise, the scales will default to 1.0.""" cpu_kvcache_space_bytes: int | None = None """(CPU backend only) CPU key-value cache space.""" mamba_page_size_padded: int | None = None """ Optional override for mamba page size; used by hybrid mamba/attention models to ensure exact alignment with attention page size.""" mamba_block_size: int | None = Field(default=None, gt=0) """Size of a contiguous cache block in number of tokens for mamba cache. Can be set only when prefix caching is enabled. Value must be a multiple of 8 to align with causal_conv1d kernel.""" mamba_cache_dtype: MambaDType = "auto" """The data type to use for the Mamba cache (both the conv as well as the ssm state). If set to 'auto', the data type will be inferred from the model config.""" mamba_ssm_cache_dtype: MambaDType = "auto" """The data type to use for the Mamba cache (ssm state only, conv state will still be controlled by mamba_cache_dtype). If set to 'auto', the data type for the ssm state will be determined by mamba_cache_dtype.""" # Will be set after profiling. num_gpu_blocks: int | None = field(default=None, init=False) """The number of blocks to allocate for GPU memory.""" num_cpu_blocks: int | None = field(default=None, init=False) """The number of blocks to allocate for CPU memory.""" kv_sharing_fast_prefill: bool = False """This feature is work in progress and no prefill optimization takes place with this flag enabled currently. In some KV sharing setups, e.g. YOCO (https://arxiv.org/abs/2405.05254), some layers can skip tokens corresponding to prefill. This flag enables attention metadata for eligible layers to be overridden with metadata necessary for implementing this optimization in some models (e.g. Gemma3n) """ kv_cache_memory_bytes: int | None = None """Size of KV Cache per GPU in bytes. By default, this is set to None and vllm can automatically infer the kv cache size based on gpu_memory_utilization. However, users may want to manually specify the kv cache memory size. kv_cache_memory_bytes allows more fine-grain control of how much memory gets used when compared with using gpu_memory_utilization. Note that kv_cache_memory_bytes (when not-None) ignores gpu_memory_utilization""" kv_offloading_size: float | None = None """Size of the KV cache offloading buffer in GiB. When TP > 1, this is the total buffer size summed across all TP ranks. By default, this is set to None, which means no KV offloading is enabled. When set with kv_offloading_backend, vLLM will enable KV cache offloading to CPU""" kv_offloading_backend: KVOffloadingBackend | None = None """The backend to use for KV cache offloading. Supported backends include 'native' (vLLM native CPU offloading), 'lmcache' This option must be used together with kv_offloading_size.""" def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ ignored_factors = { # Runtime/derived knobs that don't affect compiled graph shape "gpu_memory_utilization", "swap_space", "is_attention_free", "num_gpu_blocks_override", "enable_prefix_caching", "prefix_caching_hash_algo", "cpu_kvcache_space_bytes", "mamba_page_size_padded", # Post-init/derived counters "num_gpu_blocks", "num_cpu_blocks", # WIP feature toggle not impacting compiled graph shape "kv_sharing_fast_prefill", } from vllm.config.utils import get_hash_factors, hash_factors factors = get_hash_factors(self, ignored_factors) return hash_factors(factors) def metrics_info(self): # convert cache_config to dict(key: str, value: str) for prometheus # metrics info return {key: str(value) for key, value in self.__dict__.items()} @field_validator("cache_dtype", mode="after") @classmethod def _validate_cache_dtype(cls, cache_dtype: CacheDType) -> CacheDType: if cache_dtype.startswith("fp8"): logger.info( "Using fp8 data type to store kv cache. It reduces the GPU " "memory footprint and boosts the performance. " "Meanwhile, it may cause accuracy drop without a proper " "scaling factor." ) return cache_dtype def verify_with_parallel_config( self, parallel_config: ParallelConfig, ) -> None: swap_space_bytes = self.swap_space * GiB_bytes total_cpu_memory = get_cpu_memory() # FIXME(woosuk): Here, it is assumed that the GPUs in a tensor parallel # group are in the same node. However, the GPUs may span multiple nodes. num_gpus_per_node = parallel_config.tensor_parallel_size cpu_memory_usage = swap_space_bytes * num_gpus_per_node msg = ( f"{cpu_memory_usage / GiB_bytes:.2f} GiB out of the " f"{total_cpu_memory / GiB_bytes:.2f} GiB total CPU memory " "is allocated for the swap space." ) if cpu_memory_usage > 0.7 * total_cpu_memory: raise ValueError("Too large swap space. " + msg) elif cpu_memory_usage > 0.4 * total_cpu_memory: logger.warning("Possibly too large swap space. %s", msg)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/device.py
vllm/config/device.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import field from typing import Any, Literal import torch from pydantic import ConfigDict, SkipValidation from pydantic.dataclasses import dataclass from vllm.config.utils import config from vllm.utils.hashing import safe_hash Device = Literal["auto", "cuda", "cpu", "tpu", "xpu"] @config @dataclass(config=ConfigDict(arbitrary_types_allowed=True)) class DeviceConfig: """Configuration for the device to use for vLLM execution.""" device: SkipValidation[Device | torch.device | None] = "auto" """Device type for vLLM execution. This parameter is deprecated and will be removed in a future release. It will now be set automatically based on the current platform.""" device_type: str = field(init=False) """Device type from the current platform. This is set in `__post_init__`.""" def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # no factors to consider. # the device/platform information will be summarized # by torch/vllm automatically. factors: list[Any] = [] hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str def __post_init__(self): if self.device == "auto": # Automated device type detection from vllm.platforms import current_platform self.device_type = current_platform.device_type if not self.device_type: raise RuntimeError( "Failed to infer device type, please set " "the environment variable `VLLM_LOGGING_LEVEL=DEBUG` " "to turn on verbose logging to help debug the issue." ) else: # Device type is assigned explicitly if isinstance(self.device, str): self.device_type = self.device elif isinstance(self.device, torch.device): self.device_type = self.device.type # Some device types require processing inputs on CPU if self.device_type in ["tpu"]: self.device = None else: # Set device with device type self.device = torch.device(self.device_type)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/model_arch.py
vllm/config/model_arch.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any from pydantic import ConfigDict from pydantic.dataclasses import dataclass from vllm.logger import init_logger logger = init_logger(__name__) @dataclass(config=ConfigDict(arbitrary_types_allowed=True)) class ModelArchitectureConfig: """ Configuration for model architecture that required by vLLM runtime """ architectures: list[str] | None """List of model architecture class names (e.g., ['LlamaForCausalLM']). It can be None upon calling `vllm_config.with_hf_config(config.text_config)`""" model_type: str """Model type identifier (e.g., 'llama', 'gpt_oss').""" text_model_type: str | None """Text model type identifier (e.g., 'llama4_text').""" hidden_size: int """Hidden size of the model.""" total_num_hidden_layers: int """Number of hidden layers in the model.""" total_num_attention_heads: int """Number of attention heads in the model.""" head_size: int """Head dimension of the model.""" vocab_size: int """Vocabulary size of the model.""" total_num_kv_heads: int """Number of key value heads in the model.""" num_experts: int """Number of experts in the model.""" quantization_config: dict[str, Any] | None """Quantization configuration dictionary containing quantization parameters.""" is_deepseek_mla: bool """Whether the model is a DeepSeek MLA model.""" derived_max_model_len_and_key: tuple[float, str | None] """Derived maximum model length and key from the hf config."""
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/scheduler.py
vllm/config/scheduler.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable from dataclasses import InitVar from typing import TYPE_CHECKING, Any, ClassVar, Literal, cast from pydantic import Field, field_validator from pydantic.dataclasses import dataclass from typing_extensions import Self from vllm.config.utils import config from vllm.logger import init_logger from vllm.utils.hashing import safe_hash from vllm.utils.import_utils import resolve_obj_by_qualname if TYPE_CHECKING: from vllm.v1.core.sched.interface import SchedulerInterface logger = init_logger(__name__) RunnerType = Literal["generate", "pooling", "draft"] SchedulerPolicy = Literal["fcfs", "priority"] @config @dataclass class SchedulerConfig: """Scheduler configuration.""" max_model_len: InitVar[int] """Maximum length of a sequence (including prompt and generated text). Note: This is stored in the ModelConfig, and is used only here to provide fallbacks and validate other attributes.""" is_encoder_decoder: InitVar[bool] """True if the model is an encoder-decoder model. Note: This is stored in the ModelConfig, and is used only here to disable chunked prefill and prefix caching for encoder-decoder models. """ DEFAULT_MAX_NUM_BATCHED_TOKENS: ClassVar[int] = 2048 DEFAULT_MAX_NUM_SEQS: ClassVar[int] = 128 runner_type: RunnerType = "generate" """The runner type to launch for the model.""" max_num_batched_tokens: int = Field(default=DEFAULT_MAX_NUM_BATCHED_TOKENS, ge=1) """Maximum number of tokens to be processed in a single iteration. The default value here is mainly for convenience when testing. In real usage, this should be set in `EngineArgs.create_engine_config`. """ max_num_seqs: int = Field(default=DEFAULT_MAX_NUM_SEQS, ge=1) """Maximum number of sequences to be processed in a single iteration. The default value here is mainly for convenience when testing. In real usage, this should be set in `EngineArgs.create_engine_config`. """ max_num_partial_prefills: int = Field(default=1, ge=1) """For chunked prefill, the maximum number of sequences that can be partially prefilled concurrently.""" max_long_partial_prefills: int = Field(default=1, ge=1) """For chunked prefill, the maximum number of prompts longer than long_prefill_token_threshold that will be prefilled concurrently. Setting this less than max_num_partial_prefills will allow shorter prompts to jump the queue in front of longer prompts in some cases, improving latency.""" long_prefill_token_threshold: int = 0 """For chunked prefill, a request is considered long if the prompt is longer than this number of tokens.""" enable_chunked_prefill: bool = True """If True, prefill requests can be chunked based on the remaining `max_num_batched_tokens`. The default value here is mainly for convenience when testing. In real usage, this should be set in `EngineArgs.create_engine_config`. """ is_multimodal_model: bool = False """True if the model is multimodal.""" # TODO (ywang96): Make this configurable. max_num_encoder_input_tokens: int = Field(init=False) """Multimodal encoder compute budget, only used in V1. NOTE: This is not currently configurable. It will be overridden by max_num_batched_tokens in case max multimodal embedding size is larger.""" # TODO (ywang96): Make this configurable. encoder_cache_size: int = Field(init=False) """Multimodal encoder cache size, only used in V1. NOTE: This is not currently configurable. It will be overridden by max_num_batched_tokens in case max multimodal embedding size is larger.""" policy: SchedulerPolicy = "fcfs" """The scheduling policy to use:\n - "fcfs" means first come first served, i.e. requests are handled in order of arrival.\n - "priority" means requests are handled based on given priority (lower value means earlier handling) and time of arrival deciding any ties).""" disable_chunked_mm_input: bool = False """If set to true and chunked prefill is enabled, we do not want to partially schedule a multimodal item. Only used in V1 This ensures that if a request has a mixed prompt (like text tokens TTTT followed by image tokens IIIIIIIIII) where only some image tokens can be scheduled (like TTTTIIIII, leaving IIIII), it will be scheduled as TTTT in one step and IIIIIIIIII in the next.""" # scheduler class or path. "vllm.v1.core.sched.scheduler.Scheduler" # (default) or "mod.custom_class". scheduler_cls: str | type[object] = Field(default=None) """The scheduler class to use. "vllm.v1.core.sched.scheduler.Scheduler" is the default scheduler. Can be a class directly or the path to a class of form "mod.custom_class".""" disable_hybrid_kv_cache_manager: bool | None = None """If set to True, KV cache manager will allocate the same size of KV cache for all attention layers even if there are multiple type of attention layers like full attention and sliding window attention. If set to None, the default value will be determined based on the environment and starting configuration. """ async_scheduling: bool = Field(default=None) """If set to False, disable async scheduling. Async scheduling helps to avoid gaps in GPU utilization, leading to better latency and throughput. It is currently not supported with some features such as speculative decoding and pipeline parallelism, and will be automatically disabled in those cases. """ stream_interval: int = Field(default=1, ge=1) """The interval (or buffer size) for streaming in terms of token length. A smaller value (1) makes streaming smoother by sending each token immediately, while a larger value (e.g., 10) reduces host overhead and may increase throughput by batching multiple tokens before sending.""" @staticmethod def default_factory(**kwargs): """ Factory method to create `SchedulerConfig` with default values for `InitVar`s. """ if "max_model_len" not in kwargs: kwargs["max_model_len"] = 8192 if "is_encoder_decoder" not in kwargs: kwargs["is_encoder_decoder"] = False return SchedulerConfig(**kwargs) def get_scheduler_cls(self) -> type["SchedulerInterface"]: if self.scheduler_cls is None: if self.async_scheduling: from vllm.v1.core.sched.async_scheduler import AsyncScheduler return AsyncScheduler from vllm.v1.core.sched.scheduler import Scheduler return Scheduler # This warning can be removed once the Scheduler interface is # finalized and we can maintain support for scheduler classes that # implement it logger.warning_once( "Using custom scheduler class %s. This scheduler interface is " "not public and compatibility may not be maintained.", self.scheduler_cls, ) if not isinstance(self.scheduler_cls, str): return cast(type["SchedulerInterface"], self.scheduler_cls) return resolve_obj_by_qualname(self.scheduler_cls) def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ factors: list[Any] = [] # max_num_batched_tokens need to be included in the hash due # to two reasons: # 1. LoRA creates static buffers based on max_num_batched_tokens. # The tensor sizes and strides get captured in the torch.compile # graph explicitly. # 2. Inductor decides whether using 32-bit or 64-bit indexing integer # based on the data sizes. `max_num_batched_tokens` has an # impact on that. For more details, please check # https://github.com/vllm-project/vllm/issues/29585 factors.append(self.max_num_batched_tokens) hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str @field_validator("scheduler_cls", "async_scheduling", mode="wrap") @classmethod def _skip_none_validation(cls, value: Any, handler: Callable) -> Any: """Skip validation if the value is `None` when initialisation is delayed.""" if value is None: return value return handler(value) def __post_init__(self, max_model_len: int, is_encoder_decoder: bool) -> None: if is_encoder_decoder: # Chunked prefill should be disabled for encoder-decoder models. self.disable_chunked_mm_input = True self.enable_chunked_prefill = False self.long_prefill_token_threshold = 0 logger.info( "Encoder-decoder models do not support chunked prefill nor" " prefix caching; disabling both." ) self.max_num_encoder_input_tokens = self.max_num_batched_tokens self.encoder_cache_size = self.max_num_batched_tokens if self.enable_chunked_prefill: logger.info( "Chunked prefill is enabled with max_num_batched_tokens=%d.", self.max_num_batched_tokens, ) if self.max_num_partial_prefills > 1: if self.long_prefill_token_threshold == 0: self.long_prefill_token_threshold = int(max_model_len * 0.04) logger.info( "Concurrent partial prefills enabled with " "max_num_partial_prefills=%d, max_long_partial_prefills=%d, " "long_prefill_token_threshold=%d", self.max_num_partial_prefills, self.max_long_partial_prefills, self.long_prefill_token_threshold, ) self.verify_max_model_len(max_model_len) def verify_max_model_len(self, max_model_len: int) -> Self: if ( self.max_num_batched_tokens < max_model_len and not self.enable_chunked_prefill ): raise ValueError( f"max_num_batched_tokens ({self.max_num_batched_tokens}) is " f"smaller than max_model_len ({max_model_len}). " "This effectively limits the maximum sequence length to " "max_num_batched_tokens and makes vLLM reject longer " "sequences. Please increase max_num_batched_tokens or " "decrease max_model_len." ) if self.max_num_batched_tokens < self.max_num_seqs: raise ValueError( f"max_num_batched_tokens ({self.max_num_batched_tokens}) must " "be greater than or equal to max_num_seqs " f"({self.max_num_seqs})." ) if self.max_num_batched_tokens > self.max_num_seqs * max_model_len: logger.warning( "max_num_batched_tokens (%d) exceeds max_num_seqs " "* max_model_len (%d). This may lead to unexpected behavior.", self.max_num_batched_tokens, self.max_num_seqs * max_model_len, ) if self.max_num_partial_prefills > 1: if not self.enable_chunked_prefill: raise ValueError( "Chunked prefill must be enabled to set " "max_num_partial_prefills > 1." ) if self.long_prefill_token_threshold > max_model_len: raise ValueError( "long_prefill_token_threshold " f"({self.long_prefill_token_threshold}) cannot be greater " f"than the max_model_len ({max_model_len})." ) if self.max_long_partial_prefills > self.max_num_partial_prefills: raise ValueError( f"{self.max_long_partial_prefills=} must be less than or equal to " f"{self.max_num_partial_prefills=}." ) return self
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/config/parallel.py
vllm/config/parallel.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os from typing import TYPE_CHECKING, Any, Literal import torch from pydantic import Field, model_validator from pydantic.dataclasses import dataclass from torch.distributed import ProcessGroup, ReduceOp from typing_extensions import Self import vllm.envs as envs from vllm.config.utils import config from vllm.logger import init_logger from vllm.model_executor.layers.batch_invariant import ( vllm_is_batch_invariant, ) from vllm.platforms import current_platform from vllm.utils.network_utils import get_open_ports_list from vllm.utils.torch_utils import cuda_device_count_stateless if TYPE_CHECKING: from ray.runtime_env import RuntimeEnv from ray.util.placement_group import PlacementGroup from vllm.v1.executor import Executor else: RuntimeEnv = Any PlacementGroup = Any Executor = Any logger = init_logger(__name__) ExpertPlacementStrategy = Literal["linear", "round_robin"] DistributedExecutorBackend = Literal["ray", "mp", "uni", "external_launcher"] DataParallelBackend = Literal["ray", "mp"] EPLBPolicyOption = Literal["default"] All2AllBackend = Literal[ "naive", "pplx", "deepep_high_throughput", "deepep_low_latency", "allgather_reducescatter", "flashinfer_all2allv", ] @config @dataclass class EPLBConfig: """Configuration for Expert Parallel Load Balancing (EP).""" window_size: int = 1000 """Window size for expert load recording.""" step_interval: int = 3000 """ Interval for rearranging experts in expert parallelism. Note that if this is greater than the EPLB window size, only the metrics of the last `lb_window_size` steps will be used for rearranging experts. """ num_redundant_experts: int = Field(default=0, ge=0) """Number of redundant experts to use for expert parallelism.""" log_balancedness: bool = False """ Log the balancedness each step of expert parallelism. This is turned off by default since it will cause communication overhead. """ use_async: bool = False """ Whether to use non-blocking EPLB. """ policy: EPLBPolicyOption = "default" """The policy type for expert parallel load balancing (EPLB).""" @config @dataclass class ParallelConfig: """Configuration for the distributed execution.""" pipeline_parallel_size: int = 1 """Number of pipeline parallel groups.""" tensor_parallel_size: int = 1 """Number of tensor parallel groups.""" prefill_context_parallel_size: int = 1 """Number of prefill context parallel groups.""" data_parallel_size: int = 1 """Number of data parallel groups. MoE layers will be sharded according to the product of the tensor parallel size and data parallel size.""" data_parallel_size_local: int = 1 """Number of local data parallel groups.""" data_parallel_rank: int = 0 """Rank of the data parallel group.""" data_parallel_rank_local: int | None = None """Local rank of the data parallel group, set only in SPMD mode.""" data_parallel_master_ip: str = "127.0.0.1" """IP of the data parallel master.""" data_parallel_rpc_port: int = 29550 """Port for data parallel messaging.""" data_parallel_master_port: int = 29500 """Port of the data parallel master.""" data_parallel_backend: DataParallelBackend = "mp" """Backend to use for data parallel, either "mp" or "ray".""" data_parallel_external_lb: bool = False """Whether to use "external" DP LB mode. Applies only to online serving and when data_parallel_size > 0. This is useful for a "one-pod-per-rank" wide-EP setup in Kubernetes. Set implicitly when --data-parallel-rank is provided explicitly to vllm serve.""" data_parallel_hybrid_lb: bool = False """Whether to use "hybrid" DP LB mode. Applies only to online serving and when data_parallel_size > 0. Enables running an AsyncLLM and API server on a "per-node" basis where vLLM load balances between local data parallel ranks, but an external LB balances between vLLM nodes/replicas. Set explicitly in conjunction with --data-parallel-start-rank.""" is_moe_model: bool | None = None """Whether the deployed model is MoE (if known).""" enable_expert_parallel: bool = False """Use expert parallelism instead of tensor parallelism for MoE layers.""" enable_eplb: bool = False """Enable expert parallelism load balancing for MoE layers.""" eplb_config: EPLBConfig = Field(default_factory=EPLBConfig) """Expert parallelism configuration.""" expert_placement_strategy: ExpertPlacementStrategy = "linear" """The expert placement strategy for MoE layers:\n - "linear": Experts are placed in a contiguous manner. For example, with 4 experts and 2 ranks, rank 0 will have experts [0, 1] and rank 1 will have experts [2, 3].\n - "round_robin": Experts are placed in a round-robin manner. For example, with 4 experts and 2 ranks, rank 0 will have experts [0, 2] and rank 1 will have experts [1, 3]. This strategy can help improve load balancing for grouped expert models with no redundant experts.""" all2all_backend: All2AllBackend = "allgather_reducescatter" """All2All backend for MoE expert parallel communication. Available options: - "naive": Naive all2all implementation using broadcasts\n - "allgather_reducescatter": All2all based on allgather and reducescatter\n - "pplx": Use pplx kernels\n - "deepep_high_throughput": Use deepep high-throughput kernels\n - "deepep_low_latency": Use deepep low-latency kernels\n - "flashinfer_all2allv": Use flashinfer alltoallv kernels for mnnvl""" max_parallel_loading_workers: int | None = None """Maximum number of parallel loading workers when loading model sequentially in multiple batches. To avoid RAM OOM when using tensor parallel and large models.""" disable_custom_all_reduce: bool = False """Disable the custom all-reduce kernel and fall back to NCCL.""" enable_dbo: bool = False """Enable dual batch overlap for the model executor.""" ubatch_size: int = 0 """Number of ubatch size.""" dbo_decode_token_threshold: int = 32 """The threshold for dual batch overlap for batches only containing decodes. If the number of tokens in the request is greater than this threshold, microbatching will be used. Otherwise, the request will be processed in a single batch.""" dbo_prefill_token_threshold: int = 512 # TODO(lucas): tune """The threshold for dual batch overlap for batches that contain one or more prefills. If the number of tokens in the request is greater than this threshold, microbatching will be used. Otherwise, the request will be processed in a single batch.""" disable_nccl_for_dp_synchronization: bool = False """Forces the dp synchronization logic in vllm/v1/worker/dp_utils.py to use Gloo instead of NCCL for its all reduce""" ray_workers_use_nsight: bool = False """Whether to profile Ray workers with nsight, see https://docs.ray.io/en/latest/ray-observability/user-guides/profiling.html#profiling-nsight-profiler.""" ray_runtime_env: RuntimeEnv | None = None """Ray runtime environment to pass to distributed workers.""" placement_group: PlacementGroup | None = None """ray distributed model workers placement group.""" distributed_executor_backend: ( str | DistributedExecutorBackend | type[Executor] | None ) = None """Backend to use for distributed model workers, either "ray" or "mp" (multiprocessing). If the product of pipeline_parallel_size and tensor_parallel_size is less than or equal to the number of GPUs available, "mp" will be used to keep processing on a single host. Otherwise, an error will be raised. To use "mp" you must also set nnodes, and to use "ray" you must manually set distributed_executor_backend to "ray". Note that tpu only support Ray for distributed inference.""" worker_cls: str = "auto" """The full name of the worker class to use. If "auto", the worker class will be determined based on the platform.""" sd_worker_cls: str = "auto" """The full name of the worker class to use for speculative decoding. If "auto", the worker class will be determined based on the platform.""" worker_extension_cls: str = "" """The full name of the worker extension class to use. The worker extension class is dynamically inherited by the worker class. This is used to inject new attributes and methods to the worker class for use in collective_rpc calls.""" master_addr: str = "127.0.0.1" """distributed master address for multi-node distributed inference when distributed_executor_backend is mp.""" master_port: int = 29501 """distributed master port for multi-node distributed inference when distributed_executor_backend is mp.""" node_rank: int = 0 """distributed node rank for multi-node distributed inference when distributed_executor_backend is mp.""" nnodes: int = 1 """num of nodes for multi-node distributed inference when distributed_executor_backend is mp.""" world_size: int = Field(init=False) """world_size is TPxPP, it affects the number of workers we create.""" rank: int = 0 """Global rank in distributed setup.""" _data_parallel_master_port_list: list[int] = Field(default_factory=list) """List of open port auto-queried for data parallel messaging. Set to be private as it's not intended to be configured by users. """ decode_context_parallel_size: int = 1 """Number of decode context parallel groups, because the world size does not change by dcp, it simply reuse the GPUs of TP group, and tp_size needs to be divisible by dcp_size.""" dcp_kv_cache_interleave_size: int = 1 """ Interleave size of kv_cache storage while using DCP. dcp_kv_cache_interleave_size has been replaced by cp_kv_cache_interleave_size, and will be deprecated when PCP is fully supported. """ cp_kv_cache_interleave_size: int = 1 """Interleave size of kv_cache storage while using DCP or PCP. For `total_cp_rank = pcp_rank * dcp_world_size + dcp_rank`, and `total_cp_world_size = pcp_world_size * dcp_world_size`. store interleave_size tokens on total_cp_rank i, then store next interleave_size tokens on total_cp_rank i+1. Interleave_size=1: token-level alignment, where token `i` is stored on total_cp_rank `i % total_cp_world_size`. Interleave_size=block_size: block-level alignment, where tokens are first populated to the preceding ranks. Tokens are then stored in (rank i+1, block j) only after (rank i, block j) is fully occupied. Block_size should be greater than or equal to cp_kv_cache_interleave_size. Block_size should be divisible by cp_kv_cache_interleave_size. """ data_parallel_index: int = Field(init=False) """Equal to the data parallel rank but not used for torch process groups and not overridden for dense models.""" _api_process_count: int = Field(default=1, gt=0) """ The number of API processes initialized. Note: This is an internal config that is only valid for and should only be set by API server scale-out. """ _api_process_rank: int = Field(default=0, ge=-1) """ The rank of this API process, or `-1` for engine core processes under API server scale-out. Note: This is an internal config that is only valid for and should only be set by API server scale-out. """ @model_validator(mode="after") def _validate_parallel_config(self) -> Self: if self._api_process_rank >= self._api_process_count: raise ValueError( "Invalid value of `_api_process_rank`. " f"Expected to be `-1` or `[0, {self._api_process_count})`, " f"but found: {self._api_process_rank}" ) if self.data_parallel_size_local > self.data_parallel_size: raise ValueError( f"data_parallel_size_local ({self.data_parallel_size_local}) " f"must be <= data_parallel_size ({self.data_parallel_size})" ) if self.data_parallel_size <= 1 and self.data_parallel_external_lb: raise ValueError( "data_parallel_external_lb can only be set when data_parallel_size > 1" ) if self.enable_eplb: if not current_platform.is_cuda_alike(): raise ValueError( "Expert parallelism load balancing is only supported on " "CUDA devices or ROCm devices now." ) if not self.enable_expert_parallel: raise ValueError("enable_expert_parallel must be True to use EPLB.") if self.tensor_parallel_size * self.data_parallel_size <= 1: raise ValueError( "EPLB requires tensor_parallel_size or data_parallel_size " f"to be greater than 1, but got " f"TP={self.tensor_parallel_size},DP={self.data_parallel_size}." ) else: if self.eplb_config.num_redundant_experts != 0: raise ValueError( "num_redundant_experts is set to " f"{self.eplb_config.num_redundant_experts} but EPLB is not " "enabled. Either enable EPLB or unset " "num_redundant_experts." ) return self @property def world_size_across_dp(self) -> int: """world_size_across_dp is TPxPPxDP, it is the size of the world including data parallelism.""" return self.world_size * self.data_parallel_size @property def use_ubatching(self) -> bool: return self.enable_dbo or self.ubatch_size > 1 @property def num_ubatches(self) -> int: return 2 if self.enable_dbo else self.ubatch_size def get_next_dp_init_port(self) -> int: """ We might need to initialize process groups in multiple processes that is related to data parallelism, e.g. both in the worker and in the engine, which can live in different processes. To avoid port conflicts, we pop a new port from the prepared port list each time we need to initialize a new process group related to data parallelism. """ if self._data_parallel_master_port_list: answer = self._data_parallel_master_port_list.pop() else: answer = self.data_parallel_master_port self.data_parallel_master_port += 1 return answer def stateless_init_dp_group(self) -> ProcessGroup: # NOTE: In high-concurrency scenarios multiple processes # can pick the same (currently free) port through a race # condition when calling `get_open_port()`. When the first # process binds the port the others will subsequently fail # with `torch.distributed.DistNetworkError: EADDRINUSE`. # To make the initialization more robust we retry a few times # with a fresh port whenever this specific error is observed. from torch.distributed import DistNetworkError from vllm.distributed.utils import ( stateless_init_torch_distributed_process_group, ) max_retries = 5 last_exc: Exception | None = None for _ in range(max_retries): try: # use gloo since the engine process might not have cuda device return stateless_init_torch_distributed_process_group( self.data_parallel_master_ip, self.get_next_dp_init_port(), self.data_parallel_rank, self.data_parallel_size, backend=current_platform.dist_backend, ) except DistNetworkError as e: # We only want to retry when the root cause is EADDRINUSE. if "EADDRINUSE" in str(e): logger.warning("Address already in use. Retrying with a new port.") last_exc = e continue # try again with a new port raise e # If we get here all retries have failed. assert last_exc is not None raise last_exc # The all_reduce at the end of attention (during o_proj) means that # inputs are replicated across each rank of the tensor parallel group. # If using expert-parallelism with DeepEP All2All ops, replicated # tokens results in useless duplicate computation and communication. # # In this case, ensure the input to the experts is sequence parallel # to avoid the excess work. # # Not needed for pplx-kernels as it can handle duplicate input tokens. @property def use_sequence_parallel_moe(self) -> bool: return ( self.all2all_backend in ( "allgather_reducescatter", "naive", "deepep_high_throughput", "deepep_low_latency", ) and self.enable_expert_parallel and self.tensor_parallel_size > 1 and self.data_parallel_size > 1 ) @property def node_rank_within_dp(self) -> int: return self.node_rank % self.nnodes_within_dp @property def nnodes_within_dp(self) -> int: if self.nnodes == 1: return 1 data_parallel_node_size = ( self.data_parallel_size // self.data_parallel_size_local ) return self.nnodes // data_parallel_node_size @property def local_world_size(self) -> int: return self.world_size // self.nnodes_within_dp @staticmethod def has_unfinished_dp(dp_group: ProcessGroup, has_unfinished: bool) -> bool: tensor = torch.tensor([has_unfinished], dtype=torch.int32, device="cpu") # dp rank 0: has_unfinished_seqs=True # dp rank 1: has_unfinished_seqs=False # aggregated: has_unfinished_seqs=True # so this is an OR operation, i.e. MAX in integers torch.distributed.all_reduce(tensor, op=ReduceOp.MAX, group=dp_group) aggregated_has_unfinished = bool(tensor.item()) return aggregated_has_unfinished @staticmethod def sync_kv_cache_memory_size(dp_group: ProcessGroup, kv_cache_memory: int) -> int: if kv_cache_memory == -1: kv_cache_memory = torch.iinfo(torch.int64).max tensor = torch.tensor([kv_cache_memory], dtype=torch.int64, device="cpu") # we cannot use broadcast for stateless dp group since it depends # on global rank torch.distributed.all_reduce(tensor, op=ReduceOp.MIN, group=dp_group) return tensor.item() def compute_hash(self): """ Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. This hash is also used for DP worker configuration validation to prevent hangs from mismatched collective communication patterns. """ ignored_factors = { # Derived/runtime topology, networking, or launch details "data_parallel_rank", "data_parallel_rank_local", "data_parallel_size_local", "data_parallel_index", "data_parallel_backend", "data_parallel_external_lb", "data_parallel_hybrid_lb", "data_parallel_master_ip", "data_parallel_master_port", "_data_parallel_master_port_list", "data_parallel_rpc_port", "rank", "master_addr", "master_port", "node_rank", "nnodes", "max_parallel_loading_workers", "disable_custom_all_reduce", "ray_workers_use_nsight", "ray_runtime_env", "placement_group", "distributed_executor_backend", "worker_cls", "sd_worker_cls", "worker_extension_cls", "_api_process_count", "_api_process_rank", } from vllm.config.utils import get_hash_factors, hash_factors factors = get_hash_factors(self, ignored_factors) return hash_factors(factors) def __post_init__(self) -> None: # Set all2all_backend from env var if not specified, with deprecation warning if envs.is_set("VLLM_ALL2ALL_BACKEND"): logger.warning_once( "VLLM_ALL2ALL_BACKEND environment variable is deprecated and " "will be removed in v0.15.0. Please use the " "--all2all-backend command-line argument instead." ) self.all2all_backend = envs.VLLM_ALL2ALL_BACKEND # Continue with the rest of the initialization self.world_size = ( self.pipeline_parallel_size * self.tensor_parallel_size * self.prefill_context_parallel_size ) if self.distributed_executor_backend == "external_launcher": logger.info("Using external launcher for distributed inference.") self.world_size *= self.data_parallel_size if self.data_parallel_size > 1 or self.data_parallel_size_local == 0: # Data parallel was specified in the engine args. if self.distributed_executor_backend == "external_launcher": # For external launcher, # we need to set the data parallel rank automatically self.data_parallel_rank = int(os.environ["RANK"]) // ( self.world_size // self.data_parallel_size ) logger.info( "Set data_parallel_rank to %d automatically.", self.data_parallel_rank, ) if not self._data_parallel_master_port_list: self._data_parallel_master_port_list = get_open_ports_list(5) self.data_parallel_master_port = self._data_parallel_master_port_list.pop() if not (0 <= self.data_parallel_rank < self.data_parallel_size): raise ValueError( f"data_parallel_rank ({self.data_parallel_rank})" f" must be in the range [0, {self.data_parallel_size})" ) else: # Otherwise fall back to env vars (e.g. for offline SPMD case). self.data_parallel_size = envs.VLLM_DP_SIZE self.data_parallel_rank = envs.VLLM_DP_RANK self.data_parallel_rank_local = envs.VLLM_DP_RANK_LOCAL self.data_parallel_master_ip = envs.VLLM_DP_MASTER_IP self.data_parallel_master_port = envs.VLLM_DP_MASTER_PORT if self.data_parallel_size > 1 and self.is_moe_model is False: raise ValueError( "Offline data parallel mode is not supported/useful" " for dense models." ) self.data_parallel_index = self.data_parallel_rank if self.distributed_executor_backend == "external_launcher": os.environ["VLLM_ENABLE_V1_MULTIPROCESSING"] = "0" logger.info("Disabling V1 multiprocessing for external launcher.") if self.distributed_executor_backend is None and self.world_size > 1: # We use multiprocessing by default if world_size fits on the # current node and we aren't in a ray placement group. from vllm.v1.executor import ray_utils backend: DistributedExecutorBackend = "mp" ray_found = ray_utils.ray_is_available() if current_platform.is_tpu() and envs.VLLM_XLA_USE_SPMD: backend = "uni" elif current_platform.is_cuda() and self.nnodes > 1: backend = "mp" elif ( current_platform.is_cuda() and cuda_device_count_stateless() < self.world_size ): gpu_count = cuda_device_count_stateless() raise ValueError( f"World size ({self.world_size}) is larger than the number of " f"available GPUs ({gpu_count}) in this node. If this is " "intentional and you are using:\n" "- ray, set '--distributed-executor-backend ray'.\n" "- multiprocessing, set '--nnodes' appropriately." ) elif self.data_parallel_backend == "ray": logger.info( "Using ray distributed inference because " "data_parallel_backend is ray" ) backend = "ray" elif ray_found: if self.placement_group: backend = "ray" else: from ray import is_initialized as ray_is_initialized if ray_is_initialized(): from ray.util import get_current_placement_group if get_current_placement_group(): backend = "ray" self.distributed_executor_backend = backend logger.debug("Defaulting to use %s for distributed inference", backend) if self.distributed_executor_backend is None and self.world_size == 1: self.distributed_executor_backend = "uni" if self.max_parallel_loading_workers is not None: logger.warning( "max_parallel_loading_workers is currently " "not supported and will be ignored." ) allowed_backends = ("mp", "uni", "external_launcher") if ( self.distributed_executor_backend not in allowed_backends and self.nnodes > 1 ): raise ValueError( "nnodes > 1 can only be set when distributed executor " "backend is mp, uni or external_launcher." ) @property def use_ray(self) -> bool: return self.distributed_executor_backend == "ray" or ( isinstance(self.distributed_executor_backend, type) and getattr(self.distributed_executor_backend, "uses_ray", False) ) @model_validator(mode="after") def _verify_args(self) -> Self: # Lazy import to avoid circular import from vllm.v1.executor import Executor # Enable batch invariance settings if requested if vllm_is_batch_invariant(): self.disable_custom_all_reduce = True if ( self.distributed_executor_backend is not None and not isinstance(self.distributed_executor_backend, str) and not ( isinstance(self.distributed_executor_backend, type) and issubclass(self.distributed_executor_backend, Executor) ) ): raise ValueError( "Unrecognized distributed executor backend " f"{self.distributed_executor_backend}. Supported " "values are 'ray', 'mp' 'uni', 'external_launcher', " " custom Executor subclass or its import path." ) if self.use_ray: from vllm.v1.executor import ray_utils ray_utils.assert_ray_available() if not current_platform.use_custom_allreduce(): self.disable_custom_all_reduce = True logger.debug( "Disabled the custom all-reduce kernel because it is not " "supported on current platform." ) if self.nnodes > 1: self.disable_custom_all_reduce = True logger.debug( "Disabled the custom all-reduce since we are running on multi-node." ) if self.ray_workers_use_nsight and not self.use_ray: raise ValueError( "Unable to use nsight profiling unless workers run with Ray." ) return self
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/inputs/parse.py
vllm/inputs/parse.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Sequence from typing import TYPE_CHECKING, Literal, NamedTuple, TypeAlias, TypedDict, cast from typing_extensions import TypeIs from vllm.utils.collection_utils import is_list_of from .data import ( EmbedsPrompt, ExplicitEncoderDecoderPrompt, ProcessorInputs, PromptType, SingletonInputs, SingletonPrompt, TextPrompt, TokensPrompt, ) if TYPE_CHECKING: import torch def parse_raw_prompts( prompt: str | list[str] | list[int] | list[list[int]], ) -> Sequence[TextPrompt] | Sequence[TokensPrompt]: if isinstance(prompt, str): # case 1: a string return [TextPrompt(prompt=prompt)] if isinstance(prompt, list): if len(prompt) == 0: raise ValueError("please provide at least one prompt") # case 2: array of strings if is_list_of(prompt, str): prompt = cast(list[str], prompt) return [TextPrompt(prompt=elem) for elem in prompt] # case 3: array of tokens if is_list_of(prompt, int): prompt = cast(list[int], prompt) return [TokensPrompt(prompt_token_ids=prompt)] # case 4: array of token arrays if is_list_of(prompt, list): if len(prompt) == 1 and isinstance(prompt[0], list) and len(prompt[0]) == 0: raise ValueError("please provide at least one prompt") for elem in prompt: if not isinstance(elem, list): raise TypeError( "prompt must be a list of lists, but found a non-list element." ) if not is_list_of(elem, int): raise TypeError( "Nested lists of tokens must contain only integers." ) prompt = cast(list[list[int]], prompt) return [TokensPrompt(prompt_token_ids=elem) for elem in prompt] raise TypeError( "prompt must be a string, array of strings, " "array of tokens, or array of token arrays" ) class ParsedStrPrompt(TypedDict): type: Literal["str"] content: str class ParsedTextPrompt(TypedDict): type: Literal["text"] content: TextPrompt class ParsedTokensPrompt(TypedDict): type: Literal["tokens"] content: TokensPrompt class ParsedEmbedsPrompt(TypedDict): type: Literal["embeds"] content: EmbedsPrompt ParsedSingletonPrompt: TypeAlias = ( ParsedStrPrompt | ParsedTextPrompt | ParsedTokensPrompt | ParsedEmbedsPrompt ) def parse_singleton_prompt(prompt: SingletonPrompt) -> ParsedSingletonPrompt: if isinstance(prompt, str): return ParsedStrPrompt(type="str", content=prompt) elif isinstance(prompt, dict): # Type ignores are because mypy does not correctly infer the TypedDicts # Pyright does succeed. if "prompt_embeds" in prompt: return ParsedEmbedsPrompt(type="embeds", content=prompt) # type: ignore[typeddict-item] elif "prompt_token_ids" in prompt: return ParsedTokensPrompt(type="tokens", content=prompt) # type: ignore[typeddict-item] elif "prompt" in prompt: return ParsedTextPrompt(type="text", content=prompt) raise TypeError( "inputs must be a string, TextPrompt, TokensPrompt, or EmbedsPrompt" ) def is_explicit_encoder_decoder_prompt( prompt: PromptType, ) -> TypeIs[ExplicitEncoderDecoderPrompt]: return isinstance(prompt, dict) and "encoder_prompt" in prompt def split_enc_dec_inputs( inputs: ProcessorInputs, ) -> tuple[SingletonInputs | None, SingletonInputs]: if "encoder" in inputs and "decoder" in inputs: # NOTE: This passes pyright but not mypy return ( inputs["encoder"], # type: ignore[typeddict-item] inputs["decoder"], # type: ignore[typeddict-item] ) return None, inputs class PromptComponents(NamedTuple): text: str | None = None token_ids: list[int] | None = None embeds: "torch.Tensor | None" = None def get_prompt_components(prompt: PromptType) -> PromptComponents: if isinstance(prompt, str): return PromptComponents(text=prompt) if encoder_prompt := prompt.get("encoder_prompt"): return get_prompt_components(encoder_prompt) # type: ignore[arg-type] return PromptComponents( text=prompt.get("prompt"), # type: ignore[arg-type] token_ids=prompt.get("prompt_token_ids"), # type: ignore[arg-type] embeds=prompt.get("prompt_embeds"), )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/inputs/__init__.py
vllm/inputs/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from .data import ( DataPrompt, DecoderOnlyInputs, EmbedsInputs, EmbedsPrompt, EncoderDecoderInputs, ExplicitEncoderDecoderPrompt, ProcessorInputs, PromptType, SingletonInputs, SingletonPrompt, TextPrompt, TokenInputs, TokensPrompt, build_explicit_enc_dec_prompt, embeds_inputs, to_enc_dec_tuple_list, token_inputs, zip_enc_dec_prompts, ) __all__ = [ "DataPrompt", "TextPrompt", "TokensPrompt", "PromptType", "SingletonPrompt", "ExplicitEncoderDecoderPrompt", "TokenInputs", "EmbedsInputs", "EmbedsPrompt", "token_inputs", "embeds_inputs", "DecoderOnlyInputs", "EncoderDecoderInputs", "ProcessorInputs", "SingletonInputs", "build_explicit_enc_dec_prompt", "to_enc_dec_tuple_list", "zip_enc_dec_prompts", ]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/inputs/preprocess.py
vllm/inputs/preprocess.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Mapping from typing import Any, cast from typing_extensions import assert_never from vllm.config import ModelConfig, ObservabilityConfig from vllm.logger import init_logger from vllm.multimodal import MULTIMODAL_REGISTRY, MultiModalRegistry from vllm.multimodal.cache import BaseMultiModalProcessorCache from vllm.multimodal.inputs import ( MultiModalDataDict, MultiModalEncDecInputs, MultiModalInputs, MultiModalUUIDDict, ) from vllm.multimodal.processing import BaseMultiModalProcessor from vllm.tokenizers import TokenizerLike from vllm.utils.jsontree import json_iter_leaves from vllm.v1.metrics.stats import MultiModalCacheStats from .data import ( DecoderOnlyInputs, EmbedsInputs, EmbedsPrompt, EncoderDecoderInputs, ExplicitEncoderDecoderPrompt, ProcessorInputs, PromptType, SingletonInputs, SingletonPrompt, TextPrompt, TokenInputs, TokensPrompt, embeds_inputs, token_inputs, ) from .parse import is_explicit_encoder_decoder_prompt, parse_singleton_prompt logger = init_logger(__name__) class InputPreprocessor: def __init__( self, model_config: ModelConfig, tokenizer: TokenizerLike | None, observability_config: ObservabilityConfig | None = None, mm_registry: MultiModalRegistry = MULTIMODAL_REGISTRY, mm_processor_cache: BaseMultiModalProcessorCache | None = None, ) -> None: super().__init__() self.model_config = model_config self.tokenizer = tokenizer self.observability_config = observability_config self.mm_registry = mm_registry self.mm_processor_cache = mm_processor_cache self.mm_cache_stats = MultiModalCacheStats() if mm_processor_cache else None def get_tokenizer(self) -> TokenizerLike: if self.tokenizer is None: raise ValueError( "You cannot pass text prompts when `skip_tokenizer_init=True`" ) return self.tokenizer def get_bos_token_id(self) -> int | None: if self.tokenizer is None: logger.warning_once( "Using None for BOS token id because tokenizer is not initialized" ) return None return self.tokenizer.bos_token_id def get_eos_token_id(self) -> int | None: if self.tokenizer is None: logger.warning_once( "Using None for EOS token id because tokenizer is not initialized" ) return None return self.tokenizer.eos_token_id def get_decoder_start_token_id(self) -> int | None: """ Obtain the decoder start token id employed by an encoder/decoder model. Returns None for non-encoder/decoder models or if the model config is unavailable. """ if not self.model_config.is_encoder_decoder: logger.warning_once( "Using None for decoder start token id because " "this is not an encoder/decoder model." ) return None if self.model_config is None or self.model_config.hf_config is None: logger.warning_once( "Using None for decoder start token id because " "model config is not available." ) return None dec_start_token_id = getattr( self.model_config.hf_config, "decoder_start_token_id", None ) if dec_start_token_id is None: logger.warning_once( "Falling back on <BOS> for decoder start token " "id because decoder start token id is not " "available." ) dec_start_token_id = self.get_bos_token_id() return dec_start_token_id def _get_default_enc_dec_decoder_prompt(self) -> list[int]: """ Specifically for encoder/decoder models: generate a default decoder prompt for when the user specifies only the encoder prompt. Encoder/decoder models utilize the decoder prompt in different ways; as new models are added, it is intended that this function will be extended to produce differing default decoder prompts, depending on the model variety. Absent a special case, the default behavior of this method is to mirror the behavior of the HuggingFace (HF) GenerationMixin for a None decoder prompt, which is to employ a logit processor setting to force the first decoded token to be <BOS>. Here, this behavior is approximated by having the "default" decoder prompt be <BOS>. However, it is possible that in the future other models may have different or more complex logic for the default decoder prompt. This motivates having a special helper method for default decoder prompts. Returns: * prompt_token_ids """ bos_token_id = self.get_bos_token_id() assert bos_token_id is not None return [bos_token_id] def _prepare_decoder_input_ids_for_generation( self, decoder_input_ids: list[int] | None, ) -> list[int]: """ Prepares `decoder_input_ids` for generation with encoder-decoder models. Based on: https://github.com/huggingface/transformers/blob/4037a2b5b1278736e566aec12e169100275545ea/src/transformers/generation/utils.py specifically, `GenerationMixin._prepare_decoder_input_ids_for_generation()`. Arguments: * decoder_input_ids: input token ids to preprocess Returns: * Processed token list """ decoder_start_token_id = self.get_decoder_start_token_id() assert decoder_start_token_id is not None if decoder_input_ids is None: # no decoder prompt input -> # use decoder_start_token_id as decoder_input_ids decoder_input_ids = self._get_default_enc_dec_decoder_prompt() if ( len(decoder_input_ids) == 0 or decoder_input_ids[0] != decoder_start_token_id ): decoder_input_ids = [decoder_start_token_id] + decoder_input_ids return decoder_input_ids def _get_tokenization_kw( self, overrides: dict[str, Any] | None = None, ) -> dict[str, Any]: kwargs = dict[str, Any]() if self.model_config.is_encoder_decoder: # For Whisper, special tokens should be provided by the user based # on the task and language of their request. Also needed to avoid # appending an EOS token to the prompt which disrupts generation. kwargs["add_special_tokens"] = False if overrides: kwargs.update(overrides) return kwargs def _tokenize_prompt( self, prompt: str, tokenization_kwargs: dict[str, Any] | None = None, ) -> list[int]: """ Apply the model's tokenizer to a text prompt, returning the corresponding token IDs. """ tokenizer = self.get_tokenizer() tokenization_kwargs = self._get_tokenization_kw(tokenization_kwargs) encoder_config = self.model_config.encoder_config if encoder_config and encoder_config.get("do_lower_case", False): prompt = prompt.lower() return tokenizer.encode(prompt, **tokenization_kwargs) def _get_mm_processor(self) -> BaseMultiModalProcessor: if not hasattr(self, "_mm_processor"): self._mm_processor = self.mm_registry.create_processor( self.model_config, self.observability_config, tokenizer=self.tokenizer, cache=self.mm_processor_cache, ) return self._mm_processor def _process_multimodal( self, prompt: str | list[int], mm_data: MultiModalDataDict, mm_processor_kwargs: Mapping[str, object] | None, tokenization_kwargs: dict[str, Any] | None = None, *, mm_uuids: MultiModalUUIDDict | None = None, ) -> MultiModalInputs: """ Apply the model's multi-modal processor to a multi-modal prompt, returning the corresponding token IDs and metadata. """ mm_processor = self._get_mm_processor() if mm_processor_kwargs is None: mm_processor_kwargs = {} mm_input = mm_processor.apply( prompt, mm_data, hf_processor_mm_kwargs=mm_processor_kwargs, tokenization_kwargs=tokenization_kwargs, mm_uuids=mm_uuids, ) mm_hashes = mm_input["mm_hashes"] # Validate that all mm items have a string as their hash contains_only_strings = all( isinstance(leaf, str) for leaf in json_iter_leaves(mm_hashes) ) if not contains_only_strings: raise ValueError( f"mm_hashes must contain only strings, got: {mm_hashes}. " "This is likely due to an incorrect custom implementation of " "MultiModalProcessor.apply method." ) return mm_input def _process_embeds( self, parsed_content: EmbedsPrompt, ) -> EmbedsInputs: if not self.model_config.enable_prompt_embeds: raise ValueError( "You must set `--enable-prompt-embeds` to input `prompt_embeds`." ) prompt_embeds = parsed_content["prompt_embeds"] # prompt_embeds must be (seq_len, hidden_size), but if the user # passes in a batch of size 1, i.e. (1, seq_len, hidden_size), # we can unambiguously process the intent by squeezing the batch # dimension. if prompt_embeds.ndim == 3: prompt_embeds = prompt_embeds.squeeze(dim=0) if prompt_embeds.ndim != 2: raise ValueError("prompt_embeds must be of shape (seq_len, hidden_size).") # Tensors must be on CPU for serialization between processes # in the MsgpackEncoder. Casting to CPU here ensures that there is no # hidden device transfer in the critical path of generation. prompt_embeds = prompt_embeds.cpu() return embeds_inputs( prompt_embeds=prompt_embeds, cache_salt=parsed_content.get("cache_salt") ) def _truncate_inputs( self, inputs: list[int], tokenization_kwargs: dict[str, Any] | None = None ) -> list[int]: if ( not tokenization_kwargs or "truncation" not in tokenization_kwargs or self.tokenizer is None ): return inputs max_length = tokenization_kwargs["max_length"] if self.tokenizer.truncation_side == "left": return inputs[-max_length:] else: return inputs[:max_length] def _process_tokens( self, parsed_content: TokensPrompt, tokenization_kwargs: dict[str, Any] | None = None, *, mm_uuids: MultiModalUUIDDict | None = None, ) -> TokenInputs | MultiModalInputs: prompt_token_ids = self._truncate_inputs( parsed_content["prompt_token_ids"], tokenization_kwargs ) inputs: TokenInputs | MultiModalInputs if multi_modal_data := parsed_content.get("multi_modal_data"): inputs = self._process_multimodal( prompt_token_ids, multi_modal_data, parsed_content.get("mm_processor_kwargs") or {}, tokenization_kwargs=tokenization_kwargs, mm_uuids=mm_uuids, ) else: inputs = token_inputs(prompt_token_ids) if cache_salt := parsed_content.get("cache_salt"): inputs["cache_salt"] = cache_salt return inputs def _process_text( self, parsed_content: TextPrompt, tokenization_kwargs: dict[str, Any] | None = None, *, mm_uuids: MultiModalUUIDDict | None = None, ) -> TokenInputs | MultiModalInputs: prompt_text = parsed_content["prompt"] inputs: TokenInputs | MultiModalInputs if multi_modal_data := parsed_content.get("multi_modal_data"): inputs = self._process_multimodal( prompt_text, multi_modal_data, parsed_content.get("mm_processor_kwargs") or {}, tokenization_kwargs=tokenization_kwargs, mm_uuids=mm_uuids, ) else: prompt_token_ids = self._tokenize_prompt( prompt_text, tokenization_kwargs=tokenization_kwargs, ) inputs = token_inputs(prompt_token_ids) if cache_salt := parsed_content.get("cache_salt"): inputs["cache_salt"] = cache_salt return inputs def _prompt_to_llm_inputs( self, prompt: SingletonPrompt, tokenization_kwargs: dict[str, Any] | None = None, *, mm_uuids: MultiModalUUIDDict | None = None, ) -> SingletonInputs: """ Extract the singleton inputs from a prompt. Arguments: * prompt: single encoder or decoder input prompt Returns: * [`SingletonInputs`][vllm.inputs.data.SingletonInputs] instance """ parsed = parse_singleton_prompt(prompt) if parsed["type"] == "embeds": return self._process_embeds(parsed["content"]) if parsed["type"] == "tokens": return self._process_tokens( parsed["content"], mm_uuids=mm_uuids, ) if parsed["type"] == "text": return self._process_text( parsed["content"], tokenization_kwargs=tokenization_kwargs, mm_uuids=mm_uuids, ) if parsed["type"] == "str": return self._process_text( TextPrompt(prompt=parsed["content"]), tokenization_kwargs=tokenization_kwargs, mm_uuids=mm_uuids, ) assert_never(parsed) def _build_enc_dec_llm_inputs( self, encoder_inputs: SingletonInputs, decoder_inputs: SingletonInputs | None, ) -> EncoderDecoderInputs: if ( encoder_inputs["type"] == "embeds" or decoder_inputs and decoder_inputs["type"] == "embeds" ): raise ValueError( "Embedding inputs are not supported for encoder-decoder models" ) # Needed for mypy encoder_inputs = cast(TokenInputs | MultiModalInputs, encoder_inputs) decoder_inputs = cast(TokenInputs | MultiModalInputs | None, decoder_inputs) if decoder_inputs is None: if self.model_config.hf_config.model_type == "whisper": # For Whisper models, the text prompt should go to the decoder. # If no explicit encoder/decoder inputs, then copy the prompt # from the encoder to the decoder. The encoder tokens are later # overridden by the audio features. dec_token_ids = encoder_inputs["prompt_token_ids"].copy() else: dec_token_ids = self._prepare_decoder_input_ids_for_generation(None) decoder_inputs = token_inputs(dec_token_ids) else: if "multi_modal_data" in decoder_inputs: raise ValueError( "Multi-modal decoder inputs of encoder-" "decoder models are not supported yet" ) dec_token_ids = self._prepare_decoder_input_ids_for_generation( decoder_inputs["prompt_token_ids"] ) decoder_inputs["prompt_token_ids"] = dec_token_ids return EncoderDecoderInputs( encoder=encoder_inputs, decoder=decoder_inputs, ) def _split_enc_dec_mm_inputs( self, inputs: SingletonInputs | MultiModalEncDecInputs, decoder_inputs_to_override: SingletonInputs | None = None, ) -> tuple[SingletonInputs, SingletonInputs]: """ For encoder/decoder models only: Separate Encoder/Decoder inputs from a MultiModalEncDecInputs """ if ( inputs["type"] == "embeds" or decoder_inputs_to_override and decoder_inputs_to_override["type"] == "embeds" ): raise ValueError( "Embedding inputs are not supported for encoder-decoder models" ) # Needed for mypy inputs = cast( TokenInputs | MultiModalInputs | MultiModalEncDecInputs, inputs, ) decoder_inputs_to_override = cast( TokenInputs | MultiModalInputs | None, decoder_inputs_to_override, ) encoder_inputs: SingletonInputs decoder_inputs: SingletonInputs if inputs["type"] == "multimodal": # Multimodal data inputs if "encoder_prompt_token_ids" not in inputs: raise RuntimeError( "You should register an encoder-decoder " "multi-modal processor for encoder-decoder " "models." ) inputs = cast(MultiModalEncDecInputs, inputs) encoder_inputs = token_inputs(inputs["encoder_prompt_token_ids"]) decoder_prompt_inputs = decoder_inputs_to_override or inputs decoder_inputs = MultiModalInputs( type="multimodal", prompt_token_ids=decoder_prompt_inputs["prompt_token_ids"], mm_kwargs=inputs["mm_kwargs"], mm_hashes=inputs["mm_hashes"], mm_placeholders=inputs["mm_placeholders"], ) if cache_salt := inputs.get("cache_salt"): decoder_inputs["cache_salt"] = cache_salt elif inputs["type"] == "token": # Text-only inputs encoder_inputs = token_inputs(prompt_token_ids=[]) decoder_inputs = decoder_inputs_to_override or inputs else: assert_never(inputs) # type: ignore[arg-type] return encoder_inputs, decoder_inputs def _process_encoder_decoder_prompt( self, prompt: PromptType, tokenization_kwargs: dict[str, Any] | None = None, *, mm_uuids: MultiModalUUIDDict | None = None, ) -> EncoderDecoderInputs: """ For encoder/decoder models only: Process an input prompt into an [`EncoderDecoderInputs`][vllm.inputs.data.EncoderDecoderInputs] instance. There are two types of input prompts: singleton prompts which carry only the encoder prompt, and explicit encoder/decoder prompts which carry both the encoder and the decoder prompts as member variables. This function handles the following scenarios: * Singleton encoder prompt: extract encoder prompt token ids & infer default decoder prompt token ids * Explicit encoder/decoder prompt: extract encoder and decoder prompt token ids Note that for Explicit encoder/decoder prompts, each sub-prompt (encoder or decoder prompt) can have any possible singleton type; thus this method relies on helper functions to obtain token ids for the sub-prompts. Arguments: * prompt: an input prompt Returns: * [`EncoderDecoderInputs`][vllm.inputs.data.EncoderDecoderInputs] instance """ encoder_inputs: SingletonInputs decoder_inputs: SingletonInputs | None if is_explicit_encoder_decoder_prompt(prompt): # `cast` is needed for mypy, but not pyright prompt_ = cast(ExplicitEncoderDecoderPrompt, prompt) encoder_inputs = self._prompt_to_llm_inputs( prompt_["encoder_prompt"], tokenization_kwargs=tokenization_kwargs, mm_uuids=mm_uuids, ) if (decoder_input := prompt_["decoder_prompt"]) is None: decoder_inputs = None else: decoder_inputs = self._prompt_to_llm_inputs( decoder_input, tokenization_kwargs=tokenization_kwargs ) # For multimodal model, override decoder prompt from processor # with explicit decoder prompt. if self.model_config.is_multimodal_model: encoder_inputs, decoder_inputs = self._split_enc_dec_mm_inputs( encoder_inputs, decoder_inputs ) else: # `cast` is needed for mypy, but not pyright inputs = self._prompt_to_llm_inputs( cast(SingletonPrompt, prompt), tokenization_kwargs=tokenization_kwargs, mm_uuids=mm_uuids, ) if self.model_config.is_multimodal_model: # Encoder-Decoder Multimodal model encoder_inputs, decoder_inputs = self._split_enc_dec_mm_inputs(inputs) else: encoder_inputs = inputs decoder_inputs = None return self._build_enc_dec_llm_inputs(encoder_inputs, decoder_inputs) def _build_decoder_only_llm_inputs( self, prompt_inputs: DecoderOnlyInputs, ) -> DecoderOnlyInputs: if "prompt_token_ids" in prompt_inputs: prompt_inputs = cast( TokenInputs | MultiModalInputs, prompt_inputs ) # Needed for mypy return prompt_inputs def _process_decoder_only_prompt( self, prompt: SingletonPrompt, tokenization_kwargs: dict[str, Any] | None = None, *, mm_uuids: MultiModalUUIDDict | None = None, ) -> DecoderOnlyInputs: """ For decoder-only models: Process an input prompt into a [`DecoderOnlyInputs`][vllm.inputs.data.DecoderOnlyInputs] instance. Arguments: * prompt: input prompt Returns: * [`DecoderOnlyInputs`][vllm.inputs.data.DecoderOnlyInputs] instance """ prompt_comps = self._prompt_to_llm_inputs( prompt, tokenization_kwargs=tokenization_kwargs, mm_uuids=mm_uuids, ) return self._build_decoder_only_llm_inputs(prompt_comps) def _preprocess( self, prompt: PromptType, tokenization_kwargs: dict[str, Any] | None = None, *, mm_uuids: MultiModalUUIDDict | None = None, ) -> ProcessorInputs: if self.model_config.is_encoder_decoder: # Encoder-decoder model requires special mapping of # input prompts to encoder & decoder. return self._process_encoder_decoder_prompt( prompt, tokenization_kwargs, mm_uuids=mm_uuids, ) if is_explicit_encoder_decoder_prompt(prompt): raise ValueError( "Cannot pass encoder-decoder prompt to decoder-only models" ) # Decoder-only operation # `cast` is needed for mypy, but not pyright return self._process_decoder_only_prompt( cast(SingletonPrompt, prompt), tokenization_kwargs=tokenization_kwargs, mm_uuids=mm_uuids, ) def preprocess( self, prompt: PromptType, tokenization_kwargs: dict[str, Any] | None = None, *, mm_uuids: MultiModalUUIDDict | None = None, ) -> ProcessorInputs: """Preprocess the input prompt.""" res = self._preprocess(prompt, tokenization_kwargs, mm_uuids=mm_uuids) if self.mm_processor_cache and self.mm_cache_stats is not None: delta = self.mm_processor_cache.make_stats(delta=True) self.mm_cache_stats.requests += 1 self.mm_cache_stats.queries += delta.total self.mm_cache_stats.hits += delta.hits return res def stat_mm_cache(self) -> MultiModalCacheStats | None: mm_cache_stats = self.mm_cache_stats if mm_cache_stats is None: return None self.mm_cache_stats = MultiModalCacheStats() return mm_cache_stats def clear_mm_cache(self) -> None: if self.mm_processor_cache is not None: self.mm_processor_cache.clear_cache() if self.mm_cache_stats is not None: self.mm_cache_stats.reset = True
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/inputs/data.py
vllm/inputs/data.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Iterable from typing import TYPE_CHECKING, Any, Generic, Literal, TypeAlias, cast import torch from typing_extensions import NotRequired, TypedDict, TypeIs, TypeVar if TYPE_CHECKING: from vllm.multimodal.inputs import ( MultiModalDataDict, MultiModalInputs, MultiModalUUIDDict, ) else: MultiModalDataDict = object MultiModalInputs = object MultiModalUUIDDict = object class TextPrompt(TypedDict): """Schema for a text prompt.""" prompt: str """The input text to be tokenized before passing to the model.""" multi_modal_data: NotRequired[MultiModalDataDict | None] """ Optional multi-modal data to pass to the model, if the model supports it. """ mm_processor_kwargs: NotRequired[dict[str, Any] | None] """ Optional multi-modal processor kwargs to be forwarded to the multimodal input mapper & processor. Note that if multiple modalities have registered mappers etc for the model being considered, we attempt to pass the mm_processor_kwargs to each of them. """ multi_modal_uuids: NotRequired[MultiModalUUIDDict] """ Optional user-specified UUIDs for multimodal items, mapped by modality. Lists must match the number of items per modality and may contain `None`. For `None` entries, the hasher will compute IDs automatically; non-None entries override the default hashes for caching, and MUST be unique per multimodal item. """ cache_salt: NotRequired[str] """ Optional cache salt to be used for prefix caching. """ class TokensPrompt(TypedDict): """Schema for a tokenized prompt.""" prompt_token_ids: list[int] """A list of token IDs to pass to the model.""" prompt: NotRequired[str] """The prompt text corresponding to the token IDs, if available.""" token_type_ids: NotRequired[list[int]] """A list of token type IDs to pass to the cross encoder model.""" multi_modal_data: NotRequired[MultiModalDataDict | None] """ Optional multi-modal data to pass to the model, if the model supports it. """ mm_processor_kwargs: NotRequired[dict[str, Any] | None] """ Optional multi-modal processor kwargs to be forwarded to the multimodal input mapper & processor. Note that if multiple modalities have registered mappers etc for the model being considered, we attempt to pass the mm_processor_kwargs to each of them. """ multi_modal_uuids: NotRequired[MultiModalUUIDDict] """ Optional user-specified UUIDs for multimodal items, mapped by modality. Lists must match the number of items per modality and may contain `None`. For `None` entries, the hasher will compute IDs automatically; non-None entries override the default hashes for caching. """ cache_salt: NotRequired[str] """ Optional cache salt to be used for prefix caching. """ class EmbedsPrompt(TypedDict): """Schema for a prompt provided via token embeddings.""" prompt_embeds: torch.Tensor """The embeddings of the prompt.""" cache_salt: NotRequired[str] """ Optional cache salt to be used for prefix caching. """ class DataPrompt(TypedDict): """Represents generic inputs handled by IO processor plugins.""" data: Any """The input data""" data_format: str """The input data format""" SingletonPrompt: TypeAlias = str | TextPrompt | TokensPrompt | EmbedsPrompt """ Set of possible schemas for a single prompt: - A text prompt ([`str`][] or [`TextPrompt`][vllm.inputs.data.TextPrompt]) - A tokenized prompt ([`TokensPrompt`][vllm.inputs.data.TokensPrompt]) - An embeddings prompt ([`EmbedsPrompt`][vllm.inputs.data.EmbedsPrompt]) Note that "singleton" is as opposed to a data structure which encapsulates multiple prompts, i.e. of the sort which may be utilized for encoder/decoder models when the user desires to express both the encoder & decoder prompts explicitly, i.e. [`ExplicitEncoderDecoderPrompt`][vllm.inputs.data.ExplicitEncoderDecoderPrompt] A prompt of type [`SingletonPrompt`][vllm.inputs.data.SingletonPrompt] may be employed as (1) input to a decoder-only model, (2) input to the encoder of an encoder/decoder model, in the scenario where the decoder-prompt is not specified explicitly, or (3) as a member of a larger data structure encapsulating more than one prompt, i.e. [`ExplicitEncoderDecoderPrompt`][vllm.inputs.data.ExplicitEncoderDecoderPrompt] """ def is_tokens_prompt(prompt: SingletonPrompt) -> TypeIs[TokensPrompt]: return ( isinstance(prompt, dict) and "prompt_token_ids" in prompt and "prompt_embeds" not in prompt ) def is_embeds_prompt(prompt: SingletonPrompt) -> TypeIs[EmbedsPrompt]: return ( isinstance(prompt, dict) and "prompt_token_ids" not in prompt and "prompt_embeds" in prompt ) _T1_co = TypeVar( "_T1_co", bound=SingletonPrompt, default=SingletonPrompt, covariant=True ) _T2_co = TypeVar( "_T2_co", bound=SingletonPrompt, default=SingletonPrompt, covariant=True ) # TODO: Make fields ReadOnly once mypy supports it class ExplicitEncoderDecoderPrompt(TypedDict, Generic[_T1_co, _T2_co]): """ Represents an encoder/decoder model input prompt, comprising an explicit encoder prompt and a decoder prompt. The encoder and decoder prompts, respectively, may be formatted according to any of the [`SingletonPrompt`][vllm.inputs.data.SingletonPrompt] schemas, and are not required to have the same schema. Only the encoder prompt may have multi-modal data. mm_processor_kwargs should be at the top-level, and should not be set in the encoder/decoder prompts, since they are agnostic to the encoder/decoder. Note that an [`ExplicitEncoderDecoderPrompt`][vllm.inputs.data.ExplicitEncoderDecoderPrompt] may not be used as an input to a decoder-only model, and that the `encoder_prompt` and `decoder_prompt` fields of this data structure themselves must be [`SingletonPrompt`][vllm.inputs.data.SingletonPrompt] instances. """ encoder_prompt: _T1_co decoder_prompt: _T2_co | None mm_processor_kwargs: NotRequired[dict[str, Any]] PromptType: TypeAlias = SingletonPrompt | ExplicitEncoderDecoderPrompt """ Set of possible schemas for an LLM input, including both decoder-only and encoder/decoder input types: - A text prompt ([`str`][] or [`TextPrompt`][vllm.inputs.data.TextPrompt]) - A tokenized prompt ([`TokensPrompt`][vllm.inputs.data.TokensPrompt]) - An embeddings prompt ([`EmbedsPrompt`][vllm.inputs.data.EmbedsPrompt]) - A single data structure containing both an encoder and a decoder prompt ([`ExplicitEncoderDecoderPrompt`][vllm.inputs.data.ExplicitEncoderDecoderPrompt]) """ class TokenInputs(TypedDict): """Represents token-based inputs.""" type: Literal["token"] """The type of inputs.""" prompt_token_ids: list[int] """The token IDs of the prompt.""" cache_salt: NotRequired[str] """ Optional cache salt to be used for prefix caching. """ def token_inputs( prompt_token_ids: list[int], cache_salt: str | None = None, ) -> TokenInputs: """Construct [`TokenInputs`][vllm.inputs.data.TokenInputs] from optional values.""" inputs = TokenInputs(type="token", prompt_token_ids=prompt_token_ids) if cache_salt is not None: inputs["cache_salt"] = cache_salt return inputs class EmbedsInputs(TypedDict): """Represents embeddings-based inputs.""" type: Literal["embeds"] """The type of inputs.""" prompt_embeds: torch.Tensor """The embeddings of the prompt.""" cache_salt: NotRequired[str] """ Optional cache salt to be used for prefix caching. """ def embeds_inputs( prompt_embeds: torch.Tensor, cache_salt: str | None = None, ) -> EmbedsInputs: """Construct [`EmbedsInputs`][vllm.inputs.data.EmbedsInputs] from optional values.""" inputs = EmbedsInputs(type="embeds", prompt_embeds=prompt_embeds) if cache_salt is not None: inputs["cache_salt"] = cache_salt return inputs DecoderOnlyInputs: TypeAlias = TokenInputs | EmbedsInputs | MultiModalInputs """ The inputs in [`LLMEngine`][vllm.engine.llm_engine.LLMEngine] before they are passed to the model executor. This specifies the data required for decoder-only models. """ class EncoderDecoderInputs(TypedDict): """ The inputs in [`LLMEngine`][vllm.engine.llm_engine.LLMEngine] before they are passed to the model executor. This specifies the required data for encoder-decoder models. """ encoder: TokenInputs | MultiModalInputs """The inputs for the encoder portion.""" decoder: TokenInputs | MultiModalInputs """The inputs for the decoder portion.""" SingletonInputs: TypeAlias = TokenInputs | EmbedsInputs | MultiModalInputs """ A processed [`SingletonPrompt`][vllm.inputs.data.SingletonPrompt] which can be passed to [`Sequence`][collections.abc.Sequence]. """ ProcessorInputs: TypeAlias = DecoderOnlyInputs | EncoderDecoderInputs """ The outputs from [`vllm.inputs.preprocess.InputPreprocessor`][]. """ _T1 = TypeVar("_T1", bound=SingletonPrompt, default=SingletonPrompt) _T2 = TypeVar("_T2", bound=SingletonPrompt, default=SingletonPrompt) def build_explicit_enc_dec_prompt( encoder_prompt: _T1, decoder_prompt: _T2 | None, mm_processor_kwargs: dict[str, Any] | None = None, ) -> ExplicitEncoderDecoderPrompt[_T1, _T2]: if mm_processor_kwargs is None: mm_processor_kwargs = {} return ExplicitEncoderDecoderPrompt( encoder_prompt=encoder_prompt, decoder_prompt=decoder_prompt, mm_processor_kwargs=mm_processor_kwargs, ) def zip_enc_dec_prompts( enc_prompts: Iterable[_T1], dec_prompts: Iterable[_T2 | None], mm_processor_kwargs: Iterable[dict[str, Any]] | dict[str, Any] | None = None, ) -> list[ExplicitEncoderDecoderPrompt[_T1, _T2]]: """ Zip encoder and decoder prompts together into a list of [`ExplicitEncoderDecoderPrompt`][vllm.inputs.data.ExplicitEncoderDecoderPrompt] instances. `mm_processor_kwargs` may also be provided; if a dict is passed, the same dictionary will be used for every encoder/decoder prompt. If an iterable is provided, it will be zipped with the encoder/decoder prompts. """ if mm_processor_kwargs is None: mm_processor_kwargs = cast(dict[str, Any], {}) if isinstance(mm_processor_kwargs, dict): return [ build_explicit_enc_dec_prompt( encoder_prompt, decoder_prompt, cast(dict[str, Any], mm_processor_kwargs), ) for (encoder_prompt, decoder_prompt) in zip(enc_prompts, dec_prompts) ] return [ build_explicit_enc_dec_prompt(encoder_prompt, decoder_prompt, mm_proc_kwargs) for (encoder_prompt, decoder_prompt, mm_proc_kwargs) in zip( enc_prompts, dec_prompts, mm_processor_kwargs ) ] def to_enc_dec_tuple_list( enc_dec_prompts: Iterable[ExplicitEncoderDecoderPrompt[_T1, _T2]], ) -> list[tuple[_T1, _T2 | None]]: return [ (enc_dec_prompt["encoder_prompt"], enc_dec_prompt["decoder_prompt"]) for enc_dec_prompt in enc_dec_prompts ]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/engine/llm_engine.py
vllm/engine/llm_engine.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from vllm.v1.engine.llm_engine import LLMEngine as V1LLMEngine LLMEngine = V1LLMEngine # type: ignore
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/engine/arg_utils.py
vllm/engine/arg_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import copy import dataclasses import functools import json import sys from collections.abc import Callable from dataclasses import MISSING, dataclass, fields, is_dataclass from itertools import permutations from types import UnionType from typing import ( TYPE_CHECKING, Annotated, Any, Literal, TypeAlias, TypeVar, Union, cast, get_args, get_origin, ) import huggingface_hub import regex as re import torch from pydantic import TypeAdapter, ValidationError from pydantic.fields import FieldInfo from typing_extensions import TypeIs import vllm.envs as envs from vllm.attention.backends.registry import AttentionBackendEnum from vllm.config import ( AttentionConfig, CacheConfig, CompilationConfig, ConfigType, DeviceConfig, ECTransferConfig, EPLBConfig, KVEventsConfig, KVTransferConfig, LoadConfig, LoRAConfig, ModelConfig, MultiModalConfig, ObservabilityConfig, ParallelConfig, PoolerConfig, ProfilerConfig, SchedulerConfig, SpeculativeConfig, StructuredOutputsConfig, VllmConfig, get_attr_docs, ) from vllm.config.cache import ( BlockSize, CacheDType, KVOffloadingBackend, MambaDType, PrefixCachingHashAlgo, ) from vllm.config.device import Device from vllm.config.model import ( ConvertOption, HfOverrides, LogprobsMode, ModelDType, RunnerOption, TokenizerMode, ) from vllm.config.multimodal import MMCacheType, MMEncoderTPMode from vllm.config.observability import DetailedTraceModules from vllm.config.parallel import DistributedExecutorBackend, ExpertPlacementStrategy from vllm.config.scheduler import SchedulerPolicy from vllm.config.utils import get_field from vllm.config.vllm import OptimizationLevel from vllm.logger import init_logger, suppress_logging from vllm.platforms import CpuArchEnum, current_platform from vllm.plugins import load_general_plugins from vllm.ray.lazy_utils import is_in_ray_actor, is_ray_initialized from vllm.transformers_utils.config import ( is_interleaved, maybe_override_with_speculators, ) from vllm.transformers_utils.gguf_utils import is_gguf from vllm.transformers_utils.repo_utils import get_model_path from vllm.transformers_utils.utils import is_cloud_storage from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.mem_constants import GiB_bytes from vllm.utils.network_utils import get_ip from vllm.utils.torch_utils import resolve_kv_cache_dtype_string from vllm.v1.sample.logits_processor import LogitsProcessor if TYPE_CHECKING: from vllm.model_executor.layers.quantization import QuantizationMethods from vllm.model_executor.model_loader import LoadFormats from vllm.usage.usage_lib import UsageContext from vllm.v1.executor import Executor else: Executor = Any QuantizationMethods = Any LoadFormats = Any UsageContext = Any logger = init_logger(__name__) # object is used to allow for special typing forms T = TypeVar("T") TypeHint: TypeAlias = type[Any] | object TypeHintT: TypeAlias = type[T] | object def parse_type(return_type: Callable[[str], T]) -> Callable[[str], T]: def _parse_type(val: str) -> T: try: return return_type(val) except ValueError as e: raise argparse.ArgumentTypeError( f"Value {val} cannot be converted to {return_type}." ) from e return _parse_type def optional_type(return_type: Callable[[str], T]) -> Callable[[str], T | None]: def _optional_type(val: str) -> T | None: if val == "" or val == "None": return None return parse_type(return_type)(val) return _optional_type def union_dict_and_str(val: str) -> str | dict[str, str] | None: if not re.match(r"(?s)^\s*{.*}\s*$", val): return str(val) return optional_type(json.loads)(val) def is_type(type_hint: TypeHint, type: TypeHintT) -> TypeIs[TypeHintT]: """Check if the type hint is a specific type.""" return type_hint is type or get_origin(type_hint) is type def contains_type(type_hints: set[TypeHint], type: TypeHintT) -> bool: """Check if the type hints contain a specific type.""" return any(is_type(type_hint, type) for type_hint in type_hints) def get_type(type_hints: set[TypeHint], type: TypeHintT) -> TypeHintT: """Get the specific type from the type hints.""" return next((th for th in type_hints if is_type(th, type)), None) def literal_to_kwargs(type_hints: set[TypeHint]) -> dict[str, Any]: """Get the `type` and `choices` from a `Literal` type hint in `type_hints`. If `type_hints` also contains `str`, we use `metavar` instead of `choices`. """ type_hint = get_type(type_hints, Literal) options = get_args(type_hint) option_type = type(options[0]) if not all(isinstance(option, option_type) for option in options): raise ValueError( "All options must be of the same type. " f"Got {options} with types {[type(c) for c in options]}" ) kwarg = "metavar" if contains_type(type_hints, str) else "choices" return {"type": option_type, kwarg: sorted(options)} def collection_to_kwargs(type_hints: set[TypeHint], type: TypeHint) -> dict[str, Any]: type_hint = get_type(type_hints, type) types = get_args(type_hint) elem_type = types[0] # Handle Ellipsis assert all(t is elem_type for t in types if t is not Ellipsis), ( f"All non-Ellipsis elements must be of the same type. Got {types}." ) # Handle Union types if get_origin(elem_type) in {Union, UnionType}: # Union for Union[X, Y] and UnionType for X | Y assert str in get_args(elem_type), ( "If element can have multiple types, one must be 'str' " f"(i.e. 'list[int | str]'). Got {elem_type}." ) elem_type = str return { "type": elem_type, "nargs": "+" if type is not tuple or Ellipsis in types else len(types), } def is_not_builtin(type_hint: TypeHint) -> bool: """Check if the class is not a built-in type.""" return type_hint.__module__ != "builtins" def get_type_hints(type_hint: TypeHint) -> set[TypeHint]: """Extract type hints from Annotated or Union type hints.""" type_hints: set[TypeHint] = set() origin = get_origin(type_hint) args = get_args(type_hint) if origin is Annotated: type_hints.update(get_type_hints(args[0])) elif origin in {Union, UnionType}: # Union for Union[X, Y] and UnionType for X | Y for arg in args: type_hints.update(get_type_hints(arg)) else: type_hints.add(type_hint) return type_hints def is_online_quantization(quantization: Any) -> bool: return quantization in ["inc"] NEEDS_HELP = ( any("--help" in arg for arg in sys.argv) # vllm SUBCOMMAND --help or (argv0 := sys.argv[0]).endswith("mkdocs") # mkdocs SUBCOMMAND or argv0.endswith("mkdocs/__main__.py") # python -m mkdocs SUBCOMMAND ) @functools.lru_cache(maxsize=30) def _compute_kwargs(cls: ConfigType) -> dict[str, dict[str, Any]]: # Save time only getting attr docs if we're generating help text cls_docs = get_attr_docs(cls) if NEEDS_HELP else {} kwargs = {} for field in fields(cls): # Get the set of possible types for the field type_hints: set[TypeHint] = get_type_hints(field.type) # If the field is a dataclass, we can use the model_validate_json generator = (th for th in type_hints if is_dataclass(th)) dataclass_cls = next(generator, None) # Get the default value of the field if field.default is not MISSING: default = field.default # Handle pydantic.Field defaults if isinstance(default, FieldInfo): if default.default_factory is None: default = default.default else: # VllmConfig's Fields have default_factory set to config classes. # These could emit logs on init, which would be confusing. with suppress_logging(): default = default.default_factory() elif field.default_factory is not MISSING: default = field.default_factory() # Get the help text for the field name = field.name help = cls_docs.get(name, "").strip() # Escape % for argparse help = help.replace("%", "%%") # Initialise the kwargs dictionary for the field kwargs[name] = {"default": default, "help": help} # Set other kwargs based on the type hints json_tip = ( "Should either be a valid JSON string or JSON keys passed individually." ) if dataclass_cls is not None: def parse_dataclass(val: str, cls=dataclass_cls) -> Any: try: return TypeAdapter(cls).validate_json(val) except ValidationError as e: raise argparse.ArgumentTypeError(repr(e)) from e kwargs[name]["type"] = parse_dataclass kwargs[name]["help"] += f"\n\n{json_tip}" elif contains_type(type_hints, bool): # Creates --no-<name> and --<name> flags kwargs[name]["action"] = argparse.BooleanOptionalAction elif contains_type(type_hints, Literal): kwargs[name].update(literal_to_kwargs(type_hints)) elif contains_type(type_hints, tuple): kwargs[name].update(collection_to_kwargs(type_hints, tuple)) elif contains_type(type_hints, list): kwargs[name].update(collection_to_kwargs(type_hints, list)) elif contains_type(type_hints, set): kwargs[name].update(collection_to_kwargs(type_hints, set)) elif contains_type(type_hints, int): if name == "max_model_len": kwargs[name]["type"] = human_readable_int_or_auto kwargs[name]["help"] += f"\n\n{human_readable_int_or_auto.__doc__}" elif name in ("max_num_batched_tokens", "kv_cache_memory_bytes"): kwargs[name]["type"] = human_readable_int kwargs[name]["help"] += f"\n\n{human_readable_int.__doc__}" else: kwargs[name]["type"] = int elif contains_type(type_hints, float): kwargs[name]["type"] = float elif contains_type(type_hints, dict) and ( contains_type(type_hints, str) or any(is_not_builtin(th) for th in type_hints) ): kwargs[name]["type"] = union_dict_and_str elif contains_type(type_hints, dict): kwargs[name]["type"] = parse_type(json.loads) kwargs[name]["help"] += f"\n\n{json_tip}" elif contains_type(type_hints, str) or any( is_not_builtin(th) for th in type_hints ): kwargs[name]["type"] = str else: raise ValueError(f"Unsupported type {type_hints} for argument {name}.") # If the type hint was a sequence of literals, use the helper function # to update the type and choices if get_origin(kwargs[name].get("type")) is Literal: kwargs[name].update(literal_to_kwargs({kwargs[name]["type"]})) # If None is in type_hints, make the argument optional. # But not if it's a bool, argparse will handle this better. if type(None) in type_hints and not contains_type(type_hints, bool): kwargs[name]["type"] = optional_type(kwargs[name]["type"]) if kwargs[name].get("choices"): kwargs[name]["choices"].append("None") return kwargs def get_kwargs(cls: ConfigType) -> dict[str, dict[str, Any]]: """Return argparse kwargs for the given Config dataclass. If `--help` or `mkdocs` are not present in the command line command, the attribute documentation will not be included in the help output. The heavy computation is cached via functools.lru_cache, and a deep copy is returned so callers can mutate the dictionary without affecting the cached version. """ return copy.deepcopy(_compute_kwargs(cls)) @dataclass class EngineArgs: """Arguments for vLLM engine.""" model: str = ModelConfig.model served_model_name: str | list[str] | None = ModelConfig.served_model_name tokenizer: str | None = ModelConfig.tokenizer hf_config_path: str | None = ModelConfig.hf_config_path runner: RunnerOption = ModelConfig.runner convert: ConvertOption = ModelConfig.convert skip_tokenizer_init: bool = ModelConfig.skip_tokenizer_init enable_prompt_embeds: bool = ModelConfig.enable_prompt_embeds tokenizer_mode: TokenizerMode | str = ModelConfig.tokenizer_mode trust_remote_code: bool = ModelConfig.trust_remote_code allowed_local_media_path: str = ModelConfig.allowed_local_media_path allowed_media_domains: list[str] | None = ModelConfig.allowed_media_domains download_dir: str | None = LoadConfig.download_dir safetensors_load_strategy: str = LoadConfig.safetensors_load_strategy load_format: str | LoadFormats = LoadConfig.load_format config_format: str = ModelConfig.config_format dtype: ModelDType = ModelConfig.dtype kv_cache_dtype: CacheDType = CacheConfig.cache_dtype seed: int = ModelConfig.seed max_model_len: int | None = ModelConfig.max_model_len cudagraph_capture_sizes: list[int] | None = ( CompilationConfig.cudagraph_capture_sizes ) max_cudagraph_capture_size: int | None = get_field( CompilationConfig, "max_cudagraph_capture_size" ) # Note: Specifying a custom executor backend by passing a class # is intended for expert use only. The API may change without # notice. distributed_executor_backend: ( str | DistributedExecutorBackend | type[Executor] | None ) = ParallelConfig.distributed_executor_backend # number of P/D disaggregation (or other disaggregation) workers pipeline_parallel_size: int = ParallelConfig.pipeline_parallel_size master_addr: str = ParallelConfig.master_addr master_port: int = ParallelConfig.master_port nnodes: int = ParallelConfig.nnodes node_rank: int = ParallelConfig.node_rank tensor_parallel_size: int = ParallelConfig.tensor_parallel_size prefill_context_parallel_size: int = ParallelConfig.prefill_context_parallel_size decode_context_parallel_size: int = ParallelConfig.decode_context_parallel_size dcp_kv_cache_interleave_size: int = ParallelConfig.dcp_kv_cache_interleave_size cp_kv_cache_interleave_size: int = ParallelConfig.cp_kv_cache_interleave_size data_parallel_size: int = ParallelConfig.data_parallel_size data_parallel_rank: int | None = None data_parallel_start_rank: int | None = None data_parallel_size_local: int | None = None data_parallel_address: str | None = None data_parallel_rpc_port: int | None = None data_parallel_hybrid_lb: bool = False data_parallel_external_lb: bool = False data_parallel_backend: str = ParallelConfig.data_parallel_backend enable_expert_parallel: bool = ParallelConfig.enable_expert_parallel all2all_backend: str = ParallelConfig.all2all_backend enable_dbo: bool = ParallelConfig.enable_dbo ubatch_size: int = ParallelConfig.ubatch_size dbo_decode_token_threshold: int = ParallelConfig.dbo_decode_token_threshold dbo_prefill_token_threshold: int = ParallelConfig.dbo_prefill_token_threshold disable_nccl_for_dp_synchronization: bool = ( ParallelConfig.disable_nccl_for_dp_synchronization ) eplb_config: EPLBConfig = get_field(ParallelConfig, "eplb_config") enable_eplb: bool = ParallelConfig.enable_eplb expert_placement_strategy: ExpertPlacementStrategy = ( ParallelConfig.expert_placement_strategy ) _api_process_count: int = ParallelConfig._api_process_count _api_process_rank: int = ParallelConfig._api_process_rank max_parallel_loading_workers: int | None = ( ParallelConfig.max_parallel_loading_workers ) block_size: BlockSize | None = CacheConfig.block_size enable_prefix_caching: bool | None = None prefix_caching_hash_algo: PrefixCachingHashAlgo = ( CacheConfig.prefix_caching_hash_algo ) disable_sliding_window: bool = ModelConfig.disable_sliding_window disable_cascade_attn: bool = ModelConfig.disable_cascade_attn swap_space: float = CacheConfig.swap_space cpu_offload_gb: float = CacheConfig.cpu_offload_gb gpu_memory_utilization: float = CacheConfig.gpu_memory_utilization kv_cache_memory_bytes: int | None = CacheConfig.kv_cache_memory_bytes max_num_batched_tokens: int | None = None max_num_partial_prefills: int = SchedulerConfig.max_num_partial_prefills max_long_partial_prefills: int = SchedulerConfig.max_long_partial_prefills long_prefill_token_threshold: int = SchedulerConfig.long_prefill_token_threshold max_num_seqs: int | None = None max_logprobs: int = ModelConfig.max_logprobs logprobs_mode: LogprobsMode = ModelConfig.logprobs_mode disable_log_stats: bool = False aggregate_engine_logging: bool = False revision: str | None = ModelConfig.revision code_revision: str | None = ModelConfig.code_revision hf_token: bool | str | None = ModelConfig.hf_token hf_overrides: HfOverrides = get_field(ModelConfig, "hf_overrides") tokenizer_revision: str | None = ModelConfig.tokenizer_revision quantization: QuantizationMethods | None = ModelConfig.quantization enforce_eager: bool = ModelConfig.enforce_eager disable_custom_all_reduce: bool = ParallelConfig.disable_custom_all_reduce limit_mm_per_prompt: dict[str, int | dict[str, int]] = get_field( MultiModalConfig, "limit_per_prompt" ) enable_mm_embeds: bool = MultiModalConfig.enable_mm_embeds interleave_mm_strings: bool = MultiModalConfig.interleave_mm_strings media_io_kwargs: dict[str, dict[str, Any]] = get_field( MultiModalConfig, "media_io_kwargs" ) mm_processor_kwargs: dict[str, Any] | None = MultiModalConfig.mm_processor_kwargs mm_processor_cache_gb: float = MultiModalConfig.mm_processor_cache_gb mm_processor_cache_type: MMCacheType | None = ( MultiModalConfig.mm_processor_cache_type ) mm_shm_cache_max_object_size_mb: int = ( MultiModalConfig.mm_shm_cache_max_object_size_mb ) mm_encoder_tp_mode: MMEncoderTPMode = MultiModalConfig.mm_encoder_tp_mode mm_encoder_attn_backend: AttentionBackendEnum | str | None = ( MultiModalConfig.mm_encoder_attn_backend ) io_processor_plugin: str | None = None skip_mm_profiling: bool = MultiModalConfig.skip_mm_profiling video_pruning_rate: float = MultiModalConfig.video_pruning_rate # LoRA fields enable_lora: bool = False max_loras: int = LoRAConfig.max_loras max_lora_rank: int = LoRAConfig.max_lora_rank default_mm_loras: dict[str, str] | None = LoRAConfig.default_mm_loras fully_sharded_loras: bool = LoRAConfig.fully_sharded_loras max_cpu_loras: int | None = LoRAConfig.max_cpu_loras lora_dtype: str | torch.dtype | None = LoRAConfig.lora_dtype enable_tower_connector_lora: bool = LoRAConfig.enable_tower_connector_lora ray_workers_use_nsight: bool = ParallelConfig.ray_workers_use_nsight num_gpu_blocks_override: int | None = CacheConfig.num_gpu_blocks_override model_loader_extra_config: dict = get_field(LoadConfig, "model_loader_extra_config") ignore_patterns: str | list[str] = get_field(LoadConfig, "ignore_patterns") enable_chunked_prefill: bool | None = None disable_chunked_mm_input: bool = SchedulerConfig.disable_chunked_mm_input disable_hybrid_kv_cache_manager: bool | None = ( SchedulerConfig.disable_hybrid_kv_cache_manager ) structured_outputs_config: StructuredOutputsConfig = get_field( VllmConfig, "structured_outputs_config" ) reasoning_parser: str = StructuredOutputsConfig.reasoning_parser reasoning_parser_plugin: str | None = None logits_processor_pattern: str | None = ModelConfig.logits_processor_pattern speculative_config: dict[str, Any] | None = None show_hidden_metrics_for_version: str | None = ( ObservabilityConfig.show_hidden_metrics_for_version ) otlp_traces_endpoint: str | None = ObservabilityConfig.otlp_traces_endpoint collect_detailed_traces: list[DetailedTraceModules] | None = ( ObservabilityConfig.collect_detailed_traces ) kv_cache_metrics: bool = ObservabilityConfig.kv_cache_metrics kv_cache_metrics_sample: float = get_field( ObservabilityConfig, "kv_cache_metrics_sample" ) cudagraph_metrics: bool = ObservabilityConfig.cudagraph_metrics enable_layerwise_nvtx_tracing: bool = ( ObservabilityConfig.enable_layerwise_nvtx_tracing ) enable_mfu_metrics: bool = ObservabilityConfig.enable_mfu_metrics enable_mm_processor_stats: bool = ObservabilityConfig.enable_mm_processor_stats scheduling_policy: SchedulerPolicy = SchedulerConfig.policy scheduler_cls: str | type[object] | None = SchedulerConfig.scheduler_cls pooler_config: PoolerConfig | None = ModelConfig.pooler_config compilation_config: CompilationConfig = get_field(VllmConfig, "compilation_config") attention_config: AttentionConfig = get_field(VllmConfig, "attention_config") worker_cls: str = ParallelConfig.worker_cls worker_extension_cls: str = ParallelConfig.worker_extension_cls profiler_config: ProfilerConfig = get_field(VllmConfig, "profiler_config") kv_transfer_config: KVTransferConfig | None = None kv_events_config: KVEventsConfig | None = None ec_transfer_config: ECTransferConfig | None = None generation_config: str = ModelConfig.generation_config enable_sleep_mode: bool = ModelConfig.enable_sleep_mode override_generation_config: dict[str, Any] = get_field( ModelConfig, "override_generation_config" ) model_impl: str = ModelConfig.model_impl override_attention_dtype: str = ModelConfig.override_attention_dtype attention_backend: AttentionBackendEnum | None = AttentionConfig.backend calculate_kv_scales: bool = CacheConfig.calculate_kv_scales mamba_cache_dtype: MambaDType = CacheConfig.mamba_cache_dtype mamba_ssm_cache_dtype: MambaDType = CacheConfig.mamba_ssm_cache_dtype mamba_block_size: int | None = get_field(CacheConfig, "mamba_block_size") additional_config: dict[str, Any] = get_field(VllmConfig, "additional_config") use_tqdm_on_load: bool = LoadConfig.use_tqdm_on_load pt_load_map_location: str = LoadConfig.pt_load_map_location logits_processors: list[str | type[LogitsProcessor]] | None = ( ModelConfig.logits_processors ) """Custom logitproc types""" async_scheduling: bool | None = SchedulerConfig.async_scheduling stream_interval: int = SchedulerConfig.stream_interval kv_sharing_fast_prefill: bool = CacheConfig.kv_sharing_fast_prefill optimization_level: OptimizationLevel = VllmConfig.optimization_level kv_offloading_size: float | None = CacheConfig.kv_offloading_size kv_offloading_backend: KVOffloadingBackend | None = ( CacheConfig.kv_offloading_backend ) tokens_only: bool = False def __post_init__(self): # support `EngineArgs(compilation_config={...})` # without having to manually construct a # CompilationConfig object if isinstance(self.compilation_config, dict): self.compilation_config = CompilationConfig(**self.compilation_config) if isinstance(self.attention_config, dict): self.attention_config = AttentionConfig(**self.attention_config) if isinstance(self.eplb_config, dict): self.eplb_config = EPLBConfig(**self.eplb_config) # Setup plugins from vllm.plugins import load_general_plugins load_general_plugins() # when use hf offline,replace model and tokenizer id to local model path if huggingface_hub.constants.HF_HUB_OFFLINE: model_id = self.model self.model = get_model_path(self.model, self.revision) if model_id is not self.model: logger.info( "HF_HUB_OFFLINE is True, replace model_id [%s] to model_path [%s]", model_id, self.model, ) if self.tokenizer is not None: tokenizer_id = self.tokenizer self.tokenizer = get_model_path(self.tokenizer, self.tokenizer_revision) if tokenizer_id is not self.tokenizer: logger.info( "HF_HUB_OFFLINE is True, replace tokenizer_id [%s] " "to tokenizer_path [%s]", tokenizer_id, self.tokenizer, ) @staticmethod def add_cli_args(parser: FlexibleArgumentParser) -> FlexibleArgumentParser: """Shared CLI arguments for vLLM engine.""" # Model arguments model_kwargs = get_kwargs(ModelConfig) model_group = parser.add_argument_group( title="ModelConfig", description=ModelConfig.__doc__, ) if not ("serve" in sys.argv[1:] and "--help" in sys.argv[1:]): model_group.add_argument("--model", **model_kwargs["model"]) model_group.add_argument("--runner", **model_kwargs["runner"]) model_group.add_argument("--convert", **model_kwargs["convert"]) model_group.add_argument("--tokenizer", **model_kwargs["tokenizer"]) model_group.add_argument("--tokenizer-mode", **model_kwargs["tokenizer_mode"]) model_group.add_argument( "--trust-remote-code", **model_kwargs["trust_remote_code"] ) model_group.add_argument("--dtype", **model_kwargs["dtype"]) model_group.add_argument("--seed", **model_kwargs["seed"]) model_group.add_argument("--hf-config-path", **model_kwargs["hf_config_path"]) model_group.add_argument( "--allowed-local-media-path", **model_kwargs["allowed_local_media_path"] ) model_group.add_argument( "--allowed-media-domains", **model_kwargs["allowed_media_domains"] ) model_group.add_argument("--revision", **model_kwargs["revision"]) model_group.add_argument("--code-revision", **model_kwargs["code_revision"]) model_group.add_argument( "--tokenizer-revision", **model_kwargs["tokenizer_revision"] ) model_group.add_argument("--max-model-len", **model_kwargs["max_model_len"]) model_group.add_argument("--quantization", "-q", **model_kwargs["quantization"]) model_group.add_argument("--enforce-eager", **model_kwargs["enforce_eager"]) model_group.add_argument("--max-logprobs", **model_kwargs["max_logprobs"]) model_group.add_argument("--logprobs-mode", **model_kwargs["logprobs_mode"]) model_group.add_argument( "--disable-sliding-window", **model_kwargs["disable_sliding_window"] ) model_group.add_argument( "--disable-cascade-attn", **model_kwargs["disable_cascade_attn"] ) model_group.add_argument( "--skip-tokenizer-init", **model_kwargs["skip_tokenizer_init"] ) model_group.add_argument( "--enable-prompt-embeds", **model_kwargs["enable_prompt_embeds"] ) model_group.add_argument( "--served-model-name", **model_kwargs["served_model_name"] ) model_group.add_argument("--config-format", **model_kwargs["config_format"]) # This one is a special case because it can bool # or str. TODO: Handle this in get_kwargs model_group.add_argument( "--hf-token", type=str, nargs="?", const=True, default=model_kwargs["hf_token"]["default"], help=model_kwargs["hf_token"]["help"], ) model_group.add_argument("--hf-overrides", **model_kwargs["hf_overrides"]) model_group.add_argument("--pooler-config", **model_kwargs["pooler_config"]) model_group.add_argument( "--logits-processor-pattern", **model_kwargs["logits_processor_pattern"] ) model_group.add_argument( "--generation-config", **model_kwargs["generation_config"] ) model_group.add_argument( "--override-generation-config", **model_kwargs["override_generation_config"] ) model_group.add_argument( "--enable-sleep-mode", **model_kwargs["enable_sleep_mode"] ) model_group.add_argument("--model-impl", **model_kwargs["model_impl"]) model_group.add_argument( "--override-attention-dtype", **model_kwargs["override_attention_dtype"] ) model_group.add_argument( "--logits-processors", **model_kwargs["logits_processors"] ) model_group.add_argument( "--io-processor-plugin", **model_kwargs["io_processor_plugin"] ) # Model loading arguments load_kwargs = get_kwargs(LoadConfig) load_group = parser.add_argument_group( title="LoadConfig", description=LoadConfig.__doc__, ) load_group.add_argument("--load-format", **load_kwargs["load_format"]) load_group.add_argument("--download-dir", **load_kwargs["download_dir"]) load_group.add_argument( "--safetensors-load-strategy", **load_kwargs["safetensors_load_strategy"] ) load_group.add_argument( "--model-loader-extra-config", **load_kwargs["model_loader_extra_config"] ) load_group.add_argument("--ignore-patterns", **load_kwargs["ignore_patterns"]) load_group.add_argument("--use-tqdm-on-load", **load_kwargs["use_tqdm_on_load"]) load_group.add_argument( "--pt-load-map-location", **load_kwargs["pt_load_map_location"] ) # Attention arguments attention_kwargs = get_kwargs(AttentionConfig) attention_group = parser.add_argument_group( title="AttentionConfig", description=AttentionConfig.__doc__, ) attention_group.add_argument( "--attention-backend", **attention_kwargs["backend"] ) # Structured outputs arguments structured_outputs_kwargs = get_kwargs(StructuredOutputsConfig) structured_outputs_group = parser.add_argument_group( title="StructuredOutputsConfig", description=StructuredOutputsConfig.__doc__, ) structured_outputs_group.add_argument( "--reasoning-parser", # Choices need to be validated after parsing to include plugins **structured_outputs_kwargs["reasoning_parser"], ) structured_outputs_group.add_argument( "--reasoning-parser-plugin", **structured_outputs_kwargs["reasoning_parser_plugin"], ) # Parallel arguments parallel_kwargs = get_kwargs(ParallelConfig) parallel_group = parser.add_argument_group( title="ParallelConfig", description=ParallelConfig.__doc__, ) parallel_group.add_argument( "--distributed-executor-backend", **parallel_kwargs["distributed_executor_backend"], ) parallel_group.add_argument( "--pipeline-parallel-size", "-pp", **parallel_kwargs["pipeline_parallel_size"], ) parallel_group.add_argument("--master-addr", **parallel_kwargs["master_addr"]) parallel_group.add_argument("--master-port", **parallel_kwargs["master_port"]) parallel_group.add_argument("--nnodes", "-n", **parallel_kwargs["nnodes"]) parallel_group.add_argument("--node-rank", "-r", **parallel_kwargs["node_rank"]) parallel_group.add_argument( "--tensor-parallel-size", "-tp", **parallel_kwargs["tensor_parallel_size"] ) parallel_group.add_argument( "--decode-context-parallel-size", "-dcp", **parallel_kwargs["decode_context_parallel_size"], ) parallel_group.add_argument( "--dcp-kv-cache-interleave-size", **parallel_kwargs["dcp_kv_cache_interleave_size"], ) parallel_group.add_argument(
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/engine/__init__.py
vllm/engine/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false