sample_id
stringlengths
21
196
text
stringlengths
105
936k
metadata
dict
category
stringclasses
6 values
vllm-project/vllm:tests/compile/silly_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Shared PyTorch custom silly attention for compilation tests. Centralizes custom operation definitions to avoid duplicate registrations. """ import torch from torch.library import Library from vllm.utils.torch_utils import direct_register_custom_op # Shared library for all compilation test operations # Using "silly" namespace to match existing test expectations # import this file will automatically register # torch ops for testing (like silly.attention) silly_lib = Library("silly", "FRAGMENT") # Global counter that counts the number of times attention is invoked _global_counter = 0 def get_global_counter(): """Get the current global counter value""" return _global_counter def reset_global_counter(): """Reset the global counter to 0""" global _global_counter _global_counter = 0 def silly_attention( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, out: torch.Tensor ) -> None: """ Unified attention implementation that depends on all inputs and affects the output. Always increments a global counter that tests can use or ignore. """ global _global_counter # Always increment the global counter _global_counter += 1 # Unified implementation that depends on all inputs out.copy_(q + k + v) def silly_attention_fake( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, out: torch.Tensor ) -> None: """Fake implementation for testing""" return # Register the unified attention operation direct_register_custom_op( op_name="attention", op_func=silly_attention, mutates_args=["out"], fake_impl=silly_attention_fake, target_lib=silly_lib, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/compile/silly_attention.py", "license": "Apache License 2.0", "lines": 49, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/layers/fla/ops/chunk.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import warnings import torch from .chunk_delta_h import chunk_gated_delta_rule_fwd_h from .chunk_o import chunk_fwd_o from .chunk_scaled_dot_kkt import chunk_scaled_dot_kkt_fwd from .cumsum import chunk_local_cumsum from .l2norm import l2norm_fwd from .solve_tril import solve_tril from .utils import SUPPRESS_LEVEL, input_guard from .wy_fast import recompute_w_u_fwd def chunk_gated_delta_rule_fwd( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float, initial_state: torch.Tensor, output_final_state: bool, cu_seqlens: torch.LongTensor | None = None, ): g = chunk_local_cumsum(g, chunk_size=64, cu_seqlens=cu_seqlens) # obtain WY representation. u is actually the new v. A = chunk_scaled_dot_kkt_fwd( k=k, beta=beta, g=g, cu_seqlens=cu_seqlens, output_dtype=torch.float32 ) A = solve_tril(A=A, cu_seqlens=cu_seqlens, output_dtype=k.dtype) w, u = recompute_w_u_fwd( k=k, v=v, beta=beta, A=A, g_cumsum=g, cu_seqlens=cu_seqlens, ) h, v_new, final_state = chunk_gated_delta_rule_fwd_h( k=k, w=w, u=u, g=g, initial_state=initial_state, output_final_state=output_final_state, cu_seqlens=cu_seqlens, ) o = chunk_fwd_o( q=q, k=k, v=v_new, h=h, g=g, scale=scale, cu_seqlens=cu_seqlens, ) if SUPPRESS_LEVEL < 3: return g, o, A, final_state, None, None, None elif SUPPRESS_LEVEL >= 3: return g, o, A, final_state, w, h, v_new class ChunkGatedDeltaRuleFunction(torch.autograd.Function): @staticmethod @input_guard @torch.amp.custom_fwd(device_type="cuda") def forward( ctx, q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float, initial_state: torch.Tensor, output_final_state: bool, cu_seqlens: torch.LongTensor | None = None, use_qk_l2norm_in_kernel: bool = False, ): if use_qk_l2norm_in_kernel: q = l2norm_fwd(q) k = l2norm_fwd(k) g, o, A, final_state, w, h, v_new = chunk_gated_delta_rule_fwd( q=q, k=k, v=v, g=g, beta=beta, scale=scale, initial_state=initial_state, output_final_state=output_final_state, cu_seqlens=cu_seqlens, ) ctx.scale = scale ctx.use_qk_l2norm_in_kernel = use_qk_l2norm_in_kernel return o.to(q.dtype), final_state @torch.compiler.disable def chunk_gated_delta_rule( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float = None, initial_state: torch.Tensor = None, output_final_state: bool = False, cu_seqlens: torch.LongTensor | None = None, use_qk_l2norm_in_kernel: bool = False, ): r""" Args: q (torch.Tensor): Queries of shape `[B, T, H, K]`. k (torch.Tensor): Keys of shape `[B, T, H, K]`. v (torch.Tensor): Values of shape `[B, T, H, V]`. g (torch.Tensor): (forget) Gating tensor (in log space!) of shape `[B, T, H]`. beta (torch.Tensor): Betas of shape `[B, T, H]`. scale (Optional[int]): Scale factor for the RetNet attention scores. If not provided, it will default to `1 / sqrt(K)`. Default: `None`. initial_state (Optional[torch.Tensor]): Initial state of shape `[N, H, V, K]` for `N` input sequences. For equal-length input sequences, `N` equals the batch size `B`. Default: `None`. output_final_state (Optional[bool]): Whether to output the final state of shape `[N, H, V, K]`. Default: `False`. cu_seqlens (torch.LongTensor): Cumulative sequence lengths of shape `[N+1]` used for variable-length training, consistent with the FlashAttention API. Returns: o (torch.Tensor): Outputs of shape `[B, T, H, V]`. final_state (torch.Tensor): Final state of shape `[N, H, V, K]` if `output_final_state=True` else `None`. Examples:: >>> import torch >>> import torch.nn.functional as F >>> from einops import rearrange >>> from fla.ops.gated_delta_rule import chunk_gated_delta_rule # inputs with equal lengths >>> B, T, H, K, V = 4, 2048, 4, 512, 512 >>> q = torch.randn(B, T, H, K, dtype=torch.bfloat16, device='cuda') >>> k = F.normalize(torch.randn(B, T, H, K, dtype=torch.bfloat16, device='cuda'), p=2, dim=-1) >>> v = torch.randn(B, T, H, V, dtype=torch.bfloat16, device='cuda') >>> beta = torch.rand(B, T, H, dtype=torch.bfloat16, device='cuda').sigmoid() >>> g = F.logsigmoid(torch.rand(B, T, H, dtype=torch.bfloat16, device='cuda')) >>> h0 = torch.randn(B, H, V, K, dtype=torch.bfloat16, device='cuda') >>> o, ht = chunk_gated_delta_rule( q, k, v, g, beta, initial_state=h0, output_final_state=True ) # for variable-length inputs, the batch size `B` is expected to be 1 and `cu_seqlens` is required >>> q, k, v, beta, g = map(lambda x: rearrange(x, 'b t ... -> 1 (b t) ...'), (q, k, v, beta, g)) # for a batch with 4 sequences, `cu_seqlens` with 5 start/end positions are expected >>> cu_seqlens = q.new_tensor([0, 2048, 4096, 6144, 8192], dtype=torch.long) >>> o_var, ht_var = chunk_gated_delta_rule( q, k, v, g, beta, initial_state=h0, output_final_state=True, cu_seqlens=cu_seqlens ) """ assert q.dtype == k.dtype == v.dtype assert q.dtype != torch.float32, ( "ChunkGatedDeltaRuleFunction does not support float32. Please use bfloat16." ) assert len(beta.shape) == 3, "beta must be of shape [B, T, H]." if q.shape[1] < q.shape[2]: warnings.warn( f"Input tensor shape suggests potential format mismatch: seq_len ({q.shape[1]}) < num_heads ({q.shape[2]}). " "This may indicate the inputs were passed in head-first format [B, H, T, ...] " "Please verify your input tensor format matches the expected shape [B, T, H, ...].", stacklevel=2, ) if cu_seqlens is not None: if q.shape[0] != 1: raise ValueError( f"The batch size is expected to be 1 rather than {q.shape[0]} when using `cu_seqlens`." f"Please flatten variable-length inputs before processing." ) if initial_state is not None and initial_state.shape[0] != len(cu_seqlens) - 1: raise ValueError( f"The number of initial states is expected to be equal to the number of input sequences, " f"i.e., {len(cu_seqlens) - 1} rather than {initial_state.shape[0]}." ) if scale is None: scale = k.shape[-1] ** -0.5 o, final_state = ChunkGatedDeltaRuleFunction.apply( q, k, v, g, beta, scale, initial_state, output_final_state, cu_seqlens, use_qk_l2norm_in_kernel, ) return o, final_state
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/chunk.py", "license": "Apache License 2.0", "lines": 209, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/chunk_delta_h.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices, prepare_chunk_offsets from .op import exp from .utils import use_cuda_graph NUM_WARPS = [2, 4, 8, 16] @triton.heuristics( { "USE_G": lambda args: args["g"] is not None, "USE_GK": lambda args: args["gk"] is not None, "USE_INITIAL_STATE": lambda args: args["h0"] is not None, "STORE_FINAL_STATE": lambda args: args["ht"] is not None, "SAVE_NEW_VALUE": lambda args: args["v_new"] is not None, "IS_VARLEN": lambda args: args["cu_seqlens"] is not None, } ) @triton.autotune( configs=[ triton.Config({"BV": BV}, num_warps=num_warps, num_stages=num_stages) for num_warps in [2, 4] for num_stages in [2, 3, 4] for BV in [32, 64] ], key=["H", "K", "V", "BT"], use_cuda_graph=use_cuda_graph, ) @triton.jit(do_not_specialize=["T"]) def chunk_gated_delta_rule_fwd_kernel_h_blockdim64( k, v, w, v_new, g, gk, h, h0, ht, cu_seqlens, chunk_offsets, T, H: tl.constexpr, Hg: tl.constexpr, K: tl.constexpr, V: tl.constexpr, BT: tl.constexpr, BV: tl.constexpr, USE_G: tl.constexpr, USE_GK: tl.constexpr, USE_INITIAL_STATE: tl.constexpr, STORE_FINAL_STATE: tl.constexpr, SAVE_NEW_VALUE: tl.constexpr, IS_VARLEN: tl.constexpr, ): i_v, i_nh = tl.program_id(0), tl.program_id(1) i_n, i_h = i_nh // H, i_nh % H if IS_VARLEN: bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos NT = tl.cdiv(T, BT) boh = tl.load(chunk_offsets + i_n).to(tl.int32) else: bos, eos = i_n * T, i_n * T + T NT = tl.cdiv(T, BT) boh = i_n * NT # [BV, BK] b_h1 = tl.zeros([BV, 64], dtype=tl.float32) if K > 64: b_h2 = tl.zeros([BV, 64], dtype=tl.float32) if K > 128: b_h3 = tl.zeros([BV, 64], dtype=tl.float32) if K > 192: b_h4 = tl.zeros([BV, 64], dtype=tl.float32) # calculate offset h += ((boh * H + i_h) * V * K).to(tl.int64) v += ((bos * H + i_h) * V).to(tl.int64) k += ((bos * Hg + i_h // (H // Hg)) * K).to(tl.int64) w += ((bos * H + i_h) * K).to(tl.int64) if SAVE_NEW_VALUE: v_new += ((bos * H + i_h) * V).to(tl.int64) stride_v = H * V stride_h = H * V * K stride_k = Hg * K stride_w = H * K if USE_INITIAL_STATE: h0 = h0 + i_nh * V * K if STORE_FINAL_STATE: ht = ht + i_nh * V * K # load initial state if USE_INITIAL_STATE: p_h0_1 = tl.make_block_ptr(h0, (V, K), (K, 1), (i_v * BV, 0), (BV, 64), (1, 0)) b_h1 += tl.load(p_h0_1, boundary_check=(0, 1)).to(tl.float32) if K > 64: p_h0_2 = tl.make_block_ptr( h0, (V, K), (K, 1), (i_v * BV, 64), (BV, 64), (1, 0) ) b_h2 += tl.load(p_h0_2, boundary_check=(0, 1)).to(tl.float32) if K > 128: p_h0_3 = tl.make_block_ptr( h0, (V, K), (K, 1), (i_v * BV, 128), (BV, 64), (1, 0) ) b_h3 += tl.load(p_h0_3, boundary_check=(0, 1)).to(tl.float32) if K > 192: p_h0_4 = tl.make_block_ptr( h0, (V, K), (K, 1), (i_v * BV, 192), (BV, 64), (1, 0) ) b_h4 += tl.load(p_h0_4, boundary_check=(0, 1)).to(tl.float32) # main recurrence for i_t in range(NT): p_h1 = tl.make_block_ptr( h + i_t * stride_h, (V, K), (K, 1), (i_v * BV, 0), (BV, 64), (1, 0) ) tl.store(p_h1, b_h1.to(p_h1.dtype.element_ty), boundary_check=(0, 1)) if K > 64: p_h2 = tl.make_block_ptr( h + i_t * stride_h, (V, K), (K, 1), (i_v * BV, 64), (BV, 64), (1, 0) ) tl.store(p_h2, b_h2.to(p_h2.dtype.element_ty), boundary_check=(0, 1)) if K > 128: p_h3 = tl.make_block_ptr( h + i_t * stride_h, (V, K), (K, 1), (i_v * BV, 128), (BV, 64), (1, 0) ) tl.store(p_h3, b_h3.to(p_h3.dtype.element_ty), boundary_check=(0, 1)) if K > 192: p_h4 = tl.make_block_ptr( h + i_t * stride_h, (V, K), (K, 1), (i_v * BV, 192), (BV, 64), (1, 0) ) tl.store(p_h4, b_h4.to(p_h4.dtype.element_ty), boundary_check=(0, 1)) p_w = tl.make_block_ptr( w, (T, K), (stride_w, 1), (i_t * BT, 0), (BT, 64), (1, 0) ) b_w = tl.load(p_w, boundary_check=(0, 1)) b_v = tl.dot(b_w, tl.trans(b_h1).to(b_w.dtype)) if K > 64: p_w = tl.make_block_ptr( w, (T, K), (stride_w, 1), (i_t * BT, 64), (BT, 64), (1, 0) ) b_w = tl.load(p_w, boundary_check=(0, 1)) b_v += tl.dot(b_w, tl.trans(b_h2).to(b_w.dtype)) if K > 128: p_w = tl.make_block_ptr( w, (T, K), (stride_w, 1), (i_t * BT, 128), (BT, 64), (1, 0) ) b_w = tl.load(p_w, boundary_check=(0, 1)) b_v += tl.dot(b_w, tl.trans(b_h3).to(b_w.dtype)) if K > 192: p_w = tl.make_block_ptr( w, (T, K), (stride_w, 1), (i_t * BT, 192), (BT, 64), (1, 0) ) b_w = tl.load(p_w, boundary_check=(0, 1)) b_v += tl.dot(b_w, tl.trans(b_h4).to(b_w.dtype)) p_v = tl.make_block_ptr( v, (T, V), (stride_v, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0) ) b_v = tl.load(p_v, boundary_check=(0, 1)) - b_v if SAVE_NEW_VALUE: p_v = tl.make_block_ptr( v_new, (T, V), (stride_v, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0) ) tl.store(p_v, b_v.to(p_v.dtype.element_ty), boundary_check=(0, 1)) last_idx = min((i_t + 1) * BT, T) - 1 if USE_G: m_t = (i_t * BT + tl.arange(0, BT)) < T b_g_last = tl.load(g + bos * H + last_idx * H + i_h) p_g = tl.make_block_ptr( g + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,) ) b_g = tl.load(p_g, boundary_check=(0,)) b_v = b_v * tl.where(m_t, exp(b_g_last - b_g), 0)[:, None] b_g_last = exp(b_g_last) b_h1 *= b_g_last if K > 64: b_h2 *= b_g_last if K > 128: b_h3 *= b_g_last if K > 192: b_h4 *= b_g_last if USE_GK: o_k1 = tl.arange(0, 64) b_gk_last1 = tl.load( gk + (bos + last_idx) * H * K + i_h * K + o_k1, mask=(o_k1 < K), other=0.0, ) b_h1 *= exp(b_gk_last1)[None, :] if K > 64: o_k2 = 64 + o_k1 b_gk_last2 = tl.load( gk + (bos + last_idx) * H * K + i_h * K + o_k2, mask=(o_k2 < K), other=0.0, ) b_h2 *= exp(b_gk_last2)[None, :] if K > 128: o_k3 = 128 + o_k1 b_gk_last3 = tl.load( gk + (bos + last_idx) * H * K + i_h * K + o_k3, mask=(o_k3 < K), other=0.0, ) b_h3 *= exp(b_gk_last3)[None, :] if K > 192: o_k4 = 192 + o_k1 b_gk_last4 = tl.load( gk + (bos + last_idx) * H * K + i_h * K + o_k4, mask=(o_k4 < K), other=0.0, ) b_h4 *= exp(b_gk_last4)[None, :] b_v = b_v.to(k.dtype.element_ty) p_k = tl.make_block_ptr( k, (K, T), (1, stride_k), (0, i_t * BT), (64, BT), (0, 1) ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_h1 += tl.trans(tl.dot(b_k, b_v)) if K > 64: p_k = tl.make_block_ptr( k, (K, T), (1, stride_k), (64, i_t * BT), (64, BT), (0, 1) ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_h2 += tl.trans(tl.dot(b_k, b_v)) if K > 128: p_k = tl.make_block_ptr( k, (K, T), (1, stride_k), (128, i_t * BT), (64, BT), (0, 1) ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_h3 += tl.trans(tl.dot(b_k, b_v)) if K > 192: p_k = tl.make_block_ptr( k, (K, T), (1, stride_k), (192, i_t * BT), (64, BT), (0, 1) ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_h4 += tl.trans(tl.dot(b_k, b_v)) # epilogue if STORE_FINAL_STATE: p_ht = tl.make_block_ptr(ht, (V, K), (K, 1), (i_v * BV, 0), (BV, 64), (1, 0)) tl.store(p_ht, b_h1.to(p_ht.dtype.element_ty), boundary_check=(0, 1)) if K > 64: p_ht = tl.make_block_ptr( ht, (V, K), (K, 1), (i_v * BV, 64), (BV, 64), (1, 0) ) tl.store(p_ht, b_h2.to(p_ht.dtype.element_ty), boundary_check=(0, 1)) if K > 128: p_ht = tl.make_block_ptr( ht, (V, K), (K, 1), (i_v * BV, 128), (BV, 64), (1, 0) ) tl.store(p_ht, b_h3.to(p_ht.dtype.element_ty), boundary_check=(0, 1)) if K > 192: p_ht = tl.make_block_ptr( ht, (V, K), (K, 1), (i_v * BV, 192), (BV, 64), (1, 0) ) tl.store(p_ht, b_h4.to(p_ht.dtype.element_ty), boundary_check=(0, 1)) def chunk_gated_delta_rule_fwd_h( k: torch.Tensor, w: torch.Tensor, u: torch.Tensor, g: torch.Tensor | None = None, gk: torch.Tensor | None = None, initial_state: torch.Tensor | None = None, output_final_state: bool = False, chunk_size: int = 64, # SY: remove this argument and force chunk size 64? save_new_value: bool = True, cu_seqlens: torch.LongTensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: # This kernel is slightly different from fla to support Q/K with different head numbers. # In fla, Q/K always have the same head number, so Hg is always equal to H. B, T, Hg, K, V = *k.shape, u.shape[-1] H = u.shape[-2] BT = chunk_size chunk_indices = ( prepare_chunk_indices(cu_seqlens, chunk_size) if cu_seqlens is not None else None ) # N: the actual number of sequences in the batch with either equal or variable lengths if cu_seqlens is None: N, NT, chunk_offsets = B, triton.cdiv(T, BT), None else: N, NT, chunk_offsets = ( len(cu_seqlens) - 1, len(chunk_indices), prepare_chunk_offsets(cu_seqlens, BT), ) assert K <= 256, "current kernel does not support head dimension larger than 256." h = k.new_empty(B, NT, H, V, K) final_state = ( k.new_empty(N, H, V, K, dtype=torch.float32) if output_final_state else None ) v_new = torch.empty_like(u) if save_new_value else None def grid(meta): return (triton.cdiv(V, meta["BV"]), N * H) chunk_gated_delta_rule_fwd_kernel_h_blockdim64[grid]( k=k, v=u, w=w, v_new=v_new, g=g, gk=gk, h=h, h0=initial_state, ht=final_state, cu_seqlens=cu_seqlens, chunk_offsets=chunk_offsets, T=T, H=H, Hg=Hg, K=K, V=V, BT=BT, ) return h, v_new, final_state
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/chunk_delta_h.py", "license": "Apache License 2.0", "lines": 322, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/chunk_o.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices from .op import exp from .utils import FLA_GDN_FIX_BT, check_shared_mem, is_nvidia_hopper BKV_LIST = [64, 128] if check_shared_mem() else [32, 64] NUM_WARPS = [2, 4] if is_nvidia_hopper else [2, 4, 8] @triton.heuristics( { "USE_G": lambda args: args["g"] is not None, "IS_VARLEN": lambda args: args["cu_seqlens"] is not None, } ) @triton.autotune( configs=[ triton.Config({"BK": BK, "BV": BV}, num_warps=num_warps, num_stages=num_stages) for BK in BKV_LIST for BV in BKV_LIST for num_warps in NUM_WARPS for num_stages in [2, 3, 4] ], key=["H", "K", "V", "BT"], ) @triton.jit(do_not_specialize=["T"]) def chunk_fwd_kernel_o( q, k, v, h, g, o, cu_seqlens, chunk_indices, scale, T, H: tl.constexpr, Hg: tl.constexpr, K: tl.constexpr, V: tl.constexpr, BT: tl.constexpr, BK: tl.constexpr, BV: tl.constexpr, USE_G: tl.constexpr, IS_VARLEN: tl.constexpr, ): i_v, i_t, i_bh = tl.program_id(0), tl.program_id(1), tl.program_id(2) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_tg = i_t i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos NT = tl.cdiv(T, BT) else: NT = tl.cdiv(T, BT) i_tg = i_b * NT + i_t bos, eos = i_b * T, i_b * T + T # offset calculation q += (bos * Hg + i_h // (H // Hg)) * K k += (bos * Hg + i_h // (H // Hg)) * K v += (bos * H + i_h) * V o += (bos * H + i_h) * V h += (i_tg * H + i_h).to(tl.int64) * V * K b_o = tl.zeros([BT, BV], dtype=tl.float32) b_A = tl.zeros([BT, BT], dtype=tl.float32) for i_k in range(tl.cdiv(K, BK)): p_q = tl.make_block_ptr( q, (T, K), (Hg * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0) ) p_k = tl.make_block_ptr( k, (K, T), (1, Hg * K), (i_k * BK, i_t * BT), (BK, BT), (0, 1) ) p_h = tl.make_block_ptr( h, (V, K), (K, 1), (i_v * BV, i_k * BK), (BV, BK), (1, 0) ) # [BT, BK] b_q = tl.load(p_q, boundary_check=(0, 1)) # [BK, BT] b_k = tl.load(p_k, boundary_check=(0, 1)) # [BV, BK] b_h = tl.load(p_h, boundary_check=(0, 1)) # [BT, BK] @ [BK, BV] -> [BT, BV] b_o += tl.dot(b_q, tl.trans(b_h)) # [BT, BK] @ [BK, BT] -> [BT, BT] b_A += tl.dot(b_q, b_k) if USE_G: g += bos * H + i_h p_g = tl.make_block_ptr(g, (T,), (H,), (i_t * BT,), (BT,), (0,)) b_g = tl.load(p_g, boundary_check=(0,)) b_o = b_o * exp(b_g)[:, None] b_A = b_A * exp(b_g[:, None] - b_g[None, :]) o_t = i_t * BT + tl.arange(0, BT) m_t = o_t < T m_A = (o_t[:, None] >= o_t[None, :]) & (m_t[:, None] & m_t) b_A = tl.where(m_A, b_A, 0) p_v = tl.make_block_ptr( v, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0) ) p_o = tl.make_block_ptr( o, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0) ) b_v = tl.load(p_v, boundary_check=(0, 1)) # to fix mma -> mma layout conversion # already solved by triton v3.2 or higher b_o = b_o * scale + tl.dot(b_A.to(b_v.dtype), b_v) * scale tl.store(p_o, b_o.to(p_o.dtype.element_ty), boundary_check=(0, 1)) def chunk_fwd_o( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, h: torch.Tensor, g: torch.Tensor | None = None, # cumsum of log decay scale: float | None = None, cu_seqlens: torch.LongTensor | None = None, chunk_size: int = 64, ) -> torch.Tensor: B, T, Hg, K, V = *q.shape, v.shape[-1] H = v.shape[-2] BT = 64 if FLA_GDN_FIX_BT else min(chunk_size, max(16, triton.next_power_of_2(T))) chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = triton.cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) if scale is None: scale = k.shape[-1] ** -0.5 o = torch.empty_like(v) def grid(meta): return (triton.cdiv(V, meta["BV"]), NT, B * H) chunk_fwd_kernel_o[grid]( q, k, v, h, g, o, cu_seqlens, chunk_indices, scale, T=T, H=H, Hg=Hg, K=K, V=V, BT=BT, ) return o
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/chunk_o.py", "license": "Apache License 2.0", "lines": 161, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/chunk_scaled_dot_kkt.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices from .op import exp @triton.heuristics( { "USE_G": lambda args: args["g"] is not None, "IS_VARLEN": lambda args: args["cu_seqlens"] is not None, } ) @triton.autotune( configs=[ triton.Config({"BK": BK}, num_warps=num_warps, num_stages=num_stages) for BK in [32, 64, 128] for num_warps in [2, 4, 8] for num_stages in [2, 3, 4] ], key=["H", "K", "BT", "IS_VARLEN"], ) @triton.jit(do_not_specialize=["T"]) def chunk_scaled_dot_kkt_fwd_kernel( k, beta, g, A, cu_seqlens, chunk_indices, T, H: tl.constexpr, Hg: tl.constexpr, K: tl.constexpr, BT: tl.constexpr, BK: tl.constexpr, IS_VARLEN: tl.constexpr, USE_G: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T o_t = i_t * BT + tl.arange(0, BT) m_t = o_t < T p_beta = tl.make_block_ptr( beta + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,) ) b_beta = tl.load(p_beta, boundary_check=(0,)) b_A = tl.zeros([BT, BT], dtype=tl.float32) for i_k in range(tl.cdiv(K, BK)): p_k = tl.make_block_ptr( k + (bos * Hg + i_h // (H // Hg)) * K, (T, K), (Hg * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_kb = b_k * b_beta[:, None] b_A += tl.dot(b_kb.to(b_k.dtype), tl.trans(b_k)) if USE_G: p_g = tl.make_block_ptr(g + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,)) b_g = tl.load(p_g, boundary_check=(0,)) b_g_diff = b_g[:, None] - b_g[None, :] b_A = b_A * exp(b_g_diff) m_A = (o_t[:, None] > o_t[None, :]) & (m_t[:, None] & m_t) b_A = tl.where(m_A, b_A, 0) p_A = tl.make_block_ptr( A + (bos * H + i_h) * BT, (T, BT), (BT * H, 1), (i_t * BT, 0), (BT, BT), (1, 0) ) tl.store(p_A, b_A.to(p_A.dtype.element_ty), boundary_check=(0, 1)) def chunk_scaled_dot_kkt_fwd( k: torch.Tensor, g: torch.Tensor | None = None, beta: torch.Tensor | None = None, cu_seqlens: torch.LongTensor | None = None, chunk_size: int = 64, output_dtype: torch.dtype = torch.float32, ) -> torch.Tensor: r""" Compute beta * K * K^T. Args: k (torch.Tensor): The key tensor of shape `[B, T, H, K]`. beta (torch.Tensor): The beta tensor of shape `[B, T, H]`. g (torch.Tensor): The cumulative sum of the gate tensor of shape `[B, T, H]`. Default: `None`. cu_seqlens (torch.LongTensor): The cumulative sequence lengths of the input tensor. Default: None chunk_size (int): The chunk size. Default: 64. output_dtype (torch.dtype): The dtype of the output tensor. Default: `torch.float32` Returns: beta * K * K^T of shape `[B, T, H, BT]` where `BT` is the chunk size. """ # This kernel is slightly different from fla to support Q/K with different head numbers. # In fla, Q/K always have the same head number, so Hg is always equal to H. B, T, Hg, K = k.shape H = beta.shape[-1] BT = chunk_size chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = triton.cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) A = torch.empty(B, T, H, BT, device=k.device, dtype=output_dtype) chunk_scaled_dot_kkt_fwd_kernel[(NT, B * H)]( k=k, g=g, beta=beta, A=A, cu_seqlens=cu_seqlens, chunk_indices=chunk_indices, T=T, H=H, Hg=Hg, K=K, BT=BT, ) return A
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/chunk_scaled_dot_kkt.py", "license": "Apache License 2.0", "lines": 140, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/cumsum.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import warnings import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices from .utils import check_shared_mem, input_guard BS_LIST = [32, 64] if check_shared_mem() else [16, 32] @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[triton.Config({}, num_warps=num_warps) for num_warps in [1, 2, 4, 8]], key=["B", "H", "BT", "IS_VARLEN", "REVERSE"], ) @triton.jit(do_not_specialize=["T"]) def chunk_local_cumsum_scalar_kernel( s, o, cu_seqlens, chunk_indices, T, B: tl.constexpr, H: tl.constexpr, BT: tl.constexpr, REVERSE: tl.constexpr, IS_VARLEN: tl.constexpr, HEAD_FIRST: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T if HEAD_FIRST: p_s = tl.make_block_ptr( s + bos * H + i_h * T, (T,), (1,), (i_t * BT,), (BT,), (0,) ) p_o = tl.make_block_ptr( o + bos * H + i_h * T, (T,), (1,), (i_t * BT,), (BT,), (0,) ) else: p_s = tl.make_block_ptr(s + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,)) p_o = tl.make_block_ptr(o + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,)) # [BT] b_s = tl.load(p_s, boundary_check=(0,)).to(tl.float32) b_o = tl.cumsum(b_s, axis=0) if REVERSE: b_z = tl.sum(b_s, axis=0) b_o = -b_o + b_z[None] + b_s tl.store(p_o, b_o.to(p_o.dtype.element_ty), boundary_check=(0,)) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({"BS": BS}, num_warps=num_warps) for BS in BS_LIST for num_warps in [2, 4, 8] ], key=["B", "H", "S", "BT", "IS_VARLEN", "REVERSE"], ) @triton.jit(do_not_specialize=["T"]) def chunk_local_cumsum_vector_kernel( s, o, cu_seqlens, chunk_indices, T, B: tl.constexpr, H: tl.constexpr, S: tl.constexpr, BT: tl.constexpr, BS: tl.constexpr, REVERSE: tl.constexpr, IS_VARLEN: tl.constexpr, HEAD_FIRST: tl.constexpr, ): i_s, i_t, i_bh = tl.program_id(0), tl.program_id(1), tl.program_id(2) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T o_i = tl.arange(0, BT) if REVERSE: m_s = tl.where(o_i[:, None] <= o_i[None, :], 1.0, 0.0) else: m_s = tl.where(o_i[:, None] >= o_i[None, :], 1.0, 0.0) if HEAD_FIRST: p_s = tl.make_block_ptr( s + (bos * H + i_h * T) * S, (T, S), (S, 1), (i_t * BT, i_s * BS), (BT, BS), (1, 0), ) p_o = tl.make_block_ptr( o + (bos * H + i_h * T) * S, (T, S), (S, 1), (i_t * BT, i_s * BS), (BT, BS), (1, 0), ) else: p_s = tl.make_block_ptr( s + (bos * H + i_h) * S, (T, S), (H * S, 1), (i_t * BT, i_s * BS), (BT, BS), (1, 0), ) p_o = tl.make_block_ptr( o + (bos * H + i_h) * S, (T, S), (H * S, 1), (i_t * BT, i_s * BS), (BT, BS), (1, 0), ) # [BT, BS] b_s = tl.load(p_s, boundary_check=(0, 1)).to(tl.float32) b_o = tl.dot(m_s, b_s, allow_tf32=False) tl.store(p_o, b_o.to(p_o.dtype.element_ty), boundary_check=(0, 1)) def chunk_local_cumsum_scalar( g: torch.Tensor, chunk_size: int, reverse: bool = False, cu_seqlens: torch.Tensor | None = None, head_first: bool = False, output_dtype: torch.dtype | None = torch.float, ) -> torch.Tensor: if head_first: B, H, T = g.shape else: B, T, H = g.shape assert chunk_size == 2 ** (chunk_size.bit_length() - 1), ( "chunk_size must be a power of 2" ) BT = chunk_size chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = triton.cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) g_org, g = g, torch.empty_like(g, dtype=output_dtype or g.dtype) grid = (NT, B * H) chunk_local_cumsum_scalar_kernel[grid]( g_org, g, cu_seqlens, chunk_indices, T=T, B=B, H=H, BT=BT, HEAD_FIRST=head_first, REVERSE=reverse, ) return g def chunk_local_cumsum_vector( g: torch.Tensor, chunk_size: int, reverse: bool = False, cu_seqlens: torch.Tensor | None = None, head_first: bool = False, output_dtype: torch.dtype | None = torch.float, ) -> torch.Tensor: if head_first: B, H, T, S = g.shape else: B, T, H, S = g.shape BT = chunk_size chunk_indices = ( prepare_chunk_indices(cu_seqlens, chunk_size) if cu_seqlens is not None else None ) NT = triton.cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) assert chunk_size == 2 ** (chunk_size.bit_length() - 1), ( "chunk_size must be a power of 2" ) g_org, g = g, torch.empty_like(g, dtype=output_dtype or g.dtype) def grid(meta): return (triton.cdiv(meta["S"], meta["BS"]), NT, B * H) # keep cumulative normalizer in fp32 # this kernel is equivalent to # g = g.view(B, H, NT, BT, -1).cumsum(-2).view(B, H, T, -1) chunk_local_cumsum_vector_kernel[grid]( g_org, g, cu_seqlens, chunk_indices, T=T, B=B, H=H, S=S, BT=BT, HEAD_FIRST=head_first, REVERSE=reverse, ) return g @input_guard def chunk_local_cumsum( g: torch.Tensor, chunk_size: int, reverse: bool = False, cu_seqlens: torch.Tensor | None = None, head_first: bool = False, output_dtype: torch.dtype | None = torch.float, **kwargs, ) -> torch.Tensor: if not head_first and g.shape[1] < g.shape[2]: warnings.warn( f"Input tensor shape suggests potential format mismatch: seq_len ({g.shape[1]}) < num_heads ({g.shape[2]}). " "This may indicate the inputs were passed in head-first format [B, H, T, ...] " "when head_first=False was specified. " "Please verify your input tensor format matches the expected shape [B, T, H, ...].", stacklevel=2, ) if cu_seqlens is not None: assert g.shape[0] == 1, ( "Only batch size 1 is supported when cu_seqlens are provided" ) if len(g.shape) == 3: return chunk_local_cumsum_scalar( g, chunk_size, reverse, cu_seqlens, head_first, output_dtype ) elif len(g.shape) == 4: return chunk_local_cumsum_vector( g, chunk_size, reverse, cu_seqlens, head_first, output_dtype ) else: raise ValueError( f"Unsupported input shape {g.shape}. " f"which should be (B, T, H, D) if `head_first=False` " f"or (B, H, T, D) otherwise" )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/cumsum.py", "license": "Apache License 2.0", "lines": 260, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/fused_recurrent.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .op import exp @triton.heuristics( { "USE_INITIAL_STATE": lambda args: args["h0"] is not None, "IS_VARLEN": lambda args: args["cu_seqlens"] is not None, "IS_CONTINUOUS_BATCHING": lambda args: args["ssm_state_indices"] is not None, "IS_SPEC_DECODING": lambda args: args["num_accepted_tokens"] is not None, } ) @triton.jit(do_not_specialize=["N", "T"]) def fused_recurrent_gated_delta_rule_fwd_kernel( q, k, v, g, beta, o, h0, ht, cu_seqlens, ssm_state_indices, num_accepted_tokens, scale, N: tl.int64, # num of sequences T: tl.int64, # num of tokens B: tl.constexpr, H: tl.constexpr, HV: tl.constexpr, K: tl.constexpr, V: tl.constexpr, BK: tl.constexpr, BV: tl.constexpr, stride_init_state_token: tl.constexpr, stride_final_state_token: tl.constexpr, stride_indices_seq: tl.constexpr, stride_indices_tok: tl.constexpr, USE_INITIAL_STATE: tl.constexpr, # whether to use initial state INPLACE_FINAL_STATE: tl.constexpr, # whether to store final state inplace IS_BETA_HEADWISE: tl.constexpr, # whether beta is headwise vector or scalar, USE_QK_L2NORM_IN_KERNEL: tl.constexpr, IS_VARLEN: tl.constexpr, IS_CONTINUOUS_BATCHING: tl.constexpr, IS_SPEC_DECODING: tl.constexpr, IS_KDA: tl.constexpr, ): i_k, i_v, i_nh = tl.program_id(0), tl.program_id(1), tl.program_id(2) i_n, i_hv = i_nh // HV, i_nh % HV i_h = i_hv // (HV // H) if IS_VARLEN: bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int64), tl.load(cu_seqlens + i_n + 1).to(tl.int64), ) all = T T = eos - bos else: bos, eos = i_n * T, i_n * T + T all = B * T if T == 0: # no tokens to process for this sequence return o_k = i_k * BK + tl.arange(0, BK) o_v = i_v * BV + tl.arange(0, BV) p_q = q + (bos * H + i_h) * K + o_k p_k = k + (bos * H + i_h) * K + o_k p_v = v + (bos * HV + i_hv) * V + o_v if IS_BETA_HEADWISE: p_beta = beta + (bos * HV + i_hv) * V + o_v else: p_beta = beta + bos * HV + i_hv if not IS_KDA: p_g = g + bos * HV + i_hv else: p_gk = g + (bos * HV + i_hv) * K + o_k p_o = o + ((i_k * all + bos) * HV + i_hv) * V + o_v mask_k = o_k < K mask_v = o_v < V mask_h = mask_v[:, None] & mask_k[None, :] b_h = tl.zeros([BV, BK], dtype=tl.float32) if USE_INITIAL_STATE: if IS_CONTINUOUS_BATCHING: if IS_SPEC_DECODING: i_t = tl.load(num_accepted_tokens + i_n).to(tl.int64) - 1 else: i_t = 0 # Load state index and check for PAD_SLOT_ID (-1) state_idx = tl.load(ssm_state_indices + i_n * stride_indices_seq + i_t).to( tl.int64 ) # Skip if state index is invalid (PAD_SLOT_ID = -1) if state_idx < 0: return p_h0 = h0 + state_idx * stride_init_state_token else: p_h0 = h0 + bos * HV * V * K p_h0 = p_h0 + i_hv * V * K + o_v[:, None] * K + o_k[None, :] b_h += tl.load(p_h0, mask=mask_h, other=0).to(tl.float32) for i_t in range(0, T): b_q = tl.load(p_q, mask=mask_k, other=0).to(tl.float32) b_k = tl.load(p_k, mask=mask_k, other=0).to(tl.float32) b_v = tl.load(p_v, mask=mask_v, other=0).to(tl.float32) if USE_QK_L2NORM_IN_KERNEL: b_q = b_q / tl.sqrt(tl.sum(b_q * b_q) + 1e-6) b_k = b_k / tl.sqrt(tl.sum(b_k * b_k) + 1e-6) b_q = b_q * scale # [BV, BK] if not IS_KDA: b_g = tl.load(p_g).to(tl.float32) b_h *= exp(b_g) else: b_gk = tl.load(p_gk).to(tl.float32) b_h *= exp(b_gk[None, :]) # [BV] b_v -= tl.sum(b_h * b_k[None, :], 1) if IS_BETA_HEADWISE: b_beta = tl.load(p_beta, mask=mask_v, other=0).to(tl.float32) else: b_beta = tl.load(p_beta).to(tl.float32) b_v *= b_beta # [BV, BK] b_h += b_v[:, None] * b_k[None, :] # [BV] b_o = tl.sum(b_h * b_q[None, :], 1) tl.store(p_o, b_o.to(p_o.dtype.element_ty), mask=mask_v) # keep the states for multi-query tokens if INPLACE_FINAL_STATE: # Load state index and check for PAD_SLOT_ID (-1) final_state_idx = tl.load( ssm_state_indices + i_n * stride_indices_seq + i_t ).to(tl.int64) # Only store if state index is valid (not PAD_SLOT_ID) if final_state_idx >= 0: p_ht = ht + final_state_idx * stride_final_state_token p_ht = p_ht + i_hv * V * K + o_v[:, None] * K + o_k[None, :] tl.store(p_ht, b_h.to(p_ht.dtype.element_ty), mask=mask_h) else: p_ht = ht + (bos + i_t) * stride_final_state_token p_ht = p_ht + i_hv * V * K + o_v[:, None] * K + o_k[None, :] tl.store(p_ht, b_h.to(p_ht.dtype.element_ty), mask=mask_h) p_q += H * K p_k += H * K p_o += HV * V p_v += HV * V if not IS_KDA: p_g += HV else: p_gk += HV * K p_beta += HV * (V if IS_BETA_HEADWISE else 1) def fused_recurrent_gated_delta_rule_fwd( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float, initial_state: torch.Tensor, inplace_final_state: bool = True, cu_seqlens: torch.LongTensor | None = None, ssm_state_indices: torch.Tensor | None = None, num_accepted_tokens: torch.Tensor | None = None, use_qk_l2norm_in_kernel: bool = False, ) -> tuple[torch.Tensor, torch.Tensor]: B, T, H, K, V = *k.shape, v.shape[-1] HV = v.shape[2] N = B if cu_seqlens is None else len(cu_seqlens) - 1 BK, BV = triton.next_power_of_2(K), min(triton.next_power_of_2(V), 32) NK, NV = triton.cdiv(K, BK), triton.cdiv(V, BV) assert NK == 1, "NK > 1 is not supported yet" num_stages = 3 num_warps = 1 o = q.new_empty(NK, *v.shape) if inplace_final_state: final_state = initial_state else: final_state = q.new_empty(T, HV, V, K, dtype=initial_state.dtype) stride_init_state_token = initial_state.stride(0) stride_final_state_token = final_state.stride(0) if ssm_state_indices is None: stride_indices_seq, stride_indices_tok = 1, 1 elif ssm_state_indices.ndim == 1: stride_indices_seq, stride_indices_tok = ssm_state_indices.stride(0), 1 else: stride_indices_seq, stride_indices_tok = ssm_state_indices.stride() grid = (NK, NV, N * HV) fused_recurrent_gated_delta_rule_fwd_kernel[grid]( q=q, k=k, v=v, g=g, beta=beta, o=o, h0=initial_state, ht=final_state, cu_seqlens=cu_seqlens, ssm_state_indices=ssm_state_indices, num_accepted_tokens=num_accepted_tokens, scale=scale, N=N, T=T, B=B, H=H, HV=HV, K=K, V=V, BK=BK, BV=BV, stride_init_state_token=stride_init_state_token, stride_final_state_token=stride_final_state_token, stride_indices_seq=stride_indices_seq, stride_indices_tok=stride_indices_tok, IS_BETA_HEADWISE=beta.ndim == v.ndim, USE_QK_L2NORM_IN_KERNEL=use_qk_l2norm_in_kernel, INPLACE_FINAL_STATE=inplace_final_state, IS_KDA=False, num_warps=num_warps, num_stages=num_stages, ) o = o.squeeze(0) return o, final_state class FusedRecurrentFunction(torch.autograd.Function): @staticmethod def forward( ctx, q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor, scale: float, initial_state: torch.Tensor, inplace_final_state: bool = True, cu_seqlens: torch.LongTensor | None = None, ssm_state_indices: torch.Tensor | None = None, num_accepted_tokens: torch.Tensor | None = None, use_qk_l2norm_in_kernel: bool = False, ): o, final_state = fused_recurrent_gated_delta_rule_fwd( q=q.contiguous(), k=k.contiguous(), v=v.contiguous(), g=g.contiguous(), beta=beta.contiguous(), scale=scale, initial_state=initial_state, inplace_final_state=inplace_final_state, cu_seqlens=cu_seqlens, ssm_state_indices=ssm_state_indices, num_accepted_tokens=num_accepted_tokens, use_qk_l2norm_in_kernel=use_qk_l2norm_in_kernel, ) return o, final_state def fused_recurrent_gated_delta_rule( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, g: torch.Tensor, beta: torch.Tensor = None, scale: float = None, initial_state: torch.Tensor = None, inplace_final_state: bool = True, cu_seqlens: torch.LongTensor | None = None, ssm_state_indices: torch.Tensor | None = None, num_accepted_tokens: torch.Tensor | None = None, use_qk_l2norm_in_kernel: bool = False, ) -> tuple[torch.Tensor, torch.Tensor]: r""" Args: q (torch.Tensor): queries of shape `[B, T, H, K]`. k (torch.Tensor): keys of shape `[B, T, H, K]`. v (torch.Tensor): values of shape `[B, T, HV, V]`. GVA is applied if `HV > H`. g (torch.Tensor): g (decays) of shape `[B, T, HV]`. beta (torch.Tensor): betas of shape `[B, T, HV]`. scale (Optional[int]): Scale factor for the RetNet attention scores. If not provided, it will default to `1 / sqrt(K)`. Default: `None`. initial_state (Optional[torch.Tensor]): Initial state of shape `[N, HV, V, K]` for `N` input sequences. For equal-length input sequences, `N` equals the batch size `B`. Default: `None`. inplace_final_state: bool: Whether to store the final state in-place to save memory. Default: `True`. cu_seqlens (torch.LongTensor): Cumulative sequence lengths of shape `[N+1]` used for variable-length training, consistent with the FlashAttention API. ssm_state_indices (Optional[torch.Tensor]): Indices to map the input sequences to the initial/final states. num_accepted_tokens (Optional[torch.Tensor]): Number of accepted tokens for each sequence during decoding. Returns: o (torch.Tensor): Outputs of shape `[B, T, HV, V]`. final_state (torch.Tensor): Final state of shape `[N, HV, V, K]`. Examples:: >>> import torch >>> import torch.nn.functional as F >>> from einops import rearrange >>> from fla.ops.gated_delta_rule import fused_recurrent_gated_delta_rule # inputs with equal lengths >>> B, T, H, HV, K, V = 4, 2048, 4, 8, 512, 512 >>> q = torch.randn(B, T, H, K, device='cuda') >>> k = F.normalize(torch.randn(B, T, H, K, device='cuda'), p=2, dim=-1) >>> v = torch.randn(B, T, HV, V, device='cuda') >>> g = F.logsigmoid(torch.rand(B, T, HV, device='cuda')) >>> beta = torch.rand(B, T, HV, device='cuda').sigmoid() >>> h0 = torch.randn(B, HV, V, K, device='cuda') >>> o, ht = fused_gated_recurrent_delta_rule( q, k, v, g, beta, initial_state=h0, ) # for variable-length inputs, the batch size `B` is expected to be 1 and `cu_seqlens` is required >>> q, k, v, g, beta = map(lambda x: rearrange(x, 'b t ... -> 1 (b t) ...'), (q, k, v, g, beta)) # for a batch with 4 sequences, `cu_seqlens` with 5 start/end positions are expected >>> cu_seqlens = q.new_tensor([0, 2048, 4096, 6144, 8192], dtype=torch.long) >>> o_var, ht_var = fused_gated_recurrent_delta_rule( q, k, v, g, beta, initial_state=h0, cu_seqlens=cu_seqlens ) """ if cu_seqlens is not None and q.shape[0] != 1: raise ValueError( f"The batch size is expected to be 1 rather than {q.shape[0]} when using `cu_seqlens`." f"Please flatten variable-length inputs before processing." ) if scale is None: scale = k.shape[-1] ** -0.5 else: assert scale > 0, "scale must be positive" if beta is None: beta = torch.ones_like(q[..., 0]) o, final_state = FusedRecurrentFunction.apply( q, k, v, g, beta, scale, initial_state, inplace_final_state, cu_seqlens, ssm_state_indices, num_accepted_tokens, use_qk_l2norm_in_kernel, ) return o, final_state
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/fused_recurrent.py", "license": "Apache License 2.0", "lines": 364, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/index.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import triton from .utils import tensor_cache @tensor_cache def prepare_lens(cu_seqlens: torch.LongTensor) -> torch.LongTensor: return cu_seqlens[1:] - cu_seqlens[:-1] @tensor_cache def prepare_chunk_indices( cu_seqlens: torch.LongTensor, chunk_size: int ) -> torch.LongTensor: indices = torch.cat( [ torch.arange(n) for n in triton.cdiv(prepare_lens(cu_seqlens), chunk_size).tolist() ] ) return torch.stack([indices.eq(0).cumsum(0) - 1, indices], 1).to(cu_seqlens) @tensor_cache def prepare_chunk_offsets( cu_seqlens: torch.LongTensor, chunk_size: int ) -> torch.LongTensor: return torch.cat( [cu_seqlens.new_tensor([0]), triton.cdiv(prepare_lens(cu_seqlens), chunk_size)] ).cumsum(-1)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/index.py", "license": "Apache License 2.0", "lines": 33, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/layernorm_guard.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Tri Dao # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2024, Tri Dao. # ruff: noqa: E501 # Based on the Triton LayerNorm tutorial: https://triton-lang.org/main/getting-started/tutorials/05-layer-norm.html # For the backward pass, we keep weight_grad and bias_grad in registers and accumulate. # This backward pass is faster for dimensions up to 8k, but after that it's much slower due to register spilling. # The models we train have hidden dim up to 8k anyway (e.g. Llama 70B), so this is fine. import torch import torch.nn as nn import torch.nn.functional as F from einops import rearrange from vllm.triton_utils import tl, triton from vllm.utils.math_utils import cdiv, next_power_of_2 from vllm.utils.platform_utils import num_compute_units from .utils import input_guard def rms_norm_ref( x, weight, bias, z=None, eps=1e-6, group_size=None, norm_before_gate=True, upcast=True, ): dtype = x.dtype weight = weight.float() bias = bias.float() if bias is not None else None if upcast: x = x.float() z = z.float() if z is not None else z if z is not None and not norm_before_gate: x = x * F.silu(z) if group_size is None: rstd = 1 / torch.sqrt((x.square()).mean(dim=-1, keepdim=True) + eps) out = (x * rstd * weight) + bias if bias is not None else (x * rstd * weight) else: x_group = rearrange(x, "... (g d) -> ... g d", d=group_size) rstd = 1 / torch.sqrt((x_group.square()).mean(dim=-1, keepdim=True) + eps) out = rearrange(x_group * rstd, "... g d -> ... (g d)") * weight if bias is not None: out = out + bias if z is not None and norm_before_gate: out *= F.silu(z) return out.to(dtype) @triton.heuristics( { "HAS_BIAS": lambda args: args["B"] is not None, "HAS_Z": lambda args: args["Z"] is not None, } ) @triton.jit def layer_norm_fwd_kernel( X, # pointer to the input Y, # pointer to the output W, # pointer to the weights B, # pointer to the biases Z, # pointer to the other branch Mean, # pointer to the mean Rstd, # pointer to the 1/std stride_x_row, # how much to increase the pointer when moving by 1 row stride_y_row, stride_z_row, M, # number of rows in X N: tl.constexpr, # number of columns in X eps, # epsilon to avoid division by zero BLOCK_N: tl.constexpr, ROWS_PER_BLOCK: tl.constexpr, HAS_BIAS: tl.constexpr, HAS_Z: tl.constexpr, NORM_BEFORE_GATE: tl.constexpr, IS_RMS_NORM: tl.constexpr, ACTIVATION: tl.constexpr, ): # Map the program id to the starting row of X and Y it should compute. row_start = tl.program_id(0) * ROWS_PER_BLOCK group = tl.program_id(1) # Create 2D tile: [ROWS_PER_BLOCK, BLOCK_N] rows = row_start + tl.arange(0, ROWS_PER_BLOCK) cols = tl.arange(0, BLOCK_N) # Compute offsets for 2D tile row_offsets = rows[:, None] * stride_x_row col_offsets = cols[None, :] + group * N # Base pointers X_base = X + row_offsets + col_offsets Y_base = Y + rows[:, None] * stride_y_row + col_offsets # Create mask for valid rows and columns row_mask = rows[:, None] < M col_mask = cols[None, :] < N mask = row_mask & col_mask # Load input data with 2D tile x = tl.load(X_base, mask=mask, other=0.0).to(tl.float32) if HAS_Z and not NORM_BEFORE_GATE: Z_base = Z + rows[:, None] * stride_z_row + col_offsets z = tl.load(Z_base, mask=mask, other=0.0).to(tl.float32) if ACTIVATION == "swish" or ACTIVATION == "silu": x *= z * tl.sigmoid(z) elif ACTIVATION == "sigmoid": x *= tl.sigmoid(z) # Compute mean and variance per row (reduce along axis 1) if not IS_RMS_NORM: mean = tl.sum(x, axis=1) / N # Shape: [ROWS_PER_BLOCK] # Store mean for each row mean_offsets = group * M + rows mean_mask = rows < M tl.store(Mean + mean_offsets, mean, mask=mean_mask) # Broadcast mean back to 2D for subtraction xbar = tl.where(mask, x - mean[:, None], 0.0) var = tl.sum(xbar * xbar, axis=1) / N # Shape: [ROWS_PER_BLOCK] else: xbar = tl.where(mask, x, 0.0) var = tl.sum(xbar * xbar, axis=1) / N # Shape: [ROWS_PER_BLOCK] mean = 0.0 # Placeholder for RMS norm rstd = tl.rsqrt(var + eps) # Shape: [ROWS_PER_BLOCK] # Store rstd for each row rstd_offsets = group * M + rows rstd_mask = rows < M tl.store(Rstd + rstd_offsets, rstd, mask=rstd_mask) # Load weights and biases (broadcast across rows) w_offsets = cols + group * N w_mask = cols < N w = tl.load(W + w_offsets, mask=w_mask, other=0.0).to(tl.float32) if HAS_BIAS: b = tl.load(B + w_offsets, mask=w_mask, other=0.0).to(tl.float32) # Normalize and apply linear transformation if not IS_RMS_NORM: x_hat = (x - mean[:, None]) * rstd[:, None] else: x_hat = x * rstd[:, None] y = x_hat * w[None, :] + b[None, :] if HAS_BIAS else x_hat * w[None, :] if HAS_Z and NORM_BEFORE_GATE: Z_base = Z + rows[:, None] * stride_z_row + col_offsets z = tl.load(Z_base, mask=mask, other=0.0).to(tl.float32) if ACTIVATION == "swish" or ACTIVATION == "silu": y *= z * tl.sigmoid(z) elif ACTIVATION == "sigmoid": y *= tl.sigmoid(z) # Write output tl.store(Y_base, y, mask=mask) def calc_rows_per_block(M: int, device: torch.device) -> int: sm_count = num_compute_units(device.index) rows_per_block = next_power_of_2(cdiv(M, 2 * sm_count)) rows_per_block = min(rows_per_block, 4) return rows_per_block def layer_norm_fwd( x: torch.Tensor, weight: torch.Tensor, bias: torch.Tensor, eps: float, z: torch.Tensor = None, out: torch.Tensor = None, group_size: int = None, norm_before_gate: bool = True, is_rms_norm: bool = False, activation: str = "swish", ): M, N = x.shape if group_size is None: group_size = N assert N % group_size == 0 ngroups = N // group_size assert x.stride(-1) == 1 if z is not None: assert z.stride(-1) == 1 assert z.shape == (M, N) assert weight.shape == (N,) assert weight.stride(-1) == 1 if bias is not None: assert bias.stride(-1) == 1 assert bias.shape == (N,) # allocate output if out is not None: assert out.shape == x.shape else: out = torch.empty_like(x) assert out.stride(-1) == 1 mean = ( torch.empty((ngroups * M,), dtype=torch.float32, device=x.device) if not is_rms_norm else None ) rstd = torch.empty((ngroups * M,), dtype=torch.float32, device=x.device) # Less than 64KB per feature: enqueue fused kernel MAX_FUSED_SIZE = 65536 // x.element_size() BLOCK_N = min(MAX_FUSED_SIZE, triton.next_power_of_2(group_size)) if group_size > BLOCK_N: raise RuntimeError("This layer norm doesn't support feature dim >= 64KB.") # heuristics for number of warps num_warps = min(max(BLOCK_N // 256, 1), 8) # Calculate rows per block based on SM count rows_per_block = calc_rows_per_block(M, x.device) # Update grid to use rows_per_block grid = (cdiv(M, rows_per_block), ngroups) layer_norm_fwd_kernel[grid]( x, out, weight, bias, z, mean, rstd, x.stride(0), out.stride(0), z.stride(0) if z is not None else 0, M, group_size, eps, BLOCK_N=BLOCK_N, ROWS_PER_BLOCK=rows_per_block, HAS_BIAS=bias is not None, HAS_Z=z is not None, NORM_BEFORE_GATE=norm_before_gate, IS_RMS_NORM=is_rms_norm, num_warps=num_warps, ACTIVATION=activation, ) return out, mean, rstd class LayerNormFn(torch.autograd.Function): @input_guard @staticmethod def forward( ctx, x, weight, bias, z=None, eps=1e-6, group_size=None, norm_before_gate=True, is_rms_norm=False, activation: str = "swish", ): """If z is not None, we do norm(x) * silu(z) if norm_before_gate, else norm(x * silu(z))""" x_shape_og = x.shape # reshape input data into 2D tensor x = x.reshape(-1, x.shape[-1]) if x.stride(-1) != 1: x = x.contiguous() if z is not None: assert z.shape == x_shape_og z = z.reshape(-1, z.shape[-1]) if z.stride(-1) != 1: z = z.contiguous() weight = weight.contiguous() if bias is not None: bias = bias.contiguous() y, mean, rstd = layer_norm_fwd( x, weight, bias, eps, z=z, group_size=group_size, norm_before_gate=norm_before_gate, is_rms_norm=is_rms_norm, activation=activation, ) ctx.save_for_backward(x, weight, bias, mean, rstd, z) ctx.x_shape_og = x_shape_og ctx.eps = eps ctx.group_size = group_size ctx.norm_before_gate = norm_before_gate ctx.is_rms_norm = is_rms_norm ctx.activation = activation return y.reshape(x_shape_og) def layernorm_fn( x, weight, bias, z=None, eps=1e-6, group_size=None, norm_before_gate=True, is_rms_norm=False, activation: str = "swish", ): return LayerNormFn.apply( x, weight, bias, z, eps, group_size, norm_before_gate, is_rms_norm, activation ) def rmsnorm_fn( x, weight, bias, z=None, eps=1e-6, group_size=None, norm_before_gate=True, activation: str = "swish", ): return LayerNormFn.apply( x, weight, bias, z, eps, group_size, norm_before_gate, True, activation ) class LayerNormGated(nn.Module): def __init__( self, hidden_size, eps: float = 1e-5, group_size: int | None = None, norm_before_gate: bool = True, device: torch.device | None = None, dtype: torch.dtype | None = None, ): """If group_size is not None, we do GroupNorm with each group having group_size elements. group_size=None is equivalent to group_size=hidden_size (i.e. there's only 1 group). """ factory_kwargs = {"device": device, "dtype": dtype} super().__init__() self.eps = eps self.weight = nn.Parameter(torch.empty(hidden_size, **factory_kwargs)) self.bias = nn.Parameter(torch.empty(hidden_size, **factory_kwargs)) self.group_size = group_size self.norm_before_gate = norm_before_gate self.reset_parameters() def reset_parameters(self): torch.nn.init.ones_(self.weight) torch.nn.init.zeros_(self.bias) def forward(self, x, z=None): """If z is not None, we do norm(x) * silu(z) if norm_before_gate, else norm(x * silu(z))""" return layernorm_fn( x, self.weight, self.bias, z=z, group_size=self.group_size, eps=self.eps, norm_before_gate=self.norm_before_gate, ) class RMSNormGated(nn.Module): def __init__( self, hidden_size, eps: float = 1e-5, group_size: int | None = None, norm_before_gate: bool = False, device: torch.device | None = None, dtype: torch.dtype | None = None, activation: str = "swish", ): """If group_size is not None, we do GroupNorm with each group having group_size elements. group_size=None is equivalent to group_size=hidden_size (i.e. there's only 1 group). """ factory_kwargs = {"device": device, "dtype": dtype} super().__init__() self.eps = eps self.activation = activation self.weight = nn.Parameter(torch.empty(hidden_size, **factory_kwargs)) self.register_parameter("bias", None) self.group_size = group_size self.norm_before_gate = norm_before_gate self.reset_parameters() def reset_parameters(self): torch.nn.init.ones_(self.weight) def forward(self, x, z=None): """If z is not None, we do norm(x) * silu(z) if norm_before_gate, else norm(x * silu(z))""" return rmsnorm_fn( x, self.weight, self.bias, z=z, eps=self.eps, group_size=self.group_size, norm_before_gate=self.norm_before_gate, activation=self.activation, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/layernorm_guard.py", "license": "Apache License 2.0", "lines": 370, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/op.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang import os from vllm.triton_utils import tl, tldevice, triton from .utils import is_gather_supported if os.environ.get("FLA_USE_FAST_OPS", "0") == "1": exp = tldevice.fast_expf log = tldevice.fast_logf log2 = tldevice.fast_log2f else: exp = tl.exp log = tl.log log2 = tl.log2 if not is_gather_supported: @triton.jit def gather(src, index, axis, _builder=None): """ Gather operation that works when tl.gather is not supported. This is a fallback implementation that returns None. Just to make triton compiler happy. """ return None else: gather = tl.gather if hasattr(triton.language, "_experimental_make_tensor_descriptor"): # For Triton 3.3.x make_tensor_descriptor = triton.language._experimental_make_tensor_descriptor elif hasattr(triton.language, "make_tensor_descriptor"): # For Triton 3.4.x and later make_tensor_descriptor = triton.language.make_tensor_descriptor else: """ Fallback implementation when TMA is not supported. Returns None to indicate TMA descriptors are unavailable. Just make triton compiler happy. """ @triton.jit def make_tensor_descriptor( base, shape, strides, block_shape, _builder=None, ): return None
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/op.py", "license": "Apache License 2.0", "lines": 51, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/solve_tril.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import os import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices from .op import make_tensor_descriptor from .utils import input_guard, is_amd, is_tma_supported FLA_TRIL_PRECISION = os.environ.get("FLA_TRIL_PRECISION", "ieee") ALLOWED_TRIL_PRECISIONS = ["ieee", "tf32"] if is_amd else ["ieee", "tf32", "tf32x3"] assert FLA_TRIL_PRECISION in ALLOWED_TRIL_PRECISIONS, ( f"FLA_TRIL_PRECISION must be one of {ALLOWED_TRIL_PRECISIONS}, but got {FLA_TRIL_PRECISION}" ) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({}, num_warps=num_warps, num_stages=num_stages) for num_warps in [1, 2, 4, 8] for num_stages in [2, 3, 4, 5] ], key=["BT"], ) @triton.jit(do_not_specialize=["T"]) def solve_tril_16x16_kernel( A, Ai, cu_seqlens, chunk_indices, T, H: tl.constexpr, BT: tl.constexpr, USE_TMA: tl.constexpr, IS_VARLEN: tl.constexpr, DOT_PRECISION: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T o_i = tl.arange(0, 16) m_A = o_i[:, None] > o_i[None, :] m_I = o_i[:, None] == o_i[None, :] A = A + (bos * H + i_h) * BT Ai = Ai + (bos * H + i_h) * 16 offset = (i_t * 16) % BT if not USE_TMA: p_A = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * 16, offset), (16, 16), (1, 0) ) # [16, 16] b_A = tl.load(p_A, boundary_check=(0, 1)).to(tl.float32) else: desc = make_tensor_descriptor(A, [T, BT], [H * BT, 1], [16, 16]) desc_o = make_tensor_descriptor(Ai, [T, 16], [H * 16, 1], [16, 16]) b_A = desc.load([i_t * 16, offset]).to(tl.float32) b_A = -tl.where(m_A, b_A, 0) for i in range(2, min(16, T - i_t * 16)): # [16] b_a = -tl.load(A + (i_t * 16 + i) * H * BT + o_i + offset) b_a = b_a + tl.sum(b_a[:, None] * b_A, 0) b_A = tl.where((o_i == i)[:, None], b_a, b_A) b_A += m_I if not USE_TMA: p_Ai = tl.make_block_ptr( Ai, (T, 16), (H * 16, 1), (i_t * 16, 0), (16, 16), (1, 0) ) tl.store( p_Ai, b_A.to(p_Ai.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) else: desc_o.store([i_t * 16, 0], b_A.to(desc_o.dtype, fp_downcast_rounding="rtne")) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({}, num_warps=num_warps, num_stages=num_stages) for num_warps in [1, 2, 4, 8] for num_stages in [2, 3, 4, 5] ], key=["H", "BT", "IS_VARLEN"], ) @triton.jit(do_not_specialize=["T"]) def merge_16x16_to_32x32_inverse_kernel( A, Ai, cu_seqlens, chunk_indices, T, H: tl.constexpr, BT: tl.constexpr, USE_TMA: tl.constexpr, IS_VARLEN: tl.constexpr, DOT_PRECISION: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T o_i = tl.arange(0, 16) m_A = o_i[:, None] > o_i[None, :] m_I = o_i[:, None] == o_i[None, :] A += (bos * H + i_h) * BT Ai += (bos * H + i_h) * BT if not USE_TMA: p_A_11 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT, 0), (16, 16), (1, 0) ) p_A_22 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 16, 16), (16, 16), (1, 0) ) b_Ai_11 = tl.load(p_A_11, boundary_check=(0, 1)).to(tl.float32) b_Ai_22 = tl.load(p_A_22, boundary_check=(0, 1)).to(tl.float32) else: desc = make_tensor_descriptor(A, [T, BT], [H * BT, 1], [16, 16]) desc_o = make_tensor_descriptor(Ai, [T, BT], [H * BT, 1], [16, 16]) b_Ai_11 = desc.load([i_t * BT + 0, 0]).to(tl.float32) b_Ai_22 = desc.load([i_t * BT + 16, 16]).to(tl.float32) # [16, 16] b_Ai_11 = -tl.where(m_A, b_Ai_11, 0) b_Ai_22 = -tl.where(m_A, b_Ai_22, 0) for i in range(2, min(16, T - i_t * BT)): b_a_11 = -tl.load(A + (i_t * BT + i) * H * BT + o_i) b_a_11 += tl.sum(b_a_11[:, None] * b_Ai_11, 0) b_Ai_11 = tl.where((o_i == i)[:, None], b_a_11, b_Ai_11) for i in range(16 + 2, min(32, T - i_t * BT)): b_a_22 = -tl.load(A + (i_t * BT + i) * H * BT + o_i + 16) b_a_22 += tl.sum(b_a_22[:, None] * b_Ai_22, 0) b_Ai_22 = tl.where((o_i == i - 16)[:, None], b_a_22, b_Ai_22) b_Ai_11 += m_I b_Ai_22 += m_I if not USE_TMA: p_A_21 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 16, 0), (16, 16), (1, 0) ) b_A_21 = tl.load(p_A_21, boundary_check=(0, 1)).to(tl.float32) else: b_A_21 = desc.load([i_t * BT + 16, 0]).to(tl.float32) b_Ai_21 = -tl.dot( tl.dot(b_Ai_22, b_A_21, input_precision=DOT_PRECISION), b_Ai_11, input_precision=DOT_PRECISION, ) if not USE_TMA: p_Ai_11 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT, 0), (16, 16), (1, 0) ) p_Ai_21 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 16, 0), (16, 16), (1, 0) ) p_Ai_22 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 16, 16), (16, 16), (1, 0) ) tl.store( p_Ai_11, b_Ai_11.to(p_Ai_11.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_22, b_Ai_22.to(p_Ai_22.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_21, b_Ai_21.to(p_Ai_21.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) else: desc_o.store( [i_t * BT + 0, 0], b_Ai_11.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 16, 0], b_Ai_21.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 16, 16], b_Ai_22.to(desc_o.dtype, fp_downcast_rounding="rtne") ) @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({}, num_warps=num_warps, num_stages=num_stages) for num_warps in [2, 4, 8] for num_stages in [2, 3, 4, 5] ], key=["H", "BT", "IS_VARLEN"], ) @triton.jit(do_not_specialize=["T"]) def merge_16x16_to_64x64_inverse_kernel( A, Ai, cu_seqlens, chunk_indices, T, H: tl.constexpr, BT: tl.constexpr, USE_TMA: tl.constexpr, IS_VARLEN: tl.constexpr, DOT_PRECISION: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T o_i = tl.arange(0, 16) m_A = o_i[:, None] > o_i[None, :] m_I = o_i[:, None] == o_i[None, :] A += (bos * H + i_h) * BT Ai += (bos * H + i_h) * BT if not USE_TMA: p_A_11 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT, 0), (16, 16), (1, 0) ) p_A_22 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 16, 16), (16, 16), (1, 0) ) p_A_33 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 32, 32), (16, 16), (1, 0) ) p_A_44 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 48, 48), (16, 16), (1, 0) ) b_Ai_11 = tl.load(p_A_11, boundary_check=(0, 1)).to(tl.float32) b_Ai_22 = tl.load(p_A_22, boundary_check=(0, 1)).to(tl.float32) b_Ai_33 = tl.load(p_A_33, boundary_check=(0, 1)).to(tl.float32) b_Ai_44 = tl.load(p_A_44, boundary_check=(0, 1)).to(tl.float32) else: desc = make_tensor_descriptor(A, [T, BT], [H * BT, 1], [16, 16]) desc_o = make_tensor_descriptor(Ai, [T, BT], [H * BT, 1], [16, 16]) b_Ai_11 = desc.load([i_t * BT + 0, 0]).to(tl.float32) b_Ai_22 = desc.load([i_t * BT + 16, 16]).to(tl.float32) b_Ai_33 = desc.load([i_t * BT + 32, 32]).to(tl.float32) b_Ai_44 = desc.load([i_t * BT + 48, 48]).to(tl.float32) # [16, 16] b_Ai_11 = -tl.where(m_A, b_Ai_11, 0) b_Ai_22 = -tl.where(m_A, b_Ai_22, 0) b_Ai_33 = -tl.where(m_A, b_Ai_33, 0) b_Ai_44 = -tl.where(m_A, b_Ai_44, 0) for i in range(2, min(16, T - i_t * BT)): b_a_11 = -tl.load(A + (i_t * BT + i) * H * BT + o_i) b_a_11 += tl.sum(b_a_11[:, None] * b_Ai_11, 0) b_Ai_11 = tl.where((o_i == i)[:, None], b_a_11, b_Ai_11) for i in range(16 + 2, min(32, T - i_t * BT)): b_a_22 = -tl.load(A + (i_t * BT + i) * H * BT + o_i + 16) b_a_22 += tl.sum(b_a_22[:, None] * b_Ai_22, 0) b_Ai_22 = tl.where((o_i == i - 16)[:, None], b_a_22, b_Ai_22) for i in range(32 + 2, min(48, T - i_t * BT)): b_a_33 = -tl.load(A + (i_t * BT + i) * H * BT + o_i + 32) b_a_33 += tl.sum(b_a_33[:, None] * b_Ai_33, 0) b_Ai_33 = tl.where((o_i == i - 32)[:, None], b_a_33, b_Ai_33) for i in range(48 + 2, min(64, T - i_t * BT)): b_a_44 = -tl.load(A + (i_t * BT + i) * H * BT + o_i + 48) b_a_44 += tl.sum(b_a_44[:, None] * b_Ai_44, 0) b_Ai_44 = tl.where((o_i == i - 48)[:, None], b_a_44, b_Ai_44) b_Ai_11 += m_I b_Ai_22 += m_I b_Ai_33 += m_I b_Ai_44 += m_I if not USE_TMA: p_A_21 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 16, 0), (16, 16), (1, 0) ) p_A_31 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 32, 0), (16, 16), (1, 0) ) p_A_32 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 32, 16), (16, 16), (1, 0) ) p_A_41 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 48, 0), (16, 16), (1, 0) ) p_A_42 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 48, 16), (16, 16), (1, 0) ) p_A_43 = tl.make_block_ptr( A, (T, BT), (H * BT, 1), (i_t * BT + 48, 32), (16, 16), (1, 0) ) b_A_21 = tl.load(p_A_21, boundary_check=(0, 1)).to(tl.float32) b_A_31 = tl.load(p_A_31, boundary_check=(0, 1)).to(tl.float32) b_A_32 = tl.load(p_A_32, boundary_check=(0, 1)).to(tl.float32) b_A_41 = tl.load(p_A_41, boundary_check=(0, 1)).to(tl.float32) b_A_42 = tl.load(p_A_42, boundary_check=(0, 1)).to(tl.float32) b_A_43 = tl.load(p_A_43, boundary_check=(0, 1)).to(tl.float32) else: b_A_21 = desc.load([i_t * BT + 16, 0]).to(tl.float32) b_A_31 = desc.load([i_t * BT + 32, 0]).to(tl.float32) b_A_32 = desc.load([i_t * BT + 32, 16]).to(tl.float32) b_A_41 = desc.load([i_t * BT + 48, 0]).to(tl.float32) b_A_42 = desc.load([i_t * BT + 48, 16]).to(tl.float32) b_A_43 = desc.load([i_t * BT + 48, 32]).to(tl.float32) b_Ai_21 = -tl.dot( tl.dot(b_Ai_22, b_A_21, input_precision=DOT_PRECISION), b_Ai_11, input_precision=DOT_PRECISION, ) b_Ai_32 = -tl.dot( tl.dot(b_Ai_33, b_A_32, input_precision=DOT_PRECISION), b_Ai_22, input_precision=DOT_PRECISION, ) b_Ai_43 = -tl.dot( tl.dot(b_Ai_44, b_A_43, input_precision=DOT_PRECISION), b_Ai_33, input_precision=DOT_PRECISION, ) b_Ai_31 = -tl.dot( b_Ai_33, tl.dot(b_A_31, b_Ai_11, input_precision=DOT_PRECISION) + tl.dot(b_A_32, b_Ai_21, input_precision=DOT_PRECISION), input_precision=DOT_PRECISION, ) b_Ai_42 = -tl.dot( b_Ai_44, tl.dot(b_A_42, b_Ai_22, input_precision=DOT_PRECISION) + tl.dot(b_A_43, b_Ai_32, input_precision=DOT_PRECISION), input_precision=DOT_PRECISION, ) b_Ai_41 = -tl.dot( b_Ai_44, tl.dot(b_A_41, b_Ai_11, input_precision=DOT_PRECISION) + tl.dot(b_A_42, b_Ai_21, input_precision=DOT_PRECISION) + tl.dot(b_A_43, b_Ai_31, input_precision=DOT_PRECISION), input_precision=DOT_PRECISION, ) if not USE_TMA: p_Ai_11 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT, 0), (16, 16), (1, 0) ) p_Ai_22 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 16, 16), (16, 16), (1, 0) ) p_Ai_33 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 32, 32), (16, 16), (1, 0) ) p_Ai_44 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 48, 48), (16, 16), (1, 0) ) p_Ai_21 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 16, 0), (16, 16), (1, 0) ) p_Ai_31 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 32, 0), (16, 16), (1, 0) ) p_Ai_32 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 32, 16), (16, 16), (1, 0) ) p_Ai_41 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 48, 0), (16, 16), (1, 0) ) p_Ai_42 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 48, 16), (16, 16), (1, 0) ) p_Ai_43 = tl.make_block_ptr( Ai, (T, BT), (H * BT, 1), (i_t * BT + 48, 32), (16, 16), (1, 0) ) tl.store( p_Ai_11, b_Ai_11.to(p_Ai_11.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_22, b_Ai_22.to(p_Ai_22.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_33, b_Ai_33.to(p_Ai_33.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_44, b_Ai_44.to(p_Ai_44.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_21, b_Ai_21.to(p_Ai_21.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_31, b_Ai_31.to(p_Ai_31.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_32, b_Ai_32.to(p_Ai_32.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_41, b_Ai_41.to(p_Ai_41.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_42, b_Ai_42.to(p_Ai_42.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) tl.store( p_Ai_43, b_Ai_43.to(p_Ai_43.dtype.element_ty, fp_downcast_rounding="rtne"), boundary_check=(0, 1), ) else: desc_o.store( [i_t * BT + 0, 0], b_Ai_11.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 16, 16], b_Ai_22.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 32, 32], b_Ai_33.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 48, 48], b_Ai_44.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 16, 0], b_Ai_21.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 32, 0], b_Ai_31.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 32, 16], b_Ai_32.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 48, 0], b_Ai_41.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 48, 16], b_Ai_42.to(desc_o.dtype, fp_downcast_rounding="rtne") ) desc_o.store( [i_t * BT + 48, 32], b_Ai_43.to(desc_o.dtype, fp_downcast_rounding="rtne") ) @input_guard def solve_tril( A: torch.Tensor, cu_seqlens: torch.Tensor | None = None, output_dtype: torch.dtype = torch.float, ) -> torch.Tensor: """ Compute the inverse of the matrix I + A A should be strictly lower triangular, i.e., A.triu() == 0. Args: A (torch.Tensor): [B, T, H, BT], where BT should only be 16, 32, or 64. cu_seqlens (torch.Tensor): The cumulative sequence lengths of the input tensor. Default: `None`. output_dtype (torch.dtype): The dtype of the output tensor. Default: `torch.float`. If `None`, the output dtype will be the same as the input dtype. Returns: (I + A)^-1 with the same shape as A """ assert A.shape[-1] in [16, 32, 64] output_dtype = A.dtype if output_dtype is None else output_dtype B, T, H, BT = A.shape chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = len(chunk_indices) if cu_seqlens is not None else triton.cdiv(T, BT) Ai = torch.zeros_like(A, dtype=output_dtype) if BT == 16: merge_fn = solve_tril_16x16_kernel elif BT == 32: merge_fn = merge_16x16_to_32x32_inverse_kernel elif BT == 64: merge_fn = merge_16x16_to_64x64_inverse_kernel merge_fn[NT, B * H]( A=A, Ai=Ai, cu_seqlens=cu_seqlens, chunk_indices=chunk_indices, T=T, H=H, BT=BT, USE_TMA=is_tma_supported, DOT_PRECISION=FLA_TRIL_PRECISION, ) return Ai
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/solve_tril.py", "license": "Apache License 2.0", "lines": 519, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import contextlib import functools import logging import os from collections.abc import Callable from enum import Enum from typing import Any, Literal import torch from vllm.platforms import current_platform from vllm.triton_utils import triton logger = logging.getLogger(__name__) COMPILER_MODE = os.getenv("FLA_COMPILER_MODE") == "1" FLA_CI_ENV = os.getenv("FLA_CI_ENV") == "1" FLA_GDN_FIX_BT = os.getenv("FLA_GDN_FIX_BT", "0") == "1" SUPPRESS_LEVEL = int(os.getenv("GDN_RECOMPUTE_SUPPRESS_LEVEL", "0")) def tensor_cache(fn: Callable[..., torch.Tensor]) -> Callable[..., torch.Tensor]: """ A decorator that caches the most recent results of a function with tensor inputs. This decorator will store the output of the decorated function for the most recent set of input tensors. The cache is limited to a fixed size (default is 4). When the cache is full, the oldest entry will be removed. Args: fn (Callable[..., torch.Tensor]): The function to be decorated. It should take tensor inputs and return tensor outputs. Returns: Callable[..., torch.Tensor]: A wrapped version of the input function with single-entry caching. """ cache_entries: tuple[tuple | None, dict | None, Any] = [] cache_size = 8 @functools.wraps(fn) def wrapper(*args: Any, **kwargs: Any) -> Any: nonlocal cache_entries, cache_size for i, entry in enumerate(cache_entries): last_args, last_kwargs, last_result = entry if ( len(args) == len(last_args) and len(kwargs) == len(last_kwargs) and all(a is b for a, b in zip(args, last_args)) and all( k in last_kwargs and v is last_kwargs[k] for k, v in kwargs.items() ) ): cache_entries = ( cache_entries[:i] + cache_entries[i + 1 :] + [(args, kwargs, last_result)] ) return last_result result = fn(*args, **kwargs) if len(cache_entries) >= cache_size: cache_entries = cache_entries[1:] cache_entries.append((args, kwargs, result)) return result return wrapper def input_guard(fn: Callable[..., torch.Tensor]) -> Callable[..., torch.Tensor]: """ A decorator to make sure all input tensors are contiguous and set the device based on input tensors. """ @functools.wraps(fn) def wrapper(*args, **kwargs): contiguous_args = ( i if not isinstance(i, torch.Tensor) else i.contiguous() for i in args ) contiguous_kwargs = { k: (v if not isinstance(v, torch.Tensor) else v.contiguous()) for k, v in kwargs.items() } tensor = None for arg in args: if isinstance(arg, torch.Tensor): tensor = arg break if tensor is None: for value in kwargs.values(): if isinstance(value, torch.Tensor): tensor = value break if tensor is not None: ctx = torch.cuda.device(tensor.device.index) else: ctx = contextlib.nullcontext() with ctx: return fn(*contiguous_args, **contiguous_kwargs) return wrapper @functools.cache def get_available_device() -> str: try: return triton.runtime.driver.active.get_current_target().backend except (RuntimeError, AttributeError): return "cpu" @functools.cache def _check_platform() -> Literal["nvidia", "amd", "intel", "musa"]: device = get_available_device() mapping = { "cuda": "nvidia", "hip": "amd", "xpu": "intel", } # return the mapped value, or the original if not found return mapping.get(device, device) # For AMD GPUs, the triton backend is 'hip', while for Nvidia GPUs, the triton backend is 'cuda'. # However, the torch backend is 'cuda' for both Nvidia and AMD GPUs. # Therefore, we need to check the triton backend to determine the actual GPU vendor. device = "cuda" if current_platform.is_cuda_alike() else get_available_device() device_torch_lib = getattr(torch, device, None) device_platform = _check_platform() is_amd = device_platform == "amd" is_intel = device_platform == "intel" is_nvidia = device_platform == "nvidia" is_intel_alchemist = is_intel and "Intel(R) Arc(TM) A" in torch.xpu.get_device_name(0) is_nvidia_hopper = is_nvidia and ( "NVIDIA H" in torch.cuda.get_device_name(0) or torch.cuda.get_device_capability()[0] >= 9 ) use_cuda_graph = is_nvidia and os.environ.get("FLA_USE_CUDA_GRAPH", "0") == "1" is_gather_supported = hasattr(triton.language, "gather") is_tma_supported = (is_nvidia and torch.cuda.get_device_capability(0)[0] >= 9) and ( hasattr(triton.language, "_experimental_make_tensor_descriptor") or hasattr(triton.language, "make_tensor_descriptor") ) def get_all_max_shared_mem(): try: return [ triton.runtime.driver.active.utils.get_device_properties(i)[ "max_shared_mem" ] for i in range(device_torch_lib.device_count()) ] except BaseException: return [-1] class Backend(Enum): ADA = 101376 # RTX 4090 AMPERE = 166912 # A100 HOPPER = 232448 # H100 DEFAULT = 102400 # Default @classmethod def get_shared_memory(cls, arch: str) -> int: try: return cls[arch.upper()].value except KeyError: return cls.DEFAULT.value @functools.cache def check_shared_mem(arch: str = "none", tensor_idx: int = 0) -> bool: try: device_shared_mem_list = get_all_max_shared_mem() max_shared_memory = device_shared_mem_list[tensor_idx] return max_shared_memory >= Backend.get_shared_memory(arch) except Exception: return False
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/utils.py", "license": "Apache License 2.0", "lines": 158, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fla/ops/wy_fast.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # SPDX-FileCopyrightText: Songlin Yang, Yu Zhang # # This file contains code copied from the flash-linear-attention project. # The original source code was licensed under the MIT license and included # the following copyright notice: # Copyright (c) 2023-2025, Songlin Yang, Yu Zhang # ruff: noqa: E501 import torch from vllm.triton_utils import tl, triton from .index import prepare_chunk_indices @triton.heuristics({"IS_VARLEN": lambda args: args["cu_seqlens"] is not None}) @triton.autotune( configs=[ triton.Config({}, num_warps=num_warps, num_stages=num_stages) for num_warps in [2, 4, 8] for num_stages in [2, 3, 4] ], key=["H", "K", "V", "BT", "BK", "BV", "IS_VARLEN"], ) @triton.jit(do_not_specialize=["T"]) def recompute_w_u_fwd_kernel( k, v, beta, w, u, A, g, cu_seqlens, chunk_indices, T, H: tl.constexpr, Hg: tl.constexpr, K: tl.constexpr, V: tl.constexpr, BT: tl.constexpr, BK: tl.constexpr, BV: tl.constexpr, IS_VARLEN: tl.constexpr, ): i_t, i_bh = tl.program_id(0), tl.program_id(1) i_b, i_h = i_bh // H, i_bh % H if IS_VARLEN: i_n, i_t = ( tl.load(chunk_indices + i_t * 2).to(tl.int32), tl.load(chunk_indices + i_t * 2 + 1).to(tl.int32), ) bos, eos = ( tl.load(cu_seqlens + i_n).to(tl.int32), tl.load(cu_seqlens + i_n + 1).to(tl.int32), ) T = eos - bos else: bos, eos = i_b * T, i_b * T + T p_beta = tl.make_block_ptr( beta + bos * H + i_h, (T,), (H,), (i_t * BT,), (BT,), (0,) ) p_g = tl.make_block_ptr(g + (bos * H + i_h), (T,), (H,), (i_t * BT,), (BT,), (0,)) p_A = tl.make_block_ptr( A + (bos * H + i_h) * BT, (T, BT), (H * BT, 1), (i_t * BT, 0), (BT, BT), (1, 0) ) b_beta = tl.load(p_beta, boundary_check=(0,)) b_A = tl.load(p_A, boundary_check=(0, 1)) b_g = tl.exp(tl.load(p_g, boundary_check=(0,))) for i_v in range(tl.cdiv(V, BV)): p_v = tl.make_block_ptr( v + (bos * H + i_h) * V, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0), ) p_u = tl.make_block_ptr( u + (bos * H + i_h) * V, (T, V), (H * V, 1), (i_t * BT, i_v * BV), (BT, BV), (1, 0), ) b_v = tl.load(p_v, boundary_check=(0, 1)) b_vb = (b_v * b_beta[:, None]).to(b_v.dtype) b_u = tl.dot(b_A, b_vb, allow_tf32=False) tl.store(p_u, b_u.to(p_u.dtype.element_ty), boundary_check=(0, 1)) for i_k in range(tl.cdiv(K, BK)): p_k = tl.make_block_ptr( k + (bos * Hg + i_h // (H // Hg)) * K, (T, K), (Hg * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) p_w = tl.make_block_ptr( w + (bos * H + i_h) * K, (T, K), (H * K, 1), (i_t * BT, i_k * BK), (BT, BK), (1, 0), ) b_k = tl.load(p_k, boundary_check=(0, 1)) b_kb = (b_k * b_beta[:, None] * b_g[:, None]).to(b_k.dtype) b_w = tl.dot(b_A, b_kb) tl.store(p_w, b_w.to(p_w.dtype.element_ty), boundary_check=(0, 1)) def recompute_w_u_fwd( k: torch.Tensor, v: torch.Tensor, beta: torch.Tensor, g_cumsum: torch.Tensor, A: torch.Tensor, cu_seqlens: torch.LongTensor | None, ) -> tuple[torch.Tensor, torch.Tensor]: B, T, Hg, K, V = *k.shape, v.shape[-1] H = v.shape[-2] BT = A.shape[-1] chunk_indices = ( prepare_chunk_indices(cu_seqlens, BT) if cu_seqlens is not None else None ) NT = triton.cdiv(T, BT) if cu_seqlens is None else len(chunk_indices) BK = 64 BV = 64 u = torch.empty_like(v) w = k.new_empty(B, T, H, K) recompute_w_u_fwd_kernel[(NT, B * H)]( k=k, v=v, beta=beta, w=w, u=u, A=A, g=g_cumsum, cu_seqlens=cu_seqlens, chunk_indices=chunk_indices, T=T, H=H, Hg=Hg, K=K, V=V, BT=BT, BK=BK, BV=BV, ) return w, u
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fla/ops/wy_fast.py", "license": "Apache License 2.0", "lines": 147, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/detokenizer/test_stop_string_while_stop_model_terminates.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from vllm.sampling_params import SamplingParams from vllm.v1.engine import EngineCoreRequest from vllm.v1.engine.detokenizer import BaseIncrementalDetokenizer @pytest.fixture(params=[True, False]) def include_stop_str_in_output(request): return request.param class _DummyDetokenizer(BaseIncrementalDetokenizer): def __init__(self, request: EngineCoreRequest): super().__init__(request) def decode_next(self, next_token_id: int) -> str: # Map token id to single ASCII character for deterministic testing. return chr(next_token_id) def _make_request(stop, include_stop_str_in_output: bool, min_tokens: int = 0): params = SamplingParams( stop=stop, include_stop_str_in_output=include_stop_str_in_output, min_tokens=min_tokens, ) # Keep other fields minimal for unit test purposes. req = EngineCoreRequest( request_id="test", prompt_token_ids=[], mm_features=None, sampling_params=params, pooling_params=None, arrival_time=0.0, lora_request=None, cache_salt=None, data_parallel_rank=None, ) return req def test_stop_string_while_stop_token_terminates(include_stop_str_in_output: bool): """ This test verifies that the detokenizer correctly handles the case where the generated token sequence contains both: - a stop token - an <eos> token The detokenizer should respect the stop string and truncate the output accordingly. Imagine the following sequence: - "abcdeZ" is generated, where "Z" is the <eos> token. - "cd" is the stop string. If include_stop_str_in_output=False, the detokenizer should truncate the output to "ab" because the stop string "cd" is excluded. If include_stop_str_in_output=True, the detokenizer should include the stop string "cd" in the output, resulting in "abcd". This verifies the behavioral change introduced in BaseIncrementalDetokenizer where stop-string evaluation occurs before the early-return on stop_terminated. """ # Generate text "abcdeZ" and tokenize it. generated_text = "abcde" eos_token = "Z" stop_string = "cd" generated_text = generated_text + eos_token token_ids = [ord(c) for c in generated_text] # Create a request with the stop string and initialize the detokenizer. req = _make_request( stop=[stop_string], include_stop_str_in_output=include_stop_str_in_output ) detok = _DummyDetokenizer(req) # Simulate that the last token ('Z') is a stop token (stop_terminated=True). result = detok.update(new_token_ids=token_ids, stop_terminated=True) # The update should not report a stop string assert result == stop_string # Output text should reflect stop-string handling: # - include_stop_str_in_output=False => exclude "cd" => "ab" # - include_stop_str_in_output=True => include "cd" => "abcd" expected_text = "abcd" if include_stop_str_in_output else "ab" assert detok.output_text == expected_text # The skipped final token should still be recorded in token_ids. assert detok.output_token_ids == token_ids # get_next_output_text should return the full text when finished=True. # (Buffering only applies during streaming when finished=False.) assert detok.get_next_output_text(finished=True, delta=False) == expected_text
{ "repo_id": "vllm-project/vllm", "file_path": "tests/detokenizer/test_stop_string_while_stop_model_terminates.py", "license": "Apache License 2.0", "lines": 78, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/config/kv_transfer.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import uuid from dataclasses import field from typing import Any, Literal, get_args from vllm.config.utils import config from vllm.utils.hashing import safe_hash KVProducer = Literal["kv_producer", "kv_both"] KVConsumer = Literal["kv_consumer", "kv_both"] KVRole = Literal[KVProducer, KVConsumer] @config class KVTransferConfig: """Configuration for distributed KV cache transfer.""" kv_connector: str | None = None """The KV connector for vLLM to transmit KV caches between vLLM instances. """ engine_id: str | None = None """The engine id for KV transfers.""" kv_buffer_device: str = "cuda" """The device used by kv connector to buffer the KV cache. Choices are 'cuda' and 'cpu'.""" kv_buffer_size: float = 1e9 """The buffer size for TorchDistributedConnector. Measured in number of bytes. Recommended value: 1e9 (about 1GB).""" kv_role: KVRole | None = None """Whether this vLLM instance produces, consumes KV cache, or both. Choices are 'kv_producer', 'kv_consumer', and 'kv_both'.""" kv_rank: int | None = None """The rank of this vLLM instance in the KV cache transfer. Typical value: 0 for prefill instance, 1 for decode instance. Currently only 1P1D is supported.""" kv_parallel_size: int = 1 """The number of parallel instances for KV cache transfer. For P2pNcclConnector, this should be 2.""" kv_ip: str = "127.0.0.1" """The KV connector ip, used to build distributed connection.""" kv_port: int = 14579 """The KV connector port, used to build distributed connection.""" kv_connector_extra_config: dict[str, Any] = field(default_factory=dict) """any extra config that the connector may need.""" kv_connector_module_path: str | None = None """The Python module path to dynamically load the KV connector from. Only supported in V1.""" enable_permute_local_kv: bool = False """Experiment feature flag to enable HND to NHD KV Transfer""" kv_load_failure_policy: Literal["recompute", "fail"] = "fail" """Policy for handling KV cache load failures. 'recompute': reschedule the request to recompute failed blocks 'fail': immediately fail the request with an error finish reason (default)""" def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ # no factors to consider. # this config will not affect the computation graph. factors: list[Any] = [] hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str def __post_init__(self) -> None: if self.engine_id is None: self.engine_id = str(uuid.uuid4()) if self.kv_role is not None and self.kv_role not in get_args(KVRole): raise ValueError( f"Unsupported kv_role: {self.kv_role}. " f"Supported roles are {get_args(KVRole)}" ) if self.kv_connector is not None and self.kv_role is None: raise ValueError( "Please specify kv_role when kv_connector " f"is set, supported roles are {get_args(KVRole)}" ) @property def is_kv_transfer_instance(self) -> bool: return self.kv_connector is not None and self.kv_role in get_args(KVRole) @property def is_kv_producer(self) -> bool: return self.kv_connector is not None and self.kv_role in get_args(KVProducer) @property def is_kv_consumer(self) -> bool: return self.kv_connector is not None and self.kv_role in get_args(KVConsumer) def get_from_extra_config(self, key, default) -> Any: return self.kv_connector_extra_config.get(key, default)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/config/kv_transfer.py", "license": "Apache License 2.0", "lines": 89, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/config/kv_events.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Literal from pydantic import Field from vllm.config.utils import config @config class KVEventsConfig: """Configuration for KV event publishing.""" enable_kv_cache_events: bool = False """If True, enable KV cache events for tracking block storage and removal. Events can be published externally by zmq using the event publisher config. """ publisher: Literal["null", "zmq"] = Field(default=None) """The publisher to use for publishing kv events. Can be "null", "zmq". """ endpoint: str = "tcp://*:5557" """The zmq endpoint to use for publishing kv events. """ replay_endpoint: str | None = None """The zmq endpoint to use for replaying kv events. """ buffer_steps: int = 10_000 """The number of steps to cache for replay endpoint. Will only save events from the last N steps for the replay endpoint. """ hwm: int = 100_000 """The zmq high water mark for the event publisher. After queueing N events, events will start dropping if the consumer is not keeping up. """ max_queue_size: int = 100_000 """The maximum number of events to queue while waiting for publishing. """ topic: str = "" """The topic to use for the event publisher. Consumers can subscribe to this topic to receive events. """ def __post_init__(self): if self.publisher is None: self.publisher = "zmq" if self.enable_kv_cache_events else "null"
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/config/kv_events.py", "license": "Apache License 2.0", "lines": 39, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:benchmarks/kernels/benchmark_activation.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # benchmark custom activation op performance import itertools import torch import vllm.model_executor.layers.activation # noqa F401 from vllm.benchmarks.lib.utils import default_vllm_config from vllm.model_executor.custom_op import op_registry from vllm.triton_utils import triton from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.torch_utils import STR_DTYPE_TO_TORCH_DTYPE, set_random_seed batch_size_range = [1, 16, 128] seq_len_range = [1, 16, 64, 1024, 4096] intermediate_size = [3072, 9728, 12288] configs = list(itertools.product(batch_size_range, seq_len_range, intermediate_size)) @default_vllm_config() def benchmark_activation( batch_size: int, seq_len: int, intermediate_size: int, provider: str, func_name: str, dtype: torch.dtype, ): device = "cuda" num_tokens = batch_size * seq_len dim = intermediate_size set_random_seed(42) torch.set_default_device(device) if func_name == "gelu_and_mul": layer = op_registry[func_name](approximate="none") elif func_name == "gelu_and_mul_tanh": layer = op_registry["gelu_and_mul"](approximate="tanh") elif func_name == "fatrelu_and_mul": threshold = 0.5 layer = op_registry[func_name](threshold) else: layer = op_registry[func_name]() x = torch.randn(num_tokens, dim, dtype=dtype, device=device) compiled_layer = torch.compile(layer.forward_native) if provider == "custom": fn = lambda: layer(x) elif provider == "compiled": fn = lambda: compiled_layer(x) ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( fn, quantiles=[0.5, 0.2, 0.8] ) return ms, max_ms, min_ms if __name__ == "__main__": parser = FlexibleArgumentParser(description="Benchmark the custom activation op.") parser.add_argument( "--func-name", type=str, choices=[ "mul_and_silu", "silu_and_mul", "gelu_and_mul", "gelu_and_mul_tanh", "fatrelu_and_mul", "swigluoai_and_mul", "gelu_new", "gelu_fast", "quick_gelu", ], default="silu_and_mul", ) parser.add_argument( "--dtype", type=str, choices=["half", "bfloat16", "float"], default="bfloat16" ) args = parser.parse_args() assert args func_name = args.func_name dtype = STR_DTYPE_TO_TORCH_DTYPE[args.dtype] perf_report = triton.testing.perf_report( triton.testing.Benchmark( x_names=["batch_size", "seq_len", "intermediate_size"], x_vals=configs, line_arg="provider", line_vals=["custom", "compiled"], line_names=["Custom OP", "Compiled"], styles=[("blue", "-"), ("green", "-")], ylabel="ms", plot_name=f"{func_name}-op-performance", args={}, ) ) perf_report( lambda batch_size, seq_len, intermediate_size, provider: benchmark_activation( batch_size, seq_len, intermediate_size, provider, func_name, dtype ) ).run(print_data=True)
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/kernels/benchmark_activation.py", "license": "Apache License 2.0", "lines": 91, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/entrypoints/test_context.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from unittest.mock import MagicMock, patch import pytest from openai_harmony import Author, Message, Role, StreamState, TextContent from vllm.entrypoints.openai.responses.context import ( HarmonyContext, SimpleContext, StreamingHarmonyContext, TurnMetrics, ) from vllm.outputs import CompletionOutput, RequestOutput def create_mock_request_output( prompt_token_ids=None, output_token_ids=None, num_cached_tokens=0, finished=True, ): """Helper function to create a mock RequestOutput object for testing.""" outputs = [] token_ids = output_token_ids if output_token_ids is not None else [] outputs = [ CompletionOutput( index=0, text="Test output", token_ids=token_ids, cumulative_logprob=0.0, logprobs=None, finish_reason=None, stop_reason=None, ) ] return RequestOutput( request_id="test-id", prompt="Test prompt", prompt_token_ids=prompt_token_ids, prompt_logprobs=None, outputs=outputs, finished=finished, num_cached_tokens=num_cached_tokens, ) async def generate_mock_outputs( num_turns, prompt_token_counts, output_token_counts, cached_token_counts=None ): """Generate a sequence of mock RequestOutput objects to simulate multiple turns.""" if cached_token_counts is None: cached_token_counts = [0] * num_turns for i in range(num_turns): # Create mock prompt token IDs and output token IDs prompt_token_ids = list(range(1, prompt_token_counts[i] + 1)) output_token_ids = list(range(1, output_token_counts[i] + 1)) # Create and yield the RequestOutput yield create_mock_request_output( prompt_token_ids=prompt_token_ids, output_token_ids=output_token_ids, num_cached_tokens=cached_token_counts[i], ) @pytest.fixture def mock_parser(): """Set up a mock parser for tests.""" with patch( "vllm.entrypoints.openai.responses.context.get_streamable_parser_for_assistant" ) as mock_parser_factory: # Create a mock parser object parser = MagicMock() parser.messages = [] parser.current_channel = None parser.state = StreamState.EXPECT_START mock_parser_factory.return_value = parser yield parser def test_single_turn_token_counting(): """Test token counting behavior for a single turn.""" # Create a context context = HarmonyContext(messages=[], available_tools=[]) # Create a mock RequestOutput with specific token counts mock_output = create_mock_request_output( prompt_token_ids=[1, 2, 3, 4, 5], # 5 prompt tokens output_token_ids=[6, 7, 8], # 3 output tokens num_cached_tokens=2, # 2 cached tokens ) # Append the output to the context context.append_output(mock_output) # Verify the token counts assert context.num_prompt_tokens == 5 assert context.num_output_tokens == 3 assert context.num_cached_tokens == 2 assert context.num_tool_output_tokens == 0 # No tool tokens in first turn # Verify internal state tracking assert not context.is_first_turn assert len(context.all_turn_metrics) == 1 previous_turn = context.all_turn_metrics[0] assert previous_turn.input_tokens == 5 assert previous_turn.output_tokens == 3 assert previous_turn.cached_input_tokens == 2 assert previous_turn.tool_output_tokens == 0 @pytest.mark.asyncio async def test_multi_turn_token_counting(): """Test token counting behavior across multiple turns with tool output.""" # Create a context context = HarmonyContext(messages=[], available_tools=["browser"]) # Simulate a conversation with 3 turns # Turn 1: prefill 5, decode 3, tool 7 # Turn 2: prefill 15, cached 5, decode 4, tool 1 # Turn 3: prefill 20, cached 15, decode 5 prompt_token_counts = [5, 15, 20] output_token_counts = [3, 4, 5] cached_token_counts = [0, 5, 15] mock_generator = generate_mock_outputs( 3, prompt_token_counts, output_token_counts, cached_token_counts ) # First turn - initial prompt and response mock_output1 = await anext(mock_generator) context.append_output(mock_output1) # At this point, we should have 5 prompt tokens and 3 output tokens assert context.num_prompt_tokens == 5 assert context.num_output_tokens == 3 assert context.num_tool_output_tokens == 0 # Second turn - after tool output mock_output2 = await anext(mock_generator) context.append_output(mock_output2) # Current prompt tokens (15) - last_turn_input_tokens (5) - # last_turn_output_tokens (3) = 7 expected_tool_output = 7 assert context.num_prompt_tokens == 5 + 15 assert context.num_output_tokens == 3 + 4 assert context.num_tool_output_tokens == expected_tool_output assert context.num_cached_tokens == 5 # Third turn - final response mock_output3 = await anext(mock_generator) context.append_output(mock_output3) # Additional tool output tokens from third turn: # Current prompt (20) - last_turn_input_tokens (15) - # last_turn_output_tokens (4) = 1 expected_tool_output = 7 + 1 assert context.num_prompt_tokens == 5 + 15 + 20 assert context.num_output_tokens == 3 + 4 + 5 assert context.num_tool_output_tokens == expected_tool_output assert context.num_cached_tokens == 5 + 15 # Validate all turn metrics assert len(context.all_turn_metrics) == 3 for i, turn in enumerate(context.all_turn_metrics): assert turn.input_tokens == prompt_token_counts[i] assert turn.output_tokens == output_token_counts[i] assert turn.cached_input_tokens == cached_token_counts[i] assert context.all_turn_metrics[1].tool_output_tokens == 7 assert context.all_turn_metrics[2].tool_output_tokens == 1 def test_empty_output_tokens(): """Test behavior when RequestOutput has empty output tokens.""" context = HarmonyContext(messages=[], available_tools=[]) # Create a RequestOutput with empty output tokens mock_output = create_mock_request_output( prompt_token_ids=[1, 2, 3], # 3 prompt tokens output_token_ids=[], # Empty output tokens list num_cached_tokens=1, ) context.append_output(mock_output) # Should handle empty outputs gracefully assert context.num_prompt_tokens == 3 assert context.num_output_tokens == 0 # No output tokens assert context.num_cached_tokens == 1 assert context.num_tool_output_tokens == 0 def test_missing_prompt_token_ids(): """Test behavior when RequestOutput has None prompt_token_ids.""" context = HarmonyContext(messages=[], available_tools=[]) mock_output = create_mock_request_output( prompt_token_ids=None, # No prompt token IDs output_token_ids=[1, 2], # 2 output tokens num_cached_tokens=0, ) # Logger.error will be called, but we don't need to check for warnings # here Just ensure it doesn't raise an exception context.append_output(mock_output) # Should handle missing prompt tokens gracefully assert context.num_prompt_tokens == 0 assert context.num_output_tokens == 2 assert context.num_cached_tokens == 0 assert context.num_tool_output_tokens == 0 def test_reasoning_tokens_counting(mock_parser): """Test that reasoning tokens are counted correctly.""" context = HarmonyContext(messages=[], available_tools=[]) # Mock parser to simulate reasoning channel mock_parser.current_channel = "analysis" # Reasoning channel mock_output = create_mock_request_output( prompt_token_ids=[1, 2, 3], output_token_ids=[4, 5, 6, 7], # 4 tokens, all in reasoning num_cached_tokens=0, ) context.append_output(mock_output) # All output tokens should be counted as reasoning assert context.num_reasoning_tokens == 4 assert context.num_output_tokens == 4 def test_preamble_tokens_not_counted_as_reasoning(mock_parser): """Preambles (commentary with no recipient) are visible user text, not hidden reasoning. They must NOT inflate num_reasoning_tokens.""" context = HarmonyContext(messages=[], available_tools=[]) mock_parser.current_channel = "commentary" mock_parser.current_recipient = None # preamble mock_output = create_mock_request_output( prompt_token_ids=[1, 2, 3], output_token_ids=[4, 5, 6], num_cached_tokens=0, ) context.append_output(mock_output) assert context.num_reasoning_tokens == 0 assert context.num_output_tokens == 3 def test_commentary_with_recipient_counted_as_reasoning(mock_parser): """Commentary directed at a tool (recipient != None) is hidden from the user, so it should still count as reasoning tokens.""" context = HarmonyContext(messages=[], available_tools=[]) mock_parser.current_channel = "commentary" mock_parser.current_recipient = "python" mock_output = create_mock_request_output( prompt_token_ids=[1, 2, 3], output_token_ids=[4, 5, 6], num_cached_tokens=0, ) context.append_output(mock_output) assert context.num_reasoning_tokens == 3 assert context.num_output_tokens == 3 def test_zero_tokens_edge_case(): """Test behavior with all zero token counts.""" context = HarmonyContext(messages=[], available_tools=[]) # Create a request with empty lists (not None) for both prompt and # output tokens mock_output = create_mock_request_output( prompt_token_ids=[], # Empty prompt tokens output_token_ids=[], # Empty output tokens num_cached_tokens=0, ) context.append_output(mock_output) # All counts should be zero assert context.num_prompt_tokens == 0 assert context.num_output_tokens == 0 assert context.num_cached_tokens == 0 assert context.num_tool_output_tokens == 0 assert context.num_reasoning_tokens == 0 @pytest.mark.asyncio async def test_single_turn_no_tool_output(): """Test that first turn never generates tool output tokens.""" context = HarmonyContext( messages=[], available_tools=["browser"], # Tools available ) # Even with large prompt in first turn, no tool tokens should be counted mock_output = create_mock_request_output( prompt_token_ids=list(range(100)), # 100 tokens output_token_ids=[1, 2, 3], num_cached_tokens=0, ) context.append_output(mock_output) # First turn should never have tool output tokens assert context.num_tool_output_tokens == 0 assert context.is_first_turn is False # Should be updated after first turn @pytest.mark.asyncio async def test_negative_tool_tokens_edge_case(): """Test edge case where calculation could result in negative tool tokens. We should log an error and clamp the value to 0.""" # Use patch to check if logger.error was called with patch("vllm.entrypoints.openai.responses.context.logger.error") as mock_log: context = HarmonyContext(messages=[], available_tools=["browser"]) # First turn mock_output1 = create_mock_request_output( prompt_token_ids=list(range(10)), # 10 tokens output_token_ids=[1, 2, 3, 4, 5], # 5 tokens ) context.append_output(mock_output1) # Second turn with fewer new tokens than previous output # This could happen in edge cases with aggressive caching mock_output2 = create_mock_request_output( prompt_token_ids=list(range(12)), # 12 tokens (only 2 new) output_token_ids=[6, 7], # 2 tokens ) context.append_output(mock_output2) # Calculated negative tool tokens (12 - 10 - 5 = -3) should be clamped # to 0 and an error should be logged assert context.num_tool_output_tokens == 0 assert context.num_prompt_tokens == 10 + 12 assert context.num_output_tokens == 5 + 2 # Verify the error was logged properly mock_log.assert_called_once() # Extract the actual log message and arguments from the call args, _ = mock_log.call_args log_message = args[0] # Check for key parts of the message assert "Negative tool output tokens calculated" in log_message assert "-3" in str(args) # Check that -3 is in the arguments @pytest.mark.asyncio async def test_streaming_multi_turn_token_counting(mock_parser): """Test token counting for streaming multi-turn conversations. This test focuses on how StreamingHarmonyContext counts tokens in a multi-turn conversation with streaming (token-by-token) outputs and message boundaries. """ # Create a streaming context context = StreamingHarmonyContext(messages=[], available_tools=["browser"]) num_prompt_tokens = [3, 8, 13] num_output_tokens = [3, 3, 2] num_cached_tokens = [0, 3, 8] # Simulate three turns of conversation: # Turn 1: stream tokens one by one, then finish the message # Turn 2: new prompt, stream more tokens with a reasoning segment # Turn 3: new prompt with tool output and cached tokens # First turn: 3 tokens streamed one by one # First token of first turn context.append_output( create_mock_request_output( prompt_token_ids=[1, 2, 3], # 3 prompt tokens output_token_ids=[101], # Single token num_cached_tokens=num_cached_tokens[0], finished=False, # Not end of message yet ) ) # Second token of first turn context.append_output( create_mock_request_output( output_token_ids=[102], finished=False, ) ) # Last token of first turn (finished=True signals end of message) context.append_output( create_mock_request_output( output_token_ids=[103], finished=True, # End of message ) ) # Check token counts after first turn assert context.num_prompt_tokens == 3 # Initial prompt tokens assert context.num_output_tokens == 3 # Three output tokens assert context.num_cached_tokens == 0 assert context.num_tool_output_tokens == 0 # No tool output in first turn assert context.first_tok_of_message is True # Ready for next message # Second turn: reasoning tokens in analysis channel mock_parser.current_channel = "analysis" # Set to reasoning channel # First token of second turn context.append_output( create_mock_request_output( prompt_token_ids=[ 1, 2, 3, 101, 102, 103, 4, 5, ], # 8 tokens (includes previous) output_token_ids=[201], num_cached_tokens=num_cached_tokens[1], # Some tokens cached finished=False, ) ) # More tokens in reasoning channel context.append_output( create_mock_request_output( output_token_ids=[202], finished=False, ) ) context.append_output( create_mock_request_output( output_token_ids=[203], finished=True, # End of reasoning message ) ) # Check counts after second turn (reasoning message) assert context.num_prompt_tokens == 3 + 8 # Initial + second prompt assert context.num_output_tokens == 3 + 3 # First turn + second turn assert context.num_reasoning_tokens == 3 # All tokens in analysis channel assert context.num_cached_tokens == 3 # Cached tokens from second turn # Formula: this turn prompt tokens - last turn prompt - last turn output expected_tool_tokens = 8 - 3 - 3 # = 2 assert context.num_tool_output_tokens == expected_tool_tokens # Third turn: regular output channel mock_parser.current_channel = "final" # Switch back to regular channel # Third turn (with more cached tokens) context.append_output( create_mock_request_output( prompt_token_ids=[ 1, 2, 3, 101, 102, 103, 4, 5, 201, 202, 203, 6, 7, ], # 13 tokens output_token_ids=[301], num_cached_tokens=num_cached_tokens[2], # More cached tokens finished=False, ) ) context.append_output( create_mock_request_output( output_token_ids=[302], finished=True, ) ) # Final token counts check assert context.num_prompt_tokens == sum(num_prompt_tokens) # All prompts assert context.num_output_tokens == sum(num_output_tokens) # All outputs assert context.num_reasoning_tokens == 3 # Unchanged from second turn assert context.num_cached_tokens == sum( num_cached_tokens ) # Accumulated cached tokens # Additional tool tokens from third turn # Formula: this turn prompt - last turn prompt - last turn output additional_tool_tokens = 13 - 8 - 3 # = 2 assert ( context.num_tool_output_tokens == expected_tool_tokens + additional_tool_tokens ) # Validate all turn metrics assert len(context.all_turn_metrics) == 3 for i, turn in enumerate(context.all_turn_metrics): assert turn.input_tokens == num_prompt_tokens[i] assert turn.output_tokens == num_output_tokens[i] assert turn.cached_input_tokens == num_cached_tokens[i] assert context.all_turn_metrics[1].tool_output_tokens == 2 assert context.all_turn_metrics[2].tool_output_tokens == 2 @pytest.mark.asyncio async def test_streaming_message_synchronization(mock_parser): """Test message synchronization logic from lines 413-417 in context.py. This test verifies that when parser.messages contains more messages than the context's _messages (minus initial messages), the context properly extends its message list with the new parser messages. """ # Create a streaming context with some initial messages initial_messages = [ Message( author=Author(role=Role.USER, name="user"), content=[TextContent(text="Hello")], recipient=Role.ASSISTANT, ) ] context = StreamingHarmonyContext(messages=initial_messages, available_tools=[]) # Verify initial state assert len(context._messages) == 1 assert context.num_init_messages == 1 # Mock parser to have more messages than context # Simulate parser having processed 3 new messages mock_parser.messages = [ Message( author=Author(role=Role.ASSISTANT, name="assistant"), content=[TextContent(text="Response 1")], recipient=Role.USER, ), ] # This should trigger the message synchronization logic context.append_output( create_mock_request_output( prompt_token_ids=[1, 2, 3], output_token_ids=[101], finished=False ) ) # Verify that messages were synchronized assert len(context._messages) == 2 # Verify the new messages were added correctly assert context._messages[1].content[0].text == "Response 1" # Test the specific condition from line 413-414: # len(self._messages) - self.num_init_messages < len(self.parser.messages) messages_minus_init = len(context._messages) - context.num_init_messages parser_messages_count = len(mock_parser.messages) # After synchronization, they should be equal (no longer less than) assert messages_minus_init == parser_messages_count # Test edge case: add one more parser message mock_parser.messages.append( Message( author=Author(role=Role.ASSISTANT, name="assistant"), content=[TextContent(text="Response 4")], recipient=Role.USER, ) ) # Create another output to trigger synchronization again mock_output2 = create_mock_request_output( prompt_token_ids=[1, 2, 3], output_token_ids=[102], finished=True ) context.append_output(mock_output2) # Verify the fourth message was added, num_init_messages is still 1 assert len(context._messages) == 3 assert context.num_init_messages == 1 assert context._messages[2].content[0].text == "Response 4" def test_turn_metrics_copy_and_reset(): """Test TurnMetrics copy and reset methods work correctly.""" # Create a TurnMetrics with specific values original_metrics = TurnMetrics( input_tokens=10, output_tokens=20, cached_input_tokens=5, tool_output_tokens=3, ) # Test copy functionality copied_metrics = original_metrics.copy() # Verify copy has same values assert copied_metrics.input_tokens == 10 assert copied_metrics.output_tokens == 20 assert copied_metrics.cached_input_tokens == 5 assert copied_metrics.tool_output_tokens == 3 # Verify they are separate objects assert copied_metrics is not original_metrics # Modify copy to ensure independence copied_metrics.input_tokens = 999 assert original_metrics.input_tokens == 10 # Original unchanged assert copied_metrics.input_tokens == 999 # Test reset functionality original_metrics.reset() # Verify all fields are reset to zero assert original_metrics.input_tokens == 0 assert original_metrics.output_tokens == 0 assert original_metrics.cached_input_tokens == 0 assert original_metrics.tool_output_tokens == 0 # Verify copied metrics are unaffected by reset assert copied_metrics.input_tokens == 999 assert copied_metrics.output_tokens == 20 assert copied_metrics.cached_input_tokens == 5 assert copied_metrics.tool_output_tokens == 3 # ==================== SimpleContext Tests ==================== def create_simple_context_output( text="", token_ids=None, prompt="Test prompt", prompt_token_ids=None, num_cached_tokens=0, logprobs=None, finished=True, ): """Helper to create a RequestOutput with customizable text for SimpleContext tests.""" if token_ids is None: token_ids = [] return RequestOutput( request_id="test-id", prompt=prompt, prompt_token_ids=prompt_token_ids, prompt_logprobs=None, outputs=[ CompletionOutput( index=0, text=text, token_ids=token_ids, cumulative_logprob=0.0, logprobs=logprobs, finish_reason=None, stop_reason=None, ) ], finished=finished, num_cached_tokens=num_cached_tokens, ) def test_simple_context_output_messages_empty(): """output_messages should be empty before any output is appended.""" context = SimpleContext() assert context.output_messages == [] def test_simple_context_output_messages_single_call(): """Non-streaming: single append_output produces a single output message.""" context = SimpleContext() output = create_simple_context_output( text="Hello world", token_ids=[10, 20, 30], prompt_token_ids=[1, 2, 3], ) context.append_output(output) messages = context.output_messages assert len(messages) == 1 assert messages[0].message == "Hello world" assert messages[0].tokens == [10, 20, 30] assert messages[0].type == "raw_message_tokens" def test_simple_context_output_messages_streaming_consolidation(): """Streaming: multiple append_output calls consolidate into one message.""" context = SimpleContext() # Simulate 3 streaming deltas context.append_output( create_simple_context_output( text="Hello", token_ids=[10], prompt_token_ids=[1, 2, 3], ) ) context.append_output( create_simple_context_output( text=" world", token_ids=[20], prompt_token_ids=[1, 2, 3], ) ) context.append_output( create_simple_context_output( text="!", token_ids=[30], prompt_token_ids=[1, 2, 3], ) ) messages = context.output_messages assert len(messages) == 1 assert messages[0].message == "Hello world!" assert messages[0].tokens == [10, 20, 30] def test_simple_context_output_messages_many_deltas(): """Streaming with many small deltas still produces a single message.""" context = SimpleContext() words = ["The", " quick", " brown", " fox", " jumps"] for i, word in enumerate(words): context.append_output( create_simple_context_output( text=word, token_ids=[100 + i], prompt_token_ids=[1, 2], ) ) messages = context.output_messages assert len(messages) == 1 assert messages[0].message == "The quick brown fox jumps" assert messages[0].tokens == [100, 101, 102, 103, 104] def test_simple_context_input_messages(): """input_messages is populated on the first append_output call.""" context = SimpleContext() assert context.input_messages == [] context.append_output( create_simple_context_output( text="Hi", token_ids=[10], prompt="My prompt text", prompt_token_ids=[1, 2, 3], ) ) assert len(context.input_messages) == 1 assert context.input_messages[0].message == "My prompt text" assert context.input_messages[0].tokens == [1, 2, 3] # Second call should not add another input message context.append_output( create_simple_context_output( text=" there", token_ids=[20], prompt="My prompt text", prompt_token_ids=[1, 2, 3], ) ) assert len(context.input_messages) == 1 def test_simple_context_token_counting(): """Token counting accumulates across streaming deltas.""" context = SimpleContext() context.append_output( create_simple_context_output( text="a", token_ids=[10, 11], prompt_token_ids=[1, 2, 3, 4, 5], num_cached_tokens=2, ) ) context.append_output( create_simple_context_output( text="b", token_ids=[12], prompt_token_ids=[1, 2, 3, 4, 5], num_cached_tokens=2, ) ) assert context.num_prompt_tokens == 5 assert context.num_output_tokens == 3 # 2 + 1 assert context.num_cached_tokens == 2 def test_simple_context_final_output(): """final_output reconstructs accumulated text and token_ids.""" context = SimpleContext() context.append_output( create_simple_context_output( text="foo", token_ids=[1, 2], prompt_token_ids=[10], ) ) context.append_output( create_simple_context_output( text="bar", token_ids=[3], prompt_token_ids=[10], ) ) final = context.final_output assert final is not None assert final.outputs[0].text == "foobar" assert final.outputs[0].token_ids == (1, 2, 3) def test_simple_context_output_messages_empty_text_with_tokens(): """output_messages should be returned when tokens exist even if text is empty (e.g. special tokens).""" context = SimpleContext() context.append_output( create_simple_context_output( text="", token_ids=[99], prompt_token_ids=[1], ) ) messages = context.output_messages assert len(messages) == 1 assert messages[0].message == "" assert messages[0].tokens == [99] def test_simple_context_output_messages_no_mutation(): """Each call to output_messages returns a fresh list; callers can't corrupt internal state.""" context = SimpleContext() context.append_output( create_simple_context_output( text="hello", token_ids=[1], prompt_token_ids=[10], ) ) msgs1 = context.output_messages msgs2 = context.output_messages assert msgs1 is not msgs2 assert msgs1[0].message == msgs2[0].message # Appending more output updates the property context.append_output( create_simple_context_output( text=" world", token_ids=[2], prompt_token_ids=[10], ) ) msgs3 = context.output_messages assert len(msgs3) == 1 assert msgs3[0].message == "hello world" assert msgs3[0].tokens == [1, 2]
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/test_context.py", "license": "Apache License 2.0", "lines": 727, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/distributed/test_context_parallel.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ WARNING: This test runs in both single-node (4 GPUs) and multi-node (2 node with 2 GPUs each) modes. If the test only uses 2 GPUs, it is important to set the distributed backend to "mp" to avoid Ray scheduling all workers in a node other than the head node, which can cause the test to fail. """ import json import os from dataclasses import dataclass from typing import Literal, NamedTuple import pytest import torch from tests.evals.gsm8k.gsm8k_eval import evaluate_gsm8k from tests.utils import RemoteOpenAIServer, create_new_process_for_each_test from vllm.config.model import RunnerOption from vllm.logger import init_logger from ..models.registry import HF_EXAMPLE_MODELS logger = init_logger("test_context_parallel") VLLM_MULTI_NODE = os.getenv("VLLM_MULTI_NODE", "0") == "1" CP_TEST_MODELS = [ # TODO support other models # [LANGUAGE GENERATION] "deepseek-ai/DeepSeek-V2-Lite-Chat", "Qwen/Qwen2.5-1.5B-Instruct", ] # GSM8K eval configuration NUM_QUESTIONS = 256 # Fast eval for CI NUM_SHOTS = 5 # Few-shot examples # tp accuracy with 2% buffer MIN_ACCURACY = { # .buildkite/lm-eval-harness/configs/DeepSeek-V2-Lite-Chat.yaml "deepseek-ai/DeepSeek-V2-Lite-Chat": 0.64, # .buildkite/lm-eval-harness/configs/Qwen2.5-1.5B-Instruct.yaml "Qwen/Qwen2.5-1.5B-Instruct": 0.52, } class ParallelSetup(NamedTuple): tp_size: int pp_size: int dcp_size: int cp_kv_cache_interleave_size: int eager_mode: bool chunked_prefill: bool class CPTestOptions(NamedTuple): multi_node_only: bool attn_backend: str | None = None @dataclass class CPTestSettings: parallel_setups: list[ParallelSetup] distributed_backends: list[str] runner: RunnerOption test_options: CPTestOptions @staticmethod def detailed( *, tp_base: int = 4, pp_base: int = 1, dcp_multipliers: list[float] | None = None, cp_kv_cache_interleave_size: int = 1, multi_node_only: bool = False, runner: RunnerOption = "auto", attn_backend: str | None = None, ): parallel_setups = [] if dcp_multipliers is None: dcp_multipliers = [ 0.5, ] for eager_mode_val in [False]: for pp_multiplier in [1]: for dcp_multiplier in dcp_multipliers: for chunked_prefill_val in [True]: parallel_setups.append( ParallelSetup( tp_size=tp_base, pp_size=pp_multiplier * pp_base, dcp_size=int(dcp_multiplier * tp_base), cp_kv_cache_interleave_size=cp_kv_cache_interleave_size, eager_mode=eager_mode_val, chunked_prefill=chunked_prefill_val, ) ) return CPTestSettings( parallel_setups=parallel_setups, distributed_backends=["mp"], runner=runner, test_options=CPTestOptions( multi_node_only=multi_node_only, attn_backend=attn_backend, ), ) def iter_params(self, model_id: str): opts = self.test_options for parallel_setup in self.parallel_setups: for backend in self.distributed_backends: yield ( model_id, parallel_setup, backend, self.runner, opts, ) CP_TEXT_GENERATION_MODELS = { "deepseek-ai/DeepSeek-V2-Lite-Chat": [ CPTestSettings.detailed(dcp_multipliers=[1]), CPTestSettings.detailed( dcp_multipliers=[0.5], cp_kv_cache_interleave_size=64, attn_backend="FLASHMLA", ), ], "Qwen/Qwen2.5-1.5B-Instruct": [ CPTestSettings.detailed( cp_kv_cache_interleave_size=16, attn_backend="FLASH_ATTN" ), CPTestSettings.detailed( cp_kv_cache_interleave_size=16, attn_backend="FLASHINFER" ), ], } def _test_cp_gsm8k( model_id: str, parallel_setup: ParallelSetup, distributed_backend: str, runner: RunnerOption, test_options: CPTestOptions, num_gpus_available: int, *, method: Literal["generate"], is_multimodal: bool, ): ( tp_size, pp_size, dcp_size, cp_kv_cache_interleave_size, eager_mode, chunked_prefill, ) = parallel_setup multi_node_only, attn_backend = test_options model_info = HF_EXAMPLE_MODELS.find_hf_info(model_id) model_info.check_transformers_version(on_fail="skip") trust_remote_code = model_info.trust_remote_code tokenizer_mode = model_info.tokenizer_mode hf_overrides = model_info.hf_overrides model_info.check_available_online(on_fail="skip") if num_gpus_available < tp_size * pp_size: pytest.skip(f"Need at least {tp_size} x {pp_size} GPUs") if VLLM_MULTI_NODE and distributed_backend == "mp": pytest.skip( "Skipping multi-node pipeline parallel test for " "multiprocessing distributed backend" ) if multi_node_only and not VLLM_MULTI_NODE: pytest.skip("Not in multi-node setting") server_args = [ # use half precision for speed and memory savings in CI environment "--dtype", "bfloat16", "--max-model-len", "4096", "--max-num-seqs", "64", ] if chunked_prefill: server_args.append("--enable-chunked-prefill") if eager_mode: server_args.append("--enforce-eager") if runner != "auto": server_args.extend(["--runner", runner]) if trust_remote_code: server_args.append("--trust-remote-code") if tokenizer_mode: server_args.extend(["--tokenizer-mode", tokenizer_mode]) if hf_overrides: server_args.extend(["--hf-overrides", json.dumps(hf_overrides)]) server_args.extend( [ "--tensor-parallel-size", str(tp_size), "--pipeline-parallel-size", str(pp_size), "--decode-context-parallel-size", str(dcp_size), "--dcp-kv-cache-interleave-size", str(cp_kv_cache_interleave_size), "--distributed-executor-backend", distributed_backend, ] ) if attn_backend: server_args.append(f"--attention-backend={attn_backend}") with RemoteOpenAIServer( model_id, server_args, max_wait_seconds=720, ) as remote_server: host = f"http://{remote_server.host}" port = remote_server.port # Run GSM8K evaluation results = evaluate_gsm8k( num_questions=NUM_QUESTIONS, num_shots=NUM_SHOTS, host=host, port=port, ) # Validate accuracy is reasonable accuracy = results["accuracy"] min_accuracy = MIN_ACCURACY[model_id] assert accuracy >= min_accuracy, ( f"TP+DCP accuracy too low: {accuracy:.3f} < {min_accuracy:.3f}" ) @pytest.mark.parametrize( ( "model_id", "parallel_setup", "distributed_backend", "runner", "test_options", ), [ params for model_id, settings in CP_TEXT_GENERATION_MODELS.items() for setting in settings for params in setting.iter_params(model_id) if model_id in CP_TEST_MODELS ], ) @create_new_process_for_each_test() def test_cp_generation( model_id: str, parallel_setup: ParallelSetup, distributed_backend: str, runner: RunnerOption, test_options: CPTestOptions, num_gpus_available, ): if ( model_id == "deepseek-ai/DeepSeek-V2-Lite-Chat" and torch.cuda.get_device_capability() < (9, 0) ): pytest.skip(reason="MLA+DCP requires compute capability of 9.0 or higher") if ( model_id == "Qwen/Qwen2.5-1.5B-Instruct" and torch.cuda.get_device_capability() != (9, 0) ): pytest.skip(reason="GQA+DCP currently requires compute capability of 9.0") _test_cp_gsm8k( model_id, parallel_setup, distributed_backend, runner, test_options, num_gpus_available, method="generate", is_multimodal=False, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/distributed/test_context_parallel.py", "license": "Apache License 2.0", "lines": 259, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/metrics/test_engine_logger_apis.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import copy import pytest from tests.plugins.vllm_add_dummy_stat_logger.dummy_stat_logger.dummy_stat_logger import ( # noqa E501 DummyStatLogger, ) from vllm.v1.engine.async_llm import AsyncEngineArgs, AsyncLLM from vllm.v1.metrics.ray_wrappers import RayPrometheusStatLogger @pytest.fixture def log_stats_enabled_engine_args(): """ Shared fixture providing common AsyncEngineArgs configuration used across multiple tests. """ return AsyncEngineArgs( model="distilbert/distilgpt2", dtype="half", disable_log_stats=False, enforce_eager=True, ) @pytest.mark.asyncio async def test_async_llm_replace_default_loggers(log_stats_enabled_engine_args): """ RayPrometheusStatLogger should replace the default PrometheusStatLogger """ engine = AsyncLLM.from_engine_args( log_stats_enabled_engine_args, stat_loggers=[RayPrometheusStatLogger] ) assert isinstance(engine.logger_manager.stat_loggers[0], RayPrometheusStatLogger) engine.shutdown() @pytest.mark.asyncio async def test_async_llm_add_to_default_loggers(log_stats_enabled_engine_args): """ It's still possible to use custom stat loggers exclusively by passing disable_log_stats=True in addition to a list of custom stat loggers. """ # Create engine_args with disable_log_stats=True for this test disabled_log_engine_args = copy.deepcopy(log_stats_enabled_engine_args) disabled_log_engine_args.disable_log_stats = True # Disable default loggers; pass custom stat logger to the constructor engine = AsyncLLM.from_engine_args( disabled_log_engine_args, stat_loggers=[DummyStatLogger] ) assert len(engine.logger_manager.stat_loggers) == 2 assert len(engine.logger_manager.stat_loggers[0].per_engine_stat_loggers) == 1 assert isinstance( engine.logger_manager.stat_loggers[0].per_engine_stat_loggers[0], DummyStatLogger, ) # log_stats is still True, since custom stat loggers are used assert engine.log_stats engine.shutdown()
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/metrics/test_engine_logger_apis.py", "license": "Apache License 2.0", "lines": 53, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/v1/attention/backends/mla/flashattn_mla.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass from typing import ClassVar import torch from vllm.config import VllmConfig from vllm.config.cache import CacheDType from vllm.logger import init_logger from vllm.model_executor.layers.attention.mla_attention import ( MLACommonBackend, MLACommonDecodeMetadata, MLACommonImpl, MLACommonMetadata, MLACommonMetadataBuilder, QueryLenSupport, ) from vllm.model_executor.layers.batch_invariant import ( vllm_is_batch_invariant, ) from vllm.platforms.interface import DeviceCapability from vllm.utils.math_utils import round_up from vllm.v1.attention.backend import ( AttentionCGSupport, AttentionLayer, AttentionType, MultipleOf, is_quantized_kv_cache, ) from vllm.v1.attention.backends.fa_utils import ( flash_attn_supports_mla, get_flash_attn_version, ) from vllm.v1.kv_cache_interface import AttentionSpec from vllm.vllm_flash_attn import ( # type: ignore[attr-defined] flash_attn_varlen_func, get_scheduler_metadata, ) logger = init_logger(__name__) class FlashAttnMLABackend(MLACommonBackend): supported_dtypes: ClassVar[list[torch.dtype]] = [torch.float16, torch.bfloat16] supported_kv_cache_dtypes: ClassVar[list[CacheDType]] = [ "auto", "bfloat16", ] @staticmethod def get_supported_kernel_block_sizes() -> list[int | MultipleOf]: return [MultipleOf(16)] @staticmethod def get_name() -> str: return "FLASH_ATTN_MLA" @staticmethod def get_builder_cls() -> type["FlashAttnMLAMetadataBuilder"]: return FlashAttnMLAMetadataBuilder @staticmethod def get_impl_cls() -> type["FlashAttnMLAImpl"]: return FlashAttnMLAImpl @classmethod def supports_compute_capability(cls, capability: DeviceCapability) -> bool: return capability.major == 9 @classmethod def supports_combination( cls, head_size: int, dtype: torch.dtype, kv_cache_dtype: CacheDType | None, block_size: int, use_mla: bool, has_sink: bool, use_sparse: bool, device_capability: DeviceCapability, ) -> str | None: if not flash_attn_supports_mla(): return "FlashAttention MLA not supported on this device" return None @dataclass class FlashAttnMLADecodeMetadata(MLACommonDecodeMetadata): query_start_loc: torch.Tensor max_query_len: int max_seq_len: int scheduler_metadata: torch.Tensor | None = None max_num_splits: int = 0 @dataclass class FlashAttnMLAMetadata(MLACommonMetadata[FlashAttnMLADecodeMetadata]): pass class FlashAttnMLAMetadataBuilder(MLACommonMetadataBuilder[FlashAttnMLAMetadata]): _cudagraph_support: ClassVar[AttentionCGSupport] = AttentionCGSupport.UNIFORM_BATCH query_len_support: ClassVar[QueryLenSupport] = QueryLenSupport.VARLEN reorder_batch_threshold: int = 512 # process small prefills with decode pathway def __init__( self, kv_cache_spec: AttentionSpec, layer_names: list[str], vllm_config: VllmConfig, device: torch.device, ): interleave_size = vllm_config.parallel_config.cp_kv_cache_interleave_size super().__init__( kv_cache_spec, layer_names, vllm_config, device, FlashAttnMLAMetadata, supports_dcp_with_varlen=(interleave_size == 1), ) self.max_num_splits = 0 # No upper bound on the number of splits. self.fa_aot_schedule = get_flash_attn_version() == 3 self.use_full_cuda_graph = ( self.compilation_config.cudagraph_mode.has_full_cudagraphs() ) self.max_cudagraph_size = self.compilation_config.max_cudagraph_capture_size if self.use_full_cuda_graph and self.fa_aot_schedule: # FA3 scheduler_metadata size: 1 + round_up(batch_size, 4) * 4 # The +1 is for the tile_count_semaphore (synchronization). # The 4 slots per batch element (num_prepare_batch_vectors) are: # prepare_varlen + dynamic_split + sort_batches + head_swizzle # See: https://github.com/vllm-project/flash-attention/blob/5824e6e/hopper/flash_api.cpp#L664-L671 # noqa: E501 max_batch_size = max( vllm_config.scheduler_config.max_num_seqs, self.max_cudagraph_size or 0, ) self.scheduler_metadata = torch.zeros( 1 + round_up(max_batch_size, 4) * 4, dtype=torch.int32, device=self.device, ) # When using cuda graph, we need to set the upper bound of the # number of splits so that large enough intermediate buffers are # pre-allocated during capture. self.max_num_splits = ( vllm_config.attention_config.flash_attn_max_num_splits_for_cuda_graph ) if vllm_is_batch_invariant(): self.max_num_splits = 1 def _schedule_decode( self, num_reqs, cu_query_lens, max_query_len, seqlens, max_seq_len, causal, max_num_splits, ): if self.fa_aot_schedule: return get_scheduler_metadata( batch_size=num_reqs, max_seqlen_q=max_query_len, max_seqlen_k=max_seq_len, num_heads_q=self.num_heads * self.dcp_world_size, num_heads_kv=1, headdim=self.mla_dims.qk_rope_head_dim, cache_seqlens=seqlens, qkv_dtype=self.kv_cache_spec.dtype, headdim_v=self.mla_dims.kv_lora_rank, page_size=self.page_size, cu_seqlens_q=cu_query_lens, causal=causal, num_splits=max_num_splits, ) return None def _build_decode( self, block_table_tensor: torch.Tensor, seq_lens_device: torch.Tensor, max_seq_len: int, query_start_loc_cpu: torch.Tensor, query_start_loc_device: torch.Tensor, num_decode_tokens: int, dcp_tot_seq_lens_device: torch.Tensor | None, ) -> FlashAttnMLADecodeMetadata: query_lens_cpu = query_start_loc_cpu[1:] - query_start_loc_cpu[:-1] max_query_len = query_lens_cpu.max().item() # For Flash Attention MLA + full cudagraph max_num_splits = 0 if ( self.use_full_cuda_graph and self.max_cudagraph_size is not None and num_decode_tokens <= self.max_cudagraph_size ): # NOTE(woosuk): Setting num_splits > 1 may increase the memory # usage, because the intermediate buffers of size [num_splits, # num_heads, num_tokens, head_size] are allocated. Therefore, # we only set num_splits when using cuda graphs. max_num_splits = self.max_num_splits if vllm_is_batch_invariant(): max_num_splits = 1 scheduler_metadata = self._schedule_decode( num_reqs=seq_lens_device.shape[0], cu_query_lens=query_start_loc_device, max_query_len=max_query_len, seqlens=seq_lens_device, max_seq_len=max_seq_len, causal=True, max_num_splits=max_num_splits, ) if self.use_full_cuda_graph and scheduler_metadata is not None: n = scheduler_metadata.shape[0] # Ensure the persistent buffer is large enough assert n <= self.scheduler_metadata.shape[0], ( f"Scheduler metadata size {n} exceeds buffer size " f"{self.scheduler_metadata.shape[0]}" ) self.scheduler_metadata[:n] = scheduler_metadata # NOTE(woosuk): We should zero out the rest of the scheduler # metadata to guarantee the correctness. Otherwise, some thread # blocks may use the invalid scheduler metadata and overwrite the # output buffer. self.scheduler_metadata[n:] = 0 scheduler_metadata = self.scheduler_metadata[:n] metadata = FlashAttnMLADecodeMetadata( block_table=block_table_tensor, seq_lens=seq_lens_device, query_start_loc=query_start_loc_device, max_query_len=max_query_len, max_seq_len=max_seq_len, scheduler_metadata=scheduler_metadata, max_num_splits=max_num_splits, dcp_tot_seq_lens=dcp_tot_seq_lens_device, ) return metadata class FlashAttnMLAImpl(MLACommonImpl[FlashAttnMLAMetadata]): can_return_lse_for_decode: bool = True def __init__( self, num_heads: int, head_size: int, scale: float, num_kv_heads: int, alibi_slopes: list[float] | None, sliding_window: int | None, kv_cache_dtype: str, logits_soft_cap: float | None, attn_type: str, kv_sharing_target_layer_name: str | None, # MLA Specific Arguments **mla_args, ) -> None: super().__init__( num_heads, head_size, scale, num_kv_heads, alibi_slopes, sliding_window, kv_cache_dtype, logits_soft_cap, attn_type, kv_sharing_target_layer_name, **mla_args, ) assert flash_attn_supports_mla(), "FlashAttnMLA is not supported on this device" unsupported_features = [alibi_slopes, sliding_window, logits_soft_cap] if any(unsupported_features): raise NotImplementedError( "FlashAttnMLAImpl does not support one of the following: " "alibi_slopes, sliding_window, logits_soft_cap" ) if attn_type != AttentionType.DECODER: raise NotImplementedError( "Encoder self-attention and " "encoder/decoder cross-attention " "are not implemented for " "FlashAttnMLAImpl" ) if is_quantized_kv_cache(self.kv_cache_dtype): raise NotImplementedError( "FlashAttnMLA V1 with FP8 KV cache not yet supported" ) def forward_mqa( self, q: torch.Tensor | tuple[torch.Tensor, torch.Tensor], kv_c_and_k_pe_cache: torch.Tensor, attn_metadata: FlashAttnMLAMetadata, layer: AttentionLayer, ) -> tuple[torch.Tensor, torch.Tensor | None]: assert kv_c_and_k_pe_cache.numel() > 0 assert attn_metadata.decode is not None if type(q) is tuple: q_nope, q_pe = q else: q_nope, q_pe = torch.split( q, [self.kv_lora_rank, self.qk_rope_head_dim], dim=-1 ) if self.kv_cache_dtype.startswith("fp8"): raise NotImplementedError("FP8 FlashAttention MLA not yet supported") kv_c_cache = kv_c_and_k_pe_cache[..., : self.kv_lora_rank] k_pe_cache = kv_c_and_k_pe_cache[..., self.kv_lora_rank :] # NOTE(matt): During CUDA graph capture, max_query_len can be 0, but the # kernel uses this to calculate grid dimensions. Ensure it's at least 1 # to prevent invalid grid configuration during graph capture. max_seqlen_q = max(attn_metadata.decode.max_query_len, 1) attn_out = flash_attn_varlen_func( q=q_pe, k=k_pe_cache.unsqueeze(-2), # Add head dim of 1 v=kv_c_cache.unsqueeze(-2), # Add head dim of 1 q_v=q_nope, max_seqlen_q=max_seqlen_q, cu_seqlens_q=attn_metadata.decode.query_start_loc, max_seqlen_k=attn_metadata.decode.max_seq_len, seqused_k=attn_metadata.decode.seq_lens, block_table=attn_metadata.decode.block_table, softmax_scale=self.scale, causal=True, return_softmax_lse=self.need_to_return_lse_for_decode, fa_version=3, # only version 3 is supported scheduler_metadata=attn_metadata.decode.scheduler_metadata, num_splits=attn_metadata.decode.max_num_splits, cp_world_size=self.dcp_world_size, cp_rank=self.dcp_rank, cp_tot_seqused_k=attn_metadata.decode.dcp_tot_seq_lens, ) if self.need_to_return_lse_for_decode: o, lse = attn_out # FA returns LSE in shape [ H, B ] but DCP wants [ B, H ] return o, lse.transpose(0, 1) # [ H, B ] -> [ B, H ] else: o = attn_out return o, None
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/attention/backends/mla/flashattn_mla.py", "license": "Apache License 2.0", "lines": 318, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/mla.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass import torch from vllm.config import CacheConfig from vllm.model_executor.custom_op import PluggableLayer from vllm.model_executor.layers.attention import MLAAttention from vllm.model_executor.layers.quantization import QuantizationConfig @dataclass class MLAModules: """Modules used in MLA.""" kv_a_layernorm: torch.nn.Module kv_b_proj: torch.nn.Module rotary_emb: torch.nn.Module o_proj: torch.nn.Module fused_qkv_a_proj: torch.nn.Module | None kv_a_proj_with_mqa: torch.nn.Module | None q_a_layernorm: torch.nn.Module | None q_b_proj: torch.nn.Module | None q_proj: torch.nn.Module | None indexer: torch.nn.Module | None is_sparse: bool topk_indices_buffer: torch.Tensor | None indexer_rotary_emb: torch.nn.Module | None = None # --8<-- [start:multi_head_latent_attention] @PluggableLayer.register("multi_head_latent_attention") class MultiHeadLatentAttentionWrapper(PluggableLayer): """Pluggable MLA layer which allows OOT backends to add custom implementations of the outer MLA layer (including rope & o_proj). Note that currently oot platforms can still use CustomOp.register_oot to replace MLA layer entirly, although we use PluggableLayer to register this layer now. This class takes positions and hidden_states as input. The input tensors can either contain prefill tokens or decode tokens. The class does the following: 1. MLA Preprocess. 2. Perform multi-head attention to prefill tokens and multi-query attention to decode tokens separately. 3. Return the output tensor. """ # --8<-- [end:multi_head_latent_attention] def __init__( self, hidden_size: int, num_heads: int, scale: float, qk_nope_head_dim: int, qk_rope_head_dim: int, v_head_dim: int, q_lora_rank: int | None, kv_lora_rank: int, mla_modules: MLAModules, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.hidden_size = hidden_size self.qk_nope_head_dim = qk_nope_head_dim self.qk_rope_head_dim = qk_rope_head_dim self.qk_head_dim = qk_nope_head_dim + qk_rope_head_dim self.v_head_dim = v_head_dim self.q_lora_rank = q_lora_rank self.kv_lora_rank = kv_lora_rank self.num_heads = num_heads self.fused_qkv_a_proj = mla_modules.fused_qkv_a_proj self.kv_a_proj_with_mqa = mla_modules.kv_a_proj_with_mqa self.q_a_layernorm = mla_modules.q_a_layernorm self.q_b_proj = mla_modules.q_b_proj self.q_proj = mla_modules.q_proj self.kv_a_layernorm = mla_modules.kv_a_layernorm self.kv_b_proj = mla_modules.kv_b_proj self.rotary_emb = mla_modules.rotary_emb self.o_proj = mla_modules.o_proj self.indexer = mla_modules.indexer self.indexer_rope_emb = mla_modules.indexer_rotary_emb self.is_sparse = mla_modules.is_sparse if self.indexer is not None: assert hasattr(self.indexer, "topk_tokens") self.topk_tokens = self.indexer.topk_tokens self.topk_indices_buffer = mla_modules.topk_indices_buffer self.mla_attn = MLAAttention( num_heads=self.num_heads, scale=scale, qk_nope_head_dim=self.qk_nope_head_dim, qk_rope_head_dim=self.qk_rope_head_dim, v_head_dim=self.v_head_dim, q_lora_rank=self.q_lora_rank, kv_lora_rank=self.kv_lora_rank, cache_config=cache_config, quant_config=quant_config, prefix=f"{prefix}.attn", kv_b_proj=self.kv_b_proj, use_sparse=self.is_sparse, indexer=self.indexer, ) self.prefix = prefix def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, llama_4_scaling: torch.Tensor | None = None, ) -> torch.Tensor: q_c = None kv_lora = None if self.q_lora_rank is not None: assert self.fused_qkv_a_proj is not None, ( "fused_qkv_a_proj is required when q_lora_rank is not None" ) assert self.q_a_layernorm is not None, ( "q_a_layernorm is required when q_lora_rank is not None" ) assert self.q_b_proj is not None, ( "q_b_proj is required when q_lora_rank is not None" ) qkv_lora = self.fused_qkv_a_proj(hidden_states)[0] q_c, kv_lora = qkv_lora.split( [self.q_lora_rank, self.kv_lora_rank + self.qk_rope_head_dim], dim=-1, ) q_c = self.q_a_layernorm(q_c) q = self.q_b_proj(q_c)[0] else: assert self.kv_a_proj_with_mqa is not None, ( "kv_a_proj_with_mqa is required when q_lora_rank is None" ) assert self.q_proj is not None, ( "q_proj is required when q_lora_rank is None" ) kv_lora = self.kv_a_proj_with_mqa(hidden_states)[0] q = self.q_proj(hidden_states)[0] kv_c, k_pe = kv_lora.split([self.kv_lora_rank, self.qk_rope_head_dim], dim=-1) kv_c_normed = self.kv_a_layernorm(kv_c) q = q.view(-1, self.num_heads, self.qk_head_dim) # Add head dim of 1 to k_pe k_pe = k_pe.unsqueeze(1) if self.rotary_emb is not None: q[..., self.qk_nope_head_dim :], k_pe = self.rotary_emb( positions, q[..., self.qk_nope_head_dim :], k_pe ) if self.indexer and self.is_sparse: _topk_indices = self.indexer( hidden_states, q_c, positions, self.indexer_rope_emb ) if llama_4_scaling is not None: q *= llama_4_scaling attn_out = self.mla_attn( q, kv_c_normed, k_pe, output_shape=(hidden_states.shape[0], self.num_heads * self.v_head_dim), ) return self.o_proj(attn_out)[0]
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/mla.py", "license": "Apache License 2.0", "lines": 153, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/models/test_terratorch.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from tests.conftest import VllmRunner from tests.utils import create_new_process_for_each_test @create_new_process_for_each_test() # Memory is not cleaned up properly otherwise @pytest.mark.parametrize( "model", [ "ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11", "ibm-nasa-geospatial/Prithvi-EO-2.0-300M-BurnScars", ], ) def test_inference( vllm_runner: type[VllmRunner], model: str, ) -> None: pixel_values = torch.full((6, 512, 512), 1.0, dtype=torch.float16) location_coords = torch.full((1, 2), 1.0, dtype=torch.float16) prompt = dict( prompt_token_ids=[1], multi_modal_data={ "image": { "pixel_values": pixel_values, "location_coords": location_coords, } }, ) with vllm_runner( model, runner="pooling", dtype="half", enforce_eager=True, skip_tokenizer_init=True, enable_mm_embeds=True, # Limit the maximum number of sequences to avoid the # test going OOM during the warmup run max_num_seqs=32, default_torch_num_threads=1, ) as vllm_model: vllm_output = vllm_model.llm.encode(prompt, pooling_task="plugin") assert torch.equal( torch.isnan(vllm_output[0].outputs.data).any(), torch.tensor(False) )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/test_terratorch.py", "license": "Apache License 2.0", "lines": 45, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/transformers_utils/configs/midashenglm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright 2025 Horizon team, Xiaomi MiLM Plus. # Copyright 2024 The Qwen team. # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from transformers import PretrainedConfig from transformers.models.qwen2_5_omni.configuration_qwen2_5_omni import ( Qwen2_5OmniTextConfig, ) class DashengConfig(PretrainedConfig): model_type = "midashenglm_dasheng_encoder" def __init__( self, embed_dim: int = 768, outputdim: int = 527, patch_size: int | tuple[int, int] = 16, patch_stride: int | tuple[int, int] = 16, input_channels: int = 1, target_length: int = 1012, depth: int = 12, num_heads: int = 12, mlp_ratio: float = 4.0, qkv_bias: bool = True, init_values: float | None = None, drop_rate: float = 0.0, attn_drop_rate: float = 0.0, f_min: float = 0.0, f_max: float = 8000.0, center: bool = True, win_length: int = 512, hop_length: int = 160, sample_rate: int = 16000, n_fft: int = 512, n_mels: int = 64, **kwargs, ): self.embed_dim = embed_dim self.outputdim = outputdim self.patch_size = patch_size self.patch_stride = patch_stride self.input_channels = input_channels self.target_length = target_length self.depth = depth self.num_heads = num_heads self.mlp_ratio = mlp_ratio self.qkv_bias = qkv_bias self.init_values = init_values self.drop_rate = drop_rate self.attn_drop_rate = attn_drop_rate self.f_min = f_min self.f_max = f_max self.center = center self.win_length = win_length self.hop_length = hop_length self.sample_rate = sample_rate self.n_fft = n_fft self.n_mels = n_mels super().__init__(**kwargs) class MiDashengLMConfig(PretrainedConfig): model_type = "midashenglm" def __init__( self, audio_encoder_config: dict | None = None, subsample_factor: int = 5, text_config: dict | None = None, audio_token_id: int | None = None, **kwargs, ): self.audio_encoder_config = DashengConfig(**(audio_encoder_config or {})) self.subsample_factor = subsample_factor self.text_config = ( Qwen2_5OmniTextConfig(**text_config) if text_config else Qwen2_5OmniTextConfig() ) self.text_config.rope_parameters = None # uses_mrope is false self.audio_token_id = audio_token_id super().__init__(**kwargs)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/configs/midashenglm.py", "license": "Apache License 2.0", "lines": 96, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:examples/offline_inference/logits_processor/custom_req.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """This example demonstrates wrapping a request-level logits processor to be compatible with vLLM's batch-level logits processing For demo purposes, a dummy logits processor is employed which, if `target_token` is passed as a keyword argument to `SamplingParams.extra_args`, will mask out all tokens except `target_token`. This logits processor can be applied to a vector of logits associated with a single decode step for a single request. The logits processor cannot be applied to a request which does not pass in a `target_token` custom argument. The request-level dummy logits processor is wrapped to create a batch-level logits processor, which can apply the logits processor to output logits from all requests in the persistent batch in a given decode step. For requests which do not provide a `target_token` argument, the corresponding row of `logits` will not be modified. A batch is constructed with `temperature=0.0` and 50% of requests specifying `target_token`, and for these requests - and *only* these requests - we expect the `target_token` to be decoded in each step, yielding an output similar to that shown below: Generated Outputs: ------------------------------------------------------------ Prompt: 'Hello, my name is' Output: " ' ' ' ' ' ' ' ' ' ' ' ' ' ' ' '" ------------------------------------------------------------ Prompt: 'The president of the United States is' Output: " not a racist. He is a racist.\nHe's a racist because he" ------------------------------------------------------------ Prompt: 'The capital of France is' Output: ' also also also also also also also also also also also also also also also also' ------------------------------------------------------------ Prompt: 'The future of AI is' Output: ' in the hands of the people.\n\nThe future of AI is in the' ------------------------------------------------------------ """ from typing import Any import torch from vllm import LLM, SamplingParams from vllm.logger import init_logger from vllm.v1.sample.logits_processor import ( AdapterLogitsProcessor, RequestLogitsProcessor, ) logger = init_logger(__name__) class DummyPerReqLogitsProcessor: """The request-level logits processor masks out all logits except the token id identified by `target_token`""" def __init__(self, target_token: int) -> None: """Specify `target_token`""" self.target_token = target_token def __call__( self, output_ids: list[int], logits: torch.Tensor, ) -> torch.Tensor: val_to_keep = logits[self.target_token].item() logits[:] = float("-inf") logits[self.target_token] = val_to_keep return logits class WrappedPerReqLogitsProcessor(AdapterLogitsProcessor): """Example of wrapping a fake request-level logit processor to create a batch-level logits processor""" @classmethod def validate_params(cls, params: SamplingParams): target_token: Any | None = params.extra_args and params.extra_args.get( "target_token" ) if target_token is not None and not isinstance(target_token, int): raise ValueError(f"target_token value {target_token} is not int") def is_argmax_invariant(self) -> bool: return False def new_req_logits_processor( self, params: SamplingParams, ) -> RequestLogitsProcessor | None: """This method returns a new request-level logits processor, customized to the `target_token` value associated with a particular request. Returns None if the logits processor should not be applied to the particular request. To use the logits processor the request must have a "target_token" custom argument with an integer value. Args: params: per-request sampling params Returns: `Callable` request logits processor, or None """ target_token: Any | None = params.extra_args and params.extra_args.get( "target_token" ) if target_token is None: return None return DummyPerReqLogitsProcessor(target_token) # Sample prompts. prompts = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] # Create a mixture of requests which do and don't utilize the dummy logitproc sampling_params_list = [ SamplingParams(temperature=0.0, extra_args={"target_token": 128}), SamplingParams(temperature=0.0), SamplingParams(temperature=0.0, extra_args={"target_token": 67}), SamplingParams(temperature=0.0), ] def main(): # Create an LLM. llm = LLM( model="facebook/opt-125m", logits_processors=[WrappedPerReqLogitsProcessor], ) # Generate texts from the prompts. # The output is a list of RequestOutput objects # that contain the prompt, generated text, and other information. outputs = llm.generate(prompts, sampling_params_list) # Print the outputs. print("\nGenerated Outputs:\n" + "-" * 60) for output in outputs: prompt = output.prompt generated_text = output.outputs[0].text print(f"Prompt: {prompt!r}") print(f"Output: {generated_text!r}") print("-" * 60) if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "examples/offline_inference/logits_processor/custom_req.py", "license": "Apache License 2.0", "lines": 125, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:examples/offline_inference/logits_processor/custom_req_init.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """This example demonstrates a special case of wrapping a request-level logits processor, namely the case where it is necessary to utilize engine config or environment info passed to the constructor. The subclass must override the wrapper base class `__init__()` method to access the engine config, the device identifier, or the flag which indicates whether pinned memory is available. For demo purposes, a request-level dummy logits processor is employed which causes the same token (`target_token`) to be decoded in each step. The request-level dummy logits processor is wrapped to create a batch-level logits processor, which can apply the logits processor to output logits from all requests in the persistent batch in a given decode step. The wrapped dummy logits processor below models a scenario where we must disable the logits processor on non-"cuda" platforms. The wrapper base class `__init__()` is overridden in order to check this condition and set a flag. A batch is constructed with `temperature=0.0` and 50% of requests specifying `target_token`, and for these requests - and *only* these requests - we expect that on a "cuda" device the output will look something like: Generated Outputs: ------------------------------------------------------------ Prompt: 'Hello, my name is' Output: " ' ' ' ' ' ' ' ' ' ' ' ' ' ' ' '" ------------------------------------------------------------ Prompt: 'The president of the United States is' Output: " not a racist. He is a racist.\nHe's a racist because he" ------------------------------------------------------------ Prompt: 'The capital of France is' Output: ' also also also also also also also also also also also also also also also also' ------------------------------------------------------------ Prompt: 'The future of AI is' Output: ' in the hands of the people.\n\nThe future of AI is in the' ------------------------------------------------------------ which indicates that the logits processor is running. However, on a non-"cuda" device, the first and third requests would not repeat the same token. """ import torch from vllm import LLM, SamplingParams from vllm.config import VllmConfig from vllm.logger import init_logger from vllm.v1.sample.logits_processor import ( AdapterLogitsProcessor, RequestLogitsProcessor, ) logger = init_logger(__name__) class DummyPerReqLogitsProcessor: """The request-level logits processor masks out all logits except the token id identified by `target_token`""" def __init__(self, target_token: int) -> None: """Specify `target_token`""" self.target_token = target_token def __call__( self, output_ids: list[int], logits: torch.Tensor, ) -> torch.Tensor: val_to_keep = logits[self.target_token].item() logits[:] = float("-inf") logits[self.target_token] = val_to_keep return logits class WrappedPerReqLogitsProcessor(AdapterLogitsProcessor): """Example of overriding the wrapper class `__init__()` in order to utilize info about the device type""" @classmethod def validate_params(cls, params: SamplingParams): target_token = params.extra_args and params.extra_args.get("target_token") if target_token is not None and not isinstance(target_token, int): raise ValueError( f"`target_token` has to be an integer, got {target_token}." ) def __init__( self, vllm_config: VllmConfig, device: torch.device, is_pin_memory: bool ): super().__init__(vllm_config, device, is_pin_memory) self.is_cuda = device.type == "cuda" def is_argmax_invariant(self) -> bool: return False def new_req_logits_processor( self, params: SamplingParams, ) -> RequestLogitsProcessor | None: """This method returns a new request-level logits processor, customized to the `target_token` value associated with a particular request. Returns None if the logits processor should not be applied to the particular request. To use the logits processor the request must have a "target_token" custom argument with an integer value, and the device must be "cuda"-type Args: params: per-request sampling params Returns: `Callable` request logits processor, or None """ if ( not self.is_cuda or ( target_token := params.extra_args and params.extra_args.get("target_token") ) is None ): return None return DummyPerReqLogitsProcessor(target_token) # Sample prompts. prompts = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] # Create a mixture of requests which do and don't utilize the dummy logitproc sampling_params_list = [ SamplingParams(temperature=0.0, extra_args={"target_token": 128}), SamplingParams(temperature=0.0), SamplingParams(temperature=0.0, extra_args={"target_token": 67}), SamplingParams(temperature=0.0), ] def main(): # Create an LLM. llm = LLM( model="facebook/opt-125m", logits_processors=[WrappedPerReqLogitsProcessor], ) # Generate texts from the prompts. # The output is a list of RequestOutput objects # that contain the prompt, generated text, and other information. outputs = llm.generate(prompts, sampling_params_list) # Print the outputs. print("\nGenerated Outputs:\n" + "-" * 60) for output in outputs: prompt = output.prompt generated_text = output.outputs[0].text print(f"Prompt: {prompt!r}") print(f"Output: {generated_text!r}") print("-" * 60) if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "examples/offline_inference/logits_processor/custom_req_init.py", "license": "Apache License 2.0", "lines": 136, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/logprobs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import itertools from collections.abc import Iterable, Iterator, MutableSequence from dataclasses import dataclass, field from typing import overload # We use dataclass for now because it is used for # openai server output, and msgspec is not serializable. # TODO(sang): Fix it. @dataclass class Logprob: """Infos for supporting OpenAI compatible logprobs and token ranks. Attributes: logprob: The logprob of chosen token rank: The vocab rank of chosen token (>=1) decoded_token: The decoded chosen token index """ logprob: float rank: int | None = None decoded_token: str | None = None LogprobsOnePosition = dict[int, Logprob] @dataclass class FlatLogprobs(MutableSequence[LogprobsOnePosition | None]): """ Flat logprobs of a request into multiple primitive type lists. Compared to list[dict[int, Logprob]], this data structure reduced GC overhead significantly. As it flattened logprob information for all positions and ranks in to multiple primitive type lists (i.e. logprobs, token_ids, ranks per token_ids, decoded_tokens). So regardless of the sequence length and top_logprobs setup, FlatLogprobs would only introduce a constant amount of objects. As each position might contains different amount of ranks, start_indices_per_position would be used to access the logprob ranges for different positions. NOTE: To reduce the migration overhead and improve backward compatibility, we support the key Sequence APIs of list, so it could act as list[LogprobsOnePosition] """ # Start / end indices to indicate the range of logprobs for each position. start_indices: list[int] = field(default_factory=list) end_indices: list[int] = field(default_factory=list) # Flatten Logprob information for (each position, rank). # For position <i>, the logprobs are ranged # from self.start_indices[i] to self.end_indices[i] (exclusive). token_ids: list[int] = field(default_factory=list) logprobs: list[float] = field(default_factory=list) ranks: list[int | None] = field(default_factory=list) decoded_tokens: list[str | None] = field(default_factory=list) def append(self, logprobs_one_position: LogprobsOnePosition | None) -> None: """Appends the container with logprobs for the next position""" self.start_indices.append(len(self.logprobs)) if logprobs_one_position: for token_id, logprob in logprobs_one_position.items(): self.token_ids.append(token_id) self.logprobs.append(logprob.logprob) self.ranks.append(logprob.rank) self.decoded_tokens.append(logprob.decoded_token) self.end_indices.append(len(self.logprobs)) def append_fast( self, token_ids: list[int], logprobs: list[float], ranks: itertools.chain[int], decoded_tokens: Iterable[str | None], ) -> None: """ Appends logprobs for the next position without creating the intermediate logprob dictionary. """ self.start_indices.append(len(self.logprobs)) for token_id, logprob, rank, decoded_token in zip( token_ids, logprobs, ranks, decoded_tokens ): self.token_ids.append(token_id) self.logprobs.append(logprob) self.ranks.append(rank) self.decoded_tokens.append(decoded_token) self.end_indices.append(len(self.logprobs)) def extend(self, logprobs_multi_positions) -> None: """Extends the container with logprobs for the next multiple positions""" for logprobs_one_position in logprobs_multi_positions: self.append(logprobs_one_position) def __len__(self) -> int: """Gets number of positions stored in the container""" return len(self.start_indices) @overload def __getitem__(self, position: int) -> LogprobsOnePosition: ... @overload def __getitem__(self, s: slice, /) -> "FlatLogprobs": ... def __getitem__(self, index: int | slice): """Extracts logprobs of a given position or slice""" if isinstance(index, int): return { self.token_ids[i]: Logprob( logprob=self.logprobs[i], rank=self.ranks[i], decoded_token=self.decoded_tokens[i], ) for i in range(self.start_indices[index], self.end_indices[index]) } elif isinstance(index, slice): min_index = self.start_indices[index][0] max_index = self.end_indices[index][-1] return FlatLogprobs( # Shift updated start_indices and end_indices to # be 0-indexed start_indices=[i - min_index for i in self.start_indices[index]], end_indices=[i - min_index for i in self.end_indices[index]], token_ids=self.token_ids[min_index:max_index], logprobs=self.logprobs[min_index:max_index], ranks=self.ranks[min_index:max_index], decoded_tokens=self.decoded_tokens[min_index:max_index], ) else: raise TypeError(f"Invalid index type: {type(index)}") def __setitem__(self, item, value) -> None: raise TypeError("Cannot set logprobs in FlatLogprobs") def __delitem__(self, item) -> None: raise TypeError("Cannot delete logprobs from FlatLogprobs") def insert(self, index: int, value: dict[int, Logprob] | None) -> None: raise TypeError("Cannot insert logprobs to FlatLogprobs") def __iter__(self) -> Iterator[LogprobsOnePosition]: """ Iterates the container and yields LogprobsOnePosition for each position. """ for i in range(0, len(self.start_indices)): yield self.__getitem__(i) # {token_id -> logprob} per each sequence group. None if the corresponding # sequence group doesn't require prompt logprob. PromptLogprobs = FlatLogprobs | list[LogprobsOnePosition | None] # {token_id -> logprob} for each sequence group. SampleLogprobs = FlatLogprobs | list[LogprobsOnePosition] def create_prompt_logprobs(flat_logprobs: bool) -> PromptLogprobs: """Creates a container to store prompt logprobs for a request""" logprobs: PromptLogprobs = FlatLogprobs() if flat_logprobs else [] # NOTE: logprob of first prompt token is None. logprobs.append(None) return logprobs def create_sample_logprobs(flat_logprobs: bool) -> SampleLogprobs: """Creates a container to store decode logprobs for a request""" return FlatLogprobs() if flat_logprobs else [] def append_logprobs_for_next_position( request_logprobs: PromptLogprobs | SampleLogprobs, token_ids: list[int], logprobs: list[float], decoded_tokens: Iterable[str | None], rank: int, num_logprobs: int, ) -> None: """Appends logprobs for the next position""" if num_logprobs == -1: num_logprobs = len(logprobs) # We do not need a special case for the sampled token # being in the topk, since inserting duplicated data # into a dictionary twice is the same as doing it once. topk_ranks = range(1, num_logprobs + 1) ranks = itertools.chain((rank,), topk_ranks) if isinstance(request_logprobs, FlatLogprobs): request_logprobs.append_fast(token_ids, logprobs, ranks, decoded_tokens) else: request_logprobs.append( { token_id: Logprob( logprob=logprob, rank=rank, decoded_token=token, ) for token_id, logprob, rank, token in zip( token_ids, logprobs, ranks, decoded_tokens ) } )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/logprobs.py", "license": "Apache License 2.0", "lines": 173, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/keye_vl1_5.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import itertools from collections.abc import Mapping, Sequence from functools import partial from typing import Annotated, Any, Literal, TypeAlias import numpy as np import torch import torch.nn as nn from einops import rearrange from transformers import PretrainedConfig from transformers.activations import GELUActivation from transformers.feature_extraction_utils import BatchFeature from vllm.config import VllmConfig from vllm.logger import init_logger from vllm.model_executor.layers.linear import ColumnParallelLinear, RowParallelLinear from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import ( ImageItem, ModalityData, MultiModalFeatureSpec, MultiModalFieldConfig, MultiModalKwargsItems, VideoItem, ) from vllm.multimodal.parse import ( DictEmbeddingItems, ModalityDataItems, MultiModalDataItems, MultiModalDataParser, ) from vllm.multimodal.processing import ( PromptReplacement, PromptUpdate, PromptUpdateDetails, ) from vllm.utils.tensor_schema import TensorSchema, TensorShape from .interfaces import SupportsLoRA, SupportsMRoPE, SupportsMultiModal, SupportsPP from .keye import ( BaseKeyeModule, BaseMultiModalProcessor, KeyeBaseDummyInputsBuilder, KeyeProcessingInfo, ) logger = init_logger(__name__) def split_thw(grid_thw: torch.Tensor) -> torch.Tensor: """ Split grid_thw in t dimension. Args: grid_thw: [N, 3] tensor of [t, h, w] Returns: [Σt, 3] tensor where each row is [1, h, w] Example: >>> grid_thw = torch.tensor([[2, 3, 4], [1, 5, 6]]) >>> split_thw(grid_thw) tensor([[1, 3, 4], [1, 3, 4], [1, 5, 6]]) """ t = grid_thw[:, 0] h_w = grid_thw[:, 1:] ones = torch.ones_like(h_w[:, :1]) return torch.cat([ones, h_w], dim=1).repeat_interleave(t, dim=0) def get_num_patches( grid_thw: torch.Tensor, num_frames: list[int] | torch.Tensor ) -> list[int]: """ Return num_patches per video. Args: grid_thw: Tensor with shape [N, 3] containing temporal, height, width dimensions num_frames: List or tensor indicating the number of frames per video Returns: List of ints representing the number of patches for each video Examples: >>> # Suppose there are 2 videos with a total of 3 grids >>> grid_thw = torch.tensor( ... [ ... [2, 2, 2], # grid 0: 2*2*2=8 patches ... [2, 2, 2], # grid 1: 2*2*2=8 patches ... [1, 1, 1], ... ] ... ) # grid 2: 1*1*1=1 patches >>> num_frames = [2, 1] # The first video contains 2 grids, the second contains 1 grid. >>> get_num_patches(grid_thw, num_frames) tensor([16, 1]) # Total patches for first video: 8+8=16, second video: 1. """ assert len(grid_thw.shape) == 2 if isinstance(num_frames, torch.Tensor): num_frames = num_frames.clone().tolist() num_grids_per_frame = grid_thw.prod(dim=1) start_idx_per_video = [0, *itertools.accumulate(num_frames)] num_patches = [ num_grids_per_frame[start_idx_per_video[i] : start_idx_per_video[i + 1]].sum() for i in range(len(num_frames)) ] return ( torch.stack(num_patches) if num_patches else torch.zeros(0, dtype=grid_thw.dtype, device=grid_thw.device) ) class KeyeVL1_5ImagePixelInputs(TensorSchema): """ Dimensions: - bnp: Batch size * Number of patches - c: Number of channels - ps: Patch size - ni: Number of images - g: Grid dimensions (3 for t, h, w) """ type: Literal["pixel_values"] pixel_values: Annotated[ torch.Tensor, TensorShape("bnp", 3, "ps", "ps", dynamic_dims={"bnp"}) ] image_grid_thw: Annotated[torch.Tensor, TensorShape("ni", 3)] class KeyeVL1_5ImageEmbeddingInputs(TensorSchema): """ Dimensions: - nf: Number of image features - hs: Hidden size (must match the hidden size of language model backbone) - ni: Number of images - g: Grid dimensions (3 for t, h, w) """ type: Literal["image_embeds"] image_embeds: Annotated[torch.Tensor, TensorShape("nf", "hs")] image_grid_thw: Annotated[torch.Tensor, TensorShape("ni", 3)] KeyeVL1_5ImageInputs: TypeAlias = ( KeyeVL1_5ImagePixelInputs | KeyeVL1_5ImageEmbeddingInputs ) class KeyeVL1_5VideoPixelInputs(TensorSchema): """ Dimensions: - bnp: Batch size * Number of patches - c: Number of channels - ps: Patch size - ni: Number of images - g: Grid dimensions (3 for t, h, w) """ type: Literal["pixel_values_videos"] pixel_values_videos: Annotated[ torch.Tensor, TensorShape("bnp", 3, "ps", "ps", dynamic_dims={"bnp"}) ] video_grid_thw: Annotated[torch.Tensor, TensorShape("nv", 3)] num_frames: torch.Tensor class KeyeVL1_5VideoEmbeddingInputs(TensorSchema): """ Dimensions: - nf: Number of video features - hs: Hidden size (must match the hidden size of language model backbone) - nv: Number of videos - g: Grid dimensions (3 for t, h, w) """ type: Literal["video_embeds"] video_embeds: Annotated[torch.Tensor, TensorShape("nf", "hs")] video_grid_thw: Annotated[torch.Tensor, TensorShape("nv", 3)] num_frames: torch.Tensor KeyeVL1_5VideoInputs: TypeAlias = ( KeyeVL1_5VideoPixelInputs | KeyeVL1_5VideoEmbeddingInputs ) class KeyeVL1_5Projector(nn.Module): def __init__( self, text_config: PretrainedConfig, vision_config: PretrainedConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.text_config = text_config self.vision_config = vision_config self.merge_kernel_size = (2, 2) self.hidden_size = ( self.vision_config.hidden_size * self.merge_kernel_size[0] * self.merge_kernel_size[1] ) self.pre_norm = torch.nn.LayerNorm(self.hidden_size, eps=1e-05) self.act = GELUActivation() self.linear_1 = ColumnParallelLinear( self.hidden_size, self.hidden_size, bias=True, quant_config=quant_config, prefix=f"{prefix}.linear_1", ) self.linear_2 = RowParallelLinear( self.hidden_size, self.text_config.hidden_size, bias=True, quant_config=quant_config, prefix=f"{prefix}.linear_2", ) def forward( self, image_features: torch.Tensor | tuple[torch.Tensor] | list[torch.Tensor], image_grid_thw: list[tuple[int, int, int]], ) -> torch.Tensor | list[torch.Tensor]: m1, m2 = self.merge_kernel_size if isinstance(image_features, (list, tuple)): processed_features = list() for image_feature, image_grid in zip(image_features, image_grid_thw): t, h, w = image_grid image_feature = rearrange( image_feature, "(t h p1 w p2) d -> (t h w) (p1 p2 d)", t=t, h=h // m1, p1=m1, w=w // m2, p2=m2, ) image_feature = self.pre_norm(image_feature) hidden_states, _ = self.linear_1(image_feature) hidden_states = self.act(hidden_states) hidden_states, _ = self.linear_2(hidden_states) processed_features.append(hidden_states) return processed_features dims = image_features.shape[:-1] dim = image_features.shape[-1] image_features = image_features.view(np.prod(dims), dim) hidden_states = self.pre_norm(image_features.view(-1, self.hidden_size)) hidden_states = self.linear_1(hidden_states) hidden_states = self.act(hidden_states) hidden_states = self.linear_2(hidden_states) return hidden_states.view(*dims, -1) def _keye_field_config( hf_inputs: Mapping[str, torch.Tensor], ): image_grid_thw = hf_inputs.get( "image_grid_thw", torch.empty((0, 3), dtype=torch.int64) ) image_grid_sizes = image_grid_thw.prod(-1) video_grid_thw = hf_inputs.get( "video_grid_thw", torch.empty((0, 3), dtype=torch.int64) ) video_grid_thw = split_thw(video_grid_thw) num_frames = hf_inputs.get("num_frames", video_grid_thw[:, 0]).clone().tolist() video_num_patches = get_num_patches(video_grid_thw, num_frames) video_num_grids = [] if len(num_frames) > 0: i = 0 j = 1 cur_frames = num_frames[i] for t, _, _ in video_grid_thw.tolist(): cur_frames -= t if cur_frames == 0: video_num_grids.append(j) i += 1 if i < len(num_frames): cur_frames = num_frames[i] j = 1 else: j += 1 video_num_grids = torch.tensor(video_num_grids) return dict( pixel_values=MultiModalFieldConfig.flat_from_sizes("image", image_grid_sizes), image_embeds=MultiModalFieldConfig.flat_from_sizes("image", image_grid_sizes), image_grid_thw=MultiModalFieldConfig.batched("image"), pixel_values_videos=MultiModalFieldConfig.flat_from_sizes( "video", video_num_patches ), video_embeds=MultiModalFieldConfig.flat_from_sizes("video", video_num_patches), video_grid_thw=MultiModalFieldConfig.flat_from_sizes("video", video_num_grids), num_frames=MultiModalFieldConfig.batched("video"), ) class KeyeVL1_5MultiModalDataParser(MultiModalDataParser): def _parse_image_data( self, data: dict[str, torch.Tensor] | ModalityData[ImageItem], ) -> ModalityDataItems[Any, Any] | None: if isinstance(data, dict): return DictEmbeddingItems( data, modality="image", required_fields={ "image_embeds", "image_grid_thw", }, fields_factory=_keye_field_config, ) return super()._parse_image_data(data) def _parse_video_data( self, data: dict[str, torch.Tensor] | ModalityData[VideoItem], ) -> ModalityDataItems[Any, Any] | None: if isinstance(data, dict): return DictEmbeddingItems( data, modality="video", required_fields={ "video_embeds", "video_grid_thw", }, fields_factory=_keye_field_config, ) return super()._parse_video_data(data) class KeyeVL1_5ProcessingInfo(KeyeProcessingInfo): def get_data_parser(self): return KeyeVL1_5MultiModalDataParser( expected_hidden_size=self._get_expected_hidden_size(), ) def get_max_frame_per_video(self) -> int: return 2048 def get_supported_mm_limits( self, ) -> Mapping[str, int | None]: return {"image": None, "video": 1} class KeyeVL1_5MultiModalProcessor(BaseMultiModalProcessor[KeyeVL1_5ProcessingInfo]): def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, Any], out_mm_kwargs: MultiModalKwargsItems, ) -> Sequence[PromptUpdate]: hf_processor = self.info.get_hf_processor(**hf_processor_mm_kwargs) image_processor = self.info.get_image_processor(**hf_processor_mm_kwargs) tokenizer = self.info.get_tokenizer() vocab = tokenizer.get_vocab() image_token_id = vocab[hf_processor.image_token] video_token_id = vocab[hf_processor.video_token] placeholder = {"image": image_token_id, "video": video_token_id} merge_length = image_processor.merge_size**2 out_mm_kwargs_data = out_mm_kwargs.get_data() frame_types: list[torch.Tensor] = hf_processor_mm_kwargs.get( "frame_types", None ) timestamps: list[torch.Tensor] = hf_processor_mm_kwargs.get("timestamps", None) num_videos = mm_items.get_count("video", strict=False) if frame_types is None: frame_types = [None] * num_videos assert len(frame_types) == num_videos, ( f"Number of frame_types={len(frame_types)} " f"doesn't equal to number of videos={num_videos}" ) if timestamps is None: timestamps = [None] * num_videos assert len(timestamps) == num_videos, ( f"Number of timestamps={len(timestamps)} " f"doesn't equal to number of videos={num_videos}" ) video_grid_thw = out_mm_kwargs_data.get( "video_grid_thw", torch.empty((0, 3), dtype=torch.int64) ) num_frames = out_mm_kwargs_data.get( "num_frames", torch.tensor([], dtype=torch.int64) ) assert len(num_frames) == num_videos, ( f"Size of num_frames={len(num_frames)} " f"doesn't equal to number of videos={num_videos}" ) video_grid_hws = split_thw(video_grid_thw) assert int(num_frames.sum().tolist()) == video_grid_hws.shape[0], ( f"The first dimension of `video_grid_hws`={video_grid_hws.shape[0]}" f"doesn't equal to num of frames." ) cu_seqlens = torch.cumsum(torch.tensor([0] + num_frames.tolist()), dim=-1) def get_replacement_keye(item_idx: int, modality: str): """ Args: item_idx(int): The item index of modality to replace modality(str): The modality """ if modality == "image": out_item = out_mm_kwargs[modality][item_idx] grid_thw = out_item[f"{modality}_grid_thw"].data assert isinstance(grid_thw, torch.Tensor) num_tokens = int(grid_thw.prod()) // merge_length return [image_token_id] * num_tokens elif modality == "video": placeholders = [] video_timestamps = timestamps[item_idx] video_frame_types = frame_types[item_idx] grid_thw = video_grid_hws[ cu_seqlens[item_idx] : cu_seqlens[item_idx + 1] ] nframes = grid_thw.shape[0] if video_timestamps is None: video_timestamps = [""] * nframes else: video_timestamps = [format(ts, ".1f") for ts in video_timestamps] if video_frame_types is None: video_frame_types = [0] * nframes for i, sub_thw in enumerate(grid_thw): s = f"{hf_processor.frame_token}{video_timestamps[i]}" if video_frame_types[i] == 1: s += hf_processor.fast_start placeholders.extend(tokenizer.encode(s)) num_frame_tokens = int(sub_thw.prod()) // merge_length placeholders.extend([video_token_id] * num_frame_tokens) if video_frame_types[i] == 1: placeholders.append(vocab[hf_processor.fast_end]) return PromptUpdateDetails.select_token_id( placeholders, embed_token_id=video_token_id ) else: raise ValueError(f"Unsupported modality {modality}") return [ PromptReplacement( modality=modality, target=[placeholder[modality]], replacement=partial(get_replacement_keye, modality=modality), ) for modality in ("image", "video") ] def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: return _keye_field_config(hf_inputs) class KeyeVL1_5DummyInputsBuilder( KeyeBaseDummyInputsBuilder[KeyeVL1_5ProcessingInfo] ): ... @MULTIMODAL_REGISTRY.register_processor( KeyeVL1_5MultiModalProcessor, info=KeyeVL1_5ProcessingInfo, dummy_inputs=KeyeVL1_5DummyInputsBuilder, ) class KeyeVL1_5ForConditionalGeneration( BaseKeyeModule, SupportsMultiModal, SupportsLoRA, SupportsPP, SupportsMRoPE ): def _build_projector( self, text_config: PretrainedConfig, vision_config: PretrainedConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> nn.Module: return KeyeVL1_5Projector(text_config, vision_config, quant_config, prefix) def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): config: PretrainedConfig = vllm_config.model_config.hf_config self.merge_size = config.vision_config.spatial_merge_size super().__init__(vllm_config=vllm_config, prefix=prefix) def _parse_and_validate_image_input( self, **kwargs: object ) -> KeyeVL1_5ImageInputs | None: pixel_values = kwargs.pop("pixel_values", None) image_embeds = kwargs.pop("image_embeds", None) image_grid_thw = kwargs.pop("image_grid_thw", None) if pixel_values is None and image_embeds is None: return None if pixel_values is not None: return KeyeVL1_5ImagePixelInputs( type="pixel_values", pixel_values=pixel_values, image_grid_thw=image_grid_thw, ) if image_embeds is not None: return KeyeVL1_5ImageEmbeddingInputs( type="image_embeds", image_embeds=image_embeds, image_grid_thw=image_grid_thw, ) def _parse_and_validate_video_input( self, **kwargs: object ) -> KeyeVL1_5VideoInputs | None: pixel_values_videos = kwargs.pop("pixel_values_videos", None) video_embeds = kwargs.pop("video_embeds", None) video_grid_thw = kwargs.pop("video_grid_thw", None) num_frames = kwargs.pop("num_frames", None) if pixel_values_videos is None and video_embeds is None: return None if pixel_values_videos is not None: return KeyeVL1_5VideoPixelInputs( type="pixel_values_videos", pixel_values_videos=pixel_values_videos, video_grid_thw=video_grid_thw, num_frames=num_frames, ) if video_embeds is not None: return KeyeVL1_5VideoEmbeddingInputs( type="video_embeds", video_embeds=video_embeds, video_grid_thw=video_grid_thw, num_frames=num_frames, ) def _process_video_input( self, video_input: KeyeVL1_5VideoInputs ) -> tuple[torch.Tensor, ...]: video_type = video_input["type"] video_grid_thw = split_thw(video_input["video_grid_thw"]) pixel_values_videos = video_input.get("pixel_values_videos", None) video_embeds = self._process_video_embeds( video_type, video_grid_thw, pixel_values_videos ) video_embeds = torch.concat(video_embeds, dim=0) num_frames = video_input["num_frames"].clone().tolist() num_patches = get_num_patches(video_grid_thw, num_frames).tolist() patch_cu_seqlens = torch.cumsum( torch.tensor([0] + num_patches).detach().clone(), dim=-1 ) patch_cu_seqlens = torch.div( patch_cu_seqlens, self.merge_size**2, rounding_mode="floor" ) new_video_embeds = [] for idx in range(patch_cu_seqlens.shape[0] - 1): start = patch_cu_seqlens[idx] end = patch_cu_seqlens[idx + 1] new_video_embeds.append(video_embeds[start:end]) return tuple(new_video_embeds) def get_mrope_input_positions( self, input_tokens: list[int], mm_features: list[MultiModalFeatureSpec], ) -> tuple[torch.Tensor, int]: kwargs = MultiModalFeatureSpec.gather_kwargs( mm_features, {"image_grid_thw", "video_grid_thw"}, ) image_grid_thw = [item.tolist() for item in kwargs.get("image_grid_thw", [])] video_grid_thw = [item.tolist() for item in kwargs.get("video_grid_thw", [])] if isinstance(video_grid_thw, list) and len(video_grid_thw) > 0: video_grid_thw = video_grid_thw[0] def split_thw(grid_thw: torch.Tensor | list[int]) -> list[list[int]]: """ Split grid_thw along the t dimension. Args: grid_thw: shape [N, 3] tensor or nested list of [t, h, w]. Returns: List of [1, h, w] rows, repeated t times for each original row. """ if isinstance(grid_thw, list): grid_thw = torch.tensor(grid_thw, dtype=torch.long) if grid_thw.numel() == 0: return [] t, hw = grid_thw[:, 0], grid_thw[:, 1:] ones = torch.ones_like(hw[:, :1]) # [N,1] out = torch.cat([ones, hw], dim=1).repeat_interleave(t, dim=0) return out.tolist() video_grid_thw = split_thw(video_grid_thw) hf_config = self.config image_token_id = hf_config.image_token_id video_token_id = hf_config.video_token_id spatial_merge_size = hf_config.vision_config.spatial_merge_size image_nums = len(image_grid_thw) frame_nums = len(video_grid_thw) llm_pos_ids_list: list = [] st = 0 remain_images, remain_frames = image_nums, frame_nums image_index, video_index = 0, 0 for _ in range(image_nums + frame_nums): if remain_images > 0: try: ed_image = input_tokens.index(image_token_id, st) except ValueError: ed_image = len(input_tokens) + 1 else: ed_image = len(input_tokens) + 1 if remain_frames > 0: try: ed_video = input_tokens.index(video_token_id, st) except ValueError: ed_video = len(input_tokens) + 1 else: ed_video = len(input_tokens) + 1 if ed_image < ed_video: t, h, w = image_grid_thw[image_index] image_index += 1 remain_images -= 1 ed = ed_image else: t, h, w = video_grid_thw[video_index] video_index += 1 remain_frames -= 1 ed = ed_video llm_grid_t, llm_grid_h, llm_grid_w = ( t, h // spatial_merge_size, w // spatial_merge_size, ) text_len = ed - st st_idx = llm_pos_ids_list[-1].max() + 1 if len(llm_pos_ids_list) > 0 else 0 llm_pos_ids_list.append( torch.arange(text_len).view(1, -1).expand(3, -1) + st_idx ) t_index = ( ( torch.arange(llm_grid_t) .view(-1, 1) .expand(-1, llm_grid_h * llm_grid_w) ) .long() .flatten() ) h_index = ( torch.arange(llm_grid_h) .view(1, -1, 1) .expand(llm_grid_t, -1, llm_grid_w) .flatten() ) w_index = ( torch.arange(llm_grid_w) .view(1, 1, -1) .expand(llm_grid_t, llm_grid_h, -1) .flatten() ) llm_pos_ids_list.append( torch.stack([t_index, h_index, w_index]) + text_len + st_idx ) st = ed + llm_grid_t * llm_grid_h * llm_grid_w if st < len(input_tokens): st_idx = llm_pos_ids_list[-1].max() + 1 if len(llm_pos_ids_list) > 0 else 0 text_len = len(input_tokens) - st llm_pos_ids_list.append( torch.arange(text_len).view(1, -1).expand(3, -1) + st_idx ) llm_positions = torch.cat(llm_pos_ids_list, dim=1).reshape(3, -1) mrope_position_delta = (llm_positions.max() + 1 - len(input_tokens)).item() return llm_positions, mrope_position_delta
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/keye_vl1_5.py", "license": "Apache License 2.0", "lines": 614, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/plugins/prithvi_io_processor_plugin/prithvi_io_processor/prithvi_processor.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import base64 import datetime import os import tempfile import urllib.request from collections.abc import Sequence from typing import Any import albumentations import numpy as np import rasterio import regex as re import torch from einops import rearrange from terratorch.datamodules import Sen1Floods11NonGeoDataModule from vllm.config import VllmConfig from vllm.inputs.data import PromptType from vllm.logger import init_logger from vllm.outputs import PoolingRequestOutput from vllm.plugins.io_processors.interface import IOProcessor from vllm.renderers import BaseRenderer from .types import DataModuleConfig, ImagePrompt, ImageRequestOutput logger = init_logger(__name__) NO_DATA = -9999 NO_DATA_FLOAT = 0.0001 OFFSET = 0 PERCENTILE = 99 DEFAULT_INPUT_INDICES = [0, 1, 2, 3, 4, 5] datamodule_config: DataModuleConfig = { "bands": ["BLUE", "GREEN", "RED", "NIR_NARROW", "SWIR_1", "SWIR_2"], "batch_size": 16, "constant_scale": 0.0001, "data_root": "/dccstor/geofm-finetuning/datasets/sen1floods11", "drop_last": True, "no_data_replace": 0.0, "no_label_replace": -1, "num_workers": 8, "test_transform": [ albumentations.Resize(height=448, interpolation=1, p=1, width=448), albumentations.pytorch.ToTensorV2(transpose_mask=False, p=1.0), ], } def save_geotiff(image: torch.Tensor, meta: dict, out_format: str) -> str | bytes: """Save multi-band image in Geotiff file. Args: image: np.ndarray with shape (bands, height, width) output_path: path where to save the image meta: dict with meta info. """ if out_format == "path": # create temp file file_path = os.path.join(os.getcwd(), "prediction.tiff") with rasterio.open(file_path, "w", **meta) as dest: for i in range(image.shape[0]): dest.write(image[i, :, :], i + 1) return file_path elif out_format == "b64_json": with tempfile.NamedTemporaryFile() as tmpfile: with rasterio.open(tmpfile.name, "w", **meta) as dest: for i in range(image.shape[0]): dest.write(image[i, :, :], i + 1) file_data = tmpfile.read() return base64.b64encode(file_data) else: raise ValueError("Unknown output format") def _convert_np_uint8(float_image: torch.Tensor): image = float_image.numpy() * 255.0 image = image.astype(dtype=np.uint8) return image def read_geotiff( file_path: str | None = None, path_type: str | None = None, file_data: bytes | None = None, ) -> tuple[torch.Tensor, dict, tuple[float, float] | None]: """Read all bands from *file_path* and return image + meta info. Args: file_path: path to image file. Returns: np.ndarray with shape (bands, height, width) meta info dict """ if all([x is None for x in [file_path, path_type, file_data]]): raise Exception("All input fields to read_geotiff are None") write_to_file: bytes | None = None path: str | None = None if file_data is not None: # with tempfile.NamedTemporaryFile() as tmpfile: # tmpfile.write(file_data) # path = tmpfile.name write_to_file = file_data elif file_path is not None and path_type == "url": resp = urllib.request.urlopen(file_path) # with tempfile.NamedTemporaryFile() as tmpfile: # tmpfile.write(resp.read()) # path = tmpfile.name write_to_file = resp.read() elif file_path is not None and path_type == "path": path = file_path elif file_path is not None and path_type == "b64_json": image_data = base64.b64decode(file_path) # with tempfile.NamedTemporaryFile() as tmpfile: # tmpfile.write(image_data) # path = tmpfile.name write_to_file = image_data else: raise Exception("Wrong combination of parameters to read_geotiff") with tempfile.NamedTemporaryFile() as tmpfile: path_to_use = None if write_to_file: tmpfile.write(write_to_file) path_to_use = tmpfile.name elif path: path_to_use = path with rasterio.open(path_to_use) as src: img = src.read() meta = src.meta try: coords = src.lnglat() except Exception: # Cannot read coords coords = None return img, meta, coords def load_image( data: list[str], path_type: str, mean: list[float] | None = None, std: list[float] | None = None, indices: list[int] | None | None = None, ): """Build an input example by loading images in *file_paths*. Args: file_paths: list of file paths . mean: list containing mean values for each band in the images in *file_paths*. std: list containing std values for each band in the images in *file_paths*. Returns: np.array containing created example list of meta info for each image in *file_paths* """ imgs = [] metas = [] temporal_coords = [] location_coords = [] for file in data: # if isinstance(file, bytes): # img, meta, coords = read_geotiff(file_data=file) # else: img, meta, coords = read_geotiff(file_path=file, path_type=path_type) # Rescaling (don't normalize on nodata) img = np.moveaxis(img, 0, -1) # channels last for rescaling if indices is not None: img = img[..., indices] if mean is not None and std is not None: img = np.where(img == NO_DATA, NO_DATA_FLOAT, (img - mean) / std) imgs.append(img) metas.append(meta) if coords is not None: location_coords.append(coords) try: match = re.search(r"(\d{7,8}T\d{6})", file) if match: year = int(match.group(1)[:4]) julian_day = match.group(1).split("T")[0][4:] if len(julian_day) == 3: julian_day = int(julian_day) else: julian_day = ( datetime.datetime.strptime(julian_day, "%m%d") .timetuple() .tm_yday ) temporal_coords.append([year, julian_day]) except Exception: logger.exception("Could not extract timestamp for %s", file) imgs = np.stack(imgs, axis=0) # num_frames, H, W, C imgs = np.moveaxis(imgs, -1, 0).astype("float32") # C, num_frames, H, W imgs = np.expand_dims(imgs, axis=0) # add batch di return imgs, temporal_coords, location_coords, metas class PrithviMultimodalDataProcessor(IOProcessor[ImagePrompt, ImageRequestOutput]): indices = [0, 1, 2, 3, 4, 5] def __init__(self, vllm_config: VllmConfig, renderer: BaseRenderer): super().__init__(vllm_config, renderer) self.datamodule = Sen1Floods11NonGeoDataModule( data_root=datamodule_config["data_root"], batch_size=datamodule_config["batch_size"], num_workers=datamodule_config["num_workers"], bands=datamodule_config["bands"], drop_last=datamodule_config["drop_last"], test_transform=datamodule_config["test_transform"], ) self.img_size = 512 self.h1 = 1 self.w1 = 1 self.original_h = 512 self.original_w = 512 self.batch_size = 1 self.meta_data = None self.requests_cache: dict[str, dict[str, Any]] = {} self.indices = DEFAULT_INPUT_INDICES def parse_data(self, data: object) -> ImagePrompt: if isinstance(data, dict): return ImagePrompt(**data) raise ValueError("Prompt data should be an `ImagePrompt`") def pre_process( self, prompt: ImagePrompt, request_id: str | None = None, **kwargs, ) -> PromptType | Sequence[PromptType]: image_data = dict(prompt) if request_id: self.requests_cache[request_id] = { "out_format": image_data["out_data_format"], } input_data, temporal_coords, location_coords, meta_data = load_image( data=[image_data["data"]], indices=self.indices, path_type=image_data["data_format"], ) self.meta_data = meta_data[0] if input_data.mean() > 1: input_data = input_data / 10000 # Convert to range 0-1 self.original_h, self.original_w = input_data.shape[-2:] pad_h = (self.img_size - (self.original_h % self.img_size)) % self.img_size pad_w = (self.img_size - (self.original_w % self.img_size)) % self.img_size input_data = np.pad( input_data, ((0, 0), (0, 0), (0, 0), (0, pad_h), (0, pad_w)), mode="reflect", ) batch = torch.tensor(input_data) windows = batch.unfold(3, self.img_size, self.img_size).unfold( 4, self.img_size, self.img_size ) self.h1, self.w1 = windows.shape[3:5] windows = rearrange( windows, "b c t h1 w1 h w -> (b h1 w1) c t h w", h=self.img_size, w=self.img_size, ) # Split into batches if number of windows > batch_size num_batches = ( windows.shape[0] // self.batch_size if windows.shape[0] > self.batch_size else 1 ) windows = torch.tensor_split(windows, num_batches, dim=0) if temporal_coords: temporal_coords = torch.tensor(temporal_coords).unsqueeze(0) else: temporal_coords = None if location_coords: location_coords = torch.tensor(location_coords[0]).unsqueeze(0) else: location_coords = None prompts = [] for window in windows: # Apply standardization window = self.datamodule.test_transform( image=window.squeeze().numpy().transpose(1, 2, 0) ) window = self.datamodule.aug(window)["image"] prompts.append( { "prompt_token_ids": [1], "multi_modal_data": { "image": { "pixel_values": window.to(torch.float16)[0], "location_coords": location_coords.to(torch.float16), } }, } ) return prompts def post_process( self, model_output: Sequence[PoolingRequestOutput], request_id: str | None = None, **kwargs, ) -> ImageRequestOutput: pred_imgs_list = [] if request_id and (request_id in self.requests_cache): out_format = self.requests_cache[request_id]["out_format"] else: out_format = "b64_json" for output in model_output: y_hat = output.outputs.data.argmax(dim=0) pred = torch.nn.functional.interpolate( y_hat[None, None, ...].float(), size=self.img_size, mode="nearest", ) pred_imgs_list.append(pred) pred_imgs: torch.Tensor = torch.concat(pred_imgs_list, dim=0) # Build images from patches pred_imgs = rearrange( pred_imgs, "(b h1 w1) c h w -> b c (h1 h) (w1 w)", h=self.img_size, w=self.img_size, b=1, c=1, h1=self.h1, w1=self.w1, ) # Cut padded area back to original size pred_imgs = pred_imgs[..., : self.original_h, : self.original_w] # Squeeze (batch size 1) pred_imgs = pred_imgs[0] if not self.meta_data: raise ValueError("No metadata available for the current task") self.meta_data.update(count=1, dtype="uint8", compress="lzw", nodata=0) out_data = save_geotiff( _convert_np_uint8(pred_imgs), self.meta_data, out_format ) return ImageRequestOutput( type=out_format, format="tiff", data=out_data, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/plugins/prithvi_io_processor_plugin/prithvi_io_processor/prithvi_processor.py", "license": "Apache License 2.0", "lines": 323, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/plugins/prithvi_io_processor_plugin/prithvi_io_processor/types.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any, Literal, TypedDict import albumentations from pydantic import BaseModel class DataModuleConfig(TypedDict): bands: list[str] batch_size: int constant_scale: float data_root: str drop_last: bool no_data_replace: float no_label_replace: int num_workers: int test_transform: list[albumentations.core.transforms_interface.BasicTransform] class ImagePrompt(BaseModel): data_format: Literal["b64_json", "bytes", "url", "path"] """ This is the data type for the input image """ image_format: str """ This is the image format (e.g., jpeg, png, etc.) """ out_data_format: Literal["b64_json", "url"] data: Any """ Input image data """ class ImageRequestOutput(BaseModel): """ The output data of an image request to vLLM. Args: type (str): The data content type [path, object] format (str): The image format (e.g., jpeg, png, etc.) data (Any): The resulting data. """ type: Literal["path", "b64_json"] format: str data: str
{ "repo_id": "vllm-project/vllm", "file_path": "tests/plugins/prithvi_io_processor_plugin/prithvi_io_processor/types.py", "license": "Apache License 2.0", "lines": 40, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/plugins_tests/test_io_processor_plugins.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import base64 import io import imagehash import pytest import requests from PIL import Image from tests.utils import RemoteOpenAIServer from vllm.config import VllmConfig from vllm.entrypoints.pooling.pooling.protocol import IOProcessorResponse from vllm.plugins.io_processors import get_io_processor models_config = { "ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11": { "image_url": "https://huggingface.co/christian-pinto/Prithvi-EO-2.0-300M-TL-VLLM/resolve/main/valencia_example_2024-10-26.tiff", # noqa: E501 "out_hash": "aa6d92ad25926a5e", "plugin": "prithvi_to_tiff", }, "ibm-nasa-geospatial/Prithvi-EO-2.0-300M-BurnScars": { "image_url": "https://huggingface.co/ibm-nasa-geospatial/Prithvi-EO-2.0-300M-BurnScars/resolve/main/examples/subsetted_512x512_HLS.S30.T10SEH.2018190.v1.4_merged.tif", # noqa: E501 "out_hash": "c07f4f602da73552", "plugin": "prithvi_to_tiff", }, } def _compute_image_hash(base64_data: str) -> str: # Decode the base64 output and create image from byte stream decoded_image = base64.b64decode(base64_data) image = Image.open(io.BytesIO(decoded_image)) # Compute perceptual hash of the output image return str(imagehash.phash(image)) def test_loading_missing_plugin(): vllm_config = VllmConfig() with pytest.raises(ValueError): get_io_processor(vllm_config, None, "wrong_plugin") @pytest.fixture(scope="function") def server(model_name, plugin): args = [ "--runner", "pooling", "--enforce-eager", "--skip-tokenizer-init", # Limit the maximum number of parallel requests # to avoid the model going OOM in CI. "--max-num-seqs", "32", "--io-processor-plugin", plugin, "--enable-mm-embeds", ] with RemoteOpenAIServer(model_name, args) as remote_server: yield remote_server @pytest.mark.asyncio @pytest.mark.parametrize( "model_name, image_url, plugin, expected_hash", [ (model_name, config["image_url"], config["plugin"], config["out_hash"]) for model_name, config in models_config.items() ], ) async def test_prithvi_mae_plugin_online( server: RemoteOpenAIServer, model_name: str, image_url: str | dict, plugin: str, expected_hash: str, ): request_payload_url = { "data": { "data": image_url, "data_format": "url", "image_format": "tiff", "out_data_format": "b64_json", }, "priority": 0, "model": model_name, "softmax": False, } ret = requests.post( server.url_for("pooling"), json=request_payload_url, ) response = ret.json() # verify the request response is in the correct format assert (parsed_response := IOProcessorResponse(**response)) # verify the output is formatted as expected for this plugin plugin_data = parsed_response.data assert all(plugin_data.get(attr) for attr in ["type", "format", "data"]) # Compute the output image hash and compare it against the expected hash image_hash = _compute_image_hash(plugin_data["data"]) assert image_hash == expected_hash, ( f"Image hash mismatch: expected {expected_hash}, got {image_hash}" ) @pytest.mark.parametrize( "model_name, image_url, plugin, expected_hash", [ (model_name, config["image_url"], config["plugin"], config["out_hash"]) for model_name, config in models_config.items() ], ) def test_prithvi_mae_plugin_offline( vllm_runner, model_name: str, image_url: str | dict, plugin: str, expected_hash: str ): img_data = dict( data=image_url, data_format="url", image_format="tiff", out_data_format="b64_json", ) prompt = dict(data=img_data) with vllm_runner( model_name, runner="pooling", skip_tokenizer_init=True, enable_mm_embeds=True, enforce_eager=True, # Limit the maximum number of parallel requests # to avoid the model going OOM in CI. max_num_seqs=32, io_processor_plugin=plugin, default_torch_num_threads=1, ) as llm_runner: pooler_output = llm_runner.get_llm().encode(prompt, pooling_task="plugin") output = pooler_output[0].outputs # verify the output is formatted as expected for this plugin assert all(hasattr(output, attr) for attr in ["type", "format", "data"]) # Compute the output image hash and compare it against the expected hash image_hash = _compute_image_hash(output.data) assert image_hash == expected_hash, ( f"Image hash mismatch: expected {expected_hash}, got {image_hash}" )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/plugins_tests/test_io_processor_plugins.py", "license": "Apache License 2.0", "lines": 130, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/plugins/io_processors/interface.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import warnings from abc import ABC, abstractmethod from collections.abc import AsyncGenerator, Sequence from typing import Generic, TypeVar from vllm.config import VllmConfig from vllm.inputs.data import PromptType from vllm.outputs import PoolingRequestOutput from vllm.pooling_params import PoolingParams from vllm.renderers import BaseRenderer from vllm.sampling_params import SamplingParams IOProcessorInput = TypeVar("IOProcessorInput") IOProcessorOutput = TypeVar("IOProcessorOutput") class IOProcessor(ABC, Generic[IOProcessorInput, IOProcessorOutput]): """Abstract interface for pre/post-processing of engine I/O.""" def __init__(self, vllm_config: VllmConfig, renderer: BaseRenderer): super().__init__() self.vllm_config = vllm_config def parse_data(self, data: object) -> IOProcessorInput: if callable(parse_request := getattr(self, "parse_request", None)): warnings.warn( "`parse_request` has been renamed to `parse_data`. " "Please update your IO Processor Plugin to use the new name. " "The old name will be removed in v0.19.", DeprecationWarning, stacklevel=2, ) return parse_request(data) # type: ignore raise NotImplementedError def merge_sampling_params( self, params: SamplingParams | None = None, ) -> SamplingParams: if callable( validate_or_generate_params := getattr( self, "validate_or_generate_params", None ) ): warnings.warn( "`validate_or_generate_params` has been split into " "`merge_sampling_params` and `merge_pooling_params`." "Please update your IO Processor Plugin to use the new methods. " "The old name will be removed in v0.19.", DeprecationWarning, stacklevel=2, ) return validate_or_generate_params(params) # type: ignore return params or SamplingParams() def merge_pooling_params( self, params: PoolingParams | None = None, ) -> PoolingParams: if callable( validate_or_generate_params := getattr( self, "validate_or_generate_params", None ) ): warnings.warn( "`validate_or_generate_params` has been split into " "`merge_sampling_params` and `merge_pooling_params`." "Please update your IO Processor Plugin to use the new methods. " "The old name will be removed in v0.19.", DeprecationWarning, stacklevel=2, ) return validate_or_generate_params(params) # type: ignore return params or PoolingParams(task="plugin") @abstractmethod def pre_process( self, prompt: IOProcessorInput, request_id: str | None = None, **kwargs, ) -> PromptType | Sequence[PromptType]: raise NotImplementedError async def pre_process_async( self, prompt: IOProcessorInput, request_id: str | None = None, **kwargs, ) -> PromptType | Sequence[PromptType]: return self.pre_process(prompt, request_id, **kwargs) @abstractmethod def post_process( self, model_output: Sequence[PoolingRequestOutput], request_id: str | None = None, **kwargs, ) -> IOProcessorOutput: raise NotImplementedError async def post_process_async( self, model_output: AsyncGenerator[tuple[int, PoolingRequestOutput]], request_id: str | None = None, **kwargs, ) -> IOProcessorOutput: # We cannot guarantee outputs are returned in the same order they were # fed to vLLM. # Let's sort them by id before post_processing sorted_output = sorted( [(i, item) async for i, item in model_output], key=lambda output: output[0] ) collected_output = [output[1] for output in sorted_output] return self.post_process(collected_output, request_id=request_id, **kwargs)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/plugins/io_processors/interface.py", "license": "Apache License 2.0", "lines": 105, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/mamba/linear_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math from collections.abc import Callable import torch import torch.nn.functional as F from einops import rearrange from torch import nn from vllm.config import CacheConfig, ModelConfig, get_current_vllm_config from vllm.distributed.communication_op import tensor_model_parallel_all_reduce from vllm.distributed.parallel_state import ( get_tensor_model_parallel_rank, get_tensor_model_parallel_world_size, ) from vllm.forward_context import ForwardContext, get_forward_context from vllm.model_executor.custom_op import CustomOp from vllm.model_executor.layers.lightning_attn import ( lightning_attention, linear_decode_forward_triton, ) from vllm.model_executor.layers.linear import ColumnParallelLinear, RowParallelLinear from vllm.model_executor.layers.mamba.abstract import MambaBase from vllm.model_executor.layers.mamba.mamba_utils import ( MambaStateDtypeCalculator, MambaStateShapeCalculator, ) from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.utils.torch_utils import direct_register_custom_op from vllm.v1.attention.backend import AttentionMetadata from vllm.v1.attention.backends.linear_attn import LinearAttentionMetadata class MiniMaxText01RMSNormTP(CustomOp): name = "MiniMaxText01RMSNormTP" def __init__(self, hidden_size: int, eps: float = 1e-6) -> None: super().__init__() self.tp_world = get_tensor_model_parallel_world_size() self.tp_rank = get_tensor_model_parallel_rank() self.weight = nn.Parameter(torch.ones(int(hidden_size / self.tp_world))) self.weight.weight_loader = self.weight_loader self.variance_epsilon = eps @staticmethod def weight_loader( param: nn.Parameter, loaded_weight: torch.Tensor, ) -> None: tp_world = get_tensor_model_parallel_world_size() tp_rank = get_tensor_model_parallel_rank() shard_size = loaded_weight.shape[0] // tp_world shard = slice(tp_rank * shard_size, (tp_rank + 1) * shard_size) param.data.copy_(loaded_weight[shard]) def _forward( self, x: torch.Tensor, ) -> torch.Tensor: orig_dtype = x.dtype x = x.to(torch.float32) variance = x.pow(2).mean(dim=-1, keepdim=True, dtype=torch.float32) if self.tp_world > 1: variance = tensor_model_parallel_all_reduce(variance) / self.tp_world x = x * torch.rsqrt(variance + self.variance_epsilon) x = (x * self.weight).to(orig_dtype) return x def forward( self, x: torch.Tensor, residual: torch.Tensor | None = None, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: assert residual is None, "RMSNorm does not support residual connection." return self._forward(x) @staticmethod def forward_qk( q_norm: "MiniMaxText01RMSNormTP", k_norm: "MiniMaxText01RMSNormTP", q: torch.Tensor, k: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: orig_dtype = q.dtype q = q.to(torch.float32) k = k.to(torch.float32) q_var = q.pow(2).mean(dim=-1, keepdim=True) k_var = k.pow(2).mean(dim=-1, keepdim=True) if q_norm.tp_world > 1: qk_var = torch.cat([q_var, k_var], dim=-1) qk_var = tensor_model_parallel_all_reduce(qk_var) / q_norm.tp_world q_var, k_var = qk_var.chunk(2, dim=-1) q = q * torch.rsqrt(q_var + q_norm.variance_epsilon) * q_norm.weight k = k * torch.rsqrt(k_var + k_norm.variance_epsilon) * k_norm.weight q = q.to(orig_dtype) k = k.to(orig_dtype) return q, k def clear_linear_attention_cache_for_new_sequences( kv_cache: torch.Tensor, state_indices_tensor: torch.Tensor, attn_metadata: LinearAttentionMetadata, ) -> None: num_prefills = getattr(attn_metadata, "num_prefills", 0) if num_prefills <= 0: return num_decode_tokens = getattr(attn_metadata, "num_decode_tokens", 0) for prefill_idx in range(num_prefills): q_start = attn_metadata.query_start_loc[num_decode_tokens + prefill_idx] q_end = attn_metadata.query_start_loc[num_decode_tokens + prefill_idx + 1] query_len = q_end - q_start context_len = ( attn_metadata.seq_lens[num_decode_tokens + prefill_idx] - query_len ) if context_len == 0: block_to_clear = state_indices_tensor[num_decode_tokens + prefill_idx] kv_cache[block_to_clear, ...] = 0 def linear_attention_decode( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, kv_cache: torch.Tensor, slope_rate: torch.Tensor, state_indices_tensor: torch.Tensor, q_start: int = 0, q_end: int | None = None, slot_start: int = 0, slot_end: int | None = None, block_size: int = 32, ) -> torch.Tensor: q = q[q_start:q_end].unsqueeze(2).contiguous() k = k[q_start:q_end].unsqueeze(2).contiguous() v = v[q_start:q_end].unsqueeze(2).contiguous() slot_id = state_indices_tensor[slot_start:slot_end] return linear_decode_forward_triton( q, k, v, kv_cache, slope_rate, slot_id, block_size ) def linear_attention_prefill_and_mix( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, kv_cache: torch.Tensor, state_indices_tensor: torch.Tensor, attn_metadata: LinearAttentionMetadata, slope_rate: torch.Tensor, block_size: int, decode_fn: Callable[..., torch.Tensor], prefix_fn: Callable[..., torch.Tensor], layer_idx: int | None = None, ) -> torch.Tensor: hidden = [] for _prefill_idx in range(getattr(attn_metadata, "num_prefills", 0)): if _prefill_idx >= len(attn_metadata.query_start_loc): break if _prefill_idx >= len(state_indices_tensor): break offset = attn_metadata.num_decode_tokens _start = attn_metadata.query_start_loc[offset + _prefill_idx] _end = attn_metadata.query_start_loc[offset + _prefill_idx + 1] slot_id = state_indices_tensor[offset + _prefill_idx] qs = q[_start:_end].transpose(0, 1).contiguous() ks = k[_start:_end].transpose(0, 1).contiguous() vs = v[_start:_end].transpose(0, 1).contiguous() slice_layer_cache = kv_cache[slot_id, ...] out_slice = prefix_fn( qs, ks, vs, slice_layer_cache, slope_rate, block_size, layer_idx=layer_idx, ) hidden.append(out_slice.contiguous()) if attn_metadata.num_decode_tokens > 0: hidden_decode = decode_fn( q, k, v, kv_cache, state_indices_tensor, attn_metadata ) hidden.insert(0, hidden_decode) if not hidden: return torch.empty((0, q.size(-1)), device=q.device, dtype=q.dtype) hidden = torch.concat(hidden, dim=0).contiguous() return hidden class MiniMaxText01LinearKernel: @staticmethod def jit_linear_forward_prefix( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, kv_caches: torch.Tensor, slope_rate: torch.Tensor, block_size: int, layer_idx: int | None = None, **kwargs, ) -> torch.Tensor: slope_rate = slope_rate.to(torch.float32) should_pad_dim = q.dim() == 3 if should_pad_dim: q = q.unsqueeze(0) k = k.unsqueeze(0) v = v.unsqueeze(0) b, h, n, d = q.shape e = d kv_history = kv_caches.reshape(1, h, d, e).contiguous() output, kv_history = lightning_attention( q, k, v, slope_rate, block_size=block_size, kv_history=kv_history ) kv_caches.copy_(kv_history[:, :, -1, :, :].reshape(h, d, e)) assert output.shape[0] == 1, "batch size must be 1" return rearrange(output.squeeze(0), "h n d -> n (h d)") class MiniMaxText01LinearAttention(nn.Module, MambaBase): @property def mamba_type(self) -> str: return "linear_attention" def get_state_dtype(self) -> tuple[torch.dtype]: assert self.model_config is not None assert self.cache_config is not None return MambaStateDtypeCalculator.linear_attention_state_dtype( self.model_config.dtype, self.cache_config.mamba_cache_dtype, ) def get_state_shape(self) -> tuple[tuple[int, int, int], ...]: return MambaStateShapeCalculator.linear_attention_state_shape( num_heads=self.num_heads, tp_size=self.tp_size, head_dim=self.head_dim ) def __init__( self, hidden_size: int, hidden_inner_size: int, num_heads: int, head_dim: int, max_position: int, block_size: int, num_hidden_layer: int, model_config: ModelConfig | None = None, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, layer_idx: int = 0, linear_layer_idx: int = 0, prefix: str = "linear_attn", ) -> None: super().__init__() self.layer_idx = layer_idx self.BLOCK = block_size self.hidden_size = hidden_size self.num_heads = num_heads self.head_dim = head_dim self.total_num_heads = num_heads self.hidden_inner_size = hidden_inner_size self.tp_size = get_tensor_model_parallel_world_size() self.tp_rank = get_tensor_model_parallel_rank() assert self.total_num_heads % self.tp_size == 0 self.tp_heads = self.total_num_heads // self.tp_size self.qkv_size = self.num_heads * self.head_dim self.tp_hidden = self.head_dim * self.tp_heads self.model_config = model_config self.cache_config = cache_config self.prefix = prefix self.qkv_proj = ColumnParallelLinear( hidden_size, self.hidden_inner_size * 3, bias=False, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", ) self.output_gate = ColumnParallelLinear( hidden_size, self.hidden_inner_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.output_gate", ) self.out_proj = RowParallelLinear( self.hidden_inner_size, hidden_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.out_proj", ) self.norm = MiniMaxText01RMSNormTP( self.hidden_inner_size, eps=1e-5, ) slope_rate = MiniMaxText01LinearAttention._build_slope_tensor(self.num_heads) if num_hidden_layer <= 1: self.slope_rate = slope_rate * (1 + 1e-5) else: self.slope_rate = slope_rate * ( 1 - layer_idx / (num_hidden_layer - 1) + 1e-5 ) self.tp_slope = self.slope_rate[ self.tp_rank * self.tp_heads : (self.tp_rank + 1) * self.tp_heads ].contiguous() compilation_config = get_current_vllm_config().compilation_config if prefix in compilation_config.static_forward_context: raise ValueError(f"Duplicate layer name: {prefix}") compilation_config.static_forward_context[prefix] = self @staticmethod def weight_direct_load(param: torch.Tensor, loaded_weight: torch.Tensor) -> None: assert param.size() == loaded_weight.size() param.data.copy_(loaded_weight) return @staticmethod def _build_slope_tensor(n_attention_heads: int): def get_slopes(n): def get_slopes_power_of_2(n): start = 2 ** (-(2 ** -(math.log2(n) - 3))) ratio = start return [start * ratio**i for i in range(n)] if math.log2(n).is_integer(): return get_slopes_power_of_2(n) else: closest_power_of_2 = 2 ** math.floor(math.log2(n)) return ( get_slopes_power_of_2(closest_power_of_2) + get_slopes(2 * closest_power_of_2)[0::2][: n - closest_power_of_2] ) slopes = torch.tensor( get_slopes(n_attention_heads), dtype=torch.float32 ).reshape(n_attention_heads, 1, 1) return slopes def _prefill_and_mix_infer( self, q, k, v, kv_cache, state_indices_tensor, attn_metadata ): return linear_attention_prefill_and_mix( q=q, k=k, v=v, kv_cache=kv_cache, state_indices_tensor=state_indices_tensor, attn_metadata=attn_metadata, slope_rate=self.tp_slope, block_size=self.BLOCK, decode_fn=self._decode_infer, prefix_fn=MiniMaxText01LinearKernel.jit_linear_forward_prefix, layer_idx=self.layer_idx, ) def _decode_infer(self, q, k, v, kv_cache, state_indices_tensor, attn_metadata): hidden = linear_attention_decode( q, k, v, kv_cache, self.tp_slope, state_indices_tensor, q_start=0, q_end=attn_metadata.num_decode_tokens, slot_start=0, slot_end=attn_metadata.num_decodes, block_size=32, ) return hidden def forward( self, hidden_states: torch.Tensor, output: torch.Tensor, positions: torch.Tensor ) -> None: torch.ops.vllm.linear_attention( hidden_states, output, positions, self.prefix, ) def _forward( self, hidden_states: torch.Tensor, output: torch.Tensor, positions: torch.Tensor ) -> None: forward_context = get_forward_context() attn_metadata: AttentionMetadata = forward_context.attn_metadata if attn_metadata is not None: assert isinstance(attn_metadata, dict) attn_metadata = attn_metadata[self.prefix] assert isinstance(attn_metadata, LinearAttentionMetadata) num_actual_tokens = ( attn_metadata.num_prefill_tokens + attn_metadata.num_decode_tokens ) else: num_actual_tokens = hidden_states.shape[0] qkv, _ = self.qkv_proj(hidden_states[:num_actual_tokens]) qkv32 = qkv.to(torch.float32) qkvact = torch.nn.functional.silu(qkv32) qkvact = qkvact.view((qkv.shape[0], self.tp_heads, -1)) q, k, v = torch.split(qkvact, [self.head_dim] * 3, dim=-1) if attn_metadata is not None: kv_cache = self.kv_cache[forward_context.virtual_engine][0] state_indices_tensor = attn_metadata.state_indices_tensor clear_linear_attention_cache_for_new_sequences( kv_cache, state_indices_tensor, attn_metadata ) decode_only = getattr(attn_metadata, "num_prefills", 0) == 0 if attn_metadata is None: hidden = torch.empty( (q.shape[0], q.shape[1] * q.shape[2]), device=q.device, dtype=q.dtype ) else: if not decode_only: hidden = self._prefill_and_mix_infer( q, k, v, kv_cache, state_indices_tensor, attn_metadata ) else: hidden = self._decode_infer( q, k, v, kv_cache, state_indices_tensor, attn_metadata ) hidden = self.norm._forward(hidden) gate, _ = self.output_gate(hidden_states[:num_actual_tokens]) hidden = F.sigmoid(gate) * hidden hidden = hidden.to(hidden_states.dtype) output[:num_actual_tokens], _ = self.out_proj(hidden) def linear_attention( hidden_states: torch.Tensor, output: torch.Tensor, positions: torch.Tensor, layer_name: str, ) -> None: forward_context: ForwardContext = get_forward_context() self = forward_context.no_compile_layers[layer_name] self._forward(hidden_states=hidden_states, output=output, positions=positions) def linear_attention_fake( hidden_states: torch.Tensor, output: torch.Tensor, positions: torch.Tensor, layer_name: str, ) -> None: return direct_register_custom_op( op_name="linear_attention", op_func=linear_attention, mutates_args=["output"], fake_impl=linear_attention_fake, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/mamba/linear_attn.py", "license": "Apache License 2.0", "lines": 418, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/entrypoints/openai/test_token_in_token_out.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os import tempfile import pytest from vllm.model_executor.model_loader.weight_utils import download_weights_from_hf from vllm.tokenizers import get_tokenizer from ...utils import RemoteOpenAIServer MODEL_NAME = "Qwen/Qwen3-0.6B" MODEL_PATH = os.path.join(tempfile.gettempdir(), "qwen3_06b") @pytest.fixture(scope="module") def server(): global MODEL_PATH MODEL_PATH = download_weights_from_hf( MODEL_NAME, allow_patterns=["*"], cache_dir=MODEL_PATH, ignore_patterns=["tokenizer*", "vocab*", "*.safetensors"], ) args = [ "--max-model-len", "2048", "--max-num-seqs", "128", "--enforce-eager", "--skip-tokenizer-init", "--load-format", "dummy", ] with RemoteOpenAIServer(MODEL_PATH, args) as remote_server: yield remote_server @pytest.mark.asyncio async def test_token_in_token_out_and_logprobs(server): """ Test token-in-token-out and token_ids align with prompt_logprobs & logprobs when return_tokens_as_token_ids is enabled. """ tokenizer = get_tokenizer(tokenizer_name=MODEL_NAME) text = "Hello, world! How are you today?" token_ids = tokenizer.encode(text) async with server.get_async_client() as client: # Test with both return_token_ids and return_tokens_as_token_ids enabled completion = await client.completions.create( model=MODEL_PATH, prompt=token_ids, max_tokens=20, temperature=0, echo=True, extra_body={ "return_token_ids": True, }, ) # Verify all fields are present assert ( completion.choices[0].token_ids is not None and 0 < len(completion.choices[0].token_ids) <= 20 ) assert completion.choices[0].prompt_token_ids is not None # Decode prompt tokens if completion.choices[0].prompt_token_ids: prompt_text = tokenizer.decode(completion.choices[0].prompt_token_ids) # The decoded prompt should match or close to original prompt assert prompt_text == text
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/test_token_in_token_out.py", "license": "Apache License 2.0", "lines": 63, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/models/apertus.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Adapted from # https://github.com/huggingface/transformers/blob/v4.28.0/src/transformers/models/llama/modeling_llama.py # Copyright 2025 The Swiss AI Initiative. # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate the architectural differences made by # the Swiss AI Initiative that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Inference-only Apertus model compatible with HuggingFace weights.""" from collections.abc import Iterable from itertools import islice import torch from torch import nn from transformers import ApertusConfig from vllm.compilation.decorators import support_torch_compile from vllm.config import CacheConfig, VllmConfig from vllm.distributed import get_pp_group, get_tensor_model_parallel_world_size from vllm.model_executor.layers.activation import XIELU from vllm.model_executor.layers.attention import ( Attention, EncoderOnlyAttention, ) from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( ColumnParallelLinear, QKVParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import ( default_weight_loader, maybe_remap_kv_scale_name, ) from vllm.sequence import IntermediateTensors from vllm.v1.attention.backend import AttentionType from .interfaces import SupportsLoRA, SupportsPP from .utils import ( AutoWeightsLoader, PPMissingLayer, extract_layer_index, is_pp_missing_parameter, make_empty_intermediate_tensors_factory, make_layers, maybe_prefix, ) class ApertusMLP(nn.Module): def __init__( self, hidden_size: int, intermediate_size: int, hidden_act: str, quant_config: QuantizationConfig | None = None, bias: bool = False, prefix: str = "", reduce_results: bool = True, ) -> None: super().__init__() self.up_proj = ColumnParallelLinear( input_size=hidden_size, output_size=intermediate_size, bias=bias, quant_config=quant_config, prefix=f"{prefix}.up_proj", ) self.down_proj = RowParallelLinear( input_size=intermediate_size, output_size=hidden_size, bias=bias, quant_config=quant_config, reduce_results=reduce_results, prefix=f"{prefix}.down_proj", ) if hidden_act != "xielu": raise ValueError( f"Unsupported activation: {hidden_act}. " "Only xIELU is supported for now." ) self.act_fn = XIELU() def forward(self, x): x, _ = self.up_proj(x) x = self.act_fn(x) x, _ = self.down_proj(x) return x class ApertusAttention(nn.Module): def __init__( self, config: ApertusConfig, hidden_size: int, num_heads: int, num_kv_heads: int, max_position_embeddings: int = 8192, quant_config: QuantizationConfig | None = None, bias: bool = False, bias_o_proj: bool = False, cache_config: CacheConfig | None = None, prefix: str = "", attn_type: str = AttentionType.DECODER, ) -> None: super().__init__() layer_idx = extract_layer_index(prefix) self.hidden_size = hidden_size tp_size = get_tensor_model_parallel_world_size() self.total_num_heads = num_heads assert self.total_num_heads % tp_size == 0 self.num_heads = self.total_num_heads // tp_size self.total_num_kv_heads = num_kv_heads if self.total_num_kv_heads >= tp_size: # Number of KV heads is greater than TP size, so we partition # the KV heads across multiple tensor parallel GPUs. assert self.total_num_kv_heads % tp_size == 0 else: # Number of KV heads is less than TP size, so we replicate # the KV heads across multiple tensor parallel GPUs. assert tp_size % self.total_num_kv_heads == 0 self.num_kv_heads = max(1, self.total_num_kv_heads // tp_size) # MistralConfig has an optional head_dim introduced by Mistral-Nemo head_dim = getattr(config, "head_dim", None) if head_dim is None: head_dim = self.hidden_size // self.total_num_heads self.head_dim = head_dim self.q_size = self.num_heads * self.head_dim self.kv_size = self.num_kv_heads * self.head_dim self.scaling = self.head_dim**-0.5 self.max_position_embeddings = max_position_embeddings self.qkv_proj = QKVParallelLinear( hidden_size=hidden_size, head_size=self.head_dim, total_num_heads=self.total_num_heads, total_num_kv_heads=self.total_num_kv_heads, bias=bias, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", ) self.o_proj = RowParallelLinear( input_size=self.total_num_heads * self.head_dim, output_size=hidden_size, bias=bias_o_proj, quant_config=quant_config, prefix=f"{prefix}.o_proj", ) self._init_rotary_emb(config, quant_config=quant_config) sliding_window = None if layer_types := getattr(config, "layer_types", None): is_sliding = layer_types[layer_idx] == "sliding_attention" if is_sliding: sliding_window = config.sliding_window attn_cls = ( EncoderOnlyAttention if attn_type == AttentionType.ENCODER_ONLY else Attention ) self.attn = attn_cls( self.num_heads, self.head_dim, self.scaling, num_kv_heads=self.num_kv_heads, cache_config=cache_config, quant_config=quant_config, per_layer_sliding_window=sliding_window, attn_type=attn_type, prefix=f"{prefix}.attn", ) self.q_norm = RMSNorm(self.head_dim, eps=config.rms_norm_eps) self.k_norm = RMSNorm(self.head_dim, eps=config.rms_norm_eps) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, ) -> torch.Tensor: qkv, _ = self.qkv_proj(hidden_states) q, k, v = qkv.split([self.q_size, self.kv_size, self.kv_size], dim=-1) q = self.q_norm(q.contiguous().view(-1, self.head_dim)).view_as(q) k = self.k_norm(k.contiguous().view(-1, self.head_dim)).view_as(k) q, k = self.rotary_emb(positions, q, k) attn_output = self.attn(q, k, v) output, _ = self.o_proj(attn_output) return output def _init_rotary_emb( self, config: ApertusConfig, quant_config: QuantizationConfig | None, ) -> None: is_neox_style = True is_gguf = quant_config and quant_config.get_name() == "gguf" if is_gguf and config.model_type == "apertus": is_neox_style = False self.rotary_emb = get_rope( self.head_dim, max_position=self.max_position_embeddings, rope_parameters=config.rope_parameters, is_neox_style=is_neox_style, ) class ApertusDecoderLayer(nn.Module): def __init__( self, config: ApertusConfig, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.hidden_size = config.hidden_size max_position_embeddings = getattr(config, "max_position_embeddings", 8192) # Support abacusai/Smaug-72B-v0.1 with attention_bias # Support internlm/internlm-7b with bias attention_bias = getattr(config, "attention_bias", False) or getattr( config, "bias", False ) bias_o_proj = attention_bias # support internlm/internlm3-8b with qkv_bias if hasattr(config, "qkv_bias"): attention_bias = config.qkv_bias # Apertus defaults to causal attention as it is a decoder-only model. # You can override the HF config with `is_causal=False` to enable # bidirectional attention, which is used in some embedding models # (e.g. parasail-ai/GritLM-7B-vllm) if getattr(config, "is_causal", True): attn_type = AttentionType.DECODER else: attn_type = AttentionType.ENCODER_ONLY self.self_attn = ApertusAttention( config=config, hidden_size=self.hidden_size, num_heads=config.num_attention_heads, num_kv_heads=getattr( config, "num_key_value_heads", config.num_attention_heads ), max_position_embeddings=max_position_embeddings, quant_config=quant_config, bias=attention_bias, bias_o_proj=bias_o_proj, cache_config=cache_config, prefix=f"{prefix}.self_attn", attn_type=attn_type, ) self.mlp = ApertusMLP( hidden_size=self.hidden_size, intermediate_size=config.intermediate_size, hidden_act=config.hidden_act, quant_config=quant_config, bias=getattr(config, "mlp_bias", False), prefix=f"{prefix}.mlp", ) self.attention_layernorm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.feedforward_layernorm = RMSNorm( config.hidden_size, eps=config.rms_norm_eps ) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, residual: torch.Tensor | None, ) -> tuple[torch.Tensor, torch.Tensor]: # Self Attention if residual is None: residual = hidden_states hidden_states = self.attention_layernorm(hidden_states) else: hidden_states, residual = self.attention_layernorm(hidden_states, residual) hidden_states = self.self_attn(positions=positions, hidden_states=hidden_states) # Fully Connected hidden_states, residual = self.feedforward_layernorm(hidden_states, residual) hidden_states = self.mlp(hidden_states) return hidden_states, residual @support_torch_compile class ApertusModel(nn.Module): def __init__( self, *, vllm_config: VllmConfig, prefix: str = "", layer_type: type[nn.Module] = ApertusDecoderLayer, ): super().__init__() config = vllm_config.model_config.hf_config cache_config = vllm_config.cache_config quant_config = vllm_config.quant_config self.config = config self.quant_config = quant_config self.vocab_size = config.vocab_size if get_pp_group().is_first_rank or ( config.tie_word_embeddings and get_pp_group().is_last_rank ): self.embed_tokens = VocabParallelEmbedding( self.vocab_size, config.hidden_size, quant_config=quant_config, ) else: self.embed_tokens = PPMissingLayer() self.start_layer, self.end_layer, self.layers = make_layers( config.num_hidden_layers, lambda prefix: layer_type( config=config, cache_config=cache_config, quant_config=quant_config, prefix=prefix, ), prefix=f"{prefix}.layers", ) if get_pp_group().is_last_rank: self.norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) else: self.norm = PPMissingLayer() self.aux_hidden_state_layers = tuple[int, ...]() self.make_empty_intermediate_tensors = make_empty_intermediate_tensors_factory( ["hidden_states", "residual"], config.hidden_size ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor | IntermediateTensors | tuple[torch.Tensor, list[torch.Tensor]]: if get_pp_group().is_first_rank: if inputs_embeds is not None: hidden_states = inputs_embeds else: hidden_states = self.embed_input_ids(input_ids) residual = None else: assert intermediate_tensors is not None hidden_states = intermediate_tensors["hidden_states"] residual = intermediate_tensors["residual"] aux_hidden_states = [] for idx, layer in enumerate( islice(self.layers, self.start_layer, self.end_layer) ): if idx in self.aux_hidden_state_layers: aux_hidden_states.append(hidden_states + residual) hidden_states, residual = layer(positions, hidden_states, residual) if not get_pp_group().is_last_rank: return IntermediateTensors( {"hidden_states": hidden_states, "residual": residual} ) hidden_states, _ = self.norm(hidden_states, residual) if len(aux_hidden_states) > 0: return hidden_states, aux_hidden_states return hidden_states def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ # (param_name, shard_name, shard_id) (".qkv_proj", ".q_proj", "q"), (".qkv_proj", ".k_proj", "k"), (".qkv_proj", ".v_proj", "v"), ] params_dict = dict(self.named_parameters()) # we need to load the buffers for beta and eps (XIELU) for name, buffer in self.named_buffers(): if name.endswith(".beta") or name.endswith(".eps"): params_dict[name] = buffer loaded_params: set[str] = set() for name, loaded_weight in weights: if "rotary_emb.inv_freq" in name: continue if "rotary_emb.cos_cached" in name or "rotary_emb.sin_cached" in name: # Models trained using ColossalAI may include these tensors in # the checkpoint. Skip them. continue if self.quant_config is not None and ( scale_name := self.quant_config.get_cache_scale(name) ): # Loading kv cache quantization scales param = params_dict[scale_name] weight_loader = getattr(param, "weight_loader", default_weight_loader) loaded_weight = ( loaded_weight if loaded_weight.dim() == 0 else loaded_weight[0] ) weight_loader(param, loaded_weight) loaded_params.add(scale_name) continue if "scale" in name or "zero_point" in name: # Remapping the name of FP8 kv-scale. name = maybe_remap_kv_scale_name(name, params_dict) if name is None: continue for param_name, weight_name, shard_id in stacked_params_mapping: if weight_name not in name: continue name = name.replace(weight_name, param_name) # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params class ApertusForCausalLM(nn.Module, SupportsLoRA, SupportsPP): packed_modules_mapping = {"qkv_proj": ["q_proj", "k_proj", "v_proj"]} # LoRA specific attributes embedding_modules = { "embed_tokens": "input_embeddings", "lm_head": "output_embeddings", } def __init__( self, *, vllm_config: VllmConfig, prefix: str = "", layer_type: type[nn.Module] = ApertusDecoderLayer, ): super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config self.config = config self.model = self._init_model( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model"), layer_type=layer_type, ) if get_pp_group().is_last_rank: self.lm_head = ParallelLMHead( config.vocab_size, config.hidden_size, quant_config=quant_config, prefix=maybe_prefix(prefix, "lm_head"), ) if config.tie_word_embeddings: self.lm_head = self.lm_head.tie_weights(self.model.embed_tokens) logit_scale = getattr(config, "logit_scale", 1.0) self.logits_processor = LogitsProcessor( config.vocab_size, scale=logit_scale ) else: self.lm_head = PPMissingLayer() self.make_empty_intermediate_tensors = ( self.model.make_empty_intermediate_tensors ) def set_aux_hidden_state_layers(self, layers: tuple[int, ...]) -> None: self.model.aux_hidden_state_layers = layers def get_eagle3_aux_hidden_state_layers(self) -> tuple[int, ...]: num_layers = len(self.model.layers) return (2, num_layers // 2, num_layers - 3) def _init_model( self, vllm_config: VllmConfig, prefix: str = "", layer_type: type[nn.Module] = ApertusDecoderLayer, ): return ApertusModel( vllm_config=vllm_config, prefix=prefix, layer_type=layer_type ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor | IntermediateTensors: model_output = self.model( input_ids, positions, intermediate_tensors, inputs_embeds ) return model_output def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: logits = self.logits_processor(self.lm_head, hidden_states) return logits def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader( self, skip_prefixes=(["lm_head."] if self.config.tie_word_embeddings else None), ) return loader.load_weights(weights)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/apertus.py", "license": "Apache License 2.0", "lines": 505, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/quantization/compressed_tensors/transform/linear.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable, Generator from itertools import accumulate import torch from compressed_tensors.transform import ( TransformArgs, TransformConfig, TransformLocation, TransformScheme, ) from compressed_tensors.utils import is_match from vllm.model_executor.layers.linear import ( WEIGHT_LOADER_V2_SUPPORTED, LinearMethodBase, ) from vllm.model_executor.layers.quantization.compressed_tensors.compressed_tensors import ( # noqa: E501 CompressedTensorsScheme, ) from vllm.model_executor.layers.quantization.compressed_tensors.transform.module import ( # noqa: E501 HadamardTransform, ) from vllm.model_executor.layers.quantization.compressed_tensors.transform.utils import ( # noqa: E501 TransformTuple, ) class CompressedTensorsLinearTransformMethod(LinearMethodBase): """ Wraps `CompressedTensorsLinearMethod` or `UnquantizedLinearMethod` and adds input and output transforms to either side of the original apply method """ @classmethod def from_schemes( cls, quant_method: LinearMethodBase, quant_scheme: CompressedTensorsScheme | None, input_tfms: dict[int, TransformTuple], output_tfms: dict[int, TransformTuple], ) -> "CompressedTensorsLinearTransformMethod": from vllm.model_executor.layers.quantization.compressed_tensors.transform.schemes.linear_qutlass_nvfp4 import ( # noqa: E501 QutlassNvFP4LinearMethod, is_qutlass_fp4_scheme, ) assert input_tfms or output_tfms if is_qutlass_fp4_scheme(quant_scheme, input_tfms): return QutlassNvFP4LinearMethod(quant_method, input_tfms, output_tfms) # hadacore or dense gemm is selected by Transform module return cls(quant_method, input_tfms, output_tfms) def __init__( self, quant_method: LinearMethodBase, input_tfms: dict[int, TransformTuple], output_tfms: dict[int, TransformTuple], ): self.quant_method = quant_method self.input_tfms = input_tfms self.output_tfms = output_tfms self.input_transform: HadamardTransform | None = None self.output_transform: HadamardTransform | None = None def create_weights( self, layer: torch.nn.Module, input_size_per_partition: int, output_partition_sizes: list[int], input_size: int, output_size: int, params_dtype: torch.dtype, **extra_weight_attrs, ): # get weight loader for transforms weight_loader: Callable = extra_weight_attrs.get("weight_loader") # type: ignore[assignment] # HACK: UnquantizedLinearMethod does not support weight loader v2, but # transforms (specifically SharedWeightParameter) requires # weight loader v2. Until UnquantizedLinearMethod supports v2, we must # hack around this by getting weight loader v1 so ULM can load correctly quant_method_name = self.quant_method.__class__.__name__ if quant_method_name not in WEIGHT_LOADER_V2_SUPPORTED: weight_loader_v1 = layer.weight_loader extra_weight_attrs["weight_loader"] = weight_loader_v1 self.quant_method.create_weights( layer=layer, input_size_per_partition=input_size_per_partition, output_partition_sizes=output_partition_sizes, input_size=input_size, output_size=output_size, params_dtype=params_dtype, **extra_weight_attrs, ) # validate schemes num_partitions = len(output_partition_sizes) self._validate_tfm_schemes(num_partitions) # create submodules for weight loading if len(self.input_tfms) > 0: scheme_name = list(self.input_tfms.values())[0].scheme_name location = list(self.input_tfms.values())[0].args.location transform_name = f"{scheme_name}_{location}" transform = HadamardTransform( self.input_tfms, layer, weight_loader, input_size_per_partition, output_partition_sizes, ) layer.register_module(transform_name, transform) self.input_transform = transform if len(self.output_tfms) > 0: scheme_name = list(self.output_tfms.values())[0].scheme_name location = list(self.output_tfms.values())[0].args.location transform_name = f"{scheme_name}_{location}" transform = HadamardTransform( self.output_tfms, layer, weight_loader, input_size_per_partition, output_partition_sizes, ) layer.register_module(transform_name, transform) self.output_transform = transform # compute partition ranges for slicing activations starts = [0] + list(accumulate(output_partition_sizes))[:-1] self.partition_ranges = list(zip(starts, output_partition_sizes)) def process_weights_after_loading(self, layer): self.quant_method.process_weights_after_loading(layer) for submodule in layer.children(): if isinstance(submodule, HadamardTransform): submodule.process_weights_after_loading() def apply( self, layer: torch.nn.Module, x: torch.Tensor, bias: torch.Tensor | None = None, ) -> torch.Tensor: if self.input_transform is not None: x = self.input_transform(x) assert bias is None x = self.quant_method.apply(layer, x, bias) # In most cases, input transforms are preferred over output transforms # (@ksayers): confirm that this is done concurrently if self.output_transform is not None: for part_id, (start, length) in enumerate(self.partition_ranges): x[:, start : start + length] = self.output_transform( x[:, start : start + length].clone(), part_id=part_id ) return x def _validate_tfm_schemes(self, num_partitions: int): if len(self.input_tfms) > 0: if 0 not in self.input_tfms: raise ValueError("Must have same input") for part_index in range(num_partitions): if self.input_tfms[part_index] != self.input_tfms[0]: raise ValueError("Must have same input") if len(self.output_tfms) > 0: scheme_name = list(self.output_tfms.values())[0].scheme_name location = list(self.output_tfms.values())[0].args.location for tfm in self.output_tfms.values(): if tfm.scheme_name != scheme_name: raise ValueError("Must have same scheme name") if tfm.args.location != location: raise ValueError("Must have same location") return self.input_tfms, self.output_tfms def get_linear_transform_schemes( layer: torch.nn.Module, layer_name: str, transform_config: TransformConfig | None, packed_modules_mapping: dict[str, list[str]], ) -> tuple[ dict[int, TransformTuple], dict[int, TransformTuple] ]: # [input_transform, [output_transform, ...]] # there can only be one transform input scheme per (fused) module input_tfms = {} output_tfms = {} partition_names = get_layer_partition_names(layer_name, packed_modules_mapping) for scheme_name, scheme, args in get_schemes_args(transform_config): for part_index, part_name in enumerate(partition_names): if ( is_match(part_name, layer, args.targets, args.ignore) and args.is_online() ): if args.location == TransformLocation.INPUT: input_tfms[part_index] = TransformTuple(scheme_name, scheme, args) elif args.location == TransformLocation.OUTPUT: output_tfms[part_index] = TransformTuple(scheme_name, scheme, args) else: raise ValueError( f"Cannot apply `{args.location}` transform to `{layer_name}`" ) return (input_tfms, output_tfms) def get_schemes_args( transform_config: TransformConfig | None, ) -> Generator[tuple[str, TransformScheme, TransformArgs]]: if transform_config is None: return for scheme_name, scheme in transform_config.config_groups.items(): for args in scheme.apply: yield (scheme_name, scheme, args) def get_layer_partition_names( layer_name: str, packed_modules_mapping: dict[str, list[str]] ) -> list[str]: """ Get all partition names associated with this layer. Names are returned in order of their partition indices. ```python mapping = {"gate_up_proj", "gate_proj", "up_proj"} assert get_layer_partition_names("mlp.gate_up_proj", mapping) == [ "gate_proj", "up_proj", ] assert get_layer_partition_names("mlp.down_proj", mapping) == ["down_proj"]""" for fused_suffix, part_suffixes in packed_modules_mapping.items(): if layer_name.endswith(fused_suffix): return [ layer_name.removesuffix(fused_suffix) + part_suffix for part_suffix in part_suffixes ] return [layer_name]
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/compressed_tensors/transform/linear.py", "license": "Apache License 2.0", "lines": 214, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/quantization/compressed_tensors/transform/module.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math from collections.abc import Callable, Hashable import torch from compressed_tensors.transform import ( TransformArgs, TransformLocation, TransformScheme, ) from torch import Tensor import vllm._custom_ops as ops from vllm.distributed.parallel_state import get_tensor_model_parallel_world_size from vllm.model_executor.layers.linear import LinearBase from vllm.model_executor.layers.quantization.compressed_tensors.transform.utils import ( # noqa: E501 TransformTuple, ) from vllm.model_executor.layers.utils import dispatch_unquantized_gemm from vllm.model_executor.layers.vocab_parallel_embedding import VocabParallelEmbedding from vllm.model_executor.parameter import SharedWeightParameter class HadamardTransform(torch.nn.Module): """ Class which handles weight loading, postprocessing, and application of transforms. Meant to be used with `CompressedTensorsLinearTransformMethod` and attention transforms method (not implemented yet) """ transforms: dict[int, TransformTuple] # info parsed from transforms config weight: SharedWeightParameter # container for shared tensors scales: dict[int, float] # hadamard scale, usually sqrt(matrix.size(0)) def __init__( self, transforms: dict[int, TransformTuple], layer: torch.nn.Module, weight_loader: Callable, input_size_per_partition: int, output_partition_sizes: list[int], ): super().__init__() self.transforms = transforms self.scales = {} if get_tensor_model_parallel_world_size() > 1: raise NotImplementedError( "Online transforms with tensor parallelism is not supported" ) # Similar to row/col parallel params, but tensors are separate # to allow for loading with shared memory self.weight = SharedWeightParameter(weight_loader=weight_loader) # create shared partition data for each partition of the original weight input_size = input_size_per_partition for part_index, (_scheme_name, scheme, args) in self.transforms.items(): output_size = output_partition_sizes[part_index] weight_size = self._get_weight_size( layer, scheme, args, input_size, output_size ) data_key = self._get_data_key(scheme, weight_size) self.weight.add_partition( part_index, data_key, size=(weight_size, weight_size), dtype=scheme.precision, ) # validate that shared tensors and schemes are correct self._validate_input_transforms() def process_weights_after_loading(self): for part_id in self.weight.partitions: data = self.weight.partitions[part_id].data # required by torch.compile self.weight.process_weights_after_loading() # precompute scale as a runtime multiply, not division # do not fold into weight in order to utilize FWHT self.scales[part_id] = 1 / math.sqrt(data.size(0)) # FUTURE: avoid runtime transpose by processing weights # prior to apply def forward(self, value: Tensor, part_id: int = 0) -> Tensor: if part_id not in self.weight.partitions: return value # use hadacore if possible if self.transforms[part_id].scheme.type == "hadamard": if self.transforms[part_id].scheme.head_dim is not None: weight_size = self.transforms[part_id].scheme.head_dim value = value.unflatten(-1, (-1, weight_size)) value = ops.hadacore_transform(value) value = value.flatten(-2, -1) return value # sylvester transforms are symmetric, inv => transpose => original return ops.hadacore_transform(value) # fall back to dense else: weight = self.weight.partitions[part_id] weight = ( weight if self.transforms[part_id].args.inverse else weight.T ) # linear := x(W.T) scale = self.scales[part_id] if self.transforms[part_id].scheme.head_dim is not None: value = value.unflatten(-1, (-1, weight.size(0))) value = ( dispatch_unquantized_gemm()( self, value.to(weight.dtype), weight, None ).to(value.dtype) * scale ) value = value.flatten(-2, -1) return value return ( dispatch_unquantized_gemm()( self, value.to(weight.dtype), weight, None ).to(value.dtype) * scale ) def _get_data_key(self, scheme: TransformScheme, weight_size: int) -> Hashable: return (id(scheme), weight_size) def _get_weight_size( self, layer: torch.nn.Module, scheme: TransformScheme, args: TransformArgs, input_size: int, output_size: int, ) -> int: if scheme.head_dim is not None: return scheme.head_dim if isinstance(layer, LinearBase): if args.location == TransformLocation.INPUT: return input_size elif args.location == TransformLocation.OUTPUT: return output_size elif isinstance(layer, VocabParallelEmbedding): if args.location == TransformLocation.INPUT: return output_size elif args.location == TransformLocation.OUTPUT: return input_size raise ValueError() def _validate_input_transforms(self): assert len(self.transforms) > 0 location = list(self.transforms.values())[0].args.location if location == TransformLocation.INPUT: first_data = self.weight.partitions[0].data for partition in self.weight.partitions.values(): if partition.data.data_ptr() != first_data.data_ptr(): raise ValueError("")
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/compressed_tensors/transform/module.py", "license": "Apache License 2.0", "lines": 140, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/quantization/compressed_tensors/transform/schemes/linear_qutlass_nvfp4.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.model_executor.layers.quantization.compressed_tensors.compressed_tensors import ( # noqa: E501 CompressedTensorsScheme, CompressedTensorsW4A4Fp4, ) from vllm.model_executor.layers.quantization.compressed_tensors.transform.linear import ( # noqa: E501 CompressedTensorsLinearTransformMethod, TransformTuple, ) __all__ = ["is_qutlass_fp4_scheme", "QutlassNvFP4LinearMethod"] def is_qutlass_fp4_scheme( quant_scheme: CompressedTensorsScheme | None, input_tfms: dict[int, TransformTuple], ) -> bool: return ( isinstance(quant_scheme, (CompressedTensorsW4A4Fp4,)) and len(input_tfms) == 1 and input_tfms[0].scheme.head_dim == quant_scheme.group_size ) class QutlassNvFP4LinearMethod(CompressedTensorsLinearTransformMethod): def create_weights( self, layer, input_size_per_partition, output_partition_sizes, input_size, output_size, params_dtype, **extra_weight_attrs, ): # initializes fp4 qparams assert isinstance(layer.scheme, (CompressedTensorsW4A4Fp4,)) ret = super().create_weights( layer, input_size_per_partition, output_partition_sizes, input_size, output_size, params_dtype, **extra_weight_attrs, ) assert self.input_transform is not None assert len(self.input_transform.weight) == 1 assert self.input_transform.weight[0].size(0) == layer.scheme.group_size return ret def apply( self, layer: torch.nn.Module, x: torch.Tensor, bias: torch.Tensor | None = None, ) -> torch.Tensor: raise NotImplementedError()
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/compressed_tensors/transform/schemes/linear_qutlass_nvfp4.py", "license": "Apache License 2.0", "lines": 54, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/quantization/compressed_tensors/transform/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import NamedTuple from compressed_tensors.transform import TransformArgs, TransformScheme __all__ = ["TransformTuple"] class TransformTuple(NamedTuple): scheme_name: str scheme: TransformScheme args: TransformArgs
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/compressed_tensors/transform/utils.py", "license": "Apache License 2.0", "lines": 9, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/fused_moe/trtllm_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm.model_executor.layers.fused_moe.activation import MoEActivation from vllm.model_executor.layers.fused_moe.config import ( FusedMoEConfig, FusedMoEParallelConfig, FusedMoEQuantConfig, ) from vllm.model_executor.layers.fused_moe.topk_weight_and_reduce import ( TopKWeightAndReduceNoOP, ) from vllm.model_executor.layers.quantization.utils.quant_utils import ( QuantKey, ) class TrtLlmGenExperts(mk.FusedMoEPermuteExpertsUnpermute): """TensorRT-LLM-based fused MoE expert implementation.""" def __init__( self, moe_config: FusedMoEConfig, quant_config: FusedMoEQuantConfig, max_capture_size, ): super().__init__(moe_config, quant_config) self.device = torch.cuda.current_device() self.num_experts = moe_config.num_local_experts self.gemm1_alpha = torch.tensor( [1.702] * self.num_experts, dtype=torch.float32, device=self.device ) self.gemm1_beta = torch.tensor( [1.0] * self.num_experts, dtype=torch.float32, device=self.device ) self.gemm1_clamp_limit = torch.tensor( [7.0] * self.num_experts, dtype=torch.float32, device=self.device ) self.max_capture_size = max_capture_size @staticmethod def activation_format() -> mk.FusedMoEActivationFormat: return mk.FusedMoEActivationFormat.Standard @staticmethod def _supports_current_device() -> bool: raise NotImplementedError( "TrtLlmGenExperts is not yet used by an Oracle. " "This method should not be called." ) @staticmethod def _supports_no_act_and_mul() -> bool: raise NotImplementedError( "TrtLlmGenExperts is not yet used by an Oracle. " "This method should not be called." ) @staticmethod def _supports_quant_scheme( weight_key: QuantKey | None, activation_key: QuantKey | None, ) -> bool: raise NotImplementedError( "TrtLlmGenExperts is not yet used by an Oracle. " "This method should not be called." ) @staticmethod def _supports_activation(activation: MoEActivation) -> bool: raise NotImplementedError( "TrtLlmGenExperts is not yet used by an Oracle. " "This method should not be called." ) @staticmethod def _supports_parallel_config(moe_parallel_config: FusedMoEParallelConfig) -> bool: raise NotImplementedError( "TrtLlmGenExperts is not yet used by an Oracle. " "This method should not be called." ) def supports_chunking(self) -> bool: return True def supports_expert_map(self) -> bool: return True def finalize_weight_and_reduce_impl(self) -> mk.TopKWeightAndReduce: return TopKWeightAndReduceNoOP() def workspace_shapes( self, M: int, N: int, K: int, topk: int, global_num_experts: int, local_num_experts: int, expert_tokens_meta: mk.ExpertTokensMetadata | None, activation: MoEActivation, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: # The workspaces for this implementation are managed by flashinfer. workspace1 = (0,) workspace2 = (0,) output = (M, K) return (workspace1, workspace2, output) def apply( self, output: torch.Tensor, hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, activation: MoEActivation, global_num_experts: int, expert_map: torch.Tensor | None, a1q_scale: torch.Tensor | None, a2_scale: torch.Tensor | None, workspace13: torch.Tensor, workspace2: torch.Tensor, expert_tokens_meta: mk.ExpertTokensMetadata | None, apply_router_weight_on_input: bool, ): topk = topk_ids.size(-1) local_num_experts = w1.size(0) intermediate_size = w2.size(1) local_expert_offset = self.moe_config.ep_rank * local_num_experts x_quant = hidden_states x_scale = a1q_scale if x_scale is not None: x_scale = x_scale.view(torch.float8_e4m3fn).reshape(*x_quant.shape[:-1], -1) packed_tensor = (topk_ids.to(torch.int32) << 16) | topk_weights.to( torch.bfloat16 ).view(torch.int16) assert self.w1_scale is not None assert self.w2_scale is not None kwargs = { "topk_ids": packed_tensor, "routing_bias": None, "hidden_states": x_quant, "hidden_states_scale": x_scale, "gemm1_weights": w1, "gemm1_weights_scale": self.w1_scale, "gemm1_bias": self.w1_bias, "gemm1_alpha": self.gemm1_alpha, "gemm1_beta": self.gemm1_beta, "gemm1_clamp_limit": self.gemm1_clamp_limit, "gemm2_weights": w2, "gemm2_weights_scale": self.w2_scale, "gemm2_bias": self.w2_bias, "output1_scale_scalar": None, "output1_scale_gate_scalar": None, "output2_scale_scalar": None, "num_experts": global_num_experts, "top_k": topk, "n_group": None, "topk_group": None, "intermediate_size": intermediate_size, "local_expert_offset": local_expert_offset, "local_num_experts": local_num_experts, "routed_scaling_factor": None, "routing_method_type": 1, "do_finalize": True, "output": output, "tune_max_num_tokens": max(self.max_capture_size, 1), } from flashinfer import trtllm_fp4_block_scale_routed_moe from vllm.utils.flashinfer import autotune with autotune(False): # Enable autotune when, # https://github.com/flashinfer-ai/flashinfer/issues/2023 is # resolved. trtllm_fp4_block_scale_routed_moe(**kwargs) return output
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/fused_moe/trtllm_moe.py", "license": "Apache License 2.0", "lines": 164, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/quantization/utils/mxfp8_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from enum import Enum import torch from vllm.logger import init_logger from vllm.utils import flashinfer as vllm_flashinfer from vllm.utils.torch_utils import direct_register_custom_op logger = init_logger(__name__) class Mxfp8LinearBackend(Enum): EMULATION = "emulation" FLASHINFER_CUTLASS = "flashinfer-cutlass" # MXFP8 constants MXFP8_VALUE_DTYPE = torch.float8_e4m3fn MXFP8_SCALE_DTYPE = torch.uint8 MXFP8_BLOCK_SIZE = 32 def swizzle_mxfp8_scale(sf: torch.Tensor, M: int, K: int) -> torch.Tensor: """Swizzle MXFP8 scales from row-major 2D to F8_128x4 layout.""" scaling_vector_size = MXFP8_BLOCK_SIZE # 32 for MXFP8 factor = scaling_vector_size * 4 # 128 num_m_tiles = (M + 127) // 128 num_k_tiles = (K + factor - 1) // factor m_padded = num_m_tiles * 128 k_scale_padded = num_k_tiles * 4 scale_cols = K // scaling_vector_size sf_padded = torch.zeros( (m_padded, k_scale_padded), dtype=sf.dtype, device=sf.device ) sf_padded[:M, :scale_cols] = sf sf_reshaped = sf_padded.view(num_m_tiles, 4, 32, num_k_tiles, 4) sf_swizzled = sf_reshaped.transpose(1, 3) return sf_swizzled.contiguous().view(-1) def _mxfp8_e4m3_quantize_impl( x: torch.Tensor, is_sf_swizzled_layout: bool = False ) -> tuple[torch.Tensor, torch.Tensor]: from flashinfer import mxfp8_quantize as flashinfer_mxfp8_quantize x_q, x_scales = flashinfer_mxfp8_quantize( x, is_sf_swizzled_layout=is_sf_swizzled_layout ) if x_scales.ndim == 1 and x.ndim == 2 and not is_sf_swizzled_layout: x_scales = x_scales.view(x.size(0), -1) return x_q, x_scales def mxfp8_e4m3_quantize( x: torch.Tensor, is_sf_swizzled_layout: bool = False ) -> tuple[torch.Tensor, torch.Tensor]: return torch.ops.vllm.mxfp8_quantize(x, is_sf_swizzled_layout) def dequant_mxfp8_to_bf16(x: torch.Tensor, scales: torch.Tensor) -> torch.Tensor: """Dequantize MXFP8 tensor to BF16.""" x_float = x.to(torch.float32) num_blocks = x.shape[-1] // MXFP8_BLOCK_SIZE x_blocked = x_float.view(*x.shape[:-1], num_blocks, MXFP8_BLOCK_SIZE) descale = torch.exp2(scales.to(torch.float32) - 127.0) dequantized = x_blocked * descale.unsqueeze(-1) dequantized = dequantized.view(*x.shape) return dequantized.to(torch.bfloat16) def mxfp8_e4m3_quantize_fake( x: torch.Tensor, is_sf_swizzled_layout: bool = False ) -> tuple[torch.Tensor, torch.Tensor]: """Fake implementation for torch.compile tracing.""" fp_data = torch.empty_like(x, dtype=MXFP8_VALUE_DTYPE) block_size = MXFP8_BLOCK_SIZE if x.ndim == 2: M, N = x.shape K = (N + block_size - 1) // block_size if is_sf_swizzled_layout: M_padded = ((M + 127) // 128) * 128 K_padded = ((K + 3) // 4) * 4 scales = torch.empty( M_padded * K_padded, dtype=MXFP8_SCALE_DTYPE, device=x.device ) else: scales = torch.empty((M, K), dtype=MXFP8_SCALE_DTYPE, device=x.device) elif x.ndim == 3: B, M, N = x.shape K = (N + block_size - 1) // block_size if is_sf_swizzled_layout: M_padded = ((M + 127) // 128) * 128 K_padded = ((K + 3) // 4) * 4 scales = torch.empty( B * M_padded * K_padded, dtype=MXFP8_SCALE_DTYPE, device=x.device ) else: scales = torch.empty((B, M, K), dtype=MXFP8_SCALE_DTYPE, device=x.device) else: scale_shape = list(x.shape) scale_shape[-1] = (x.shape[-1] + block_size - 1) // block_size scales = torch.empty(scale_shape, dtype=MXFP8_SCALE_DTYPE, device=x.device) return fp_data, scales direct_register_custom_op( op_name="mxfp8_quantize", op_func=_mxfp8_e4m3_quantize_impl, fake_impl=mxfp8_e4m3_quantize_fake, ) class Mxfp8LinearOp: def __init__(self, backend: Mxfp8LinearBackend): if backend not in Mxfp8LinearBackend: raise ValueError(f"Unsupported backend: {backend}") self.backend = backend def _apply_emulation( self, input: torch.Tensor, weight: torch.Tensor, weight_scale: torch.Tensor, out_dtype: torch.dtype, bias: torch.Tensor | None = None, ) -> torch.Tensor: # Validate weight_scale dtype and shape (must be 2D for TORCH backend) if weight_scale.dtype != MXFP8_SCALE_DTYPE: raise ValueError( f"TORCH backend requires {MXFP8_SCALE_DTYPE} weight_scale dtype, " f"got {weight_scale.dtype}." ) if weight_scale.ndim != 2: raise ValueError( f"TORCH backend requires 2D weight_scale, got {weight_scale.ndim}D. " f"Ensure process_weights_after_loading was called." ) weight_bf16 = dequant_mxfp8_to_bf16(weight, weight_scale) output = torch.nn.functional.linear(input, weight_bf16, bias) return output.to(out_dtype) def _apply_flashinfer_cutlass( self, input: torch.Tensor, weight: torch.Tensor, weight_scale: torch.Tensor, out_dtype: torch.dtype, bias: torch.Tensor | None = None, ) -> torch.Tensor: N, K = weight.shape input_shape = input.shape input_2d = input.view(-1, K) M_orig = input_2d.shape[0] # Minimum dimension size for F8_128x4 block scaling layout min_dim = 128 assert min_dim <= K, ( f"mm_mxfp8 requires K >= {min_dim}, got K={K}. " f"in_features is too small for mm_mxfp8." ) assert K % MXFP8_BLOCK_SIZE == 0, ( f"mm_mxfp8 requires K to be divisible by {MXFP8_BLOCK_SIZE}, got K={K}." ) assert min_dim <= N, ( f"mm_mxfp8 requires N >= {min_dim}, got N={N}. " f"out_features is too small for mm_mxfp8." ) M_padded = ((M_orig + min_dim - 1) // min_dim) * min_dim if M_padded != M_orig: pad_rows = M_padded - M_orig input_2d = torch.nn.functional.pad(input_2d, (0, 0, 0, pad_rows)) input_mxfp8, input_scale = mxfp8_e4m3_quantize( input_2d, is_sf_swizzled_layout=True, # Swizzled for best accuracy ) if not weight.is_contiguous(): weight = weight.contiguous() output = vllm_flashinfer.mm_mxfp8( input_mxfp8, weight.t(), input_scale, weight_scale, out_dtype=out_dtype, backend="cutlass", ) if M_padded != M_orig: output = output[:M_orig, :] if bias is not None: output = output + bias output_shape = (*input_shape[:-1], N) return output.view(output_shape) def apply( self, input: torch.Tensor, weight: torch.Tensor, weight_scale: torch.Tensor, out_dtype: torch.dtype, bias: torch.Tensor | None = None, ) -> torch.Tensor: if self.backend == Mxfp8LinearBackend.EMULATION: return self._apply_emulation(input, weight, weight_scale, out_dtype, bias) assert self.backend == Mxfp8LinearBackend.FLASHINFER_CUTLASS return self._apply_flashinfer_cutlass( input, weight, weight_scale, out_dtype, bias )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/utils/mxfp8_utils.py", "license": "Apache License 2.0", "lines": 182, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/rotary_embedding/ernie45_vl_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from .mrope import MRotaryEmbedding class Ernie4_5_VLRotaryEmbedding(MRotaryEmbedding): """3D rotary positional embedding. 3D is t:time h:height w:width""" def forward_native( # type: ignore[override] self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: assert positions.ndim == 1 or positions.ndim == 2 assert key is not None num_tokens = positions.shape[-1] cos_sin = self.cos_sin_cache[positions] cos, sin = cos_sin.chunk(2, dim=-1) if positions.ndim == 2: assert self.mrope_section section_h = self.mrope_section[0] # 22 section_w = self.mrope_section[1] # 22 section_t = self.mrope_section[2] # 20 assert section_h == section_w # Split according to [h w h w h w h w... t t t...] section_cos_t = cos[..., -section_t:] section_cos_h = cos[..., : section_h + section_w : 2] section_cos_w = cos[..., 1 : section_h + section_w : 2] cos_t, cos_h, cos_w = section_cos_t[0], section_cos_h[1], section_cos_w[2] cos_hw = torch.stack([cos_h, cos_w], dim=-1).reshape( cos_h.shape[:-1] + (cos_h.shape[-1] * 2,) ) cos = torch.cat([cos_hw, cos_t], dim=-1) section_sin_t = sin[..., -section_t:] section_sin_h = sin[..., : section_h + section_w : 2] section_sin_w = sin[..., 1 : section_h + section_w : 2] sin_t, sin_h, sin_w = section_sin_t[0], section_sin_h[1], section_sin_w[2] sin_hw = torch.stack([sin_h, sin_w], dim=-1).reshape( sin_h.shape[:-1] + (sin_h.shape[-1] * 2,) ) sin = torch.cat([sin_hw, sin_t], dim=-1) query_shape = query.shape query = query.view(num_tokens, -1, self.head_size) query_rot = query[..., : self.rotary_dim] query_pass = query[..., self.rotary_dim :] query_rot = self.apply_rotary_emb.forward_native( query_rot, cos, sin, ) query = torch.cat((query_rot, query_pass), dim=-1).reshape(query_shape) key_shape = key.shape key = key.view(num_tokens, -1, self.head_size) key_rot = key[..., : self.rotary_dim] key_pass = key[..., self.rotary_dim :] key_rot = self.apply_rotary_emb.forward_native( key_rot, cos, sin, ) key = torch.cat((key_rot, key_pass), dim=-1).reshape(key_shape) return query, key def forward_cuda( # type: ignore[override] self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: return self.forward_native(positions, query, key)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/rotary_embedding/ernie45_vl_rope.py", "license": "Apache License 2.0", "lines": 68, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/ernie45_vl.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright 2025 The Baidu team. # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Inference-only Ernie VL model compatible with HuggingFace weights.""" import itertools import math from collections.abc import Callable, Iterable, Mapping, Sequence from functools import partial from typing import Annotated, Any, Literal import numpy as np import torch import torch.nn as nn import torch.nn.functional as F from einops import rearrange from transformers import BaseImageProcessor, BatchFeature from vllm.config import VllmConfig from vllm.config.multimodal import BaseDummyOptions, VideoDummyOptions from vllm.distributed import parallel_state from vllm.distributed import utils as dist_utils from vllm.logger import init_logger from vllm.model_executor.layers.activation import QuickGELU from vllm.model_executor.layers.attention import ( MMEncoderAttention, ) from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( ColumnParallelLinear, QKVParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.rotary_embedding.common import ( ApplyRotaryEmb, ) from vllm.model_executor.model_loader.weight_utils import default_weight_loader from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import ( MultiModalDataDict, MultiModalFeatureSpec, MultiModalFieldConfig, MultiModalKwargsItems, ) from vllm.multimodal.parse import ImageSize, MultiModalDataItems, MultiModalDataParser from vllm.multimodal.processing import ( BaseDummyInputsBuilder, BaseMultiModalProcessor, BaseProcessingInfo, PromptReplacement, PromptUpdate, ) from vllm.sequence import IntermediateTensors from vllm.utils.tensor_schema import TensorSchema, TensorShape from vllm.v1.attention.backends.registry import AttentionBackendEnum from .ernie45_vl_moe import Ernie4_5_VLMoeForCausalLM from .interfaces import ( MultiModalEmbeddings, SupportsLoRA, SupportsMRoPE, SupportsMultiModal, SupportsPP, ) from .utils import AutoWeightsLoader, WeightsMapper, maybe_prefix from .vision import get_vit_attn_backend logger = init_logger(__name__) # === Vision Transformer === # def all_gather_interleave(local_tensor, hidden_size: int, tp_size: int): """All-gather the input tensor interleavely across model parallel group.""" import torch.distributed as dist gathered_tensors = [torch.zeros_like(local_tensor) for _ in range(tp_size)] dist.all_gather( gathered_tensors, local_tensor, group=parallel_state.get_tp_group().device_group ) gathered_tensors_split = [ torch.split(tensor, hidden_size // tp_size, -1) for tensor in gathered_tensors ] ordered_tensors = [ tensor for pair in zip(*gathered_tensors_split) for tensor in pair ] result_tensor = torch.cat(ordered_tensors, dim=-1) return result_tensor class Ernie4_5_VisionAttention(nn.Module): """VisionAttention using VLLM framework APIs""" def __init__( self, embed_dim: int, num_heads: int, projection_size: int, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() # Per attention head and per partition values. self.tp_size = parallel_state.get_tensor_model_parallel_world_size() self.tp_rank = parallel_state.get_tensor_model_parallel_rank() self.hidden_size_per_attention_head = dist_utils.divide( projection_size, num_heads ) self.num_attention_heads_per_partition = dist_utils.divide( num_heads, self.tp_size ) self.qkv = QKVParallelLinear( hidden_size=embed_dim, head_size=self.hidden_size_per_attention_head, total_num_heads=num_heads, total_num_kv_heads=num_heads, bias=True, quant_config=quant_config, prefix=f"{prefix}.qkv", ) self.proj = RowParallelLinear( input_size=projection_size, output_size=embed_dim, quant_config=quant_config, prefix=f"{prefix}.proj", ) self.attn = MMEncoderAttention( num_heads=self.num_attention_heads_per_partition, head_size=self.hidden_size_per_attention_head, scale=self.hidden_size_per_attention_head**-0.5, prefix=f"{prefix}.attn", ) self.apply_rotary_emb = ApplyRotaryEmb( enforce_enable=True, enable_fp32_compute=True, ) def split_qkv(self, qkv: torch.Tensor) -> tuple[torch.Tensor, ...]: # [s, b, 3 * head * head_dim] seq_len, bs, _ = qkv.shape if self.tp_size > 1: qkv = all_gather_interleave(qkv, self.qkv.hidden_size, self.tp_size) # [s, b, 3 * head * head_dim] -> 3 * [s, b, head * head_dim] q, k, v = qkv.chunk(3, dim=2) # 3 * [s, b, head * head_dim] if self.tp_size > 1: splitter = partial( dist_utils.split_tensor_along_last_dim, num_partitions=self.tp_size ) q = splitter(q)[self.tp_rank] k = splitter(k)[self.tp_rank] v = splitter(v)[self.tp_rank] # 3 * [s, b, head * head_dim] -> 3 * [s, b, head, head_dim] new_shape = ( seq_len, bs, self.num_attention_heads_per_partition, self.hidden_size_per_attention_head, ) q, k, v = (x.view(*new_shape) for x in (q, k, v)) return q, k, v def forward( self, x: torch.Tensor, cu_seqlens: torch.Tensor, rotary_pos_emb: torch.Tensor, max_seqlen: torch.Tensor | None = None, # Only used for Flash Attention ) -> torch.Tensor: # [s, b, c] --> [s, b, head * 3 * head_dim] x, _ = self.qkv(x) # [s, b, 3 * head * head_dim] -> 3 * [s, b, head, head_dim] q, k, v = self.split_qkv(x) q, k, v = (rearrange(x, "s b ... -> b s ...").contiguous() for x in (q, k, v)) if rotary_pos_emb is not None: qk_concat = torch.cat([q, k], dim=0) qk_rotated = self.apply_rotary_emb( qk_concat, rotary_pos_emb.cos(), rotary_pos_emb.sin(), ) q, k = torch.chunk(qk_rotated, 2, dim=0) output = self.attn( query=q, key=k, value=v, cu_seqlens=cu_seqlens, max_seqlen=max_seqlen, ) context_layer = rearrange(output, "b s h d -> s b (h d)").contiguous() output, _ = self.proj(context_layer) return output class Ernie4_5_VisionMLP(nn.Module): def __init__( self, in_features: int, hidden_features: int, act_layer: type[nn.Module] = QuickGELU, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.fc1 = ColumnParallelLinear( in_features, hidden_features, quant_config=quant_config, prefix=f"{prefix}.fc1", ) self.act = act_layer() self.fc2 = RowParallelLinear( hidden_features, in_features, quant_config=quant_config, prefix=f"{prefix}.fc2", ) def forward(self, x: torch.Tensor) -> torch.Tensor: x_parallel, _ = self.fc1(x) x_parallel = self.act(x_parallel) x, _ = self.fc2(x_parallel) return x class Ernie4_5_VisionBlock(nn.Module): def __init__( self, dim: int, num_heads: int, mlp_ratio: float, act_layer: type[nn.Module] = QuickGELU, norm_layer: Callable[[int], nn.Module] | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() if norm_layer is None: norm_layer = partial(nn.LayerNorm, eps=1e-6) self.norm1 = norm_layer(dim) self.norm2 = norm_layer(dim) mlp_hidden_dim = int(dim * mlp_ratio) self.attn = Ernie4_5_VisionAttention( embed_dim=dim, num_heads=num_heads, projection_size=dim, quant_config=quant_config, prefix=f"{prefix}.attn", ) self.mlp = Ernie4_5_VisionMLP( dim, mlp_hidden_dim, act_layer=act_layer, quant_config=quant_config, prefix=f"{prefix}.mlp", ) def forward( self, hidden_states: torch.Tensor, cu_seqlens: torch.Tensor, rotary_pos_emb: torch.Tensor, max_seqlen: torch.Tensor | None = None, # Only used for Flash Attention ) -> torch.Tensor: hidden_states = hidden_states + self.attn( self.norm1(hidden_states), cu_seqlens=cu_seqlens, rotary_pos_emb=rotary_pos_emb, max_seqlen=max_seqlen, ) hidden_states = hidden_states + self.mlp(self.norm2(hidden_states)) return hidden_states class Ernie4_5_VisionPatchEmbed(nn.Module): def __init__( self, patch_size: int = 14, in_channels: int = 3, embed_dim: int = 1280, prefix="", ) -> None: super().__init__() self.patch_size = patch_size self.in_channels = in_channels self.embed_dim = embed_dim self.proj = nn.Linear( in_channels * patch_size * patch_size, embed_dim, bias=False ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: target_dtype = self.proj.weight.dtype hidden_states = hidden_states.to(target_dtype) hidden_states = self.proj(hidden_states) return hidden_states class Ernie4_5_VisionRotaryEmbedding(nn.Module): def __init__(self, dim: int, theta: float = 10000.0) -> None: super().__init__() self.inv_freq = 1.0 / theta ** ( torch.arange(start=0, end=dim, step=2, dtype=torch.float32) / dim ) def forward(self, seqlen: int) -> torch.Tensor: seq = torch.arange( seqlen, device=self.inv_freq.device, dtype=self.inv_freq.dtype ) freqs = torch.outer(input=seq, vec2=self.inv_freq) return freqs class Ernie4_5_VisionTransformer(nn.Module): def __init__( self, vision_config, norm_eps: float = 1e-6, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() patch_size = vision_config.patch_size spatial_merge_size = vision_config.spatial_merge_size in_channels = vision_config.in_channels hidden_size = vision_config.hidden_size embed_dim = vision_config.embed_dim depth = vision_config.depth num_heads = vision_config.num_heads mlp_ratio = vision_config.mlp_ratio self.spatial_merge_size = spatial_merge_size self.num_heads = num_heads self.embed_dim = embed_dim self.patch_embed = Ernie4_5_VisionPatchEmbed( patch_size=patch_size, in_channels=in_channels, embed_dim=embed_dim, prefix=f"{prefix}.patch_embed", ) norm_layer = partial(nn.LayerNorm, eps=norm_eps) head_dim = embed_dim // num_heads self.rotary_pos_emb = Ernie4_5_VisionRotaryEmbedding(head_dim // 2) self.blocks = nn.ModuleList( [ Ernie4_5_VisionBlock( dim=embed_dim, num_heads=num_heads, mlp_ratio=mlp_ratio, norm_layer=norm_layer, quant_config=quant_config, prefix=f"{prefix}.blocks.{layer_idx}", ) for layer_idx in range(depth) ] ) assert hidden_size == embed_dim, ( "vit's config.hidden must be equal to config.embed_dim" ) self.ln = nn.LayerNorm(hidden_size, eps=1e-6) self.attn_backend = get_vit_attn_backend( head_size=head_dim, dtype=torch.get_default_dtype(), ) @property def dtype(self) -> torch.dtype: return self.patch_embed.proj.weight.dtype @property def device(self) -> torch.device: return self.patch_embed.proj.weight.device def rot_pos_emb(self, grid_thw: torch.Tensor) -> torch.Tensor: pos_ids = [] for t, h, w in grid_thw: hpos_ids = torch.arange(h).unsqueeze(1).expand(-1, w) wpos_ids = torch.arange(w).unsqueeze(0).expand(h, -1) hpos_ids = ( hpos_ids.reshape( h // self.spatial_merge_size, self.spatial_merge_size, w // self.spatial_merge_size, self.spatial_merge_size, ) .permute(0, 2, 1, 3) .flatten() ) wpos_ids = ( wpos_ids.reshape( h // self.spatial_merge_size, self.spatial_merge_size, w // self.spatial_merge_size, self.spatial_merge_size, ) .permute(0, 2, 1, 3) .flatten() ) pos_ids.append(torch.stack([hpos_ids, wpos_ids], dim=-1).repeat(t, 1)) pos_ids = torch.cat(pos_ids, dim=0) max_grid_size = grid_thw[:, 1:].max() rotary_pos_emb_full = self.rotary_pos_emb(max_grid_size) rotary_pos_emb = rotary_pos_emb_full[pos_ids].flatten(1) return rotary_pos_emb def compute_attn_mask_seqlen(self, cu_seqlens: torch.Tensor) -> torch.Tensor | None: max_seqlen = None if self.attn_backend in { AttentionBackendEnum.FLASH_ATTN, AttentionBackendEnum.ROCM_AITER_FA, AttentionBackendEnum.TRITON_ATTN, }: max_seqlen = (cu_seqlens[1:] - cu_seqlens[:-1]).max() return max_seqlen def forward( self, hidden_states: torch.Tensor, grid_thw: torch.Tensor, num_pad=0 ) -> torch.Tensor: hidden_states = self.patch_embed(hidden_states) rotary_pos_emb = self.rot_pos_emb(grid_thw) rotary_pos_emb = rotary_pos_emb.to(hidden_states.device) cu_seqlens = torch.repeat_interleave( grid_thw[:, 1] * grid_thw[:, 2], grid_thw[:, 0] ).cumsum(dim=0, dtype=torch.int32) zeros = cu_seqlens.new_zeros(1) if num_pad > 0: cu_seqlens = torch.cat([zeros, cu_seqlens, zeros]) cu_seqlens[-1] = cu_seqlens[-2] + num_pad else: cu_seqlens = torch.cat([zeros, cu_seqlens]) # add batch size if hidden_states.ndim == 2: hidden_states = hidden_states.unsqueeze(dim=1) # pre-compute max_seqlen for attn mask to reduce cuMemcpy operations max_seqlen = self.compute_attn_mask_seqlen(cu_seqlens) for i, blk in enumerate(self.blocks): hidden_states = blk( hidden_states, cu_seqlens=cu_seqlens, rotary_pos_emb=rotary_pos_emb, max_seqlen=max_seqlen, ) final_output = self.ln(hidden_states) if final_output.ndim == 3: final_output = final_output.squeeze(dim=1) return final_output def load_weights(self, weights) -> set[str]: params_dict = dict(self.named_parameters(remove_duplicate=False)) loaded_params: set[str] = set() for name, loaded_weight in weights: param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params # === Vision Inputs === # class Ernie4_5_VLImagePixelInputs(TensorSchema): """ Dimensions: - np: The total number of patches over each image over each prompt in the batch - ni: Number of images - cps: Number of channels * patch_size * patch_size """ type: Literal["pixel_values"] pixel_values: Annotated[torch.Tensor, TensorShape("np", "cps")] image_grid_thw: Annotated[torch.Tensor, TensorShape("ni", 3)] Ernie4_5_VLImageInputs = Ernie4_5_VLImagePixelInputs class Ernie4_5_VLVideoPixelInputs(TensorSchema): """ Dimensions: - np: The total number of patches over each image over each prompt in the batch - ni: Number of images - cps: Number of channels * temporal_patch_size * patch_size * patch_size """ type: Literal["pixel_values_videos"] pixel_values_videos: Annotated[torch.Tensor, TensorShape("np", "cps")] video_grid_thw: Annotated[torch.Tensor, TensorShape("ni", 3)] Ernie4_5_VLVideoInputs = Ernie4_5_VLVideoPixelInputs # === Vision Processor === # def round_by_factor(number: int | float, factor: int) -> int: return round(number / factor) * factor def ceil_by_factor(number: int | float, factor: int) -> int: return math.ceil(number / factor) * factor def floor_by_factor(number: int | float, factor: int) -> int: return math.floor(number / factor) * factor def smart_resize( height: int, width: int, factor: int = 28, min_pixels: int = 4 * 28 * 28, max_pixels: int = 16384 * 28 * 28, ): MAX_RATIO = 200 if max(height, width) / min(height, width) > MAX_RATIO: if height > width: new_width = max(factor, round_by_factor(width, factor)) new_height = floor_by_factor(new_width * MAX_RATIO, factor) else: new_height = max(factor, round_by_factor(height, factor)) new_width = floor_by_factor(new_height * MAX_RATIO, factor) height = new_height width = new_width h_bar = max(factor, round_by_factor(height, factor)) w_bar = max(factor, round_by_factor(width, factor)) if h_bar * w_bar > max_pixels: beta = math.sqrt((height * width) / max_pixels) h_bar = floor_by_factor(height / beta, factor) w_bar = floor_by_factor(width / beta, factor) elif h_bar * w_bar < min_pixels: beta = math.sqrt(min_pixels / (height * width)) h_bar = ceil_by_factor(height * beta, factor) w_bar = ceil_by_factor(width * beta, factor) if min_pixels > h_bar * w_bar or h_bar * w_bar > max_pixels: raise ValueError( f"Invalid h_bar={h_bar}, w_bar={w_bar}: " f"h_bar * w_bar must be >= min_pixels ({min_pixels}) " f"and <= max_pixels ({max_pixels})." ) return h_bar, w_bar class VariableResolutionResamplerModel(nn.Module): def __init__( self, in_dim, out_dim, spatial_conv_size, temporal_conv_size, config, prefix: str = "", ) -> None: super().__init__() self.in_dim = in_dim self.out_dim = out_dim self.config = config self.spatial_conv_size = spatial_conv_size self.temporal_conv_size = temporal_conv_size self.use_temporal_conv = config.use_temporal_conv # compress 2d conv(picture) to 1d self.spatial_dim = self.in_dim * self.spatial_conv_size * self.spatial_conv_size # compress 3d conv(video) to 1d self.temporal_dim = ( self.in_dim * self.spatial_conv_size * self.spatial_conv_size * self.temporal_conv_size ) self.spatial_linear1 = ColumnParallelLinear( self.spatial_dim, self.spatial_dim, bias=True, gather_output=True, quant_config=getattr(config, "quant_config", None), prefix=f"{prefix}.spatial_linear1", ) self.spatial_gelu = nn.GELU() self.spatial_linear2 = ColumnParallelLinear( self.spatial_dim, self.spatial_dim, bias=True, gather_output=True, quant_config=getattr(config, "quant_config", None), prefix=f"{prefix}.spatial_linear2", ) self.spatial_norm = nn.LayerNorm(self.spatial_dim, eps=1e-6) if self.use_temporal_conv: self.temporal_linear1 = ColumnParallelLinear( self.temporal_dim, self.spatial_dim, bias=True, gather_output=True, quant_config=getattr(config, "quant_config", None), prefix=f"{prefix}.temporal_linear1", ) self.temporal_gelu = nn.GELU() self.temporal_linear2 = ColumnParallelLinear( self.spatial_dim, self.spatial_dim, bias=True, gather_output=True, quant_config=getattr(config, "quant_config", None), prefix=f"{prefix}.temporal_linear2", ) self.temporal_norm = nn.LayerNorm(self.spatial_dim, eps=1e-6) self.mlp = ColumnParallelLinear( self.spatial_dim, self.out_dim, bias=True, gather_output=True, quant_config=getattr(config, "quant_config", None), prefix=f"{prefix}.mlp", ) self.after_norm = RMSNorm( hidden_size=out_dim, eps=getattr(config, "rms_norm_eps", 1e-6) ) def spatial_conv_reshape(self, x, spatial_conv_size): S, C = x.shape x = x.reshape([-1, C * (spatial_conv_size**2)]) return x def forward(self, x, grid_thw): def fwd_spatial(x): x = self.spatial_conv_reshape(x, self.spatial_conv_size) x, _ = self.spatial_linear1(x) x = self.spatial_gelu(x) x, _ = self.spatial_linear2(x) x = self.spatial_norm(x) return x def fwd_placeholder(x, grid_thw, to_tensor=False): grid_thw_cpu = grid_thw.cpu().numpy() grid_t, grid_hw = grid_thw_cpu[:, 0], grid_thw_cpu[:, 1:] grid_hw_after_conv = grid_hw.prod(-1) // (self.spatial_conv_size**2) tokens_per_img_or_vid = grid_thw_cpu.prod(-1) // (self.spatial_conv_size**2) batch_offset = np.empty( tokens_per_img_or_vid.size, dtype=tokens_per_img_or_vid.dtype ) batch_offset[0] = 0 batch_offset[1:] = tokens_per_img_or_vid.cumsum()[:-1] slice_offsets = [] for temporoal_size, spatial_size, b_offset in zip( grid_t, grid_hw_after_conv, batch_offset ): for temp_offset in range(0, temporoal_size, 2): slice_offsets.append( np.arange( b_offset + (temp_offset) * spatial_size, b_offset + (temp_offset + 1) * spatial_size, ) ) slice_offsets = torch.tensor(np.concatenate(slice_offsets, axis=-1)).to( x.device ) slice_offsets2 = [] for temporoal_size, spatial_size, b_offset in zip( grid_t, grid_hw_after_conv, batch_offset ): for temp_offset in range( 1 if temporoal_size > 1 else 0, temporoal_size, 2 ): slice_offsets2.append( np.arange( b_offset + (temp_offset) * spatial_size, b_offset + (temp_offset + 1) * spatial_size, ) ) slice_offsets2 = torch.tensor(np.concatenate(slice_offsets2, axis=-1)).to( x.device ) x_timestep_1 = torch.index_select(x, dim=0, index=slice_offsets) x_timestep_2 = torch.index_select(x, dim=0, index=slice_offsets2) x = torch.concat([x_timestep_1, x_timestep_2], dim=-1) return x def fwd_temporal(x): x, _ = self.temporal_linear1(x) x = self.temporal_gelu(x) x, _ = self.temporal_linear2(x) x = self.temporal_norm(x) return x def fwd_mlp(x): x, _ = self.mlp(x) x = self.after_norm(x) return x x = fwd_spatial(x) if self.use_temporal_conv: x = fwd_placeholder(x, grid_thw) x = fwd_temporal(x) x = fwd_mlp(x) return x def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: params_dict = dict(self.named_parameters(remove_duplicate=False)) loaded_params: set[str] = set() for name, loaded_weight in weights: if name not in params_dict: continue param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params class Ernie4_5_VLProcessingInfo(BaseProcessingInfo): def get_hf_config(self): return self.ctx.model_config.hf_config def get_hf_processor(self, **kwargs: object): return self.ctx.get_hf_processor(use_fast=True, **kwargs) def get_image_processor(self, **kwargs: object): return self.get_hf_processor(**kwargs).image_processor def get_data_parser(self): return MultiModalDataParser( video_needs_metadata=True, expected_hidden_size=self._get_expected_hidden_size(), ) def get_supported_mm_limits(self) -> Mapping[str, int | None]: return {"image": None, "video": None} def get_mm_max_tokens_per_item( self, seq_len: int, mm_counts: Mapping[str, int], ) -> Mapping[str, int]: max_image_tokens = self.get_max_image_tokens() max_video_tokens = self.get_max_video_tokens(seq_len, mm_counts) return {"image": max_image_tokens, "video": max_video_tokens} def _get_vision_info( self, *, image_width: int, image_height: int, num_frames: int = 1, do_resize: bool = True, image_processor: BaseImageProcessor, mm_kwargs: Mapping[str, object], ) -> tuple[ImageSize, int]: hf_config = self.get_hf_config() vision_config = hf_config.vision_config patch_size = vision_config.patch_size spatial_conv_size = hf_config.spatial_conv_size temporal_conv_size = hf_config.temporal_conv_size if self.ctx.model_config.trust_remote_code: # Defined in HF Hub repo min_pixels_key = "min_pixels" max_pixels_key = "max_pixels" else: # Defined in Transformers library (requires v5.0 or above) min_pixels_key = "shortest_edge" max_pixels_key = "longest_edge" mm_kwargs = self.ctx.get_merged_mm_kwargs(mm_kwargs) size = image_processor.size if override_size := mm_kwargs.get("size"): size = size | override_size if (override_min_pixels := mm_kwargs.get("min_pixels")) is not None: size = size | {min_pixels_key: override_min_pixels} if (override_max_pixels := mm_kwargs.get("max_pixels")) is not None: size = size | {max_pixels_key: override_max_pixels} if do_resize: resized_height, resized_width = smart_resize( height=image_height, width=image_width, factor=patch_size * spatial_conv_size, min_pixels=size[min_pixels_key], max_pixels=size[max_pixels_key], ) preprocessed_size = ImageSize(width=resized_width, height=resized_height) else: preprocessed_size = ImageSize(width=image_width, height=image_height) grid_t = max(num_frames // temporal_conv_size, 1) grid_h = preprocessed_size.height // patch_size grid_w = preprocessed_size.width // patch_size num_patches = grid_t * grid_h * grid_w num_vision_tokens = num_patches // (spatial_conv_size**2) return preprocessed_size, num_vision_tokens def get_num_image_tokens( self, *, image_width: int, image_height: int, image_processor: BaseImageProcessor, mm_kwargs: Mapping[str, object], ) -> int: _, num_image_tokens = self._get_vision_info( image_width=image_width, image_height=image_height, image_processor=image_processor, mm_kwargs=mm_kwargs, ) return num_image_tokens def get_num_video_tokens( self, *, image_width: int, image_height: int, num_frames: int, image_processor: BaseImageProcessor, mm_kwargs: Mapping[str, object], ) -> int: _, num_video_tokens = self._get_vision_info( image_width=image_width, image_height=image_height, num_frames=num_frames, image_processor=image_processor, mm_kwargs=mm_kwargs, ) return num_video_tokens def get_image_size_with_most_features(self) -> ImageSize: image_processor = self.get_image_processor() max_image_size, _ = self._get_vision_info( image_width=9999999, image_height=9999999, image_processor=image_processor, mm_kwargs={}, ) return max_image_size def get_max_image_tokens(self) -> int: image_processor = self.get_image_processor() target_width, target_height = self.get_image_size_with_most_features() num_image_tokens = self.get_num_image_tokens( image_width=target_width, image_height=target_height, image_processor=image_processor, mm_kwargs={}, ) return num_image_tokens def _get_max_video_frames(self, max_tokens: int) -> int: image_processor = self.get_image_processor() target_width, target_height = self.get_image_size_with_most_features() num_frames = 0 while True: next_num_frames = num_frames + 1 next_max_tokens = self.get_num_video_tokens( image_width=target_width, image_height=target_height, num_frames=next_num_frames, image_processor=image_processor, mm_kwargs={}, ) if next_max_tokens > max_tokens: break num_frames = next_num_frames # If the number of frames is odd, discard one frame. if num_frames % 2 != 0: num_frames -= 1 return num_frames def get_num_frames_with_most_features( self, seq_len: int, mm_counts: Mapping[str, int], ) -> int: max_images = mm_counts.get("image", 0) max_videos = mm_counts.get("video", 0) max_image_tokens = self.get_max_image_tokens() * max_images max_total_frames = self._get_max_video_frames(seq_len - max_image_tokens) max_frames_per_video = max_total_frames // max(max_videos, 1) return max(max_frames_per_video, 2) def get_max_video_tokens( self, seq_len: int, mm_counts: Mapping[str, int], ) -> int: image_processor = self.get_image_processor() target_width, target_height = self.get_image_size_with_most_features() return self.get_num_video_tokens( image_width=target_width, image_height=target_height, num_frames=self.get_num_frames_with_most_features(seq_len, mm_counts), image_processor=image_processor, mm_kwargs={}, ) class Ernie4_5VLMultiModalProcessor(BaseMultiModalProcessor[Ernie4_5_VLProcessingInfo]): def _pixel_values_norm( self, pixel_values: torch.Tensor, mm_kwargs: object, ) -> torch.Tensor: hf_config = self.info.get_hf_config() vision_config = hf_config.vision_config image_processor = self.info.get_image_processor(**mm_kwargs) image_mean_tensor = torch.tensor( image_processor.image_mean, dtype=torch.float32 ).reshape([1, 3, 1, 1]) image_std_tensor = torch.tensor( image_processor.image_std, dtype=torch.float32 ).reshape([1, 3, 1, 1]) rescale_factor = torch.tensor( image_processor.rescale_factor, dtype=torch.float32 ) patch_size_squared = vision_config.patch_size**2 image_mean_tensor = image_mean_tensor.squeeze([-2, -1]).repeat_interleave( patch_size_squared, -1 ) image_std_tensor = image_std_tensor.squeeze([-2, -1]).repeat_interleave( patch_size_squared, -1 ) if not image_mean_tensor.is_contiguous(): image_mean_tensor = image_mean_tensor.contiguous() if not image_std_tensor.is_contiguous(): image_std_tensor = image_std_tensor.contiguous() pixel_values = ( rescale_factor * pixel_values.to(torch.float32) - image_mean_tensor ) / image_std_tensor pixel_values = pixel_values.to(hf_config.dtype) return pixel_values def _call_hf_processor( self, prompt: str, mm_data: Mapping[str, object], mm_kwargs: Mapping[str, object], tok_kwargs: Mapping[str, object], ) -> BatchFeature: # when the prompt is not empty but the multimodal data is empty, # directly invoke the tokenizer. if "images" not in mm_data and "videos" not in mm_data and prompt != "": tokenizer = self.info.get_tokenizer() prompt_ids = tokenizer.encode(prompt) tokenizer_output = BatchFeature( dict(input_ids=[prompt_ids]), tensor_type="pt" ) return tokenizer_output if "images" not in mm_data: mm_data["images"] = [] if "videos" not in mm_data: mm_data["videos"] = [] # Check if HF processor supports video metadata hf_processor = self.info.get_hf_processor(**mm_kwargs) supports_video_metadata = getattr( hf_processor, "supports_video_metadata", False ) if mm_data["videos"] and not supports_video_metadata: # Old HF processor, unwrap tuple to pure frames logger.warning_once( "HF processor doesn't support video metadata. " "Timestamps will NOT be rendered. Please upgrade the model." ) mm_data["videos"] = [ v[0] if isinstance(v, tuple) else v for v in mm_data["videos"] ] processor_output = self.info.ctx.call_hf_processor( hf_processor, dict(text=[prompt], images=mm_data["images"], videos=mm_data["videos"]), dict(**mm_kwargs, **tok_kwargs), ) # Divide the processor_output into two modalities: image and video. if processor_output is not None: pixel_values = processor_output["images"] if pixel_values is not None: processor_output["images"] = self._pixel_values_norm( pixel_values, mm_kwargs ) for key in list(processor_output.keys()): if processor_output[key] is None: del processor_output[key] continue if key == "grid_thw": grid_thw = processor_output["grid_thw"] pixel_values_all = processor_output["images"] # Identify elements where the first # dimension is greater than 1 and # treat them as the video modality mask = grid_thw[:, 0] > 1 processor_output["video_grid_thw"] = grid_thw[mask] processor_output["image_grid_thw"] = grid_thw[~mask] image_patch_num = ( processor_output["image_grid_thw"].prod(dim=1).sum() ) processor_output["pixel_values"] = pixel_values_all[ :image_patch_num ] processor_output["pixel_values_videos"] = pixel_values_all[ image_patch_num: ] del processor_output["images"] return processor_output def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, Any], out_mm_kwargs: MultiModalKwargsItems, ) -> Sequence[PromptUpdate]: hf_processor = self.info.get_hf_processor(**hf_processor_mm_kwargs) before_placeholder = { "image": "<|image@placeholder|>", "video": "<|video@placeholder|>", } after_placeholder = { # image and video have same placeholder "image": "<|IMAGE_PLACEHOLDER|>", "video": "<|IMAGE_PLACEHOLDER|>", } merge_length = hf_processor.spatial_conv_size**2 def get_replacement_ernie45vl(item_idx: int, modality: str): out_item = out_mm_kwargs[modality][item_idx] grid_thw = out_item[f"{modality}_grid_thw"].data assert isinstance(grid_thw, torch.Tensor) if modality == "video": num_tokens = ( int(grid_thw.prod()) // hf_processor.temporal_conv_size // merge_length ) else: num_tokens = int(grid_thw.prod()) // merge_length return after_placeholder[modality] * num_tokens return [ PromptReplacement( modality=modality, target=before_placeholder[modality], replacement=partial(get_replacement_ernie45vl, modality=modality), ) for modality in ("image", "video") ] def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: image_grid_thw = hf_inputs.get("image_grid_thw", torch.empty((0, 3))) image_grid_sizes = image_grid_thw.prod(-1) video_grid_thw = hf_inputs.get("video_grid_thw", torch.empty((0, 3))) video_grid_sizes = video_grid_thw.prod(-1) return dict( pixel_values=MultiModalFieldConfig.flat_from_sizes( "image", image_grid_sizes ), image_grid_thw=MultiModalFieldConfig.batched("image"), pixel_values_videos=MultiModalFieldConfig.flat_from_sizes( "video", video_grid_sizes ), video_grid_thw=MultiModalFieldConfig.batched("video"), ) class Ernie4_5_VLDummyInputsBuilder(BaseDummyInputsBuilder[Ernie4_5_VLProcessingInfo]): def get_dummy_text(self, mm_counts: Mapping[str, int]) -> str: num_images = mm_counts.get("image", 0) num_videos = mm_counts.get("video", 0) prompt = "" for i in range(num_images): prompt += ( f"Picture {i + 1}:<|IMAGE_START|><|image@placeholder|><|IMAGE_END|>" ) for i in range(num_videos): prompt += f"Video {i + 1}:<|VIDEO_START|><|video@placeholder|><|VIDEO_END|>" return prompt def get_dummy_mm_data( self, seq_len: int, mm_counts: Mapping[str, int], mm_options: Mapping[str, BaseDummyOptions], ) -> MultiModalDataDict: num_images = mm_counts.get("image", 0) num_videos = mm_counts.get("video", 0) target_width, target_height = self.info.get_image_size_with_most_features() target_num_frames = self.info.get_num_frames_with_most_features( seq_len, mm_counts ) image_overrides = mm_options.get("image") video_overrides = mm_options.get("video") return { "image": self._get_dummy_images( width=target_width, height=target_height, num_images=num_images, overrides=image_overrides, ), "video": self._get_dummy_videos( width=target_width, height=target_height, num_frames=target_num_frames, num_videos=num_videos, overrides=video_overrides, ), } def _get_dummy_videos( self, *, width: int, height: int, num_frames: int, num_videos: int, overrides: VideoDummyOptions | None = None, ): if overrides: if overrides.num_frames: if overrides.num_frames > num_frames: logger.warning( "video.num_frames override (%d) exceeds model's " "maximum number of frames (%d), will be ignored", overrides.num_frames, num_frames, ) num_frames = min(num_frames, overrides.num_frames) if overrides.width: if overrides.width > width: logger.warning( "video.width override (%d) exceeds model's " "maximum width (%d), will be ignored", overrides.width, width, ) width = min(width, overrides.width) if overrides.height: if overrides.height > height: logger.warning( "video.height override (%d) exceeds model's " "maximum height (%d), will be ignored", overrides.height, height, ) height = min(height, overrides.height) num_frames = max(num_frames, 2) # ernie4.5-vl requires at least 2 frames video = np.full((num_frames, width, height, 3), 255, dtype=np.uint8) video_items = [] for i in range(num_videos): video_metadata = { "fps": 2.0, "duration": num_frames / 2.0, "total_num_frames": num_frames, "frames_indices": [i for i in range(num_frames)], "video_backend": "opencv", "do_sample_frames": False, } video_item = (video.copy(), video_metadata) video_items.append(video_item) return video_items @MULTIMODAL_REGISTRY.register_processor( Ernie4_5VLMultiModalProcessor, info=Ernie4_5_VLProcessingInfo, dummy_inputs=Ernie4_5_VLDummyInputsBuilder, ) class Ernie4_5_VLMoeForConditionalGeneration( nn.Module, SupportsMultiModal, SupportsLoRA, SupportsPP, SupportsMRoPE ): packed_modules_mapping = { "qkv_proj": [ "q_proj", "k_proj", "v_proj", ], "gate_up_proj": [ "gate_proj", "up_proj", ], } # To ensure correct weight loading and mapping. hf_to_vllm_mapper = WeightsMapper( orig_to_new_prefix={ "lm_head.": "language_model.lm_head.", "model.": "language_model.model.", # model.resampler_model.-> language_model.model.resampler_model. # language_model.model.resampler_model. -> resampler_model. "language_model.model.resampler_model.": "resampler_model.", }, # resampler_weight_mappings orig_to_new_substr={ "spatial_linear.0.": "spatial_linear1.", "spatial_linear.2.": "spatial_linear2.", "spatial_linear.3.": "spatial_norm.", "temporal_linear.0.": "temporal_linear1.", "temporal_linear.2.": "temporal_linear2.", "temporal_linear.3.": "temporal_norm.", }, ) @classmethod def get_placeholder_str(cls, modality: str, i: int) -> str | None: if modality.startswith("image"): return "<|IMAGE_START|><|image@placeholder|><|IMAGE_END|>" if modality.startswith("video"): return "<|VIDEO_START|><|video@placeholder|><|VIDEO_END|>" raise ValueError("Only image or video modality is supported") def __init__(self, vllm_config: VllmConfig, prefix: str = "") -> None: super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config multimodal_config = vllm_config.model_config.multimodal_config self.config = config self.multimodal_config = multimodal_config with self._mark_tower_model(vllm_config, {"image", "video"}): self.vision_model = Ernie4_5_VisionTransformer( config.vision_config, norm_eps=getattr(config, "rms_norm_eps", 1e-6), quant_config=quant_config, prefix=maybe_prefix(prefix, "vision_model"), ) self.resampler_model = VariableResolutionResamplerModel( self.config.pixel_hidden_size, self.config.hidden_size, self.config.spatial_conv_size, self.config.temporal_conv_size, config=self.config, prefix=maybe_prefix(prefix, "resampler_model"), ) with self._mark_language_model(vllm_config): self.language_model = Ernie4_5_VLMoeForCausalLM( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "language_model"), ) self.visual_token_mask = None self.make_empty_intermediate_tensors = ( self.language_model.make_empty_intermediate_tensors ) if getattr(self.config, "im_patch_id", None): visual_token_ids = [ token_id for token_id in [ self.config.im_patch_id, getattr(self.config, "image_start_token_id", None), getattr(self.config, "image_end_token_id", None), getattr(self.config, "video_start_token_id", None), getattr(self.config, "video_end_token_id", None), ] if token_id is not None ] self._visual_token_ids_tensor_cache = torch.tensor( visual_token_ids, dtype=torch.long ) else: self._visual_token_ids_tensor_cache = None def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: """compute logits""" return self.language_model.compute_logits(hidden_states) def _vision_forward( self, pixel_values: torch.Tensor, grid_thw: torch.Tensor, ) -> torch.Tensor: if grid_thw is not None: grid_thw = grid_thw[grid_thw > 0] if grid_thw.numel() % 3 != 0: raise ValueError( f"grid_thw has {grid_thw.numel()} elements after filtering," "which is not divisible by 3." ) grid_thw = grid_thw.reshape(-1, 3) # example: [[1,64,64],[2,80,80]] -> [[1,64,64],[1,80,80],[1,80,80]] grid_thw = F.pad( torch.repeat_interleave(grid_thw[:, 1:], grid_thw[:, 0], 0), [1, 0, 0, 0], value=1, ) image_features = self.vision_model(pixel_values, grid_thw) return image_features def _set_visual_token_mask(self, input_ids: torch.Tensor) -> None: """Set mask for visual tokens (image/video patches and delimiters).""" if self._visual_token_ids_tensor_cache is None: self.visual_token_mask = None return # Create tensor on the correct device visual_token_ids_tensor = self._visual_token_ids_tensor_cache.to( device=input_ids.device, dtype=input_ids.dtype, ) self.visual_token_mask = torch.isin(input_ids, visual_token_ids_tensor).reshape( -1, 1 ) def get_mrope_input_positions( self, input_tokens: list[int], mm_features: list[MultiModalFeatureSpec], ) -> tuple[torch.Tensor, int]: kwargs = MultiModalFeatureSpec.gather_kwargs( mm_features, {"image_grid_thw", "video_grid_thw"}, ) image_grid_thw = [item.tolist() for item in kwargs.get("image_grid_thw", [])] video_grid_thw = [item.tolist() for item in kwargs.get("video_grid_thw", [])] hf_config = self.config image_token_id = hf_config.im_patch_id video_start_token_id = hf_config.video_start_token_id video_end_token_id = hf_config.video_end_token_id spatial_conv_size = hf_config.spatial_conv_size temporal_conv_size = hf_config.temporal_conv_size llm_pos_ids_list: list = [] if image_grid_thw or video_grid_thw: input_token_type: list[str] = [] video_check_flg = False for token in input_tokens: if token == video_start_token_id: video_check_flg = True elif token == video_end_token_id: video_check_flg = False if (token == image_token_id) and (video_check_flg is False): input_token_type.append("image") elif (token == image_token_id) and (video_check_flg is True): input_token_type.append("video") else: input_token_type.append("text") input_type_group: list[tuple[str, int, int]] = [] for key, group_iter in itertools.groupby( enumerate(input_token_type), lambda x: x[1] ): group_list = list(group_iter) start_index = group_list[0][0] end_index = group_list[-1][0] + 1 input_type_group.append((key, start_index, end_index)) video_frame_num = 1 mm_data_idx = 0 for modality_type, start_idx, end_idx in input_type_group: st_idx = ( llm_pos_ids_list[-1].max() + 1 if len(llm_pos_ids_list) > 0 else 0 ) if modality_type == "image": t, h, w = image_grid_thw[mm_data_idx] llm_grid_t, llm_grid_h, llm_grid_w = ( t, h // spatial_conv_size, w // spatial_conv_size, ) t_index = ( torch.arange(llm_grid_t) .view(-1, 1) .expand(-1, llm_grid_h * llm_grid_w) .flatten() ) h_index = ( torch.arange(llm_grid_h) .view(1, -1, 1) .expand(llm_grid_t, -1, llm_grid_w) .flatten() ) w_index = ( torch.arange(llm_grid_w) .view(1, 1, -1) .expand(llm_grid_t, llm_grid_h, -1) .flatten() ) llm_pos_ids_list.append( torch.stack([t_index, h_index, w_index]) + st_idx ) mm_data_idx += 1 elif modality_type == "video": t, h, w = video_grid_thw[mm_data_idx] llm_grid_t, llm_grid_h, llm_grid_w = ( t // temporal_conv_size, h // spatial_conv_size, w // spatial_conv_size, ) for t_idx in range(llm_grid_t): t_index = ( torch.tensor(t_idx) .view(-1, 1) .expand(-1, llm_grid_h * llm_grid_w) .flatten() ) h_index = ( torch.arange(llm_grid_h) .view(1, -1, 1) .expand(1, -1, llm_grid_w) .flatten() ) w_index = ( torch.arange(llm_grid_w) .view(1, 1, -1) .expand(1, llm_grid_h, -1) .flatten() ) llm_pos_ids_list.append( torch.stack([t_index, h_index, w_index]) + st_idx ) mm_data_idx += 1 video_frame_num += 1 else: text_len = end_idx - start_idx llm_pos_ids_list.append( torch.arange(text_len).view(1, -1).expand(3, -1) + st_idx ) video_frame_num = 1 else: text_len = len(input_tokens) llm_pos_ids_list.append(torch.arange(text_len).view(1, -1).expand(3, -1)) llm_positions = torch.cat(llm_pos_ids_list, dim=1).reshape(3, -1) mrope_position_delta = (llm_positions.max() + 1 - len(input_tokens)).item() return llm_positions, mrope_position_delta def _parse_and_validate_image_input( self, **kwargs: object ) -> Ernie4_5_VLImageInputs | None: pixel_values = kwargs.pop("pixel_values", None) image_grid_thw = kwargs.pop("image_grid_thw", None) if pixel_values is None: return None if pixel_values is not None: return Ernie4_5_VLImagePixelInputs( type="pixel_values", pixel_values=pixel_values, image_grid_thw=image_grid_thw, ) def _parse_and_validate_video_input( self, **kwargs: object ) -> Ernie4_5_VLVideoInputs | None: pixel_values_videos = kwargs.pop("pixel_values_videos", None) video_grid_thw = kwargs.pop("video_grid_thw", None) if pixel_values_videos is None: return None if pixel_values_videos is not None: return Ernie4_5_VLVideoPixelInputs( type="pixel_values_videos", pixel_values_videos=pixel_values_videos, video_grid_thw=video_grid_thw, ) def _process_image_input( self, image_input: Ernie4_5_VLImageInputs ) -> tuple[torch.Tensor, ...]: grid_thw = image_input["image_grid_thw"] assert grid_thw.ndim == 2 pixel_values = image_input["pixel_values"].type(self.vision_model.dtype) image_features = self._vision_forward( pixel_values=pixel_values, grid_thw=grid_thw ) image_embeds = self.resampler_model(image_features, grid_thw) merge_size = self.vision_model.spatial_merge_size sizes = grid_thw.prod(-1) // merge_size // merge_size return image_embeds.split(sizes.tolist()) def _process_video_input( self, video_input: Ernie4_5_VLVideoInputs ) -> tuple[torch.Tensor, ...]: grid_thw = video_input["video_grid_thw"] assert grid_thw.ndim == 2 pixel_values_videos = video_input["pixel_values_videos"].type( self.vision_model.dtype ) video_features = self._vision_forward( pixel_values=pixel_values_videos, grid_thw=grid_thw ) video_embeds = self.resampler_model(video_features, grid_thw) merge_size = self.vision_model.spatial_merge_size sizes = ( (grid_thw.prod(-1) // self.config.temporal_conv_size) // merge_size // merge_size ) return video_embeds.split(sizes.tolist()) def _parse_and_validate_multimodal_inputs(self, **kwargs: object) -> dict: modalities = {} # Preserve the order of modalities if there are multiple of them # from the order of kwargs. for input_key in kwargs: if ( input_key in ("pixel_values", "image_embeds") and "images" not in modalities ): modalities["images"] = self._parse_and_validate_image_input(**kwargs) if ( input_key in ("pixel_values_videos", "video_embeds") and "videos" not in modalities ): modalities["videos"] = self._parse_and_validate_video_input(**kwargs) return modalities def embed_multimodal(self, **kwargs: object) -> MultiModalEmbeddings | None: modalities = self._parse_and_validate_multimodal_inputs(**kwargs) if not modalities: return None # The result multimodal_embeddings is tuple of tensors, with each # tensor corresponding to a multimodal data item (image or video). multimodal_embeddings: tuple[torch.Tensor, ...] = () # NOTE: It is important to iterate over the keys in this dictionary # to preserve the order of the modalities. for modality in modalities: if modality == "images": image_input = modalities["images"] image_embeddings = self._process_image_input(image_input) multimodal_embeddings += tuple(image_embeddings) if modality == "videos": video_input = modalities["videos"] video_embeddings = self._process_video_input(video_input) multimodal_embeddings += tuple(video_embeddings) return multimodal_embeddings def embed_input_ids( self, input_ids: torch.Tensor, multimodal_embeddings: MultiModalEmbeddings | None = None, *, is_multimodal: torch.Tensor | None = None, handle_oov_mm_token: bool = False, ) -> torch.Tensor: if multimodal_embeddings is not None and len(multimodal_embeddings) > 0: self._set_visual_token_mask(input_ids) # This is to satisfy the type checker for each overload if multimodal_embeddings is None or is_multimodal is None: return super().embed_input_ids(input_ids) return super().embed_input_ids( input_ids, multimodal_embeddings=multimodal_embeddings, is_multimodal=is_multimodal, handle_oov_mm_token=handle_oov_mm_token, ) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs, ): forward_kwargs = { "input_ids": input_ids, "positions": positions, "intermediate_tensors": intermediate_tensors, "inputs_embeds": inputs_embeds, } if self.visual_token_mask is not None: if self.visual_token_mask.shape[0] != inputs_embeds.shape[0]: padding_len = inputs_embeds.shape[0] - self.visual_token_mask.shape[0] # right pad False pad = torch.zeros( (padding_len, self.visual_token_mask.shape[1]), dtype=self.visual_token_mask.dtype, device=self.visual_token_mask.device, ) self.visual_token_mask = torch.cat([self.visual_token_mask, pad], dim=0) forward_kwargs.update({"visual_token_mask": self.visual_token_mask}) self.visual_token_mask = None hidden_states = self.language_model.model( **forward_kwargs, **kwargs, ) return hidden_states def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader(self) return loader.load_weights(weights, mapper=self.hf_to_vllm_mapper)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/ernie45_vl.py", "license": "Apache License 2.0", "lines": 1481, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/ernie45_vl_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright 2025 The Baidu team. # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Inference-only Erine VL model compatible with HuggingFace weights.""" from collections.abc import Iterable from itertools import islice from typing import Any import torch from torch import nn from transformers import PretrainedConfig # from vllm.compilation.decorators import support_torch_compile from vllm.config import CacheConfig, VllmConfig from vllm.distributed import get_pp_group, get_tensor_model_parallel_world_size from vllm.logger import init_logger from vllm.model_executor.layers.attention import Attention from vllm.model_executor.layers.fused_moe import SharedFusedMoE from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( QKVParallelLinear, ReplicatedLinear, RowParallelLinear, ) from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.rotary_embedding.ernie45_vl_rope import ( Ernie4_5_VLRotaryEmbedding, ) from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import ( default_weight_loader, maybe_remap_kv_scale_name, ) from vllm.sequence import IntermediateTensors from vllm.transformers_utils.config import set_default_rope_theta from .ernie45_moe import Ernie4_5_MoeMLP from .interfaces import SupportsPP from .utils import ( PPMissingLayer, extract_layer_index, is_pp_missing_parameter, make_empty_intermediate_tensors_factory, make_layers, maybe_prefix, ) logger = init_logger(__name__) class Ernie4_5_VLMoeMLP(Ernie4_5_MoeMLP): def __init__(self, shared_experts: torch.nn.Module | None = None, **kwargs): super().__init__(**kwargs) self.shared_experts = shared_experts def forward(self, x): if self.shared_experts is not None: return self.shared_experts(x) + super().forward(x) else: return super().forward(x) class Ernie4_5_VLMoeAttention(nn.Module): def __init__( self, hidden_size: int, num_heads: int, num_kv_heads: int, rope_parameters: dict[str, Any], head_dim: int | None = None, freq_allocation: int = 20, max_position_embeddings: int = 131072, rms_norm_eps: float = 1e-05, qkv_bias: bool = False, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() layer_idx = extract_layer_index(prefix) if len(prefix) > 0 else 0 self.layer_idx = layer_idx self.hidden_size = hidden_size tp_size = get_tensor_model_parallel_world_size() self.total_num_heads = num_heads assert self.total_num_heads % tp_size == 0 self.num_heads = self.total_num_heads // tp_size self.total_num_kv_heads = num_kv_heads if self.total_num_kv_heads >= tp_size: # Number of KV heads is greater than TP size, so we partition # the KV heads across multiple tensor parallel GPUs. assert self.total_num_kv_heads % tp_size == 0 else: # Number of KV heads is less than TP size, so we replicate # the KV heads across multiple tensor parallel GPUs. assert tp_size % self.total_num_kv_heads == 0 self.num_kv_heads = max(1, self.total_num_kv_heads // tp_size) self.head_dim = head_dim or (hidden_size // self.total_num_heads) self.q_size = self.num_heads * self.head_dim self.kv_size = self.num_kv_heads * self.head_dim self.scaling = self.head_dim**-0.5 self.max_position_embeddings = max_position_embeddings self.qkv_proj = QKVParallelLinear( hidden_size, self.head_dim, self.total_num_heads, self.total_num_kv_heads, bias=qkv_bias, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", ) self.o_proj = RowParallelLinear( self.total_num_heads * self.head_dim, hidden_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.o_proj", ) t_rope = freq_allocation h_rope = (self.head_dim // 2 - freq_allocation) // 2 w_rope = (self.head_dim // 2 - freq_allocation) // 2 self.rotary_emb = Ernie4_5_VLRotaryEmbedding( head_size=self.head_dim, rotary_dim=self.head_dim, max_position_embeddings=max_position_embeddings, base=rope_parameters["rope_theta"], is_neox_style=False, dtype=torch.get_default_dtype(), mrope_section=[h_rope, w_rope, t_rope], ) self.attn = Attention( self.num_heads, self.head_dim, self.scaling, num_kv_heads=self.num_kv_heads, cache_config=cache_config, quant_config=quant_config, prefix=f"{prefix}.attn", ) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, ) -> torch.Tensor: qkv, _ = self.qkv_proj(hidden_states) q, k, v = qkv.split([self.q_size, self.kv_size, self.kv_size], dim=-1) q, k = self.rotary_emb(positions, q, k) # Attention attn_output = self.attn(q, k, v) # Output projection output, _ = self.o_proj(attn_output) return output class Ernie4_5_VLMoeMoE(nn.Module): def __init__( self, config: PretrainedConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() layer_idx = extract_layer_index(prefix) self.layer_idx = layer_idx self.tp_size = get_tensor_model_parallel_world_size() self.has_shared_experts = getattr(config, "moe_num_shared_experts", 0) > 0 self.hidden_size = config.hidden_size moe_num_experts = config.moe_num_experts max_moe_num_experts = max(moe_num_experts) if self.tp_size > max_moe_num_experts: raise ValueError( f"Tensor parallel size {self.tp_size} is greater than " f"the number of experts {moe_num_experts}." ) moe_layer_start_index = config.moe_layer_start_index text_moe_layer_start_index = moe_layer_start_index[0] vision_moe_layer_start_index = moe_layer_start_index[1] moe_layer_end_index = config.moe_layer_end_index moe_layer_end_index = getattr( config, "moe_layer_end_index", [config.num_hidden_layers - 1, config.num_hidden_layers - 1], ) text_moe_layer_end_index = moe_layer_end_index[0] vision_moe_layer_end_index = moe_layer_end_index[1] assert config.moe_num_experts[0] == config.moe_num_experts[1] self.e_score_correction_bias = nn.Parameter( torch.empty(2, config.moe_num_experts[0], dtype=torch.float32) ) assert text_moe_layer_start_index <= text_moe_layer_end_index if self.has_shared_experts: intermediate_size = ( config.moe_intermediate_size[0] * config.moe_num_shared_experts ) self.shared_experts = Ernie4_5_VLMoeMLP( hidden_size=config.hidden_size, intermediate_size=intermediate_size, hidden_act=config.hidden_act, quant_config=quant_config, prefix=f"{prefix}.shared_experts", reduce_results=False, ) else: self.shared_experts = None if ( layer_idx >= text_moe_layer_start_index and layer_idx <= text_moe_layer_end_index ): self.text_experts_gate = ReplicatedLinear( config.hidden_size, config.moe_num_experts[0], bias=False, params_dtype=torch.float32, quant_config=quant_config, prefix=f"{prefix}.text_experts_gate", ) self.text_experts = SharedFusedMoE( shared_experts=self.shared_experts, num_experts=config.moe_num_experts[0], top_k=config.moe_k, hidden_size=config.hidden_size, intermediate_size=config.moe_intermediate_size[0], reduce_results=False, renormalize=True, quant_config=quant_config, e_score_correction_bias=self.e_score_correction_bias[0], prefix=f"{prefix}.text_experts", router_logits_dtype=torch.float32, ) else: self.text_experts = Ernie4_5_VLMoeMLP( shared_experts=self.shared_experts, hidden_size=config.hidden_size, intermediate_size=config.intermediate_size, hidden_act=config.hidden_act, use_bias=getattr(config, "use_bias", False), quant_config=quant_config, prefix=f"{prefix}.mlp", ) assert vision_moe_layer_start_index <= vision_moe_layer_end_index if ( layer_idx >= vision_moe_layer_start_index and layer_idx <= vision_moe_layer_end_index ): self.vision_experts_gate = ReplicatedLinear( config.hidden_size, config.moe_num_experts[1], bias=False, params_dtype=torch.float32, quant_config=quant_config, prefix=f"{prefix}.vision_experts_gate", ) self.vision_experts = SharedFusedMoE( shared_experts=self.shared_experts, num_experts=config.moe_num_experts[1], top_k=config.moe_k, hidden_size=config.hidden_size, intermediate_size=config.moe_intermediate_size[1], reduce_results=False, renormalize=True, quant_config=quant_config, e_score_correction_bias=self.e_score_correction_bias[1], prefix=f"{prefix}.vision_experts", router_logits_dtype=torch.float32, ) else: self.vision_experts = Ernie4_5_VLMoeMLP( shared_experts=self.shared_experts, hidden_size=config.hidden_size, intermediate_size=config.intermediate_size, hidden_act=config.hidden_act, use_bias=getattr(config, "use_bias", False), quant_config=quant_config, prefix=f"{prefix}.mlp", ) def forward( self, hidden_states: torch.Tensor, visual_token_mask: torch.Tensor, **kwargs: object, ) -> torch.Tensor: orig_shape = hidden_states.shape hidden_dim = hidden_states.shape[-1] hidden_states = hidden_states.view(-1, hidden_dim) if visual_token_mask is not None and visual_token_mask.all(): # only vision modal input router_logits, _ = self.vision_experts_gate( hidden_states.to(dtype=torch.float32) ) final_hidden_states = self.vision_experts( hidden_states=hidden_states, router_logits=router_logits ) elif visual_token_mask is not None and visual_token_mask.any(): # text and vision modals input visual_token_mask = visual_token_mask.repeat(1, self.hidden_size).bool() text_token_mask = ~visual_token_mask final_experts_hidden_states = torch.zeros_like(hidden_states) final_shared_ouput = ( torch.zeros_like(hidden_states) if self.has_shared_experts else None ) text_hidden_states = hidden_states[text_token_mask].reshape( -1, self.hidden_size ) vision_hidden_states = hidden_states[visual_token_mask].reshape( -1, self.hidden_size ) text_router_logits, _ = self.text_experts_gate( text_hidden_states.to(dtype=torch.float32) ) text_shared_ouput, text_experts_output = self.text_experts( hidden_states=text_hidden_states, router_logits=text_router_logits ) final_experts_hidden_states[text_token_mask] = text_experts_output.flatten() if self.has_shared_experts: final_shared_ouput[text_token_mask] = text_shared_ouput.flatten() vision_router_logits, _ = self.vision_experts_gate( vision_hidden_states.to(dtype=torch.float32) ) vision_shared_ouput, vision_experts_output = self.vision_experts( hidden_states=vision_hidden_states, router_logits=vision_router_logits ) final_experts_hidden_states[visual_token_mask] = ( vision_experts_output.flatten() ) if self.has_shared_experts: final_shared_ouput[visual_token_mask] = vision_shared_ouput.flatten() final_hidden_states = (final_shared_ouput, final_experts_hidden_states) else: # only text modal input text_router_logits, _ = self.text_experts_gate( hidden_states.to(dtype=torch.float32) ) final_hidden_states = self.text_experts( hidden_states=hidden_states, router_logits=text_router_logits ) if self.has_shared_experts: # for shared_experts model final_hidden_states = final_hidden_states[0] + final_hidden_states[1] else: # for not shared_experts model final_hidden_states = final_hidden_states[1] if self.tp_size > 1: final_hidden_states = ( self.text_experts.maybe_all_reduce_tensor_model_parallel( final_hidden_states ) ) return final_hidden_states.view(orig_shape) class Ernie4_5_VLMoeDecoderLayer(nn.Module): def __init__( self, config: PretrainedConfig, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.hidden_size = config.hidden_size set_default_rope_theta(config, default_theta=500000) freq_allocation = getattr(config, "freq_allocation", 20) max_position_embeddings = getattr(config, "max_position_embeddings", 131072) self.self_attn = Ernie4_5_VLMoeAttention( hidden_size=self.hidden_size, num_heads=config.num_attention_heads, num_kv_heads=config.num_key_value_heads, head_dim=getattr(config, "head_dim", None), rope_parameters=config.rope_parameters, freq_allocation=freq_allocation, max_position_embeddings=max_position_embeddings, rms_norm_eps=config.rms_norm_eps, qkv_bias=getattr(config, "use_bias", False), cache_config=cache_config, quant_config=quant_config, prefix=f"{prefix}.self_attn", ) layer_idx = extract_layer_index(prefix) self.layer_idx = layer_idx # MoE moe_layer_start_index = config.moe_layer_start_index min_moe_layer_start_index = min(moe_layer_start_index) moe_layer_end_index = getattr( config, "moe_layer_end_index", [config.num_hidden_layers - 1, config.num_hidden_layers - 1], ) max_moe_layer_end_index = max(moe_layer_end_index) assert min_moe_layer_start_index <= max_moe_layer_end_index moe_num_experts = config.moe_num_experts max_moe_num_experts = max(moe_num_experts) moe_layer_interval = getattr(config, "moe_layer_interval", 1) use_moe = getattr(config, "use_moe", max_moe_num_experts > 0) if ( use_moe and ((layer_idx + 1) % moe_layer_interval == 0) and layer_idx >= min_moe_layer_start_index and layer_idx <= max_moe_layer_end_index ): self.mlp = Ernie4_5_VLMoeMoE( config=config, quant_config=quant_config, prefix=f"{prefix}.mlp" ) else: self.mlp = Ernie4_5_VLMoeMLP( hidden_size=config.hidden_size, intermediate_size=config.intermediate_size, hidden_act=config.hidden_act, use_bias=getattr(config, "use_bias", False), quant_config=quant_config, prefix=f"{prefix}.mlp", ) self.input_layernorm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.post_attention_layernorm = RMSNorm( config.hidden_size, eps=config.rms_norm_eps ) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, residual: torch.Tensor | None, visual_token_mask: torch.Tensor | None, **kwargs: object, ) -> torch.Tensor: # Self Attention if residual is None: residual = hidden_states hidden_states = self.input_layernorm(hidden_states) else: hidden_states, residual = self.input_layernorm(hidden_states, residual) hidden_states = self.self_attn( positions=positions, hidden_states=hidden_states, ) # Fully Connected hidden_states, residual = self.post_attention_layernorm(hidden_states, residual) if isinstance(self.mlp, Ernie4_5_VLMoeMoE): hidden_states = self.mlp(hidden_states, visual_token_mask, **kwargs) else: hidden_states = self.mlp(hidden_states) return hidden_states, residual # Since Ernie VL distinguishes between text experts and vision experts, # enabling torch.compile will cause errors. # @support_torch_compile( # dynamic_arg_dims={ # "input_ids": 0, # "positions": -1, # "intermediate_tensors": 0, # "inputs_embeds": 0, # "visual_token_mask": 0, # }) class Ernie4_5_VLMoeModel(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config cache_config = vllm_config.cache_config quant_config = vllm_config.quant_config self.vocab_size = config.vocab_size self.config = config self.im_patch_id = config.im_patch_id if get_pp_group().is_first_rank: self.embed_tokens = VocabParallelEmbedding( config.vocab_size, config.hidden_size, quant_config=quant_config, prefix=f"{prefix}.embed_tokens", ) else: self.embed_tokens = PPMissingLayer() self.start_layer, self.end_layer, self.layers = make_layers( config.num_hidden_layers, lambda prefix: Ernie4_5_VLMoeDecoderLayer( config=config, cache_config=cache_config, quant_config=quant_config, prefix=prefix, ), prefix=f"{prefix}.layers", ) if get_pp_group().is_last_rank: self.norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) else: self.norm = PPMissingLayer() self.make_empty_intermediate_tensors = make_empty_intermediate_tensors_factory( ["hidden_states", "residual"], config.hidden_size ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, visual_token_mask: torch.Tensor | None = None, **kwargs: object, ) -> torch.Tensor | IntermediateTensors: if get_pp_group().is_first_rank: if inputs_embeds is not None: hidden_states = inputs_embeds else: hidden_states = self.embed_input_ids(input_ids) residual = None else: assert intermediate_tensors is not None hidden_states = intermediate_tensors["hidden_states"] residual = intermediate_tensors["residual"] for layer in islice(self.layers, self.start_layer, self.end_layer): hidden_states, residual = layer( positions, hidden_states, residual, visual_token_mask, **kwargs ) if not get_pp_group().is_last_rank: return IntermediateTensors( {"hidden_states": hidden_states, "residual": residual} ) hidden_states, _ = self.norm(hidden_states, residual) return hidden_states # only used as text backbone for ernie4.5-vl class Ernie4_5_VLMoeForCausalLM(nn.Module, SupportsPP): packed_modules_mapping = { "qkv_proj": [ "q_proj", "k_proj", "v_proj", ], "gate_up_proj": [ "gate_proj", "up_proj", ], } fall_back_to_pt_during_load = False def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config self.config = config self.quant_config = quant_config self.model = Ernie4_5_VLMoeModel( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model") ) if get_pp_group().is_last_rank: self.lm_head = ParallelLMHead( config.vocab_size, config.hidden_size, quant_config=quant_config, prefix=maybe_prefix(prefix, "lm_head"), ) else: self.lm_head = PPMissingLayer() if self.config.tie_word_embeddings: self.lm_head.weight = self.model.embed_tokens.weight self.logits_processor = LogitsProcessor(config.vocab_size) self.make_empty_intermediate_tensors = ( self.model.make_empty_intermediate_tensors ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs: object, ) -> torch.Tensor | IntermediateTensors: hidden_states = self.model( input_ids, positions, intermediate_tensors, inputs_embeds, **kwargs ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: logits = self.logits_processor(self.lm_head, hidden_states) return logits def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ # (param_name, shard_name, shard_id) ("qkv_proj", "q_proj", "q"), ("qkv_proj", "k_proj", "k"), ("qkv_proj", "v_proj", "v"), ("gate_up_proj", "gate_proj", 0), ("gate_up_proj", "up_proj", 1), ] # Params for weights, fp8 weight scales, fp8 activation scales # (param_name, weight_name, expert_id, shard_id) expert_params_mapping = SharedFusedMoE.make_expert_params_mapping( self, ckpt_gate_proj_name="gate_proj", ckpt_down_proj_name="down_proj", ckpt_up_proj_name="up_proj", num_experts=max(self.config.moe_num_experts), ) params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() for name, loaded_weight in weights: if self.config.tie_word_embeddings and name.endswith("lm_head.weight"): loaded_params.add("lm_head.weight") continue # MTP will be supported soon. if "mtp" in name or "vision_model" in name or "resampler_model" in name: continue for param_name, weight_name, shard_id in stacked_params_mapping: # Skip non-stacked layers and experts (experts handled below). if weight_name not in name: continue if ("mlp.experts." in name) and name not in params_dict: continue name = name.replace(weight_name, param_name) # Skip loading extra bias for GPTQ models. if ( name.endswith(".bias") or name.endswith("_bias") ) and name not in params_dict: continue # Skip layers on other devices. if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: # Distinguish between vision experts and text experts if "mlp.experts" in name: moe_offset = int(name.split(".")[-3]) vision_expert_start_idx = self.config.moe_num_experts[0] is_text_expert = moe_offset <= vision_expert_start_idx - 1 if is_text_expert: name = name.replace(".experts.", ".text_experts.") else: name = name.replace( f".experts.{moe_offset}", f".vision_experts.{moe_offset - vision_expert_start_idx}", ) for mapping in expert_params_mapping: param_name, weight_name, expert_id, shard_id = mapping if weight_name not in name: continue # Distinguish between vision experts and text experts moe_offset = int(name.split(".")[-3]) is_text_expert = moe_offset <= self.config.moe_num_experts[0] - 1 name = name.replace(weight_name, param_name) if is_text_expert: name = name.replace(".experts.", ".text_experts.") else: name = name.replace(".experts.", ".vision_experts.") # Skip layers on other devices. if is_pp_missing_parameter(name, self): continue # Skip loading extra bias for GPTQ models. if ( name.endswith(".bias") or name.endswith("_bias") ) and name not in params_dict: continue param = params_dict[name] weight_loader = param.weight_loader weight_loader( param, loaded_weight, name, shard_id=shard_id, expert_id=expert_id, ) break else: # Distinguish between vision expert gate # and text expert gate if name.endswith("mlp.gate.weight"): name = name.replace("gate.weight", "text_experts_gate.weight") loaded_weight = loaded_weight.T elif name.endswith("mlp.gate.weight_1"): name = name.replace( "gate.weight_1", "vision_experts_gate.weight" ) loaded_weight = loaded_weight.T if "e_score_correction_bias" in name: name = name.replace(".moe_statics.", ".") # Skip loading extra bias for GPTQ models. if ( name.endswith(".bias") or name.endswith("_bias") ) and name not in params_dict: continue # Skip layers on other devices. if is_pp_missing_parameter(name, self): continue # Remapping the name of FP8 kv-scale. name = maybe_remap_kv_scale_name(name, params_dict) if name is None: continue param = params_dict[name] weight_loader = getattr( param, "weight_loader", default_weight_loader ) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/ernie45_vl_moe.py", "license": "Apache License 2.0", "lines": 704, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/quantization/test_flashinfer_scaled_mm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from vllm import _custom_ops as ops from vllm.platforms import current_platform from vllm.utils.flashinfer import flashinfer_scaled_fp8_mm from vllm.utils.torch_utils import set_random_seed if not current_platform.has_device_capability(100): pytest.skip( reason="Flashinfer FP8 gemms requires compute capability of 10.0 or above.", allow_module_level=True, ) DTYPES = [torch.float16, torch.bfloat16] # m, n, k SHAPES = [(128, 128, 64), (128, 128, 128), (256, 128, 64), (128, 256, 128)] PAD_SHAPES = [(150, 128, 64), (128, 128, 96)] SHAPES.extend(PAD_SHAPES) SEEDS = [42] CUDA_DEVICES = ["cuda:0"] @pytest.mark.parametrize("dtype", DTYPES) @pytest.mark.parametrize("shape", SHAPES) @pytest.mark.parametrize("use_bias", [True, False]) @pytest.mark.parametrize("seed", SEEDS) @pytest.mark.parametrize("device", CUDA_DEVICES) @pytest.mark.parametrize("autotune", [False, True]) @torch.inference_mode() def test_flashinfer_fp8_gemm( dtype: torch.dtype, shape: tuple[int, int, int], use_bias: bool, seed: int, device: str, autotune: bool, ) -> None: set_random_seed(seed) m, n, k = shape a = torch.randn((m, k), dtype=dtype, device=device) b = torch.randn((n, k), dtype=dtype, device=device) / k a_fp8, a_scale = ops.scaled_fp8_quant(a) b_fp8, b_scale = ops.scaled_fp8_quant(b) expected_out = torch.mm( a_scale * a_fp8.to(dtype=torch.float32), b_scale * b_fp8.to(dtype=torch.float32).t(), ).to(dtype=dtype) if use_bias: bias = torch.randn((n,), dtype=dtype, device=device) expected_out = expected_out + bias else: bias = None import flashinfer with flashinfer.autotune(autotune): out = flashinfer_scaled_fp8_mm( a_fp8, b_fp8.t(), a_scale, b_scale, dtype, bias=bias, ) torch.testing.assert_close(out, expected_out, atol=1e-2, rtol=1e-2)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/quantization/test_flashinfer_scaled_mm.py", "license": "Apache License 2.0", "lines": 61, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/kernels/moe/test_grouped_topk.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests for the MoE grouped topk kernel Run `pytest tests/kernels/moe/test_grouped_topk.py`. """ import pytest import torch import vllm.model_executor.layers.batch_invariant as batch_invariant from vllm.config import ( CompilationConfig, VllmConfig, get_cached_compilation_config, set_current_vllm_config, ) from vllm.model_executor.layers.fused_moe.router.grouped_topk_router import ( GroupedTopk, fused_grouped_topk, ) from vllm.platforms import current_platform from vllm.utils.torch_utils import set_random_seed @pytest.mark.skipif( not current_platform.is_cuda(), reason="This test is skipped on non-CUDA platform." ) @pytest.mark.parametrize("n_token", [1, 33, 64]) @pytest.mark.parametrize("n_hidden", [1024, 2048]) @pytest.mark.parametrize( "n_expert,topk,num_expert_group,topk_group", [ (16, 2, 8, 2), (128, 2, 8, 2), (256, 8, 8, 4), (384, 8, 1, 1), (512, 22, 1, 1), ], ) @pytest.mark.parametrize("renormalize", [True, False]) @pytest.mark.parametrize("scoring_func", ["softmax", "sigmoid"]) @pytest.mark.parametrize("routed_scaling_factor", [1.0, 2.5]) @pytest.mark.parametrize("input_dtype", [torch.bfloat16, torch.float32]) @pytest.mark.parametrize("bias_dtype", [torch.float32]) def test_grouped_topk( monkeypatch: pytest.MonkeyPatch, n_token: int, n_hidden: int, n_expert: int, topk: int, num_expert_group: int, topk_group: int, renormalize: bool, scoring_func: str, routed_scaling_factor: float, input_dtype: torch.dtype, bias_dtype: torch.dtype, ): vllm_config = VllmConfig( compilation_config=CompilationConfig(custom_ops=["all", "+grouped_topk"]) ) get_cached_compilation_config.cache_clear() set_random_seed(0) hidden_states = torch.randn((n_token, n_hidden), dtype=input_dtype, device="cuda") gating_output = torch.randn((n_token, n_expert), dtype=input_dtype, device="cuda") e_score_correction_bias = torch.randn((n_expert,), dtype=bias_dtype, device="cuda") with set_current_vllm_config(vllm_config), monkeypatch.context() as m: m.setenv("VLLM_USE_FUSED_MOE_GROUPED_TOPK", "0") m.setattr(batch_invariant, "VLLM_BATCH_INVARIANT", True) grouped_topk = GroupedTopk( topk=topk, renormalize=renormalize, num_expert_group=num_expert_group, topk_group=topk_group, scoring_func=scoring_func, routed_scaling_factor=routed_scaling_factor, ) assert grouped_topk._forward_method.__name__ == "forward_cuda" baseline_topk_weights, baseline_topk_ids = grouped_topk( hidden_states=hidden_states, gating_output=gating_output, e_score_correction_bias=e_score_correction_bias, ) test_topk_weights, test_topk_ids = fused_grouped_topk( hidden_states=hidden_states, gating_output=gating_output, topk=topk, renormalize=renormalize, num_expert_group=num_expert_group, topk_group=topk_group, scoring_func=scoring_func, routed_scaling_factor=routed_scaling_factor, e_score_correction_bias=e_score_correction_bias, ) torch.testing.assert_close( baseline_topk_weights, test_topk_weights, atol=2e-2, rtol=0 ) torch.testing.assert_close(baseline_topk_ids, test_topk_ids, atol=0, rtol=0)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/moe/test_grouped_topk.py", "license": "Apache License 2.0", "lines": 94, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/attention/test_attention_backends_selection.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests for mamba attention backend selectors.""" from types import SimpleNamespace import pytest from vllm.model_executor.layers.mamba.mamba_mixer import MambaMixer from vllm.model_executor.layers.mamba.mamba_mixer2 import MambaMixer2 from vllm.model_executor.layers.mamba.short_conv import ShortConv from vllm.model_executor.models.minimax_text_01 import MiniMaxText01LinearAttention from vllm.v1.attention.backends.linear_attn import LinearAttentionBackend from vllm.v1.attention.backends.mamba1_attn import Mamba1AttentionBackend from vllm.v1.attention.backends.mamba2_attn import Mamba2AttentionBackend from vllm.v1.attention.backends.short_conv_attn import ShortConvAttentionBackend @pytest.mark.parametrize( "layer_class, init_kwargs, expected_backend, expected_mamba_type", [ ( MambaMixer, dict( hidden_size=128, ssm_state_size=16, conv_kernel_size=4, intermediate_size=256, time_step_rank=8, use_conv_bias=True, use_bias=False, use_rms_norm=True, ), Mamba1AttentionBackend, "mamba1", ), ( MambaMixer2, dict( hidden_size=128, ssm_state_size=16, conv_kernel_size=4, intermediate_size=256, use_conv_bias=True, use_bias=False, n_groups=1, num_heads=8, head_dim=32, ), Mamba2AttentionBackend, "mamba2", ), ( MiniMaxText01LinearAttention, dict( hidden_size=128, hidden_inner_size=256, num_heads=8, head_dim=32, max_position=2048, block_size=64, num_hidden_layer=12, layer_idx=0, linear_layer_idx=0, ), LinearAttentionBackend, "linear_attention", ), ( ShortConv, dict( config=SimpleNamespace(conv_L_cache=32, conv_bias=True), dim=128, layer_idx=0, ), ShortConvAttentionBackend, "short_conv", ), ], ) def test_mamba_layers_get_attn_backend( default_vllm_config, dist_init, layer_class, init_kwargs, expected_backend, expected_mamba_type, ): """Test that Mamba-like layers return the correct attention backend.""" layer = layer_class(**init_kwargs) backend_class = layer.get_attn_backend() assert backend_class is expected_backend assert layer.mamba_type == expected_mamba_type @pytest.mark.parametrize( "layer_class,expected_backend,expected_mamba_type", [ (MambaMixer, Mamba1AttentionBackend, "mamba1"), (MambaMixer2, Mamba2AttentionBackend, "mamba2"), (MiniMaxText01LinearAttention, LinearAttentionBackend, "linear_attention"), (ShortConv, ShortConvAttentionBackend, "short_conv"), ], ) def test_mamba_layers_have_unified_interface( layer_class, expected_backend, expected_mamba_type ): """Test that all Mamba layers have the unified get_attn_backend interface.""" assert hasattr(layer_class, "get_attn_backend"), ( f"{layer_class.__name__} should have get_attn_backend method" ) assert hasattr(layer_class, "mamba_type"), ( f"{layer_class.__name__} should have mamba_type property" )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/attention/test_attention_backends_selection.py", "license": "Apache License 2.0", "lines": 108, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/layers/attention_layer_base.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Base class for attention-like layers.""" from abc import ABC, abstractmethod from vllm.config import VllmConfig from vllm.v1.attention.backend import AttentionBackend, AttentionImpl from vllm.v1.kv_cache_interface import KVCacheSpec class AttentionLayerBase(ABC): """ Base class for attention-like layers (Attention, Mamba, etc.) that support the v1 engine. This provides a common interface for getting attention backends from different layer types. """ impl: "AttentionImpl" @abstractmethod def get_attn_backend(self) -> type[AttentionBackend]: """Get the attention backend class for this layer.""" pass @abstractmethod def get_kv_cache_spec(self, vllm_config: VllmConfig) -> KVCacheSpec | None: """ Get the KV cache spec for this layer. May be None if the layer does not need KV cache. """ pass
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/attention_layer_base.py", "license": "Apache License 2.0", "lines": 26, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/benchmarks/test_random_dataset.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import random from typing import Any, NamedTuple, cast import numpy as np import pytest from transformers import AutoTokenizer, PreTrainedTokenizerBase from vllm.benchmarks.datasets import ( RandomDataset, RandomMultiModalDataset, SampleRequest, ) @pytest.fixture(scope="session") def hf_tokenizer() -> PreTrainedTokenizerBase: # Use a small, commonly available tokenizer return AutoTokenizer.from_pretrained("gpt2") class Params(NamedTuple): num_requests: int prefix_len: int range_ratio: float input_len: int output_len: int @pytest.fixture(scope="session") def random_dataset_params() -> Params: return Params( num_requests=16, prefix_len=7, range_ratio=0.3, input_len=50, output_len=20 ) def _fingerprint_sample(req: SampleRequest) -> tuple[str, int, int]: """Project a SampleRequest into a comparable tuple.""" return (req.prompt, req.prompt_len, req.expected_output_len) def _collect_samples( dataset: RandomDataset, tokenizer: PreTrainedTokenizerBase, num_requests: int = 16, prefix_len: int = 7, range_ratio: float = 0.3, input_len: int = 50, output_len: int = 20, ) -> list[tuple[str, int, int]]: samples = dataset.sample( tokenizer=tokenizer, num_requests=num_requests, prefix_len=prefix_len, range_ratio=range_ratio, input_len=input_len, output_len=output_len, ) return [_fingerprint_sample(s) for s in samples] @pytest.mark.benchmark def test_random_dataset_same_seed( hf_tokenizer: PreTrainedTokenizerBase, random_dataset_params: Params ) -> None: """Same seed should yield identical outputs, even if global RNGs change. This guards against accidental reliance on Python's random or np.random in RandomDataset after moving to numpy.default_rng. """ p = random_dataset_params common_seed = 123 dataset_a = RandomDataset(random_seed=common_seed) dataset_b = RandomDataset(random_seed=common_seed) a = _collect_samples( dataset_a, hf_tokenizer, num_requests=p.num_requests, prefix_len=p.prefix_len, range_ratio=p.range_ratio, input_len=p.input_len, output_len=p.output_len, ) # Perturb global RNG state to ensure isolation random.seed(999) _ = [random.random() for _ in range(100)] np.random.seed(888) _ = [np.random.random() for _ in range(100)] b = _collect_samples( dataset_b, hf_tokenizer, num_requests=p.num_requests, prefix_len=p.prefix_len, range_ratio=p.range_ratio, input_len=p.input_len, output_len=p.output_len, ) assert a == b @pytest.mark.benchmark def test_random_dataset_different_seeds( hf_tokenizer: PreTrainedTokenizerBase, random_dataset_params: Params ) -> None: """Different seeds should change outputs with overwhelming likelihood.""" p = random_dataset_params seed_a = 0 dataset_a = RandomDataset(random_seed=seed_a) a = _collect_samples( dataset_a, hf_tokenizer, num_requests=p.num_requests, prefix_len=p.prefix_len, range_ratio=p.range_ratio, input_len=p.input_len, output_len=p.output_len, ) seed_b = 999 dataset_b = RandomDataset(random_seed=seed_b) # Perturb global RNG with same seed as dataset_a to ensure isolation random.seed(seed_a) np.random.seed(seed_a) b = _collect_samples( dataset_b, hf_tokenizer, num_requests=p.num_requests, prefix_len=p.prefix_len, range_ratio=p.range_ratio, input_len=p.input_len, output_len=p.output_len, ) assert a != b # ----------------------------- # RandomMultiModalDataset tests # ----------------------------- def _mm_fingerprint_sample( req: SampleRequest, ) -> tuple[str, int, int, int, list[str]]: """Create a compact fingerprint for multimodal samples. Includes: - prompt string - prompt_len - expected_output_len - count of multimodal items - per-item type and URL prefix (e.g., 'data:image/jpeg;base64,') """ items = req.multi_modal_data or [] item_prefixes: list[str] = [] for it in items: if isinstance(it, dict) and it.get("type") == "image_url": url = it.get("image_url", {}).get("url", "") # Only keep a short identifying prefix to avoid huge strings item_prefixes.append(f"image:{url[:22]}") elif isinstance(it, dict) and it.get("type") == "video_url": url = it.get("video_url", {}).get("url", "") item_prefixes.append(f"video:{url[:22]}") else: item_prefixes.append("unknown:") return ( req.prompt, req.prompt_len, req.expected_output_len, len(items), item_prefixes, ) def _collect_mm_samples( dataset: RandomMultiModalDataset, tokenizer: PreTrainedTokenizerBase, *, num_requests: int = 8, prefix_len: int = 3, range_ratio: float = 0.0, input_len: int = 20, output_len: int = 5, base_items_per_request: int = 2, num_mm_items_range_ratio: float = 0.0, limit_mm_per_prompt: dict[str, int] | None = None, bucket_config: dict[tuple[int, int, int], float] | None = None, enable_multimodal_chat: bool = False, ) -> list[SampleRequest]: if limit_mm_per_prompt is None: limit_mm_per_prompt = {"image": 5, "video": 0} if bucket_config is None: bucket_config = {(32, 32, 1): 0.5, (52, 64, 1): 0.5} return dataset.sample( tokenizer=tokenizer, num_requests=num_requests, prefix_len=prefix_len, range_ratio=range_ratio, input_len=input_len, output_len=output_len, base_items_per_request=base_items_per_request, num_mm_items_range_ratio=num_mm_items_range_ratio, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, enable_multimodal_chat=enable_multimodal_chat, ) @pytest.mark.benchmark def test_random_mm_same_seed(hf_tokenizer: PreTrainedTokenizerBase) -> None: seed = 42 ds_a = RandomMultiModalDataset(random_seed=seed) ds_b = RandomMultiModalDataset(random_seed=seed) a = _collect_mm_samples(ds_a, hf_tokenizer) b = _collect_mm_samples(ds_b, hf_tokenizer) fa = [_mm_fingerprint_sample(s) for s in a] fb = [_mm_fingerprint_sample(s) for s in b] assert fa == fb @pytest.mark.benchmark def test_random_mm_different_seeds( hf_tokenizer: PreTrainedTokenizerBase, ) -> None: ds_a = RandomMultiModalDataset(random_seed=0) ds_b = RandomMultiModalDataset(random_seed=999) a = _collect_mm_samples(ds_a, hf_tokenizer) b = _collect_mm_samples(ds_b, hf_tokenizer) fa = [_mm_fingerprint_sample(s) for s in a] fb = [_mm_fingerprint_sample(s) for s in b] assert fa != fb @pytest.mark.benchmark def test_random_mm_respects_limits( hf_tokenizer: PreTrainedTokenizerBase, ) -> None: ds = RandomMultiModalDataset(random_seed=0) # Requesting 3 items with a per-prompt limit of 1 should error per current # design (dataset refuses to silently clamp below the requested baseline). with pytest.raises(ValueError): _collect_mm_samples( ds, hf_tokenizer, num_requests=12, base_items_per_request=3, num_mm_items_range_ratio=0.0, limit_mm_per_prompt={"image": 1, "video": 0}, bucket_config={(32, 32, 1): 1.0}, ) @pytest.mark.benchmark def test_random_mm_zero_prob_entries_are_removed( hf_tokenizer: PreTrainedTokenizerBase, ) -> None: ds = RandomMultiModalDataset(random_seed=0) # Second bucket has zero probability and should be ignored after # normalization samples = _collect_mm_samples( ds, hf_tokenizer, num_requests=6, base_items_per_request=2, num_mm_items_range_ratio=0.0, limit_mm_per_prompt={"image": 10, "video": 0}, bucket_config={(32, 32, 1): 1.0, (52, 64, 1): 0.0}, ) for s in samples: assert isinstance(s.multi_modal_data, list) typed_mm = cast(list[dict[str, Any]], s.multi_modal_data) for it in typed_mm: assert it.get("type") == "image_url" @pytest.mark.benchmark def test_random_mm_zero_items(hf_tokenizer: PreTrainedTokenizerBase) -> None: ds = RandomMultiModalDataset(random_seed=0) samples = _collect_mm_samples( ds, hf_tokenizer, num_requests=5, base_items_per_request=0, num_mm_items_range_ratio=0.0, limit_mm_per_prompt={"image": 5, "video": 0}, bucket_config={(32, 32, 1): 1.0}, ) for s in samples: assert s.multi_modal_data == [] @pytest.mark.benchmark def test_random_mm_num_items_per_prompt(hf_tokenizer: PreTrainedTokenizerBase) -> None: ds = RandomMultiModalDataset(random_seed=0) # Fixed number of images per prompt # set num_mm_items_range_ratio to 0.0 # TODO: modify video values when video sampling is implemented samples_fixed_items = _collect_mm_samples( ds, hf_tokenizer, num_requests=5, base_items_per_request=3, num_mm_items_range_ratio=0.0, limit_mm_per_prompt={"image": 3, "video": 0}, bucket_config={(32, 32, 1): 1.0}, ) # Must have 5 requests each with 3 mm items per prompt assert len(samples_fixed_items) == 5 for s in samples_fixed_items: mm_data = cast(list[dict[str, Any]], s.multi_modal_data) assert len(mm_data) == 3 for it in mm_data: assert it.get("type") == "image_url" @pytest.mark.benchmark def test_random_mm_bucket_config_not_mutated( hf_tokenizer: PreTrainedTokenizerBase, ) -> None: ds = RandomMultiModalDataset(random_seed=0) # This bucket config is not normalized to sum to 1 # and has more buckets than requested images original = {(32, 32, 1): 0.2, (52, 64, 1): 6, (25, 64, 1): 3} # Keep a snapshot to compare after sampling snapshot = dict(original) _ = _collect_mm_samples( ds, hf_tokenizer, num_requests=4, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt={"image": 1, "video": 0}, bucket_config=original, ) # Ensure the original dict content is unchanged assert original == snapshot # Vary number of mm items per prompt # set num_mm_items_range_ratio to 0.5 samples_varying_items = _collect_mm_samples( ds, hf_tokenizer, num_requests=5, base_items_per_request=2, num_mm_items_range_ratio=0.5, limit_mm_per_prompt={"image": 4, "video": 0}, bucket_config={(32, 32, 1): 1.0}, ) # Must have 5 requests each with less than 4 mm items per prompt # but at least 1 mm item per prompt assert len(samples_varying_items) == 5 for s in samples_varying_items: mm_data = cast(list[dict[str, Any]], s.multi_modal_data) assert len(mm_data) <= 4 assert len(mm_data) >= 1 for it in mm_data: assert it.get("type") == "image_url" @pytest.mark.benchmark def test_random_mm_video_sampling(hf_tokenizer: PreTrainedTokenizerBase) -> None: """Test video sampling functionality in RandomMultiModalDataset.""" ds = RandomMultiModalDataset(random_seed=42) # Test with video bucket configuration bucket_config = { (64, 64, 1): 0.3, # Images (64, 64, 8): 0.7, # Videos } limit_mm_per_prompt = {"image": 2, "video": 2} samples = _collect_mm_samples( ds, hf_tokenizer, num_requests=5, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, ) assert len(samples) == 5 # Check that we have both images and videos video_count = 0 image_count = 0 for s in samples: mm_data = cast(list[dict[str, Any]], s.multi_modal_data) assert len(mm_data) == 1 item = mm_data[0] if item.get("type") == "video_url": video_count += 1 # Verify video URL format url = item.get("video_url", {}).get("url", "") assert url.startswith("data:video/mp4;base64,") elif item.get("type") == "image_url": image_count += 1 # Verify image URL format url = item.get("image_url", {}).get("url", "") assert url.startswith("data:image/jpeg;base64,") # Should have some videos due to 0.7 probability assert video_count > 0 assert image_count > 0 @pytest.mark.benchmark def test_random_mm_video_only_sampling(hf_tokenizer: PreTrainedTokenizerBase) -> None: """Test sampling with only video buckets.""" ds = RandomMultiModalDataset(random_seed=42) bucket_config = { (64, 64, 8): 1.0, # Only videos } limit_mm_per_prompt = {"image": 0, "video": 1} samples = _collect_mm_samples( ds, hf_tokenizer, num_requests=3, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, ) assert len(samples) == 3 for s in samples: mm_data = cast(list[dict[str, Any]], s.multi_modal_data) assert len(mm_data) == 1 item = mm_data[0] assert item.get("type") == "video_url" url = item.get("video_url", {}).get("url", "") assert url.startswith("data:video/mp4;base64,") @pytest.mark.benchmark def test_random_mm_video_deterministic_sampling( hf_tokenizer: PreTrainedTokenizerBase, ) -> None: """Test that video sampling is deterministic with same seed.""" seed = 123 ds_a = RandomMultiModalDataset(random_seed=seed) ds_b = RandomMultiModalDataset(random_seed=seed) bucket_config = { (64, 64, 8): 1.0, # Only videos } limit_mm_per_prompt = {"image": 0, "video": 1} a = _collect_mm_samples( ds_a, hf_tokenizer, num_requests=3, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, ) b = _collect_mm_samples( ds_b, hf_tokenizer, num_requests=3, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, ) fa = [_mm_fingerprint_sample(s) for s in a] fb = [_mm_fingerprint_sample(s) for s in b] assert fa == fb
{ "repo_id": "vllm-project/vllm", "file_path": "tests/benchmarks/test_random_dataset.py", "license": "Apache License 2.0", "lines": 415, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/core/test_encoder_cache_manager.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from vllm.multimodal.inputs import MultiModalFeatureSpec, PlaceholderRange from vllm.v1.core.encoder_cache_manager import ( EncoderCacheManager, EncoderDecoderCacheManager, ) pytestmark = pytest.mark.cpu_test # ------------------ Mock Classes ------------------ # class MockRequest: def __init__(self, request_id, mm_hashes, token_counts): self.request_id = request_id self._token_counts = token_counts self.mm_features = [] for i, mm_hash in enumerate(mm_hashes): feature = MultiModalFeatureSpec( data=None, modality="image", identifier=mm_hash, mm_position=PlaceholderRange(offset=0, length=self._token_counts[i]), ) self.mm_features.append(feature) def get_num_encoder_embeds(self, input_id: int) -> int: return self._token_counts[input_id] # ------------------ Unit Tests ------------------ # def test_basic_allocate_and_reuse(): cache = EncoderCacheManager(cache_size=10) req = MockRequest("r1", ["imgA"], [4]) assert not cache.check_and_update_cache(req, 0) assert cache.can_allocate(req, 0, int(1e9), 0) cache.allocate(req, 0) assert cache.check_and_update_cache(req, 0) assert "r1" in cache.cached["imgA"] assert cache.num_free_slots == 6 # Free twice to bring refcount to 0. cache.free_encoder_input(req, 0) cache.free_encoder_input(req, 0) assert not cache.cached["imgA"] assert "imgA" in cache.freeable assert cache.num_freeable_slots == 10 assert cache.num_free_slots == 6 def test_freeing_decreases_refcount_and_moves_to_freeable(): manager = EncoderCacheManager(cache_size=10) req = MockRequest("req2", ["img3"], [5]) assert manager.can_allocate(req, 0, int(1e9), 0) manager.allocate(req, 0) assert len(manager.cached["img3"]) == 1 manager.free_encoder_input(req, 0) assert not manager.cached["img3"] assert "img3" in manager.freeable assert manager.num_freeable_slots == 10 def test_free_request_frees_all_inputs(): manager = EncoderCacheManager(cache_size=10) req = MockRequest("req3", ["a", "b"], [2, 3]) assert manager.can_allocate(req, 0, int(1e9), 0) manager.allocate(req, 0) assert manager.can_allocate(req, 1, int(1e9), 0) manager.allocate(req, 1) assert len(manager.cached["a"]) == 1 assert len(manager.cached["b"]) == 1 manager.free(req) assert not manager.cached["a"] assert not manager.cached["b"] assert "a" in manager.freeable assert "b" in manager.freeable assert manager.num_freeable_slots == 10 def test_eviction_when_cache_is_full(): manager = EncoderCacheManager(cache_size=10) req1 = MockRequest("req1", ["x"], [6]) req2 = MockRequest("req2", ["y"], [5]) assert manager.can_allocate(req1, 0, int(1e9), 0) manager.allocate(req1, 0) manager.free_encoder_input(req1, 0) assert manager.can_allocate(req2, 0, int(1e9), 0) manager.allocate(req2, 0) # 'x' should have been evicted. assert "x" not in manager.cached assert "x" in manager.get_freed_mm_hashes() def test_get_cached_input_ids(): manager = EncoderCacheManager(cache_size=10) req = MockRequest("reqX", ["m", "n", "o"], [2, 4, 3]) assert manager.can_allocate(req, 0, int(1e9), 0) manager.allocate(req, 0) assert manager.can_allocate(req, 2, int(1e9), 0) manager.allocate(req, 2) cached_ids = manager.get_cached_input_ids(req) assert cached_ids == {0, 2} def test_has_cache_restores_from_freeable(): manager = EncoderCacheManager(cache_size=10) req = MockRequest("reqY", ["imgZ"], [4]) assert manager.can_allocate(req, 0, int(1e9), 0) manager.allocate(req, 0) manager.free_encoder_input(req, 0) # Should restore from freeable. assert manager.check_and_update_cache(req, 0) assert len(manager.cached["imgZ"]) == 1 assert "imgZ" not in manager.freeable assert manager.num_freeable_slots == 6 def test_get_freed_mm_hashes_clears_freed_list(): manager = EncoderCacheManager(cache_size=10) req1 = MockRequest("reqA", ["a"], [5]) req2 = MockRequest("reqB", ["b"], [6]) assert manager.can_allocate(req1, 0, int(1e9), 0) manager.allocate(req1, 0) manager.free_encoder_input(req1, 0) # Should trigger eviction of 'a'. assert manager.can_allocate(req2, 0, int(1e9), 0) manager.allocate(req2, 0) freed = manager.get_freed_mm_hashes() assert "a" in freed assert manager.get_freed_mm_hashes() == [] def test_schedule_request_multi_images_respect_space_limit(): manager = EncoderCacheManager(cache_size=10) req = MockRequest("reqA", ["a", "b"], [5, 6]) compute_budget = 100 num_tokens_to_schedule = 0 assert manager.can_allocate(req, 0, compute_budget, num_tokens_to_schedule) num_tokens_to_schedule += req.get_num_encoder_embeds(0) compute_budget -= req.get_num_encoder_embeds(0) assert not manager.can_allocate(req, 1, compute_budget, num_tokens_to_schedule) def test_schedule_request_multi_images_respect_compute_limit(): manager = EncoderCacheManager(cache_size=100) req = MockRequest("reqA", ["a", "b"], [5, 6]) compute_budget = 10 num_tokens_to_schedule = 0 assert manager.can_allocate(req, 0, compute_budget, num_tokens_to_schedule) num_tokens_to_schedule += req.get_num_encoder_embeds(0) compute_budget -= req.get_num_encoder_embeds(0) assert not manager.can_allocate(req, 1, compute_budget, num_tokens_to_schedule) def test_encoder_cache_with_is_embed_mask(): class MockRequestWithMask(MockRequest): def get_num_encoder_embeds(self, input_id: int) -> int: return self.mm_features[input_id].mm_position.get_num_embeds() is_embed = torch.zeros(100, dtype=torch.bool) is_embed[torch.tensor([5, 15, 25, 35, 45, 55, 65, 75])] = True request = MockRequestWithMask("r1", ["img1"], [100]) request.mm_features[0] = MultiModalFeatureSpec( data=None, modality="image", identifier="img1", mm_position=PlaceholderRange(offset=0, length=100, is_embed=is_embed), ) manager = EncoderCacheManager(cache_size=100) manager.allocate(request, 0) assert manager.num_free_slots == 92 assert "img1" in manager.cached old_size = 100 new_size = request.mm_features[0].mm_position.get_num_embeds() assert new_size == 8 savings_ratio = old_size / new_size assert savings_ratio == 12.5 def test_encoder_cache_mask_based_retrieval(): class MockRequestWithMask(MockRequest): def get_num_encoder_embeds(self, input_id: int) -> int: return self.mm_features[input_id].mm_position.get_num_embeds() is_embed = torch.tensor( [False, False, True, True, False, True, True, True, False, False] ) request = MockRequestWithMask("r1", ["img1"], [10]) request.mm_features[0] = MultiModalFeatureSpec( data=None, modality="image", identifier="img1", mm_position=PlaceholderRange(offset=0, length=10, is_embed=is_embed), ) manager = EncoderCacheManager(cache_size=50) manager.allocate(request, 0) assert request.mm_features[0].mm_position.get_num_embeds() == 5 start_idx = 2 end_idx = 8 num_embeds_before = is_embed[:start_idx].sum().item() num_embeds_in_range = is_embed[start_idx:end_idx].sum().item() assert num_embeds_before == 0 assert num_embeds_in_range == 5 start_idx = 0 end_idx = 5 num_embeds_before = is_embed[:start_idx].sum().item() if start_idx > 0 else 0 num_embeds_in_range = is_embed[start_idx:end_idx].sum().item() assert num_embeds_before == 0 assert num_embeds_in_range == 2 def test_reset_clears_all_state(): """Test that reset() clears all cached entries and restores capacity.""" manager = EncoderCacheManager(cache_size=20) req1 = MockRequest("req1", ["img1", "img2"], [5, 3]) req2 = MockRequest("req2", ["img3"], [4]) manager.allocate(req1, 0) manager.allocate(req1, 1) manager.allocate(req2, 0) manager.free_encoder_input(req1, 0) req3 = MockRequest("req3", ["img4"], [10]) manager.free_encoder_input(req1, 1) manager.free_encoder_input(req2, 0) manager.can_allocate(req3, 0, int(1e9), 0) manager.allocate(req3, 0) assert len(manager.cached) > 0 assert manager.num_free_slots < 20 manager.reset() assert len(manager.cached) == 0 assert len(manager.freeable) == 0 assert len(manager.freed) == 0 assert manager.num_free_slots == 20 assert manager.num_freeable_slots == 20 def test_reset_allows_fresh_allocations(): manager = EncoderCacheManager(cache_size=10) req1 = MockRequest("req1", ["img1"], [10]) manager.allocate(req1, 0) assert manager.num_free_slots == 0 manager.reset() req2 = MockRequest("req2", ["img2"], [8]) assert manager.can_allocate(req2, 0, int(1e9), 0) manager.allocate(req2, 0) assert manager.num_free_slots == 2 assert "img2" in manager.cached assert "img1" not in manager.cached def test_encoder_decoder_cache_manager_reset(): manager = EncoderDecoderCacheManager(cache_size=20) req1 = MockRequest("req1", ["img1"], [5]) req2 = MockRequest("req2", ["img2"], [3]) manager.allocate(req1, 0) manager.allocate(req2, 0) manager.free(req1) manager.get_freed_mm_hashes() assert manager.num_free_slots < 20 manager.reset() assert len(manager.allocated) == 0 assert len(manager.to_free) == 0 assert manager.num_free_slots == 20 def test_encoder_decoder_cache_manager_reset_allows_fresh_allocations(): manager = EncoderDecoderCacheManager(cache_size=10) req1 = MockRequest("req1", ["img1"], [10]) manager.allocate(req1, 0) assert manager.num_free_slots == 0 manager.reset() req2 = MockRequest("req2", ["img2"], [8]) assert manager.can_allocate(req2, 0, int(1e9), 0) manager.allocate(req2, 0) assert manager.num_free_slots == 2 assert "img2" in manager.allocated
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/core/test_encoder_cache_manager.py", "license": "Apache License 2.0", "lines": 240, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/v1/structured_output/backend_lm_format_enforcer.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import ast import json from dataclasses import dataclass, field from functools import lru_cache from typing import TYPE_CHECKING import torch from transformers import PreTrainedTokenizerBase from vllm.sampling_params import SamplingParams from vllm.utils.import_utils import LazyLoader from vllm.v1.structured_output.backend_types import ( StructuredOutputBackend, StructuredOutputGrammar, StructuredOutputOptions, ) if TYPE_CHECKING: import lmformatenforcer import lmformatenforcer.integrations.vllm as lmfe_vllm else: lmformatenforcer = LazyLoader("lmformatenforcer", globals(), "lmformatenforcer") lmfe_vllm = LazyLoader( "lmformatenforcer.integrations.vllm", globals(), "lmformatenforcer.integrations.vllm", ) @lru_cache def _cached_build_vllm_token_enforcer_tokenizer_data( tokenizer: PreTrainedTokenizerBase, vocab_size: int ) -> "lmfe_vllm.TokenEnforcerTokenizerData": return lmfe_vllm.build_vllm_token_enforcer_tokenizer_data( tokenizer, use_bitmask=True, vocab_size=vocab_size ) @dataclass class LMFormatEnforcerGrammar(StructuredOutputGrammar): token_enforcer: lmformatenforcer.TokenEnforcer current_tokens_prefix: list[int] = field(default_factory=list) def accept_tokens(self, request_id: str, tokens: list[int]) -> bool: original_len = len(self.current_tokens_prefix) for token in tokens: if not self.token_enforcer.get_allowed_tokens( self.current_tokens_prefix ).is_token_allowed(token): # Rollback partial updates to ensure atomicity. del self.current_tokens_prefix[original_len:] return False self.current_tokens_prefix.append(token) return True def validate_tokens(self, tokens: list[int]) -> list[int]: for prefix_length in range(len(tokens)): prefix = tokens[:prefix_length] next_token = tokens[prefix_length] if not self.token_enforcer.get_allowed_tokens( self.current_tokens_prefix + prefix ).is_token_allowed(next_token): break else: return tokens return tokens[:prefix_length] def rollback(self, num_tokens: int) -> None: self.current_tokens_prefix = self.current_tokens_prefix[:-num_tokens] def fill_bitmask(self, bitmask: torch.Tensor, batch_index: int) -> None: allowed_tokens = self.token_enforcer.get_allowed_tokens( self.current_tokens_prefix ) bitmask[batch_index] = allowed_tokens.allowed_tokens def is_terminated(self) -> bool: # We are considered terminated if the prefix ends with eos_token_id return_value = ( len(self.current_tokens_prefix) > 0 and self.current_tokens_prefix[-1] == self.token_enforcer.eos_token_id ) return return_value def reset(self): self.current_tokens_prefix = [] @dataclass class LMFormatEnforcerBackend(StructuredOutputBackend): def __post_init__(self): self.tokenizer_data = _cached_build_vllm_token_enforcer_tokenizer_data( self.tokenizer, self.vocab_size ) def compile_grammar( self, request_type: StructuredOutputOptions, grammar_spec: str ) -> StructuredOutputGrammar: character_level_parser: lmformatenforcer.CharacterLevelParser if request_type == StructuredOutputOptions.JSON: spec_dict = json.loads(grammar_spec) character_level_parser = lmformatenforcer.JsonSchemaParser(spec_dict) elif request_type == StructuredOutputOptions.JSON_OBJECT: character_level_parser = lmformatenforcer.JsonSchemaParser(None) elif request_type == StructuredOutputOptions.REGEX: character_level_parser = lmformatenforcer.RegexParser(grammar_spec) elif request_type == StructuredOutputOptions.CHOICE: choices = ast.literal_eval(grammar_spec) character_level_parser = lmformatenforcer.UnionParser( [lmformatenforcer.StringParser(choice) for choice in choices] ) else: raise ValueError( f"Invalid request type for LM Format Enforcer backend({request_type!s})" ) max_rollback_tokens = ( self.vllm_config.speculative_config.num_speculative_tokens if self.vllm_config.speculative_config is not None else 0 ) if max_rollback_tokens > 0: raise ValueError( "LM Format Enforcer backend does not support speculative tokens" ) token_enforcer = lmformatenforcer.TokenEnforcer( tokenizer_data=self.tokenizer_data, parser=character_level_parser, ) return LMFormatEnforcerGrammar(token_enforcer) def allocate_token_bitmask(self, max_num_seqs: int) -> torch.Tensor: return torch.full( (max_num_seqs, (self.vocab_size + 31) // 32), -1, dtype=torch.int32, pin_memory=torch.cuda.is_available(), ) def destroy(self): pass def validate_structured_output_request_lm_format_enforcer(params: SamplingParams): if params.structured_outputs is None: return so_params = params.structured_outputs if so_params.regex: return elif so_params.json: if isinstance(so_params.json, str): try: # make sure schema is valid json json.loads(so_params.json) except json.JSONDecodeError as e: raise ValueError("Invalid JSON grammar specification.") from e else: try: json.dumps(so_params.json) except Exception as e: raise ValueError( f"Error serializing structured outputs jsonschema: {e}" ) from e return elif so_params.choice: return elif so_params.grammar: raise ValueError( "LM Format Enforcer structured outputs backend " "does not support grammar specifications" )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/structured_output/backend_lm_format_enforcer.py", "license": "Apache License 2.0", "lines": 152, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/swin.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Iterable import torch import torch.nn as nn from transformers import SwinConfig from transformers.models.swin.modeling_swin import SwinEmbeddings, SwinPatchMerging from transformers.models.swin.modeling_swin import SwinLayer as HFSwinLayer from transformers.pytorch_utils import meshgrid from vllm.model_executor.layers.activation import get_act_fn from vllm.model_executor.layers.linear import ( ColumnParallelLinear, QKVParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.model_loader.weight_utils import default_weight_loader class SwinSelfAttention(nn.Module): def __init__( self, config: SwinConfig, dim: int, num_heads: int, window_size: int, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() if dim % num_heads != 0: raise ValueError( f"The hidden size ({dim}) is not a multiple of the number of " f"attention heads ({num_heads})" ) self.num_attention_heads = num_heads self.attention_head_size = int(dim / num_heads) self.all_head_size = self.num_attention_heads * self.attention_head_size self.window_size = ( window_size if isinstance(window_size, Iterable) else (window_size, window_size) ) self.scale = self.attention_head_size**-0.5 self.relative_position_bias_table = nn.Parameter( torch.zeros( (2 * self.window_size[0] - 1) * (2 * self.window_size[1] - 1), num_heads ) ) # get pair-wise relative position index for each token inside the window coords_h = torch.arange(self.window_size[0]) coords_w = torch.arange(self.window_size[1]) coords = torch.stack(meshgrid([coords_h, coords_w], indexing="ij")) coords_flatten = torch.flatten(coords, 1) relative_coords = coords_flatten[:, :, None] - coords_flatten[:, None, :] relative_coords = relative_coords.permute(1, 2, 0).contiguous() relative_coords[:, :, 0] += self.window_size[0] - 1 relative_coords[:, :, 1] += self.window_size[1] - 1 relative_coords[:, :, 0] *= 2 * self.window_size[1] - 1 relative_position_index = relative_coords.sum(-1) self.relative_position_index = nn.Parameter( relative_position_index, requires_grad=False ) self.qkv = QKVParallelLinear( hidden_size=dim, head_size=self.attention_head_size, total_num_heads=self.num_attention_heads, bias=config.qkv_bias, quant_config=quant_config, prefix=f"{prefix}.qkv", ) def transpose_for_scores(self, x): new_x_shape = x.size()[:-1] + ( self.num_attention_heads, self.attention_head_size, ) x = x.view(new_x_shape) return x.permute(0, 2, 1, 3) def _get_rel_pos_bias(self) -> torch.Tensor: relative_position_bias = self.relative_position_bias_table[ self.relative_position_index.view(-1) ] relative_position_bias = relative_position_bias.view( self.window_size[0] * self.window_size[1], self.window_size[0] * self.window_size[1], -1, ) relative_position_bias = relative_position_bias.permute(2, 0, 1).contiguous() return relative_position_bias.unsqueeze(0) def forward( self, hidden_states: torch.Tensor, attention_mask: torch.FloatTensor | None = None, output_attentions: bool | None = False, ) -> tuple[torch.Tensor, ...]: batch_size, dim, num_channels = hidden_states.shape qkv_output, _ = self.qkv(hidden_states) query_layer, key_layer, value_layer = qkv_output.chunk(3, dim=-1) key_layer = self.transpose_for_scores(key_layer) value_layer = self.transpose_for_scores(value_layer) query_layer = self.transpose_for_scores(query_layer) attention_scores = self._get_rel_pos_bias() if attention_mask is not None: mask_shape = attention_mask.shape[0] attention_mask_expanded = attention_mask.view( 1, mask_shape, 1, dim, dim ).expand( batch_size // mask_shape, mask_shape, self.num_attention_heads, dim, dim ) attention_scores = attention_scores + attention_mask_expanded.unsqueeze( 1 ).unsqueeze(0) attention_scores = attention_scores.view( -1, self.num_attention_heads, dim, dim ) context_layer = torch.nn.functional.scaled_dot_product_attention( query_layer, key_layer, value_layer, attn_mask=attention_scores, dropout_p=0.0, ) attention_probs = None context_layer = context_layer.permute(0, 2, 1, 3).contiguous() new_context_layer_shape = context_layer.size()[:-2] + (self.all_head_size,) context_layer = context_layer.view(new_context_layer_shape) outputs = ( (context_layer, attention_probs) if output_attentions else (context_layer,) ) return outputs class SwinSelfOutput(nn.Module): def __init__( self, config: SwinConfig, dim: int, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.dense = RowParallelLinear( input_size=dim, output_size=dim, quant_config=quant_config, prefix=f"{prefix}.dense", ) def forward( self, hidden_states: torch.Tensor, input_tensor: torch.Tensor ) -> torch.Tensor: hidden_states, _ = self.dense(hidden_states) return hidden_states class SwinAttention(nn.Module): def __init__( self, config: SwinConfig, dim: int, num_heads: int, window_size: int, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.self = SwinSelfAttention( config, dim, num_heads, window_size, quant_config=quant_config, prefix=f"{prefix}.self", ) self.output = SwinSelfOutput( config, dim, quant_config=quant_config, prefix=f"{prefix}.output" ) self.pruned_heads = set() def forward( self, hidden_states: torch.Tensor, attention_mask: torch.FloatTensor | None = None, output_attentions: bool | None = False, ) -> tuple[torch.Tensor]: self_outputs = self.self(hidden_states, attention_mask, output_attentions) attention_output = self.output(self_outputs[0], hidden_states) outputs = (attention_output,) + self_outputs[1:] return outputs class SwinIntermediate(nn.Module): def __init__( self, config: SwinConfig, dim: int, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.dense = ColumnParallelLinear( dim, int(config.mlp_ratio * dim), quant_config=quant_config, prefix=f"{prefix}.dense", ) self.intermediate_act_fn = get_act_fn(config.hidden_act) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states, _ = self.dense(hidden_states) hidden_states = self.intermediate_act_fn(hidden_states) return hidden_states class SwinOutput(nn.Module): def __init__( self, config: SwinConfig, dim: int, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.dense = RowParallelLinear( int(config.mlp_ratio * dim), dim, quant_config=quant_config, prefix=f"{prefix}.dense", ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states, _ = self.dense(hidden_states) return hidden_states class SwinLayer(HFSwinLayer): def __init__( self, config: SwinConfig, dim: int, input_resolution: int, num_heads: int, drop_path_rate: float = 0.0, shift_size: int = 0, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__( config=config, dim=dim, input_resolution=input_resolution, num_heads=num_heads, drop_path_rate=drop_path_rate, shift_size=shift_size, ) self.attention = SwinAttention( config, dim, num_heads, window_size=self.window_size, quant_config=quant_config, prefix=f"{prefix}.attention", ) self.intermediate = SwinIntermediate( config, dim, quant_config=quant_config, prefix=f"{prefix}.intermediate" ) self.output = SwinOutput( config, dim, quant_config=quant_config, prefix=f"{prefix}.output" ) class SwinStage(nn.Module): def __init__( self, config: SwinConfig, dim: int, input_resolution: int, depth: int, num_heads: int, drop_path: list[float], downsample: SwinPatchMerging | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.config = config self.dim = dim self.blocks = nn.ModuleList( [ SwinLayer( config=config, dim=dim, input_resolution=input_resolution, num_heads=num_heads, drop_path_rate=drop_path[layer_idx], shift_size=0 if (layer_idx % 2 == 0) else config.window_size // 2, quant_config=quant_config, prefix=f"{prefix}.blocks.{layer_idx}", ) for layer_idx in range(depth) ] ) # patch merging layer if downsample is not None: self.downsample = downsample( input_resolution, dim=dim, norm_layer=nn.LayerNorm ) else: self.downsample = None self.pointing = False def forward( self, hidden_states: torch.Tensor, input_dimensions: tuple[int, int], output_attentions: bool | None = False, always_partition: bool | None = False, ) -> tuple[torch.Tensor]: height, width = input_dimensions for i, layer_module in enumerate(self.blocks): layer_outputs = layer_module( hidden_states, input_dimensions, output_attentions, always_partition, ) hidden_states = layer_outputs[0] hidden_states_before_downsampling = hidden_states if self.downsample is not None: height_downsampled, width_downsampled = (height + 1) // 2, (width + 1) // 2 output_dimensions = (height, width, height_downsampled, width_downsampled) hidden_states = self.downsample( hidden_states_before_downsampling, input_dimensions ) else: output_dimensions = (height, width, height, width) stage_outputs = ( hidden_states, hidden_states_before_downsampling, output_dimensions, ) if output_attentions: stage_outputs += layer_outputs[1:] return stage_outputs class SwinEncoder(nn.Module): def __init__( self, config: SwinConfig, grid_size: int, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.num_layers = len(config.depths) self.config = config dpr = [ x.item() for x in torch.linspace( 0, config.drop_path_rate, sum(config.depths), device="cpu" ) ] self.layers = nn.ModuleList( [ SwinStage( config=config, dim=int(config.embed_dim * 2**layer_idx), input_resolution=( grid_size[0] // (2**layer_idx), grid_size[1] // (2**layer_idx), ), depth=config.depths[layer_idx], num_heads=config.num_heads[layer_idx], drop_path=dpr[ sum(config.depths[:layer_idx]) : sum( config.depths[: layer_idx + 1] ) ], downsample=SwinPatchMerging if (layer_idx < self.num_layers - 1) else None, quant_config=quant_config, prefix=f"{prefix}.layers.{layer_idx}", ) for layer_idx in range(self.num_layers) ] ) def forward( self, hidden_states: torch.Tensor, input_dimensions: tuple[int, int], output_attentions: bool | None = False, always_partition: bool | None = False, ) -> tuple[torch.Tensor]: for i, layer_module in enumerate(self.layers): layer_outputs = layer_module( hidden_states, input_dimensions, output_attentions, always_partition, ) hidden_states = layer_outputs[0] output_dimensions = layer_outputs[2] input_dimensions = (output_dimensions[-2], output_dimensions[-1]) return hidden_states class SwinModel(nn.Module): config_class: SwinConfig def __init__( self, config: SwinConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.config = config self.num_layers = len(config.depths) self.num_features = int(config.embed_dim * 2 ** (self.num_layers - 1)) self.embeddings = SwinEmbeddings(config) self.encoder = SwinEncoder( config, self.embeddings.patch_grid, quant_config=quant_config, prefix=f"{prefix}.encoder", ) def forward( self, pixel_values: torch.FloatTensor | None = None, output_attentions: bool | None = None, ) -> tuple[torch.Tensor]: embedding_output, input_dimensions = self.embeddings(pixel_values) encoder_outputs = self.encoder( embedding_output, input_dimensions, output_attentions=output_attentions, ) return encoder_outputs def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ ("qkv", "query", "q"), ("qkv", "key", "k"), ("qkv", "value", "v"), ] params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() for name, loaded_weight in weights: for param_name, weight_name, shard_id in stacked_params_mapping: if weight_name not in name: continue name = name.replace(weight_name, param_name) param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/swin.py", "license": "Apache License 2.0", "lines": 439, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/quantization/test_cutlass_w4a8.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests for the CUTLASS W4A8 kernel. Run `pytest tests/kernels/quantization/test_cutlass_w4a8.py`. """ from dataclasses import dataclass import pytest import torch from vllm import _custom_ops as ops from vllm.model_executor.layers.quantization.utils.quant_utils import ( convert_packed_uint4b8_to_signed_int4_inplace, pack_cols, pack_rows, quantize_weights, unpack_quantized_values_into_int32, ) from vllm.platforms import current_platform from vllm.scalar_type import ScalarType, scalar_types if not current_platform.is_cuda(): pytest.skip("These tests use CUTLASS which requires CUDA", allow_module_level=True) # TODO: in future PR refactor this and `is_quant_method_supported` in the kernel # unit tests to a common utility function. Currently the use of # `is_quant_method_supported` conflates kernels with quantization methods # an assumption which is breaking down as quantizations methods can have # have kernels and some kernels support multiple quantization methods. IS_SUPPORTED_BY_GPU = current_platform.get_device_capability()[0] >= 9 MNK_SHAPES = [ (1, 128, 128), (1, 512, 1024), (1, 4096, 4096), (1, 8192, 28672), (13, 8192, 4096), (26, 4096, 8192), (64, 4096, 4096), (64, 8192, 28672), (257, 128, 4096), (257, 4096, 4096), (1024, 4096, 8192), (1024, 8192, 4096), ] # TODO(czhu): get supported schedules from fn SCHEDULES = [ "128x16_1x1x1", "256x16_1x1x1", "128x32_1x1x1", "256x32_1x1x1", "128x64_1x1x1", "256x64_1x1x1", "128x128_1x1x1", "256x128_1x1x1", "128x256_1x1x1", "128x256_2x1x1", ] @dataclass class TypeConfig: act_type: torch.dtype weight_type: ScalarType output_type: torch.dtype | None group_scale_type: torch.dtype | None channel_scale_type: torch.dtype | None token_scale_type: torch.dtype | None @dataclass class Tensors: w_ref: torch.Tensor a_ref: torch.Tensor a: torch.Tensor w_q: torch.Tensor w_g_s: torch.Tensor w_ch_s: torch.Tensor w_tok_s: torch.Tensor # (Act Type, Weight Type, Output Type, Scale Type, ZeroPoints, # Ch Scales Type, Tok Scales Type) TestTypeTuple = tuple[ list[torch.dtype], ScalarType, torch.dtype | None, torch.dtype | None, bool ] TEST_TYPES = [ *( TypeConfig( act_type=torch.float8_e4m3fn, weight_type=w_type, output_type=o_type, group_scale_type=torch.float8_e4m3fn, channel_scale_type=torch.float32, token_scale_type=torch.float32, ) for w_type in [scalar_types.int4] # TODO(czhu): fp16 out type for o_type in [torch.bfloat16] ), ] # TODO: in future PR refactor this and `is_quant_method_supported` in the kernel # unit tests to a common utility function. Currently the use of # `is_quant_method_supported` conflates kernels with quantization methods # an assumption which is breaking down as quantizations methods can have # have kernels and some kernels support multiple quantization methods. IS_SUPPORTED_BY_GPU = current_platform.has_device_capability(90) # For testing quantized linear kernels def to_fp8(tensor: torch.Tensor): finfo = torch.finfo(torch.float8_e4m3fn) return tensor.clamp(min=finfo.min, max=finfo.max).to(dtype=torch.float8_e4m3fn) def cutlass_quantize_and_pack( atype: torch.dtype, w: torch.Tensor, wtype: ScalarType, stype: torch.dtype | None, group_size: int | None, zero_points: bool = False, ): assert wtype.is_integer(), "TODO: support floating point weights" w_ref, w_q, w_s, w_zp = quantize_weights( w, wtype, group_size=group_size, zero_points=zero_points ) # since scales are cast to fp8, we need to compute w_ref this way w_ref = ( (w_q).to(torch.float32) * w_s.to(atype).to(torch.float32).repeat_interleave(group_size, dim=0) ).to(atype) # bit mask prevents sign extending int4 when packing w_q = pack_rows(w_q & 0x0F, wtype.size_bits, *w_q.shape) w_q = w_q.t().contiguous().t() # convert to col major w_q_packed = ops.cutlass_encode_and_reorder_int4b(w_q) w_s_packed = ops.cutlass_pack_scale_fp8(w_s.to(atype)) return w_ref, w_q_packed, w_s_packed, w_zp def create_test_tensors( shape: tuple[int, int, int], types: TypeConfig, group_size: int | None ) -> Tensors: m, n, k = shape print( "create_test_tensors, shape:", shape, "types:", types, "group_size:", group_size ) a = to_fp8(torch.randn((m, k), device="cuda")) w = to_fp8(torch.randn((k, n), device="cuda")) if types.group_scale_type is not None: w = w.to(types.group_scale_type) if w.dtype.itemsize == 1: w = w.to(torch.float16) w_ref, w_q_packed, w_s, _ = cutlass_quantize_and_pack( a.dtype, w, types.weight_type, types.group_scale_type, group_size, False ) a_ref = a.to(torch.float32) w_ref = w_ref.to(torch.float32) # for the practical use case we need per-tok scales for fp8 activations w_tok_s = torch.randn((m,), device="cuda", dtype=types.token_scale_type) w_ch_s = torch.randn((n,), device="cuda", dtype=types.channel_scale_type) return Tensors( w_ref=w_ref, a_ref=a_ref, a=a, w_q=w_q_packed, w_g_s=w_s, w_ch_s=w_ch_s, w_tok_s=w_tok_s, ) def mm_test_helper( types: TypeConfig, tensors: Tensors, group_size: int | None = None, schedule: str | None = None, ): # CUTLASS upstream uses fp8 with fastaccum as reference # https://github.com/NVIDIA/cutlass/blob/main/examples/55_hopper_mixed_dtype_gemm/55_hopper_int4_fp8_gemm.cu#L406 output_ref = torch._scaled_mm( tensors.a_ref.to(types.act_type), tensors.w_ref.to(types.act_type).t().contiguous().t(), # col major tensors.w_tok_s.unsqueeze(1), tensors.w_ch_s.unsqueeze(0), out_dtype=types.output_type, use_fast_accum=True, ) output = ops.cutlass_w4a8_mm( a=tensors.a, b_q=tensors.w_q, b_group_scales=tensors.w_g_s, b_group_size=group_size, b_channel_scales=tensors.w_ch_s, a_token_scales=tensors.w_tok_s, ) print(output) print(output_ref) torch.testing.assert_close( output, output_ref.to(output.dtype), rtol=1e-2, atol=1e-2 ) @pytest.mark.skipif( not IS_SUPPORTED_BY_GPU, reason="CUTLASS W4A8 is not supported on this GPU type." ) @pytest.mark.parametrize("shape", MNK_SHAPES, ids=lambda x: "x".join(str(v) for v in x)) @pytest.mark.parametrize("types", TEST_TYPES) @pytest.mark.parametrize("schedule", SCHEDULES) def test_cutlass_w4a8(shape, types: TypeConfig, schedule): group_sizes = [128] for group_size in group_sizes: tensors = create_test_tensors(shape, types, group_size) mm_test_helper(types, tensors, group_size, schedule) # Test to make sure cuda graphs work class W4A8Layer(torch.nn.Module): def __init__(self, **kwargs): super().__init__() self.kwargs = kwargs def forward(self, a): return ops.cutlass_w4a8_mm(a=a, **self.kwargs) @pytest.mark.skipif( not IS_SUPPORTED_BY_GPU, reason="CUTLASS W4A8 is not supported on this GPU type." ) def test_w4a8_cuda_graph(): m, n, k = 512, 4096, 4096 a = to_fp8(torch.randn((m, k), device="cuda")) b = to_fp8(torch.randn((k, n), device="cuda")) wtype = scalar_types.int4 stype = torch.float8_e4m3fn group_size = 128 zero_points = False w_ref, w_q_packed, w_s, _ = cutlass_quantize_and_pack( a.dtype, b.to(torch.float16), wtype, stype, group_size, zero_points ) w_tok_s = torch.randn((m,), device="cuda", dtype=torch.float32) w_ch_s = torch.randn((n,), device="cuda", dtype=torch.float32) # Construct a trivial model with a single layer that calls the kernel model = W4A8Layer( b_q=w_q_packed, b_group_scales=w_s, b_group_size=group_size, b_channel_scales=w_ch_s, a_token_scales=w_tok_s, ) output_ref = torch._scaled_mm( a, w_ref.to(a.dtype).t().contiguous().t(), # col major w_tok_s.unsqueeze(1), w_ch_s.unsqueeze(0), out_dtype=torch.bfloat16, use_fast_accum=True, ) # Run the model with a cuda graph stream = torch.cuda.Stream() with torch.cuda.stream(stream): g = torch.cuda.CUDAGraph() with torch.cuda.graph(g): output = model(a) output.zero_() g.replay() torch.testing.assert_close(output, output_ref, rtol=1e-2, atol=1e-2) @pytest.mark.skipif( not IS_SUPPORTED_BY_GPU, reason="CUTLASS W4A8 is not supported on this GPU type." ) @pytest.mark.parametrize("shape", MNK_SHAPES) def test_convert_packed_uint4b8_to_signed_int4_inplace(shape): """ The W4A16 checkpoints encode the weights as int4b8 packed to int32. The CUTLASS kernels expect signed int4 packed to int32. This tests checks that the runtime int4b8 -> signed int4 conversion matches the offline conversion step exactly. """ _, N, K = shape # random weights packed to int32 t = torch.randint( low=torch.iinfo(torch.int32).min, high=torch.iinfo(torch.int32).max + 1, size=(N, K // 8), dtype=torch.int32, device="cuda", ) # compute reference unpacked = unpack_quantized_values_into_int32( t.clone(), scalar_types.uint4b8, packed_dim=1 ) unpacked = unpacked - 8 # int4b8 -> signed int4 ref = pack_cols(unpacked & 0x0F, 4, *unpacked.shape) out = convert_packed_uint4b8_to_signed_int4_inplace(t.clone()) assert torch.equal(ref, out) assert not torch.equal(ref, t)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/quantization/test_cutlass_w4a8.py", "license": "Apache License 2.0", "lines": 270, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/layers/quantization/compressed_tensors/schemes/compressed_tensors_w4a8_fp8.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable import torch from compressed_tensors.quantization import ActivationOrdering from vllm.logger import init_logger from vllm.model_executor.kernels.linear import ( MPLinearLayerConfig, choose_mp_linear_kernel, ) from vllm.model_executor.layers.quantization.compressed_tensors.schemes import ( CompressedTensorsScheme, ) from vllm.model_executor.layers.quantization.utils.marlin_utils import ( marlin_repeat_scales_on_all_ranks, ) from vllm.model_executor.parameter import ( BasevLLMParameter, ChannelQuantScaleParameter, GroupQuantScaleParameter, PackedvLLMParameter, ) from vllm.scalar_type import scalar_types logger = init_logger(__name__) __all__ = ["CompressedTensorsW4A8Fp8"] W4A8_SUPPORTED_TYPES_MAP = { 4: scalar_types.int4, } W4A8_SUPPORTED_BITS = list(W4A8_SUPPORTED_TYPES_MAP.keys()) class CompressedTensorsW4A8Fp8(CompressedTensorsScheme): _kernel_backends_being_used: set[str] = set() def __init__( self, strategy: str, num_bits: int, group_size: int | None = None, symmetric: bool | None = True, actorder: ActivationOrdering | None = None, ): self.pack_factor = 32 // num_bits self.strategy = strategy self.symmetric = symmetric self.group_size = -1 if group_size is None else group_size self.has_g_idx = actorder == ActivationOrdering.GROUP if self.group_size != 128 or self.strategy != "group": raise ValueError( "W4A8 kernels require group quantization with group size 128" ) if num_bits not in W4A8_SUPPORTED_TYPES_MAP: raise ValueError( f"Unsupported num_bits = {num_bits}. " f"Supported num_bits = {W4A8_SUPPORTED_TYPES_MAP.keys()}" ) self.quant_type = W4A8_SUPPORTED_TYPES_MAP[num_bits] @classmethod def get_min_capability(cls) -> int: # hopper return 90 def create_weights( self, layer: torch.nn.Module, output_size: int, input_size: int, output_partition_sizes: list[int], input_size_per_partition: int, params_dtype: torch.dtype, weight_loader: Callable, **kwargs, ): output_size_per_partition = sum(output_partition_sizes) mp_linear_kernel_config = MPLinearLayerConfig( full_weight_shape=(input_size, output_size), partition_weight_shape=( input_size_per_partition, output_size_per_partition, ), weight_type=self.quant_type, act_type=torch.float8_e4m3fn, # always use fp8(e4m3) group_size=self.group_size, zero_points=not self.symmetric, has_g_idx=self.has_g_idx, out_type=params_dtype, ) kernel_type = choose_mp_linear_kernel(mp_linear_kernel_config) if kernel_type.__name__ not in self._kernel_backends_being_used: logger.info("Using %s for CompressedTensorsW4A8Fp8", kernel_type.__name__) self._kernel_backends_being_used.add(kernel_type.__name__) # If group_size is -1, we are in channelwise case. group_size = self.group_size if self.group_size != -1 else input_size row_parallel = input_size != input_size_per_partition partition_scales = not marlin_repeat_scales_on_all_ranks( self.has_g_idx, self.group_size, row_parallel ) scales_and_zp_size = input_size // group_size if partition_scales: assert input_size_per_partition % group_size == 0 scales_and_zp_size = input_size_per_partition // group_size weight = PackedvLLMParameter( input_dim=1, output_dim=0, weight_loader=weight_loader, packed_factor=self.pack_factor, packed_dim=1, data=torch.empty( output_size_per_partition, input_size_per_partition // self.pack_factor, dtype=torch.int32, ), ) # After loading, we will transform bf16 -> fp8 -> # expand by 8x via `cutlass_pack_scale_fp8` # and construct per-channel fp32 scales. weight_scale_args = { "weight_loader": weight_loader, "data": torch.empty( output_size_per_partition, scales_and_zp_size, dtype=params_dtype, ), } if not partition_scales: weight_scale = ChannelQuantScaleParameter(output_dim=0, **weight_scale_args) else: weight_scale = GroupQuantScaleParameter( output_dim=0, input_dim=1, **weight_scale_args ) # A 2D array defining the original shape of the weights # before packing weight_shape = BasevLLMParameter( data=torch.empty(2, dtype=torch.int64), weight_loader=weight_loader ) layer.register_parameter("weight_packed", weight) layer.register_parameter("weight_scale", weight_scale) layer.register_parameter("weight_shape", weight_shape) self.kernel = kernel_type( mp_linear_kernel_config, w_q_param_name="weight_packed", w_s_param_name="weight_scale", w_zp_param_name="weight_zero_point", w_gidx_param_name="weight_g_idx", ) # Checkpoints are serialized in compressed-tensors format, which is # different from the format the kernel may want. Handle repacking here. def process_weights_after_loading(self, layer: torch.nn.Module) -> None: self.kernel.process_weights_after_loading(layer) def apply_weights( self, layer: torch.nn.Module, x: torch.Tensor, bias: torch.Tensor | None ) -> torch.Tensor: return self.kernel.apply_weights(layer, x, bias)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/compressed_tensors/schemes/compressed_tensors_w4a8_fp8.py", "license": "Apache License 2.0", "lines": 149, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/quantization/utils/petit_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import TYPE_CHECKING import torch # TYPE_CHECKING is used for static type analysis to prevent circular imports. if TYPE_CHECKING: from types import ModuleType # 1. Create a global variable as a placeholder for the module _petit_kernel: "ModuleType | None" = None _PETIT_INSTALL_MSG = ( "Petit is not installed. Please install it with `pip install petit-kernel`." ) def _import_petit_kernel() -> "ModuleType": """ A helper function to handle the lazy import. The first time this function is called, it will import the petit_kernel library and store it in the global _petit_kernel variable. Subsequent calls will return the already-loaded module directly. """ global _petit_kernel if _petit_kernel is not None: return _petit_kernel try: import petit_kernel _petit_kernel = petit_kernel return _petit_kernel except ImportError: # The 'from None' syntax prevents chaining the original ImportError, # making the traceback cleaner. raise ImportError(_PETIT_INSTALL_MSG) from None def _check_petit_nvfp4_supported( quant_method: str, group_size: int | None ) -> tuple[bool, str | None]: if quant_method != "NVFP4": return ( False, ( "Petit currently only supports: NVFP4 quantizations in sglang. " "Please check the `hf_quant_config.json` file for your model's " "quant configuration." ), ) if group_size is not None and group_size != 16: return ( False, "Petit currently only supports: group_size=16 quantizations.", ) return (True, None) def verify_petit_nvfp4_supported(quant_method: str, group_size: int | None) -> None: supported, error_msg = _check_petit_nvfp4_supported(quant_method, group_size) if not supported: assert error_msg is not None raise ValueError(error_msg) def prepare_nvfp4_layer_for_petit(layer: torch.nn.Module) -> None: # 2. Call _import_petit_kernel() to trigger (or get) the import. petit_kernel = _import_petit_kernel() # Repack weights to petit format part_size_n = layer.output_size_per_partition part_size_k = layer.input_size_per_partition qweight = layer.weight.view(torch.int32).contiguous() # 3. Call functions through the imported module variable. petit_qweight = petit_kernel.repack_nvfp4( qweight, size_n=part_size_n, size_k=part_size_k ) layer.weight = torch.nn.Parameter(petit_qweight, requires_grad=False) # Permute scales weight_scale = petit_kernel.process_nvfp4_scales( scales=layer.weight_scale, size_k=part_size_k, size_n=part_size_n ) layer.weight_scale = torch.nn.Parameter(weight_scale, requires_grad=False) def apply_petit_nvfp4_linear( input: torch.Tensor, weight: torch.Tensor, weight_scale: torch.Tensor, weight_scale_2: torch.Tensor, size_n: int, size_k: int, bias: torch.Tensor | None = None, ) -> torch.Tensor: # Trigger (or get) the import here as well. petit_kernel = _import_petit_kernel() reshaped_x = input.reshape(-1, input.shape[-1]) out_shape = input.shape[:-1] + (size_n,) # TODO: Use auto-tuning to find the performant solution_id # Call the function via the module variable. output = petit_kernel.mul_nvfp4_a16( a=reshaped_x, b=weight, s=weight_scale, global_scale=weight_scale_2, size_m=reshaped_x.size(0), size_n=size_n, size_k=size_k, solution_id=-1, ) if bias is not None: output.add_(bias) # In-place add return output.reshape(out_shape)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/quantization/utils/petit_utils.py", "license": "Apache License 2.0", "lines": 98, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/distributed/test_symm_mem_allreduce.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import queue import random import typing import pytest import torch import torch.distributed as dist import torch.multiprocessing as mp import vllm.envs as envs from vllm.config import ParallelConfig, VllmConfig, set_current_vllm_config from vllm.distributed import cleanup_dist_env_and_memory from vllm.distributed.communication_op import tensor_model_parallel_all_reduce from vllm.distributed.device_communicators.cuda_communicator import CudaCommunicator from vllm.distributed.parallel_state import ( get_tp_group, init_distributed_environment, initialize_model_parallel, ) from vllm.engine.arg_utils import EngineArgs from vllm.engine.llm_engine import LLMEngine from vllm.platforms import current_platform from vllm.utils.system_utils import update_environment_variables torch.manual_seed(42) random.seed(44) test_size_elements = 1024 * 1024 def symm_mem_allreduce_worker(local_rank: int, world_size: int, q: mp.Queue): monkeypatch = pytest.MonkeyPatch() config = VllmConfig(parallel_config=ParallelConfig(tensor_parallel_size=world_size)) with monkeypatch.context() as m, set_current_vllm_config(config): m.delenv("CUDA_VISIBLE_DEVICES", raising=False) dtype = torch.bfloat16 device = torch.device(f"cuda:{local_rank}") torch.cuda.set_device(device) torch.set_default_device(device) torch.set_default_dtype(dtype) update_environment_variables( { "RANK": str(local_rank), "LOCAL_RANK": str(local_rank), "WORLD_SIZE": str(world_size), "MASTER_ADDR": "localhost", "MASTER_PORT": "12345", } ) init_distributed_environment() initialize_model_parallel(tensor_model_parallel_size=world_size) cuda_communicator = typing.cast( CudaCommunicator, get_tp_group().device_communicator ) symm_mem_comm = cuda_communicator.symm_mem_comm if symm_mem_comm is None or symm_mem_comm.disabled: # can't use skip under multiprocessing q.put("SymmMemCommunicator is not available or disabled.") return inp_direct_symm_mem = torch.randint( 1, 23, (test_size_elements,), dtype=dtype, device=device ) if not symm_mem_comm.should_use_symm_mem(inp_direct_symm_mem): # can't use skip under multiprocessing q.put("SymmMemCommunicator isn't used for this world and input size.") return original_inp_direct_symm_mem = inp_direct_symm_mem.clone() out_direct_symm_mem = symm_mem_comm.all_reduce(inp_direct_symm_mem) assert out_direct_symm_mem is not None group = get_tp_group().device_group dist.all_reduce(original_inp_direct_symm_mem, group=group) torch.testing.assert_close( out_direct_symm_mem, original_inp_direct_symm_mem, atol=2.5, rtol=0.1 ) # Test tensor_model_parallel_all_reduce which should use symm_mem inp_tensor_parallel = torch.randint( -23, 1, (test_size_elements,), dtype=dtype, device=device ) original_inp_tensor_parallel = inp_tensor_parallel.clone() out_tensor_parallel = tensor_model_parallel_all_reduce(inp_tensor_parallel) dist.all_reduce(original_inp_tensor_parallel, group=group) torch.testing.assert_close( out_tensor_parallel, original_inp_tensor_parallel, atol=2.5, rtol=0.1 ) @pytest.mark.skipif( not current_platform.is_cuda(), reason="SymmMemAllreduce is only available for CUDA platforms.", ) @pytest.mark.parametrize("tp_size", [2]) @pytest.mark.parametrize("pipeline_parallel_size", [1]) @pytest.mark.skipif(envs.VLLM_TARGET_DEVICE not in ["cuda"], reason="Only test on CUDA") def test_symm_mem_allreduce( monkeypatch: pytest.MonkeyPatch, tp_size, pipeline_parallel_size ): world_size = tp_size * pipeline_parallel_size if world_size > torch.cuda.device_count(): pytest.skip("Not enough GPUs to run the test.") q = mp.get_context("spawn").Queue() mp.spawn(symm_mem_allreduce_worker, args=(world_size, q), nprocs=world_size) try: val = q.get(timeout=1) except queue.Empty: val = None finally: cleanup_dist_env_and_memory() if val is not None: pytest.skip(val) @pytest.mark.skipif( not current_platform.is_cuda(), reason="SymmMemAllreduce is only available for CUDA platforms.", ) @pytest.mark.skipif(envs.VLLM_TARGET_DEVICE not in ["cuda"], reason="Only test on CUDA") def test_dp_with_symm_mem_allreduce(monkeypatch: pytest.MonkeyPatch): world_size = 4 if world_size > torch.cuda.device_count(): pytest.skip("Not enough GPUs to run the test.") # Verify that the DataParallel runs without error engine_args = EngineArgs( model="distilbert/distilgpt2", enforce_eager=True, enable_prefix_caching=True, data_parallel_size=2, tensor_parallel_size=2, data_parallel_backend="mp", ) LLMEngine.from_engine_args(engine_args)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/distributed/test_symm_mem_allreduce.py", "license": "Apache License 2.0", "lines": 122, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/distributed/device_communicators/symm_mem.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import torch.distributed as dist from torch.distributed import ProcessGroup from vllm.distributed.device_communicators.all_reduce_utils import ( SYMM_MEM_ALL_REDUCE_MAX_SIZES, ) from vllm.logger import init_logger from vllm.model_executor.layers.batch_invariant import ( vllm_is_batch_invariant, ) from vllm.platforms import current_platform try: import torch.distributed._symmetric_memory as torch_symm_mem symm_mem_available = True except ImportError: symm_mem_available = False logger = init_logger(__name__) class SymmMemCommunicator: _WORLD_SIZES_MULTIMEM = { "9.0": [4, 6, 8], "10.0": [6, 8], } def __init__( self, group: ProcessGroup, device: int | str | torch.device, # add options for testing force_multimem: bool | None = None, max_size_override: int | None = None, ): self.disabled = True if not symm_mem_available: return if not current_platform.is_cuda(): logger.warning("SymmMemCommunicator: symmetric memory is not available.") return if isinstance(device, int): device = torch.device(f"cuda:{device}") elif isinstance(device, str): device = torch.device(device) torch.cuda.set_device(device) self.dtype = torch.bfloat16 self.device = device self.group = group self.world_size = dist.get_world_size(self.group) capability = current_platform.get_device_capability() if capability is None: logger.warning( "SymmMemCommunicator: device capability is unknown, " "communicator is not available." ) return self.device_capability = capability.as_version_str() if self.device_capability not in SYMM_MEM_ALL_REDUCE_MAX_SIZES: logger.warning( "SymmMemCommunicator: Device capability %s not supported, " "communicator is not available.", self.device_capability, ) return if self.world_size not in SYMM_MEM_ALL_REDUCE_MAX_SIZES[self.device_capability]: logger.warning( "SymmMemCommunicator: World size %d not supported, " "communicator is not available.", self.world_size, ) return # Use override max_size if provided, otherwise use default if max_size_override is not None: self.max_size = max_size_override logger.info( "SymmMemCommunicator: Using override max_size: %s bytes", self.max_size, ) else: self.max_size = SYMM_MEM_ALL_REDUCE_MAX_SIZES[self.device_capability][ self.world_size ] try: self.buffer = torch_symm_mem.empty( self.max_size // self.dtype.itemsize, device=self.device, dtype=self.dtype, ) handle = torch_symm_mem.rendezvous(self.buffer, self.group.group_name) except RuntimeError as e: logger.warning_once( "SymmMemCommunicator: symmetric memory initialization failed: %s " "Communicator is not available. To suppress this warning set " "VLLM_ALLREDUCE_USE_SYMM_MEM=0", str(e), ) return if handle.multicast_ptr == 0: logger.warning( "SymmMemCommunicator: symmetric memory " "multicast operations are not supported." ) return self.force_multimem = force_multimem self.disabled = False if vllm_is_batch_invariant(): self.disabled = True def should_use_symm_mem(self, inp: torch.Tensor): if self.disabled: return False if inp.dtype != self.dtype: return False inp_size = inp.numel() * inp.element_size() if inp_size % 4 != 0: return False return inp_size < self.max_size def all_reduce( self, inp: torch.Tensor, *, out: torch.Tensor | None = None ) -> torch.Tensor | None: if not self.should_use_symm_mem(inp): return None if out is None: out = torch.empty_like(inp) self.buffer[: inp.numel()].copy_(inp.view(-1)) # Determine which algorithm to use use_multimem = False if self.force_multimem is not None: # Test override: use forced setting use_multimem = self.force_multimem else: # Normal logic: use multimem for supported world sizes use_multimem = ( self.world_size in self._WORLD_SIZES_MULTIMEM[self.device_capability] ) if use_multimem: torch.ops.symm_mem.multimem_all_reduce_( self.buffer[: inp.numel()], "sum", self.group.group_name ) else: torch.ops.symm_mem.two_shot_all_reduce_( self.buffer[: inp.numel()], "sum", self.group.group_name ) out.copy_(self.buffer[: inp.numel()].view(out.shape)) return out
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/distributed/device_communicators/symm_mem.py", "license": "Apache License 2.0", "lines": 142, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/seed_oss.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright 2025 The Seed team. # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Inference-only SeedOss model compatible with HuggingFace weights.""" from collections.abc import Iterable from itertools import islice import torch from torch import nn from transformers import PretrainedConfig as SeedOssConfig from vllm.compilation.decorators import support_torch_compile from vllm.config import CacheConfig, VllmConfig from vllm.distributed import get_pp_group, get_tensor_model_parallel_world_size from vllm.logger import init_logger from vllm.model_executor.layers.activation import SiluAndMul from vllm.model_executor.layers.attention import Attention from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( MergedColumnParallelLinear, QKVParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import ( default_weight_loader, maybe_remap_kv_scale_name, ) from vllm.sequence import IntermediateTensors from vllm.transformers_utils.config import set_default_rope_theta from vllm.v1.attention.backend import AttentionType from .interfaces import SupportsLoRA, SupportsPP from .utils import ( AutoWeightsLoader, PPMissingLayer, is_pp_missing_parameter, make_empty_intermediate_tensors_factory, make_layers, maybe_prefix, ) logger = init_logger(__name__) class SeedOssMLP(nn.Module): def __init__( self, hidden_size: int, intermediate_size: int, hidden_act: str, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.gate_up_proj = MergedColumnParallelLinear( hidden_size, [intermediate_size] * 2, bias=False, quant_config=quant_config, prefix=f"{prefix}.gate_up_proj", ) self.down_proj = RowParallelLinear( intermediate_size, hidden_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.down_proj", ) if hidden_act != "silu": raise ValueError( f"Unsupported activation: {hidden_act}. Only silu is supported for now." ) self.act_fn = SiluAndMul() def forward(self, x): gate_up, _ = self.gate_up_proj(x) x = self.act_fn(gate_up) x, _ = self.down_proj(x) return x class SeedOssAttention(nn.Module): def __init__( self, hidden_size: int, num_heads: int, num_kv_heads: int, head_dim: int, rope_parameters: dict, max_position: int = 4096 * 32, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", attn_type: str = AttentionType.DECODER, ) -> None: super().__init__() self.hidden_size = hidden_size tp_size = get_tensor_model_parallel_world_size() self.total_num_heads = num_heads assert self.total_num_heads % tp_size == 0 self.num_heads = self.total_num_heads // tp_size self.total_num_kv_heads = num_kv_heads self.head_dim = head_dim if self.total_num_kv_heads >= tp_size: # Number of KV heads is greater than TP size, so we partition # the KV heads across multiple tensor parallel GPUs. assert self.total_num_kv_heads % tp_size == 0 else: # Number of KV heads is less than TP size, so we replicate # the KV heads across multiple tensor parallel GPUs. assert tp_size % self.total_num_kv_heads == 0 self.num_kv_heads = max(1, self.total_num_kv_heads // tp_size) self.q_size = self.num_heads * self.head_dim self.kv_size = self.num_kv_heads * self.head_dim self.scaling = self.head_dim**-0.5 self.qkv_proj = QKVParallelLinear( hidden_size, self.head_dim, self.total_num_heads, self.total_num_kv_heads, bias=True, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", ) self.o_proj = RowParallelLinear( self.total_num_heads * self.head_dim, hidden_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.o_proj", ) self.rotary_emb = get_rope( self.head_dim, max_position=max_position, rope_parameters=rope_parameters, ) self.attn = Attention( self.num_heads, self.head_dim, self.scaling, num_kv_heads=self.num_kv_heads, cache_config=cache_config, quant_config=quant_config, attn_type=attn_type, prefix=f"{prefix}.attn", ) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, ) -> torch.Tensor: qkv, _ = self.qkv_proj(hidden_states) q, k, v = qkv.split([self.q_size, self.kv_size, self.kv_size], dim=-1) q, k = self.rotary_emb(positions, q, k) attn_output = self.attn(q, k, v) output, _ = self.o_proj(attn_output) return output class SeedOssDecoderLayer(nn.Module): def __init__( self, config: SeedOssConfig, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.hidden_size = config.hidden_size set_default_rope_theta(config, default_theta=1000000) # By default, SeedOss uses causal attention as it is a # decoder-only model. # You can override the HF config with `is_causal=False` to enable # bidirectional attention, which is used in some embedding models if getattr(config, "is_causal", True): attn_type = AttentionType.DECODER else: attn_type = AttentionType.ENCODER_ONLY self.self_attn = SeedOssAttention( hidden_size=self.hidden_size, num_heads=config.num_attention_heads, max_position=config.max_position_embeddings, num_kv_heads=config.num_key_value_heads, head_dim=config.head_dim, cache_config=cache_config, quant_config=quant_config, rope_parameters=config.rope_parameters, prefix=f"{prefix}.self_attn", attn_type=attn_type, ) self.mlp = SeedOssMLP( hidden_size=self.hidden_size, intermediate_size=config.intermediate_size, hidden_act=config.hidden_act, quant_config=quant_config, prefix=f"{prefix}.mlp", ) self.input_layernorm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.post_attention_layernorm = RMSNorm( config.hidden_size, eps=config.rms_norm_eps ) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, residual: torch.Tensor | None, ) -> tuple[torch.Tensor, torch.Tensor]: # Self Attention if residual is None: residual = hidden_states hidden_states = self.input_layernorm(hidden_states) else: hidden_states, residual = self.input_layernorm(hidden_states, residual) hidden_states = self.self_attn( positions=positions, hidden_states=hidden_states, ) # Fully Connected hidden_states, residual = self.post_attention_layernorm(hidden_states, residual) hidden_states = self.mlp(hidden_states) return hidden_states, residual @support_torch_compile( dynamic_arg_dims={ "input_ids": 0, "positions": -1, "intermediate_tensors": 0, "inputs_embeds": 0, } ) class SeedOssModel(nn.Module): def __init__( self, *, vllm_config: VllmConfig, prefix: str = "", decoder_layer_type: type[nn.Module] = SeedOssDecoderLayer, ): super().__init__() config = vllm_config.model_config.hf_config cache_config = vllm_config.cache_config quant_config = vllm_config.quant_config # TODO (@robertgshaw2): see if this can be moved out if cache_config.sliding_window is not None and hasattr( config, "max_window_layers" ): assert config.max_window_layers == config.num_hidden_layers, ( "Sliding window for some but all layers is not supported. " "This model uses sliding window but `max_window_layers` = {} " "is less than `num_hidden_layers` = {}. Please open an issue " "to discuss this feature.".format( config.max_window_layers, config.num_hidden_layers, ) ) self.config = config self.quant_config = quant_config self.vocab_size = config.vocab_size if get_pp_group().is_first_rank or ( config.tie_word_embeddings and get_pp_group().is_last_rank ): self.embed_tokens = VocabParallelEmbedding( config.vocab_size, config.hidden_size, quant_config=quant_config, prefix=f"{prefix}.embed_tokens", ) else: self.embed_tokens = PPMissingLayer() # Use the provided decoder layer type or default to SeedDecoderLayer decoder_layer_type = decoder_layer_type or SeedOssDecoderLayer self.start_layer, self.end_layer, self.layers = make_layers( config.num_hidden_layers, lambda prefix: decoder_layer_type( config=config, cache_config=cache_config, quant_config=quant_config, prefix=prefix, ), prefix=f"{prefix}.layers", ) self.make_empty_intermediate_tensors = make_empty_intermediate_tensors_factory( ["hidden_states", "residual"], config.hidden_size ) if get_pp_group().is_last_rank: self.norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) else: self.norm = PPMissingLayer() def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor | IntermediateTensors: if get_pp_group().is_first_rank: if inputs_embeds is not None: hidden_states = inputs_embeds else: hidden_states = self.embed_input_ids(input_ids) residual = None else: assert intermediate_tensors is not None hidden_states = intermediate_tensors["hidden_states"] residual = intermediate_tensors["residual"] for layer in islice(self.layers, self.start_layer, self.end_layer): hidden_states, residual = layer( positions, hidden_states, residual, ) if not get_pp_group().is_last_rank: return IntermediateTensors( {"hidden_states": hidden_states, "residual": residual} ) hidden_states, _ = self.norm(hidden_states, residual) return hidden_states def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ # (param_name, shard_name, shard_id) ("qkv_proj", "q_proj", "q"), ("qkv_proj", "k_proj", "k"), ("qkv_proj", "v_proj", "v"), ("gate_up_proj", "gate_proj", 0), ("gate_up_proj", "up_proj", 1), ] params_dict = dict(self.named_parameters(remove_duplicate=False)) loaded_params: set[str] = set() for name, loaded_weight in weights: if "rotary_emb.inv_freq" in name: continue if self.quant_config is not None and ( scale_name := self.quant_config.get_cache_scale(name) ): # Loading kv cache quantization scales param = params_dict[scale_name] weight_loader = getattr(param, "weight_loader", default_weight_loader) loaded_weight = ( loaded_weight if loaded_weight.dim() == 0 else loaded_weight[0] ) weight_loader(param, loaded_weight) loaded_params.add(scale_name) continue for param_name, weight_name, shard_id in stacked_params_mapping: if weight_name not in name: continue name = name.replace(weight_name, param_name) # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue # Remapping the name of FP8 kv-scale. name = maybe_remap_kv_scale_name(name, params_dict) if name is None: continue if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params class SeedOssForCausalLM(nn.Module, SupportsLoRA, SupportsPP): packed_modules_mapping = { "qkv_proj": [ "q_proj", "k_proj", "v_proj", ], "gate_up_proj": [ "gate_proj", "up_proj", ], } def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config self.config = config self.quant_config = quant_config self.model = SeedOssModel( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model") ) if get_pp_group().is_last_rank: if config.tie_word_embeddings: self.lm_head = self.model.embed_tokens else: self.lm_head = ParallelLMHead( config.vocab_size, config.hidden_size, quant_config=quant_config, prefix=maybe_prefix(prefix, "lm_head"), ) else: self.lm_head = PPMissingLayer() self.logits_processor = LogitsProcessor(config.vocab_size) self.make_empty_intermediate_tensors = ( self.model.make_empty_intermediate_tensors ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor | IntermediateTensors: hidden_states = self.model( input_ids, positions, intermediate_tensors, inputs_embeds ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: logits = self.logits_processor(self.lm_head, hidden_states) return logits def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader( self, skip_prefixes=(["lm_head."] if self.config.tie_word_embeddings else None), ) return loader.load_weights(weights)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/seed_oss.py", "license": "Apache License 2.0", "lines": 449, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/v1/e2e/test_min_tokens.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Comprehensive end-to-end tests for `min_tokens` in the V1 engine. Addresses #21950: verify and add CI coverage. Covers: 1) Basic functionality 2) Stop strings with `min_tokens` (bug #21987; fix in PR #22014) 3) EOS behavior with `min_tokens` (potential logits-processor bug) 4) Edge cases (min_tokens == max_tokens, min_tokens == 0) 5) Multiple stop conditions """ import pytest from vllm import LLM, SamplingParams from vllm.outputs import RequestOutput # Test configuration TEST_MODEL = "facebook/opt-125m" # Small model for fast CI execution GREEDY = 0.0 # Deterministic generation for consistent testing class MinTokensTestCase: """Data class for min_tokens test scenarios""" def __init__( self, name: str, min_tokens: int, max_tokens: int, stop: str | list[str] | None = None, expected_min_len: int | None = None, expected_exact_len: int | None = None, ): self.name = name self.min_tokens = min_tokens self.max_tokens = max_tokens self.stop = stop self.expected_min_len = expected_min_len or min_tokens self.expected_exact_len = expected_exact_len def __str__(self): return ( f"{self.name}: min={self.min_tokens}, " f"max={self.max_tokens}, stop={self.stop}" ) # Test scenarios covering all critical cases MIN_TOKENS_TEST_CASES = [ # === BASIC FUNCTIONALITY (should work) === MinTokensTestCase( name="basic_min_tokens_no_stop", min_tokens=8, max_tokens=20, stop=None, expected_min_len=8, ), MinTokensTestCase( name="min_tokens_zero", min_tokens=0, max_tokens=10, stop=None, expected_min_len=0, ), MinTokensTestCase( name="min_equals_max_no_stop", min_tokens=15, max_tokens=15, stop=None, expected_exact_len=15, ), # === STOP STRINGS WITH MIN_TOKENS === # These tests expose the detokenizer bug where stop strings # bypass min_tokens # Using mathematically guaranteed approach with wide stop nets pytest.param( MinTokensTestCase( name="min_tokens_with_comprehensive_stops", min_tokens=5, max_tokens=20, stop=[ "a", "e", "i", "o", "u", "t", "n", "s", "r", "l", " ", ], expected_min_len=5, ), marks=pytest.mark.xfail( reason=( "Known bug #21987: stop strings bypass min_tokens (fixed by PR #22014)" ), strict=False, ), id="min_tokens_with_comprehensive_stops", ), pytest.param( MinTokensTestCase( name="min_tokens_with_simple_char_stop", min_tokens=3, max_tokens=15, stop=["e", "a", " "], expected_min_len=3, ), marks=pytest.mark.xfail( reason=( "Known bug #21987: stop strings bypass min_tokens (fixed by PR #22014)" ), strict=False, ), id="min_tokens_with_simple_char_stop", ), # === EOS TOKEN WITH MIN_TOKENS (potential LogitsProcessor bug) === # These test the MinTokensLogitsProcessor handling of EOS tokens pytest.param( MinTokensTestCase( name="min_equals_max_eos_only", min_tokens=20, max_tokens=20, stop=None, # Relies on default EOS token behavior expected_exact_len=20, ), marks=pytest.mark.xfail( reason=("Potential logits-processor bug: EOS tokens may bypass min_tokens"), strict=False, ), id="min_equals_max_eos_only", ), # === EDGE CASES === MinTokensTestCase( name="large_min_tokens", min_tokens=50, max_tokens=60, stop=None, expected_min_len=50, ), MinTokensTestCase( name="min_tokens_with_empty_stop_list", min_tokens=5, max_tokens=15, stop=[], # Empty stop list expected_min_len=5, ), ] @pytest.fixture(scope="module") def llm_v1(): """Create V1 LLM instance for testing""" llm = LLM( model=TEST_MODEL, tensor_parallel_size=1, max_model_len=1024, # Small context for fast testing enforce_eager=True, # Avoid graph compilation overhead ) return llm def get_token_count(output: RequestOutput) -> int: """Extract token count from LLM output""" if not output.outputs: return 0 return len(output.outputs[0].token_ids) def assert_min_tokens_satisfied( output: RequestOutput, test_case: MinTokensTestCase ) -> None: """Assert that min_tokens requirement is satisfied""" token_count = get_token_count(output) stop_reason = output.outputs[0].stop_reason if output.outputs else "no output" if test_case.expected_exact_len is not None: # Exact length requirement assert token_count == test_case.expected_exact_len, ( f"Expected exactly {test_case.expected_exact_len} tokens, " f"got {token_count} tokens. " f"Stop reason: {stop_reason}" ) else: # Minimum length requirement assert token_count >= (test_case.expected_min_len or 0), ( f"Expected at least {test_case.expected_min_len} tokens, " f"got {token_count} tokens. " f"Stop reason: {stop_reason}" ) @pytest.mark.parametrize( "test_case", MIN_TOKENS_TEST_CASES, ids=lambda tc: tc.name, ) def test_min_tokens_comprehensive(llm_v1: LLM, test_case: MinTokensTestCase): """ Comprehensive test for min_tokens functionality in V1 engine. This test covers all critical scenarios for min_tokens: - Basic functionality (should work) - Stop strings with min_tokens (known bug) - EOS tokens with min_tokens (potential bug) - Edge cases Args: llm_v1: V1 LLM instance test_case: Test scenario parameters """ # Known failing cases are handled via param-level xfail marks above. # Create sampling parameters sampling_params = SamplingParams( min_tokens=test_case.min_tokens, max_tokens=test_case.max_tokens, stop=test_case.stop, temperature=GREEDY, include_stop_str_in_output=True, # Include stop strings for debugging ) # Use simple prompt. Comprehensive stop lists should catch any generation prompt = "Hello" # Generate output outputs = llm_v1.generate([prompt], sampling_params) assert len(outputs) == 1, "Expected exactly one output" output = outputs[0] # Debug information token_count = get_token_count(output) generated_text = output.outputs[0].text if output.outputs else "" stop_reason = output.outputs[0].stop_reason if output.outputs else "unknown" print(f"\nTest: {test_case.name}") print(f"Generated {token_count} tokens") print(f"Stop reason: {stop_reason}") print(f"Generated text: {repr(generated_text)}") print(f"Expected min: {test_case.expected_min_len}") if test_case.expected_exact_len: print(f"Expected exact: {test_case.expected_exact_len}") # Validate min_tokens requirement assert_min_tokens_satisfied(output, test_case) def test_min_tokens_basic_functionality(llm_v1: LLM): """ Test basic min_tokens functionality without stop conditions. This is a baseline test that should always pass and validates that min_tokens works correctly in the simple case. """ sampling_params = SamplingParams(min_tokens=10, max_tokens=20, temperature=GREEDY) prompt = "Once upon a time" outputs = llm_v1.generate([prompt], sampling_params) assert len(outputs) == 1 token_count = get_token_count(outputs[0]) assert token_count >= 10, f"Expected at least 10 tokens, got {token_count}" assert token_count <= 20, f"Expected at most 20 tokens, got {token_count}" @pytest.mark.xfail( reason=("Known bug #21987: stop strings bypass min_tokens (fixed by PR #22014)"), strict=False, ) def test_min_tokens_stop_strings_bug(llm_v1: LLM): """ Test the specific bug where stop strings bypass min_tokens. This test specifically reproduces the bug Calvin is fixing in PR #22014. It should fail until that fix is merged. Strategy: Use guaranteed stop characters that will appear in any generated text. """ # If the bug is fixed upstream, this test will XPASS sampling_params = SamplingParams( min_tokens=15, max_tokens=50, # Common letter; likely appears early stop=["e"], temperature=GREEDY, include_stop_str_in_output=True, ) # Simple prompt that will generate text containing "e" prompt = "The quick brown fox" outputs = llm_v1.generate([prompt], sampling_params) assert len(outputs) == 1 token_count = get_token_count(outputs[0]) generated_text = outputs[0].outputs[0].text if outputs[0].outputs else "" # Debug info to understand what happened print(f"Generated text: {repr(generated_text)}") print(f"Token count: {token_count}") print(f"Contains 'e': {'e' in generated_text}") # This assertion should fail due to the bug - if stop string is found early, # the model should still continue generating until min_tokens is reached stop_reason = ( outputs[0].outputs[0].stop_reason if outputs[0].outputs else "no output" ) assert token_count >= 15, ( "Bug confirmed: " f"{token_count} tokens < min_tokens=15. " f"Reason: {stop_reason}. " f"Text: {repr(generated_text)}" ) @pytest.mark.xfail( reason=("Known bug #21987: stop strings bypass min_tokens (fixed by PR #22014)"), strict=False, ) def test_min_tokens_stop_strings_guaranteed_early_trigger(llm_v1: LLM): """ Guaranteed test for stop strings bypassing min_tokens bug. Strategy: Use very low temperature and multiple common stop strings to virtually guarantee early detection, combined with long min_tokens to ensure the bug is exposed regardless of model behavior. """ # If the bug is fixed upstream, this test will XPASS sampling_params = SamplingParams( min_tokens=50, # Set high min_tokens to ensure bug detection max_tokens=200, # Use multiple very common patterns - at least one will appear stop=["e", "a", "i", "o", "u", " ", "t", "n", "s", "r"], temperature=GREEDY, include_stop_str_in_output=True, ) # Simple prompt that will generate some text prompt = "The cat" outputs = llm_v1.generate([prompt], sampling_params) assert len(outputs) == 1 token_count = get_token_count(outputs[0]) generated_text = outputs[0].outputs[0].text if outputs[0].outputs else "" stop_reason = outputs[0].outputs[0].stop_reason if outputs[0].outputs else "unknown" print(f"Generated text: {repr(generated_text)}") print(f"Token count: {token_count}") print(f"Stop reason: {stop_reason}") # With the bug, this will fail because ANY of the common characters # will trigger early termination before min_tokens=50 is reached # It's virtually impossible to generate 50 tokens without hitting # at least one of: e, a, i, o, u, space, t, n, s, r finish_reason = ( outputs[0].outputs[0].finish_reason if outputs[0].outputs else "unknown" ) print(f"Finish reason: {finish_reason}") if finish_reason == "stop": assert token_count >= 50, ( "Bug confirmed: " f"{token_count} tokens < min_tokens=50. " f"Reason: {finish_reason}. " f"Text: {repr(generated_text)}" ) @pytest.mark.xfail( reason=("Potential logits-processor bug: EOS tokens may bypass min_tokens"), strict=False, ) def test_min_tokens_eos_behavior(llm_v1: LLM): """ Verify EOS handling with and without min_tokens. - Without min_tokens: expect early EOS -> finish_reason == "stop", stop_reason is None, and generated tokens < max_tokens (25). - With min_tokens: EOS should be blocked until min_tokens is reached (finish_reason == "length"); verify that eos_token_id does not appear in generated token_ids. """ # tokenizer + eos id tokenizer = llm_v1.get_tokenizer() eos_token_id = tokenizer.eos_token_id prompt = "Give a file extension." max_toks = 32 # Case 1: WITHOUT min_tokens sp_no_min = SamplingParams( max_tokens=max_toks, temperature=GREEDY, ) out_no_min = llm_v1.generate([prompt], sp_no_min) assert len(out_no_min) == 1 choice_no_min = out_no_min[0].outputs[0] ids_no_min = choice_no_min.token_ids or [] finish_no_min = choice_no_min.finish_reason stop_no_min = choice_no_min.stop_reason print( "[no-min] tokens=", len(ids_no_min), " finish=", finish_no_min, " stop_reason=", stop_no_min, ) assert finish_no_min == "stop", ( f"Expected finish_reason 'stop' without min_tokens, got {finish_no_min}" ) assert stop_no_min is None, ( "For EOS-based stop (no user stop strings), stop_reason should be None." ) assert len(ids_no_min) < max_toks, ( f"Expected early EOS with < {max_toks} tokens, got {len(ids_no_min)}" ) # Case 2: WITH min_tokens sp_with_min = SamplingParams( min_tokens=max_toks, max_tokens=max_toks, temperature=GREEDY, ) out_with_min = llm_v1.generate([prompt], sp_with_min) assert len(out_with_min) == 1 choice_with_min = out_with_min[0].outputs[0] ids_with_min = choice_with_min.token_ids or [] finish_with_min = choice_with_min.finish_reason stop_with_min = choice_with_min.stop_reason print( "[with-min] tokens=", len(ids_with_min), " finish=", finish_with_min, " stop_reason=", stop_with_min, ) # Exact length reached; EOS should have been blocked assert len(ids_with_min) == max_toks, ( f"Expected exactly {max_toks} tokens with min_tokens; got {len(ids_with_min)}" ) assert finish_with_min == "length", ( f"Expected finish_reason 'length'; got {finish_with_min}" ) assert eos_token_id not in ids_with_min, ( "EOS token id should not appear when min_tokens prevents early EOS." ) def test_min_tokens_validation(): """ Test that SamplingParams correctly validates min_tokens parameters. This tests the parameter validation logic in SamplingParams. """ # Valid cases SamplingParams(min_tokens=0, max_tokens=10) SamplingParams(min_tokens=5, max_tokens=10) SamplingParams(min_tokens=10, max_tokens=10) # Invalid cases with pytest.raises( ValueError, match="min_tokens must be greater than or equal to 0", ): SamplingParams(min_tokens=-1, max_tokens=10) with pytest.raises( ValueError, match="min_tokens must be less than or equal to max_tokens", ): SamplingParams(min_tokens=15, max_tokens=10) if __name__ == "__main__": """ Run tests locally for development. Usage: cd vllm/ python -m pytest tests/v1/e2e/test_min_tokens.py -v """ pytest.main([__file__, "-v"])
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/e2e/test_min_tokens.py", "license": "Apache License 2.0", "lines": 428, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/executor/test_executor.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import asyncio import os from collections.abc import Callable from concurrent.futures import Future from typing import Any import pytest from vllm.distributed.kv_transfer.kv_connector.utils import KVOutputAggregator from vllm.engine.arg_utils import AsyncEngineArgs, EngineArgs from vllm.sampling_params import SamplingParams from vllm.v1.engine.async_llm import AsyncLLM from vllm.v1.engine.llm_engine import LLMEngine from vllm.v1.executor.multiproc_executor import MultiprocExecutor class Mock: ... class CustomMultiprocExecutor(MultiprocExecutor): def collective_rpc( self, method: str | Callable, timeout: float | None = None, args: tuple = (), kwargs: dict | None = None, non_block: bool = False, unique_reply_rank: int | None = None, kv_output_aggregator: KVOutputAggregator = None, ) -> Any | list[Any] | Future[Any | list[Any]]: # Drop marker to show that this was run with open(".marker", "w"): ... return super().collective_rpc( method, timeout, args, kwargs, non_block, unique_reply_rank, kv_output_aggregator, ) CustomMultiprocExecutorAsync = CustomMultiprocExecutor MODEL = "Qwen/Qwen3-0.6B" def test_custom_executor_type_checking(): with pytest.raises(ValueError): engine_args = EngineArgs( model=MODEL, gpu_memory_utilization=0.2, max_model_len=8192, distributed_executor_backend=Mock, ) LLMEngine.from_engine_args(engine_args) with pytest.raises(ValueError): engine_args = AsyncEngineArgs( model=MODEL, gpu_memory_utilization=0.2, max_model_len=8192, distributed_executor_backend=Mock, ) AsyncLLM.from_engine_args(engine_args) @pytest.mark.parametrize( "distributed_executor_backend", [ CustomMultiprocExecutor, "tests.v1.executor.test_executor.CustomMultiprocExecutor", ], ) def test_custom_executor(distributed_executor_backend, tmp_path): cwd = os.path.abspath(".") os.chdir(tmp_path) try: assert not os.path.exists(".marker") engine_args = EngineArgs( model=MODEL, gpu_memory_utilization=0.2, max_model_len=8192, distributed_executor_backend=distributed_executor_backend, enforce_eager=True, # reduce test time ) engine = LLMEngine.from_engine_args(engine_args) sampling_params = SamplingParams(max_tokens=1) engine.add_request("0", "foo", sampling_params) engine.step() assert os.path.exists(".marker") finally: os.chdir(cwd) @pytest.mark.parametrize( "distributed_executor_backend", [ CustomMultiprocExecutorAsync, "tests.v1.executor.test_executor.CustomMultiprocExecutorAsync", ], ) def test_custom_executor_async(distributed_executor_backend, tmp_path): cwd = os.path.abspath(".") os.chdir(tmp_path) try: assert not os.path.exists(".marker") engine_args = AsyncEngineArgs( model=MODEL, gpu_memory_utilization=0.2, max_model_len=8192, distributed_executor_backend=distributed_executor_backend, enforce_eager=True, # reduce test time ) engine = AsyncLLM.from_engine_args(engine_args) sampling_params = SamplingParams(max_tokens=1) async def t(): stream = engine.generate( request_id="0", prompt="foo", sampling_params=sampling_params ) async for x in stream: ... asyncio.run(t()) assert os.path.exists(".marker") finally: os.chdir(cwd)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/executor/test_executor.py", "license": "Apache License 2.0", "lines": 114, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:benchmarks/kernels/benchmark_silu_mul_fp8_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Comprehensive 3-way SiLU Benchmark Suite This benchmark compares three SiLU implementations: 1. SiLU V2 (CUDA) - Optimized CUDA kernel implementation 2. Triton Kernel - Triton-based implementation The suite generates detailed performance comparisons including: - Memory bandwidth utilization - Speedup ratios (baseline vs optimized implementations) - Performance across different expert configurations and token distributions """ from collections.abc import Callable import matplotlib.pyplot as plt import numpy as np import torch from vllm.model_executor.layers.fused_moe.batched_deep_gemm_moe import ( persistent_masked_m_silu_mul_quant, ) from vllm.triton_utils import tl, triton from vllm.utils.deep_gemm import is_deep_gemm_e8m0_used from vllm.utils.torch_utils import set_random_seed @triton.jit def _silu_mul_fp8_quant_deep_gemm( # Pointers ------------------------------------------------------------ input_ptr, # 16-bit activations (E, T, 2*H) y_q_ptr, # fp8 quantized activations (E, T, H) y_s_ptr, # 16-bit scales (E, T, G) counts_ptr, # int32 num tokens per expert (E) # Sizes --------------------------------------------------------------- H: tl.constexpr, # hidden dimension (per output) GROUP_SIZE: tl.constexpr, # elements per group (usually 128) # Strides for input (elements) --------------------------------------- stride_i_e, stride_i_t, stride_i_h, # Strides for y_q (elements) ----------------------------------------- stride_yq_e, stride_yq_t, stride_yq_h, # Strides for y_s (elements) ----------------------------------------- stride_ys_e, stride_ys_t, stride_ys_g, # Stride for counts (elements) stride_counts_e, # Numeric params ------------------------------------------------------ eps: tl.constexpr, fp8_min: tl.constexpr, fp8_max: tl.constexpr, use_ue8m0: tl.constexpr, # Meta --------------------------------------------------------------- BLOCK: tl.constexpr, NUM_STAGES: tl.constexpr, ): G = H // GROUP_SIZE # map program id -> (e, g) pid = tl.program_id(0) e = pid // G g = pid % G e = e.to(tl.int64) g = g.to(tl.int64) # number of valid tokens for this expert n_tokens = tl.load(counts_ptr + e * stride_counts_e).to(tl.int64) cols = tl.arange(0, BLOCK).to(tl.int64) mask = cols < BLOCK base_input_offset = e * stride_i_e + g * GROUP_SIZE * stride_i_h base_gate_offset = base_input_offset + cols * stride_i_h base_up_offset = base_input_offset + H * stride_i_h + cols * stride_i_h base_yq_offset = e * stride_yq_e + g * GROUP_SIZE * stride_yq_h + cols * stride_yq_h base_ys_offset = e * stride_ys_e + g * stride_ys_g for t in tl.range(0, n_tokens, num_stages=NUM_STAGES): gate = tl.load( input_ptr + base_gate_offset + t * stride_i_t, mask=mask, other=0.0 ).to(tl.float32) up = tl.load(input_ptr + base_up_offset + t * stride_i_t, mask=mask, other=0.0) gate = gate * (1.0 / (1.0 + tl.exp(-gate))) y = gate * up y_s = tl.maximum(tl.max(tl.abs(y)), eps) / fp8_max if use_ue8m0: y_s = tl.exp2(tl.ceil(tl.log2(y_s))) y_q = tl.clamp(y / y_s, fp8_min, fp8_max).to(y_q_ptr.dtype.element_ty) tl.store(y_q_ptr + base_yq_offset + t * stride_yq_t, y_q, mask=mask) tl.store(y_s_ptr + base_ys_offset + t * stride_ys_t, y_s) def silu_mul_fp8_quant_deep_gemm_triton( y: torch.Tensor, # (E, T, 2*H) tokens_per_expert: torch.Tensor, # (E,) number of valid tokens per expert num_parallel_tokens, group_size: int = 128, eps: float = 1e-10, expert_offsets: torch.Tensor = None, ) -> tuple[torch.Tensor, torch.Tensor]: """Quantize silu(y[..., :H]) * y[..., H:] to FP8 with group per-token scales y has shape (E, T, 2*H). The first half of the last dimension is silu-activated, multiplied by the second half, then quantized into FP8. Returns `(y_q, y_s)` where * `y_q`: FP8 tensor, shape (E, T, H), same layout as y[..., :H] * `y_s`: FP32 tensor, shape (E, T, H // group_size), strides (T*G, 1, T) """ assert y.ndim == 3, "y must be (E, T, 2*H)" E, T, H2 = y.shape assert H2 % 2 == 0, "last dim of y must be even (2*H)" H = H2 // 2 G = (H + group_size - 1) // group_size assert H % group_size == 0, "H must be divisible by group_size" assert tokens_per_expert.ndim == 1 and tokens_per_expert.shape[0] == E, ( "tokens_per_expert must be shape (E,)" ) tokens_per_expert = tokens_per_expert.to(device=y.device, dtype=torch.int32) # allocate outputs fp8_dtype = torch.float8_e4m3fn y_q = torch.empty((E, T, H), dtype=fp8_dtype, device=y.device) # strides (elements) stride_i_e, stride_i_t, stride_i_h = y.stride() stride_yq_e, stride_yq_t, stride_yq_h = y_q.stride() # desired scale strides (elements): (T*G, 1, T) stride_ys_e = T * G stride_ys_t = 1 stride_ys_g = T y_s = torch.empty_strided( (E, T, G), (stride_ys_e, stride_ys_t, stride_ys_g), dtype=torch.float32, device=y.device, ) stride_cnt_e = tokens_per_expert.stride()[0] # Static grid over experts and H-groups. # A loop inside the kernel handles the token dim grid = (E * G,) f_info = torch.finfo(fp8_dtype) fp8_max = f_info.max fp8_min = f_info.min _silu_mul_fp8_quant_deep_gemm[grid]( y, y_q, y_s, tokens_per_expert, H, group_size, stride_i_e, stride_i_t, stride_i_h, stride_yq_e, stride_yq_t, stride_yq_h, stride_ys_e, stride_ys_t, stride_ys_g, stride_cnt_e, eps, fp8_min, fp8_max, is_deep_gemm_e8m0_used(), BLOCK=group_size, NUM_STAGES=4, num_warps=1, ) return y_q, y_s # Parse generation strategies strategies = ["random_imbalanced", "uniform", "max_t"] def benchmark( kernel: Callable, E: int, T: int, H: int, total_tokens: int, num_parallel_tokens: int = 64, G: int = 128, runs: int = 200, num_warmups: int = 20, gen_strategy: str = "default", iterations_per_run: int = 20, ): def generate_data(seed_offset=0): """Generate input data with given seed offset""" set_random_seed(42 + seed_offset) y = torch.rand((E, T, 2 * H), dtype=torch.bfloat16, device="cuda").contiguous() if gen_strategy == "random_imbalanced": def generate_expert_loads(n_e, total_tokens, ratio, device="cuda"): mean = total_tokens // n_e min_max = mean // ratio e = torch.ones(size=(E,), dtype=torch.int64, device=device) * mean e[0] = min_max r = torch.rand(size=(E - 1,)) r /= r.sum() r *= total_tokens - min_max r = r.round().long() e[1:] = r.to(device=device) return e tokens_per_expert = generate_expert_loads(E, total_tokens, 0.7, "cuda") elif gen_strategy == "uniform": r = torch.rand(size=(E,)) r /= r.sum() r *= total_tokens r = r.round().long() tokens_per_expert = r elif gen_strategy == "max_t": tokens_per_expert = torch.empty(size=(E,), dtype=torch.int32, device="cuda") tokens_per_expert.fill_(total_tokens / E) elif gen_strategy == "first_t": tokens_per_expert = torch.zeros(size=(E,), dtype=torch.int32, device="cuda") tokens_per_expert[0] = min(T, total_tokens) else: raise ValueError(f"Unknown generation strategy: {gen_strategy}") return y, tokens_per_expert dataset_count = 4 # Pre-generate different input matrices for each iteration to avoid cache effects data_sets = [generate_data(i) for i in range(dataset_count)] # Warmup y, tokens_per_expert = data_sets[0] for _ in range(num_warmups): kernel( y, tokens_per_expert, num_parallel_tokens=num_parallel_tokens, group_size=G ) torch.cuda.synchronize() start_event = torch.Event(enable_timing=True) end_event = torch.Event(enable_timing=True) # Benchmark latencies: list[float] = [] for _ in range(runs): torch.cuda.synchronize() start_event.record() for i in range(iterations_per_run): y, tokens_per_expert = data_sets[i % dataset_count] kernel( y, tokens_per_expert, num_parallel_tokens=num_parallel_tokens, group_size=G, ) end_event.record() end_event.synchronize() total_time_ms = start_event.elapsed_time(end_event) per_iter_time_ms = total_time_ms / iterations_per_run latencies.append(per_iter_time_ms) # Use median instead of average for better outlier handling median_time_ms = np.median(latencies) median_time_s = median_time_ms / 1000 # Calculate actual work done (using first dataset for consistency) _, tokens_per_expert = data_sets[0] actual_tokens = tokens_per_expert.sum().item() actual_elements = actual_tokens * H # GFLOPS: operations per element = exp + 3 muls + 1 div + quantization ops ≈ 8 ops ops_per_element = 8 total_ops = actual_elements * ops_per_element gflops = total_ops / median_time_s / 1e9 # Memory bandwidth: bfloat16 inputs (2 bytes), fp8 output (1 byte), scales (4 bytes) input_bytes = actual_tokens * 2 * H * 2 # 2*H bfloat16 inputs output_bytes = actual_tokens * H * 1 # H fp8 outputs scale_bytes = actual_tokens * (H // G) * 4 # scales in float32 total_bytes = input_bytes + output_bytes + scale_bytes memory_bw = total_bytes / median_time_s / 1e9 HOPPER_BANDWIDTH_TBPS = 3.35 return ( median_time_ms, gflops, memory_bw, (memory_bw / (HOPPER_BANDWIDTH_TBPS * 1024)) * 100, ) def create_comparison_plot( ratios, silu_v2_times, triton_times, config_labels, strategy_name, id ): fig, ax = plt.subplots(1, 1, figsize=(18, 6)) # Configure x-axis positions x = np.arange(len(config_labels)) width = 0.25 # Execution Time plot (lower is better) ax.bar(x, silu_v2_times, width, label="SiLU V2 (CUDA)", alpha=0.8, color="blue") ax.bar( x + width, triton_times, width, label="Triton Kernel", alpha=0.8, color="green" ) # Add speedup labels over each bar trio for i in range(len(x)): triton_v2_speedup = ratios[i][1] # triton/v2 max_height = max(silu_v2_times[i], triton_times[i]) # Triton/V2 speedup ax.text( x[i] + width / 2, max_height + max_height * 0.02, f"{triton_v2_speedup:.2f}x", ha="center", va="bottom", fontweight="bold", fontsize=8, ) ax.set_xlabel("Configuration") ax.set_ylabel("% Utilization") ax.set_title( f"Memory Bandwidth Utilization (%) - {strategy_name}\n(Higher is Better)" ) ax.set_xticks(x) ax.set_xticklabels(config_labels, rotation=45, ha="right") ax.legend() ax.grid(True, alpha=0.3) plt.tight_layout() return fig, ax def create_combined_plot(all_results): num_strategies = len(all_results) fig, axes = plt.subplots(num_strategies, 1, figsize=(22, 7 * num_strategies)) if num_strategies == 1: axes = [axes] for idx, ( strategy_name, all_ratios, all_silu_v2_results, all_triton_results, config_labels, config_x_axis, ) in enumerate(all_results): ax = axes[idx] # Flatten the nested results to get bandwidth percentages for plotting silu_v2_bandwidths = [] triton_bandwidths = [] flat_ratios = [] for config_results in all_silu_v2_results: for result in config_results: silu_v2_bandwidths.append(result[3]) # bandwidth percentage for config_results in all_triton_results: for result in config_results: triton_bandwidths.append(result[3]) # bandwidth percentage for config_ratios in all_ratios: for ratio in config_ratios: flat_ratios.append(ratio) # Configure x-axis positions x = np.arange(len(config_labels)) width = 0.25 # Bandwidth utilization plot (higher is better) ax.bar( x, silu_v2_bandwidths, width, label="SiLU V2 (CUDA)", alpha=0.8, color="blue", ) ax.bar( x + width, triton_bandwidths, width, label="Triton Kernel", alpha=0.8, color="green", ) # Add speedup labels over each bar trio for i in range(len(x)): triton_v2_speedup = flat_ratios[i] # triton/v2 max_height = max(silu_v2_bandwidths[i], triton_bandwidths[i]) # Triton/V2 speedup ax.text( x[i] + width / 2, max_height + max_height * 0.02, f"{triton_v2_speedup:.2f}x", ha="center", va="bottom", fontweight="bold", fontsize=8, ) ax.set_xlabel("Configuration") ax.set_ylabel("% Utilization") ax.set_title( f"Memory Bandwidth Utilization (%) - {strategy_name}\n(Higher is Better)" ) ax.set_xticks(x) ax.set_xticklabels(config_labels, rotation=45, ha="right") ax.legend() ax.grid(True, alpha=0.3) plt.tight_layout() filename = "silu_benchmark_combined_3way.png" plt.savefig(filename, dpi=300, bbox_inches="tight") plt.show() return filename outer_dim = 7168 configs = [ # DeepSeekV3 Configs # (1, 56, 7168), (8, 1024, 7168), # (32, 56, 7168), # DeepSeekV3 Configs (32, 1024, 7168), # DeepSeekV3 Configs (256, 1024, 7168), ] runs = 100 num_warmups = 20 strategy_descriptions = { "uniform": "Uniform Random", "random_imbalanced": "Imbalanced Random", "max_t": "Even Assignment", "first_t": "experts[0] = T, experts[1:] = 0", } print(f"GPU: {torch.cuda.get_device_name()}") print(f"Testing strategies: {', '.join(strategies)}") print(f"Configurations: {len(configs)} configs") all_results = [] # Run benchmarks for each strategy for id, strategy in enumerate(strategies): print(f"\n{'=' * 60}") print(f"Testing strategy: {strategy_descriptions[strategy]}") print(f"{'=' * 60}") # Collect benchmark data for all three algorithms config_labels = [] config_x_axis = [] all_silu_v2_results = [] all_triton_results = [] all_ratios = [] for E, T, H in configs: total_tokens_config = [] for i in [8, 16, 32, 64, 128, 256, 512]: if i <= T: total_tokens_config.append(i * E) config_x_axis.append(total_tokens_config) silu_v2_results = [] triton_results = [] ratios = [] for total_tokens in total_tokens_config: config_label = f"E={E},T={T},H={H},TT={total_tokens}" config_labels.append(config_label) # SiLU V2 (CUDA kernel) results time_ms_silu_v2, gflops, gbps, perc = benchmark( persistent_masked_m_silu_mul_quant, E, T, H, total_tokens, runs=runs, num_warmups=num_warmups, gen_strategy=strategy, ) silu_v2_results.append((time_ms_silu_v2, gflops, gbps, perc)) # Triton kernel results time_ms_triton, gflops, gbps, perc = benchmark( silu_mul_fp8_quant_deep_gemm_triton, E, T, H, total_tokens, runs=runs, num_warmups=num_warmups, gen_strategy=strategy, ) triton_results.append((time_ms_triton, gflops, gbps, perc)) # Calculate speedup ratios (triton baseline / implementation) triton_v2_ratio = time_ms_triton / time_ms_silu_v2 ratios.append(triton_v2_ratio) print( f"Completed: {config_label}:" f" V2: {time_ms_silu_v2:.3f}ms," f" Triton: {time_ms_triton:.3f}ms" ) all_silu_v2_results.append(silu_v2_results) all_triton_results.append(triton_results) all_ratios.append(ratios) # Store results for combined plotting all_results.append( ( strategy_descriptions[strategy], all_ratios, all_silu_v2_results, all_triton_results, config_labels, config_x_axis, ) ) # Print summary table for this strategy print(f"\nSummary Table - {strategy_descriptions[strategy]}:") print(f" {'V2 Time(ms)':<12} {'Triton Time(ms)':<14} {'Triton/V2':<10}") print("-" * 90) for i, (E, T, H) in enumerate(configs): # Get the first result for each config (simplifying for summary) v2_time = silu_v2_results[i][0] triton_time = triton_results[i][0] triton_v2_speedup = triton_time / v2_time config_label = f"E={E:3d},T={T:4d},H={H:4d}" print( f"{config_label:<20} {v2_time:8.5f} {triton_time:10.5f} " f"{triton_v2_speedup:8.2f}x" ) def create_total_tokens_plot(all_results): num_strategies = len(all_results) num_configs = len(configs) fig, axs = plt.subplots( num_strategies, num_configs * 2, figsize=(32, 8 * num_strategies) ) # Add main title to the entire figure fig.suptitle( "Performance Analysis: Speedup vs Bandwidth Utilization (SiLU V2, and Triton)", fontsize=18, fontweight="bold", y=0.98, ) # Handle single strategy case if num_strategies == 1: axs = axs.reshape(1, -1) # Handle single config case if num_configs == 1: axs = axs.reshape(-1, 2) for strategy_idx, result in enumerate(all_results): ( strategy_name, all_ratios, all_silu_v2_results, all_triton_results, config_labels, config_x_axis, ) = result for config_idx in range(num_configs): # Speedup plot (left column) ax_speedup = axs[strategy_idx, config_idx * 2] # Bandwidth plot (right column) ax_bandwidth = axs[strategy_idx, config_idx * 2 + 1] E, T, H = configs[config_idx] ratios = all_ratios[config_idx] total_tokens_values = config_x_axis[config_idx] # Extract speedup ratios triton_v2_ratios = [ratio for ratio in ratios] # Extract bandwidth percentages for all implementations v2_bandwidth_percentages = [ result[3] for result in all_silu_v2_results[config_idx] ] triton_bandwidth_percentages = [ result[3] for result in all_triton_results[config_idx] ] # Plot speedup ratios vs total tokens (left plot) ax_speedup.plot( total_tokens_values, triton_v2_ratios, "go-", linewidth=3, markersize=8, label="Triton/V2 Speedup", ) ax_speedup.set_title( f"{strategy_name}\nSpeedup vs Baseline (Triton)\nE={E}, T={T}, H={H}", fontsize=12, fontweight="bold", ) ax_speedup.set_xlabel("Total Tokens", fontweight="bold", fontsize=11) ax_speedup.set_ylabel("Speedup Ratio", fontweight="bold", fontsize=11) ax_speedup.legend(prop={"weight": "bold"}) ax_speedup.grid(True, alpha=0.3) # Plot bandwidth utilization (right plot) ax_bandwidth.plot( total_tokens_values, v2_bandwidth_percentages, "o-", linewidth=3, markersize=8, label="SiLU V2", color="blue", ) ax_bandwidth.plot( total_tokens_values, triton_bandwidth_percentages, "o-", linewidth=3, markersize=8, label="Triton", color="green", ) ax_bandwidth.set_title( f"{strategy_name}\nBandwidth Utilization (Hopper)\nE={E}, T={T}, H={H}", fontsize=12, fontweight="bold", ) ax_bandwidth.set_xlabel("Total Tokens", fontweight="bold", fontsize=11) ax_bandwidth.set_ylabel( "% of Peak Bandwidth", fontweight="bold", fontsize=11 ) ax_bandwidth.legend(prop={"weight": "bold"}) ax_bandwidth.grid(True, alpha=0.3) # Format x-axis labels for both plots for ax in [ax_speedup, ax_bandwidth]: ax.set_xticks(total_tokens_values) ax.set_xticklabels( [ f"{tt // 1000}K" if tt >= 1000 else str(tt) for tt in total_tokens_values ], fontweight="bold", ) # Make tick labels bold for label in ax.get_xticklabels() + ax.get_yticklabels(): label.set_fontweight("bold") # Add value labels on Triton/V2 speedup points for x, y in zip(total_tokens_values, triton_v2_ratios): ax_speedup.annotate( f"{y:.2f}x", (x, y), textcoords="offset points", xytext=(0, -15), ha="center", fontsize=9, fontweight="bold", bbox=dict(boxstyle="round,pad=0.2", facecolor="green", alpha=0.3), ) plt.tight_layout() plt.subplots_adjust(top=0.93) # Make room for main title filename = "silu_benchmark_total_tokens_3way.png" plt.savefig(filename, dpi=300, bbox_inches="tight") plt.show() return filename # Create comprehensive 3-way comparison plots combined_plot_filename = create_combined_plot(all_results) total_tokens_plot_filename = create_total_tokens_plot(all_results) print(f"\n{'=' * 80}") print("3-Way Benchmark Suite Complete!") print(f"Generated combined comparison plot: {combined_plot_filename}") print(f"Generated total tokens analysis plot: {total_tokens_plot_filename}") print("Compared: SiLU V2 (CUDA), and Triton implementations") print(f"{'=' * 80}")
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/kernels/benchmark_silu_mul_fp8_quant.py", "license": "Apache License 2.0", "lines": 610, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/layers/mamba/short_conv.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.config import CacheConfig, ModelConfig, get_current_vllm_config from vllm.distributed import get_tensor_model_parallel_world_size from vllm.forward_context import ForwardContext, get_forward_context from vllm.model_executor.custom_op import CustomOp from vllm.model_executor.layers.linear import ( ColumnParallelLinear, MergedColumnParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.mamba.abstract import MambaBase from vllm.model_executor.layers.mamba.mamba_utils import ( MambaStateDtypeCalculator, MambaStateShapeCalculator, ) from vllm.model_executor.layers.mamba.ops.causal_conv1d import ( causal_conv1d_fn, causal_conv1d_update, ) from vllm.utils.torch_utils import direct_register_custom_op from vllm.v1.attention.backend import AttentionMetadata from vllm.v1.attention.backends.short_conv_attn import ShortConvAttentionMetadata # --8<-- [start:short_conv] @CustomOp.register("short_conv") class ShortConv(MambaBase, CustomOp): # --8<-- [end:short_conv] def __init__( self, config, dim: int, layer_idx: int, model_config: ModelConfig | None = None, cache_config: CacheConfig | None = None, prefix: str = "", ): super().__init__() self.config = config self.layer_idx = layer_idx self.conv_dim = dim self.L_cache = config.conv_L_cache self.bias = config.conv_bias self.conv = ColumnParallelLinear( input_size=self.L_cache, output_size=dim, bias=self.bias, prefix=f"{prefix}.conv1d", ) # unsqueeze to fit conv1d weights shape into the linear weights shape. # Can't do this in `weight_loader` since it already exists in # `ColumnParallelLinear` and `set_weight_attrs` # doesn't allow to override it self.conv.weight.data = self.conv.weight.data.unsqueeze(1) self.in_proj = MergedColumnParallelLinear( input_size=dim, output_sizes=[dim] * 3, bias=self.bias, prefix=f"{prefix}.in_proj", ) self.out_proj = RowParallelLinear( input_size=dim, output_size=dim, bias=self.bias, prefix=f"{prefix}.out_proj", ) compilation_config = get_current_vllm_config().compilation_config if prefix in compilation_config.static_forward_context: raise ValueError(f"Duplicate layer name: {prefix}") compilation_config.static_forward_context[prefix] = self self.kv_cache = (torch.tensor([]),) self.model_config = model_config self.cache_config = cache_config self.prefix = prefix def forward_native( self, hidden_states: torch.Tensor, output: torch.Tensor, ): return def forward( self, hidden_states: torch.Tensor, output: torch.Tensor, ): torch.ops.vllm.short_conv( hidden_states, output, self.prefix, ) def forward_cuda( self, hidden_states: torch.Tensor, output: torch.Tensor, ): forward_context = get_forward_context() # ShortConvAttentionMetadata contains metadata necessary for the # short_conv triton kernels to operate in continuous batching and in # chunked prefill modes; they are computed at top-level model forward # since they stay the same and reused for all mamba layers in the same # iteration. attn_metadata: AttentionMetadata = forward_context.attn_metadata if attn_metadata is not None: assert isinstance(attn_metadata, dict) attn_metadata = attn_metadata[self.prefix] assert isinstance(attn_metadata, ShortConvAttentionMetadata) self_kv_cache = self.kv_cache[forward_context.virtual_engine] conv_state = self_kv_cache[0].transpose(-1, -2) state_indices_tensor_p = attn_metadata.state_indices_tensor_p state_indices_tensor_d = attn_metadata.state_indices_tensor_d has_initial_states_p = attn_metadata.has_initial_states_p query_start_loc_p = attn_metadata.query_start_loc_p BCx, _ = self.in_proj(hidden_states) B, C, x = BCx.chunk(3, dim=-1) conv_weights = self.conv.weight.view( self.conv.weight.size(0), self.conv.weight.size(2) ) if attn_metadata is None: # V1 profile run Bx = (B * x).contiguous() hidden_states = C * Bx contextualized_states, _ = self.out_proj(hidden_states) return contextualized_states num_prefills = attn_metadata.num_prefills # request count num_decodes = attn_metadata.num_decode_tokens # token count (=request) num_prefill_tokens = attn_metadata.num_prefill_tokens # token count has_prefill = num_prefills > 0 has_decode = num_decodes > 0 num_actual_tokens = num_decodes + num_prefill_tokens # NOTE: V1 puts decode before prefill # Separate prefill and decode by splitting varlen input # Split along token dimension B_d, B_p = torch.split( B[:num_actual_tokens], [num_decodes, num_prefill_tokens], dim=0, ) C_d, C_p = torch.split( C[:num_actual_tokens], [num_decodes, num_prefill_tokens], dim=0, ) x_d, x_p = torch.split( x[:num_actual_tokens], [num_decodes, num_prefill_tokens], dim=0, ) conv_output_list = [] if has_prefill: Bx_p = (B_p * x_p).transpose(0, 1) Bx = causal_conv1d_fn( Bx_p, conv_weights, self.conv.bias, activation=None, conv_states=conv_state, has_initial_state=has_initial_states_p, cache_indices=state_indices_tensor_p, metadata=attn_metadata, query_start_loc=query_start_loc_p, ).transpose(0, 1)[:num_prefill_tokens] y = C_p * Bx conv_output_list.append(y) if has_decode: Bx_d = (B_d * x_d).contiguous() Bx = causal_conv1d_update( Bx_d, conv_state, conv_weights, self.conv.bias, activation=None, conv_state_indices=state_indices_tensor_d, ) y = C_d * Bx conv_output_list.insert(0, y) # Merge prefill and decode outputs before passing to gated MLP hidden_states = torch.vstack(conv_output_list) # Final linear projection output[:num_actual_tokens], _ = self.out_proj(hidden_states) def get_state_dtype(self) -> tuple[torch.dtype, ...]: assert self.model_config is not None assert self.cache_config is not None return MambaStateDtypeCalculator.short_conv_state_dtype( self.model_config.dtype, self.cache_config.mamba_cache_dtype, ) def get_state_shape(self) -> tuple[tuple[int, ...]]: return MambaStateShapeCalculator.short_conv_state_shape( tp_world_size=get_tensor_model_parallel_world_size(), intermediate_size=self.conv_dim, conv_kernel=self.L_cache, ) @property def mamba_type(self) -> str: return "short_conv" def short_conv( hidden_states: torch.Tensor, output: torch.Tensor, layer_name: str, ) -> None: forward_context: ForwardContext = get_forward_context() self = forward_context.no_compile_layers[layer_name] self.forward_cuda(hidden_states=hidden_states, output=output) def short_conv_fake( hidden_states: torch.Tensor, output: torch.Tensor, layer_name: str, ) -> None: return direct_register_custom_op( op_name="short_conv", op_func=short_conv, mutates_args=["output"], fake_impl=short_conv_fake, )
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/layers/mamba/short_conv.py", "license": "Apache License 2.0", "lines": 215, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/lfm2.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Iterable from itertools import islice import torch import torch.nn as nn from transformers import Lfm2Config from vllm.compilation.decorators import support_torch_compile from vllm.config import CacheConfig, ModelConfig, VllmConfig from vllm.distributed import get_pp_group, get_tensor_model_parallel_world_size from vllm.model_executor.layers.activation import SiluAndMul from vllm.model_executor.layers.attention import Attention from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.linear import ( MergedColumnParallelLinear, QKVParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.mamba.mamba_utils import ( MambaStateCopyFunc, MambaStateCopyFuncCalculator, MambaStateDtypeCalculator, MambaStateShapeCalculator, ) from vllm.model_executor.layers.mamba.short_conv import ShortConv from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import default_weight_loader from vllm.sequence import IntermediateTensors from .interfaces import HasInnerState, IsHybrid, SupportsLoRA, SupportsPP, SupportsQuant from .utils import ( AutoWeightsLoader, PPMissingLayer, WeightsMapper, extract_layer_index, is_pp_missing_parameter, make_empty_intermediate_tensors_factory, make_layers, maybe_prefix, ) class Lfm2MLP(nn.Module): def __init__( self, dim: int, ff_dim: int, multiple_of: int, auto_adjust_ff_dim: bool, ffn_dim_multiplier: float | None, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() if auto_adjust_ff_dim: ff_dim = int(2 * ff_dim / 3) # custom dim factor multiplier if ffn_dim_multiplier is not None: ff_dim = int(ffn_dim_multiplier * ff_dim) ff_dim = multiple_of * ((ff_dim + multiple_of - 1) // multiple_of) self.w13 = MergedColumnParallelLinear( input_size=dim, output_sizes=[ff_dim] * 2, bias=False, quant_config=quant_config, prefix=f"{prefix}.w13", ) self.w2 = RowParallelLinear( input_size=ff_dim, output_size=dim, bias=False, quant_config=quant_config, prefix=f"{prefix}.w2", ) self.act_fn = SiluAndMul() def forward(self, x: torch.Tensor) -> torch.Tensor: gate_up, _ = self.w13(x) x = self.act_fn(gate_up) x, _ = self.w2(x) return x class Lfm2Attention(nn.Module): def __init__( self, config: Lfm2Config, layer_idx: int, hidden_size: int, num_heads: int, num_kv_heads: int, max_position_embeddings: int = 8192, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.layer_idx = layer_idx self.hidden_size = hidden_size self.num_kv_heads = num_kv_heads tp_size = get_tensor_model_parallel_world_size() self.total_num_heads = num_heads assert self.total_num_heads % tp_size == 0 self.num_heads = self.total_num_heads // tp_size self.total_num_kv_heads = num_kv_heads if self.total_num_kv_heads >= tp_size: # Number of KV heads is greater than TP size, so we partition # the KV heads across multiple tensor parallel GPUs. assert self.total_num_kv_heads % tp_size == 0 else: # Number of KV heads is less than TP size, so we replicate # the KV heads across multiple tensor parallel GPUs. assert tp_size % self.total_num_kv_heads == 0 self.num_kv_heads = max(1, self.total_num_kv_heads // tp_size) self.head_dim = self.hidden_size // self.total_num_heads self.q_size = self.num_heads * self.head_dim self.kv_size = self.num_kv_heads * self.head_dim self.scaling = self.head_dim**-0.5 self.max_position_embeddings = max_position_embeddings self.qkv_proj = QKVParallelLinear( hidden_size=self.hidden_size, head_size=self.head_dim, total_num_heads=self.total_num_heads, total_num_kv_heads=self.total_num_kv_heads, bias=False, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", ) self.out_proj = RowParallelLinear( input_size=self.total_num_heads * self.head_dim, output_size=self.hidden_size, bias=False, quant_config=quant_config, prefix=f"{prefix}.out_proj", ) self.rotary_emb = get_rope( self.head_dim, max_position=self.max_position_embeddings, rope_parameters=config.rope_parameters, is_neox_style=True, ) self.attn = Attention( self.num_heads, self.head_dim, self.scaling, num_kv_heads=self.num_kv_heads, cache_config=cache_config, prefix=f"{prefix}.attn", ) self.q_layernorm = RMSNorm(self.head_dim, eps=config.norm_eps) self.k_layernorm = RMSNorm(self.head_dim, eps=config.norm_eps) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, ) -> torch.Tensor: n_tokens, _ = hidden_states.shape qkv, _ = self.qkv_proj(hidden_states) q, k, v = qkv.split([self.q_size, self.kv_size, self.kv_size], dim=-1) q = q.view(n_tokens, self.num_heads, self.head_dim).contiguous() k = k.view(n_tokens, self.num_kv_heads, self.head_dim).contiguous() q = self.q_layernorm(q) k = self.k_layernorm(k) q, k = self.rotary_emb(positions, q, k) q = q.view(n_tokens, self.num_heads * self.head_dim) k = k.view(n_tokens, self.num_kv_heads * self.head_dim) attn_output = self.attn(q, k, v) output, _ = self.out_proj(attn_output) return output class Lfm2AttentionDecoderLayer(nn.Module): def __init__( self, config: Lfm2Config, layer_idx: int, model_config: ModelConfig | None = None, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.prefix = prefix self.config = config self.layer_idx = layer_idx max_position_embeddings = getattr(config, "max_position_embeddings", 8192) self.self_attn = Lfm2Attention( config=config, layer_idx=layer_idx, hidden_size=config.hidden_size, num_heads=config.num_attention_heads, num_kv_heads=config.num_key_value_heads, max_position_embeddings=max_position_embeddings, cache_config=cache_config, quant_config=quant_config, prefix=f"{prefix}.self_attn", ) self.feed_forward = Lfm2MLP( dim=config.block_dim, ff_dim=config.block_ff_dim, multiple_of=config.block_multiple_of, auto_adjust_ff_dim=config.block_auto_adjust_ff_dim, ffn_dim_multiplier=config.block_ffn_dim_multiplier, quant_config=quant_config, prefix=f"{prefix}.feed_forward", ) self.operator_norm = RMSNorm(config.hidden_size, eps=config.norm_eps) self.ffn_norm = RMSNorm(config.hidden_size, eps=config.norm_eps) def forward( self, positions: torch.Tensor, hidden_states: torch.Tensor, residual: torch.Tensor | None, **kwargs, ) -> tuple[torch.Tensor, torch.Tensor]: if residual is None: residual = hidden_states hidden_states = self.operator_norm(hidden_states) else: hidden_states, residual = self.operator_norm(hidden_states, residual) hidden_states = self.self_attn(positions=positions, hidden_states=hidden_states) hidden_states, residual = self.ffn_norm(hidden_states, residual) return self.feed_forward(hidden_states), residual class Lfm2ShortConvDecoderLayer(nn.Module): def __init__( self, config: Lfm2Config, layer_idx: int, model_config: ModelConfig | None = None, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", ) -> None: super().__init__() self.layer_idx = layer_idx self.short_conv = ShortConv( config=config, dim=config.conv_dim, layer_idx=layer_idx, model_config=model_config, cache_config=cache_config, prefix=f"{prefix}.conv", ) self.feed_forward = Lfm2MLP( dim=config.block_dim, ff_dim=config.block_ff_dim, multiple_of=config.block_multiple_of, auto_adjust_ff_dim=config.block_auto_adjust_ff_dim, ffn_dim_multiplier=config.block_ffn_dim_multiplier, quant_config=quant_config, prefix=f"{prefix}.feed_forward", ) self.operator_norm = RMSNorm(config.hidden_size, eps=config.norm_eps) self.ffn_norm = RMSNorm(config.hidden_size, eps=config.norm_eps) def forward( self, hidden_states: torch.Tensor, residual: torch.Tensor | None, **kwargs, ): if residual is None: residual = hidden_states hidden_states = self.operator_norm(hidden_states) else: hidden_states, residual = self.operator_norm(hidden_states, residual) output = torch.empty_like(hidden_states) self.short_conv( hidden_states, output, ) hidden_states, residual = self.ffn_norm(output, residual) hidden_states = self.feed_forward(hidden_states) return hidden_states, residual @support_torch_compile class Lfm2Model(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config model_config = vllm_config.model_config cache_config = vllm_config.cache_config quant_config = vllm_config.quant_config self.config = config self.vocab_size = config.vocab_size self.embed_tokens = VocabParallelEmbedding( self.vocab_size, config.hidden_size, org_num_embeddings=config.vocab_size ) def get_layer(prefix: str): layer_idx = extract_layer_index(prefix) is_attn = self.config.layer_types[layer_idx] == "full_attention" layer_class = ( Lfm2AttentionDecoderLayer if is_attn else Lfm2ShortConvDecoderLayer ) return layer_class( config, layer_idx, model_config, cache_config, quant_config=quant_config, prefix=prefix, ) self.start_layer, self.end_layer, self.layers = make_layers( config.num_hidden_layers, get_layer, prefix=f"{prefix}.layers" ) self.make_empty_intermediate_tensors = make_empty_intermediate_tensors_factory( ["hidden_states", "residual"], config.hidden_size ) if get_pp_group().is_last_rank: self.embedding_norm = RMSNorm(config.hidden_size, eps=config.norm_eps) else: self.embedding_norm = PPMissingLayer() def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, ) -> torch.Tensor: if get_pp_group().is_first_rank: if inputs_embeds is not None: hidden_states = inputs_embeds else: hidden_states = self.embed_input_ids(input_ids) residual = None else: assert intermediate_tensors is not None hidden_states = intermediate_tensors["hidden_states"] residual = intermediate_tensors["residual"] for layer in islice(self.layers, self.start_layer, self.end_layer): hidden_states, residual = layer( positions=positions, hidden_states=hidden_states, residual=residual, ) if not get_pp_group().is_last_rank: return IntermediateTensors( {"hidden_states": hidden_states, "residual": residual} ) hidden_states, _ = self.embedding_norm(hidden_states, residual) return hidden_states def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ (".qkv_proj", ".q_proj", "q"), (".qkv_proj", ".k_proj", "k"), (".qkv_proj", ".v_proj", "v"), (".w13", ".w1", 0), (".w13", ".w3", 1), ] params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() for name, loaded_weight in weights: if ".conv." in name: name = name.replace(".conv.", ".short_conv.", 1) for param_name, weight_name, shard_id in stacked_params_mapping: # Use segment-boundary matching (trailing dot) to prevent # e.g. ".w1" from matching inside ".w13" in pre-fused keys. if weight_name + "." not in name: continue name = name.replace(weight_name + ".", param_name + ".") if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params class Lfm2ForCausalLM( nn.Module, HasInnerState, SupportsLoRA, SupportsPP, IsHybrid, SupportsQuant ): packed_modules_mapping = { "qkv_proj": [ "q_proj", "k_proj", "v_proj", ], "w13": [ "w1", "w3", ], "in_proj": ["in_proj"], } # HF uses .conv. but vLLM uses .short_conv. to avoid LoRA regex collision # with the inner .conv.conv child (ShortConv has a child self.conv, so # naming the container .conv too makes _match_target_modules match both) hf_to_vllm_mapper = WeightsMapper( orig_to_new_substr={".conv.": ".short_conv."}, ) # LoRA specific attributes embedding_modules = { "embed_tokens": "input_embeddings", "lm_head": "output_embeddings", } @classmethod def get_mamba_state_dtype_from_config( cls, vllm_config: "VllmConfig", ) -> tuple[torch.dtype, ...]: return MambaStateDtypeCalculator.short_conv_state_dtype( vllm_config.model_config.dtype, vllm_config.cache_config.mamba_cache_dtype, ) @classmethod def get_mamba_state_shape_from_config( cls, vllm_config: "VllmConfig", ) -> tuple[tuple[int, int]]: """Calculate shapes for LFM2's convolutional cache. Args: vllm_config: vLLM config Returns: Tuple containing: - conv_state_shape: Shape for convolutional state cache """ parallel_config = vllm_config.parallel_config hf_config = vllm_config.model_config.hf_config return MambaStateShapeCalculator.short_conv_state_shape( tp_world_size=parallel_config.tensor_parallel_size, intermediate_size=hf_config.conv_dim, conv_kernel=hf_config.conv_L_cache, ) @classmethod def get_mamba_state_copy_func(cls) -> tuple[MambaStateCopyFunc]: return MambaStateCopyFuncCalculator.short_conv_state_copy_func() def __init__(self, *, vllm_config: VllmConfig, prefix: str = "") -> None: config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config cache_config = vllm_config.cache_config if cache_config.mamba_cache_mode == "all": raise NotImplementedError( "Lfm2 currently does not support 'all' prefix caching, " "please use '--mamba-cache-mode=align' instead" ) super().__init__() self.config = config self.model = Lfm2Model( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model") ) if get_pp_group().is_last_rank: self.lm_head = ParallelLMHead( config.vocab_size, config.hidden_size, quant_config=quant_config, prefix=maybe_prefix(prefix, "lm_head"), ) self.lm_head = self.lm_head.tie_weights(self.model.embed_tokens) else: self.lm_head = PPMissingLayer() self.logits_processor = LogitsProcessor(config.vocab_size) self.make_empty_intermediate_tensors = ( self.model.make_empty_intermediate_tensors ) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs, ) -> torch.Tensor: hidden_states = self.model( input_ids, positions, intermediate_tensors, inputs_embeds ) return hidden_states def compute_logits(self, hidden_states: torch.Tensor) -> torch.Tensor: logits = self.logits_processor(self.lm_head, hidden_states) return logits def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader( self, skip_prefixes=(["lm_head."] if self.config.tie_word_embeddings else None), ) return loader.load_weights(weights)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/lfm2.py", "license": "Apache License 2.0", "lines": 480, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/attention/backends/short_conv_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass from vllm.v1.attention.backend import AttentionBackend from vllm.v1.attention.backends.mamba_attn import ( BaseMambaAttentionMetadata, BaseMambaAttentionMetadataBuilder, ) class ShortConvAttentionBackend(AttentionBackend): @staticmethod def get_name() -> str: return "SHORT_CONV_ATTN" @staticmethod def get_builder_cls() -> type["ShortConvAttentionMetadataBuilder"]: return ShortConvAttentionMetadataBuilder @dataclass class ShortConvAttentionMetadata(BaseMambaAttentionMetadata): pass class ShortConvAttentionMetadataBuilder( BaseMambaAttentionMetadataBuilder[ShortConvAttentionMetadata] ): metadata_cls = ShortConvAttentionMetadata
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/attention/backends/short_conv_attn.py", "license": "Apache License 2.0", "lines": 22, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/rvl.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Mapping import torch import torch.nn as nn from transformers.activations import GELUActivation from vllm.config import VllmConfig from vllm.config.multimodal import BaseDummyOptions from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import MultiModalDataDict from .llava_next import ( LlavaDummyInputsBuilder, LlavaNextMultiModalProcessor, LlavaNextProcessingInfo, ) from .llava_onevision import LlavaOnevisionForConditionalGeneration from .utils import WeightsMapper class RVLProcessingInfo(LlavaNextProcessingInfo): def get_hf_config(self): return self.ctx.get_hf_config() def get_hf_processor(self, **kwargs: object): return self.ctx.get_hf_processor(**kwargs) class RVLDummyInputsBuilder(LlavaDummyInputsBuilder[RVLProcessingInfo]): def get_dummy_text(self, mm_counts: Mapping[str, int]) -> str: num_images = mm_counts.get("image", 0) image_token = "<image>" return image_token * num_images def get_dummy_mm_data( self, seq_len: int, mm_counts: Mapping[str, int], mm_options: Mapping[str, BaseDummyOptions], ) -> MultiModalDataDict: num_images = mm_counts.get("image", 0) target_width, target_height = self.info.get_image_size_with_most_features() image_overrides = mm_options.get("image") return { "image": self._get_dummy_images( width=target_width, height=target_height, num_images=num_images, overrides=image_overrides, ), } class RVLMultiModalProjector(nn.Module): def __init__(self, config): super().__init__() self.pre_norm = nn.LayerNorm(config.vision_config.hidden_size, eps=1e-06) self.linear_1 = nn.Linear( config.vision_config.hidden_size, config.text_config.hidden_size, bias=True, ) self.act = GELUActivation() self.linear_2 = nn.Linear( config.text_config.hidden_size, config.text_config.hidden_size, bias=True, ) def forward(self, image_feature: torch.Tensor) -> torch.Tensor: image_feature = self.pre_norm(image_feature) hidden_states = self.linear_1(image_feature) hidden_states = self.act(hidden_states) hidden_states = self.linear_2(hidden_states) return hidden_states @MULTIMODAL_REGISTRY.register_processor( LlavaNextMultiModalProcessor, info=RVLProcessingInfo, dummy_inputs=RVLDummyInputsBuilder, ) class RForConditionalGeneration(LlavaOnevisionForConditionalGeneration): hf_to_vllm_mapper = WeightsMapper( orig_to_new_prefix={ # mapping for new names in checkpoint saved after transformers # v4.52 "model.language_model.": "language_model.model.", "model.vision_tower.": "vision_tower.", "model.multi_modal_projector.": "multi_modal_projector.", "model.image_newline": "image_newline", "lm_head.": "language_model.lm_head.", } ) def __init__(self, *, vllm_config: VllmConfig, prefix: str = "") -> None: super().__init__(vllm_config=vllm_config, prefix=prefix) config = vllm_config.model_config.hf_config self.multi_modal_projector = RVLMultiModalProjector(config)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/rvl.py", "license": "Apache License 2.0", "lines": 86, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/attention/backends/mamba_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import abc from dataclasses import dataclass, replace from typing import Any, ClassVar, TypeVar import torch from vllm.config import VllmConfig from vllm.utils.math_utils import cdiv from vllm.v1.attention.backend import ( AttentionCGSupport, AttentionMetadataBuilder, CommonAttentionMetadata, ) from vllm.v1.attention.backends.utils import ( PAD_SLOT_ID, compute_causal_conv1d_metadata, mamba_get_block_table_tensor, split_decodes_and_prefills, ) from vllm.v1.kv_cache_interface import AttentionSpec, MambaSpec M = TypeVar("M", bound="BaseMambaAttentionMetadata") @dataclass class BaseMambaAttentionMetadata: num_prefills: int num_prefill_tokens: int num_decodes: int num_decode_tokens: int num_reqs: int # The following tensors only contain prefill requests and will be None if # the batch has no prefill requests. has_initial_states_p: torch.Tensor | None query_start_loc_p: torch.Tensor | None num_computed_tokens_p: torch.Tensor | None state_indices_tensor_p: torch.Tensor | None # The following tensors are used for decode requests and # speculative decoding compatibility, and will be None if the batch # has no decode requests. state_indices_tensor_d: torch.Tensor | None query_start_loc_d: torch.Tensor | None # shape: [num_decodes + 1,] # Number of accepted tokens for each spec sequence (for loading correct checkpoint) # Includes the bonus token (so minimum is 1) num_accepted_tokens: torch.Tensor | None # shape: [batch,] # The following tensors are only used for prefix caching in all mode and # are None if disabled block_idx_last_scheduled_token: torch.Tensor | None block_idx_first_scheduled_token_p: torch.Tensor | None block_idx_last_computed_token: torch.Tensor | None # The following tensor is only used for prefix caching in align mode seq_lens: torch.Tensor # cu_chunk_seqlen_p is a tensor of shape (nchunks+1,) that contains, for # each chunk, its offsets into the varlen sequence dimension. It is defined # such that the i-th chunk contains tokens from cu_chunk_seqlen_p[i] to # cu_chunk_seqlen_p[i+1]. cu_chunk_seqlen_p: torch.Tensor | None = None # last_chunk_indices_p is a tensor of shape (batch,) that contains the # index of the last chunk for every sequence in the (prefill) batch. last_chunk_indices_p: torch.Tensor | None = None # The following attributes are for triton implementation of causal_conv1d nums_dict: dict | None = None batch_ptr: torch.Tensor | None = None token_chunk_offset_ptr: torch.Tensor | None = None class BaseMambaAttentionMetadataBuilder(AttentionMetadataBuilder[M], abc.ABC): metadata_cls: type[M] reorder_batch_threshold: int = 1 _cudagraph_support: ClassVar[AttentionCGSupport] = AttentionCGSupport.UNIFORM_BATCH # Will be disabled if speculative decoding is used supports_update_block_table: bool = True def __init__( self, kv_cache_spec: AttentionSpec, layer_names: list[str], vllm_config: VllmConfig, device: torch.device, ): super().__init__(kv_cache_spec, layer_names, vllm_config, device) # Enable speculative decoding support self.speculative_config = vllm_config.speculative_config self.compilation_config = vllm_config.compilation_config self.num_spec_tokens: int = vllm_config.num_speculative_tokens self.use_spec_decode = self.num_spec_tokens > 0 assert isinstance(kv_cache_spec, MambaSpec) self.compilation_config = vllm_config.compilation_config self.decode_cudagraph_max_bs = self.vllm_config.scheduler_config.max_num_seqs if self.compilation_config.max_cudagraph_capture_size is not None: self.decode_cudagraph_max_bs = min( self.decode_cudagraph_max_bs, self.compilation_config.max_cudagraph_capture_size, ) if self.vllm_config.cache_config.mamba_cache_mode == "all": max_num_blocks = cdiv( self.vllm_config.model_config.max_model_len, self.kv_cache_spec.block_size, ) # Speculative decoding not supported with prefix caching, # so keep shape consistent with prefill buffer # TODO: reduce this size as needed for decode-only cudagraph capture self.state_indices_tensor_d = torch.empty( ( self.decode_cudagraph_max_bs, max_num_blocks, ), dtype=torch.int32, device=device, ) self.block_idx_last_scheduled_token = torch.empty( (self.decode_cudagraph_max_bs,), dtype=torch.int32, device=device, ) self.block_idx_last_computed_token = torch.empty( (self.decode_cudagraph_max_bs,), dtype=torch.int32, device=device, ) else: self.state_indices_tensor_d = torch.empty( (self.decode_cudagraph_max_bs, 1 + self.num_spec_tokens), dtype=torch.int32, device=device, ) # For speculative decoding, we need to store the following buffers # for CUDA graph capture during decode if self.num_spec_tokens > 0: self.decode_num_accepted_tokens = torch.empty( (self.decode_cudagraph_max_bs,), dtype=torch.int32, device=device, ) self._init_reorder_batch_threshold(1, self.use_spec_decode) if self.use_spec_decode: self.supports_update_block_table = False def build_for_cudagraph_capture( self, common_attn_metadata: CommonAttentionMetadata ) -> M: """ This method builds the metadata for full cudagraph capture. Currently, only decode is supported for full cudagraphs with Mamba. """ m = common_attn_metadata assert ( m.max_query_len <= 1 + self.num_spec_tokens and m.num_reqs <= self.decode_cudagraph_max_bs ), ( "Mamba only supports decode-only full CUDAGraph capture. " "Make sure all cudagraph capture sizes <= max_num_seq." ) assert m.max_query_len == 1 + self.num_spec_tokens # decode-only num_accepted_tokens = None if self.num_spec_tokens > 0: num_accepted_tokens = torch.diff(m.query_start_loc) return self.build(0, m, num_accepted_tokens=num_accepted_tokens) def build( self, common_prefix_len: int, common_attn_metadata: CommonAttentionMetadata, fast_build: bool = False, *, num_accepted_tokens: torch.Tensor | None = None, **kwargs: Any, ) -> M: """ Default build implementation for Mamba-like attention backends. Subclasses (e.g., Mamba2) can override to add additional metadata. """ return self._compute_common_metadata( common_attn_metadata, num_accepted_tokens=num_accepted_tokens ) def _compute_chunk_metadata( self, chunk_size: int, num_prefills: int, num_computed_tokens_p_cpu: torch.Tensor, query_start_loc_p_cpu: torch.Tensor, ) -> tuple[list[int], list[int], list[int]]: """ Compute chunk-specific metadata for Mamba models. The code below carefully constructs the chunks such that: 1. Chunks contain tokens from a *single* sequence only. 2. For every sequence, we are guaranteed that we can retrieve the mamba state *every* chunk_size tokens. Constraint (1) dramatically simplifies the mamba kernels. Constraint (2) dramatically simplifies the implementation of prefix caching for mamba (wip). We need to take care of the interaction with chunked prefill in order to satisfy constraint (2). """ # TODO (tdoublep): This code could probably be optimized. cu_chunk_seqlen = [] seq_idx = [] last_chunk_indices = [] seqlen_pos = 0 for req_idx in range(num_prefills): this_num_computed = num_computed_tokens_p_cpu[req_idx].item() this_new_tokens = ( query_start_loc_p_cpu[req_idx + 1].item() - query_start_loc_p_cpu[req_idx].item() ) # if computed tokens are not chunk-aligned, use the first # chunk to finish it off if this_num_computed % chunk_size != 0: seq_idx.append(req_idx) cu_chunk_seqlen.append(seqlen_pos) # how many tokens to finish the chunk? chunk_len = ( cdiv(this_num_computed, chunk_size) * chunk_size - this_num_computed ) # we can only use at most this_new_tokens chunk_len = min(chunk_len, this_new_tokens) seqlen_pos += chunk_len this_new_tokens -= chunk_len n_chunks = cdiv(this_new_tokens, chunk_size) for chunk in range(n_chunks): seq_idx.append(req_idx) cu_chunk_seqlen.append(seqlen_pos) chunk_len = min(chunk_size, this_new_tokens) seqlen_pos += chunk_len this_new_tokens -= chunk_len assert this_new_tokens == 0 last_chunk_indices.append(len(cu_chunk_seqlen) - 1) cu_chunk_seqlen.append(seqlen_pos) return cu_chunk_seqlen, seq_idx, last_chunk_indices def _build_chunk_metadata_tensors( self, chunk_size: int, common: M, common_attn_metadata: CommonAttentionMetadata, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Compute chunk metadata and return as device tensors. Returns (cu_chunk_seqlen_p, seq_idx_p, last_chunk_indices_p). """ num_reqs = common.num_reqs num_prefills = common.num_prefills num_decode_tokens = common.num_decode_tokens num_computed_tokens_cpu = ( common_attn_metadata.compute_num_computed_tokens().cpu() ) num_computed_tokens_p_cpu = num_computed_tokens_cpu[ num_reqs - num_prefills : num_reqs ] query_start_loc_p_cpu = ( common_attn_metadata.query_start_loc_cpu[-num_prefills - 1 :] - num_decode_tokens ) cu_chunk_seqlen, seq_idx, last_chunk_indices = self._compute_chunk_metadata( chunk_size, num_prefills, num_computed_tokens_p_cpu, query_start_loc_p_cpu, ) device = common_attn_metadata.query_start_loc.device cu_chunk_seqlen_p = torch.as_tensor( cu_chunk_seqlen, device=device, dtype=torch.int32, ) seq_idx_p = torch.as_tensor( seq_idx, device=device, dtype=torch.int32, ) last_chunk_indices_p = torch.as_tensor( last_chunk_indices, device=device, dtype=torch.int32, ) return cu_chunk_seqlen_p, seq_idx_p, last_chunk_indices_p def _compute_prefix_caching_block_indices( self, common_attn_metadata: CommonAttentionMetadata, mamba_block_size: int, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: num_computed_tokens = common_attn_metadata.compute_num_computed_tokens() # Block index of the last computed token block_idx_last_computed_token = cdiv(num_computed_tokens, mamba_block_size) - 1 # which is <= block index for the first scheduled token block_idx_first_scheduled_token = ( cdiv(num_computed_tokens + 1, mamba_block_size) - 1 ) # which is <= block index of the last scheduled token block_idx_last_scheduled_token = ( cdiv(common_attn_metadata.seq_lens, mamba_block_size) - 1 ) # -1 in case it's non-computed and causes later issues with indexing block_idx_last_computed_token = torch.clamp( block_idx_last_computed_token, min=0 ) # -1 in the case we have a padded request (0 seq-len) block_idx_last_scheduled_token = torch.clamp( block_idx_last_scheduled_token, min=0 ) return ( block_idx_last_computed_token, block_idx_first_scheduled_token, block_idx_last_scheduled_token, ) def _compute_common_metadata( self, common_attn_metadata: CommonAttentionMetadata, *, num_accepted_tokens: torch.Tensor | None = None, ) -> M: """ Compute metadata common to both Mamba1 and Mamba2. """ num_reqs = common_attn_metadata.num_reqs # Treat multi-token queries as decode requests when # speculative decoding is enabled. Otherwise, use the # default decode threshold to prevent misclassification # of prefill queries as decode requests. decode_threshold = ( self.reorder_batch_threshold if num_accepted_tokens is not None else 1 ) num_decodes, num_prefills, num_decode_tokens, num_prefill_tokens = ( split_decodes_and_prefills( common_attn_metadata, decode_threshold=decode_threshold ) ) # Need flags to indicate if there are initial states has_initial_states_p = None query_start_loc_p = None query_start_loc_d = None num_computed_tokens = None num_computed_tokens_p = None # for prefix caching block_idx_first_scheduled_token = None block_idx_first_scheduled_token_p = None block_idx_last_computed_token = None block_idx_last_scheduled_token = None # for causal_conv1d nums_dict, batch_ptr, token_chunk_offset_ptr = None, None, None if self.vllm_config.cache_config.mamba_cache_mode == "all": num_computed_tokens = common_attn_metadata.compute_num_computed_tokens() # Return a tensor of shape (#requests, #max blocks) state_indices_tensor = common_attn_metadata.block_table_tensor # Additional cache-related varaiables: mamba_block_size = self.kv_cache_spec.block_size ( block_idx_last_computed_token, block_idx_first_scheduled_token, block_idx_last_scheduled_token, ) = self._compute_prefix_caching_block_indices( common_attn_metadata, mamba_block_size ) else: state_indices_tensor = mamba_get_block_table_tensor( common_attn_metadata.block_table_tensor, common_attn_metadata.seq_lens, self.kv_cache_spec, self.vllm_config.cache_config.mamba_cache_mode, ) if state_indices_tensor.dim() == 1: state_indices_tensor = state_indices_tensor.unsqueeze(-1) state_indices_tensor_d, state_indices_tensor_p = torch.split( state_indices_tensor, [num_decodes, num_prefills], dim=0, ) if self.vllm_config.cache_config.mamba_cache_mode != "all": state_indices_tensor_d = state_indices_tensor_d[ :, : 1 + self.num_spec_tokens ] state_indices_tensor_p = state_indices_tensor_p[:, 0] if num_decodes > 0 and self.use_spec_decode: assert num_accepted_tokens is not None query_start_loc_d = common_attn_metadata.query_start_loc[: num_decodes + 1] num_accepted_tokens = num_accepted_tokens[:num_decodes] if num_prefills > 0: if num_computed_tokens is None: num_computed_tokens = common_attn_metadata.compute_num_computed_tokens() query_start_loc_p_cpu = ( common_attn_metadata.query_start_loc_cpu[-num_prefills - 1 :] - num_decode_tokens ) query_start_loc_p = ( common_attn_metadata.query_start_loc[-num_prefills - 1 :] - num_decode_tokens ) has_initial_states_p = ( num_computed_tokens[num_reqs - num_prefills : num_reqs] > 0 ) nums_dict, batch_ptr, token_chunk_offset_ptr = ( compute_causal_conv1d_metadata( query_start_loc_p_cpu, device=common_attn_metadata.query_start_loc.device, ) ) if self.vllm_config.cache_config.mamba_cache_mode == "all": assert num_computed_tokens is not None num_computed_tokens_p = num_computed_tokens[ num_reqs - num_prefills : num_reqs ] assert block_idx_first_scheduled_token is not None block_idx_first_scheduled_token_p = block_idx_first_scheduled_token[ num_reqs - num_prefills : num_reqs ] metadata = self.metadata_cls( num_prefills=num_prefills, num_prefill_tokens=num_prefill_tokens, num_decodes=num_decodes, num_decode_tokens=num_decode_tokens, query_start_loc_p=query_start_loc_p, has_initial_states_p=has_initial_states_p, state_indices_tensor_p=state_indices_tensor_p, state_indices_tensor_d=state_indices_tensor_d, num_accepted_tokens=num_accepted_tokens, query_start_loc_d=query_start_loc_d, block_idx_last_scheduled_token=block_idx_last_scheduled_token, block_idx_first_scheduled_token_p=block_idx_first_scheduled_token_p, block_idx_last_computed_token=block_idx_last_computed_token, num_computed_tokens_p=num_computed_tokens_p, num_reqs=num_reqs, seq_lens=common_attn_metadata.seq_lens, nums_dict=nums_dict, batch_ptr=batch_ptr, token_chunk_offset_ptr=token_chunk_offset_ptr, ) return self._update_metadata_for_cudagraph_capture(metadata) def _update_metadata_for_cudagraph_capture( self, metadata: M, ) -> M: """ Update the metadata for cudagraph capture. Currently, only decode is supported for full cudagraphs with Mamba. """ state_indices_tensor_d = metadata.state_indices_tensor_d query_start_loc_d = metadata.query_start_loc_d num_accepted_tokens = metadata.num_accepted_tokens block_idx_last_scheduled_token = metadata.block_idx_last_scheduled_token block_idx_last_computed_token = metadata.block_idx_last_computed_token if ( metadata.num_prefills == 0 and metadata.num_decodes <= self.decode_cudagraph_max_bs and self.compilation_config.cudagraph_mode.has_full_cudagraphs() ): padded_bs = metadata.num_reqs self.state_indices_tensor_d[: metadata.num_decodes].copy_( state_indices_tensor_d, non_blocking=True ) state_indices_tensor_d = self.state_indices_tensor_d[:padded_bs] state_indices_tensor_d[metadata.num_decodes :] = PAD_SLOT_ID if self.use_spec_decode: assert query_start_loc_d is not None assert num_accepted_tokens is not None query_start_loc_d = query_start_loc_d[: padded_bs + 1] self.decode_num_accepted_tokens[: metadata.num_decodes].copy_( num_accepted_tokens, non_blocking=True ) num_accepted_tokens = self.decode_num_accepted_tokens[:padded_bs] num_accepted_tokens[metadata.num_decodes :] = ( 1 # pad with 1st slot index ) if self.vllm_config.cache_config.mamba_cache_mode == "all": assert block_idx_last_scheduled_token is not None assert block_idx_last_computed_token is not None self.block_idx_last_scheduled_token[: metadata.num_decodes].copy_( block_idx_last_scheduled_token[: metadata.num_decodes], non_blocking=True, ) block_idx_last_scheduled_token = self.block_idx_last_scheduled_token[ : metadata.num_decode_tokens ] self.block_idx_last_computed_token[: metadata.num_decodes].copy_( block_idx_last_computed_token[: metadata.num_decodes], non_blocking=True, ) block_idx_last_computed_token = self.block_idx_last_computed_token[ : metadata.num_decode_tokens ] return replace( metadata, state_indices_tensor_d=state_indices_tensor_d, query_start_loc_d=query_start_loc_d, num_accepted_tokens=num_accepted_tokens, block_idx_last_scheduled_token=block_idx_last_scheduled_token, block_idx_last_computed_token=block_idx_last_computed_token, ) def update_block_table( self, metadata: M, blk_table: torch.Tensor, slot_mapping: torch.Tensor, ) -> M: state_indices_tensor = mamba_get_block_table_tensor( blk_table, metadata.seq_lens, self.kv_cache_spec, self.vllm_config.cache_config.mamba_cache_mode, ) if state_indices_tensor.dim() == 1: state_indices_tensor = state_indices_tensor.unsqueeze(-1) assert ( metadata.num_prefills + metadata.num_decodes == state_indices_tensor.shape[0] ), ( "Mismatch in number of requests when updating block table." f" Expected {metadata.num_prefills + metadata.num_decodes}, " f"got {state_indices_tensor.shape[0]}." ) state_indices_tensor_d, state_indices_tensor_p = torch.split( state_indices_tensor, [metadata.num_decodes, metadata.num_prefills], dim=0, ) if self.vllm_config.cache_config.mamba_cache_mode != "all": state_indices_tensor_d = state_indices_tensor_d[ :, : 1 + self.num_spec_tokens ] state_indices_tensor_p = state_indices_tensor_p[:, 0] new_metadata = replace( metadata, state_indices_tensor_d=state_indices_tensor_d, state_indices_tensor_p=state_indices_tensor_p, ) return self._update_metadata_for_cudagraph_capture(new_metadata)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/attention/backends/mamba_attn.py", "license": "Apache License 2.0", "lines": 516, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/test_onednn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from tests.kernels.utils import to_int8 from vllm import _custom_ops as ops from vllm.platforms import current_platform if not current_platform.is_cpu(): pytest.skip("skipping CPU-only tests", allow_module_level=True) NK_FACTORS = [ (256, 128), (4096, 4096), (16384, 4096), (1023, 491), (1001, 15), ] M_FACTORS = [ (16, 1, 32, 128, 64), (1, 17, 1, 31, 17), ] CACHE_SIZES = [2] DTYPE = [torch.bfloat16] def rand_int8(shape: tuple, device: str = "cpu"): return to_int8(torch.rand(shape, device=device) * 255 - 128) def ref_int8_scaled_mm( a: torch.Tensor, b: torch.Tensor, scale_a: torch.Tensor, scale_b: torch.Tensor, azp: torch.Tensor | None, bias: torch.Tensor | None, output_type: torch.dtype, ): if azp is not None: a = a.to(dtype=torch.float32) - azp.to(dtype=torch.float32) output = torch.mm( (scale_a * a.to(dtype=torch.float32)), (scale_b * b.to(dtype=torch.float32)) ) if bias is not None: output += bias.float() return output.to(dtype=output_type) def onednn_int8_gemm_test_helper( primitive_cache_size: int, m: int, n: int, k: int, per_tensor_a_quant: bool, per_tensor_b_quant: bool, use_azp: bool, use_bias: bool, out_dtype: torch.dtype = torch.bfloat16, device: str = "cpu", ): # Test for a oneDNN kernel with per-tensor / per-token activation # quantization and per-tensor / per-output channel weight quantization. a = to_int8(torch.randn((m, k), device=device) * 5) b = to_int8(torch.randn((n, k), device=device).t() * 5) a_scales_shape = (1, 1) if per_tensor_a_quant else (m, 1) b_scales_shape = (1, 1) if per_tensor_b_quant else (1, n) scale_a = torch.randn(a_scales_shape, device=device, dtype=torch.float32) scale_b = torch.randn(b_scales_shape, device=device, dtype=torch.float32) if use_azp: azp = torch.rand(a_scales_shape, dtype=torch.float32) * 10 + 1.5 azp = (azp / scale_a).round().to(dtype=torch.int32) azp_adj = scale_b * b.sum(dim=0, keepdim=True, dtype=torch.float32) else: azp = None azp_adj = None bias = torch.rand((n,), device=device, dtype=out_dtype) * 10 if use_bias else None handler = ops.create_onednn_scaled_mm( b, scale_b, out_dtype, not per_tensor_a_quant, use_azp, primitive_cache_size, ) out = torch.zeros((m, n), dtype=out_dtype) ops.onednn_scaled_mm(handler, a, out, scale_a, azp, azp_adj, bias) baseline = ref_int8_scaled_mm(a, b, scale_a, scale_b, azp, bias, out_dtype) torch.testing.assert_close(out, baseline, rtol=1e-1, atol=1e0) if use_bias: # To test runtime bias setting out = torch.zeros((m, n), dtype=out_dtype) ops.onednn_scaled_mm(handler, a, out, scale_a, azp, azp_adj, None) baseline = ref_int8_scaled_mm(a, b, scale_a, scale_b, azp, None, out_dtype) torch.testing.assert_close(out, baseline, rtol=1e-1, atol=1e0) def onednn_gemm_test_helper( primitive_cache_size: int, m: int, n: int, k: int, use_bias: bool, use_stride: bool, dtype: torch.dtype = torch.bfloat16, device: str = "cpu", ): if use_stride: a = torch.rand((m, 2 * k), dtype=dtype, device=device) * 1.5 a = a[:, :k] else: a = torch.rand((m, k), dtype=dtype, device=device) * 1.5 b = torch.rand((n, k), dtype=dtype, device=device) * 1.5 if use_bias: bias = torch.rand((n,), device=device, dtype=dtype) * 5 bias_f32 = bias.float() else: bias = None bias_f32 = None handler = ops.create_onednn_mm( b.t(), primitive_cache_size, ) out = ops.onednn_mm(handler, a, bias) baseline = torch.nn.functional.linear(a.float(), b.float(), bias_f32).to( dtype=a.dtype ) torch.testing.assert_close(out, baseline) if use_bias: # To test runtime bias setting out = ops.onednn_mm(handler, a, None) baseline = torch.nn.functional.linear(a.float(), b.float(), None).to( dtype=a.dtype ) torch.testing.assert_close(out, baseline) @pytest.mark.parametrize("n,k", NK_FACTORS) @pytest.mark.parametrize("m_list", M_FACTORS) @pytest.mark.parametrize("per_tensor_a_scale", [True, False]) @pytest.mark.parametrize("per_tensor_b_scale", [True, False]) @pytest.mark.parametrize("use_bias", [True, False]) @pytest.mark.parametrize("use_azp", [True, False]) @pytest.mark.parametrize("output_type", DTYPE) @pytest.mark.parametrize("primitive_cache_size", CACHE_SIZES) def test_onednn_int8_scaled_gemm( n: int, k: int, m_list: tuple[int, ...], per_tensor_a_scale: bool, per_tensor_b_scale: bool, use_bias: bool, use_azp: bool, output_type: torch.dtype, primitive_cache_size: int, ): for m in m_list: onednn_int8_gemm_test_helper( primitive_cache_size=primitive_cache_size, m=m, n=n, k=k, per_tensor_a_quant=per_tensor_a_scale, per_tensor_b_quant=per_tensor_b_scale, use_bias=use_bias, use_azp=use_azp, out_dtype=output_type, ) @pytest.mark.parametrize("n,k", NK_FACTORS) @pytest.mark.parametrize("m_list", M_FACTORS) @pytest.mark.parametrize("use_bias", [True, False]) @pytest.mark.parametrize("use_stride", [True, False]) @pytest.mark.parametrize("dtype", DTYPE) @pytest.mark.parametrize("primitive_cache_size", CACHE_SIZES) def test_onednn_gemm( n: int, k: int, m_list: tuple[int, ...], use_bias: bool, use_stride: bool, dtype: torch.dtype, primitive_cache_size: int, ): for m in m_list: onednn_gemm_test_helper( primitive_cache_size=primitive_cache_size, m=m, n=n, k=k, use_bias=use_bias, use_stride=use_stride, dtype=dtype, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/test_onednn.py", "license": "Apache License 2.0", "lines": 181, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/attention/test_mla_backends.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests for v1 MLA backends without GPUModelRunner dependency. Known Issues: - FLASH_ATTN_MLA backend occasionally produces NaN values in test_backend_correctness[mixed_small] when run after test_backend_correctness[small_prefill], but passes when run alone. """ import pytest import torch from tests.v1.attention.utils import ( BatchSpec, create_common_attn_metadata, create_vllm_config, try_get_attention_backend, ) from vllm import _custom_ops as ops from vllm.config.vllm import set_current_vllm_config from vllm.model_executor.layers.attention.mla_attention import ( QueryLenSupport, _DecodeConcatQuantFP8, ) from vllm.model_executor.layers.attention_layer_base import AttentionLayerBase from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape from vllm.platforms import current_platform from vllm.utils.math_utils import cdiv from vllm.utils.torch_utils import STR_DTYPE_TO_TORCH_DTYPE from vllm.v1.attention.backend import CommonAttentionMetadata from vllm.v1.attention.backends.fa_utils import flash_attn_supports_mla from vllm.v1.attention.backends.registry import AttentionBackendEnum from vllm.v1.attention.ops.flashmla import is_flashmla_dense_supported from vllm.v1.kv_cache_interface import MLAAttentionSpec BACKENDS_TO_TEST = [ AttentionBackendEnum.CUTLASS_MLA, AttentionBackendEnum.FLASHMLA, AttentionBackendEnum.FLASH_ATTN_MLA, AttentionBackendEnum.FLASHINFER_MLA, AttentionBackendEnum.TRITON_MLA, ] # Remove sm100 backends from the list if not using sm100 if not torch.cuda.is_available() or torch.cuda.get_device_properties(0).major < 10: BACKENDS_TO_TEST.remove(AttentionBackendEnum.CUTLASS_MLA) BACKENDS_TO_TEST.remove(AttentionBackendEnum.FLASHINFER_MLA) # Remove FLASH_ATTN_MLA from the list if not supported if not flash_attn_supports_mla(): BACKENDS_TO_TEST.remove(AttentionBackendEnum.FLASH_ATTN_MLA) # Remove FLASHMLA from the list if not supported if not is_flashmla_dense_supported()[0]: BACKENDS_TO_TEST.remove(AttentionBackendEnum.FLASHMLA) SPEC_DECODE_BACKENDS = [] for backend in BACKENDS_TO_TEST: builder_cls, _ = try_get_attention_backend(backend) query_len_support = getattr( builder_cls, "query_len_support", QueryLenSupport.SINGLE_ONLY ) if query_len_support != QueryLenSupport.SINGLE_ONLY: SPEC_DECODE_BACKENDS.append(backend) BACKEND_BLOCK_SIZES = {} for backend in BACKENDS_TO_TEST: supported_sizes = backend.get_class().get_supported_kernel_block_sizes() if supported_sizes: default_size = supported_sizes[0] block_size = ( default_size if isinstance(default_size, int) else default_size.base ) else: block_size = 16 BACKEND_BLOCK_SIZES[backend] = block_size torch.manual_seed(42) def _convert_dtype_to_torch(dtype): """Convert ModelDType to torch.dtype.""" if isinstance(dtype, str): if dtype == "auto": return torch.float16 # Default dtype for testing elif dtype in STR_DTYPE_TO_TORCH_DTYPE: return STR_DTYPE_TO_TORCH_DTYPE[dtype] else: raise ValueError(f"Unknown dtype: {dtype}") elif isinstance(dtype, torch.dtype): return dtype else: raise ValueError(f"Unknown dtype: {dtype}") # Define common batch configurations BATCH_SPECS = { "small_decode": BatchSpec(seq_lens=[32, 40], query_lens=[1, 1]), "small_prefill": BatchSpec(seq_lens=[32, 40], query_lens=[8, 8]), "mixed_small": BatchSpec(seq_lens=[32, 40, 48, 56], query_lens=[1, 1, 5, 5]), "medium_decode": BatchSpec( seq_lens=[128, 256, 512, 1024, 128, 256, 512, 1024], query_lens=[1, 1, 1, 1, 1, 1, 1, 1], ), "medium_prefill": BatchSpec( seq_lens=[256, 512, 1024, 2048], query_lens=[16, 16, 16, 16] ), "mixed_medium": BatchSpec( seq_lens=[512, 1024, 2048, 512, 1024, 2048], query_lens=[1, 1, 1, 7, 7, 7] ), "large_decode": BatchSpec(seq_lens=[2048] * 32, query_lens=[1] * 32), "large_prefill": BatchSpec(seq_lens=[4096] * 8, query_lens=[32] * 8), "single_decode": BatchSpec(seq_lens=[1024], query_lens=[1]), "single_prefill": BatchSpec(seq_lens=[1024], query_lens=[64]), "spec_decode_small": BatchSpec( seq_lens=[128, 256, 512, 1024], query_lens=[4, 4, 4, 4] ), "spec_decode_medium": BatchSpec( seq_lens=[512, 1024, 2048, 512, 1024, 2048], query_lens=[8, 8, 8, 8, 8, 8] ), } def create_and_prepopulate_kv_cache( kv_c_contexts: list[torch.Tensor], k_pe_contexts: list[torch.Tensor], block_size: int, head_size: int, dtype: torch.dtype, device: torch.device, num_blocks: int, common_attn_metadata: CommonAttentionMetadata, randomize_blocks: bool = True, kv_cache_dtype: str | None = None, scale: float | torch.Tensor = 1.0, ) -> torch.Tensor: """Create and prepopulate an MLA KV cache with context data. Args: kv_c_contexts: List of latent KV context tensors for each sequence k_pe_contexts: List of key positional embedding context tensors for each sequence block_size: Size of each block head_size: Size of each head (latent dimension) dtype: Data type for the cache device: Device to create the cache on num_blocks: Total number of blocks in the cache common_attn_metadata: Common attention metadata randomize_blocks: Whether to randomly permute blocks or use sequential order kv_cache_dtype: Optional kv cache dtype string. For fp8 cache dtype, the cache is populated via concat_and_cache_mla. scale: Scaling factor forwarded to concat_and_cache_mla when the fp8 cache layout is requested. Returns: MLA KV cache tensor """ batch_size = len(kv_c_contexts) seq_lens = common_attn_metadata.seq_lens.cpu() query_lens = ( common_attn_metadata.query_start_loc_cpu[1:] - common_attn_metadata.query_start_loc_cpu[:-1] ) context_lens = seq_lens - query_lens block_table = common_attn_metadata.block_table_tensor slot_mapping = common_attn_metadata.slot_mapping fp8_attention = kv_cache_dtype and kv_cache_dtype.startswith("fp8") use_fp8_ds_mla = kv_cache_dtype == "fp8_ds_mla" if fp8_attention: if use_fp8_ds_mla: kv_lora_rank = kv_c_contexts[0].shape[-1] rope_dim = k_pe_contexts[0].shape[-1] # 4 * 4: 4 float32 scale values for 128-element tiles # 2 * rope_dim: 16-bit RoPE values kv_entry_size = kv_lora_rank + 4 * 4 + 2 * rope_dim else: kv_entry_size = head_size kv_cache = torch.zeros( num_blocks, block_size, kv_entry_size, dtype=torch.uint8, device=device ) scale_tensor = ( scale if isinstance(scale, torch.Tensor) else torch.tensor(scale, dtype=torch.float32, device=device) ) scale_tensor = scale_tensor.to(device=device, dtype=torch.float32) else: # Create MLA KV cache: (num_blocks, block_size, head_size) kv_cache = torch.zeros( num_blocks, block_size, head_size, dtype=dtype, device=device ) kv_cache_flat = kv_cache.view(-1, head_size) # Populate the cache with the context tokens # Start from block_id=1 since block_id=0 is considered the null block start_block_idx = 1 for i in range(batch_size): kv_c_context, k_pe_context = kv_c_contexts[i], k_pe_contexts[i] context_len = kv_c_context.shape[0] if context_len == 0: start_block_idx += cdiv(int(seq_lens[i]), block_size) continue start = start_block_idx * block_size if fp8_attention: slots = torch.arange(context_len, device=device, dtype=torch.long) + start ops.concat_and_cache_mla( kv_c_context, k_pe_context.squeeze(1), kv_cache, slots, kv_cache_dtype=kv_cache_dtype, scale=scale_tensor, ) else: kv_context = torch.cat([kv_c_context, k_pe_context.squeeze(1)], dim=-1) end = start + kv_context.shape[0] kv_cache_flat[start:end, ...] = kv_context # Stay block aligned and allocate enough blocks for the new tokens start_block_idx += cdiv(int(seq_lens[i]), block_size) blocks_end = start_block_idx # Permute the context blocks (excluding block 0 which is null) if randomize_blocks: perm = ( torch.randperm(blocks_end - 1) + 1 ) # Random permutation starting from block 1 else: perm = torch.arange(1, blocks_end) # Sequential order starting from block 1 inv_perm = torch.zeros(blocks_end, dtype=torch.long, device=device) inv_perm[1:] = torch.argsort(perm) + 1 # Add 1 to account for starting from block 1 kv_cache[1:blocks_end, ...] = kv_cache[perm, ...] # Construct the right block table # Start from block_id=1 since block_id=0 is considered the null block start_block_idx = 1 for i in range(batch_size): num_blocks_for_seq = cdiv(int(seq_lens[i]), block_size) start = start_block_idx end = start + num_blocks_for_seq block_table[i, :num_blocks_for_seq] = inv_perm[start:end] block_table[i, num_blocks_for_seq:] = 0 start_block_idx += num_blocks_for_seq # Create a realistic slot mapping that corresponds to the block table for i in range(batch_size): token_offsets = torch.arange(int(query_lens[i])) + int(context_lens[i]) block_indices = token_offsets // block_size token_inter_block_offsets = token_offsets % block_size start = common_attn_metadata.query_start_loc_cpu[i] end = common_attn_metadata.query_start_loc_cpu[i + 1] slot_mapping[start:end] = block_table[ i, block_indices ] * block_size + token_inter_block_offsets.to(device) return kv_cache class MockAttentionLayer: """A mock attention layer for testing.""" def __init__(self, device: torch.device): self._q_scale = torch.tensor(1.0, device=device) self._k_scale = torch.tensor(1.0, device=device) self._v_scale = torch.tensor(1.0, device=device) self._prob_scale = torch.tensor(1.0, device=device) self._q_scale_float = 1.0 self._k_scale_float = 1.0 self._v_scale_float = 1.0 def forward(self, *_args, **_kwargs): raise NotImplementedError class MockSparseMLAAttentionLayer: """A mock sparse MLA attention layer for testing. Sparse MLA implementations only support forward_mqa (decode-style attention) for all tokens, so this class only implements that path. Unlike regular MLA impls, sparse MLA impls don't have W_UK_T and W_UV attributes. These transformations are done by the layer (MLAAttention), not the impl. This mock layer accepts these weight matrices directly. """ def __init__( self, impl, num_heads: int, qk_nope_head_dim: int, qk_rope_head_dim: int, v_head_dim: int, kv_lora_rank: int, device: torch.device, W_UK: torch.Tensor, W_UV: torch.Tensor, ): self.impl = impl self.num_heads = num_heads self.qk_nope_head_dim = qk_nope_head_dim self.qk_rope_head_dim = qk_rope_head_dim self.v_head_dim = v_head_dim self.kv_lora_rank = kv_lora_rank # Compute weight matrices in the format expected by forward_impl # W_UK shape: (L, N, P) -> W_UK_T shape: (N, P, L) self.W_UK_T = W_UK.permute(1, 2, 0) # W_UV shape: (L, N, V) -> (N, L, V) self.W_UV = W_UV.transpose(0, 1) # Scale attributes needed by attention backends self._q_scale = torch.tensor(1.0, device=device) self._k_scale = torch.tensor(1.0, device=device) self._v_scale = torch.tensor(1.0, device=device) self._prob_scale = torch.tensor(1.0, device=device) self._q_scale_float = 1.0 self._k_scale_float = 1.0 self._v_scale_float = 1.0 def forward_impl( self, q: torch.Tensor, kv_c: torch.Tensor, k_pe: torch.Tensor, kv_cache: torch.Tensor, attn_metadata, output: torch.Tensor, ) -> torch.Tensor: """Forward for sparse MLA - uses forward_mqa for all tokens.""" kv_cache_dtype = getattr(self.impl, "kv_cache_dtype", "auto") # Write to KV cache if kv_cache.numel() > 0: ops.concat_and_cache_mla( kv_c, k_pe.squeeze(1), kv_cache, attn_metadata.slot_mapping.flatten(), kv_cache_dtype=kv_cache_dtype, scale=self._k_scale, ) num_tokens = q.shape[0] # Sparse MLA uses forward_mqa for all tokens # Split q into nope and pe parts mqa_q_nope, mqa_q_pe = q.split( [self.qk_nope_head_dim, self.qk_rope_head_dim], dim=-1 ) # Convert from (B, N, P) to (N, B, P) mqa_q_nope = mqa_q_nope.transpose(0, 1) # Multiply (N, B, P) x (N, P, L) -> (N, B, L) mqa_ql_nope = torch.bmm(mqa_q_nope, self.W_UK_T) # Convert from (N, B, L) to (B, N, L) mqa_ql_nope = mqa_ql_nope.transpose(0, 1) # Pass as tuple to forward_mqa mqa_q = (mqa_ql_nope, mqa_q_pe) attn_out, _ = self.impl.forward_mqa(mqa_q, kv_cache, attn_metadata, self) # v_up projection: multiply by W_UV # attn_out shape: (B, N, L) where L = kv_lora_rank # W_UV shape: (N, L, V) # output shape: (B, N, V) -> flatten to (B, N*V) decode_output = torch.bmm(attn_out.transpose(0, 1), self.W_UV).transpose(0, 1) output[:num_tokens] = decode_output.reshape( num_tokens, self.num_heads * self.v_head_dim ) return output class MockMLAAttentionLayer(AttentionLayerBase): """A mock MLA attention layer for testing. This replicates the forward_impl logic from MLAAttention to allow testing MLA backends without the full layer infrastructure. The W_UK_T and W_UV weight matrices are created on the layer (like in MLAAttention.process_weights_after_loading), not on the impl. """ def __init__( self, impl, num_heads: int, qk_nope_head_dim: int, qk_rope_head_dim: int, v_head_dim: int, kv_lora_rank: int, device: torch.device, kv_b_proj, ): self.impl = impl self.num_heads = num_heads self.qk_nope_head_dim = qk_nope_head_dim self.qk_rope_head_dim = qk_rope_head_dim self.v_head_dim = v_head_dim self.kv_lora_rank = kv_lora_rank # Compute weight matrices from kv_b_proj (like MLAAttention does) # This replicates MLAAttention.process_weights_after_loading logic kv_b_proj_weight = kv_b_proj.weight.T kv_b_proj_weight = kv_b_proj_weight.view( kv_lora_rank, num_heads, qk_nope_head_dim + v_head_dim, ) W_UK, W_UV = kv_b_proj_weight.split([qk_nope_head_dim, v_head_dim], dim=-1) # Convert from (L, N, V) to (N, L, V) self.W_UV = W_UV.transpose(0, 1) # Convert from (L, N, P) to (N, P, L) self.W_UK_T = W_UK.permute(1, 2, 0) # Scale attributes needed by attention backends self._q_scale = torch.tensor(1.0, device=device) self._k_scale = torch.tensor(1.0, device=device) self._v_scale = torch.tensor(1.0, device=device) self._prob_scale = torch.tensor(1.0, device=device) self._q_scale_float = 1.0 self._k_scale_float = 1.0 self._v_scale_float = 1.0 self._decode_concat_quant_fp8_op = _DecodeConcatQuantFP8( static=True, group_shape=GroupShape.PER_TENSOR, compile_native=True, ) def get_attn_backend(self): raise NotImplementedError def get_kv_cache_spec(self, vllm_config): raise NotImplementedError def forward_impl( self, q: torch.Tensor, kv_c: torch.Tensor, k_pe: torch.Tensor, kv_cache: torch.Tensor, attn_metadata, output: torch.Tensor, ) -> torch.Tensor: """Replicates MLAAttention.forward_impl logic for testing.""" # Write to KV cache kv_cache_dtype = getattr(self.impl, "kv_cache_dtype", "auto") fp8_attention = kv_cache_dtype.startswith("fp8") if kv_cache.numel() > 0: ops.concat_and_cache_mla( kv_c, k_pe.squeeze(1), kv_cache, attn_metadata.slot_mapping.flatten(), kv_cache_dtype=kv_cache_dtype, scale=self._k_scale, ) if fp8_attention and kv_cache_dtype != "fp8_ds_mla": kv_cache = kv_cache.view(current_platform.fp8_dtype()) # Determine decode vs prefill split num_decode_tokens = attn_metadata.num_decode_tokens or 0 has_decode = (attn_metadata.num_decodes or 0) > 0 has_prefill = (attn_metadata.num_prefills or 0) > 0 # Run prefill with forward_mha if has_prefill: prefill_q = q[num_decode_tokens:] prefill_k_pe = k_pe[num_decode_tokens:] prefill_k_c = kv_c[num_decode_tokens:] self.impl.forward_mha( prefill_q, prefill_k_c, prefill_k_pe, kv_cache, attn_metadata, self._k_scale, output=output[num_decode_tokens:], ) # Run decode with forward_mqa if has_decode: decode_q = q[:num_decode_tokens] # Split q into nope and pe parts mqa_q_nope, mqa_q_pe = decode_q.split( [self.qk_nope_head_dim, self.qk_rope_head_dim], dim=-1 ) # Convert from (B, N, P) to (N, B, P) mqa_q_nope = mqa_q_nope.transpose(0, 1) # Multiply (N, B, P) x (N, P, L) -> (N, B, L) mqa_ql_nope = torch.bmm(mqa_q_nope, self.W_UK_T) # Convert from (N, B, L) to (B, N, L) mqa_ql_nope = mqa_ql_nope.transpose(0, 1) if fp8_attention and self.impl.supports_quant_query_input: assert mqa_ql_nope.shape[0] == mqa_q_pe.shape[0] assert mqa_ql_nope.shape[1] == mqa_q_pe.shape[1] mqa_q = self._decode_concat_quant_fp8_op( mqa_ql_nope, mqa_q_pe, self._q_scale ) else: mqa_q = (mqa_ql_nope, mqa_q_pe) attn_out, _ = self.impl.forward_mqa(mqa_q, kv_cache, attn_metadata, self) # v_up projection: multiply by W_UV # attn_out shape: (B, N, L) where L = kv_lora_rank # W_UV shape: (N, L, V) # output shape: (B, N, V) -> flatten to (B, N*V) decode_output = torch.bmm(attn_out.transpose(0, 1), self.W_UV).transpose( 0, 1 ) output[:num_decode_tokens] = decode_output.reshape( num_decode_tokens, self.num_heads * self.v_head_dim ) return output def run_attention_backend( backend: AttentionBackendEnum, kv_cache_spec: MLAAttentionSpec, layer_names: list[str], vllm_config, device: torch.device, common_attn_metadata: CommonAttentionMetadata, query: torch.Tensor, kv_c: torch.Tensor, k_pe: torch.Tensor, kv_cache: torch.Tensor, kv_lora_rank: int, qk_nope_head_dim: int, qk_rope_head_dim: int, v_head_dim: int, mock_kv_b_proj, kv_cache_dtype: str = "auto", ) -> torch.Tensor: """Run attention computation using the specified backend's AttentionImpl.""" builder_cls, impl_cls = try_get_attention_backend(backend) # Set the current vllm config so that get_current_vllm_config() works # in the backend implementations with set_current_vllm_config(vllm_config): # Instantiate MLA implementation num_heads = vllm_config.model_config.get_num_attention_heads( vllm_config.parallel_config ) num_kv_heads = vllm_config.model_config.get_num_kv_heads( vllm_config.parallel_config ) head_size = vllm_config.model_config.get_head_size() scale = 1.0 / (head_size**0.5) impl = impl_cls( num_heads=num_heads, head_size=head_size, scale=scale, num_kv_heads=num_kv_heads, alibi_slopes=None, sliding_window=None, kv_cache_dtype=kv_cache_dtype, logits_soft_cap=None, attn_type="decoder", kv_sharing_target_layer_name=None, q_lora_rank=None, kv_lora_rank=kv_lora_rank, qk_nope_head_dim=qk_nope_head_dim, qk_rope_head_dim=qk_rope_head_dim, qk_head_dim=qk_nope_head_dim + qk_rope_head_dim, v_head_dim=v_head_dim, kv_b_proj=mock_kv_b_proj, ) # Process weights on the impl act_dtype = _convert_dtype_to_torch(vllm_config.model_config.dtype) impl.process_weights_after_loading(act_dtype) # Initialize DCP attributes (normally set by MLAAttention.forward # before calling forward_mha, see mla_attention.py:511-512) if impl.dcp_world_size == -1: impl.dcp_world_size = 1 # Create mock MLA layer mock_layer = MockMLAAttentionLayer( impl=impl, num_heads=num_heads, qk_nope_head_dim=qk_nope_head_dim, qk_rope_head_dim=qk_rope_head_dim, v_head_dim=v_head_dim, kv_lora_rank=kv_lora_rank, device=device, kv_b_proj=mock_kv_b_proj, ) # Populate static_forward_context with mock attention layers for layer_name in layer_names: vllm_config.compilation_config.static_forward_context[layer_name] = ( mock_layer ) # Build metadata builder = builder_cls(kv_cache_spec, layer_names, vllm_config, device) attn_metadata = builder.build( common_prefix_len=0, common_attn_metadata=common_attn_metadata, ) # Create output buffer num_tokens = query.shape[0] output = torch.empty( num_tokens, num_heads * v_head_dim, dtype=query.dtype, device=query.device ) # Run forward pass output = mock_layer.forward_impl( query, kv_c, k_pe, kv_cache, attn_metadata, output ) return output @pytest.mark.parametrize( "batch_spec_name", [ "small_decode", "small_prefill", "mixed_small", "medium_decode", "medium_prefill", "mixed_medium", "large_decode", "large_prefill", "single_decode", "single_prefill", "spec_decode_small", "spec_decode_medium", ], ) @pytest.mark.parametrize("model", ["deepseek-ai/DeepSeek-R1"]) @pytest.mark.parametrize("tensor_parallel_size", [1, 4, 8, 16]) @pytest.mark.parametrize("kv_cache_dtype", ["auto", "fp8", "fp8_e4m3"]) def test_backend_correctness( default_vllm_config, dist_init, batch_spec_name: str, model: str, tensor_parallel_size: int, kv_cache_dtype: str, ): """ Test that all backends produce similar outputs to a reference implementation using torch.nn.functional.scaled_dot_product_attention. This test works by: 1. Generating a batch of sequences with specified context and query lengths. 2. Computing a ground-truth attention output using torch.sdpa on contiguous Q, K, and V tensors. 3. Simulating vLLM's paged KV cache: It takes the context portion of the K/V tensors and manually places them into a paged buffer according to the test's (randomly generated) block table. 4. Running each vLLM attention backend with the new queries and the simulated paged KV cache. 5. Comparing the vLLM backend's output to the ground-truth SDPA output. Note: When tensor_parallel_size > 1, we simulate the head partitioning by overriding the model config to use fewer heads, without requiring multiple GPUs. This tests that backends work correctly with different head counts. """ # Filter backends to those that support the requested kv_cache_dtype backends_to_test = [ b for b in BACKENDS_TO_TEST if kv_cache_dtype in b.get_class().supported_kv_cache_dtypes ] if not backends_to_test: pytest.skip(f"No backends support kv_cache_dtype={kv_cache_dtype}") batch_spec = BATCH_SPECS[batch_spec_name] is_spec_decode_test = batch_spec_name.startswith("spec_decode") unique_block_sizes = sorted(set(BACKEND_BLOCK_SIZES[b] for b in backends_to_test)) default_block_size = unique_block_sizes[0] required_blocks = sum( (seq_len + default_block_size - 1) // default_block_size for seq_len in batch_spec.seq_lens ) # Add 1 for null block at index 0, and some buffer num_gpu_blocks = required_blocks + 1 + 100 hf_config_override = None if tensor_parallel_size > 1: from vllm.config import ModelConfig temp_config = ModelConfig(model=model, max_model_len=1) original_num_heads = temp_config.hf_text_config.num_attention_heads original_num_kv_heads = getattr( temp_config.hf_text_config, "num_key_value_heads", None ) hf_config_override = { "num_attention_heads": original_num_heads // tensor_parallel_size, } if original_num_kv_heads is not None: hf_config_override["num_key_value_heads"] = max( 1, original_num_kv_heads // tensor_parallel_size ) vllm_config = create_vllm_config( model_name=model, tensor_parallel_size=1, # Always use TP=1 to avoid multi-GPU requirements max_model_len=max(batch_spec.seq_lens), num_gpu_blocks=num_gpu_blocks, block_size=default_block_size, hf_config_override=hf_config_override, ) vllm_config.cache_config.cache_dtype = kv_cache_dtype # For spec decode tests, add a speculative_config to set the reorder_batch_threshold if is_spec_decode_test: from vllm.config import SpeculativeConfig # Get the query length from the batch spec (they should all be uniform) query_len = batch_spec.query_lens[0] # Set num_speculative_tokens to query_len - 1 # (since threshold is 1 + num_spec_tokens) # Use ngram method which doesn't require a draft model vllm_config.speculative_config = SpeculativeConfig( method="ngram", num_speculative_tokens=query_len - 1 ) device = torch.device("cuda:0") # 1. Setup batch_size = batch_spec.batch_size seq_lens = batch_spec.seq_lens query_lens = batch_spec.query_lens num_q_heads = vllm_config.model_config.get_num_attention_heads( vllm_config.parallel_config ) head_size = vllm_config.model_config.get_head_size() dtype = _convert_dtype_to_torch(vllm_config.model_config.dtype) kv_lora_rank = 512 qk_rope_head_dim = 64 qk_nope_head_dim = 128 v_head_dim = 128 total_head_size = kv_lora_rank + qk_rope_head_dim assert kv_lora_rank + qk_rope_head_dim == head_size, ( f"MLA dimensions don't match: {total_head_size} != {head_size}" ) scale = 1.0 / (total_head_size**0.5) # 2. Generate data and compute SDPA reference output for MLA all_q_vllm, all_kv_c_vllm, all_k_pe_vllm = [], [], [] all_sdpa_outputs: list[list[torch.Tensor]] = [] kv_c_contexts, k_pe_contexts = [], [] # Create shared MLA weight matrices for consistency across all sequences W_UK = torch.randn( kv_lora_rank, num_q_heads, qk_nope_head_dim, dtype=dtype, device=device ) W_UV = torch.randn( kv_lora_rank, num_q_heads, v_head_dim, dtype=dtype, device=device ) # Scale weights to produce realistic magnitude outputs. # Without scaling, projection output has std ~sqrt(kv_lora_rank) ≈ 22.6, # causing extreme attention scores and numerical instability in LSE merging. weight_scale = 1.0 / (kv_lora_rank**0.5) W_UK = W_UK * weight_scale W_UV = W_UV * weight_scale kv_b_proj_weight = torch.cat([W_UK, W_UV], dim=-1) for i, backend in enumerate(backends_to_test): all_sdpa_outputs.append([]) for i in range(batch_size): s_len = seq_lens[i] q_len = query_lens[i] context_len = s_len - q_len # Generate MLA tensors # Q has both nope and rope components: # [q_len, num_heads, qk_nope_head_dim + qk_rope_head_dim] q_c = torch.randn( q_len, num_q_heads, qk_nope_head_dim + qk_rope_head_dim, dtype=dtype, device=device, ) # KV_C (latent K/V): [s_len, kv_lora_rank] kv_c_full = torch.randn(s_len, kv_lora_rank, dtype=dtype, device=device) # K_PE (rope component): [s_len, 1, qk_rope_head_dim] k_pe_full = torch.randn(s_len, 1, qk_rope_head_dim, dtype=dtype, device=device) # Determine if this sequence uses the decode pipeline or prefill # pipeline for each backend # NOTE: For spec decode tests with uniform query_len > 1, backends that # support spec decode (FLASH_ATTN_MLA with varlen support, FLASHMLA with # uniform support) will use the decode pipeline (MQA-style), while # backends that only support single-token queries will use the prefill # pipeline (MHA-style). This ensures the reference implementation # matches each backend's actual decode/prefill pipeline path. is_decode = [] for backend_idx, backend in enumerate(backends_to_test): builder_cls, _ = try_get_attention_backend(backend) if is_spec_decode_test: query_len_support = getattr( builder_cls, "query_len_support", QueryLenSupport.SINGLE_ONLY ) supports_spec = query_len_support != QueryLenSupport.SINGLE_ONLY is_decode.append(supports_spec) else: threshold = getattr(builder_cls, "reorder_batch_threshold", None) query_len_support = getattr( builder_cls, "query_len_support", QueryLenSupport.SINGLE_ONLY ) within_threshold = q_len <= threshold if threshold else False if ( within_threshold and query_len_support == QueryLenSupport.UNIFORM and i > 0 ): first_q_len = query_lens[0] within_threshold = q_len == first_q_len is_decode.append(within_threshold) # Split q into nope and rope components q_nope, q_pe = q_c.split([qk_nope_head_dim, qk_rope_head_dim], dim=-1) ####################################################### # Decode path: MQA-style attention in latent space # Transform q_nope to latent space: q_nope @ W_UK # q_nope: [1, num_heads, qk_nope_head_dim] # W_UK: [kv_lora_rank, num_heads, qk_nope_head_dim] ql_nope = torch.einsum( "qnh,lnh->qnl", q_nope, W_UK ) # [1, num_heads, kv_lora_rank] # Build MQA attention inputs # Q: [1, num_heads, kv_lora_rank + qk_rope_head_dim] q_mqa = torch.cat([ql_nope, q_pe], dim=-1) # K: [s_len, kv_lora_rank + qk_rope_head_dim] # (broadcasted to all heads) k_mqa = torch.cat([kv_c_full, k_pe_full.squeeze(1)], dim=-1) k_mqa = k_mqa.unsqueeze(1).expand(-1, num_q_heads, -1) # V: [s_len, kv_lora_rank] (broadcasted to all heads) v_mqa = kv_c_full.unsqueeze(1).expand(-1, num_q_heads, -1) # Create custom attention mask for decode path: # - Query tokens can attend to all context tokens # - Query tokens can only attend to query tokens up to their position attn_mask = torch.ones(q_len, s_len, dtype=torch.bool, device=device) # Apply causal mask only to the query portion (context_len onwards) causal_mask = torch.tril(torch.ones(q_len, q_len, device=device)) attn_mask[:, context_len:] = causal_mask # SDPA expects (N, H, L, D) q_sdpa_in = q_mqa.unsqueeze(0).transpose(1, 2) k_sdpa_in = k_mqa.unsqueeze(0).transpose(1, 2) v_sdpa_in = v_mqa.unsqueeze(0).transpose(1, 2) sdpa_out_i_decode = torch.nn.functional.scaled_dot_product_attention( q_sdpa_in, k_sdpa_in, v_sdpa_in, attn_mask=attn_mask, scale=scale ) sdpa_out_i_decode = sdpa_out_i_decode.transpose(1, 2).squeeze( 0 ) # [1, num_heads, kv_lora_rank] # Project back to output space: sdpa_out @ W_UV sdpa_out_i_decode = torch.einsum("qnl,lnv->qnv", sdpa_out_i_decode, W_UV) sdpa_out_i_decode = sdpa_out_i_decode.flatten(start_dim=-2) ####################################################### # Prefill path: MHA-style attention with full sequence # Apply kv_b_proj to the full kv_c tensor kv_nope_full = torch.einsum("sl,lnh->snh", kv_c_full, kv_b_proj_weight) k_nope_full, v_full = kv_nope_full.split([qk_nope_head_dim, v_head_dim], dim=-1) # Build attention inputs for full sequence q_mha = torch.cat([q_nope, q_pe], dim=-1) # [q_len, num_heads, total_dim] k_pe_full_expanded = k_pe_full.expand(-1, num_q_heads, -1) k_full = torch.cat([k_nope_full, k_pe_full_expanded], dim=-1) # Create custom attention mask: # - Query tokens can attend to all context tokens # - Query tokens can only attend to query tokens up to their pos attn_mask = torch.ones(q_len, s_len, dtype=torch.bool, device=device) # Apply causal mask only to the query portion (context_len onwards) causal_mask = torch.tril(torch.ones(q_len, q_len, device=device)) attn_mask[:, context_len:] = causal_mask # SDPA expects (N, H, L, D) q_sdpa_in = q_mha.unsqueeze(0).transpose(1, 2) k_sdpa_in = k_full.unsqueeze(0).transpose(1, 2) v_sdpa_in = v_full.unsqueeze(0).transpose(1, 2) # Single attention call with custom mask sdpa_out_i_prefill = torch.nn.functional.scaled_dot_product_attention( q_sdpa_in, k_sdpa_in, v_sdpa_in, attn_mask=attn_mask, scale=scale ) sdpa_out_i_prefill = sdpa_out_i_prefill.transpose(1, 2).squeeze(0) sdpa_out_i_prefill = sdpa_out_i_prefill.flatten(start_dim=-2) for backend_idx, backend in enumerate(backends_to_test): if is_decode[backend_idx]: all_sdpa_outputs[backend_idx].append(sdpa_out_i_decode) else: all_sdpa_outputs[backend_idx].append(sdpa_out_i_prefill) # Inputs for vLLM MLA backends are just the new tokens all_q_vllm.append(q_c) all_kv_c_vllm.append(kv_c_full[context_len:]) # New kv_c tokens all_k_pe_vllm.append(k_pe_full[context_len:]) # New k_pe tokens # Contextual K/V data used to populate the paged cache (MLA format) kv_c_contexts.append(kv_c_full[:context_len]) k_pe_contexts.append(k_pe_full[:context_len]) # Concatenate all sequences (no reordering needed) query_vllm = torch.cat(all_q_vllm, dim=0) kv_c_vllm = torch.cat(all_kv_c_vllm, dim=0) k_pe_vllm = torch.cat(all_k_pe_vllm, dim=0) sdpa_outputs = {} for backend_idx, backend in enumerate(backends_to_test): sdpa_outputs[backend] = torch.cat(all_sdpa_outputs[backend_idx], dim=0) # Create mock kv_b_proj using the same weights as reference implementation from vllm.model_executor.layers.linear import ColumnParallelLinear mock_kv_b_proj = ColumnParallelLinear( input_size=kv_lora_rank, output_size=num_q_heads * (qk_nope_head_dim + v_head_dim), bias=False, ).to(device=device, dtype=dtype) # Set the mock weights to match our reference implementation # Reshape W_UK and W_UV to match the expected kv_b_proj format # [kv_lora_rank, num_heads, qk_nope_head_dim + v_head_dim] kv_b_proj_weight = kv_b_proj_weight.view( kv_lora_rank, num_q_heads * (qk_nope_head_dim + v_head_dim) ) mock_kv_b_proj.weight = torch.nn.Parameter(kv_b_proj_weight.T, requires_grad=False) # 3. Create metadata and KV caches for each block size # Group backends by block size and test each group metadata_per_block_size = {} kv_cache_per_block_size = {} for block_size in unique_block_sizes: # Create metadata for this block size common_attn_metadata = create_common_attn_metadata( batch_spec, block_size, device ) # Pad block table to meet requirement: # block_num % (128 / block_size) == 0 required_divisor = int(128 / block_size) current_block_num = common_attn_metadata.block_table_tensor.shape[1] if current_block_num % required_divisor != 0: # Pad to next multiple of required_divisor padded_block_num = ( (current_block_num + required_divisor - 1) // required_divisor ) * required_divisor padding_cols = padded_block_num - current_block_num padding = torch.zeros( (common_attn_metadata.block_table_tensor.shape[0], padding_cols), dtype=torch.int32, device=device, ) common_attn_metadata.block_table_tensor = torch.cat( [common_attn_metadata.block_table_tensor, padding], dim=1 ) metadata_per_block_size[block_size] = common_attn_metadata # Create KV cache for this block size required_blocks_for_size = sum( (seq_len + block_size - 1) // block_size for seq_len in batch_spec.seq_lens ) num_blocks_for_size = required_blocks_for_size + 1 + 100 kv_cache = create_and_prepopulate_kv_cache( kv_c_contexts=kv_c_contexts, k_pe_contexts=k_pe_contexts, block_size=block_size, head_size=head_size, dtype=dtype, device=device, num_blocks=num_blocks_for_size, common_attn_metadata=common_attn_metadata, randomize_blocks=True, kv_cache_dtype=kv_cache_dtype, ) kv_cache_per_block_size[block_size] = kv_cache # 4. Run vLLM backends and compare failures = [] for backend_idx, backend_name in enumerate(backends_to_test): # Skip backends that don't support spec decode for spec decode tests if is_spec_decode_test and backend_name not in SPEC_DECODE_BACKENDS: continue # Get the appropriate block_size, metadata, and cache for this backend block_size = BACKEND_BLOCK_SIZES[backend_name] common_attn_metadata = metadata_per_block_size[block_size] kv_cache = kv_cache_per_block_size[block_size] # Create kv_cache_spec with the correct block_size for this backend backend_kv_cache_spec = MLAAttentionSpec( block_size=block_size, num_kv_heads=vllm_config.model_config.get_num_kv_heads( vllm_config.parallel_config ), head_size=vllm_config.model_config.get_head_size(), dtype=vllm_config.model_config.dtype, sliding_window=vllm_config.model_config.get_sliding_window(), cache_dtype_str=kv_cache_dtype, ) backend_output = run_attention_backend( backend_name, backend_kv_cache_spec, ["placeholder"], vllm_config, device, common_attn_metadata, query_vllm, kv_c_vllm, k_pe_vllm, kv_cache, kv_lora_rank, qk_nope_head_dim, qk_rope_head_dim, v_head_dim, mock_kv_b_proj, kv_cache_dtype=kv_cache_dtype, ) # Use backend_idx to get the correct SDPA output for this backend expected_output = sdpa_outputs[backend_name] # Check shape and dtype consistency try: assert backend_output.shape == expected_output.shape, ( f"[{backend_name}] shape {backend_output.shape} != " f"SDPA shape {expected_output.shape}" ) assert backend_output.dtype == expected_output.dtype, ( f"[{backend_name}] dtype {backend_output.dtype} != " f"SDPA dtype {expected_output.dtype}" ) assert torch.isfinite(backend_output).all(), ( f"[{backend_name}] produced non-finite values" ) # Check numerical similarity rtol = 1e-2 atol = 5e-1 max_diff = torch.max(torch.abs(backend_output - expected_output)).item() max_rel_diff = torch.max( torch.abs(backend_output - expected_output) / torch.abs(expected_output) ).item() all_close = torch.allclose( backend_output, expected_output, rtol=rtol, atol=atol ) assert all_close, ( f"[{backend_name}] output differs from SDPA baseline. " f"Max diff: {max_diff:.6f}, max rel diff: {max_rel_diff:.6f})" ) except AssertionError as e: failures.append(str(e)) # Report all failures at once if failures: # Create a summary for the single-line failure message backend_names = [] for f in failures: if "[AttentionBackendEnum." in f: backend_name = f.split("[")[1].split("]")[0] backend_names.append(backend_name) summary = f"{len(failures)} backend(s) failed: {', '.join(backend_names)}" detailed_msg = "\n".join(failures) pytest.fail(f"{summary}\n{detailed_msg}")
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/attention/test_mla_backends.py", "license": "Apache License 2.0", "lines": 961, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/entrypoints/constants.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Shared constants for vLLM entrypoints. """ # HTTP header limits for h11 parser # These constants help mitigate header abuse attacks H11_MAX_INCOMPLETE_EVENT_SIZE_DEFAULT = 4194304 # 4 MB H11_MAX_HEADER_COUNT_DEFAULT = 256 MCP_PREFIX = "mcp_"
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/entrypoints/constants.py", "license": "Apache License 2.0", "lines": 10, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/compile/test_decorator.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from torch import nn from vllm.compilation.counter import compilation_counter from vllm.compilation.decorators import ignore_torch_compile, support_torch_compile from vllm.config import ( CacheConfig, CompilationConfig, CompilationMode, CUDAGraphMode, VllmConfig, set_current_vllm_config, ) from vllm.forward_context import BatchDescriptor, set_forward_context from vllm.utils.torch_utils import is_torch_equal_or_newer # This import automatically registers `torch.ops.silly.attention` from . import silly_attention # noqa: F401 BATCH_SIZE = 32 MLP_SIZE = 128 @torch.inference_mode def run_model( vllm_config: VllmConfig, model: nn.Module, cudagraph_runtime_mode: CUDAGraphMode ): with set_forward_context({}, vllm_config=vllm_config): # warmup for the model with cudagraph_mode NONE model(torch.randn(BATCH_SIZE, MLP_SIZE).cuda()) # simulate cudagraphs capturing with set_forward_context( {}, vllm_config=vllm_config, cudagraph_runtime_mode=cudagraph_runtime_mode, batch_descriptor=BatchDescriptor( num_tokens=2, ), ): model(torch.randn(2, MLP_SIZE).cuda()) with set_forward_context( {}, vllm_config=vllm_config, cudagraph_runtime_mode=cudagraph_runtime_mode, batch_descriptor=BatchDescriptor( num_tokens=1, ), ): model(torch.randn(1, MLP_SIZE).cuda()) # simulate cudagraphs replay with set_forward_context( {}, vllm_config=vllm_config, cudagraph_runtime_mode=cudagraph_runtime_mode, batch_descriptor=BatchDescriptor( num_tokens=2, ), ): output = model(torch.randn(2, MLP_SIZE).cuda()) output = output.cpu() return output.cpu() @pytest.mark.parametrize("use_inductor_graph_partition", [True, False]) def test_ignore_torch_compile_decorator(use_inductor_graph_partition, monkeypatch): # disable compile cache so that we can count the number of compilations # appropriately monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", "1") if use_inductor_graph_partition and not is_torch_equal_or_newer("2.9.0.dev"): pytest.skip("inductor graph partition is only available in PyTorch 2.9+") # piecewise vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, splitting_ops=["silly::attention"], cudagraph_capture_sizes=[1, 2], use_inductor_graph_partition=use_inductor_graph_partition, ) ) cudagraph_runtime_mode = CUDAGraphMode.PIECEWISE expected_num_graphs_seen = 1 expected_num_cudagraph_captured = ( 4 # num_cudagraph_sizes * num cudagraphs to capture ) if use_inductor_graph_partition: expected_num_piecewise_graphs_seen = 1 expected_num_piecewise_capturable_graphs_seen = 1 expected_num_backend_compilations = 1 else: expected_num_piecewise_graphs_seen = 3 expected_num_piecewise_capturable_graphs_seen = 2 expected_num_backend_compilations = 2 @support_torch_compile class A(nn.Module): def __init__( self, *, vllm_config: VllmConfig, prefix: str = "", **kwargs ) -> None: super().__init__() def forward(self, x: torch.Tensor) -> torch.Tensor: x = x + x attn_output = torch.empty_like(x) torch.ops.silly.attention(x, x, x, attn_output) x = attn_output x = x * 3 return x @ignore_torch_compile class B(A): ... @support_torch_compile class C(B): ... with set_current_vllm_config(vllm_config): mod_A = A(vllm_config=vllm_config, prefix="").eval().cuda() # A has support_torch_compile with compilation_counter.expect( num_graphs_seen=expected_num_graphs_seen, num_piecewise_graphs_seen=expected_num_piecewise_graphs_seen, num_piecewise_capturable_graphs_seen=expected_num_piecewise_capturable_graphs_seen, num_backend_compilations=expected_num_backend_compilations, num_cudagraph_captured=expected_num_cudagraph_captured, ): run_model(vllm_config, mod_A, cudagraph_runtime_mode) with set_current_vllm_config(vllm_config): mod_B = B(vllm_config=vllm_config, prefix="").eval().cuda() # B's ignore_torch_compile should override A's support_torch_compile with compilation_counter.expect( num_graphs_seen=0, num_piecewise_graphs_seen=0, num_piecewise_capturable_graphs_seen=0, num_backend_compilations=0, num_cudagraph_captured=0, ): run_model(vllm_config, mod_B, cudagraph_runtime_mode) with set_current_vllm_config(vllm_config): mod_C = C(vllm_config=vllm_config, prefix="").eval().cuda() # C's support_torch_compile should override B's ignore_torch_compile with compilation_counter.expect( num_graphs_seen=expected_num_graphs_seen, num_piecewise_graphs_seen=expected_num_piecewise_graphs_seen, num_piecewise_capturable_graphs_seen=expected_num_piecewise_capturable_graphs_seen, num_backend_compilations=expected_num_backend_compilations, num_cudagraph_captured=expected_num_cudagraph_captured, ): run_model(vllm_config, mod_C, cudagraph_runtime_mode) # Only enable torch.compile if # vllm_config.cache_config.kv_sharing_fast_prefill=True @support_torch_compile( enable_if=lambda vllm_config: vllm_config.cache_config.kv_sharing_fast_prefill ) class B(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = "", **kwargs) -> None: super().__init__() def forward(self, x: torch.Tensor) -> torch.Tensor: x = x + x attn_output = torch.empty_like(x) torch.ops.silly.attention(x, x, x, attn_output) x = attn_output x = x + x return x # Only enable torch.compile if # vllm_config.cache_config.kv_sharing_fast_prefill=False @support_torch_compile( enable_if=lambda vllm_config: not vllm_config.cache_config.kv_sharing_fast_prefill ) class A(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = "", **kwargs) -> None: super().__init__() self.mod1 = B(vllm_config=vllm_config, prefix=prefix, **kwargs) self.mod2 = B(vllm_config=vllm_config, prefix=prefix, **kwargs) def forward(self, x: torch.Tensor) -> torch.Tensor: x = self.mod1(x) attn_output = torch.empty_like(x) torch.ops.silly.attention(x, x, x, attn_output) x = attn_output x = self.mod2(x) return x @pytest.mark.parametrize("use_inductor_graph_partition", [True, False]) def test_conditional_compile_enable_if(use_inductor_graph_partition, monkeypatch): # disable compile cache so that we can count the number of compilations # appropriately monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", "1") if use_inductor_graph_partition and not is_torch_equal_or_newer("2.9.0.dev"): pytest.skip("inductor graph partition is only available in PyTorch 2.9+") vllm_config = VllmConfig( cache_config=CacheConfig( kv_sharing_fast_prefill=True, ), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, splitting_ops=["silly::attention"], cudagraph_capture_sizes=[1, 2], use_inductor_graph_partition=use_inductor_graph_partition, ), ) cudagraph_runtime_mode = CUDAGraphMode.PIECEWISE with set_current_vllm_config(vllm_config): mod_A = A(vllm_config=vllm_config, prefix="").eval().cuda() if use_inductor_graph_partition: expected_num_piecewise_graphs_seen = 2 expected_num_piecewise_capturable_graphs_seen = 2 expected_num_backend_compilations = 2 else: expected_num_piecewise_graphs_seen = 6 expected_num_piecewise_capturable_graphs_seen = 4 expected_num_backend_compilations = 4 # A has support_torch_compile but enable_if fn returns False # enalbe_if will be True for B, so we expect mod1 and mod2 # to be compiled with compilation_counter.expect( num_graphs_seen=2, num_piecewise_graphs_seen=expected_num_piecewise_graphs_seen, # 3 piecewise graphs per instance of B() num_piecewise_capturable_graphs_seen=expected_num_piecewise_capturable_graphs_seen, num_backend_compilations=expected_num_backend_compilations, num_cudagraph_captured=8, # num_cudagraph_sizes * num cudagraphable graphs to capture ): run_model(vllm_config, mod_A, cudagraph_runtime_mode) # Set kv_sharing_fast_prefill=False # which will cause A to be compiled and B to not be compiled vllm_config = VllmConfig( cache_config=CacheConfig( kv_sharing_fast_prefill=False, ), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, splitting_ops=["silly::attention"], cudagraph_capture_sizes=[1, 2], use_inductor_graph_partition=use_inductor_graph_partition, ), ) with set_current_vllm_config(vllm_config): mod_A = A(vllm_config=vllm_config, prefix="").eval().cuda() if use_inductor_graph_partition: expected_num_piecewise_graphs_seen = 1 expected_num_piecewise_capturable_graphs_seen = 1 expected_num_backend_compilations = 1 else: # 3 attn ops and 4 non-attn ops expected_num_piecewise_graphs_seen = 7 expected_num_piecewise_capturable_graphs_seen = 4 expected_num_backend_compilations = 4 with compilation_counter.expect( num_graphs_seen=1, num_piecewise_graphs_seen=expected_num_piecewise_graphs_seen, # 3 attn ops and 4 non-attn ops num_piecewise_capturable_graphs_seen=expected_num_piecewise_capturable_graphs_seen, num_backend_compilations=expected_num_backend_compilations, num_cudagraph_captured=8, # num_cudagraph_sizes * num cudagraphable graphs to capture ): run_model(vllm_config, mod_A, cudagraph_runtime_mode)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/compile/test_decorator.py", "license": "Apache License 2.0", "lines": 246, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/models/ernie_mtp.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright 2025 The Baidu team. # Copyright 2023 The vLLM team. # Copyright 2022 EleutherAI and the HuggingFace Inc. team. All rights reserved. # # This code is based on EleutherAI's GPT-NeoX library and the GPT-NeoX # and OPT implementations in this library. It has been modified from its # original forms to accommodate minor architectural differences compared # to GPT-NeoX and OPT used by the Meta AI team that trained the model. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Inference-only Ernie-MTP model.""" from collections.abc import Iterable import torch import torch.nn as nn from transformers import PretrainedConfig from vllm.config import VllmConfig from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import default_weight_loader from vllm.sequence import IntermediateTensors from .llama import LlamaDecoderLayer from .utils import is_pp_missing_parameter, maybe_prefix class ErnieMultiTokenPredictorLayer(nn.Module): def __init__( self, vllm_config: VllmConfig, prefix: str, ) -> None: super().__init__() config = vllm_config.model_config.hf_config self.mtp_emb_norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.mtp_hidden_norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps) self.mtp_linear_proj = nn.Linear( config.hidden_size * 2, config.hidden_size, bias=False ) self.mtp_block = LlamaDecoderLayer(vllm_config, prefix) def forward( self, inputs_embeds: torch.Tensor, positions: torch.Tensor, previous_hidden_states: torch.Tensor, spec_step_index: int = 0, ) -> torch.Tensor: assert inputs_embeds is not None # masking inputs at position 0, as not needed by MTP inputs_embeds[positions == 0] = 0 inputs_embeds = self.mtp_emb_norm(inputs_embeds) previous_hidden_states = self.mtp_hidden_norm(previous_hidden_states) hidden_states = self.mtp_linear_proj( torch.cat([inputs_embeds, previous_hidden_states], dim=-1) ) hidden_states, residual = self.mtp_block( positions=positions, hidden_states=hidden_states, residual=None ) hidden_states = residual + hidden_states return hidden_states class ErnieMultiTokenPredictor(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config self.mtp_start_layer_idx = config.num_hidden_layers self.num_mtp_layers = config.num_nextn_predict_layers # to map the exact layer index from weights self.layers = torch.nn.ModuleDict( { str(idx): ErnieMultiTokenPredictorLayer( vllm_config, f"{prefix}.layers.{idx}", ) for idx in range( self.mtp_start_layer_idx, self.mtp_start_layer_idx + self.num_mtp_layers, ) } ) self.embed_tokens = VocabParallelEmbedding( config.vocab_size, config.hidden_size, ) self.logits_processor = LogitsProcessor(config.vocab_size) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor, positions: torch.Tensor, previous_hidden_states: torch.Tensor, inputs_embeds: torch.Tensor | None = None, spec_step_idx: int = 0, ) -> torch.Tensor: if inputs_embeds is None: inputs_embeds = self.embed_tokens(input_ids) return self.layers[str(self.mtp_start_layer_idx + spec_step_idx)]( inputs_embeds, positions, previous_hidden_states, spec_step_idx, ) def compute_logits( self, hidden_states: torch.Tensor, lm_head: ParallelLMHead, spec_step_idx: int = 0, ) -> torch.Tensor: self.layers[str(self.mtp_start_layer_idx + spec_step_idx)] logits = self.logits_processor(lm_head, hidden_states) return logits class ErnieMTP(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() self.config = vllm_config.model_config.hf_config self.model = ErnieMultiTokenPredictor( vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model") ) self.lm_head = ParallelLMHead( self.config.vocab_size, self.config.hidden_size, prefix=maybe_prefix(prefix, "lm_head"), ) if self.config.tie_word_embeddings: self.lm_head.weight = self.model.embed_tokens.weight def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, hidden_states: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, spec_step_idx: int = 0, ) -> torch.Tensor: assert spec_step_idx == 0, "ernie_mtp only support predict one token" hidden_states = self.model( input_ids, positions, hidden_states, inputs_embeds, spec_step_idx ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, spec_step_idx: int = 0, ) -> torch.Tensor | None: return self.model.compute_logits(hidden_states, self.lm_head, spec_step_idx) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ ("qkv_proj", "q_proj", "q"), ("qkv_proj", "k_proj", "k"), ("qkv_proj", "v_proj", "v"), ("gate_up_proj", "gate_proj", 0), ("gate_up_proj", "up_proj", 1), ] params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() for name, loaded_weight in weights: if self.config.tie_word_embeddings and name.endswith("lm_head.weight"): continue if "rotary_emb.inv_freq" in name: continue if "mtp" in name: name = self._rewrite_spec_layer_name(self.config, name) for param_name, weight_name, shard_id in stacked_params_mapping: # Skip non-stacked layers and experts (experts handled below). if weight_name not in name: continue if "mtp" not in name: continue # We have mlp.experts[0].gate_proj in the checkpoint. # Since we handle the experts below in expert_params_mapping, # we need to skip here BEFORE we update the name, otherwise # name will be updated to mlp.experts[0].gate_up_proj, which # will then be updated below in expert_params_mapping # for mlp.experts[0].gate_gate_up_proj, which breaks load. if ("mlp.experts." in name) and name not in params_dict: continue name = name.replace(weight_name, param_name) # Skip loading extra bias for GPTQ models. if ( name.endswith(".bias") or name.endswith("_bias") ) and name not in params_dict: continue # Skip layers on other devices. if is_pp_missing_parameter(name, self): continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: # Skip loading extra bias for GPTQ models. if ( name.endswith(".bias") or name.endswith("_bias") ) and name not in params_dict: continue # Skip layers on other devices. if is_pp_missing_parameter(name, self): continue # According to DeepSeek-V3 Technical Report, MTP modules # shares embedding layer. We only load the first weights. if "mtp_" not in name and ( "embed_tokens" not in name and "lm_head" not in name ): continue param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params def _rewrite_spec_layer_name(self, config: PretrainedConfig, name: str) -> str: """ Rewrite the weight name to match the format of the original model. """ spec_layer_weight_names = [ "embed_tokens", "mtp_emb_norm", "mtp_hidden_norm", "mtp_linear_proj", ] layer_idx = config.num_hidden_layers for weight_name in spec_layer_weight_names: if weight_name in name: name = name.replace( f"model.{weight_name}.0.", f"model.layers.{layer_idx}.{weight_name}.", ) return name name = name.replace( "model.mtp_block.0.", f"model.layers.{layer_idx}.mtp_block." ) return name
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/ernie_mtp.py", "license": "Apache License 2.0", "lines": 245, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/deepseek_eagle.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Iterable import torch import torch.nn as nn from vllm.compilation.decorators import support_torch_compile from vllm.config import VllmConfig from vllm.model_executor.layers.fused_moe import FusedMoE from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.logits_processor import LogitsProcessor from vllm.model_executor.layers.vocab_parallel_embedding import ( ParallelLMHead, VocabParallelEmbedding, ) from vllm.model_executor.model_loader.weight_utils import ( default_weight_loader, maybe_remap_kv_scale_name, ) from vllm.model_executor.models.deepseek_v2 import ( DeepseekV2DecoderLayer, DeepseekV3ForCausalLM, ) from .utils import AutoWeightsLoader, maybe_prefix, process_eagle_weight @support_torch_compile class DeepseekV2Model(nn.Module): def __init__( self, *, vllm_config: VllmConfig, prefix: str = "", start_layer_id: int = 0, ) -> None: super().__init__() self.config = vllm_config.speculative_config.draft_model_config.hf_config quant_config = vllm_config.quant_config self.vocab_size = self.config.vocab_size self.embed_tokens = VocabParallelEmbedding( self.config.vocab_size, self.config.hidden_size, quant_config=quant_config, prefix=maybe_prefix(prefix, "embed_tokens"), ) self.layers = nn.ModuleList( [ DeepseekV2DecoderLayer( vllm_config, prefix=maybe_prefix(prefix, f"layers.{i + start_layer_id}"), config=self.config, ) for i in range(self.config.num_hidden_layers) ] ) self.fc = nn.Linear( self.config.model.hidden_size * 2, self.config.model.hidden_size, bias=False, ) self.enorm = RMSNorm(self.config.hidden_size, eps=self.config.rms_norm_eps) self.hnorm = RMSNorm(self.config.hidden_size, eps=self.config.rms_norm_eps) self.norm = RMSNorm(self.config.hidden_size, eps=self.config.rms_norm_eps) def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.embed_tokens(input_ids) def forward( self, input_ids: torch.Tensor, positions: torch.Tensor, hidden_states: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: input_embeds = self.embed_tokens(input_ids) inputs = torch.cat( [self.enorm(input_embeds), self.hnorm(hidden_states)], dim=-1 ) hidden_states = self.fc(inputs) residual = None for layer in self.layers: hidden_states, residual = layer( positions, hidden_states, residual, ) hidden_states, _ = self.norm(hidden_states, residual) return hidden_states, hidden_states def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ # (param_name, shard_name, shard_id) ("gate_up_proj", "gate_proj", 0), ("gate_up_proj", "up_proj", 1), ("fused_qkv_a_proj", "q_a_proj", 0), ("fused_qkv_a_proj", "kv_a_proj_with_mqa", 1), ] # Params for weights, fp8 weight scales, fp8 activation scales # (param_name, weight_name, expert_id, shard_id) expert_params_mapping = FusedMoE.make_expert_params_mapping( self, ckpt_gate_proj_name="gate_proj", ckpt_down_proj_name="down_proj", ckpt_up_proj_name="up_proj", num_experts=self.config.n_routed_experts, ) params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() for name, loaded_weight in weights: if "rotary_emb.inv_freq" in name: continue for param_name, weight_name, shard_id in stacked_params_mapping: # Skip non-stacked layers and experts (experts handled below). if weight_name not in name: continue # We have mlp.experts[0].gate_proj in the checkpoint. # Since we handle the experts below in expert_params_mapping, # we need to skip here BEFORE we update the name, otherwise # name will be updated to mlp.experts[0].gate_up_proj, which # will then be updated below in expert_params_mapping # for mlp.experts[0].gate_gate_up_proj, which breaks load. if ("mlp.experts." in name) and name not in params_dict: continue name_mapped = name.replace(weight_name, param_name) # QKV fusion is optional, fall back to normal # weight loading if it's not enabled # if go with fusion option, then update name if ( param_name == "fused_qkv_a_proj" ) and name_mapped not in params_dict: continue else: name = name_mapped # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: for mapping in expert_params_mapping: param_name, weight_name, expert_id, shard_id = mapping if weight_name not in name: continue name = name.replace(weight_name, param_name) param = params_dict[name] weight_loader = param.weight_loader weight_loader( param, loaded_weight, name, shard_id=shard_id, expert_id=expert_id, ) break else: # Skip loading extra bias for GPTQ models. if name.endswith(".bias") and name not in params_dict: continue # Remapping the name of FP8 kv-scale. name = maybe_remap_kv_scale_name(name, params_dict) if name is None: continue param = params_dict[name] weight_loader = getattr( param, "weight_loader", default_weight_loader ) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params class EagleDeepseekV3ForCausalLM(DeepseekV3ForCausalLM): def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): nn.Module.__init__(self) self.config = vllm_config.speculative_config.draft_model_config.hf_config quant_config = vllm_config.quant_config target_layer_num = vllm_config.model_config.get_num_layers( vllm_config.parallel_config ) self.model = DeepseekV2Model( vllm_config=vllm_config, prefix="model", start_layer_id=target_layer_num ) self.lm_head = ParallelLMHead( self.config.vocab_size, self.config.hidden_size, quant_config=quant_config, prefix=maybe_prefix(prefix, "lm_head"), ) logit_scale = getattr(self.config, "logit_scale", 1.0) self.logits_processor = LogitsProcessor( self.config.vocab_size, scale=logit_scale ) # Set MoE hyperparameters self.num_moe_layers = self.config.num_hidden_layers self.set_moe_parameters() def embed_input_ids(self, input_ids: torch.Tensor) -> torch.Tensor: return self.model.embed_input_ids(input_ids) def forward( self, input_ids: torch.Tensor, positions: torch.Tensor, hidden_states: torch.Tensor, inputs_embeds: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: if inputs_embeds is not None: raise NotImplementedError( f"{type(self).__name__} does not support multimodal inputs yet." ) return self.model(input_ids, positions, hidden_states) def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: logits = self.logits_processor(self.lm_head, hidden_states) return logits def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]): def transform(inputs): name, loaded_weight = inputs if "lm_head" not in name: name = "model." + name process_eagle_weight(self, name) return name, loaded_weight loader = AutoWeightsLoader( self, skip_prefixes=None, ) loader.load_weights(map(transform, weights))
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/deepseek_eagle.py", "license": "Apache License 2.0", "lines": 220, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/evals/gsm8k/gsm8k_eval.py
#!/usr/bin/env python3 # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Isolated GSM8K evaluation script for vLLM serve endpoint. """ import argparse import ast import asyncio import json import os import time from collections.abc import Generator import aiohttp import numpy as np import regex as re import requests from tqdm.asyncio import tqdm INVALID = -9999999 def download_and_cache_file(url: str, filename: str | None = None) -> str: """Download and cache a file from a URL.""" if filename is None: filename = os.path.join("/tmp", url.split("/")[-1]) if os.path.exists(filename): return filename print(f"Downloading from {url} to {filename}") response = requests.get(url, stream=True) response.raise_for_status() with open(filename, "wb") as f: for chunk in response.iter_content(chunk_size=1024): f.write(chunk) return filename def load_gsm8k_data() -> tuple[list[dict], list[dict]]: """Load GSM8K train and test data""" train_url = "https://raw.githubusercontent.com/openai/grade-school-math/master/grade_school_math/data/train.jsonl" test_url = "https://raw.githubusercontent.com/openai/grade-school-math/master/grade_school_math/data/test.jsonl" train_file = download_and_cache_file(train_url) test_file = download_and_cache_file(test_url) train_data = list(read_jsonl(train_file)) test_data = list(read_jsonl(test_file)) return train_data, test_data def read_jsonl(filename: str) -> Generator[dict, None, None]: """Read a JSONL file.""" with open(filename) as fin: for line in fin: if not line.startswith("#"): yield json.loads(line) def get_answer_value(answer_str: str) -> int: """Extract the numerical answer from the response.""" answer_str = answer_str.replace(",", "") numbers = re.findall(r"\d+", answer_str) if len(numbers) < 1: return INVALID try: return ast.literal_eval(numbers[-1]) except SyntaxError: return INVALID async def call_vllm_api( session: aiohttp.ClientSession, prompt: str, temperature: float, max_tokens: int, stop: list[str] | None = None, url: str | None = None, seed: int | None = None, ) -> tuple[str, int]: """Call vLLM's OpenAI-compatible completions endpoint. Returns: Tuple of (response_text, completion_tokens) """ data = { "prompt": prompt, "temperature": temperature, "max_tokens": max_tokens, "stop": stop, } if seed is not None: data["seed"] = seed try: async with session.post(f"{url}/v1/completions", json=data) as response: response.raise_for_status() result = await response.json() text = result["choices"][0]["text"] completion_tokens = result.get("usage", {}).get("completion_tokens", 0) return text, completion_tokens except Exception as e: print(f"Error calling vLLM API: {e}") return "", 0 def _build_gsm8k_prompts( num_questions: int = 1319, num_shots: int = 5, ) -> tuple[list[str], list[int]]: """Build few-shot GSM8K completion prompts and ground-truth labels.""" if num_questions == 0: return [], [] train_data, test_data = load_gsm8k_data() num_questions = min(num_questions, len(test_data)) few_shot_examples = "" for i in range(num_shots): few_shot_examples += ( f"Question: {train_data[i]['question']}\n" f"Answer: {train_data[i]['answer']}\n\n" ) prompts = [] labels = [] for i in range(num_questions): prompts.append( few_shot_examples + f"Question: {test_data[i]['question']}\nAnswer:" ) labels.append(get_answer_value(test_data[i]["answer"])) assert all(label != INVALID for label in labels), "Some labels are invalid" return prompts, labels def _score_gsm8k( states: list[str], output_tokens: list[int], labels: list[int], num_shots: int, max_tokens: int, latency: float, ) -> dict[str, float | int]: """Score GSM8K responses and return a results dict.""" num_questions = len(labels) preds = [get_answer_value(state) for state in states] accuracy = np.mean(np.array(preds) == np.array(labels)) invalid_rate = np.mean(np.array(preds) == INVALID) total_output_tokens = sum(output_tokens) tokens_per_second = total_output_tokens / latency if latency > 0 else 0.0 return { "accuracy": accuracy, "invalid_rate": invalid_rate, "latency": latency, "questions_per_second": num_questions / latency if latency > 0 else 0.0, "total_output_tokens": total_output_tokens, "tokens_per_second": tokens_per_second, "num_questions": num_questions, "num_shots": num_shots, "max_tokens": max_tokens, "timestamp": time.time(), } def evaluate_gsm8k( num_questions: int = 1319, num_shots: int = 5, max_tokens: int = 256, host: str = "http://127.0.0.1", port: int = 8000, temperature: float = 0.0, seed: int | None = 42, ) -> dict[str, float | int]: """ Evaluate GSM8K accuracy using vLLM serve endpoint. Returns dict with accuracy, invalid_rate, latency, etc. """ base_url = f"{host}:{port}" prompts, labels = _build_gsm8k_prompts(num_questions, num_shots) num_questions = len(prompts) async def run_async_evaluation(): states: list[str] = [""] * num_questions output_tokens: list[int] = [0] * num_questions async def get_answer(session: aiohttp.ClientSession, i: int) -> tuple[str, int]: answer, tokens = await call_vllm_api( session=session, prompt=prompts[i], temperature=temperature, max_tokens=max_tokens, stop=["Question", "Assistant:", "<|separator|>"], url=base_url, seed=seed, ) states[i] = answer output_tokens[i] = tokens return answer, tokens async with aiohttp.ClientSession( timeout=aiohttp.ClientTimeout(total=600) ) as session: tasks = [get_answer(session, i) for i in range(num_questions)] await tqdm.gather(*tasks, desc="Evaluating") return states, output_tokens print(f"Running GSM8K evaluation: {num_questions} questions, {num_shots}-shot") tic = time.perf_counter() states, output_tokens = asyncio.run(run_async_evaluation()) latency = time.perf_counter() - tic return _score_gsm8k(states, output_tokens, labels, num_shots, max_tokens, latency) def evaluate_gsm8k_offline( llm, num_questions: int = 1319, num_shots: int = 5, max_tokens: int = 256, temperature: float = 0.0, ) -> dict[str, float | int]: """Evaluate GSM8K accuracy using an offline vllm.LLM object. Same prompts and scoring as evaluate_gsm8k(), but runs generation directly via llm.generate() instead of calling a server over HTTP. """ from vllm import SamplingParams prompts, labels = _build_gsm8k_prompts(num_questions, num_shots) sampling_params = SamplingParams( temperature=temperature, max_tokens=max_tokens, stop=["Question", "Assistant:", "<|separator|>"], ) print( f"Running offline GSM8K evaluation: {len(prompts)} questions, {num_shots}-shot" ) tic = time.perf_counter() outputs = llm.generate(prompts, sampling_params) latency = time.perf_counter() - tic states = [o.outputs[0].text for o in outputs] output_tokens = [len(o.outputs[0].token_ids) for o in outputs] return _score_gsm8k(states, output_tokens, labels, num_shots, max_tokens, latency) def main() -> None: parser = argparse.ArgumentParser(description="GSM8K evaluation for vLLM serve") parser.add_argument( "--num-shots", type=int, default=5, help="Number of few-shot examples" ) parser.add_argument( "--num-questions", type=int, default=1319, help="Number of questions to evaluate", ) parser.add_argument( "--max-tokens", type=int, default=256, help="Max tokens for generation" ) parser.add_argument("--host", type=str, default="http://127.0.0.1", help="Host URL") parser.add_argument("--port", type=int, default=8000, help="Port number") parser.add_argument( "--temperature", type=float, default=0.0, help="Temperature for generation" ) parser.add_argument( "--seed", type=int, default=42, help="Random seed for reproducibility" ) parser.add_argument("--save-results", type=str, help="Save results to JSON file") args = parser.parse_args() result = evaluate_gsm8k( num_questions=args.num_questions, num_shots=args.num_shots, max_tokens=args.max_tokens, host=args.host, port=args.port, temperature=args.temperature, seed=args.seed, ) # Print results to terminal print("\nResults:") print(f"Accuracy: {result['accuracy']:.3f}") print(f"Invalid responses: {result['invalid_rate']:.3f}") print(f"Total latency: {result['latency']:.3f} s") print(f"Questions per second: {result['questions_per_second']:.3f}") print(f"Total output tokens: {result['total_output_tokens']}") print(f"Output tokens per second: {result['tokens_per_second']:.3f}") # Optional file saving if args.save_results: with open(args.save_results, "w") as f: json.dump(result, f, indent=2) print(f"Results saved to {args.save_results}") if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "tests/evals/gsm8k/gsm8k_eval.py", "license": "Apache License 2.0", "lines": 257, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/evals/gsm8k/test_gsm8k_correctness.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ GSM8K evaluation using vLLM server and isolated GSM8K script. Replacement for lm-eval-harness with better performance and control. Usage: pytest -s -v tests/evals/gsm8k/test_gsm8k_correctness.py \ --config-list-file=configs/models-small.txt """ import shlex import pytest import yaml from tests.utils import RemoteOpenAIServer from vllm.platforms import current_platform from .gsm8k_eval import evaluate_gsm8k TOL = 0.08 # Absolute tolerance for accuracy comparison def run_gsm8k_eval(eval_config: dict, server_url: str) -> dict: """Run GSM8K evaluation using our isolated script.""" # Extract host and port from server URL if "://" in server_url: server_url = server_url.split("://")[1] host_port = server_url.split("/")[0] # Remove path if present if ":" in host_port: host, p = host_port.split(":") port = int(p) else: host = host_port port = 8000 # Add http:// prefix if not present if not host.startswith("http"): host = f"http://{host}" # Run GSM8K evaluation results = evaluate_gsm8k( num_questions=eval_config["num_questions"], num_shots=eval_config["num_fewshot"], host=host, port=port, ) return results def test_gsm8k_correctness(config_filename): """Test GSM8K correctness for a given model configuration.""" eval_config = yaml.safe_load(config_filename.read_text(encoding="utf-8")) if ( not current_platform.is_cuda() and "Qwen3-30B-A3B-MXFP4A16" in eval_config["model_name"] ): pytest.skip( "Skipping Qwen3-30B-A3B-MXFP4A16 on non-CUDA platforms. " "Marlin kernels are not supported." ) # Parse server arguments from config (use shlex to handle quoted strings) server_args_str = eval_config.get("server_args", "") server_args = shlex.split(server_args_str) if server_args_str else [] # Add standard server arguments server_args.extend( [ "--trust-remote-code", "--disable-uvicorn-access-log", ] ) env_dict = eval_config.get("env", None) print(f"Starting GSM8K evaluation for model: {eval_config['model_name']}") print(f"Expected metric threshold: {eval_config['accuracy_threshold']}") print(f"Number of questions: {eval_config['num_questions']}") print(f"Number of few-shot examples: {eval_config['num_fewshot']}") print(f"Server args: {' '.join(server_args)}") print(f"Environment variables: {env_dict}") # Launch server and run evaluation with RemoteOpenAIServer( eval_config["model_name"], server_args, env_dict=env_dict, max_wait_seconds=eval_config.get("startup_max_wait_seconds", 600), ) as remote_server: server_url = remote_server.url_for("v1") print(f"Server started at: {server_url}") results = run_gsm8k_eval(eval_config, server_url) measured_metric = results["accuracy"] expected_metric = eval_config["accuracy_threshold"] print(f"GSM8K Results for {eval_config['model_name']}:") print(f" Measured metric: {measured_metric:.4f}") print(f" Expected metric: {expected_metric:.4f}") print(f" Tolerance: {TOL:.4f}") print(f" Questions: {results['num_questions']}") print(f" Invalid rate: {results['invalid_rate']:.3f}") print(f" Latency: {results['latency']:.1f}s") print(f" QPS: {results['questions_per_second']:.1f}") # Verify metric is within tolerance assert measured_metric >= expected_metric - TOL, ( f"GSM8K metric too low: {measured_metric:.4f} < " f"{expected_metric:.4f} - {TOL:.4f} = {expected_metric - TOL:.4f}" ) print(f"✅ GSM8K test passed for {eval_config['model_name']}")
{ "repo_id": "vllm-project/vllm", "file_path": "tests/evals/gsm8k/test_gsm8k_correctness.py", "license": "Apache License 2.0", "lines": 93, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/kernels/moe/test_flashinfer.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass import pytest import torch import vllm.model_executor.layers.fused_moe.modular_kernel as mk from vllm.config import ParallelConfig, VllmConfig, set_current_vllm_config from vllm.model_executor.layers.fused_moe.activation import MoEActivation from vllm.model_executor.layers.fused_moe.config import ( FusedMoEConfig, FusedMoEParallelConfig, FusedMoEQuantConfig, RoutingMethodType, fp8_w8a8_moe_quant_config, ) from vllm.model_executor.layers.fused_moe.flashinfer_cutlass_moe import ( FlashInferExperts, ) from vllm.model_executor.layers.fused_moe.fused_moe import fused_experts from vllm.model_executor.layers.fused_moe.prepare_finalize import ( MoEPrepareAndFinalizeNoEP, ) from vllm.model_executor.layers.quantization.utils.flashinfer_utils import ( apply_fi_trtllm_fp8_per_tensor_moe, register_scales_for_trtllm_fp8_per_tensor_moe, rotate_weights_for_fi_trtllm_fp8_per_tensor_moe, swap_w13_to_w31, ) from vllm.model_executor.layers.quantization.utils.fp8_utils import input_to_float8 from vllm.model_executor.models.llama4 import Llama4MoE from vllm.platforms import current_platform from vllm.utils.torch_utils import set_random_seed try: from vllm.utils.flashinfer import has_flashinfer_cutlass_fused_moe except ImportError: if current_platform.is_rocm(): pytest.skip( "flashinfer not supported for vLLM on ROCm", allow_module_level=True ) if not has_flashinfer_cutlass_fused_moe() or not current_platform.has_device_capability( 90 ): pytest.skip( "Supported for sm >= 90", allow_module_level=True, ) NUM_EXPERTS = [16] TOP_KS = [1] MNK_FACTORS = [ (256, 8192, 5120), (127, 4096, 5120), (10, 8192, 5120), (10, 4096, 5120), (1, 8192, 5120), (1, 4096, 5120), ] vllm_config = VllmConfig(parallel_config=ParallelConfig(pipeline_parallel_size=1)) def quant_fp8_per_tensor_batches(a): num_batches = a.size(0) a_quant = [] a_scales = [] for i in range(num_batches): a_fp8, a_global_sf = input_to_float8(a[i]) if a_global_sf.numel() == 1: a_global_sf = a_global_sf.view(1, 1) a_quant.append(a_fp8) a_scales.append(a_global_sf) result_a_quant = torch.stack(a_quant) result_a_scales = torch.stack(a_scales) return result_a_quant, result_a_scales def check_accuracy(ref_output, actual_output, atol=0.1, rtol=0.85, percent=0.925): close = torch.isclose(ref_output, actual_output, atol=atol, rtol=rtol) match_ratio = close.float().mean() assert match_ratio >= percent, ( f"Match ratio {match_ratio:.4f} is below the threshold {percent:.4f}" ) mismatch_percent = 1.0 - match_ratio.item() assert mismatch_percent <= 1 - percent, ( f"Mismatch percentage {mismatch_percent:.4f} is above the threshold " f"{1 - percent:.4f}" ) @dataclass class TestData: hidden_states: torch.Tensor w13_quantized: torch.Tensor w2_quantized: torch.Tensor a1_scale: torch.Tensor a2_scale: torch.Tensor w13_weight_scale: torch.Tensor w2_weight_scale: torch.Tensor layer: torch.nn.Module @staticmethod def make_moe_tensors_8bit( m: int, k: int, n: int, e: int, is_trtllm: bool, activation: MoEActivation = MoEActivation.SILU, ) -> "TestData": is_gated = activation.is_gated hidden_states = torch.randn((m, k), device="cuda", dtype=torch.bfloat16) / 10 w13 = ( torch.randn( (e, (2 * n) if is_gated else n, k), device="cuda", dtype=torch.bfloat16 ) / 10 ) w2 = torch.randn((e, k, n), device="cuda", dtype=torch.bfloat16) / 10 # Scale to fp8 _, a1_scale = input_to_float8(hidden_states) a2_scale = torch.scalar_tensor(1.0).to(device="cuda").to(dtype=torch.float32) w13_quantized, w13_weight_scale = quant_fp8_per_tensor_batches(w13) w2_quantized, w2_weight_scale = quant_fp8_per_tensor_batches(w2) layer = torch.nn.Module() layer.orig_dtype = torch.bfloat16 layer.w13_weight = w13_quantized.clone() layer.w2_weight = w2_quantized.clone() layer.w13_input_scale = a1_scale layer.w2_input_scale = a2_scale layer.w13_weight_scale = w13_weight_scale layer.w2_weight_scale = w2_weight_scale layer.activation = activation # Setup dummy config. layer.moe_parallel_config = mk.FusedMoEParallelConfig.make_no_parallel() # flashinfer expects swapped rows for w13 if is_gated: layer.w13_weight.data = swap_w13_to_w31(layer.w13_weight.data) if is_trtllm: rotate_weights_for_fi_trtllm_fp8_per_tensor_moe( layer.w13_weight, layer.w2_weight, is_gated ) register_scales_for_trtllm_fp8_per_tensor_moe( layer, layer.w13_weight_scale, layer.w13_input_scale, layer.w2_weight_scale, layer.w2_input_scale, ) layer.custom_routing_function = Llama4MoE.custom_routing_function layer.routing_method_type = RoutingMethodType.Llama4 layer.renormalize = False layer.intermediate_size_per_partition = n layer.ep_rank = 0 layer.local_num_experts = e return TestData( hidden_states=hidden_states, w13_quantized=w13_quantized, w2_quantized=w2_quantized, a1_scale=a1_scale, a2_scale=a2_scale, w13_weight_scale=w13_weight_scale, w2_weight_scale=w2_weight_scale, layer=layer, ) @pytest.mark.parametrize("m,n,k", MNK_FACTORS) @pytest.mark.parametrize("e", NUM_EXPERTS) @pytest.mark.parametrize("topk", TOP_KS) @pytest.mark.parametrize("activation", [MoEActivation.SILU, MoEActivation.RELU2_NO_MUL]) def test_flashinfer_per_tensor_moe_fp8_no_graph( m: int, n: int, k: int, e: int, topk: int, activation: MoEActivation, monkeypatch, ): if not current_platform.has_device_capability(100): pytest.skip("Test is only supported for sm >= 100") set_random_seed(7) monkeypatch.setenv("VLLM_FUSED_MOE_CHUNK_SIZE", "8192") with set_current_vllm_config(vllm_config): td = TestData.make_moe_tensors_8bit( m, k, n, e, is_trtllm=True, activation=activation ) score = torch.randn((m, e), device="cuda", dtype=torch.bfloat16) topk_weights, topk_ids = Llama4MoE.custom_routing_function( hidden_states=td.hidden_states, gating_output=score, topk=topk, renormalize=False, ) quant_config = fp8_w8a8_moe_quant_config( w1_scale=td.w13_weight_scale, w2_scale=td.w2_weight_scale, a1_scale=td.a1_scale, a2_scale=td.a2_scale, per_act_token_quant=False, ) output = fused_experts( td.hidden_states, td.w13_quantized, td.w2_quantized, topk_weights=topk_weights, topk_ids=topk_ids, inplace=False, activation=activation, global_num_experts=e, expert_map=None, apply_router_weight_on_input=True, quant_config=quant_config, ) flashinfer_output = apply_fi_trtllm_fp8_per_tensor_moe( layer=td.layer, hidden_states=td.hidden_states, router_logits=score, routing_bias=None, global_num_experts=e, top_k=topk, num_expert_group=None, topk_group=None, apply_router_weight_on_input=True, ) check_accuracy( ref_output=output, actual_output=flashinfer_output, atol=0.1, rtol=0.85, percent=0.925, ) @pytest.mark.parametrize("m,n,k", MNK_FACTORS) @pytest.mark.parametrize("e", NUM_EXPERTS) @pytest.mark.parametrize("topk", TOP_KS) @pytest.mark.parametrize("activation", [MoEActivation.SILU, MoEActivation.RELU2_NO_MUL]) def test_flashinfer_cutlass_moe_fp8_no_graph( m: int, n: int, k: int, e: int, topk: int, activation: MoEActivation, monkeypatch, workspace_init, ): set_random_seed(7) monkeypatch.setenv("VLLM_FUSED_MOE_CHUNK_SIZE", "8192") with set_current_vllm_config(vllm_config): td = TestData.make_moe_tensors_8bit( m, k, n, e, is_trtllm=False, activation=activation ) score = torch.randn((m, e), device="cuda", dtype=torch.bfloat16) topk_weights, topk_ids = Llama4MoE.custom_routing_function( hidden_states=td.hidden_states, gating_output=score, topk=topk, renormalize=False, ) quant_config = fp8_w8a8_moe_quant_config( w1_scale=td.w13_weight_scale, g1_alphas=(td.w13_weight_scale * td.a1_scale).squeeze(), w2_scale=td.w2_weight_scale, g2_alphas=(td.w2_weight_scale * td.a2_scale).squeeze(), a1_scale=td.a1_scale, a1_gscale=td.a1_scale, a2_scale=td.a2_scale, a2_gscale=1.0 / td.a2_scale, per_act_token_quant=False, ) output = fused_experts( td.hidden_states, td.w13_quantized, td.w2_quantized, topk_weights=topk_weights, topk_ids=topk_ids, inplace=False, activation=activation, global_num_experts=e, expert_map=None, apply_router_weight_on_input=True, quant_config=quant_config, ) td.layer.dp_size = 1 def get_fused_moe_quant_config(n: torch.nn.Module) -> FusedMoEQuantConfig: return quant_config td.layer.get_fused_moe_quant_config = get_fused_moe_quant_config td.layer.quant_method = td.layer moe_config = FusedMoEConfig( num_experts=e, experts_per_token=topk, hidden_dim=k, intermediate_size_per_partition=n, num_local_experts=e, num_logical_experts=e, activation=activation, device="cuda", moe_parallel_config=FusedMoEParallelConfig.make_no_parallel(), in_dtype=torch.bfloat16, is_act_and_mul=activation.is_gated, routing_method=RoutingMethodType.TopK, ) kernel = mk.FusedMoEModularKernel( MoEPrepareAndFinalizeNoEP(), FlashInferExperts( moe_config=moe_config, quant_config=quant_config, ), inplace=False, ) flashinfer_cutlass_output = kernel( td.hidden_states, td.layer.w13_weight, td.layer.w2_weight, topk_weights, topk_ids, activation=activation, global_num_experts=e, expert_map=None, apply_router_weight_on_input=True, ) check_accuracy( ref_output=output, actual_output=flashinfer_cutlass_output, atol=0.1, rtol=0.85, percent=0.925, ) @pytest.mark.parametrize( "num_experts,intermediate,hidden", [ (8, 2048, 1536), (64, 4096, 4096), ], ) def test_convert_moe_weights_to_flashinfer_trtllm_block_layout( num_experts, intermediate, hidden ): from vllm.model_executor.layers.quantization.utils.flashinfer_utils import ( convert_moe_weights_to_flashinfer_trtllm_block_layout, ) w13 = torch.randn( (num_experts, 2 * intermediate, hidden), dtype=torch.bfloat16, device="cuda" ) w2 = torch.randn( (num_experts, hidden, intermediate), dtype=torch.bfloat16, device="cuda" ) cache: dict[torch.Size, torch.Tensor] = {} w13_converted, w2_converted = convert_moe_weights_to_flashinfer_trtllm_block_layout( cache, w13, w2 ) assert w13_converted.ndim == 4, ( f"Expected 4D tensor, got shape {w13_converted.shape}" ) assert w2_converted.ndim == 4, f"Expected 4D tensor, got shape {w2_converted.shape}" assert w13_converted.numel() == w13.numel(), "W13 element count should be preserved" assert w2_converted.numel() == w2.numel(), "W2 element count should be preserved" assert w13_converted.dtype == torch.bfloat16 assert w2_converted.dtype == torch.bfloat16 assert w13_converted.shape[0] == num_experts assert w2_converted.shape[0] == num_experts
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/moe/test_flashinfer.py", "license": "Apache License 2.0", "lines": 350, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/entrypoints/openai/test_return_token_ids.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from vllm.tokenizers import get_tokenizer from ...utils import RemoteOpenAIServer MODEL_NAME = "Qwen/Qwen2.5-1.5B-Instruct" @pytest.fixture(scope="module") def server(): args = [ "--max-model-len", "2048", "--max-num-seqs", "128", "--enable-auto-tool-choice", "--tool-call-parser", "hermes", "--enforce-eager", ] with RemoteOpenAIServer(MODEL_NAME, args) as remote_server: yield remote_server @pytest.mark.asyncio @pytest.mark.parametrize("return_token_ids", [True, False, None]) async def test_basic_completion_with_emoji(server, return_token_ids: bool | None): """Test basic completion with emoji to verify token_ids field.""" extra_body = None if return_token_ids is not None: extra_body = {"return_token_ids": return_token_ids} async with server.get_async_client() as client: # Test with return_token_ids enabled completion = await client.completions.create( model=MODEL_NAME, prompt="Complete this sentence with emojis: I love coding 🚀", max_tokens=10, temperature=0, logprobs=1, extra_body=extra_body, ) # Check the raw response to see the structure completion_dict = completion.model_dump() # Verify prompt_token_ids field is present in the completion response assert "prompt_token_ids" in completion_dict["choices"][0] if not return_token_ids: # If return_token_ids is False, token_ids should not be present assert completion_dict["choices"][0].get("token_ids") is None assert completion_dict["choices"][0].get("prompt_token_ids") is None # Skip further checks return assert isinstance(completion.choices[0].prompt_token_ids, list) # Check against the expected prompt token IDs tokenizer = get_tokenizer(tokenizer_name=MODEL_NAME) encoded_tokens = tokenizer.encode( "Complete this sentence with emojis: I love coding 🚀" ) # Check that encoded_tokens is a subsequence of prompt_token_ids assert any( completion.choices[0].prompt_token_ids[i : i + len(encoded_tokens)] == encoded_tokens for i in range( len(completion.choices[0].prompt_token_ids) - len(encoded_tokens) + 1 ) ) # Verify token_ids field is present in the choice assert completion.choices[0].token_ids is not None assert isinstance(completion.choices[0].token_ids, list) assert len(completion.choices[0].token_ids) > 0 # Verify decoding works correctly decoded_text = tokenizer.decode(completion.choices[0].token_ids) # The decoded text should contain a <|im_end|> at the end assert decoded_text.startswith(completion.choices[0].text) # Test without return_token_ids (should be None) completion_without = await client.completions.create( model=MODEL_NAME, prompt="Complete this sentence with emojis: I love coding 🚀", max_tokens=10, temperature=0, logprobs=1, extra_body={"return_token_ids": False}, ) completion_without_dict = completion_without.model_dump() assert completion_without_dict["choices"][0].get("token_ids") is None assert completion_without_dict.get("prompt_token_ids") is None @pytest.mark.asyncio async def test_chat_completion_with_tool_use(server): """Test chat completion with tool use (get_weather function).""" tools = [ { "type": "function", "function": { "name": "get_weather", "description": "Get the current weather in a given location", "parameters": { "type": "object", "properties": { "location": { "type": "string", "description": "The city and state, e.g. San Francisco, CA", }, "unit": { "type": "string", "enum": ["celsius", "fahrenheit"], "description": "The unit of temperature", }, }, "required": ["location"], }, }, } ] async with server.get_async_client() as client: # Test with return_token_ids enabled response = await client.chat.completions.create( model=MODEL_NAME, messages=[ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": "What's the weather like in Paris?"}, ], tools=tools, tool_choice="auto", max_tokens=100, temperature=0, logprobs=True, extra_body={"return_token_ids": True}, ) # Verify token_ids field is present in choices assert response.choices[0].token_ids is not None assert isinstance(response.choices[0].token_ids, list) # Verify prompt_token_ids field is present assert response.prompt_token_ids is not None assert isinstance(response.prompt_token_ids, list) # Verify the prompt texts and response texts tokenizer = get_tokenizer(tokenizer_name=MODEL_NAME) prompt_text = tokenizer.decode(response.prompt_token_ids) assert prompt_text.startswith( "<|im_start|>system\nYou are a helpful assistant." ) assert prompt_text.endswith( "What's the weather like in Paris?<|im_end|>\n<|im_start|>assistant\n" ) response_text = tokenizer.decode(response.choices[0].token_ids) assert response_text.startswith('<tool_call>\n{"name": "get_weather"') assert response_text.endswith("</tool_call><|im_end|>") # If tool call was made, verify the response structure if response.choices[0].message.tool_calls: assert len(response.choices[0].message.tool_calls) > 0 tool_call = response.choices[0].message.tool_calls[0] assert tool_call.function.name == "get_weather" # Test without return_token_ids response_without = await client.chat.completions.create( model=MODEL_NAME, messages=[ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": "What's the weather like in Paris?"}, ], tools=tools, tool_choice="auto", max_tokens=100, temperature=0, logprobs=True, extra_body={"return_token_ids": False}, ) assert response_without.choices[0].token_ids is None assert response_without.prompt_token_ids is None @pytest.mark.asyncio async def test_comparison_with_prompt_logprobs_and_logprobs(server): """ Test that token_ids align with prompt_logprobs and logprobs when return_tokens_as_token_ids is enabled. """ async with server.get_async_client() as client: # Test with both return_token_ids and return_tokens_as_token_ids enabled completion = await client.completions.create( model=MODEL_NAME, prompt="Hello, world! How are you today?", max_tokens=20, temperature=0, echo=True, logprobs=1, extra_body={ "return_token_ids": True, "return_tokens_as_token_ids": True, "prompt_logprobs": 1, }, ) # Verify all fields are present assert completion.choices[0].token_ids is not None assert completion.choices[0].prompt_token_ids is not None assert completion.choices[0].prompt_logprobs is not None assert completion.choices[0].logprobs is not None # Extract token IDs from logprobs # (when return_tokens_as_token_ids is True) logprobs_token_ids = [] for token_str in completion.choices[0].logprobs.tokens: # Token format is "token_id:12345" when # return_tokens_as_token_ids is True if token_str.startswith("token_id:"): token_id = int(token_str.removeprefix("token_id:")) logprobs_token_ids.append(token_id) # When echo=True, the logprobs include both prompt and response tokens # The token_ids field should match the suffix of response portion # The prompt_token_ids should match the prompt portion assert len(completion.choices[0].token_ids) < len(logprobs_token_ids) response_token_ids_length = len(completion.choices[0].token_ids) assert ( logprobs_token_ids[-response_token_ids_length:] == completion.choices[0].token_ids ) # Verify tokenizer consistency tokenizer = get_tokenizer(tokenizer_name=MODEL_NAME) # Decode prompt tokens if completion.choices[0].prompt_token_ids: prompt_text = tokenizer.decode(completion.choices[0].prompt_token_ids) # The decoded prompt should match or close to original prompt assert "Hello, world" in prompt_text # Decode response tokens if completion.choices[0].token_ids: response_text = tokenizer.decode(completion.choices[0].token_ids) assert completion.choices[0].text.endswith(response_text) # Test streaming mode stream = await client.completions.create( model=MODEL_NAME, prompt="Tell me a short fact about Python:", max_tokens=30, temperature=0, stream=True, echo=False, logprobs=1, extra_body={"return_token_ids": True, "return_tokens_as_token_ids": True}, ) # Collect streamed tokens streamed_prompt_token_ids = [] streamed_token_ids = [] streamed_logprob_token_ids = [] first_chunk = True async for chunk in stream: for token_str in chunk.choices[0].logprobs.tokens: # Token format is "token_id:12345" when # return_tokens_as_token_ids is True if token_str.startswith("token_id:"): token_id = int(token_str.removeprefix("token_id:")) streamed_logprob_token_ids.append(token_id) if first_chunk: streamed_prompt_token_ids = chunk.choices[0].prompt_token_ids first_chunk = False streamed_token_ids += chunk.choices[0].token_ids # Verify we collected some tokens and first chunk had prompt_token_ids assert len(streamed_prompt_token_ids) > 0 assert streamed_token_ids == streamed_logprob_token_ids @pytest.mark.asyncio async def test_chat_completion_with_emoji_and_token_ids(server): """Test chat completion with emojis to verify token_ids handling.""" chat_messages = [ {"role": "system", "content": "You like to use emojis in your responses."}, {"role": "user", "content": "Repeat after me: I love cats 🐱"}, ] async with server.get_async_client() as client: response = await client.chat.completions.create( model=MODEL_NAME, messages=chat_messages, max_tokens=50, temperature=0, logprobs=True, extra_body={"return_token_ids": True}, ) # Verify token_ids are present response_dict = response.model_dump() assert response.choices[0].token_ids is not None assert "prompt_token_ids" in response_dict # Verify the response contains the expected fields assert response.choices[0].message.content is not None # Decode token_ids and verify consistency tokenizer = get_tokenizer(tokenizer_name=MODEL_NAME) decoded_prompt = tokenizer.decode(response.prompt_token_ids) assert decoded_prompt.startswith( "<|im_start|>system\nYou like to use emojis in your responses." ) assert decoded_prompt.endswith( "I love cats 🐱<|im_end|>\n<|im_start|>assistant\n" ) decoded_response = tokenizer.decode(response.choices[0].token_ids) # The content should match the response text # except the ending <|im_end|> assert decoded_response == response.choices[0].message.content + "<|im_end|>" # Test with streaming stream = await client.chat.completions.create( model=MODEL_NAME, messages=chat_messages, max_tokens=50, temperature=0, stream=True, extra_body={"return_token_ids": True}, ) collected_content = "" collected_token_ids = [] first_chunk = True async for chunk in stream: if first_chunk: assert chunk.prompt_token_ids is not None assert isinstance(chunk.prompt_token_ids, list) # Check the prompt_token_ids match the initial prompt decoded_prompt_stream = tokenizer.decode(chunk.prompt_token_ids) assert decoded_prompt_stream == decoded_prompt first_chunk = False else: chunk_dump = chunk.model_dump() assert "prompt_token_ids" not in chunk_dump, ( "Subsequent chunks should not have prompt_token_ids" ) if chunk.choices: if chunk.choices[0].delta.content: collected_content += chunk.choices[0].delta.content # token_ids may not present in all chunks choice_dump = chunk.choices[0].model_dump() if "token_ids" in choice_dump: collected_token_ids.extend(chunk.choices[0].token_ids) # Verify we got response and token_ids assert len(collected_content) > 0 assert len(collected_token_ids) > 0 # Verify token_ids decode properly decoded_response = tokenizer.decode(collected_token_ids) assert decoded_response == collected_content + "<|im_end|>"
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/test_return_token_ids.py", "license": "Apache License 2.0", "lines": 320, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/model_executor/models/ovis2_5.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """PyTorch Ovis model.""" from collections.abc import Iterable, Mapping from functools import partial from typing import Annotated, Literal import torch import torch.nn as nn from transformers import BaseImageProcessor, BatchFeature, PretrainedConfig from vllm.config import VllmConfig from vllm.config.multimodal import BaseDummyOptions from vllm.model_executor.layers.linear import ReplicatedLinear from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.models.ovis import VisualEmbedding from vllm.model_executor.models.siglip2navit import Siglip2NavitModel from vllm.model_executor.models.utils import ( AutoWeightsLoader, flatten_bn, init_vllm_registered_model, maybe_prefix, ) from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import ( MultiModalDataDict, MultiModalFieldConfig, MultiModalKwargsItems, ) from vllm.multimodal.parse import ImageSize, MultiModalDataItems from vllm.multimodal.processing import ( BaseDummyInputsBuilder, BaseMultiModalProcessor, BaseProcessingInfo, PromptReplacement, ) from vllm.renderers import TokenizeParams from vllm.sequence import IntermediateTensors from vllm.transformers_utils.processors.ovis2_5 import Ovis2_5Processor from vllm.utils.tensor_schema import TensorSchema, TensorShape from .interfaces import MultiModalEmbeddings, SupportsMultiModal, SupportsPP IMAGE_TOKEN = "<image>" VIDEO_TOKEN = "<video>" INDICATOR_IDS = [151672, 151673, 151674, 151675] IMAGE_PAD_TOKEN_ID = 151655 class Ovis2_5ImagePatchInputs(TensorSchema): """ Dimensions: - bnp: Batch size * number of images * number of patches - patch_size: patch_size_x * patch_size_y * num_channels - patch_indicators: Batch size * (number of patches + 1) - bn: Batch size * number of images """ type: Literal["image_patches"] flat_data: Annotated[torch.Tensor, TensorShape("bnp", "patch_size")] indicator_tokens: Annotated[torch.Tensor, TensorShape("patch_indicators")] patches_per_item: Annotated[list[int], TensorShape("bn")] grids: Annotated[torch.Tensor, TensorShape("bn", 3)] # This is used to restore the first two dimensions of `flat_data`. class Ovis2_5VideoPatchInputs(TensorSchema): """ Dimensions: - bnp: Batch size * number of videos * number of patches - patch_size: patch_size_x * patch_size_y * num_channels - patch_indicators: Batch size * (number of patches + 1) - bn: Batch size * number of videos """ type: Literal["video_patches"] flat_data: Annotated[torch.Tensor, TensorShape("bnp", "patch_size")] indicator_tokens: Annotated[torch.Tensor, TensorShape("patch_indicators")] patches_per_item: Annotated[list[int], TensorShape("bn")] grids: Annotated[torch.Tensor, TensorShape("bn", 3)] # This is used to restore the first two dimensions of `flat_data`. class VisualTokenizer(torch.nn.Module): """ VIT """ def __init__( self, config: PretrainedConfig, visual_vocab_size: int, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.config = config self.vit = self._init_backbone( config=config, quant_config=quant_config, prefix=f"{prefix}.vit", ) # reserved tokens for INDICATOR_IDS head_dim = visual_vocab_size - len(INDICATOR_IDS) self.head = torch.nn.Sequential( ReplicatedLinear( self.config.hidden_size * self.config.hidden_stride**2, head_dim, bias=False, return_bias=False, ), torch.nn.LayerNorm(head_dim), ) def _init_backbone( self, config: PretrainedConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): model_type = config.model_type if model_type == "siglip2_navit": return Siglip2NavitModel( config=config, quant_config=quant_config, prefix=prefix, ) raise ValueError(f"Unsupported visual tokenizer model_type: {model_type}") @property def dtype(self) -> torch.dtype: return next(self.head.parameters()).dtype @property def device(self) -> torch.device: return next(self.head.parameters()).device def tokenize(self, logits: torch.Tensor) -> torch.Tensor: tokens = torch.softmax(logits, dim=-1, dtype=torch.float32).to(logits.dtype) return tokens def encode( self, pixel_values: torch.Tensor, grid_thws: torch.Tensor ) -> torch.Tensor: features = self.vit(pixel_values, grid_thws) # refer to qwen2.5-vl patchmerger seq_len, _ = features.shape features = features.reshape(seq_len // (self.config.hidden_stride**2), -1) return features def forward( self, pixel_values: torch.Tensor, grid_thws: torch.Tensor ) -> torch.Tensor: features = self.encode(pixel_values, grid_thws) logits = self.head(features) tokens = self.tokenize(logits) # tokens' shape is [#Token, VocabSize-4], # so padding with [#Token, 4], after which, # tokens' shape should become [#Token, VocabSize]; tokens = torch.nn.functional.pad( tokens, (0, len(INDICATOR_IDS)), mode="constant", value=0, ) return tokens class Ovis2_5ProcessingInfo(BaseProcessingInfo): def get_hf_config(self): return self.ctx.get_hf_config() def get_hf_processor(self, **kwargs): vit_config = self.get_hf_config().vit_config return self.ctx.get_hf_processor( Ovis2_5Processor, patch_size=vit_config.patch_size, hidden_stride=vit_config.hidden_stride, temporal_patch_size=vit_config.temporal_patch_size, ) def get_default_tok_params(self) -> TokenizeParams: return super().get_default_tok_params().with_kwargs(add_special_tokens=False) def get_image_processor(self) -> BaseImageProcessor: return self.get_hf_processor().image_processor # type: ignore def get_supported_mm_limits(self) -> Mapping[str, int | None]: return {"image": None, "video": 1} def get_image_size_with_most_features(self) -> ImageSize: # NOTE(myselvess): max_pixels 1792 * 1792 hardcoded in original code # TODO(myselvess): Be adjusted based on the max_pixels return ImageSize(width=1792, height=1792) def get_num_image_tokens( self, *, image_width: int, image_height: int, num_frames: int = 1, ) -> int: hf_config = self.get_hf_config() vit_config = hf_config.vit_config patch_size = vit_config.patch_size temporal_patch_size = vit_config.temporal_patch_size # NOTE: Frames are padded to be divisible by `temporal_patch_size` # https://github.com/huggingface/transformers/blob/v4.48.3/src/transformers/models/qwen2_vl/image_processing_qwen2_vl.py#L294 padded_num_frames = num_frames + (-num_frames % temporal_patch_size) grid_t = max(padded_num_frames // temporal_patch_size, 1) grid_h = image_height // patch_size grid_w = image_width // patch_size num_patches = grid_t * grid_h * grid_w num_vision_tokens = num_patches return num_vision_tokens def get_max_image_tokens(self) -> int: target_width, target_height = self.get_image_size_with_most_features() return self.get_num_image_tokens( image_width=target_width, image_height=target_height ) def _get_max_video_frames(self, max_tokens: int) -> int: target_width, target_height = self.get_image_size_with_most_features() num_frames = 0 while True: next_num_frames = num_frames + 1 next_max_tokens = self.get_num_video_tokens( image_width=target_width, image_height=target_height, num_frames=next_num_frames, ) if next_max_tokens > max_tokens: break num_frames = next_num_frames return num_frames def get_num_frames_with_most_features( self, seq_len: int, mm_counts: Mapping[str, int], ) -> int: max_images = mm_counts.get("image", 0) max_videos = mm_counts.get("video", 0) max_image_tokens = self.get_max_image_tokens() * max_images max_total_frames = self._get_max_video_frames(seq_len - max_image_tokens) max_frames_per_video = max_total_frames // max(max_videos, 1) return max(max_frames_per_video, 1) def get_num_video_tokens( self, *, image_width: int, image_height: int, num_frames: int, ) -> int: num_video_tokens = self.get_num_image_tokens( image_width=image_width, image_height=image_height, num_frames=num_frames ) return num_video_tokens def get_max_video_tokens( self, seq_len: int, mm_counts: Mapping[str, int], ) -> int: target_width, target_height = self.get_image_size_with_most_features() return self.get_num_video_tokens( image_width=target_width, image_height=target_height, num_frames=self.get_num_frames_with_most_features(seq_len, mm_counts), ) class Ovis2_5DummyInputsBuilder(BaseDummyInputsBuilder[Ovis2_5ProcessingInfo]): def get_dummy_text(self, mm_counts: Mapping[str, int]) -> str: num_images = mm_counts.get("image", 0) num_videos = mm_counts.get("video", 0) return IMAGE_TOKEN * num_images + VIDEO_TOKEN * num_videos def get_dummy_mm_data( self, seq_len: int, mm_counts: Mapping[str, int], mm_options: Mapping[str, BaseDummyOptions], ) -> MultiModalDataDict: num_images = mm_counts.get("image", 0) num_videos = mm_counts.get("video", 0) target_width, target_height = self.info.get_image_size_with_most_features() target_num_frames = self.info.get_num_frames_with_most_features( seq_len, mm_counts ) image_overrides = mm_options.get("image") video_overrides = mm_options.get("video") mm_data = { "image": self._get_dummy_images( width=target_width, height=target_height, num_images=num_images, overrides=image_overrides, ), "video": self._get_dummy_videos( width=target_width, height=target_height, num_frames=target_num_frames, num_videos=num_videos, overrides=video_overrides, ), } return mm_data class Ovis2_5MultiModalProcessor(BaseMultiModalProcessor[Ovis2_5ProcessingInfo]): def visual_indicators_to_visual_tokens( self, visual_indicators: list[int], ) -> list[int]: """ Filter image indicators placeholders and convert them to corresponding tokens in visual tokenizer. """ hf_config = self.info.get_hf_config() vte_vocab_size = hf_config.visual_vocab_size return [ vte_vocab_size - len(INDICATOR_IDS) + (x - INDICATOR_IDS[0]) for x in visual_indicators if x >= INDICATOR_IDS[0] ] def _call_hf_processor( self, prompt: str, mm_data: Mapping[str, object], mm_kwargs: Mapping[str, object], tok_kwargs: Mapping[str, object], ) -> BatchFeature: if not mm_data: # Avoid warning from HF logger for text-only input tokenizer = self.info.get_tokenizer() prompt_ids = tokenizer.encode(prompt, add_special_tokens=False) return BatchFeature(dict(input_ids=[prompt_ids]), tensor_type="pt") processed_outputs = super()._call_hf_processor( prompt=prompt, mm_data=mm_data, mm_kwargs=mm_kwargs, tok_kwargs=tok_kwargs, ) hf_processor = self.info.get_hf_processor() if "videos" in mm_data: visual_indicators = [ hf_processor.construct_visual_indicators((1, 1, 1), True) for grid in processed_outputs["video_grids"] ] indicator_tokens = [ self.visual_indicators_to_visual_tokens(indicator) for indicator in visual_indicators ] processed_outputs["video_indicator_tokens"] = torch.tensor(indicator_tokens) if "images" in mm_data: visual_indicators = [ hf_processor.construct_visual_indicators((1, 1, 1), False) for grid in processed_outputs["grids"] ] indicator_tokens = [ self.visual_indicators_to_visual_tokens(indicator) for indicator in visual_indicators ] processed_outputs["indicator_tokens"] = torch.tensor(indicator_tokens) return processed_outputs def _apply_hf_processor_tokens_only( self, prompt_tokens: list[int], ) -> list[int]: return prompt_tokens def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: return dict( pixel_values=MultiModalFieldConfig.batched("image"), grids=MultiModalFieldConfig.batched("image"), indicator_tokens=MultiModalFieldConfig.batched("image"), video_pixel_values=MultiModalFieldConfig.batched("video"), video_indicator_tokens=MultiModalFieldConfig.batched("video"), video_grids=MultiModalFieldConfig.batched("video"), ) def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, object], out_mm_kwargs: MultiModalKwargsItems, ) -> list[PromptReplacement]: tokenizer = self.info.get_tokenizer() vocab = tokenizer.get_vocab() placeholder = { "image": vocab[IMAGE_TOKEN], "video": vocab[VIDEO_TOKEN], } def get_replacement_ovis(item_idx, modality: str): if modality == "image": out_item = out_mm_kwargs["image"][item_idx] grid = out_item["grids"].data elif modality == "video": out_item = out_mm_kwargs["video"][item_idx] grid = out_item["video_grids"].data hf_processor = self.info.get_hf_processor() return hf_processor.construct_visual_placeholders( grid[0], ) return [ PromptReplacement( modality=modality, target=[placeholder[modality]], replacement=partial(get_replacement_ovis, modality=modality), ) for modality in ("image", "video") ] @MULTIMODAL_REGISTRY.register_processor( Ovis2_5MultiModalProcessor, info=Ovis2_5ProcessingInfo, dummy_inputs=Ovis2_5DummyInputsBuilder, ) class Ovis2_5(nn.Module, SupportsMultiModal, SupportsPP): @classmethod def get_placeholder_str(cls, modality: str, i: int) -> str | None: if modality.startswith("image"): return IMAGE_TOKEN if modality.startswith("video"): return VIDEO_TOKEN raise ValueError("Only image or video modality is supported") def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config self.config: PretrainedConfig = config with self._mark_language_model(vllm_config): self.llm = init_vllm_registered_model( vllm_config=vllm_config.with_hf_config(config.text_config), prefix=maybe_prefix(prefix, "llm"), ) with self._mark_tower_model(vllm_config, {"image", "video"}): self.visual_tokenizer = VisualTokenizer( config=config.vit_config, visual_vocab_size=config.visual_vocab_size, quant_config=quant_config, prefix=f"{prefix}.visual_tokenizer", ) self.vte = VisualEmbedding(config.visual_vocab_size, config.hidden_size) self.image_pad_token_id: int = IMAGE_PAD_TOKEN_ID self.make_empty_intermediate_tensors = ( self.get_language_model().make_empty_intermediate_tensors ) def _parse_and_validate_image_input( self, **kwargs: object ) -> Ovis2_5ImagePatchInputs | None: pixel_values = kwargs.pop("pixel_values", None) indicator_tokens = kwargs.pop("indicator_tokens", None) grids = kwargs.pop("grids", None) if pixel_values is None and indicator_tokens is None: return None if pixel_values is not None and indicator_tokens is not None: if not isinstance(pixel_values, (torch.Tensor, list)): raise ValueError( f"Incorrect type of pixel values. Got type: {type(pixel_values)}" ) if not isinstance(indicator_tokens, (torch.Tensor, list)): raise ValueError( "Incorrect type of indicator_tokens. " f"Got type: {type(indicator_tokens)}" ) return Ovis2_5ImagePatchInputs( type="image_patches", flat_data=flatten_bn(pixel_values, concat=True), patches_per_item=[ x.shape[0] // (self.config.vit_config.hidden_stride**2) for x in pixel_values ], indicator_tokens=flatten_bn(indicator_tokens, concat=True), grids=flatten_bn(grids, concat=True), ) raise AssertionError("This line should be unreachable.") def _parse_and_validate_video_input( self, **kwargs: object ) -> Ovis2_5VideoPatchInputs | None: pixel_values = kwargs.pop("video_pixel_values", None) indicator_tokens = kwargs.pop("video_indicator_tokens", None) grids = kwargs.pop("video_grids", None) if pixel_values is None and indicator_tokens is None: return None if pixel_values is not None and indicator_tokens is not None: if not isinstance(pixel_values, (torch.Tensor, list)): raise ValueError( f"Incorrect type of pixel values. Got type: {type(pixel_values)}" ) if not isinstance(indicator_tokens, (torch.Tensor, list)): raise ValueError( "Incorrect type of indicator_tokens. " f"Got type: {type(indicator_tokens)}" ) return Ovis2_5VideoPatchInputs( type="video_patches", flat_data=flatten_bn(pixel_values, concat=True), patches_per_item=[ x.shape[0] // (self.config.vit_config.hidden_stride**2) for x in pixel_values ], indicator_tokens=flatten_bn(indicator_tokens, concat=True), grids=flatten_bn(grids, concat=True), ) raise AssertionError("This line should be unreachable.") def _process_visual_input( self, visual_input: Ovis2_5ImagePatchInputs | Ovis2_5VideoPatchInputs ) -> MultiModalEmbeddings: image_patches_flat = visual_input["flat_data"] patches_per_image = visual_input["patches_per_item"] indicator_tokens = visual_input["indicator_tokens"] grid_thws = visual_input["grids"] indicator_per_image = list( map(lambda x: 2 if x > 1 else x + 2, patches_per_image) ) target_dtype = self.visual_tokenizer.dtype visual_tokens = self.visual_tokenizer( image_patches_flat.to(target_dtype), grid_thws ) visual_embeds = self.vte(visual_tokens) # 1:1 numeric eq. indicator_embeds = self.vte(indicator_tokens) visual_embeds_per_image = visual_embeds.split(patches_per_image, dim=0) indicator_embeds_per_image = indicator_embeds.split(indicator_per_image) vision_embeddings = [] for indicator, visual in zip( indicator_embeds_per_image, visual_embeds_per_image ): vision_embeddings_per_image = [] visual = visual.unsqueeze(0) for i in range(visual.shape[0]): vision_embeddings_per_image.append( torch.cat([indicator[i : i + 1], visual[i]], dim=0) ) vision_embeddings_per_image.append(indicator[i + 1 :]) vision_embeddings.append(torch.cat(vision_embeddings_per_image, dim=0)) return tuple(vision_embeddings) def _parse_and_validate_multimodal_inputs(self, **kwargs: object) -> dict: modalities = {} # Preserve the order of modalities if there are multiple of them # from the order of kwargs. for input_key in kwargs: if ( input_key in ("pixel_values", "indicator_tokens", "grids") and "images" not in modalities ): modalities["images"] = self._parse_and_validate_image_input(**kwargs) if ( input_key in ("video_pixel_values", "video_indicator_tokens", "video_grids") and "videos" not in modalities ): modalities["videos"] = self._parse_and_validate_video_input(**kwargs) return modalities def embed_multimodal(self, **kwargs: object) -> MultiModalEmbeddings: modalities = self._parse_and_validate_multimodal_inputs(**kwargs) if not modalities: return [] multimodal_embeddings: tuple[torch.Tensor, ...] = () # NOTE: It is important to iterate over the keys in this dictionary # to preserve the order of the modalities. for modality in modalities: if modality == "images": image_input = modalities["images"] image_embeddings = self._process_visual_input(image_input) multimodal_embeddings += tuple(image_embeddings) if modality == "videos": video_input = modalities["videos"] video_embeddings = self._process_visual_input(video_input) multimodal_embeddings += tuple(video_embeddings) return multimodal_embeddings def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs: object, ) -> torch.Tensor | IntermediateTensors: if intermediate_tensors is not None: inputs_embeds = None # up until here we have a inputs_embeds 100% numerical identity # between the OG HF Transformers implementation and ours hidden_states = self.llm( input_ids=input_ids, positions=positions, intermediate_tensors=intermediate_tensors, inputs_embeds=inputs_embeds, ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: return self.llm.compute_logits(hidden_states) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader(self) return loader.load_weights(weights)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/ovis2_5.py", "license": "Apache License 2.0", "lines": 565, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/siglip2navit.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Implementation of SiglipVisionModel intended to be only used within a vision language model.""" from collections.abc import Iterable import torch from torch import nn from torch.nn import functional as F from transformers import Siglip2VisionConfig from transformers.configuration_utils import PretrainedConfig from vllm.distributed import divide, get_tensor_model_parallel_world_size from vllm.model_executor.layers.activation import get_act_fn from vllm.model_executor.layers.attention import MMEncoderAttention from vllm.model_executor.layers.conv import Conv2dLayer from vllm.model_executor.layers.linear import ( ColumnParallelLinear, LinearBase, QKVParallelLinear, ReplicatedLinear, RowParallelLinear, ) from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.rotary_embedding.common import ( ApplyRotaryEmb, ) from vllm.model_executor.model_loader.weight_utils import default_weight_loader from vllm.platforms import current_platform from .vision import is_vit_use_data_parallel class VisionRotaryEmbedding(nn.Module): def __init__(self, dim: int, theta: float = 10000.0) -> None: super().__init__() inv_freq = 1.0 / (theta ** (torch.arange(0, dim, 2, dtype=torch.float) / dim)) self.register_buffer("inv_freq", inv_freq, persistent=False) def forward(self, seqlen: int) -> torch.Tensor: seq = torch.arange( seqlen, device=self.inv_freq.device, dtype=self.inv_freq.dtype ) freqs = torch.outer(seq, self.inv_freq) return freqs class Siglip2VisionEmbeddings(nn.Module): def __init__(self, config: PretrainedConfig): super().__init__() self.config = config self.embed_dim = config.hidden_size self.patch_size = config.patch_size self.image_size = config.image_size self.num_patches = config.num_patches self.preserve_original_pe = config.preserve_original_pe self.hidden_stride = config.hidden_stride # siglip2 naflex if self.num_patches > 0: self.patch_embedding = ReplicatedLinear( input_size=config.num_channels * self.patch_size * self.patch_size, output_size=self.embed_dim, return_bias=False, ) if self.preserve_original_pe: self.position_embedding_size = int(self.num_patches**0.5) self.position_embedding = nn.Embedding(self.num_patches, self.embed_dim) else: self.patch_embedding = Conv2dLayer( in_channels=config.num_channels, out_channels=self.embed_dim, kernel_size=self.patch_size, stride=self.patch_size, padding="valid", ) if self.preserve_original_pe: self.num_patches = (self.image_size // self.patch_size) ** 2 self.position_embedding_size = self.image_size // self.patch_size self.position_embedding = nn.Embedding(self.num_patches, self.embed_dim) def forward( self, pixel_values: torch.FloatTensor, grid_thws: torch.LongTensor | None = None, ) -> torch.Tensor: """ Args: pixel_values (`torch.FloatTensor`): Pixel values of shape ( num_patches, num_channels * temporal_patch_size * patch_size * patch_size ) grid_thws: (`torch.LongTensor`): grid shape (num_patches, 3) """ # Apply patch embeddings to already patchified pixel values target_dtype = self.patch_embedding.weight.dtype if isinstance(self.patch_embedding, LinearBase): patch_embeds = self.patch_embedding(pixel_values.to(dtype=target_dtype)) elif isinstance(self.patch_embedding, Conv2dLayer): pixel_values = pixel_values.view( -1, self.config.num_channels * self.config.temporal_patch_size, self.patch_size, self.patch_size, ) patch_embeds = self.patch_embedding(pixel_values.to(dtype=target_dtype)) patch_embeds = patch_embeds.reshape(-1, self.embed_dim) if self.preserve_original_pe: assert grid_thws is not None pos_embed_new = torch.zeros_like(patch_embeds) positional_embeddings = ( self.position_embedding.weight.reshape( self.position_embedding_size, self.position_embedding_size, -1 ) .unsqueeze(0) .permute(0, 3, 1, 2) ) cnt = 0 for t, h, w in grid_thws: volume = t * h * w pe = F.interpolate( positional_embeddings, size=(h, w), mode="bicubic", align_corners=False, ) pe = pe.permute(0, 2, 3, 1).reshape(1, h * w, -1) pe = pe[0].repeat(t, 1) pe = pe.reshape( t, h // self.hidden_stride, self.hidden_stride, w // self.hidden_stride, self.hidden_stride, -1, ) pe = pe.permute(0, 1, 3, 2, 4, 5).reshape(volume, -1) pos_embed_new[cnt : cnt + volume] = pe cnt += volume patch_embeds = patch_embeds + pos_embed_new return patch_embeds def apply_rotary_pos_emb( q: torch.Tensor, k: torch.Tensor, cos: torch.Tensor, sin: torch.Tensor, is_flash_attn_backend: bool, apply_rotary_emb: ApplyRotaryEmb, ) -> tuple[torch.Tensor, torch.Tensor]: cos = cos.chunk(2, dim=-1)[0].contiguous() sin = sin.chunk(2, dim=-1)[0].contiguous() if is_flash_attn_backend and current_platform.is_cuda(): apply_rotary_emb_func = apply_rotary_emb.forward_cuda elif is_flash_attn_backend and current_platform.is_rocm(): apply_rotary_emb_func = apply_rotary_emb.forward_hip else: apply_rotary_emb_func = apply_rotary_emb.forward_native q_embed = apply_rotary_emb_func(q, cos, sin) k_embed = apply_rotary_emb_func(k, cos, sin) return q_embed, k_embed class Siglip2Attention(nn.Module): """Multi-headed attention from 'Attention Is All You Need' paper""" def __init__( self, config: Siglip2VisionConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.config = config self.embed_dim = config.hidden_size self.num_heads = config.num_attention_heads self.head_dim = self.embed_dim // self.num_heads if self.head_dim * self.num_heads != self.embed_dim: raise ValueError( f"embed_dim must be divisible by num_heads " f"(got `embed_dim`: {self.embed_dim} and " f"`num_heads`: {self.num_heads})." ) self.scale = self.head_dim**-0.5 self.dropout = config.attention_dropout use_data_parallel = is_vit_use_data_parallel() self.qkv_proj = QKVParallelLinear( hidden_size=self.embed_dim, head_size=self.head_dim, total_num_heads=self.num_heads, quant_config=quant_config, prefix=f"{prefix}.qkv_proj", disable_tp=use_data_parallel, ) self.out_proj = RowParallelLinear( input_size=self.embed_dim, output_size=self.embed_dim, quant_config=quant_config, prefix=f"{prefix}.out_proj", disable_tp=use_data_parallel, ) self.tp_size = ( 1 if use_data_parallel else get_tensor_model_parallel_world_size() ) self.num_heads_per_partition = divide(self.num_heads, self.tp_size) self.use_rope = config.use_rope self.attn = MMEncoderAttention( num_heads=self.num_heads_per_partition, head_size=self.head_dim, scale=self.scale, prefix=f"{prefix}.attn", ) self.apply_rotary_emb = ApplyRotaryEmb( enforce_enable=True, enable_fp32_compute=True, ) def forward( self, hidden_states: torch.Tensor, cu_seqlens: torch.Tensor, position_embeddings: tuple[torch.Tensor, torch.Tensor] | None = None, ) -> tuple[torch.Tensor, torch.Tensor | None]: """Input shape: Batch x Time x Channel""" seq_length, embed_dim = hidden_states.shape qkv_states, _ = self.qkv_proj(hidden_states) queries, keys, values = qkv_states.chunk(3, dim=-1) queries = queries.view(seq_length, self.num_heads_per_partition, self.head_dim) keys = keys.view(seq_length, self.num_heads_per_partition, self.head_dim) values = values.view(seq_length, self.num_heads_per_partition, self.head_dim) if self.use_rope: cos, sin = position_embeddings queries, keys = apply_rotary_pos_emb( queries.unsqueeze(0), keys.unsqueeze(0), cos, sin, self.attn.is_flash_attn_backend, self.apply_rotary_emb, ) queries = queries.squeeze(0) keys = keys.squeeze(0) max_seqlen = (cu_seqlens[1:] - cu_seqlens[:-1]).max() attn_output = self.attn( query=queries.unsqueeze(0), key=keys.unsqueeze(0), value=values.unsqueeze(0), cu_seqlens=cu_seqlens, max_seqlen=max_seqlen, ) attn_output = attn_output.reshape( seq_length, self.num_heads_per_partition * self.head_dim ) attn_output, _ = self.out_proj(attn_output) return attn_output class Siglip2MLP(nn.Module): def __init__( self, config: Siglip2VisionConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.config = config use_data_parallel = is_vit_use_data_parallel() self.activation_fn = get_act_fn(config.hidden_act) self.fc1 = ColumnParallelLinear( config.hidden_size, config.intermediate_size, quant_config=quant_config, prefix=f"{prefix}.fc1", disable_tp=use_data_parallel, ) self.fc2 = RowParallelLinear( config.intermediate_size, config.hidden_size, quant_config=quant_config, prefix=f"{prefix}.fc2", disable_tp=use_data_parallel, ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states, _ = self.fc1(hidden_states) hidden_states = self.activation_fn(hidden_states) hidden_states, _ = self.fc2(hidden_states) return hidden_states class Siglip2EncoderLayer(nn.Module): def __init__( self, config: Siglip2VisionConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.embed_dim = config.hidden_size self.layer_norm1 = nn.LayerNorm(self.embed_dim, eps=config.layer_norm_eps) self.self_attn = Siglip2Attention( config, quant_config=quant_config, prefix=f"{prefix}.self_attn", ) self.layer_norm2 = nn.LayerNorm(self.embed_dim, eps=config.layer_norm_eps) self.mlp = Siglip2MLP( config, quant_config=quant_config, prefix=f"{prefix}.mlp", ) def forward( self, hidden_states: torch.Tensor, cu_seqlens: torch.Tensor, position_embeddings: torch.Tensor, ) -> tuple[torch.FloatTensor]: """ Args: hidden_states: Input tensor of shape (batch, seq_len, embed_dim). cu_seqlens: Cumulative sequence lengths tensor. position_embeddings: Position embeddings tensor. """ residual = hidden_states hidden_states = self.layer_norm1(hidden_states) hidden_states = self.self_attn( hidden_states=hidden_states, cu_seqlens=cu_seqlens, position_embeddings=position_embeddings, ) hidden_states = residual + hidden_states residual = hidden_states hidden_states = self.layer_norm2(hidden_states) hidden_states = self.mlp(hidden_states) hidden_states = residual + hidden_states return hidden_states class Siglip2Encoder(nn.Module): """ Transformer encoder consisting of `config.num_hidden_layers` self attention layers. Each layer is a [`Siglip2EncoderLayer`]. Args: config: PretrainedConfig """ def __init__( self, config: Siglip2VisionConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.config = config self.layers = nn.ModuleList( [ Siglip2EncoderLayer( config, quant_config=quant_config, prefix=f"{prefix}.layers.{idx}", ) for idx in range(config.num_hidden_layers) ] ) self.rotary_pos_emb = VisionRotaryEmbedding( config.hidden_size // config.num_attention_heads // 2 ) self.patch_size = config.patch_size self.hidden_stride = config.hidden_stride self.window_size = config.window_size self.spatial_merge_unit = config.hidden_stride * config.hidden_stride if config.fullatt_block_indexes is None: self.fullatt_block_indexes = None else: self.fullatt_block_indexes = [ int(i) for i in config.fullatt_block_indexes.split("|") ] # copied from qwen2.5_vl def rot_pos_emb(self, grid_thw): pos_ids = [] for t, h, w in grid_thw: hpos_ids = torch.arange(h).unsqueeze(1).expand(-1, w) hpos_ids = hpos_ids.reshape( h // self.hidden_stride, self.hidden_stride, w // self.hidden_stride, self.hidden_stride, ) hpos_ids = hpos_ids.permute(0, 2, 1, 3) hpos_ids = hpos_ids.flatten() wpos_ids = torch.arange(w).unsqueeze(0).expand(h, -1) wpos_ids = wpos_ids.reshape( h // self.hidden_stride, self.hidden_stride, w // self.hidden_stride, self.hidden_stride, ) wpos_ids = wpos_ids.permute(0, 2, 1, 3) wpos_ids = wpos_ids.flatten() pos_ids.append(torch.stack([hpos_ids, wpos_ids], dim=-1).repeat(t, 1)) pos_ids = torch.cat(pos_ids, dim=0) max_grid_size = grid_thw[:, 1:].max() rotary_pos_emb_full = self.rotary_pos_emb(max_grid_size) rotary_pos_emb = rotary_pos_emb_full[pos_ids].flatten(1) return rotary_pos_emb def get_window_index(self, grid_thw): window_index: list = [] cu_window_seqlens: list = [0] window_index_id = 0 # patch (after merge) number in each window vit_merger_window_size = ( self.window_size // self.hidden_stride // self.patch_size ) for grid_t, grid_h, grid_w in grid_thw: llm_grid_h, llm_grid_w = ( grid_h // self.hidden_stride, # number of patch after merge grid_w // self.hidden_stride, ) index = torch.arange(grid_t * llm_grid_h * llm_grid_w).reshape( grid_t, llm_grid_h, llm_grid_w ) pad_h = vit_merger_window_size - llm_grid_h % vit_merger_window_size pad_w = vit_merger_window_size - llm_grid_w % vit_merger_window_size num_windows_h = (llm_grid_h + pad_h) // vit_merger_window_size num_windows_w = (llm_grid_w + pad_w) // vit_merger_window_size index_padded = F.pad(index, (0, pad_w, 0, pad_h), "constant", -100) index_padded = index_padded.reshape( grid_t, num_windows_h, vit_merger_window_size, num_windows_w, vit_merger_window_size, ) index_padded = index_padded.permute(0, 1, 3, 2, 4).reshape( grid_t, num_windows_h * num_windows_w, vit_merger_window_size, vit_merger_window_size, ) seqlens = (index_padded != -100).sum([2, 3]).reshape(-1) index_padded = index_padded.reshape(-1) index_new = index_padded[index_padded != -100] window_index.append(index_new + window_index_id) cu_seqlens_tmp = ( seqlens.cumsum(0) * self.spatial_merge_unit + cu_window_seqlens[-1] ) cu_window_seqlens.extend(cu_seqlens_tmp.tolist()) window_index_id += (grid_t * llm_grid_h * llm_grid_w).item() window_index = torch.cat(window_index, dim=0) return window_index, cu_window_seqlens def forward( self, inputs_embeds: torch.Tensor, grid_thws: torch.Tensor, ) -> torch.Tensor: r""" Args: inputs_embeds: Input tensor of shape (batch_size, sequence_length, hidden_size). Embedded representation of the input tokens. grid_thws: Grid tensor of shape (num_patches, 3) containing grid dimensions. Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ rotary_pos_emb = self.rot_pos_emb(grid_thws) window_index, cu_window_seqlens = self.get_window_index(grid_thws) cu_window_seqlens = torch.tensor( cu_window_seqlens, device=inputs_embeds.device, dtype=grid_thws.dtype if torch.jit.is_tracing() else torch.int32, ) cu_window_seqlens = torch.unique_consecutive(cu_window_seqlens) seq_len, _ = inputs_embeds.size() inputs_embeds = inputs_embeds.reshape( seq_len // self.spatial_merge_unit, self.spatial_merge_unit, -1 ) inputs_embeds = inputs_embeds[window_index, :, :] inputs_embeds = inputs_embeds.reshape(seq_len, -1) rotary_pos_emb = rotary_pos_emb.reshape( seq_len // self.spatial_merge_unit, self.spatial_merge_unit, -1 ) rotary_pos_emb = rotary_pos_emb[window_index, :, :] rotary_pos_emb = rotary_pos_emb.reshape(seq_len, -1) emb = torch.cat((rotary_pos_emb, rotary_pos_emb), dim=-1) position_embeddings = (emb.cos(), emb.sin()) cu_seqlens = torch.repeat_interleave( grid_thws[:, 1] * grid_thws[:, 2], grid_thws[:, 0] ).cumsum( dim=0, # Select dtype based on the following factors: # - FA2 requires that cu_seqlens_q must have dtype int32 # - torch.onnx.export requires that cu_seqlens_q must have # same dtype as grid_thw # See https://github.com/huggingface/transformers/pull/34852 # for more information dtype=grid_thws.dtype if torch.jit.is_tracing() else torch.int32, ) cu_seqlens = torch.cat([cu_seqlens.new_zeros(1), cu_seqlens]) reverse_indices = torch.argsort(window_index) hidden_states = inputs_embeds for index, block in enumerate(self.layers): if not self.fullatt_block_indexes or index in self.fullatt_block_indexes: cu_seqlens_tmp = cu_seqlens else: cu_seqlens_tmp = cu_window_seqlens hidden_states = block(hidden_states, cu_seqlens_tmp, position_embeddings) hidden_states = hidden_states.reshape( seq_len // self.spatial_merge_unit, self.spatial_merge_unit, -1 ) hidden_states = hidden_states[reverse_indices, :].reshape(seq_len, -1) return hidden_states class Siglip2VisionTransformer(nn.Module): def __init__( self, config: Siglip2VisionConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.config = config embed_dim = config.hidden_size self.embeddings = Siglip2VisionEmbeddings(config) self.encoder = Siglip2Encoder( config, quant_config=quant_config, prefix=f"{prefix}.encoder", ) self.post_layernorm = nn.LayerNorm(embed_dim, eps=config.layer_norm_eps) def forward( self, pixel_values: torch.FloatTensor, grid_thws: torch.LongTensor, ) -> torch.Tensor: r""" spatial_shapes (`torch.LongTensor` of shape `(batch_size, 2)`): Tensor containing the spatial dimensions (height, width) of the input images. """ hidden_states = self.embeddings(pixel_values, grid_thws) last_hidden_state = self.encoder(hidden_states, grid_thws) return last_hidden_state class Siglip2NavitModel(torch.nn.Module): def __init__( self, config: Siglip2VisionConfig, quant_config: QuantizationConfig | None = None, prefix: str = "", ): super().__init__() self.vision_model = Siglip2VisionTransformer( config, quant_config=quant_config, prefix=f"{prefix}.vision_model", ) def forward( self, pixel_values: torch.FloatTensor, grid_thws: torch.LongTensor, ) -> torch.Tensor: return self.vision_model( pixel_values=pixel_values, grid_thws=grid_thws, ) def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: stacked_params_mapping = [ # (param_name, shard_name, shard_id) ("qkv_proj", "q_proj", "q"), ("qkv_proj", "k_proj", "k"), ("qkv_proj", "v_proj", "v"), ] params_dict = dict(self.named_parameters()) loaded_params: set[str] = set() for name, loaded_weight in weights: for param_name, weight_name, shard_id in stacked_params_mapping: if weight_name not in name: continue name = name.replace(weight_name, param_name) param = params_dict[name] weight_loader = param.weight_loader weight_loader(param, loaded_weight, shard_id) break else: param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) loaded_params.add(name) return loaded_params
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/siglip2navit.py", "license": "Apache License 2.0", "lines": 567, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/transformers_utils/processors/ovis2_5.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math from functools import cached_property import numpy as np import PIL import torch from transformers import AutoProcessor, BatchFeature from transformers.image_utils import ImageInput from transformers.processing_utils import ProcessingKwargs, ProcessorMixin, Unpack from transformers.tokenization_utils_base import PreTokenizedInput, TextInput __all__ = ["Ovis2_5Processor"] IMAGE_TOKEN = "<image>" VIDEO_TOKEN = "<video>" MIN_PIXELS = 448 * 448 MAX_PIXELS = 1792 * 1792 class Ovis2_5ProcessorKwargs(ProcessingKwargs, total=False): # type: ignore[call-arg] _defaults = { "text_kwargs": { "padding": False, }, "images_kwargs": { "do_convert_rgb": True, }, "videos_kwargs": { "do_convert_rgb": True, }, } class Ovis2_5Processor(ProcessorMixin): r""" Constructs an Ovis processor which wraps an Ovis image processor and a Qwen2 tokenizer into a single processor. [`OvisProcessor`] offers all the functionalities of [`Qwen2VLImageProcessor`] and [`Qwen2TokenizerFast`]. See the [`~OvisProcessor.__call__`] and [`~OvisProcessor.decode`] for more information. Args: image_processor ([`Qwen2VLImageProcessor`], *optional*): The image processor is a required input. tokenizer ([`Qwen2TokenizerFast`], *optional*): The tokenizer is a required input. chat_template (`str`, *optional*): A Jinja template which will be used to convert lists of messages in a chat into a tokenizable string. """ attributes = ["image_processor", "tokenizer"] valid_kwargs = ["chat_template", "image_pad_token"] image_processor_class = "AutoImageProcessor" tokenizer_class = "AutoTokenizer" def __init__( self, image_processor=None, tokenizer=None, chat_template=None, image_pad_token=None, patch_size=16, hidden_stride=2, temporal_patch_size=1, **kwargs, ): self.image_token = IMAGE_TOKEN self.video_token = VIDEO_TOKEN self.image_pad_token = "<|image_pad|>" self.patch_size = patch_size self.hidden_stride = hidden_stride self.temporal_patch_size = temporal_patch_size super().__init__(image_processor, tokenizer, chat_template=chat_template) @cached_property def extra_special_tokens(self): vocab = self.tokenizer.get_vocab() required_tokens = { "image_token": "<image>", "video_token": "<video>", "visual_atom": "<ovis_visual_atom>", "image_start": "<ovis_image_start>", "image_end": "<ovis_image_end>", "video_start": "<ovis_video_start>", "video_end": "<ovis_video_end>", "image_pad": "<|image_pad|>", } extra_special_tokens = {} suggestion = ( "please add '<image>', '<video>', '<ovis_visual_atom>', " "'<ovis_image_start>', '<ovis_image_end>', '<ovis_video_start>', " "'<ovis_video_end>' in 'additional_special_tokens' of " "tokenizer_config.json, You can refer to " "https://huggingface.co/AIDC-AI/Ovis2.6-30B-A3B/blob/main/tokenizer_config.json" ) for key, token_name in required_tokens.items(): if token_name not in vocab: raise ValueError(f"Can not find {token_name}, {suggestion}") extra_special_tokens[key] = vocab[token_name] return extra_special_tokens def __call__( self, images: ImageInput = None, videos: np.ndarray | list[ImageInput] = None, text: TextInput | PreTokenizedInput | list[TextInput] | list[PreTokenizedInput] = None, **kwargs: Unpack[Ovis2_5ProcessorKwargs], ) -> BatchFeature: """ Main method to prepare for the model one or several sequences(s) and image(s). This method forwards the `text`and `kwargs` arguments to Qwen2TokenizerFast's [`~Qwen2TokenizerFast.__call__`] if `text` is not `None` to encode the text. To prepare the vision inputs, this method forwards the `vision_infos` and `kwrags` arguments to Qwen2VLImageProcessor's [`~Qwen2VLImageProcessor.__call__`] if `vision_infos` is not `None`. Args: images (`PIL.Image.Image`, `np.ndarray`, `torch.Tensor`, `list[PIL.Image.Image]`, `list[np.ndarray]`, `list[torch.Tensor]`): The image or batch of images to be prepared. Each image can be a PIL image, NumPy array or PyTorch tensor. Both channels-first and channels-last formats are supported. text (`str`, `list[str]`, `list[list[str]]`): The sequence or batch of sequences to be encoded. Each sequence can be a string or a list of strings (pretokenized string). If the sequences are provided as list of strings (pretokenized), you must set `is_split_into_words=True` (to lift the ambiguity with a batch of sequences). videos (`np.ndarray`, `torch.Tensor`, `list[np.ndarray]`, `list[torch.Tensor]`): The image or batch of videos to be prepared. Each video can be a 4D NumPy array or PyTorch tensor, or a nested list of 3D frames. Both channels-first and channels-last formats are supported. return_tensors (`str` or [`~utils.TensorType`], *optional*): If set, will return tensors of a particular framework. Acceptable values are: - `'tf'`: Return TensorFlow `tf.constant` objects. - `'pt'`: Return PyTorch `torch.Tensor` objects. - `'np'`: Return NumPy `np.ndarray` objects. - `'jax'`: Return JAX `jnp.ndarray` objects. Returns: [`BatchFeature`]: A [`BatchFeature`] with the following fields: - **input_ids** -- list of token ids to be fed to a model. Returned when `text` is not `None`. - **attention_mask** -- list of indices specifying which tokens should be attended to by the model (when `return_attention_mask=True` or if *"attention_mask"* is in `self.model_input_names` and if `text` is not `None`). - **pixel_values** -- Pixel values to be fed to a model. Returned when `images` is not `None`. - **pixel_values_videos** -- Pixel values of videos to be fed to a model. Returned when `videos` is not `None`. - **image_grid_thw** -- list of image 3D grid in LLM. Returned when `images` is not `None`. - **video_grid_thw** -- list of video 3D grid in LLM. Returned when `videos` is not `None`. - **second_per_grid_ts** -- list of video seconds per time grid. Returned when `videos` is not `None`. """ output_kwargs = self._merge_kwargs( Ovis2_5ProcessorKwargs, tokenizer_init_kwargs=self.tokenizer.init_kwargs, **kwargs, ) # Process all images first visual_features = {} output = BatchFeature() if images is not None: processed_images = [] image_placeholders_list = [] grids = [] # Process each image for image in images if isinstance(images, list) else [images]: pixel_values, image_placeholders, grid = self.preprocess_multidata( images=image, **output_kwargs["images_kwargs"], ) processed_images.append(pixel_values) image_placeholders_list.append(image_placeholders) grids.append(grid) # assign all processed images if processed_images: visual_features["image_placeholders"] = image_placeholders_list output["pixel_values"] = processed_images output["grids"] = grids if videos is not None: processed_videos = [] videos_placeholders_list = [] grids = [] # Process each video for video in videos if isinstance(videos, list) else [videos]: pixel_values, video_placeholders, grid = self.preprocess_multidata( video=video, **output_kwargs["videos_kwargs"], ) processed_videos.append(pixel_values) videos_placeholders_list.append(video_placeholders) grids.append(grid) # assign all processed videos if processed_videos: visual_features["video_placeholders"] = videos_placeholders_list output["video_pixel_values"] = processed_videos output["video_grids"] = grids # Process text input if text is not None: if not isinstance(text, list): text = [text] tokenized_batched_text = self._tokenize_with_visual_symbol(text) image_token_id = self.get_token_value("image_token") video_token_id = self.get_token_value("video_token") replaced_ids_list = [] image_idx = 0 video_idx = 0 for ids_tensor in tokenized_batched_text: has_image_tokens = ( image_token_id in ids_tensor and "image_placeholders" in visual_features and image_idx < len(visual_features["image_placeholders"]) ) has_video_tokens = ( video_token_id in ids_tensor and "video_placeholders" in visual_features and video_idx < len(visual_features["video_placeholders"]) ) if has_image_tokens or has_video_tokens: # Convert to list for easier manipulation ids_list = ids_tensor.tolist() new_ids = [] # Replace placeholders for token_id in ids_list: if token_id == image_token_id: new_ids.extend( visual_features["image_placeholders"][image_idx] ) image_idx += 1 elif token_id == video_token_id: new_ids.extend( visual_features["video_placeholders"][video_idx] ) video_idx += 1 else: new_ids.append(token_id) # Convert back to tensor ids_tensor = torch.tensor(new_ids, dtype=torch.long) replaced_ids_list.append(ids_tensor) if replaced_ids_list: replaced_and_tokenized_ids = torch.stack(replaced_ids_list) else: replaced_and_tokenized_ids = torch.tensor([], dtype=torch.long) output["input_ids"] = replaced_and_tokenized_ids return output # If only images were provided return BatchFeature(data=visual_features) def _tokenize_with_visual_symbol(self, text_list: list[str]) -> torch.LongTensor: batch_token_ids = [] for text in text_list: token_ids = [] video_token_id = self.get_token_value("video_token") image_token_id = self.get_token_value("image_token") video_split_texts = text.split(self.video_token) for j, video_segment in enumerate(video_split_texts): image_split_texts = video_segment.split(self.image_token) text_chunks = [ self.tokenizer(chunk, add_special_tokens=False).input_ids for chunk in image_split_texts ] segment_tokens = [] for i, chunk in enumerate(text_chunks): segment_tokens.extend(chunk) if i < len(text_chunks) - 1: segment_tokens.append(image_token_id) token_ids.extend(segment_tokens) if j < len(video_split_texts) - 1: token_ids.append(video_token_id) batch_token_ids.append(token_ids) return torch.tensor(batch_token_ids, dtype=torch.long) # Copied from qwen2_vl def smart_resize( self, height: int, width: int, factor: int = 28, min_pixels: int = MIN_PIXELS, max_pixels: int = MAX_PIXELS, ): """Rescales the image so that the following conditions are met: 1. Both dimensions (height and width) are divisible by 'factor'. 2. The total number of pixels is within the range ['min_pixels', 'max_pixels']. 3. The aspect ratio of the image is maintained as closely as possible. """ if height < factor or width < factor: print( f"height:{height} or width:{width} must be larger than factor:{factor}" ) if height < width: width = round(factor / height * width) height = factor else: height = round(factor / width * height) width = factor elif max(height, width) / min(height, width) > 200: print( f"absolute aspect ratio must be smaller than 200, " f"got {max(height, width) / min(height, width)}" ) if height > width: height = 200 * width else: width = 200 * height h_bar = round(height / factor) * factor w_bar = round(width / factor) * factor if h_bar * w_bar > max_pixels: beta = math.sqrt((height * width) / max_pixels) h_bar = math.floor(height / beta / factor) * factor w_bar = math.floor(width / beta / factor) * factor elif h_bar * w_bar < min_pixels: beta = math.sqrt(min_pixels / (height * width)) h_bar = math.ceil(height * beta / factor) * factor w_bar = math.ceil(width * beta / factor) * factor return h_bar, w_bar def get_token_value(self, tok): return self.extra_special_tokens[tok] def construct_visual_indicators(self, grid, is_video: bool = False): if is_video: start_token = self.get_token_value("video_start") end_token = self.get_token_value("video_end") else: start_token = self.get_token_value("image_start") end_token = self.get_token_value("image_end") image_placeholders = [start_token, self.get_token_value("visual_atom")] if grid[0] * grid[1] > 1: for r in range(grid[0]): for c in range(grid[1]): image_placeholders.append(self.get_token_value("visual_atom")) image_placeholders.append(end_token) return image_placeholders def construct_visual_placeholders(self, grid, is_video: bool = False): visual_placeholders = self.construct_visual_indicators((1, 1), is_video) image_atom_token_id = self.get_token_value("visual_atom") # Extract the padding token ID from tokenizer image_padding_token_id = self.get_token_value("image_pad") num_image_atoms = grid[0] * grid[1] * grid[2] num_image_atoms //= self.hidden_stride**2 num_image_atoms //= self.temporal_patch_size # Create a new list with padding tokens inserted padded_placeholder_tokens = [] for token in visual_placeholders: if token == image_atom_token_id: padded_placeholder_tokens.extend( [image_padding_token_id] * num_image_atoms ) else: padded_placeholder_tokens.append(image_padding_token_id) return padded_placeholder_tokens def preprocess_multidata( self, images: PIL.Image.Image | list[PIL.Image.Image] | None = None, video: list[PIL.Image.Image] | np.ndarray | None = None, do_convert_rgb: bool | None = True, min_pixels: int = MIN_PIXELS, max_pixels: int = MAX_PIXELS, return_tensors: str | None = "pt", ): is_video = False if images is not None: if not isinstance(images, list): images = [images] elif video is not None: is_video = True # type of vidoe in dummy_mm_data is np.ndarray if isinstance(video, np.ndarray): images = [] for i in range(video.shape[0]): image = PIL.Image.fromarray(video[i].astype(np.uint8)) images.append(image) elif isinstance(video, list): images = video else: raise ValueError("Either images or video should be provided.") min_pixels = min( max_pixels if max_pixels is not None else MAX_PIXELS, min_pixels if min_pixels is not None else MIN_PIXELS, ) images = [ image.convert("RGB") if do_convert_rgb and image.mode != "RGB" else image for image in images ] width, height = images[0].size resized_height, resized_width = height, width processed_images = [] for image in images: resized_height, resized_width = self.smart_resize( height, width, factor=self.patch_size * self.hidden_stride, min_pixels=min_pixels, max_pixels=max_pixels, ) new_size = dict(height=resized_height, width=resized_width) image_pt = self.image_processor.preprocess(image, size=new_size)[ "pixel_values" ][0] processed_images.append(image_pt) patches = np.array(processed_images) if patches.shape[0] % self.temporal_patch_size != 0: num_to_pad = self.temporal_patch_size - ( patches.shape[0] % self.temporal_patch_size ) repeats = np.repeat(patches[-1][np.newaxis], num_to_pad, axis=0) patches = np.concatenate([patches, repeats], axis=0) channel = patches.shape[1] grid_t = patches.shape[0] // self.temporal_patch_size grid_h = resized_height // self.patch_size grid_w = resized_width // self.patch_size patches = patches.reshape( grid_t, self.temporal_patch_size, channel, grid_h // self.hidden_stride, self.hidden_stride, self.patch_size, grid_w // self.hidden_stride, self.hidden_stride, self.patch_size, ) patches = patches.transpose(0, 3, 6, 4, 7, 2, 1, 5, 8) flatten_patches = patches.reshape( grid_t * grid_h * grid_w, channel * self.temporal_patch_size * self.patch_size * self.patch_size, ) visual_placeholders = self.construct_visual_placeholders( [grid_t, grid_h, grid_w], is_video ) return ( torch.tensor(flatten_patches), visual_placeholders, torch.tensor([[grid_t, grid_h, grid_w]]), ) AutoProcessor.register("Ovis2_5Processor", Ovis2_5Processor)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/transformers_utils/processors/ovis2_5.py", "license": "Apache License 2.0", "lines": 439, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/models/language/pooling/test_multilabel_classification_support.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from transformers import AutoModelForSequenceClassification @pytest.mark.parametrize( "model", ["Rami/multi-label-class-classification-on-github-issues"], ) @pytest.mark.parametrize("dtype", ["half"]) def test_classify_models( hf_runner, vllm_runner, example_prompts, model: str, dtype: str, ) -> None: with vllm_runner(model, max_model_len=512, dtype=dtype) as vllm_model: vllm_outputs = vllm_model.classify(example_prompts) with hf_runner( model, dtype=dtype, auto_cls=AutoModelForSequenceClassification ) as hf_model: hf_outputs = hf_model.classify(example_prompts) for hf_output, vllm_output in zip(hf_outputs, vllm_outputs): hf_output = torch.tensor(hf_output) vllm_output = torch.tensor(vllm_output) assert torch.allclose( hf_output, vllm_output, 1e-3 if dtype == "float" else 1e-2 )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/language/pooling/test_multilabel_classification_support.py", "license": "Apache License 2.0", "lines": 29, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/logits_processors/test_custom_offline.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any import pytest from tests.utils import create_new_process_for_each_test, set_random_seed from tests.v1.logits_processors.utils import ( DUMMY_LOGITPROC_ARG, DUMMY_LOGITPROC_FQCN, MAX_TOKENS, MODEL_NAME, POOLING_MODEL_NAME, TEMP_GREEDY, CustomLogitprocSource, DummyLogitsProcessor, WrappedPerReqLogitsProcessor, prompts, ) from tests.v1.logits_processors.utils import entry_points as fake_entry_points from vllm import LLM, SamplingParams from vllm.v1.sample.logits_processor import ( STR_POOLING_REJECTS_LOGITSPROCS, STR_SPEC_DEC_REJECTS_LOGITSPROCS, LogitsProcessor, ) # Create a mixture of requests which do and don't utilize the dummy logitproc sampling_params_list = [ SamplingParams( temperature=TEMP_GREEDY, max_tokens=MAX_TOKENS, extra_args={DUMMY_LOGITPROC_ARG: 128}, ), SamplingParams(temperature=TEMP_GREEDY, max_tokens=MAX_TOKENS), SamplingParams( temperature=TEMP_GREEDY, max_tokens=MAX_TOKENS, extra_args={DUMMY_LOGITPROC_ARG: 67}, ), SamplingParams(temperature=TEMP_GREEDY, max_tokens=MAX_TOKENS), ] def _run_test(kwargs: dict, logitproc_loaded: bool) -> None: """Compare `LLM` instance initialized with specified `kwargs` against reference `LLM` instance. Two scenarios: 1. Server has loaded dummy logitproc; test that requests which specify dummy logitproc arg value behave as if logitproc is operating (output token value should repeat), while requests that don't specify dummy logitproc arg value should match reference `LLM` output. 2. Server has *not* loaded dummy logitproc; test that all requests behave as if logitproc is *not* operating (output matches reference `LLM` output.) Args: kwargs: `LLM` constructor kwargs logitproc_loaded: server has loaded dummy logitproc if True """ # Create a vLLM instance and load custom logitproc llm_logitproc = LLM( model=MODEL_NAME, gpu_memory_utilization=0.1, **kwargs, ) # Create a reference vLLM instance without custom logitproc llm_ref = LLM(model=MODEL_NAME, gpu_memory_utilization=0.1) # Run inference with logitproc loaded outputs_logitproc = llm_logitproc.generate(prompts, sampling_params_list) # Reference run outputs_ref = llm_ref.generate(prompts, sampling_params_list) # Validate outputs for bdx, (out_lp, out_ref, params) in enumerate( zip(outputs_logitproc, outputs_ref, sampling_params_list) ): lp_toks = out_lp.outputs[0].token_ids if logitproc_loaded and params.extra_args: # This request exercises custom logitproc; validate that logitproc # forces `target_token` to be decoded in each step target_token = params.extra_args[DUMMY_LOGITPROC_ARG] if not all(x == target_token for x in lp_toks): raise AssertionError( f"Request {bdx} generated {lp_toks}, should all be {target_token}" ) else: # This request does not exercise custom logitproc (or custom # logitproc is not enabled on this server); validate against # reference result ref_toks = out_ref.outputs[0].token_ids if lp_toks != ref_toks: raise AssertionError( f"Request {bdx} generated {lp_toks}, should match {ref_toks}" ) @create_new_process_for_each_test() @pytest.mark.parametrize("logitproc_source", list(CustomLogitprocSource)) def test_custom_logitsprocs(monkeypatch, logitproc_source: CustomLogitprocSource): """Test offline Python interface for passing custom logitsprocs Construct an `LLM` instance which loads a custom logitproc that has a well-defined behavior (mask out all tokens except one `target_token`) Construct a reference `LLM` instance with no custom logitproc Pass in a batch of requests, 50% of which pass a `target_token` value in through `SamplingParams.extra_args`, 50% of which do not. Validate that * Requests which do not activate the custom logitproc, yield the same results for both `LLM` instances * Requests which activate the custom logitproc, only output `target_token` Test four scenarios, corresponding to `logitproc_source` value * No logitsprocs loaded - test that generated tokens match reference `LLM` instance output * Logitproc passed in via {entrypoint, class object, fully-qualified class name (FQCN)} - test that dummy logitproc is utilized correctly when provided via any of these three possible sources Args: monkeypatch: for setting env vars logitproc_source: what source (entrypoint, fully-qualified class name (FQCN), class object, or None) the user pulls the logitproc from """ # Test that logitproc info is passed to workers monkeypatch.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "1") set_random_seed(40) # Choose LLM args based on logitproc source if logitproc_source == CustomLogitprocSource.LOGITPROC_SOURCE_NONE: # Scenario: the server does not load any custom logitproc # Every other scenario is a different way of loading a custom logitproc _run_test({}, logitproc_loaded=False) return if logitproc_source == CustomLogitprocSource.LOGITPROC_SOURCE_ENTRYPOINT: # Scenario: vLLM loads a logitproc from a preconfigured entrypoint # To that end, mock a dummy logitproc entrypoint import importlib.metadata importlib.metadata.entry_points = fake_entry_points # type: ignore # fork is required for workers to see entrypoint patch monkeypatch.setenv("VLLM_WORKER_MULTIPROC_METHOD", "fork") _run_test({}, logitproc_loaded=True) return kwargs: dict[str, list[str | type[LogitsProcessor]]] = {} if logitproc_source == CustomLogitprocSource.LOGITPROC_SOURCE_FQCN: # Scenario: load logitproc based on fully-qualified class name (FQCN) kwargs["logits_processors"] = [DUMMY_LOGITPROC_FQCN] elif logitproc_source == CustomLogitprocSource.LOGITPROC_SOURCE_CLASS: # Scenario: load logitproc from provided class object kwargs["logits_processors"] = [DummyLogitsProcessor] _run_test(kwargs, logitproc_loaded=True) @create_new_process_for_each_test() def test_custom_logitsprocs_req(monkeypatch): """Test passing request-level logits processor to offline Python interface Wrap a request-level logits processor to create a batch level logits processor that has a well-defined behavior (mask out all tokens except one `target_token`) Construct an `LLM` instance which loads the wrapped logits processor. Pass the custom logitproc as a class object. Construct a reference `LLM` instance with no custom logitproc Pass in a batch of requests, 50% of which pass a `target_token` value in through `SamplingParams.extra_args`, 50% of which do not. Validate that * Requests which do not activate the custom logitproc, yield the same results for both `LLM` instances * Requests which activate the custom logitproc, only output `target_token` Args: monkeypatch: for setting env vars """ # Test that logitproc info is passed to workers monkeypatch.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "1") set_random_seed(40) _run_test( {"logits_processors": [WrappedPerReqLogitsProcessor]}, logitproc_loaded=True ) @create_new_process_for_each_test() @pytest.mark.parametrize("model_scenario", ["pooling", "spec_dec"]) @pytest.mark.parametrize( "logitproc_source", [ CustomLogitprocSource.LOGITPROC_SOURCE_ENTRYPOINT, CustomLogitprocSource.LOGITPROC_SOURCE_FQCN, CustomLogitprocSource.LOGITPROC_SOURCE_CLASS, ], ) def test_rejects_custom_logitsprocs( monkeypatch, model_scenario: str, logitproc_source: CustomLogitprocSource ): """Validate that vLLM engine initialization properly rejects custom logitsprocs when the model is a pooling model or speculative decoding enabled. Use `LLM` entrypoint. We expect `LLM` initialization to fail before the logitproc is actually loaded. Scenario 1: * Mock a logitproc entrypoint * Validate that `LLM` does not load the logitproc Scenario 2: * Pass custom logitproc to `LLM` constructor * Scenario 2a: via FQCN * Scenario 2b: via class object * Validate that initialization fails with appropriate exception Args: monkeypatch: used to set environment variables logitproc_source: what source (entrypoint, fully-qualified class name (FQCN), or class object) the user pulls the logitproc from """ monkeypatch.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "0") set_random_seed(40) test_params: dict[str, dict[str, Any]] = { "pooling": { "runner": "pooling", "model": POOLING_MODEL_NAME, "error_message": STR_POOLING_REJECTS_LOGITSPROCS, "speculative_config": None, }, "spec_dec": { "runner": "auto", "model": MODEL_NAME, "error_message": STR_SPEC_DEC_REJECTS_LOGITSPROCS, "speculative_config": {"model": "ngram", "num_speculative_tokens": 1}, }, } config = test_params[model_scenario] llm_kwargs: dict[str, Any] = { "runner": config["runner"], "model": config["model"], "gpu_memory_utilization": 0.1, "speculative_config": config["speculative_config"], } if logitproc_source == CustomLogitprocSource.LOGITPROC_SOURCE_ENTRYPOINT: # Scenario: vLLM loads a model and ignores a logitproc that is # available at a preconfigured entrypoint # Patch in dummy logitproc entrypoint import importlib.metadata importlib.metadata.entry_points = fake_entry_points # type: ignore # fork is required for entrypoint patch to be visible to workers, # although they should ignore the entrypoint patch anyway monkeypatch.setenv("VLLM_WORKER_MULTIPROC_METHOD", "fork") llm = LLM(**llm_kwargs) # Require that no custom logitsprocs have been loaded # (built-in processors may exist: MinTokensLogitsProcessor, # LogitBiasLogitsProcessor, MinPLogitsProcessor) worker = llm.llm_engine.model_executor.driver_worker.worker for proc in worker.model_runner.input_batch.logitsprocs.all: assert not isinstance(proc, DummyLogitsProcessor) return if logitproc_source == CustomLogitprocSource.LOGITPROC_SOURCE_FQCN: # Scenario: load logitproc based on fully-qualified class name (FQCN) llm_kwargs["logits_processors"] = [DUMMY_LOGITPROC_FQCN] elif logitproc_source == CustomLogitprocSource.LOGITPROC_SOURCE_CLASS: # Scenario: load logitproc from provided class object llm_kwargs["logits_processors"] = [DummyLogitsProcessor] with pytest.raises(ValueError, match=config["error_message"]): # Require that loading a model alongside the logitproc raises # the appropriate exception. LLM(**llm_kwargs)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/logits_processors/test_custom_offline.py", "license": "Apache License 2.0", "lines": 245, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/logits_processors/test_custom_online.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os import random import sys from typing import Any import openai import pytest import pytest_asyncio from tests.utils import RemoteOpenAIServerCustom, create_new_process_for_each_test from tests.v1.logits_processors.utils import ( DUMMY_LOGITPROC_ARG, DUMMY_LOGITPROC_FQCN, MAX_TOKENS, MODEL_NAME, TEMP_GREEDY, prompts, ) from tests.v1.logits_processors.utils import entry_points as fake_entry_points def _server_with_logitproc_entrypoint( env_dict: dict[str, str] | None, model: str, vllm_serve_args: list[str], ) -> None: """Start vLLM server, inject dummy logitproc entrypoint""" # Patch `entry_points` to inject logitproc entrypoint import importlib.metadata importlib.metadata.entry_points = fake_entry_points # type: ignore from vllm.entrypoints.cli import main # fork is required for workers to see entrypoint patch os.environ["VLLM_WORKER_MULTIPROC_METHOD"] = "fork" if env_dict is not None: os.environ.update(env_dict) # Emulate `vllm serve <model> <CLI args>` sys.argv = ["vllm", "serve", model] + vllm_serve_args main.main() def _server_with_logitproc_fqcn( env_dict: dict[str, str] | None, model: str, vllm_serve_args: list[str], ) -> None: """Start vLLM server, inject module with dummy logitproc""" from vllm.entrypoints.cli import main if env_dict is not None: os.environ.update(env_dict) # Emulate `vllm serve <model> <CLI args>` sys.argv = ["vllm", "serve", model] + vllm_serve_args main.main() @pytest.fixture(scope="module") def default_server_args(): return [ # use half precision for speed and memory savings in CI environment "--dtype", "bfloat16", "--max-model-len", "2048", "--max-num-seqs", "128", ] @pytest.fixture( scope="function", params=[[], ["--logits-processors", DUMMY_LOGITPROC_FQCN]] ) def server(default_server_args, request, monkeypatch): """Consider two server configurations: (1) --logits-processors cli arg specifies dummy logits processor via fully- qualified class name (FQCN); patch in a dummy logits processor module (2) No --logits-processors cli arg; patch in a dummy logits processor entrypoint """ # Test that logitproc info is passed to workers monkeypatch.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "1") if request.param: # Launch server, append FQCN argument, inject dummy logitproc module args = default_server_args + request.param _server_fxn = _server_with_logitproc_fqcn else: # Launch server, inject dummy logitproc entrypoint args = default_server_args _server_fxn = _server_with_logitproc_entrypoint with RemoteOpenAIServerCustom(MODEL_NAME, args, _server_fxn) as remote_server: yield remote_server @pytest_asyncio.fixture async def client(server): async with server.get_async_client() as async_client: yield async_client # General request argument values for these tests api_keyword_args = { # Greedy sampling ensures that requests which receive the `target_token` # arg will decode it in every step "temperature": TEMP_GREEDY, # Since EOS will never be decoded (unless `target_token` is EOS) "max_tokens": MAX_TOKENS, # Return decoded token logprobs (as a way of getting token id) "logprobs": 0, } @create_new_process_for_each_test() @pytest.mark.asyncio @pytest.mark.parametrize( "model_name", [MODEL_NAME], ) async def test_custom_logitsprocs(client: openai.AsyncOpenAI, model_name: str): """Test custom logitsprocs when starting OpenAI server from CLI Launch vLLM OpenAI-compatible server, configured to load a custom logitproc that has a well-defined behavior (mask out all tokens except one `target_token`). Pass in requests, 50% of which pass a `target_token` value in through `extra_body["vllm_xargs"]`, 50% of which do not. Validate that requests which activate the custom logitproc, repeat the same token """ use_dummy_logitproc = True for prompt in prompts: # Build request arguments request_keyword_args: dict[str, Any] = { **api_keyword_args, } if use_dummy_logitproc: # 50% of requests pass target_token custom arg target_token = random.choice([128, 67]) # For requests which activate the dummy logitproc, choose one of # two `target_token` values which are known not to be EOS tokens request_keyword_args["extra_body"] = { "vllm_xargs": {DUMMY_LOGITPROC_ARG: target_token} } batch = await client.completions.create( model=model_name, prompt=prompt, **request_keyword_args, ) if use_dummy_logitproc: # Only for requests which activate dummy logitproc - validate that # output token is repeated choices: openai.types.CompletionChoice = batch.choices toks = choices[0].logprobs.tokens if not all([x == toks[0] for x in toks]): raise AssertionError(f"Generated {toks} should all be {toks[0]}") # Alternate whether to activate dummy logitproc for each request use_dummy_logitproc = not use_dummy_logitproc @pytest.mark.asyncio @pytest.mark.parametrize( "model_name", [MODEL_NAME], ) async def test_invalid_custom_logitsproc_arg( client: openai.AsyncOpenAI, model_name: str ): """Test that request with invalid custom logitsproc is rejected""" prompt = "Hello, my name is" # Pass invalid (non-int) target_token value to dummy logits processor request_keyword_args: dict[str, Any] = { **api_keyword_args, "extra_body": { "vllm_xargs": {DUMMY_LOGITPROC_ARG: "invalid_target_token_value"} }, } with pytest.raises(openai.OpenAIError) as exc_info: await client.completions.create( model=model_name, prompt=prompt, **request_keyword_args, ) assert "is not int" in str(exc_info.value)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/logits_processors/test_custom_online.py", "license": "Apache License 2.0", "lines": 163, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/logits_processors/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import types from enum import Enum, auto from typing import Any import torch from vllm.config import VllmConfig from vllm.logger import init_logger from vllm.sampling_params import SamplingParams from vllm.v1.sample.logits_processor import ( LOGITSPROCS_GROUP, AdapterLogitsProcessor, BatchUpdate, LogitsProcessor, RequestLogitsProcessor, ) from vllm.v1.sample.logits_processor.builtin import process_dict_updates logger = init_logger(__name__) MODEL_NAME = "facebook/opt-125m" POOLING_MODEL_NAME = "BAAI/bge-base-en-v1.5" DUMMY_LOGITPROC_ARG = "target_token" TEMP_GREEDY = 0.0 MAX_TOKENS = 20 DUMMY_LOGITPROC_ENTRYPOINT = "dummy_logitproc" DUMMY_LOGITPROC_MODULE = "tests.v1.logits_processors.utils" DUMMY_LOGITPROC_FQCN = f"{DUMMY_LOGITPROC_MODULE}:DummyLogitsProcessor" class CustomLogitprocSource(Enum): """How to source a logitproc for testing purposes""" LOGITPROC_SOURCE_NONE = auto() # No custom logitproc LOGITPROC_SOURCE_ENTRYPOINT = auto() # Via entrypoint LOGITPROC_SOURCE_FQCN = auto() # Via fully-qualified class name (FQCN) LOGITPROC_SOURCE_CLASS = auto() # Via provided class object # Sample prompts. prompts = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] class DummyLogitsProcessor(LogitsProcessor): """Fake logit processor to support unit testing and examples""" @classmethod def validate_params(cls, params: SamplingParams): target_token: int | None = params.extra_args and params.extra_args.get( "target_token" ) if target_token is not None and not isinstance(target_token, int): raise ValueError( f"target_token value {target_token} {type(target_token)} is not int" ) def __init__( self, vllm_config: "VllmConfig", device: torch.device, is_pin_memory: bool ): self.req_info: dict[int, int] = {} def is_argmax_invariant(self) -> bool: """Never impacts greedy sampling""" return False def update_state(self, batch_update: BatchUpdate | None): def extract_extra_arg(params: SamplingParams) -> int | None: self.validate_params(params) return params.extra_args and params.extra_args.get("target_token") process_dict_updates( self.req_info, batch_update, lambda params, _, __: extract_extra_arg(params), ) def apply(self, logits: torch.Tensor) -> torch.Tensor: if not self.req_info: return logits # Save target values before modification cols = torch.tensor( list(self.req_info.values()), dtype=torch.long, device=logits.device ) rows = torch.tensor( list(self.req_info.keys()), dtype=torch.long, device=logits.device ) values_to_keep = logits[rows, cols].clone() # Mask all but target tokens logits[rows] = float("-inf") logits[rows, cols] = values_to_keep return logits """Dummy module with dummy logitproc class""" dummy_module = types.ModuleType(DUMMY_LOGITPROC_MODULE) dummy_module.DummyLogitsProcessor = DummyLogitsProcessor # type: ignore class EntryPoint: """Dummy entrypoint class for logitsprocs testing""" def __init__(self): self.name = DUMMY_LOGITPROC_ENTRYPOINT self.value = DUMMY_LOGITPROC_FQCN def load(self): return DummyLogitsProcessor class EntryPoints(list): """Dummy EntryPoints class for logitsprocs testing""" def __init__(self, group: str): # Emulate list-like functionality eps = [EntryPoint()] if group == LOGITSPROCS_GROUP else [] super().__init__(eps) # Extra attributes self.names = [ep.name for ep in eps] class DummyPerReqLogitsProcessor: """The request-level logits processor masks out all logits except the token id identified by `target_token`""" def __init__(self, target_token: int) -> None: """Specify `target_token`""" self.target_token = target_token def __call__( self, output_ids: list[int], logits: torch.Tensor, ) -> torch.Tensor: val_to_keep = logits[self.target_token].item() logits[:] = float("-inf") logits[self.target_token] = val_to_keep return logits class WrappedPerReqLogitsProcessor(AdapterLogitsProcessor): """Example of wrapping a fake request-level logit processor to create a batch-level logits processor""" def is_argmax_invariant(self) -> bool: return False def new_req_logits_processor( self, params: SamplingParams, ) -> RequestLogitsProcessor | None: """This method returns a new request-level logits processor, customized to the `target_token` value associated with a particular request. Returns None if the logits processor should not be applied to the particular request. To use the logits processor the request must have a "target_token" custom argument with an integer value. Args: params: per-request sampling params Returns: `Callable` request logits processor, or None """ target_token: Any | None = params.extra_args and params.extra_args.get( "target_token" ) if target_token is None: return None if not isinstance(target_token, int): logger.warning( "target_token value %s is not int; not applying logits" " processor to request.", target_token, ) return None return DummyPerReqLogitsProcessor(target_token) """Fake version of importlib.metadata.entry_points""" entry_points = lambda group: EntryPoints(group)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/logits_processors/utils.py", "license": "Apache License 2.0", "lines": 148, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/v1/sample/logits_processor/interface.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import ABC, abstractmethod from collections.abc import Sequence from dataclasses import dataclass from enum import Enum, auto from typing import TYPE_CHECKING import torch from vllm import SamplingParams if TYPE_CHECKING: from vllm.config import VllmConfig class MoveDirectionality(Enum): # One-way i1->i2 req move within batch UNIDIRECTIONAL = auto() # Two-way i1<->i2 req swap within batch SWAP = auto() # Batch indices of any removed requests. RemovedRequest = int # (index, params, prompt_tok_ids, output_tok_ids) tuples for new # requests added to the batch. AddedRequest = tuple[int, SamplingParams, list[int] | None, list[int]] # (index 1, index 2, directionality) tuples representing # one-way moves or two-way swaps of requests in batch MovedRequest = tuple[int, int, MoveDirectionality] @dataclass(frozen=True) class BatchUpdate: """Persistent batch state change info for logitsprocs""" batch_size: int # Current num reqs in batch # Metadata for requests added to, removed from, and moved # within the persistent batch. # # Key assumption: the `output_tok_ids` list (which is an element of each # tuple in `added`) is a reference to the request's running output tokens # list; via this reference, the logits processors always see the latest # list of generated output tokens. # # NOTE: # * Added or moved requests may replace existing requests with the same # index. # * Operations should be processed in the following order: # - removed, added, moved removed: Sequence[RemovedRequest] added: Sequence[AddedRequest] moved: Sequence[MovedRequest] class LogitsProcessor(ABC): @classmethod def validate_params(cls, sampling_params: SamplingParams): """Validate sampling params for this logits processor. Raise ValueError for invalid ones. """ return None @abstractmethod def __init__( self, vllm_config: "VllmConfig", device: torch.device, is_pin_memory: bool ) -> None: raise NotImplementedError @abstractmethod def apply(self, logits: torch.Tensor) -> torch.Tensor: """Apply LogitsProcessor to batch logits tensor. The updated tensor must be returned but may be modified in-place. """ raise NotImplementedError @abstractmethod def is_argmax_invariant(self) -> bool: """True if logits processor has no impact on the argmax computation in greedy sampling. NOTE: may or may not have the same value for all instances of a given LogitsProcessor subclass, depending on subclass implementation. """ raise NotImplementedError @abstractmethod def update_state( self, batch_update: "BatchUpdate | None", ) -> None: """Called when there are new output tokens, prior to each forward pass. Args: batch_update: Non-None iff there have been changes to the batch makeup. """ raise NotImplementedError
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/sample/logits_processor/interface.py", "license": "Apache License 2.0", "lines": 84, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/sample/logits_processor/state.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Iterable, Iterator from itertools import chain from typing import TYPE_CHECKING from vllm.v1.sample.logits_processor.interface import ( AddedRequest, BatchUpdate, MovedRequest, RemovedRequest, ) if TYPE_CHECKING: from vllm.v1.sample.logits_processor.interface import LogitsProcessor class BatchUpdateBuilder: """Helps track persistent batch state changes and build a batch update data structure for logitsprocs Assumptions: * All information about requests removed from persistent batch during a step is aggregated in self._removed through calls to self.removed_append() at the beginning of a step. This must happen before the first time that self.removed, self.pop_removed() or self.peek_removed() are invoked in a given step * After the first time that self.removed, self.pop_removed() or self.peek_removed() are read in a step, no new removals are registered using self.removed_append() * Elements of self._removed are never directly modified, added or removed (i.e. modification is only via self.removed_append() and self.pop_removed()) Guarantees under above assumptions: * self.removed is always sorted in descending order * self.pop_removed() and self.peek_removed() both return the lowest removed request index in the current step """ _removed: list[RemovedRequest] _is_removed_sorted: bool added: list[AddedRequest] moved: list[MovedRequest] def __init__( self, removed: list[RemovedRequest] | None = None, added: list[AddedRequest] | None = None, moved: list[MovedRequest] | None = None, ) -> None: self._removed = removed or [] self.added = added or [] self.moved = moved or [] self._is_removed_sorted = False # Used to track changes in the pooling case # where we don't populate the added list. self.batch_changed = False def _ensure_removed_sorted(self) -> None: """Sort removed request indices in descending order. Idempotent after first call in a given step, until reset. """ if not self._is_removed_sorted: self._removed.sort(reverse=True) self._is_removed_sorted = True @property def removed(self) -> list[RemovedRequest]: """Removed request indices sorted in descending order""" self._ensure_removed_sorted() return self._removed def removed_append(self, index: int) -> None: """Register the removal of a request from the persistent batch. Must not be called after the first time self.removed, self.pop_removed() or self.peek_removed() are invoked. Args: index: request index """ if self._is_removed_sorted: raise RuntimeError( "Cannot register new removed request after self.removed has been read." ) self._removed.append(index) self.batch_changed = True def has_removed(self) -> bool: return bool(self._removed) def peek_removed(self) -> int | None: """Return lowest removed request index""" if self.has_removed(): self._ensure_removed_sorted() return self._removed[-1] return None def pop_removed(self) -> int | None: """Pop lowest removed request index""" if self.has_removed(): self._ensure_removed_sorted() return self._removed.pop() return None def reset(self) -> bool: """Returns True if there were any changes to the batch.""" self._is_removed_sorted = False self._removed.clear() self.added.clear() self.moved.clear() batch_changed = self.batch_changed self.batch_changed = False return batch_changed def get_and_reset(self, batch_size: int) -> BatchUpdate | None: """Generate a logitsprocs batch update data structure and reset internal batch update builder state. Args: batch_size: current persistent batch size Returns: Frozen logitsprocs batch update instance; `None` if no updates """ # Reset removal-sorting logic self._is_removed_sorted = False self.batch_changed = False if not any((self._removed, self.moved, self.added)): # No update; short-circuit return None # Build batch state update batch_update = BatchUpdate( batch_size=batch_size, removed=self._removed, moved=self.moved, added=self.added, ) self._removed = [] self.moved = [] self.added = [] return batch_update class LogitsProcessors: """Encapsulates initialized logitsproc objects.""" def __init__(self, logitsprocs: Iterable["LogitsProcessor"] | None = None) -> None: self.argmax_invariant: list[LogitsProcessor] = [] self.non_argmax_invariant: list[LogitsProcessor] = [] if logitsprocs: for logitproc in logitsprocs: ( self.argmax_invariant if logitproc.is_argmax_invariant() else self.non_argmax_invariant ).append(logitproc) @property def all(self) -> Iterator["LogitsProcessor"]: """Iterator over all logits processors.""" return chain(self.argmax_invariant, self.non_argmax_invariant)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/sample/logits_processor/state.py", "license": "Apache License 2.0", "lines": 142, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/entrypoints/openai/tool_parsers/test_hermes_tool_parser.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json import pytest from vllm.entrypoints.openai.chat_completion.protocol import ChatCompletionRequest from vllm.tokenizers import TokenizerLike from vllm.tool_parsers.hermes_tool_parser import Hermes2ProToolParser from ....utils import RemoteOpenAIServer MODEL_NAME = "meta-llama/Llama-3.2-1B-Instruct" LORA_MODEL = "minpeter/LoRA-Llama-3.2-1B-tool-vllm-ci" SERVER_ARGS = [ "--enforce-eager", "--enable-auto-tool-choice", "--tool-call-parser", "hermes", "--enable-lora", "--lora-modules", f"{LORA_MODEL}={LORA_MODEL}", "--tokenizer", f"{LORA_MODEL}", ] TOOLS = [ { "type": "function", "function": { "name": "get_current_weather", "description": "Get the current weather in a given location", "parameters": { "type": "object", "properties": { "location": { "type": "string", "description": "The city and state, e.g. San Francisco, CA", }, "unit": { "type": "string", "enum": ["celsius", "fahrenheit"], }, }, "required": ["location"], }, }, } ] PRODUCT_TOOLS = [ { "type": "function", "function": { "name": "get_product_info", "description": "Get detailed information of a product based on its " "product ID.", "parameters": { "type": "object", "properties": { "inserted": { "type": "boolean", "description": "inserted.", }, "product_id": { "type": "integer", "description": "The product ID of the product.", }, }, "required": ["product_id", "inserted"], }, }, } ] MESSAGES = [{"role": "user", "content": "What's the weather like in Boston?"}] PRODUCT_MESSAGES = [ { "role": "user", "content": "Hi! Do you have any detailed information about the product id " "7355608 and inserted true?", } ] @pytest.mark.asyncio async def test_non_streaming_tool_call(): """Test tool call in non-streaming mode.""" with RemoteOpenAIServer(MODEL_NAME, SERVER_ARGS) as server: client = server.get_async_client() response = await client.chat.completions.create( model=LORA_MODEL, messages=MESSAGES, tools=TOOLS, tool_choice="auto", temperature=0.0, ) assert response.choices choice = response.choices[0] message = choice.message assert choice.finish_reason == "tool_calls" assert message.tool_calls is not None tool_call = message.tool_calls[0] assert tool_call.type == "function" assert tool_call.function.name == "get_current_weather" arguments = json.loads(tool_call.function.arguments) assert "location" in arguments assert "Boston" in arguments["location"] print("\n[Non-Streaming Test Passed]") print(f"Tool Call: {tool_call.function.name}") print(f"Arguments: {arguments}") @pytest.mark.asyncio async def test_streaming_tool_call(): """Test tool call in streaming mode.""" with RemoteOpenAIServer(MODEL_NAME, SERVER_ARGS) as server: client = server.get_async_client() stream = await client.chat.completions.create( model=LORA_MODEL, messages=MESSAGES, tools=TOOLS, tool_choice="auto", temperature=0.0, stream=True, ) tool_call_chunks = {} async for chunk in stream: if not chunk.choices: continue delta = chunk.choices[0].delta if not delta or not delta.tool_calls: continue for tool_chunk in delta.tool_calls: index = tool_chunk.index if index not in tool_call_chunks: tool_call_chunks[index] = {"name": "", "arguments": ""} if tool_chunk.function.name: tool_call_chunks[index]["name"] += tool_chunk.function.name if tool_chunk.function.arguments: tool_call_chunks[index]["arguments"] += ( tool_chunk.function.arguments ) assert len(tool_call_chunks) == 1 reconstructed_tool_call = tool_call_chunks[0] assert reconstructed_tool_call["name"] == "get_current_weather" arguments = json.loads(reconstructed_tool_call["arguments"]) assert "location" in arguments assert "Boston" in arguments["location"] print("\n[Streaming Test Passed]") print(f"Reconstructed Tool Call: {reconstructed_tool_call['name']}") print(f"Reconstructed Arguments: {arguments}") @pytest.mark.asyncio async def test_non_streaming_product_tool_call(): """Test tool call integer and boolean parameters in non-streaming mode.""" with RemoteOpenAIServer(MODEL_NAME, SERVER_ARGS) as server: client = server.get_async_client() response = await client.chat.completions.create( model=LORA_MODEL, messages=PRODUCT_MESSAGES, tools=PRODUCT_TOOLS, tool_choice="auto", temperature=0.66, ) assert response.choices choice = response.choices[0] message = choice.message assert choice.finish_reason == "tool_calls" assert message.tool_calls is not None tool_call = message.tool_calls[0] assert tool_call.type == "function" assert tool_call.function.name == "get_product_info" arguments = json.loads(tool_call.function.arguments) assert "product_id" in arguments assert "inserted" in arguments product_id = arguments.get("product_id") inserted = arguments.get("inserted") assert isinstance(product_id, int) assert product_id == 7355608 assert isinstance(inserted, bool) assert inserted is True print("\n[Non-Streaming Product Test Passed]") print(f"Tool Call: {tool_call.function.name}") print(f"Arguments: {arguments}") @pytest.mark.asyncio async def test_streaming_product_tool_call(): """Test tool call integer and boolean parameters in streaming mode.""" with RemoteOpenAIServer(MODEL_NAME, SERVER_ARGS) as server: client = server.get_async_client() stream = await client.chat.completions.create( model=LORA_MODEL, messages=PRODUCT_MESSAGES, tools=PRODUCT_TOOLS, tool_choice="auto", temperature=0.66, stream=True, ) tool_call_chunks = {} async for chunk in stream: if not chunk.choices: continue delta = chunk.choices[0].delta if not delta or not delta.tool_calls: continue for tool_chunk in delta.tool_calls: index = tool_chunk.index if index not in tool_call_chunks: tool_call_chunks[index] = {"name": "", "arguments": ""} if tool_chunk.function.name: tool_call_chunks[index]["name"] += tool_chunk.function.name if tool_chunk.function.arguments: tool_call_chunks[index]["arguments"] += ( tool_chunk.function.arguments ) assert len(tool_call_chunks) == 1 reconstructed_tool_call = tool_call_chunks[0] assert reconstructed_tool_call["name"] == "get_product_info" arguments = json.loads(reconstructed_tool_call["arguments"]) assert "product_id" in arguments assert "inserted" in arguments # Handle type coercion for streaming test as well product_id = arguments.get("product_id") inserted = arguments.get("inserted") assert isinstance(product_id, int) assert product_id == 7355608 assert isinstance(inserted, bool) assert inserted is True print("\n[Streaming Product Test Passed]") print(f"Reconstructed Tool Call: {reconstructed_tool_call['name']}") print(f"Reconstructed Arguments: {arguments}") @pytest.fixture def qwen_tokenizer() -> TokenizerLike: from vllm.tokenizers import get_tokenizer return get_tokenizer("Qwen/Qwen3-32B") @pytest.fixture def hermes_parser(qwen_tokenizer: TokenizerLike) -> Hermes2ProToolParser: return Hermes2ProToolParser(qwen_tokenizer) @pytest.fixture def any_chat_request() -> ChatCompletionRequest: return ChatCompletionRequest( seed=42, model="Qwen/Qwen3-32B", messages=[], ) def test_hermes_parser_streaming_just_forward_text( qwen_tokenizer: TokenizerLike, hermes_parser: Hermes2ProToolParser, any_chat_request: ChatCompletionRequest, ) -> None: text = """This is some prior text that has nothing to do with tool calling.""" tokens = qwen_tokenizer.encode(text) previous_text = "" delta_messages = [] for token in tokens: delta_text = qwen_tokenizer.decode([token]) current_text = previous_text + delta_text delta = hermes_parser.extract_tool_calls_streaming( previous_text=previous_text, current_text=current_text, delta_text=delta_text, previous_token_ids=[], current_token_ids=[], delta_token_ids=[], request=any_chat_request, ) previous_text = current_text delta_messages.append(delta) for delta in delta_messages: assert delta is not None assert not delta.tool_calls print(delta_messages) assert "".join([delta.content for delta in delta_messages]) == text def test_hermes_parser_streaming_failure_case_bug_19056( qwen_tokenizer: TokenizerLike, hermes_parser: Hermes2ProToolParser, any_chat_request: ChatCompletionRequest, ) -> None: text = """<tool_call> {"name": "final_answer", "arguments": {"trigger": true}} </tool_call>""" tokens = qwen_tokenizer.encode(text) previous_text = "" delta_messages = [] for token in tokens: text = qwen_tokenizer.decode([token]) current_text = previous_text + text delta = hermes_parser.extract_tool_calls_streaming( previous_text=previous_text, current_text=current_text, delta_text=text, previous_token_ids=[], current_token_ids=[], delta_token_ids=[], request=any_chat_request, ) previous_text = current_text if delta is not None: delta_messages.append(delta) assert delta_messages[0].tool_calls[0].function.name == "final_answer" tool_call_args = "".join( delta.tool_calls[0].function.arguments or "" for delta in delta_messages ) assert tool_call_args == '{"trigger": true}' def test_hermes_parser_streaming( qwen_tokenizer: TokenizerLike, hermes_parser: Hermes2ProToolParser, any_chat_request: ChatCompletionRequest, ) -> None: text = '<tool_call>\ {"name": "get_current_temperature",\ "arguments": {"location":\ "San Francisco, California, United States", "unit": "celsius"}}\ </tool_call>' tokens = qwen_tokenizer.encode(text) previous_text = "" delta_messages = [] for token in tokens: text = qwen_tokenizer.decode([token]) current_text = previous_text + text delta = hermes_parser.extract_tool_calls_streaming( previous_text=previous_text, current_text=current_text, delta_text=text, previous_token_ids=[], current_token_ids=[], delta_token_ids=[], request=any_chat_request, ) previous_text = current_text if delta is not None: delta_messages.append(delta) print(delta_messages) assert delta_messages[0].tool_calls[0].function.name == "get_current_temperature" tool_call_args = "".join( delta.tool_calls[0].function.arguments or "" for delta in delta_messages ) assert tool_call_args == ( '{"location":"San Francisco, California, United States", "unit": "celsius"}' ) def test_hermes_parser_non_streaming_no_tool_call( hermes_parser: Hermes2ProToolParser, any_chat_request: ChatCompletionRequest, ) -> None: text = """This is not a tool call.""" tool_call = hermes_parser.extract_tool_calls( model_output=text, request=any_chat_request, ) assert tool_call is not None assert not tool_call.tools_called def test_hermes_parser_non_streaming_tool_call_between_tags( hermes_parser: Hermes2ProToolParser, any_chat_request: ChatCompletionRequest, ) -> None: text = """<tool_call> {"name": "final_answer", "arguments": {"trigger": true}} </tool_call>""" tool_call = hermes_parser.extract_tool_calls( model_output=text, request=any_chat_request, ) assert tool_call is not None assert tool_call.tools_called assert tool_call.tool_calls[0].function.name == "final_answer" assert tool_call.tool_calls[0].function.arguments == '{"trigger": true}' def test_hermes_parser_non_streaming_tool_call_until_eos( hermes_parser: Hermes2ProToolParser, any_chat_request: ChatCompletionRequest, ) -> None: text = """<tool_call> {"name": "final_answer", "arguments": {"trigger": true}}""" tool_call = hermes_parser.extract_tool_calls( model_output=text, request=any_chat_request, ) assert tool_call is not None assert tool_call.tools_called assert tool_call.tool_calls[0].function.name == "final_answer" assert tool_call.tool_calls[0].function.arguments == '{"trigger": true}' def test_hermes_parser_non_streaming_tool_call_invalid_json( hermes_parser: Hermes2ProToolParser, any_chat_request: ChatCompletionRequest, ) -> None: # Missing closing brace to trigger exception text = """<tool_call> {"name": "final_answer", "arguments": {"trigger": true}""" tool_call = hermes_parser.extract_tool_calls( model_output=text, request=any_chat_request, ) assert tool_call is not None assert not tool_call.tools_called
{ "repo_id": "vllm-project/vllm", "file_path": "tests/entrypoints/openai/tool_parsers/test_hermes_tool_parser.py", "license": "Apache License 2.0", "lines": 381, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tools/profiler/nsys_profile_tools/gputrc2graph.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ This generates gpu kernel analysis output from nsys rep. Will call nsys stats -r cuda_gpu_kern_trace, get non-overlapped gpu cycles, then generate csv and html output for analysis """ import argparse import logging import os import regex as re logger = logging.getLogger(__name__) # helper data class for annotating kernels def load_engine_model(): """returns engine_model built from all json files in the current dir""" import glob import json engine_model = {} json_files = glob.glob(os.path.join(os.path.dirname(__file__) or ".", "*.json")) for fname in json_files: with open(fname, encoding="utf-8") as f: engine_model.update(json.load(f)) return engine_model class GPUTrace2Graph: """ Parses output of nsys report, generates csv and bar chart output """ def __init__(self): import pandas as pd # avoid importing till needed self.pd = pd self.pd.options.mode.copy_on_write = True # helper functions for generating trace->summary csvs def gen_nonoverlapped_sum_from_gputrace(self, in_file, out_file): logger.info("loading %s", in_file) df = self.pd.read_csv( in_file, usecols=["Start (ns)", "Duration (ns)", "Device", "Strm", "Name"] ) df["End (ns)"] = df["Start (ns)"] + df["Duration (ns)"] df = self.sum_non_overlapping_intervals(df) # get ready to print table with elapsed times per kernel df["Instances"] = 1 df_sum = df.groupby("Name", as_index=False).agg( {"Elapsed Time (ns)": "sum", "Duration (ns)": "sum", "Instances": "size"} ) # generate csv df_sum["Total Time (sec)"] = df_sum["Duration (ns)"] / 1e9 df_sum["Elapsed Time (sec)"] = df_sum["Elapsed Time (ns)"] / 1e9 df_sum = df_sum.sort_values(by="Elapsed Time (sec)", ascending=False) df_sum[["Elapsed Time (sec)", "Total Time (sec)", "Instances", "Name"]].to_csv( out_file, index=False ) def sum_non_overlapping_intervals(self, df): """ returns new sorted df with Elapsed Time (ns) column using vectorized operations """ logger.info("sorting %s trace records by start time", str(df.shape)) # Sort by start time and reset index df = df.sort_values(by="Start (ns)").reset_index(drop=True) # Initialize elapsed time as duration df["Elapsed Time (ns)"] = df["Duration (ns)"] # Get numpy arrays for faster operations starts = df["Start (ns)"].values ends = df["End (ns)"].values # Keep track of current interval end current_end = ends[0] display_units = int(len(df) / 100) # Update current_end for overlapping intervals for i in range(1, len(df)): if i % display_units == 0: print(f"processing trace: {int(i / len(df) * 100)} %", end="\r") if starts[i] <= current_end: if ends[i] > current_end: # Partial overlap df.iloc[i, df.columns.get_loc("Elapsed Time (ns)")] = ( ends[i] - current_end ) current_end = ends[i] else: # Complete overlap df.iloc[i, df.columns.get_loc("Elapsed Time (ns)")] = 0 else: # No overlap current_end = ends[i] return df # functions for generating html files def make_html(self, df, output_dir, title): """make html graph from df""" import plotly.express as px if df.empty: return output_name = output_dir + "/result" if not title: title = "Model_Engine" x = "Model_Engine" y = "Elapsed Time (sec)" color = "Category" """ generate kernel mapping table """ # Sort Model_Engine categories by last field after underscore df["Model_Engine"] = self.pd.Categorical( df["Model_Engine"], sorted(df["Model_Engine"].unique(), key=lambda x: x.split("_")[-1]), ) df[["Model_Engine", color, "Instances", "Name", y]].sort_values( by=color ).to_csv(f"{output_name}.csv", index=False) graph = px.histogram( df.round(2), x=x, y=y, title=(f"{y} for {title}"), color=color, text_auto=True, ) # wrap x axis labels graph.update_xaxes(automargin=True) graph.write_html(f"{output_name}.html") """ Generate data table with columns per Model_Engine into result.html """ pivot_df = df.pivot_table( values="Elapsed Time (sec)", index="Category", columns="Model_Engine", aggfunc="sum", observed=False, ).round(2) # Add sum row at bottom pivot_df.loc["total_elapsed_sec"] = pivot_df.sum() pivot_df.fillna("").to_html("temp.html") with ( open(f"{output_name}.html", "a", encoding="utf-8") as outfile, open("temp.html", encoding="utf-8") as infile, ): outfile.write(infile.read()) os.remove("temp.html") print( f"Finished generating: \n" f" {output_name}.html for stack bar chart \n" f" {output_name}.csv for Kernel-Category mapping" ) def anno_gpu_kernname(self, df, mapping): """add "Category" column""" def anno_gpu_kernname_helper(name): for kern_name, val in mapping.items(): if re.search(kern_name, name): return val df["Category"] = df["Name"].apply(anno_gpu_kernname_helper) def make_nongpu_row(self, df, nongpu_sec): """this will append non-gpu time entry at end of df""" nongpu_row = self.pd.DataFrame([df.iloc[-1]]) nongpu_row["Category"] = nongpu_row["Name"] = "CPU(non-GPU)" nongpu_row["Instances"] = 1 nongpu_row["Elapsed Time (sec)"] = nongpu_sec return nongpu_row def is_valid_file(self, base_file): """asserts if base_file is non-existent or is empty""" assert os.path.isfile(base_file) and os.path.getsize(base_file) > 0, ( f"{base_file} doesn't exist or is empty" ) def should_gen_file(self, new_file, base_file): """figure out if new file should be generated from base_file""" self.is_valid_file(base_file) if ( os.path.exists(new_file) and (os.path.getmtime(new_file) > os.path.getmtime(base_file)) and (os.path.getsize(base_file) > 0) ): logger.info("reusing %s", new_file) return False else: logger.info("generating %s", new_file) return True def gen_sum_file(self, file, nsys_cmd): """ generates sum file from nsys trace with times per kernel and returns the name of the sum file """ import subprocess file_dir = os.path.dirname(file) file_name = os.path.basename(file) if not file_dir: file_dir = "." # Walk through trace and get the total non-overlapped time nsys_stats_file = f"{file_dir}/{file_name}_cuda_gpu_trace.csv" sum_file = f"{file_dir}/{file_name}_cuda_gpu_kernel_tracesum.csv" if self.should_gen_file(nsys_stats_file, file): cmd = [ nsys_cmd, "stats", "-r", "cuda_gpu_trace", file, "-o", f"{file_dir}/{file_name}", ] cmd_str = " ".join(cmd) logger.info("+ %s", cmd_str) # estimate time based on calibrated 240M/min file_size_mb = os.path.getsize(file) / 1e6 logger.info( "nsys stats for %.2f MB file expected to take %.2f min", file_size_mb, file_size_mb / 240, ) try: subprocess.run(cmd, check=True) except Exception: logger.error("%s failed; Use --nsys_cmd to specify nsys path", cmd_str) exit(1) logger.info("generating non-overalapped sum %s", sum_file) self.gen_nonoverlapped_sum_from_gputrace(nsys_stats_file, sum_file) self.is_valid_file(sum_file) logger.info("Finished generating %s", sum_file) return sum_file def gen_graph(self, in_file, out_dir, title, nsys_cmd, engine_model): """generates graph and csv file from in_file into out_dir""" # Initialize an empty DataFrame to store combined data combined_df = self.pd.DataFrame() for idx, (file, engine, model, total_sec) in enumerate(in_file): file_dir = os.path.dirname(file) file_name = os.path.basename(file) if not file_dir: file_dir = "." sum_file = self.gen_sum_file(file, nsys_cmd) # read kernel summary file df = self.pd.read_csv(sum_file) # annotate kernel to their categories assert engine_model.get(engine), f"engine {engine} unknown" assert engine_model[engine].get(model), f"model {model} unknown" # remove nsys-rep from file_name for shorter x-label file_name = file_name.replace(".nsys-rep", "") df["Model_Engine"] = f"{model}_{engine}_{file_name}_{idx}" self.anno_gpu_kernname(df, engine_model[engine][model]) # patch in non-gpu time gpu_sec = round(df["Elapsed Time (sec)"].sum(), 1) total_sec = round(float(total_sec), 1) if total_sec < gpu_sec: logger.warning( "Elapsed sec %.2f < GPU sec %.2f resetting Elapsed sec ", total_sec, gpu_sec, ) total_sec = gpu_sec nongpu_row = self.make_nongpu_row(df, total_sec - gpu_sec) df = self.pd.concat([df, nongpu_row], ignore_index=True) combined_df = self.pd.concat([combined_df, df], ignore_index=True) if out_dir is None: out_dir = "." else: os.makedirs(out_dir, exist_ok=True) # generate html file self.make_html(combined_df, out_dir, title) def parse_tuple(s): return tuple(s.split(",")) def main(): logging.basicConfig( format=("%(asctime)s - %(levelname)s - %(message)s"), level=logging.INFO ) parser = argparse.ArgumentParser( description=( "Process nsys rep and generate kernel non-overlapped cycles. \n" "Example:\n" "gputrc2graph.py --in_file d1.nsys-rep,vllm,llama,100 \n" "d2.nsys-rep,vllm,gpt-oss,102 " '--out_dir results/ --title "Model=gpt-oss vLLM chart"' ), formatter_class=argparse.RawDescriptionHelpFormatter, ) # load supported engine_model engine_model_supported = load_engine_model() # Get a string representation of supported engine/model combinations engine_model_supported_str = ", ".join( f"{engine}:[{', '.join(models.keys())}]" for engine, models in engine_model_supported.items() ) parser.add_argument( "--in_file", type=parse_tuple, nargs="+", help=( "list of (nsys-rep, engine, model, elapsed_nonprofiled_sec) " "separated by space. Elapsed_nonprofiled_sec is runtime without " "profiling used to calculate non-gpu time. Specify 0 to use " "elapsed time from nsys-rep but that might inflate non-gpu time. " f"Available engine:[model] are: {engine_model_supported_str} " f"Example: --infile d1.nsys-rep,vllm,llama,100 " "d2.nsys-rep,vllm,gpt-oss,102" ), required=True, ) parser.add_argument("--out_dir", help=("output dir for result.csv/html")) parser.add_argument("--title", help=("title for html chart")) parser.add_argument( "--nsys_cmd", help=("nsys cmd, e.g. /usr/bin/nsys, Default: nsys"), default="nsys", ) args = parser.parse_args() gputrace = GPUTrace2Graph() gputrace.gen_graph( args.in_file, args.out_dir, args.title, args.nsys_cmd, engine_model_supported ) if __name__ == "__main__": main()
{ "repo_id": "vllm-project/vllm", "file_path": "tools/profiler/nsys_profile_tools/gputrc2graph.py", "license": "Apache License 2.0", "lines": 305, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/detokenizer/test_min_tokens.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from transformers import AutoTokenizer from vllm import SamplingParams from vllm.v1.engine import EngineCoreRequest from vllm.v1.engine.detokenizer import FastIncrementalDetokenizer PROMPT = "Hello, my name is Lee, and I'm a student in the " + "college of engineering" @pytest.mark.parametrize( "min_tokens,stop,truth", [ (0, None, " is Lee, and I'm a student in the college of engineering"), (0, "e", " is L"), (5, "e", " is Lee, and I'm a stud"), ], ) def test_min_tokens_with_stop(min_tokens: int, stop: str, truth: str): """Test for a specific min_tokens and stop. See https://github.com/vllm-project/vllm/pull/22014 """ tokenizer = AutoTokenizer.from_pretrained("facebook/opt-125m") all_prompt_ids = tokenizer(PROMPT, add_special_tokens=False).input_ids # The prompt is "Hello, my name is" prompt_token_ids = all_prompt_ids[:4] params = SamplingParams( stop=stop, min_tokens=min_tokens, ) request = EngineCoreRequest( request_id="", prompt_token_ids=prompt_token_ids, mm_features=None, sampling_params=params, pooling_params=None, arrival_time=0.0, lora_request=None, cache_salt=None, data_parallel_rank=None, ) detokenizer = FastIncrementalDetokenizer(tokenizer, request) detokenizer.update(all_prompt_ids[4:], False) assert detokenizer.output_text == truth
{ "repo_id": "vllm-project/vllm", "file_path": "tests/detokenizer/test_min_tokens.py", "license": "Apache License 2.0", "lines": 42, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/kernels/moe/test_flashinfer_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from tests.kernels.moe.utils import make_test_quant_config from tests.kernels.quantization.nvfp4_utils import ( FLOAT4_E2M1_MAX, FLOAT8_E4M3_MAX, dequantize_nvfp4_to_dtype, ) from tests.kernels.utils import torch_moe from vllm import _custom_ops as ops from vllm.config import ParallelConfig, VllmConfig, set_current_vllm_config from vllm.model_executor.layers.fused_moe import fused_topk from vllm.model_executor.layers.fused_moe.activation import MoEActivation from vllm.model_executor.layers.fused_moe.config import ( FusedMoEConfig, FusedMoEParallelConfig, RoutingMethodType, ) from vllm.model_executor.layers.fused_moe.flashinfer_cutlass_moe import ( FlashInferExperts, is_valid_flashinfer_cutlass_fused_moe, ) from vllm.model_executor.layers.fused_moe.modular_kernel import FusedMoEModularKernel from vllm.model_executor.layers.fused_moe.prepare_finalize import ( MoEPrepareAndFinalizeNoEP, ) from vllm.platforms import current_platform from vllm.utils.flashinfer import has_flashinfer_cutlass_fused_moe from vllm.utils.torch_utils import set_random_seed if not has_flashinfer_cutlass_fused_moe() or not current_platform.has_device_capability( 100 ): pytest.skip( "Requires flashinfer_cutlass_fused_moe and nvfp4 support", allow_module_level=True, ) MNK_FACTORS = [ (2, 1024, 1024), (2, 3072, 1024), (2, 3072, 1536), (64, 1024, 1536), (64, 3072, 1024), (64, 2048, 1536), (224, 1024, 1024), (224, 1024, 1536), ] @pytest.mark.parametrize("m,n,k", MNK_FACTORS) @pytest.mark.parametrize("e", [40, 64, 256]) @pytest.mark.parametrize("topk", [1, 6, 8]) @pytest.mark.parametrize("dtype", [torch.bfloat16]) @pytest.mark.parametrize("activation", [MoEActivation.SILU, MoEActivation.RELU2_NO_MUL]) @torch.inference_mode() def test_flashinfer_fp4_moe_no_graph( m: int, n: int, k: int, e: int, topk: int, dtype: torch.dtype, activation: MoEActivation, workspace_init, ): set_random_seed(7) with set_current_vllm_config( VllmConfig(parallel_config=ParallelConfig(pipeline_parallel_size=1)) ): a = torch.randn((m, k), device="cuda", dtype=dtype) / 10 quant_blocksize = 16 is_gated_act = activation.is_gated w1_q, w2_q, quant_config = make_test_quant_config( e, n, k, in_dtype=dtype, quant_dtype="nvfp4", block_shape=None, per_act_token_quant=False, make_gate=is_gated_act, ) score = torch.randn((m, e), device="cuda", dtype=dtype) topk_weights, topk_ids, _ = fused_topk(a, score, topk, renormalize=False) assert is_valid_flashinfer_cutlass_fused_moe(a, w1_q, w2_q) moe_config = FusedMoEConfig( num_experts=e, experts_per_token=topk, hidden_dim=k, intermediate_size_per_partition=n, num_local_experts=e, num_logical_experts=e, activation=activation, device="cuda", moe_parallel_config=FusedMoEParallelConfig.make_no_parallel(), in_dtype=dtype, is_act_and_mul=is_gated_act, routing_method=RoutingMethodType.TopK, ) flashinfer_experts = FusedMoEModularKernel( MoEPrepareAndFinalizeNoEP(), FlashInferExperts(moe_config=moe_config, quant_config=quant_config), inplace=False, ) flashinfer_output = flashinfer_experts( hidden_states=a, w1=w1_q, w2=w2_q, topk_weights=topk_weights, topk_ids=topk_ids, activation=activation, ) # Reference check: a_global_scale = ( (FLOAT8_E4M3_MAX * FLOAT4_E2M1_MAX) / torch.amax(a.flatten(), dim=-1) ).to(torch.float32) a_fp4, a_scale_interleaved = ops.scaled_fp4_quant(a, a_global_scale) _, m_k = a_fp4.shape a_in_dtype = dequantize_nvfp4_to_dtype( a_fp4, a_scale_interleaved, a_global_scale, dtype=a.dtype, device=a.device, block_size=quant_blocksize, ) w1_d = torch.empty( (e, (2 if is_gated_act else 1) * n, k), device="cuda", dtype=dtype ) w2_d = torch.empty((e, k, n), device="cuda", dtype=dtype) for idx in range(0, e): w1_d[idx] = dequantize_nvfp4_to_dtype( w1_q[idx], quant_config.w1_scale[idx], (1 / quant_config.g1_alphas[idx]), dtype=dtype, device=w1_q.device, block_size=quant_blocksize, ) w2_d[idx] = dequantize_nvfp4_to_dtype( w2_q[idx], quant_config.w2_scale[idx], (1 / quant_config.g2_alphas[idx]), dtype=dtype, device=w2_q.device, block_size=quant_blocksize, ) torch_output = torch_moe( a_in_dtype, w1_d, w2_d, score, topk, activation=activation ) torch.testing.assert_close( torch_output, flashinfer_output, atol=1e-1, rtol=1e-1 ) if __name__ == "__main__": test_flashinfer_fp4_moe_no_graph((2, 1024, 1024), 40, 1, torch.half)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/moe/test_flashinfer_moe.py", "license": "Apache License 2.0", "lines": 154, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/cudagraph/test_cudagraph_dispatch.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import replace from unittest.mock import MagicMock, patch import pytest import torch import torch.nn as nn from tests.utils import create_new_process_for_each_test from vllm.compilation.cuda_graph import CUDAGraphWrapper from vllm.compilation.monitor import set_cudagraph_capturing_enabled from vllm.config import ( CompilationConfig, CompilationMode, CUDAGraphMode, ParallelConfig, SchedulerConfig, VllmConfig, ) from vllm.config.lora import LoRAConfig from vllm.forward_context import BatchDescriptor, set_forward_context from vllm.platforms import current_platform from vllm.v1.cudagraph_dispatcher import CudagraphDispatcher # Helper MLP for testing class SimpleMLP(nn.Module): def __init__(self): super().__init__() self.fc1 = nn.Linear(10, 10) self.fc2 = nn.Linear(10, 10) def forward(self, x): return self.fc2(self.fc1(x)) def _create_vllm_config( compilation_config: CompilationConfig, max_num_seqs: int = 8, lora_config: bool = False, ) -> MagicMock: mock_config = MagicMock(spec=VllmConfig) mock_config.compilation_config = compilation_config mock_config.scheduler_config = SchedulerConfig.default_factory( max_num_seqs=max_num_seqs, ) mock_config.parallel_config = ParallelConfig() mock_config.speculative_config = None # No speculative decoding if not lora_config: mock_config.lora_config = None else: # Create a real LoRAConfig with specialize_active_lora enabled mock_config.lora_config = LoRAConfig( max_loras=4, specialize_active_lora=True, ) # Mimic the behavior of VllmConfig.__post_init__() if compilation_config.mode == CompilationMode.VLLM_COMPILE: compilation_config.set_splitting_ops_for_v1( all2all_backend=mock_config.parallel_config.all2all_backend, data_parallel_size=mock_config.parallel_config.data_parallel_size, ) # mimic VllmConfig.__post_init__ if compilation_config.cudagraph_capture_sizes: compilation_config.max_cudagraph_capture_size = ( compilation_config.cudagraph_capture_sizes[-1] ) compilation_config.post_init_cudagraph_sizes() return mock_config class TestCudagraphDispatcher: @pytest.mark.parametrize( "cudagraph_mode_str,compilation_mode,lora_config", [ # Test case 0: Full CG for mixed batches, no separate routine ("FULL", CompilationMode.NONE, False), # Test case 1: Full CG for uniform batches, piecewise for mixed ("FULL_AND_PIECEWISE", CompilationMode.NONE, False), # Test case 2: Full CG for uniform batches, no CG for mixed ("FULL_DECODE_ONLY", CompilationMode.NONE, False), # Test case 3: PIECEWISE for all ("PIECEWISE", CompilationMode.VLLM_COMPILE, False), # Test case 4: PIECEWISE for all, specialize LoRA cases ("PIECEWISE", CompilationMode.VLLM_COMPILE, True), ], ) def test_dispatcher(self, cudagraph_mode_str, compilation_mode, lora_config): # Setup dispatcher comp_config = CompilationConfig( cudagraph_mode=cudagraph_mode_str, mode=compilation_mode, cudagraph_capture_sizes=[1, 8], ) config = _create_vllm_config( comp_config, max_num_seqs=8, lora_config=lora_config ) if ( cudagraph_mode_str == "FULL_AND_PIECEWISE" and compilation_mode == CompilationMode.NONE ): with pytest.raises(AssertionError): dispatcher = CudagraphDispatcher(config) return dispatcher = CudagraphDispatcher(config) dispatcher.initialize_cudagraph_keys( cudagraph_mode=comp_config.cudagraph_mode, uniform_decode_query_len=1 ) # Verify the key is initialized correctly # With LoRA specialization (max_loras=4, specialize_active_lora=True): # - lora_cases = [0, 1, 2, 4, 5] (no-lora + powers of 2 up to 4 + max_loras+1) # - capture_sizes = [1, 8] # - Total keys = 2 sizes × 5 lora_cases = 10 if cudagraph_mode_str in ["FULL_AND_PIECEWISE", "PIECEWISE"]: assert len(dispatcher.cudagraph_keys[CUDAGraphMode.PIECEWISE]) == ( 10 if lora_config else 2 ) else: assert len(dispatcher.cudagraph_keys[CUDAGraphMode.PIECEWISE]) == 0 if cudagraph_mode_str not in ["NONE", "PIECEWISE"]: assert len(dispatcher.cudagraph_keys[CUDAGraphMode.FULL]) == ( 10 if lora_config else 2 ) else: assert len(dispatcher.cudagraph_keys[CUDAGraphMode.FULL]) == 0 # Test dispatch logic # 1. non-uniform batch, size in cudagraph size list # FULL mode uses exact keys with num_reqs set desc_full_with_reqs = BatchDescriptor(num_tokens=8, num_reqs=8, uniform=False) # PIECEWISE mode uses relaxed keys with num_reqs=None desc_piecewise = BatchDescriptor(num_tokens=8, num_reqs=None, uniform=False) rt_mode, key = dispatcher.dispatch( num_tokens=8, uniform_decode=False, has_lora=False ) if cudagraph_mode_str == "FULL": assert rt_mode == CUDAGraphMode.FULL assert key == desc_full_with_reqs elif cudagraph_mode_str in ["FULL_AND_PIECEWISE", "PIECEWISE"]: assert rt_mode == CUDAGraphMode.PIECEWISE assert key == desc_piecewise else: assert rt_mode == CUDAGraphMode.NONE # 2. uniform decode batch, size in cudagraph size list desc_uniform_exact = BatchDescriptor(num_tokens=8, num_reqs=8, uniform=True) desc_non_uniform = BatchDescriptor(num_tokens=8, num_reqs=8, uniform=False) rt_mode, key = dispatcher.dispatch( num_tokens=8, uniform_decode=True, has_lora=False ) if cudagraph_mode_str == "FULL": # Pure FULL mode uses non-uniform keys for all batches assert rt_mode == CUDAGraphMode.FULL assert key == desc_non_uniform elif cudagraph_mode_str in ["FULL_DECODE_ONLY", "FULL_AND_PIECEWISE"]: # These modes have separate uniform decode keys assert rt_mode == CUDAGraphMode.FULL assert key == desc_uniform_exact elif cudagraph_mode_str == "PIECEWISE": assert rt_mode == CUDAGraphMode.PIECEWISE assert key == replace(desc_uniform_exact, num_reqs=None, uniform=False) else: assert rt_mode == CUDAGraphMode.NONE # 3. No key match rt_mode, key = dispatcher.dispatch( num_tokens=15, uniform_decode=False, has_lora=False ) assert rt_mode == CUDAGraphMode.NONE assert key == BatchDescriptor(num_tokens=15) # 4. invalid_modes={FULL} should have a fall back mode # (e.g., cascade attention) desc_full_exact = BatchDescriptor(num_tokens=8, uniform=False) rt_mode, key = dispatcher.dispatch( num_tokens=8, uniform_decode=False, has_lora=False, invalid_modes={CUDAGraphMode.FULL}, ) if "PIECEWISE" in cudagraph_mode_str: # string contains check assert rt_mode == CUDAGraphMode.PIECEWISE assert key == replace(desc_full_exact, num_reqs=None, uniform=False) else: assert rt_mode == CUDAGraphMode.NONE # 5. valid_modes={NONE} always returns NONE even when keys exist rt_mode, key = dispatcher.dispatch( num_tokens=8, uniform_decode=False, has_lora=False, valid_modes={CUDAGraphMode.NONE}, ) assert rt_mode == CUDAGraphMode.NONE assert key == BatchDescriptor(num_tokens=8) @pytest.mark.parametrize( "cudagraph_mode_str,compilation_mode,expected_modes", [ # FULL mode: only FULL keys, no PIECEWISE ("FULL", CompilationMode.NONE, [CUDAGraphMode.FULL]), # PIECEWISE mode: only PIECEWISE keys ("PIECEWISE", CompilationMode.VLLM_COMPILE, [CUDAGraphMode.PIECEWISE]), # FULL_DECODE_ONLY: only FULL keys for uniform decode ("FULL_DECODE_ONLY", CompilationMode.NONE, [CUDAGraphMode.FULL]), # NONE mode: no keys ("NONE", CompilationMode.NONE, []), ], ) def test_get_capture_descs( self, cudagraph_mode_str, compilation_mode, expected_modes ): """Test get_capture_descs returns correctly grouped and ordered descs.""" comp_config = CompilationConfig( cudagraph_mode=cudagraph_mode_str, mode=compilation_mode, cudagraph_capture_sizes=[1, 4, 8, 16], ) config = _create_vllm_config(comp_config, max_num_seqs=16) dispatcher = CudagraphDispatcher(config) dispatcher.initialize_cudagraph_keys( cudagraph_mode=comp_config.cudagraph_mode, uniform_decode_query_len=1 ) capture_descs = dispatcher.get_capture_descs() # Verify we get the expected modes actual_modes = [mode for mode, _ in capture_descs] assert actual_modes == expected_modes # Verify each group is sorted largest-first for mode, descs in capture_descs: assert len(descs) > 0, "Each group should have at least one descriptor" num_tokens_list = [d.num_tokens for d in descs] assert num_tokens_list == sorted(num_tokens_list, reverse=True), ( f"Descriptors for {mode} should be sorted largest-first" ) # All descriptors in a group should have same uniform value uniform_values = [d.uniform for d in descs] assert len(set(uniform_values)) == 1, ( "All descriptors in a group should have the same uniform value" ) def test_get_capture_descs_empty_when_not_initialized(self): """Test that get_capture_descs returns empty list when keys not initialized.""" comp_config = CompilationConfig( cudagraph_mode="FULL", mode=CompilationMode.NONE, cudagraph_capture_sizes=[1, 8], ) config = _create_vllm_config(comp_config, max_num_seqs=8) dispatcher = CudagraphDispatcher(config) # Don't initialize keys assert dispatcher.get_capture_descs() == [] @pytest.mark.skipif(not current_platform.is_cuda(), reason="Skip if not cuda") class TestCUDAGraphWrapper: def setup_method(self): self.vllm_config = _create_vllm_config(CompilationConfig()) self.model = SimpleMLP().to("cuda") self.persistent_input_buffer = torch.zeros(1, 10, device="cuda") self.input_tensor = torch.randn(1, 10, device="cuda") def test_capture_and_replay(self): wrapper = CUDAGraphWrapper( self.model, self.vllm_config, runtime_mode=CUDAGraphMode.FULL ) batch_descriptor = BatchDescriptor(num_tokens=10) # 0. global warmup with set_forward_context( attn_metadata=None, vllm_config=self.vllm_config, cudagraph_runtime_mode=CUDAGraphMode.NONE, batch_descriptor=None, ): wrapper(self.input_tensor) # 1. Capture with ( set_forward_context( attn_metadata=None, vllm_config=self.vllm_config, cudagraph_runtime_mode=CUDAGraphMode.FULL, batch_descriptor=batch_descriptor, ), patch("torch.cuda.graph", wraps=torch.cuda.graph) as mock_cuda_graph, ): output1 = wrapper(self.input_tensor) # capturing phase should generate a zero output assert torch.allclose(output1, torch.zeros_like(output1)) mock_cuda_graph.assert_called_once() assert batch_descriptor in wrapper.concrete_cudagraph_entries entry = wrapper.concrete_cudagraph_entries[batch_descriptor] assert entry.cudagraph is not None # 2. Replay with ( set_forward_context( attn_metadata=None, vllm_config=self.vllm_config, cudagraph_runtime_mode=CUDAGraphMode.FULL, batch_descriptor=batch_descriptor, ), patch.object( entry.cudagraph, "replay", wraps=entry.cudagraph.replay ) as mock_replay, ): output2 = wrapper(self.input_tensor) mock_replay.assert_called_once() # Compare with eager output eager_output = self.model(self.input_tensor) torch.testing.assert_close(eager_output, output2) def test_bypass_on_mode_mismatch(self): wrapper = CUDAGraphWrapper( self.model, self.vllm_config, runtime_mode=CUDAGraphMode.FULL ) batch_descriptor = BatchDescriptor(num_tokens=10) with ( set_forward_context( attn_metadata=None, vllm_config=self.vllm_config, cudagraph_runtime_mode=CUDAGraphMode.PIECEWISE, batch_descriptor=batch_descriptor, ), patch("torch.cuda.graph", wraps=torch.cuda.graph) as mock_cuda_graph, patch.object( self.model, "forward", wraps=self.model.forward ) as mock_forward, ): wrapper(self.input_tensor) mock_cuda_graph.assert_not_called() mock_forward.assert_called_once() assert not wrapper.concrete_cudagraph_entries def test_bypass_on_mode_none(self): wrapper = CUDAGraphWrapper( self.model, self.vllm_config, runtime_mode=CUDAGraphMode.FULL ) batch_descriptor = BatchDescriptor(num_tokens=10) with ( set_forward_context( attn_metadata=None, vllm_config=self.vllm_config, cudagraph_runtime_mode=CUDAGraphMode.NONE, batch_descriptor=batch_descriptor, ), patch("torch.cuda.graph", wraps=torch.cuda.graph) as mock_cuda_graph, ): wrapper(self.input_tensor) mock_cuda_graph.assert_not_called() assert not wrapper.concrete_cudagraph_entries @pytest.mark.skipif(not current_platform.is_cuda(), reason="Skip if not cuda") class TestCudagraphIntegration: def setup_method(self): # only FULL mode for non-uniform batches self.comp_config = CompilationConfig( mode=CompilationMode.VLLM_COMPILE, cudagraph_mode="FULL", cudagraph_capture_sizes=[10, 20], ) self.vllm_config = _create_vllm_config(self.comp_config) self.dispatcher = CudagraphDispatcher(self.vllm_config) self.dispatcher.initialize_cudagraph_keys( self.comp_config.cudagraph_mode, uniform_decode_query_len=1 ) def _run_and_monitor_call( self, wrapper, input_tensor, runtime_mode, batch_descriptor ): """Helper to run a single call and monitor the action.""" with ( patch("torch.cuda.graph", wraps=torch.cuda.graph) as mock_graph_context, patch.object(wrapper, "runnable", wraps=wrapper.runnable) as mock_runnable, ): entry = wrapper.concrete_cudagraph_entries.get(batch_descriptor, None) context = set_forward_context( attn_metadata=None, vllm_config=self.vllm_config, cudagraph_runtime_mode=runtime_mode, batch_descriptor=batch_descriptor, ) mock_replay = MagicMock() if entry and entry.cudagraph: with ( context, patch.object( entry.cudagraph, "replay", new_callable=MagicMock ) as mock_replay, ): wrapper(input_tensor) else: with context: wrapper(input_tensor) if mock_graph_context.called: # note that this is globally mocked, so it will be detected # even whether called by the inner or outer wrapper return "capture_global" if mock_replay.called: # only for outer wrapper return "replay" if mock_runnable.call_count > 0: # only for outer wrapper return "bypass" return "unknown" @create_new_process_for_each_test("spawn") def test_capture_replay_bypass_logic(self): model = SimpleMLP().to("cuda") full_wrapper = CUDAGraphWrapper(model, self.vllm_config, CUDAGraphMode.FULL) max_bs = 16 persistent_input_buffer = torch.zeros(max_bs, 10, device="cuda") input_1 = persistent_input_buffer[:1] input_2 = persistent_input_buffer[:2] input_3 = persistent_input_buffer[:3] desc_1 = BatchDescriptor(num_tokens=1) desc_2 = BatchDescriptor(num_tokens=2) desc_3_unseen = BatchDescriptor(num_tokens=3) # 0. global warmup with set_forward_context( attn_metadata=None, vllm_config=self.vllm_config, cudagraph_runtime_mode=CUDAGraphMode.NONE, batch_descriptor=None, ): full_wrapper(input_1) rt_mode, key = self.dispatcher.dispatch(num_tokens=desc_1.num_tokens) # 1. Capture first shape action = self._run_and_monitor_call(full_wrapper, input_1, rt_mode, key) assert action == "capture_global" # 2. Replay first shape action = self._run_and_monitor_call(full_wrapper, input_1, rt_mode, key) assert action == "replay" rt_mode, key = self.dispatcher.dispatch(num_tokens=desc_2.num_tokens) # 3. Capture second shape action = self._run_and_monitor_call(full_wrapper, input_2, rt_mode, key) assert action == "capture_global" # 4. Replay second shape action = self._run_and_monitor_call( full_wrapper, input_2, CUDAGraphMode.FULL, desc_2 ) assert action == "replay" # 5. Bypass if no key match rt_mode, key = self.dispatcher.dispatch(num_tokens=desc_3_unseen.num_tokens) assert rt_mode == CUDAGraphMode.NONE action = self._run_and_monitor_call(full_wrapper, input_3, rt_mode, key) assert action == "bypass" # capture unseen shape is not allowed after disable set_cudagraph_capturing_enabled(False) with pytest.raises(RuntimeError): self._run_and_monitor_call( full_wrapper, input_3, CUDAGraphMode.FULL, desc_3_unseen ) set_cudagraph_capturing_enabled(True) @create_new_process_for_each_test("spawn") def test_nested_wrappers(self): """Tests a scenario with a PIECEWISE wrapper inside a FULL one.""" model = SimpleMLP().to("cuda") full_wrapper = CUDAGraphWrapper(model, self.vllm_config, CUDAGraphMode.FULL) input_1 = torch.randn(1, 10, device="cuda") # Setup: Inner model is wrapped with PIECEWISE, outer with FULL inner_model = SimpleMLP().to("cuda") piecewise_wrapper = CUDAGraphWrapper( inner_model, self.vllm_config, CUDAGraphMode.PIECEWISE ) inner_model.forward = MagicMock(wraps=inner_model.forward) outer_model = SimpleMLP().to("cuda") # When outer model is called, it calls the piecewise_wrapper outer_model.forward = MagicMock( wraps=outer_model.forward, side_effect=piecewise_wrapper ) full_wrapper = CUDAGraphWrapper( outer_model, self.vllm_config, CUDAGraphMode.FULL ) desc_1 = BatchDescriptor(num_tokens=1) # 0. global warmup with set_forward_context( attn_metadata=None, vllm_config=self.vllm_config, cudagraph_runtime_mode=CUDAGraphMode.NONE, batch_descriptor=None, ): full_wrapper(input_1) # --- Test runtime mode FULL--- # Run with FULL mode context. Expect outer wrapper to capture. # The inner mock should be called once inside the graph capture. outer_model.forward.reset_mock() inner_model.forward.reset_mock() action = self._run_and_monitor_call( full_wrapper, input_1, CUDAGraphMode.FULL, desc_1 ) assert action == "capture_global" assert outer_model.forward.call_count == 1 assert inner_model.forward.call_count == 1 # Run again. Expect outer wrapper to replay. # The outer model should NOT be called because the whole graph # is replayed. action = self._run_and_monitor_call( full_wrapper, input_1, CUDAGraphMode.FULL, desc_1 ) assert action == "replay" assert outer_model.forward.call_count == 1 # No new call assert inner_model.forward.call_count == 1 # --- Test runtime mode PIECEWISE --- outer_model.forward.reset_mock() inner_model.forward.reset_mock() # Run with PIECEWISE mode context. # Expect outer wrapper to bypass and call inner wrapper. # Inner wrapper should capture. action = self._run_and_monitor_call( full_wrapper, input_1, CUDAGraphMode.PIECEWISE, desc_1 ) assert action == "capture_global" assert outer_model.forward.call_count == 1 assert inner_model.forward.call_count == 1 # Run again with PIECEWISE. # Outer bypasses, inner replays. action = self._run_and_monitor_call( full_wrapper, input_1, CUDAGraphMode.PIECEWISE, desc_1 ) assert action == "bypass" assert outer_model.forward.call_count == 2 assert inner_model.forward.call_count == 1
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/cudagraph/test_cudagraph_dispatch.py", "license": "Apache License 2.0", "lines": 497, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:tests/v1/cudagraph/test_cudagraph_mode.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import weakref from contextlib import ExitStack import pytest from tests.utils import wait_for_gpu_memory_to_clear from tests.v1.attention.utils import full_cg_backend_configs as backend_configs from vllm import LLM from vllm.config import CompilationConfig, CompilationMode from vllm.platforms import current_platform # test attention backend and cudagraph_mode combo # (backend_name, cudagraph_mode, supported) if current_platform.is_rocm(): combo_cases_1 = [ ("RocmAttn", "FULL", True), ("RocmAttn", "FULL_AND_PIECEWISE", True), ("TritonAttn", "FULL", True), ("TritonAttn", "FULL_AND_PIECEWISE", True), ] else: combo_cases_1 = [ ("FA3", "FULL", True), ("FA3", "FULL_AND_PIECEWISE", True), ("FA2", "FULL", True), # Should fallback to FULL_AND_PIECEWISE ("FA2", "FULL_AND_PIECEWISE", True), ("FlashInfer", "FULL", True), # Should fallback to FULL_AND_PIECEWISE ("FlashInfer", "FULL_AND_PIECEWISE", True), ] @pytest.mark.parametrize("backend_name, cudagraph_mode, supported", combo_cases_1) def test_backend_and_cudagraph_mode_combo(backend_name, cudagraph_mode, supported): if backend_name == "FlashInfer": try: import flashinfer # noqa: F401 except ImportError: pytest.skip("FlashInfer is not installed") backend_config = backend_configs[backend_name] # Dynamically skip test if GPU capability is not met if ( backend_config.specific_gpu_arch and backend_config.specific_gpu_arch != current_platform.get_device_capability() ): pytest.skip("Only Hopper GPUs support FA3 and FlashMLA") attention_config = backend_config.attention_config with ExitStack() as stack: if not supported: stack.enter_context(pytest.raises(Exception)) llm = LLM( model="Qwen/Qwen2-1.5B-Instruct", max_num_seqs=256, trust_remote_code=True, gpu_memory_utilization=0.45, max_model_len=1024, attention_config=attention_config, compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, cudagraph_mode=cudagraph_mode ), ) llm.generate(["Hello, my name is"] * 10) # when above code raises, `llm` may be undefined, so we need to catch that try: llm = weakref.proxy(llm) del llm except UnboundLocalError: pass wait_for_gpu_memory_to_clear( devices=[0], threshold_ratio=0.1, ) # test cudagraph_mode with different compilation mode. # (backend_name, cudagraph_mode, compilation_mode, supported) attn_backend = "RocmAttn" if current_platform.is_rocm() else "FA2" combo_cases_2 = [ (attn_backend, "FULL", CompilationMode.NONE, True), (attn_backend, "FULL", CompilationMode.VLLM_COMPILE, True), (attn_backend, "PIECEWISE", CompilationMode.NONE, True), (attn_backend, "PIECEWISE", CompilationMode.VLLM_COMPILE, True), (attn_backend, "FULL_AND_PIECEWISE", CompilationMode.NONE, True), (attn_backend, "FULL_AND_PIECEWISE", CompilationMode.VLLM_COMPILE, True), (attn_backend, "FULL_DECODE_ONLY", CompilationMode.NONE, True), (attn_backend, "FULL_DECODE_ONLY", CompilationMode.VLLM_COMPILE, True), (attn_backend, "NONE", CompilationMode.NONE, True), (attn_backend, "NONE", CompilationMode.VLLM_COMPILE, True), ] @pytest.mark.parametrize( "backend_name,cudagraph_mode,compilation_mode,supported", combo_cases_2 ) def test_cudagraph_compilation_combo( backend_name, cudagraph_mode, compilation_mode, supported ): backend_config = backend_configs[backend_name] attention_config = backend_config.attention_config with ExitStack() as stack: if not supported: stack.enter_context(pytest.raises(Exception)) llm = LLM( model="Qwen/Qwen2-1.5B-Instruct", max_num_seqs=256, trust_remote_code=True, gpu_memory_utilization=0.45, max_model_len=1024, attention_config=attention_config, compilation_config=CompilationConfig( mode=compilation_mode, cudagraph_mode=cudagraph_mode ), ) llm.generate(["Hello, my name is"] * 10) # when above code raises, `llm` may be undefined, so we need to catch that try: llm = weakref.proxy(llm) del llm except UnboundLocalError: pass finally: wait_for_gpu_memory_to_clear( devices=[0], threshold_ratio=0.1, )
{ "repo_id": "vllm-project/vllm", "file_path": "tests/v1/cudagraph/test_cudagraph_mode.py", "license": "Apache License 2.0", "lines": 117, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:vllm/compilation/base_static_graph.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable from typing import Any, Protocol from vllm.config import CUDAGraphMode, VllmConfig class AbstractStaticGraphWrapper(Protocol): """ StaticGraphWrapper interface that allows platforms to wrap a callable to be captured as a static graph. """ def __init__( self, runnable: Callable[..., Any], vllm_config: VllmConfig, runtime_mode: CUDAGraphMode, **kwargs: Any, ) -> None: """ Initializes the StaticGraphWrapper class with graph capturing and execution-related configurations. Args: runnable (Callable): The callable to be wrapped and captured. vllm_config (VllmConfig): Global configuration for vLLM. runtime_mode (CUDAGraphMode): The style of the static graph runtime. See CUDAGraphMode in vllm/config.py. Note that only the subset enum `NONE`, `PIECEWISE` and `FULL` are used as concrete runtime mode for cudagraph dispatching. Keyword Args: kwargs: Additional keyword arguments for platform-specific configurations. """ raise NotImplementedError def __call__(self, *args: Any, **kwargs: Any) -> Any: """ Executes the wrapped callable. If the current runtime mode in the ForwardContext matches the runtime mode of this instance, it replays the CUDAGraph or captures it using the callable if it hasn't been captured yet. Otherwise, it calls the original callable directly. Args: *args: Variable length input arguments to be passed into the callable. **kwargs: Keyword arguments to be passed into the callable. Returns: Any: Output of the executed callable. """ raise NotImplementedError
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/compilation/base_static_graph.py", "license": "Apache License 2.0", "lines": 47, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/compilation/cuda_graph.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import dataclasses from collections import Counter from collections.abc import Callable from contextlib import ExitStack from typing import Any from unittest.mock import patch import torch import vllm.envs as envs from vllm.compilation.counter import compilation_counter from vllm.compilation.monitor import validate_cudagraph_capturing_enabled from vllm.config import CUDAGraphMode, VllmConfig from vllm.distributed.device_communicators.pynccl_allocator import set_graph_pool_id from vllm.forward_context import BatchDescriptor, get_forward_context from vllm.logger import init_logger from vllm.model_executor.offloader.base import get_offloader from vllm.platforms import current_platform from vllm.utils.torch_utils import current_stream, weak_ref_tensors logger = init_logger(__name__) @dataclasses.dataclass(frozen=True) class CUDAGraphStat: num_unpadded_tokens: int num_padded_tokens: int num_paddings: int runtime_mode: str class CUDAGraphLogging: """Aggregate and log cudagraph metrics""" COLUMN_HEADERS = [ "Unpadded Tokens", "Padded Tokens", "Num Paddings", "Runtime Mode", "Count", ] def __init__( self, cg_mode: CUDAGraphMode, cg_capture_sizes: list[int] | None ) -> None: self.reset() self.cg_mode = str(cg_mode) self.cg_capture_sizes = str(cg_capture_sizes or []) self.settings_header = ( "**CUDAGraph Config Settings:**\n\n" f"- Mode: {self.cg_mode}\n" f"- Capture sizes: {self.cg_capture_sizes}\n\n" "**CUDAGraph Stats:**\n\n" ) def reset(self) -> None: self.stats: list[CUDAGraphStat] = [] def observe(self, cudagraph_stat: CUDAGraphStat) -> None: self.stats.append(cudagraph_stat) def generate_metric_table(self) -> str: stats_counts = Counter(self.stats) # Convert stats to rows of strings, in descending order of observed frequencies rows = [] for stat, count in sorted( stats_counts.items(), key=lambda item: item[1], reverse=True ): rows.append( [ str(stat.num_unpadded_tokens), str(stat.num_padded_tokens), str(stat.num_paddings), stat.runtime_mode, str(count), ] ) # Calculate column widths (max of header and data) col_widths = [] for i, header_text in enumerate(self.COLUMN_HEADERS): max_width = len(header_text) for row in rows: max_width = max(max_width, len(row[i])) col_widths.append(max_width) table_header_list = [ h.ljust(w) for h, w in zip(self.COLUMN_HEADERS, col_widths) ] table_header = "| " + " | ".join(table_header_list) + " |\n" table_separator = "|" + "|".join("-" * (w + 2) for w in col_widths) + "|\n" # Create data rows with proper alignment data_rows = [] for row in rows: formatted_row = [ str(val).ljust(width) for val, width in zip(row, col_widths) ] data_rows.append("| " + " | ".join(formatted_row) + " |") return ( self.settings_header + table_header + table_separator + "\n".join(data_rows) + "\n" ) def log(self, log_fn: Callable[..., Any] = logger.info) -> None: if not self.stats: return log_fn(self.generate_metric_table()) self.reset() @dataclasses.dataclass class CUDAGraphEntry: batch_descriptor: BatchDescriptor cudagraph: torch.cuda.CUDAGraph | None = None output: Any | None = None # for cudagraph debugging, track the input addresses # during capture, and check if they are the same during replay input_addresses: list[int] | None = None @dataclasses.dataclass class CUDAGraphOptions: debug_log_enable: bool = True gc_disable: bool = False weak_ref_output: bool = True class CUDAGraphWrapper: """Wraps a runnable to add CUDA graph capturing and replaying ability. And provide attribute access to the underlying `runnable` via `__getattr__`. The workflow of this wrapper in the cudagraph dispatching is as follows: 1. At initialization, a runtime mode is assigned to the wrapper (FULL or PIECEWISE). 2. At runtime, the wrapper receives a runtime_mode and a batch_descriptor(key) from the forward context and blindly trust them for cudagraph dispatching. 3. If runtime_mode is NONE or runtime_mode does not match the mode of the wrapper, just call the runnable directly. 4. Otherwise, i.e., the runtime_mode matches the mode of the wrapper, the wrapper will perform cudagraph capture(if key does not exist, create a new entry and cache it) or replay (if key exists in the cache). Note: CUDAGraphWrapper does not store persistent buffers or copy any runtime inputs into that buffers for replay. We assume implementing them is done outside of the wrapper. That is because we do not make any assumption on the dynamic shape (batch size) of the runtime inputs, as a trade-off for staying orthogonal to compilation logic. Nevertheless, tracing and checking the input addresses to be consistent during replay is guaranteed when VLLM_LOGGING_LEVEL == "DEBUG". """ def __init__( self, runnable: Callable[..., Any], vllm_config: VllmConfig, runtime_mode: CUDAGraphMode, cudagraph_options: CUDAGraphOptions | None = None, ) -> None: self.runnable = runnable self.vllm_config = vllm_config self.runtime_mode = runtime_mode self.compilation_config = vllm_config.compilation_config self.first_run_finished = False self.is_debugging_mode = envs.VLLM_LOGGING_LEVEL == "DEBUG" # assert runtime_mode is not NONE(no cudagraph), otherwise, we don't # need to initialize a CUDAGraphWrapper. assert self.runtime_mode != CUDAGraphMode.NONE # TODO: in the future, if we want to use multiple # streams, it might not be safe to share a global pool. # only investigate this when we use multiple streams self.graph_pool = current_platform.get_global_graph_pool() if cudagraph_options is None: cudagraph_options = CUDAGraphOptions() self.cudagraph_options = cudagraph_options # the entries for different batch descriptors that we need to capture # cudagraphs for. self.concrete_cudagraph_entries: dict[BatchDescriptor, CUDAGraphEntry] = {} def __getattr__(self, key: str) -> Any: # allow accessing the attributes of the runnable. if hasattr(self.runnable, key): return getattr(self.runnable, key) raise AttributeError( f"Attribute {key} not exists in the runnable of " f"cudagraph wrapper: {self.runnable}" ) def unwrap(self) -> Callable[..., Any]: # in case we need to access the original runnable. return self.runnable def __call__(self, *args: Any, **kwargs: Any) -> Any | None: forward_context = get_forward_context() batch_descriptor = forward_context.batch_descriptor cudagraph_runtime_mode = forward_context.cudagraph_runtime_mode if ( cudagraph_runtime_mode == CUDAGraphMode.NONE or cudagraph_runtime_mode != self.runtime_mode ): # CUDAGraphMode.NONE could mean the profile run, a warmup run, or # running without cudagraphs. # We do not trigger capture/replay if the runtime mode is not # matches. This enables properly dispatching to the correct # CUDAGraphWrapper when nesting multiple instances with different # runtime modes. return self.runnable(*args, **kwargs) assert batch_descriptor is not None if batch_descriptor not in self.concrete_cudagraph_entries: # create a new entry for this batch descriptor self.concrete_cudagraph_entries[batch_descriptor] = CUDAGraphEntry( batch_descriptor=batch_descriptor ) entry = self.concrete_cudagraph_entries[batch_descriptor] if entry.cudagraph is None: if self.cudagraph_options.debug_log_enable: # Since we capture cudagraph for many different shapes and # capturing is fast, we don't need to log it for every # shape. E.g. we only log it for the first subgraph in # piecewise mode. logger.debug( "Capturing a cudagraph on (%s,%s)", self.runtime_mode.name, entry.batch_descriptor, ) # validate that cudagraph capturing is legal at this point. validate_cudagraph_capturing_enabled() input_addresses = [ x.data_ptr() for x in args if isinstance(x, torch.Tensor) ] entry.input_addresses = input_addresses cudagraph = torch.cuda.CUDAGraph() with ExitStack() as stack: if self.cudagraph_options.gc_disable: # during every model forward for piecewise cudagraph # mode, we will capture many pieces of cudagraphs # (roughly one per layer). running gc again and again # across layers will make the cudagraph capture very slow. # therefore, we only run gc for the first graph, # and disable gc for the rest of the graphs. stack.enter_context(patch("gc.collect", lambda: None)) stack.enter_context(patch("torch.cuda.empty_cache", lambda: None)) if self.graph_pool is not None: set_graph_pool_id(self.graph_pool) else: set_graph_pool_id(current_platform.graph_pool_handle()) # Sync offloader's copy stream before capture. # Ensure any pre-capture prefetches from offloader are complete. get_offloader().sync_prev_onload() # mind-exploding: carefully manage the reference and memory. with torch.cuda.graph( cudagraph, pool=self.graph_pool, stream=current_stream(), ): # `output` is managed by pytorch's cudagraph pool output = self.runnable(*args, **kwargs) # Join offloader's copy stream after forward to avoid # unjoined stream error. The last layer's start_prefetch # forks copy_stream, but wait_prefetch only happens in # the next forward pass. get_offloader().join_after_forward() if self.cudagraph_options.weak_ref_output: # by converting it to weak ref, # the original `output` will immediately be released # to save memory. It is only safe to do this for # the last graph in piecewise cuadgraph mode, because # the output of the last graph will not be used by # any other cuda graph. output = weak_ref_tensors(output) # here we always use weak ref for the output # to save memory entry.output = weak_ref_tensors(output) entry.cudagraph = cudagraph compilation_counter.num_cudagraph_captured += 1 # important: we need to return the output, rather than # the weak ref of the output, so that pytorch can correctly # manage the memory during cuda graph capture return output if self.is_debugging_mode: # check if the input addresses are the same new_input_addresses = [ x.data_ptr() for x in args if isinstance(x, torch.Tensor) ] assert new_input_addresses == entry.input_addresses, ( f"Input addresses for cudagraphs are different " f"during replay. Expected {entry.input_addresses}, " f"got {new_input_addresses}" ) # Sync offloader before replay - ensures any external dependencies # from pre-capture prefetches are satisfied. get_offloader().sync_prev_onload() entry.cudagraph.replay() return entry.output
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/compilation/cuda_graph.py", "license": "Apache License 2.0", "lines": 272, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/v1/cudagraph_dispatcher.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Set as AbstractSet from dataclasses import replace from itertools import product from vllm.config import CUDAGraphMode, VllmConfig from vllm.forward_context import BatchDescriptor from vllm.logger import init_logger from vllm.lora.utils import get_captured_lora_counts logger = init_logger(__name__) class CudagraphDispatcher: """ Runtime cudagraph dispatcher to dispatch keys for multiple set of cudagraphs. The dispatcher stores two sets of dispatch keys, one for PIECEWISE and one for FULL cudagraph runtime mode. The keys are initialized depending on attention support and what cudagraph mode is set in CompilationConfig. The keys stored in dispatcher are the only source of truth for valid cudagraphs that can be dispatched at runtime. At runtime, the dispatch method generates the runtime cudagraph mode (FULL, PIECEWISE, or NONE for no cudagraph) and the valid key (batch descriptor) based on the input key. After dispatching (communicated via forward context), the cudagraph wrappers will trust the dispatch key to either capture or replay (if the mode matches), or pass through to the underlying runnable without cudagraph (if the mode does not match or mode is NONE). """ def __init__(self, vllm_config: VllmConfig): self.vllm_config = vllm_config self.compilation_config = vllm_config.compilation_config self.uniform_decode_query_len = ( 1 if not self.vllm_config.speculative_config else 1 + self.vllm_config.speculative_config.num_speculative_tokens ) # Dict to store valid cudagraph dispatching keys. self.cudagraph_keys: dict[CUDAGraphMode, set[BatchDescriptor]] = { CUDAGraphMode.PIECEWISE: set(), CUDAGraphMode.FULL: set(), } assert ( not self.compilation_config.cudagraph_mode.requires_piecewise_compilation() or self.compilation_config.is_attention_compiled_piecewise() ), ( "Compilation mode should be CompilationMode.VLLM_COMPILE when " "cudagraph_mode piecewise cudagraphs is used, " "and attention should be in splitting_ops or " "inductor splitting should be used. " f"cudagraph_mode={self.compilation_config.cudagraph_mode}, " f"compilation_mode={self.compilation_config.mode}, " f"splitting_ops={self.compilation_config.splitting_ops}" ) self.keys_initialized = False self.specialize_lora_count = ( self.vllm_config.lora_config.specialize_active_lora if self.vllm_config.lora_config is not None else False ) # Default cudagraph_mode to NONE until initialize_cudagraph_keys is called self.cudagraph_mode = CUDAGraphMode.NONE def _compute_bs_to_padded_graph_size(self) -> None: """Pre-compute the mapping from batch size to padded graph size.""" max_size = self.compilation_config.max_cudagraph_capture_size capture_sizes = self.compilation_config.cudagraph_capture_sizes assert capture_sizes is not None, ( "Cudagraph capture sizes must be set when cudagraphs are enabled." ) self._bs_to_padded_graph_size: list[int] = [0] * (max_size + 1) for end, start in zip( capture_sizes + [max_size + 1], [0] + capture_sizes, ): for bs in range(start, end): if bs == start: self._bs_to_padded_graph_size[bs] = start else: self._bs_to_padded_graph_size[bs] = end # Validate that compile_sizes won't be changed by padding. # Only validate when cudagraphs are actually being used. if ( self.compilation_config.compile_sizes and self.cudagraph_mode != CUDAGraphMode.NONE ): for size in self.compilation_config.compile_sizes: size = int(size) if size <= self.compilation_config.max_cudagraph_capture_size: padded = self._bs_to_padded_graph_size[size] if padded != size: raise ValueError( f"compile_sizes contains {size} which would be " f"padded to {padded}. All compile_sizes must be " "values that won't be changed by cudagraph padding. " "Use values from cudagraph_capture_sizes." ) def _get_lora_cases(self) -> list[int]: """ Returns list of has_lora values for CUDA graph capture. This is the single source of truth for LoRA capture cases. """ lora_config = self.vllm_config.lora_config if lora_config is None: # No LoRA configured - single case with no LoRA return [0] # LoRA is enabled - capture graphs based on cudagraph_specialize_lora if self.compilation_config.cudagraph_specialize_lora: captured_counts = get_captured_lora_counts( lora_config.max_loras, self.specialize_lora_count ) # Specialize: capture separate graphs for with and without LoRA return [0] + captured_counts else: # No specialization: only capture graphs with LoRA active return [lora_config.max_loras + 1] def _create_padded_batch_descriptor( self, num_tokens: int, uniform_decode: bool, has_lora: bool, num_active_loras: int = 0, ) -> BatchDescriptor: max_num_seqs = self.vllm_config.scheduler_config.max_num_seqs uniform_decode_query_len = self.uniform_decode_query_len num_tokens_padded = self._bs_to_padded_graph_size[num_tokens] if uniform_decode and self.cudagraph_mode.has_mode(CUDAGraphMode.FULL): num_reqs = min(num_tokens_padded // uniform_decode_query_len, max_num_seqs) assert num_tokens_padded % uniform_decode_query_len == 0 else: uniform_decode = False num_reqs = min(num_tokens_padded, max_num_seqs) return BatchDescriptor( num_tokens=num_tokens_padded, num_reqs=num_reqs, uniform=uniform_decode, has_lora=has_lora, num_active_loras=num_active_loras, ) def add_cudagraph_key( self, runtime_mode: CUDAGraphMode, batch_descriptor: BatchDescriptor ): assert runtime_mode in [CUDAGraphMode.PIECEWISE, CUDAGraphMode.FULL], ( f"Invalid cudagraph runtime mode for keys: {runtime_mode}" ) self.cudagraph_keys[runtime_mode].add(batch_descriptor) def initialize_cudagraph_keys( self, cudagraph_mode: CUDAGraphMode, uniform_decode_query_len: int = 1 ): # This should be called only after attention backend is initialized. So we can # get the correct cudagraph mode after backend support is resolved. self.cudagraph_mode = cudagraph_mode # Early exit if cudagraphs are disabled if cudagraph_mode == CUDAGraphMode.NONE: self.keys_initialized = True return self._compute_bs_to_padded_graph_size() # Get LoRA cases to capture lora_cases = self._get_lora_cases() self.captured_lora_counts = [ lora_count for lora_count in lora_cases if lora_count ] # Note: we create all valid keys for cudagraph here but do not # guarantee all keys would be used. For example, if we allow lazy # capturing in future PR, some keys may never be triggered. if cudagraph_mode.mixed_mode() != CUDAGraphMode.NONE: assert self.compilation_config.cudagraph_capture_sizes is not None, ( "Cudagraph capture sizes must be set when mixed mode is enabled." ) for bs, num_active_loras in product( self.compilation_config.cudagraph_capture_sizes, lora_cases ): batch_desc = self._create_padded_batch_descriptor( bs, False, num_active_loras > 0, num_active_loras ) # Only relax for PIECEWISE mode. FULL mode needs exact num_reqs # because FA3's scheduler_metadata computation depends on it. if cudagraph_mode.mixed_mode() == CUDAGraphMode.PIECEWISE: batch_desc = replace(batch_desc, num_reqs=None, uniform=False) self.add_cudagraph_key(cudagraph_mode.mixed_mode(), batch_desc) # if decode cudagraph mode is FULL, and we don't already have mixed # mode full cudagraphs then add them here. if ( cudagraph_mode.decode_mode() == CUDAGraphMode.FULL and cudagraph_mode.separate_routine() ): max_num_tokens = ( uniform_decode_query_len * self.vllm_config.scheduler_config.max_num_seqs ) assert self.compilation_config.cudagraph_capture_sizes is not None, ( "Cudagraph capture sizes must be set when full mode is enabled." ) cudagraph_capture_sizes_for_decode = [ x for x in self.compilation_config.cudagraph_capture_sizes if x <= max_num_tokens and x >= uniform_decode_query_len ] for bs, num_active_loras in product( cudagraph_capture_sizes_for_decode, lora_cases ): self.add_cudagraph_key( CUDAGraphMode.FULL, self._create_padded_batch_descriptor( bs, True, num_active_loras > 0, num_active_loras ), ) self.keys_initialized = True def dispatch( self, num_tokens: int, uniform_decode: bool = False, has_lora: bool = False, num_active_loras: int = 0, valid_modes: AbstractSet[CUDAGraphMode] | None = None, invalid_modes: AbstractSet[CUDAGraphMode] | None = None, ) -> tuple[CUDAGraphMode, BatchDescriptor]: """ Given conditions(e.g.,batch descriptor and if using piecewise only), dispatch to a cudagraph runtime mode and the valid batch descriptor. A new batch descriptor is returned as we might dispatch a uniform batch to a graph that supports a more general batch (uniform to non-uniform). Args: num_tokens: Number of tokens in the batch. uniform_decode: Whether the batch is uniform decode (i.e. uniform and query length is uniform_decode_query_len). has_lora: Whether LoRA is active. num_active_loras: Number of distinct active LoRA adapters. valid_modes: Set of cudagraph modes that are allowed. None means all modes are allowed. invalid_modes: Set of cudagraph modes to exclude. Subtracted from valid_modes to compute allowed modes. (e.g., {FULL} for features like cascade attention not supported by full cudagraphs). None means no modes are excluded. """ allowed_modes = valid_modes or CUDAGraphMode.valid_runtime_modes() if invalid_modes: allowed_modes -= invalid_modes assert len(allowed_modes) >= 1, ( f"No allowed cudagraph modes: valid_modes={valid_modes}, " f"invalid_modes={invalid_modes}" ) if ( not self.keys_initialized or self.cudagraph_mode == CUDAGraphMode.NONE or num_tokens > self.compilation_config.max_cudagraph_capture_size or allowed_modes <= {CUDAGraphMode.NONE} ): return CUDAGraphMode.NONE, BatchDescriptor(num_tokens) effective_num_active_loras = num_active_loras if has_lora and num_active_loras > 0: if self.specialize_lora_count: # Find the smallest captured `num_active_loras` that is >= the current # `num_active_loras`. This is because we only capture graphs for # a subset of possible `num_active_loras` values (powers of 2). import bisect idx = bisect.bisect_left(self.captured_lora_counts, num_active_loras) if idx < len(self.captured_lora_counts): effective_num_active_loras = self.captured_lora_counts[idx] else: # When not specializing, graphs are captured only with max_loras + 1, # so we must use max_loras + 1 for dispatch to find a matching graph. assert self.vllm_config.lora_config is not None, ( "LoRA config must be set when has_lora is True." ) effective_num_active_loras = self.vllm_config.lora_config.max_loras + 1 batch_desc = self._create_padded_batch_descriptor( num_tokens, uniform_decode, has_lora, effective_num_active_loras ) if CUDAGraphMode.FULL in allowed_modes: # check if key exists for full cudagraph # For pure FULL mode, keys are registered with uniform=False. batch_desc_to_check = batch_desc if self.cudagraph_mode == CUDAGraphMode.FULL: batch_desc_to_check = replace(batch_desc, uniform=False) if batch_desc_to_check in self.cudagraph_keys[CUDAGraphMode.FULL]: return CUDAGraphMode.FULL, batch_desc_to_check if CUDAGraphMode.PIECEWISE in allowed_modes: # also check if the relaxed key exists for more "general" # piecewise cudagraph batch_desc_to_check = replace(batch_desc, num_reqs=None, uniform=False) if batch_desc_to_check in self.cudagraph_keys[CUDAGraphMode.PIECEWISE]: return CUDAGraphMode.PIECEWISE, batch_desc_to_check assert CUDAGraphMode.NONE in allowed_modes, ( f"No matching cudagraph found and NONE is not in " f"allowed_modes={allowed_modes}" ) return CUDAGraphMode.NONE, BatchDescriptor(num_tokens) def get_capture_descs(self) -> list[tuple[CUDAGraphMode, list[BatchDescriptor]]]: """ Returns capture descriptors for cudagraph capturing. Returns: List of (runtime_mode, batch_descriptors) tuples, ordered PIECEWISE first then FULL. Batch descriptors are sorted largest-first for memory efficiency. """ if not self.keys_initialized or self.cudagraph_mode == CUDAGraphMode.NONE: return [] result = [] # Return in order: PIECEWISE first, then FULL for mode in [CUDAGraphMode.PIECEWISE, CUDAGraphMode.FULL]: descs = list(self.cudagraph_keys[mode]) if descs: # Sort by num_tokens descending (largest first) descs.sort(key=lambda d: d.num_tokens, reverse=True) result.append((mode, descs)) return result
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/v1/cudagraph_dispatcher.py", "license": "Apache License 2.0", "lines": 303, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/kernels/quantization/test_flashinfer_nvfp4_scaled_mm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from nvfp4_utils import ( FLOAT4_E2M1_MAX, FLOAT8_E4M3_MAX, convert_swizzled_to_linear, dequantize_nvfp4_to_dtype, ) from vllm import _custom_ops as ops from vllm.platforms import current_platform from vllm.utils.flashinfer import ( flashinfer_scaled_fp4_mm, ) from vllm.utils.torch_utils import set_random_seed if not current_platform.has_device_capability(100): pytest.skip( reason="Nvfp4 Requires compute capability of 10 or above.", allow_module_level=True, ) DTYPES = [torch.float16, torch.bfloat16] # m, n, k SHAPES = [ (128, 128, 64), (128, 128, 128), (256, 128, 64), (128, 256, 128), (1, 128, 128), ] PAD_SHAPES = [(150, 128, 64), (128, 128, 96), (2, 128, 64), (3, 128, 96)] SHAPES.extend(PAD_SHAPES) SEEDS = [42] CUDA_DEVICES = ["cuda:0"] def get_ref_results( a_fp4, b_fp4, a_sf, b_sf, a_global_scale, b_global_scale, m, n, dtype, block_size, device, is_sf_128x4_layout, ): _, m_k = a_fp4.shape _, n_k = b_fp4.shape assert m_k == n_k a_in_dtype = dequantize_nvfp4_to_dtype( a_fp4, a_sf, a_global_scale, dtype=dtype, device=device, block_size=block_size, is_sf_128x4_layout=is_sf_128x4_layout, ) b_in_dtype = dequantize_nvfp4_to_dtype( b_fp4, b_sf, b_global_scale, dtype=dtype, device=device, block_size=block_size ) return torch.matmul(a_in_dtype, b_in_dtype.t()) @pytest.mark.parametrize("dtype", DTYPES) @pytest.mark.parametrize("shape", SHAPES) @pytest.mark.parametrize("seed", SEEDS) @pytest.mark.parametrize("device", CUDA_DEVICES) @pytest.mark.parametrize("backend", ["cutlass", "cudnn", "trtllm"]) @pytest.mark.parametrize("autotune", [False, True]) @torch.inference_mode() def test_flashinfer_nvfp4_gemm( dtype: torch.dtype, shape: tuple[int, int, int], seed: int, device: str, backend: str, autotune: bool, ) -> None: if "trtllm" in backend and dtype == torch.float16: pytest.skip("Only torch.bfloat16 is supported for TRTLLM FP4 GEMM operations") set_random_seed(seed) m, n, packed_k = shape k = packed_k * 2 block_size = 16 a_dtype = torch.randn((m, k), dtype=dtype, device=device) b_dtype = torch.randn((n, k), dtype=dtype, device=device) a_global_scale = ( (FLOAT8_E4M3_MAX * FLOAT4_E2M1_MAX) / torch.amax(a_dtype.flatten(), dim=-1) ).to(torch.float32) b_global_scale = ( (FLOAT8_E4M3_MAX * FLOAT4_E2M1_MAX) / torch.amax(b_dtype.flatten(), dim=-1) ).to(torch.float32) alpha = 1.0 / (a_global_scale * b_global_scale) # ops.scaled_fp4_quant returns swizzled scales, while weights # from checkpoints are in linear scales. # So instead of needing to swizzle for cutlass as in modelopt.py, # we need to unswizzle for trtllm here. a_fp4, a_scale_interleaved = ops.scaled_fp4_quant( a_dtype, a_global_scale, is_sf_swizzled_layout=True, backend=backend ) is_sf_128x4_layout = not (backend == "trtllm" and m <= 32) b_fp4, b_scale_interleaved = ops.scaled_fp4_quant( b_dtype, b_global_scale, is_sf_swizzled_layout=True ) # get_ref_results unswizzles the scales internally. expected_out = get_ref_results( a_fp4, b_fp4, a_scale_interleaved, b_scale_interleaved, a_global_scale, b_global_scale, m, n, dtype, block_size, device, is_sf_128x4_layout, ) import flashinfer if "trtllm" in backend: epilogue_tile_m = 128 b_fp4 = flashinfer.shuffle_matrix_a(b_fp4.view(torch.uint8), epilogue_tile_m) b_scale_interleaved = convert_swizzled_to_linear( b_scale_interleaved, n, k, block_size ) b_scale_interleaved = ( flashinfer.shuffle_matrix_sf_a( b_scale_interleaved.view(torch.uint8), epilogue_tile_m ) .reshape(b_scale_interleaved.shape) .view(torch.float8_e4m3fn) ) with flashinfer.autotune(autotune): out = flashinfer_scaled_fp4_mm( a_fp4, b_fp4, a_scale_interleaved, b_scale_interleaved, alpha, dtype, backend=backend, ) torch.testing.assert_close(out, expected_out.to(dtype=dtype), atol=1e-1, rtol=1e-1)
{ "repo_id": "vllm-project/vllm", "file_path": "tests/kernels/quantization/test_flashinfer_nvfp4_scaled_mm.py", "license": "Apache License 2.0", "lines": 145, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
vllm-project/vllm:benchmarks/benchmark_block_pool.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import gc from benchmark_utils import TimeCollector from tabulate import tabulate from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.v1.core.block_pool import BlockPool def main(args): rows = [] for allocate_block in args.allocate_blocks: # Enforce a GC collect ahead to minimize the impact among runs gc.collect() block_pool = BlockPool(num_gpu_blocks=args.num_gpu_blocks, enable_caching=True) get_blocks_times = TimeCollector(TimeCollector.US) free_blocks_times = TimeCollector(TimeCollector.US) for _ in range(args.num_iteration): with get_blocks_times: blocks = block_pool.get_new_blocks(allocate_block) with free_blocks_times: block_pool.free_blocks(blocks) rows.append( [get_blocks_times.cnt, args.num_gpu_blocks, allocate_block] + get_blocks_times.dump_avg_max() + free_blocks_times.dump_avg_max() ) print( tabulate( rows, headers=[ "Iterations", "Total\nBlocks", "Allocated\nBlocks", "Get Blocks\nAvg (us)", "Get Blocks\nMax (us)", "Free Blocks\nAvg (us)", "Free Blocks\nMax (us)", ], tablefmt="grid", floatfmt=".3f", ) ) def invoke_main() -> None: parser = FlexibleArgumentParser( description="Benchmark the performance of BlockPool for KV Cache." ) parser.add_argument("--num-gpu-blocks", type=int, default=100000) parser.add_argument( "--num-iteration", type=int, default=1000, help="Number of iterations to run to stabilize final data readings", ) parser.add_argument( "--allocate-blocks", type=int, nargs="*", default=[10, 50, 100, 500, 1000], help="Number of blocks to allocate", ) args = parser.parse_args() main(args) if __name__ == "__main__": invoke_main() # pragma: no cover
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/benchmark_block_pool.py", "license": "Apache License 2.0", "lines": 63, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:benchmarks/benchmark_ngram_proposer.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import gc import time from unittest import mock import numpy as np from benchmark_utils import TimeCollector from tabulate import tabulate from vllm.config import ( CacheConfig, DeviceConfig, LoadConfig, ModelConfig, ParallelConfig, SchedulerConfig, SpeculativeConfig, VllmConfig, ) from vllm.platforms import current_platform from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.v1.spec_decode.ngram_proposer import NgramProposer from vllm.v1.worker.gpu_input_batch import InputBatch from vllm.v1.worker.gpu_model_runner import GPUModelRunner def benchmark_propose(args): rows = [] for max_ngram in args.max_ngram: collector = TimeCollector(TimeCollector.US) model_config = ModelConfig( model="facebook/opt-125m", max_model_len=args.num_token + args.num_spec_token, tokenizer="facebook/opt-125m", tokenizer_mode="auto", dtype="auto", seed=0, trust_remote_code=False, ) proposer = NgramProposer( vllm_config=VllmConfig( model_config=model_config, speculative_config=SpeculativeConfig( prompt_lookup_min=args.min_ngram, prompt_lookup_max=max_ngram, num_speculative_tokens=args.num_spec_token, method="ngram", ), ) ) # Warm up proposer.propose(np.random.randint(0, 20, (args.num_token,))) gc.collect() for _ in range(args.num_iteration): tokens = np.random.randint(0, 20, (args.num_req, args.num_token)) with collector: for i in range(args.num_req): proposer.propose(tokens[i, :]) rows.append( [args.num_req, args.num_token, args.min_ngram, max_ngram] + collector.dump_avg_max() ) print( tabulate( rows, headers=[ "# Request", "# Token", "Min Ngram", "Max Ngram", "Avg (us)", "Max (us)", ], tablefmt="grid", floatfmt=".3f", ) ) def benchmark_batched_propose(args): NUM_SPECULATIVE_TOKENS_NGRAM = 10 PROMPT_LOOKUP_MIN = 5 PROMPT_LOOKUP_MAX = 15 MAX_MODEL_LEN = int(1e7) DEVICE = current_platform.device_type model_config = ModelConfig(model="facebook/opt-125m", runner="generate") speculative_config = SpeculativeConfig( target_model_config=model_config, target_parallel_config=ParallelConfig(), method="ngram", num_speculative_tokens=NUM_SPECULATIVE_TOKENS_NGRAM, prompt_lookup_max=PROMPT_LOOKUP_MAX, prompt_lookup_min=PROMPT_LOOKUP_MIN, ) vllm_config = VllmConfig( model_config=model_config, cache_config=CacheConfig(), speculative_config=speculative_config, device_config=DeviceConfig(device=current_platform.device_type), parallel_config=ParallelConfig(), load_config=LoadConfig(), scheduler_config=SchedulerConfig( max_model_len=model_config.max_model_len, is_encoder_decoder=model_config.is_encoder_decoder, ), ) # monkey patch vllm.v1.worker.gpu_model_runner.get_pp_group mock_pp_group = mock.MagicMock() mock_pp_group.world_size = 1 with mock.patch( "vllm.v1.worker.gpu_model_runner.get_pp_group", return_value=mock_pp_group ): runner = GPUModelRunner(vllm_config, DEVICE) # hack max model len runner.max_model_len = MAX_MODEL_LEN runner.drafter.max_model_len = MAX_MODEL_LEN dummy_input_batch = InputBatch( max_num_reqs=args.num_req, max_model_len=MAX_MODEL_LEN, max_num_batched_tokens=args.num_req * args.num_token, device=DEVICE, pin_memory=False, vocab_size=256000, block_sizes=[16], ) dummy_input_batch._req_ids = list(str(id) for id in range(args.num_req)) dummy_input_batch.num_tokens_no_spec = [args.num_token] * args.num_req dummy_input_batch.token_ids_cpu = np.random.randint( 0, 20, (args.num_req, args.num_token) ) runner.input_batch = dummy_input_batch sampled_token_ids = [[0]] * args.num_req print("Starting benchmark") # first run is warmup so ignore it for _ in range(args.num_iteration): start = time.time() runner.drafter.propose( sampled_token_ids, dummy_input_batch.num_tokens_no_spec, dummy_input_batch.token_ids_cpu, ) end = time.time() print(f"Iteration time (s): {end - start}") def invoke_main() -> None: parser = FlexibleArgumentParser( description="Benchmark the performance of N-gram speculative decode drafting" ) parser.add_argument( "--batched", action="store_true", help="consider time to prepare batch" ) parser.add_argument( "--num-iteration", type=int, default=100, help="Number of iterations to run to stabilize final data readings", ) parser.add_argument( "--num-req", type=int, default=128, help="Number of requests in the batch" ) parser.add_argument( "--num-token", type=int, default=1500, help="Number of tokens for each request" ) parser.add_argument( "--min-ngram", type=int, default=3, help="Minimum n-gram to match", ) parser.add_argument( "--max-ngram", type=int, nargs="*", default=[5, 7, 10, 15, 20], help="Maximum n-gram to match", ) parser.add_argument( "--num-spec-token", type=int, default=3, help="Number of speculative tokens to generate", ) args = parser.parse_args() if not args.batched: benchmark_propose(args) else: benchmark_batched_propose(args) """ # Example command lines: # time python3 benchmarks/benchmark_ngram_proposer.py # time python3 benchmarks/benchmark_ngram_proposer.py --batched --num-iteration 4 --num-token 1000000 --num-req 128 """ # noqa: E501 if __name__ == "__main__": invoke_main() # pragma: no cover
{ "repo_id": "vllm-project/vllm", "file_path": "benchmarks/benchmark_ngram_proposer.py", "license": "Apache License 2.0", "lines": 188, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/model_executor/models/cohere2_vision.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Adapted from vllm/model_executor/models/aya_vision.py """Command-A-Vision (Cohere2Vision) multimodal model implementation for vLLM.""" from collections.abc import Iterable, Mapping, Sequence from typing import Annotated, Literal import torch from torch import nn from transformers import BatchFeature, PretrainedConfig from transformers.models.cohere2_vision import Cohere2VisionConfig from transformers.models.cohere2_vision.image_processing_cohere2_vision_fast import ( # noqa: E501 Cohere2VisionImageProcessorFast, ) from transformers.models.cohere2_vision.processing_cohere2_vision import ( Cohere2VisionProcessor, ) from vllm.config import VllmConfig from vllm.config.multimodal import BaseDummyOptions from vllm.model_executor.layers.activation import MulAndSilu from vllm.model_executor.layers.linear import ( MergedColumnParallelLinear, RowParallelLinear, ) from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.quantization.awq import AWQConfig from vllm.multimodal import MULTIMODAL_REGISTRY from vllm.multimodal.inputs import ( MultiModalDataDict, MultiModalFieldConfig, MultiModalKwargsItems, ) from vllm.multimodal.parse import ImageProcessorItems, ImageSize, MultiModalDataItems from vllm.multimodal.processing import ( BaseDummyInputsBuilder, BaseMultiModalProcessor, BaseProcessingInfo, PromptReplacement, PromptUpdate, PromptUpdateDetails, ) from vllm.sequence import IntermediateTensors from vllm.utils.tensor_schema import TensorSchema, TensorShape from .interfaces import MultiModalEmbeddings, SupportsMultiModal, SupportsPP from .siglip import SiglipVisionModel from .utils import ( AutoWeightsLoader, WeightsMapper, init_vllm_registered_model, maybe_prefix, ) class Cohere2VisionImagePixelInputs(TensorSchema): """ Dimensions: - np: The total number of patches over each image over each prompt in the batch - c: Number of channels - h: Height of each image patch - w: Width of each image patch - bn: Batch size * number of images """ type: Literal["pixel_values"] pixel_values: Annotated[ torch.Tensor, TensorShape("np", 3, "h", "w"), ] num_patches: Annotated[ torch.Tensor, TensorShape("bn"), ] class Cohere2VisionMultiModalProjector(nn.Module): """Multimodal projector that maps vision features to text embedding space. Uses pixel shuffle downsampling followed by SwiGLU activation. """ def __init__(self, config: Cohere2VisionConfig, prefix: str = ""): super().__init__() self.downsample_factor = config.downsample_factor # Input dimension after pixel shuffle downsampling input_dim = config.vision_config.hidden_size * (config.downsample_factor**2) # MergedColumnParallelLinear expects the intermediate size to be a list # of sizes, so that it will load the weights as two separate linear # layers before applying any parallelism. # We need to divide the alignment intermediate size by 2 because # the weights are merged weights of two linear layers for SwiGLU. self.intermediate_size = config.alignment_intermediate_size // 2 self.linear_1 = MergedColumnParallelLinear( input_dim, [self.intermediate_size] * 2, bias=True, return_bias=False, prefix=f"{prefix}.linear_1", ) self.act = MulAndSilu() self.linear_2 = RowParallelLinear( self.intermediate_size, config.text_config.hidden_size, bias=True, return_bias=False, prefix=f"{prefix}.linear_2", ) def forward(self, image_features): image_features = self.pixel_shuffle(image_features) hidden_states = self.linear_1(image_features) hidden_states = self.act(hidden_states) hidden_states = self.linear_2(hidden_states) return hidden_states def pixel_shuffle(self, image_features: torch.Tensor) -> torch.Tensor: """Apply pixel shuffle downsampling to reduce spatial dimensions. Args: image_features: Input tensor of shape [B, S, D] where S = H*W Returns: Downsampled tensor with increased channel dimension """ height = width = int(image_features.shape[1] ** 0.5) x = image_features.reshape(image_features.shape[0], width, height, -1) n, h, w, c = x.size() scale_factor = 1.0 / self.downsample_factor nh = int(h * scale_factor) nw = int(w * scale_factor) x = x.reshape(n, nh, self.downsample_factor, nw, self.downsample_factor, c) x = x.permute(0, 1, 3, 2, 4, 5).contiguous() x = x.reshape(n, nh, nw, -1) return x class Cohere2VisionProcessingInfo(BaseProcessingInfo): def get_hf_config(self) -> Cohere2VisionConfig: return self.ctx.get_hf_config(Cohere2VisionConfig) def get_hf_processor(self, **kwargs: object) -> Cohere2VisionProcessor: return self.ctx.get_hf_processor(Cohere2VisionProcessor, **kwargs) def get_image_processor(self, **kwargs: object): return self.get_hf_processor(**kwargs).image_processor def get_supported_mm_limits(self) -> Mapping[str, int | None]: return {"image": None} def get_image_size_with_most_features(self) -> ImageSize: image_processor = self.get_image_processor() height = image_processor.size["height"] width = image_processor.size["width"] max_patches = image_processor.max_patches return ImageSize(height=height * max_patches, width=width) def get_num_patches( self, *, image_width: int, image_height: int, processor: Cohere2VisionProcessor, mm_kwargs: Mapping[str, object], ) -> int: """ Calculate the number of image patches for a given image. Uses the HF processor to determine the actual number of patches. """ image_processor: Cohere2VisionImageProcessorFast = processor.image_processor return image_processor.get_number_of_image_patches( image_height, image_width, self.ctx.get_merged_mm_kwargs(mm_kwargs), ) class Cohere2VisionDummyInputsBuilder( BaseDummyInputsBuilder[Cohere2VisionProcessingInfo] ): def get_dummy_text(self, mm_counts: Mapping[str, int]) -> str: num_images = mm_counts.get("image", 0) processor = self.info.get_hf_processor() image_token = processor.image_token return image_token * num_images def get_dummy_mm_data( self, seq_len: int, mm_counts: Mapping[str, int], mm_options: Mapping[str, BaseDummyOptions], ) -> MultiModalDataDict: num_images = mm_counts.get("image", 0) image_size = self.info.get_image_size_with_most_features() image_overrides = mm_options.get("image") return { "image": self._get_dummy_images( width=image_size.width, height=image_size.height, num_images=num_images, overrides=image_overrides, ) } class Cohere2VisionMultiModalProcessor( BaseMultiModalProcessor[Cohere2VisionProcessingInfo] ): def _call_hf_processor( self, prompt: str, mm_data: Mapping[str, object], mm_kwargs: Mapping[str, object], tok_kwargs: Mapping[str, object], ) -> BatchFeature: processed_outputs = super()._call_hf_processor( prompt, mm_data, mm_kwargs, tok_kwargs, ) # Ensure num_patches is available for proper tensor splitting if ( "num_patches" not in processed_outputs and (images := mm_data.get("images")) is not None ): hf_processor = self.info.get_hf_processor(**mm_kwargs) # Fallback calculation if HF processor didn't provide num_patches mm_items = self.info.parse_mm_data({"image": images}, validate=False) parsed_images = mm_items.get_items("image", ImageProcessorItems) num_patches = [ self.info.get_num_patches( image_width=parsed_images.get_image_size(i).width, image_height=parsed_images.get_image_size(i).height, processor=hf_processor, mm_kwargs=mm_kwargs, ) for i in range(len(parsed_images)) ] processed_outputs["num_patches"] = torch.tensor(num_patches) return processed_outputs def _get_mm_fields_config( self, hf_inputs: BatchFeature, hf_processor_mm_kwargs: Mapping[str, object], ) -> Mapping[str, MultiModalFieldConfig]: num_patches = hf_inputs.get("num_patches", torch.empty(0)) return dict( pixel_values=MultiModalFieldConfig.flat_from_sizes("image", num_patches), num_patches=MultiModalFieldConfig.batched("image"), image_embeds=MultiModalFieldConfig.batched("image"), ) def _get_prompt_updates( self, mm_items: MultiModalDataItems, hf_processor_mm_kwargs: Mapping[str, object], out_mm_kwargs: MultiModalKwargsItems, ) -> Sequence[PromptUpdate]: hf_processor = self.info.get_hf_processor(**hf_processor_mm_kwargs) image_token = hf_processor.image_token img_tokens_per_tile = int(hf_processor.patch_size**2) img_line_break_token = hf_processor.img_line_break_token boi_token = hf_processor.boi_token eoi_token = hf_processor.eoi_token def get_replacement(item_idx: int): images = mm_items.get_items("image", ImageProcessorItems) image_size: ImageSize = images.get_image_size(item_idx) num_patches = self.info.get_num_patches( image_width=image_size.width, image_height=image_size.height, processor=hf_processor, mm_kwargs=hf_processor_mm_kwargs, ) patch_tokens = image_token * img_tokens_per_tile + img_line_break_token repl = f"{boi_token}{patch_tokens * num_patches}{eoi_token}" return PromptUpdateDetails.select_text(repl, image_token) return [ PromptReplacement( modality="image", target=image_token, replacement=get_replacement, ) ] @MULTIMODAL_REGISTRY.register_processor( Cohere2VisionMultiModalProcessor, info=Cohere2VisionProcessingInfo, dummy_inputs=Cohere2VisionDummyInputsBuilder, ) class Cohere2VisionForConditionalGeneration(nn.Module, SupportsMultiModal, SupportsPP): hf_to_vllm_mapper = WeightsMapper( orig_to_new_prefix={ "model.vision_tower.": "vision_tower.", "model.multi_modal_projector.": "multi_modal_projector.", "model.language_model.": "language_model.model.", "lm_head.": "language_model.lm_head.", } ) def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__() config: Cohere2VisionConfig = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config multimodal_config = vllm_config.model_config.multimodal_config self.config = config self.quant_config = quant_config self.multimodal_config = multimodal_config self._patch_quant_config(config, quant_config) with self._mark_tower_model(vllm_config, "image"): self.vision_tower = SiglipVisionModel( config.vision_config, quant_config, prefix=maybe_prefix(prefix, "vision_tower"), ) self.multi_modal_projector = Cohere2VisionMultiModalProjector( config, prefix=maybe_prefix(prefix, "multi_modal_projector") ) with self._mark_language_model(vllm_config): self.language_model = init_vllm_registered_model( vllm_config=vllm_config, hf_config=config.text_config, prefix=maybe_prefix(prefix, "language_model"), architectures=config.text_config.architectures, ) @property def dtype(self): return next(self.parameters()).dtype def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]: loader = AutoWeightsLoader(self) return loader.load_weights(weights, mapper=self.hf_to_vllm_mapper) def _process_image_input( self, image_input: Cohere2VisionImagePixelInputs, **kwargs ) -> list[torch.Tensor]: """Process image pixels through vision tower and projector. Args: image_input: Validated image input containing pixel values and patch counts Returns: List of flattened image embeddings, one per image """ pixel_values = image_input["pixel_values"] num_patches = image_input["num_patches"] # Extract visual features image_features = self.vision_tower(pixel_values) # Project to text embedding space image_embeds = self.multi_modal_projector(image_features) # Split and flatten embeddings per image return [e.flatten(0, 2) for e in image_embeds.split(num_patches.tolist())] def _parse_and_validate_image_input( self, **kwargs: object ) -> Cohere2VisionImagePixelInputs | None: pixel_values = kwargs.pop("pixel_values", None) num_patches = kwargs.pop("num_patches", None) image_embeds = kwargs.pop("image_embeds", None) assert image_embeds is None, "Cohere2Vision does not support image_embeds." if pixel_values is None: return None return Cohere2VisionImagePixelInputs( type="pixel_values", pixel_values=pixel_values, num_patches=num_patches, resolve_bindings={ "h": self.config.vision_config.image_size, "w": self.config.vision_config.image_size, }, ) def _patch_quant_config( self, config: PretrainedConfig, quant_config: QuantizationConfig ): # the awq models from OpenGVLab missing `modules_to_not_convert` # patch the quant_config to add `modules_to_not_convert` back if isinstance(quant_config, AWQConfig): text_config = config.text_config llm_quant_config = getattr(text_config, "quantization_config", None) if (not quant_config.modules_to_not_convert) and ( llm_quant_config is not None ): quant_config.modules_to_not_convert.append("vision_tower") def embed_multimodal(self, **kwargs: object) -> MultiModalEmbeddings: image_input = self._parse_and_validate_image_input(**kwargs) if image_input is None: return [] return self._process_image_input(image_input, **kwargs) def forward( self, input_ids: torch.Tensor | None, positions: torch.Tensor, intermediate_tensors: IntermediateTensors | None = None, inputs_embeds: torch.Tensor | None = None, **kwargs: object, ) -> torch.Tensor | IntermediateTensors: if intermediate_tensors is not None: inputs_embeds = None hidden_states = self.language_model.model( input_ids=input_ids, positions=positions, intermediate_tensors=intermediate_tensors, inputs_embeds=inputs_embeds, ) return hidden_states def compute_logits( self, hidden_states: torch.Tensor, ) -> torch.Tensor | None: return self.language_model.compute_logits(hidden_states)
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/model_executor/models/cohere2_vision.py", "license": "Apache License 2.0", "lines": 378, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:vllm/config/scheduler.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable from dataclasses import InitVar from typing import TYPE_CHECKING, Any, ClassVar, Literal, cast from pydantic import Field, field_validator from typing_extensions import Self from vllm.config.utils import config from vllm.logger import init_logger from vllm.utils.hashing import safe_hash from vllm.utils.import_utils import resolve_obj_by_qualname if TYPE_CHECKING: from vllm.v1.core.sched.interface import SchedulerInterface logger = init_logger(__name__) RunnerType = Literal["generate", "pooling", "draft"] SchedulerPolicy = Literal["fcfs", "priority"] @config class SchedulerConfig: """Scheduler configuration.""" max_model_len: InitVar[int] """Maximum length of a sequence (including prompt and generated text). Note: This is stored in the ModelConfig, and is used only here to provide fallbacks and validate other attributes.""" is_encoder_decoder: InitVar[bool] """True if the model is an encoder-decoder model. Note: This is stored in the ModelConfig, and is used only here to disable chunked prefill and prefix caching for encoder-decoder models. """ DEFAULT_MAX_NUM_BATCHED_TOKENS: ClassVar[int] = 2048 DEFAULT_MAX_NUM_SEQS: ClassVar[int] = 128 runner_type: RunnerType = "generate" """The runner type to launch for the model.""" max_num_batched_tokens: int = Field(default=DEFAULT_MAX_NUM_BATCHED_TOKENS, ge=1) """Maximum number of tokens that can be processed in a single iteration. The default value here is mainly for convenience when testing. In real usage, this should be set in `EngineArgs.create_engine_config`. """ max_num_scheduled_tokens: int | None = Field(default=None) """Maximum number of tokens that the scheduler may issue in a single iteration. This is usually equal to max_num_batched_tokens, but can be smaller in cases when the model might append tokens into the batch (such as speculative decoding). Defaults to max_num_batched_tokens.""" max_num_seqs: int = Field(default=DEFAULT_MAX_NUM_SEQS, ge=1) """Maximum number of sequences to be processed in a single iteration. The default value here is mainly for convenience when testing. In real usage, this should be set in `EngineArgs.create_engine_config`. """ max_num_partial_prefills: int = Field(default=1, ge=1) """For chunked prefill, the maximum number of sequences that can be partially prefilled concurrently.""" max_long_partial_prefills: int = Field(default=1, ge=1) """For chunked prefill, the maximum number of prompts longer than long_prefill_token_threshold that will be prefilled concurrently. Setting this less than max_num_partial_prefills will allow shorter prompts to jump the queue in front of longer prompts in some cases, improving latency.""" long_prefill_token_threshold: int = 0 """For chunked prefill, a request is considered long if the prompt is longer than this number of tokens.""" enable_chunked_prefill: bool = True """If True, prefill requests can be chunked based on the remaining `max_num_batched_tokens`. The default value here is mainly for convenience when testing. In real usage, this should be set in `EngineArgs.create_engine_config`. """ is_multimodal_model: bool = False """True if the model is multimodal.""" # TODO (ywang96): Make this configurable. max_num_encoder_input_tokens: int = Field(init=False) """Multimodal encoder compute budget, only used in V1. NOTE: This is not currently configurable. It will be overridden by max_num_batched_tokens in case max multimodal embedding size is larger.""" # TODO (ywang96): Make this configurable. encoder_cache_size: int = Field(init=False) """Multimodal encoder cache size, only used in V1. NOTE: This is not currently configurable. It will be overridden by max_num_batched_tokens in case max multimodal embedding size is larger.""" policy: SchedulerPolicy = "fcfs" """The scheduling policy to use:\n - "fcfs" means first come first served, i.e. requests are handled in order of arrival.\n - "priority" means requests are handled based on given priority (lower value means earlier handling) and time of arrival deciding any ties).""" disable_chunked_mm_input: bool = False """If set to true and chunked prefill is enabled, we do not want to partially schedule a multimodal item. Only used in V1 This ensures that if a request has a mixed prompt (like text tokens TTTT followed by image tokens IIIIIIIIII) where only some image tokens can be scheduled (like TTTTIIIII, leaving IIIII), it will be scheduled as TTTT in one step and IIIIIIIIII in the next.""" # scheduler class or path. "vllm.v1.core.sched.scheduler.Scheduler" # (default) or "mod.custom_class". scheduler_cls: str | type[object] | None = Field(default=None) """The scheduler class to use. "vllm.v1.core.sched.scheduler.Scheduler" is the default scheduler. Can be a class directly or the path to a class of form "mod.custom_class".""" disable_hybrid_kv_cache_manager: bool | None = None """If set to True, KV cache manager will allocate the same size of KV cache for all attention layers even if there are multiple type of attention layers like full attention and sliding window attention. If set to None, the default value will be determined based on the environment and starting configuration. """ async_scheduling: bool | None = Field(default=None) """If set to False, disable async scheduling. Async scheduling helps to avoid gaps in GPU utilization, leading to better latency and throughput. """ stream_interval: int = Field(default=1, ge=1) """The interval (or buffer size) for streaming in terms of token length. A smaller value (1) makes streaming smoother by sending each token immediately, while a larger value (e.g., 10) reduces host overhead and may increase throughput by batching multiple tokens before sending.""" @staticmethod def default_factory(**kwargs): """ Factory method to create `SchedulerConfig` with default values for `InitVar`s. """ if "max_model_len" not in kwargs: kwargs["max_model_len"] = 8192 if "is_encoder_decoder" not in kwargs: kwargs["is_encoder_decoder"] = False return SchedulerConfig(**kwargs) def get_scheduler_cls(self) -> type["SchedulerInterface"]: if self.scheduler_cls is None: if self.async_scheduling: from vllm.v1.core.sched.async_scheduler import AsyncScheduler return AsyncScheduler from vllm.v1.core.sched.scheduler import Scheduler return Scheduler # This warning can be removed once the Scheduler interface is # finalized and we can maintain support for scheduler classes that # implement it logger.warning_once( "Using custom scheduler class %s. This scheduler interface is " "not public and compatibility may not be maintained.", self.scheduler_cls, ) if not isinstance(self.scheduler_cls, str): return cast(type["SchedulerInterface"], self.scheduler_cls) return resolve_obj_by_qualname(self.scheduler_cls) def compute_hash(self) -> str: """ WARNING: Whenever a new field is added to this config, ensure that it is included in the factors list if it affects the computation graph. Provide a hash that uniquely identifies all the configs that affect the structure of the computation graph from input ids/embeddings to the final hidden states, excluding anything before input ids/embeddings and after the final hidden states. """ factors: list[Any] = [] # max_num_batched_tokens need to be included in the hash due # to two reasons: # 1. LoRA creates static buffers based on max_num_batched_tokens. # The tensor sizes and strides get captured in the torch.compile # graph explicitly. # 2. Inductor decides whether using 32-bit or 64-bit indexing integer # based on the data sizes. `max_num_batched_tokens` has an # impact on that. For more details, please check # https://github.com/vllm-project/vllm/issues/29585 factors.append(self.max_num_batched_tokens) hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest() return hash_str @field_validator("scheduler_cls", "async_scheduling", mode="wrap") @classmethod def _skip_none_validation(cls, value: Any, handler: Callable) -> Any: """Skip validation if the value is `None` when initialisation is delayed.""" return None if value is None else handler(value) def __post_init__(self, max_model_len: int, is_encoder_decoder: bool) -> None: if is_encoder_decoder: # Chunked prefill should be disabled for encoder-decoder models. self.disable_chunked_mm_input = True self.enable_chunked_prefill = False self.long_prefill_token_threshold = 0 logger.info( "Encoder-decoder models do not support chunked prefill nor" " prefix caching; disabling both." ) self.max_num_encoder_input_tokens = self.max_num_batched_tokens self.encoder_cache_size = self.max_num_batched_tokens if self.enable_chunked_prefill: logger.info( "Chunked prefill is enabled with max_num_batched_tokens=%d.", self.max_num_batched_tokens, ) if self.max_num_partial_prefills > 1: if self.long_prefill_token_threshold == 0: self.long_prefill_token_threshold = int(max_model_len * 0.04) logger.info( "Concurrent partial prefills enabled with " "max_num_partial_prefills=%d, max_long_partial_prefills=%d, " "long_prefill_token_threshold=%d", self.max_num_partial_prefills, self.max_long_partial_prefills, self.long_prefill_token_threshold, ) self.verify_max_model_len(max_model_len) def verify_max_model_len(self, max_model_len: int) -> Self: if ( self.max_num_batched_tokens < max_model_len and not self.enable_chunked_prefill ): raise ValueError( f"max_num_batched_tokens ({self.max_num_batched_tokens}) is " f"smaller than max_model_len ({max_model_len}). " "This effectively limits the maximum sequence length to " "max_num_batched_tokens and makes vLLM reject longer " "sequences. Please increase max_num_batched_tokens or " "decrease max_model_len." ) if self.max_num_batched_tokens < self.max_num_seqs: raise ValueError( f"max_num_batched_tokens ({self.max_num_batched_tokens}) must " "be greater than or equal to max_num_seqs " f"({self.max_num_seqs})." ) if self.max_num_batched_tokens > self.max_num_seqs * max_model_len: logger.warning( "max_num_batched_tokens (%d) exceeds max_num_seqs " "* max_model_len (%d). This may lead to unexpected behavior.", self.max_num_batched_tokens, self.max_num_seqs * max_model_len, ) if self.max_num_partial_prefills > 1: if not self.enable_chunked_prefill: raise ValueError( "Chunked prefill must be enabled to set " "max_num_partial_prefills > 1." ) if self.long_prefill_token_threshold > max_model_len: raise ValueError( "long_prefill_token_threshold " f"({self.long_prefill_token_threshold}) cannot be greater " f"than the max_model_len ({max_model_len})." ) if self.max_long_partial_prefills > self.max_num_partial_prefills: raise ValueError( f"{self.max_long_partial_prefills=} must be less than or equal to " f"{self.max_num_partial_prefills=}." ) return self
{ "repo_id": "vllm-project/vllm", "file_path": "vllm/config/scheduler.py", "license": "Apache License 2.0", "lines": 241, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
vllm-project/vllm:tests/models/language/pooling/test_auto_prefix_cache_support.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from transformers import AutoModelForSequenceClassification from tests.models.language.pooling.embed_utils import run_embedding_correctness_test @pytest.mark.parametrize( "model", ["jason9693/Qwen2.5-1.5B-apeach"], ) @pytest.mark.parametrize("dtype", ["half"]) def test_classify_models( hf_runner, vllm_runner, example_prompts, model: str, dtype: str, ) -> None: # example_prompts is too short for testing prefix_caching example_prompts = [s * 10 for s in example_prompts] with vllm_runner( model, max_model_len=512, dtype=dtype, enable_prefix_caching=True ) as vllm_model: vllm_config = vllm_model.llm.llm_engine.vllm_config cache_config = vllm_config.cache_config assert cache_config.enable_prefix_caching # First Run vllm_model.classify(example_prompts) # assert prefix_caching works pooling_outputs = vllm_model.llm.encode( example_prompts, pooling_task="classify" ) for output in pooling_outputs: assert output.num_cached_tokens > 0 vllm_outputs = [req_output.outputs.data for req_output in pooling_outputs] with hf_runner( model, dtype=dtype, auto_cls=AutoModelForSequenceClassification ) as hf_model: hf_outputs = hf_model.classify(example_prompts) for hf_output, vllm_output in zip(hf_outputs, vllm_outputs): hf_output = torch.tensor(hf_output) vllm_output = torch.tensor(vllm_output) assert torch.allclose( hf_output, vllm_output, 1e-3 if dtype == "float" else 1e-2 ) @pytest.mark.parametrize( "model", ["Qwen/Qwen3-Embedding-0.6B"], ) @pytest.mark.parametrize("dtype", ["half"]) def test_embed_models( hf_runner, vllm_runner, example_prompts, model: str, dtype: str, ): # example_prompts is too short for testing prefix_caching example_prompts = [str(s).strip() * 10 for s in example_prompts] with vllm_runner( model, runner="pooling", max_model_len=None, enable_prefix_caching=True, ) as vllm_model: vllm_config = vllm_model.llm.llm_engine.vllm_config cache_config = vllm_config.cache_config assert cache_config.enable_prefix_caching # First Run vllm_model.embed(example_prompts) # assert prefix_caching works pooling_outputs = vllm_model.llm.encode(example_prompts, pooling_task="embed") for output in pooling_outputs: assert output.num_cached_tokens > 0 vllm_outputs = [req_output.outputs.data for req_output in pooling_outputs] with hf_runner( model, is_sentence_transformer=True, ) as hf_model: run_embedding_correctness_test(hf_model, example_prompts, vllm_outputs) @pytest.mark.parametrize( "model", [ "intfloat/e5-small", "Alibaba-NLP/gte-Qwen2-1.5B-instruct", # is_causal == False "papluca/xlm-roberta-base-language-detection", ], ) @pytest.mark.parametrize("dtype", ["half"]) def test_non_causal_models( hf_runner, vllm_runner, example_prompts, model: str, dtype: str ) -> None: with vllm_runner(model, max_model_len=512, dtype=dtype) as vllm_model: vllm_config = vllm_model.llm.llm_engine.vllm_config cache_config = vllm_config.cache_config assert not cache_config.enable_prefix_caching
{ "repo_id": "vllm-project/vllm", "file_path": "tests/models/language/pooling/test_auto_prefix_cache_support.py", "license": "Apache License 2.0", "lines": 96, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test