repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/examples/others/lmcache/disagg_prefill_lmcache_v0.py
examples/others/lmcache/disagg_prefill_lmcache_v0.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ This file demonstrates the example usage of disaggregated prefilling with LMCache. We will launch 2 vllm instances (GPU 0 for prefill and GPU 1 for decode), and launch an additional LMCache server. KV cache is transferred in the following manner: vLLM prefill node -> LMCache server -> vLLM decode node. Note that `pip install lmcache` is needed to run this example. Learn more about LMCache in https://github.com/LMCache/LMCache. """ import os import subprocess import time from multiprocessing import Event, Process from lmcache.experimental.cache_engine import LMCacheEngineBuilder from lmcache.integration.vllm.utils import ENGINE_NAME from vllm import LLM, SamplingParams from vllm.config import KVTransferConfig # LMCache-related environment variables # The port to start LMCache server port = 8100 # Use experimental features in LMCache os.environ["LMCACHE_USE_EXPERIMENTAL"] = "True" # LMCache is set to use 256 tokens per chunk os.environ["LMCACHE_CHUNK_SIZE"] = "256" # Disable local CPU backend in LMCache os.environ["LMCACHE_LOCAL_CPU"] = "False" # Set local CPU memory buffer limit to 5.0 GB os.environ["LMCACHE_MAX_LOCAL_CPU_SIZE"] = "5.0" # Set the remote URL for LMCache server os.environ["LMCACHE_REMOTE_URL"] = f"lm://localhost:{port}" # Set the serializer/deserializer between vllm and LMCache server # `naive` indicates using raw bytes of the tensor without any compression os.environ["LMCACHE_REMOTE_SERDE"] = "naive" prompts = [ "Hello, how are you?" * 1000, ] def run_prefill(prefill_done, prompts): # We use GPU 0 for prefill node. os.environ["CUDA_VISIBLE_DEVICES"] = "0" sampling_params = SamplingParams(temperature=0, top_p=0.95, max_tokens=1) ktc = KVTransferConfig( kv_connector="LMCacheConnector", kv_role="kv_producer", kv_rank=0, kv_parallel_size=2, ) # Set GPU memory utilization to 0.8 for an A40 GPU with 40GB # memory. Reduce the value if your GPU has less memory. llm = LLM( model="mistralai/Mistral-7B-Instruct-v0.2", kv_transfer_config=ktc, max_model_len=8000, gpu_memory_utilization=0.8, enforce_eager=True, ) # llm.generate(prompts, sampling_params) outputs = llm.generate(prompts, sampling_params) for output in outputs: generated_text = output.outputs[0].text print(f"Generated text: {generated_text!r}") print("Prefill node is finished.") prefill_done.set() # Clean up lmcache backend LMCacheEngineBuilder.destroy(ENGINE_NAME) def run_decode(prefill_done, prompts, timeout=1): # We use GPU 1 for decode node. os.environ["CUDA_VISIBLE_DEVICES"] = "1" sampling_params = SamplingParams(temperature=0, top_p=0.95, max_tokens=10) ktc = KVTransferConfig( kv_connector="LMCacheConnector", kv_role="kv_consumer", kv_rank=1, kv_parallel_size=2, ) # Set GPU memory utilization to 0.8 for an A40 GPU with 40GB # of memory. Reduce the value if your GPU has less memory. llm = LLM( model="mistralai/Mistral-7B-Instruct-v0.2", kv_transfer_config=ktc, max_model_len=8000, gpu_memory_utilization=0.8, enforce_eager=True, ) print("Waiting for prefill node to finish...") prefill_done.wait() time.sleep(timeout) outputs = llm.generate(prompts, sampling_params) for output in outputs: generated_text = output.outputs[0].text print(f"Generated text: {generated_text!r}") # Clean up lmcache backend LMCacheEngineBuilder.destroy(ENGINE_NAME) def run_lmcache_server(port): server_proc = subprocess.Popen( ["python", "-m", "lmcache.experimental.server", "localhost", str(port)] ) return server_proc def main(): prefill_done = Event() prefill_process = Process(target=run_prefill, args=(prefill_done, prompts)) decode_process = Process(target=run_decode, args=(prefill_done, prompts)) lmcache_server_process = run_lmcache_server(port) # Start prefill node prefill_process.start() # Start decode node decode_process.start() # Clean up the processes decode_process.join() prefill_process.terminate() lmcache_server_process.terminate() lmcache_server_process.wait() if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/examples/others/lmcache/disagg_prefill_lmcache_v1/disagg_proxy_server.py
examples/others/lmcache/disagg_prefill_lmcache_v1/disagg_proxy_server.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import os import time from contextlib import asynccontextmanager import httpx import numpy as np from fastapi import FastAPI, Request from fastapi.responses import StreamingResponse @asynccontextmanager async def lifespan(app: FastAPI): """ Lifespan context manager to handle startup and shutdown events. """ # Startup: Initialize clients prefiller_base_url = ( f"http://{global_args.prefiller_host}:{global_args.prefiller_port}/v1" ) decoder_base_url = ( f"http://{global_args.decoder_host}:{global_args.decoder_port}/v1" ) app.state.prefill_client = httpx.AsyncClient( timeout=None, base_url=prefiller_base_url, limits=httpx.Limits( max_connections=None, max_keepalive_connections=None, ), ) app.state.decode_client = httpx.AsyncClient( timeout=None, base_url=decoder_base_url, limits=httpx.Limits( max_connections=None, max_keepalive_connections=None, ), ) yield # Shutdown: Close clients await app.state.prefill_client.aclose() await app.state.decode_client.aclose() # Update FastAPI app initialization to use lifespan app = FastAPI(lifespan=lifespan) class StatsCalculator: def __init__(self): self._stats = [] self._last_log_time = time.time() def add(self, value): self._stats.append(value) if time.time() - self._last_log_time > 5: self._log_stats() self._last_log_time = time.time() def _log_stats(self): # Print average, median, and 99th percentile np_arr = np.array(self._stats) output_str = ( f"\nNum requests: {len(self._stats)}" + "\nPrefill node TTFT stats:" + f"\n - Average (ms): {np.mean(np_arr)}" + f"\n - Median (ms): {np.median(np_arr)}" + f"\n - 99th Percentile (ms): {np.percentile(np_arr, 99)}\n" ) print( "===============================", output_str, "===============================", ) stats_calculator = StatsCalculator() counter = 0 def parse_args(): parser = argparse.ArgumentParser() parser.add_argument("--port", type=int, default=8000) parser.add_argument("--host", type=str, default="localhost") parser.add_argument("--prefiller-host", type=str, default="localhost") parser.add_argument("--prefiller-port", type=int, default=8100) parser.add_argument("--decoder-host", type=str, default="localhost") parser.add_argument("--decoder-port", type=int, default=8200) args = parser.parse_args() return args # Initialize variables to hold the persistent clients app.state.prefill_client = None app.state.decode_client = None async def send_request_to_service( client: httpx.AsyncClient, endpoint: str, req_data: dict ): """ Send a request to a service using a persistent client. """ req_data = req_data.copy() req_data["max_tokens"] = 1 if "max_completion_tokens" in req_data: req_data["max_completion_tokens"] = 1 headers = {"Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}"} response = await client.post(endpoint, json=req_data, headers=headers) response.raise_for_status() # read/consume the response body to release the connection # otherwise, it would http.ReadError await response.aread() return response async def stream_service_response( client: httpx.AsyncClient, endpoint: str, req_data: dict ): """ Asynchronously stream the response from a service using a persistent client. """ headers = {"Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}"} async with client.stream( "POST", endpoint, json=req_data, headers=headers ) as response: response.raise_for_status() async for chunk in response.aiter_bytes(): yield chunk @app.post("/v1/completions") async def handle_completions(request: Request): global counter, stats_calculator counter += 1 st = time.time() try: req_data = await request.json() # Send request to prefill service, ignore the response await send_request_to_service( app.state.prefill_client, "/completions", req_data ) et = time.time() stats_calculator.add(et - st) # Stream response from decode service async def generate_stream(): async for chunk in stream_service_response( app.state.decode_client, "/completions", req_data ): yield chunk return StreamingResponse(generate_stream(), media_type="text/event-stream") except Exception as e: import sys import traceback exc_info = sys.exc_info() print("Error occurred in disagg prefill proxy server - completions endpoint") print(e) print("".join(traceback.format_exception(*exc_info))) raise @app.post("/v1/chat/completions") async def handle_chat_completions(request: Request): global counter, stats_calculator counter += 1 st = time.time() try: req_data = await request.json() # Send request to prefill service, ignore the response await send_request_to_service( app.state.prefill_client, "/chat/completions", req_data ) et = time.time() stats_calculator.add(et - st) # Stream response from decode service async def generate_stream(): async for chunk in stream_service_response( app.state.decode_client, "/chat/completions", req_data ): yield chunk return StreamingResponse(generate_stream(), media_type="text/event-stream") except Exception as e: import sys import traceback exc_info = sys.exc_info() print( "Error occurred in disagg prefill proxy server - chat completions endpoint" ) print(e) print("".join(traceback.format_exception(*exc_info))) raise if __name__ == "__main__": global global_args global_args = parse_args() import uvicorn uvicorn.run(app, host=global_args.host, port=global_args.port)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/cmake/hipify.py
cmake/hipify.py
#!/usr/bin/env python3 # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # # A command line tool for running pytorch's hipify preprocessor on CUDA # source files. # # See https://github.com/ROCm/hipify_torch # and <torch install dir>/utils/hipify/hipify_python.py # import argparse import os import shutil from torch.utils.hipify.hipify_python import hipify if __name__ == "__main__": parser = argparse.ArgumentParser() # Project directory where all the source + include files live. parser.add_argument( "-p", "--project_dir", help="The project directory.", ) # Directory where hipified files are written. parser.add_argument( "-o", "--output_dir", help="The output directory.", ) # Source files to convert. parser.add_argument( "sources", help="Source files to hipify.", nargs="*", default=[] ) args = parser.parse_args() # Limit include scope to project_dir only includes = [os.path.join(args.project_dir, "*")] # Get absolute path for all source files. extra_files = [os.path.abspath(s) for s in args.sources] # Copy sources from project directory to output directory. # The directory might already exist to hold object files so we ignore that. shutil.copytree(args.project_dir, args.output_dir, dirs_exist_ok=True) hipify_result = hipify( project_directory=args.project_dir, output_directory=args.output_dir, header_include_dirs=[], includes=includes, extra_files=extra_files, show_detailed=True, is_pytorch_extension=True, hipify_extra_files_only=True, ) hipified_sources = [] for source in args.sources: s_abs = os.path.abspath(source) hipified_s_abs = ( hipify_result[s_abs].hipified_path if ( s_abs in hipify_result and hipify_result[s_abs].hipified_path is not None ) else s_abs ) hipified_sources.append(hipified_s_abs) assert len(hipified_sources) == len(args.sources) # Print hipified source files. print("\n".join(hipified_sources))
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/connections.py
vllm/connections.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Mapping, MutableMapping from pathlib import Path from urllib.parse import urlparse import aiohttp import requests from vllm.version import __version__ as VLLM_VERSION class HTTPConnection: """Helper class to send HTTP requests.""" def __init__(self, *, reuse_client: bool = True) -> None: super().__init__() self.reuse_client = reuse_client self._sync_client: requests.Session | None = None self._async_client: aiohttp.ClientSession | None = None def get_sync_client(self) -> requests.Session: if self._sync_client is None or not self.reuse_client: self._sync_client = requests.Session() return self._sync_client # NOTE: We intentionally use an async function even though it is not # required, so that the client is only accessible inside async event loop async def get_async_client(self) -> aiohttp.ClientSession: if self._async_client is None or not self.reuse_client: self._async_client = aiohttp.ClientSession(trust_env=True) return self._async_client def _validate_http_url(self, url: str): parsed_url = urlparse(url) if parsed_url.scheme not in ("http", "https"): raise ValueError( "Invalid HTTP URL: A valid HTTP URL must have scheme 'http' or 'https'." ) def _headers(self, **extras: str) -> MutableMapping[str, str]: return {"User-Agent": f"vLLM/{VLLM_VERSION}", **extras} def get_response( self, url: str, *, stream: bool = False, timeout: float | None = None, extra_headers: Mapping[str, str] | None = None, allow_redirects: bool = True, ): self._validate_http_url(url) client = self.get_sync_client() extra_headers = extra_headers or {} return client.get( url, headers=self._headers(**extra_headers), stream=stream, timeout=timeout, allow_redirects=allow_redirects, ) async def get_async_response( self, url: str, *, timeout: float | None = None, extra_headers: Mapping[str, str] | None = None, allow_redirects: bool = True, ): self._validate_http_url(url) client = await self.get_async_client() extra_headers = extra_headers or {} return client.get( url, headers=self._headers(**extra_headers), timeout=timeout, allow_redirects=allow_redirects, ) def get_bytes( self, url: str, *, timeout: float | None = None, allow_redirects: bool = True ) -> bytes: with self.get_response( url, timeout=timeout, allow_redirects=allow_redirects ) as r: r.raise_for_status() return r.content async def async_get_bytes( self, url: str, *, timeout: float | None = None, allow_redirects: bool = True, ) -> bytes: async with await self.get_async_response( url, timeout=timeout, allow_redirects=allow_redirects ) as r: r.raise_for_status() return await r.read() def get_text(self, url: str, *, timeout: float | None = None) -> str: with self.get_response(url, timeout=timeout) as r: r.raise_for_status() return r.text async def async_get_text( self, url: str, *, timeout: float | None = None, ) -> str: async with await self.get_async_response(url, timeout=timeout) as r: r.raise_for_status() return await r.text() def get_json(self, url: str, *, timeout: float | None = None) -> str: with self.get_response(url, timeout=timeout) as r: r.raise_for_status() return r.json() async def async_get_json( self, url: str, *, timeout: float | None = None, ) -> str: async with await self.get_async_response(url, timeout=timeout) as r: r.raise_for_status() return await r.json() def download_file( self, url: str, save_path: Path, *, timeout: float | None = None, chunk_size: int = 128, ) -> Path: with self.get_response(url, timeout=timeout) as r: r.raise_for_status() with save_path.open("wb") as f: for chunk in r.iter_content(chunk_size): f.write(chunk) return save_path async def async_download_file( self, url: str, save_path: Path, *, timeout: float | None = None, chunk_size: int = 128, ) -> Path: async with await self.get_async_response(url, timeout=timeout) as r: r.raise_for_status() with save_path.open("wb") as f: async for chunk in r.content.iter_chunked(chunk_size): f.write(chunk) return save_path global_http_connection = HTTPConnection() """ The global [`HTTPConnection`][vllm.connections.HTTPConnection] instance used by vLLM. """
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/logits_process.py
vllm/logits_process.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable, Sequence from typing import TypeAlias import torch from vllm.tokenizers import TokenizerLike LogitsProcessor: TypeAlias = ( Callable[[list[int], torch.Tensor], torch.Tensor] | Callable[[list[int], list[int], torch.Tensor], torch.Tensor] ) """LogitsProcessor is a function that takes a list of previously generated tokens, the logits tensor for the next token and, optionally, prompt tokens as a first argument, and returns a modified tensor of logits to sample from.""" def get_bad_words_logits_processors( bad_words: list[str], tokenizer: TokenizerLike ) -> list[LogitsProcessor]: bad_words_ids: list[list[int]] = list() for bad_word in bad_words: # To prohibit words both at the beginning # and in the middle of text # (related to add_prefix_space tokenizer parameter) for add_prefix_space in [False, True]: prefix = " " if add_prefix_space else "" prompt = prefix + bad_word.lstrip() prompt_token_ids = tokenizer.encode(text=prompt, add_special_tokens=False) # If no space at the beginning # or if prefix space produces a new word token if (not add_prefix_space) or ( add_prefix_space and prompt_token_ids[0] != bad_words_ids[-1][0] and len(prompt_token_ids) == len(bad_words_ids[-1]) ): bad_words_ids.append(prompt_token_ids) return [NoBadWordsLogitsProcessor(bad_words_ids=bad_words_ids)] class NoBadWordsLogitsProcessor: _SMALLEST_LOGIT = float("-inf") _NEUTRAL_LOGIT = 0.0 def __init__(self, bad_words_ids: list[list[int]]): self.bad_words_ids = bad_words_ids self.word_bias: torch.FloatTensor = None def __call__( self, past_tokens_ids: Sequence[int], logits: torch.FloatTensor, ) -> torch.Tensor: if self.word_bias is None: self._init_word_bias(logits=logits) last_token_bias = torch.zeros_like(logits) for bad_word_ids in self.bad_words_ids: if len(bad_word_ids) == 1: # 1-token words already processed continue if len(bad_word_ids) > len(past_tokens_ids) + 1: continue prefix_length = len(bad_word_ids) - 1 last_token_id = bad_word_ids[-1] actual_prefix = past_tokens_ids[-prefix_length:] expected_prefix = bad_word_ids[:prefix_length] assert len(actual_prefix) == len(expected_prefix) is_match = tuple(actual_prefix) == tuple(expected_prefix) last_token_bias[last_token_id] += ( self._SMALLEST_LOGIT if is_match else self._NEUTRAL_LOGIT ) logits = logits + self.word_bias + last_token_bias return logits def _init_word_bias(self, logits: torch.FloatTensor) -> None: # Code based on NoBadWordsLogitsProcessor and SequenceBiasLogitsProcessor # noqa: E501 # from https://github.com/huggingface/transformers/blob/main/src/transformers/generation/logits_process.py vocab_size = logits.shape[-1] self._check_token_ids_bounds(vocab_size=vocab_size) self.word_bias = torch.zeros( (vocab_size,), dtype=torch.float, device=logits.device ) for bad_word_ids in self.bad_words_ids: if len(bad_word_ids) == 1: bad_word_id = bad_word_ids[-1] self.word_bias[bad_word_id] = self._SMALLEST_LOGIT def _check_token_ids_bounds(self, vocab_size: int) -> None: invalid_token_ids = [] for bad_word_ids in self.bad_words_ids: for token_id in bad_word_ids: if token_id < 0 or token_id >= vocab_size: invalid_token_ids.append(token_id) if len(invalid_token_ids) > 0: raise ValueError( f"The model vocabulary size is {vocab_size}," f" but the following tokens" f" were specified as bad: {invalid_token_ids}." f" All token id values should be integers satisfying:" f" 0 <= token_id < {vocab_size}." )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/pooling_params.py
vllm/pooling_params.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from copy import deepcopy from typing import Annotated, Any, Optional import msgspec from vllm.config import ModelConfig, PoolerConfig from vllm.config.pooler import get_use_activation from vllm.sampling_params import RequestOutputKind from vllm.tasks import PoolingTask class PoolingParams( msgspec.Struct, omit_defaults=True, # type: ignore[call-arg] array_like=True, ): # type: ignore[call-arg] """API parameters for pooling models. Attributes: truncate_prompt_tokens: Controls prompt truncation. Set to -1 to use the model's default truncation size. Set to k to keep only the last k tokens (left truncation). Set to None to disable truncation. dimensions: Reduce the dimensions of embeddings if model support matryoshka representation. normalize: Whether to normalize the embeddings outputs. softmax: softmax will be deprecated, please use use_activation instead. activation: activation will be deprecated, please use use_activation instead. use_activation: Whether to apply activation function to the classification outputs. """ # --8<-- [start:common-pooling-params] truncate_prompt_tokens: Annotated[int, msgspec.Meta(ge=-1)] | None = None # --8<-- [end:common-pooling-params] ## for embeddings models # --8<-- [start:embedding-pooling-params] dimensions: int | None = None normalize: bool | None = None # --8<-- [end:embedding-pooling-params] ## for classification, scoring and rerank # --8<-- [start:classification-pooling-params] softmax: bool | None = None activation: bool | None = None use_activation: bool | None = None # --8<-- [end:classification-pooling-params] ## for step pooling models step_tag_id: int | None = None returned_token_ids: list[int] | None = None ## Internal use only task: PoolingTask | None = None requires_token_ids: bool = False skip_reading_prefix_cache: bool | None = None extra_kwargs: dict[str, Any] | None = None output_kind: RequestOutputKind = RequestOutputKind.FINAL_ONLY @property def all_parameters(self) -> list[str]: return ["dimensions", "normalize", "use_activation"] @property def valid_parameters(self): return { "embed": ["dimensions", "normalize"], "classify": ["use_activation"], "score": ["use_activation"], "token_embed": ["dimensions", "normalize"], "token_classify": ["use_activation"], } def clone(self) -> "PoolingParams": """Returns a deep copy of the PoolingParams instance.""" return deepcopy(self) def verify( self, task: PoolingTask, model_config: Optional["ModelConfig"] = None ) -> None: if self.task is None: self.task = task elif self.task != task: msg = f"You cannot overwrite {self.task=!r} with {task=!r}!" raise ValueError(msg) # raise deprecated warning for softmax and activation self.use_activation = get_use_activation(self) # plugin task uses io_processor.parse_request to verify inputs, # skipping PoolingParams verify if self.task == "plugin": if self.skip_reading_prefix_cache is None: self.skip_reading_prefix_cache = True return # NOTE: Task validation needs to done against the model instance, # which is not available in model config. So, it's not included # in this method self._merge_default_parameters(model_config) self._set_default_parameters(model_config) self._verify_valid_parameters() def _merge_default_parameters( self, model_config: Optional["ModelConfig"] = None ) -> None: if model_config is None: return pooler_config = model_config.pooler_config if pooler_config is None: return assert self.task is not None, "task must be set" valid_parameters = self.valid_parameters[self.task] for k in valid_parameters: if getattr(pooler_config, k, None) is None: continue if getattr(self, k, None) is None: setattr(self, k, getattr(pooler_config, k)) if self.skip_reading_prefix_cache is None: # If prefix caching is enabled, # the output of all pooling may less than n_prompt_tokens, # we need to skip reading cache at this request. if self.task in ["token_embed", "token_classify"]: self.skip_reading_prefix_cache = True else: self.skip_reading_prefix_cache = False self._verify_step_pooling(pooler_config, valid_parameters) def _verify_step_pooling( self, pooler_config: "PoolerConfig", valid_parameters: list[str] ): step_pooling_parameters = ["step_tag_id", "returned_token_ids"] if pooler_config.pooling_type != "STEP": invalid_parameters = [] for k in step_pooling_parameters: if getattr(self, k, None) is not None: invalid_parameters.append(k) if invalid_parameters: raise ValueError( f"Task {self.task} only supports {valid_parameters} " f"parameters, does not support " f"{invalid_parameters} parameters" ) else: for k in step_pooling_parameters: if getattr(pooler_config, k, None) is None: continue if getattr(self, k, None) is None: setattr(self, k, getattr(pooler_config, k)) def _set_default_parameters(self, model_config: Optional["ModelConfig"]): if self.task in ["embed", "token_embed"]: if self.normalize is None: self.normalize = True if self.dimensions is not None and model_config is not None: if not model_config.is_matryoshka: raise ValueError( f'Model "{model_config.served_model_name}" does not ' f"support matryoshka representation, " f"changing output dimensions will lead to poor results." ) mds = model_config.matryoshka_dimensions if mds is not None: if self.dimensions not in mds: raise ValueError( f'Model "{model_config.served_model_name}" ' f"only supports {str(mds)} matryoshka dimensions, " f"use other output dimensions will " f"lead to poor results." ) elif self.dimensions < 1: raise ValueError("Dimensions must be greater than 0") elif self.task in ["classify", "score", "token_classify"]: if self.use_activation is None: self.use_activation = True else: raise ValueError(f"Unknown pooling task: {self.task}") def _verify_valid_parameters(self): assert self.task is not None, "task must be set" valid_parameters = self.valid_parameters[self.task] invalid_parameters = [] for k in self.all_parameters: if k in valid_parameters: continue if getattr(self, k, None) is not None: invalid_parameters.append(k) if invalid_parameters: raise ValueError( f"Task {self.task} only supports {valid_parameters} " f"parameters, does not support " f"{invalid_parameters} parameters" ) def __repr__(self) -> str: return ( f"PoolingParams(" f"task={self.task}, " f"normalize={self.normalize}, " f"dimensions={self.dimensions}, " f"use_activation={self.use_activation}, " f"step_tag_id={self.step_tag_id}, " f"returned_token_ids={self.returned_token_ids}, " f"requires_token_ids={self.requires_token_ids}, " f"skip_reading_prefix_cache={self.skip_reading_prefix_cache}, " f"truncate_prompt_tokens={self.truncate_prompt_tokens}, " f"extra_kwargs={self.extra_kwargs})" ) def __post_init__(self) -> None: assert self.output_kind == RequestOutputKind.FINAL_ONLY, ( "For pooling output_kind has to be FINAL_ONLY" )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/tasks.py
vllm/tasks.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Literal, get_args GenerationTask = Literal["generate", "transcription"] GENERATION_TASKS = get_args(GenerationTask) PoolingTask = Literal[ "embed", "classify", "score", "token_embed", "token_classify", "plugin" ] POOLING_TASKS = get_args(PoolingTask) SupportedTask = Literal[GenerationTask, PoolingTask]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/sequence.py
vllm/sequence.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Sequence and its related classes.""" from dataclasses import dataclass from typing import TYPE_CHECKING, Any import torch if TYPE_CHECKING: from vllm.v1.worker.kv_connector_model_runner_mixin import KVConnectorOutput else: KVConnectorOutput = Any VLLM_TOKEN_ID_ARRAY_TYPE = "l" VLLM_INVALID_TOKEN_ID = -1 @dataclass class RequestMetrics: """Metrics associated with a request. Attributes: arrival_time: The time when the request arrived. first_scheduled_time: The time when the request was first scheduled. first_token_time: The time when the first token was generated. time_in_queue: The time the request spent in the queue. finished_time: The time when the request was finished. scheduler_time: The time spent in the scheduler when this request was being considered by the scheduler. model_forward_time: The time spent in the model forward pass when this request was in the batch. model_execute_time: The time spent in the model execute function. This will include model forward, block/sync across workers, cpu-gpu sync time and sampling time. """ arrival_time: float last_token_time: float first_scheduled_time: float | None first_token_time: float | None time_in_queue: float | None finished_time: float | None = None scheduler_time: float | None = None model_forward_time: float | None = None model_execute_time: float | None = None # cannot use msgspec.Struct here because Dynamo does not support it @dataclass class IntermediateTensors: """For all pipeline stages except the last, we need to return the hidden states and residuals to be sent to the next stage. This data structure contains the hidden states and residuals for a request. Each stage also needs to handle its own kv_connector_output. """ tensors: dict[str, torch.Tensor] kv_connector_output: KVConnectorOutput | None def __init__( self, tensors: dict[str, torch.Tensor], kv_connector_output: KVConnectorOutput | None = None, ) -> None: # manually define this function, so that # Dynamo knows `IntermediateTensors()` comes from this file. # Otherwise, dataclass will generate this function by evaluating # a string, and we will lose the information about the source file. self.tensors = tensors self.kv_connector_output = kv_connector_output def __getitem__(self, key: str | slice): if isinstance(key, str): return self.tensors[key] elif isinstance(key, slice): return self.__class__({k: v[key] for k, v in self.tensors.items()}) def __setitem__(self, key: str, value: torch.Tensor): self.tensors[key] = value def items(self): return self.tensors.items() def __len__(self): return len(self.tensors) def __eq__(self, other: object): if not isinstance(other, self.__class__): return False if self.tensors.keys() != other.tensors.keys(): return False return all(torch.equal(self.tensors[k], other.tensors[k]) for k in self.tensors) def __repr__(self) -> str: return f"IntermediateTensors(tensors={self.tensors})"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/_ipex_ops.py
vllm/_ipex_ops.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.logger import init_logger from vllm.platforms import current_platform logger = init_logger(__name__) try: import intel_extension_for_pytorch as ipex except ImportError as e: logger.debug("Import error msg: %s", e.msg) class ipex_ops: @staticmethod def _reshape_activation_tensor( x: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: num = x.size(0) d = x.size(1) // 2 x = x.reshape(num, 2, d) x1, x2 = torch.chunk(x, chunks=2, dim=1) x1 = x1.reshape(num, d) x2 = x2.reshape(num, d) return x1, x2 @staticmethod def silu_and_mul(out: torch.Tensor, x: torch.Tensor) -> None: ipex.llm.functional.silu_and_mul(x, out) @staticmethod def gelu_and_mul(out: torch.Tensor, x: torch.Tensor) -> None: ipex.llm.functional.gelu_and_mul(x, out) @staticmethod def gelu_tanh_and_mul(out: torch.Tensor, x: torch.Tensor) -> None: ipex.llm.functional.gelu_and_mul(x, out) @staticmethod def gelu_fast(x: torch.Tensor) -> torch.Tensor: return torch.nn.functional.gelu(x) @staticmethod def gelu_new(x: torch.Tensor) -> torch.Tensor: return torch.nn.functional.gelu(x) @staticmethod def gelu_quick(out: torch.Tensor, x: torch.Tensor) -> None: ipex.llm.functional.gelu_quick(x, out) @staticmethod def paged_attention_v1( out: torch.Tensor, query: torch.Tensor, key_cache: torch.Tensor, value_cache: torch.Tensor, num_kv_heads: int, scale: float, block_tables: torch.Tensor, context_lens: torch.Tensor, block_size: int, max_context_len: int, alibi_slopes: torch.Tensor | None, kv_cache_dtype: str, k_scale: float, v_scale: float, tp_rank: int = 0, blocksparse_local_blocks: int = 0, blocksparse_vert_stride: int = 0, blocksparse_block_size: int = 64, blocksparse_head_sliding_step: int = 0, ) -> None: assert kv_cache_dtype == "auto" num_heads = out.size(1) num_queries_per_tokens = num_heads // num_kv_heads ipex.llm.modules.PagedAttention.single_query_kv_attention( out, query.contiguous(), key_cache.view_as(value_cache), value_cache, num_queries_per_tokens, scale, block_tables, context_lens, block_size, max_context_len, alibi_slopes, ) @staticmethod def paged_attention_v2( out: torch.Tensor, exp_sum: torch.Tensor, max_logits: torch.Tensor, tmp_out: torch.Tensor, query: torch.Tensor, key_cache: torch.Tensor, value_cache: torch.Tensor, num_kv_heads: int, scale: float, block_tables: torch.Tensor, context_lens: torch.Tensor, block_size: int, max_context_len: int, alibi_slopes: torch.Tensor | None, kv_cache_dtype: str, k_scale: float, v_scale: float, tp_rank: int = 0, blocksparse_local_blocks: int = 0, blocksparse_vert_stride: int = 0, blocksparse_block_size: int = 64, blocksparse_head_sliding_step: int = 0, ) -> None: assert kv_cache_dtype == "auto" num_heads = out.size(1) num_queries_per_tokens = num_heads // num_kv_heads ipex.llm.modules.PagedAttention.single_query_kv_attention( out, query.contiguous(), key_cache.view_as(value_cache), value_cache, num_queries_per_tokens, scale, block_tables, context_lens, block_size, max_context_len, alibi_slopes, ) @staticmethod def rotary_embedding( positions: torch.Tensor, # [batch_size, seq_len] query: torch.Tensor, # [batch_size, seq_len, num_heads*head_size] key: torch.Tensor, # [batch_size, seq_len, num_kv_heads*head_size] head_size: int, cos_sin_cache: torch.Tensor, # [cos_sin_dim, rot_dim] is_neox: bool, ) -> None: rot_dim = cos_sin_cache.size(1) ipex.llm.functional.rotary_embedding_batched( positions, query, key, head_size, cos_sin_cache, is_neox, rot_dim ) @staticmethod def rms_norm( input: torch.Tensor, weight: torch.Tensor, epsilon: float ) -> torch.Tensor: out = torch.empty_like(input) torch.ops.torch_ipex.rms_norm_vllm(out, input.contiguous(), weight, epsilon) return out @staticmethod def fused_add_rms_norm( input: torch.Tensor, residual: torch.Tensor, weight: torch.Tensor, epsilon: float, ) -> None: torch.ops.torch_ipex.fused_add_rms_norm_vllm(input, residual, weight, epsilon) @staticmethod def varlen_attention( query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, out: torch.Tensor, seqlen_q: torch.Tensor, seqlen_k: torch.Tensor, alibi_slopes: torch.Tensor | None, max_seqlen_q: int, max_seqlen_k: int, pdropout: float, softmax_scale: float, zero_tensors: bool, is_causal: bool, return_softmax: bool, gen_: torch.Generator, window_size_left: float, window_size_right: float, logits_soft_cap: float, ) -> None: if ipex.__version__.endswith("cpu"): if logits_soft_cap != 0.0: raise ValueError("IPEX CPU does not support logits_soft_cap") assert alibi_slopes is None assert window_size_left < 0 and window_size_right < 0 ipex.llm.functional.varlen_attention( query.contiguous(), key.contiguous(), value.contiguous(), out, seqlen_q.int(), seqlen_k.int(), max_seqlen_q, max_seqlen_k, pdropout, softmax_scale, zero_tensors, is_causal, return_softmax, gen_, ) else: # XPU build ipex.llm.functional.varlen_attention( query.contiguous(), key.contiguous(), value.contiguous(), out, seqlen_q.int(), seqlen_k.int(), alibi_slopes, max_seqlen_q, max_seqlen_k, pdropout, softmax_scale, zero_tensors, is_causal, return_softmax, gen_, window_size_left, window_size_right, logits_soft_cap, ) @staticmethod def reshape_and_cache( key: torch.Tensor, value: torch.Tensor, key_cache: torch.Tensor, value_cache: torch.Tensor, slot_mapping: torch.Tensor, kv_cache_dtype: str, k_scale: float, v_scale: float, ) -> None: assert kv_cache_dtype == "auto" ipex.llm.modules.PagedAttention.reshape_and_cache( key, value, key_cache, value_cache, slot_mapping ) @staticmethod def reshape_and_cache_flash( key: torch.Tensor, value: torch.Tensor, key_cache: torch.Tensor, value_cache: torch.Tensor, slot_mapping: torch.Tensor, kv_cache_dtype: str, k_scale: torch.Tensor | None = None, v_scale: torch.Tensor | None = None, k_scale_float: float = 1.0, v_scale_float: float = 1.0, ) -> None: ipex.llm.modules.PagedAttention.reshape_and_cache_flash( key, value, key_cache, value_cache, slot_mapping, kv_cache_dtype, k_scale_float, v_scale_float, ) @staticmethod def flash_attn_varlen_func( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, cu_seqlens_q: torch.Tensor, max_seqlen_q: int, max_seqlen_k: int, softmax_scale: float | None = None, causal: bool = False, out: torch.Tensor | None = None, block_table: torch.Tensor | None = None, alibi_slopes: torch.Tensor | None = None, window_size: list[int] | None = None, softcap: float | None = 0.0, seqused_k: torch.Tensor | None = None, cu_seqlens_k: torch.Tensor | None = None, # passed in qwen vl dropout_p: float = 0.0, # The following parameters are not used in ipex kernel currently, # we keep API compatible to CUDA's. scheduler_metadata=None, fa_version: int = 2, q_descale=None, k_descale=None, v_descale=None, num_splits=0, s_aux: torch.Tensor | None = None, ): if out is None: out = torch.empty(q.shape, dtype=q.dtype, device=q.device) real_window_size: tuple[int, int] if window_size is None: real_window_size = (-1, -1) else: assert len(window_size) == 2 real_window_size = (window_size[0], window_size[1]) if block_table is None: assert cu_seqlens_k is not None, ( "cu_seqlens_k can't be None when calling varlen_attention." ) if softmax_scale is None: softmax_scale = q.shape[-1] ** (-0.5) ipex_ops.varlen_attention( q.contiguous(), k.contiguous(), v.contiguous(), out, cu_seqlens_q, cu_seqlens_k, None, max_seqlen_q, max_seqlen_k, 0.0, softmax_scale, False, causal, False, None, real_window_size[0], real_window_size[1], -1, ) return out else: return ipex.llm.modules.PagedAttention.flash_attn_varlen_func( out, q.contiguous(), k, v, cu_seqlens_q, seqused_k, max_seqlen_q, max_seqlen_k, softmax_scale, causal, block_table, alibi_slopes, sink=s_aux, softcap=softcap, window_size_left=real_window_size[0], window_size_right=real_window_size[1], k_scale=1.0, v_scale=1.0, ) @staticmethod def get_scheduler_metadata( batch_size, max_seqlen_q, max_seqlen_k, num_heads_q, num_heads_kv, headdim, cache_seqlens: torch.Tensor, qkv_dtype=torch.bfloat16, headdim_v=None, cu_seqlens_q: torch.Tensor | None = None, cu_seqlens_k_new: torch.Tensor | None = None, cache_leftpad: torch.Tensor | None = None, page_size: int | None = None, max_seqlen_k_new=0, causal=False, window_size=(-1, -1), # -1 means infinite context window has_softcap=False, num_splits=0, # Can be tuned for speed pack_gqa=None, # Can be tuned for speed sm_margin=0, # Can be tuned if some SMs are used for communication ) -> None: logger.warning_once( "get_scheduler_metadata is not implemented for ipex_ops, returning None." ) return None @staticmethod def swap_blocks( src: torch.Tensor, dst: torch.Tensor, block_mapping: torch.Tensor ) -> None: torch.xpu.swap_blocks(src, dst, block_mapping) # type: ignore @staticmethod def scaled_fp8_quant( input: torch.Tensor, scale: torch.Tensor | None = None, num_token_padding: int | None = None, scale_ub: torch.Tensor | None = None, use_per_token_if_dynamic: bool = False, output: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: """ Quantize input tensor to FP8 and return quantized tensor and scale. This function is designed for both static and dynamic quantization: If you provide the scale, it will use static scaling and if you omit it, the scale will be determined dynamically. Currently, XPU platform only supports dynamic quantization. The function also allows optional padding of the output tensors for downstream kernels that will benefit from padding. Args: input: The input tensor to be quantized to FP8 scale: Optional scaling factor for the FP8 quantization scale_ub: Optional upper bound for scaling factor in dynamic per token case num_token_padding: If specified, pad the first dimension of the output to at least this value. use_per_token_if_dynamic: Whether to do per_tensor or per_token in the dynamic quantization case. Returns: tuple[torch.Tensor, torch.Tensor]: The output tensor in FP8 and scaling factor. """ # This code assumes batch_dim and num_tokens are flattened assert input.ndim == 2 shape: tuple[int, int] | torch.Size = input.shape out_dtype: torch.dtype = current_platform.fp8_dtype() if num_token_padding: shape = (max(num_token_padding, input.shape[0]), shape[1]) if output is None: output = torch.empty(shape, device=input.device, dtype=out_dtype) else: assert num_token_padding is None, ( "padding not supported if output passed in" ) assert output.dtype == out_dtype assert scale is None, "only dynamic fp8 quantization supported on XPU" assert not use_per_token_if_dynamic, ( "per token dynamic fp8 quantization not supported on XPU" ) scale = torch.zeros(1, device=input.device, dtype=torch.float32) torch.ops.torch_ipex.dynamic_scaled_fp8_quant(output, input, scale) return output, scale
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/forward_context.py
vllm/forward_context.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import time from collections import defaultdict from contextlib import contextmanager from dataclasses import dataclass from typing import Any, NamedTuple import torch import vllm.envs as envs from vllm.attention.backends.abstract import AttentionMetadata from vllm.config import CUDAGraphMode, ParallelConfig, VllmConfig from vllm.logger import init_logger from vllm.v1.worker.dp_utils import coordinate_batch_across_dp from vllm.v1.worker.ubatch_utils import UBatchSlices logger = init_logger(__name__) track_batchsize: bool = envs.VLLM_LOG_BATCHSIZE_INTERVAL >= 0 last_logging_time: float = 0 forward_start_time: float = 0 batchsize_logging_interval: float = envs.VLLM_LOG_BATCHSIZE_INTERVAL batchsize_forward_time: defaultdict = defaultdict(list) class BatchDescriptor(NamedTuple): """ Batch descriptor for cudagraph dispatching. We should keep the num of items as minimal as possible to properly and uniquely describe the padded batch for cudagraph. """ num_tokens: int num_reqs: int | None = None """ Number of requests in the batch. Can be None for PIECEWISE cudagraphs where the cudagraphs can handle any number of requests. """ uniform: bool = False """ True if all the requests in the batch have the same number of tokens. """ has_lora: bool = False """ Whether this batch has active LoRA adapters. """ def relax_for_mixed_batch_cudagraphs(self) -> "BatchDescriptor": """ Return a relaxed version of current batch descriptor that is still compatible with PIECEWISE cudagraphs (or mixed prefill-decode FA cudagraphs). """ return BatchDescriptor( self.num_tokens, num_reqs=None, uniform=False, has_lora=self.has_lora ) def _compute_sp_num_tokens( num_tokens_across_dp_cpu: torch.Tensor, sequence_parallel_size: int ) -> list[int]: sp_tokens = ( num_tokens_across_dp_cpu + sequence_parallel_size - 1 ) // sequence_parallel_size sp_tokens = sp_tokens.repeat_interleave(sequence_parallel_size) return sp_tokens.tolist() def _compute_chunked_local_num_tokens( num_tokens_across_dp_cpu: torch.Tensor, sequence_parallel_size: int, max_num_tokens: int, chunk_idx: int, ) -> list[int]: sp_tokens = _compute_sp_num_tokens(num_tokens_across_dp_cpu, sequence_parallel_size) sp_size = len(sp_tokens) local_size = [-1] * sp_size for i in range(sp_size): # Take into account sharding if MoE activation is sequence parallel. local_size[i] = min(max_num_tokens, sp_tokens[i] - (max_num_tokens * chunk_idx)) if local_size[i] <= 0: local_size[i] = 1 # ensure lockstep even if done return local_size @dataclass class DPMetadata: max_tokens_across_dp_cpu: torch.Tensor num_tokens_across_dp_cpu: torch.Tensor # NOTE: local_sizes should only be set by the chunked_sizes context manager local_sizes: list[int] | None = None @staticmethod def make( parallel_config: ParallelConfig, num_tokens: int, num_tokens_across_dp_cpu: torch.Tensor, ) -> "DPMetadata": assert num_tokens_across_dp_cpu is not None assert parallel_config.data_parallel_size > 1 assert parallel_config.is_moe_model is not False dp_rank = parallel_config.data_parallel_rank batchsize = num_tokens # If num_tokens_across_dp is None, it will be computed by all_reduce # Otherwise, num_tokens_across_dp[dp_rank] should be equal to batchsize assert num_tokens_across_dp_cpu[dp_rank] == batchsize, ( f"{num_tokens_across_dp_cpu[dp_rank]} {batchsize}" ) max_tokens_across_dp_cpu = torch.max(num_tokens_across_dp_cpu) return DPMetadata(max_tokens_across_dp_cpu, num_tokens_across_dp_cpu) @contextmanager def chunked_sizes( self, sequence_parallel_size: int, max_chunk_size_per_rank: int, chunk_idx: int ): """ Context manager to compute and temporarily set the per-rank local token sizes for a specific chunk during chunked forward execution. This is necessary to ensure each DP (data parallel) rank processes its designated portion of tokens in lockstep with others, even when the token counts are uneven or some ranks have completed their input early. For chunked execution, we break up the total tokens on each rank into multiple chunks (of at most `max_chunk_size_per_rank`), and for a given `chunk_idx`, this context manager sets `self.local_sizes` to the number of tokens to process in that chunk on each rank. `self.local_sizes` is only valid inside the context. Args: sequence_parallel_size: When Attn is TP and MoE layers are EP, we use SP between the layers to avoid redundant ops. We need this value to compute the chunked sizes. max_chunk_size_per_rank: The max number of tokens each rank is allowed to process in this chunk. chunk_idx: The index of the chunk to compute sizes for. """ self.local_sizes = _compute_chunked_local_num_tokens( self.num_tokens_across_dp_cpu, sequence_parallel_size, max_chunk_size_per_rank, chunk_idx, ) try: yield self.local_sizes finally: self.local_sizes = None @contextmanager def sp_local_sizes(self, sequence_parallel_size: int): """ Context manager for setting self.local_sizes. Same as self.chunked_sizes but without any chunking. """ self.local_sizes = _compute_sp_num_tokens( self.num_tokens_across_dp_cpu, sequence_parallel_size ) try: yield self.local_sizes finally: self.local_sizes = None def get_chunk_sizes_across_dp_rank(self) -> list[int] | None: assert self.local_sizes is not None return self.local_sizes # Get the cumulative tokens across sequence parallel ranks. # In this case the input to the MoEs will be distributed w.r.t both # DP and TP rank. # When sp_size==1, this is just the cummulative num tokens across DP. def cu_tokens_across_sp(self, sp_size: int) -> torch.Tensor: num_tokens_across_sp_cpu = ( self.num_tokens_across_dp_cpu - 1 + sp_size ) // sp_size num_tokens_across_sp_cpu = num_tokens_across_sp_cpu.repeat_interleave(sp_size) return torch.cumsum(num_tokens_across_sp_cpu, dim=0) @dataclass class ForwardContext: # copy from vllm_config.compilation_config.static_forward_context no_compile_layers: dict[str, Any] attn_metadata: dict[str, AttentionMetadata] | list[dict[str, AttentionMetadata]] """ Type Dict[str, AttentionMetadata] for v1, map from layer_name of each attention layer to its attention metadata Type List[Dict[str, AttentionMetadata]] for DBO. List of size two, one for each microbatch. Set dynamically for each forward pass """ # TODO: remove after making all virtual_engines share the same kv cache virtual_engine: int # set dynamically for each forward pass # set dynamically for each forward pass dp_metadata: DPMetadata | None = None # determine the cudagraph style at runtime to be FULL, PIECEWISE, or NONE. # by default NONE, no cudagraph is used. cudagraph_runtime_mode: CUDAGraphMode = CUDAGraphMode.NONE batch_descriptor: BatchDescriptor | None = None ubatch_slices: UBatchSlices | None = None def __post_init__(self): assert self.cudagraph_runtime_mode.valid_runtime_modes(), ( f"Invalid cudagraph runtime mode: {self.cudagraph_runtime_mode}" ) _forward_context: ForwardContext | None = None def get_forward_context() -> ForwardContext: """Get the current forward context.""" assert _forward_context is not None, ( "Forward context is not set. " "Please use `set_forward_context` to set the forward context." ) return _forward_context def is_forward_context_available() -> bool: return _forward_context is not None def create_forward_context( attn_metadata: Any, vllm_config: VllmConfig, virtual_engine: int = 0, dp_metadata: DPMetadata | None = None, cudagraph_runtime_mode: CUDAGraphMode = CUDAGraphMode.NONE, batch_descriptor: BatchDescriptor | None = None, ubatch_slices: UBatchSlices | None = None, ): return ForwardContext( no_compile_layers=vllm_config.compilation_config.static_forward_context, virtual_engine=virtual_engine, attn_metadata=attn_metadata, dp_metadata=dp_metadata, cudagraph_runtime_mode=cudagraph_runtime_mode, batch_descriptor=batch_descriptor, ubatch_slices=ubatch_slices, ) @contextmanager def override_forward_context(forward_context: ForwardContext | None): """A context manager that overrides the current forward context. This is used to override the forward context for a specific forward pass. """ global _forward_context prev_context = _forward_context _forward_context = forward_context try: yield finally: _forward_context = prev_context @contextmanager def set_forward_context( attn_metadata: Any, vllm_config: VllmConfig, virtual_engine: int = 0, num_tokens: int | None = None, num_tokens_across_dp: torch.Tensor | None = None, cudagraph_runtime_mode: CUDAGraphMode = CUDAGraphMode.NONE, batch_descriptor: BatchDescriptor | None = None, ubatch_slices: UBatchSlices | None = None, ): """A context manager that stores the current forward context, can be attention metadata, etc. Here we can inject common logic for every model forward pass. """ global forward_start_time need_to_track_batchsize = track_batchsize and attn_metadata is not None if need_to_track_batchsize: forward_start_time = time.perf_counter() dp_metadata: DPMetadata | None = None if vllm_config.parallel_config.data_parallel_size > 1 and ( attn_metadata is not None or num_tokens is not None ): # If num_tokens_across_dp hasn't already been initialized, then # initialize it here. Both DP padding and Microbatching will be # disabled. if num_tokens_across_dp is None: assert ubatch_slices is None assert num_tokens is not None _, num_tokens_across_dp, _ = coordinate_batch_across_dp( num_tokens_unpadded=num_tokens, parallel_config=vllm_config.parallel_config, allow_microbatching=False, allow_dp_padding=False, ) assert num_tokens_across_dp is not None dp_metadata = DPMetadata.make( vllm_config.parallel_config, num_tokens or 0, num_tokens_across_dp ) # Convenience: if cudagraph is used and num_tokens is given, we can just # create a batch descriptor here if not given (there's no harm since if it # doesn't match in the wrapper it'll fall through). if cudagraph_runtime_mode != CUDAGraphMode.NONE and num_tokens is not None: batch_descriptor = batch_descriptor or BatchDescriptor(num_tokens=num_tokens) forward_context = create_forward_context( attn_metadata, vllm_config, virtual_engine, dp_metadata, cudagraph_runtime_mode, batch_descriptor, ubatch_slices, ) try: with override_forward_context(forward_context): yield finally: global last_logging_time, batchsize_logging_interval if need_to_track_batchsize: batchsize = num_tokens # we use synchronous scheduling right now, # adding a sync point here should not affect # scheduling of the next batch from vllm.platforms import current_platform synchronize = current_platform.synchronize if synchronize is not None: synchronize() now = time.perf_counter() # time measurement is in milliseconds batchsize_forward_time[batchsize].append((now - forward_start_time) * 1000) if now - last_logging_time > batchsize_logging_interval: last_logging_time = now forward_stats = [] for bs, times in batchsize_forward_time.items(): if len(times) <= 1: # can be cudagraph / profiling run continue medium = torch.quantile(torch.tensor(times), q=0.5).item() medium = round(medium, 2) forward_stats.append((bs, len(times), medium)) forward_stats.sort(key=lambda x: x[1], reverse=True) if forward_stats: logger.info( ( "Batchsize forward time stats " "(batchsize, count, median_time(ms)): %s" ), forward_stats, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/scripts.py
vllm/scripts.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from vllm.entrypoints.cli.main import main as vllm_main from vllm.logger import init_logger logger = init_logger(__name__) # Backwards compatibility for the move from vllm.scripts to # vllm.entrypoints.cli.main def main(): logger.warning( "vllm.scripts.main() is deprecated. Please re-install " "vllm or use vllm.entrypoints.cli.main.main() instead." ) vllm_main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/_bc_linter.py
vllm/_bc_linter.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # vllm/_bc_linter.py from collections.abc import Callable from typing import Any, TypeVar, overload T = TypeVar("T") @overload def bc_linter_skip(obj: T) -> T: ... @overload def bc_linter_skip(*, reason: str | None = ...) -> Callable[[T], T]: ... def bc_linter_skip(obj: Any = None, *, reason: str | None = None): """ No-op decorator to mark symbols/files for BC-linter suppression. Usage: @bc_linter_skip def legacy_api(...): ... """ def _wrap(x: T) -> T: return x return _wrap if obj is None else obj @overload def bc_linter_include(obj: T) -> T: ... @overload def bc_linter_include(*, reason: str | None = ...) -> Callable[[T], T]: ... def bc_linter_include(obj: Any = None, *, reason: str | None = None): """ Usage: @bc_linter_include def public_api(...): ... """ def _wrap(x: T) -> T: return x return _wrap if obj is None else obj __all__ = ["bc_linter_skip", "bc_linter_include"]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/tracing.py
vllm/tracing.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os from collections.abc import Mapping from vllm.logger import init_logger from vllm.utils.func_utils import run_once TRACE_HEADERS = ["traceparent", "tracestate"] logger = init_logger(__name__) _is_otel_imported = False otel_import_error_traceback: str | None = None try: from opentelemetry.context.context import Context from opentelemetry.sdk.environment_variables import ( OTEL_EXPORTER_OTLP_TRACES_PROTOCOL, ) from opentelemetry.sdk.trace import TracerProvider from opentelemetry.sdk.trace.export import BatchSpanProcessor from opentelemetry.trace import SpanKind, Tracer, set_tracer_provider from opentelemetry.trace.propagation.tracecontext import ( TraceContextTextMapPropagator, ) _is_otel_imported = True except ImportError: # Capture and format traceback to provide detailed context for the import # error. Only the string representation of the error is retained to avoid # memory leaks. # See https://github.com/vllm-project/vllm/pull/7266#discussion_r1707395458 import traceback otel_import_error_traceback = traceback.format_exc() class Context: # type: ignore pass class BaseSpanAttributes: # type: ignore pass class SpanKind: # type: ignore pass class Tracer: # type: ignore pass def is_otel_available() -> bool: return _is_otel_imported def init_tracer( instrumenting_module_name: str, otlp_traces_endpoint: str ) -> Tracer | None: if not is_otel_available(): raise ValueError( "OpenTelemetry is not available. Unable to initialize " "a tracer. Ensure OpenTelemetry packages are installed. " f"Original error:\n{otel_import_error_traceback}" ) trace_provider = TracerProvider() span_exporter = get_span_exporter(otlp_traces_endpoint) trace_provider.add_span_processor(BatchSpanProcessor(span_exporter)) set_tracer_provider(trace_provider) tracer = trace_provider.get_tracer(instrumenting_module_name) return tracer def get_span_exporter(endpoint): protocol = os.environ.get(OTEL_EXPORTER_OTLP_TRACES_PROTOCOL, "grpc") if protocol == "grpc": from opentelemetry.exporter.otlp.proto.grpc.trace_exporter import ( OTLPSpanExporter, ) elif protocol == "http/protobuf": from opentelemetry.exporter.otlp.proto.http.trace_exporter import ( OTLPSpanExporter, # type: ignore ) else: raise ValueError(f"Unsupported OTLP protocol '{protocol}' is configured") return OTLPSpanExporter(endpoint=endpoint) def extract_trace_context(headers: Mapping[str, str] | None) -> Context | None: if is_otel_available(): headers = headers or {} return TraceContextTextMapPropagator().extract(headers) else: return None def extract_trace_headers(headers: Mapping[str, str]) -> Mapping[str, str]: return {h: headers[h] for h in TRACE_HEADERS if h in headers} class SpanAttributes: # Attribute names copied from here to avoid version conflicts: # https://github.com/open-telemetry/semantic-conventions/blob/main/docs/gen-ai/gen-ai-spans.md GEN_AI_USAGE_COMPLETION_TOKENS = "gen_ai.usage.completion_tokens" GEN_AI_USAGE_PROMPT_TOKENS = "gen_ai.usage.prompt_tokens" GEN_AI_REQUEST_MAX_TOKENS = "gen_ai.request.max_tokens" GEN_AI_REQUEST_TOP_P = "gen_ai.request.top_p" GEN_AI_REQUEST_TEMPERATURE = "gen_ai.request.temperature" GEN_AI_RESPONSE_MODEL = "gen_ai.response.model" # Attribute names added until they are added to the semantic conventions: GEN_AI_REQUEST_ID = "gen_ai.request.id" GEN_AI_REQUEST_N = "gen_ai.request.n" GEN_AI_USAGE_NUM_SEQUENCES = "gen_ai.usage.num_sequences" GEN_AI_LATENCY_TIME_IN_QUEUE = "gen_ai.latency.time_in_queue" GEN_AI_LATENCY_TIME_TO_FIRST_TOKEN = "gen_ai.latency.time_to_first_token" GEN_AI_LATENCY_E2E = "gen_ai.latency.e2e" GEN_AI_LATENCY_TIME_IN_SCHEDULER = "gen_ai.latency.time_in_scheduler" # Time taken in the forward pass for this across all workers GEN_AI_LATENCY_TIME_IN_MODEL_FORWARD = "gen_ai.latency.time_in_model_forward" # Time taken in the model execute function. This will include model # forward, block/sync across workers, cpu-gpu sync time and sampling time. GEN_AI_LATENCY_TIME_IN_MODEL_EXECUTE = "gen_ai.latency.time_in_model_execute" GEN_AI_LATENCY_TIME_IN_MODEL_PREFILL = "gen_ai.latency.time_in_model_prefill" GEN_AI_LATENCY_TIME_IN_MODEL_DECODE = "gen_ai.latency.time_in_model_decode" GEN_AI_LATENCY_TIME_IN_MODEL_INFERENCE = "gen_ai.latency.time_in_model_inference" def contains_trace_headers(headers: Mapping[str, str]) -> bool: return any(h in headers for h in TRACE_HEADERS) @run_once def log_tracing_disabled_warning() -> None: logger.warning("Received a request with trace context but tracing is disabled")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/version.py
vllm/version.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project try: from ._version import __version__, __version_tuple__ except Exception as e: import warnings warnings.warn(f"Failed to read commit hash:\n{e}", RuntimeWarning, stacklevel=2) __version__ = "dev" __version_tuple__ = (0, 0, __version__) def _prev_minor_version_was(version_str): """Check whether a given version matches the previous minor version. Return True if version_str matches the previous minor version. For example - return True if the current version if 0.7.4 and the supplied version_str is '0.6'. Used for --show-hidden-metrics-for-version. """ # Match anything if this is a dev tree if __version_tuple__[0:2] == (0, 0): return True # Note - this won't do the right thing when we release 1.0! assert __version_tuple__[0] == 0 assert isinstance(__version_tuple__[1], int) return version_str == f"{__version_tuple__[0]}.{__version_tuple__[1] - 1}" def _prev_minor_version(): """For the purpose of testing, return a previous minor version number.""" # In dev tree, this will return "0.-1", but that will work fine" assert isinstance(__version_tuple__[1], int) return f"{__version_tuple__[0]}.{__version_tuple__[1] - 1}"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/logger.py
vllm/logger.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Logging configuration for vLLM.""" import datetime import json import logging import os import sys from collections.abc import Generator, Hashable from contextlib import contextmanager from functools import lru_cache, partial from logging import Logger from logging.config import dictConfig from os import path from types import MethodType from typing import Any, Literal, cast import vllm.envs as envs _FORMAT = ( f"{envs.VLLM_LOGGING_PREFIX}%(levelname)s %(asctime)s " "[%(fileinfo)s:%(lineno)d] %(message)s" ) _DATE_FORMAT = "%m-%d %H:%M:%S" def _use_color() -> bool: if envs.NO_COLOR or envs.VLLM_LOGGING_COLOR == "0": return False if envs.VLLM_LOGGING_COLOR == "1": return True if envs.VLLM_LOGGING_STREAM == "ext://sys.stdout": # stdout return hasattr(sys.stdout, "isatty") and sys.stdout.isatty() elif envs.VLLM_LOGGING_STREAM == "ext://sys.stderr": # stderr return hasattr(sys.stderr, "isatty") and sys.stderr.isatty() return False DEFAULT_LOGGING_CONFIG = { "formatters": { "vllm": { "class": "vllm.logging_utils.NewLineFormatter", "datefmt": _DATE_FORMAT, "format": _FORMAT, }, "vllm_color": { "class": "vllm.logging_utils.ColoredFormatter", "datefmt": _DATE_FORMAT, "format": _FORMAT, }, }, "handlers": { "vllm": { "class": "logging.StreamHandler", # Choose formatter based on color setting. "formatter": "vllm_color" if _use_color() else "vllm", "level": envs.VLLM_LOGGING_LEVEL, "stream": envs.VLLM_LOGGING_STREAM, }, }, "loggers": { "vllm": { "handlers": ["vllm"], "level": envs.VLLM_LOGGING_LEVEL, "propagate": False, }, }, "version": 1, "disable_existing_loggers": False, } @lru_cache def _print_debug_once(logger: Logger, msg: str, *args: Hashable) -> None: # Set the stacklevel to 3 to print the original caller's line info logger.debug(msg, *args, stacklevel=3) @lru_cache def _print_info_once(logger: Logger, msg: str, *args: Hashable) -> None: # Set the stacklevel to 3 to print the original caller's line info logger.info(msg, *args, stacklevel=3) @lru_cache def _print_warning_once(logger: Logger, msg: str, *args: Hashable) -> None: # Set the stacklevel to 3 to print the original caller's line info logger.warning(msg, *args, stacklevel=3) LogScope = Literal["process", "global", "local"] def _should_log_with_scope(scope: LogScope) -> bool: """Decide whether to log based on scope""" if scope == "global": from vllm.distributed.parallel_state import is_global_first_rank return is_global_first_rank() if scope == "local": from vllm.distributed.parallel_state import is_local_first_rank return is_local_first_rank() # default "process" scope: always log return True class _VllmLogger(Logger): """ Note: This class is just to provide type information. We actually patch the methods directly on the [`logging.Logger`][] instance to avoid conflicting with other libraries such as `intel_extension_for_pytorch.utils._logger`. """ def debug_once( self, msg: str, *args: Hashable, scope: LogScope = "process" ) -> None: """ As [`debug`][logging.Logger.debug], but subsequent calls with the same message are silently dropped. """ if not _should_log_with_scope(scope): return _print_debug_once(self, msg, *args) def info_once(self, msg: str, *args: Hashable, scope: LogScope = "process") -> None: """ As [`info`][logging.Logger.info], but subsequent calls with the same message are silently dropped. """ if not _should_log_with_scope(scope): return _print_info_once(self, msg, *args) def warning_once( self, msg: str, *args: Hashable, scope: LogScope = "process" ) -> None: """ As [`warning`][logging.Logger.warning], but subsequent calls with the same message are silently dropped. """ if not _should_log_with_scope(scope): return _print_warning_once(self, msg, *args) # Pre-defined methods mapping to avoid repeated dictionary creation _METHODS_TO_PATCH = { "debug_once": _VllmLogger.debug_once, "info_once": _VllmLogger.info_once, "warning_once": _VllmLogger.warning_once, } def _configure_vllm_root_logger() -> None: logging_config = dict[str, Any]() if not envs.VLLM_CONFIGURE_LOGGING and envs.VLLM_LOGGING_CONFIG_PATH: raise RuntimeError( "VLLM_CONFIGURE_LOGGING evaluated to false, but " "VLLM_LOGGING_CONFIG_PATH was given. VLLM_LOGGING_CONFIG_PATH " "implies VLLM_CONFIGURE_LOGGING. Please enable " "VLLM_CONFIGURE_LOGGING or unset VLLM_LOGGING_CONFIG_PATH." ) if envs.VLLM_CONFIGURE_LOGGING: logging_config = DEFAULT_LOGGING_CONFIG vllm_handler = logging_config["handlers"]["vllm"] # Refresh these values in case env vars have changed. vllm_handler["level"] = envs.VLLM_LOGGING_LEVEL vllm_handler["stream"] = envs.VLLM_LOGGING_STREAM vllm_handler["formatter"] = "vllm_color" if _use_color() else "vllm" vllm_loggers = logging_config["loggers"]["vllm"] vllm_loggers["level"] = envs.VLLM_LOGGING_LEVEL if envs.VLLM_LOGGING_CONFIG_PATH: if not path.exists(envs.VLLM_LOGGING_CONFIG_PATH): raise RuntimeError( "Could not load logging config. File does not exist: %s", envs.VLLM_LOGGING_CONFIG_PATH, ) with open(envs.VLLM_LOGGING_CONFIG_PATH, encoding="utf-8") as file: custom_config = json.loads(file.read()) if not isinstance(custom_config, dict): raise ValueError( "Invalid logging config. Expected dict, got %s.", type(custom_config).__name__, ) logging_config = custom_config for formatter in logging_config.get("formatters", {}).values(): # This provides backwards compatibility after #10134. if formatter.get("class") == "vllm.logging.NewLineFormatter": formatter["class"] = "vllm.logging_utils.NewLineFormatter" if logging_config: dictConfig(logging_config) def init_logger(name: str) -> _VllmLogger: """The main purpose of this function is to ensure that loggers are retrieved in such a way that we can be sure the root vllm logger has already been configured.""" logger = logging.getLogger(name) for method_name, method in _METHODS_TO_PATCH.items(): setattr(logger, method_name, MethodType(method, logger)) return cast(_VllmLogger, logger) @contextmanager def suppress_logging(level: int = logging.INFO) -> Generator[None, Any, None]: current_level = logging.root.manager.disable logging.disable(level) yield logging.disable(current_level) # The root logger is initialized when the module is imported. # This is thread-safe as the module is only imported once, # guaranteed by the Python GIL. _configure_vllm_root_logger() # Transformers uses httpx to access the Hugging Face Hub. httpx is quite verbose, # so we set its logging level to WARNING when vLLM's logging level is INFO. if envs.VLLM_LOGGING_LEVEL == "INFO": logging.getLogger("httpx").setLevel(logging.WARNING) logger = init_logger(__name__) def _trace_calls(log_path, root_dir, frame, event, arg=None): if event in ["call", "return"]: # Extract the filename, line number, function name, and the code object filename = frame.f_code.co_filename lineno = frame.f_lineno func_name = frame.f_code.co_name if not filename.startswith(root_dir): # only log the functions in the vllm root_dir return # Log every function call or return try: last_frame = frame.f_back if last_frame is not None: last_filename = last_frame.f_code.co_filename last_lineno = last_frame.f_lineno last_func_name = last_frame.f_code.co_name else: # initial frame last_filename = "" last_lineno = 0 last_func_name = "" with open(log_path, "a") as f: ts = datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S.%f") if event == "call": f.write( f"{ts} Call to" f" {func_name} in {filename}:{lineno}" f" from {last_func_name} in {last_filename}:" f"{last_lineno}\n" ) else: f.write( f"{ts} Return from" f" {func_name} in {filename}:{lineno}" f" to {last_func_name} in {last_filename}:" f"{last_lineno}\n" ) except NameError: # modules are deleted during shutdown pass return partial(_trace_calls, log_path, root_dir) def enable_trace_function_call(log_file_path: str, root_dir: str | None = None): """ Enable tracing of every function call in code under `root_dir`. This is useful for debugging hangs or crashes. `log_file_path` is the path to the log file. `root_dir` is the root directory of the code to trace. If None, it is the vllm root directory. Note that this call is thread-level, any threads calling this function will have the trace enabled. Other threads will not be affected. """ logger.warning( "VLLM_TRACE_FUNCTION is enabled. It will record every" " function executed by Python. This will slow down the code. It " "is suggested to be used for debugging hang or crashes only." ) logger.info("Trace frame log is saved to %s", log_file_path) if root_dir is None: # by default, this is the vllm root directory root_dir = os.path.dirname(os.path.dirname(__file__)) sys.settrace(partial(_trace_calls, log_file_path, root_dir))
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/_aiter_ops.py
vllm/_aiter_ops.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools from collections.abc import Callable import torch from torch._ops import OpOverload import vllm.envs as envs from vllm.platforms import current_platform from vllm.utils.torch_utils import direct_register_custom_op, is_torch_equal_or_newer _FP8_DTYPE = current_platform.fp8_dtype() def is_aiter_found() -> bool: from importlib.util import find_spec return find_spec("aiter") is not None # `find_spec` is not torch.compile compatible. # In cases where aiter availability might have # been checked in forward passes that are torch compiled. # we keep this global outside to not cause torch compile breaks. IS_AITER_FOUND = is_aiter_found() def is_aiter_found_and_supported() -> bool: if current_platform.is_rocm() and IS_AITER_FOUND: from vllm.platforms.rocm import on_gfx9 return on_gfx9() return False def if_aiter_supported(func: Callable) -> Callable: """Decorator that only executes the function if ROCm AITER package is supported on gfx9 archs. """ @functools.wraps(func) def wrapper(*args, **kwargs): # checks the platform, device arch and aiter library existence. if is_aiter_found_and_supported(): return func(*args, **kwargs) return None return wrapper # Can't use dtypes.fp8 directly inside an op # because it returns wrong result on gfx942. # This is a workaround to get the correct FP8 dtype. # This might because that the get_gfx() is wrapped as a custom op. if is_aiter_found_and_supported(): from aiter import dtypes AITER_FP8_DTYPE = dtypes.fp8 def _rocm_aiter_fused_moe_impl( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weight: torch.Tensor, topk_ids: torch.Tensor, expert_mask: torch.Tensor | None = None, activation_method: int = 0, quant_method: int = 0, doweight_stage1: bool = False, w1_scale: torch.Tensor | None = None, w2_scale: torch.Tensor | None = None, a1_scale: torch.Tensor | None = None, a2_scale: torch.Tensor | None = None, ) -> torch.Tensor: from aiter import ActivationType, QuantType from aiter.fused_moe import fused_moe activation = ActivationType(activation_method) quant_type = QuantType(quant_method) return fused_moe( hidden_states, w1, w2, topk_weight, topk_ids, expert_mask, activation, quant_type, doweight_stage1, w1_scale, w2_scale, a1_scale, a2_scale, ) def _rocm_aiter_fused_moe_fake( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weight: torch.Tensor, topk_ids: torch.Tensor, expert_mask: torch.Tensor | None = None, activation_method: int = 0, quant_method: int = 0, doweight_stage1: bool = False, w1_scale: torch.Tensor | None = None, w2_scale: torch.Tensor | None = None, a1_scale: torch.Tensor | None = None, a2_scale: torch.Tensor | None = None, ) -> torch.Tensor: return torch.empty_like(hidden_states) def _rocm_aiter_asm_moe_tkw1_impl( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, fc1_scale: torch.Tensor | None = None, fc2_scale: torch.Tensor | None = None, fc1_smooth_scale: torch.Tensor | None = None, fc2_smooth_scale: torch.Tensor | None = None, a16: bool = False, per_tensor_quant_scale: torch.Tensor | None = None, expert_mask: torch.Tensor | None = None, activation_method: int = 0, ) -> torch.Tensor: from aiter import ActivationType from aiter.fused_moe_bf16_asm import asm_moe_tkw1 activation = ActivationType(activation_method) return asm_moe_tkw1( hidden_states, w1, w2, topk_weights, topk_ids, fc1_scale=fc1_scale, fc2_scale=fc2_scale, fc1_smooth_scale=fc1_smooth_scale, fc2_smooth_scale=fc2_smooth_scale, a16=a16, per_tensor_quant_scale=per_tensor_quant_scale, expert_mask=expert_mask, activation=activation, ) def _rocm_aiter_asm_moe_tkw1_fake( hidden_states: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, fc1_scale: torch.Tensor | None = None, fc2_scale: torch.Tensor | None = None, fc1_smooth_scale: torch.Tensor | None = None, fc2_smooth_scale: torch.Tensor | None = None, a16: bool = False, per_tensor_quant_scale: torch.Tensor | None = None, expert_mask: torch.Tensor | None = None, activation_method: int = 0, ) -> torch.Tensor: return torch.empty_like(hidden_states) def _rocm_aiter_topk_softmax_impl( topk_weights: torch.Tensor, topk_indices: torch.Tensor, token_expert_indices: torch.Tensor, gating_output: torch.Tensor, renormalize: bool, ) -> None: from aiter import topk_softmax topk_softmax( topk_weights, topk_indices, token_expert_indices, gating_output, renormalize ) def _rocm_aiter_topk_softmax_fake( topk_weights: torch.Tensor, topk_indices: torch.Tensor, token_expert_indices: torch.Tensor, gating_output: torch.Tensor, renormalize: bool, ) -> None: pass def _rocm_aiter_biased_grouped_topk_impl( gating_output: torch.Tensor, correction_bias: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_expert_group: int, topk_group: int, need_renorm: bool, routed_scaling_factor: float = 1.0, # mul to topk_weights ) -> None: from aiter import biased_grouped_topk biased_grouped_topk( gating_output, correction_bias, topk_weights, topk_ids, num_expert_group, topk_group, need_renorm, routed_scaling_factor, ) def _rocm_aiter_biased_grouped_topk_fake( gating_output: torch.Tensor, correction_bias: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_expert_group: int, topk_group: int, need_renorm: bool, routed_scaling_factor: float = 1.0, # mul to topk_weights ) -> None: pass def _rocm_aiter_grouped_topk_impl( gating_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_expert_group: int, topk_group: int, need_renorm: bool, scoring_func: str = "softmax", routed_scaling_factor: float = 1.0, # mul to topk_weights ) -> None: is_softmax = scoring_func == "softmax" from aiter import grouped_topk grouped_topk( gating_output, topk_weights, topk_ids, num_expert_group, topk_group, need_renorm, is_softmax, routed_scaling_factor, ) def _rocm_aiter_grouped_topk_fake( gating_output: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, num_expert_group: int, topk_group: int, need_renorm: bool, scoring_func: str = "softmax", routed_scaling_factor: float = 1.0, # mul to topk_weights ) -> None: pass # Cache whether aiter supports FP8 MLA parameters _AITER_MLA_SUPPORTS_FP8: bool | None = None def _check_aiter_mla_fp8_support() -> bool: """Check if aiter.mla.mla_decode_fwd supports q_scale and kv_scale parameters.""" global _AITER_MLA_SUPPORTS_FP8 if _AITER_MLA_SUPPORTS_FP8 is None: try: import inspect from aiter.mla import mla_decode_fwd sig = inspect.signature(mla_decode_fwd) _AITER_MLA_SUPPORTS_FP8 = ( "q_scale" in sig.parameters and "kv_scale" in sig.parameters ) except Exception: _AITER_MLA_SUPPORTS_FP8 = False return _AITER_MLA_SUPPORTS_FP8 def _rocm_aiter_mla_decode_fwd_impl( q: torch.Tensor, kv_buffer: torch.Tensor, o: torch.Tensor, qo_indptr: torch.Tensor, max_seqlen_qo: int, kv_indptr: torch.Tensor | None = None, kv_indices: torch.Tensor | None = None, kv_last_page_lens: torch.Tensor | None = None, sm_scale: float = 1.0, logit_cap: float = 0.0, q_scale: torch.Tensor | None = None, kv_scale: torch.Tensor | None = None, ) -> None: from aiter.mla import mla_decode_fwd kwargs = { "sm_scale": sm_scale, "logit_cap": logit_cap, } # Only pass q_scale and kv_scale if the aiter library supports them if _check_aiter_mla_fp8_support(): kwargs["q_scale"] = q_scale kwargs["kv_scale"] = kv_scale mla_decode_fwd( q, kv_buffer.view(-1, 1, 1, q.shape[-1]), o, qo_indptr, kv_indptr, kv_indices, kv_last_page_lens, max_seqlen_qo, **kwargs, ) def _rocm_aiter_mla_decode_fwd_fake( q: torch.Tensor, kv_buffer: torch.Tensor, o: torch.Tensor, qo_indptr: torch.Tensor, max_seqlen_qo: int, kv_indptr: torch.Tensor | None = None, kv_indices: torch.Tensor | None = None, kv_last_page_lens: torch.Tensor | None = None, sm_scale: float = 1.0, logit_cap: float = 0.0, q_scale: torch.Tensor | None = None, kv_scale: torch.Tensor | None = None, ) -> None: pass def _rocm_aiter_gemm_a8w8_impl( A: torch.Tensor, B: torch.Tensor, As: torch.Tensor, Bs: torch.Tensor, bias: torch.Tensor | None = None, output_dtype: torch.dtype = torch.float16, ) -> torch.Tensor: from aiter import gemm_a8w8_CK # gemm_a8w8_CK(a, b, scale_a, scale_b, bias) expects # a to be [M, K] # b to be [N, K] # CutlassScaledMMLinearKernel prepare weight `w_q` in [K, N] format return gemm_a8w8_CK(A, B, As, Bs, bias, output_dtype) def _rocm_aiter_gemm_a8w8_fake( A: torch.Tensor, B: torch.Tensor, As: torch.Tensor, Bs: torch.Tensor, bias: torch.Tensor | None = None, output_dtype: torch.dtype = torch.float16, ) -> torch.Tensor: m = A.shape[0] n = B.shape[0] Y = torch.empty(m, n, dtype=output_dtype, device=A.device) return Y def _rocm_aiter_triton_gemm_a8w8_blockscale_impl( A: torch.Tensor, B: torch.Tensor, As: torch.Tensor, Bs: torch.Tensor, output_dtype: torch.dtype = torch.float16, ) -> torch.Tensor: from aiter.ops.triton.gemm_a8w8_blockscale import gemm_a8w8_blockscale return gemm_a8w8_blockscale(A, B, As, Bs, dtype=output_dtype) def _rocm_aiter_triton_gemm_a8w8_blockscale_fake( A: torch.Tensor, B: torch.Tensor, As: torch.Tensor, Bs: torch.Tensor, output_dtype: torch.dtype = torch.float16, ) -> torch.Tensor: m = A.shape[0] n = B.shape[0] Y = torch.empty(m, n, dtype=output_dtype, device=A.device) return Y def _rocm_aiter_gemm_a8w8_blockscale_impl( A: torch.Tensor, B: torch.Tensor, As: torch.Tensor, Bs: torch.Tensor, output_dtype: torch.dtype = torch.float16, ) -> torch.Tensor: from aiter import gemm_a8w8_blockscale return gemm_a8w8_blockscale(A, B, As, Bs, dtype=output_dtype) def _rocm_aiter_gemm_a8w8_blockscale_fake( A: torch.Tensor, B: torch.Tensor, As: torch.Tensor, Bs: torch.Tensor, output_dtype: torch.dtype = torch.float16, ) -> torch.Tensor: m = A.shape[0] n = B.shape[0] Y = torch.empty(m, n, dtype=output_dtype, device=A.device) return Y def _rocm_aiter_rms_norm_impl( x: torch.Tensor, weight: torch.Tensor, variance_epsilon: float ) -> torch.Tensor: from aiter import rms_norm if x.dim() > 2: x_original_shape = x.shape x = x.reshape(-1, x_original_shape[-1]) x = rms_norm(x, weight, variance_epsilon) return x.reshape(x_original_shape) return rms_norm(x, weight, variance_epsilon) def _rocm_aiter_rms_norm_fake( x: torch.Tensor, weight: torch.Tensor, variance_epsilon: float ) -> torch.Tensor: return torch.empty_like(x) def _rocm_aiter_rmsnorm2d_fwd_with_add_impl( x: torch.Tensor, residual: torch.Tensor, weight: torch.Tensor, variance_epsilon: float, ) -> tuple[torch.Tensor, torch.Tensor]: from aiter import rmsnorm2d_fwd_with_add residual_out = torch.empty_like(residual) out = torch.empty_like(x) rmsnorm2d_fwd_with_add( out, # output x, # input residual, # residual input residual_out, # residual output weight, variance_epsilon, ) return out, residual_out def _rocm_aiter_rmsnorm2d_fwd_with_add_fake( x: torch.Tensor, residual: torch.Tensor, weight: torch.Tensor, variance_epsilon: float, ) -> tuple[torch.Tensor, torch.Tensor]: residual_out = torch.empty_like(residual) out = torch.empty_like(x) return out, residual_out def _rocm_aiter_rmsnorm_fused_add_dynamic_quant_impl( x: torch.Tensor, residual: torch.Tensor, weight: torch.Tensor, epsilon: float, quant_dtype: torch.dtype, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: import aiter as rocm_aiter assert quant_dtype in [torch.int8, _FP8_DTYPE] y_scale = torch.empty(x.shape[0], 1, dtype=torch.float32, device=x.device) out = torch.empty(x.shape, dtype=quant_dtype, device=x.device) residual_out = torch.empty_like(x) rocm_aiter.rmsnorm2d_fwd_with_add_dynamicquant( out, x, residual, residual_out, y_scale, weight, epsilon, use_model_sensitive_rmsnorm=0, ) return out, residual_out, y_scale def _rocm_aiter_rmsnorm_fused_add_dynamic_quant_fake( x: torch.Tensor, residual: torch.Tensor, weight: torch.Tensor, epsilon: float, quant_dtype: torch.dtype, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: y_scale = torch.empty(x.shape[0], 1, dtype=torch.float32, device=x.device) out = torch.empty(x.shape, dtype=quant_dtype, device=x.device) residual_out = torch.empty_like(x) return out, residual_out, y_scale def _rocm_aiter_rmsnorm_fused_dynamic_quant_impl( x: torch.Tensor, weight: torch.Tensor, epsilon: float, quant_dtype: torch.dtype, ) -> tuple[torch.Tensor, torch.Tensor]: import aiter as rocm_aiter assert quant_dtype in [torch.int8, _FP8_DTYPE] y_scale = torch.empty(x.shape[0], 1, dtype=torch.float32, device=x.device) out = torch.empty(x.shape, dtype=quant_dtype, device=x.device) rocm_aiter.rmsnorm2d_fwd_with_dynamicquant( out, x, y_scale, weight, epsilon, use_model_sensitive_rmsnorm=0 ) return out, y_scale def _rocm_aiter_rmsnorm_fused_dynamic_quant_fake( x: torch.Tensor, weight: torch.Tensor, epsilon: float, quant_dtype: torch.dtype, ) -> tuple[torch.Tensor, torch.Tensor]: y_scale = torch.empty(x.shape[0], 1, dtype=torch.float32, device=x.device) out = torch.empty(x.shape, dtype=quant_dtype, device=x.device) return out, y_scale def _rocm_aiter_per_tensor_quant_impl( x: torch.Tensor, quant_dtype: torch.dtype, scale: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: from aiter.ops.quant import per_tensor_quant_hip return per_tensor_quant_hip(x, scale, quant_dtype) def _rocm_aiter_per_tensor_quant_fake( x: torch.Tensor, quant_dtype: torch.dtype, scale: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: return torch.empty_like(x, dtype=quant_dtype), torch.empty( 1, dtype=torch.float32, device=x.device ) def _rocm_aiter_per_token_quant_impl( x: torch.Tensor, quant_dtype: torch.dtype, scale: torch.Tensor | None = None ) -> tuple[torch.Tensor, torch.Tensor]: from aiter.ops.quant import dynamic_per_token_scaled_quant assert quant_dtype in [torch.int8, _FP8_DTYPE] out_shape = x.shape out = torch.empty(x.shape, dtype=_FP8_DTYPE, device=x.device) if scale is None: scale = torch.empty((*out_shape[:-1], 1), dtype=torch.float32, device=x.device) dynamic_per_token_scaled_quant( out, x, scale, scale_ub=None, shuffle_scale=False, num_rows=None, num_rows_factor=1, ) return out, scale def _rocm_aiter_per_token_quant_fake( x: torch.Tensor, quant_dtype: torch.dtype, scale: torch.Tensor | None = None ) -> tuple[torch.Tensor, torch.Tensor]: out_shape = x.shape return ( torch.empty(x.shape, dtype=_FP8_DTYPE, device=x.device), torch.empty((*out_shape[:-1], 1), dtype=torch.float32, device=x.device), ) def _rocm_aiter_rmsnorm_with_add_fp8_group_quant_impl( x: torch.Tensor, residual: torch.Tensor, weight: torch.Tensor, variance_epsilon: float, group_size: int, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: from aiter.ops.triton.fused_fp8_quant import fused_rms_fp8_group_quant (x_quant, x_quant_scales), _, _, res = fused_rms_fp8_group_quant( x, weight, variance_epsilon, None, None, None, group_size=group_size, dtype_quant=AITER_FP8_DTYPE, res1=residual, ) return ( x_quant, res, x_quant_scales, ) def _rocm_aiter_rmsnorm_with_add_fp8_group_quant_fake( x: torch.Tensor, residual: torch.Tensor, weight: torch.Tensor, variance_epsilon: float, group_size: int, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: M, N = x.shape scale_shape = (M, (N + group_size - 1) // group_size) return ( torch.empty_like(x, dtype=AITER_FP8_DTYPE, device=x.device), torch.empty_like(residual, device=residual.device), torch.empty(scale_shape, dtype=torch.float32, device=x.device), ) def _rocm_aiter_rmsnorm_fp8_group_quant_impl( x: torch.Tensor, weight: torch.Tensor, variance_epsilon: float, group_size: int, ) -> tuple[torch.Tensor, torch.Tensor]: from aiter.ops.triton.fused_fp8_quant import fused_rms_fp8_group_quant (x_quant, x_quant_scales), _, _, res = fused_rms_fp8_group_quant( x, weight, variance_epsilon, None, None, None, group_size=group_size, dtype_quant=AITER_FP8_DTYPE, res1=None, ) return (x_quant, x_quant_scales) def _rocm_aiter_rmsnorm_fp8_group_quant_fake( x: torch.Tensor, weight: torch.Tensor, variance_epsilon: float, group_size: int, ) -> tuple[torch.Tensor, torch.Tensor]: M, N = x.shape scale_shape = (M, (N + group_size - 1) // group_size) return ( torch.empty_like(x, dtype=AITER_FP8_DTYPE, device=x.device), torch.empty(scale_shape, dtype=torch.float32, device=x.device), ) def _rocm_aiter_group_fp8_quant_impl( x: torch.Tensor, group_size: int, ) -> tuple[torch.Tensor, torch.Tensor]: assert x.shape[-1] % group_size == 0, "Input shape must be divisible by group size" from aiter import QuantType, get_hip_quant aiter_per1x128_quant = get_hip_quant(QuantType.per_1x128) return aiter_per1x128_quant(x.contiguous(), quant_dtype=AITER_FP8_DTYPE) def _rocm_aiter_group_fp8_quant_fake( x: torch.Tensor, group_size: int, ) -> tuple[torch.Tensor, torch.Tensor]: M, N = x.shape x_fp8 = torch.empty((M, N), dtype=AITER_FP8_DTYPE, device=x.device) out_bs = torch.empty( ( M, (N + group_size - 1) // group_size, ), dtype=torch.float32, device=x.device, ) return x_fp8, out_bs def _rocm_aiter_act_mul_and_fp8_group_quant_impl( x: torch.Tensor, group_size: int, ) -> tuple[torch.Tensor, torch.Tensor]: from aiter.ops.triton.activation import act_mul_and_fp8_group_quant return act_mul_and_fp8_group_quant( x, activation="silu", group_size=group_size, dtype_quant=AITER_FP8_DTYPE, ) def _rocm_aiter_act_mul_and_fp8_group_quant_fake( x: torch.Tensor, group_size: int, ) -> tuple[torch.Tensor, torch.Tensor]: M, N = x.shape assert N % 2 == 0 N_half = N // 2 x_fp8 = torch.empty((M, N_half), dtype=AITER_FP8_DTYPE, device=x.device) out_bs = torch.empty( ( M, (N_half + group_size - 1) // group_size, ), dtype=torch.float32, device=x.device, ) return x_fp8, out_bs # Global flag to ensure ops are registered only once _OPS_REGISTERED = False class rocm_aiter_ops: """ROCm AITER operations wrapper for AMD GPU acceleration in vLLM. This class centralizes the import and registration of AITER ops, and provides a unified interface for checking if AITER is enabled. Operations are only available on supported gfx9 architectures when aiter is installed. The class uses environment variables to control which features are enabled, allowing fine-grained control over which AITER optimizations are used. Environment Variables: VLLM_ROCM_USE_AITER: Main toggle for all AITER operations. VLLM_ROCM_USE_AITER_LINEAR: Controls GEMM and quantization ops. VLLM_ROCM_USE_AITER_RMSNORM: Controls RMSNorm operations. VLLM_ROCM_USE_AITER_MOE: Controls MoE (Mixture of Experts) ops. VLLM_ROCM_USE_AITER_MLA: Controls MLA (Multi-head Latent Attention) ops. VLLM_ROCM_USE_AITER_MHA: Controls MHA ops including flash_attn_varlen. VLLM_ROCM_USE_AITER_UNIFIED_ATTENTION: Controls Triton unified attention. VLLM_ROCM_USE_AITER_FP8BMM: Controls FP8 batched matrix multiply. VLLM_ROCM_USE_AITER_FP4_ASM_GEMM: Controls FP4 assembly GEMM. VLLM_ROCM_USE_AITER_TRITON_ROPE: Controls Triton rotary embeddings. VLLM_ROCM_USE_AITER_FUSION_SHARED_EXPERTS: Controls shared expert fusion. VLLM_ROCM_USE_AITER_TRITON_GEMM: Controls Triton unquantized GEMM. Note: The environment variables are assigned when the module is imported, so you can't change the environment variables after the module is imported. This is done out of performance consideration. Accessing environment variables is expensive as described in issue https://github.com/vllm-project/vllm/issues/17067 so we don't want to do it repeatedly, especially in the hot path (the forward pass). You can call the refresh_env_variables() function to reload the env variables after monkey patching the env variables in the unit test. Check Functions: All check functions (is_*_enabled) are decorated with @if_aiter_supported, which verifies: (1) platform is ROCm, (2) device arch is gfx9, and (3) aiter library is installed. The check function then also verifies the corresponding environment variable is enabled. i.e. ___ is_enabled() == current_platform.is_rocm() and | checked by current_platform.is_on_gfx9() and | @if_aiter_supported IS_AITER_FOUND and _______________| cls._AITER_ENABLED -----> Check by the logic in `is_enabled()` Example: from vllm._aiter_ops import rocm_aiter_ops # Check if aiter is enabled before using operations if rocm_aiter_ops.is_enabled(): result = rocm_aiter_ops.rms_norm(x, weight, epsilon) Operations: - RMS normalization: rms_norm, rms_norm2d_with_add - GEMM operations: gemm_a8w8, gemm_a8w8_blockscale - Fused MoE: fused_moe, asm_moe_tkw1 - Routing: topk_softmax, biased_grouped_topk, grouped_topk - MLA decode: mla_decode_fwd - Quantization: per_tensor_quant, per_token_quant, group_fp8_quant - Triton ops: triton_rotary_embed, triton_fp8_bmm, triton_gemm_a8w8_blockscale """ # Check if the env variable is set _AITER_ENABLED = envs.VLLM_ROCM_USE_AITER _LINEAR_ENABLED = envs.VLLM_ROCM_USE_AITER_LINEAR _RMSNORM_ENABLED = envs.VLLM_ROCM_USE_AITER_RMSNORM _FMOE_ENABLED = envs.VLLM_ROCM_USE_AITER_MOE _MLA_ENABLED = envs.VLLM_ROCM_USE_AITER_MLA _MHA_ENABLED = envs.VLLM_ROCM_USE_AITER_MHA _TRITON_UNIFIED_ATTN_ENABLED = envs.VLLM_ROCM_USE_AITER_UNIFIED_ATTENTION # TODO: Consolidate under _LINEAR_ENABLED _FP8BMM_ENABLED = envs.VLLM_ROCM_USE_AITER_FP8BMM # TODO: Consolidate under _LINEAR_ENABLED _FP4_GEMM_DYNAMIC_QUANT_ASM = envs.VLLM_ROCM_USE_AITER_FP4_ASM_GEMM # TODO: Consolidate under VLLM_ROCM_USE_AITER_ROPE _TRITON_ROTARY_EMBED = envs.VLLM_ROCM_USE_AITER_TRITON_ROPE _MOE_SHARED_EXPERTS_ENABLED = envs.VLLM_ROCM_USE_AITER_FUSION_SHARED_EXPERTS # TODO: Consolidate under _LINEAR_ENABLED _TRITON_UNQUANT_GEMM = envs.VLLM_ROCM_USE_AITER_TRITON_GEMM @classmethod def refresh_env_variables(cls): """ Since the environment variables are assigned when the module is imported, This is a helper function to reload all the env variables from the environment variables. for example, after monkey patching the env variables in the unit test, you can call this function to reload the env variables. """ cls._AITER_ENABLED = envs.VLLM_ROCM_USE_AITER cls._LINEAR_ENABLED = envs.VLLM_ROCM_USE_AITER_LINEAR cls._RMSNORM_ENABLED = envs.VLLM_ROCM_USE_AITER_RMSNORM cls._FMOE_ENABLED = envs.VLLM_ROCM_USE_AITER_MOE cls._MLA_ENABLED = envs.VLLM_ROCM_USE_AITER_MLA cls._MHA_ENABLED = envs.VLLM_ROCM_USE_AITER_MHA cls._TRITON_UNIFIED_ATTN_ENABLED = envs.VLLM_ROCM_USE_AITER_UNIFIED_ATTENTION cls._FP8BMM_ENABLED = envs.VLLM_ROCM_USE_AITER_FP8BMM cls._FP4_GEMM_DYNAMIC_QUANT_ASM = envs.VLLM_ROCM_USE_AITER_FP4_ASM_GEMM cls._TRITON_ROTARY_EMBED = envs.VLLM_ROCM_USE_AITER_TRITON_ROPE cls._MOE_SHARED_EXPERTS_ENABLED = envs.VLLM_ROCM_USE_AITER_FUSION_SHARED_EXPERTS cls._TRITON_UNQUANT_GEMM = envs.VLLM_ROCM_USE_AITER_TRITON_GEMM @classmethod @if_aiter_supported def is_enabled(cls) -> bool: return cls._AITER_ENABLED @classmethod @if_aiter_supported def is_linear_enabled(cls) -> bool: return cls._AITER_ENABLED and cls._LINEAR_ENABLED @classmethod @if_aiter_supported def is_linear_fp8_enabled(cls) -> bool: return cls.is_linear_enabled() @classmethod @if_aiter_supported def is_rmsnorm_enabled(cls) -> bool: return cls._AITER_ENABLED and cls._RMSNORM_ENABLED @classmethod @if_aiter_supported def is_fused_moe_enabled(cls) -> bool: return cls._AITER_ENABLED and cls._FMOE_ENABLED @classmethod @if_aiter_supported def is_fusion_moe_shared_experts_enabled(cls) -> bool: return cls.is_fused_moe_enabled() and cls._MOE_SHARED_EXPERTS_ENABLED @classmethod @if_aiter_supported def is_mla_enabled(cls) -> bool: return cls._AITER_ENABLED and cls._MLA_ENABLED @classmethod @if_aiter_supported def is_mha_enabled(cls) -> bool: return cls._AITER_ENABLED and cls._MHA_ENABLED @classmethod @if_aiter_supported def is_triton_unified_attn_enabled(cls) -> bool: return cls._AITER_ENABLED and cls._TRITON_UNIFIED_ATTN_ENABLED @classmethod @if_aiter_supported def is_fp8bmm_enabled(cls) -> bool: return cls._AITER_ENABLED and cls._FP8BMM_ENABLED @classmethod @if_aiter_supported def is_asm_fp4_gemm_dynamic_quant_enabled(cls) -> bool: return cls._AITER_ENABLED and cls._FP4_GEMM_DYNAMIC_QUANT_ASM @classmethod @if_aiter_supported def is_triton_rotary_embed_enabled(cls) -> bool: return cls._AITER_ENABLED and cls._TRITON_ROTARY_EMBED @classmethod @if_aiter_supported def is_triton_gemm_enabled(cls) -> bool: return cls._AITER_ENABLED and cls._TRITON_UNQUANT_GEMM @staticmethod @if_aiter_supported def register_ops_once() -> None: global _OPS_REGISTERED if not _OPS_REGISTERED: tags = ( tuple() if is_torch_equal_or_newer("2.7.0") else (torch.Tag.needs_fixed_stride_order,) ) # register all the custom ops here direct_register_custom_op( op_name="rocm_aiter_asm_moe_tkw1", op_func=_rocm_aiter_asm_moe_tkw1_impl, mutates_args=[], fake_impl=_rocm_aiter_asm_moe_tkw1_fake, dispatch_key=current_platform.dispatch_key, ) direct_register_custom_op( op_name="rocm_aiter_fused_moe", op_func=_rocm_aiter_fused_moe_impl, mutates_args=[], fake_impl=_rocm_aiter_fused_moe_fake, dispatch_key=current_platform.dispatch_key, ) direct_register_custom_op( op_name="rocm_aiter_topk_softmax", op_func=_rocm_aiter_topk_softmax_impl, mutates_args=["topk_weights", "topk_indices", "token_expert_indices"], fake_impl=_rocm_aiter_topk_softmax_fake, dispatch_key=current_platform.dispatch_key, ) direct_register_custom_op( op_name="rocm_aiter_biased_grouped_topk", op_func=_rocm_aiter_biased_grouped_topk_impl, mutates_args=["topk_weights", "topk_ids"], fake_impl=_rocm_aiter_biased_grouped_topk_fake, dispatch_key=current_platform.dispatch_key, ) direct_register_custom_op( op_name="rocm_aiter_grouped_topk", op_func=_rocm_aiter_grouped_topk_impl, mutates_args=["topk_weights", "topk_ids"], fake_impl=_rocm_aiter_grouped_topk_fake, dispatch_key=current_platform.dispatch_key, ) direct_register_custom_op( op_name="rocm_aiter_mla_decode_fwd", op_func=_rocm_aiter_mla_decode_fwd_impl, mutates_args=["o"], fake_impl=_rocm_aiter_mla_decode_fwd_fake, tags=tags, ) direct_register_custom_op( op_name="rocm_aiter_gemm_a8w8", op_func=_rocm_aiter_gemm_a8w8_impl, mutates_args=[], fake_impl=_rocm_aiter_gemm_a8w8_fake, dispatch_key=current_platform.dispatch_key, ) direct_register_custom_op( op_name="rocm_aiter_triton_gemm_a8w8_blockscale", op_func=_rocm_aiter_triton_gemm_a8w8_blockscale_impl, fake_impl=_rocm_aiter_triton_gemm_a8w8_blockscale_fake, ) direct_register_custom_op( op_name="rocm_aiter_gemm_a8w8_blockscale", op_func=_rocm_aiter_gemm_a8w8_blockscale_impl, fake_impl=_rocm_aiter_gemm_a8w8_blockscale_fake, ) direct_register_custom_op( op_name="rocm_aiter_rms_norm", op_func=_rocm_aiter_rms_norm_impl, fake_impl=_rocm_aiter_rms_norm_fake, ) direct_register_custom_op( op_name="rocm_aiter_rmsnorm2d_fwd_with_add", op_func=_rocm_aiter_rmsnorm2d_fwd_with_add_impl, fake_impl=_rocm_aiter_rmsnorm2d_fwd_with_add_fake, dispatch_key=current_platform.dispatch_key, ) direct_register_custom_op( op_name="rocm_aiter_rmsnorm_fused_dynamic_quant", op_func=_rocm_aiter_rmsnorm_fused_dynamic_quant_impl, fake_impl=_rocm_aiter_rmsnorm_fused_dynamic_quant_fake, dispatch_key=current_platform.dispatch_key, ) direct_register_custom_op( op_name="rocm_aiter_rmsnorm_fused_add_dynamic_quant", op_func=_rocm_aiter_rmsnorm_fused_add_dynamic_quant_impl, fake_impl=_rocm_aiter_rmsnorm_fused_add_dynamic_quant_fake, dispatch_key=current_platform.dispatch_key, ) direct_register_custom_op( op_name="rocm_aiter_rmsnorm_fp8_group_quant", op_func=_rocm_aiter_rmsnorm_fp8_group_quant_impl, fake_impl=_rocm_aiter_rmsnorm_fp8_group_quant_fake, ) direct_register_custom_op( op_name="rocm_aiter_rmsnorm_with_add_fp8_group_quant", op_func=_rocm_aiter_rmsnorm_with_add_fp8_group_quant_impl, fake_impl=_rocm_aiter_rmsnorm_with_add_fp8_group_quant_fake, ) direct_register_custom_op(
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/_custom_ops.py
vllm/_custom_ops.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import TYPE_CHECKING, Literal import torch import vllm.envs as envs from vllm.logger import init_logger from vllm.platforms import current_platform from vllm.scalar_type import ScalarType logger = init_logger(__name__) current_platform.import_kernels() if TYPE_CHECKING: def register_fake(fn): return lambda name: fn else: try: from torch.library import register_fake except ImportError: from torch.library import impl_abstract as register_fake # page attention ops def paged_attention_v1( out: torch.Tensor, query: torch.Tensor, key_cache: torch.Tensor, value_cache: torch.Tensor, num_kv_heads: int, scale: float, block_tables: torch.Tensor, seq_lens: torch.Tensor, block_size: int, max_seq_len: int, alibi_slopes: torch.Tensor | None, kv_cache_dtype: str, k_scale: torch.Tensor, v_scale: torch.Tensor, tp_rank: int = 0, blocksparse_local_blocks: int = 0, blocksparse_vert_stride: int = 0, blocksparse_block_size: int = 64, blocksparse_head_sliding_step: int = 0, ) -> None: torch.ops._C.paged_attention_v1( out, query, key_cache, value_cache, num_kv_heads, scale, block_tables, seq_lens, block_size, max_seq_len, alibi_slopes, kv_cache_dtype, k_scale, v_scale, tp_rank, blocksparse_local_blocks, blocksparse_vert_stride, blocksparse_block_size, blocksparse_head_sliding_step, ) def paged_attention_v2( out: torch.Tensor, exp_sum: torch.Tensor, max_logits: torch.Tensor, tmp_out: torch.Tensor, query: torch.Tensor, key_cache: torch.Tensor, value_cache: torch.Tensor, num_kv_heads: int, scale: float, block_tables: torch.Tensor, seq_lens: torch.Tensor, block_size: int, max_seq_len: int, alibi_slopes: torch.Tensor | None, kv_cache_dtype: str, k_scale: torch.Tensor, v_scale: torch.Tensor, tp_rank: int = 0, blocksparse_local_blocks: int = 0, blocksparse_vert_stride: int = 0, blocksparse_block_size: int = 64, blocksparse_head_sliding_step: int = 0, ) -> None: torch.ops._C.paged_attention_v2( out, exp_sum, max_logits, tmp_out, query, key_cache, value_cache, num_kv_heads, scale, block_tables, seq_lens, block_size, max_seq_len, alibi_slopes, kv_cache_dtype, k_scale, v_scale, tp_rank, blocksparse_local_blocks, blocksparse_vert_stride, blocksparse_block_size, blocksparse_head_sliding_step, ) def paged_attention_rocm( out: torch.Tensor, exp_sum: torch.Tensor, max_logits: torch.Tensor, tmp_out: torch.Tensor, query: torch.Tensor, key_cache: torch.Tensor, value_cache: torch.Tensor, num_kv_heads: int, scale: float, block_tables: torch.Tensor, seq_lens: torch.Tensor, query_start_loc: torch.Tensor | None, block_size: int, max_seq_len: int, alibi_slopes: torch.Tensor | None, kv_cache_dtype: str, k_scale: torch.Tensor, v_scale: torch.Tensor, fp8_out_scale: torch.Tensor | None = None, mfma_type: str = "fp8" if envs.VLLM_ROCM_FP8_MFMA_PAGE_ATTN else "f16", ) -> None: torch.ops._rocm_C.paged_attention( out, exp_sum, max_logits, tmp_out, query, key_cache, value_cache, num_kv_heads, scale, block_tables, seq_lens, query_start_loc, block_size, max_seq_len, alibi_slopes, kv_cache_dtype, k_scale, v_scale, fp8_out_scale, mfma_type, ) def mla_decode_kvcache_cpu( out: torch.Tensor, query: torch.Tensor, kv_cache: torch.Tensor, scale: float, block_tables: torch.Tensor, seq_lens: torch.Tensor, ) -> None: torch.ops._C_cpu.mla_decode_kvcache( out, query, kv_cache, scale, block_tables, seq_lens ) # merge attn states ops def merge_attn_states( output: torch.Tensor, prefix_output: torch.Tensor, prefix_lse: torch.Tensor, suffix_output: torch.Tensor, suffix_lse: torch.Tensor, output_lse: torch.Tensor | None = None, ) -> None: torch.ops._C.merge_attn_states( output, output_lse, prefix_output, prefix_lse, suffix_output, suffix_lse ) def convert_vertical_slash_indexes( q_seqlens: torch.Tensor, # [BATCH, ] kv_seqlens: torch.Tensor, # [BATCH, ] vertical_indexes: torch.Tensor, # [BATCH, N_HEADS, NNZ_V] slash_indexes: torch.Tensor, # [BATCH, N_HEADS, NNZ_S] context_size: int, block_size_M: int, block_size_N: int, causal: bool = True, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: batch_size = slash_indexes.size(0) num_heads = slash_indexes.size(1) nnz_slash = slash_indexes.size(2) nnz_vertical = vertical_indexes.size(2) num_rows = (context_size + block_size_M - 1) // block_size_M block_count = torch.zeros( batch_size, num_heads, num_rows, dtype=q_seqlens.dtype, device=q_seqlens.device ) block_offset = torch.zeros( batch_size, num_heads, num_rows, nnz_slash, dtype=q_seqlens.dtype, device=q_seqlens.device, ) column_count = torch.zeros( batch_size, num_heads, num_rows, dtype=q_seqlens.dtype, device=q_seqlens.device ) column_index = torch.zeros( batch_size, num_heads, num_rows, nnz_vertical, dtype=q_seqlens.dtype, device=q_seqlens.device, ) torch.ops._C.convert_vertical_slash_indexes( block_count, block_offset, column_count, column_index, q_seqlens, kv_seqlens, vertical_indexes, slash_indexes, context_size, block_size_M, block_size_N, causal, ) return block_count, block_offset, column_count, column_index def convert_vertical_slash_indexes_mergehead( q_seqlens: torch.Tensor, # [BATCH, ] kv_seqlens: torch.Tensor, # [BATCH, ] vertical_indexes: torch.Tensor, # [BATCH, N_HEADS, NNZ_V] slash_indexes: torch.Tensor, # [BATCH, N_HEADS, NNZ_S] # [N_HEADS] : different head use different number of indices vertical_indices_count: torch.Tensor, slash_indices_count: torch.Tensor, context_size: int, block_size_M: int, block_size_N: int, causal: bool = True, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: batch_size = slash_indexes.size(0) num_heads = slash_indexes.size(1) nnz_slash = slash_indexes.size(2) nnz_vertical = vertical_indexes.size(2) num_rows = (context_size + block_size_M - 1) // block_size_M block_count = torch.empty( batch_size, num_heads, num_rows, dtype=q_seqlens.dtype, device=q_seqlens.device ) block_offset = torch.empty( batch_size, num_heads, num_rows, nnz_slash, dtype=q_seqlens.dtype, device=q_seqlens.device, ) column_count = torch.empty( batch_size, num_heads, num_rows, dtype=q_seqlens.dtype, device=q_seqlens.device ) column_index = torch.empty( batch_size, num_heads, num_rows, nnz_vertical, dtype=q_seqlens.dtype, device=q_seqlens.device, ) torch.ops._C.convert_vertical_slash_indexes_mergehead( block_count, block_offset, column_count, column_index, q_seqlens, kv_seqlens, vertical_indexes, slash_indexes, vertical_indices_count, slash_indices_count, context_size, block_size_M, block_size_N, causal, ) return block_count, block_offset, column_count, column_index # pos encoding ops def rotary_embedding( positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None, head_size: int, cos_sin_cache: torch.Tensor, is_neox: bool, ) -> None: torch.ops._C.rotary_embedding( positions, query, key, head_size, cos_sin_cache, is_neox ) # layer norm ops def rms_norm( out: torch.Tensor, input: torch.Tensor, weight: torch.Tensor, epsilon: float ) -> None: torch.ops._C.rms_norm(out, input, weight, epsilon) def fused_add_rms_norm( input: torch.Tensor, residual: torch.Tensor, weight: torch.Tensor, epsilon: float ) -> None: torch.ops._C.fused_add_rms_norm(input, residual, weight, epsilon) def fused_qk_norm_rope( qkv: torch.Tensor, num_heads_q: int, num_heads_k: int, num_heads_v: int, head_dim: int, eps: float, q_weight: torch.Tensor, k_weight: torch.Tensor, cos_sin_cache: torch.Tensor, is_neox: bool, position_ids: torch.Tensor, ) -> None: torch.ops._C.fused_qk_norm_rope( qkv, num_heads_q, num_heads_k, num_heads_v, head_dim, eps, q_weight, k_weight, cos_sin_cache, is_neox, position_ids, ) def apply_repetition_penalties_torch( logits: torch.Tensor, prompt_mask: torch.Tensor, output_mask: torch.Tensor, repetition_penalties: torch.Tensor, ) -> None: repetition_penalties = repetition_penalties.unsqueeze(dim=1).repeat( 1, logits.size(1) ) # If token appears in prompt or output, apply, otherwise use 1.0 for no-op. penalties = torch.where(prompt_mask | output_mask, repetition_penalties, 1.0) # If logits are positive, divide by penalty, otherwise multiply by penalty. scaling = torch.where(logits > 0, 1.0 / penalties, penalties) logits *= scaling def apply_repetition_penalties_cuda( logits: torch.Tensor, prompt_mask: torch.Tensor, output_mask: torch.Tensor, repetition_penalties: torch.Tensor, ) -> None: torch.ops._C.apply_repetition_penalties_( logits, prompt_mask, output_mask, repetition_penalties ) def apply_repetition_penalties( logits: torch.Tensor, prompt_mask: torch.Tensor, output_mask: torch.Tensor, repetition_penalties: torch.Tensor, ) -> None: """Apply repetition penalties to logits in-place. Args: logits: The logits tensor of shape [num_seqs, vocab_size]. prompt_mask: A boolean tensor indicating which tokens appear in the prompt. output_mask: A boolean tensor indicating which tokens appear in the output. repetition_penalties: The repetition penalties of shape (num_seqs, ). """ if logits.is_cuda and logits.is_contiguous(): apply_repetition_penalties_cuda( logits, prompt_mask, output_mask, repetition_penalties ) else: apply_repetition_penalties_torch( logits, prompt_mask, output_mask, repetition_penalties ) # fused quant layer norm ops def rms_norm_dynamic_per_token_quant( input: torch.Tensor, weight: torch.Tensor, epsilon: float, quant_dtype: torch.dtype, scale_ub: torch.Tensor | None = None, residual: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: output = torch.empty_like(input, dtype=quant_dtype) scales = torch.empty( (input.numel() // input.shape[-1], 1), device=input.device, dtype=torch.float32 ) torch.ops._C.rms_norm_dynamic_per_token_quant( output, input, weight, scales, epsilon, scale_ub, residual ) return output, scales # fused quant layer norm ops blocked def rms_norm_per_block_quant( input: torch.Tensor, weight: torch.Tensor, epsilon: float, quant_dtype: torch.dtype, group_size: list[int], scale_ub: torch.Tensor | None = None, residual: torch.Tensor | None = None, is_scale_transposed: bool = False, ) -> tuple[torch.Tensor, torch.Tensor]: assert len(group_size) == 2 output = torch.empty_like(input, dtype=quant_dtype) if is_scale_transposed: scales = torch.empty( (input.shape[-1] // group_size[1], input.numel() // input.shape[-1]), device=input.device, dtype=torch.float32, ).transpose(0, 1) else: scales = torch.empty( (input.numel() // input.shape[-1], input.shape[-1] // group_size[1]), device=input.device, dtype=torch.float32, ) torch.ops._C.rms_norm_per_block_quant( output, input, weight, scales, epsilon, scale_ub, residual, group_size[1], is_scale_transposed, ) return output, scales # quantization ops # awq def awq_dequantize( qweight: torch.Tensor, scales: torch.Tensor, zeros: torch.Tensor, split_k_iters: int, thx: int, thy: int, ) -> torch.Tensor: if envs.VLLM_USE_TRITON_AWQ: from vllm.model_executor.layers.quantization.awq_triton import ( awq_dequantize_triton, ) return awq_dequantize_triton(qweight, scales, zeros) return torch.ops._C.awq_dequantize(qweight, scales, zeros, split_k_iters, thx, thy) def awq_gemm( input: torch.Tensor, qweight: torch.Tensor, scales: torch.Tensor, qzeros: torch.Tensor, split_k_iters: int, ) -> torch.Tensor: if envs.VLLM_USE_TRITON_AWQ: from vllm.model_executor.layers.quantization.awq_triton import awq_gemm_triton return awq_gemm_triton(input, qweight, scales, qzeros, split_k_iters) return torch.ops._C.awq_gemm(input, qweight, scales, qzeros, split_k_iters) # gptq def gptq_gemm( a: torch.Tensor, b_q_weight: torch.Tensor, b_gptq_qzeros: torch.Tensor, b_gptq_scales: torch.Tensor, b_g_idx: torch.Tensor, use_exllama: bool, use_v2_format: bool, bit: int, ) -> torch.Tensor: return torch.ops._C.gptq_gemm( a, b_q_weight, b_gptq_qzeros, b_gptq_scales, b_g_idx, use_exllama, use_v2_format, bit, ) if hasattr(torch.ops._C, "gptq_gemm"): @register_fake("_C::gptq_gemm") def _gptq_gemm_fake( a: torch.Tensor, b_q_weight: torch.Tensor, b_gptq_qzeros: torch.Tensor, b_gptq_scales: torch.Tensor, b_g_idx: torch.Tensor, use_exllama: bool, use_v2_format: bool, bit: int, ) -> torch.Tensor: return torch.empty( (a.size(0), b_q_weight.size(1)), dtype=a.dtype, device=a.device ) def gptq_shuffle(q_weight: torch.Tensor, q_perm: torch.Tensor, bit: int) -> None: torch.ops._C.gptq_shuffle(q_weight, q_perm, bit) # marlin_24 def gptq_marlin_24_gemm( a: torch.Tensor, b_q_weight: torch.Tensor, b_meta: torch.Tensor, b_scales: torch.Tensor, workspace: torch.Tensor, b_q_type: ScalarType, size_m: int, size_n: int, size_k: int, ) -> torch.Tensor: return torch.ops._C.gptq_marlin_24_gemm( a, b_q_weight, b_meta, b_scales, workspace, b_q_type.id, size_m, size_n, size_k ) if hasattr(torch.ops._C, "gptq_marlin_24_gemm"): @register_fake("_C::gptq_marlin_24_gemm") def _gptq_marlin_24_gemm_fake( a: torch.Tensor, b_q_weight: torch.Tensor, b_meta: torch.Tensor, b_scales: torch.Tensor, workspace: torch.Tensor, b_q_type: ScalarType, size_m: torch.SymInt, size_n: torch.SymInt, size_k: torch.SymInt, ) -> torch.Tensor: return torch.empty((size_m, size_n), device=a.device, dtype=a.dtype) @register_fake("_C::gptq_marlin_gemm") def _gptq_marlin_gemm_fake( a: torch.Tensor, c: torch.Tensor | None, b_q_weight: torch.Tensor, b_bias: torch.Tensor | None, b_scales: torch.Tensor, a_scales: torch.Tensor | None, global_scale: torch.Tensor | None, b_zeros: torch.Tensor | None, g_idx: torch.Tensor | None, perm: torch.Tensor | None, workspace: torch.Tensor, b_q_type_id: int, size_m: torch.SymInt, size_n: torch.SymInt, size_k: torch.SymInt, is_k_full: bool = True, use_atomic_add: bool = False, use_fp32_reduce: bool = False, is_zp_float: bool = False, ) -> torch.Tensor: dtype = a.dtype if dtype not in [torch.half, torch.bfloat16]: dtype = b_scales.dtype return torch.empty((size_m, size_n), device=a.device, dtype=dtype) @register_fake("_C::awq_dequantize") def _awq_dequantize_fake( qweight: torch.Tensor, scales: torch.Tensor, zeros: torch.Tensor, split_k_iters: torch.SymInt, thx: int, thy: int, ) -> torch.Tensor: in_c = qweight.size(0) qout_c = qweight.size(1) out_c = qout_c * 8 return torch.empty((in_c, out_c), dtype=scales.dtype, device=scales.device) @register_fake("_C::awq_gemm") def _awq_gemm_fake( input: torch.Tensor, qweight: torch.Tensor, scales: torch.Tensor, qzeros: torch.Tensor, split_k_iters: torch.SymInt, ) -> torch.Tensor: num_in_feats = input.size(0) return torch.empty( (split_k_iters, num_in_feats, qweight.size(1) * 8), dtype=input.dtype, device=input.device, ).sum(0) @register_fake("_C::machete_mm") def machete_mm_fake( a: torch.Tensor, # b_q Should be the tensor returned by machete_prepack_B b_q: torch.Tensor, b_type: ScalarType, out_type: torch.dtype | None = None, b_group_scales: torch.Tensor | None = None, b_group_zeros: torch.Tensor | None = None, b_group_size: int | None = None, b_channel_scales: torch.Tensor | None = None, a_token_scales: torch.Tensor | None = None, schedule: str | None = None, ) -> torch.Tensor: m = a.size(0) n = b_q.size(1) return torch.empty((m, n), device=a.device, dtype=a.dtype) @register_fake("_C::machete_prepack_B") def machete_prepack_B_fake( b_q_weight: torch.Tensor, a_type: torch.dtype, b_type: ScalarType, group_scales_type: torch.dtype | None, ) -> torch.Tensor: return torch.empty_like(b_q_weight, memory_format=torch.contiguous_format) @register_fake("_C::cutlass_w4a8_mm") def cutlass_w4a8_mm_fake( a: torch.Tensor, # b_q Should be the tensor returned by cutlass_encode_and_reorder_int4b b_q: torch.Tensor, b_group_scales: torch.Tensor, b_group_size: int, b_channel_scales: torch.Tensor, a_token_scales: torch.Tensor, out_type: torch.dtype | None = None, maybe_schedule: str | None = None, ) -> torch.Tensor: m = a.size(0) n = b_q.size(1) out_dtype = out_type if out_type is not None else torch.bfloat16 return torch.empty((m, n), device=a.device, dtype=out_dtype) @register_fake("_C::cutlass_pack_scale_fp8") def cutlass_pack_scale_fp8_fake(scales: torch.Tensor) -> torch.Tensor: return torch.empty_like(scales, memory_format=torch.contiguous_format) @register_fake("_C::cutlass_encode_and_reorder_int4b") def cutlass_encode_and_reorder_int4b_fake(b: torch.Tensor) -> torch.Tensor: return torch.empty_like(b, memory_format=torch.contiguous_format) @register_fake("_C::cutlass_encode_and_reorder_int4b_grouped") def cutlass_encode_and_reorder_int4b_grouped_fake(b: torch.Tensor) -> torch.Tensor: return torch.empty_like(b, memory_format=torch.contiguous_format) if hasattr(torch.ops._C, "allspark_w8a16_gemm"): @register_fake("_C::allspark_w8a16_gemm") def _allspark_w8a16_gemm_fake( a: torch.Tensor, b_qweight: torch.Tensor, b_scales: torch.Tensor, b_qzeros: torch.Tensor | None, n: torch.SymInt, group_size: torch.SymInt, sm_count: torch.SymInt, sm_version: torch.SymInt, CUBLAS_M_THRESHOLD: torch.SymInt, has_zp: bool, n32k16_reorder: bool, ) -> torch.Tensor: m = a.size(0) return torch.empty((m, n), device=a.device, dtype=a.dtype) if hasattr(torch.ops._C, "ggml_dequantize"): @register_fake("_C::ggml_dequantize") def _ggml_dequantize_fake( W: torch.Tensor, quant_type: int, m: torch.SymInt, n: torch.SymInt, dtype: torch.dtype | None = None, ) -> torch.Tensor: return torch.empty((m, n), dtype=torch.float16, device=W.device) @register_fake("_C::ggml_mul_mat_vec_a8") def _ggml_mul_mat_vec_a8_fake( W: torch.Tensor, X: torch.Tensor, quant_type: int, row: torch.SymInt, ) -> torch.Tensor: return torch.empty((X.shape[0], row), dtype=X.dtype, device=W.device) @register_fake("_C::ggml_mul_mat_a8") def _ggml_mul_mat_a8_fake( W: torch.Tensor, X: torch.Tensor, quant_type: int, row: torch.SymInt, ) -> torch.Tensor: batch = X.size(0) return torch.empty((batch, row), dtype=X.dtype, device=W.device) @register_fake("_C::ggml_moe_a8") def _ggml_moe_a8_fake( X: torch.Tensor, W: torch.Tensor, sorted_token_ids: torch.Tensor, expert_ids: torch.Tensor, num_tokens_post_padded: torch.Tensor, quant_type: int, row: torch.SymInt, top_k: torch.SymInt, tokens: torch.SymInt, ) -> torch.Tensor: tokens = X.size(0) return torch.empty((tokens * top_k, row), dtype=torch.float16, device=W.device) if hasattr(torch.ops._C, "ggml_moe_a8_vec"): @register_fake("_C::ggml_moe_a8_vec") def _ggml_moe_a8_vec_fake( X: torch.Tensor, W: torch.Tensor, topk_ids: torch.Tensor, top_k: int, quant_type: int, row: torch.SymInt, tokens: torch.SymInt, ) -> torch.Tensor: tokens = X.size(0) return torch.empty((tokens * top_k, row), dtype=X.dtype, device=W.device) # cutlass def cutlass_scaled_mm_supports_fp4(cuda_device_capability: int) -> bool: return torch.ops._C.cutlass_scaled_mm_supports_fp4(cuda_device_capability) def cutlass_scaled_fp4_mm( a: torch.Tensor, b: torch.Tensor, block_scale_a: torch.Tensor, block_scale_b: torch.Tensor, alpha: torch.Tensor, out_dtype: torch.dtype, ) -> torch.Tensor: assert a.ndim == 2 and b.ndim == 2 m, n = a.shape[0], b.shape[0] out = torch.empty((m, n), dtype=out_dtype, device=a.device) torch.ops._C.cutlass_scaled_fp4_mm(out, a, b, block_scale_a, block_scale_b, alpha) return out def cutlass_scaled_mm_supports_fp8(cuda_device_capability: int) -> bool: return torch.ops._C.cutlass_scaled_mm_supports_fp8(cuda_device_capability) def cutlass_scaled_mm_supports_block_fp8(cuda_device_capability: int) -> bool: return torch.ops._C.cutlass_scaled_mm_supports_block_fp8(cuda_device_capability) def cutlass_scaled_mm( a: torch.Tensor, b: torch.Tensor, scale_a: torch.Tensor, scale_b: torch.Tensor, out_dtype: torch.dtype, bias: torch.Tensor | None = None, ) -> torch.Tensor: """ `cutlass_scaled_mm` implements a fused version of `output = torch.mm((scale_a * a), (scale_b * b)).to(out_dtype)` where scale_a * a and scale_b * b are implemented using numpy-style broadcasting. In order to support blockwise scaling like found in DeepSeek V3 we also support extended "group" broadcast rules. We extend the numpy-style broadcasting rules with the following rule: "if the extent of a dimension in the source shape is between 1 and corresponding extent in the target shape we repeat each element along that dimension src_shape[dim] // target_shape[dim] times consecutively" example if we have: a = [[1, 2], and target_shape = (2, 4) [3, 4]] then we would expand a to: a = [[1, 1, 2, 2], [3, 3, 4, 4]] currently we only support the case: scale_a.shape * [1, 128] == a.shape scale_b.shape * [128, 128] == b.shape """ assert out_dtype is torch.bfloat16 or out_dtype is torch.float16 assert bias is None or bias.numel() == b.shape[1] and bias.dtype == out_dtype # Massage the input to be 2D target_shape = (*a.shape[:-1], b.shape[1]) a = a.view(-1, a.shape[-1]) cutlass_compatible_b = b.shape[0] % 16 == 0 and b.shape[1] % 16 == 0 if current_platform.is_rocm() or not cutlass_compatible_b: from vllm.model_executor.layers.quantization.compressed_tensors.triton_scaled_mm import ( # noqa triton_scaled_mm, ) out = triton_scaled_mm(a, b, scale_a, scale_b, out_dtype, bias) else: out = torch.empty((a.shape[0], b.shape[1]), dtype=out_dtype, device=a.device) torch.ops._C.cutlass_scaled_mm(out, a, b, scale_a, scale_b, bias) return out.view(*target_shape) def cutlass_scaled_mm_azp( a: torch.Tensor, b: torch.Tensor, scale_a: torch.Tensor, scale_b: torch.Tensor, out_dtype: torch.dtype, azp_adj: torch.Tensor, azp: torch.Tensor | None = None, bias: torch.Tensor | None = None, ) -> torch.Tensor: """ :param azp_adj: In the per-tensor case, this should include the azp. Always per-channel. :param azp: Only set in the per-token case. Per-token if set. """ assert b.shape[0] % 16 == 0 and b.shape[1] % 16 == 0 assert out_dtype is torch.bfloat16 or out_dtype is torch.float16 assert bias is None or bias.numel() == b.shape[1] and bias.dtype == out_dtype # Massage the input to be 2D target_shape = (*a.shape[:-1], b.shape[1]) a = a.view(-1, a.shape[-1]) assert azp is None or azp.numel() == a.shape[0] out = torch.empty((a.shape[0], b.shape[1]), dtype=out_dtype, device=a.device) torch.ops._C.cutlass_scaled_mm_azp(out, a, b, scale_a, scale_b, azp_adj, azp, bias) return out.view(*target_shape) def cutlass_sparse_scaled_mm_supported(cuda_device_capability: int) -> bool: return torch.ops._C.cutlass_sparse_scaled_mm_supported(cuda_device_capability) def cutlass_group_gemm_supported(cuda_device_capability: int) -> bool: try: return torch.ops._C.cutlass_group_gemm_supported(cuda_device_capability) except AttributeError: # Return False on non-CUDA platforms where it is not available return False def cutlass_sparse_compress(a: torch.Tensor) -> tuple[torch.Tensor, torch.Tensor]: """ Compresses a sparse matrix for use with Cutlass sparse operations. This function takes a dense tensor and compresses it into two components: non-zero elements and metadata. The compressed representation is compatible with Cutlass sparse kernels. Args: a (torch.Tensor): The input tensor to be compressed. Must have one of the following data types: - `torch.int8` - `torch.float8_e4m3fn` - `torch.bfloat16` - `torch.float16` Returns: tuple[torch.Tensor, torch.Tensor]: A tuple containing: - `a_nzs` (torch.Tensor): A tensor containing non-zero elements of `a`. - `a_meta` (torch.Tensor): A tensor containing metadata for the sparse representation. Raises: ValueError: If the compression operation fails. Notes: - The `a_meta` tensor has a data type of `torch.uint8`. - Each metadata element encodes the sparsity of 4 non-zero elements (i.e., `elemsPerMetaElem = 4`). - The shape of `a_nzs` is `(m, k // 2)`, where `m` and `k` are the dimensions of the input tensor. - The shape of `a_meta` is `(m, k // 2 // elemsPerMetaElem)`. """ assert a.dtype in [torch.int8, torch.float8_e4m3fn, torch.bfloat16, torch.float16] assert a.is_contiguous() # a_meta.dtype: torch.uint8 so elemsPerMetaElem = 8b / 2b_per_nz = 4 elemsPerMetaElem = 4 assert a.shape[1] % (2 * elemsPerMetaElem) == 0 return torch.ops._C.cutlass_sparse_compress(a) def cutlass_scaled_sparse_mm( a: torch.Tensor, bt_nzs: torch.Tensor, bt_meta: torch.Tensor, scale_a: torch.Tensor, scale_b: torch.Tensor, out_dtype: torch.dtype, bias: torch.Tensor | None = None, ) -> torch.Tensor: """ Performs a scaled sparse matrix multiplication using Cutlass. Steps: 1. Create a dense matrix `a` of shape (m, k) on the CUDA device: `a = torch.randn((m, k), device='cuda')`. 2. Create a dense matrix `b` of shape (k, n) on the CUDA device: `b = torch.randn((k, n), device='cuda')`. 3. Prune matrix `b` to 2:4 sparsity along the specified dimension: `b = prune_to_2_4(b, dim=0)`. 4. Compress the transposed sparse matrix `b.t()`: `bt_nzs, bt_meta = cutlass_sparse_compress(b.t())`. 5. Perform sparse matrix multiplication using the compressed matrix, applying scaling factors for `a` and `b`, and the output data type: `out = cutlass_scaled_sparse_mm(a, bt_nzs, bt_meta, scale_a, scale_b, out_dtype)`. Returns: - The result of the scaled sparse matrix multiplication. """ assert bt_nzs.shape[0] % 16 == 0 and bt_nzs.shape[1] % 16 == 0 assert out_dtype is torch.bfloat16 or out_dtype is torch.float16 assert bias is None or bias.shape[0] == bt_nzs.shape[0] and bias.dtype == out_dtype m = a.shape[0] n = bt_nzs.shape[0] out = torch.empty((m, n), dtype=out_dtype, device=a.device) torch.ops._C.cutlass_scaled_sparse_mm( out, a, bt_nzs, bt_meta, scale_a, scale_b, bias ) return out def get_cutlass_moe_mm_data( topk_ids: torch.Tensor, expert_offsets: torch.Tensor, problem_sizes1: torch.Tensor, problem_sizes2: torch.Tensor, input_permutation: torch.Tensor, output_permutation: torch.Tensor, num_experts: int, n: int, k: int, blockscale_offsets: torch.Tensor | None = None, ): """ Prepare data necessary to perform CUTLASS grouped matrix multiplications used in CUTLASS-based fused MoE. The function takes in topk_ids (token-expert mapping) and uses it to compute: - expert_offsets: Indices that mark at which token index each expert begins its computation after the input is sorted with input_permutation. The number of tokens computed with expert E is expert_offsets[E + 1] - expert_offsets[E] - problem_sizes1, problem_sizes2: MxNxK sizes of each expert's multiplication in two grouped MMs used in the fused MoE operation. - input_permutation: Permutation that must be used to shuffle the input before executing the MMs. - output_permutation: Permutation that must be used to shuffle the output after executing the MMs. - blockscale_offsets: Optional argument passed for fp4 moe. Indices that mark at which block scale index each expert begins its computation. The number of block scale rows computed with expert E is blockscale_offsets[E + 1] - blockscale_offsets[E] """ return torch.ops._C.get_cutlass_moe_mm_data( topk_ids, expert_offsets, problem_sizes1, problem_sizes2, input_permutation, output_permutation, num_experts, n, k, blockscale_offsets, ) def get_cutlass_moe_mm_problem_sizes( topk_ids: torch.Tensor, problem_sizes1: torch.Tensor, problem_sizes2: torch.Tensor, num_experts: int, n: int, k: int, blockscale_offsets: torch.Tensor | None = None, force_swap_ab: bool | None = None, ): """ Compute only the per-expert problem sizes needed by the two grouped matrix multiplications used in CUTLASS-based fused MoE. The function takes in topk_ids (token→expert mapping) and computes: - problem_sizes1, problem_sizes2: M×N×K sizes of each expert's multiplication for the two grouped MMs used in the fused MoE operation. Optional: - force_swap_ab: If set to True or False, explicitly enable or disable the A/B input swap optimization. If None (default), the swap
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/logprobs.py
vllm/logprobs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import itertools from collections.abc import Iterable, Iterator, MutableSequence from dataclasses import dataclass, field from typing import overload # We use dataclass for now because it is used for # openai server output, and msgspec is not serializable. # TODO(sang): Fix it. @dataclass class Logprob: """Infos for supporting OpenAI compatible logprobs and token ranks. Attributes: logprob: The logprob of chosen token rank: The vocab rank of chosen token (>=1) decoded_token: The decoded chosen token index """ logprob: float rank: int | None = None decoded_token: str | None = None LogprobsOnePosition = dict[int, Logprob] @dataclass class FlatLogprobs(MutableSequence[LogprobsOnePosition]): """ Flat logprobs of a request into multiple primitive type lists. Compared to list[dict[int, Logprob]], this data structure reduced GC overhead significantly. As it flattened logprob information for all positions and ranks in to multiple primitive type lists (i.e. logprobs, token_ids, ranks per token_ids, decoded_tokens). So regardless of the sequence length and top_logprobs setup, FlatLogprobs would only introduce a constant amount of objects. As each position might contains different amount of ranks, start_indices_per_position would be used to access the logprob ranges for different positions. NOTE: To reduce the migration overhead and improve backward compatibility, we support the key Sequence APIs of list, so it could act as list[LogprobsOnePosition] """ # Start / end indices to indicate the range of logprobs for each position. start_indices: list[int] = field(default_factory=list) end_indices: list[int] = field(default_factory=list) # Flatten Logprob information for (each position, rank). # For position <i>, the logprobs are ranged # from self.start_indices[i] to self.end_indices[i] (exclusive). token_ids: list[int] = field(default_factory=list) logprobs: list[float] = field(default_factory=list) ranks: list[int | None] = field(default_factory=list) decoded_tokens: list[str | None] = field(default_factory=list) def append(self, logprobs_one_position: LogprobsOnePosition | None) -> None: """Appends the container with logprobs for the next position""" self.start_indices.append(len(self.logprobs)) if logprobs_one_position: for token_id, logprob in logprobs_one_position.items(): self.token_ids.append(token_id) self.logprobs.append(logprob.logprob) self.ranks.append(logprob.rank) self.decoded_tokens.append(logprob.decoded_token) self.end_indices.append(len(self.logprobs)) def append_fast( self, token_ids: list[int], logprobs: list[float], ranks: itertools.chain[int], decoded_tokens: Iterable[str | None], ) -> None: """ Appends logprobs for the next position without creating the intermediate logprob dictionary. """ self.start_indices.append(len(self.logprobs)) for token_id, logprob, rank, decoded_token in zip( token_ids, logprobs, ranks, decoded_tokens ): self.token_ids.append(token_id) self.logprobs.append(logprob) self.ranks.append(rank) self.decoded_tokens.append(decoded_token) self.end_indices.append(len(self.logprobs)) def extend(self, logprobs_multi_positions) -> None: """Extends the container with logprobs for the next multiple positions""" for logprobs_one_position in logprobs_multi_positions: self.append(logprobs_one_position) def __len__(self) -> int: """Gets number of positions stored in the container""" return len(self.start_indices) @overload def __getitem__(self, position: int) -> LogprobsOnePosition: ... @overload def __getitem__(self, s: slice, /) -> "FlatLogprobs": ... def __getitem__(self, index: int | slice): """Extracts logprobs of a given position or slice""" if isinstance(index, int): return { self.token_ids[i]: Logprob( logprob=self.logprobs[i], rank=self.ranks[i], decoded_token=self.decoded_tokens[i], ) for i in range(self.start_indices[index], self.end_indices[index]) } elif isinstance(index, slice): min_index = self.start_indices[index][0] max_index = self.end_indices[index][-1] return FlatLogprobs( # Shift updated start_indices and end_indices to # be 0-indexed start_indices=[i - min_index for i in self.start_indices[index]], end_indices=[i - min_index for i in self.end_indices[index]], token_ids=self.token_ids[min_index:max_index], logprobs=self.logprobs[min_index:max_index], ranks=self.ranks[min_index:max_index], decoded_tokens=self.decoded_tokens[min_index:max_index], ) else: raise TypeError(f"Invalid index type: {type(index)}") def __setitem__(self, item, value) -> None: raise TypeError("Cannot set logprobs in FlatLogprobs") def __delitem__(self, item) -> None: raise TypeError("Cannot delete logprobs from FlatLogprobs") def insert(self, item) -> None: raise TypeError("Cannot insert logprobs to FlatLogprobs") def __iter__(self) -> Iterator[LogprobsOnePosition]: """ Iterates the container and yields LogprobsOnePosition for each position. """ for i in range(0, len(self.start_indices)): yield self.__getitem__(i) # {token_id -> logprob} per each sequence group. None if the corresponding # sequence group doesn't require prompt logprob. PromptLogprobs = FlatLogprobs | list[LogprobsOnePosition | None] # {token_id -> logprob} for each sequence group. SampleLogprobs = FlatLogprobs | list[LogprobsOnePosition] def create_prompt_logprobs(flat_logprobs: bool) -> PromptLogprobs: """Creates a container to store prompt logprobs for a request""" logprobs = FlatLogprobs() if flat_logprobs else [] # NOTE: logprob of first prompt token is None. logprobs.append(None) return logprobs def create_sample_logprobs(flat_logprobs: bool) -> SampleLogprobs: """Creates a container to store decode logprobs for a request""" return FlatLogprobs() if flat_logprobs else [] def append_logprobs_for_next_position( request_logprobs: PromptLogprobs | SampleLogprobs, token_ids: list[int], logprobs: list[float], decoded_tokens: Iterable[str | None], rank: int, num_logprobs: int, ) -> None: """Appends logprobs for the next position""" if num_logprobs == -1: num_logprobs = len(logprobs) # We do not need a special case for the sampled token # being in the topk, since inserting duplicated data # into a dictionary twice is the same as doing it once. topk_ranks = range(1, num_logprobs + 1) ranks = itertools.chain((rank,), topk_ranks) if isinstance(request_logprobs, FlatLogprobs): request_logprobs.append_fast(token_ids, logprobs, ranks, decoded_tokens) else: request_logprobs.append( { token_id: Logprob( logprob=logprob, rank=rank, decoded_token=token, ) for token_id, logprob, rank, token in zip( token_ids, logprobs, ranks, decoded_tokens ) } )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/collect_env.py
vllm/collect_env.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # ruff: noqa # code borrowed from https://github.com/pytorch/pytorch/blob/main/torch/utils/collect_env.py import datetime import locale import os import subprocess import sys # Unlike the rest of the PyTorch this file must be python2 compliant. # This script outputs relevant system environment info # Run it with `python collect_env.py` or `python -m torch.utils.collect_env` from collections import namedtuple import regex as re from vllm.envs import environment_variables try: import torch TORCH_AVAILABLE = True except (ImportError, NameError, AttributeError, OSError): TORCH_AVAILABLE = False # System Environment Information SystemEnv = namedtuple( "SystemEnv", [ "torch_version", "is_debug_build", "cuda_compiled_version", "gcc_version", "clang_version", "cmake_version", "os", "libc_version", "python_version", "python_platform", "is_cuda_available", "cuda_runtime_version", "cuda_module_loading", "nvidia_driver_version", "nvidia_gpu_models", "cudnn_version", "pip_version", # 'pip' or 'pip3' "pip_packages", "conda_packages", "hip_compiled_version", "hip_runtime_version", "miopen_runtime_version", "caching_allocator_config", "is_xnnpack_available", "cpu_info", "rocm_version", # vllm specific field "vllm_version", # vllm specific field "vllm_build_flags", # vllm specific field "gpu_topo", # vllm specific field "env_vars", ], ) DEFAULT_CONDA_PATTERNS = { "torch", "numpy", "cudatoolkit", "soumith", "mkl", "magma", "triton", "optree", "nccl", "transformers", "zmq", "nvidia", "pynvml", "flashinfer-python", } DEFAULT_PIP_PATTERNS = { "torch", "numpy", "mypy", "flake8", "triton", "optree", "onnx", "nccl", "transformers", "zmq", "nvidia", "pynvml", "flashinfer-python", } def run(command): """Return (return-code, stdout, stderr).""" shell = True if type(command) is str else False try: p = subprocess.Popen( command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=shell ) raw_output, raw_err = p.communicate() rc = p.returncode if get_platform() == "win32": enc = "oem" else: enc = locale.getpreferredencoding() output = raw_output.decode(enc) if command == "nvidia-smi topo -m": # don't remove the leading whitespace of `nvidia-smi topo -m` # because they are meaningful output = output.rstrip() else: output = output.strip() err = raw_err.decode(enc) return rc, output, err.strip() except FileNotFoundError: cmd_str = command if isinstance(command, str) else command[0] return 127, "", f"Command not found: {cmd_str}" def run_and_read_all(run_lambda, command): """Run command using run_lambda; reads and returns entire output if rc is 0.""" rc, out, _ = run_lambda(command) if rc != 0: return None return out def run_and_parse_first_match(run_lambda, command, regex): """Run command using run_lambda, returns the first regex match if it exists.""" rc, out, _ = run_lambda(command) if rc != 0: return None match = re.search(regex, out) if match is None: return None return match.group(1) def run_and_return_first_line(run_lambda, command): """Run command using run_lambda and returns first line if output is not empty.""" rc, out, _ = run_lambda(command) if rc != 0: return None return out.split("\n")[0] def get_conda_packages(run_lambda, patterns=None): if patterns is None: patterns = DEFAULT_CONDA_PATTERNS conda = os.environ.get("CONDA_EXE", "conda") out = run_and_read_all(run_lambda, [conda, "list"]) if out is None: return out return "\n".join( line for line in out.splitlines() if not line.startswith("#") and any(name in line for name in patterns) ) def get_gcc_version(run_lambda): return run_and_parse_first_match(run_lambda, "gcc --version", r"gcc (.*)") def get_clang_version(run_lambda): return run_and_parse_first_match( run_lambda, "clang --version", r"clang version (.*)" ) def get_cmake_version(run_lambda): return run_and_parse_first_match(run_lambda, "cmake --version", r"cmake (.*)") def get_nvidia_driver_version(run_lambda): if get_platform() == "darwin": cmd = "kextstat | grep -i cuda" return run_and_parse_first_match( run_lambda, cmd, r"com[.]nvidia[.]CUDA [(](.*?)[)]" ) smi = get_nvidia_smi() return run_and_parse_first_match(run_lambda, smi, r"Driver Version: (.*?) ") def get_gpu_info(run_lambda): if get_platform() == "darwin" or ( TORCH_AVAILABLE and hasattr(torch.version, "hip") and torch.version.hip is not None ): if TORCH_AVAILABLE and torch.cuda.is_available(): if torch.version.hip is not None: prop = torch.cuda.get_device_properties(0) if hasattr(prop, "gcnArchName"): gcnArch = " ({})".format(prop.gcnArchName) else: gcnArch = "NoGCNArchNameOnOldPyTorch" else: gcnArch = "" return torch.cuda.get_device_name(None) + gcnArch return None smi = get_nvidia_smi() uuid_regex = re.compile(r" \(UUID: .+?\)") rc, out, _ = run_lambda(smi + " -L") if rc != 0: return None # Anonymize GPUs by removing their UUID return re.sub(uuid_regex, "", out) def get_running_cuda_version(run_lambda): return run_and_parse_first_match(run_lambda, "nvcc --version", r"release .+ V(.*)") def get_cudnn_version(run_lambda): """Return a list of libcudnn.so; it's hard to tell which one is being used.""" if get_platform() == "win32": system_root = os.environ.get("SYSTEMROOT", "C:\\Windows") cuda_path = os.environ.get("CUDA_PATH", "%CUDA_PATH%") where_cmd = os.path.join(system_root, "System32", "where") cudnn_cmd = '{} /R "{}\\bin" cudnn*.dll'.format(where_cmd, cuda_path) elif get_platform() == "darwin": # CUDA libraries and drivers can be found in /usr/local/cuda/. See # https://docs.nvidia.com/cuda/cuda-installation-guide-mac-os-x/index.html#install # https://docs.nvidia.com/deeplearning/sdk/cudnn-install/index.html#installmac # Use CUDNN_LIBRARY when cudnn library is installed elsewhere. cudnn_cmd = "ls /usr/local/cuda/lib/libcudnn*" else: cudnn_cmd = 'ldconfig -p | grep libcudnn | rev | cut -d" " -f1 | rev' rc, out, _ = run_lambda(cudnn_cmd) # find will return 1 if there are permission errors or if not found if len(out) == 0 or (rc != 1 and rc != 0): l = os.environ.get("CUDNN_LIBRARY") if l is not None and os.path.isfile(l): return os.path.realpath(l) return None files_set = set() for fn in out.split("\n"): fn = os.path.realpath(fn) # eliminate symbolic links if os.path.isfile(fn): files_set.add(fn) if not files_set: return None # Alphabetize the result because the order is non-deterministic otherwise files = sorted(files_set) if len(files) == 1: return files[0] result = "\n".join(files) return "Probably one of the following:\n{}".format(result) def get_nvidia_smi(): # Note: nvidia-smi is currently available only on Windows and Linux smi = "nvidia-smi" if get_platform() == "win32": system_root = os.environ.get("SYSTEMROOT", "C:\\Windows") program_files_root = os.environ.get("PROGRAMFILES", "C:\\Program Files") legacy_path = os.path.join( program_files_root, "NVIDIA Corporation", "NVSMI", smi ) new_path = os.path.join(system_root, "System32", smi) smis = [new_path, legacy_path] for candidate_smi in smis: if os.path.exists(candidate_smi): smi = '"{}"'.format(candidate_smi) break return smi def get_rocm_version(run_lambda): """Returns the ROCm version if available, otherwise 'N/A'.""" return run_and_parse_first_match( run_lambda, "hipcc --version", r"HIP version: (\S+)" ) def get_vllm_version(): from vllm import __version__, __version_tuple__ if __version__ == "dev": return "N/A (dev)" version_str = __version_tuple__[-1] if isinstance(version_str, str) and version_str.startswith("g"): # it's a dev build if "." in version_str: # it's a dev build containing local changes git_sha = version_str.split(".")[0][1:] date = version_str.split(".")[-1][1:] return f"{__version__} (git sha: {git_sha}, date: {date})" else: # it's a dev build without local changes git_sha = version_str[1:] # type: ignore return f"{__version__} (git sha: {git_sha})" return __version__ def summarize_vllm_build_flags(): # This could be a static method if the flags are constant, or dynamic if you need to check environment variables, etc. return "CUDA Archs: {}; ROCm: {}".format( os.environ.get("TORCH_CUDA_ARCH_LIST", "Not Set"), "Enabled" if os.environ.get("ROCM_HOME") else "Disabled", ) def get_gpu_topo(run_lambda): output = None if get_platform() == "linux": output = run_and_read_all(run_lambda, "nvidia-smi topo -m") if output is None: output = run_and_read_all(run_lambda, "rocm-smi --showtopo") return output # example outputs of CPU infos # * linux # Architecture: x86_64 # CPU op-mode(s): 32-bit, 64-bit # Address sizes: 46 bits physical, 48 bits virtual # Byte Order: Little Endian # CPU(s): 128 # On-line CPU(s) list: 0-127 # Vendor ID: GenuineIntel # Model name: Intel(R) Xeon(R) Platinum 8375C CPU @ 2.90GHz # CPU family: 6 # Model: 106 # Thread(s) per core: 2 # Core(s) per socket: 32 # Socket(s): 2 # Stepping: 6 # BogoMIPS: 5799.78 # Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr # sse sse2 ss ht syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon rep_good nopl # xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq monitor ssse3 fma cx16 # pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand # hypervisor lahf_lm abm 3dnowprefetch invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced # fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap # avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 # xsaves wbnoinvd ida arat avx512vbmi pku ospke avx512_vbmi2 gfni vaes vpclmulqdq # avx512_vnni avx512_bitalg tme avx512_vpopcntdq rdpid md_clear flush_l1d arch_capabilities # Virtualization features: # Hypervisor vendor: KVM # Virtualization type: full # Caches (sum of all): # L1d: 3 MiB (64 instances) # L1i: 2 MiB (64 instances) # L2: 80 MiB (64 instances) # L3: 108 MiB (2 instances) # NUMA: # NUMA node(s): 2 # NUMA node0 CPU(s): 0-31,64-95 # NUMA node1 CPU(s): 32-63,96-127 # Vulnerabilities: # Itlb multihit: Not affected # L1tf: Not affected # Mds: Not affected # Meltdown: Not affected # Mmio stale data: Vulnerable: Clear CPU buffers attempted, no microcode; SMT Host state unknown # Retbleed: Not affected # Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp # Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization # Spectre v2: Mitigation; Enhanced IBRS, IBPB conditional, RSB filling, PBRSB-eIBRS SW sequence # Srbds: Not affected # Tsx async abort: Not affected # * win32 # Architecture=9 # CurrentClockSpeed=2900 # DeviceID=CPU0 # Family=179 # L2CacheSize=40960 # L2CacheSpeed= # Manufacturer=GenuineIntel # MaxClockSpeed=2900 # Name=Intel(R) Xeon(R) Platinum 8375C CPU @ 2.90GHz # ProcessorType=3 # Revision=27142 # # Architecture=9 # CurrentClockSpeed=2900 # DeviceID=CPU1 # Family=179 # L2CacheSize=40960 # L2CacheSpeed= # Manufacturer=GenuineIntel # MaxClockSpeed=2900 # Name=Intel(R) Xeon(R) Platinum 8375C CPU @ 2.90GHz # ProcessorType=3 # Revision=27142 def get_cpu_info(run_lambda): rc, out, err = 0, "", "" if get_platform() == "linux": rc, out, err = run_lambda("lscpu") elif get_platform() == "win32": rc, out, err = run_lambda( "wmic cpu get Name,Manufacturer,Family,Architecture,ProcessorType,DeviceID, \ CurrentClockSpeed,MaxClockSpeed,L2CacheSize,L2CacheSpeed,Revision /VALUE" ) elif get_platform() == "darwin": rc, out, err = run_lambda("sysctl -n machdep.cpu.brand_string") cpu_info = "None" if rc == 0: cpu_info = out else: cpu_info = err return cpu_info def get_platform(): if sys.platform.startswith("linux"): return "linux" elif sys.platform.startswith("win32"): return "win32" elif sys.platform.startswith("cygwin"): return "cygwin" elif sys.platform.startswith("darwin"): return "darwin" else: return sys.platform def get_mac_version(run_lambda): return run_and_parse_first_match(run_lambda, "sw_vers -productVersion", r"(.*)") def get_windows_version(run_lambda): system_root = os.environ.get("SYSTEMROOT", "C:\\Windows") wmic_cmd = os.path.join(system_root, "System32", "Wbem", "wmic") findstr_cmd = os.path.join(system_root, "System32", "findstr") return run_and_read_all( run_lambda, "{} os get Caption | {} /v Caption".format(wmic_cmd, findstr_cmd) ) def get_lsb_version(run_lambda): return run_and_parse_first_match( run_lambda, "lsb_release -a", r"Description:\t(.*)" ) def check_release_file(run_lambda): return run_and_parse_first_match( run_lambda, "cat /etc/*-release", r'PRETTY_NAME="(.*)"' ) def get_os(run_lambda): from platform import machine platform = get_platform() if platform == "win32" or platform == "cygwin": return get_windows_version(run_lambda) if platform == "darwin": version = get_mac_version(run_lambda) if version is None: return None return "macOS {} ({})".format(version, machine()) if platform == "linux": # Ubuntu/Debian based desc = get_lsb_version(run_lambda) if desc is not None: return "{} ({})".format(desc, machine()) # Try reading /etc/*-release desc = check_release_file(run_lambda) if desc is not None: return "{} ({})".format(desc, machine()) return "{} ({})".format(platform, machine()) # Unknown platform return platform def get_python_platform(): import platform return platform.platform() def get_libc_version(): import platform if get_platform() != "linux": return "N/A" return "-".join(platform.libc_ver()) def is_uv_venv(): if os.environ.get("UV"): return True pyvenv_cfg_path = os.path.join(sys.prefix, "pyvenv.cfg") if os.path.exists(pyvenv_cfg_path): with open(pyvenv_cfg_path, "r") as f: return any(line.startswith("uv = ") for line in f) return False def get_pip_packages(run_lambda, patterns=None): """Return `pip list` output. Note: will also find conda-installed pytorch and numpy packages.""" if patterns is None: patterns = DEFAULT_PIP_PATTERNS def run_with_pip(): try: import importlib.util pip_spec = importlib.util.find_spec("pip") pip_available = pip_spec is not None except ImportError: pip_available = False if pip_available: cmd = [sys.executable, "-mpip", "list", "--format=freeze"] elif is_uv_venv(): print("uv is set") cmd = ["uv", "pip", "list", "--format=freeze"] else: raise RuntimeError( "Could not collect pip list output (pip or uv module not available)" ) out = run_and_read_all(run_lambda, cmd) return "\n".join( line for line in out.splitlines() if any(name in line for name in patterns) ) pip_version = "pip3" if sys.version[0] == "3" else "pip" out = run_with_pip() return pip_version, out def get_cachingallocator_config(): ca_config = os.environ.get("PYTORCH_CUDA_ALLOC_CONF", "") return ca_config def get_cuda_module_loading_config(): if TORCH_AVAILABLE and torch.cuda.is_available(): torch.cuda.init() config = os.environ.get("CUDA_MODULE_LOADING", "") return config else: return "N/A" def is_xnnpack_available(): if TORCH_AVAILABLE: import torch.backends.xnnpack return str(torch.backends.xnnpack.enabled) # type: ignore[attr-defined] else: return "N/A" def get_env_vars(): env_vars = "" secret_terms = ("secret", "token", "api", "access", "password") report_prefix = ( "TORCH", "NCCL", "PYTORCH", "CUDA", "CUBLAS", "CUDNN", "OMP_", "MKL_", "NVIDIA", ) for k, v in os.environ.items(): if any(term in k.lower() for term in secret_terms): continue if k in environment_variables: env_vars = env_vars + "{}={}".format(k, v) + "\n" if k.startswith(report_prefix): env_vars = env_vars + "{}={}".format(k, v) + "\n" return env_vars def get_env_info(): run_lambda = run pip_version, pip_list_output = get_pip_packages(run_lambda) if TORCH_AVAILABLE: version_str = torch.__version__ debug_mode_str = str(torch.version.debug) cuda_available_str = str(torch.cuda.is_available()) cuda_version_str = torch.version.cuda if ( not hasattr(torch.version, "hip") or torch.version.hip is None ): # cuda version hip_compiled_version = hip_runtime_version = miopen_runtime_version = "N/A" else: # HIP version def get_version_or_na(cfg, prefix): _lst = [s.rsplit(None, 1)[-1] for s in cfg if prefix in s] return _lst[0] if _lst else "N/A" cfg = torch._C._show_config().split("\n") hip_runtime_version = get_version_or_na(cfg, "HIP Runtime") miopen_runtime_version = get_version_or_na(cfg, "MIOpen") cuda_version_str = "N/A" hip_compiled_version = torch.version.hip else: version_str = debug_mode_str = cuda_available_str = cuda_version_str = "N/A" hip_compiled_version = hip_runtime_version = miopen_runtime_version = "N/A" sys_version = sys.version.replace("\n", " ") conda_packages = get_conda_packages(run_lambda) rocm_version = get_rocm_version(run_lambda) vllm_version = get_vllm_version() vllm_build_flags = summarize_vllm_build_flags() gpu_topo = get_gpu_topo(run_lambda) return SystemEnv( torch_version=version_str, is_debug_build=debug_mode_str, python_version="{} ({}-bit runtime)".format( sys_version, sys.maxsize.bit_length() + 1 ), python_platform=get_python_platform(), is_cuda_available=cuda_available_str, cuda_compiled_version=cuda_version_str, cuda_runtime_version=get_running_cuda_version(run_lambda), cuda_module_loading=get_cuda_module_loading_config(), nvidia_gpu_models=get_gpu_info(run_lambda), nvidia_driver_version=get_nvidia_driver_version(run_lambda), cudnn_version=get_cudnn_version(run_lambda), hip_compiled_version=hip_compiled_version, hip_runtime_version=hip_runtime_version, miopen_runtime_version=miopen_runtime_version, pip_version=pip_version, pip_packages=pip_list_output, conda_packages=conda_packages, os=get_os(run_lambda), libc_version=get_libc_version(), gcc_version=get_gcc_version(run_lambda), clang_version=get_clang_version(run_lambda), cmake_version=get_cmake_version(run_lambda), caching_allocator_config=get_cachingallocator_config(), is_xnnpack_available=is_xnnpack_available(), cpu_info=get_cpu_info(run_lambda), rocm_version=rocm_version, vllm_version=vllm_version, vllm_build_flags=vllm_build_flags, gpu_topo=gpu_topo, env_vars=get_env_vars(), ) env_info_fmt = """ ============================== System Info ============================== OS : {os} GCC version : {gcc_version} Clang version : {clang_version} CMake version : {cmake_version} Libc version : {libc_version} ============================== PyTorch Info ============================== PyTorch version : {torch_version} Is debug build : {is_debug_build} CUDA used to build PyTorch : {cuda_compiled_version} ROCM used to build PyTorch : {hip_compiled_version} ============================== Python Environment ============================== Python version : {python_version} Python platform : {python_platform} ============================== CUDA / GPU Info ============================== Is CUDA available : {is_cuda_available} CUDA runtime version : {cuda_runtime_version} CUDA_MODULE_LOADING set to : {cuda_module_loading} GPU models and configuration : {nvidia_gpu_models} Nvidia driver version : {nvidia_driver_version} cuDNN version : {cudnn_version} HIP runtime version : {hip_runtime_version} MIOpen runtime version : {miopen_runtime_version} Is XNNPACK available : {is_xnnpack_available} ============================== CPU Info ============================== {cpu_info} ============================== Versions of relevant libraries ============================== {pip_packages} {conda_packages} """.strip() # both the above code and the following code use `strip()` to # remove leading/trailing whitespaces, so we need to add a newline # in between to separate the two sections env_info_fmt += "\n\n" env_info_fmt += """ ============================== vLLM Info ============================== ROCM Version : {rocm_version} vLLM Version : {vllm_version} vLLM Build Flags: {vllm_build_flags} GPU Topology: {gpu_topo} ============================== Environment Variables ============================== {env_vars} """.strip() def pretty_str(envinfo): def replace_nones(dct, replacement="Could not collect"): for key in dct.keys(): if dct[key] is not None: continue dct[key] = replacement return dct def replace_bools(dct, true="Yes", false="No"): for key in dct.keys(): if dct[key] is True: dct[key] = true elif dct[key] is False: dct[key] = false return dct def prepend(text, tag="[prepend]"): lines = text.split("\n") updated_lines = [tag + line for line in lines] return "\n".join(updated_lines) def replace_if_empty(text, replacement="No relevant packages"): if text is not None and len(text) == 0: return replacement return text def maybe_start_on_next_line(string): # If `string` is multiline, prepend a \n to it. if string is not None and len(string.split("\n")) > 1: return "\n{}\n".format(string) return string mutable_dict = envinfo._asdict() # If nvidia_gpu_models is multiline, start on the next line mutable_dict["nvidia_gpu_models"] = maybe_start_on_next_line( envinfo.nvidia_gpu_models ) # If the machine doesn't have CUDA, report some fields as 'No CUDA' dynamic_cuda_fields = [ "cuda_runtime_version", "nvidia_gpu_models", "nvidia_driver_version", ] all_cuda_fields = dynamic_cuda_fields + ["cudnn_version"] all_dynamic_cuda_fields_missing = all( mutable_dict[field] is None for field in dynamic_cuda_fields ) if ( TORCH_AVAILABLE and not torch.cuda.is_available() and all_dynamic_cuda_fields_missing ): for field in all_cuda_fields: mutable_dict[field] = "No CUDA" if envinfo.cuda_compiled_version is None: mutable_dict["cuda_compiled_version"] = "None" # Replace True with Yes, False with No mutable_dict = replace_bools(mutable_dict) # Replace all None objects with 'Could not collect' mutable_dict = replace_nones(mutable_dict) # If either of these are '', replace with 'No relevant packages' mutable_dict["pip_packages"] = replace_if_empty(mutable_dict["pip_packages"]) mutable_dict["conda_packages"] = replace_if_empty(mutable_dict["conda_packages"]) # Tag conda and pip packages with a prefix # If they were previously None, they'll show up as ie '[conda] Could not collect' if mutable_dict["pip_packages"]: mutable_dict["pip_packages"] = prepend( mutable_dict["pip_packages"], "[{}] ".format(envinfo.pip_version) ) if mutable_dict["conda_packages"]: mutable_dict["conda_packages"] = prepend( mutable_dict["conda_packages"], "[conda] " ) mutable_dict["cpu_info"] = envinfo.cpu_info return env_info_fmt.format(**mutable_dict) def get_pretty_env_info(): return pretty_str(get_env_info()) def main(): print("Collecting environment information...") output = get_pretty_env_info() print(output) if ( TORCH_AVAILABLE and hasattr(torch, "utils") and hasattr(torch.utils, "_crash_handler") ): minidump_dir = torch.utils._crash_handler.DEFAULT_MINIDUMP_DIR if sys.platform == "linux" and os.path.exists(minidump_dir): dumps = [ os.path.join(minidump_dir, dump) for dump in os.listdir(minidump_dir) ] latest = max(dumps, key=os.path.getctime) ctime = os.path.getctime(latest) creation_time = datetime.datetime.fromtimestamp(ctime).strftime( "%Y-%m-%d %H:%M:%S" ) msg = ( "\n*** Detected a minidump at {} created on {}, ".format( latest, creation_time ) + "if this is related to your bug please include it when you file a report ***" ) print(msg, file=sys.stderr) if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/scalar_type.py
vllm/scalar_type.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools import struct from dataclasses import dataclass from enum import Enum _SCALAR_TYPES_ID_MAP = {} # Mirrors enum in `core/scalar_type.hpp` class NanRepr(Enum): NONE = 0 # nans are not supported IEEE_754 = 1 # nans are: Exp all 1s, mantissa not all 0s EXTD_RANGE_MAX_MIN = 2 # nans are: Exp all 1s, mantissa all 1s # This ScalarType class is a parallel implementation of the C++ ScalarType # class found in csrc/core/scalar_type.hpp. These two classes should be kept # in sync until the inductor fully supports custom C++ classes. @dataclass(frozen=True) class ScalarType: """ ScalarType can represent a wide range of floating point and integer types, in particular it can be used to represent sub-byte data types (something that torch.dtype currently does not support). It is also capable of representing types with a bias, i.e.: `stored_value = value + bias`, this is useful for quantized types (e.g. standard GPTQ 4bit uses a bias of 8). The implementation for this class can be found in csrc/core/scalar_type.hpp, these type signatures should be kept in sync with that file. """ exponent: int """ Number of bits in the exponent if this is a floating point type (zero if this an integer type) """ mantissa: int """ Number of bits in the mantissa if this is a floating point type, or the number bits representing an integer excluding the sign bit if this an integer type. """ signed: bool "If the type is signed (i.e. has a sign bit)" bias: int """ bias used to encode the values in this scalar type (value = stored_value - bias, default 0) for example if we store the type as an unsigned integer with a bias of 128 then the value 0 will be stored as 128 and -1 will be stored as 127 and 1 will be stored as 129. """ _finite_values_only: bool = False """ Private: if infs are supported, used `has_infs()` instead. """ nan_repr: NanRepr = NanRepr.IEEE_754 """ How NaNs are represent in this scalar type, returns NanRepr value. (not applicable for integer types) """ def _floating_point_max_int(self) -> int: assert self.mantissa <= 52 and self.exponent <= 11, ( f"Cannot represent max/min as a double for type {self.__str__()}" ) max_mantissa = (1 << self.mantissa) - 1 if self.nan_repr == NanRepr.EXTD_RANGE_MAX_MIN: max_mantissa = max_mantissa - 1 max_exponent = (1 << self.exponent) - 2 if self.nan_repr == NanRepr.EXTD_RANGE_MAX_MIN or self.nan_repr == NanRepr.NONE: assert self.exponent < 11, ( f"Cannot represent max/min as a double for type {self.__str__()}" ) max_exponent = max_exponent + 1 # adjust the exponent to match that of a double # for now we assume the exponent bias is the standard 2^(e-1) -1, (where # e is the exponent bits), there is some precedent for non-standard # biases, example `float8_e4m3b11fnuz` here: # https://github.com/jax-ml/ml_dtypes but to avoid premature over # complication we are just assuming the standard exponent bias until # there is a need to support non-standard biases exponent_bias = (1 << (self.exponent - 1)) - 1 exponent_bias_double = (1 << 10) - 1 # double e = 11 max_exponent_double = max_exponent - exponent_bias + exponent_bias_double # shift the mantissa and exponent into the proper positions for an # IEEE double and bitwise-or them together. return (max_mantissa << (52 - self.mantissa)) | (max_exponent_double << 52) def _floating_point_max(self) -> float: double_raw = self._floating_point_max_int() return struct.unpack("!d", struct.pack("!Q", double_raw))[0] def _raw_max(self) -> int | float: if self.is_floating_point(): return self._floating_point_max() else: assert self.size_bits < 64 or self.size_bits == 64 and self.is_signed(), ( "Cannot represent max as an int" ) return (1 << self.mantissa) - 1 def _raw_min(self) -> int | float: if self.is_floating_point(): assert self.is_signed(), ( "We currently assume all floating point types are signed" ) sign_bit_double = 1 << 63 max_raw = self._floating_point_max_int() min_raw = max_raw | sign_bit_double return struct.unpack("!d", struct.pack("!Q", min_raw))[0] else: assert not self.is_signed() or self.size_bits <= 64, ( "Cannot represent min as a int64_t" ) if self.is_signed(): return -(1 << (self.size_bits - 1)) else: return 0 @functools.cached_property def id(self) -> int: """ Convert the ScalarType to an int which can be passed to pytorch custom ops. This layout of the int must be kept in sync with the C++ ScalarType's from_id method. """ val = 0 offset = 0 def or_and_advance(member, bit_width): nonlocal val nonlocal offset bit_mask = (1 << bit_width) - 1 val = val | (int(member) & bit_mask) << offset offset = offset + bit_width or_and_advance(self.exponent, 8) or_and_advance(self.mantissa, 8) or_and_advance(self.signed, 1) or_and_advance(self.bias, 32) or_and_advance(self._finite_values_only, 1) or_and_advance(self.nan_repr.value, 8) assert offset <= 64, f"ScalarType fields too big {offset} to fit into an int64" _SCALAR_TYPES_ID_MAP[val] = self return val @property def size_bits(self) -> int: return self.exponent + self.mantissa + int(self.signed) def min(self) -> int | float: """ Min representable value for this scalar type. (accounting for bias if there is one) """ return self._raw_min() - self.bias def max(self) -> int | float: """ Max representable value for this scalar type. (accounting for bias if there is one) """ return self._raw_max() - self.bias def is_signed(self) -> bool: """ If the type is signed (i.e. has a sign bit), same as `signed` added for consistency with: https://pytorch.org/docs/stable/generated/torch.Tensor.is_signed.html """ return self.signed def is_floating_point(self) -> bool: "If the type is a floating point type" return self.exponent != 0 def is_integer(self) -> bool: "If the type is an integer type" return self.exponent == 0 def has_bias(self) -> bool: "If the type has a non-zero bias" return self.bias != 0 def has_infs(self) -> bool: "If the type is floating point and supports infinity" return not self._finite_values_only def has_nans(self) -> bool: return self.nan_repr != NanRepr.NONE.value def is_ieee_754(self) -> bool: """ If the type is a floating point type that follows IEEE 754 conventions """ return self.nan_repr == NanRepr.IEEE_754.value and not self._finite_values_only def __str__(self) -> str: """ naming generally follows: https://github.com/jax-ml/ml_dtypes for floating point types (leading f) the scheme is: `float<size_bits>_e<exponent_bits>m<mantissa_bits>[flags]` flags: - no-flags: means it follows IEEE 754 conventions - f: means finite values only (no infinities) - n: means nans are supported (non-standard encoding) for integer types the scheme is: `[u]int<size_bits>[b<bias>]` - if bias is not present it means its zero """ if self.is_floating_point(): ret = ( "float" + str(self.size_bits) + "_e" + str(self.exponent) + "m" + str(self.mantissa) ) if not self.is_ieee_754(): if self._finite_values_only: ret = ret + "f" if self.nan_repr != NanRepr.NONE: ret = ret + "n" return ret else: ret = ("int" if self.is_signed() else "uint") + str(self.size_bits) if self.has_bias(): ret = ret + "b" + str(self.bias) return ret def __repr__(self) -> str: return "ScalarType." + self.__str__() # __len__ needs to be defined (and has to throw TypeError) for pytorch's # opcheck to work. def __len__(self) -> int: raise TypeError # # Convenience Constructors # @classmethod def int_(cls, size_bits: int, bias: int | None) -> "ScalarType": "Create a signed integer scalar type (size_bits includes sign-bit)." ret = cls(0, size_bits - 1, True, bias if bias else 0) ret.id # noqa B018: make sure the id is cached return ret @classmethod def uint(cls, size_bits: int, bias: int | None) -> "ScalarType": """Create an unsigned integer scalar type.""" ret = cls(0, size_bits, False, bias if bias else 0) ret.id # noqa B018: make sure the id is cached return ret @classmethod def float_IEEE754(cls, exponent: int, mantissa: int) -> "ScalarType": """ Create a standard floating point type (i.e. follows IEEE 754 conventions). """ assert mantissa > 0 and exponent > 0 ret = cls(exponent, mantissa, True, 0) ret.id # noqa B018: make sure the id is cached return ret @classmethod def float_( cls, exponent: int, mantissa: int, finite_values_only: bool, nan_repr: NanRepr ) -> "ScalarType": """ Create a non-standard floating point type (i.e. does not follow IEEE 754 conventions). """ assert mantissa > 0 and exponent > 0 assert nan_repr != NanRepr.IEEE_754, ( "use `float_IEEE754` constructor for floating point types that " "follow IEEE 754 conventions" ) ret = cls(exponent, mantissa, True, 0, finite_values_only, nan_repr) ret.id # noqa B018: make sure the id is cached return ret @classmethod def from_id(cls, scalar_type_id: int): if scalar_type_id not in _SCALAR_TYPES_ID_MAP: raise ValueError(f"scalar_type_id {scalar_type_id} doesn't exists.") return _SCALAR_TYPES_ID_MAP[scalar_type_id] # naming generally follows: https://github.com/jax-ml/ml_dtypes # for floating point types (leading f) the scheme is: # `float<size_bits>_e<exponent_bits>m<mantissa_bits>[flags]` # flags: # - no-flags: means it follows IEEE 754 conventions # - f: means finite values only (no infinities) # - n: means nans are supported (non-standard encoding) # for integer types the scheme is: # `[u]int<size_bits>[b<bias>]` # - if bias is not present it means its zero class scalar_types: int4 = ScalarType.int_(4, None) uint4 = ScalarType.uint(4, None) int8 = ScalarType.int_(8, None) uint8 = ScalarType.uint(8, None) float8_e4m3fn = ScalarType.float_(4, 3, True, NanRepr.EXTD_RANGE_MAX_MIN) float8_e5m2 = ScalarType.float_IEEE754(5, 2) float8_e8m0fnu = ScalarType(8, 0, False, 0, True, NanRepr.EXTD_RANGE_MAX_MIN) float16_e8m7 = ScalarType.float_IEEE754(8, 7) float16_e5m10 = ScalarType.float_IEEE754(5, 10) # fp6, https://github.com/usyd-fsalab/fp6_llm/tree/main # and https://www.opencompute.org/documents/ocp-microscaling-formats-mx-v1-0-spec-final-pdf float6_e3m2f = ScalarType.float_(3, 2, True, NanRepr.NONE) float6_e2m3f = ScalarType.float_(2, 3, True, NanRepr.NONE) # fp4, https://www.opencompute.org/documents/ocp-microscaling-formats-mx-v1-0-spec-final-pdf float4_e2m1f = ScalarType.float_(2, 1, True, NanRepr.NONE) # "gptq" types uint2b2 = ScalarType.uint(2, 2) uint3b4 = ScalarType.uint(3, 4) uint4b8 = ScalarType.uint(4, 8) uint8b128 = ScalarType.uint(8, 128) # colloquial names bfloat16 = float16_e8m7 float16 = float16_e5m10
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/__init__.py
vllm/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """vLLM: a high-throughput and memory-efficient inference engine for LLMs""" # The version.py should be independent library, and we always import the # version library first. Such assumption is critical for some customization. from .version import __version__, __version_tuple__ # isort:skip import typing # The environment variables override should be imported before any other # modules to ensure that the environment variables are set before any # other modules are imported. import vllm.env_override # noqa: F401 MODULE_ATTRS = { "bc_linter_skip": "._bc_linter:bc_linter_skip", "bc_linter_include": "._bc_linter:bc_linter_include", "AsyncEngineArgs": ".engine.arg_utils:AsyncEngineArgs", "EngineArgs": ".engine.arg_utils:EngineArgs", "AsyncLLMEngine": ".engine.async_llm_engine:AsyncLLMEngine", "LLMEngine": ".engine.llm_engine:LLMEngine", "LLM": ".entrypoints.llm:LLM", "initialize_ray_cluster": ".v1.executor.ray_utils:initialize_ray_cluster", "PromptType": ".inputs:PromptType", "TextPrompt": ".inputs:TextPrompt", "TokensPrompt": ".inputs:TokensPrompt", "ModelRegistry": ".model_executor.models:ModelRegistry", "SamplingParams": ".sampling_params:SamplingParams", "PoolingParams": ".pooling_params:PoolingParams", "ClassificationOutput": ".outputs:ClassificationOutput", "ClassificationRequestOutput": ".outputs:ClassificationRequestOutput", "CompletionOutput": ".outputs:CompletionOutput", "EmbeddingOutput": ".outputs:EmbeddingOutput", "EmbeddingRequestOutput": ".outputs:EmbeddingRequestOutput", "PoolingOutput": ".outputs:PoolingOutput", "PoolingRequestOutput": ".outputs:PoolingRequestOutput", "RequestOutput": ".outputs:RequestOutput", "ScoringOutput": ".outputs:ScoringOutput", "ScoringRequestOutput": ".outputs:ScoringRequestOutput", } if typing.TYPE_CHECKING: from vllm.engine.arg_utils import AsyncEngineArgs, EngineArgs from vllm.engine.async_llm_engine import AsyncLLMEngine from vllm.engine.llm_engine import LLMEngine from vllm.entrypoints.llm import LLM from vllm.inputs import PromptType, TextPrompt, TokensPrompt from vllm.model_executor.models import ModelRegistry from vllm.outputs import ( ClassificationOutput, ClassificationRequestOutput, CompletionOutput, EmbeddingOutput, EmbeddingRequestOutput, PoolingOutput, PoolingRequestOutput, RequestOutput, ScoringOutput, ScoringRequestOutput, ) from vllm.pooling_params import PoolingParams from vllm.sampling_params import SamplingParams from vllm.v1.executor.ray_utils import initialize_ray_cluster from ._bc_linter import bc_linter_include, bc_linter_skip else: def __getattr__(name: str) -> typing.Any: from importlib import import_module if name in MODULE_ATTRS: module_name, attr_name = MODULE_ATTRS[name].split(":") module = import_module(module_name, __package__) return getattr(module, attr_name) else: raise AttributeError(f"module {__package__} has no attribute {name}") __all__ = [ "__version__", "bc_linter_skip", "bc_linter_include", "__version_tuple__", "LLM", "ModelRegistry", "PromptType", "TextPrompt", "TokensPrompt", "SamplingParams", "RequestOutput", "CompletionOutput", "PoolingOutput", "PoolingRequestOutput", "EmbeddingOutput", "EmbeddingRequestOutput", "ClassificationOutput", "ClassificationRequestOutput", "ScoringOutput", "ScoringRequestOutput", "LLMEngine", "EngineArgs", "AsyncLLMEngine", "AsyncEngineArgs", "initialize_ray_cluster", "PoolingParams", ]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/outputs.py
vllm/outputs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import MutableSequence from collections.abc import Sequence as GenericSequence from dataclasses import dataclass from typing import Any, Generic import torch from typing_extensions import TypeVar from vllm.logger import init_logger from vllm.logprobs import PromptLogprobs, SampleLogprobs from vllm.lora.request import LoRARequest from vllm.multimodal.inputs import MultiModalPlaceholderDict from vllm.sequence import RequestMetrics from vllm.v1.metrics.stats import RequestStateStats logger = init_logger(__name__) @dataclass class CompletionOutput: """The output data of one completion output of a request. Args: index: The index of the output in the request. text: The generated output text. token_ids: The token IDs of the generated output text. cumulative_logprob: The cumulative log probability of the generated output text. logprobs: The log probabilities of the top probability words at each position if the logprobs are requested. finish_reason: The reason why the sequence is finished. stop_reason: The stop string or token id that caused the completion to stop, None if the completion finished for some other reason including encountering the EOS token. lora_request: The LoRA request that was used to generate the output. """ index: int text: str token_ids: GenericSequence[int] cumulative_logprob: float | None logprobs: SampleLogprobs | None finish_reason: str | None = None stop_reason: int | str | None = None lora_request: LoRARequest | None = None def finished(self) -> bool: return self.finish_reason is not None def __repr__(self) -> str: return ( f"CompletionOutput(index={self.index}, " f"text={self.text!r}, " f"token_ids={self.token_ids}, " f"cumulative_logprob={self.cumulative_logprob}, " f"logprobs={self.logprobs}, " f"finish_reason={self.finish_reason}, " f"stop_reason={self.stop_reason})" ) @dataclass class PoolingOutput: """The output data of one pooling output of a request. Args: data: The extracted hidden states. """ data: torch.Tensor def __repr__(self) -> str: return f"PoolingOutput(data={self.data})" def __eq__(self, other: object) -> bool: return isinstance(other, self.__class__) and bool( (self.data == other.data).all() ) class RequestOutput: """The output data of a completion request to the LLM. Args: request_id: The unique ID of the request. prompt: The prompt string of the request. For encoder/decoder models, this is the decoder input prompt. prompt_token_ids: The token IDs of the prompt. For encoder/decoder models, this is the decoder input prompt token ids. prompt_logprobs: The log probabilities to return per prompt token. outputs: The output sequences of the request. finished: Whether the whole request is finished. metrics: Metrics associated with the request. lora_request: The LoRA request that was used to generate the output. encoder_prompt: The encoder prompt string of the request. None if decoder-only. encoder_prompt_token_ids: The token IDs of the encoder prompt. None if decoder-only. num_cached_tokens: The number of tokens with prefix cache hit. kv_transfer_params: The params for remote K/V transfer. """ def __init__( self, request_id: str, prompt: str | None, prompt_token_ids: list[int] | None, prompt_logprobs: PromptLogprobs | None, outputs: list[CompletionOutput], finished: bool, metrics: RequestMetrics | RequestStateStats | None = None, lora_request: LoRARequest | None = None, encoder_prompt: str | None = None, encoder_prompt_token_ids: list[int] | None = None, num_cached_tokens: int | None = None, *, multi_modal_placeholders: MultiModalPlaceholderDict | None = None, kv_transfer_params: dict[str, Any] | None = None, # Forward compatibility, code that uses args added in new release can # still run with older versions of vLLM without breaking. **kwargs: Any, ) -> None: if kwargs: logger.warning_once( "RequestOutput: Ignoring extra arguments: %s", str(kwargs) ) self.request_id = request_id self.prompt = prompt self.prompt_token_ids = prompt_token_ids self.multi_modal_placeholders = multi_modal_placeholders or {} self.prompt_logprobs = prompt_logprobs self.outputs = outputs self.finished = finished self.metrics = metrics self.lora_request = lora_request self.encoder_prompt = encoder_prompt self.encoder_prompt_token_ids = encoder_prompt_token_ids self.num_cached_tokens = num_cached_tokens self.kv_transfer_params = kv_transfer_params def add(self, next_output: "RequestOutput", aggregate: bool) -> None: """Merge subsequent RequestOutput into this one""" self.finished |= next_output.finished self.kv_transfer_params = next_output.kv_transfer_params for next_completion in next_output.outputs: for i, completion in enumerate(self.outputs): if completion.index == next_completion.index: if aggregate: # Merge outputs with same index completion.text += next_completion.text if not isinstance(completion.token_ids, MutableSequence): completion.token_ids = list(completion.token_ids) completion.token_ids.extend(next_completion.token_ids) if next_completion.logprobs: assert completion.logprobs is not None completion.logprobs.extend(next_completion.logprobs) completion.cumulative_logprob = ( next_completion.cumulative_logprob ) completion.finish_reason = next_completion.finish_reason completion.stop_reason = next_completion.stop_reason else: # Replace the output with the new one self.outputs[i] = next_completion break else: self.outputs.append(next_completion) def __repr__(self) -> str: return ( f"RequestOutput(request_id={self.request_id}, " f"prompt={self.prompt!r}, " f"prompt_token_ids={self.prompt_token_ids}, " f"encoder_prompt={self.encoder_prompt!r}, " f"encoder_prompt_token_ids={self.encoder_prompt_token_ids}, " f"prompt_logprobs={self.prompt_logprobs}, " f"outputs={self.outputs}, " f"finished={self.finished}, " f"metrics={self.metrics}, " f"lora_request={self.lora_request}, " f"num_cached_tokens={self.num_cached_tokens}, " f"multi_modal_placeholders={self.multi_modal_placeholders})" ) _O = TypeVar("_O", default=PoolingOutput) class PoolingRequestOutput(Generic[_O]): """ The output data of a pooling request to the LLM. Args: request_id (str): A unique identifier for the pooling request. outputs (PoolingOutput): The pooling results for the given input. prompt_token_ids (list[int]): A list of token IDs used in the prompt. num_cached_tokens: The number of tokens with prefix cache hit. finished (bool): A flag indicating whether the pooling is completed. """ def __init__( self, request_id: str, outputs: _O, prompt_token_ids: list[int], num_cached_tokens: int, finished: bool, ): self.request_id = request_id self.prompt_token_ids = prompt_token_ids self.num_cached_tokens = num_cached_tokens self.finished = finished self.outputs = outputs def __repr__(self): return ( f"{type(self).__name__}(request_id={self.request_id!r}, " f"outputs={self.outputs!r}, " f"prompt_token_ids={self.prompt_token_ids}, " f"num_cached_tokens={self.num_cached_tokens}, " f"finished={self.finished})" ) @dataclass class EmbeddingOutput: """The output data of one embedding output of a request. Args: embedding: The embedding vector, which is a list of floats. Its length depends on the hidden dimension of the model. """ embedding: list[float] @staticmethod def from_base(pooling_output: PoolingOutput): pooled_data = pooling_output.data if pooled_data.ndim != 1: raise ValueError("pooled_data should be a 1-D embedding vector") return EmbeddingOutput(pooled_data.tolist()) @property def hidden_size(self) -> int: return len(self.embedding) def __repr__(self) -> str: return f"EmbeddingOutput(hidden_size={self.hidden_size})" class EmbeddingRequestOutput(PoolingRequestOutput[EmbeddingOutput]): @staticmethod def from_base(request_output: PoolingRequestOutput): return EmbeddingRequestOutput( request_id=request_output.request_id, outputs=EmbeddingOutput.from_base(request_output.outputs), prompt_token_ids=request_output.prompt_token_ids, num_cached_tokens=request_output.num_cached_tokens, finished=request_output.finished, ) @dataclass class ClassificationOutput: """The output data of one classification output of a request. Args: probs: The probability vector, which is a list of floats. Its length depends on the number of classes. """ probs: list[float] @staticmethod def from_base(pooling_output: PoolingOutput): # pooling_output shape: (num_classes) pooled_data = pooling_output.data if pooled_data.ndim != 1: raise ValueError("pooled_data should be a 1-D probability vector") return ClassificationOutput(pooled_data.tolist()) @property def num_classes(self) -> int: return len(self.probs) def __repr__(self) -> str: return f"ClassificationOutput(num_classes={self.num_classes})" class ClassificationRequestOutput(PoolingRequestOutput[ClassificationOutput]): @staticmethod def from_base(request_output: PoolingRequestOutput): return ClassificationRequestOutput( request_id=request_output.request_id, outputs=ClassificationOutput.from_base(request_output.outputs), prompt_token_ids=request_output.prompt_token_ids, num_cached_tokens=request_output.num_cached_tokens, finished=request_output.finished, ) @dataclass class ScoringOutput: """The output data of one scoring output of a request. Args: score: The similarity score, which is a scalar value. """ score: float @staticmethod def from_base(pooling_output: PoolingOutput): # pooling_output shape: # classify task: (num_classes) num_classes == 1 # embed task: a scalar value pooled_data = pooling_output.data.squeeze() if pooled_data.ndim != 0: raise ValueError("pooled_data should be a scalar score") return ScoringOutput(pooled_data.item()) def __repr__(self) -> str: return f"ScoringOutput(score={self.score})" class ScoringRequestOutput(PoolingRequestOutput[ScoringOutput]): @staticmethod def from_base(request_output: PoolingRequestOutput): return ScoringRequestOutput( request_id=request_output.request_id, outputs=ScoringOutput.from_base(request_output.outputs), prompt_token_ids=request_output.prompt_token_ids, num_cached_tokens=request_output.num_cached_tokens, finished=request_output.finished, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/envs.py
vllm/envs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools import json import logging import os import sys import tempfile from collections.abc import Callable from typing import TYPE_CHECKING, Any, Literal if TYPE_CHECKING: VLLM_HOST_IP: str = "" VLLM_PORT: int | None = None VLLM_RPC_BASE_PATH: str = tempfile.gettempdir() VLLM_USE_MODELSCOPE: bool = False VLLM_RINGBUFFER_WARNING_INTERVAL: int = 60 VLLM_NCCL_SO_PATH: str | None = None LD_LIBRARY_PATH: str | None = None VLLM_ROCM_SLEEP_MEM_CHUNK_SIZE: int = 256 VLLM_V1_USE_PREFILL_DECODE_ATTENTION: bool = False VLLM_FLASH_ATTN_VERSION: int | None = None LOCAL_RANK: int = 0 CUDA_VISIBLE_DEVICES: str | None = None VLLM_ENGINE_ITERATION_TIMEOUT_S: int = 60 VLLM_ENGINE_READY_TIMEOUT_S: int = 600 VLLM_API_KEY: str | None = None VLLM_DEBUG_LOG_API_SERVER_RESPONSE: bool = False S3_ACCESS_KEY_ID: str | None = None S3_SECRET_ACCESS_KEY: str | None = None S3_ENDPOINT_URL: str | None = None VLLM_MODEL_REDIRECT_PATH: str | None = None VLLM_CACHE_ROOT: str = os.path.expanduser("~/.cache/vllm") VLLM_CONFIG_ROOT: str = os.path.expanduser("~/.config/vllm") VLLM_USAGE_STATS_SERVER: str = "https://stats.vllm.ai" VLLM_NO_USAGE_STATS: bool = False VLLM_DISABLE_FLASHINFER_PREFILL: bool = False VLLM_DO_NOT_TRACK: bool = False VLLM_USAGE_SOURCE: str = "" VLLM_CONFIGURE_LOGGING: bool = True VLLM_LOGGING_LEVEL: str = "INFO" VLLM_LOGGING_PREFIX: str = "" VLLM_LOGGING_STREAM: str = "ext://sys.stdout" VLLM_LOGGING_CONFIG_PATH: str | None = None VLLM_LOGGING_COLOR: str = "auto" NO_COLOR: bool = False VLLM_LOG_STATS_INTERVAL: float = 10.0 VLLM_TRACE_FUNCTION: int = 0 VLLM_ATTENTION_BACKEND: str | None = None VLLM_USE_FLASHINFER_SAMPLER: bool | None = None VLLM_PP_LAYER_PARTITION: str | None = None VLLM_CPU_KVCACHE_SPACE: int | None = 0 VLLM_CPU_OMP_THREADS_BIND: str = "" VLLM_CPU_NUM_OF_RESERVED_CPU: int | None = None VLLM_CPU_SGL_KERNEL: bool = False VLLM_XLA_CACHE_PATH: str = os.path.join(VLLM_CACHE_ROOT, "xla_cache") VLLM_XLA_CHECK_RECOMPILATION: bool = False VLLM_FUSED_MOE_CHUNK_SIZE: int = 16 * 1024 VLLM_ENABLE_FUSED_MOE_ACTIVATION_CHUNKING: bool = True VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE: Literal["auto", "nccl", "shm"] = "auto" VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM: bool = False VLLM_USE_RAY_WRAPPED_PP_COMM: bool = True VLLM_XLA_USE_SPMD: bool = False VLLM_WORKER_MULTIPROC_METHOD: Literal["fork", "spawn"] = "fork" VLLM_ASSETS_CACHE: str = os.path.join(VLLM_CACHE_ROOT, "assets") VLLM_ASSETS_CACHE_MODEL_CLEAN: bool = False VLLM_IMAGE_FETCH_TIMEOUT: int = 5 VLLM_VIDEO_FETCH_TIMEOUT: int = 30 VLLM_AUDIO_FETCH_TIMEOUT: int = 10 VLLM_MEDIA_URL_ALLOW_REDIRECTS: bool = True VLLM_MEDIA_LOADING_THREAD_COUNT: int = 8 VLLM_MAX_AUDIO_CLIP_FILESIZE_MB: int = 25 VLLM_VIDEO_LOADER_BACKEND: str = "opencv" VLLM_MEDIA_CONNECTOR: str = "http" VLLM_TARGET_DEVICE: str = "cuda" VLLM_MAIN_CUDA_VERSION: str = "12.9" VLLM_FLOAT32_MATMUL_PRECISION: Literal["ieee", "tf32"] = "ieee" MAX_JOBS: str | None = None NVCC_THREADS: str | None = None VLLM_USE_PRECOMPILED: bool = False VLLM_SKIP_PRECOMPILED_VERSION_SUFFIX: bool = False VLLM_DOCKER_BUILD_CONTEXT: bool = False VLLM_KEEP_ALIVE_ON_ENGINE_DEATH: bool = False CMAKE_BUILD_TYPE: Literal["Debug", "Release", "RelWithDebInfo"] | None = None VERBOSE: bool = False VLLM_ALLOW_LONG_MAX_MODEL_LEN: bool = False VLLM_RPC_TIMEOUT: int = 10000 # ms VLLM_HTTP_TIMEOUT_KEEP_ALIVE: int = 5 # seconds VLLM_PLUGINS: list[str] | None = None VLLM_LORA_RESOLVER_CACHE_DIR: str | None = None # Deprecated env variables for profiling, kept for backward compatibility # See also vllm/config/profiler.py and `--profiler-config` argument VLLM_TORCH_CUDA_PROFILE: str | None = None VLLM_TORCH_PROFILER_DIR: str | None = None VLLM_TORCH_PROFILER_RECORD_SHAPES: str | None = None VLLM_TORCH_PROFILER_WITH_PROFILE_MEMORY: str | None = None VLLM_TORCH_PROFILER_DISABLE_ASYNC_LLM: str | None = None VLLM_TORCH_PROFILER_WITH_STACK: str | None = None VLLM_TORCH_PROFILER_WITH_FLOPS: str | None = None VLLM_TORCH_PROFILER_USE_GZIP: str | None = None VLLM_TORCH_PROFILER_DUMP_CUDA_TIME_TOTAL: str | None = None VLLM_PROFILER_DELAY_ITERS: str | None = None VLLM_PROFILER_MAX_ITERS: str | None = None # End of deprecated env variables for profiling VLLM_USE_AOT_COMPILE: bool = False VLLM_USE_BYTECODE_HOOK: bool = False VLLM_FORCE_AOT_LOAD: bool = False VLLM_USE_TRITON_AWQ: bool = False VLLM_ALLOW_RUNTIME_LORA_UPDATING: bool = False VLLM_SKIP_P2P_CHECK: bool = False VLLM_DISABLED_KERNELS: list[str] = [] VLLM_DISABLE_PYNCCL: bool = False VLLM_ROCM_USE_AITER: bool = False VLLM_ROCM_USE_AITER_PAGED_ATTN: bool = False VLLM_ROCM_USE_AITER_LINEAR: bool = True VLLM_ROCM_USE_AITER_MOE: bool = True VLLM_ROCM_USE_AITER_RMSNORM: bool = True VLLM_ROCM_USE_AITER_MLA: bool = True VLLM_ROCM_USE_AITER_MHA: bool = True VLLM_ROCM_USE_AITER_FP4_ASM_GEMM: bool = False VLLM_ROCM_USE_AITER_TRITON_ROPE: bool = False VLLM_ROCM_USE_AITER_FP8BMM: bool = True VLLM_ROCM_USE_AITER_UNIFIED_ATTENTION: bool = False VLLM_ROCM_USE_AITER_FUSION_SHARED_EXPERTS: bool = False VLLM_ROCM_USE_AITER_TRITON_GEMM: bool = True VLLM_ROCM_USE_SKINNY_GEMM: bool = True VLLM_ROCM_FP8_PADDING: bool = True VLLM_ROCM_MOE_PADDING: bool = True VLLM_ROCM_CUSTOM_PAGED_ATTN: bool = True VLLM_ENABLE_V1_MULTIPROCESSING: bool = True VLLM_LOG_BATCHSIZE_INTERVAL: float = -1 VLLM_DISABLE_COMPILE_CACHE: bool = False Q_SCALE_CONSTANT: int = 200 K_SCALE_CONSTANT: int = 200 V_SCALE_CONSTANT: int = 100 VLLM_SERVER_DEV_MODE: bool = False VLLM_V1_OUTPUT_PROC_CHUNK_SIZE: int = 128 VLLM_MLA_DISABLE: bool = False VLLM_FLASH_ATTN_MAX_NUM_SPLITS_FOR_CUDA_GRAPH: int = 32 VLLM_RAY_PER_WORKER_GPUS: float = 1.0 VLLM_RAY_BUNDLE_INDICES: str = "" VLLM_CUDART_SO_PATH: str | None = None VLLM_DP_RANK: int = 0 VLLM_DP_RANK_LOCAL: int = -1 VLLM_DP_SIZE: int = 1 VLLM_USE_STANDALONE_COMPILE: bool = True VLLM_DP_MASTER_IP: str = "" VLLM_DP_MASTER_PORT: int = 0 VLLM_MOE_DP_CHUNK_SIZE: int = 256 VLLM_ENABLE_MOE_DP_CHUNK: bool = True VLLM_RANDOMIZE_DP_DUMMY_INPUTS: bool = False VLLM_RAY_DP_PACK_STRATEGY: Literal["strict", "fill", "span"] = "strict" VLLM_MARLIN_USE_ATOMIC_ADD: bool = False VLLM_MARLIN_INPUT_DTYPE: Literal["int8", "fp8"] | None = None VLLM_MXFP4_USE_MARLIN: bool | None = None VLLM_DEEPEPLL_NVFP4_DISPATCH: bool = False VLLM_V1_USE_OUTLINES_CACHE: bool = False VLLM_TPU_BUCKET_PADDING_GAP: int = 0 VLLM_TPU_MOST_MODEL_LEN: int | None = None VLLM_TPU_USING_PATHWAYS: bool = False VLLM_USE_DEEP_GEMM: bool = True VLLM_MOE_USE_DEEP_GEMM: bool = True VLLM_USE_DEEP_GEMM_E8M0: bool = True VLLM_DEEP_GEMM_WARMUP: Literal[ "skip", "full", "relax", ] = "relax" VLLM_USE_FUSED_MOE_GROUPED_TOPK: bool = True VLLM_USE_FLASHINFER_MOE_FP16: bool = False VLLM_USE_FLASHINFER_MOE_FP8: bool = False VLLM_USE_FLASHINFER_MOE_FP4: bool = False VLLM_FLASHINFER_MOE_BACKEND: Literal["throughput", "latency", "masked_gemm"] = ( "latency" ) VLLM_FLASHINFER_WORKSPACE_BUFFER_SIZE: int = 394 * 1024 * 1024 VLLM_XGRAMMAR_CACHE_MB: int = 0 VLLM_MSGPACK_ZERO_COPY_THRESHOLD: int = 256 VLLM_ALLOW_INSECURE_SERIALIZATION: bool = False VLLM_NIXL_SIDE_CHANNEL_HOST: str = "localhost" VLLM_NIXL_SIDE_CHANNEL_PORT: int = 5600 VLLM_MOONCAKE_BOOTSTRAP_PORT: int = 8998 VLLM_ALL2ALL_BACKEND: Literal[ "naive", "pplx", "deepep_high_throughput", "deepep_low_latency", "allgather_reducescatter", "flashinfer_all2allv", ] = "allgather_reducescatter" VLLM_MAX_TOKENS_PER_EXPERT_FP4_MOE: int = 163840 VLLM_TOOL_PARSE_REGEX_TIMEOUT_SECONDS: int = 1 VLLM_SLEEP_WHEN_IDLE: bool = False VLLM_MQ_MAX_CHUNK_BYTES_MB: int = 16 VLLM_EXECUTE_MODEL_TIMEOUT_SECONDS: int = 300 VLLM_KV_CACHE_LAYOUT: Literal["NHD", "HND"] | None = None VLLM_COMPUTE_NANS_IN_LOGITS: bool = False VLLM_USE_NVFP4_CT_EMULATIONS: bool = False VLLM_ROCM_QUICK_REDUCE_QUANTIZATION: Literal[ "FP", "INT8", "INT6", "INT4", "NONE" ] = "NONE" VLLM_ROCM_QUICK_REDUCE_CAST_BF16_TO_FP16: bool = True VLLM_ROCM_QUICK_REDUCE_MAX_SIZE_BYTES_MB: int | None = None VLLM_NIXL_ABORT_REQUEST_TIMEOUT: int = 480 VLLM_MOONCAKE_ABORT_REQUEST_TIMEOUT: int = 480 VLLM_USE_CUDNN_PREFILL: bool = False VLLM_USE_TRTLLM_RAGGED_DEEPSEEK_PREFILL: bool = False VLLM_ENABLE_CUDAGRAPH_GC: bool = False VLLM_LOOPBACK_IP: str = "" VLLM_ALLOW_CHUNKED_LOCAL_ATTN_WITH_HYBRID_KV_CACHE: bool = True VLLM_ENABLE_RESPONSES_API_STORE: bool = False VLLM_USE_TRTLLM_ATTENTION: str | None = None VLLM_NVFP4_GEMM_BACKEND: str | None = None VLLM_FLASHINFER_DISABLE_Q_QUANTIZATION: bool = False VLLM_HAS_FLASHINFER_CUBIN: bool = False VLLM_USE_FLASHINFER_MOE_MXFP4_MXFP8: bool = False VLLM_USE_FLASHINFER_MOE_MXFP4_BF16: bool = False VLLM_ROCM_FP8_MFMA_PAGE_ATTN: bool = False VLLM_USE_FLASHINFER_MOE_MXFP4_MXFP8_CUTLASS: bool = False VLLM_ALLREDUCE_USE_SYMM_MEM: bool = True VLLM_TUNED_CONFIG_FOLDER: str | None = None VLLM_GPT_OSS_SYSTEM_TOOL_MCP_LABELS: set[str] = set() VLLM_USE_EXPERIMENTAL_PARSER_CONTEXT: bool = False VLLM_GPT_OSS_HARMONY_SYSTEM_INSTRUCTIONS: bool = False VLLM_TOOL_JSON_ERROR_AUTOMATIC_RETRY: bool = False VLLM_CUSTOM_SCOPES_FOR_PROFILING: bool = False VLLM_NVTX_SCOPES_FOR_PROFILING: bool = False VLLM_KV_EVENTS_USE_INT_BLOCK_HASHES: bool = True VLLM_OBJECT_STORAGE_SHM_BUFFER_NAME: str = "VLLM_OBJECT_STORAGE_SHM_BUFFER" VLLM_DEEPEP_BUFFER_SIZE_MB: int = 1024 VLLM_DEEPEP_HIGH_THROUGHPUT_FORCE_INTRA_NODE: bool = False VLLM_DEEPEP_LOW_LATENCY_USE_MNNVL: bool = False VLLM_DBO_COMM_SMS: int = 20 VLLM_PATTERN_MATCH_DEBUG: str | None = None VLLM_DEBUG_DUMP_PATH: str | None = None VLLM_ENABLE_INDUCTOR_MAX_AUTOTUNE: bool = True VLLM_ENABLE_INDUCTOR_COORDINATE_DESCENT_TUNING: bool = True VLLM_USE_NCCL_SYMM_MEM: bool = False VLLM_NCCL_INCLUDE_PATH: str | None = None VLLM_USE_FBGEMM: bool = False VLLM_GC_DEBUG: str = "" VLLM_DEBUG_WORKSPACE: bool = False VLLM_DISABLE_SHARED_EXPERTS_STREAM: bool = False VLLM_SHARED_EXPERTS_STREAM_TOKEN_THRESHOLD: int = 256 VLLM_COMPILE_CACHE_SAVE_FORMAT: Literal["binary", "unpacked"] = "binary" VLLM_USE_V2_MODEL_RUNNER: bool = False VLLM_DEBUG_MFU_METRICS: bool = False def get_default_cache_root(): return os.getenv( "XDG_CACHE_HOME", os.path.join(os.path.expanduser("~"), ".cache"), ) def get_default_config_root(): return os.getenv( "XDG_CONFIG_HOME", os.path.join(os.path.expanduser("~"), ".config"), ) def maybe_convert_int(value: str | None) -> int | None: if value is None: return None return int(value) def maybe_convert_bool(value: str | None) -> bool | None: if value is None: return None return bool(int(value)) def disable_compile_cache() -> bool: return bool(int(os.getenv("VLLM_DISABLE_COMPILE_CACHE", "0"))) def use_aot_compile() -> bool: from vllm.model_executor.layers.batch_invariant import ( vllm_is_batch_invariant, ) from vllm.utils.torch_utils import is_torch_equal_or_newer default_value = ( "1" if is_torch_equal_or_newer("2.10.0.dev") and not disable_compile_cache() else "0" ) return ( not vllm_is_batch_invariant() and os.environ.get("VLLM_USE_AOT_COMPILE", default_value) == "1" ) def env_with_choices( env_name: str, default: str | None, choices: list[str] | Callable[[], list[str]], case_sensitive: bool = True, ) -> Callable[[], str | None]: """ Create a lambda that validates environment variable against allowed choices Args: env_name: Name of the environment variable default: Default value if not set (can be None) choices: List of valid string options or callable that returns list case_sensitive: Whether validation should be case sensitive Returns: Lambda function for environment_variables dict """ def _get_validated_env() -> str | None: value = os.getenv(env_name) if value is None: return default # Resolve choices if it's a callable (for lazy loading) actual_choices = choices() if callable(choices) else choices if not case_sensitive: check_value = value.lower() check_choices = [choice.lower() for choice in actual_choices] else: check_value = value check_choices = actual_choices if check_value not in check_choices: raise ValueError( f"Invalid value '{value}' for {env_name}. " f"Valid options: {actual_choices}." ) return value return _get_validated_env def env_list_with_choices( env_name: str, default: list[str], choices: list[str] | Callable[[], list[str]], case_sensitive: bool = True, ) -> Callable[[], list[str]]: """ Create a lambda that validates environment variable containing comma-separated values against allowed choices Args: env_name: Name of the environment variable default: Default list of values if not set choices: List of valid string options or callable that returns list case_sensitive: Whether validation should be case sensitive Returns: Lambda function for environment_variables dict that returns list of strings """ def _get_validated_env_list() -> list[str]: value = os.getenv(env_name) if value is None: return default # Split comma-separated values and strip whitespace values = [v.strip() for v in value.split(",") if v.strip()] if not values: return default # Resolve choices if it's a callable (for lazy loading) actual_choices = choices() if callable(choices) else choices # Validate each value for val in values: if not case_sensitive: check_value = val.lower() check_choices = [choice.lower() for choice in actual_choices] else: check_value = val check_choices = actual_choices if check_value not in check_choices: raise ValueError( f"Invalid value '{val}' in {env_name}. " f"Valid options: {actual_choices}." ) return values return _get_validated_env_list def env_set_with_choices( env_name: str, default: list[str], choices: list[str] | Callable[[], list[str]], case_sensitive: bool = True, ) -> Callable[[], set[str]]: """ Creates a lambda which that validates environment variable containing comma-separated values against allowed choices which returns choices as a set. """ def _get_validated_env_set() -> set[str]: return set(env_list_with_choices(env_name, default, choices, case_sensitive)()) return _get_validated_env_set def get_vllm_port() -> int | None: """Get the port from VLLM_PORT environment variable. Returns: The port number as an integer if VLLM_PORT is set, None otherwise. Raises: ValueError: If VLLM_PORT is a URI, suggest k8s service discovery issue. """ if "VLLM_PORT" not in os.environ: return None port = os.getenv("VLLM_PORT", "0") try: return int(port) except ValueError as err: from urllib.parse import urlparse parsed = urlparse(port) if parsed.scheme: raise ValueError( f"VLLM_PORT '{port}' appears to be a URI. " "This may be caused by a Kubernetes service discovery issue," "check the warning in: https://docs.vllm.ai/en/stable/serving/env_vars.html" ) from None raise ValueError(f"VLLM_PORT '{port}' must be a valid integer") from err # The start-* and end* here are used by the documentation generator # to extract the used env vars. # --8<-- [start:env-vars-definition] logger = logging.getLogger(__name__) environment_variables: dict[str, Callable[[], Any]] = { # ================== Installation Time Env Vars ================== # Target device of vLLM, supporting [cuda (by default), # rocm, cpu] "VLLM_TARGET_DEVICE": lambda: os.getenv("VLLM_TARGET_DEVICE", "cuda").lower(), # Main CUDA version of vLLM. This follows PyTorch but can be overridden. "VLLM_MAIN_CUDA_VERSION": lambda: os.getenv("VLLM_MAIN_CUDA_VERSION", "").lower() or "12.9", # Controls PyTorch float32 matmul precision mode within vLLM workers. # Accepted values: # - "ieee" (default): force full IEEE FP32 matmul precision. # - "tf32": enable TensorFloat32-based fast matmul. "VLLM_FLOAT32_MATMUL_PRECISION": env_with_choices( "VLLM_FLOAT32_MATMUL_PRECISION", "ieee", ["ieee", "tf32"], case_sensitive=False, ), # Maximum number of compilation jobs to run in parallel. # By default this is the number of CPUs "MAX_JOBS": lambda: os.getenv("MAX_JOBS", None), # Number of threads to use for nvcc # By default this is 1. # If set, `MAX_JOBS` will be reduced to avoid oversubscribing the CPU. "NVCC_THREADS": lambda: os.getenv("NVCC_THREADS", None), # If set, vllm will use precompiled binaries (*.so) "VLLM_USE_PRECOMPILED": lambda: os.environ.get("VLLM_USE_PRECOMPILED", "") .strip() .lower() in ("1", "true") or bool(os.environ.get("VLLM_PRECOMPILED_WHEEL_LOCATION")), # If set, skip adding +precompiled suffix to version string "VLLM_SKIP_PRECOMPILED_VERSION_SUFFIX": lambda: bool( int(os.environ.get("VLLM_SKIP_PRECOMPILED_VERSION_SUFFIX", "0")) ), # Used to mark that setup.py is running in a Docker build context, # in order to force the use of precompiled binaries. "VLLM_DOCKER_BUILD_CONTEXT": lambda: os.environ.get("VLLM_DOCKER_BUILD_CONTEXT", "") .strip() .lower() in ("1", "true"), # CMake build type # If not set, defaults to "Debug" or "RelWithDebInfo" # Available options: "Debug", "Release", "RelWithDebInfo" "CMAKE_BUILD_TYPE": env_with_choices( "CMAKE_BUILD_TYPE", None, ["Debug", "Release", "RelWithDebInfo"] ), # If set, vllm will print verbose logs during installation "VERBOSE": lambda: bool(int(os.getenv("VERBOSE", "0"))), # Root directory for vLLM configuration files # Defaults to `~/.config/vllm` unless `XDG_CONFIG_HOME` is set # Note that this not only affects how vllm finds its configuration files # during runtime, but also affects how vllm installs its configuration # files during **installation**. "VLLM_CONFIG_ROOT": lambda: os.path.expanduser( os.getenv( "VLLM_CONFIG_ROOT", os.path.join(get_default_config_root(), "vllm"), ) ), # ================== Runtime Env Vars ================== # Root directory for vLLM cache files # Defaults to `~/.cache/vllm` unless `XDG_CACHE_HOME` is set "VLLM_CACHE_ROOT": lambda: os.path.expanduser( os.getenv( "VLLM_CACHE_ROOT", os.path.join(get_default_cache_root(), "vllm"), ) ), # used in distributed environment to determine the ip address # of the current node, when the node has multiple network interfaces. # If you are using multi-node inference, you should set this differently # on each node. "VLLM_HOST_IP": lambda: os.getenv("VLLM_HOST_IP", ""), # used in distributed environment to manually set the communication port # Note: if VLLM_PORT is set, and some code asks for multiple ports, the # VLLM_PORT will be used as the first port, and the rest will be generated # by incrementing the VLLM_PORT value. "VLLM_PORT": get_vllm_port, # path used for ipc when the frontend api server is running in # multi-processing mode to communicate with the backend engine process. "VLLM_RPC_BASE_PATH": lambda: os.getenv( "VLLM_RPC_BASE_PATH", tempfile.gettempdir() ), # If true, will load models from ModelScope instead of Hugging Face Hub. # note that the value is true or false, not numbers "VLLM_USE_MODELSCOPE": lambda: os.environ.get( "VLLM_USE_MODELSCOPE", "False" ).lower() == "true", # Interval in seconds to log a warning message when the ring buffer is full "VLLM_RINGBUFFER_WARNING_INTERVAL": lambda: int( os.environ.get("VLLM_RINGBUFFER_WARNING_INTERVAL", "60") ), # path to cudatoolkit home directory, under which should be bin, include, # and lib directories. "CUDA_HOME": lambda: os.environ.get("CUDA_HOME", None), # Path to the NCCL library file. It is needed because nccl>=2.19 brought # by PyTorch contains a bug: https://github.com/NVIDIA/nccl/issues/1234 "VLLM_NCCL_SO_PATH": lambda: os.environ.get("VLLM_NCCL_SO_PATH", None), # when `VLLM_NCCL_SO_PATH` is not set, vllm will try to find the nccl # library file in the locations specified by `LD_LIBRARY_PATH` "LD_LIBRARY_PATH": lambda: os.environ.get("LD_LIBRARY_PATH", None), # flag to control the chunk size (in MB) for sleeping memory allocations under ROCm "VLLM_ROCM_SLEEP_MEM_CHUNK_SIZE": lambda: int( os.environ.get("VLLM_ROCM_SLEEP_MEM_CHUNK_SIZE", "256") ), # Use separate prefill and decode kernels for V1 attention instead of # the unified triton kernel. "VLLM_V1_USE_PREFILL_DECODE_ATTENTION": lambda: ( os.getenv("VLLM_V1_USE_PREFILL_DECODE_ATTENTION", "False").lower() in ("true", "1") ), # Force vllm to use a specific flash-attention version (2 or 3), only valid # when using the flash-attention backend. "VLLM_FLASH_ATTN_VERSION": lambda: maybe_convert_int( os.environ.get("VLLM_FLASH_ATTN_VERSION", None) ), # Feature flag to enable/disable Inductor standalone compile. # In torch <= 2.7 we ignore this flag; in torch >= 2.9 this is # enabled by default. "VLLM_USE_STANDALONE_COMPILE": lambda: os.environ.get( "VLLM_USE_STANDALONE_COMPILE", "1" ) == "1", # Debug pattern matching inside custom passes. # Should be set to the fx.Node name (e.g. 'getitem_34' or 'scaled_mm_3'). "VLLM_PATTERN_MATCH_DEBUG": lambda: os.environ.get( "VLLM_PATTERN_MATCH_DEBUG", None ), # Dump fx graphs to the given directory. # It will override CompilationConfig.debug_dump_path if set. "VLLM_DEBUG_DUMP_PATH": lambda: os.environ.get("VLLM_DEBUG_DUMP_PATH", None), # Feature flag to enable/disable AOT compilation. This will ensure # compilation is done in warmup phase and the compilation will be # reused in subsequent calls. "VLLM_USE_AOT_COMPILE": use_aot_compile, # Feature flag to enable/disable bytecode in # TorchCompileWithNoGuardsWrapper. "VLLM_USE_BYTECODE_HOOK": lambda: bool( int(os.environ.get("VLLM_USE_BYTECODE_HOOK", "1")) ), # Force vllm to always load AOT compiled models from disk. Failure # to load will result in a hard error when this is enabled. # Will be ignored when VLLM_USE_AOT_COMPILE is disabled. "VLLM_FORCE_AOT_LOAD": lambda: os.environ.get("VLLM_FORCE_AOT_LOAD", "0") == "1", # local rank of the process in the distributed setting, used to determine # the GPU device id "LOCAL_RANK": lambda: int(os.environ.get("LOCAL_RANK", "0")), # used to control the visible devices in the distributed setting "CUDA_VISIBLE_DEVICES": lambda: os.environ.get("CUDA_VISIBLE_DEVICES", None), # timeout for each iteration in the engine "VLLM_ENGINE_ITERATION_TIMEOUT_S": lambda: int( os.environ.get("VLLM_ENGINE_ITERATION_TIMEOUT_S", "60") ), # Timeout in seconds for waiting for engine cores to become ready # during startup. Default is 600 seconds (10 minutes). "VLLM_ENGINE_READY_TIMEOUT_S": lambda: int( os.environ.get("VLLM_ENGINE_READY_TIMEOUT_S", "600") ), # API key for vLLM API server "VLLM_API_KEY": lambda: os.environ.get("VLLM_API_KEY", None), # Whether to log responses from API Server for debugging "VLLM_DEBUG_LOG_API_SERVER_RESPONSE": lambda: os.environ.get( "VLLM_DEBUG_LOG_API_SERVER_RESPONSE", "False" ).lower() == "true", # S3 access information, used for tensorizer to load model from S3 "S3_ACCESS_KEY_ID": lambda: os.environ.get("S3_ACCESS_KEY_ID", None), "S3_SECRET_ACCESS_KEY": lambda: os.environ.get("S3_SECRET_ACCESS_KEY", None), "S3_ENDPOINT_URL": lambda: os.environ.get("S3_ENDPOINT_URL", None), # Usage stats collection "VLLM_USAGE_STATS_SERVER": lambda: os.environ.get( "VLLM_USAGE_STATS_SERVER", "https://stats.vllm.ai" ), "VLLM_NO_USAGE_STATS": lambda: os.environ.get("VLLM_NO_USAGE_STATS", "0") == "1", "VLLM_DISABLE_FLASHINFER_PREFILL": lambda: os.environ.get( "VLLM_DISABLE_FLASHINFER_PREFILL", "0" ) == "1", "VLLM_DO_NOT_TRACK": lambda: ( os.environ.get("VLLM_DO_NOT_TRACK", None) or os.environ.get("DO_NOT_TRACK", None) or "0" ) == "1", "VLLM_USAGE_SOURCE": lambda: os.environ.get("VLLM_USAGE_SOURCE", "production"), # Logging configuration # If set to 0, vllm will not configure logging # If set to 1, vllm will configure logging using the default configuration # or the configuration file specified by VLLM_LOGGING_CONFIG_PATH "VLLM_CONFIGURE_LOGGING": lambda: bool( int(os.getenv("VLLM_CONFIGURE_LOGGING", "1")) ), "VLLM_LOGGING_CONFIG_PATH": lambda: os.getenv("VLLM_LOGGING_CONFIG_PATH"), # this is used for configuring the default logging level "VLLM_LOGGING_LEVEL": lambda: os.getenv("VLLM_LOGGING_LEVEL", "INFO").upper(), # this is used for configuring the default logging stream "VLLM_LOGGING_STREAM": lambda: os.getenv("VLLM_LOGGING_STREAM", "ext://sys.stdout"), # if set, VLLM_LOGGING_PREFIX will be prepended to all log messages "VLLM_LOGGING_PREFIX": lambda: os.getenv("VLLM_LOGGING_PREFIX", ""), # Controls colored logging output. Options: "auto" (default, colors when terminal), # "1" (always use colors), "0" (never use colors) "VLLM_LOGGING_COLOR": lambda: os.getenv("VLLM_LOGGING_COLOR", "auto"), # Standard unix flag for disabling ANSI color codes "NO_COLOR": lambda: os.getenv("NO_COLOR", "0") != "0", # If set, vllm will log stats at this interval in seconds # If not set, vllm will log stats every 10 seconds. "VLLM_LOG_STATS_INTERVAL": lambda: val if (val := float(os.getenv("VLLM_LOG_STATS_INTERVAL", "10."))) > 0.0 else 10.0, # Trace function calls # If set to 1, vllm will trace function calls # Useful for debugging "VLLM_TRACE_FUNCTION": lambda: int(os.getenv("VLLM_TRACE_FUNCTION", "0")), # Backend for attention computation # Example options: # - "TORCH_SDPA": use torch.nn.MultiheadAttention # - "FLASH_ATTN": use FlashAttention # - "FLASHINFER": use flashinfer # - "FLASHMLA": use FlashMLA # - "FLASH_ATTN_MLA": use FlashAttention for MLA # - "FLASHINFER_MLA": use FlashInfer for MLA # - "CUTLASS_MLA": use CUTLASS for MLA # All possible options loaded dynamically from AttentionBackendEnum "VLLM_ATTENTION_BACKEND": env_with_choices( "VLLM_ATTENTION_BACKEND", None, lambda: list( __import__( "vllm.attention.backends.registry", fromlist=["AttentionBackendEnum"] ).AttentionBackendEnum.__members__.keys() ), ), # If set, vllm will use flashinfer sampler "VLLM_USE_FLASHINFER_SAMPLER": lambda: bool( int(os.environ["VLLM_USE_FLASHINFER_SAMPLER"]) ) if "VLLM_USE_FLASHINFER_SAMPLER" in os.environ else None, # Pipeline stage partition strategy "VLLM_PP_LAYER_PARTITION": lambda: os.getenv("VLLM_PP_LAYER_PARTITION", None), # (CPU backend only) CPU key-value cache space. # default is None and will be set as 4 GB "VLLM_CPU_KVCACHE_SPACE": lambda: int(os.getenv("VLLM_CPU_KVCACHE_SPACE", "0")) if "VLLM_CPU_KVCACHE_SPACE" in os.environ else None, # (CPU backend only) CPU core ids bound by OpenMP threads, e.g., "0-31", # "0,1,2", "0-31,33". CPU cores of different ranks are separated by '|'. "VLLM_CPU_OMP_THREADS_BIND": lambda: os.getenv("VLLM_CPU_OMP_THREADS_BIND", "auto"), # (CPU backend only) CPU cores not used by OMP threads . # Those CPU cores will not be used by OMP threads of a rank. "VLLM_CPU_NUM_OF_RESERVED_CPU": lambda: int( os.getenv("VLLM_CPU_NUM_OF_RESERVED_CPU", "0") ) if "VLLM_CPU_NUM_OF_RESERVED_CPU" in os.environ else None, # (CPU backend only) whether to use SGL kernels, optimized for small batch. "VLLM_CPU_SGL_KERNEL": lambda: bool(int(os.getenv("VLLM_CPU_SGL_KERNEL", "0"))), # If the env var is set, Ray Compiled Graph uses the specified # channel type to communicate between workers belonging to # different pipeline-parallel stages. # Available options: # - "auto": use the default channel type # - "nccl": use NCCL for communication # - "shm": use shared memory and gRPC for communication "VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE": env_with_choices( "VLLM_USE_RAY_COMPILED_DAG_CHANNEL_TYPE", "auto", ["auto", "nccl", "shm"] ), # If the env var is set, it enables GPU communication overlap # (experimental feature) in Ray's Compiled Graph. "VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM": lambda: bool( int(os.getenv("VLLM_USE_RAY_COMPILED_DAG_OVERLAP_COMM", "0")) ), # If the env var is set, it uses a Ray Communicator wrapping # vLLM's pipeline parallelism communicator to interact with Ray's # Compiled Graph. Otherwise, it uses Ray's NCCL communicator. "VLLM_USE_RAY_WRAPPED_PP_COMM": lambda: bool( int(os.getenv("VLLM_USE_RAY_WRAPPED_PP_COMM", "1")) ), # Use dedicated multiprocess context for workers. # Both spawn and fork work "VLLM_WORKER_MULTIPROC_METHOD": env_with_choices( "VLLM_WORKER_MULTIPROC_METHOD", "fork", ["spawn", "fork"] ), # Path to the cache for storing downloaded assets "VLLM_ASSETS_CACHE": lambda: os.path.expanduser( os.getenv( "VLLM_ASSETS_CACHE", os.path.join(get_default_cache_root(), "vllm", "assets"), ) ), # If the env var is set, we will clean model file in # this path $VLLM_ASSETS_CACHE/model_streamer/$model_name "VLLM_ASSETS_CACHE_MODEL_CLEAN": lambda: bool( int(os.getenv("VLLM_ASSETS_CACHE_MODEL_CLEAN", "0")) ), # Timeout for fetching images when serving multimodal models # Default is 5 seconds "VLLM_IMAGE_FETCH_TIMEOUT": lambda: int(os.getenv("VLLM_IMAGE_FETCH_TIMEOUT", "5")), # Timeout for fetching videos when serving multimodal models # Default is 30 seconds "VLLM_VIDEO_FETCH_TIMEOUT": lambda: int( os.getenv("VLLM_VIDEO_FETCH_TIMEOUT", "30") ), # Timeout for fetching audio when serving multimodal models # Default is 10 seconds "VLLM_AUDIO_FETCH_TIMEOUT": lambda: int( os.getenv("VLLM_AUDIO_FETCH_TIMEOUT", "10") ), # Whether to allow HTTP redirects when fetching from media URLs. # Default to True "VLLM_MEDIA_URL_ALLOW_REDIRECTS": lambda: bool( int(os.getenv("VLLM_MEDIA_URL_ALLOW_REDIRECTS", "1")) ), # Max number of workers for the thread pool handling # media bytes loading. Set to 1 to disable parallel processing. # Default is 8 "VLLM_MEDIA_LOADING_THREAD_COUNT": lambda: int( os.getenv("VLLM_MEDIA_LOADING_THREAD_COUNT", "8") ), # Maximum filesize in MB for a single audio file when processing # speech-to-text requests. Files larger than this will be rejected. # Default is 25 MB "VLLM_MAX_AUDIO_CLIP_FILESIZE_MB": lambda: int( os.getenv("VLLM_MAX_AUDIO_CLIP_FILESIZE_MB", "25") ),
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/env_override.py
vllm/env_override.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os import torch from vllm.logger import init_logger from vllm.utils.torch_utils import is_torch_equal logger = init_logger(__name__) # set some common config/environment variables that should be set # for all processes created by vllm and all processes # that interact with vllm workers. # they are executed whenever `import vllm` is called. # see https://github.com/vllm-project/vllm/pull/15951 # it avoids unintentional cuda initialization from torch.cuda.is_available() os.environ["PYTORCH_NVML_BASED_CUDA_CHECK"] = "1" # see https://github.com/vllm-project/vllm/issues/10480 os.environ["TORCHINDUCTOR_COMPILE_THREADS"] = "1" # see https://github.com/vllm-project/vllm/issues/10619 torch._inductor.config.compile_threads = 1 # =================================================== # torch 2.9 Inductor PythonWrapperCodegen monkeypatch # =================================================== # This change monkeypatches memory_plan_reuse in pytorch 2.9.0 to work around # a test failure for test_multi_graph_piecewise_compile_outputs_equal. # For more context, see https://github.com/pytorch/pytorch/pull/165514. def memory_plan_reuse_patched(self): import torch._inductor.ir as ir from torch._inductor.codegen.wrapper import ( EnterSubgraphLine, ExitSubgraphLine, MemoryPlanningLine, MemoryPlanningState, SubgraphPythonWrapperCodegen, ) from torch._inductor.virtualized import V def get_output_names(graph_outputs) -> list[str]: import itertools names = [] shape_counter = itertools.count(0) none_counter = itertools.count(0) for node in graph_outputs: if isinstance(node, ir.NoneAsConstantBuffer): names.append(f"{V.graph.name}_none{next(none_counter)}") elif isinstance(node, ir.ShapeAsConstantBuffer): names.append(f"{V.graph.name}_shape{next(shape_counter)}") else: names.append(node.get_name()) return names if ( isinstance(V.graph.wrapper_code, SubgraphPythonWrapperCodegen) and V.graph.wrapper_code.partition_signatures is not None ): out_names = get_output_names( V.graph.wrapper_code.partition_signatures.output_nodes ) else: out_names = V.graph.get_output_names() while ( self.lines and isinstance(self.lines[-1], MemoryPlanningLine) and self.lines[-1].node.name not in out_names # type: ignore[attr-defined] ): # these lines will be pointless self.lines.pop() # codegen allocations in two passes planning_states = [MemoryPlanningState()] past_planning_states = [] for i in range(len(self.lines)): line = self.lines[i] if isinstance(line, MemoryPlanningLine): self.lines[i] = line.plan(planning_states[-1]) elif isinstance(line, EnterSubgraphLine): planning_states.append(MemoryPlanningState()) elif isinstance(line, ExitSubgraphLine): past_planning_states.append(planning_states.pop()) past_planning_states.append(planning_states.pop()) assert len(planning_states) == 0 # =================================================== # torch 2.9 Inductor get_graph_partition_signature monkeypatch # =================================================== # This change monkeypatches get_graph_partition_signature in pytorch 2.9.0 to # fix inductor partition + attention-nvfp4 quant fusion, tested in # `tests/compile/distributed/test_fusions_e2e.py::test_attn_quant`. # For more context, see https://github.com/pytorch/pytorch/pull/165815. def get_graph_partition_signature_patched( self, partitions, skip_cudagraphs: list[bool] ): """ Gets signature for each graph partition, including input nodes, output nodes, and whether deallocating an input within graph partition. """ from torch._inductor import dependencies from torch._inductor.ir import GraphPartitionSignature, MutationOutput, NoneLayout from torch._inductor.virtualized import V from torch.utils._ordered_set import OrderedSet signatures = [] unmet_output_names = OrderedSet(V.graph.get_output_names()) name_to_node = self.get_name_to_nodes() def is_none_layout(buf_name: str) -> bool: """ Checks if buf_name is NoneLayout. Buffers with NoneLayout is not allocated so graph partition should not take it as inputs or outputs. """ buf = self.name_to_buf.get(buf_name, None) if buf is None: return False if isinstance(buf.node.layout, NoneLayout): if isinstance(buf.node, MutationOutput) and ( real_name := self.mutation_real_name.get(buf_name, None) ): return is_none_layout(real_name) return True return False for partition, skip_cudagraph in zip( reversed(partitions), reversed(skip_cudagraphs) ): output_names: OrderedSet[str] = OrderedSet() for node in partition: output_names.update(node.outputs_by_name.keys()) returned_output_names = output_names.intersection(unmet_output_names) # all reads/writes are partition inputs except those generated # within the partition and tensor constants read_writes = dependencies.ReadWrites.merge_list( [node.read_writes for node in partition] ) # WeakDep is fake dependency on unused buffer. It should not appear # in partition_input_names for inputs that are actually read or written. partition_input_names = ( OrderedSet( [ x.name for x in read_writes.reads | read_writes.writes if not is_none_layout(x.name) ] ) - output_names ) partition_input_names = OrderedSet( self.mutation_real_name.get(name, name) for name in partition_input_names ) buffer_names_to_free: OrderedSet[str] = OrderedSet() for node in partition: buffer_names_to_free.update(node.last_usage) # buffer_names_to_free may contain buffers allocated in previous # graph partitions. These buffers should also be a partition # input. extra_input_names = [ name for name in (buffer_names_to_free - output_names) if name in name_to_node ] partition_input_names.update(extra_input_names) input_nodes = { name: name_to_node[name] for name in partition_input_names if name in name_to_node } input_deallocation = { name: name in buffer_names_to_free for name in partition_input_names if name in name_to_node } # if an input tensor is not freed in the partition function, it should # also be returned as an output. This brings benefits to cudagraph # since the returned output tensor is a cudagraph managed tensor with # a static tensor address. extra_output_names = [ name for name in partition_input_names if name in name_to_node and name not in buffer_names_to_free ] returned_output_names.update(extra_output_names) returned_output_names = OrderedSet( self.mutation_real_name.get(name, name) for name in returned_output_names ) output_nodes = [ name_to_node[name] for name in returned_output_names if not is_none_layout(name) ] constant_names = [ name for name in partition_input_names if name in V.graph.constants ] symbol_inputs = self.get_graph_partition_symbol_inputs(partition, input_nodes) partition_signature = GraphPartitionSignature( symbol_inputs, input_nodes, output_nodes, input_deallocation, skip_cudagraph, constant_names, ) signatures.append(partition_signature) unmet_output_names = partition_input_names.union( unmet_output_names - returned_output_names ) return signatures[::-1] # ======================================== # torch 2.9 Inductor Scheduler monkeypatch # ======================================== # This change monkeypatches a function in Inductor to work around the following # bug: https://github.com/vllm-project/vllm/issues/26678 # # The bug occurs when `use_inductor_graph_partition` is turned on and there # exists operators inside of `splitting_ops` that have an in-place mutation. In # vllm, this specifically occurs on the operator # vllm.unified_attention_with_output. In this case, inductor does not populate # the inductor IR's `origin_node` field, causing an assertion error when trying # to access the node's `origin_node` field. # # So, we will monkeypatch torch._inductor.scheduler.Scheduler.should_partition # so that it does not access the inductor IR node's `origin_node` field and just # returns True if a node is registered as having a custom partition function. # This is ok for now since vllm's implementation of the custom partition # functions just return True. # ======================================== def should_partition_patched(self, node, should_log: bool = False) -> bool: # This is a patched version of # torch._inductor.scheduler.Scheduler.should_partition that modifies # the following piece of code so that we always return True: # https://github.com/pytorch/pytorch/blob/ecb53078faf86ca1b33277df33b82985675bb011/torch/_inductor/scheduler.py#L4712-L4724 """Return True if we should partition the inductor graph on this node""" import torch._inductor.ir as ir from torch._inductor.scheduler import ( BaseSchedulerNode, FusedSchedulerNode, ) from torch._inductor.utils import ( _unstable_customized_partition_wrapper, is_cudagraph_unsafe_op, maybe_log_cudagraph_partition, ) # Allow users to manually specify if a node should be partitioned # Can only do this for FallbackKernels ir_node = node.node if isinstance(ir_node, torch._inductor.ir.FallbackKernel) and ( op := ir_node.op_overload ): op_overload_packet_name = op.name() op_overload_name = ( f"{op_overload_packet_name}.{op._overloadname}" if isinstance(op, torch._ops.OpOverload) else op_overload_packet_name ) if ( op_overload_packet_name in torch._inductor.config.custom_should_partition_ops or op_overload_name in torch._inductor.config.custom_should_partition_ops ): assert isinstance(op, torch._ops.OpOverload) return True # When not using cudagraphs, keep all kernels in the `call` function # instead of graph partition functions, since graph partition only brings # benefit to cudagraph if ( not torch._inductor.config.triton.cudagraphs and _unstable_customized_partition_wrapper.wrapper is None ): return True # avoid duplicating logs when should_partition is called multiple times # on the same node def noop_log(msg: str, node: BaseSchedulerNode | None) -> None: return log_partition_reason = maybe_log_cudagraph_partition if should_log else noop_log if isinstance(node, FusedSchedulerNode): return any(self.should_partition(snode) for snode in node.snodes) assert node.node is not None if not node.is_gpu(): log_partition_reason("non gpu ops", node=node) return True if isinstance(node.node, ir.DeviceCopy): log_partition_reason("DeviceCopy ops", node=node) return True if isinstance(node.node, ir.Conditional): log_partition_reason("Conditional ops", node=node) return True if getattr(node.node, "unbacked_bindings", None): log_partition_reason("unbacked binding ops", node=node) return True if is_cudagraph_unsafe_op(node.node): log_partition_reason("CUDAGraph-unsafe custom ops", node=node) return True return False def _update_scheduler_patched(self) -> None: # Copied from torch._inductor.graph.GrahLowering._update_scheduler. Patches # this method so that we can patch Scheduler.should_partition with the # function above """ (Re)initializes the scheduler member. When initializing the scheduler, no CUBIN files should be generated (to avoid biasing any benchmarks and pessimizing fusion decisions). """ import torch._inductor.config as config from torch._inductor.scheduler import Scheduler Scheduler.should_partition = should_partition_patched Scheduler.get_graph_partition_signature = get_graph_partition_signature_patched with config.patch("triton.store_cubin", False): self.scheduler = Scheduler(self.operations) # =================================================== # torch 2.9 Inductor get_raw_stream workaround # =================================================== # Workaround for TorchInductor autotune using get_raw_stream() without defining it. # This occurs when compile_sizes > 1 in compilation_config. # For more context, see https://github.com/vllm-project/vllm/issues/30905. def _patch_get_raw_stream_if_needed(): """Workaround for TorchInductor autotune get_raw_stream() bug.""" from vllm.utils.torch_utils import is_torch_equal # Only apply the patch for torch 2.9.0 or 2.9.1 if is_torch_equal("2.9.0") or is_torch_equal("2.9.1"): import builtins # Check if CUDA functionality is available without initializing CUDA # _cuda_getCurrentRawStream only exists in CUDA builds of PyTorch if hasattr(torch._C, "_cuda_getCurrentRawStream"): from torch._C import _cuda_getCurrentRawStream as _get_raw_stream builtins.get_raw_stream = _get_raw_stream _patch_get_raw_stream_if_needed() if is_torch_equal("2.9.0"): from torch._inductor.codegen.wrapper import PythonWrapperCodegen from torch._inductor.graph import GraphLowering from torch.utils._config_module import _Config, _ConfigEntry # `custom_should_partition_ops` is a new config after 2.9.0. So this would # not overwrite any user configs. torch._inductor.config._config["custom_should_partition_ops"] = _ConfigEntry( _Config(default=[]) ) PythonWrapperCodegen.memory_plan_reuse = memory_plan_reuse_patched GraphLowering._update_scheduler = _update_scheduler_patched
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/beam_search.py
vllm/beam_search.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass from typing import TYPE_CHECKING, Any, Optional from vllm.logprobs import Logprob from vllm.lora.request import LoRARequest if TYPE_CHECKING: from vllm.multimodal import MultiModalDataDict @dataclass class BeamSearchSequence: """A sequence for beam search. It keeps track of the tokens and the log probability of the sequence. The text field is optional and will only be filled when the sequence is about to be returned to the user. """ # The tokens include the prompt. tokens: list[int] logprobs: list[dict[int, Logprob]] lora_request: LoRARequest | None = None cum_logprob: float = 0.0 text: str | None = None finish_reason: str | None = None stop_reason: int | str | None = None multi_modal_data: Optional["MultiModalDataDict"] = None mm_processor_kwargs: dict[str, Any] | None = None @dataclass class BeamSearchOutput: """The output of beam search. It contains the list of the best beam search sequences. The length of the list is equal to the beam width. """ sequences: list[BeamSearchSequence] class BeamSearchInstance: def __init__( self, prompt_tokens: list[int], lora_request: LoRARequest | None = None, logprobs: list[dict[int, Logprob]] | None = None, **kwargs, ): self.beams: list[BeamSearchSequence] = [ BeamSearchSequence( tokens=prompt_tokens, logprobs=[] if logprobs is None else list(logprobs), lora_request=lora_request, **kwargs, ) ] self.completed: list[BeamSearchSequence] = [] def get_beam_search_score( tokens: list[int], cumulative_logprob: float, eos_token_id: int, length_penalty: float = 1.0, ) -> float: """Calculate the beam search score with length penalty. Adapted from https://github.com/huggingface/transformers/blob/ccb92be23def445f2afdea94c31286f84b89eb5b/src/transformers/generation/beam_search.py#L938 """ seq_len = len(tokens) if tokens[-1] == eos_token_id: seq_len -= 1 return cumulative_logprob / (seq_len**length_penalty) def create_sort_beams_key_function(eos_token_id: int, length_penalty: float): def sort_beams_key(x: BeamSearchSequence) -> float: return get_beam_search_score( x.tokens, x.cum_logprob, eos_token_id, length_penalty ) return sort_beams_key
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/sampling_params.py
vllm/sampling_params.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Sampling parameters for text generation.""" import copy from dataclasses import field from enum import Enum, IntEnum from functools import cached_property from typing import Annotated, Any import msgspec from pydantic.dataclasses import dataclass from vllm.logger import init_logger from vllm.logits_process import LogitsProcessor from vllm.tokenizers import TokenizerLike from vllm.v1.serial_utils import PydanticMsgspecMixin logger = init_logger(__name__) _SAMPLING_EPS = 1e-5 _MAX_TEMP = 1e-2 class SamplingType(IntEnum): GREEDY = 0 RANDOM = 1 RANDOM_SEED = 2 # maybe make msgspec? @dataclass class StructuredOutputsParams: # One of these fields will be used to build a logit processor. json: str | dict | None = None regex: str | None = None choice: list[str] | None = None grammar: str | None = None json_object: bool | None = None # These are other options that can be set. disable_fallback: bool = False disable_any_whitespace: bool = False disable_additional_properties: bool = False whitespace_pattern: str | None = None structural_tag: str | None = None _backend: str | None = field(default=None, init=False) """CAUTION: Should only be set by Processor._validate_structured_output""" _backend_was_auto: bool = field(default=False, init=False) """CAUTION: Should only be set by Processor._validate_structured_output""" def __post_init__(self): """Validate that some fields are mutually exclusive.""" count = sum( [ self.json is not None, self.regex is not None, self.choice is not None, self.grammar is not None, self.json_object is not None, self.structural_tag is not None, ] ) if count > 1: raise ValueError( "You can only use one kind of structured outputs constraint " f"but multiple are specified: {self.__dict__}" ) def all_constraints_none(self) -> bool: """ Returns True if all structured-output constraint fields are None. """ return all( getattr(self, field) is None for field in ( "json", "regex", "choice", "grammar", "json_object", "structural_tag", ) ) def all_non_structural_tag_constraints_none(self) -> bool: """ Returns True if all structured-output constraint fields are None. """ return all( getattr(self, field) is None for field in ( "json", "regex", "choice", "grammar", "json_object", ) ) class RequestOutputKind(Enum): # Return entire output so far in every RequestOutput CUMULATIVE = 0 # Return only deltas in each RequestOutput DELTA = 1 # Do not return intermediate RequestOutput FINAL_ONLY = 2 class SamplingParams( PydanticMsgspecMixin, msgspec.Struct, omit_defaults=True, # type: ignore[call-arg] # required for @cached_property. dict=True, ): # type: ignore[call-arg] """Sampling parameters for text generation. Overall, we follow the sampling parameters from the OpenAI text completion API (https://platform.openai.com/docs/api-reference/completions/create). In addition, we support beam search, which is not supported by OpenAI. """ n: int = 1 """Number of outputs to return for the given prompt request. NOTE: `AsyncLLM` streams outputs by default. When `n > 1`, all `n` outputs are generated and streamed cumulatively per request. To see all `n` outputs upon completion, use `output_kind=RequestOutputKind.FINAL_ONLY` in `SamplingParams`.""" presence_penalty: float = 0.0 """Penalizes new tokens based on whether they appear in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.""" frequency_penalty: float = 0.0 """Penalizes new tokens based on their frequency in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.""" repetition_penalty: float = 1.0 """Penalizes new tokens based on whether they appear in the prompt and the generated text so far. Values > 1 encourage the model to use new tokens, while values < 1 encourage the model to repeat tokens.""" temperature: float = 1.0 """Controls the randomness of the sampling. Lower values make the model more deterministic, while higher values make the model more random. Zero means greedy sampling.""" top_p: float = 1.0 """Controls the cumulative probability of the top tokens to consider. Must be in (0, 1]. Set to 1 to consider all tokens.""" top_k: int = 0 """Controls the number of top tokens to consider. Set to 0 (or -1) to consider all tokens.""" min_p: float = 0.0 """Represents the minimum probability for a token to be considered, relative to the probability of the most likely token. Must be in [0, 1]. Set to 0 to disable this.""" seed: int | None = None """Random seed to use for the generation.""" stop: str | list[str] | None = None """String(s) that stop the generation when they are generated. The returned output will not contain the stop strings.""" stop_token_ids: list[int] | None = None """Token IDs that stop the generation when they are generated. The returned output will contain the stop tokens unless the stop tokens are special tokens.""" ignore_eos: bool = False """Whether to ignore the EOS token and continue generating tokens after the EOS token is generated.""" max_tokens: int | None = 16 """Maximum number of tokens to generate per output sequence.""" min_tokens: int = 0 """Minimum number of tokens to generate per output sequence before EOS or `stop_token_ids` can be generated""" logprobs: int | None = None """Number of log probabilities to return per output token. When set to `None`, no probability is returned. If set to a non-`None` value, the result includes the log probabilities of the specified number of most likely tokens, as well as the chosen tokens. Note that the implementation follows the OpenAI API: The API will always return the log probability of the sampled token, so there may be up to `logprobs+1` elements in the response. When set to -1, return all `vocab_size` log probabilities.""" prompt_logprobs: int | None = None """Number of log probabilities to return per prompt token. When set to -1, return all `vocab_size` log probabilities.""" flat_logprobs: bool = False """Whether to return logprobs in flatten format (i.e. FlatLogprob) for better performance. NOTE: GC costs of FlatLogprobs is significantly smaller than list[dict[int, Logprob]]. After enabled, PromptLogprobs and SampleLogprobs would populated as FlatLogprobs.""" # NOTE: This parameter is only exposed at the engine level for now. # It is not exposed in the OpenAI API server, as the OpenAI API does # not support returning only a list of token IDs. detokenize: bool = True """Whether to detokenize the output.""" skip_special_tokens: bool = True """Whether to skip special tokens in the output.""" spaces_between_special_tokens: bool = True """Whether to add spaces between special tokens in the output.""" # `list[LogitsProcessor] | None` type. We use Any here because # `list[LogitsProcessor] | None` type is not supported by msgspec. logits_processors: Any | None = None """Functions that modify logits based on previously generated tokens, and optionally prompt tokens as a first argument.""" include_stop_str_in_output: bool = False """Whether to include the stop strings in output text.""" truncate_prompt_tokens: Annotated[int, msgspec.Meta(ge=-1)] | None = None """If set to -1, will use the truncation size supported by the model. If set to an integer k, will use only the last k tokens from the prompt (i.e., left truncation). If set to `None`, truncation is disabled.""" output_kind: RequestOutputKind = RequestOutputKind.CUMULATIVE skip_clone: bool = False """Internal flag indicating that this SamplingParams instance is safe to reuse without cloning. When True, clone() will return self without performing a deep copy. This should only be set when the params object is guaranteed to be dedicated to a single request and won't be modified in ways that would affect other uses.""" # The below fields are not supposed to be used as an input. # They are set in post_init. output_text_buffer_length: int = 0 _all_stop_token_ids: set[int] = msgspec.field(default_factory=set) # Fields used to construct logits processors structured_outputs: StructuredOutputsParams | None = None """Parameters for configuring structured outputs.""" logit_bias: dict[int, float] | None = None """If provided, the engine will construct a logits processor that applies these logit biases.""" allowed_token_ids: list[int] | None = None """If provided, the engine will construct a logits processor which only retains scores for the given token ids.""" extra_args: dict[str, Any] | None = None """Arbitrary additional args, that can be used by custom sampling implementations, plugins, etc. Not used by any in-tree sampling implementations.""" # Fields used for bad words bad_words: list[str] | None = None """Words that are not allowed to be generated. More precisely, only the last token of a corresponding token sequence is not allowed when the next generated token can complete the sequence.""" _bad_words_token_ids: list[list[int]] | None = None skip_reading_prefix_cache: bool | None = None @staticmethod def from_optional( n: int | None = 1, presence_penalty: float | None = 0.0, frequency_penalty: float | None = 0.0, repetition_penalty: float | None = 1.0, temperature: float | None = 1.0, top_p: float | None = 1.0, top_k: int = 0, min_p: float = 0.0, seed: int | None = None, stop: str | list[str] | None = None, stop_token_ids: list[int] | None = None, bad_words: list[str] | None = None, include_stop_str_in_output: bool = False, ignore_eos: bool = False, max_tokens: int | None = 16, min_tokens: int = 0, logprobs: int | None = None, prompt_logprobs: int | None = None, detokenize: bool = True, skip_special_tokens: bool = True, spaces_between_special_tokens: bool = True, logits_processors: list[LogitsProcessor] | None = None, truncate_prompt_tokens: Annotated[int, msgspec.Meta(ge=-1)] | None = None, output_kind: RequestOutputKind = RequestOutputKind.CUMULATIVE, structured_outputs: StructuredOutputsParams | None = None, logit_bias: dict[int, float] | dict[str, float] | None = None, allowed_token_ids: list[int] | None = None, extra_args: dict[str, Any] | None = None, skip_clone: bool = False, ) -> "SamplingParams": if logit_bias is not None: # Convert token_id to integer # Clamp the bias between -100 and 100 per OpenAI API spec logit_bias = { int(token): min(100.0, max(-100.0, bias)) for token, bias in logit_bias.items() } return SamplingParams( n=1 if n is None else n, presence_penalty=0.0 if presence_penalty is None else presence_penalty, frequency_penalty=0.0 if frequency_penalty is None else frequency_penalty, repetition_penalty=1.0 if repetition_penalty is None else repetition_penalty, temperature=1.0 if temperature is None else temperature, top_p=1.0 if top_p is None else top_p, top_k=top_k, min_p=min_p, seed=seed, stop=stop, stop_token_ids=stop_token_ids, bad_words=bad_words, include_stop_str_in_output=include_stop_str_in_output, ignore_eos=ignore_eos, max_tokens=max_tokens, min_tokens=min_tokens, logprobs=logprobs, prompt_logprobs=prompt_logprobs, detokenize=detokenize, skip_special_tokens=skip_special_tokens, spaces_between_special_tokens=spaces_between_special_tokens, logits_processors=logits_processors, truncate_prompt_tokens=truncate_prompt_tokens, output_kind=output_kind, structured_outputs=structured_outputs, logit_bias=logit_bias, allowed_token_ids=allowed_token_ids, extra_args=extra_args, skip_clone=skip_clone, ) def __post_init__(self) -> None: if 0 < self.temperature < _MAX_TEMP: logger.warning( "temperature %s is less than %s, which may cause numerical " "errors nan or inf in tensors. We have maxed it out to %s.", self.temperature, _MAX_TEMP, _MAX_TEMP, ) self.temperature = max(self.temperature, _MAX_TEMP) if self.seed == -1: self.seed = None if self.stop is None: self.stop = [] elif isinstance(self.stop, str): self.stop = [self.stop] if self.stop_token_ids is None: self.stop_token_ids = [] if self.bad_words is None: self.bad_words = [] if self.logprobs is True: self.logprobs = 1 if self.prompt_logprobs is True: self.prompt_logprobs = 1 # Number of characters to hold back for stop string evaluation # until sequence is finished. if self.stop and not self.include_stop_str_in_output: self.output_text_buffer_length = max(len(s) for s in self.stop) - 1 self._verify_args() if self.temperature < _SAMPLING_EPS: # Zero temperature means greedy sampling. self.top_p = 1.0 self.top_k = 0 self.min_p = 0.0 self._verify_greedy_sampling() # eos_token_id is added to this by the engine self._all_stop_token_ids.update(self.stop_token_ids) if self.skip_reading_prefix_cache is None: # If prefix caching is enabled, # the output of prompt logprobs may less than n_prompt_tokens, # we need to skip reading cache at this request. self.skip_reading_prefix_cache = self.prompt_logprobs is not None def _verify_args(self) -> None: if not isinstance(self.n, int): raise ValueError(f"n must be an int, but is of type {type(self.n)}") if self.n < 1: raise ValueError(f"n must be at least 1, got {self.n}.") if not -2.0 <= self.presence_penalty <= 2.0: raise ValueError( f"presence_penalty must be in [-2, 2], got {self.presence_penalty}." ) if not -2.0 <= self.frequency_penalty <= 2.0: raise ValueError( f"frequency_penalty must be in [-2, 2], got {self.frequency_penalty}." ) if self.repetition_penalty <= 0.0: raise ValueError( "repetition_penalty must be greater than zero, got " f"{self.repetition_penalty}." ) if self.temperature < 0.0: raise ValueError( f"temperature must be non-negative, got {self.temperature}." ) if not 0.0 < self.top_p <= 1.0: raise ValueError(f"top_p must be in (0, 1], got {self.top_p}.") # quietly accept -1 as disabled, but prefer 0 if self.top_k < -1: raise ValueError( f"top_k must be 0 (disable), or at least 1, got {self.top_k}." ) if not isinstance(self.top_k, int): raise TypeError( f"top_k must be an integer, got {type(self.top_k).__name__}" ) if not 0.0 <= self.min_p <= 1.0: raise ValueError(f"min_p must be in [0, 1], got {self.min_p}.") if self.max_tokens is not None and self.max_tokens < 1: raise ValueError(f"max_tokens must be at least 1, got {self.max_tokens}.") if self.min_tokens < 0: raise ValueError( f"min_tokens must be greater than or equal to 0, got {self.min_tokens}." ) if self.max_tokens is not None and self.min_tokens > self.max_tokens: raise ValueError( f"min_tokens must be less than or equal to " f"max_tokens={self.max_tokens}, got {self.min_tokens}." ) if self.logprobs is not None and self.logprobs != -1 and self.logprobs < 0: raise ValueError( f"logprobs must be non-negative or -1, got {self.logprobs}." ) if ( self.prompt_logprobs is not None and self.prompt_logprobs != -1 and self.prompt_logprobs < 0 ): raise ValueError( f"prompt_logprobs must be non-negative or -1, got " f"{self.prompt_logprobs}." ) if self.truncate_prompt_tokens is not None and ( self.truncate_prompt_tokens == 0 or self.truncate_prompt_tokens < -1 ): raise ValueError( f"truncate_prompt_tokens must be an integer >= 1 or -1, " f"got {self.truncate_prompt_tokens}" ) assert isinstance(self.stop_token_ids, list) if not all(isinstance(st_id, int) for st_id in self.stop_token_ids): raise ValueError( f"stop_token_ids must contain only integers, got {self.stop_token_ids}." ) assert isinstance(self.stop, list) if any(not stop_str for stop_str in self.stop): raise ValueError("stop cannot contain an empty string.") if self.stop and not self.detokenize: raise ValueError( "stop strings are only supported when detokenize is True. " "Set detokenize=True to use stop." ) def _verify_greedy_sampling(self) -> None: if self.n > 1: raise ValueError(f"n must be 1 when using greedy sampling, got {self.n}.") def update_from_generation_config( self, generation_config: dict[str, Any], model_eos_token_id: int | None = None, ) -> None: """Update if there are non-default values from generation_config""" if model_eos_token_id is not None: # Add the eos token id into the sampling_params to support # min_tokens processing. self._all_stop_token_ids.add(model_eos_token_id) # Update eos_token_id for generation if (eos_ids := generation_config.get("eos_token_id")) is not None: # it can be either int or list of int eos_ids = {eos_ids} if isinstance(eos_ids, int) else set(eos_ids) if model_eos_token_id is not None: # We don't need to include the primary eos_token_id in # stop_token_ids since it's handled separately for stopping # purposes. eos_ids.discard(model_eos_token_id) if eos_ids: self._all_stop_token_ids.update(eos_ids) if not self.ignore_eos: eos_ids.update(self.stop_token_ids) self.stop_token_ids = list(eos_ids) def update_from_tokenizer(self, tokenizer: TokenizerLike) -> None: if not self.bad_words: return self._bad_words_token_ids = [] for bad_word in self.bad_words: # To prohibit words both at the beginning # and in the middle of text # (related to add_prefix_space tokenizer parameter) for add_prefix_space in [False, True]: prefix = " " if add_prefix_space else "" prompt = prefix + bad_word.lstrip() prompt_token_ids = tokenizer.encode( text=prompt, add_special_tokens=False ) # If no space at the beginning # or if prefix space produces a new word token if (not add_prefix_space) or ( add_prefix_space and prompt_token_ids[0] != self._bad_words_token_ids[-1][0] and len(prompt_token_ids) == len(self._bad_words_token_ids[-1]) ): self._bad_words_token_ids.append(prompt_token_ids) invalid_token_ids = [ token_id for bad_words_token_ids in self._bad_words_token_ids for token_id in bad_words_token_ids if token_id < 0 or token_id > tokenizer.max_token_id ] if len(invalid_token_ids) > 0: raise ValueError( f"The model vocabulary size is {tokenizer.max_token_id + 1}," f" but the following tokens" f" were specified as bad: {invalid_token_ids}." f" All token id values should be integers satisfying:" f" 0 <= token_id <= {tokenizer.max_token_id}." ) @cached_property def sampling_type(self) -> SamplingType: if self.temperature < _SAMPLING_EPS: return SamplingType.GREEDY if self.seed is not None: return SamplingType.RANDOM_SEED return SamplingType.RANDOM @property def all_stop_token_ids(self) -> set[int]: return self._all_stop_token_ids @property def bad_words_token_ids(self) -> list[list[int]] | None: # For internal use only. Backward compatibility not guaranteed return self._bad_words_token_ids def clone(self) -> "SamplingParams": """Deep copy, but maybe not the LogitsProcessor objects. LogitsProcessor objects may contain an arbitrary, nontrivial amount of data that is expensive to copy. However, if not copied, the processor needs to support parallel decoding for multiple sequences See https://github.com/vllm-project/vllm/issues/3087 If skip_clone is True, uses shallow copy instead of deep copy. """ if self.skip_clone: return copy.copy(self) logit_processor_refs = ( None if self.logits_processors is None else { id(lp): lp.clone() if hasattr(lp, "clone") else lp for lp in self.logits_processors } ) return copy.deepcopy(self, memo=logit_processor_refs) def __repr__(self) -> str: return ( f"SamplingParams(n={self.n}, " f"presence_penalty={self.presence_penalty}, " f"frequency_penalty={self.frequency_penalty}, " f"repetition_penalty={self.repetition_penalty}, " f"temperature={self.temperature}, " f"top_p={self.top_p}, " f"top_k={self.top_k}, " f"min_p={self.min_p}, " f"seed={self.seed}, " f"stop={self.stop}, " f"stop_token_ids={self.stop_token_ids}, " f"bad_words={self.bad_words}, " f"include_stop_str_in_output={self.include_stop_str_in_output}, " f"ignore_eos={self.ignore_eos}, " f"max_tokens={self.max_tokens}, " f"min_tokens={self.min_tokens}, " f"logprobs={self.logprobs}, " f"prompt_logprobs={self.prompt_logprobs}, " f"skip_special_tokens={self.skip_special_tokens}, " "spaces_between_special_tokens=" f"{self.spaces_between_special_tokens}, " f"truncate_prompt_tokens={self.truncate_prompt_tokens}, " f"structured_outputs={self.structured_outputs}, " f"extra_args={self.extra_args})" ) class BeamSearchParams( msgspec.Struct, omit_defaults=True, # type: ignore[call-arg] # required for @cached_property. dict=True, ): # type: ignore[call-arg] """Beam search parameters for text generation.""" beam_width: int max_tokens: int ignore_eos: bool = False temperature: float = 0.0 length_penalty: float = 1.0 include_stop_str_in_output: bool = False
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/third_party/__init__.py
vllm/third_party/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/third_party/pynvml.py
vllm/third_party/pynvml.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # copied from https://pypi.org/project/nvidia-ml-py # version 12.570.86 ##### # Copyright (c) 2011-2023, NVIDIA Corporation. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright notice, # this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # * Neither the name of the NVIDIA Corporation nor the names of its # contributors may be used to endorse or promote products derived from # this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE # ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE # LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR # CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF # SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS # INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN # CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) # ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF # THE POSSIBILITY OF SUCH DAMAGE. ##### ## # Python bindings for the NVML library ## from ctypes import * from ctypes.util import find_library from functools import wraps import sys import os import threading import string ## C Type mappings ## ## Enums _nvmlEnableState_t = c_uint NVML_FEATURE_DISABLED = 0 NVML_FEATURE_ENABLED = 1 _nvmlBrandType_t = c_uint NVML_BRAND_UNKNOWN = 0 NVML_BRAND_QUADRO = 1 NVML_BRAND_TESLA = 2 NVML_BRAND_NVS = 3 NVML_BRAND_GRID = 4 # Deprecated from API reporting. Keeping definition for backward compatibility. NVML_BRAND_GEFORCE = 5 NVML_BRAND_TITAN = 6 NVML_BRAND_NVIDIA_VAPPS = 7 # NVIDIA Virtual Applications NVML_BRAND_NVIDIA_VPC = 8 # NVIDIA Virtual PC NVML_BRAND_NVIDIA_VCS = 9 # NVIDIA Virtual Compute Server NVML_BRAND_NVIDIA_VWS = 10 # NVIDIA RTX Virtual Workstation NVML_BRAND_NVIDIA_CLOUD_GAMING = 11 # NVIDIA Cloud Gaming NVML_BRAND_NVIDIA_VGAMING = NVML_BRAND_NVIDIA_CLOUD_GAMING # Deprecated from API reporting. Keeping definition for backward compatibility. NVML_BRAND_QUADRO_RTX = 12 NVML_BRAND_NVIDIA_RTX = 13 NVML_BRAND_NVIDIA = 14 NVML_BRAND_GEFORCE_RTX = 15 # Unused NVML_BRAND_TITAN_RTX = 16 # Unused NVML_BRAND_COUNT = 17 _nvmlTemperatureThresholds_t = c_uint NVML_TEMPERATURE_THRESHOLD_SHUTDOWN = 0 NVML_TEMPERATURE_THRESHOLD_SLOWDOWN = 1 NVML_TEMPERATURE_THRESHOLD_MEM_MAX = 2 NVML_TEMPERATURE_THRESHOLD_GPU_MAX = 3 NVML_TEMPERATURE_THRESHOLD_ACOUSTIC_MIN = 4 NVML_TEMPERATURE_THRESHOLD_ACOUSTIC_CURR = 5 NVML_TEMPERATURE_THRESHOLD_ACOUSTIC_MAX = 6 NVML_TEMPERATURE_THRESHOLD_GPS_CURR = 7 NVML_TEMPERATURE_THRESHOLD_COUNT = 8 _nvmlTemperatureSensors_t = c_uint NVML_TEMPERATURE_GPU = 0 NVML_TEMPERATURE_COUNT = 1 _nvmlComputeMode_t = c_uint NVML_COMPUTEMODE_DEFAULT = 0 NVML_COMPUTEMODE_EXCLUSIVE_THREAD = 1 ## Support Removed NVML_COMPUTEMODE_PROHIBITED = 2 NVML_COMPUTEMODE_EXCLUSIVE_PROCESS = 3 NVML_COMPUTEMODE_COUNT = 4 _nvmlMemoryLocation_t = c_uint NVML_MEMORY_LOCATION_L1_CACHE = 0 NVML_MEMORY_LOCATION_L2_CACHE = 1 NVML_MEMORY_LOCATION_DEVICE_MEMORY = 2 NVML_MEMORY_LOCATION_DRAM = 2 NVML_MEMORY_LOCATION_REGISTER_FILE = 3 NVML_MEMORY_LOCATION_TEXTURE_MEMORY = 4 NVML_MEMORY_LOCATION_TEXTURE_SHM = 5 NVML_MEMORY_LOCATION_CBU = 6 NVML_MEMORY_LOCATION_SRAM = 7 NVML_MEMORY_LOCATION_COUNT = 8 NVML_NVLINK_MAX_LINKS = 18 # For backwards compatibility, maintain the incorrectly-named "LANES" define NVML_NVLINK_MAX_LANES = NVML_NVLINK_MAX_LINKS _nvmlNvLinkErrorCounter_t = c_uint NVML_NVLINK_ERROR_DL_REPLAY = 0 NVML_NVLINK_ERROR_DL_RECOVERY = 1 NVML_NVLINK_ERROR_DL_CRC_FLIT = 2 NVML_NVLINK_ERROR_DL_CRC_DATA = 3 NVML_NVLINK_ERROR_DL_ECC_DATA = 4 NVML_NVLINK_ERROR_COUNT = 5 _nvmlNvLinkEccLaneErrorCounter_t = c_uint NVML_NVLINK_ERROR_DL_ECC_LANE0 = 0 NVML_NVLINK_ERROR_DL_ECC_LANE1 = 1 NVML_NVLINK_ERROR_DL_ECC_LANE2 = 2 NVML_NVLINK_ERROR_DL_ECC_LANE3 = 3 NVML_NVLINK_ERROR_DL_ECC_COUNT = 5 _nvmlNvLinkCapability_t = c_uint NVML_NVLINK_CAP_P2P_SUPPORTED = 0 NVML_NVLINK_CAP_SYSMEM_ACCESS = 1 NVML_NVLINK_CAP_P2P_ATOMICS = 2 NVML_NVLINK_CAP_SYSMEM_ATOMICS= 3 NVML_NVLINK_CAP_SLI_BRIDGE = 4 NVML_NVLINK_CAP_VALID = 5 NVML_NVLINK_CAP_COUNT = 6 _nvmlNvLinkUtilizationCountPktTypes_t = c_uint NVML_NVLINK_COUNTER_PKTFILTER_NOP = 0x1 NVML_NVLINK_COUNTER_PKTFILTER_READ = 0x2 NVML_NVLINK_COUNTER_PKTFILTER_WRITE = 0x4 NVML_NVLINK_COUNTER_PKTFILTER_RATOM = 0x8 NVML_NVLINK_COUNTER_PKTFILTER_NRATOM = 0x10 NVML_NVLINK_COUNTER_PKTFILTER_FLUSH = 0x20 NVML_NVLINK_COUNTER_PKTFILTER_RESPDATA = 0x40 NVML_NVLINK_COUNTER_PKTFILTER_RESPNODATA = 0x80 NVML_NVLINK_COUNTER_PKTFILTER_ALL = 0xFF _nvmlNvLinkUtilizationCountUnits_t = c_uint NVML_NVLINK_COUNTER_UNIT_CYCLES = 0 NVML_NVLINK_COUNTER_UNIT_PACKETS = 1 NVML_NVLINK_COUNTER_UNIT_BYTES = 2 NVML_NVLINK_COUNTER_UNIT_RESERVED = 3 NVML_NVLINK_COUNTER_UNIT_COUNT = 4 _nvmlNvLinkDeviceType_t = c_uint NVML_NVLINK_DEVICE_TYPE_GPU = 0x00 NVML_NVLINK_DEVICE_TYPE_IBMNPU = 0x01 NVML_NVLINK_DEVICE_TYPE_SWITCH = 0x02 NVML_NVLINK_DEVICE_TYPE_UNKNOWN = 0xFF # These are deprecated, instead use _nvmlMemoryErrorType_t _nvmlEccBitType_t = c_uint NVML_SINGLE_BIT_ECC = 0 NVML_DOUBLE_BIT_ECC = 1 NVML_ECC_ERROR_TYPE_COUNT = 2 _nvmlEccCounterType_t = c_uint NVML_VOLATILE_ECC = 0 NVML_AGGREGATE_ECC = 1 NVML_ECC_COUNTER_TYPE_COUNT = 2 _nvmlMemoryErrorType_t = c_uint NVML_MEMORY_ERROR_TYPE_CORRECTED = 0 NVML_MEMORY_ERROR_TYPE_UNCORRECTED = 1 NVML_MEMORY_ERROR_TYPE_COUNT = 2 _nvmlClockType_t = c_uint NVML_CLOCK_GRAPHICS = 0 NVML_CLOCK_SM = 1 NVML_CLOCK_MEM = 2 NVML_CLOCK_VIDEO = 3 NVML_CLOCK_COUNT = 4 _nvmlClockId_t = c_uint NVML_CLOCK_ID_CURRENT = 0 NVML_CLOCK_ID_APP_CLOCK_TARGET = 1 NVML_CLOCK_ID_APP_CLOCK_DEFAULT = 2 NVML_CLOCK_ID_CUSTOMER_BOOST_MAX = 3 NVML_CLOCK_ID_COUNT = 4 _nvmlDriverModel_t = c_uint NVML_DRIVER_WDDM = 0 NVML_DRIVER_WDM = 1 NVML_DRIVER_MCDM = 2 NVML_MAX_GPU_PERF_PSTATES = 16 _nvmlPstates_t = c_uint NVML_PSTATE_0 = 0 NVML_PSTATE_1 = 1 NVML_PSTATE_2 = 2 NVML_PSTATE_3 = 3 NVML_PSTATE_4 = 4 NVML_PSTATE_5 = 5 NVML_PSTATE_6 = 6 NVML_PSTATE_7 = 7 NVML_PSTATE_8 = 8 NVML_PSTATE_9 = 9 NVML_PSTATE_10 = 10 NVML_PSTATE_11 = 11 NVML_PSTATE_12 = 12 NVML_PSTATE_13 = 13 NVML_PSTATE_14 = 14 NVML_PSTATE_15 = 15 NVML_PSTATE_UNKNOWN = 32 _nvmlInforomObject_t = c_uint NVML_INFOROM_OEM = 0 NVML_INFOROM_ECC = 1 NVML_INFOROM_POWER = 2 NVML_INFOROM_DEN = 3 NVML_INFOROM_COUNT = 4 _nvmlReturn_t = c_uint NVML_SUCCESS = 0 NVML_ERROR_UNINITIALIZED = 1 NVML_ERROR_INVALID_ARGUMENT = 2 NVML_ERROR_NOT_SUPPORTED = 3 NVML_ERROR_NO_PERMISSION = 4 NVML_ERROR_ALREADY_INITIALIZED = 5 NVML_ERROR_NOT_FOUND = 6 NVML_ERROR_INSUFFICIENT_SIZE = 7 NVML_ERROR_INSUFFICIENT_POWER = 8 NVML_ERROR_DRIVER_NOT_LOADED = 9 NVML_ERROR_TIMEOUT = 10 NVML_ERROR_IRQ_ISSUE = 11 NVML_ERROR_LIBRARY_NOT_FOUND = 12 NVML_ERROR_FUNCTION_NOT_FOUND = 13 NVML_ERROR_CORRUPTED_INFOROM = 14 NVML_ERROR_GPU_IS_LOST = 15 NVML_ERROR_RESET_REQUIRED = 16 NVML_ERROR_OPERATING_SYSTEM = 17 NVML_ERROR_LIB_RM_VERSION_MISMATCH = 18 NVML_ERROR_IN_USE = 19 NVML_ERROR_MEMORY = 20 NVML_ERROR_NO_DATA = 21 NVML_ERROR_VGPU_ECC_NOT_SUPPORTED = 22 NVML_ERROR_INSUFFICIENT_RESOURCES = 23 NVML_ERROR_FREQ_NOT_SUPPORTED = 24 NVML_ERROR_ARGUMENT_VERSION_MISMATCH = 25 NVML_ERROR_DEPRECATED = 26 NVML_ERROR_NOT_READY = 27 NVML_ERROR_GPU_NOT_FOUND = 28 NVML_ERROR_INVALID_STATE = 29 NVML_ERROR_UNKNOWN = 999 _nvmlFanState_t = c_uint NVML_FAN_NORMAL = 0 NVML_FAN_FAILED = 1 _nvmlFanControlPolicy_t = c_uint NVML_FAN_POLICY_TEMPERATURE_CONTINOUS_SW = 0 NVML_FAN_POLICY_MANUAL = 1 _nvmlLedColor_t = c_uint NVML_LED_COLOR_GREEN = 0 NVML_LED_COLOR_AMBER = 1 _nvmlGpuOperationMode_t = c_uint NVML_GOM_ALL_ON = 0 NVML_GOM_COMPUTE = 1 NVML_GOM_LOW_DP = 2 _nvmlPageRetirementCause_t = c_uint NVML_PAGE_RETIREMENT_CAUSE_MULTIPLE_SINGLE_BIT_ECC_ERRORS = 0 NVML_PAGE_RETIREMENT_CAUSE_DOUBLE_BIT_ECC_ERROR = 1 NVML_PAGE_RETIREMENT_CAUSE_COUNT = 2 _nvmlRestrictedAPI_t = c_uint NVML_RESTRICTED_API_SET_APPLICATION_CLOCKS = 0 NVML_RESTRICTED_API_SET_AUTO_BOOSTED_CLOCKS = 1 NVML_RESTRICTED_API_COUNT = 2 _nvmlBridgeChipType_t = c_uint NVML_BRIDGE_CHIP_PLX = 0 NVML_BRIDGE_CHIP_BRO4 = 1 NVML_MAX_PHYSICAL_BRIDGE = 128 _nvmlValueType_t = c_uint NVML_VALUE_TYPE_DOUBLE = 0 NVML_VALUE_TYPE_UNSIGNED_INT = 1 NVML_VALUE_TYPE_UNSIGNED_LONG = 2 NVML_VALUE_TYPE_UNSIGNED_LONG_LONG = 3 NVML_VALUE_TYPE_SIGNED_LONG_LONG = 4 NVML_VALUE_TYPE_SIGNED_INT = 5 NVML_VALUE_TYPE_UNSIGNED_SHORT = 6 NVML_VALUE_TYPE_COUNT = 7 _nvmlNvlinkVersion_t = c_uint NVML_NVLINK_VERSION_INVALID = 0 NVML_NVLINK_VERSION_1_0 = 1 NVML_NVLINK_VERSION_2_0 = 2 NVML_NVLINK_VERSION_2_2 = 3 NVML_NVLINK_VERSION_3_0 = 4 NVML_NVLINK_VERSION_3_1 = 5 NVML_NVLINK_VERSION_4_0 = 6 NVML_NVLINK_VERSION_5_0 = 7 _nvmlPerfPolicyType_t = c_uint NVML_PERF_POLICY_POWER = 0 NVML_PERF_POLICY_THERMAL = 1 NVML_PERF_POLICY_SYNC_BOOST = 2 NVML_PERF_POLICY_BOARD_LIMIT = 3 NVML_PERF_POLICY_LOW_UTILIZATION = 4 NVML_PERF_POLICY_RELIABILITY = 5 NVML_PERF_POLICY_TOTAL_APP_CLOCKS = 10 NVML_PERF_POLICY_TOTAL_BASE_CLOCKS = 11 NVML_PERF_POLICY_COUNT = 12 _nvmlEncoderQueryType_t = c_uint NVML_ENCODER_QUERY_H264 = 0 NVML_ENCODER_QUERY_HEVC = 1 NVML_ENCODER_QUERY_AV1 = 2 NVML_ENCODER_QUERY_UNKNOWN = 255 _nvmlFBCSessionType_t = c_uint NVML_FBC_SESSION_TYPE_UNKNOWN = 0 NVML_FBC_SESSION_TYPE_TOSYS = 1 NVML_FBC_SESSION_TYPE_CUDA = 2 NVML_FBC_SESSION_TYPE_VID = 3 NVML_FBC_SESSION_TYPE_HWENC = 4 _nvmlDetachGpuState_t = c_uint NVML_DETACH_GPU_KEEP = 0 NVML_DETACH_GPU_REMOVE = 1 _nvmlPcieLinkState_t = c_uint NVML_PCIE_LINK_KEEP = 0 NVML_PCIE_LINK_SHUT_DOWN = 1 _nvmlSamplingType_t = c_uint NVML_TOTAL_POWER_SAMPLES = 0 NVML_GPU_UTILIZATION_SAMPLES = 1 NVML_MEMORY_UTILIZATION_SAMPLES = 2 NVML_ENC_UTILIZATION_SAMPLES = 3 NVML_DEC_UTILIZATION_SAMPLES = 4 NVML_PROCESSOR_CLK_SAMPLES = 5 NVML_MEMORY_CLK_SAMPLES = 6 NVML_MODULE_POWER_SAMPLES = 7 NVML_JPG_UTILIZATION_SAMPLES = 8 NVML_OFA_UTILIZATION_SAMPLES = 9 NVML_SAMPLINGTYPE_COUNT = 10 _nvmlPcieUtilCounter_t = c_uint NVML_PCIE_UTIL_TX_BYTES = 0 NVML_PCIE_UTIL_RX_BYTES = 1 NVML_PCIE_UTIL_COUNT = 2 _nvmlGpuTopologyLevel_t = c_uint NVML_TOPOLOGY_INTERNAL = 0 NVML_TOPOLOGY_SINGLE = 10 NVML_TOPOLOGY_MULTIPLE = 20 NVML_TOPOLOGY_HOSTBRIDGE = 30 NVML_TOPOLOGY_NODE = 40 NVML_TOPOLOGY_CPU = NVML_TOPOLOGY_NODE NVML_TOPOLOGY_SYSTEM = 50 _nvmlGpuP2PCapsIndex_t = c_uint NVML_P2P_CAPS_INDEX_READ = 0, NVML_P2P_CAPS_INDEX_WRITE = 1 NVML_P2P_CAPS_INDEX_NVLINK =2 NVML_P2P_CAPS_INDEX_ATOMICS = 3 # # NVML_P2P_CAPS_INDEX_PROP is deprecated. # Use NVML_P2P_CAPS_INDEX_PCI instead. # NVML_P2P_CAPS_INDEX_PROP = 4 NVML_P2P_CAPS_INDEX_PCI = 4 NVML_P2P_CAPS_INDEX_UNKNOWN = 5 _nvmlGpuP2PStatus_t = c_uint NVML_P2P_STATUS_OK = 0 NVML_P2P_STATUS_CHIPSET_NOT_SUPPORED = 1 NVML_P2P_STATUS_CHIPSET_NOT_SUPPORTED = NVML_P2P_STATUS_CHIPSET_NOT_SUPPORED NVML_P2P_STATUS_GPU_NOT_SUPPORTED = 2 NVML_P2P_STATUS_IOH_TOPOLOGY_NOT_SUPPORTED =3 NVML_P2P_STATUS_DISABLED_BY_REGKEY =4 NVML_P2P_STATUS_NOT_SUPPORTED =5 NVML_P2P_STATUS_UNKNOWN =6 _nvmlDeviceArchitecture_t = c_uint NVML_DEVICE_ARCH_KEPLER = 2 NVML_DEVICE_ARCH_MAXWELL = 3 NVML_DEVICE_ARCH_PASCAL = 4 NVML_DEVICE_ARCH_VOLTA = 5 NVML_DEVICE_ARCH_TURING = 6 NVML_DEVICE_ARCH_AMPERE = 7 NVML_DEVICE_ARCH_ADA = 8 NVML_DEVICE_ARCH_HOPPER = 9 NVML_DEVICE_ARCH_BLACKWELL = 10 NVML_DEVICE_ARCH_T23X = 11 NVML_DEVICE_ARCH_UNKNOWN = 0xffffffff # PCI bus Types _nvmlBusType_t = c_uint NVML_BUS_TYPE_UNKNOWN = 0 NVML_BUS_TYPE_PCI = 1 NVML_BUS_TYPE_PCIE = 2 NVML_BUS_TYPE_FPCI = 3 NVML_BUS_TYPE_AGP = 4 _nvmlPowerSource_t = c_uint NVML_POWER_SOURCE_AC = 0x00000000 NVML_POWER_SOURCE_BATTERY = 0x00000001 NVML_POWER_SOURCE_UNDERSIZED = 0x00000002 _nvmlAdaptiveClockInfoStatus_t = c_uint NVML_ADAPTIVE_CLOCKING_INFO_STATUS_DISABLED = 0x00000000 NVML_ADAPTIVE_CLOCKING_INFO_STATUS_ENABLED = 0x00000001 _nvmlClockLimitId_t = c_uint NVML_CLOCK_LIMIT_ID_RANGE_START = 0xffffff00 NVML_CLOCK_LIMIT_ID_TDP = 0xffffff01 NVML_CLOCK_LIMIT_ID_UNLIMITED = 0xffffff02 _nvmlPcieLinkMaxSpeed_t = c_uint NVML_PCIE_LINK_MAX_SPEED_INVALID = 0x00000000 NVML_PCIE_LINK_MAX_SPEED_2500MBPS = 0x00000001 NVML_PCIE_LINK_MAX_SPEED_5000MBPS = 0x00000002 NVML_PCIE_LINK_MAX_SPEED_8000MBPS = 0x00000003 NVML_PCIE_LINK_MAX_SPEED_16000MBPS = 0x00000004 NVML_PCIE_LINK_MAX_SPEED_32000MBPS = 0x00000005 NVML_PCIE_LINK_MAX_SPEED_64000MBPS = 0x00000006 _nvmlPcieAtomicsCapability_t = c_uint NVML_PCIE_ATOMICS_CAP_FETCHADD32 = 0x01 NVML_PCIE_ATOMICS_CAP_FETCHADD64 = 0x02 NVML_PCIE_ATOMICS_CAP_SWAP32 = 0x04 NVML_PCIE_ATOMICS_CAP_SWAP64 = 0x08 NVML_PCIE_ATOMICS_CAP_CAS32 = 0x10 NVML_PCIE_ATOMICS_CAP_CAS64 = 0x20 NVML_PCIE_ATOMICS_CAP_CAS128 = 0x40 NVML_PCIE_ATOMICS_OPS_MAX = 7 _nvmlAffinityScope_t = c_uint NVML_AFFINITY_SCOPE_NODE = 0 NVML_AFFINITY_SCOPE_SOCKET = 1 _nvmlDeviceGpuRecoveryAction_t = c_uint NVML_GPU_RECOVERY_ACTION_NONE = 0 NVML_GPU_RECOVERY_ACTION_GPU_RESET = 1 NVML_GPU_RECOVERY_ACTION_NODE_REBOOT = 2 NVML_GPU_RECOVERY_ACTION_DRAIN_P2P = 3 NVML_GPU_RECOVERY_ACTION_DRAIN_AND_RESET = 4 # C preprocessor defined values nvmlFlagDefault = 0 nvmlFlagForce = 1 NVML_INIT_FLAG_NO_GPUS = 1 NVML_INIT_FLAG_NO_ATTACH = 2 NVML_MAX_GPC_COUNT = 32 # buffer size NVML_DEVICE_INFOROM_VERSION_BUFFER_SIZE = 16 NVML_DEVICE_UUID_BUFFER_SIZE = 80 NVML_DEVICE_UUID_V2_BUFFER_SIZE = 96 NVML_SYSTEM_DRIVER_VERSION_BUFFER_SIZE = 80 NVML_SYSTEM_NVML_VERSION_BUFFER_SIZE = 80 NVML_DEVICE_NAME_BUFFER_SIZE = 64 NVML_DEVICE_NAME_V2_BUFFER_SIZE = 96 NVML_DEVICE_SERIAL_BUFFER_SIZE = 30 NVML_DEVICE_PART_NUMBER_BUFFER_SIZE = 80 NVML_DEVICE_GPU_PART_NUMBER_BUFFER_SIZE = 80 NVML_DEVICE_VBIOS_VERSION_BUFFER_SIZE = 32 NVML_DEVICE_PCI_BUS_ID_BUFFER_SIZE = 32 NVML_DEVICE_PCI_BUS_ID_BUFFER_V2_SIZE = 16 NVML_GRID_LICENSE_BUFFER_SIZE = 128 NVML_VGPU_NAME_BUFFER_SIZE = 64 NVML_GRID_LICENSE_FEATURE_MAX_COUNT = 3 NVML_VGPU_METADATA_OPAQUE_DATA_SIZE = sizeof(c_uint) + 256 NVML_VGPU_PGPU_METADATA_OPAQUE_DATA_SIZE = 256 NVML_DEVICE_GPU_FRU_PART_NUMBER_BUFFER_SIZE = 0x14 # NV2080_GPU_MAX_PRODUCT_PART_NUMBER_LENGTH NVML_PERF_MODES_BUFFER_SIZE = 2048 # Format strings NVML_DEVICE_PCI_BUS_ID_LEGACY_FMT = "%04X:%02X:%02X.0" NVML_DEVICE_PCI_BUS_ID_FMT = "%08X:%02X:%02X.0" NVML_VALUE_NOT_AVAILABLE_ulonglong = c_ulonglong(-1) NVML_VALUE_NOT_AVAILABLE_uint = c_uint(-1) ''' Field Identifiers. All Identifiers pertain to a device. Each ID is only used once and is guaranteed never to change. ''' NVML_FI_DEV_ECC_CURRENT = 1 # Current ECC mode. 1=Active. 0=Inactive NVML_FI_DEV_ECC_PENDING = 2 # Pending ECC mode. 1=Active. 0=Inactive #ECC Count Totals NVML_FI_DEV_ECC_SBE_VOL_TOTAL = 3 # Total single bit volatile ECC errors NVML_FI_DEV_ECC_DBE_VOL_TOTAL = 4 # Total double bit volatile ECC errors NVML_FI_DEV_ECC_SBE_AGG_TOTAL = 5 # Total single bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_DBE_AGG_TOTAL = 6 # Total double bit aggregate (persistent) ECC errors #Individual ECC locations NVML_FI_DEV_ECC_SBE_VOL_L1 = 7 # L1 cache single bit volatile ECC errors NVML_FI_DEV_ECC_DBE_VOL_L1 = 8 # L1 cache double bit volatile ECC errors NVML_FI_DEV_ECC_SBE_VOL_L2 = 9 # L2 cache single bit volatile ECC errors NVML_FI_DEV_ECC_DBE_VOL_L2 = 10 # L2 cache double bit volatile ECC errors NVML_FI_DEV_ECC_SBE_VOL_DEV = 11 # Device memory single bit volatile ECC errors NVML_FI_DEV_ECC_DBE_VOL_DEV = 12 # Device memory double bit volatile ECC errors NVML_FI_DEV_ECC_SBE_VOL_REG = 13 # Register file single bit volatile ECC errors NVML_FI_DEV_ECC_DBE_VOL_REG = 14 # Register file double bit volatile ECC errors NVML_FI_DEV_ECC_SBE_VOL_TEX = 15 # Texture memory single bit volatile ECC errors NVML_FI_DEV_ECC_DBE_VOL_TEX = 16 # Texture memory double bit volatile ECC errors NVML_FI_DEV_ECC_DBE_VOL_CBU = 17 # CBU double bit volatile ECC errors NVML_FI_DEV_ECC_SBE_AGG_L1 = 18 # L1 cache single bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_DBE_AGG_L1 = 19 # L1 cache double bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_SBE_AGG_L2 = 20 # L2 cache single bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_DBE_AGG_L2 = 21 # L2 cache double bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_SBE_AGG_DEV = 22 # Device memory single bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_DBE_AGG_DEV = 23 # Device memory double bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_SBE_AGG_REG = 24 # Register File single bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_DBE_AGG_REG = 25 # Register File double bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_SBE_AGG_TEX = 26 # Texture memory single bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_DBE_AGG_TEX = 27 # Texture memory double bit aggregate (persistent) ECC errors NVML_FI_DEV_ECC_DBE_AGG_CBU = 28 # CBU double bit aggregate ECC errors # Page Retirement NVML_FI_DEV_RETIRED_SBE = 29 # Number of retired pages because of single bit errors NVML_FI_DEV_RETIRED_DBE = 30 # Number of retired pages because of double bit errors NVML_FI_DEV_RETIRED_PENDING = 31 # If any pages are pending retirement. 1=yes. 0=no. # NvLink Flit Error Counters NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L0 = 32 # NVLink flow control CRC Error Counter for Lane 0 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L1 = 33 # NVLink flow control CRC Error Counter for Lane 1 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L2 = 34 # NVLink flow control CRC Error Counter for Lane 2 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L3 = 35 # NVLink flow control CRC Error Counter for Lane 3 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L4 = 36 # NVLink flow control CRC Error Counter for Lane 4 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L5 = 37 # NVLink flow control CRC Error Counter for Lane 5 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_TOTAL = 38 # NVLink flow control CRC Error Counter total for all Lanes # NvLink CRC Data Error Counters NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L0 = 39 # NVLink data CRC Error Counter for Lane 0 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L1 = 40 # NVLink data CRC Error Counter for Lane 1 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L2 = 41 # NVLink data CRC Error Counter for Lane 2 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L3 = 42 # NVLink data CRC Error Counter for Lane 3 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L4 = 43 # NVLink data CRC Error Counter for Lane 4 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L5 = 44 # NVLink data CRC Error Counter for Lane 5 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_TOTAL = 45 # NvLink data CRC Error Counter total for all Lanes # NvLink Replay Error Counters NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L0 = 46 # NVLink Replay Error Counter for Lane 0 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L1 = 47 # NVLink Replay Error Counter for Lane 1 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L2 = 48 # NVLink Replay Error Counter for Lane 2 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L3 = 49 # NVLink Replay Error Counter for Lane 3 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L4 = 50 # NVLink Replay Error Counter for Lane 4 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L5 = 51 # NVLink Replay Error Counter for Lane 5 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_TOTAL = 52 # NVLink Replay Error Counter total for all Lanes # NvLink Recovery Error Counters NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L0 = 53 # NVLink Recovery Error Counter for Lane 0 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L1 = 54 # NVLink Recovery Error Counter for Lane 1 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L2 = 55 # NVLink Recovery Error Counter for Lane 2 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L3 = 56 # NVLink Recovery Error Counter for Lane 3 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L4 = 57 # NVLink Recovery Error Counter for Lane 4 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L5 = 58 # NVLink Recovery Error Counter for Lane 5 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_TOTAL = 59 # NVLink Recovery Error Counter total for all Lanes # NvLink Bandwidth Counters NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L0 = 60 # NVLink Bandwidth Counter for Counter Set 0, Lane 0 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L1 = 61 # NVLink Bandwidth Counter for Counter Set 0, Lane 1 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L2 = 62 # NVLink Bandwidth Counter for Counter Set 0, Lane 2 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L3 = 63 # NVLink Bandwidth Counter for Counter Set 0, Lane 3 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L4 = 64 # NVLink Bandwidth Counter for Counter Set 0, Lane 4 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L5 = 65 # NVLink Bandwidth Counter for Counter Set 0, Lane 5 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_TOTAL = 66 # NVLink Bandwidth Counter Total for Counter Set 0, All Lanes # NvLink Bandwidth Counters NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L0 = 67 # NVLink Bandwidth Counter for Counter Set 1, Lane 0 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L1 = 68 # NVLink Bandwidth Counter for Counter Set 1, Lane 1 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L2 = 69 # NVLink Bandwidth Counter for Counter Set 1, Lane 2 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L3 = 70 # NVLink Bandwidth Counter for Counter Set 1, Lane 3 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L4 = 71 # NVLink Bandwidth Counter for Counter Set 1, Lane 4 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L5 = 72 # NVLink Bandwidth Counter for Counter Set 1, Lane 5 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_TOTAL = 73 # NVLink Bandwidth Counter Total for Counter Set 1, All Lanes # Perf Policy Counters NVML_FI_DEV_PERF_POLICY_POWER = 74 # Perf Policy Counter for Power Policy NVML_FI_DEV_PERF_POLICY_THERMAL = 75 # Perf Policy Counter for Thermal Policy NVML_FI_DEV_PERF_POLICY_SYNC_BOOST = 76 # Perf Policy Counter for Sync boost Policy NVML_FI_DEV_PERF_POLICY_BOARD_LIMIT = 77 # Perf Policy Counter for Board Limit NVML_FI_DEV_PERF_POLICY_LOW_UTILIZATION = 78 # Perf Policy Counter for Low GPU Utilization Policy NVML_FI_DEV_PERF_POLICY_RELIABILITY = 79 # Perf Policy Counter for Reliability Policy NVML_FI_DEV_PERF_POLICY_TOTAL_APP_CLOCKS = 80 # Perf Policy Counter for Total App Clock Policy NVML_FI_DEV_PERF_POLICY_TOTAL_BASE_CLOCKS = 81 # Perf Policy Counter for Total Base Clocks Policy # Memory temperatures NVML_FI_DEV_MEMORY_TEMP = 82 # Memory temperature for the device # Energy Counter NVML_FI_DEV_TOTAL_ENERGY_CONSUMPTION = 83 # Total energy consumption for the GPU in mJ since the driver was last reloaded # NVLink Speed NVML_FI_DEV_NVLINK_SPEED_MBPS_L0 = 84 NVML_FI_DEV_NVLINK_SPEED_MBPS_L1 = 85 NVML_FI_DEV_NVLINK_SPEED_MBPS_L2 = 86 NVML_FI_DEV_NVLINK_SPEED_MBPS_L3 = 87 NVML_FI_DEV_NVLINK_SPEED_MBPS_L4 = 88 NVML_FI_DEV_NVLINK_SPEED_MBPS_L5 = 89 NVML_FI_DEV_NVLINK_SPEED_MBPS_COMMON = 90 # NVLink Link Count NVML_FI_DEV_NVLINK_LINK_COUNT = 91 # Page Retirement pending fields NVML_FI_DEV_RETIRED_PENDING_SBE = 92 NVML_FI_DEV_RETIRED_PENDING_DBE = 93 # PCIe replay and replay rollover counters NVML_FI_DEV_PCIE_REPLAY_COUNTER = 94 NVML_FI_DEV_PCIE_REPLAY_ROLLOVER_COUNTER = 95 # NvLink Flit Error Counters NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L6 = 96 # NVLink flow control CRC Error Counter for Lane 6 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L7 = 97 # NVLink flow control CRC Error Counter for Lane 7 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L8 = 98 # NVLink flow control CRC Error Counter for Lane 8 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L9 = 99 # NVLink flow control CRC Error Counter for Lane 9 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L10 = 100 # NVLink flow control CRC Error Counter for Lane 10 NVML_FI_DEV_NVLINK_CRC_FLIT_ERROR_COUNT_L11 = 101 # NVLink flow control CRC Error Counter for Lane 11 # NvLink CRC Data Error Counters NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L6 = 102 # NVLink data CRC Error Counter for Lane 6 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L7 = 103 # NVLink data CRC Error Counter for Lane 7 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L8 = 104 # NVLink data CRC Error Counter for Lane 8 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L9 = 105 # NVLink data CRC Error Counter for Lane 9 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L10 = 106 # NVLink data CRC Error Counter for Lane 10 NVML_FI_DEV_NVLINK_CRC_DATA_ERROR_COUNT_L11 = 107 # NVLink data CRC Error Counter for Lane 11 # NvLink Replay Error Counters NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L6 = 108 # NVLink Replay Error Counter for Lane 6 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L7 = 109 # NVLink Replay Error Counter for Lane 7 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L8 = 110 # NVLink Replay Error Counter for Lane 8 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L9 = 111 # NVLink Replay Error Counter for Lane 9 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L10 = 112 # NVLink Replay Error Counter for Lane 10 NVML_FI_DEV_NVLINK_REPLAY_ERROR_COUNT_L11 = 113 # NVLink Replay Error Counter for Lane 11 # NvLink Recovery Error Counters NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L6 = 114 # NVLink Recovery Error Counter for Lane 6 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L7 = 115 # NVLink Recovery Error Counter for Lane 7 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L8 = 116 # NVLink Recovery Error Counter for Lane 8 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L9 = 117 # NVLink Recovery Error Counter for Lane 9 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L10 = 118 # NVLink Recovery Error Counter for Lane 10 NVML_FI_DEV_NVLINK_RECOVERY_ERROR_COUNT_L11 = 119 # NVLink Recovery Error Counter for Lane 11 # NvLink Bandwidth Counters NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L6 = 120 # NVLink Bandwidth Counter for Counter Set 0, Lane 6 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L7 = 121 # NVLink Bandwidth Counter for Counter Set 0, Lane 7 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L8 = 122 # NVLink Bandwidth Counter for Counter Set 0, Lane 8 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L9 = 123 # NVLink Bandwidth Counter for Counter Set 0, Lane 9 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L10 = 124 # NVLink Bandwidth Counter for Counter Set 0, Lane 10 NVML_FI_DEV_NVLINK_BANDWIDTH_C0_L11 = 125 # NVLink Bandwidth Counter for Counter Set 0, Lane 11 # NvLink Bandwidth Counters NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L6 = 126 # NVLink Bandwidth Counter for Counter Set 1, Lane 6 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L7 = 127 # NVLink Bandwidth Counter for Counter Set 1, Lane 7 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L8 = 128 # NVLink Bandwidth Counter for Counter Set 1, Lane 8 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L9 = 129 # NVLink Bandwidth Counter for Counter Set 1, Lane 9 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L10 = 130 # NVLink Bandwidth Counter for Counter Set 1, Lane 10 NVML_FI_DEV_NVLINK_BANDWIDTH_C1_L11 = 131 # NVLink Bandwidth Counter for Counter Set 1, Lane 11 # NVLink Speed NVML_FI_DEV_NVLINK_SPEED_MBPS_L6 = 132 NVML_FI_DEV_NVLINK_SPEED_MBPS_L7 = 133 NVML_FI_DEV_NVLINK_SPEED_MBPS_L8 = 134 NVML_FI_DEV_NVLINK_SPEED_MBPS_L9 = 135 NVML_FI_DEV_NVLINK_SPEED_MBPS_L10 = 136 NVML_FI_DEV_NVLINK_SPEED_MBPS_L11 = 137 # NVLink Throughput Counters NVML_FI_DEV_NVLINK_THROUGHPUT_DATA_TX = 138 # NVLink TX Data throughput in KiB NVML_FI_DEV_NVLINK_THROUGHPUT_DATA_RX = 139 # NVLink RX Data throughput in KiB NVML_FI_DEV_NVLINK_THROUGHPUT_RAW_TX = 140 # NVLink TX Data + protocol overhead in KiB NVML_FI_DEV_NVLINK_THROUGHPUT_RAW_RX = 141 # NVLink RX Data + protocol overhead in KiB # Row Remapper NVML_FI_DEV_REMAPPED_COR = 142 NVML_FI_DEV_REMAPPED_UNC = 143 NVML_FI_DEV_REMAPPED_PENDING = 144 NVML_FI_DEV_REMAPPED_FAILURE = 145 #Remote device NVLink ID NVML_FI_DEV_NVLINK_REMOTE_NVLINK_ID = 146 # Number of NVLinks connected to NVSwitch NVML_FI_DEV_NVSWITCH_CONNECTED_LINK_COUNT = 147 # NvLink ECC Data Error Counters NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L0 = 148 #< NVLink data ECC Error Counter for Link 0 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L1 = 149 #< NVLink data ECC Error Counter for Link 1 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L2 = 150 #< NVLink data ECC Error Counter for Link 2 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L3 = 151 #< NVLink data ECC Error Counter for Link 3 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L4 = 152 #< NVLink data ECC Error Counter for Link 4 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L5 = 153 #< NVLink data ECC Error Counter for Link 5 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L6 = 154 #< NVLink data ECC Error Counter for Link 6 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L7 = 155 #< NVLink data ECC Error Counter for Link 7 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L8 = 156 #< NVLink data ECC Error Counter for Link 8 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L9 = 157 #< NVLink data ECC Error Counter for Link 9 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L10 = 158 #< NVLink data ECC Error Counter for Link 10 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_L11 = 159 #< NVLink data ECC Error Counter for Link 11 NVML_FI_DEV_NVLINK_ECC_DATA_ERROR_COUNT_TOTAL = 160 #< NvLink data ECC Error Counter total for all Links NVML_FI_DEV_NVLINK_ERROR_DL_REPLAY = 161 NVML_FI_DEV_NVLINK_ERROR_DL_RECOVERY = 162 NVML_FI_DEV_NVLINK_ERROR_DL_CRC = 163 NVML_FI_DEV_NVLINK_GET_SPEED = 164 NVML_FI_DEV_NVLINK_GET_STATE = 165 NVML_FI_DEV_NVLINK_GET_VERSION = 166 NVML_FI_DEV_NVLINK_GET_POWER_STATE = 167 NVML_FI_DEV_NVLINK_GET_POWER_THRESHOLD = 168 NVML_FI_DEV_PCIE_L0_TO_RECOVERY_COUNTER = 169 NVML_FI_DEV_C2C_LINK_COUNT = 170 NVML_FI_DEV_C2C_LINK_GET_STATUS = 171 NVML_FI_DEV_C2C_LINK_GET_MAX_BW = 172 NVML_FI_DEV_PCIE_COUNT_CORRECTABLE_ERRORS = 173 NVML_FI_DEV_PCIE_COUNT_NAKS_RECEIVED = 174 NVML_FI_DEV_PCIE_COUNT_RECEIVER_ERROR = 175 NVML_FI_DEV_PCIE_COUNT_BAD_TLP = 176 NVML_FI_DEV_PCIE_COUNT_NAKS_SENT = 177 NVML_FI_DEV_PCIE_COUNT_BAD_DLLP = 178 NVML_FI_DEV_PCIE_COUNT_NON_FATAL_ERROR = 179 NVML_FI_DEV_PCIE_COUNT_FATAL_ERROR = 180 NVML_FI_DEV_PCIE_COUNT_UNSUPPORTED_REQ = 181 NVML_FI_DEV_PCIE_COUNT_LCRC_ERROR = 182 NVML_FI_DEV_PCIE_COUNT_LANE_ERROR = 183 NVML_FI_DEV_IS_RESETLESS_MIG_SUPPORTED = 184 NVML_FI_DEV_POWER_AVERAGE = 185 NVML_FI_DEV_POWER_INSTANT = 186 NVML_FI_DEV_POWER_MIN_LIMIT = 187 NVML_FI_DEV_POWER_MAX_LIMIT = 188 NVML_FI_DEV_POWER_DEFAULT_LIMIT = 189
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/layer.py
vllm/attention/layer.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Attention layer.""" from typing import cast import torch import torch.nn as nn import vllm.envs as envs from vllm.attention.backends.abstract import ( AttentionBackend, AttentionType, MLAAttentionImpl, ) from vllm.attention.backends.registry import AttentionBackendEnum from vllm.attention.selector import get_attn_backend from vllm.attention.utils.kv_sharing_utils import validate_kv_sharing_target from vllm.attention.utils.kv_transfer_utils import maybe_transfer_kv_layer from vllm.config import CacheConfig, get_current_vllm_config from vllm.config.vllm import VllmConfig from vllm.forward_context import ForwardContext, get_forward_context from vllm.logger import init_logger from vllm.model_executor.layers.attention_layer_base import AttentionLayerBase from vllm.model_executor.layers.batch_invariant import vllm_is_batch_invariant from vllm.model_executor.layers.linear import ( ColumnParallelLinear, UnquantizedLinearMethod, ) from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.model_executor.layers.quantization.base_config import QuantizeMethodBase from vllm.model_executor.layers.quantization.input_quant_fp8 import QuantFP8 from vllm.model_executor.layers.quantization.kv_cache import BaseKVCacheMethod from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape from vllm.platforms import current_platform from vllm.utils.torch_utils import ( direct_register_custom_op, kv_cache_dtype_str_to_dtype, ) from vllm.v1.kv_cache_interface import ( FullAttentionSpec, KVCacheSpec, MLAAttentionSpec, SlidingWindowSpec, ) logger = init_logger(__name__) def should_load_quant_weights(quant_method: QuantizeMethodBase | None) -> bool: """Returns whether the quantization method should load quantized weights.""" return quant_method is not None and not isinstance( quant_method, UnquantizedLinearMethod ) def set_default_quant_scales(layer: nn.Module, register_buffer: bool = False) -> None: """Sets default quantization scales for the layer.""" if register_buffer: layer.register_buffer("_k_scale", torch.tensor(1.0, dtype=torch.float32)) layer.register_buffer("_v_scale", torch.tensor(1.0, dtype=torch.float32)) layer.register_buffer("_q_scale", torch.tensor(1.0, dtype=torch.float32)) layer.register_buffer("_prob_scale", torch.tensor(1.0, dtype=torch.float32)) else: layer._k_scale.fill_(1.0) layer._v_scale.fill_(1.0) layer._q_scale.fill_(1.0) layer._prob_scale.fill_(1.0) # We also keep q/k/v_scale on host (cpu) memory for attention # backends that require the scales to be on host instead of on device. # e.g. Flashinfer layer._q_scale_float = 1.0 layer._k_scale_float = 1.0 layer._v_scale_float = 1.0 layer._prob_scale_float = 1.0 def _init_kv_cache_quant( layer: nn.Module, quant_config: QuantizationConfig | None, prefix: str, kv_cache_dtype: str, calculate_kv_scales: bool, ) -> None: """Initializes KV cache scaling factors and quantization method. This helper function sets up the KV cache quantization attributes that are shared between Attention and MLAAttention layers. It initializes scale tensors for query, key, value, and probability, and configures the quantization method if applicable. Args: layer: The attention layer instance to initialize. quant_config: Optional quantization configuration. prefix: Layer name prefix for quantization method lookup. kv_cache_dtype: The KV cache data type string. calculate_kv_scales: Whether to calculate KV scales dynamically. """ # The default k/v_scale is set to 1.0. This is ignored # when kv-cache is not fp8, and should be used with # kv-cache in fp8_e5m2. For kv-cache in fp8_e4m3, we # expect the pre-quantized k/v_scale to be loaded along # with the model weights. layer.kv_cache_dtype = kv_cache_dtype layer.calculate_kv_scales = calculate_kv_scales # Note [Register q/k/v/prob scales in state dict] # When calling model.to(device), only parameters/buffers in state dict are # moved. If not registering q/k/v/prob scales in state dict, there would # be an IMA error when a cuda kernel (e.g., quant_fp8) accesses the tensor # on cpu. # Registering in state dict means it interacts with weight loading. One edge # case is when quant_method is None, or quant_method is UnquantizedLinearMethod # (i.e., should_load_quant_weights(quant_method) == False). # In this case, the checkpoint does not have the scales. We need to # initialize the scales to 1.0 and update the scales after weight loading. # This is espectially important when we load dummy weights first (providing # wrong scales) and then load real weights (which misses scales and keeps the # wrong scales from dummy load). set_default_quant_scales(layer, register_buffer=True) # The output scale on host memory. This should be the input scale of # the quant op after this attention layer. layer._o_scale_float = None quant_method = ( quant_config.get_quant_method(layer, prefix=prefix) if quant_config else None ) # See [Note: Register q/k/v/prob scales in state dict] if should_load_quant_weights(quant_method): assert isinstance(quant_method, BaseKVCacheMethod) # TODO (mgoin): kv cache dtype should be specified in the FP8 # checkpoint config and become the "auto" behavior if kv_cache_dtype == "fp8_e5m2": raise ValueError("fp8_e5m2 kv-cache is not supported with fp8 checkpoints.") # If quantization is enabled, we make "k_scale" and "v_scale" # parameters so that it can be loaded from the model checkpoint. # The k/v_scale will then be converted back to native float32 # values after weight loading. layer.quant_method = quant_method layer.quant_method.create_weights(layer) class Attention(nn.Module, AttentionLayerBase): """Attention layer. This class takes query, key, and value tensors as input. The input tensors can either contain prompt tokens or generation tokens. The class does the following: 1. Store the input key and value tensors in the KV cache. 2. Perform (multi-head/multi-query/grouped-query) attention. 3. Return the output tensor. """ def __init__( self, num_heads: int, head_size: int, scale: float, num_kv_heads: int | None = None, alibi_slopes: list[float] | None = None, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, logits_soft_cap: float | None = None, per_layer_sliding_window: int | None = None, prefix: str = "", attn_type: str = AttentionType.DECODER, kv_sharing_target_layer_name: str | None = None, attn_backend: type[AttentionBackend] | None = None, head_size_v: int | None = None, **extra_impl_args, ) -> None: """ The KV cache is stored inside this class and is accessed via `self.kv_cache`. """ super().__init__() if per_layer_sliding_window is not None: # per-layer sliding window sliding_window = per_layer_sliding_window elif cache_config is not None: # model-level sliding window sliding_window = cache_config.sliding_window else: sliding_window = None vllm_config = get_current_vllm_config() if cache_config is not None: kv_cache_dtype = cache_config.cache_dtype block_size = cache_config.block_size calculate_kv_scales = cache_config.calculate_kv_scales else: kv_cache_dtype = "auto" block_size = 16 calculate_kv_scales = False self.kv_cache_torch_dtype = kv_cache_dtype_str_to_dtype( kv_cache_dtype, vllm_config.model_config ) if num_kv_heads is None: num_kv_heads = num_heads assert num_heads % num_kv_heads == 0, ( f"num_heads ({num_heads}) is not divisible by num_kv_heads ({num_kv_heads})" ) self.quant_config = quant_config self.layer_name = prefix # Initialize KV cache quantization attributes _init_kv_cache_quant( self, self.quant_config, self.layer_name, kv_cache_dtype, calculate_kv_scales, ) self.num_heads = num_heads self.head_size = head_size self.head_size_v = self.head_size if head_size_v is None else head_size_v self.num_kv_heads = num_kv_heads self.sliding_window = sliding_window self.has_sink = extra_impl_args.get("sinks") is not None # NOTE: model_config may be None during certain tests model_config = vllm_config.model_config self.use_mm_prefix = model_config is not None and model_config.is_mm_prefix_lm # During model initialization, the default dtype is set as the model # weight and activation dtype. dtype = torch.get_default_dtype() if attn_backend is None: self.attn_backend = get_attn_backend( head_size, dtype, kv_cache_dtype, block_size, use_mla=False, has_sink=self.has_sink, use_mm_prefix=self.use_mm_prefix, attn_type=attn_type, ) else: self.attn_backend = attn_backend # prefix caching + batch invariance is currently not supported for # FLASHINFER and TRITON_MLA. if ( cache_config is not None and cache_config.enable_prefix_caching and vllm_is_batch_invariant() and ( self.attn_backend.get_name() == "FLASHINFER" or self.attn_backend.get_name() == "TRITON_MLA" ) ): logger.warning_once( "Disabling prefix caching for FLASHINFER/TRITON_MLA " "with batch invariance, as it is not yet supported.", scope="local", ) cache_config.enable_prefix_caching = False impl_cls = self.attn_backend.get_impl_cls() self.impl = impl_cls( num_heads, head_size, scale, num_kv_heads, alibi_slopes, sliding_window, kv_cache_dtype, logits_soft_cap, attn_type, kv_sharing_target_layer_name, **extra_impl_args, ) self.backend = AttentionBackendEnum[self.attn_backend.get_name()] self.dtype = dtype # For cuda-alike (CUDA and ROCM) and cpu platforms, we control how # torch.compile works by registering the attention as one giant # opaque custom op. For other platforms, we directly call them # and let torch.compile handle them. self.use_direct_call = not current_platform.opaque_attention_op() self.use_output = self.attn_backend.accept_output_buffer compilation_config = vllm_config.compilation_config if prefix in compilation_config.static_forward_context: raise ValueError(f"Duplicate layer name: {prefix}") compilation_config.static_forward_context[prefix] = self self.attn_type = attn_type if kv_sharing_target_layer_name is not None: validate_kv_sharing_target( prefix, kv_sharing_target_layer_name, compilation_config.static_forward_context, ) self.kv_sharing_target_layer_name = kv_sharing_target_layer_name # use a placeholder kv cache tensor during init, which will be replaced # by bind_kv_cache # this variable will not be accessed if use_direct_call is True self.kv_cache = [ torch.tensor([]) for _ in range(vllm_config.parallel_config.pipeline_parallel_size) ] # Initialize q/k/v range constants. self.q_range = torch.tensor(envs.Q_SCALE_CONSTANT, dtype=torch.float32) self.k_range = torch.tensor(envs.K_SCALE_CONSTANT, dtype=torch.float32) self.v_range = torch.tensor(envs.V_SCALE_CONSTANT, dtype=torch.float32) # for attn backends supporting query quantization self.query_quant = None if ( self.kv_cache_dtype.startswith("fp8") and self.impl.supports_quant_query_input ): self.query_quant = QuantFP8(static=True, group_shape=GroupShape.PER_TENSOR) def forward( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, # For some alternate attention backends like MLA the attention output # shape does not match the query shape, so we optionally let the model # definition specify the output tensor shape. output_shape: torch.Size | None = None, ) -> torch.Tensor: """ The KV cache is stored inside this class and is accessed via `self.kv_cache`. Attention metadata (`attn_metadata`) is set using a context manager in the model runner's `execute_model` method. It is accessed via forward context using `vllm.forward_context.get_forward_context().attn_metadata`. """ if self.calculate_kv_scales: torch.ops.vllm.maybe_calc_kv_scales(query, key, value, self.layer_name) output_dtype = query.dtype if self.query_quant is not None: # quantizing with a simple torch operation enables # torch.compile to fuse this into previous ops # which reduces overheads during decoding. # Otherwise queries are quantized using custom ops # which causes decoding overheads assert self.kv_cache_dtype in {"fp8", "fp8_e4m3"} # check if query quantization is supported if self.impl.supports_quant_query_input: query, _ = self.query_quant(query, self._q_scale) if self.use_output: if output_shape is None: # Handle both 2D [num_tokens, hidden] and # 3D [num_tokens, heads, head_dim] query num_tokens = query.shape[0] output_shape = torch.Size( (num_tokens, self.num_heads * self.head_size_v) ) output_shape = output_shape if output_shape is not None else query.shape output = torch.empty(output_shape, dtype=output_dtype, device=query.device) hidden_size = output_shape[-1] # Reshape the query, key, and value tensors. # NOTE(woosuk): We do this outside the custom op to minimize the # CPU overheads from the non-CUDA-graph regions. query = query.view(-1, self.num_heads, self.head_size) output = output.view(-1, self.num_heads, self.head_size_v) if key is not None: key = key.view(-1, self.num_kv_heads, self.head_size) if value is not None: value = value.view(-1, self.num_kv_heads, self.head_size_v) if self.use_direct_call: forward_context: ForwardContext = get_forward_context() attn_metadata = forward_context.attn_metadata if isinstance(attn_metadata, dict): attn_metadata = attn_metadata[self.layer_name] self_kv_cache = self.kv_cache[forward_context.virtual_engine] self.impl.forward( self, query, key, value, self_kv_cache, attn_metadata, output=output ) else: torch.ops.vllm.unified_attention_with_output( query, key, value, output, self.layer_name ) return output.view(-1, hidden_size) else: if self.use_direct_call: forward_context = get_forward_context() attn_metadata = forward_context.attn_metadata if isinstance(attn_metadata, dict): attn_metadata = attn_metadata[self.layer_name] self_kv_cache = self.kv_cache[forward_context.virtual_engine] return self.impl.forward( self, query, key, value, self_kv_cache, attn_metadata ) else: return torch.ops.vllm.unified_attention( query, key, value, self.layer_name ) def calc_kv_scales(self, query, key, value): self._q_scale.copy_(torch.abs(query).max() / self.q_range) self._k_scale.copy_(torch.abs(key).max() / self.k_range) self._v_scale.copy_(torch.abs(value).max() / self.v_range) self._q_scale_float = self._q_scale.item() self._k_scale_float = self._k_scale.item() self._v_scale_float = self._v_scale.item() # We only calculate the scales once self.calculate_kv_scales = False def extra_repr(self) -> str: s = f"head_size={self.impl.head_size}" # type: ignore s += f", num_heads={self.impl.num_heads}" # type: ignore s += f", num_kv_heads={self.impl.num_kv_heads}" # type: ignore s += f", scale={self.impl.scale}" # type: ignore s += f", backend={self.impl.__class__.__name__}" return s def process_weights_after_loading(self, act_dtype: torch.dtype): self.impl.process_weights_after_loading(act_dtype) # If we should not load quant weights, we initialize the scales to 1.0 # as the default value. See [Note: Register q/k/v/prob scales in state dict] # for more details. quant_method = ( self.quant_config.get_quant_method(self, prefix=self.layer_name) if self.quant_config else None ) if not should_load_quant_weights(quant_method): set_default_quant_scales(self, register_buffer=False) def get_attn_backend(self) -> type[AttentionBackend]: return self.attn_backend def get_kv_cache_spec(self, vllm_config: VllmConfig) -> KVCacheSpec: # Block size may get updated after model loading, refresh it block_size = vllm_config.cache_config.block_size # Should not be called for enc-dec or encoder-only attention. assert self.attn_type == AttentionType.DECODER if self.sliding_window is not None: assert not vllm_config.model_config.use_mla, ( "MLA is not supported for slidingwindow" ) return SlidingWindowSpec( block_size=block_size, num_kv_heads=self.num_kv_heads, head_size=self.head_size, dtype=self.kv_cache_torch_dtype, sliding_window=self.sliding_window, ) else: return FullAttentionSpec( block_size=block_size, num_kv_heads=self.num_kv_heads, head_size=self.head_size, head_size_v=self.head_size_v, dtype=self.kv_cache_torch_dtype, ) class MLAAttention(nn.Module, AttentionLayerBase): """Multi-Head Latent Attention layer. This class takes query, and compressed key/value tensors as input. The class does the following: 1. Store the input key and value tensors in the KV cache. 2. Perform (multi-head/multi-query/grouped-query) attention. 3. Return the output tensor. """ def __init__( self, num_heads: int, scale: float, qk_nope_head_dim: int, qk_rope_head_dim: int, v_head_dim: int, q_lora_rank: int | None, kv_lora_rank: int, kv_b_proj: ColumnParallelLinear, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, prefix: str = "", use_sparse: bool = False, indexer: object | None = None, **extra_impl_args, ): super().__init__() self.num_heads = num_heads self.scale = scale self.qk_nope_head_dim = qk_nope_head_dim self.qk_rope_head_dim = qk_rope_head_dim self.v_head_dim = v_head_dim self.q_lora_rank = q_lora_rank self.kv_lora_rank = kv_lora_rank self.head_size = kv_lora_rank + qk_rope_head_dim self.layer_name = prefix if cache_config is not None: kv_cache_dtype = cache_config.cache_dtype block_size = cache_config.block_size calculate_kv_scales = cache_config.calculate_kv_scales else: kv_cache_dtype = "auto" block_size = 16 calculate_kv_scales = False self.quant_config = quant_config # Initialize KV cache quantization attributes _init_kv_cache_quant( self, self.quant_config, self.layer_name, kv_cache_dtype, calculate_kv_scales, ) dtype = torch.get_default_dtype() self.attn_backend = get_attn_backend( self.head_size, dtype, kv_cache_dtype, block_size, use_mla=True, use_sparse=use_sparse, ) if ( cache_config is not None and cache_config.enable_prefix_caching and vllm_is_batch_invariant() and ( self.attn_backend.get_name() == "TRITON_MLA" or self.attn_backend.get_name() == "FLASHINFER" ) ): logger.warning_once( "Disabling prefix caching for TRITON_MLA / FLASHINFER " "with batch invariance, as it is not yet supported.", scope="local", ) cache_config.enable_prefix_caching = False impl_cls = cast(type[MLAAttentionImpl], self.attn_backend.get_impl_cls()) self.impl = impl_cls( num_heads=self.num_heads, head_size=self.head_size, scale=self.scale, num_kv_heads=1, alibi_slopes=None, sliding_window=None, kv_cache_dtype=self.kv_cache_dtype, logits_soft_cap=None, attn_type=AttentionType.DECODER, kv_sharing_target_layer_name=None, # MLA Args q_lora_rank=self.q_lora_rank, kv_lora_rank=self.kv_lora_rank, qk_nope_head_dim=self.qk_nope_head_dim, qk_rope_head_dim=self.qk_rope_head_dim, qk_head_dim=self.qk_nope_head_dim + self.qk_rope_head_dim, v_head_dim=self.v_head_dim, kv_b_proj=kv_b_proj, indexer=indexer, **extra_impl_args, ) self.use_direct_call = not current_platform.opaque_attention_op() compilation_config = get_current_vllm_config().compilation_config if prefix in compilation_config.static_forward_context: raise ValueError(f"Duplicate layer name: {prefix}") compilation_config.static_forward_context[prefix] = self self.kv_cache = [ torch.tensor([]) for _ in range( get_current_vllm_config().parallel_config.pipeline_parallel_size ) ] self.use_sparse = use_sparse # Initialize q/k/v range constants. self.q_range = torch.tensor(envs.Q_SCALE_CONSTANT, dtype=torch.float32) self.k_range = torch.tensor(envs.K_SCALE_CONSTANT, dtype=torch.float32) self.v_range = torch.tensor(envs.V_SCALE_CONSTANT, dtype=torch.float32) def forward( self, q: torch.Tensor, kv_c_normed: torch.Tensor, k_pe: torch.Tensor, output_shape: torch.Size | None = None, ) -> torch.Tensor: if self.calculate_kv_scales: torch.ops.vllm.maybe_calc_kv_scales(q, kv_c_normed, k_pe, self.layer_name) if self.use_direct_call: forward_context: ForwardContext = get_forward_context() attn_metadata = forward_context.attn_metadata if isinstance(attn_metadata, dict): attn_metadata = attn_metadata[self.layer_name] self_kv_cache = self.kv_cache[forward_context.virtual_engine] if self.attn_backend.accept_output_buffer: output = torch.empty(output_shape, dtype=q.dtype, device=q.device) self.impl.forward( self, q, kv_c_normed, k_pe, self_kv_cache, attn_metadata, output=output, ) return output else: return self.impl.forward( self, q, kv_c_normed, k_pe, self_kv_cache, attn_metadata ) else: if self.attn_backend.accept_output_buffer: output = torch.empty(output_shape, dtype=q.dtype, device=q.device) torch.ops.vllm.unified_mla_attention_with_output( q, kv_c_normed, k_pe, output, self.layer_name, ) return output else: return torch.ops.vllm.unified_mla_attention( q, kv_c_normed, k_pe, self.layer_name, ) def process_weights_after_loading(self, act_dtype: torch.dtype): if hasattr(self.impl, "process_weights_after_loading"): self.impl.process_weights_after_loading(act_dtype) # If we should not load quant weights, we initialize the scales to 1.0 # as the default value. See [Note: Register q/k/v/prob scales in state dict] # for more details. quant_method = ( self.quant_config.get_quant_method(self, prefix=self.layer_name) if self.quant_config else None ) if not should_load_quant_weights(quant_method): set_default_quant_scales(self, register_buffer=False) def calc_kv_scales( self, q: torch.Tensor, kv_c_normed: torch.Tensor, k_pe: torch.Tensor ) -> None: """Optional scale calculation for MLA inputs. Mirrors Attention.calc_kv_scales. Not all MLA backends require this """ # Use safe defaults if ranges are not present q_range = getattr(self, "q_range", torch.tensor(1.0)) k_range = getattr(self, "k_range", torch.tensor(1.0)) v_range = getattr(self, "v_range", torch.tensor(1.0)) self._q_scale.copy_(torch.abs(q).max() / q_range) # kv_c_normed is the compressed KV representation; use it for k/v kv_abs_max = torch.abs(kv_c_normed).max() self._k_scale.copy_(kv_abs_max / k_range) self._v_scale.copy_(kv_abs_max / v_range) self._q_scale_float = self._q_scale.item() self._k_scale_float = self._k_scale.item() self._v_scale_float = self._v_scale.item() self.calculate_kv_scales = False def get_attn_backend(self) -> type[AttentionBackend]: return self.attn_backend def get_kv_cache_spec(self, vllm_config: VllmConfig) -> KVCacheSpec: kv_cache_dtype = kv_cache_dtype_str_to_dtype( self.kv_cache_dtype, vllm_config.model_config ) return MLAAttentionSpec( block_size=vllm_config.cache_config.block_size, num_kv_heads=1, head_size=self.head_size, dtype=kv_cache_dtype, cache_dtype_str=vllm_config.cache_config.cache_dtype, ) def maybe_calc_kv_scales( query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, layer_name: str, ) -> None: forward_context: ForwardContext = get_forward_context() self = forward_context.no_compile_layers[layer_name] # Only calculate if the layer's calculate_kv_scales flag is True # This flag gets set to False after the first forward pass if not self.calculate_kv_scales: return self.calc_kv_scales(query, key, value) def maybe_calc_kv_scales_fake( query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, layer_name: str, ) -> None: return direct_register_custom_op( op_name="maybe_calc_kv_scales", op_func=maybe_calc_kv_scales, mutates_args=["query", "key", "value"], fake_impl=maybe_calc_kv_scales_fake, ) def get_attention_context( layer_name: str, ) -> tuple[dict | object | None, Attention | MLAAttention, torch.Tensor]: """Extract attention context for a given layer. This helper function extracts the attention metadata, attention layer instance, and KV cache tensor for a specific layer. Args: layer_name: The name/identifier of the attention layer. Returns: A tuple containing: - attn_metadata: Attention metadata for this specific layer, or None if no metadata available - attn_layer: The attention layer instance (Attention or MLAAttention) - kv_cache: The KV cache tensor for current virtual engine Note: attn_metadata may be None, but attn_layer and kv_cache are always extracted from the forward context. """ forward_context: ForwardContext = get_forward_context() attn_metadata = forward_context.attn_metadata if isinstance(attn_metadata, dict): attn_metadata = attn_metadata[layer_name] attn_layer: Attention | MLAAttention = forward_context.no_compile_layers[layer_name] kv_cache = attn_layer.kv_cache[forward_context.virtual_engine] return attn_metadata, attn_layer, kv_cache @maybe_transfer_kv_layer def unified_attention( query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, layer_name: str, ) -> torch.Tensor: attn_metadata, self, kv_cache = get_attention_context(layer_name) output = self.impl.forward(self, query, key, value, kv_cache, attn_metadata) return output def unified_attention_fake( query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, layer_name: str, ) -> torch.Tensor: return torch.empty_like(query).contiguous() direct_register_custom_op( op_name="unified_attention", op_func=unified_attention, fake_impl=unified_attention_fake, ) @maybe_transfer_kv_layer def unified_attention_with_output( query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, output: torch.Tensor, layer_name: str, output_scale: torch.Tensor | None = None, output_block_scale: torch.Tensor | None = None, ) -> None: attn_metadata, self, kv_cache = get_attention_context(layer_name) self.impl.forward( self, query, key, value, kv_cache, attn_metadata, output=output, output_scale=output_scale, output_block_scale=output_block_scale, ) def unified_attention_with_output_fake( query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, output: torch.Tensor, layer_name: str, output_scale: torch.Tensor | None = None, output_block_scale: torch.Tensor | None = None, ) -> None: return direct_register_custom_op( op_name="unified_attention_with_output", op_func=unified_attention_with_output, mutates_args=["output", "output_block_scale"], fake_impl=unified_attention_with_output_fake, ) @maybe_transfer_kv_layer def unified_mla_attention( q: torch.Tensor, kv_c_normed: torch.Tensor, k_pe: torch.Tensor, layer_name: str, ) -> torch.Tensor: attn_metadata, self, kv_cache = get_attention_context(layer_name) output = self.impl.forward(self, q, kv_c_normed, k_pe, kv_cache, attn_metadata) return output def unified_mla_attention_fake( q: torch.Tensor, kv_c_normed: torch.Tensor, k_pe: torch.Tensor, layer_name: str, ) -> torch.Tensor: return torch.empty_like(q).contiguous() direct_register_custom_op( op_name="unified_mla_attention", op_func=unified_mla_attention, mutates_args=[], fake_impl=unified_mla_attention_fake, dispatch_key=current_platform.dispatch_key, ) @maybe_transfer_kv_layer def unified_mla_attention_with_output( q: torch.Tensor, kv_c_normed: torch.Tensor,
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/__init__.py
vllm/attention/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/selector.py
vllm/attention/selector.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from functools import cache from typing import NamedTuple, cast, get_args import torch from vllm.attention.backends.abstract import AttentionBackend, AttentionType from vllm.attention.backends.registry import ( MAMBA_TYPE_TO_BACKEND_MAP, MambaAttentionBackendEnum, ) from vllm.config.cache import CacheDType from vllm.logger import init_logger from vllm.utils.import_utils import resolve_obj_by_qualname logger = init_logger(__name__) class AttentionSelectorConfig(NamedTuple): head_size: int dtype: torch.dtype kv_cache_dtype: CacheDType | None block_size: int | None use_mla: bool = False has_sink: bool = False use_sparse: bool = False use_mm_prefix: bool = False attn_type: str = AttentionType.DECODER def __repr__(self): return ( f"AttentionSelectorConfig(head_size={self.head_size}, " f"dtype={self.dtype}, " f"kv_cache_dtype={self.kv_cache_dtype}, " f"block_size={self.block_size}, " f"use_mla={self.use_mla}, " f"has_sink={self.has_sink}, " f"use_sparse={self.use_sparse}, " f"use_mm_prefix={self.use_mm_prefix}, " f"attn_type={self.attn_type})" ) def get_attn_backend( head_size: int, dtype: torch.dtype, kv_cache_dtype: str | None, block_size: int | None, use_mla: bool = False, has_sink: bool = False, use_sparse: bool = False, use_mm_prefix: bool = False, attn_type: str | None = None, ) -> type[AttentionBackend]: """Selects which attention backend to use and lazily imports it.""" if kv_cache_dtype is not None: valid_cache_dtypes = get_args(CacheDType) assert kv_cache_dtype in valid_cache_dtypes, ( f"Invalid kv_cache_dtype: {kv_cache_dtype}. " f"Valid values are: {valid_cache_dtypes}" ) from vllm.config import get_current_vllm_config vllm_config = get_current_vllm_config() backend_enum = vllm_config.attention_config.backend attn_selector_config = AttentionSelectorConfig( head_size=head_size, dtype=dtype, kv_cache_dtype=cast(CacheDType | None, kv_cache_dtype), block_size=block_size, use_mla=use_mla, has_sink=has_sink, use_sparse=use_sparse, use_mm_prefix=use_mm_prefix, attn_type=attn_type or AttentionType.DECODER, ) return _cached_get_attn_backend( backend=backend_enum, attn_selector_config=attn_selector_config, ) @cache def _cached_get_attn_backend( backend, attn_selector_config: AttentionSelectorConfig, ) -> type[AttentionBackend]: from vllm.platforms import current_platform attention_cls = current_platform.get_attn_backend_cls( backend, attn_selector_config=attn_selector_config, ) if not attention_cls: raise ValueError( f"Invalid attention backend for {current_platform.device_name}" ) backend = resolve_obj_by_qualname(attention_cls) # Adjust kv cache layout if the selected backend requires a specific one required_layout = backend.get_required_kv_cache_layout() if required_layout is not None: from vllm.v1.attention.backends.utils import set_kv_cache_layout set_kv_cache_layout(required_layout) logger.info( "Using %s KV cache layout for %s backend.", required_layout, backend.get_name(), ) return backend def get_mamba_attn_backend( mamba_type: str, ) -> type[AttentionBackend]: """Select which mamba attention backend to use and lazily import it.""" return _cached_get_mamba_attn_backend(mamba_type) @cache def _cached_get_mamba_attn_backend( mamba_type: str, ) -> type[AttentionBackend]: assert mamba_type and isinstance(mamba_type, str) selected_backend = None try: backend_name = MAMBA_TYPE_TO_BACKEND_MAP[mamba_type] selected_backend = MambaAttentionBackendEnum[backend_name] except KeyError as e: raise ValueError( f"Invalid mamba attention backend type: '{backend_name}'. Valid " f"backends are: {list(MambaAttentionBackendEnum.__members__.keys())}" ) from e mamba_attn_backend = selected_backend.get_class() return mamba_attn_backend
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/utils/kv_sharing_utils.py
vllm/attention/utils/kv_sharing_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project def validate_kv_sharing_target( current_layer_name, target_layer_name, static_forward_context ): error_msg = ( f"Specified KV sharing target layer for {current_layer_name} " f"is not valid: target layer {target_layer_name} " ) if current_layer_name == target_layer_name: raise ValueError(error_msg + "cannot be the same as the current layer.") if target_layer_name not in static_forward_context: from vllm.model_executor.models.utils import extract_layer_index # If target layer name is not in the static fwd context, it means either # a) the target layer does not come BEFORE the current layer, or # b) the target layer is not an Attention layer that exists in the model current_layer_idx = extract_layer_index(current_layer_name) target_layer_idx = extract_layer_index(target_layer_name) if current_layer_idx <= target_layer_idx: raise ValueError(error_msg + "must come before the current layer.") else: raise ValueError(error_msg + "is not a valid Attention layer in the model.") # Currently KV sharing is only supported between layers of the same type target_layer_attn_type = static_forward_context[target_layer_name].attn_type expected = static_forward_context[current_layer_name].attn_type if target_layer_attn_type != expected: raise ValueError( error_msg + f"must be the same type as the current layer ({expected})." )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/utils/__init__.py
vllm/attention/utils/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/utils/kv_transfer_utils.py
vllm/attention/utils/kv_transfer_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import inspect from collections.abc import Callable from functools import wraps from vllm.distributed.kv_transfer import ( get_kv_transfer_group, has_kv_transfer_group, is_v1_kv_transfer_group, ) def maybe_transfer_kv_layer(func: Callable) -> Callable: """Decorator that handles KV layer transfer prior and after execution of an attention layer, if enabled. Otherwise, the wrapper is a no-op. On entry: waits for the KV layer from the connector. On exit: saves the KV layer to the connector. """ # Import at runtime to avoid circular dependency from vllm.attention.layer import get_attention_context # Inspect the signature ONCE when the decorator is applied. sig = inspect.signature(func) param_names = list(sig.parameters.keys()) # Find the index of 'layer_name' parameter. try: layer_name_index = param_names.index("layer_name") except ValueError as e: raise TypeError( f"Function {func.__name__} must have a 'layer_name' parameter" ) from e @wraps(func) def wrapper(*args, **kwargs): if not has_kv_transfer_group() or not is_v1_kv_transfer_group(): return func(*args, **kwargs) layer_name: str = args[layer_name_index] # Extract attention context (layer-specific metadata, layer, and kv_cache) attn_metadata, attn_layer, kv_cache = get_attention_context(layer_name) connector = get_kv_transfer_group() if attn_metadata is None or not connector.has_connector_metadata(): return func(*args, **kwargs) # Wait for KV layer on entry connector.wait_for_layer_load(layer_name) # Execute the function result = func(*args, **kwargs) # Save KV cache layer on exit connector.save_kv_layer(layer_name, kv_cache, attn_metadata) return result return wrapper
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/utils/fa_utils.py
vllm/attention/utils/fa_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from vllm.logger import init_logger from vllm.platforms import current_platform logger = init_logger(__name__) if current_platform.is_cuda(): from vllm import _custom_ops as ops reshape_and_cache_flash = ops.reshape_and_cache_flash from vllm.vllm_flash_attn import flash_attn_varlen_func, get_scheduler_metadata elif current_platform.is_xpu(): from vllm._ipex_ops import ipex_ops as ops reshape_and_cache_flash = ops.reshape_and_cache_flash flash_attn_varlen_func = ops.flash_attn_varlen_func get_scheduler_metadata = ops.get_scheduler_metadata elif current_platform.is_rocm(): try: from flash_attn import flash_attn_varlen_func # noqa: F401 except ImportError as e: raise ImportError( "Rocm platform requires upstream flash-attn " "to be installed. Please install flash-attn first." ) from e def get_flash_attn_version(requires_alibi: bool = False) -> int | None: # import here to avoid circular dependencies from vllm.platforms import current_platform if current_platform.is_xpu(): return 2 try: from vllm.vllm_flash_attn.flash_attn_interface import ( fa_version_unsupported_reason, is_fa_version_supported, ) device_capability = current_platform.get_device_capability() assert device_capability is not None # 1. default version depending on platform fa_version = ( 3 if (device_capability.major == 9 and is_fa_version_supported(3)) else 2 ) # 2. override if passed by environment or config from vllm.config import get_current_vllm_config vllm_config = get_current_vllm_config() if vllm_config.attention_config.flash_attn_version is not None: fa_version = vllm_config.attention_config.flash_attn_version # 3. fallback for unsupported combinations if device_capability.major == 10 and fa_version == 3: logger.warning_once( "Cannot use FA version 3 on Blackwell platform " "defaulting to FA version 2." ) fa_version = 2 if requires_alibi and fa_version == 3: logger.warning_once( "Cannot use FA version 3 with ALiBi, defaulting to FA version 2." ) fa_version = 2 if not is_fa_version_supported(fa_version): logger.error( "Cannot use FA version %d is not supported due to %s", fa_version, fa_version_unsupported_reason(fa_version), ) assert is_fa_version_supported(fa_version) return fa_version except (ImportError, AssertionError): return None def flash_attn_supports_fp8() -> bool: return ( get_flash_attn_version() == 3 and current_platform.get_device_capability().major == 9 ) def flash_attn_supports_sinks() -> bool: if current_platform.is_xpu(): return True else: return get_flash_attn_version() == 3 def flash_attn_supports_mla(): from vllm.platforms import current_platform if current_platform.is_cuda(): try: from vllm.vllm_flash_attn.flash_attn_interface import ( is_fa_version_supported, ) return ( is_fa_version_supported(3) and current_platform.get_device_capability()[0] == 9 ) except (ImportError, AssertionError): pass return False def is_flash_attn_varlen_func_available() -> bool: return current_platform.is_cuda() or current_platform.is_xpu()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/layers/static_sink_attention.py
vllm/attention/layers/static_sink_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools import torch from vllm.attention.backends.abstract import ( AttentionBackend, AttentionMetadata, AttentionType, ) from vllm.attention.layer import Attention from vllm.attention.ops.triton_reshape_and_cache_flash import ( triton_reshape_and_cache_flash_diffkv, ) from vllm.attention.selector import get_attn_backend from vllm.config import CacheConfig, VllmConfig from vllm.forward_context import ForwardContext, get_forward_context from vllm.logger import init_logger from vllm.model_executor.custom_op import CustomOp from vllm.utils.math_utils import cdiv from vllm.utils.torch_utils import direct_register_custom_op from vllm.v1.attention.backends.utils import ( CommonAttentionMetadata, subclass_attention_backend, ) from vllm.v1.kv_cache_interface import ( AttentionSpec, KVCacheSpec, SinkFullAttentionSpec, ) logger = init_logger(__name__) @functools.lru_cache def create_static_sink_attention_backend( underlying_attn_backend: type[AttentionBackend], sink_len: int = 0, ) -> type[AttentionBackend]: prefix = "StaticSink_" underlying_builder = underlying_attn_backend.get_builder_cls() class StaticSinkAttentionBuilder(underlying_builder): # type: ignore def __init__( self, kv_cache_spec: AttentionSpec, layer_names: list[str], vllm_config: VllmConfig, device: torch.device, ): super().__init__(kv_cache_spec, layer_names, vllm_config, device) model_config = vllm_config.model_config scheduler_config = vllm_config.scheduler_config self.sink_len = sink_len self.block_size = vllm_config.cache_config.block_size self.num_sink_blocks = self.sink_len // vllm_config.cache_config.block_size self.max_num_blocks = cdiv( model_config.max_model_len, vllm_config.cache_config.block_size ) self.block_table_with_sink = torch.zeros( ( scheduler_config.max_num_seqs, self.max_num_blocks + self.num_sink_blocks, ), device=device, dtype=torch.int32, ) self.block_table_with_sink[:, : self.num_sink_blocks] = torch.arange( 1, self.num_sink_blocks + 1, device=device, dtype=torch.int32, ) def build( self, common_prefix_len: int, common_attn_metadata: CommonAttentionMetadata, fast_build: bool = False, ) -> AttentionMetadata: common_attn_metadata.seq_lens[:] = ( common_attn_metadata.seq_lens + self.sink_len ) common_attn_metadata.seq_lens[ common_attn_metadata.seq_lens == self.sink_len ] = 0 common_attn_metadata.max_seq_len = ( common_attn_metadata.max_seq_len + self.sink_len ) max_num_blocks = cdiv(common_attn_metadata.max_seq_len, self.block_size) num_reqs = common_attn_metadata.num_reqs self.block_table_with_sink[ :num_reqs, self.num_sink_blocks : self.num_sink_blocks + max_num_blocks ] = common_attn_metadata.block_table_tensor[:, :max_num_blocks] common_attn_metadata.block_table_tensor = self.block_table_with_sink[ :num_reqs ] return super().build(common_prefix_len, common_attn_metadata, fast_build) attn_backend = subclass_attention_backend( name_prefix=prefix, attention_backend_cls=underlying_attn_backend, builder_cls=StaticSinkAttentionBuilder, ) return attn_backend @CustomOp.register("static_sink_attention") class StaticSinkAttention(Attention, CustomOp): """ Attention with static sink tokens """ def __init__( self, num_heads: int, head_size: int, scale: float, sink_len: int, attn_backend: type[AttentionBackend] | None = None, cache_config: CacheConfig | None = None, **kwargs, ): dtype = torch.get_default_dtype() if cache_config is not None: kv_cache_dtype = cache_config.cache_dtype block_size = cache_config.block_size else: kv_cache_dtype = "auto" block_size = 16 if attn_backend is not None: underlying_attn_backend = attn_backend else: underlying_attn_backend = get_attn_backend( head_size, dtype, kv_cache_dtype, block_size ) attn_backend = create_static_sink_attention_backend( underlying_attn_backend, sink_len=sink_len, ) Attention.__init__( self=self, num_heads=num_heads, head_size=head_size, scale=scale, cache_config=cache_config, attn_backend=attn_backend, **kwargs, ) CustomOp.__init__(self) self.sink_len = sink_len self.block_size = block_size self.sink_populated = False self.sink_key = None self.sink_value = None def update_sink_kv(self, sink_key, sink_value) -> None: self.sink_key = sink_key self.sink_value = sink_value def forward_native( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, output_shape: torch.Size | None = None, ) -> torch.Tensor: assert self.sink_key is not None and self.sink_value is not None, ( "sink_key and sink_value have not been prepared" ) if not self.sink_populated: forward_context: ForwardContext = get_forward_context() self_kv_cache = self.kv_cache[forward_context.virtual_engine] torch.ops.vllm.maybe_populate_sink(self_kv_cache, self.layer_name) return super().forward(query, key, value, output_shape) def forward_cuda( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, output_shape: torch.Size | None = None, ) -> torch.Tensor: return self.forward_native(query, key, value, output_shape) def forward(self, *args, **kwargs): return self._forward_method(*args, **kwargs) def populate_sink_kv(self, self_kv_cache): sink_kv_slot_mapping = torch.arange( self.block_size, self.sink_len + self.block_size, device=torch.cuda.current_device(), dtype=torch.long, ) triton_reshape_and_cache_flash_diffkv( self.sink_key, self.sink_value, self_kv_cache, sink_kv_slot_mapping, self.kv_cache_dtype, self._k_scale, self._v_scale, ) # We only populate the sink_key and sink_value once self.sink_populated = True def get_kv_cache_spec(self, vllm_config: VllmConfig) -> KVCacheSpec: # Block size may get updated after model loading, refresh it block_size = vllm_config.cache_config.block_size # Should not be called for enc-dec or encoder-only attention. assert self.attn_type == AttentionType.DECODER return SinkFullAttentionSpec( block_size=block_size, num_kv_heads=self.num_kv_heads, head_size=self.head_size, head_size_v=self.head_size_v, sink_len=self.sink_len, dtype=self.kv_cache_torch_dtype, ) def maybe_populate_sink( self_kv_cache: torch.Tensor, layer_name: str, ) -> None: forward_context: ForwardContext = get_forward_context() self = forward_context.no_compile_layers[layer_name] if self.sink_populated or self_kv_cache.numel() == 0: return self.populate_sink_kv(self_kv_cache) def maybe_populate_sink_fake( self_kv_cache: torch.Tensor, layer_name: str, ) -> None: return direct_register_custom_op( op_name="maybe_populate_sink", op_func=maybe_populate_sink, mutates_args=["self_kv_cache"], fake_impl=maybe_populate_sink_fake, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/layers/chunked_local_attention.py
vllm/attention/layers/chunked_local_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools import torch from vllm.attention.backends.abstract import AttentionBackend from vllm.attention.layer import Attention from vllm.attention.selector import get_attn_backend from vllm.config import CacheConfig from vllm.config.vllm import VllmConfig from vllm.model_executor.layers.quantization import QuantizationConfig from vllm.v1.attention.backends.utils import ( AttentionCGSupport, AttentionMetadataBuilder, CommonAttentionMetadata, make_local_attention_virtual_batches, subclass_attention_backend, ) from vllm.v1.kv_cache_interface import ( AttentionSpec, ChunkedLocalAttentionSpec, KVCacheSpec, ) @functools.lru_cache def create_chunked_local_attention_backend( underlying_attn_backend: AttentionBackend, attention_chunk_size: int, block_size: int, ) -> type[AttentionBackend]: prefix = f"ChunkedLocalAttention_{attention_chunk_size}_{block_size}_" underlying_builder = underlying_attn_backend.get_builder_cls() assert issubclass(underlying_builder, AttentionMetadataBuilder) class ChunkedLocalAttentionBuilder(underlying_builder): # type: ignore @classmethod def get_cudagraph_support( cls: type["AttentionMetadataBuilder"], vllm_config: VllmConfig, kv_cache_spec: AttentionSpec, ) -> AttentionCGSupport: # Explicit override in case the underlying builder specialized this getter. # @override omitted only because of mypy limitation due to type variable. return AttentionCGSupport.NEVER def build( self, common_prefix_len: int, common_attn_metadata: CommonAttentionMetadata, fast_build: bool = False, ): cm, make_virtual_batches_block_table = make_local_attention_virtual_batches( attention_chunk_size, common_attn_metadata, block_size ) metadata = super().build(common_prefix_len, cm, fast_build) metadata.make_virtual_batches_block_table = make_virtual_batches_block_table return metadata def update_block_table( self, metadata, blk_table: torch.Tensor, slot_mapping: torch.Tensor ): blk_table = metadata.make_virtual_batches_block_table(blk_table) return super().update_block_table(metadata, blk_table, slot_mapping) attn_backend = subclass_attention_backend( name_prefix=prefix, attention_backend_cls=underlying_attn_backend, builder_cls=ChunkedLocalAttentionBuilder, ) return attn_backend class ChunkedLocalAttention(Attention): def __init__( self, num_heads: int, head_size: int, scale: float, attention_chunk_size: int, num_kv_heads: int | None = None, alibi_slopes: list[float] | None = None, cache_config: CacheConfig | None = None, quant_config: QuantizationConfig | None = None, kv_sharing_target_layer_name: str | None = None, prefix: str = "", ): self.attention_chunk_size = attention_chunk_size dtype = torch.get_default_dtype() if cache_config is not None: kv_cache_dtype = cache_config.cache_dtype block_size = cache_config.block_size else: kv_cache_dtype = "auto" block_size = 16 underlying_attn_backend = get_attn_backend( head_size, dtype, kv_cache_dtype, block_size ) attn_backend = create_chunked_local_attention_backend( underlying_attn_backend, attention_chunk_size, block_size ) super().__init__( num_heads=num_heads, head_size=head_size, scale=scale, num_kv_heads=num_kv_heads, alibi_slopes=alibi_slopes, cache_config=cache_config, quant_config=quant_config, prefix=prefix, kv_sharing_target_layer_name=kv_sharing_target_layer_name, attn_backend=attn_backend, ) def get_kv_cache_spec(self, vllm_config: VllmConfig) -> KVCacheSpec: assert self.attention_chunk_size return ChunkedLocalAttentionSpec( block_size=vllm_config.cache_config.block_size, num_kv_heads=self.num_kv_heads, head_size=self.head_size, dtype=self.kv_cache_torch_dtype, attention_chunk_size=self.attention_chunk_size, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/layers/__init__.py
vllm/attention/layers/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/layers/cross_attention.py
vllm/attention/layers/cross_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools from copy import copy import numpy as np import torch from vllm.attention.backends.abstract import ( AttentionBackend, AttentionMetadata, AttentionType, ) from vllm.attention.layer import Attention from vllm.attention.selector import get_attn_backend from vllm.config import CacheConfig, VllmConfig from vllm.logger import init_logger from vllm.utils.math_utils import cdiv from vllm.v1.attention.backends.utils import ( CommonAttentionMetadata, subclass_attention_backend, ) from vllm.v1.kv_cache_interface import CrossAttentionSpec, KVCacheSpec logger = init_logger(__name__) def _get_cross_slot_mapping( encoder_seq_lens: np.ndarray, block_table_tensor: torch.Tensor, kv_cache_spec: CrossAttentionSpec, device: torch.device, ) -> torch.Tensor: """Get cross-attention slot mappings.""" block_size = kv_cache_spec.block_size slot_mappings = [] # Find indices with non-zero encoder sequence lengths # The majority of parallel requests will be running the # decoder, so this list should be relatively small. active_indices = np.nonzero(encoder_seq_lens)[0] for req_index in active_indices: encoder_seq_len = encoder_seq_lens[req_index].item() # Calculate the number of blocks needed for this request num_blocks_needed = cdiv(encoder_seq_len, block_size) # Get the block IDs for this request from the tensor req_block_ids = block_table_tensor[req_index] # Get only the blocks we need (first num_blocks_needed blocks) needed_block_ids = req_block_ids[:num_blocks_needed] # All needed blocks are allocated i_values = torch.arange(encoder_seq_len, dtype=torch.int64, device=device) block_indices = i_values // block_size block_offsets = i_values % block_size block_numbers = needed_block_ids[block_indices] slot_mapping = block_numbers * block_size + block_offsets slot_mappings.append(slot_mapping) if slot_mappings: return torch.cat(slot_mappings) else: return torch.empty(0, dtype=torch.int64, device=device) @functools.lru_cache def create_cross_attention_backend( underlying_attn_backend: AttentionBackend, ) -> type[AttentionBackend]: prefix = "CrossAttention_" underlying_builder = underlying_attn_backend.get_builder_cls() class CrossAttentionBuilder(underlying_builder): # type: ignore def build( self, common_prefix_len: int, common_attn_metadata: CommonAttentionMetadata, fast_build: bool = False, ) -> AttentionMetadata: new_metadata = copy(common_attn_metadata) new_metadata.causal = False max_encoder_len = int(new_metadata.encoder_seq_lens_cpu.max()) new_metadata.max_seq_len = max_encoder_len # Any computed tokens indicated decode step>1 (no chunked prefill) num_cache_decodes = ( (common_attn_metadata.num_computed_tokens_cpu > 0).sum().item() ) if num_cache_decodes > 0: # CrossAttn KV cache has already been populated on first decoder step, # skip slot_mapping calculation for requests that do not need # reshape_and_cache. num_tokens = common_attn_metadata.num_computed_tokens_cpu.numpy() new_metadata.encoder_seq_lens_cpu = np.where( num_tokens > 0, 0, new_metadata.encoder_seq_lens_cpu ) # seq_lens is provided by model runner: initial encoder input length is # needed here to know how many tokens to attend to from the cached # cross-attention KV cache. new_metadata.seq_lens = common_attn_metadata.encoder_seq_lens new_metadata._seq_lens_cpu = torch.from_numpy( common_attn_metadata.encoder_seq_lens_cpu ) # NOTE (NickLucche) use `new_metadata` instead of `common_*` (initial) here new_metadata.slot_mapping = _get_cross_slot_mapping( new_metadata.encoder_seq_lens_cpu, new_metadata.block_table_tensor, self.kv_cache_spec, self.device, ) return super().build(common_prefix_len, new_metadata, fast_build) attn_backend = subclass_attention_backend( name_prefix=prefix, attention_backend_cls=underlying_attn_backend, builder_cls=CrossAttentionBuilder, ) return attn_backend class CrossAttention(Attention): """ Cross-attention for encoder-decoder models. Handles attention between decoder queries and encoder keys/values. """ def __init__( self, num_heads: int, head_size: int, scale: float, cache_config: CacheConfig | None = None, attn_type: str | None = None, **kwargs, ): dtype = torch.get_default_dtype() if cache_config is not None: kv_cache_dtype = cache_config.cache_dtype block_size = cache_config.block_size else: kv_cache_dtype = "auto" block_size = 16 underlying_attn_backend = get_attn_backend( head_size, dtype, kv_cache_dtype, block_size ) attn_backend = create_cross_attention_backend(underlying_attn_backend) if attn_type is not None: assert attn_type == AttentionType.ENCODER_DECODER, ( "CrossAttention only supports AttentionType.ENCODER_DECODER" ) super().__init__( num_heads=num_heads, head_size=head_size, scale=scale, cache_config=cache_config, attn_backend=attn_backend, attn_type=AttentionType.ENCODER_DECODER, **kwargs, ) def get_kv_cache_spec(self, vllm_config: VllmConfig) -> KVCacheSpec: return CrossAttentionSpec( block_size=vllm_config.cache_config.block_size, num_kv_heads=self.num_kv_heads, head_size=self.head_size, dtype=self.kv_cache_torch_dtype, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/layers/encoder_only_attention.py
vllm/attention/layers/encoder_only_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools from copy import copy import torch from vllm.attention.backends.abstract import ( AttentionBackend, AttentionMetadata, AttentionType, ) from vllm.attention.layer import Attention from vllm.attention.selector import get_attn_backend from vllm.config import CacheConfig from vllm.config.vllm import VllmConfig from vllm.v1.attention.backends.utils import ( CommonAttentionMetadata, subclass_attention_backend, ) from vllm.v1.kv_cache_interface import KVCacheSpec @functools.lru_cache def create_encoder_only_attention_backend( underlying_attn_backend: AttentionBackend, ) -> type[AttentionBackend]: prefix = "EncoderOnlyAttention_" underlying_builder = underlying_attn_backend.get_builder_cls() class EncoderOnlyAttentionBuilder(underlying_builder): # type: ignore def build( self, common_prefix_len: int, common_attn_metadata: CommonAttentionMetadata, fast_build: bool = False, ) -> AttentionMetadata: new_common_attn_metadata = copy(common_attn_metadata) new_common_attn_metadata.causal = False return super().build( common_prefix_len, new_common_attn_metadata, fast_build ) attn_backend = subclass_attention_backend( name_prefix=prefix, attention_backend_cls=underlying_attn_backend, builder_cls=EncoderOnlyAttentionBuilder, ) return attn_backend class EncoderOnlyAttention(Attention): """ Encoder attention is a special case that doesn't need a KV Cache. """ def __init__( self, num_heads: int, head_size: int, scale: float, cache_config: CacheConfig | None = None, attn_type: str | None = None, **kwargs, ): dtype = torch.get_default_dtype() if cache_config is not None: kv_cache_dtype = cache_config.cache_dtype block_size = cache_config.block_size else: kv_cache_dtype = "auto" block_size = 16 underlying_attn_backend = get_attn_backend( head_size, dtype, kv_cache_dtype, block_size, attn_type=AttentionType.ENCODER_ONLY, ) attn_backend = create_encoder_only_attention_backend(underlying_attn_backend) if attn_type is not None: assert attn_type == AttentionType.ENCODER_ONLY, ( "EncoderOnlyAttention only supports AttentionType.ENCODER_ONLY" ) super().__init__( num_heads=num_heads, head_size=head_size, scale=scale, cache_config=cache_config, attn_backend=attn_backend, attn_type=AttentionType.ENCODER_ONLY, **kwargs, ) def get_kv_cache_spec(self, vllm_config: VllmConfig) -> KVCacheSpec: # Does not need KV cache return None
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/layers/mm_encoder_attention.py
vllm/attention/layers/mm_encoder_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.attention.backends.registry import AttentionBackendEnum from vllm.attention.ops.vit_attn_wrappers import ( vit_flash_attn_wrapper, vit_torch_sdpa_wrapper, ) from vllm.attention.utils.fa_utils import get_flash_attn_version from vllm.config import MultiModalConfig from vllm.logger import init_logger from vllm.model_executor.custom_op import CustomOp from vllm.model_executor.models.vision import get_vit_attn_backend logger = init_logger(__name__) @CustomOp.register("mm_encoder_attn") class MMEncoderAttention(CustomOp): """Multi-headed attention without any cache, used for multimodal encoder.""" def __init__( self, num_heads: int, head_size: int, scale: float | None = None, num_kv_heads: int | None = None, prefix: str = "", multimodal_config: MultiModalConfig | None = None, ) -> None: """ Args: num_heads: number of attention heads per partition. head_size: hidden_size per attention head. scale: scale factor. num_kv_heads: number of kv heads. prefix: This has no effect, it is only here to make it easier to swap between Attention and MultiHeadAttention multimodal_config: configs for multi-modal. """ super().__init__() self.num_heads = num_heads self.head_size = head_size self.scale = scale self.num_kv_heads = num_heads if num_kv_heads is None else num_kv_heads self.layer_name = prefix assert self.num_heads % self.num_kv_heads == 0, ( f"num_heads ({self.num_heads}) is not " f"divisible by num_kv_heads ({self.num_kv_heads})" ) self.num_queries_per_kv = self.num_heads // self.num_kv_heads # During model initialization, the default dtype is set as the model # weight and activation dtype. dtype = torch.get_default_dtype() # Try to get vision attention backend from multimodal_config. attn_backend_override = None if multimodal_config is not None: attn_backend_override = multimodal_config.mm_encoder_attn_backend # Get device-specific vision attention backend. self.attn_backend = get_vit_attn_backend( head_size=head_size, dtype=dtype, attn_backend_override=attn_backend_override, ) self.is_flash_attn_backend = self.attn_backend in { AttentionBackendEnum.FLASH_ATTN, AttentionBackendEnum.ROCM_AITER_FA, } self._fa_version = ( get_flash_attn_version() if self.is_flash_attn_backend else None ) logger.info_once(f"Using {self.attn_backend} for MMEncoderAttention.") @classmethod def enabled(cls) -> bool: return True def maybe_reshape_qkv_to_4d( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, bsz: int, q_len: int, kv_len: int, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Reshape query, key, value to 4D tensors: (batch_size, seq_len, num_heads, head_size) """ query = query.view(bsz, q_len, self.num_heads, self.head_size) key = key.view(bsz, kv_len, self.num_kv_heads, self.head_size) value = value.view(bsz, kv_len, self.num_kv_heads, self.head_size) if (num_repeat := self.num_queries_per_kv) > 1: # Handle MQA and GQA key = torch.repeat_interleave(key, num_repeat, dim=2) value = torch.repeat_interleave(value, num_repeat, dim=2) return query, key, value def _forward_sdpa( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, cu_seqlens: torch.Tensor | None = None, ) -> torch.Tensor: """Input shape: (batch_size x seq_len x hidden_size) or (batch_size x seq_len x num_heads x head_size) """ bsz, q_len = query.size()[:2] kv_len = key.size(1) is_reshaped = query.dim() != 4 query, key, value = self.maybe_reshape_qkv_to_4d( query, key, value, bsz, q_len, kv_len ) output = vit_torch_sdpa_wrapper( q=query, k=key, v=value, cu_seqlens=cu_seqlens, ) if is_reshaped: output = output.reshape(bsz, q_len, -1) return output def _forward_fa( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, cu_seqlens: torch.Tensor | None = None, max_seqlen: torch.Tensor | None = None, # Only used for Flash Attention ) -> torch.Tensor: """Input shape: (batch_size x seq_len x hidden_size) or (batch_size x seq_len x num_heads x head_size) """ assert (cu_seqlens is not None and max_seqlen is not None) or ( cu_seqlens is None and max_seqlen is None ), "cu_seqlens and max_seqlen should be both set or both None." bsz, q_len = query.size()[:2] kv_len = key.size(1) is_reshaped = query.dim() != 4 query, key, value = self.maybe_reshape_qkv_to_4d( query, key, value, bsz, q_len, kv_len ) output = vit_flash_attn_wrapper( q=query, k=key, v=value, cu_seqlens=cu_seqlens, max_seqlen=max_seqlen, batch_size=bsz, is_rocm_aiter=(self.attn_backend == AttentionBackendEnum.ROCM_AITER_FA), fa_version=self._fa_version, ) if is_reshaped: output = output.reshape(bsz, q_len, -1) return output def forward_native( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, cu_seqlens: torch.Tensor | None = None, max_seqlen: torch.Tensor | None = None, # Only used for Flash Attention ) -> torch.Tensor: return self._forward_sdpa(query, key, value, cu_seqlens) def forward_cuda( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, cu_seqlens: torch.Tensor | None = None, max_seqlen: torch.Tensor | None = None, # Only used for Flash Attention ) -> torch.Tensor: if self.is_flash_attn_backend: return self._forward_fa(query, key, value, cu_seqlens, max_seqlen) elif self.attn_backend == AttentionBackendEnum.TORCH_SDPA: return self._forward_sdpa(query, key, value, cu_seqlens) else: raise ValueError( f"Unsupported multi-modal encoder attention backend for CUDA: " f"{self.attn_backend}." ) def forward_cpu( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, cu_seqlens: torch.Tensor | None = None, max_seqlen: torch.Tensor | None = None, # Only used for Flash Attention ) -> torch.Tensor: return self._forward_sdpa(query, key, value, cu_seqlens) def forward_xpu( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, cu_seqlens: torch.Tensor | None = None, max_seqlen: torch.Tensor | None = None, # Only used for Flash Attention ) -> torch.Tensor: assert self.is_flash_attn_backend, ( "XPU only supports FLASH_ATTN for vision attention." ) return self._forward_fa(query, key, value, cu_seqlens, max_seqlen) def forward_tpu( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, cu_seqlens: torch.Tensor | None = None, max_seqlen: torch.Tensor | None = None, # Only used for Flash Attention ) -> torch.Tensor: assert self.attn_backend == AttentionBackendEnum.PALLAS, ( f"MMEncoderAttention on TPU only supports PALLAS backend, " f"but got {self.attn_backend}." ) if cu_seqlens is None: query, key, value = (x.transpose(1, 2) for x in (query, key, value)) from torch_xla.experimental.custom_kernel import flash_attention out = flash_attention(query, key, value, sm_scale=self.scale) out = out.transpose(1, 2) return out logger.warning_once( "PALLAS backend with cu_seqlens is not supported for ViT yet. ", "Falling back to SDPA implementation.", ) return self._forward_sdpa(query, key, value, cu_seqlens)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/backends/registry.py
vllm/attention/backends/registry.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Attention backend registry""" from collections.abc import Callable from enum import Enum, EnumMeta from typing import TYPE_CHECKING, cast from vllm.logger import init_logger from vllm.utils.import_utils import resolve_obj_by_qualname if TYPE_CHECKING: from vllm.attention.backends.abstract import AttentionBackend logger = init_logger(__name__) class _AttentionBackendEnumMeta(EnumMeta): """Metaclass for AttentionBackendEnum to provide better error messages.""" def __getitem__(cls, name: str): """Get backend by name with helpful error messages.""" try: return super().__getitem__(name) except KeyError: members = cast("dict[str, Enum]", cls.__members__).keys() valid_backends = ", ".join(members) raise ValueError( f"Unknown attention backend: '{name}'. " f"Valid options are: {valid_backends}" ) from None class AttentionBackendEnum(Enum, metaclass=_AttentionBackendEnumMeta): """Enumeration of all supported attention backends. The enum value is the default class path, but this can be overridden at runtime using register_backend(). To get the actual backend class (respecting overrides), use: backend.get_class() """ FLASH_ATTN = "vllm.v1.attention.backends.flash_attn.FlashAttentionBackend" FLASH_ATTN_DIFFKV = ( "vllm.v1.attention.backends.flash_attn_diffkv.FlashAttentionDiffKVBackend" ) TRITON_ATTN = "vllm.v1.attention.backends.triton_attn.TritonAttentionBackend" ROCM_ATTN = "vllm.v1.attention.backends.rocm_attn.RocmAttentionBackend" ROCM_AITER_MLA = "vllm.v1.attention.backends.mla.rocm_aiter_mla.AiterMLABackend" ROCM_AITER_TRITON_MLA = ( "vllm.v1.attention.backends.mla.aiter_triton_mla.AiterTritonMLABackend" ) ROCM_AITER_FA = ( "vllm.v1.attention.backends.rocm_aiter_fa.AiterFlashAttentionBackend" ) ROCM_AITER_MLA_SPARSE = ( "vllm.v1.attention.backends.mla.rocm_aiter_mla_sparse.ROCMAiterMLASparseBackend" ) TORCH_SDPA = "" # this tag is only used for ViT FLASHINFER = "vllm.v1.attention.backends.flashinfer.FlashInferBackend" FLASHINFER_MLA = ( "vllm.v1.attention.backends.mla.flashinfer_mla.FlashInferMLABackend" ) TRITON_MLA = "vllm.v1.attention.backends.mla.triton_mla.TritonMLABackend" CUTLASS_MLA = "vllm.v1.attention.backends.mla.cutlass_mla.CutlassMLABackend" FLASHMLA = "vllm.v1.attention.backends.mla.flashmla.FlashMLABackend" FLASHMLA_SPARSE = ( "vllm.v1.attention.backends.mla.flashmla_sparse.FlashMLASparseBackend" ) FLASH_ATTN_MLA = "vllm.v1.attention.backends.mla.flashattn_mla.FlashAttnMLABackend" PALLAS = "vllm.v1.attention.backends.pallas.PallasAttentionBackend" IPEX = "vllm.v1.attention.backends.ipex.IpexAttentionBackend" NO_ATTENTION = "vllm.v1.attention.backends.no_attention.NoAttentionBackend" FLEX_ATTENTION = "vllm.v1.attention.backends.flex_attention.FlexAttentionBackend" TREE_ATTN = "vllm.v1.attention.backends.tree_attn.TreeAttentionBackend" ROCM_AITER_UNIFIED_ATTN = ( "vllm.v1.attention.backends.rocm_aiter_unified_attn." "RocmAiterUnifiedAttentionBackend" ) CPU_ATTN = "vllm.v1.attention.backends.cpu_attn.CPUAttentionBackend" # Placeholder for third-party/custom backends - must be registered before use # set to None to avoid alias with other backend, whose value is an empty string CUSTOM = None def get_path(self, include_classname: bool = True) -> str: """Get the class path for this backend (respects overrides). Returns: The fully qualified class path string Raises: ValueError: If Backend.CUSTOM is used without being registered """ path = _ATTN_OVERRIDES.get(self, self.value) if not path: raise ValueError( f"Backend {self.name} must be registered before use. " f"Use register_backend(Backend.{self.name}, 'your.module.YourClass')" ) if not include_classname: path = path.rsplit(".", 1)[0] return path def get_class(self) -> "type[AttentionBackend]": """Get the backend class (respects overrides). Returns: The backend class Raises: ImportError: If the backend class cannot be imported ValueError: If Backend.CUSTOM is used without being registered """ return resolve_obj_by_qualname(self.get_path()) def is_overridden(self) -> bool: """Check if this backend has been overridden. Returns: True if the backend has a registered override """ return self in _ATTN_OVERRIDES def clear_override(self) -> None: """Clear any override for this backend, reverting to the default.""" _ATTN_OVERRIDES.pop(self, None) class MambaAttentionBackendEnum(Enum, metaclass=_AttentionBackendEnumMeta): """Enumeration of all supported mamba attention backends. The enum value is the default class path, but this can be overridden at runtime using register_backend(). To get the actual backend class (respecting overrides), use: backend.get_class() """ MAMBA1 = "vllm.v1.attention.backends.mamba1_attn.Mamba1AttentionBackend" MAMBA2 = "vllm.v1.attention.backends.mamba2_attn.Mamba2AttentionBackend" SHORT_CONV = "vllm.v1.attention.backends.short_conv_attn.ShortConvAttentionBackend" LINEAR = "vllm.v1.attention.backends.linear_attn.LinearAttentionBackend" GDN_ATTN = "vllm.v1.attention.backends.gdn_attn.GDNAttentionBackend" # Placeholder for third-party/custom backends - must be registered before use # set to None to avoid alias with other backend, whose value is an empty string CUSTOM = None def get_path(self, include_classname: bool = True) -> str: """Get the class path for this backend (respects overrides). Returns: The fully qualified class path string Raises: ValueError: If Backend.CUSTOM is used without being registered """ path = _MAMBA_ATTN_OVERRIDES.get(self, self.value) if not path: raise ValueError( f"Backend {self.name} must be registered before use. " f"Use register_backend(Backend.{self.name}, 'your.module.YourClass')" ) if not include_classname: path = path.rsplit(".", 1)[0] return path def get_class(self) -> "type[AttentionBackend]": """Get the backend class (respects overrides). Returns: The backend class Raises: ImportError: If the backend class cannot be imported ValueError: If Backend.CUSTOM is used without being registered """ return resolve_obj_by_qualname(self.get_path()) def is_overridden(self) -> bool: """Check if this backend has been overridden. Returns: True if the backend has a registered override """ return self in _MAMBA_ATTN_OVERRIDES def clear_override(self) -> None: """Clear any override for this backend, reverting to the default.""" _MAMBA_ATTN_OVERRIDES.pop(self, None) MAMBA_TYPE_TO_BACKEND_MAP = { "mamba1": MambaAttentionBackendEnum.MAMBA1.name, "mamba2": MambaAttentionBackendEnum.MAMBA2.name, "short_conv": MambaAttentionBackendEnum.SHORT_CONV.name, "linear_attention": MambaAttentionBackendEnum.LINEAR.name, "gdn_attention": MambaAttentionBackendEnum.GDN_ATTN.name, "custom": MambaAttentionBackendEnum.CUSTOM.name, } _ATTN_OVERRIDES: dict[AttentionBackendEnum, str] = {} _MAMBA_ATTN_OVERRIDES: dict[MambaAttentionBackendEnum, str] = {} def register_backend( backend: AttentionBackendEnum | MambaAttentionBackendEnum, class_path: str | None = None, is_mamba: bool = False, ) -> Callable[[type], type]: """Register or override a backend implementation. Args: backend: The AttentionBackendEnum member to register class_path: Optional class path. If not provided and used as decorator, will be auto-generated from the class. Returns: Decorator function if class_path is None, otherwise a no-op Examples: # Override an existing attention backend @register_backend(AttentionBackendEnum.FLASH_ATTN) class MyCustomFlashAttn: ... # Override an existing mamba attention backend @register_backend(MambaAttentionBackendEnum.LINEAR, is_mamba=True) class MyCustomMambaAttn: ... # Register a custom third-party attention backend @register_backend(AttentionBackendEnum.CUSTOM) class MyCustomBackend: ... # Direct registration register_backend( AttentionBackendEnum.CUSTOM, "my.module.MyCustomBackend" ) """ def decorator(cls: type) -> type: if is_mamba: _MAMBA_ATTN_OVERRIDES[backend] = f"{cls.__module__}.{cls.__qualname__}" # type: ignore[index] else: _ATTN_OVERRIDES[backend] = f"{cls.__module__}.{cls.__qualname__}" # type: ignore[index] return cls if class_path is not None: if is_mamba: _MAMBA_ATTN_OVERRIDES[backend] = class_path # type: ignore[index] else: _ATTN_OVERRIDES[backend] = class_path # type: ignore[index] return lambda x: x return decorator
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/backends/abstract.py
vllm/attention/backends/abstract.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import ABC, abstractmethod from typing import TYPE_CHECKING, ClassVar, Generic, Protocol, TypeVar, get_args import torch if TYPE_CHECKING: from vllm.config.cache import CacheDType from vllm.model_executor.layers.linear import ColumnParallelLinear from vllm.model_executor.layers.quantization.utils.quant_utils import QuantKey from vllm.platforms.interface import DeviceCapability from vllm.v1.attention.backends.utils import KVCacheLayoutType class AttentionType: """ Attention type. Use string to be compatible with `torch.compile`. """ DECODER = "decoder" """Decoder attention between previous layer Q/K/V.""" ENCODER = "encoder" """Encoder attention between previous layer Q/K/V for encoder-decoder.""" ENCODER_ONLY = "encoder_only" """Encoder attention between previous layer Q/K/V.""" ENCODER_DECODER = "encoder_decoder" """Attention between dec. Q and enc. K/V for encoder-decoder.""" class MultipleOf: base: int def __init__(self, base: int): self.base = base class AttentionBackend(ABC): """Abstract class for attention backends.""" # For some attention backends, we allocate an output tensor before # calling the custom op. When piecewise cudagraph is enabled, this # makes sure the output tensor is allocated inside the cudagraph. accept_output_buffer: bool = False supported_dtypes: ClassVar[list[torch.dtype]] = [torch.float16, torch.bfloat16] supported_kv_cache_dtypes: ClassVar[list["CacheDType"]] = ["auto"] @staticmethod def get_supported_kernel_block_sizes() -> list[int | MultipleOf]: return [MultipleOf(1)] @staticmethod @abstractmethod def get_name() -> str: raise NotImplementedError @staticmethod @abstractmethod def get_impl_cls() -> type["AttentionImpl"]: raise NotImplementedError @staticmethod @abstractmethod def get_builder_cls(): # -> Type["AttentionMetadataBuilder"]: raise NotImplementedError @staticmethod @abstractmethod def get_kv_cache_shape( num_blocks: int, block_size: int, num_kv_heads: int, head_size: int, cache_dtype_str: str = "auto", ) -> tuple[int, ...]: raise NotImplementedError @staticmethod def get_kv_cache_stride_order( include_num_layers_dimension: bool = False, ) -> tuple[int, ...]: """ Get the physical (memory layout) ordering of the kv cache dimensions. e.g. if the KV cache shape is [2, num_blocks, block_size, num_heads, head_size], and get_kv_cache_stride_order returns (1, 3, 0, 2, 4) then the physical ordering of dimensions is [num_blocks, num_heads, 2, block_size, head_size]. If this function is unimplemented / raises NotImplementedError, the physical layout of the KV cache will match the logical shape. Args: include_num_layers_dimension: if True, includes an additional num_layers dimension, which is assumed to be prepended to the logical KV cache shape. With the above example, a return value (2, 4, 0, 1, 3, 5) corresponds to [num_blocks, num_heads, num_layers, 2, block_size, head_size]. If an additional dimension is NOT included in the returned tuple, the physical layout will not include a layers dimension. Returns: A tuple of ints which is a permutation of range(len(shape)). """ raise NotImplementedError @classmethod def full_cls_name(cls) -> tuple[str, str]: return (cls.__module__, cls.__qualname__) @classmethod def get_supported_head_sizes(cls) -> list[int]: return [] @classmethod def supports_head_size(cls, head_size: int) -> bool: supported_head_sizes = cls.get_supported_head_sizes() return (not supported_head_sizes) or head_size in supported_head_sizes @classmethod def supports_dtype(cls, dtype: torch.dtype) -> bool: return dtype in cls.supported_dtypes @classmethod def supports_kv_cache_dtype(cls, kv_cache_dtype: "CacheDType | None") -> bool: if kv_cache_dtype is None: return True return (not cls.supported_kv_cache_dtypes) or ( kv_cache_dtype in cls.supported_kv_cache_dtypes ) @classmethod def supports_block_size(cls, block_size: int | None) -> bool: from vllm.config.cache import BlockSize if block_size is None: return True valid_sizes = get_args(BlockSize) if block_size not in valid_sizes: return False supported_kernel_block_sizes = cls.get_supported_kernel_block_sizes() if not supported_kernel_block_sizes: return True for supported_size in supported_kernel_block_sizes: if isinstance(supported_size, MultipleOf): supported_size = supported_size.base # With hybrid_blocks feature, the framework-level block size # only needs to be a multiple of the kernel's requirement, # even if the kernel requires a fixed block_size. if block_size % supported_size == 0: return True return False @classmethod def is_mla(cls) -> bool: return False @classmethod def supports_sink(cls) -> bool: return False @classmethod def supports_mm_prefix(cls) -> bool: return False @classmethod def is_sparse(cls) -> bool: return False @classmethod def supports_attn_type(cls, attn_type: str) -> bool: """Check if backend supports a given attention type. By default, only supports decoder attention. Backends should override this to support other attention types. """ return attn_type == AttentionType.DECODER @classmethod def supports_compute_capability(cls, capability: "DeviceCapability") -> bool: return True @classmethod def supports_combination( cls, head_size: int, dtype: torch.dtype, kv_cache_dtype: "CacheDType | None", block_size: int | None, use_mla: bool, has_sink: bool, use_sparse: bool, device_capability: "DeviceCapability", ) -> str | None: return None @classmethod def validate_configuration( cls, head_size: int, dtype: torch.dtype, kv_cache_dtype: "CacheDType | None", block_size: int | None, use_mla: bool, has_sink: bool, use_sparse: bool, use_mm_prefix: bool, device_capability: "DeviceCapability", attn_type: str, ) -> list[str]: invalid_reasons = [] if not cls.supports_head_size(head_size): invalid_reasons.append("head_size not supported") if not cls.supports_dtype(dtype): invalid_reasons.append("dtype not supported") if not cls.supports_kv_cache_dtype(kv_cache_dtype): invalid_reasons.append("kv_cache_dtype not supported") if not cls.supports_block_size(block_size): invalid_reasons.append("block_size not supported") if use_mm_prefix and not cls.supports_mm_prefix(): invalid_reasons.append( "partial multimodal token full attention not supported" ) if use_mla != cls.is_mla(): if use_mla: invalid_reasons.append("MLA not supported") else: invalid_reasons.append("non-MLA not supported") if has_sink and not cls.supports_sink(): invalid_reasons.append("sink setting not supported") if use_sparse != cls.is_sparse(): if use_sparse: invalid_reasons.append("sparse not supported") else: invalid_reasons.append("non-sparse not supported") if not cls.supports_compute_capability(device_capability): invalid_reasons.append("compute capability not supported") if not cls.supports_attn_type(attn_type): invalid_reasons.append(f"attention type {attn_type} not supported") combination_reason = cls.supports_combination( head_size, dtype, kv_cache_dtype, block_size, use_mla, has_sink, use_sparse, device_capability, ) if combination_reason is not None: invalid_reasons.append(combination_reason) return invalid_reasons @classmethod def get_required_kv_cache_layout(cls) -> "KVCacheLayoutType | None": return None class AttentionMetadata: pass T = TypeVar("T", bound=AttentionMetadata) class AttentionLayer(Protocol): _q_scale: torch.Tensor _k_scale: torch.Tensor _v_scale: torch.Tensor _q_scale_float: float _k_scale_float: float _v_scale_float: float _prob_scale: torch.Tensor def forward( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, kv_cache: torch.Tensor, attn_metadata: AttentionMetadata, ) -> torch.Tensor: ... class AttentionImpl(ABC, Generic[T]): # Whether the attention impl can return the softmax lse for decode. # Some features like decode context parallelism require the softmax lse. can_return_lse_for_decode: bool = False # Whether the attention impl supports Prefill Context Parallelism. supports_pcp: bool = False # Whether the attention impl(or ops) supports MTP # when cp_kv_cache_interleave_size > 1 supports_mtp_with_cp_non_trivial_interleave_size: bool = False # some attention backends might not always want to return lse # even if they can return lse (for efficiency reasons) need_to_return_lse_for_decode: bool = False # Whether this attention implementation supports pre-quantized query input. # When True, the attention layer will quantize queries before passing them # to this backend, allowing torch.compile to fuse the quantization with # previous operations. This is typically supported when using FP8 KV cache # with compatible attention kernels (e.g., TRT-LLM). # Subclasses should set this in __init__. # TODO add support to more backends: # https://github.com/vllm-project/vllm/issues/25584 supports_quant_query_input: bool = False dcp_world_size: int dcp_rank: int pcp_world_size: int pcp_rank: int total_cp_world_size: int total_cp_rank: int def __new__(cls, *args, **kwargs): # use __new__ so that all subclasses will call this self = super().__new__(cls) try: from vllm.distributed.parallel_state import get_dcp_group self.dcp_world_size = get_dcp_group().world_size self.dcp_rank = get_dcp_group().rank_in_group except AssertionError: # DCP might not be initialized in testing self.dcp_world_size = 1 self.dcp_rank = 0 try: from vllm.distributed.parallel_state import get_pcp_group self.pcp_world_size = get_pcp_group().world_size self.pcp_rank = get_pcp_group().rank_in_group except AssertionError: self.pcp_world_size = 1 self.pcp_rank = 0 self.total_cp_world_size = self.pcp_world_size * self.dcp_world_size self.total_cp_rank = self.pcp_rank * self.dcp_world_size + self.dcp_rank self.need_to_return_lse_for_decode = ( self.dcp_world_size > 1 and self.can_return_lse_for_decode ) return self @abstractmethod def __init__( self, num_heads: int, head_size: int, scale: float, num_kv_heads: int | None = None, alibi_slopes: list[float] | None = None, sliding_window: int | None = None, kv_cache_dtype: str = "auto", logits_soft_cap: float | None = None, attn_type: str = AttentionType.DECODER, kv_sharing_target_layer_name: str | None = None, ) -> None: raise NotImplementedError @abstractmethod def forward( self, layer: AttentionLayer, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, kv_cache: torch.Tensor, attn_metadata: T, output: torch.Tensor | None = None, output_scale: torch.Tensor | None = None, output_block_scale: torch.Tensor | None = None, ) -> torch.Tensor: raise NotImplementedError def fused_output_quant_supported(self, quant_key: "QuantKey"): """ Does this attention implementation support fused output quantization. This is used by the AttnFusionPass to only fuse output quantization onto implementations that support it. :param quant_key: QuantKey object that describes the quantization op :return: is fusion supported for this type of quantization """ return False def process_weights_after_loading(self, act_dtype: torch.dtype): pass class MLAAttentionImpl(AttentionImpl[T], Generic[T]): @abstractmethod def __init__( self, num_heads: int, head_size: int, scale: float, num_kv_heads: int, alibi_slopes: list[float] | None, sliding_window: int | None, kv_cache_dtype: str, logits_soft_cap: float | None, attn_type: str, kv_sharing_target_layer_name: str | None, # MLA Specific Arguments q_lora_rank: int | None, kv_lora_rank: int, qk_nope_head_dim: int, qk_rope_head_dim: int, qk_head_dim: int, v_head_dim: int, kv_b_proj: "ColumnParallelLinear", indexer: object | None = None, ) -> None: raise NotImplementedError @abstractmethod def forward( self, layer: AttentionLayer, hidden_states_or_cq: torch.Tensor, kv_c_normed: torch.Tensor, k_pe: torch.Tensor, kv_cache: torch.Tensor, attn_metadata: T, output: torch.Tensor | None = None, output_scale: torch.Tensor | None = None, output_block_scale: torch.Tensor | None = None, ) -> torch.Tensor: raise NotImplementedError def is_quantized_kv_cache(kv_cache_dtype: str) -> bool: return kv_cache_dtype != "auto"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/backends/utils.py
vllm/attention/backends/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Attention backend utils""" from dataclasses import dataclass from vllm.config import ModelConfig from vllm.logger import init_logger logger = init_logger(__name__) PAD_SLOT_ID = -1 @dataclass class MLADims: q_lora_rank: int | None kv_lora_rank: int qk_nope_head_dim: int qk_rope_head_dim: int v_head_dim: int def get_mla_dims(model_config: ModelConfig) -> MLADims: hf_text_config = model_config.hf_text_config return MLADims( q_lora_rank=getattr(hf_text_config, "q_lora_rank", None), kv_lora_rank=hf_text_config.kv_lora_rank, qk_nope_head_dim=hf_text_config.qk_nope_head_dim, qk_rope_head_dim=hf_text_config.qk_rope_head_dim, v_head_dim=hf_text_config.v_head_dim, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/backends/__init__.py
vllm/attention/backends/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/vit_attn_wrappers.py
vllm/attention/ops/vit_attn_wrappers.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ This file contains ops for ViT attention to be compatible with torch.compile as there are operations here not supported by torch.compile (for instance, `.item()` in flash attention) Using these ops and wrapping vision blocks with `torch.compile` can speed up throughput in vision models by ~5% relative on H100, and improve token latencies by ~7% (see qwen2_5_vl for example usage) To use these ops, you must have a recent version of PyTorch installed (>= 2.4.0) """ import einops import torch import torch.nn.functional as F from vllm.platforms import current_platform from vllm.utils.torch_utils import direct_register_custom_op def flash_attn_maxseqlen_wrapper( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, batch_size: int, is_rocm_aiter: bool, fa_version: int | None, cu_seqlens: torch.Tensor | None = None, max_seqlen: torch.Tensor | None = None, ) -> torch.Tensor: kwargs = {} if is_rocm_aiter: from aiter import flash_attn_varlen_func else: from vllm.attention.utils.fa_utils import flash_attn_varlen_func if not current_platform.is_rocm() and fa_version is not None: kwargs["fa_version"] = fa_version q_len = q.size(1) if cu_seqlens is None: cu_seqlens = torch.arange( 0, (batch_size + 1) * q_len, step=q_len, dtype=torch.int32, device=q.device ) max_seqlen = q_len if max_seqlen is None else max_seqlen.item() q, k, v = (einops.rearrange(x, "b s ... -> (b s) ...") for x in [q, k, v]) output = flash_attn_varlen_func( q, k, v, cu_seqlens_q=cu_seqlens, cu_seqlens_k=cu_seqlens, max_seqlen_q=max_seqlen, max_seqlen_k=max_seqlen, dropout_p=0.0, causal=False, **kwargs, ) context_layer = einops.rearrange(output, "(b s) h d -> b s h d", b=batch_size) return context_layer def flash_attn_maxseqlen_wrapper_fake( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, cu_seqlens: torch.Tensor, max_seqlen: torch.Tensor, batch_size: int, is_rocm_aiter: bool, fa_version: int | None, ) -> torch.Tensor: return torch.empty_like(q) direct_register_custom_op( op_name="flash_attn_maxseqlen_wrapper", op_func=flash_attn_maxseqlen_wrapper, fake_impl=flash_attn_maxseqlen_wrapper_fake, ) def vit_flash_attn_wrapper( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, batch_size: int, is_rocm_aiter: bool, fa_version: int | None, cu_seqlens: torch.Tensor | None = None, max_seqlen: torch.Tensor | None = None, ) -> torch.Tensor: return torch.ops.vllm.flash_attn_maxseqlen_wrapper( q, k, v, batch_size, is_rocm_aiter, fa_version, cu_seqlens, max_seqlen, ) def apply_sdpa(q: torch.Tensor, k: torch.Tensor, v: torch.Tensor) -> torch.Tensor: """ Input shape: (batch_size x seq_len x num_heads x head_size) """ q, k, v = (einops.rearrange(x, "b s h d -> b h s d") for x in [q, k, v]) output = F.scaled_dot_product_attention(q, k, v, dropout_p=0.0) output = einops.rearrange(output, "b h s d -> b s h d ") return output # TODO: Once we have a torch 2.10, we can use tensor slices # so we won't need to wrap this in custom ops def torch_sdpa_wrapper( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, cu_seqlens: torch.Tensor | None = None, ) -> torch.Tensor: # Never remove the contiguous logic for ROCm # Without it, hallucinations occur with the backend if current_platform.is_rocm(): q = q.contiguous() k = k.contiguous() v = v.contiguous() if cu_seqlens is None: return apply_sdpa(q, k, v) outputs = [] lens = (cu_seqlens[1:] - cu_seqlens[:-1]).tolist() q_chunks = torch.split(q, lens, dim=1) k_chunks = torch.split(k, lens, dim=1) v_chunks = torch.split(v, lens, dim=1) for q_i, k_i, v_i in zip(q_chunks, k_chunks, v_chunks): output_i = apply_sdpa(q_i, k_i, v_i) outputs.append(output_i) context_layer = torch.cat(outputs, dim=1) return context_layer def torch_sdpa_wrapper_fake( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, cu_seqlens: torch.Tensor, ) -> torch.Tensor: return torch.empty_like(q) direct_register_custom_op( op_name="torch_sdpa_wrapper", op_func=torch_sdpa_wrapper, fake_impl=torch_sdpa_wrapper_fake, ) def vit_torch_sdpa_wrapper( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, cu_seqlens: torch.Tensor | None = None, ) -> torch.Tensor: return torch.ops.vllm.torch_sdpa_wrapper(q, k, v, cu_seqlens)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/triton_merge_attn_states.py
vllm/attention/ops/triton_merge_attn_states.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.triton_utils import tl, triton # Implements section 2.2 of https://www.arxiv.org/pdf/2501.01005 # can be used to combine partial attention results (in the split-KV case) def merge_attn_states( output: torch.Tensor, prefix_output: torch.Tensor, prefix_lse: torch.Tensor, suffix_output: torch.Tensor, suffix_lse: torch.Tensor, output_lse: torch.Tensor | None = None, ) -> None: num_tokens = output.shape[0] num_query_heads = output.shape[1] head_size = output.shape[2] padded_head_size = triton.next_power_of_2(head_size) # We assume the output stride on num_head is not always as same as the # `suffix_output` and `prefix_output`, as them might be padded by the attention # backend. prefix_head_stride = prefix_output.stride(1) output_head_stride = output.stride(1) # TODO(woosuk): Use CUDA kernel instead of Triton to minimize CPU overhead. merge_attn_states_kernel[(num_tokens, num_query_heads)]( output, output_lse, prefix_output, prefix_lse, suffix_output, suffix_lse, prefix_head_stride, output_head_stride, head_size, padded_head_size, output_lse is not None, ) @triton.jit def merge_attn_states_kernel( output, # [NUM_TOKENS, NUM_HEADS, HEAD_SIZE] output_lse, # [NUM_HEADS, NUM_TOKENS] prefix_output, # [NUM_TOKENS, NUM_HEADS, HEAD_SIZE] prefix_lse, # [NUM_HEADS, NUM_TOKENS] suffix_output, # [NUM_TOKENS, NUM_HEADS, HEAD_SIZE] suffix_lse, # [NUM_HEADS, NUM_TOKENS] prefix_head_stride, output_head_stride, HEAD_SIZE: tl.constexpr, PADDED_HEAD_SIZE: tl.constexpr, OUTPUT_LSE: tl.constexpr, ): token_idx = tl.program_id(0) num_tokens = tl.num_programs(0) head_idx = tl.program_id(1) num_heads = tl.num_programs(1) p_lse = tl.load(prefix_lse + head_idx * num_tokens + token_idx) s_lse = tl.load(suffix_lse + head_idx * num_tokens + token_idx) # FA2 and FA3 have different behavior for when the sum-exp is 0, this namely # arises with 0 len seqlens. FA3 returns -inf here while FA2 returns inf. # If we see an inf assume FA2 and convert inf to -inf for consistency # and correctness. Inf generally doesn't make sense in this context outside # of undefined-behavior/FA2-case, so I think this a safe assumption. p_lse = float("-inf") if p_lse == float("inf") else p_lse s_lse = float("-inf") if s_lse == float("inf") else s_lse max_lse = tl.maximum(p_lse, s_lse) p_lse = p_lse - max_lse s_lse = s_lse - max_lse # Will reuse precomputed Exp values for scale factor computation. p_se = tl.exp(p_lse) s_se = tl.exp(s_lse) out_se = p_se + s_se if OUTPUT_LSE: out_lse = tl.log(out_se) + max_lse tl.store(output_lse + head_idx * num_tokens + token_idx, out_lse) head_arange = tl.arange(0, PADDED_HEAD_SIZE) head_mask = head_arange < HEAD_SIZE p_out = tl.load( prefix_output + token_idx * num_heads * prefix_head_stride + head_idx * prefix_head_stride + head_arange, mask=head_mask, ) s_out = tl.load( suffix_output + token_idx * num_heads * prefix_head_stride + head_idx * prefix_head_stride + head_arange, mask=head_mask, ) # NOTE(woosuk): Be careful with the numerical stability. # We should compute the scale first, and then multiply it with the output. # Do not multiply the output with tl.exp(p_lse) or tl.exp(s_lse) directly. p_scale = p_se / out_se s_scale = s_se / out_se out = p_out * p_scale + s_out * s_scale tl.store( output + token_idx * num_heads * output_head_stride + head_idx * output_head_stride + head_arange, out, mask=head_mask, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/flashmla.py
vllm/attention/ops/flashmla.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # adapted from: https://github.com/deepseek-ai/FlashMLA/blob/main/flash_mla/flash_mla_interface.py import torch from vllm.logger import init_logger from vllm.platforms import current_platform logger = init_logger(__name__) if current_platform.is_cuda(): try: import vllm._flashmla_C # noqa: F401 _flashmla_C_AVAILABLE = True except ImportError: _flashmla_C_AVAILABLE = False else: _flashmla_C_AVAILABLE = False if current_platform.is_cuda(): try: import vllm._flashmla_extension_C # noqa: F401 _flashmla_extension_C_AVAILABLE = True except ImportError: _flashmla_extension_C_AVAILABLE = False else: _flashmla_extension_C_AVAILABLE = False def _is_flashmla_available() -> tuple[bool, str | None]: if not _flashmla_C_AVAILABLE: return ( False, "vllm._flashmla_C is not available, likely was not " "compiled due to insufficient nvcc version or a supported arch " "was not in the list of target arches to compile for.", ) if not _flashmla_extension_C_AVAILABLE: return ( False, "vllm._flashmla_extension_C is not available, likely " "was not compiled due to a build error.", ) return True, None def is_flashmla_dense_supported() -> tuple[bool, str | None]: """ Return: is_supported_flag, unsupported_reason (optional). """ is_availble, maybe_reason = _is_flashmla_available() if not is_availble: return False, maybe_reason if current_platform.get_device_capability()[0] != 9: return False, "FlashMLA Dense is only supported on Hopper devices." return True, None def is_flashmla_sparse_supported() -> tuple[bool, str | None]: """ Return: is_supported_flag, unsupported_reason (optional). """ is_availble, maybe_reason = _is_flashmla_available() if not is_availble: return False, maybe_reason if current_platform.get_device_capability()[0] not in (9, 10): return ( False, "FlashMLA Sparse is only supported on Hopper and Blackwell devices.", ) return True, None def get_mla_metadata( cache_seqlens: torch.Tensor, num_q_tokens_per_head_k: int, num_heads_k: int, num_heads_q: int | None = None, is_fp8_kvcache: bool = False, topk: int | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: """ Arguments: - cache_seqlens: (batch_size), dtype torch.int32. - num_q_tokens_per_head_k: Equals to num_q_tokens_per_q_seq * num_heads_q // num_heads_k. - num_heads_k: The number of k heads. - num_heads_q: The number of q heads. This argument is optional when sparse attention is not enabled - is_fp8_kvcache: Whether the k_cache and v_cache are in fp8 format. - topk: If not None, sparse attention will be enabled, and only tokens in the `indices` array passed to `flash_mla_with_kvcache_sm90` will be attended to. Returns: - tile_scheduler_metadata: (num_sm_parts, TileSchedulerMetaDataSize), dtype torch.int32. - num_splits: (batch_size + 1), dtype torch.int32. """ if is_fp8_kvcache and topk is None: return torch.ops._flashmla_extension_C.get_mla_decoding_metadata_dense_fp8( cache_seqlens, num_q_tokens_per_head_k, num_heads_k, ) return torch.ops._flashmla_C.get_mla_decoding_metadata( cache_seqlens, num_q_tokens_per_head_k, num_heads_k, num_heads_q, is_fp8_kvcache, topk, ) def flash_mla_with_kvcache( q: torch.Tensor, k_cache: torch.Tensor, block_table: torch.Tensor, cache_seqlens: torch.Tensor, head_dim_v: int, tile_scheduler_metadata: torch.Tensor, num_splits: torch.Tensor, softmax_scale: float | None = None, causal: bool = False, descale_q: torch.Tensor | None = None, descale_k: torch.Tensor | None = None, is_fp8_kvcache: bool = False, indices: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: """ Arguments: - q: (batch_size, seq_len_q, num_heads_q, head_dim). - k_cache: (num_blocks, page_block_size, num_heads_k, head_dim). - block_table: (batch_size, max_num_blocks_per_seq), torch.int32. - cache_seqlens: (batch_size), torch.int32. - head_dim_v: Head dimension of v. - tile_scheduler_metadata: (num_sm_parts, TileSchedulerMetaDataSize), torch.int32, returned by get_mla_metadata. - num_splits: (batch_size + 1), torch.int32, returned by get_mla_metadata. - softmax_scale: float. The scale of QK^T before applying softmax. Default to 1 / sqrt(head_dim). - causal: bool. Whether to apply causal attention mask. - descale_q: (batch_size), torch.float32. Descaling factors for Q, used for fp8 quantization. - descale_k: (batch_size), torch.float32. Descaling factors for K, used for fp8 quantization. - is_fp8_kvcache: bool. Whether the k_cache and v_cache are in fp8 format. For the format of FP8 KV cache, please refer to README.md - indices: (batch_size, seq_len_q, topk), torch.int32. If not None, sparse attention will be enabled, and only tokens in the `indices` array will be attended to. Invalid indices should be set to -1 or numbers >= total_seq_len_kv. For details about how to set up `indices`, please refer to README.md. Returns: - out: (batch_size, seq_len_q, num_heads_q, head_dim_v). - softmax_lse: (batch_size, num_heads_q, seq_len_q), torch.float32. """ if softmax_scale is None: softmax_scale = q.shape[-1] ** (-0.5) if indices is not None: # NOTE (zyongye): sparse attention is also causal # since it only attend to the tokens before # but here `causal` should not be specified assert not causal, "causal must be `false` if sparse attention is enabled." assert (descale_q is None) == (descale_k is None), ( "descale_q and descale_k should be both None or both not None" ) if indices is None and q.element_size() == 1: out, softmax_lse = torch.ops._flashmla_extension_C.fwd_kvcache_mla_fp8( q, k_cache, head_dim_v, cache_seqlens, block_table, softmax_scale, causal, tile_scheduler_metadata, num_splits, descale_q, descale_k, ) else: out, softmax_lse = torch.ops._flashmla_C.fwd_kvcache_mla( q, k_cache, head_dim_v, cache_seqlens, block_table, softmax_scale, causal, tile_scheduler_metadata, num_splits, is_fp8_kvcache, indices, ) return out, softmax_lse def flash_mla_sparse_prefill( q: torch.Tensor, kv: torch.Tensor, indices: torch.Tensor, sm_scale: float, d_v: int = 512, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Sparse attention prefill kernel Args: - q: [s_q, h_q, d_qk], bfloat16 - kv: [s_kv, h_kv, d_qk], bfloat16 - indices: [s_q, h_kv, topk], int32. Invalid indices should be set to -1 or numbers >= s_kv - sm_scale: float - d_v: The dimension of value vectors. Can only be 512 Returns: - (output, max_logits, lse) About the definition of output, max_logits and lse, please refer to README.md - output: [s_q, h_q, d_v], bfloat16 - max_logits: [s_q, h_q], float - lse: [s_q, h_q], float, 2-based log-sum-exp """ results = torch.ops._flashmla_C.sparse_prefill_fwd(q, kv, indices, sm_scale, d_v) return results # # TODO: Add fake functions # # @register_fake("_flashmla_C::get_mla_metadata") # def _get_mla_metadata_fake(....) -> Tuple[torch.Tensor, torch.Tensor]: # return .... # # @register_fake("_flashmla_C::fwd_kvcache_mla") # def _fwd_kvcache_mla_fake(....) -> Tuple[torch.Tensor, torch.Tensor]: # return .... #
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/common.py
vllm/attention/ops/common.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.distributed.parallel_state import GroupCoordinator from vllm.triton_utils import tl, triton @triton.jit def _correct_attn_cp_out_kernel( outputs_ptr, new_output_ptr, lses_ptr, vlse_ptr, outputs_stride_B, outputs_stride_H, outputs_stride_D, lses_stride_N, lses_stride_B, lses_stride_H, lse_idx, HEAD_DIM: tl.constexpr, N_ROUNDED: tl.constexpr, IS_BASE_E: tl.constexpr, ): """ Apply the all-gathered lses to correct each local rank's attention output. we still need perform a cross-rank reduction to obtain the final attention output. Args: outputs_ptr (triton.PointerType): Pointer to input tensor of shape [ B, H, D ] lses_ptr (triton.PointerType): Pointer to input tensor of shape [ N, B, H ] new_output_ptr (triton.PointerType): Pointer to output tensor of shape [ B, H, D ] vlse_ptr (triton.PointerType): Pointer to output tensor of shape [ B, H ] """ batch_idx = tl.program_id(axis=0).to(tl.int64) head_idx = tl.program_id(axis=1).to(tl.int64) d_offsets = tl.arange(0, HEAD_DIM) num_n_offsets = tl.arange(0, N_ROUNDED) # shape = [N] lse_offsets = ( num_n_offsets * lses_stride_N + batch_idx * lses_stride_B + head_idx * lses_stride_H ) # calc final lse lse = tl.load(lses_ptr + lse_offsets) lse = tl.where((lse != lse) | (lse == float("inf")), -float("inf"), lse) lse_max = tl.max(lse, axis=0) lse_max = tl.where(lse_max == -float("inf"), 0, lse_max) lse -= lse_max if IS_BASE_E: lse_exp = tl.exp(lse) lse_acc = tl.sum(lse_exp, axis=0) lse = tl.log(lse_acc) else: lse_exp = tl.exp2(lse) lse_acc = tl.sum(lse_exp, axis=0) lse = tl.log2(lse_acc) lse += lse_max lse_offsets = batch_idx * lses_stride_B + head_idx * lses_stride_H tl.store(vlse_ptr + lse_offsets, lse) # shape = [D] output_offsets = ( batch_idx * outputs_stride_B + head_idx * outputs_stride_H + d_offsets * outputs_stride_D ) # correct output lse_offset = ( lse_idx * lses_stride_N + batch_idx * lses_stride_B + head_idx * lses_stride_H ) lse_tmp = tl.load(lses_ptr + lse_offset) lse_finally = lse_tmp - lse lse_finally = tl.where( (lse_finally != lse_finally) | (lse_finally == float("inf")), -float("inf"), lse_finally, ) factor = tl.exp(lse_finally) if IS_BASE_E else tl.exp2(lse_finally) output = tl.load(outputs_ptr + output_offsets) output = output * factor tl.store(new_output_ptr + output_offsets, output) class CPTritonContext: """The CPTritonContext is used to avoid recompilation of the Triton JIT.""" def __init__(self): self.inner_kernel = None def call_kernel(self, kernel, grid, *regular_args, **const_args): if self.inner_kernel is None: self.inner_kernel = kernel[grid](*regular_args, **const_args) else: self.inner_kernel[grid](*regular_args) def correct_attn_out( out: torch.Tensor, lses: torch.Tensor, cp_rank: int, ctx: CPTritonContext, is_lse_base_on_e: bool = True, ) -> tuple[torch.Tensor, torch.Tensor]: """Correct the attention output using the all-gathered lses. Args: out: Tensor of shape [ B, H, D ] lses: Tensor of shape [ N, B, H ] cp_rank: Current rank in the context-parallel group ctx: Triton context to avoid recompilation Returns: Tuple of (out, lse) with corrected attention and final log-sum-exp. """ if ctx is None: ctx = CPTritonContext() # --- Normalize to 3D views --- if out.ndim == 4 and out.shape[1] == 1: out = out.squeeze(1) assert out.ndim == 3, f"expected out [B,H,D] or [B,1,H,D], got {tuple(out.shape)}" if lses.ndim == 4 and lses.shape[-1] == 1: lses = lses.squeeze(-1) if lses.ndim == 4 and lses.shape[1] == 1: lses = lses.squeeze(1) assert lses.ndim == 3, ( f"expected lses [N,B,H] (optionally with a 1-sized extra dim), " f"got {tuple(lses.shape)}" ) B, H, D = out.shape N = lses.shape[0] # Strides after we normalized shapes to 3-D views. The kernel computes # offsets for `vlse_ptr` using lses_stride_B/H, so the output buffer must # have the same B/H stride layout as a slice of `lses`. o_sB, o_sH, o_sD = out.stride() l_sN, l_sB, l_sH = lses.stride() # Allocate LSE with the same B/H strides as `lses` so writes land correctly # even when `lses` is a non-contiguous view (e.g., 4-D to 3-D squeeze). lse = torch.empty_strided( (B, H), (l_sB, l_sH), device=lses.device, dtype=lses.dtype ) # Kernel launch config grid = (B, H, 1) regular_args = ( out, out, lses, lse, o_sB, o_sH, o_sD, l_sN, l_sB, l_sH, cp_rank, ) const_args = {"HEAD_DIM": D, "N_ROUNDED": N, "IS_BASE_E": is_lse_base_on_e} ctx.call_kernel(_correct_attn_cp_out_kernel, grid, *regular_args, **const_args) return out, lse def _cp_lse_common( cp_attn_out: torch.Tensor, cp_attn_lse: torch.Tensor, cp_group: GroupCoordinator, ctx: CPTritonContext | None = None, is_lse_base_on_e=True, ): """ cp_attn_out: [ B, H, D ] cp_attn_lse: [ B, H ] """ if cp_group.world_size == 1: return cp_attn_out if ctx is None: ctx = CPTritonContext() lses = torch.empty( (cp_group.world_size,) + cp_attn_lse.shape, dtype=cp_attn_lse.dtype, device=cp_attn_lse.device, ) cp_attn_lse = cp_attn_lse.contiguous() lses = cp_group.all_gather(cp_attn_lse, dim=0).view_as(lses) out, lse = correct_attn_out( cp_attn_out, lses, cp_group.rank_in_group, ctx, is_lse_base_on_e=is_lse_base_on_e, ) return out, lse def cp_lse_ag_out_rs( cp_attn_out: torch.Tensor, cp_attn_lse: torch.Tensor, cp_group: GroupCoordinator, ctx: CPTritonContext | None = None, return_lse: bool = False, is_lse_base_on_e=True, ): """ cp_attn_out: [ B, H, D ] cp_attn_lse: [ B, H ] """ out, lse = _cp_lse_common( cp_attn_out, cp_attn_lse, cp_group, ctx=ctx, is_lse_base_on_e=is_lse_base_on_e ) out = cp_group.reduce_scatter(out, dim=1) if return_lse: cp_num_heads = lse.shape[1] // cp_group.world_size cp_rank = cp_group.rank_in_group lse = lse[:, cp_num_heads * cp_rank : cp_num_heads * (cp_rank + 1)] return out, lse return out def cp_lse_ag_out_ar( cp_attn_out: torch.Tensor, cp_attn_lse: torch.Tensor, cp_group: GroupCoordinator, ctx: CPTritonContext | None = None, return_lse: bool = False, is_lse_base_on_e=True, ): """ cp_attn_out: [ B, H, D ] cp_attn_lse: [ B, H ] """ out, lse = _cp_lse_common( cp_attn_out, cp_attn_lse, cp_group, ctx=ctx, is_lse_base_on_e=is_lse_base_on_e ) out = cp_group.all_reduce(out) if return_lse: return out, lse return out @triton.jit def _pack_seq_kernel( x_ptr, # [N, D] out_ptr, # [B, Lmax, D] lengths_ptr, # *i32, [B] N: tl.constexpr, D: tl.constexpr, Lmax: tl.constexpr, PAD_VALUE: tl.constexpr, BLOCK_T: tl.constexpr, # timesteps per program BLOCK_D: tl.constexpr, # features per program ): pid_b = tl.program_id(0) # batch id pid_t = tl.program_id(1) # block over time dimension pid_d = tl.program_id(2) # block over feature dimension off_t = pid_t * BLOCK_T + tl.arange(0, BLOCK_T) # [BLOCK_T] off_d = pid_d * BLOCK_D + tl.arange(0, BLOCK_D) # [BLOCK_D] # Compute start index and sequence length from cumulative lengths in_start = 0 for i in range(pid_b): in_start += tl.load(lengths_ptr + i) seq_len = tl.load(lengths_ptr + pid_b) # valid time positions for this block t_mask = off_t < Lmax # compute input row indices for valid (b, t) in_row = in_start + off_t valid_row = (off_t < seq_len) & t_mask # Pointers # x_ptr: row-major [N, D] x_row_ptr = x_ptr + in_row[:, None] * D + off_d[None, :] # out_ptr: row-major [B, Lmax, D] out_row_ptr = out_ptr + (pid_b * Lmax + off_t)[:, None] * D + off_d[None, :] # Initialize with PAD (cast will occur as needed based on out_ptr dtype) d_mask = off_d[None, :] < D pad_vals = tl.full([BLOCK_T, BLOCK_D], PAD_VALUE, tl.float32) tl.store(out_row_ptr, pad_vals, mask=t_mask[:, None] & d_mask) # Load & write only where within seq_len x_vals = tl.load(x_row_ptr, mask=valid_row[:, None] & d_mask) tl.store(out_row_ptr, x_vals, mask=valid_row[:, None] & d_mask) def pack_seq_triton( x: torch.Tensor, lengths: torch.Tensor, pad_value: float = -float("inf"), block_t: int = 64, block_d: int = 64, ) -> torch.Tensor: """ Pack sequences of different lengths into a batched tensor. Args: x: [N, ...] - input tensor where N is total number of tokens lengths: [B] - sequence lengths for each batch pad_value: value to use for padding block_t: block size for time dimension block_d: block size for feature dimension Returns: packed: [B, Lmax, ...] - packed tensor """ # Handle multi-dimensional input by reshaping to (N, -1) original_shape = x.shape if len(original_shape) > 2: N = original_shape[0] x_reshaped = x.reshape(N, -1) D = x_reshaped.shape[1] else: N, D = x.shape x_reshaped = x B = lengths.numel() Lmax = int(lengths.max().item()) # Starts are computed inside the kernel from lengths out = torch.empty((B, Lmax, D), device=x.device, dtype=x.dtype) grid = (B, triton.cdiv(Lmax, block_t), triton.cdiv(D, block_d)) _pack_seq_kernel[grid]( x_reshaped, out, lengths.int(), N, D, Lmax, PAD_VALUE=float(pad_value), BLOCK_T=block_t, BLOCK_D=block_d, num_warps=4, num_stages=2, ) # Reshape output back to original dimensions (except first dimension) if len(original_shape) > 2: output_shape = (B, Lmax) + original_shape[1:] out = out.reshape(output_shape) return out @triton.jit def _unpack_seq_triton_kernel( packed_ptr, # [B, Lmax, D] out_ptr, # [N, D] lengths_ptr, # *i32, [B] B: tl.constexpr, Lmax: tl.constexpr, D: tl.constexpr, BLOCK_T: tl.constexpr, # timesteps per program BLOCK_D: tl.constexpr, # features per program ): pid_b = tl.program_id(0) # batch id pid_t = tl.program_id(1) # block over time dimension pid_d = tl.program_id(2) # block over feature dimension off_t = pid_t * BLOCK_T + tl.arange(0, BLOCK_T) # [BLOCK_T] off_d = pid_d * BLOCK_D + tl.arange(0, BLOCK_D) # [BLOCK_D] # bounds: compute start from cumulative lengths in_start = 0 for i in range(pid_b): in_start += tl.load(lengths_ptr + i) seq_len = tl.load(lengths_ptr + pid_b) # valid time positions for this block t_mask = off_t < Lmax valid_row = (off_t < seq_len) & t_mask # compute output row indices for valid (b, t) out_row = in_start + off_t # Pointers # packed_ptr: row-major [B, Lmax, D] packed_row_ptr = packed_ptr + (pid_b * Lmax + off_t)[:, None] * D + off_d[None, :] # out_ptr: row-major [N, D] out_row_ptr = out_ptr + out_row[:, None] * D + off_d[None, :] # Load from packed tensor and store to output d_mask = off_d[None, :] < D packed_vals = tl.load(packed_row_ptr, mask=valid_row[:, None] & d_mask) tl.store(out_row_ptr, packed_vals, mask=valid_row[:, None] & d_mask) def unpack_seq_triton( packed_tensor: torch.Tensor, lengths: torch.Tensor, block_t: int = 64, block_d: int = 64, ) -> torch.Tensor: """ Unpack a packed decode query tensor back to the original format. Efficient Triton implementation. Args: packed_tensor: [B, Lmax, ...] - packed tensor from pack_seq_triton lengths: [B] - sequence lengths for each batch block_t: block size for time dimension block_d: block size for feature dimension Returns: unpacked_tensor: [N, ...] where N = sum(lengths) """ # Handle multi-dimensional input by reshaping to (B, Lmax, -1) original_shape = packed_tensor.shape if len(original_shape) > 3: B, Lmax = original_shape[:2] packed_reshaped = packed_tensor.reshape(B, Lmax, -1) D = packed_reshaped.shape[2] else: B, Lmax, D = packed_tensor.shape packed_reshaped = packed_tensor # Calculate total number of elements N = int(lengths.sum().item()) out = torch.empty((N, D), device=packed_tensor.device, dtype=packed_tensor.dtype) grid = (B, triton.cdiv(Lmax, block_t), triton.cdiv(D, block_d)) _unpack_seq_triton_kernel[grid]( packed_reshaped, out, lengths.int(), B, Lmax, D, BLOCK_T=block_t, BLOCK_D=block_d, num_warps=4, num_stages=2, ) # Reshape output back to original dimensions (except first dimension) if len(original_shape) > 3: output_shape = (N,) + original_shape[2:] out = out.reshape(output_shape) return out
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/triton_unified_attention.py
vllm/attention/ops/triton_unified_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Authors: # - Burkhard Ringlein <ngl@zurich.ibm.com> # - Jan van Lunteren <jvl@zurich.ibm.com> # - Chih-Chieh Yang <chih.chieh.yang@ibm.com> # - Thomas Parnell <tpa@zurich.ibm.com> import torch from vllm.logger import init_logger from vllm.platforms import current_platform from vllm.triton_utils import tl, triton logger = init_logger(__name__) float8_info = torch.finfo(current_platform.fp8_dtype()) @triton.jit def cdiv_fn(x, y): return (x + y - 1) // y @triton.jit def apply_softcap(S, x): Sdiv = S / x p1 = tl.exp(Sdiv) p2 = tl.exp(-Sdiv) return x * (p1 - p2) / (p1 + p2) @triton.jit def find_seq_idx( query_start_len_ptr, target_idx, num_seqs, BLOCK_Q: tl.constexpr, use_q_block_mode: tl.constexpr, ): left: tl.int32 = 0 right = num_seqs while left < right: mid = (left + right) // 2 val = tl.load(query_start_len_ptr + mid) mid_val = val // BLOCK_Q + mid if use_q_block_mode else val if mid_val <= target_idx: left = mid + 1 else: right = mid return left - 1 @triton.jit def kernel_unified_attention_2d( output_ptr, # [num_tokens, num_query_heads, head_size] query_ptr, # [num_tokens, num_query_heads, head_size] key_cache_ptr, # [num_blks, blk_size, num_kv_heads, head_size] value_cache_ptr, # [num_blks, blk_size, num_kv_heads, head_size] sink_ptr, # [num_query_heads] block_tables_ptr, # [num_seqs, max_num_blocks_per_seq] seq_lens_ptr, # [num_seqs] alibi_slopes_ptr, # [num_query_heads] qq_bias_ptr, # [num_query_tokens, num_query_tokens] scale, # float32 k_scale, # float32 v_scale, # float32 out_scale, # float32 softcap, # float32 num_query_heads: tl.constexpr, # int num_queries_per_kv: tl.constexpr, # int block_table_stride: tl.int64, # int query_stride_0: tl.int64, # int query_stride_1: tl.int64, # int, should be equal to head_size output_stride_0: tl.int64, # int output_stride_1: tl.int64, # int, should be equal to head_size qq_bias_stride_0: tl.int64, # int BLOCK_SIZE: tl.constexpr, # int TILE_SIZE: tl.constexpr, # int must be power of 2 HEAD_SIZE: tl.constexpr, # int HEAD_SIZE_PADDED: tl.constexpr, # int, must be power of 2 USE_ALIBI_SLOPES: tl.constexpr, # bool USE_QQ_BIAS: tl.constexpr, # bool USE_SOFTCAP: tl.constexpr, # bool USE_SINKS: tl.constexpr, # bool SLIDING_WINDOW: tl.constexpr, # int USE_MM_PREFIX: tl.constexpr, # bool MAX_MM_RANGES: tl.constexpr, # int mm_prefix_range_ptr, # [num_seqs] - prefix length for each sequence stride_k_cache_0: tl.int64, # int stride_k_cache_1: tl.int64, # int stride_k_cache_2: tl.int64, # int stride_k_cache_3: tl.constexpr, # int stride_v_cache_0: tl.int64, # int stride_v_cache_1: tl.int64, # int stride_v_cache_2: tl.int64, # int stride_v_cache_3: tl.constexpr, # int query_start_len_ptr, # [num_seqs+1] BLOCK_Q: tl.constexpr, # int num_seqs: tl.int32, BLOCK_M: tl.constexpr, # int USE_FP8: tl.constexpr, # bool FP8_MIN: tl.constexpr = float8_info.min, FP8_MAX: tl.constexpr = float8_info.max, ): q_block_global_idx = tl.program_id(0) kv_head_idx = tl.program_id(1) seq_idx = find_seq_idx( query_start_len_ptr, q_block_global_idx, num_seqs, BLOCK_Q, True ) q_block_start_idx = tl.load(query_start_len_ptr + seq_idx) // BLOCK_Q + seq_idx q_block_local_idx = q_block_global_idx - q_block_start_idx cur_batch_in_all_start_index = tl.load(query_start_len_ptr + seq_idx) cur_batch_in_all_stop_index = tl.load(query_start_len_ptr + seq_idx + 1) cur_batch_query_len = cur_batch_in_all_stop_index - cur_batch_in_all_start_index if q_block_local_idx * BLOCK_Q >= cur_batch_query_len: return offs_m = tl.arange(0, BLOCK_M) offs_d = tl.arange(0, HEAD_SIZE_PADDED) offs_t = tl.arange(0, TILE_SIZE) query_pos = q_block_local_idx * BLOCK_Q + offs_m // num_queries_per_kv query_offset_0 = cur_batch_in_all_start_index + query_pos query_offset_1 = kv_head_idx * num_queries_per_kv + offs_m % num_queries_per_kv query_offset = ( query_offset_0[:, None] * query_stride_0 + query_offset_1[:, None] * query_stride_1 + offs_d[None, :] ) dim_mask = tl.where(offs_d < HEAD_SIZE, 1, 0).to(tl.int1) query_mask_0 = tl.where(query_pos < cur_batch_query_len, 1, 0).to(tl.int1) query_mask_1 = tl.where(query_offset_1 < num_query_heads, 1, 0).to(tl.int1) # Q : (BLOCK_M, HEAD_SIZE_PADDED) Q = tl.load( query_ptr + query_offset, mask=dim_mask[None, :] & query_mask_0[:, None] & query_mask_1[:, None], other=0.0, ) block_table_offset = seq_idx * block_table_stride if not USE_SINKS: M = tl.full([BLOCK_M], float("-inf"), dtype=tl.float32) else: M = tl.load( sink_ptr + query_offset_1, mask=query_mask_1, other=float("-inf"), ).to(dtype=tl.float32) L = tl.full([BLOCK_M], 1.0, dtype=tl.float32) acc = tl.zeros([BLOCK_M, HEAD_SIZE_PADDED], dtype=tl.float32) # sequence len for this particular sequence seq_len = tl.load(seq_lens_ptr + seq_idx) # context length for this particular sequences context_len = seq_len - cur_batch_query_len # alibi slope for this head if USE_ALIBI_SLOPES: alibi_slope = tl.load( alibi_slopes_ptr + query_offset_1, mask=query_mask_1, other=0.0 ) # query-query attention bias if USE_QQ_BIAS: qq_bias_row_ptrs = ( qq_bias_ptr + query_pos[:, None] * qq_bias_stride_0 ) # shape: [BLOCK_M] # compute the length of the longest sequence prefix spanned by any # query token in the current q_block (q_block_local_idx) max_seq_prefix_len = ( context_len + q_block_local_idx * BLOCK_Q + (BLOCK_M - 1) // num_queries_per_kv + 1 ) if USE_MM_PREFIX: # image bidirectional attention ranges require a full range # including q_block padding to make sure doc mask is correct max_seq_prefix_len = tl.maximum(max_seq_prefix_len, seq_len) else: # adjust for potential padding in the last q_block by considering the # actual sequence length max_seq_prefix_len = tl.minimum(max_seq_prefix_len, seq_len) # calculate the number of tiles that need to be processed to # cover the longest sequence prefix (due to causal masking, tiles beyond # this prefix can be skipped) num_tiles = cdiv_fn(max_seq_prefix_len, TILE_SIZE) # ---- Sliding-window tile pruning -------------------- # Default: keep previous global behavior tile_start = 0 tile_end = num_tiles # TODO(Isotr0py): sliding window pruning with image bidirectional mask if SLIDING_WINDOW > 0 and not USE_MM_PREFIX: # Query rows covered by this Q-block qpos_lo = q_block_local_idx * BLOCK_Q qpos_hi = tl.minimum( qpos_lo + (BLOCK_M - 1) // num_queries_per_kv, cur_batch_query_len - 1, ) # For sliding window, each query position q can only attend to # keys in the range [q_abs - SLIDING_WINDOW + 1, q_abs] # where q_abs = context_len + q # The union of allowed key positions for this Q-block is: # [context_len + qpos_lo - SLIDING_WINDOW + 1, context_len + qpos_hi] first_allowed_key = context_len + qpos_lo - SLIDING_WINDOW + 1 last_allowed_key = context_len + qpos_hi # Convert to tile indices and clamp tile_start = tl.maximum(0, first_allowed_key // TILE_SIZE) tile_end = tl.minimum((last_allowed_key // TILE_SIZE) + 1, num_tiles) # iterate through tiles (now limited to the sliding window range) for j in range(tile_start, tile_end): seq_offset = j * TILE_SIZE + offs_t tile_mask = seq_offset < max_seq_prefix_len physical_block_idx = tl.load( block_tables_ptr + block_table_offset + seq_offset // BLOCK_SIZE ).to(tl.int64) v_offset = ( physical_block_idx[:, None] * stride_v_cache_0 + kv_head_idx * stride_v_cache_2 + offs_d[None, :] * stride_v_cache_3 + (seq_offset % BLOCK_SIZE)[:, None] * stride_v_cache_1 ) k_offset = ( physical_block_idx[None, :] * stride_k_cache_0 + kv_head_idx * stride_k_cache_2 + offs_d[:, None] * stride_k_cache_3 + (seq_offset % BLOCK_SIZE)[None, :] * stride_k_cache_1 ) # K : (HEAD_SIZE, TILE_SIZE) K_load = tl.load( key_cache_ptr + k_offset, mask=dim_mask[:, None] & tile_mask[None, :], other=0.0, ) if K_load.dtype.is_fp8(): if Q.dtype.is_fp8(): K = K_load else: K = (K_load.to(tl.float32) * tl.load(k_scale)).to(Q.dtype) else: K = K_load # V : (TILE_SIZE, HEAD_SIZE) V_load = tl.load( value_cache_ptr + v_offset, mask=dim_mask[None, :] & tile_mask[:, None], other=0.0, ) if V_load.dtype.is_fp8(): if Q.dtype.is_fp8(): V = V_load else: V = (V_load.to(tl.float32) * tl.load(v_scale)).to(Q.dtype) else: V = V_load # Compute attention mask: causal by default (key <= query) query_abs_pos = context_len + query_pos[:, None] seq_mask = seq_offset[None, :] <= query_abs_pos # Apply sliding window to base mask BEFORE mm_prefix OR. # Order must match FlexAttention: (causal AND sliding_window) OR mm_prefix if SLIDING_WINDOW > 0: seq_mask = seq_mask & ((query_abs_pos - seq_offset) < SLIDING_WINDOW) # PrefixLM: extend mask with bidirectional ranges for multimodal tokens. # Applied AFTER sliding window so mm_prefix ranges override SW restriction. if USE_MM_PREFIX: for i in range(MAX_MM_RANGES): range_start = tl.load( mm_prefix_range_ptr + seq_idx * MAX_MM_RANGES * 2 + i * 2 ) range_end = tl.load( mm_prefix_range_ptr + seq_idx * MAX_MM_RANGES * 2 + i * 2 + 1 ) is_valid = range_start < range_end q_in_range = ( (query_abs_pos >= range_start) & (query_abs_pos <= range_end) & is_valid ) k_in_range = ( (seq_offset[None, :] >= range_start) & (seq_offset[None, :] <= range_end) & is_valid ) seq_mask |= q_in_range & k_in_range # S : (BLOCK_M, TILE_SIZE) S = tl.zeros(shape=(BLOCK_M, TILE_SIZE), dtype=tl.float32) S += scale * tl.dot(Q, K) if USE_SOFTCAP: S = apply_softcap(S, softcap) S = tl.where( query_mask_1[:, None] & query_mask_0[:, None] & seq_mask, S, float("-inf") ) if USE_ALIBI_SLOPES: S += alibi_slope[:, None] * (seq_offset - context_len) if USE_QQ_BIAS: # compute key positions relative to query section key_rel_pos = seq_offset - context_len # shape: [BLOCK_SIZE] # load bias only for keys that correspond to queries is_query_key = key_rel_pos >= 0 and key_rel_pos < qq_bias_stride_0 qq_bias = tl.load( qq_bias_row_ptrs + key_rel_pos[None, :], mask=is_query_key[None, :], # avoid OOB for context keys other=0.0, ) S += qq_bias # compute running maximum # m_j : (BLOCK_M,) m_j = tl.maximum(M, tl.max(S, axis=1)) # For sliding window there's a chance the max is -inf due to masking of # the entire row. In this case we need to set m_j 0 to avoid NaN m_j = tl.where(m_j > float("-inf"), m_j, 0.0) # P : (BLOCK_M, TILE_SIZE) P = tl.exp(S - m_j[:, None]) # l_j : (BLOCK_M,) l_j = tl.sum(P, axis=1) # alpha : (BLOCK_M, ) alpha = tl.exp(M - m_j) # acc : (BLOCK_M, HEAD_SIZE_PADDED) acc = acc * alpha[:, None] # update constants L = L * alpha + l_j M = m_j if SLIDING_WINDOW: qpos_lo = q_block_local_idx * BLOCK_Q V = tl.where( (context_len + qpos_lo - seq_offset[:, None]) < SLIDING_WINDOW, V, 0.0 ) # acc : (BLOCK_M, HEAD_SIZE_PADDED) acc += tl.dot(P.to(V.dtype), V) # epilogue acc = acc / L[:, None] if USE_FP8: acc = acc * tl.load(out_scale) acc = tl.clamp(acc, FP8_MIN, FP8_MAX) output_offset = ( query_offset_0[:, None] * output_stride_0 + query_offset_1[:, None] * output_stride_1 + offs_d[None, :] ) tl.store( output_ptr + output_offset, acc, mask=dim_mask[None, :] & query_mask_0[:, None] & query_mask_1[:, None], ) @triton.jit def kernel_unified_attention_3d( segm_output_ptr, # [num_tokens, num_query_heads, num_segments, head_size_padded] segm_max_ptr, # [num_tokens, num_query_heads, num_segments] segm_expsum_ptr, # [num_tokens, num_query_heads, num_segments] query_ptr, # [num_tokens, num_query_heads, head_size] key_cache_ptr, # [num_blks, num_kv_heads, head_size // x, blk_size, x] value_cache_ptr, # [num_blks, num_kv_heads, head_size, blk_size] sink_ptr, # [num_query_heads] block_tables_ptr, # [num_seqs, max_num_blocks_per_seq] seq_lens_ptr, # [num_seqs] alibi_slopes_ptr, # [num_query_heads] qq_bias_ptr, # [num_query_tokens, num_query_tokens] scale, # float32 k_scale, # float32 v_scale, # float32 softcap, # float32 num_query_heads: tl.constexpr, # int num_queries_per_kv: tl.constexpr, # int block_table_stride: tl.int64, # int query_stride_0: tl.int64, # int query_stride_1: tl.int64, # int, should be equal to head_size qq_bias_stride_0: tl.int64, # int BLOCK_SIZE: tl.constexpr, # int TILE_SIZE: tl.constexpr, # int, must be power of 2 HEAD_SIZE: tl.constexpr, # int HEAD_SIZE_PADDED: tl.constexpr, # int, must be power of 2 USE_ALIBI_SLOPES: tl.constexpr, # bool USE_QQ_BIAS: tl.constexpr, # bool USE_SOFTCAP: tl.constexpr, # bool USE_SINKS: tl.constexpr, # bool SLIDING_WINDOW: tl.constexpr, # int stride_k_cache_0: tl.int64, # int stride_k_cache_1: tl.int64, # int stride_k_cache_2: tl.int64, # int stride_k_cache_3: tl.constexpr, # int stride_v_cache_0: tl.int64, # int stride_v_cache_1: tl.int64, # int stride_v_cache_2: tl.int64, # int stride_v_cache_3: tl.constexpr, # int query_start_len_ptr, # [num_seqs+1] BLOCK_Q: tl.constexpr, # int num_seqs: tl.int32, BLOCK_M: tl.constexpr, # int NUM_SEGMENTS_PER_SEQ: tl.constexpr, # int USE_MM_PREFIX: tl.constexpr, # bool MAX_MM_RANGES: tl.constexpr, # int mm_prefix_range_ptr, # [num_seqs] - prefix length for each sequence ): q_block_global_idx = tl.program_id(0) kv_head_idx = tl.program_id(1) segm_idx = tl.program_id(2) seq_idx = find_seq_idx( query_start_len_ptr, q_block_global_idx, num_seqs, BLOCK_Q, True ) q_block_start_idx = tl.load(query_start_len_ptr + seq_idx) // BLOCK_Q + seq_idx q_block_local_idx = q_block_global_idx - q_block_start_idx cur_batch_in_all_start_index = tl.load(query_start_len_ptr + seq_idx) cur_batch_in_all_stop_index = tl.load(query_start_len_ptr + seq_idx + 1) cur_batch_query_len = cur_batch_in_all_stop_index - cur_batch_in_all_start_index if q_block_local_idx * BLOCK_Q >= cur_batch_query_len: return # sequence len for this particular sequence seq_len = tl.load(seq_lens_ptr + seq_idx) # number of segments for this particular sequence num_segments = NUM_SEGMENTS_PER_SEQ tiles_per_segment = cdiv_fn(seq_len, num_segments * TILE_SIZE) if segm_idx * tiles_per_segment * TILE_SIZE >= seq_len: return offs_m = tl.arange(0, BLOCK_M) offs_d = tl.arange(0, HEAD_SIZE_PADDED) offs_t = tl.arange(0, TILE_SIZE) query_pos = q_block_local_idx * BLOCK_Q + offs_m // num_queries_per_kv query_offset_0 = cur_batch_in_all_start_index + query_pos query_offset_1 = kv_head_idx * num_queries_per_kv + offs_m % num_queries_per_kv query_offset = ( query_offset_0[:, None] * query_stride_0 + query_offset_1[:, None] * query_stride_1 + offs_d[None, :] ) dim_mask = tl.where(offs_d < HEAD_SIZE, 1, 0).to(tl.int1) query_mask_0 = tl.where(query_pos < cur_batch_query_len, 1, 0).to(tl.int1) query_mask_1 = tl.where(query_offset_1 < num_query_heads, 1, 0).to(tl.int1) # Q : (BLOCK_M, HEAD_SIZE_PADDED) Q = tl.load( query_ptr + query_offset, mask=dim_mask[None, :] & query_mask_0[:, None] & query_mask_1[:, None], other=0.0, ) block_table_offset = seq_idx * block_table_stride if USE_SINKS: if segm_idx == 0: M = tl.load( sink_ptr + query_offset_1, mask=query_mask_1, other=float("-inf"), ).to(dtype=tl.float32) else: M = tl.full([BLOCK_M], float("-inf"), dtype=tl.float32) else: M = tl.full([BLOCK_M], float("-inf"), dtype=tl.float32) L = tl.full([BLOCK_M], 1.0, dtype=tl.float32) acc = tl.zeros([BLOCK_M, HEAD_SIZE_PADDED], dtype=tl.float32) # context length for this particular sequences context_len = seq_len - cur_batch_query_len # alibi slope for this head if USE_ALIBI_SLOPES: alibi_slope = tl.load( alibi_slopes_ptr + query_offset_1, mask=query_mask_1, other=0.0 ) # query-query attention bias if USE_QQ_BIAS: qq_bias_row_ptrs = ( qq_bias_ptr + query_pos[:, None] * qq_bias_stride_0 ) # shape: [BLOCK_M] # compute the length of the longest sequence prefix spanned by any # query token in the current q_block (q_block_local_idx) max_seq_prefix_len = ( context_len + q_block_local_idx * BLOCK_Q + (BLOCK_M - 1) // num_queries_per_kv + 1 ) # adjust for potential padding in the last q_block by considering the # actual sequence length max_seq_prefix_len = tl.minimum(max_seq_prefix_len, seq_len) # calculate the number of tiles that need to be processed to # cover the longest sequence prefix (due to causal masking, tiles beyond # this prefix can be skipped) num_tiles = cdiv_fn(max_seq_prefix_len, TILE_SIZE) # iterate through tiles within current segment for j in range( segm_idx * tiles_per_segment, min((segm_idx + 1) * tiles_per_segment, num_tiles), ): seq_offset = j * TILE_SIZE + offs_t tile_mask = seq_offset < max_seq_prefix_len physical_block_idx = tl.load( block_tables_ptr + block_table_offset + seq_offset // BLOCK_SIZE ).to(tl.int64) v_offset = ( physical_block_idx[:, None] * stride_v_cache_0 + kv_head_idx * stride_v_cache_2 + offs_d[None, :] * stride_v_cache_3 + (seq_offset % BLOCK_SIZE)[:, None] * stride_v_cache_1 ) k_offset = ( physical_block_idx[None, :] * stride_k_cache_0 + kv_head_idx * stride_k_cache_2 + offs_d[:, None] * stride_k_cache_3 + (seq_offset % BLOCK_SIZE)[None, :] * stride_k_cache_1 ) # K : (HEAD_SIZE, TILE_SIZE) K_load = tl.load( key_cache_ptr + k_offset, mask=dim_mask[:, None] & tile_mask[None, :], other=0.0, ) if K_load.dtype.is_fp8(): if Q.dtype.is_fp8(): K = K_load else: K = (K_load.to(tl.float32) * tl.load(k_scale)).to(Q.dtype) else: K = K_load # V : (TILE_SIZE, HEAD_SIZE) V_load = tl.load( value_cache_ptr + v_offset, mask=dim_mask[None, :] & tile_mask[:, None], other=0.0, ) if V_load.dtype.is_fp8(): if Q.dtype.is_fp8(): V = V_load else: V = (V_load.to(tl.float32) * tl.load(v_scale)).to(Q.dtype) else: V = V_load # Compute attention mask: causal by default (key <= query) query_abs_pos = context_len + query_pos[:, None] seq_mask = seq_offset[None, :] <= query_abs_pos # Apply sliding window to base mask BEFORE mm_prefix OR. # Order must match FlexAttention: (causal AND sliding_window) OR mm_prefix if SLIDING_WINDOW > 0: seq_mask = seq_mask & ((query_abs_pos - seq_offset) < SLIDING_WINDOW) # PrefixLM: extend mask with bidirectional ranges for multimodal tokens. # Applied AFTER sliding window so mm_prefix ranges override SW restriction. if USE_MM_PREFIX: for i in range(MAX_MM_RANGES): range_start = tl.load( mm_prefix_range_ptr + seq_idx * MAX_MM_RANGES * 2 + i * 2 ) range_end = tl.load( mm_prefix_range_ptr + seq_idx * MAX_MM_RANGES * 2 + i * 2 + 1 ) is_valid = range_start < range_end q_in_range = ( (query_abs_pos >= range_start) & (query_abs_pos <= range_end) & is_valid ) k_in_range = ( (seq_offset[None, :] >= range_start) & (seq_offset[None, :] <= range_end) & is_valid ) seq_mask |= q_in_range & k_in_range # S : (BLOCK_M, TILE_SIZE) S = tl.zeros(shape=(BLOCK_M, TILE_SIZE), dtype=tl.float32) S += scale * tl.dot(Q, K) if USE_SOFTCAP: S = apply_softcap(S, softcap) S = tl.where( query_mask_1[:, None] & query_mask_0[:, None] & seq_mask, S, float("-inf") ) if USE_ALIBI_SLOPES: S += alibi_slope[:, None] * (seq_offset - context_len) if USE_QQ_BIAS: # compute key positions relative to query section key_rel_pos = seq_offset - context_len # shape: [BLOCK_SIZE] # load bias only for keys that correspond to queries is_query_key = key_rel_pos >= 0 and key_rel_pos < qq_bias_stride_0 qq_bias = tl.load( qq_bias_row_ptrs + key_rel_pos[None, :], mask=is_query_key[None, :], # avoid OOB for context keys other=0.0, ) S += qq_bias # compute running maximum # m_j : (BLOCK_M,) m_j = tl.maximum(M, tl.max(S, axis=1)) # For sliding window there's a chance the max is -inf due to masking of # the entire row. In this case we need to set m_j 0 to avoid NaN m_j = tl.where(m_j > float("-inf"), m_j, 0.0) # P : (BLOCK_M, TILE_SIZE,) P = tl.exp(S - m_j[:, None]) # l_j : (BLOCK_M,) l_j = tl.sum(P, axis=1) # alpha : (BLOCK_M, ) alpha = tl.exp(M - m_j) # acc : (BLOCK_M, HEAD_SIZE_PADDED) acc = acc * alpha[:, None] # update constants L = L * alpha + l_j M = m_j if SLIDING_WINDOW: qpos_lo = q_block_local_idx * BLOCK_Q V = tl.where( (context_len + qpos_lo - seq_offset[:, None]) < SLIDING_WINDOW, V, 0.0 ) # acc : (BLOCK_M, HEAD_SIZE_PADDED) acc += tl.dot(P.to(V.dtype), V) segm_output_offset = ( query_offset_0[:, None].to(tl.int64) * (num_query_heads * NUM_SEGMENTS_PER_SEQ * HEAD_SIZE_PADDED) + query_offset_1[:, None] * (NUM_SEGMENTS_PER_SEQ * HEAD_SIZE_PADDED) + segm_idx * HEAD_SIZE_PADDED + tl.arange(0, HEAD_SIZE_PADDED)[None, :] ) tl.store( segm_output_ptr + segm_output_offset, acc, mask=dim_mask[None, :] & query_mask_0[:, None] & query_mask_1[:, None], ) segm_offset = ( query_offset_0.to(tl.int64) * (num_query_heads * NUM_SEGMENTS_PER_SEQ) + query_offset_1 * NUM_SEGMENTS_PER_SEQ + segm_idx ) tl.store(segm_max_ptr + segm_offset, M, mask=query_mask_0 & query_mask_1) tl.store(segm_expsum_ptr + segm_offset, L, mask=query_mask_0 & query_mask_1) @triton.jit def reduce_segments( output_ptr, # [num_tokens, num_query_heads, head_size] segm_output_ptr, # [num_tokens, num_query_heads, max_num_segments, head_size] segm_max_ptr, # [num_tokens, num_query_heads, max_num_segments] segm_expsum_ptr, # [num_tokens, num_query_heads, max_num_segments] seq_lens_ptr, # [num_seqs] num_seqs, # int num_query_heads: tl.constexpr, # int out_scale_inv, # float32 output_stride_0: tl.int64, # int output_stride_1: tl.int64, # int, should be equal to head_size block_table_stride: tl.int64, # int TILE_SIZE: tl.constexpr, # int HEAD_SIZE: tl.constexpr, # int, must be power of 2 HEAD_SIZE_PADDED: tl.constexpr, # int, must be power of 2 query_start_len_ptr, # [num_seqs+1] BLOCK_Q: tl.constexpr, # int NUM_SEGMENTS_PER_SEQ: tl.constexpr, # int USE_FP8: tl.constexpr, # bool FP8_MIN: tl.constexpr = float8_info.min, FP8_MAX: tl.constexpr = float8_info.max, ): query_token_idx = tl.program_id(0) query_head_idx = tl.program_id(1) seq_idx = find_seq_idx( query_start_len_ptr, query_token_idx, num_seqs, BLOCK_Q, False ) # sequence len for this particular sequence seq_len = tl.load(seq_lens_ptr + seq_idx) # number of segments for this particular sequence num_segments = NUM_SEGMENTS_PER_SEQ tiles_per_segment = cdiv_fn(seq_len, num_segments * TILE_SIZE) # create masks for subsequent loads act_num_segments = cdiv_fn(seq_len, tiles_per_segment * TILE_SIZE) segm_mask = tl.arange(0, NUM_SEGMENTS_PER_SEQ) < tl.full( [NUM_SEGMENTS_PER_SEQ], act_num_segments, dtype=tl.int32 ) dim_mask = tl.where(tl.arange(0, HEAD_SIZE_PADDED) < HEAD_SIZE, 1, 0).to(tl.int1) # load segment maxima segm_offset = ( query_token_idx.to(tl.int64) * (num_query_heads * NUM_SEGMENTS_PER_SEQ) + query_head_idx * NUM_SEGMENTS_PER_SEQ + tl.arange(0, NUM_SEGMENTS_PER_SEQ) ) segm_max = tl.load(segm_max_ptr + segm_offset, mask=segm_mask, other=float("-inf")) overall_max = tl.max(segm_max) # load and rescale segment exp sums segm_expsum = tl.load(segm_expsum_ptr + segm_offset, mask=segm_mask, other=0.0) segm_expsum = segm_expsum * tl.exp(segm_max - overall_max) overall_expsum = tl.sum(segm_expsum) # load, rescale, and add segment attention outputs segm_output_offset = ( query_token_idx.to(tl.int64) * (num_query_heads * NUM_SEGMENTS_PER_SEQ * HEAD_SIZE_PADDED) + query_head_idx * (NUM_SEGMENTS_PER_SEQ * HEAD_SIZE_PADDED) + tl.arange(0, NUM_SEGMENTS_PER_SEQ)[:, None] * HEAD_SIZE_PADDED + tl.arange(0, HEAD_SIZE_PADDED)[None, :] ) segm_output = tl.load( segm_output_ptr + segm_output_offset, mask=segm_mask[:, None] & dim_mask[None, :], other=0.0, ) segm_output *= tl.exp(segm_max - overall_max)[:, None] acc_sum = tl.sum(segm_output, axis=0) # safely divide by overall_expsum, returning 0.0 if overall_expsum is 0 acc = tl.where(overall_expsum == 0.0, 0.0, acc_sum / overall_expsum) if USE_FP8: acc = acc * tl.load(out_scale_inv) acc = tl.clamp(acc, FP8_MIN, FP8_MAX) # write result output_offset = ( query_token_idx * output_stride_0 + query_head_idx * output_stride_1 + tl.arange(0, HEAD_SIZE_PADDED) ) tl.store(output_ptr + output_offset, acc, mask=dim_mask) def _is_gemma3_attention(head_size: int, sliding_window: int) -> bool: """Detect Gemma3 models via unique (head_size, sliding_window) signature. Gemma3 models are the only ones using sliding_window=1024 with head_size 128 (27B) or 256 (1B, 4B, 12B). Other SWA models use different window sizes (Mistral=4096, Phi-3=2047). """ return sliding_window == 1024 and head_size in (128, 256) def _get_tile_size( head_size: int, sliding_window: int, element_size: int, is_prefill: bool, ) -> int: """Select tile size with Gemma3-specific optimization. For Gemma3, use 32 for both prefill and decode to better utilize the larger head dimension (128/256). For other models, use the default vLLM behavior. """ if _is_gemma3_attention(head_size, sliding_window): # Gemma3: use 32 for decode (default is 16) return 32 # Default behavior if is_prefill: return 32 return 16 if element_size >= 2 else 32 def unified_attention( q, k, v, out, cu_seqlens_q, max_seqlen_q, seqused_k, max_seqlen_k, softmax_scale, causal, window_size, block_table, softcap, q_descale, k_descale, v_descale, seq_threshold_3D=None, num_par_softmax_segments=None, softmax_segm_output=None, softmax_segm_max=None, softmax_segm_expsum=None, alibi_slopes=None, output_scale=None, qq_bias=None, # Optional tensor for sinks sinks=None, # Optional tensor for prefix lengths (PrefixLM support) mm_prefix_range=None, ): assert causal, "Only causal attention is supported" assert q_descale is None, "Q scales not supported" if sinks is not None: assert sinks.shape[0] == q.shape[1], "Sinks must be num_query_heads size" use_mm_prefix = False max_mm_ranges = 0 if mm_prefix_range is not None: if mm_prefix_range.ndim == 3: use_mm_prefix = True max_mm_ranges = mm_prefix_range.shape[1] else: raise ValueError( f"Unsupported mm_prefix_range shape: {mm_prefix_range.shape}" ) use_alibi_slopes = alibi_slopes is not None use_qq_bias = qq_bias is not None block_size = v.shape[1] num_seqs = len(seqused_k) num_query_heads = q.shape[1] num_kv_heads = k.shape[2] num_queries_per_kv = num_query_heads // num_kv_heads head_size = q.shape[2] BLOCK_M = ( 16 if num_queries_per_kv <= 16 else triton.next_power_of_2(num_queries_per_kv) ) BLOCK_Q = BLOCK_M // num_queries_per_kv # Ideally we would launch with kernel with: # \sum_i[ceil(query_len[i] / BLOCK_Q)] blocks. # However, it is slow to realize the query_lens on cpu. # Instead we use upper-bound: # \sum_i[ceil(query_len[i] / BLOCK_Q)] # <= \sum_i[floor(query_len[i] / BLOCK_Q) + 1] # = \sum_i[floor(query_len[i] / BLOCK_Q)] + num_seqs # <= floor(\sum_i(query_len[i]) / BLOCK_Q) + num_seqs # = floor(q.shape[0] / BLOCK_Q) + num_seqs total_num_q_blocks = q.shape[0] // BLOCK_Q + num_seqs # Tile sizes for prefill and decode. Gemma3 models use optimized values. # Note: tile size must be at least 32 for fp8 (element_size == 1). sliding_window_val = 1 + window_size[0] if window_size[0] >= 0 else 0 TILE_SIZE_PREFILL = _get_tile_size( head_size, sliding_window_val, q.element_size(), is_prefill=True, ) TILE_SIZE_DECODE = _get_tile_size( head_size, sliding_window_val, q.element_size(), is_prefill=False, ) # Launch the 2D kernel if # 1. No intermediate tiled softmax buffers for the 3D kernel have been allocated, or # 2. The batch includes at least one prefill request, or # 3. The number of sequences exceeds the configured threshold if ( seq_threshold_3D is None or num_par_softmax_segments is None or softmax_segm_output is None or softmax_segm_max is None or softmax_segm_expsum is None or max_seqlen_q > 1 or num_seqs > seq_threshold_3D ): kernel_unified_attention_2d[ (
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/triton_decode_attention.py
vllm/attention/ops/triton_decode_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Adapted from # https://github.com/sgl-project/sglang/blob/9f635ea50de920aa507f486daafba26a5b837574/python/sglang/srt/layers/attention/triton_ops/decode_attention.py # which was originally adapted from # https://github.com/ModelTC/lightllm/blob/96353e868a840db4d103138caf15ed9dbea8c186/lightllm/models/deepseek2/triton_kernel/gqa_flash_decoding_stage1.py # https://github.com/ModelTC/lightllm/blob/96353e868a840db4d103138caf15ed9dbea8c186/lightllm/models/deepseek2/triton_kernel/gqa_flash_decoding_stage2.py # Changes: # - Add support for page size >= 1. # Copyright 2025 vLLM Team # Copyright 2023-2024 SGLang Team # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """ Memory-efficient attention for decoding. It supports page size >= 1. """ import logging from packaging import version from vllm.platforms import current_platform from vllm.triton_utils import tl, triton is_hip_ = current_platform.is_rocm() logger = logging.getLogger(__name__) # Only print the following warnings when triton version < 3.2.0. # The issue won't affect performance or accuracy. if version.parse(triton.__version__) < version.parse("3.2.0"): logger.warning( "The following error message 'operation scheduled before its operands' " "can be ignored." ) @triton.jit def tanh(x): # Tanh is just a scaled sigmoid return 2 * tl.sigmoid(2 * x) - 1 @triton.jit def _fwd_kernel_stage1( Q, K_Buffer, V_Buffer, sm_scale, Req_to_tokens, B_Seqlen, Att_Out, stride_req_to_tokens_b, stride_qbs, stride_qh, stride_buf_kbs, stride_buf_kh, stride_buf_vbs, stride_buf_vh, stride_mid_ob, stride_mid_oh, stride_mid_os, kv_group_num: tl.constexpr, BLOCK_DMODEL: tl.constexpr, BLOCK_DV: tl.constexpr, BLOCK_N: tl.constexpr, NUM_KV_SPLITS: tl.constexpr, PAGE_SIZE: tl.constexpr, logit_cap: tl.constexpr, Lk: tl.constexpr, Lv: tl.constexpr, ): cur_batch = tl.program_id(0) cur_head = tl.program_id(1) split_kv_id = tl.program_id(2) cur_kv_head = cur_head // kv_group_num offs_d = tl.arange(0, BLOCK_DMODEL) offs_dv = tl.arange(0, BLOCK_DV) mask_d = offs_d < Lk mask_dv = offs_dv < Lv cur_batch_seq_len = tl.load(B_Seqlen + cur_batch) cur_batch_req_idx = cur_batch off_q = cur_batch * stride_qbs + cur_head * stride_qh + offs_d q = tl.load(Q + off_q, mask=mask_d, other=0.0) kv_len_per_split = tl.cdiv(cur_batch_seq_len, NUM_KV_SPLITS) split_kv_start = kv_len_per_split * split_kv_id split_kv_end = tl.minimum(split_kv_start + kv_len_per_split, cur_batch_seq_len) e_max = -float("inf") e_sum = 0.0 acc = tl.zeros([BLOCK_DV], dtype=tl.float32) if split_kv_end > split_kv_start: for start_n in range(split_kv_start, split_kv_end, BLOCK_N): offs_n = start_n + tl.arange(0, BLOCK_N) kv_page_number = tl.load( Req_to_tokens + stride_req_to_tokens_b * cur_batch_req_idx + offs_n // PAGE_SIZE, mask=offs_n < split_kv_end, other=0, ) kv_loc = kv_page_number * PAGE_SIZE + offs_n % PAGE_SIZE offs_buf_k = ( kv_loc[:, None] * stride_buf_kbs + cur_kv_head * stride_buf_kh + offs_d[None, :] ) k = tl.load( K_Buffer + offs_buf_k, mask=(offs_n[:, None] < split_kv_end) & (mask_d[None, :]), other=0.0, ) qk = tl.sum(q[None, :] * k, 1) qk *= sm_scale if logit_cap > 0: qk = logit_cap * tanh(qk / logit_cap) qk = tl.where(offs_n < split_kv_end, qk, float("-inf")) offs_buf_v = ( kv_loc[:, None] * stride_buf_vbs + cur_kv_head * stride_buf_vh + offs_dv[None, :] ) v = tl.load( V_Buffer + offs_buf_v, mask=(offs_n[:, None] < split_kv_end) & (mask_dv[None, :]), other=0.0, ) n_e_max = tl.maximum(tl.max(qk, 0), e_max) re_scale = tl.exp(e_max - n_e_max) p = tl.exp(qk - n_e_max) acc *= re_scale acc += tl.sum(p[:, None] * v, 0) e_sum = e_sum * re_scale + tl.sum(p, 0) e_max = n_e_max offs_mid_o = ( cur_batch * stride_mid_ob + cur_head * stride_mid_oh + split_kv_id * stride_mid_os + offs_dv ) tl.store( Att_Out + offs_mid_o, acc / e_sum, mask=(mask_dv), ) offs_mid_o_1 = ( cur_batch * stride_mid_ob + cur_head * stride_mid_oh + split_kv_id * stride_mid_os + Lv ) tl.store( Att_Out + offs_mid_o_1, e_max + tl.log(e_sum), ) def _decode_att_m_fwd( q, k_buffer, v_buffer, att_out, Req_to_tokens, B_Seqlen, num_kv_splits, sm_scale, page_size, logit_cap, ): BLOCK = 64 if not is_hip_ else 8 NUM_KV_SPLITS = num_kv_splits Lk = k_buffer.shape[-1] Lv = v_buffer.shape[-1] batch, head_num = q.shape[0], q.shape[1] grid = (batch, head_num, NUM_KV_SPLITS) kv_group_num = q.shape[1] // k_buffer.shape[-2] num_warps = 4 if kv_group_num != 1: num_warps = 1 if is_hip_ else 2 BLOCK_DMODEL = triton.next_power_of_2(Lk) BLOCK_DV = triton.next_power_of_2(Lv) _fwd_kernel_stage1[grid]( q, k_buffer, v_buffer, sm_scale, Req_to_tokens, B_Seqlen, att_out, Req_to_tokens.stride(0), q.stride(0), q.stride(1), k_buffer.stride(-3), # Assume (..., PAGE_SIZE, NUM_HEADS, HEAD_DIM) k_buffer.stride(-2), # Assume (..., PAGE_SIZE, NUM_HEADS, HEAD_DIM) v_buffer.stride(-3), # Assume (..., PAGE_SIZE, NUM_HEADS, HEAD_DIM) v_buffer.stride(-2), # Assume (..., PAGE_SIZE, NUM_HEADS, HEAD_DIM) att_out.stride(0), att_out.stride(1), att_out.stride(2), kv_group_num=kv_group_num, BLOCK_DMODEL=BLOCK_DMODEL, BLOCK_DV=BLOCK_DV, BLOCK_N=BLOCK, NUM_KV_SPLITS=NUM_KV_SPLITS, PAGE_SIZE=page_size, logit_cap=logit_cap, num_warps=num_warps, num_stages=2, Lk=Lk, Lv=Lv, ) @triton.jit def _fwd_grouped_kernel_stage1( Q, K_Buffer, V_Buffer, sm_scale, Req_to_tokens, B_Seqlen, Att_Out, stride_req_to_tokens_b, stride_qbs, stride_qh, stride_buf_kbs, stride_buf_kh, stride_buf_vbs, stride_buf_vh, stride_mid_ob, stride_mid_oh, stride_mid_os, kv_group_num: tl.constexpr, q_head_num: tl.constexpr, BLOCK_DMODEL: tl.constexpr, BLOCK_DPE: tl.constexpr, BLOCK_DV: tl.constexpr, BLOCK_N: tl.constexpr, BLOCK_H: tl.constexpr, NUM_KV_SPLITS: tl.constexpr, PAGE_SIZE: tl.constexpr, logit_cap: tl.constexpr, Lk: tl.constexpr, Lv: tl.constexpr, ): cur_batch = tl.program_id(0) cur_head_id = tl.program_id(1) cur_kv_head = cur_head_id // tl.cdiv(kv_group_num, BLOCK_H) split_kv_id = tl.program_id(2) if kv_group_num > BLOCK_H: VALID_BLOCK_H: tl.constexpr = BLOCK_H else: VALID_BLOCK_H: tl.constexpr = kv_group_num cur_head = cur_head_id * VALID_BLOCK_H + tl.arange(0, BLOCK_H) mask_h = cur_head < (cur_head_id + 1) * VALID_BLOCK_H mask_h = mask_h & (cur_head < q_head_num) offs_d = tl.arange(0, BLOCK_DMODEL) offs_dv = tl.arange(0, BLOCK_DV) mask_d = offs_d < Lk mask_dv = offs_dv < Lv cur_batch_seq_len = tl.load(B_Seqlen + cur_batch) cur_batch_req_idx = cur_batch offs_q = cur_batch * stride_qbs + cur_head[:, None] * stride_qh + offs_d[None, :] q = tl.load(Q + offs_q, mask=(mask_h[:, None]) & (mask_d[None, :]), other=0.0) if BLOCK_DPE > 0: offs_dpe = BLOCK_DMODEL + tl.arange(0, BLOCK_DPE) mask_dpe = offs_dpe < Lk off_qpe = ( cur_batch * stride_qbs + cur_head[:, None] * stride_qh + offs_dpe[None, :] ) qpe = tl.load( Q + off_qpe, mask=(mask_h[:, None]) & (mask_dpe[None, :]), other=0.0 ) kv_len_per_split = tl.cdiv(cur_batch_seq_len, NUM_KV_SPLITS) split_kv_start = kv_len_per_split * split_kv_id split_kv_end = tl.minimum(split_kv_start + kv_len_per_split, cur_batch_seq_len) e_max = tl.zeros([BLOCK_H], dtype=tl.float32) - float("inf") e_sum = tl.zeros([BLOCK_H], dtype=tl.float32) acc = tl.zeros([BLOCK_H, BLOCK_DV], dtype=tl.float32) if split_kv_end > split_kv_start: for start_n in range(split_kv_start, split_kv_end, BLOCK_N): offs_n = start_n + tl.arange(0, BLOCK_N) kv_page_number = tl.load( Req_to_tokens + stride_req_to_tokens_b * cur_batch_req_idx + offs_n // PAGE_SIZE, mask=offs_n < split_kv_end, other=0, ) kv_loc = kv_page_number * PAGE_SIZE + offs_n % PAGE_SIZE offs_buf_k = ( kv_loc[None, :] * stride_buf_kbs + cur_kv_head * stride_buf_kh + offs_d[:, None] ) k = tl.load( K_Buffer + offs_buf_k, mask=(offs_n[None, :] < split_kv_end) & (mask_d[:, None]), other=0.0, ) qk = tl.dot(q, k.to(q.dtype)) if BLOCK_DPE > 0: offs_buf_kpe = ( kv_loc[None, :] * stride_buf_kbs + cur_kv_head * stride_buf_kh + offs_dpe[:, None] ) kpe = tl.load( K_Buffer + offs_buf_kpe, mask=(offs_n[None, :] < split_kv_end) & (mask_dpe[:, None]), other=0.0, ) qk += tl.dot(qpe, kpe.to(qpe.dtype)) qk *= sm_scale if logit_cap > 0: qk = logit_cap * tanh(qk / logit_cap) qk = tl.where( mask_h[:, None] & (offs_n[None, :] < split_kv_end), qk, float("-inf") ) offs_buf_v = ( kv_loc[:, None] * stride_buf_vbs + cur_kv_head * stride_buf_vh + offs_dv[None, :] ) v = tl.load( V_Buffer + offs_buf_v, mask=(offs_n[:, None] < split_kv_end) & (mask_dv[None, :]), other=0.0, ) n_e_max = tl.maximum(tl.max(qk, 1), e_max) re_scale = tl.exp(e_max - n_e_max) p = tl.exp(qk - n_e_max[:, None]) acc *= re_scale[:, None] acc += tl.dot(p.to(v.dtype), v) e_sum = e_sum * re_scale + tl.sum(p, 1) e_max = n_e_max offs_mid_o = ( cur_batch * stride_mid_ob + cur_head[:, None] * stride_mid_oh + split_kv_id * stride_mid_os + offs_dv[None, :] ) tl.store( Att_Out + offs_mid_o, acc / e_sum[:, None], mask=(mask_h[:, None]) & (mask_dv[None, :]), ) offs_mid_o_1 = ( cur_batch * stride_mid_ob + cur_head * stride_mid_oh + split_kv_id * stride_mid_os + Lv ) tl.store( Att_Out + offs_mid_o_1, e_max + tl.log(e_sum), mask=mask_h, ) def _decode_grouped_att_m_fwd( q, k_buffer, v_buffer, att_out, Req_to_tokens, B_Seqlen, num_kv_splits, sm_scale, page_size, logit_cap, ): BLOCK = 32 Lk = k_buffer.shape[-1] Lv = v_buffer.shape[-1] # [TODO] work around shmem limit on MI3xx if is_hip_ and Lk >= 576: BLOCK = 16 if Lk == 576: BLOCK_DMODEL = 512 BLOCK_DPE = 64 elif Lk == 288: BLOCK_DMODEL = 256 BLOCK_DPE = 32 else: BLOCK_DMODEL = triton.next_power_of_2(Lk) BLOCK_DPE = 0 BLOCK_DV = triton.next_power_of_2(Lv) batch, head_num = q.shape[0], q.shape[1] kv_group_num = q.shape[1] // k_buffer.shape[-2] BLOCK_H = 16 NUM_KV_SPLITS = num_kv_splits grid = ( batch, triton.cdiv(head_num, min(BLOCK_H, kv_group_num)), NUM_KV_SPLITS, ) extra_kargs = {} num_stages = 2 if is_hip_: # https://rocm.docs.amd.com/en/latest/how-to/rocm-for-ai/inference-optimization/workload.html#mi300x-triton-kernel-performance-optimization # https://github.com/triton-lang/triton/blob/main/third_party/amd/backend/compiler.py extra_kargs = {"waves_per_eu": 1, "matrix_instr_nonkdim": 16, "kpack": 2} num_stages = 1 _fwd_grouped_kernel_stage1[grid]( q, k_buffer, v_buffer, sm_scale, Req_to_tokens, B_Seqlen, att_out, Req_to_tokens.stride(0), q.stride(0), q.stride(1), k_buffer.stride(-3), # Assume (..., PAGE_SIZE, NUM_HEADS, HEAD_DIM) k_buffer.stride(-2), # Assume (..., PAGE_SIZE, NUM_HEADS, HEAD_DIM) v_buffer.stride(-3), # Assume (..., PAGE_SIZE, NUM_HEADS, HEAD_DIM) v_buffer.stride(-2), # Assume (..., PAGE_SIZE, NUM_HEADS, HEAD_DIM) att_out.stride(0), att_out.stride(1), att_out.stride(2), kv_group_num=kv_group_num, q_head_num=head_num, BLOCK_DMODEL=BLOCK_DMODEL, BLOCK_DPE=BLOCK_DPE, BLOCK_DV=BLOCK_DV, BLOCK_N=BLOCK, BLOCK_H=BLOCK_H, NUM_KV_SPLITS=NUM_KV_SPLITS, PAGE_SIZE=page_size, logit_cap=logit_cap, num_warps=4, num_stages=num_stages, Lk=Lk, Lv=Lv, **extra_kargs, ) @triton.jit def _fwd_kernel_stage2( Mid_O, o, lse, B_Seqlen, stride_mid_ob, stride_mid_oh, stride_mid_os, stride_obs, stride_oh, stride_lse_bs, NUM_KV_SPLITS: tl.constexpr, BLOCK_DV: tl.constexpr, Lv: tl.constexpr, ): cur_batch = tl.program_id(0) cur_head = tl.program_id(1) cur_batch_seq_len = tl.load(B_Seqlen + cur_batch) offs_d = tl.arange(0, BLOCK_DV) mask_d = offs_d < Lv e_sum = 0.0 e_max = -float("inf") acc = tl.zeros([BLOCK_DV], dtype=tl.float32) offs_v = cur_batch * stride_mid_ob + cur_head * stride_mid_oh + offs_d offs_logic = cur_batch * stride_mid_ob + cur_head * stride_mid_oh + Lv for split_kv_id in range(0, NUM_KV_SPLITS): kv_len_per_split = tl.cdiv(cur_batch_seq_len, NUM_KV_SPLITS) split_kv_start = kv_len_per_split * split_kv_id split_kv_end = tl.minimum(split_kv_start + kv_len_per_split, cur_batch_seq_len) if split_kv_end > split_kv_start: tv = tl.load( Mid_O + offs_v + split_kv_id * stride_mid_os, mask=mask_d, other=0.0 ) tlogic = tl.load(Mid_O + offs_logic + split_kv_id * stride_mid_os) n_e_max = tl.maximum(tlogic, e_max) old_scale = tl.exp(e_max - n_e_max) acc *= old_scale exp_logic = tl.exp(tlogic - n_e_max) acc += exp_logic * tv e_sum = e_sum * old_scale + exp_logic e_max = n_e_max tl.store( o + cur_batch * stride_obs + cur_head * stride_oh + offs_d, acc / e_sum, mask=mask_d, ) lse_val = e_max + tl.log(e_sum) tl.store( lse + cur_batch * stride_lse_bs + cur_head, lse_val, ) def _decode_softmax_reducev_fwd( logits, q, o, lse, v_buffer, b_seq_len, num_kv_splits, ): batch, head_num = q.shape[0], q.shape[1] Lv = v_buffer.shape[-1] BLOCK_DV = triton.next_power_of_2(Lv) NUM_KV_SPLITS = num_kv_splits extra_kargs = {} if is_hip_: # https://rocm.docs.amd.com/en/docs-6.2.0/how-to/llm-fine-tuning-optimization/optimizing-triton-kernel.html # https://github.com/triton-lang/triton/blob/main/third_party/amd/backend/compiler.py extra_kargs = {"waves_per_eu": 4, "matrix_instr_nonkdim": 16, "kpack": 2} grid = (batch, head_num) _fwd_kernel_stage2[grid]( logits, o, lse, b_seq_len, logits.stride(0), logits.stride(1), logits.stride(2), o.stride(0), o.stride(1), lse.stride(0), NUM_KV_SPLITS=NUM_KV_SPLITS, BLOCK_DV=BLOCK_DV, Lv=Lv, num_warps=4, num_stages=2, **extra_kargs, ) def decode_attention_fwd_normal( q, k_buffer, v_buffer, o, lse, req_to_token, b_seq_len, attn_logits, num_kv_splits, sm_scale, page_size, logit_cap=0.0, ): _decode_att_m_fwd( q, k_buffer, v_buffer, attn_logits, req_to_token, b_seq_len, num_kv_splits, sm_scale, page_size, logit_cap, ) _decode_softmax_reducev_fwd( attn_logits, q, o, lse, v_buffer, b_seq_len, num_kv_splits ) def decode_attention_fwd_grouped( q, k_buffer, v_buffer, o, lse, req_to_token, b_seq_len, attn_logits, num_kv_splits, sm_scale, page_size, logit_cap=0.0, ): _decode_grouped_att_m_fwd( q, k_buffer, v_buffer, attn_logits, req_to_token, b_seq_len, num_kv_splits, sm_scale, page_size, logit_cap, ) _decode_softmax_reducev_fwd( attn_logits, q, o, lse, v_buffer, b_seq_len, num_kv_splits ) def decode_attention_fwd( q, k_buffer, v_buffer, o, lse, req_to_token, b_seq_len, attn_logits, num_kv_splits, sm_scale, page_size=1, logit_cap=0.0, ): assert num_kv_splits == attn_logits.shape[2] kv_group_num = q.shape[1] // v_buffer.shape[-2] if kv_group_num == 1: # MHA decode_attention_fwd_normal( q, k_buffer, v_buffer, o, lse, req_to_token, b_seq_len, attn_logits, num_kv_splits, sm_scale, page_size, logit_cap, ) else: # GQA/MQA/MLA decode_attention_fwd_grouped( q, k_buffer, v_buffer, o, lse, req_to_token, b_seq_len, attn_logits, num_kv_splits, sm_scale, page_size, logit_cap, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/paged_attn.py
vllm/attention/ops/paged_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.platforms import current_platform if current_platform.is_cuda_alike(): from vllm import _custom_ops as ops elif current_platform.is_xpu(): from vllm._ipex_ops import ipex_ops as ops class PagedAttention: @staticmethod def split_kv_cache( kv_cache: torch.Tensor, num_kv_heads: int, head_size: int, ) -> tuple[torch.Tensor, torch.Tensor]: x = 16 // kv_cache.element_size() num_blocks = kv_cache.shape[1] key_cache = kv_cache[0] key_cache = key_cache.view(num_blocks, num_kv_heads, head_size // x, -1, x) value_cache = kv_cache[1] value_cache = value_cache.view(num_blocks, num_kv_heads, head_size, -1) return key_cache, value_cache @staticmethod def write_to_paged_cache( key: torch.Tensor, value: torch.Tensor, key_cache: torch.Tensor, value_cache: torch.Tensor, slot_mapping: torch.Tensor, kv_cache_dtype: str, k_scale: torch.Tensor, v_scale: torch.Tensor, ) -> None: ops.reshape_and_cache( key, value, key_cache, value_cache, slot_mapping.flatten(), kv_cache_dtype, k_scale, v_scale, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/prefix_prefill.py
vllm/attention/ops/prefix_prefill.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # The kernels in this file are adapted from LightLLM's context_attention_fwd: # https://github.com/ModelTC/lightllm/blob/main/lightllm/models/llama/triton_kernel/context_flashattention_nopad.py import torch from vllm.platforms import current_platform from vllm.triton_utils import tl, triton # Static kernels parameters BASE_BLOCK = 128 if current_platform.has_device_capability(80) else 64 NUM_WARPS = 4 if current_platform.is_rocm() else 8 # To check compatibility IS_TURING = current_platform.get_device_capability() == (7, 5) float8_info = torch.finfo(current_platform.fp8_dtype()) # Here's an example autotuner config for this kernel. This config does provide # a performance improvement, but dramatically increases first call latency in # triton 3.2. Because of this tradeoff, it's currently commented out. # @triton.autotune( # configs=[ # triton.Config({'BLOCK_M': 128, 'BLOCK_N': 64, \ # "num_unroll_cache": 4, \ # "num_unroll_request": 1 } | \ # ({"kpack": 2, "waves_per_eu": 2} \ # if current_platform.is_rocm() else {}), \ # num_warps=4, \ # num_stages=1) # ], # key=["BLOCK_SIZE", "MAX_Q_LEN", "MAX_CTX_LEN"] # ) @triton.jit def _fwd_kernel( Q, K, V, K_cache, V_cache, sink_ptr, B_Loc, sm_scale, k_scale, v_scale, out_scale_inv, B_Start_Loc, B_Seqlen, x: tl.constexpr, Out, stride_b_loc_b, stride_b_loc_s, stride_qbs, stride_qh, stride_qd, stride_kbs, stride_kh, stride_kd, stride_vbs, stride_vh, stride_vd, stride_obs, stride_oh, stride_od, stride_k_cache_bs, stride_k_cache_h, stride_k_cache_d, stride_k_cache_bl: tl.constexpr, stride_k_cache_x, stride_v_cache_bs, stride_v_cache_h, stride_v_cache_d, stride_v_cache_bl, num_queries_per_kv: tl.constexpr, IN_PRECISION: tl.constexpr, BLOCK_M: tl.constexpr, BLOCK_DMODEL: tl.constexpr, BLOCK_DMODEL_PADDED: tl.constexpr, BLOCK_SIZE: tl.constexpr, BLOCK_N: tl.constexpr, SLIDING_WINDOW: tl.constexpr, num_unroll_cache: tl.constexpr, num_unroll_request: tl.constexpr, SKIP_DECODE: tl.constexpr, USE_SINKS: tl.constexpr, USE_FP8: tl.constexpr, MAX_Q_LEN: tl.constexpr = 0, MAX_CTX_LEN: tl.constexpr = 0, FP8_MIN: tl.constexpr = float8_info.min, FP8_MAX: tl.constexpr = float8_info.max, ): cur_batch = tl.program_id(0) cur_head = tl.program_id(1) start_m = tl.program_id(2) cur_kv_head = cur_head // num_queries_per_kv cur_batch_seq_len = tl.load(B_Seqlen + cur_batch) cur_batch_in_all_start_index = tl.load(B_Start_Loc + cur_batch) cur_batch_in_all_stop_index = tl.load(B_Start_Loc + cur_batch + 1) cur_batch_query_len = cur_batch_in_all_stop_index - cur_batch_in_all_start_index cur_batch_ctx_len = cur_batch_seq_len - cur_batch_query_len if SKIP_DECODE and cur_batch_query_len == 1: return # start position inside of the query # generally, N goes over kv, while M goes over query_len block_start_loc = BLOCK_M * start_m # initialize offsets # [BLOCK_SIZE]; starts at 0 offs_bs_n = tl.arange(0, BLOCK_SIZE) # [N]; starts at 0 offs_n = tl.arange(0, BLOCK_N) # [D]; starts at 0 offs_d = tl.arange(0, BLOCK_DMODEL_PADDED) # [M]; starts at current position in query offs_m = start_m * BLOCK_M + tl.arange(0, BLOCK_M) # [M,D] off_q = ( (cur_batch_in_all_start_index + offs_m[:, None]) * stride_qbs + cur_head * stride_qh + offs_d[None, :] * stride_qd ) dim_mask = tl.where(tl.arange(0, BLOCK_DMODEL_PADDED) < BLOCK_DMODEL, 1, 0).to( tl.int1 ) # [D] q = tl.load( Q + off_q, mask=dim_mask[None, :] & (offs_m[:, None] < cur_batch_query_len), other=0.0, ) # [M,D] # initialize pointer to m and l if not USE_SINKS: m_i = tl.full([BLOCK_M], float("-inf"), dtype=tl.float32) else: m_i = tl.load( sink_ptr + tl.full([BLOCK_M], cur_head, dtype=tl.int64), mask=(offs_m < cur_batch_query_len), other=float("-inf"), ).to(dtype=tl.float32) l_i = tl.full([BLOCK_M], 1.0, dtype=tl.float32) acc = tl.zeros([BLOCK_M, BLOCK_DMODEL_PADDED], dtype=tl.float32) # [M,D] # compute query against context (no causal mask here) for start_n in tl.range( 0, cur_batch_ctx_len, BLOCK_SIZE, loop_unroll_factor=num_unroll_cache ): start_n = tl.multiple_of(start_n, BLOCK_SIZE) # -- compute qk ---- bn = tl.load( B_Loc + cur_batch * stride_b_loc_b + (start_n // BLOCK_SIZE) * stride_b_loc_s ).to(tl.int64) # [D,BLOCK_SIZE] off_k = ( bn[None, :] * stride_k_cache_bs + cur_kv_head * stride_k_cache_h + (offs_d[:, None] // x) * stride_k_cache_d + ((start_n + offs_bs_n[None, :]) % BLOCK_SIZE) * stride_k_cache_bl + (offs_d[:, None] % x) * stride_k_cache_x ) # [BLOCK_SIZE,D] off_v = ( bn[:, None] * stride_v_cache_bs + cur_kv_head * stride_v_cache_h + offs_d[None, :] * stride_v_cache_d + offs_bs_n[:, None] * stride_v_cache_bl ) if ( start_n + BLOCK_SIZE > cur_batch_ctx_len or BLOCK_DMODEL != BLOCK_DMODEL_PADDED ): k_load = tl.load( K_cache + off_k, mask=dim_mask[:, None] & ((start_n + offs_bs_n[None, :]) < cur_batch_ctx_len), other=0.0, ) # [D,N] else: k_load = tl.load(K_cache + off_k) if k_load.dtype.is_fp8(): k = (k_load.to(tl.float32) * tl.load(k_scale)).to(q.dtype) else: k = k_load qk = tl.zeros([BLOCK_M, BLOCK_SIZE], dtype=tl.float32) # [M,N] qk = tl.dot(q, k, acc=qk, input_precision=IN_PRECISION) qk = tl.where( (start_n + offs_bs_n[None, :]) < cur_batch_ctx_len, qk, float("-inf") ) qk *= sm_scale if SLIDING_WINDOW > 0: # (cur_batch_ctx_len + offs_m[:, None]) are the positions of # Q entries in sequence # (start_n + offs_bs_n[None, :]) are the positions of # KV entries in sequence # So the condition makes sure each entry in Q only attends # to KV entries not more than SLIDING_WINDOW away. # # We can't use -inf here, because the # sliding window may lead to the entire row being masked. # This then makes m_ij contain -inf, which causes NaNs in # exp(). qk = tl.where( (cur_batch_ctx_len + offs_m[:, None]) - (start_n + offs_bs_n[None, :]) < SLIDING_WINDOW, qk, -10000, ) # compute running maximum m_ij = tl.maximum(m_i, tl.max(qk, axis=1)) p = tl.exp(qk - m_ij[:, None]) l_ij = tl.sum(p, axis=1) alpha = tl.exp(m_i - m_ij) acc = acc * alpha[:, None] # update acc if ( start_n + BLOCK_SIZE > cur_batch_ctx_len or BLOCK_DMODEL != BLOCK_DMODEL_PADDED ): v_load = tl.load( V_cache + off_v, mask=dim_mask[None, :] & ((start_n + offs_bs_n[:, None]) < cur_batch_ctx_len), other=0.0, ) # [N,D] else: v_load = tl.load(V_cache + off_v) if v_load.dtype.is_fp8(): v = (v_load.to(tl.float32) * tl.load(v_scale)).to(q.dtype) else: v = v_load p = p.to(v.dtype) acc = tl.dot(p, v, acc=acc, input_precision=IN_PRECISION) # # update m_i and l_i l_i = l_i * alpha + l_ij m_i = m_ij off_k = ( offs_n[None, :] * stride_kbs + cur_kv_head * stride_kh + offs_d[:, None] * stride_kd ) off_v = ( offs_n[:, None] * stride_vbs + cur_kv_head * stride_vh + offs_d[None, :] * stride_vd ) k_ptrs = K + off_k v_ptrs = V + off_v # block_mask is 0 when we're already past the current query length block_mask = tl.where(block_start_loc < cur_batch_query_len, 1, 0) # compute query against itself (with causal mask) for start_n in tl.range( 0, block_mask * (start_m + 1) * BLOCK_M, BLOCK_N, loop_unroll_factor=num_unroll_request, ): start_n = tl.multiple_of(start_n, BLOCK_N) # -- compute qk ---- k = tl.load( k_ptrs + (cur_batch_in_all_start_index + start_n) * stride_kbs, mask=dim_mask[:, None] & ((start_n + offs_n[None, :]) < cur_batch_query_len), other=0.0, ) qk = tl.zeros([BLOCK_M, BLOCK_N], dtype=tl.float32) qk = tl.dot(q, k, acc=qk, input_precision=IN_PRECISION) qk *= sm_scale # apply causal mask qk = tl.where(offs_m[:, None] >= (start_n + offs_n[None, :]), qk, float("-inf")) if SLIDING_WINDOW > 0: qk = tl.where( offs_m[:, None] - (start_n + offs_n[None, :]) < SLIDING_WINDOW, qk, -10000, ) # compute running maximum m_ij = tl.maximum(m_i, tl.max(qk, axis=1)) p = tl.exp(qk - m_ij[:, None]) l_ij = tl.sum(p, axis=1) alpha = tl.exp(m_i - m_ij) acc = acc * alpha[:, None] # update acc v = tl.load( v_ptrs + (cur_batch_in_all_start_index + start_n) * stride_vbs, mask=dim_mask[None, :] & ((start_n + offs_n[:, None]) < cur_batch_query_len), other=0.0, ) p = p.to(v.dtype) acc = tl.dot(p, v, acc=acc, input_precision=IN_PRECISION) # update m_i and l_i l_i = l_i * alpha + l_ij m_i = m_ij acc = acc / l_i[:, None] # initialize pointers to output off_o = ( (cur_batch_in_all_start_index + offs_m[:, None]) * stride_obs + cur_head * stride_oh + offs_d[None, :] * stride_od ) out_ptrs = Out + off_o if USE_FP8: acc = acc * tl.load(out_scale_inv) acc = tl.clamp(acc, FP8_MIN, FP8_MAX) tl.store( out_ptrs, acc, mask=dim_mask[None, :] & (offs_m[:, None] < cur_batch_query_len) ) return @triton.jit def _fwd_kernel_alibi( Q, K, V, K_cache, V_cache, B_Loc, sm_scale, k_scale, v_scale, B_Start_Loc, B_Seqlen, Alibi_slopes, block_size, x, Out, stride_b_loc_b, stride_b_loc_s, stride_qbs, stride_qh, stride_qd, stride_kbs, stride_kh, stride_kd, stride_vbs, stride_vh, stride_vd, stride_obs, stride_oh, stride_od, stride_k_cache_bs, stride_k_cache_h, stride_k_cache_d, stride_k_cache_bl, stride_k_cache_x, stride_v_cache_bs, stride_v_cache_h, stride_v_cache_d, stride_v_cache_bl, num_queries_per_kv: int, IN_PRECISION: tl.constexpr, BLOCK_M: tl.constexpr, BLOCK_DMODEL: tl.constexpr, # head size BLOCK_DMODEL_PADDED: tl.constexpr, # head size padded to a power of 2 BLOCK_N: tl.constexpr, SKIP_DECODE: tl.constexpr, ): # attn_bias[] cur_batch = tl.program_id(0) cur_head = tl.program_id(1) start_m = tl.program_id(2) cur_kv_head = cur_head // num_queries_per_kv # cur_batch_seq_len: the length of prompts # cur_batch_ctx_len: the length of prefix # cur_batch_in_all_start_index: the start id of the dim=0 cur_batch_seq_len = tl.load(B_Seqlen + cur_batch) cur_batch_in_all_start_index = tl.load(B_Start_Loc + cur_batch) cur_batch_in_all_stop_index = tl.load(B_Start_Loc + cur_batch + 1) cur_batch_query_len = cur_batch_in_all_stop_index - cur_batch_in_all_start_index cur_batch_ctx_len = cur_batch_seq_len - cur_batch_query_len if SKIP_DECODE and cur_batch_query_len == 1: return block_start_loc = BLOCK_M * start_m # initialize offsets offs_n = tl.arange(0, BLOCK_N) offs_d = tl.arange(0, BLOCK_DMODEL_PADDED) offs_m = start_m * BLOCK_M + tl.arange(0, BLOCK_M) off_q = ( (cur_batch_in_all_start_index + offs_m[:, None]) * stride_qbs + cur_head * stride_qh + offs_d[None, :] * stride_qd ) dim_mask = tl.where(tl.arange(0, BLOCK_DMODEL_PADDED) < BLOCK_DMODEL, 1, 0).to( tl.int1 ) q = tl.load( Q + off_q, mask=dim_mask[None, :] & (offs_m[:, None] < cur_batch_seq_len - cur_batch_ctx_len), other=0.0, ) # # initialize pointer to m and l m_i = tl.zeros([BLOCK_M], dtype=tl.float32) - float("inf") l_i = tl.zeros([BLOCK_M], dtype=tl.float32) acc = tl.zeros([BLOCK_M, BLOCK_DMODEL_PADDED], dtype=tl.float32) alibi_slope = tl.load(Alibi_slopes + cur_head) alibi_start_q = tl.arange(0, BLOCK_M) + block_start_loc + cur_batch_ctx_len alibi_start_k = 0 for start_n in range(0, cur_batch_ctx_len, BLOCK_N): start_n = tl.multiple_of(start_n, BLOCK_N) # -- compute qk ---- bn = tl.load( B_Loc + cur_batch * stride_b_loc_b + ((start_n + offs_n) // block_size) * stride_b_loc_s, mask=(start_n + offs_n) < cur_batch_ctx_len, other=0, ).to(tl.int64) off_k = ( bn[None, :] * stride_k_cache_bs + cur_kv_head * stride_k_cache_h + (offs_d[:, None] // x) * stride_k_cache_d + ((start_n + offs_n[None, :]) % block_size) * stride_k_cache_bl + (offs_d[:, None] % x) * stride_k_cache_x ) off_v = ( bn[:, None] * stride_v_cache_bs + cur_kv_head * stride_v_cache_h + offs_d[None, :] * stride_v_cache_d + (start_n + offs_n[:, None]) % block_size * stride_v_cache_bl ) k_load = tl.load( K_cache + off_k, mask=dim_mask[:, None] & ((start_n + offs_n[None, :]) < cur_batch_ctx_len), other=0.0, ) # [D,N] if k_load.dtype.is_fp8(): k = (k_load.to(tl.float32) * tl.load(k_scale)).to(q.dtype) else: k = k_load qk = tl.zeros([BLOCK_M, BLOCK_N], dtype=tl.float32) qk = tl.dot(q, k, acc=qk, input_precision=IN_PRECISION) qk = tl.where( (start_n + offs_n[None, :]) < cur_batch_ctx_len, qk, float("-inf") ) qk *= sm_scale # load alibi alibi = ( tl.arange(0, BLOCK_N)[None, :] + alibi_start_k - alibi_start_q[:, None] ) * alibi_slope alibi = tl.where( (alibi <= 0) & (alibi_start_q[:, None] < cur_batch_seq_len), alibi, float("-inf"), ) qk += alibi alibi_start_k += BLOCK_N # -- compute m_ij, p, l_ij m_ij = tl.max(qk, 1) m_i_new = tl.maximum(m_i, m_ij) p = tl.math.exp(qk - m_i_new[:, None]) l_ij = tl.sum(p, 1) # -- update m_i and l_i alpha = tl.math.exp(m_i - m_i_new) l_i_new = alpha * l_i + l_ij # -- update output accumulator -- # scale p # scale acc acc_scale = alpha # acc_scale = l_i / l_i_new * alpha acc = acc * acc_scale[:, None] # update acc v_load = tl.load( V_cache + off_v, mask=dim_mask[None, :] & ((start_n + offs_n[:, None]) < cur_batch_ctx_len), other=0.0, ) if v_load.dtype.is_fp8(): v = (v_load.to(tl.float32) * tl.load(v_scale)).to(q.dtype) else: v = v_load p = p.to(v.dtype) acc = tl.dot(p, v, acc=acc, input_precision="ieee") # update m_i and l_i l_i = l_i_new m_i = m_i_new off_k = ( offs_n[None, :] * stride_kbs + cur_kv_head * stride_kh + offs_d[:, None] * stride_kd ) off_v = ( offs_n[:, None] * stride_vbs + cur_kv_head * stride_vh + offs_d[None, :] * stride_vd ) k_ptrs = K + off_k v_ptrs = V + off_v block_mask = tl.where(block_start_loc < cur_batch_seq_len - cur_batch_ctx_len, 1, 0) # init alibi alibi_slope = tl.load(Alibi_slopes + cur_head) alibi_start_q = tl.arange(0, BLOCK_M) + block_start_loc + cur_batch_ctx_len alibi_start_k = cur_batch_ctx_len # # init debugger # offset_db_q = tl.arange(0, BLOCK_M) + block_start_loc # offset_db_k = tl.arange(0, BLOCK_N) # calc q[BLOCK_M, BLOCK_MODEL] mul k[prefix_len: , BLOCK_DMODEL] for start_n in range(0, block_mask * (start_m + 1) * BLOCK_M, BLOCK_N): start_n = tl.multiple_of(start_n, BLOCK_N) # -- compute qk ---- k = tl.load( k_ptrs + (cur_batch_in_all_start_index + start_n) * stride_kbs, mask=dim_mask[:, None] & ((start_n + offs_n[None, :]) < cur_batch_seq_len - cur_batch_ctx_len), other=0.0, ) qk = tl.zeros([BLOCK_M, BLOCK_N], dtype=tl.float32) qk = tl.dot(q, k, acc=qk, input_precision="ieee") qk *= sm_scale qk = tl.where(offs_m[:, None] >= (start_n + offs_n[None, :]), qk, float("-inf")) # load alibi alibi = ( tl.arange(0, BLOCK_N)[None, :] + alibi_start_k - alibi_start_q[:, None] ) * alibi_slope alibi = tl.where( (alibi <= 0) & (alibi_start_q[:, None] < cur_batch_seq_len), alibi, float("-inf"), ) qk += alibi alibi_start_k += BLOCK_N # -- compute m_ij, p, l_ij m_ij = tl.max(qk, 1) m_i_new = tl.maximum(m_i, m_ij) p = tl.math.exp(qk - m_i_new[:, None]) l_ij = tl.sum(p, 1) # -- update m_i and l_i alpha = tl.math.exp(m_i - m_i_new) l_i_new = alpha * l_i + l_ij # -- update output accumulator -- # scale p # scale acc acc_scale = alpha # acc_scale = l_i / l_i_new * alpha acc = acc * acc_scale[:, None] # update acc v = tl.load( v_ptrs + (cur_batch_in_all_start_index + start_n) * stride_vbs, mask=dim_mask[None, :] & ((start_n + offs_n[:, None]) < cur_batch_seq_len - cur_batch_ctx_len), other=0.0, ) p = p.to(v.dtype) acc = tl.dot(p, v, acc=acc, input_precision="ieee") # update m_i and l_i l_i = l_i_new m_i = m_i_new acc = acc / l_i[:, None] # initialize pointers to output off_o = ( (cur_batch_in_all_start_index + offs_m[:, None]) * stride_obs + cur_head * stride_oh + offs_d[None, :] * stride_od ) out_ptrs = Out + off_o tl.store( out_ptrs, acc, mask=dim_mask[None, :] & (offs_m[:, None] < cur_batch_seq_len - cur_batch_ctx_len), ) return @torch.inference_mode() def context_attention_fwd( q, k, v, o, kv_cache_dtype: str, k_cache, v_cache, b_loc, b_start_loc, b_seq_len, max_seq_len, max_input_len, k_scale: torch.Tensor, v_scale: torch.Tensor, alibi_slopes=None, sliding_window=None, sm_scale=None, skip_decode=False, fp8_out_scale=None, sinks=None, ): q_dtype_is_f32 = q.dtype is torch.float32 # Turing does have tensor core for float32 multiplication # use ieee as fallback for triton kernels work. There is also # warning on vllm/config.py to inform users this fallback # implementation IN_PRECISION = "ieee" if IS_TURING and q_dtype_is_f32 else None # Conversion of FP8 Tensor from uint8 storage to # appropriate torch.dtype for interpretation by Triton if "fp8" in kv_cache_dtype: assert k_cache.dtype in [torch.uint8, current_platform.fp8_dtype()] assert v_cache.dtype in [torch.uint8, current_platform.fp8_dtype()] if kv_cache_dtype in ("fp8", "fp8_e4m3"): target_dtype = current_platform.fp8_dtype() elif kv_cache_dtype == "fp8_e5m2": target_dtype = torch.float8_e5m2 else: raise ValueError("Unsupported FP8 dtype:", kv_cache_dtype) k_cache = k_cache.view(target_dtype) v_cache = v_cache.view(target_dtype) if ( k_cache.dtype == torch.uint8 or v_cache.dtype == torch.uint8 and kv_cache_dtype == "auto" ): raise ValueError( "kv_cache_dtype='auto' unsupported for\ FP8 KV Cache prefill kernel" ) # shape constraints Lq, Lk, Lv = q.shape[-1], k.shape[-1], v.shape[-1] assert Lq == Lk and Lk == Lv # round up Lk to a power of 2 - this is required for Triton block size Lk_padded = triton.next_power_of_2(Lk) if sm_scale is None: sm_scale = 1.0 / (Lq**0.5) batch, head = b_seq_len.shape[0], q.shape[1] num_queries_per_kv = q.shape[1] // k.shape[1] assert batch + 1 == len(b_start_loc) # 0 means "disable" if sliding_window is None or sliding_window <= 0: sliding_window = 0 if alibi_slopes is not None: assert sinks is None, "Sinks arg is not supported with alibi" assert fp8_out_scale is None, "FP8 output not supported with alibi" # need to reduce num. blocks when using fp32 # due to increased use of GPU shared memory # if q.dtype is torch.float32: BLOCK = BASE_BLOCK // 2 if q_dtype_is_f32 else BASE_BLOCK # batch, head, grid = (batch, head, triton.cdiv(max_input_len, BLOCK)) _fwd_kernel_alibi[grid]( q, k, v, k_cache, v_cache, b_loc, sm_scale, k_scale, v_scale, b_start_loc, b_seq_len, alibi_slopes, v_cache.shape[3], k_cache.shape[4], o, b_loc.stride(0), b_loc.stride(1), q.stride(0), q.stride(1), q.stride(2), k.stride(0), k.stride(1), k.stride(2), v.stride(0), v.stride(1), v.stride(2), o.stride(0), o.stride(1), o.stride(2), k_cache.stride(0), k_cache.stride(1), k_cache.stride(2), k_cache.stride(3), k_cache.stride(4), # [num_blocks, num_kv_heads, head_size/x, block_size, x] v_cache.stride(0), v_cache.stride(1), v_cache.stride(2), v_cache.stride(3), # [num_blocks, num_kv_heads, head_size, block_size] num_queries_per_kv=num_queries_per_kv, IN_PRECISION=IN_PRECISION, BLOCK_M=BLOCK, BLOCK_DMODEL=Lk, BLOCK_DMODEL_PADDED=Lk_padded, BLOCK_N=BLOCK, SKIP_DECODE=skip_decode, num_warps=NUM_WARPS, num_stages=1, ) return max_seq_len = 0 if max_seq_len is None else max_seq_len extra_kargs = {} if current_platform.is_rocm(): extra_kargs = {"kpack": 1, "waves_per_eu": 2} grid = lambda META: (batch, head, triton.cdiv(max_input_len, META["BLOCK_M"])) _fwd_kernel[grid]( q, k, v, k_cache, v_cache, sinks, b_loc, sm_scale, k_scale, v_scale, 1.0 / fp8_out_scale if fp8_out_scale is not None else 1.0, b_start_loc, b_seq_len, k_cache.shape[4], o, b_loc.stride(0), b_loc.stride(1), q.stride(0), q.stride(1), q.stride(2), k.stride(0), k.stride(1), k.stride(2), v.stride(0), v.stride(1), v.stride(2), o.stride(0), o.stride(1), o.stride(2), k_cache.stride(0), k_cache.stride(1), k_cache.stride(2), k_cache.stride(3), k_cache.stride(4), # [num_blocks, num_kv_heads, head_size/x, block_size, x] v_cache.stride(0), v_cache.stride(1), v_cache.stride(2), v_cache.stride(3), # [num_blocks, num_kv_heads, head_size, block_size] BLOCK_SIZE=v_cache.shape[3], num_queries_per_kv=num_queries_per_kv, IN_PRECISION=IN_PRECISION, BLOCK_DMODEL=Lk, BLOCK_DMODEL_PADDED=Lk_padded, SLIDING_WINDOW=sliding_window, SKIP_DECODE=skip_decode, USE_FP8=fp8_out_scale is not None, BLOCK_M=128, BLOCK_N=64, num_unroll_cache=4, num_unroll_request=1, num_warps=4, num_stages=1, USE_SINKS=sinks is not None, **extra_kargs, ) return
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/rocm_aiter_mla_sparse.py
vllm/attention/ops/rocm_aiter_mla_sparse.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import importlib from functools import lru_cache import torch from vllm._aiter_ops import rocm_aiter_ops from vllm.logger import init_logger from vllm.platforms import current_platform logger = init_logger(__name__) # Take from https://github.com/deepseek-ai/DeepGEMM/blob/main/tests/test_attention.py#L84 def fp8_mqa_logits_torch( q: torch.Tensor, kv: tuple[torch.Tensor, torch.Tensor], weights: torch.Tensor, cu_seqlen_ks: torch.Tensor, cu_seqlen_ke: torch.Tensor, ) -> torch.Tensor: """Compute FP8 MQA logits for a single sequence without KV paging. Args: q: Query tensor of shape [M, H, D]. Casted to `torch.float8_e4m3fn` by caller. kv: Tuple `(k_fp8, k_scales)` where `k_fp8` has shape [N, D] with dtype `torch.float8_e4m3fn` and `k_scales` has shape [N] (or [N, 1]) with dtype `torch.float32`. weights: weights of shape [M, H], dtype `torch.float32`. cu_seqlen_ks: Start indices (inclusive) for valid K per query position, shape [M], dtype int32. cu_seqlen_ke: End indices (exclusive) for valid K per query position, shape [M], dtype int32. Returns: Logits tensor of shape [M, N], dtype `torch.float32`. """ kv, scale = kv seq_len_kv = kv.shape[0] k = kv.to(torch.bfloat16) q = q.to(torch.bfloat16) mask_lo = ( torch.arange(0, seq_len_kv, device="cuda")[None, :] >= cu_seqlen_ks[:, None] ) mask_hi = ( torch.arange(0, seq_len_kv, device="cuda")[None, :] < cu_seqlen_ke[:, None] ) mask = mask_lo & mask_hi score = torch.einsum("mhd,nd->hmn", q, k).float() * scale logits = (score.relu() * weights.unsqueeze(-1).transpose(0, 1)).sum(dim=0) logits = logits.masked_fill(~mask, float("-inf")) return logits def rocm_fp8_mqa_logits( q: torch.Tensor, kv: tuple[torch.Tensor, torch.Tensor], weights: torch.Tensor, cu_seqlen_ks: torch.Tensor, cu_seqlen_ke: torch.Tensor, ) -> torch.Tensor: """Compute FP8 MQA logits for a single sequence without KV paging. Args: q: Query tensor of shape [M, H, D]. Casted to `torch.float8_e4m3fn` by caller. kv: Tuple `(k_fp8, k_scales)` where `k_fp8` has shape [N, D] with dtype `torch.float8_e4m3fn` and `k_scales` has shape [N] (or [N, 1]) with dtype `torch.float32`. weights: weights of shape [M, H], dtype `torch.float32`. cu_seqlen_ks: Start indices (inclusive) for valid K per query position, shape [M], dtype int32. cu_seqlen_ke: End indices (exclusive) for valid K per query position, shape [M], dtype int32. Returns: Logits tensor of shape [M, N], dtype `torch.float32`. """ # TODO(ganyi): Temporarily workaround, will remove the module check and reference # path after aiter merge this kernel into main @lru_cache def has_mqa_logits_module(): return importlib.util.find_spec("aiter.ops.triton.fp8_mqa_logits") is not None if rocm_aiter_ops.is_enabled() and has_mqa_logits_module(): from aiter.ops.triton.fp8_mqa_logits import fp8_mqa_logits kv, scale = kv return fp8_mqa_logits(q, kv, scale, weights, cu_seqlen_ks, cu_seqlen_ke) else: return fp8_mqa_logits_torch(q, kv, weights, cu_seqlen_ks, cu_seqlen_ke) # Taken from https://github.com/deepseek-ai/DeepGEMM/blob/main/tests/test_attention.py#L156 def fp8_paged_mqa_logits_torch( q: torch.Tensor, kv_cache: torch.Tensor, weights: torch.Tensor, context_lens: torch.Tensor, block_tables: torch.Tensor, max_model_len: int, ): from vllm.utils.math_utils import cdiv fp8_dtype = current_platform.fp8_dtype() batch_size, next_n, _, dim = q.size() kv_cache, scale = kv_cache[..., :dim], kv_cache[..., dim:] scale = scale.contiguous().view(torch.float) q = q.float() kv_cache = kv_cache.view(fp8_dtype).float() * scale num_block, block_size, _, dim = kv_cache.size() logits = torch.full( [batch_size * next_n, max_model_len], float("-inf"), device=q.device, dtype=torch.float32, ) context_lens = context_lens.tolist() for i in range(batch_size): context_len = context_lens[i] q_offsets = torch.arange(context_len - next_n, context_len, device="cuda") weight_slice = ( weights[i * next_n : (i + 1) * next_n, :].transpose(0, 1).contiguous() ) for block_rk in range(cdiv(context_len, block_size)): block_idx = block_tables[i][block_rk] qx, kx = q[i], kv_cache[block_idx] k_offsets = torch.arange( block_rk * block_size, (block_rk + 1) * block_size, device="cuda" ) mask = (k_offsets[None, :] < context_len) & ( k_offsets[None, :] <= q_offsets[:, None] ) s = torch.where( mask[None, :, :], (qx.transpose(0, 1) @ kx.transpose(0, 1).transpose(1, 2)).to( logits.dtype ), float("-inf"), ) s = torch.relu(s) * weight_slice[..., None] s = s.sum(dim=0) logits[ i * next_n : (i + 1) * next_n, block_rk * block_size : (block_rk + 1) * block_size, ] = torch.where(k_offsets[None, :] <= q_offsets[:, None], s, float("-inf")) return logits def rocm_fp8_paged_mqa_logits( q_fp8: torch.Tensor, kv_cache_fp8: torch.Tensor, weights: torch.Tensor, context_lens: torch.Tensor, block_tables: torch.Tensor, schedule_metadata: torch.Tensor, max_model_len: int, ) -> torch.Tensor: """Compute FP8 MQA logits using paged KV-cache. Args: q_fp8: Query tensor of shape [B, next_n, H, D]. Casted to `torch.float8_e4m3fn` by caller. kv_cache_fp8: Paged KV-cache in packed FP8+scale layout with shape [num_blocks, block_size, 1, D+4], dtype `torch.uint8`. The last 4 bytes per (block,pos) store the `float` dequant scale. weights: Tensor of shape [B * next_n, H], dtype `torch.float32`. context_lens: Tensor of shape [B], dtype int32; effective context length for each batch element. block_tables: Tensor of shape [B, max_blocks], dtype int32; maps logical block indices to physical blocks in the paged cache. schedule_metadata: Returned by `get_paged_mqa_logits_metadata`; used to distribute work across SMs. max_model_len: Maximum sequence length used to size the logits output. Returns: Logits tensor of shape [B * next_n, max_model_len], dtype `torch.float32`. """ if rocm_aiter_ops.is_enabled(): from aiter.ops.triton.pa_mqa_logits import deepgemm_fp8_paged_mqa_logits_stage1 batch_size, next_n, heads, _ = q_fp8.shape out_qk = torch.full( (heads, batch_size * next_n, max_model_len), float("-inf"), device="cuda", dtype=torch.float32, ) deepgemm_fp8_paged_mqa_logits_stage1( q_fp8, kv_cache_fp8, weights, out_qk, context_lens, block_tables, max_model_len, ) return out_qk.sum(dim=0) else: return fp8_paged_mqa_logits_torch( q_fp8, kv_cache_fp8, weights, context_lens, block_tables, max_model_len )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/merge_attn_states.py
vllm/attention/ops/merge_attn_states.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.platforms import current_platform def merge_attn_states( output: torch.Tensor, prefix_output: torch.Tensor, prefix_lse: torch.Tensor, suffix_output: torch.Tensor, suffix_lse: torch.Tensor, output_lse: torch.Tensor | None = None, ) -> None: # NOTE(DefTruth): Currently, custom merge_attn_states CUDA kernel # does not support FP8 dtype, fallback to use Triton kernel. def supported_dtypes(o: torch.Tensor) -> bool: return o.dtype in [torch.float32, torch.half, torch.bfloat16] # NOTE(DefTruth): Currently, custom merge_attn_states CUDA # kernel load/store 128b(16 bytes) per memory issue within # thread. Namely, the headsize(headdim) must be multiple of # pack_size (float32 -> 4, half/bfloat16 -> 8). def supported_headdim(o: torch.Tensor) -> bool: headdim = o.shape[2] # [NUM_TOKENS, NUM_HEADS, HEAD_SIZE] if o.dtype == torch.float32: return headdim % 4 == 0 return headdim % 8 == 0 if ( current_platform.is_cuda() and supported_dtypes(output) and supported_headdim(output) ): from vllm._custom_ops import merge_attn_states return merge_attn_states( output, prefix_output, prefix_lse, suffix_output, suffix_lse, output_lse ) else: from vllm.attention.ops.triton_merge_attn_states import merge_attn_states return merge_attn_states( output, prefix_output, prefix_lse, suffix_output, suffix_lse, output_lse )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/triton_reshape_and_cache_flash.py
vllm/attention/ops/triton_reshape_and_cache_flash.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.platforms import current_platform from vllm.triton_utils import tl, triton @triton.jit def reshape_and_cache_kernel_flash( key_ptr, # [num_tokens, num_heads, head_size] value_ptr, # [num_tokens, num_heads, head_size] key_cache_ptr, # [num_blocks, block_size, num_heads, head_size] value_cache_ptr, # [num_blocks, block_size, num_heads, head_size] slot_mapping_ptr, # [num_tokens] k_scale, # float32 v_scale, # float32 # strides key_stride: tl.int64, value_stride: tl.int64, block_stride: tl.int64, page_stride: tl.int64, num_heads: tl.constexpr, head_size: tl.constexpr, block_size: tl.constexpr, # FP8 flags FP8_KV_CACHE: tl.constexpr, # tune parameters TILE_SIZE: tl.constexpr, ): token_idx = tl.program_id(axis=0) slot_idx = tl.load(slot_mapping_ptr + token_idx).to(tl.int64) if slot_idx < 0: # Padding token that should be ignored. return tile_i = tl.program_id(axis=1) tile_offs = tl.arange(0, TILE_SIZE) tile_pos = tile_i * TILE_SIZE + tile_offs block_idx = slot_idx // block_size block_offset = slot_idx % block_size src_key_idx = token_idx * key_stride src_value_idx = token_idx * value_stride tgt_idx = block_idx * block_stride + block_offset * page_stride # [TILE_SIZE] key_load = tl.load( key_ptr + src_key_idx + tile_pos, mask=tile_pos < (num_heads * head_size) ) if FP8_KV_CACHE: # tl.store will do the correct implicit cast to fp8, # based on the key_cache_ptr.dtype.element_ty key_tile = key_load if key_load.dtype.is_fp8() else key_load / tl.load(k_scale) else: key_tile = key_load # [TILE_SIZE] value_load = tl.load( value_ptr + src_value_idx + tile_pos, mask=tile_pos < (num_heads * head_size) ) if FP8_KV_CACHE: if value_load.dtype.is_fp8(): value_tile = value_load else: # tl.store will do the correct implicit cast to fp8, # based on the value_cache_ptr.dtype.element_ty value_tile = value_load / tl.load(v_scale) else: value_tile = value_load tl.store( key_cache_ptr + tgt_idx + tile_pos, key_tile, mask=tile_pos < (num_heads * head_size), ) tl.store( value_cache_ptr + tgt_idx + tile_pos, value_tile, mask=tile_pos < (num_heads * head_size), ) return def triton_reshape_and_cache_flash( key: torch.Tensor, # [num_tokens, num_heads, head_size] value: torch.Tensor, # [num_tokens, num_heads, head_size] # [num_blocks, block_size, num_heads, head_size] key_cache: torch.Tensor, # [num_blocks, block_size, num_heads, head_size] value_cache: torch.Tensor, slot_mapping: torch.Tensor, # [num_tokens] kv_cache_dtype: str, # "auto", "fp8" k_scale: torch.Tensor, # float32 v_scale: torch.Tensor, # float32 ): num_heads = key.shape[1] head_size = key.shape[2] block_size = key_cache.shape[1] n = num_heads * head_size key_stride = key.stride()[0] value_stride = value.stride()[0] block_stride = key_cache.stride()[0] page_stride = key_cache.stride()[1] head_stride = key_cache.stride()[2] assert head_stride == head_size, "only continous heads are supported" assert kv_cache_dtype == "auto" or kv_cache_dtype.startswith("fp8"), ( f"unsupported kv_cache_dtype (str), got {kv_cache_dtype}." ) kv_cache_torch_dtype = ( current_platform.fp8_dtype() if kv_cache_dtype.startswith("fp8") else key_cache.dtype ) if key_cache.dtype != kv_cache_torch_dtype and kv_cache_dtype.startswith("fp8"): # to avoid erounous implicit cast in triton kernel (tl.store to uint8) # (e.g. explicit cast to fp8e4m3fnuz is not supported in triton 3.4) key_cache = key_cache.view(kv_cache_torch_dtype) value_cache = value_cache.view(kv_cache_torch_dtype) assert kv_cache_dtype != torch.uint8, ( "explicit fp8 cast and store to " "uint8 is not supported by triton reshape_and_cache_flash" ) FP8_KV_CACHE = kv_cache_dtype.startswith("fp8") assert (not FP8_KV_CACHE) or kv_cache_torch_dtype in [ torch.float8_e4m3fn, torch.float8_e5m2, torch.uint8, torch.float8_e4m3fnuz, ], ( "unsupported dtype of KV cache tensor, got " "{kv_cache_torch_dtype}. Supported kv cache dtypes: fp8e4m3fn, " "fp8e5m2, uint8, bfloat16, float16, float32, fp8e4m3fnuz." ) # heuristics instead of autotuning TILE_SIZE = min(2048, triton.next_power_of_2(n)) if current_platform.is_rocm() or current_platform.is_xpu(): num_stages = 4 num_warps = 8 else: # cuda num_stages = 10 num_warps = 16 if torch.cuda.get_device_capability(key.device)[0] < 9: TILE_SIZE = min(512, TILE_SIZE) # TODO(ngl): maybe replace with static launch grid to avoid overhead if # using cudagraphs grid = lambda meta: ( slot_mapping.shape[0], triton.cdiv(n, meta["TILE_SIZE"]), ) reshape_and_cache_kernel_flash[grid]( key_ptr=key, value_ptr=value, key_cache_ptr=key_cache, value_cache_ptr=value_cache, slot_mapping_ptr=slot_mapping, k_scale=k_scale, v_scale=v_scale, # strides key_stride=key_stride, value_stride=value_stride, block_stride=block_stride, page_stride=page_stride, num_heads=num_heads, head_size=head_size, block_size=block_size, # FP8 flags FP8_KV_CACHE=FP8_KV_CACHE, # autotune parameters TILE_SIZE=TILE_SIZE, num_warps=num_warps, num_stages=num_stages, ) @triton.jit def reshape_and_cache_kernel_flash_diffkv( key_ptr, # [num_tokens, num_heads, head_size] value_ptr, # [num_tokens, num_heads, head_size_v] kv_cache_ptr, # [num_blocks, block_size, num_heads, head_size + head_size_v] slot_mapping_ptr, # [num_tokens] k_scale, # float32 v_scale, # float32 # strides key_stride: tl.int64, value_stride: tl.int64, block_stride: tl.int64, page_stride: tl.int64, num_heads: tl.constexpr, head_size_k: tl.constexpr, head_size_v: tl.constexpr, block_size: tl.constexpr, # FP8 flags FP8_KV_CACHE: tl.constexpr, # tune parameters TILE_SIZE: tl.constexpr, ): token_idx = tl.program_id(axis=0) slot_idx = tl.load(slot_mapping_ptr + token_idx).to(tl.int64) if slot_idx < 0: # Padding token that should be ignored. return tile_i = tl.program_id(axis=1) tile_offs = tl.arange(0, TILE_SIZE) block_idx = slot_idx // block_size block_offset = slot_idx % block_size src_key_idx = token_idx * key_stride + tile_i * head_size_k src_value_idx = token_idx * value_stride + tile_i * head_size_v tgt_idx = ( block_idx * block_stride + block_offset * page_stride + tile_i * (head_size_k + head_size_v) ) # [TILE_SIZE] key_load = tl.load(key_ptr + src_key_idx + tile_offs, mask=tile_offs < head_size_k) if FP8_KV_CACHE: # tl.store will do the correct implicit cast to fp8, # based on the key_cache_ptr.dtype.element_ty key_tile = key_load if key_load.dtype.is_fp8() else key_load / tl.load(k_scale) else: key_tile = key_load # [TILE_SIZE] value_load = tl.load( value_ptr + src_value_idx + tile_offs, mask=tile_offs < head_size_v ) if FP8_KV_CACHE: if value_load.dtype.is_fp8(): value_tile = value_load else: # tl.store will do the correct implicit cast to fp8, # based on the value_cache_ptr.dtype.element_ty value_tile = value_load / tl.load(v_scale) else: value_tile = value_load tl.store( kv_cache_ptr + tgt_idx + tile_offs, key_tile, mask=tile_offs < head_size_k, ) tl.store( kv_cache_ptr + tgt_idx + head_size_k + tile_offs, value_tile, mask=tile_offs < head_size_v, ) return def triton_reshape_and_cache_flash_diffkv( key: torch.Tensor, # [num_tokens, num_heads, head_size] value: torch.Tensor, # [num_tokens, num_heads, head_size_v] # [num_blocks, block_size, num_heads, head_size + head_size_v] kv_cache: torch.Tensor, slot_mapping: torch.Tensor, # [num_tokens] kv_cache_dtype: str, # "auto", "fp8" k_scale: torch.Tensor, # float32 v_scale: torch.Tensor, # float32 ): num_heads = key.shape[1] head_size_k = key.shape[2] head_size_v = value.shape[2] block_size = kv_cache.shape[1] k_stride = key.stride()[0] v_stride = value.stride()[0] block_stride = kv_cache.stride()[0] page_stride = kv_cache.stride()[1] assert kv_cache_dtype == "auto" or kv_cache_dtype.startswith("fp8"), ( f"unsupported kv_cache_dtype (str), got {kv_cache_dtype}." ) kv_cache_torch_dtype = ( current_platform.fp8_dtype() if kv_cache_dtype.startswith("fp8") else kv_cache.dtype ) if kv_cache.dtype != kv_cache_torch_dtype and kv_cache_dtype.startswith("fp8"): # to avoid erounous implicit cast in triton kernel (tl.store to uint8) # (e.g. explicit cast to fp8e4m3fnuz is not supported in triton 3.4) kv_cache = kv_cache.view(kv_cache_torch_dtype) assert kv_cache_dtype != torch.uint8, ( "explicit fp8 cast and store to " "uint8 is not supported by triton reshape_and_cache_flash_diffkv" ) FP8_KV_CACHE = kv_cache_dtype.startswith("fp8") assert (not FP8_KV_CACHE) or kv_cache_torch_dtype in [ torch.float8_e4m3fn, torch.float8_e5m2, torch.uint8, torch.float8_e4m3fnuz, ], ( "unsupported dtype of KV cache tensor, got " "{kv_cache_torch_dtype}. Supported kv cache dtypes: fp8e4m3fn, " "fp8e5m2, uint8, bfloat16, float16, float32, fp8e4m3fnuz." ) # heuristics instead of autotuning TILE_SIZE = max(head_size_k, head_size_v) TILE_SIZE = triton.next_power_of_2(TILE_SIZE) if current_platform.is_rocm() or current_platform.is_xpu(): num_stages = 4 num_warps = 8 else: # cuda num_stages = 10 num_warps = 16 # TODO(ngl): maybe replace with static launch grid to avoid overhead if # using cudagraphs grid = lambda meta: ( slot_mapping.shape[0], num_heads, ) reshape_and_cache_kernel_flash_diffkv[grid]( key_ptr=key, value_ptr=value, kv_cache_ptr=kv_cache, slot_mapping_ptr=slot_mapping, k_scale=k_scale, v_scale=v_scale, # strides key_stride=k_stride, value_stride=v_stride, block_stride=block_stride, page_stride=page_stride, num_heads=num_heads, head_size_k=head_size_k, head_size_v=head_size_v, block_size=block_size, # FP8 flags FP8_KV_CACHE=FP8_KV_CACHE, # autotune parameters TILE_SIZE=TILE_SIZE, num_warps=num_warps, num_stages=num_stages, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/__init__.py
vllm/attention/ops/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/chunked_prefill_paged_decode.py
vllm/attention/ops/chunked_prefill_paged_decode.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Authors: # - Burkhard Ringlein <ngl@zurich.ibm.com> # - Jan van Lunteren <jvl@zurich.ibm.com> # - Chih-Chieh Yang <chih.chieh.yang@ibm.com> # - Thomas Parnell <tpa@zurich.ibm.com> import torch from vllm import _custom_ops as ops from vllm.platforms import current_platform from vllm.triton_utils import tl, triton from .prefix_prefill import context_attention_fwd float8_info = torch.finfo(current_platform.fp8_dtype()) @triton.jit def cdiv_fn(x, y): return (x + y - 1) // y @triton.jit def kernel_paged_attention_2d( output_ptr, # [num_tokens, num_query_heads, head_size] query_ptr, # [num_tokens, num_query_heads, head_size] key_cache_ptr, # [num_blks, num_kv_heads, head_size // x, blk_size, x] value_cache_ptr, # [num_blks, num_kv_heads, head_size, blk_size] sink_ptr, # [num_query_heads] block_tables_ptr, # [num_seqs, max_num_blocks_per_seq] seq_lens_ptr, # [num_seqs] alibi_slopes_ptr, # [num_query_heads] scale, # float32 k_scale, # float32 v_scale, # float32 out_scale_inv, num_query_heads: tl.constexpr, # int num_queries_per_kv: tl.constexpr, # int num_queries_per_kv_padded: tl.constexpr, # int block_table_stride: tl.int64, # int query_stride_0: tl.int64, # int query_stride_1: tl.int64, # int, should be equal to head_size output_stride_0: tl.int64, # int output_stride_1: tl.int64, # int, should be equal to head_size BLOCK_SIZE: tl.constexpr, # int HEAD_SIZE: tl.constexpr, # int HEAD_SIZE_PADDED: tl.constexpr, # int, must be power of 2 USE_ALIBI_SLOPES: tl.constexpr, # bool SLIDING_WINDOW: tl.constexpr, # int x: tl.constexpr, # int stride_k_cache_0: tl.int64, # int stride_k_cache_1: tl.int64, # int stride_k_cache_2: tl.int64, # int stride_k_cache_3: tl.int64, # int stride_k_cache_4: tl.int64, # int stride_v_cache_0: tl.int64, # int stride_v_cache_1: tl.int64, # int stride_v_cache_2: tl.int64, # int stride_v_cache_3: tl.int64, # int filter_by_query_len: tl.constexpr, # bool query_start_len_ptr, # [num_seqs+1] USE_SINKS: tl.constexpr, # bool USE_FP8: tl.constexpr, FP8_MIN: tl.constexpr = float8_info.min, FP8_MAX: tl.constexpr = float8_info.max, ): seq_idx = tl.program_id(0) kv_head_idx = tl.program_id(1) if filter_by_query_len: cur_batch_in_all_start_index = tl.load(query_start_len_ptr + seq_idx) cur_batch_in_all_stop_index = tl.load(query_start_len_ptr + seq_idx + 1) cur_batch_query_len = cur_batch_in_all_stop_index - cur_batch_in_all_start_index if cur_batch_query_len > 1: return else: cur_batch_in_all_start_index = seq_idx query_head_idx = kv_head_idx * num_queries_per_kv + tl.arange( 0, num_queries_per_kv_padded ) query_offset = ( cur_batch_in_all_start_index * query_stride_0 + query_head_idx[:, None] * query_stride_1 ) head_mask = query_head_idx < (kv_head_idx + 1) * num_queries_per_kv head_mask = head_mask & (query_head_idx < num_query_heads) dim_mask = tl.where(tl.arange(0, HEAD_SIZE_PADDED) < HEAD_SIZE, 1, 0).to(tl.int1) # Q : (num_queries_per_kv, HEAD_SIZE,) Q = tl.load( query_ptr + query_offset + tl.arange(0, HEAD_SIZE_PADDED)[None, :], mask=dim_mask[None, :] & head_mask[:, None], other=0.0, ) block_table_offset = seq_idx * block_table_stride if not USE_SINKS: M = tl.full([num_queries_per_kv_padded], float("-inf"), dtype=tl.float32) else: M = tl.load( sink_ptr + query_head_idx, mask=head_mask, other=float("-inf"), ).to(dtype=tl.float32) L = tl.full([num_queries_per_kv_padded], 1.0, dtype=tl.float32) acc = tl.zeros([num_queries_per_kv_padded, HEAD_SIZE_PADDED], dtype=tl.float32) # sequence len for this particular sequence seq_len = tl.load(seq_lens_ptr + seq_idx) # alibi slope for this head if USE_ALIBI_SLOPES: alibi_slope = tl.load( alibi_slopes_ptr + query_head_idx, mask=head_mask, other=0.0 ) num_blocks = cdiv_fn(seq_len, BLOCK_SIZE) # iterate through tiles for j in range(0, num_blocks): physical_block_idx = tl.load(block_tables_ptr + block_table_offset + j) offs_n = tl.arange(0, BLOCK_SIZE) offs_d = tl.arange(0, HEAD_SIZE_PADDED) v_offset = ( physical_block_idx * stride_v_cache_0 + kv_head_idx * stride_v_cache_1 + offs_d[None, :] * stride_v_cache_2 + offs_n[:, None] * stride_v_cache_3 ) k_offset = ( physical_block_idx * stride_k_cache_0 + kv_head_idx * stride_k_cache_1 + (offs_d[:, None] // x) * stride_k_cache_2 + offs_n[None, :] * stride_k_cache_3 + (offs_d[:, None] % x) * stride_k_cache_4 ) # K : (HEAD_SIZE, BLOCK_SIZE) K_load = tl.load(key_cache_ptr + k_offset, mask=dim_mask[:, None], other=0.0) if K_load.dtype.is_fp8(): K = (K_load.to(tl.float32) * tl.load(k_scale)).to(Q.dtype) else: K = K_load # V : (BLOCK_SIZE, HEAD_SIZE) V_load = tl.load(value_cache_ptr + v_offset, mask=dim_mask[None, :], other=0.0) if V_load.dtype.is_fp8(): V = (V_load.to(tl.float32) * tl.load(v_scale)).to(Q.dtype) else: V = V_load seq_offset = j * BLOCK_SIZE + tl.arange(0, BLOCK_SIZE) boundary = tl.full([BLOCK_SIZE], seq_len, dtype=tl.int32) seq_mask = seq_offset[None, :] < boundary # S : (num_queries_per_kv, BLOCK_SIZE,) S = tl.where(head_mask[:, None] & seq_mask, 0.0, float("-inf")).to(tl.float32) S += scale * tl.dot(Q, K) context_len = seq_len - 1 if SLIDING_WINDOW > 0: S = tl.where((context_len - seq_offset) < SLIDING_WINDOW, S, -10000) if USE_ALIBI_SLOPES: S += alibi_slope[:, None] * (seq_offset - context_len) # compute running maximum # m_j : (num_queries_per_kv,) m_j = tl.maximum(M, tl.max(S, axis=1)) # P : (num_queries_per_kv, BLOCK_SIZE,) P = tl.exp(S - m_j[:, None]) # l_j : (num_queries_per_kv,) l_j = tl.sum(P, axis=1) # alpha : (num_queries_per_kv, ) alpha = tl.exp(M - m_j) # acc : (num_queries_per_kv, BLOCK_SIZE,) acc = acc * alpha[:, None] # update constants L = L * alpha + l_j M = m_j # acc : (num_queries_per_kv, BLOCK_SIZE,) acc += tl.dot(P.to(V.dtype), V) # epilogue acc = acc / L[:, None] if USE_FP8: acc = acc * tl.load(out_scale_inv) acc = tl.clamp(acc, FP8_MIN, FP8_MAX) output_offset = ( cur_batch_in_all_start_index * output_stride_0 + query_head_idx * output_stride_1 ) tl.store( output_ptr + output_offset[:, None] + tl.arange(0, HEAD_SIZE_PADDED)[None, :], acc, mask=dim_mask[None, :] & head_mask[:, None], ) def chunked_prefill_paged_decode( query, key, value, output, kv_cache_dtype, key_cache, value_cache, block_table, query_start_loc, seq_lens, max_seq_len, max_query_len, k_scale, v_scale, alibi_slopes=None, sliding_window=None, sm_scale=None, output_scale=None, # Optional tensor for sinks sinks=None, ): if sm_scale is None: sm_scale = 1.0 / (query.shape[1] ** 0.5) use_alibi_slopes = alibi_slopes is not None if sliding_window is None or sliding_window <= 0: sliding_window = 0 if max_query_len > 1: context_attention_fwd( q=query, k=key, v=value, o=output, kv_cache_dtype=kv_cache_dtype, k_cache=key_cache, v_cache=value_cache, b_loc=block_table, b_start_loc=query_start_loc, b_seq_len=seq_lens, max_seq_len=max_seq_len, max_input_len=max_query_len, k_scale=k_scale, v_scale=v_scale, alibi_slopes=alibi_slopes, sliding_window=sliding_window, sm_scale=sm_scale, skip_decode=True, fp8_out_scale=output_scale, sinks=sinks, ) block_size = value_cache.shape[3] num_seqs = len(seq_lens) num_query_heads = query.shape[1] num_kv_heads = key.shape[1] num_queries_per_kv = query.shape[1] // key.shape[1] head_size = query.shape[2] # Conversion of FP8 Tensor from uint8 storage to # appropriate torch.dtype for interpretation by Triton if "fp8" in kv_cache_dtype: assert key_cache.dtype in [torch.uint8, current_platform.fp8_dtype()] assert value_cache.dtype in [torch.uint8, current_platform.fp8_dtype()] if kv_cache_dtype in ("fp8", "fp8_e4m3"): target_dtype = current_platform.fp8_dtype() elif kv_cache_dtype == "fp8_e5m2": target_dtype = torch.float8_e5m2 else: raise ValueError("Unsupported FP8 dtype:", kv_cache_dtype) key_cache = key_cache.view(target_dtype) value_cache = value_cache.view(target_dtype) num_queries_per_kv_padded = max(triton.next_power_of_2(num_queries_per_kv), 16) from vllm.platforms.rocm import use_rocm_custom_paged_attention use_custom = use_rocm_custom_paged_attention( query.dtype, head_size, block_size, num_queries_per_kv, max_seq_len, sliding_window, kv_cache_dtype, alibi_slopes, sinks, ) if use_custom: _PARTITION_SIZE_ROCM = 256 max_num_partitions = ( max_seq_len + _PARTITION_SIZE_ROCM - 1 ) // _PARTITION_SIZE_ROCM assert _PARTITION_SIZE_ROCM % block_size == 0 total_num_seq = block_table.shape[0] tmp_output = torch.empty( size=(total_num_seq, num_query_heads, max_num_partitions, head_size), dtype=query.dtype, device=output.device, ) exp_sums = torch.empty( size=(total_num_seq, num_query_heads, max_num_partitions), dtype=torch.float32, device=output.device, ) max_logits = torch.empty_like(exp_sums) ops.paged_attention_rocm( output, exp_sums, max_logits, tmp_output, query, key_cache, value_cache, num_kv_heads, scale=sm_scale, block_tables=block_table, seq_lens=seq_lens, query_start_loc=query_start_loc, block_size=block_size, max_seq_len=max_seq_len, alibi_slopes=alibi_slopes, kv_cache_dtype=kv_cache_dtype, k_scale=k_scale, v_scale=v_scale, fp8_out_scale=output_scale, ) else: kernel_paged_attention_2d[ ( num_seqs, num_kv_heads, ) ]( output_ptr=output, query_ptr=query, key_cache_ptr=key_cache, value_cache_ptr=value_cache, sink_ptr=sinks, block_tables_ptr=block_table, seq_lens_ptr=seq_lens, alibi_slopes_ptr=alibi_slopes, scale=sm_scale, k_scale=k_scale, v_scale=v_scale, out_scale_inv=1.0 / output_scale if output_scale is not None else 1.0, num_query_heads=num_query_heads, num_queries_per_kv=num_queries_per_kv, num_queries_per_kv_padded=num_queries_per_kv_padded, block_table_stride=block_table.stride(0), query_stride_0=query.stride(0), query_stride_1=query.stride(1), output_stride_0=output.stride(0), output_stride_1=output.stride(1), BLOCK_SIZE=block_size, HEAD_SIZE=head_size, HEAD_SIZE_PADDED=triton.next_power_of_2(head_size), USE_ALIBI_SLOPES=use_alibi_slopes, SLIDING_WINDOW=sliding_window, x=key_cache.shape[4], stride_k_cache_0=key_cache.stride(0), stride_k_cache_1=key_cache.stride(1), stride_k_cache_2=key_cache.stride(2), stride_k_cache_3=key_cache.stride(3), stride_k_cache_4=key_cache.stride(4), stride_v_cache_0=value_cache.stride(0), stride_v_cache_1=value_cache.stride(1), stride_v_cache_2=value_cache.stride(2), stride_v_cache_3=value_cache.stride(3), filter_by_query_len=True, query_start_len_ptr=query_start_loc, USE_SINKS=sinks is not None, USE_FP8=output_scale is not None, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/attention/ops/pallas_kv_cache_update.py
vllm/attention/ops/pallas_kv_cache_update.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools import jax from jax.experimental import pallas as pl from jax.experimental.pallas import tpu as pltpu from vllm.utils.math_utils import cdiv def _kv_cache_update_kernel( # Prefetch slices_ref, # [3, padded_num_slices], list of (kv_cache_start, # new_kv_start, slice_len) num_slices_ref, # [1] # Input new_kv_hbm_ref, # [num_tokens, num_combined_kv_heads, head_dim] kv_cache_hbm_ref, # [total_num_pages * page_size, num_combined_kv_heads, # head_dim] # Output _, # [total_num_pages * page_size, num_combined_kv_heads, head_dim] # Scratch scratch, # [num_slices_per_block, page_size, num_combined_kv_heads, # head_dim] sem, ): async_copies = [] block_idx = pl.program_id(0) num_slices_per_block = scratch.shape[0] # Copy from new_kv_hbm_ref to scratch for i in range(num_slices_per_block): offset_i = i + block_idx * num_slices_per_block new_kv_start = jax.lax.select( offset_i < num_slices_ref[0], slices_ref[1, offset_i], 0 ) length = jax.lax.select( offset_i < num_slices_ref[0], slices_ref[2, offset_i], 0 ) async_copy = pltpu.make_async_copy( new_kv_hbm_ref.at[pl.ds(new_kv_start, length), ...], scratch.at[i, pl.ds(0, length), ...], sem, ) async_copy.start() async_copies.append(async_copy) for async_copy in async_copies: async_copy.wait() # Copy from scratch to kv_cache_hbm_ref async_copies.clear() for i in range(num_slices_per_block): offset_i = i + block_idx * num_slices_per_block kv_cache_start = jax.lax.select( offset_i < num_slices_ref[0], slices_ref[0, offset_i], 0 ) length = jax.lax.select( offset_i < num_slices_ref[0], slices_ref[2, offset_i], 0 ) async_copy = pltpu.make_async_copy( scratch.at[i, pl.ds(0, length), ...], kv_cache_hbm_ref.at[pl.ds(kv_cache_start, length), ...], sem, ) async_copy.start() async_copies.append(async_copy) for async_copy in async_copies: async_copy.wait() @functools.partial( jax.jit, static_argnames=["page_size", "num_slices_per_block"], ) def kv_cache_update( # [total_num_token, num_combined_kv_heads, head_dim] new_kv: jax.Array, # [3, slices], list of (kv_cache_start, new_kv_start, slice_len) slices: jax.Array, # [total_num_pages * page_size, num_combined_kv_heads, head_dim] kv_cache: jax.Array, # [1] num_kv_update_slices: jax.Array, *, page_size: int = 32, num_slices_per_block: int = 8, ): _, num_combined_kv_heads, head_dim = new_kv.shape assert kv_cache.shape[1] == num_combined_kv_heads assert kv_cache.shape[2] == head_dim assert head_dim % 128 == 0 # TODO: Add dynamic check to make sure that the all the slice lengths are # smaller or equal to page_size in_specs = [ pl.BlockSpec(memory_space=pltpu.TPUMemorySpace.ANY), pl.BlockSpec(memory_space=pltpu.TPUMemorySpace.ANY), ] out_specs = [pl.BlockSpec(memory_space=pltpu.TPUMemorySpace.ANY)] out_shape = [jax.ShapeDtypeStruct(kv_cache.shape, dtype=kv_cache.dtype)] scalar_prefetches = [slices, num_kv_update_slices] scratch = pltpu.VMEM( (num_slices_per_block, page_size, num_combined_kv_heads, head_dim), new_kv.dtype, ) scratch_shapes = [ scratch, pltpu.SemaphoreType.DMA, ] kernel = pl.pallas_call( _kv_cache_update_kernel, grid_spec=pltpu.PrefetchScalarGridSpec( num_scalar_prefetch=len(scalar_prefetches), in_specs=in_specs, out_specs=out_specs, grid=(cdiv(num_kv_update_slices[0], num_slices_per_block),), scratch_shapes=scratch_shapes, ), out_shape=out_shape, input_output_aliases={len(scalar_prefetches) + 1: 0}, ) return kernel(*scalar_prefetches, new_kv, kv_cache)[0]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/profiler/utils.py
vllm/profiler/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import dataclasses from collections.abc import Callable from torch._C._profiler import _EventType, _ProfilerEvent, _TensorMetadata # # String / Print Manipulation # def trim_string_front(string, width): if len(string) > width: offset = len(string) - width + 3 string = string[offset:] if len(string) > 3: string = "..." + string[3:] return string def trim_string_back(string, width): if len(string) > width: offset = len(string) - width + 3 string = string[:-offset] if len(string) > 3: string = string + "..." return string class TablePrinter: def __init__( self, row_cls: type[dataclasses.dataclass], column_widths: dict[str, int] ): self.row_cls = row_cls self.fieldnames = [x.name for x in dataclasses.fields(row_cls)] self.column_widths = column_widths assert set(self.column_widths.keys()) == set(self.fieldnames) def print_table(self, rows: list[dataclasses.dataclass]): self._print_header() self._print_line() for row in rows: self._print_row(row) def _print_header(self): for i, f in enumerate(self.fieldnames): last = i == len(self.fieldnames) - 1 col_width = self.column_widths[f] print( trim_string_back(f, col_width).ljust(col_width), end=" | " if not last else "\n", ) def _print_row(self, row): assert isinstance(row, self.row_cls) for i, f in enumerate(self.fieldnames): last = i == len(self.fieldnames) - 1 col_width = self.column_widths[f] val = getattr(row, f) val_str = "" if isinstance(val, str): val_str = trim_string_back(val, col_width).ljust(col_width) elif type(val) in [float, int]: val_str = f"{float(val):>.2f}".rjust(col_width) else: val_str = f"{val}".rjust(col_width) print(val_str, end=" | " if not last else "\n") def _print_line(self): total_col_width = 0 for column_width in self.column_widths.values(): total_col_width += column_width print("=" * (total_col_width + 3 * (len(self.column_widths) - 1))) def indent_string( string: str, indent: int, indent_style: Callable[[int], str] | str = " " ) -> str: if indent: if isinstance(indent_style, str): return indent_style * indent + string else: return indent_style(indent) + string else: return string # # _ProfilerEvent utils # def event_has_module(event: _ProfilerEvent) -> bool: event_type, typed_event = event.typed if event_type == _EventType.PyCall: return typed_event.module is not None return False def event_is_torch_op(event: _ProfilerEvent) -> bool: return event.tag == _EventType.TorchOp def event_arg_repr(arg) -> str: if arg is None or type(arg) in [float, int, bool, str]: return f"{arg}" elif isinstance(arg, list): return f"[{', '.join([event_arg_repr(x) for x in arg])}]" elif isinstance(arg, tuple): return f"({', '.join([event_arg_repr(x) for x in arg])})" else: assert isinstance(arg, _TensorMetadata), f"Unsupported type: {type(arg)}" sizes_str = ", ".join([str(x) for x in arg.sizes]) return f"{str(arg.dtype).replace('torch.', '')}[{sizes_str}]" def event_torch_op_repr(event: _ProfilerEvent) -> str: assert event.tag == _EventType.TorchOp args_str = ", ".join([event_arg_repr(x) for x in event.typed[1].inputs]) return f"{event.name}({args_str})".replace("aten::", "") def event_module_repr(event: _ProfilerEvent) -> str: assert event_has_module(event) module = event.typed[1].module if module.parameters and len(module.parameters) > 0: args_str = ", ".join( [f"{x[0]}={event_arg_repr(x[1])}" for x in module.parameters] ) return f"{module.cls_name}({args_str})" else: return module.cls_name def event_torch_op_stack_trace( curr_event: _ProfilerEvent, until: Callable[[_ProfilerEvent], bool] ) -> str: trace = "" curr_event = curr_event.parent while curr_event and not until(curr_event): if event_is_torch_op(curr_event): if len(trace) > 0: trace += " <- " trace += event_torch_op_repr(curr_event) curr_event = curr_event.parent return trace
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/profiler/wrapper.py
vllm/profiler/wrapper.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import ABC, abstractmethod from contextlib import nullcontext from typing import Literal import torch from typing_extensions import override from vllm.config import ProfilerConfig from vllm.logger import init_logger logger = init_logger(__name__) class WorkerProfiler(ABC): def __init__(self, profiler_config: ProfilerConfig) -> None: self._delay_iters = profiler_config.delay_iterations if self._delay_iters > 0: logger.info_once( "GPU profiling will start " f"{self._delay_iters} steps after start_profile." ) self._max_iters = profiler_config.max_iterations if self._max_iters > 0: logger.info_once( "GPU profiling will stop " f"after {self._max_iters} worker steps, " "or when stop_profile is received." ) # Track when the profiler gets triggered by start_profile self._active_iteration_count = 0 self._active = False # Track when the profiler is actually running self._profiling_for_iters = 0 self._running = False @abstractmethod def _start(self) -> None: """Start the profiler.""" pass @abstractmethod def _stop(self) -> None: """Stop the profiler.""" pass def _call_start(self) -> None: """Call _start with error handling but no safeguards.""" try: self._start() self._running = True # Only mark as running if start succeeds except Exception as e: logger.warning("Failed to start profiler: %s", e) def _call_stop(self) -> None: """Call _stop with error handling but no safeguards.""" try: self._stop() logger.info_once("Profiler stopped successfully.", scope="local") except Exception as e: logger.warning("Failed to stop profiler: %s", e) self._running = False # Always mark as not running, assume stop worked def start(self) -> None: """Attempt to start the profiler, accounting for delayed starts.""" if self._active: logger.debug( "start_profile received when profiler is already active. " "Ignoring request." ) return self._active = True if self._delay_iters == 0: self._call_start() def step(self) -> None: """Update the profiler state at each worker step, to handle delayed starts and max iteration limits.""" if not self._active: return self._active_iteration_count += 1 if ( not self._running and self._delay_iters > 0 and self._active_iteration_count == self._delay_iters ): logger.info_once("Starting profiler after delay...", scope="local") self._call_start() if self._running: self._profiling_for_iters += 1 if ( self._max_iters > 0 and self._running and self._profiling_for_iters > self._max_iters ): # Automatically stop the profiler after max iters # will be marked as not running, but leave as active so that stop # can clean up properly logger.info_once( "Max profiling iterations reached. Stopping profiler...", scope="local" ) self._call_stop() return def stop(self) -> None: """Attempt to stop the profiler, accounting for overlapped calls.""" if not self._active: logger.debug( "stop_profile received when profiler is not active. Ignoring request." ) return self._active = False self._active_iteration_count = 0 self._profiling_for_iters = 0 if self._running: self._call_stop() def shutdown(self) -> None: """Ensure profiler is stopped when shutting down.""" logger.info_once("Shutting down profiler", scope="local") if self._running: self.stop() def annotate_context_manager(self, name: str): """Return a context manager to annotate profiler traces.""" return nullcontext() TorchProfilerActivity = Literal["CPU", "CUDA", "XPU"] TorchProfilerActivityMap = { "CPU": torch.profiler.ProfilerActivity.CPU, "CUDA": torch.profiler.ProfilerActivity.CUDA, "XPU": torch.profiler.ProfilerActivity.XPU, } class TorchProfilerWrapper(WorkerProfiler): def __init__( self, profiler_config: ProfilerConfig, worker_name: str, local_rank: int, activities: list[TorchProfilerActivity], ) -> None: super().__init__(profiler_config) self.local_rank = local_rank self.profiler_config = profiler_config torch_profiler_trace_dir = profiler_config.torch_profiler_dir if local_rank in (None, 0): logger.info_once( "Torch profiling enabled. Traces will be saved to: %s", torch_profiler_trace_dir, scope="local", ) logger.debug( "Profiler config: record_shapes=%s," "profile_memory=%s,with_stack=%s,with_flops=%s", profiler_config.torch_profiler_record_shapes, profiler_config.torch_profiler_with_memory, profiler_config.torch_profiler_with_stack, profiler_config.torch_profiler_with_flops, ) self.dump_cpu_time_total = "CPU" in activities and len(activities) == 1 self.profiler = torch.profiler.profile( activities=[TorchProfilerActivityMap[activity] for activity in activities], record_shapes=profiler_config.torch_profiler_record_shapes, profile_memory=profiler_config.torch_profiler_with_memory, with_stack=profiler_config.torch_profiler_with_stack, with_flops=profiler_config.torch_profiler_with_flops, on_trace_ready=torch.profiler.tensorboard_trace_handler( torch_profiler_trace_dir, worker_name=worker_name, use_gzip=profiler_config.torch_profiler_use_gzip, ), ) @override def _start(self) -> None: self.profiler.start() @override def _stop(self) -> None: self.profiler.stop() profiler_config = self.profiler_config rank = self.local_rank if profiler_config.torch_profiler_dump_cuda_time_total: profiler_dir = profiler_config.torch_profiler_dir profiler_out_file = f"{profiler_dir}/profiler_out_{rank}.txt" sort_key = "self_cuda_time_total" table = self.profiler.key_averages().table(sort_by=sort_key) with open(profiler_out_file, "w") as f: print(table, file=f) # only print profiler results on rank 0 if rank == 0: print(table) if self.dump_cpu_time_total and rank == 0: logger.info( self.profiler.key_averages().table( sort_by="self_cpu_time_total", row_limit=50 ) ) @override def annotate_context_manager(self, name: str): return torch.profiler.record_function(name) class CudaProfilerWrapper(WorkerProfiler): def __init__(self, profiler_config: ProfilerConfig) -> None: super().__init__(profiler_config) # Note: lazy import to avoid dependency issues if CUDA is not available. import torch.cuda.profiler as cuda_profiler self._cuda_profiler = cuda_profiler @override def _start(self) -> None: self._cuda_profiler.start() @override def _stop(self) -> None: self._cuda_profiler.stop() @override def annotate_context_manager(self, name: str): return torch.cuda.nvtx.range(name)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/profiler/__init__.py
vllm/profiler/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/profiler/layerwise_profile.py
vllm/profiler/layerwise_profile.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import copy from collections import defaultdict from collections.abc import Callable from dataclasses import asdict, dataclass, field from typing import Any, Optional, TypeAlias from torch._C._autograd import DeviceType, _KinetoEvent, _ProfilerResult from torch._C._profiler import _EventType, _ExperimentalConfig, _ProfilerEvent from torch.autograd.profiler import FunctionEvent from torch.profiler import ProfilerActivity, profile from vllm.profiler.utils import ( TablePrinter, event_has_module, event_is_torch_op, event_module_repr, event_torch_op_stack_trace, indent_string, ) from vllm.utils.import_utils import PlaceholderModule try: import pandas as pd except ImportError: pd = PlaceholderModule("pandas") @dataclass class _ModuleTreeNode: event: _ProfilerEvent parent: Optional["_ModuleTreeNode"] = None children: list["_ModuleTreeNode"] = field(default_factory=list) trace: str = "" @property def is_leaf(self): return self.event.children is None or len(self.event.children) == 0 @property def is_torch_op(self): return event_is_torch_op(self.event) @property def is_cuda(self): return ( self.event.tag == _EventType.Kineto and self.event.typed[1].device_type == DeviceType.CUDA ) @dataclass class SummaryStatsEntry: name: str cuda_time_us: float pct_cuda_time: float invocations: int @dataclass class ModelStatsEntry: name: str cpu_time_us: float cuda_time_us: float pct_cuda_time: float trace: str StatsEntry: TypeAlias = ModelStatsEntry | SummaryStatsEntry @dataclass class _StatsTreeNode: entry: StatsEntry children: list[StatsEntry] parent: StatsEntry | None @dataclass class LayerwiseProfileResults(profile): _kineto_results: _ProfilerResult _kineto_event_correlation_map: dict[int, list[_KinetoEvent]] = field(init=False) _event_correlation_map: dict[int, list[FunctionEvent]] = field(init=False) _module_tree: list[_ModuleTreeNode] = field(init=False) _model_stats_tree: list[_StatsTreeNode] = field(init=False) _summary_stats_tree: list[_StatsTreeNode] = field(init=False) # profile metadata num_running_seqs: int | None = None def __post_init__(self): self._build_correlation_map() self._build_module_tree() self._build_stats_trees() def print_model_table(self, column_widths: dict[str, int] = None): _column_widths = dict( name=60, cpu_time_us=12, cuda_time_us=12, pct_cuda_time=12, trace=60 ) if column_widths: _column_widths.update(**column_widths) filtered_model_table = [ (depth, row) for depth, row in self._flatten_stats_tree(self._model_stats_tree) if row.cuda_time_us > 0 or row.cpu_time_us > 0 ] TablePrinter(ModelStatsEntry, _column_widths).print_table( self._indent_row_names_based_on_depth( filtered_model_table, indent_style=lambda indent: "|" + "-" * indent + " ", ) ) def print_summary_table(self, column_widths: dict[str, int] = None): _column_widths = dict( name=80, cuda_time_us=12, pct_cuda_time=12, invocations=15 ) if column_widths: _column_widths.update(**column_widths) filtered_summary_table = [ (depth, row) for depth, row in self._flatten_stats_tree(self._summary_stats_tree) if row.cuda_time_us > 0 ] TablePrinter(SummaryStatsEntry, _column_widths).print_table( self._indent_row_names_based_on_depth( filtered_summary_table, indent_style=lambda indent: "|" + "-" * indent + " ", ) ) def export_model_stats_table_csv(self, filename: str): df = pd.DataFrame( [asdict(row) for _, row in self._flatten_stats_tree(self._model_stats_tree)] ) df.to_csv(filename) def export_summary_stats_table_csv(self, filename: str): df = pd.DataFrame( [ asdict(row) for _, row in self._flatten_stats_tree(self._summary_stats_tree) ] ) df.to_csv(filename) def convert_stats_to_dict(self) -> dict[str, Any]: return { "metadata": {"num_running_seqs": self.num_running_seqs}, "summary_stats": self._convert_stats_tree_to_dict(self._summary_stats_tree), "model_stats": self._convert_stats_tree_to_dict(self._model_stats_tree), } @staticmethod def _indent_row_names_based_on_depth( depths_rows: list[tuple[int, StatsEntry]], indent_style: Callable[[int], str] | str = " ", ): indented_rows = [] for depth, row in depths_rows: if row.cuda_time_us == 0: continue indented_row = copy.deepcopy(row) indented_row.name = indent_string(indented_row.name, depth, indent_style) indented_rows.append(indented_row) return indented_rows def _build_correlation_map(self): self._kineto_event_correlation_map = defaultdict(list) for event in self._kineto_results.events(): self._kineto_event_correlation_map[event.correlation_id()].append(event) def _build_module_tree(self): self._module_tree = [] event_tree = self._kineto_results.experimental_event_tree() def _df_traversal( event: _ProfilerEvent, curr_node: _ModuleTreeNode | None = None ): # For the tensor parallel case for now only look at task 1 if event.start_tid != 1: return if event_has_module(event): node = _ModuleTreeNode(event=event, parent=curr_node) if curr_node: curr_node.children.append(node) else: self._module_tree.append(node) curr_node = node is_leaf = event.children is None or len(event.children) == 0 if is_leaf and curr_node: node = _ModuleTreeNode( event=event, parent=curr_node, trace=event_torch_op_stack_trace( event, until=lambda x: event_has_module(x) ), ) curr_node.children.append(node) curr_node = node for child in event.children: _df_traversal(child, curr_node) for root in event_tree: _df_traversal(root) def _get_kineto_gpu_event(self, node: _ModuleTreeNode): if node.event.tag != _EventType.Kineto: return None correlated_kineto_events = self._kineto_event_correlation_map.get( node.event.correlation_id, [] ) iterator = ( x for x in correlated_kineto_events if x.device_type() == DeviceType.CUDA and x.name() == node.event.name ) return next(iterator, None) def _cumulative_cuda_time(self, node: _ModuleTreeNode): "Return cuda time in microseconds" def _cumulative_cuda_time_recursive(node: _ModuleTreeNode): if node.is_leaf and (gpu_kineto_event := self._get_kineto_gpu_event(node)): return gpu_kineto_event.duration_ns() / 1000.0 else: cumulative_cuda_time = 0 for child in node.children: cumulative_cuda_time += _cumulative_cuda_time_recursive(child) return cumulative_cuda_time return _cumulative_cuda_time_recursive(node) def _total_cuda_time(self): return sum([self._cumulative_cuda_time(root) for root in self._module_tree]) def _build_stats_trees(self): summary_dict: dict[str, _StatsTreeNode] = {} total_cuda_time = self._total_cuda_time() def pct_cuda_time(cuda_time_us): return (cuda_time_us / total_cuda_time) * 100 def build_summary_stats_tree_df( node: _ModuleTreeNode, parent: _StatsTreeNode | None = None, summary_trace: tuple[str] = (), ): if event_has_module(node.event): name = event_module_repr(node.event) cuda_time_us = self._cumulative_cuda_time(node) elif gpu_kineto_event := self._get_kineto_gpu_event(node): name = gpu_kineto_event.name() cuda_time_us = gpu_kineto_event.duration_ns() / 1000.0 else: return None summary_trace = summary_trace + (name,) if summary_trace in summary_dict: entry = summary_dict[summary_trace].entry entry.cuda_time_us += cuda_time_us entry.invocations += 1 entry.pct_cuda_time = pct_cuda_time(entry.cuda_time_us) else: new_node = _StatsTreeNode( entry=SummaryStatsEntry( name=name, cuda_time_us=cuda_time_us, pct_cuda_time=pct_cuda_time(cuda_time_us), invocations=1, ), children=[], parent=parent, ) if parent: parent.children.append(new_node) summary_dict[summary_trace] = new_node for child in node.children: build_summary_stats_tree_df( child, summary_dict[summary_trace], summary_trace ) return summary_dict[summary_trace] self._summary_stats_tree = [] for root in self._module_tree: self._summary_stats_tree.append(build_summary_stats_tree_df(root)) def build_model_stats_tree_df( node: _ModuleTreeNode, parent: _StatsTreeNode | None = None ): if event_has_module( node.event, ): name = event_module_repr(node.event) cuda_time_us = self._cumulative_cuda_time(node) cpu_time_us = node.event.duration_time_ns / 1000 trace = "" elif gpu_kineto_event := self._get_kineto_gpu_event(node): name = gpu_kineto_event.name() cuda_time_us = gpu_kineto_event.duration_ns() / 1000.0 cpu_time_us = 0 trace = node.trace else: return None new_node = _StatsTreeNode( entry=ModelStatsEntry( name=name, cpu_time_us=cpu_time_us, cuda_time_us=cuda_time_us, pct_cuda_time=pct_cuda_time(cuda_time_us), trace=trace, ), parent=parent, children=[], ) if parent: parent.children.append(new_node) for child in node.children: build_model_stats_tree_df(child, new_node) return new_node self._model_stats_tree = [] for root in self._module_tree: self._model_stats_tree.append(build_model_stats_tree_df(root)) def _flatten_stats_tree( self, tree: list[_StatsTreeNode] ) -> list[tuple[int, StatsEntry]]: entries: list[tuple[int, StatsEntry]] = [] def df_traversal(node: _StatsTreeNode, depth=0): entries.append((depth, node.entry)) for child in node.children: df_traversal(child, depth=depth + 1) for root in tree: df_traversal(root) return entries def _convert_stats_tree_to_dict(self, tree: list[_StatsTreeNode]) -> list[dict]: root_dicts: list[dict] = [] def df_traversal(node: _StatsTreeNode, curr_json_list: list[dict]): curr_json_list.append({"entry": asdict(node.entry), "children": []}) for child in node.children: df_traversal(child, curr_json_list[-1]["children"]) for root in tree: df_traversal(root, root_dicts) return root_dicts class layerwise_profile(profile): def __init__(self, num_running_seqs: int | None = None): """ layerwise profile constructor. Args: num_running_seqs (Optional[int], optional): When given, num_running_seqs will be passed to LayerProfileResults for metadata update. Defaults to None. """ super().__init__( activities=[ProfilerActivity.CPU, ProfilerActivity.CUDA], record_shapes=True, with_stack=True, with_modules=True, experimental_config=_ExperimentalConfig(verbose=True), ) self.num_running_seqs = num_running_seqs def __enter__(self): return super().__enter__() def __exit__(self, exc_type, exc_val, exc_tb): super().__exit__(exc_type, exc_val, exc_tb) self.results = LayerwiseProfileResults( self.profiler.kineto_results, num_running_seqs=self.num_running_seqs )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/mm_processor.py
vllm/benchmarks/mm_processor.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project r"""Benchmark multimodal processor latency. This benchmark measures the latency of the mm processor module using multimodal prompts from datasets. MM processor stats are automatically enabled. Run: vllm bench mm-processor \ --model <your_model> \ --dataset-name random-mm \ --num-prompts 10 \ """ import argparse import dataclasses import json import time from datetime import datetime from typing import Any import numpy as np from vllm.benchmarks.throughput import get_requests from vllm.engine.arg_utils import EngineArgs from vllm.multimodal.processing import ( get_timing_stats_from_engine_client, ) from vllm.utils.gc_utils import freeze_gc_heap from vllm.utils.import_utils import PlaceholderModule try: import pandas as pd except ImportError: pd = PlaceholderModule("pandas") def collect_mm_processor_stats( llm_engine: Any, ) -> dict[str, list[float]]: """ Collect multimodal processor timing stats. Returns a dictionary mapping stage names to lists of timing values (in seconds). """ all_stats = get_timing_stats_from_engine_client(llm_engine) stats_by_stage = { "hf_processor_time": [], "hashing_time": [], "cache_lookup_time": [], "prompt_update_time": [], "total_time": [], } for stats_dict in all_stats.values(): stats_by_stage["hf_processor_time"].append( stats_dict.get("hf_processor_time", 0.0) ) stats_by_stage["hashing_time"].append(stats_dict.get("hashing_time", 0.0)) stats_by_stage["cache_lookup_time"].append( stats_dict.get("cache_lookup_time", 0.0) ) stats_by_stage["prompt_update_time"].append( stats_dict.get("prompt_update_time", 0.0) ) stats_by_stage["total_time"].append(stats_dict.get("total_time", 0.0)) return stats_by_stage def calculate_mm_processor_metrics( stats_by_stage: dict[str, list[float]], selected_percentiles: list[float], ) -> dict[str, dict[str, float]]: """ Calculate aggregate metrics from stats by stage. """ metrics = {} for stage_name, times in stats_by_stage.items(): if not times: metrics[stage_name] = { "mean": 0.0, "median": 0.0, "std": 0.0, **{f"p{p}": 0.0 for p in selected_percentiles}, } continue times_ms = [t * 1000 for t in times] metrics[stage_name] = { "mean": float(np.mean(times_ms)), "median": float(np.median(times_ms)), "std": float(np.std(times_ms)), **{ f"p{p}": float(np.percentile(times_ms, p)) for p in selected_percentiles }, } return metrics def validate_args(args): """ Validate command-line arguments for mm_processor benchmark. """ if not getattr(args, "tokenizer", None): args.tokenizer = args.model if not hasattr(args, "dataset_path"): args.dataset_path = None if not hasattr(args, "lora_path"): args.lora_path = None if not hasattr(args, "max_loras"): args.max_loras = None def benchmark_multimodal_processor( args: argparse.Namespace, ) -> dict[str, Any]: """ Run the multimodal processor benchmark. """ from vllm import LLM, SamplingParams validate_args(args) if args.seed is None: args.seed = 0 engine_args = EngineArgs.from_cli_args(args) llm = LLM(**dataclasses.asdict(engine_args)) tokenizer = llm.get_tokenizer() requests = get_requests(args, tokenizer) assert all( llm.llm_engine.model_config.max_model_len >= (request.prompt_len + request.expected_output_len) for request in requests ), ( "Please ensure that max_model_len is greater than the sum of " "prompt_len and expected_output_len for all requests." ) prompts = [request.prompt for request in requests] expected_output_lens = [request.expected_output_len for request in requests] sampling_params = [ SamplingParams( n=1, temperature=0.0, max_tokens=output_len, detokenize=True, ) for output_len in expected_output_lens ] selected_percentiles = [ float(p) for p in getattr(args, "metric_percentiles", "99").split(",") ] freeze_gc_heap() print(f"Processing {len(prompts)} requests...") start_time = time.perf_counter() outputs = llm.chat( prompts, sampling_params, use_tqdm=not getattr(args, "disable_tqdm", False) ) end_time = time.perf_counter() total_time = end_time - start_time mm_stats_by_stage = collect_mm_processor_stats( llm.llm_engine, ) if not any(mm_stats_by_stage.values()): print( "\n⚠️ Warning: No MM processor stats found in registry.\n" " This may indicate that:\n" " - No multimodal requests were processed\n" " - Stats were already retrieved (registry is cleared after retrieval)\n" ) mm_processor_metrics = calculate_mm_processor_metrics( mm_stats_by_stage, selected_percentiles ) completed = len([o for o in outputs if o.finished]) failed = len(outputs) - completed e2el_times = [] for output in outputs: if not output.finished or output.metrics is None: continue metrics = output.metrics for attr in ("finished_time", "last_token_time"): if ( getattr(metrics, attr, None) is not None and getattr(metrics, "arrival_time", None) is not None ): e2el_times.append( (getattr(metrics, attr) - metrics.arrival_time) * 1000 ) break if not e2el_times and completed > 0: avg_time_per_request = total_time / completed e2el_times = [avg_time_per_request * 1000] * completed if e2el_times: mean_e2el_ms = float(np.mean(e2el_times)) median_e2el_ms = float(np.median(e2el_times)) std_e2el_ms = float(np.std(e2el_times)) percentiles_e2el_ms = [ (p, float(np.percentile(e2el_times, p))) for p in selected_percentiles ] else: mean_e2el_ms = 0.0 median_e2el_ms = 0.0 std_e2el_ms = 0.0 percentiles_e2el_ms = [(p, 0.0) for p in selected_percentiles] benchmark_result = { "completed": completed, "failed": failed, "mean_e2el_ms": mean_e2el_ms, "median_e2el_ms": median_e2el_ms, "std_e2el_ms": std_e2el_ms, "percentiles_e2el_ms": percentiles_e2el_ms, "mm_processor_stats": mm_processor_metrics, } return benchmark_result def add_cli_args(parser: argparse.ArgumentParser) -> None: """Add CLI arguments for the multimodal processor benchmark.""" from vllm.engine.arg_utils import EngineArgs EngineArgs.add_cli_args(parser) parser.set_defaults(enable_mm_processor_stats=True) parser.add_argument( "--dataset-name", type=str, default="random-mm", choices=["random-mm", "random-rerank"], help="Name of the dataset to benchmark on. Defaults to 'random-mm'.", ) parser.add_argument( "--num-prompts", type=int, default=10, help="Number of prompts to process.", ) from vllm.benchmarks.datasets import ( add_random_dataset_base_args, add_random_multimodal_dataset_args, ) add_random_dataset_base_args(parser) add_random_multimodal_dataset_args(parser) parser.add_argument( "--output-json", type=str, default=None, help="Path to save the benchmark results in JSON format.", ) parser.add_argument( "--metric-percentiles", type=str, default="99", help="Comma-separated list of percentiles to calculate (e.g., '50,90,99').", ) parser.add_argument( "--disable-tqdm", action="store_true", help="Disable tqdm progress bar.", ) def main(args: argparse.Namespace) -> None: """Main entry point for the multimodal processor benchmark.""" print("Starting multimodal processor benchmark...") result = benchmark_multimodal_processor(args) print("\n" + "=" * 80) print("Multimodal Processor Benchmark Results") print("=" * 80) if "mm_processor_stats" in result: print("\nMM Processor Timing (ms):") selected_percentiles = [ float(p) for p in getattr(args, "metric_percentiles", "99").split(",") ] mm_data = [] for stage, metrics in result["mm_processor_stats"].items(): row = { "Stage": stage, "Mean": f"{metrics['mean']:.2f}", "Median": f"{metrics['median']:.2f}", "Std": f"{metrics['std']:.2f}", } for p in selected_percentiles: row[f"P{p}"] = f"{metrics.get(f'p{p}', 0.0):.2f}" mm_data.append(row) mm_df = pd.DataFrame(mm_data) print(mm_df.to_string(index=False)) if "mean_e2el_ms" in result: print("\nEnd-to-End Latency (ms):") selected_percentiles = [ float(p) for p in getattr(args, "metric_percentiles", "99").split(",") ] e2el_data = [ {"Metric": "Mean", "Value (ms)": f"{result['mean_e2el_ms']:.2f}"}, {"Metric": "Median", "Value (ms)": f"{result['median_e2el_ms']:.2f}"}, {"Metric": "Std", "Value (ms)": f"{result['std_e2el_ms']:.2f}"}, ] for p in selected_percentiles: percentile_value = next( (val for pct, val in result["percentiles_e2el_ms"] if pct == p), 0.0, ) e2el_data.append( { "Metric": f"P{p}", "Value (ms)": f"{percentile_value:.2f}", } ) e2el_df = pd.DataFrame(e2el_data) print(e2el_df.to_string(index=False)) if args.output_json: result["config"] = { "model": args.model, "num_prompts": args.num_prompts, "input_len": getattr(args, "random_input_len", None), "output_len": getattr(args, "random_output_len", None), } result["timestamp"] = datetime.now().isoformat() with open(args.output_json, "w") as f: json.dump(result, f, indent=2) print(f"\nResults saved to {args.output_json}") if __name__ == "__main__": parser = argparse.ArgumentParser(description="Benchmark mm processor latency") add_cli_args(parser) args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/latency.py
vllm/benchmarks/latency.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Benchmark the latency of processing a single batch of requests.""" import argparse import dataclasses import json import os import time from typing import Any import numpy as np from tqdm import tqdm from vllm.benchmarks.lib.utils import convert_to_pytorch_benchmark_format, write_to_json from vllm.engine.arg_utils import EngineArgs from vllm.inputs import PromptType from vllm.sampling_params import BeamSearchParams def save_to_pytorch_benchmark_format( args: argparse.Namespace, results: dict[str, Any] ) -> None: pt_records = convert_to_pytorch_benchmark_format( args=args, metrics={"latency": results["latencies"]}, extra_info={k: results[k] for k in ["avg_latency", "percentiles"]}, ) if pt_records: pt_file = f"{os.path.splitext(args.output_json)[0]}.pytorch.json" write_to_json(pt_file, pt_records) def add_cli_args(parser: argparse.ArgumentParser): parser.add_argument("--input-len", type=int, default=32) parser.add_argument("--output-len", type=int, default=128) parser.add_argument("--batch-size", type=int, default=8) parser.add_argument( "--n", type=int, default=1, help="Number of generated sequences per prompt.", ) parser.add_argument("--use-beam-search", action="store_true") parser.add_argument( "--num-iters-warmup", type=int, default=10, help="Number of iterations to run for warmup.", ) parser.add_argument( "--num-iters", type=int, default=30, help="Number of iterations to run." ) parser.add_argument( "--profile", action="store_true", help="profile the generation process of a single batch", ) parser.add_argument( "--output-json", type=str, default=None, help="Path to save the latency results in JSON format.", ) parser.add_argument( "--disable-detokenize", action="store_true", help=( "Do not detokenize responses (i.e. do not include " "detokenization time in the latency measurement)" ), ) parser = EngineArgs.add_cli_args(parser) # V1 enables prefix caching by default which skews the latency # numbers. We need to disable prefix caching by default. parser.set_defaults(enable_prefix_caching=False) def main(args: argparse.Namespace): engine_args = EngineArgs.from_cli_args(args) # Lazy import to avoid importing LLM when the bench command is not selected. from vllm import LLM, SamplingParams # NOTE(woosuk): If the request cannot be processed in a single batch, # the engine will automatically process the request in multiple batches. llm = LLM(**dataclasses.asdict(engine_args)) assert llm.llm_engine.model_config.max_model_len >= ( args.input_len + args.output_len ), ( "Please ensure that max_model_len is greater than" " the sum of input_len and output_len." ) sampling_params = SamplingParams( n=args.n, temperature=1.0, top_p=1.0, ignore_eos=True, max_tokens=args.output_len, detokenize=not args.disable_detokenize, ) dummy_prompt_token_ids = np.random.randint( 10000, size=(args.batch_size, args.input_len) ) dummy_prompts: list[PromptType] = [ {"prompt_token_ids": batch} for batch in dummy_prompt_token_ids.tolist() ] def llm_generate(): if not args.use_beam_search: llm.generate(dummy_prompts, sampling_params=sampling_params, use_tqdm=False) else: llm.beam_search( dummy_prompts, BeamSearchParams( beam_width=args.n, max_tokens=args.output_len, ignore_eos=True, ), ) def run_to_completion(do_profile: bool = False): if do_profile: llm.start_profile() llm_generate() llm.stop_profile() else: start_time = time.perf_counter() llm_generate() end_time = time.perf_counter() latency = end_time - start_time return latency print("Warming up...") for _ in tqdm(range(args.num_iters_warmup), desc="Warmup iterations"): run_to_completion(do_profile=False) if args.profile: profiler_config = engine_args.profiler_config if profiler_config.profiler == "torch": print( "Profiling with torch profiler (results will be saved to" f" {profiler_config.torch_profiler_dir})..." ) elif profiler_config.profiler == "cuda": print("Profiling with cuda profiler ...") run_to_completion(do_profile=True) return # Benchmark. latencies = [] for _ in tqdm(range(args.num_iters), desc="Bench iterations"): latencies.append(run_to_completion(do_profile=False)) latencies = np.array(latencies) percentages = [10, 25, 50, 75, 90, 99] percentiles = np.percentile(latencies, percentages) print(f"Avg latency: {np.mean(latencies)} seconds") for percentage, percentile in zip(percentages, percentiles): print(f"{percentage}% percentile latency: {percentile} seconds") # Output JSON results if specified if args.output_json: results = { "avg_latency": np.mean(latencies), "latencies": latencies.tolist(), "percentiles": dict(zip(percentages, percentiles.tolist())), } with open(args.output_json, "w") as f: json.dump(results, f, indent=4) save_to_pytorch_benchmark_format(args, results)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/datasets.py
vllm/benchmarks/datasets.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ This module defines a framework for sampling benchmark requests from various datasets. Each dataset subclass of BenchmarkDataset must implement sample generation. Supported dataset types include: - ShareGPT - Random (synthetic) - Sonnet - BurstGPT - HuggingFace - VisionArena """ import argparse import ast import base64 import io import json import logging import math import random from abc import ABC, abstractmethod from collections.abc import Callable, Iterator, Mapping from contextlib import suppress from copy import deepcopy from dataclasses import dataclass from functools import cache from io import BytesIO from tempfile import NamedTemporaryFile from typing import Any, cast import numpy as np from PIL import Image from typing_extensions import deprecated from vllm.lora.request import LoRARequest from vllm.lora.utils import get_adapter_absolute_path from vllm.multimodal import MultiModalDataDict from vllm.multimodal.image import convert_image_mode from vllm.tokenizers import TokenizerLike from vllm.utils.import_utils import PlaceholderModule try: from datasets import load_dataset except ImportError: datasets = PlaceholderModule("datasets") load_dataset = datasets.placeholder_attr("load_dataset") try: import pandas as pd except ImportError: pd = PlaceholderModule("pandas") try: import librosa except ImportError: librosa = PlaceholderModule("librosa") try: from vllm.utils.argparse_utils import FlexibleArgumentParser except ImportError: from argparse import ArgumentParser as FlexibleArgumentParser logger = logging.getLogger(__name__) # ----------------------------------------------------------------------------- # Data Classes # ----------------------------------------------------------------------------- @dataclass class SampleRequest: """ Represents a single inference request for benchmarking. """ prompt: str | list[str] prompt_len: int expected_output_len: int multi_modal_data: MultiModalDataDict | dict | list[dict] | None = None lora_request: LoRARequest | None = None request_id: str | None = None # ----------------------------------------------------------------------------- # Benchmark Dataset Base Class # ----------------------------------------------------------------------------- class BenchmarkDataset(ABC): DEFAULT_SEED = 0 IS_MULTIMODAL = False def __init__( self, dataset_path: str | None = None, random_seed: int = DEFAULT_SEED, disable_shuffle: bool = False, **kwargs, ) -> None: """ Initialize the BenchmarkDataset with an optional dataset path and random seed. Args: dataset_path (Optional[str]): Path to the dataset. If None, it indicates that a default or random dataset might be used. random_seed (int): Seed value for reproducible shuffling or sampling. Defaults to DEFAULT_SEED. """ self.dataset_path = dataset_path # Set the random seed, ensuring that a None value is replaced with the # default seed. self.random_seed = random_seed if random_seed is not None else self.DEFAULT_SEED self.disable_shuffle = disable_shuffle self.data = None def apply_multimodal_chat_transformation( self, prompt: str, mm_content: MultiModalDataDict | dict | list[dict] | None = None, ) -> list[dict]: """ Transform a prompt and optional multimodal content into a chat format. This method is used for chat models that expect a specific conversation format. """ content = [{"text": prompt, "type": "text"}] if mm_content is not None: if isinstance(mm_content, list): content.extend(cast(list[dict[str, Any]], mm_content)) elif isinstance(mm_content, dict): content.append(mm_content) else: raise TypeError( "Could not process multimodal content of type: " + f"{type(mm_content)}" ) return [{"role": "user", "content": content}] def load_data(self) -> None: """ Load data from the dataset path into self.data. This method must be overridden by subclasses since the method to load data will vary depending on the dataset format and source. Raises: NotImplementedError: If a subclass does not implement this method. """ # TODO (jenniferzhao): add support for downloading data raise NotImplementedError("load_data must be implemented in subclasses.") def get_random_lora_request( self, max_loras: int | None = None, lora_path: str | None = None, ) -> LoRARequest | None: """ Optionally select a random LoRA request. This method is used when LoRA parameters are provided. It randomly selects a LoRA based on max_loras. Args: max_loras (Optional[int]): The maximum number of LoRAs available. If `None`, LoRA is not used. lora_path (Optional[str]): Path to the LoRA parameters on disk. If `None`, LoRA is not used. Returns: A new [`LoRARequest`][vllm.lora.request.LoRARequest] (or `None` if not applicable). """ if max_loras is None or lora_path is None: return None # Generate a random LoRA ID in the range [1, max_loras]. lora_id = random.randint(1, max_loras) lora_request = LoRARequest( lora_name=str(lora_id), lora_int_id=lora_id, lora_path=lora_path_on_disk(lora_path), ) return lora_request @abstractmethod def sample( self, tokenizer: TokenizerLike, num_requests: int, request_id_prefix: str = "", no_oversample: bool = False, ) -> list[SampleRequest]: """ Abstract method to generate sample requests from the dataset. Subclasses must override this method to implement dataset-specific logic for generating a list of SampleRequest objects. Args: tokenizer (TokenizerLike): The tokenizer to be used for processing the dataset's text. num_requests (int): The number of sample requests to generate. request_id_prefix (str): The prefix of request_id. Returns: list[SampleRequest]: A list of sample requests generated from the dataset. """ raise NotImplementedError("sample must be implemented in subclasses.") def maybe_oversample_requests( self, requests: list[SampleRequest], num_requests: int, request_id_prefix: str = "", no_oversample: bool = False, ) -> None: """ Oversamples the list of requests if its size is less than the desired number. Args: requests (List[SampleRequest]): The current list of sampled requests. num_requests (int): The target number of requests. request_id_prefix (str): The prefix applied to generated request identifiers. """ if no_oversample: logger.info("Skipping oversampling. Total samples: %d.", len(requests)) return if len(requests) < num_requests: random.seed(self.random_seed) needed = num_requests - len(requests) additional = [] for i in range(needed): req = deepcopy(random.choice(requests)) req.request_id = request_id_prefix + str(len(requests) + i) additional.append(req) requests.extend(additional) logger.info("Oversampled requests to reach %d total samples.", num_requests) ids = [req.request_id for req in requests] if len(ids) != len(set(ids)): raise ValueError( "Duplicate request_id found in the sampled " "requests. Please ensure that each request_id " "is unique." ) # ----------------------------------------------------------------------------- # Utility Functions and Global Caches # ----------------------------------------------------------------------------- def is_valid_sequence( prompt_len: int, output_len: int, min_len: int = 4, max_prompt_len: int = 1024, max_total_len: int = 2048, skip_min_output_len_check: bool = False, ) -> bool: """ Validate a sequence based on prompt and output lengths. Default pruning criteria are copied from the original `sample_hf_requests` and `sample_sharegpt_requests` functions in benchmark_serving.py, as well as from `sample_requests` in benchmark_throughput.py. """ # Check for invalid conditions prompt_too_short = prompt_len < min_len output_too_short = (not skip_min_output_len_check) and (output_len < min_len) prompt_too_long = prompt_len > max_prompt_len combined_too_long = (prompt_len + output_len) > max_total_len # Return True if none of the invalid conditions are met return not ( prompt_too_short or output_too_short or prompt_too_long or combined_too_long ) @cache def lora_path_on_disk(lora_path: str) -> str: return get_adapter_absolute_path(lora_path) # Global cache for LoRA tokenizers. lora_tokenizer_cache: dict[int, TokenizerLike] = {} def process_image(image: Any) -> Mapping[str, Any]: """ Process a single image input and return a multimedia content dictionary. Supports the following input types: 1. Dictionary with raw image bytes: - Expects a dict with a 'bytes' key containing raw image data. - Loads the bytes as a PIL.Image.Image. 2. PIL.Image.Image input: - Converts the image to RGB. - Saves the image as a JPEG in memory. - Encodes the JPEG data as a base64 string. - Returns a dictionary with the image as a base64 data URL. 3. String input: - Treats the string as a URL or local file path. - Prepends "file://" if the string doesn't start with "http://" or "file://". - Returns a dictionary with the image URL. Raises: ValueError: If the input is not a supported type. """ if isinstance(image, dict) and "bytes" in image: image = Image.open(BytesIO(image["bytes"])) if isinstance(image, Image.Image): image = convert_image_mode(image, "RGB") with io.BytesIO() as image_data: image.save(image_data, format="JPEG") image_base64 = base64.b64encode(image_data.getvalue()).decode("utf-8") return { "type": "image_url", "image_url": {"url": f"data:image/jpeg;base64,{image_base64}"}, } if isinstance(image, str): image_url = ( image if image.startswith(("http://", "https://", "file://")) else f"file://{image}" ) return {"type": "image_url", "image_url": {"url": image_url}} raise ValueError( f"Invalid image input {image}. Must be a PIL.Image.Image" " or str or dictionary with raw image bytes." ) def process_video(video: Any) -> Mapping[str, Any]: """ Process a single video input and return a multimedia content dictionary. Supports the following input types: 1. Dictionary with raw video bytes: - Expects a dict with a 'bytes' key containing raw video data. 2. String input: - Treats the string as a URL or local file path. - Prepends "file://" if the string doesn't start with "http://" or "file://". - Returns a dictionary with the image URL. Raises: ValueError: If the input is not a supported type. """ if isinstance(video, dict) and "bytes" in video: video_bytes = video["bytes"] video_base64 = base64.b64encode(video_bytes).decode("utf-8") return { "type": "video_url", "video_url": {"url": f"data:video/mp4;base64,{video_base64}"}, } if isinstance(video, str): video_url = ( video if video.startswith(("http://", "https://", "file://")) else f"file://{video}" ) return {"type": "video_url", "video_url": {"url": video_url}} raise ValueError( f"Invalid video input {video}. Must be a string of local path/remote url, or a dictionary with raw video bytes in the form of `{{'bytes': raw_video_bytes}}`." # noqa: E501 ) def gen_prompt_decode_to_target_len( tokenizer: TokenizerLike, token_sequence: list[int], target_token_len: int, max_retry: int = 10, add_special_tokens: bool = False, rng: np.random.Generator | None = None, ) -> tuple[str, list[int]]: """ Ensure decoded-then-encoded prompt length matches the target token length. This function decodes an initial token sequence to text and re-encodes it , iteratively adjusting the token sequence length to match a target. This is necessary because some tokenizers do not guarantee a 1:1 mapping between consecutive tokens and the decoded-then-encoded sequence length. For example, for GPT2Tokenizer: [6880, 6881] -> ['Ġcalls', 'here'] -> [1650, 939, 486] -> ['Ġcall', 'sh', 'ere'] Returns a tuple of the final prompt string and the adjusted token sequence. """ remain_num_try = max_retry token_mismatch = 0 while True: prompt = tokenizer.decode(token_sequence) token_sequence = tokenizer.encode(prompt, add_special_tokens=add_special_tokens) if remain_num_try <= 0: if len(token_sequence) != target_token_len: token_mismatch = len(token_sequence) - target_token_len break if len(token_sequence) == target_token_len: break elif len(token_sequence) < target_token_len: if rng is not None: extra_tokens = rng.integers( 0, tokenizer.vocab_size, size=target_token_len - len(token_sequence), ).tolist() else: extra_tokens = np.random.randint( 0, tokenizer.vocab_size, size=target_token_len - len(token_sequence), ).tolist() token_sequence.extend(extra_tokens) elif len(token_sequence) > target_token_len: token_sequence = token_sequence[:target_token_len] remain_num_try -= 1 return prompt, token_sequence, token_mismatch # ----------------------------------------------------------------------------- # Random Dataset Implementation (Synthetic Data) # ----------------------------------------------------------------------------- class RandomDataset(BenchmarkDataset): """ Synthetic text-only dataset for serving/throughput benchmarks. Strategy: - Sample input/output token lengths per request from integer-uniform ranges around configured means (controlled by range_ratio). - Prepend a fixed random prefix of length prefix_len. - Generate the remaining tokens as a reproducible sequence: (offset + index + arange(input_len)) % vocab_size. - Decode then re-encode/truncate to ensure prompt token counts match. - Uses numpy.default_rng seeded with random_seed for reproducible sampling. """ # Default values copied from benchmark_serving.py for the random dataset. DEFAULT_PREFIX_LEN = 0 DEFAULT_RANGE_RATIO = 0.0 DEFAULT_INPUT_LEN = 1024 DEFAULT_OUTPUT_LEN = 128 def __init__(self, **kwargs) -> None: super().__init__(**kwargs) # Use numpy's default_rng for deterministic sampling # Do not use random.seed() or np.random.seed() elsewhere in this class. # This ensures that the RNG is isolated from global RNG state. self._rng = np.random.default_rng(self.random_seed) def sample( self, tokenizer: TokenizerLike, num_requests: int, request_id_prefix: str = "", no_oversample: bool = False, prefix_len: int = DEFAULT_PREFIX_LEN, range_ratio: float = DEFAULT_RANGE_RATIO, input_len: int = DEFAULT_INPUT_LEN, output_len: int = DEFAULT_OUTPUT_LEN, batchsize: int = 1, **kwargs, ) -> list[SampleRequest]: # validate total input tokens (prefix + sampled) is at least 1. num_special = int(tokenizer.num_special_tokens_to_add()) real_input_len = max(0, int(input_len) - num_special) min_sampled_input = math.floor(real_input_len * (1.0 - float(range_ratio))) min_total_input = int(prefix_len) + min_sampled_input if min_total_input < 1: raise ValueError( "--random-input-len is too small: with tokenizer special " f"tokens {num_special} and --random-range-ratio {range_ratio}, " "the minimum possible total input tokens (prefix + sampled) is " f"{min_total_input}. Increase --random-input-len and/or " "--random-prefix-len, or decrease --random-range-ratio so that " "prefix_len + floor(max(0, random_input_len - num_special)) " "* (1 - range_ratio) >= 1." ) input_lens, output_lens, offsets = self.get_sampling_params( num_requests, range_ratio, input_len, output_len, tokenizer ) vocab_size = tokenizer.vocab_size prohibited_tokens = tokenizer.all_special_ids all_tokens = np.arange(vocab_size) allowed_tokens = np.array(list(set(all_tokens) - set(prohibited_tokens))) # Generate prefix once prefix_token_ids = self.get_prefix(allowed_tokens, prefix_len) requests = [] token_mismatch_total = 0 for i in range(num_requests): prompt, total_input_len, token_mismatch = self.generate_token_sequence( # noqa: E501 tokenizer=tokenizer, prefix_token_ids=prefix_token_ids, prefix_len=prefix_len, vocab_size=vocab_size, input_len=int(input_lens[i]), offset=int(offsets[i]), index=i, allowed_tokens=allowed_tokens, ) token_mismatch_total += token_mismatch requests.append( SampleRequest( prompt=prompt, prompt_len=total_input_len, expected_output_len=int(output_lens[i]), request_id=request_id_prefix + str(i), ) ) # only used for embeddings benchmark. if batchsize > 1: batch_requests = [] # Create batched requests for i in range(0, num_requests, batchsize): batch = requests[i : i + batchsize] batch_requests.append( SampleRequest( prompt=[req.prompt for req in batch], prompt_len=sum(req.prompt_len for req in batch), expected_output_len=0, request_id=request_id_prefix + str(i // batchsize), ) ) requests = batch_requests if token_mismatch_total != 0: sign = "more" if token_mismatch_total > 0 else "fewer" logger.warning( "Across all generated prompts, there were %d %s tokens " "than expected after decoding and re-encoding. This is " "expected due to the imperfect nature of the sampling " "procedure.", abs(token_mismatch_total), sign, ) return requests def get_prefix( self, allowed_tokens: np.ndarray, prefix_len: int, ) -> list[int]: """ Get the prefix for the dataset. """ return ( allowed_tokens[ self._rng.integers(0, len(allowed_tokens), size=prefix_len) ].tolist() if prefix_len > 0 else [] ) def get_sampling_params( self, num_requests: int, range_ratio: float, input_len: int, output_len: int, tokenizer: TokenizerLike, ) -> tuple[np.ndarray, np.ndarray, np.ndarray]: """ Get the sampling parameters for the dataset. """ # Enforce range_ratio < 1 if not (0.0 <= range_ratio < 1.0): raise ValueError("range_ratio must be in [0, 1).") num_special_tokens = int(tokenizer.num_special_tokens_to_add()) real_input_len = max(0, int(input_len) - num_special_tokens) # Bounds use floor for low and ceil for high input_low = math.floor(real_input_len * (1 - range_ratio)) input_high = math.ceil(real_input_len * (1 + range_ratio)) output_low = math.floor(output_len * (1 - range_ratio)) output_high = math.ceil(output_len * (1 + range_ratio)) # Ensure the lower bound for output length is at least 1 to # prevent sampling 0 tokens. output_low = max(output_low, 1) output_high = max(output_high, 1) if input_low > input_high: raise ValueError( f"Invalid input sampling interval: low={input_low} > high={input_high}" ) if output_low > output_high: raise ValueError( "Invalid output sampling interval: " f"low={output_low} > high={output_high}" ) logger.info( "Sampling input_len from [%s, %s] and output_len from [%s, %s]", input_low, input_high, output_low, output_high, ) input_lens = self._rng.integers(input_low, input_high + 1, size=num_requests) output_lens = self._rng.integers(output_low, output_high + 1, size=num_requests) offsets = self._rng.integers(0, tokenizer.vocab_size, size=num_requests) return input_lens, output_lens, offsets def generate_token_sequence( self, *, tokenizer: TokenizerLike, prefix_token_ids: list[int], prefix_len: int, vocab_size: int, input_len: int, offset: int, index: int, allowed_tokens: np.ndarray, ) -> tuple[str, int, int]: """ Returns (prompt, total_input_len). NOTE: After decoding the prompt we have to encode and decode it again. This is done because in some cases N consecutive tokens give a string tokenized into != N number of tokens. For example for GPT2Tokenizer: [6880, 6881] -> ['Ġcalls', 'here'] -> [1650, 939, 486] -> ['Ġcall', 'sh', 'ere'] To avoid uncontrolled change of the prompt length, the encoded sequence is truncated before being decoded again. """ # Build the inner sequence by sampling # sequentially from the allowed tokens inner_seq = allowed_tokens[ (offset + index + np.arange(input_len)) % len(allowed_tokens) ].tolist() token_sequence = prefix_token_ids + inner_seq # Decode, then re-encode and truncate to preserve token count invariants total_input_len = prefix_len + int(input_len) prompt, adjusted_token_sequence, token_mismatch = ( gen_prompt_decode_to_target_len( tokenizer=tokenizer, token_sequence=token_sequence, target_token_len=total_input_len, add_special_tokens=False, rng=self._rng, ) ) total_input_len = len(adjusted_token_sequence) return prompt, total_input_len, token_mismatch # ----------------------------------------------------------------------------- # Random Dataset Implementation (Synthetic Data) # ----------------------------------------------------------------------------- class RandomDatasetForReranking(RandomDataset): """ Random dataset specialized for the needs of scoring: - Batches of inputs - Inputs composed of pairs """ def __init__(self, **kwargs) -> None: super().__init__(**kwargs) def sample( self, tokenizer: TokenizerLike, num_requests: int, request_id_prefix: str = "", range_ratio: float = RandomDataset.DEFAULT_RANGE_RATIO, input_len: int = RandomDataset.DEFAULT_INPUT_LEN, batchsize: int = 1, is_reranker: bool = True, **kwargs, ) -> list[SampleRequest]: n_sep_tokens = int(is_reranker) query_len_param = (input_len // 2) - n_sep_tokens if is_reranker else input_len query_lens, _, query_offsets = self.get_sampling_params( 1, range_ratio, query_len_param, 0, tokenizer ) query_len = int(query_lens[0]) if not is_reranker: assert num_requests > 1 and batchsize > 1 num_requests -= 1 batchsize -= 1 doc_len_param = input_len else: doc_len_param = input_len - query_len - n_sep_tokens doc_lens, _, doc_offsets = self.get_sampling_params( num_requests, range_ratio, doc_len_param, 0, tokenizer ) vocab_size = tokenizer.vocab_size prohibited_tokens = tokenizer.all_special_ids all_tokens = np.arange(vocab_size) allowed_tokens = np.array(list(set(all_tokens) - set(prohibited_tokens))) query_prompt, query_input_len, token_mismatch_total = ( self.generate_token_sequence( tokenizer=tokenizer, prefix_token_ids=[], prefix_len=0, vocab_size=vocab_size, input_len=query_len, offset=int(query_offsets[0]), index=0, allowed_tokens=allowed_tokens, ) ) requests = [] for i in range(num_requests): prompt, total_input_len, token_mismatch = self.generate_token_sequence( # noqa: E501 tokenizer=tokenizer, prefix_token_ids=[], prefix_len=0, vocab_size=vocab_size, input_len=int(doc_lens[i]), offset=int(doc_offsets[i]), index=i + 1, allowed_tokens=allowed_tokens, ) token_mismatch_total += token_mismatch requests.append((prompt, total_input_len)) batch_requests = [] # Create batched requests for i in range(0, num_requests, batchsize): batch = requests[i : i + batchsize] query_contrib = ( (query_input_len + n_sep_tokens) * len(batch) if is_reranker else query_input_len ) batch_requests.append( SampleRequest( prompt=[query_prompt] + [req[0] for req in batch], prompt_len=query_contrib + sum(req[1] for req in batch), expected_output_len=0, request_id=request_id_prefix + str(i // batchsize), ) ) if token_mismatch_total != 0: logger.warning( "Across all generated prompts, there were %d %s tokens " "than expected after decoding and re-encoding. This is " "expected due to the imperfect nature of the sampling " "procedure.", abs(token_mismatch_total), "more" if token_mismatch_total > 0 else "fewer", ) return batch_requests # ----------------------------------------------------------------------------- # MultiModalDataset Implementation # ----------------------------------------------------------------------------- class RandomMultiModalDataset(RandomDataset): """ Synthetic multimodal dataset (text + images) that extends RandomDataset. Status: - Images: supported via synthetic RGB data. - Video: supported via synthetic RGB data. - Audio: not yet supported. Sampling overview: 1) Number of items per request is sampled uniformly from the integer range [floor(n·(1−r)), ceil(n·(1+r))], where n is the base count and r is `num_mm_items_range_ratio` in [0, 1]. r=0 keeps it fixed; r=1 allows 0. The maximum is further clamped to the sum of per-modality limits. 2) Each item’s modality and shape is sampled from `bucket_config`, a dict mapping (height, width, num_frames) → probability. We treat `num_frames`=1 as image and `num_frames` > 1 as video. Entries with zero probability are removed and the rest are renormalized to sum to 1. 3) Per-modality hard caps are enforced via `limit_mm_per_prompt`. When a modality reaches its cap, all of its buckets are excluded and the remaining probabilities are renormalized. Example bucket configuration: {(256, 256, 1): 0.5, (720, 1280, 1): 0.4, (720, 1280, 16): 0.1} - Two image buckets (`num_frames`=1) and one video bucket (`num_frames`=16). OBS.: Only image sampling is supported for now. """ IS_MULTIMODAL = True DEFAULT_LIMIT_MM_PER_PROMPT = {"image": 255, "video": 1} DEFAULT_BASE_ITEMS_PER_REQUEST = 1 DEFAULT_NUM_MM_ITEMS_RANGE_RATIO = 0.0 DEFAULT_MM_ITEM_BUCKET_CONFIG = { (256, 256, 1): 0.5, (720, 1280, 1): 0.5, (720, 1280, 16): 0.0, } DEFAULT_ENABLE_MULTIMODAL_CHAT = False def __init__(self, **kwargs) -> None: super().__init__(**kwargs) def generate_synthetic_image(self, width: int, height: int) -> Image.Image: """Generate synthetic PIL image with random RGB values. NOTE: iid pixel sampling results in worst-case compression (good for stressing I/O), but very unlike real photos. We could consider a “low-freq” mode (e.g., noise blur) to emulate network realism instead of max stress. """ random_pixels = self._rng.integers( 0, 256, (height, width, 3), dtype=np.uint8, ) return Image.fromarray(random_pixels) def generate_synthetic_video( self, width: int, height: int, num_frames: int ) -> dict: """Generate synthetic video with random values. Creates a video with random pixel values, encodes it to MP4 format, and returns the content as bytes. """ import cv2 random_pixels = self._rng.integers( 0, 256, (num_frames, height, width, 3), dtype=np.uint8, ) # Create a temporary video file in memory fourcc = cv2.VideoWriter_fourcc(*"mp4v") fps = 30 # frames per second with NamedTemporaryFile(suffix=".mp4", delete_on_close=False) as temp_file: temp_path = temp_file.name # Create video writer video_writer = cv2.VideoWriter( temp_path, fourcc=fourcc, fps=fps, frameSize=(width, height) ) if not video_writer.isOpened(): raise RuntimeError("Failed to create video writer") for frame in random_pixels: video_writer.write(frame) video_writer.release() temp_file.close() # Read the video file content with open(temp_path, "rb") as f: video_content = f.read() return {"bytes": video_content} def map_config_to_modality(self, config: tuple[int, int, int]) -> str: """Map the configuration to the modality.""" if config[-1] == 1: return "image" elif config[-1] > 1: return "video" else: raise ValueError(f"Invalid multimodal item configuration: {config}") def normalize_bucket_config( self, bucket_config: dict[tuple[int, int, int], float] ) -> dict[tuple[int, int, int], float]: """ Remove zero probability entries and normalize the bucket config to sum to 1. """ # Raise error if value is negative if any(v < 0 for v in bucket_config.values()):
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/startup.py
vllm/benchmarks/startup.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Benchmark the cold and warm startup time of vLLM models. This script measures total startup time (including model loading, compilation, and cache operations) for both cold and warm scenarios: - Cold startup: Fresh start with no caches (temporary cache directories) - Warm startup: Using cached compilation and model info """ import argparse import dataclasses import json import multiprocessing import os import shutil import tempfile import time from contextlib import contextmanager from typing import Any import numpy as np from tqdm import tqdm from vllm.benchmarks.lib.utils import ( convert_to_pytorch_benchmark_format, write_to_json, ) from vllm.engine.arg_utils import EngineArgs @contextmanager def cold_startup(): """ Context manager to measure cold startup time: 1. Uses a temporary directory for vLLM cache to avoid any pollution between cold startup iterations. 2. Uses inductor's fresh_cache to clear torch.compile caches. """ from torch._inductor.utils import fresh_cache # Use temporary directory for caching to avoid any pollution between cold startups original_cache_root = os.environ.get("VLLM_CACHE_ROOT") temp_cache_dir = tempfile.mkdtemp(prefix="vllm_startup_bench_cold_") try: os.environ["VLLM_CACHE_ROOT"] = temp_cache_dir with fresh_cache(): yield finally: # Clean up temporary cache directory shutil.rmtree(temp_cache_dir, ignore_errors=True) if original_cache_root: os.environ["VLLM_CACHE_ROOT"] = original_cache_root else: os.environ.pop("VLLM_CACHE_ROOT", None) def run_startup_in_subprocess(engine_args_dict, result_queue): """ Run LLM startup in a subprocess and return timing metrics via a queue. This ensures complete isolation between iterations. """ try: # Import inside the subprocess to avoid issues with forking from vllm import LLM from vllm.engine.arg_utils import EngineArgs engine_args = EngineArgs(**engine_args_dict) # Measure total startup time start_time = time.perf_counter() llm = LLM(**dataclasses.asdict(engine_args)) total_startup_time = time.perf_counter() - start_time # Extract compilation time if available compilation_time = 0.0 if hasattr(llm.llm_engine, "vllm_config"): vllm_config = llm.llm_engine.vllm_config if ( hasattr(vllm_config, "compilation_config") and vllm_config.compilation_config is not None ): compilation_time = vllm_config.compilation_config.compilation_time result_queue.put( { "total_startup_time": total_startup_time, "compilation_time": compilation_time, } ) except Exception as e: result_queue.put(None) result_queue.put(str(e)) def save_to_pytorch_benchmark_format( args: argparse.Namespace, results: dict[str, Any] ) -> None: base_name = os.path.splitext(args.output_json)[0] cold_startup_records = convert_to_pytorch_benchmark_format( args=args, metrics={ "avg_cold_startup_time": results["avg_cold_startup_time"], }, extra_info={ "cold_startup_times": results["cold_startup_times"], "cold_startup_percentiles": results["cold_startup_percentiles"], }, ) if cold_startup_records: write_to_json(f"{base_name}.cold_startup.pytorch.json", cold_startup_records) cold_compilation_records = convert_to_pytorch_benchmark_format( args=args, metrics={ "avg_cold_compilation_time": results["avg_cold_compilation_time"], }, extra_info={ "cold_compilation_times": results["cold_compilation_times"], "cold_compilation_percentiles": results["cold_compilation_percentiles"], }, ) if cold_compilation_records: write_to_json( f"{base_name}.cold_compilation.pytorch.json", cold_compilation_records ) warm_startup_records = convert_to_pytorch_benchmark_format( args=args, metrics={ "avg_warm_startup_time": results["avg_warm_startup_time"], }, extra_info={ "warm_startup_times": results["warm_startup_times"], "warm_startup_percentiles": results["warm_startup_percentiles"], }, ) if warm_startup_records: write_to_json(f"{base_name}.warm_startup.pytorch.json", warm_startup_records) warm_compilation_records = convert_to_pytorch_benchmark_format( args=args, metrics={ "avg_warm_compilation_time": results["avg_warm_compilation_time"], }, extra_info={ "warm_compilation_times": results["warm_compilation_times"], "warm_compilation_percentiles": results["warm_compilation_percentiles"], }, ) if warm_compilation_records: write_to_json( f"{base_name}.warm_compilation.pytorch.json", warm_compilation_records ) def add_cli_args(parser: argparse.ArgumentParser): parser.add_argument( "--num-iters-cold", type=int, default=5, help="Number of cold startup iterations.", ) parser.add_argument( "--num-iters-warmup", type=int, default=3, help="Number of warmup iterations before benchmarking warm startups.", ) parser.add_argument( "--num-iters-warm", type=int, default=5, help="Number of warm startup iterations.", ) parser.add_argument( "--output-json", type=str, default=None, help="Path to save the startup time results in JSON format.", ) parser = EngineArgs.add_cli_args(parser) return parser def main(args: argparse.Namespace): # Set multiprocessing start method to 'spawn' for clean process isolation # This ensures each subprocess starts fresh without inheriting state multiprocessing.set_start_method("spawn", force=True) engine_args = EngineArgs.from_cli_args(args) def create_llm_and_measure_startup(): """ Create LLM instance in a subprocess and measure startup time. Returns timing metrics, using subprocess for complete isolation. """ # Convert engine_args to dictionary for pickling engine_args_dict = dataclasses.asdict(engine_args) # Create a queue for inter-process communication result_queue = multiprocessing.Queue() process = multiprocessing.Process( target=run_startup_in_subprocess, args=( engine_args_dict, result_queue, ), ) process.start() process.join() if not result_queue.empty(): result = result_queue.get() if result is None: if not result_queue.empty(): error_msg = result_queue.get() raise RuntimeError(f"Subprocess failed: {error_msg}") else: raise RuntimeError("Subprocess failed with unknown error") return result else: raise RuntimeError("Subprocess did not return a result") os.environ["VLLM_ENABLE_V1_MULTIPROCESSING"] = "0" print("Setting VLLM_ENABLE_V1_MULTIPROCESSING=0 to collect startup metrics.\n") print("Measuring cold startup time...\n") cold_startup_times = [] cold_compilation_times = [] for i in tqdm(range(args.num_iters_cold), desc="Cold startup iterations"): with cold_startup(): metrics = create_llm_and_measure_startup() cold_startup_times.append(metrics["total_startup_time"]) cold_compilation_times.append(metrics["compilation_time"]) # Warmup for warm startup print("\nWarming up for warm startup measurement...\n") for _ in tqdm(range(args.num_iters_warmup), desc="Warmup iterations"): create_llm_and_measure_startup() print("\nMeasuring warm startup time...\n") warm_startup_times = [] warm_compilation_times = [] for i in tqdm(range(args.num_iters_warm), desc="Warm startup iterations"): metrics = create_llm_and_measure_startup() warm_startup_times.append(metrics["total_startup_time"]) warm_compilation_times.append(metrics["compilation_time"]) # Calculate statistics cold_startup_array = np.array(cold_startup_times) cold_compilation_array = np.array(cold_compilation_times) warm_startup_array = np.array(warm_startup_times) warm_compilation_array = np.array(warm_compilation_times) avg_cold_startup = np.mean(cold_startup_array) avg_cold_compilation = np.mean(cold_compilation_array) avg_warm_startup = np.mean(warm_startup_array) avg_warm_compilation = np.mean(warm_compilation_array) percentages = [10, 25, 50, 75, 90, 99] cold_startup_percentiles = np.percentile(cold_startup_array, percentages) cold_compilation_percentiles = np.percentile(cold_compilation_array, percentages) warm_startup_percentiles = np.percentile(warm_startup_array, percentages) warm_compilation_percentiles = np.percentile(warm_compilation_array, percentages) print("\n" + "=" * 60) print("STARTUP TIME BENCHMARK RESULTS") print("=" * 60) # Cold startup statistics print("\nCOLD STARTUP:") print(f"Avg total startup time: {avg_cold_startup:.2f} seconds") print(f"Avg compilation time: {avg_cold_compilation:.2f} seconds") print("Startup time percentiles:") for percentage, percentile in zip(percentages, cold_startup_percentiles): print(f" {percentage}%: {percentile:.2f} seconds") print("Compilation time percentiles:") for percentage, percentile in zip(percentages, cold_compilation_percentiles): print(f" {percentage}%: {percentile:.2f} seconds") # Warm startup statistics print("\nWARM STARTUP:") print(f"Avg total startup time: {avg_warm_startup:.2f} seconds") print(f"Avg compilation time: {avg_warm_compilation:.2f} seconds") print("Startup time percentiles:") for percentage, percentile in zip(percentages, warm_startup_percentiles): print(f" {percentage}%: {percentile:.2f} seconds") print("Compilation time percentiles:") for percentage, percentile in zip(percentages, warm_compilation_percentiles): print(f" {percentage}%: {percentile:.2f} seconds") print("=" * 60) # Output JSON results if specified if args.output_json: results = { "avg_cold_startup_time": float(avg_cold_startup), "avg_cold_compilation_time": float(avg_cold_compilation), "cold_startup_times": cold_startup_times, "cold_compilation_times": cold_compilation_times, "cold_startup_percentiles": dict( zip(percentages, cold_startup_percentiles.tolist()) ), "cold_compilation_percentiles": dict( zip(percentages, cold_compilation_percentiles.tolist()) ), "avg_warm_startup_time": float(avg_warm_startup), "avg_warm_compilation_time": float(avg_warm_compilation), "warm_startup_times": warm_startup_times, "warm_compilation_times": warm_compilation_times, "warm_startup_percentiles": dict( zip(percentages, warm_startup_percentiles.tolist()) ), "warm_compilation_percentiles": dict( zip(percentages, warm_compilation_percentiles.tolist()) ), } with open(args.output_json, "w") as f: json.dump(results, f, indent=4) save_to_pytorch_benchmark_format(args, results)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/__init__.py
vllm/benchmarks/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/serve.py
vllm/benchmarks/serve.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project r"""Benchmark online serving throughput. On the server side, run one of the following commands to launch the vLLM OpenAI API server: vllm serve <your_model> <engine arguments> On the client side, run: vllm bench serve \ --backend <backend or endpoint type. Default 'openai'> \ --label <benchmark result label. Default using backend> \ --model <your_model. Optional, defaults to first model from server> \ --dataset-name <dataset_name. Default 'random'> \ --input-len <general input length. Optional, maps to dataset-specific args> \ --output-len <general output length. Optional, maps to dataset-specific args> \ --request-rate <request_rate. Default inf> \ --num-prompts <num_prompts. Default 1000> """ import argparse import asyncio import contextlib import importlib.util import json import os import random import shutil import time import uuid import warnings from collections.abc import AsyncGenerator, Iterable from dataclasses import dataclass from datetime import datetime from enum import Enum from typing import Any, Literal import aiohttp import numpy as np from tqdm.asyncio import tqdm from vllm.benchmarks.datasets import SampleRequest, add_dataset_parser, get_samples from vllm.benchmarks.lib.endpoint_request_func import ( ASYNC_REQUEST_FUNCS, OPENAI_COMPATIBLE_BACKENDS, RequestFuncInput, RequestFuncOutput, ) from vllm.benchmarks.lib.ready_checker import wait_for_endpoint from vllm.benchmarks.lib.utils import convert_to_pytorch_benchmark_format, write_to_json from vllm.tokenizers import TokenizerLike, get_tokenizer from vllm.utils.gc_utils import freeze_gc_heap from vllm.utils.network_utils import join_host_port MILLISECONDS_TO_SECONDS_CONVERSION = 1000 TERM_PLOTLIB_AVAILABLE = (importlib.util.find_spec("termplotlib") is not None) and ( shutil.which("gnuplot") is not None ) async def get_first_model_from_server( base_url: str, headers: dict | None = None ) -> tuple[str, str]: """Fetch the first model from the server's /v1/models endpoint.""" models_url = f"{base_url}/v1/models" async with aiohttp.ClientSession() as session: try: async with session.get(models_url, headers=headers) as response: response.raise_for_status() data = await response.json() if "data" in data and len(data["data"]) > 0: return data["data"][0]["id"], data["data"][0]["root"] else: raise ValueError( f"No models found on the server at {base_url}. " "Make sure the server is running and has models loaded." ) except (aiohttp.ClientError, json.JSONDecodeError) as e: raise RuntimeError( f"Failed to fetch models from server at {models_url}. " "Check that:\n" "1. The server is running\n" "2. The server URL is correct\n" f"Error: {e}" ) from e class TaskType(Enum): GENERATION = "generation" POOLING = "pooling" @dataclass class BenchmarkMetrics: completed: int failed: int total_input: int total_output: int request_throughput: float request_goodput: float output_throughput: float total_token_throughput: float mean_ttft_ms: float median_ttft_ms: float std_ttft_ms: float percentiles_ttft_ms: list[tuple[float, float]] mean_tpot_ms: float median_tpot_ms: float std_tpot_ms: float percentiles_tpot_ms: list[tuple[float, float]] mean_itl_ms: float median_itl_ms: float std_itl_ms: float percentiles_itl_ms: list[tuple[float, float]] # E2EL stands for end-to-end latency per request. # It is the time taken on the client side from sending # a request to receiving a complete response. mean_e2el_ms: float median_e2el_ms: float std_e2el_ms: float percentiles_e2el_ms: list[tuple[float, float]] # Max output tokens per second and concurrent requests at that peak max_output_tokens_per_s: float max_concurrent_requests: int @dataclass class EmbedBenchmarkMetrics: completed: int failed: int total_input: int request_throughput: float total_token_throughput: float mean_e2el_ms: float std_e2el_ms: float median_e2el_ms: float percentiles_e2el_ms: float def _get_current_request_rate( ramp_up_strategy: Literal["linear", "exponential"] | None, ramp_up_start_rps: int | None, ramp_up_end_rps: int | None, request_index: int, total_requests: int, request_rate: float, ) -> float: if ( ramp_up_strategy and ramp_up_start_rps is not None and ramp_up_end_rps is not None ): progress = request_index / max(total_requests - 1, 1) if ramp_up_strategy == "linear": increase = (ramp_up_end_rps - ramp_up_start_rps) * progress return ramp_up_start_rps + increase elif ramp_up_strategy == "exponential": ratio = ramp_up_end_rps / ramp_up_start_rps return ramp_up_start_rps * (ratio**progress) else: raise ValueError(f"Unknown ramp-up strategy: {ramp_up_strategy}") return request_rate async def get_request( input_requests: list[SampleRequest], request_rate: float, burstiness: float = 1.0, ramp_up_strategy: Literal["linear", "exponential"] | None = None, ramp_up_start_rps: int | None = None, ramp_up_end_rps: int | None = None, ) -> AsyncGenerator[tuple[SampleRequest, float], None]: """ Asynchronously generates requests at a specified rate with OPTIONAL burstiness and OPTIONAL ramp-up strategy. Args: input_requests: A list of input requests, each represented as a SampleRequest. request_rate: The rate at which requests are generated (requests/s). burstiness (optional): The burstiness factor of the request generation. Only takes effect when request_rate is not inf. Default value is 1, which follows a Poisson process. Otherwise, the request intervals follow a gamma distribution. A lower burstiness value (0 < burstiness < 1) results in more bursty requests, while a higher burstiness value (burstiness > 1) results in a more uniform arrival of requests. ramp_up_strategy (optional): The ramp-up strategy. Can be "linear" or "exponential". If None, uses constant request rate (specified by request_rate). ramp_up_start_rps (optional): The starting request rate for ramp-up. ramp_up_end_rps (optional): The ending request rate for ramp-up. """ assert burstiness > 0, ( f"A positive burstiness factor is expected, but given {burstiness}." ) # Convert to list to get length for ramp-up calculations if isinstance(input_requests, Iterable) and not isinstance(input_requests, list): input_requests = list(input_requests) total_requests = len(input_requests) assert total_requests > 0, "No requests provided." # Precompute delays among requests to minimize request send laggings request_rates = [] delay_ts = [] for request_index, request in enumerate(input_requests): current_request_rate = _get_current_request_rate( ramp_up_strategy, ramp_up_start_rps, ramp_up_end_rps, request_index, total_requests, request_rate, ) assert current_request_rate > 0.0, ( f"Obtained non-positive request rate {current_request_rate}." ) request_rates.append(current_request_rate) if current_request_rate == float("inf"): delay_ts.append(0) elif burstiness == float("inf"): # when burstiness tends to infinity, the delay time becomes constant # and tends to the inverse of the request rate delay_ts.append(1.0 / current_request_rate) else: theta = 1.0 / (current_request_rate * burstiness) # Sample the request interval from the gamma distribution. # If burstiness is 1, it follows exponential distribution. delay_ts.append(np.random.gamma(shape=burstiness, scale=theta)) # Calculate the cumulative delay time from the first sent out requests. for i in range(1, len(delay_ts)): delay_ts[i] += delay_ts[i - 1] if ramp_up_strategy is None and delay_ts[-1] != 0: # When ramp_up_strategy is not set, we assume the request rate is fixed # and all requests should be sent in target_total_delay_s, the following # logic would re-scale delay time to ensure the final delay_ts # align with target_total_delay_s. # # NOTE: If we simply accumulate the random delta values # from the gamma distribution, their sum would have 1-2% gap # from target_total_delay_s. The purpose of the following logic is to # close the gap for stabilizing the throughput data # from different random seeds. target_total_delay_s = total_requests / request_rate normalize_factor = target_total_delay_s / delay_ts[-1] delay_ts = [delay * normalize_factor for delay in delay_ts] start_ts = time.time() for request_index, request in enumerate(input_requests): if delay_ts[request_index] > 0: current_ts = time.time() sleep_interval_s = start_ts + delay_ts[request_index] - current_ts if sleep_interval_s > 0: await asyncio.sleep(sleep_interval_s) yield request, request_rates[request_index] def calculate_metrics_for_embeddings( outputs: list[RequestFuncOutput], dur_s: float, selected_percentiles: list[float], ) -> EmbedBenchmarkMetrics: """Calculate the metrics for the embedding requests. Args: outputs: The outputs of the requests. dur_s: The duration of the benchmark. selected_percentiles: The percentiles to select. Returns: The calculated benchmark metrics. """ total_input = 0 completed = 0 failed = 0 e2els: list[float] = [] for i in range(len(outputs)): if outputs[i].success: e2els.append(outputs[i].latency) completed += 1 total_input += outputs[i].prompt_len else: failed += 1 if completed == 0: warnings.warn( "All requests failed. This is likely due to a misconfiguration " "on the benchmark arguments.", stacklevel=2, ) metrics = EmbedBenchmarkMetrics( completed=completed, failed=failed, total_input=total_input, request_throughput=completed / dur_s, total_token_throughput=total_input / dur_s, mean_e2el_ms=np.mean(e2els or 0) * 1000, std_e2el_ms=np.std(e2els or 0) * 1000, median_e2el_ms=np.median(e2els or 0) * 1000, percentiles_e2el_ms=[ (p, np.percentile(e2els or 0, p) * 1000) for p in selected_percentiles ], ) return metrics def calculate_metrics( input_requests: list[SampleRequest], outputs: list[RequestFuncOutput], dur_s: float, tokenizer: TokenizerLike, selected_percentiles: list[float], goodput_config_dict: dict[str, float], ) -> tuple[BenchmarkMetrics, list[int]]: """Calculate the metrics for the benchmark. Args: input_requests: The input requests. outputs: The outputs of the requests. dur_s: The duration of the benchmark. tokenizer: The tokenizer to use. selected_percentiles: The percentiles to select. goodput_config_dict: The goodput configuration. Returns: A tuple of the benchmark metrics and the actual output lengths. """ actual_output_lens: list[int] = [] total_input = 0 completed = 0 good_completed = 0 itls: list[float] = [] tpots: list[float] = [] all_tpots: list[float] = [] ttfts: list[float] = [] e2els: list[float] = [] for i in range(len(outputs)): if outputs[i].success: output_len = outputs[i].output_tokens if not output_len: # We use the tokenizer to count the number of output tokens # for some serving backends instead of looking at # len(outputs[i].itl) since multiple output tokens may be # bundled together # Note : this may inflate the output token count slightly output_len = len( tokenizer( outputs[i].generated_text, add_special_tokens=False ).input_ids ) actual_output_lens.append(output_len) total_input += input_requests[i].prompt_len tpot = 0 if output_len > 1: latency_minus_ttft = outputs[i].latency - outputs[i].ttft tpot = latency_minus_ttft / (output_len - 1) tpots.append(tpot) # Note: if output_len <= 1, we regard tpot as 0 for goodput all_tpots.append(tpot) itls += outputs[i].itl ttfts.append(outputs[i].ttft) e2els.append(outputs[i].latency) completed += 1 else: actual_output_lens.append(0) if goodput_config_dict: valid_metrics = [] slo_values = [] if "ttft" in goodput_config_dict: valid_metrics.append(ttfts) slo_values.append( goodput_config_dict["ttft"] / MILLISECONDS_TO_SECONDS_CONVERSION ) if "tpot" in goodput_config_dict: valid_metrics.append(all_tpots) slo_values.append( goodput_config_dict["tpot"] / MILLISECONDS_TO_SECONDS_CONVERSION ) if "e2el" in goodput_config_dict: valid_metrics.append(e2els) slo_values.append( goodput_config_dict["e2el"] / MILLISECONDS_TO_SECONDS_CONVERSION ) for req_metric in zip(*valid_metrics): is_good_req = all([s >= r for s, r in zip(slo_values, req_metric)]) if is_good_req: good_completed += 1 if completed == 0: warnings.warn( "All requests failed. This is likely due to a misconfiguration " "on the benchmark arguments.", stacklevel=2, ) # Calculate max output tokens per second metric max_output_tokens_per_s = 0.0 max_concurrent_requests = 0 # Find the time range across all successful requests successful_outputs = [output for output in outputs if output.success] failed_outputs = [output for output in outputs if not output.success] if successful_outputs: min_start_time = min(output.start_time for output in successful_outputs) max_end_time = max( output.start_time + output.latency for output in successful_outputs ) # Create second buckets (ceiling to ensure we capture all time) duration_seconds = int(np.ceil(max_end_time - min_start_time)) + 1 tokens_per_second = np.zeros(duration_seconds) concurrent_requests_per_second = np.zeros(duration_seconds) for i, output in enumerate(successful_outputs): # Calculate token generation timestamp using # start_time, ttft, and itl token_times = [output.start_time + output.ttft] current_time = token_times[0] for itl_value in output.itl: current_time += itl_value token_times.append(current_time) # Add tokens to second buckets for token_time in token_times: second_bucket = int(token_time - min_start_time) if 0 <= second_bucket < duration_seconds: tokens_per_second[second_bucket] += 1 # Track concurrent requests for each second this request was active request_start_second = int(output.start_time - min_start_time) request_end_second = int( (output.start_time + output.latency) - min_start_time ) for second in range(request_start_second, request_end_second + 1): concurrent_requests_per_second[second] += 1 # Find the maximum tokens per second and corresponding # concurrent requests if len(tokens_per_second) > 0: max_output_tokens_per_s = float(np.max(tokens_per_second)) max_concurrent_requests = int(np.max(concurrent_requests_per_second)) if TERM_PLOTLIB_AVAILABLE: import termplotlib as tpl fig = tpl.figure() fig.plot( np.arange(len(tokens_per_second)), tokens_per_second, title="Output tokens per second", ) fig.plot( np.arange(len(concurrent_requests_per_second)), concurrent_requests_per_second, title="Concurrent requests per second", ) fig.show() else: print("tip: install termplotlib and gnuplot to plot the metrics") metrics = BenchmarkMetrics( completed=completed, failed=len(failed_outputs), total_input=total_input, total_output=sum(actual_output_lens), request_throughput=completed / dur_s, request_goodput=good_completed / dur_s, output_throughput=sum(actual_output_lens) / dur_s, total_token_throughput=(total_input + sum(actual_output_lens)) / dur_s, mean_ttft_ms=np.mean(ttfts or 0) * 1000, # ttfts is empty if streaming is not supported by the endpoint std_ttft_ms=np.std(ttfts or 0) * 1000, median_ttft_ms=np.median(ttfts or 0) * 1000, percentiles_ttft_ms=[ (p, np.percentile(ttfts or 0, p) * 1000) for p in selected_percentiles ], mean_tpot_ms=np.mean(tpots or 0) * 1000, std_tpot_ms=np.std(tpots or 0) * 1000, median_tpot_ms=np.median(tpots or 0) * 1000, percentiles_tpot_ms=[ (p, np.percentile(tpots or 0, p) * 1000) for p in selected_percentiles ], mean_itl_ms=np.mean(itls or 0) * 1000, std_itl_ms=np.std(itls or 0) * 1000, median_itl_ms=np.median(itls or 0) * 1000, percentiles_itl_ms=[ (p, np.percentile(itls or 0, p) * 1000) for p in selected_percentiles ], mean_e2el_ms=np.mean(e2els or 0) * 1000, std_e2el_ms=np.std(e2els or 0) * 1000, median_e2el_ms=np.median(e2els or 0) * 1000, percentiles_e2el_ms=[ (p, np.percentile(e2els or 0, p) * 1000) for p in selected_percentiles ], max_output_tokens_per_s=max_output_tokens_per_s, max_concurrent_requests=max_concurrent_requests, ) return metrics, actual_output_lens async def benchmark( task_type: TaskType, endpoint_type: str, api_url: str, base_url: str, model_id: str, model_name: str, tokenizer: TokenizerLike, input_requests: list[SampleRequest], logprobs: int | None, request_rate: float, burstiness: float, disable_tqdm: bool, num_warmups: int, profile: bool, selected_percentile_metrics: list[str], selected_percentiles: list[float], ignore_eos: bool, goodput_config_dict: dict[str, float], max_concurrency: int | None, lora_modules: Iterable[str] | None, extra_headers: dict | None, extra_body: dict | None, ramp_up_strategy: Literal["linear", "exponential"] | None = None, ramp_up_start_rps: int | None = None, ramp_up_end_rps: int | None = None, ready_check_timeout_sec: int = 600, ): try: request_func = ASYNC_REQUEST_FUNCS[endpoint_type] except KeyError: raise ValueError(f"Unknown backend: {endpoint_type}") from None # Reuses connections across requests to reduce TLS handshake overhead. connector = aiohttp.TCPConnector( limit=max_concurrency or 0, limit_per_host=max_concurrency or 0, ttl_dns_cache=300, use_dns_cache=True, keepalive_timeout=60, enable_cleanup_closed=True, force_close=False, ssl=("https://" in api_url), ) session = aiohttp.ClientSession( connector=connector, trust_env=True, timeout=aiohttp.ClientTimeout(total=6 * 60 * 60), ) print("Starting initial single prompt test run...") test_prompt, test_prompt_len, test_output_len, test_mm_content = ( input_requests[0].prompt, input_requests[0].prompt_len, input_requests[0].expected_output_len, input_requests[0].multi_modal_data, ) assert ( test_mm_content is None or isinstance(test_mm_content, dict) or ( isinstance(test_mm_content, list) and all(isinstance(item, dict) for item in test_mm_content) ) ), "multi_modal_data must be a dict or list[dict]" test_input = RequestFuncInput( model=model_id, model_name=model_name, prompt=test_prompt, api_url=api_url, prompt_len=test_prompt_len, output_len=test_output_len, logprobs=logprobs, multi_modal_content=test_mm_content, ignore_eos=ignore_eos, extra_headers=extra_headers, extra_body=extra_body, ) if ready_check_timeout_sec > 0: test_output = await wait_for_endpoint( request_func, test_input, session, timeout_seconds=ready_check_timeout_sec, ) if not test_output.success: raise ValueError( "Initial test run failed - Please make sure benchmark " "arguments are correctly specified. " f"Error: {test_output.error}" ) else: print("Initial test run completed.") else: print("Skipping endpoint ready check.") if num_warmups > 0: print(f"Warming up with {num_warmups} requests...") warmup_pbar = None if disable_tqdm else tqdm(total=num_warmups) warmup_semaphore = ( asyncio.Semaphore(max_concurrency) if max_concurrency else contextlib.nullcontext() ) warmup_tasks = [] async def warmup_limited_request_func(): async with warmup_semaphore: return await request_func( request_func_input=test_input, session=session, pbar=warmup_pbar ) for _ in range(num_warmups): request_task = asyncio.create_task(warmup_limited_request_func()) warmup_tasks.append(request_task) _ = await asyncio.gather(*warmup_tasks) if warmup_pbar is not None: warmup_pbar.close() print("Warmup run completed.") print("Starting main benchmark run...") if lora_modules: # For each input request, choose a LoRA module at random. lora_modules = iter( [random.choice(lora_modules) for _ in range(len(input_requests))] ) if profile: print("Starting profiler...") profile_input = RequestFuncInput( model=model_id, model_name=model_name, prompt=test_prompt, api_url=base_url + "/start_profile", prompt_len=test_prompt_len, output_len=test_output_len, logprobs=logprobs, multi_modal_content=test_mm_content, ignore_eos=ignore_eos, extra_headers=extra_headers, extra_body=extra_body, ) profile_output = await request_func( request_func_input=profile_input, session=session ) if profile_output.success: print("Profiler started") distribution = "Poisson process" if burstiness == 1.0 else "Gamma distribution" if ramp_up_strategy is not None: print(f"Traffic ramp-up strategy: {ramp_up_strategy}.") print( f"Will increase RPS from {ramp_up_start_rps} to " f"{ramp_up_end_rps} RPS over the duration of the benchmark." ) else: print(f"Traffic request rate: {request_rate}") print(f"Burstiness factor: {burstiness} ({distribution})") print(f"Maximum request concurrency: {max_concurrency}") pbar = None if disable_tqdm else tqdm(total=len(input_requests)) semaphore = ( asyncio.Semaphore(max_concurrency) if max_concurrency else contextlib.nullcontext() ) async def limited_request_func(request_func_input, session, pbar): async with semaphore: return await request_func( request_func_input=request_func_input, session=session, pbar=pbar ) benchmark_start_time = time.perf_counter() tasks: list[asyncio.Task] = [] rps_change_events = [] last_int_rps = -1 if ramp_up_strategy is not None and ramp_up_start_rps is not None: last_int_rps = ramp_up_start_rps rps_change_events.append( { "rps": last_int_rps, "timestamp": datetime.now().isoformat(), } ) async for request, current_request_rate in get_request( input_requests, request_rate, burstiness, ramp_up_strategy, ramp_up_start_rps, ramp_up_end_rps, ): if ramp_up_strategy is not None: current_int_rps = int(current_request_rate) if current_int_rps > last_int_rps: timestamp = datetime.now().isoformat() for rps_val in range(last_int_rps + 1, current_int_rps + 1): rps_change_events.append({"rps": rps_val, "timestamp": timestamp}) last_int_rps = current_int_rps prompt, prompt_len, output_len, mm_content, request_id = ( request.prompt, request.prompt_len, request.expected_output_len, request.multi_modal_data, request.request_id, ) req_model_id, req_model_name = model_id, model_name if lora_modules: req_lora_module = next(lora_modules) req_model_id, req_model_name = req_lora_module, req_lora_module request_func_input = RequestFuncInput( model=req_model_id, model_name=req_model_name, prompt=prompt, api_url=api_url, prompt_len=prompt_len, output_len=output_len, logprobs=logprobs, multi_modal_content=mm_content, ignore_eos=ignore_eos, extra_headers=extra_headers, extra_body=extra_body, request_id=request_id, ) tasks.append( asyncio.create_task( limited_request_func( request_func_input=request_func_input, session=session, pbar=pbar ) ) ) outputs: list[RequestFuncOutput] = await asyncio.gather(*tasks) if pbar is not None: pbar.close() benchmark_duration = time.perf_counter() - benchmark_start_time if task_type == TaskType.GENERATION: metrics, actual_output_lens = calculate_metrics( input_requests=input_requests, outputs=outputs, dur_s=benchmark_duration, tokenizer=tokenizer, selected_percentiles=selected_percentiles, goodput_config_dict=goodput_config_dict, ) else: metrics = calculate_metrics_for_embeddings( outputs=outputs, dur_s=benchmark_duration, selected_percentiles=selected_percentiles, ) actual_output_lens = 0 print("{s:{c}^{n}}".format(s=" Serving Benchmark Result ", n=50, c="=")) print("{:<40} {:<10}".format("Successful requests:", metrics.completed)) print("{:<40} {:<10}".format("Failed requests:", metrics.failed)) if max_concurrency is not None: print("{:<40} {:<10}".format("Maximum request concurrency:", max_concurrency)) if request_rate != float("inf"): print("{:<40} {:<10.2f}".format("Request rate configured (RPS):", request_rate)) print("{:<40} {:<10.2f}".format("Benchmark duration (s):", benchmark_duration)) print("{:<40} {:<10}".format("Total input tokens:", metrics.total_input)) if isinstance(metrics, BenchmarkMetrics): print("{:<40} {:<10}".format("Total generated tokens:", metrics.total_output)) print( "{:<40} {:<10.2f}".format( "Request throughput (req/s):", metrics.request_throughput ) ) if goodput_config_dict: print( "{:<40} {:<10.2f}".format( "Request goodput (req/s):", metrics.request_goodput ) ) if isinstance(metrics, BenchmarkMetrics): print( "{:<40} {:<10.2f}".format( "Output token throughput (tok/s):", metrics.output_throughput ) ) print( "{:<40} {:<10.2f}".format( "Peak output token throughput (tok/s):", metrics.max_output_tokens_per_s ) ) print( "{:<40} {:<10.2f}".format( "Peak concurrent requests:", metrics.max_concurrent_requests ) ) print( "{:<40} {:<10.2f}".format( "Total token throughput (tok/s):", metrics.total_token_throughput ) ) if isinstance(metrics, BenchmarkMetrics): result = { "duration": benchmark_duration, "completed": metrics.completed, "failed": metrics.failed, "total_input_tokens": metrics.total_input, "total_output_tokens": metrics.total_output, "request_throughput": metrics.request_throughput, "request_goodput": metrics.request_goodput if goodput_config_dict else None, "output_throughput": metrics.output_throughput, "total_token_throughput": metrics.total_token_throughput, "input_lens": [output.prompt_len for output in outputs], "output_lens": actual_output_lens, "ttfts": [output.ttft for output in outputs], "itls": [output.itl for output in outputs], "generated_texts": [output.generated_text for output in outputs], "errors": [output.error for output in outputs], "max_output_tokens_per_s": metrics.max_output_tokens_per_s, "max_concurrent_requests": metrics.max_concurrent_requests, } else: result = { "duration": benchmark_duration, "completed": metrics.completed, "total_input_tokens": metrics.total_input, "request_throughput": metrics.request_throughput, "total_token_throughput": metrics.total_token_throughput, "input_lens": [output.prompt_len for output in outputs], "errors": [output.error for output in outputs], } if rps_change_events: result["rps_change_events"] = rps_change_events def process_one_metric( # E.g., "ttft" metric_attribute_name: str, # E.g., "TTFT" metric_name: str, # E.g., "Time to First Token" metric_header: str, ): # This function prints and adds statistics of the specified # metric. if metric_attribute_name not in selected_percentile_metrics: return print("{s:{c}^{n}}".format(s=metric_header, n=50, c="-")) print( "{:<40} {:<10.2f}".format( f"Mean {metric_name} (ms):", getattr(metrics, f"mean_{metric_attribute_name}_ms"), ) ) print( "{:<40} {:<10.2f}".format( f"Median {metric_name} (ms):", getattr(metrics, f"median_{metric_attribute_name}_ms"), ) ) result[f"mean_{metric_attribute_name}_ms"] = getattr( metrics, f"mean_{metric_attribute_name}_ms" ) result[f"median_{metric_attribute_name}_ms"] = getattr( metrics, f"median_{metric_attribute_name}_ms" ) result[f"std_{metric_attribute_name}_ms"] = getattr( metrics, f"std_{metric_attribute_name}_ms"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/throughput.py
vllm/benchmarks/throughput.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Benchmark offline inference throughput.""" import argparse import dataclasses import json import os import random import time import warnings from typing import Any import torch import uvloop from tqdm import tqdm from transformers import AutoModelForCausalLM, PreTrainedTokenizerBase from vllm.benchmarks.datasets import ( AIMODataset, BurstGPTDataset, ConversationDataset, InstructCoderDataset, MultiModalConversationDataset, PrefixRepetitionRandomDataset, RandomDataset, RandomDatasetForReranking, RandomMultiModalDataset, SampleRequest, ShareGPTDataset, SonnetDataset, VisionArenaDataset, add_random_dataset_base_args, add_random_multimodal_dataset_args, ) from vllm.benchmarks.lib.utils import convert_to_pytorch_benchmark_format, write_to_json from vllm.engine.arg_utils import AsyncEngineArgs, EngineArgs from vllm.inputs import TextPrompt, TokensPrompt from vllm.lora.request import LoRARequest from vllm.outputs import RequestOutput from vllm.sampling_params import BeamSearchParams from vllm.tokenizers import TokenizerLike, get_tokenizer from vllm.utils.async_utils import merge_async_iterators def run_vllm( requests: list[SampleRequest], n: int, engine_args: EngineArgs, do_profile: bool, disable_detokenize: bool = False, ) -> tuple[float, list[RequestOutput] | None]: from vllm import LLM, SamplingParams llm = LLM(**dataclasses.asdict(engine_args)) assert all( llm.llm_engine.model_config.max_model_len >= (request.prompt_len + request.expected_output_len) for request in requests ), ( "Please ensure that max_model_len is greater than the sum of" " prompt_len and expected_output_len for all requests." ) # Add the requests to the engine. prompts: list[TextPrompt | TokensPrompt] = [] sampling_params: list[SamplingParams] = [] for request in requests: prompt = ( TokensPrompt(prompt_token_ids=request.prompt["prompt_token_ids"]) if "prompt_token_ids" in request.prompt else TextPrompt(prompt=request.prompt) ) if request.multi_modal_data: assert isinstance(request.multi_modal_data, dict) prompt["multi_modal_data"] = request.multi_modal_data prompts.append(prompt) sampling_params.append( SamplingParams( n=n, temperature=1.0, top_p=1.0, ignore_eos=True, max_tokens=request.expected_output_len, detokenize=not disable_detokenize, ) ) lora_requests: list[LoRARequest] | None = None if engine_args.enable_lora: lora_requests = [request.lora_request for request in requests] use_beam_search = False outputs = None if not use_beam_search: start = time.perf_counter() if do_profile: llm.start_profile() outputs = llm.generate( prompts, sampling_params, lora_request=lora_requests, use_tqdm=True ) if do_profile: llm.stop_profile() end = time.perf_counter() else: assert lora_requests is None, "BeamSearch API does not support LoRA" prompts = [request.prompt for request in requests] # output_len should be the same for all requests. output_len = requests[0].expected_output_len for request in requests: assert request.expected_output_len == output_len start = time.perf_counter() if do_profile: llm.start_profile() llm.beam_search( prompts, BeamSearchParams( beam_width=n, max_tokens=output_len, ignore_eos=True, ), ) if do_profile: llm.stop_profile() end = time.perf_counter() return end - start, outputs def run_vllm_chat( requests: list[SampleRequest], n: int, engine_args: EngineArgs, do_profile: bool, disable_detokenize: bool = False, ) -> tuple[float, list[RequestOutput]]: """ Run vLLM chat benchmark. This function is recommended ONLY for benchmarking multimodal models as it properly handles multimodal inputs and chat formatting. For non-multimodal models, use run_vllm() instead. """ from vllm import LLM, SamplingParams llm = LLM(**dataclasses.asdict(engine_args)) assert all( llm.llm_engine.model_config.max_model_len >= (request.prompt_len + request.expected_output_len) for request in requests ), ( "Please ensure that max_model_len is greater than the sum of " "prompt_len and expected_output_len for all requests." ) prompts = [] sampling_params: list[SamplingParams] = [] for request in requests: prompts.append(request.prompt) sampling_params.append( SamplingParams( n=n, temperature=1.0, top_p=1.0, ignore_eos=True, max_tokens=request.expected_output_len, detokenize=not disable_detokenize, ) ) start = time.perf_counter() if do_profile: llm.start_profile() outputs = llm.chat(prompts, sampling_params, use_tqdm=True) if do_profile: llm.stop_profile() end = time.perf_counter() return end - start, outputs async def run_vllm_async( requests: list[SampleRequest], n: int, engine_args: AsyncEngineArgs, do_profile: bool, disable_frontend_multiprocessing: bool = False, disable_detokenize: bool = False, ) -> float: from vllm import SamplingParams from vllm.entrypoints.openai.api_server import ( build_async_engine_client_from_engine_args, ) async with build_async_engine_client_from_engine_args( engine_args, disable_frontend_multiprocessing=disable_frontend_multiprocessing, ) as llm: model_config = llm.model_config assert all( model_config.max_model_len >= (request.prompt_len + request.expected_output_len) for request in requests ), ( "Please ensure that max_model_len is greater than the sum of" " prompt_len and expected_output_len for all requests." ) # Add the requests to the engine. prompts: list[TextPrompt | TokensPrompt] = [] sampling_params: list[SamplingParams] = [] lora_requests: list[LoRARequest | None] = [] for request in requests: prompt = ( TokensPrompt(prompt_token_ids=request.prompt["prompt_token_ids"]) if "prompt_token_ids" in request.prompt else TextPrompt(prompt=request.prompt) ) if request.multi_modal_data: assert isinstance(request.multi_modal_data, dict) prompt["multi_modal_data"] = request.multi_modal_data sampling_params.append( SamplingParams( n=n, temperature=1.0, top_p=1.0, ignore_eos=True, max_tokens=request.expected_output_len, detokenize=not disable_detokenize, ) ) prompts.append(prompt) lora_requests.append(request.lora_request) generators = [] start = time.perf_counter() if do_profile: await llm.start_profile() for i, (prompt, sp, lr) in enumerate( zip(prompts, sampling_params, lora_requests) ): generator = llm.generate(prompt, sp, lora_request=lr, request_id=f"test{i}") generators.append(generator) all_gens = merge_async_iterators(*generators) async for i, res in all_gens: pass if do_profile: await llm.stop_profile() end = time.perf_counter() return end - start def run_hf( requests: list[SampleRequest], model: str, tokenizer: TokenizerLike, n: int, max_batch_size: int, trust_remote_code: bool, disable_detokenize: bool = False, ) -> float: assert isinstance(tokenizer, PreTrainedTokenizerBase), ( "the hf backend only supports HF tokenizers" ) llm = AutoModelForCausalLM.from_pretrained( model, dtype=torch.float16, trust_remote_code=trust_remote_code ) if llm.config.model_type == "llama": # To enable padding in the HF backend. tokenizer.pad_token = tokenizer.eos_token llm = llm.cuda() pbar = tqdm(total=len(requests)) start = time.perf_counter() batch: list[str] = [] max_prompt_len = 0 max_output_len = 0 for i in range(len(requests)): prompt = requests[i].prompt prompt_len = requests[i].prompt_len output_len = requests[i].expected_output_len # Add the prompt to the batch. batch.append(prompt) max_prompt_len = max(max_prompt_len, prompt_len) max_output_len = max(max_output_len, output_len) if len(batch) < max_batch_size and i != len(requests) - 1: # Check if we can add more requests to the batch. next_prompt_len = requests[i + 1].prompt_len next_output_len = requests[i + 1].expected_output_len if ( max(max_prompt_len, next_prompt_len) + max(max_output_len, next_output_len) ) <= 2048: # We can add more requests to the batch. continue # Generate the sequences. input_ids = tokenizer(batch, return_tensors="pt", padding=True).input_ids llm_outputs = llm.generate( input_ids=input_ids.cuda(), do_sample=True, num_return_sequences=n, temperature=1.0, top_p=1.0, use_cache=True, max_new_tokens=max_output_len, ) if not disable_detokenize: # Include the decoding time. tokenizer.batch_decode(llm_outputs, skip_special_tokens=True) pbar.update(len(batch)) # Clear the batch. batch = [] max_prompt_len = 0 max_output_len = 0 end = time.perf_counter() return end - start def save_to_pytorch_benchmark_format( args: argparse.Namespace, results: dict[str, Any] ) -> None: pt_records = convert_to_pytorch_benchmark_format( args=args, metrics={ "requests_per_second": [results["requests_per_second"]], "tokens_per_second": [results["tokens_per_second"]], }, extra_info={ k: results[k] for k in ["elapsed_time", "num_requests", "total_num_tokens"] }, ) if pt_records: # Don't use json suffix here as we don't want CI to pick it up pt_file = f"{os.path.splitext(args.output_json)[0]}.pytorch.json" write_to_json(pt_file, pt_records) def get_requests(args, tokenizer): # Common parameters for all dataset types. common_kwargs = { "dataset_path": args.dataset_path, "random_seed": args.seed, } sample_kwargs = { "tokenizer": tokenizer, "lora_path": args.lora_path, "max_loras": args.max_loras, "num_requests": args.num_prompts, } if args.dataset_name == "random" or ( args.dataset_path is None and args.dataset_name not in {"prefix_repetition", "random-mm", "random-rerank"} ): sample_kwargs["range_ratio"] = args.random_range_ratio # prefer random_* arguments, fall back to regular arguments random_prefix_len = getattr(args, "random_prefix_len", None) sample_kwargs["prefix_len"] = ( random_prefix_len if random_prefix_len is not None else args.prefix_len ) random_input_len = getattr(args, "random_input_len", None) sample_kwargs["input_len"] = ( random_input_len if random_input_len is not None else args.input_len ) random_output_len = getattr(args, "random_output_len", None) sample_kwargs["output_len"] = ( random_output_len if random_output_len is not None else args.output_len ) dataset_cls = RandomDataset elif args.dataset_name == "sharegpt": dataset_cls = ShareGPTDataset if args.backend == "vllm-chat": sample_kwargs["enable_multimodal_chat"] = True if args.output_len is not None: sample_kwargs["output_len"] = args.output_len elif args.dataset_name == "sonnet": assert tokenizer.chat_template or tokenizer.default_chat_template, ( "Tokenizer/model must have chat template for sonnet dataset." ) dataset_cls = SonnetDataset sample_kwargs["prefix_len"] = args.prefix_len sample_kwargs["return_prompt_formatted"] = True if args.input_len is not None: sample_kwargs["input_len"] = args.input_len if args.output_len is not None: sample_kwargs["output_len"] = args.output_len elif args.dataset_name == "burstgpt": dataset_cls = BurstGPTDataset elif args.dataset_name == "hf": if args.output_len is not None: sample_kwargs["output_len"] = args.output_len if args.dataset_path in VisionArenaDataset.SUPPORTED_DATASET_PATHS: dataset_cls = VisionArenaDataset common_kwargs["dataset_subset"] = None common_kwargs["dataset_split"] = "train" sample_kwargs["enable_multimodal_chat"] = True elif args.dataset_path in InstructCoderDataset.SUPPORTED_DATASET_PATHS: dataset_cls = InstructCoderDataset common_kwargs["dataset_split"] = "train" elif args.dataset_path in MultiModalConversationDataset.SUPPORTED_DATASET_PATHS: dataset_cls = MultiModalConversationDataset common_kwargs["dataset_subset"] = args.hf_subset common_kwargs["dataset_split"] = args.hf_split sample_kwargs["enable_multimodal_chat"] = True elif args.dataset_path in ConversationDataset.SUPPORTED_DATASET_PATHS: dataset_cls = ConversationDataset common_kwargs["dataset_subset"] = args.hf_subset common_kwargs["dataset_split"] = args.hf_split sample_kwargs["enable_multimodal_chat"] = True elif args.dataset_path in AIMODataset.SUPPORTED_DATASET_PATHS: dataset_cls = AIMODataset common_kwargs["dataset_subset"] = None common_kwargs["dataset_split"] = "train" elif args.dataset_name == "prefix_repetition": dataset_cls = PrefixRepetitionRandomDataset sample_kwargs["prefix_len"] = args.prefix_repetition_prefix_len sample_kwargs["suffix_len"] = args.prefix_repetition_suffix_len sample_kwargs["num_prefixes"] = args.prefix_repetition_num_prefixes sample_kwargs["output_len"] = args.prefix_repetition_output_len elif args.dataset_name == "random-mm": dataset_cls = RandomMultiModalDataset # prefer random_* arguments, fall back to regular arguments random_input_len = getattr(args, "random_input_len", None) sample_kwargs["input_len"] = ( random_input_len if random_input_len is not None else getattr(args, "input_len", None) ) random_output_len = getattr(args, "random_output_len", None) sample_kwargs["output_len"] = ( random_output_len if random_output_len is not None else getattr(args, "output_len", None) ) sample_kwargs["base_items_per_request"] = getattr( args, "random_mm_base_items_per_request", None ) sample_kwargs["num_mm_items_range_ratio"] = getattr( args, "random_mm_num_mm_items_range_ratio", None ) sample_kwargs["limit_mm_per_prompt"] = getattr( args, "random_mm_limit_mm_per_prompt", None ) sample_kwargs["bucket_config"] = getattr(args, "random_mm_bucket_config", None) sample_kwargs["enable_multimodal_chat"] = True random_prefix_len = getattr(args, "random_prefix_len", None) prefix_len = getattr(args, "prefix_len", None) sample_kwargs["prefix_len"] = ( random_prefix_len if random_prefix_len is not None else prefix_len ) sample_kwargs["range_ratio"] = args.random_range_ratio elif args.dataset_name == "random-rerank": dataset_cls = RandomDatasetForReranking # prefer random_* arguments, fall back to regular arguments random_input_len = getattr(args, "random_input_len", None) sample_kwargs["input_len"] = ( random_input_len if random_input_len is not None else getattr(args, "input_len", None) ) random_output_len = getattr(args, "random_output_len", None) sample_kwargs["output_len"] = ( random_output_len if random_output_len is not None else getattr(args, "output_len", None) ) sample_kwargs["batchsize"] = getattr(args, "random_batch_size", 1) sample_kwargs["is_reranker"] = not getattr(args, "no_reranker", False) sample_kwargs["range_ratio"] = args.random_range_ratio else: raise ValueError(f"Unknown dataset name: {args.dataset_name}") # Remove None values sample_kwargs = {k: v for k, v in sample_kwargs.items() if v is not None} requests = dataset_cls(**common_kwargs).sample(**sample_kwargs) requests = filter_requests_for_dp(requests, args.data_parallel_size) return requests def filter_requests_for_dp(requests, data_parallel_size): # Note(zhuohan): The way we get data_parallel_rank is hacky and only # works for external launcher mode. Should be cleaned up and deprecated # in the future with a better vLLM distributed process design. if data_parallel_size == 1: return requests global_rank = int(os.environ["RANK"]) world_size = int(os.environ["WORLD_SIZE"]) data_parallel_rank = global_rank // (world_size // data_parallel_size) return [ r for i, r in enumerate(requests) if i % data_parallel_size == data_parallel_rank ] def validate_args(args): """ Validate command-line arguments. """ # === Deprecation and Defaulting === if args.dataset is not None: warnings.warn( "The '--dataset' argument will be deprecated in the next release. " "Please use '--dataset-name' and '--dataset-path' instead.", stacklevel=2, ) args.dataset_path = args.dataset if not getattr(args, "tokenizer", None): args.tokenizer = args.model # === Backend Validation === valid_backends = {"vllm", "hf", "mii", "vllm-chat"} if args.backend not in valid_backends: raise ValueError(f"Unsupported backend: {args.backend}") # === Dataset Configuration === if ( not args.dataset and not args.dataset_path and args.dataset_name not in {"prefix_repetition"} ): print("When dataset path is not set, it will default to random dataset") args.dataset_name = "random" random_input_len = getattr(args, "random_input_len", None) if args.input_len is None and random_input_len is None: raise ValueError( "Either --input-len or --random-input-len must be provided " "for a random dataset" ) # === Dataset Name Specific Checks === # --hf-subset and --hf-split: only used # when dataset_name is 'hf' if args.dataset_name != "hf" and ( getattr(args, "hf_subset", None) is not None or getattr(args, "hf_split", None) is not None ): warnings.warn( "--hf-subset and --hf-split will be ignored \ since --dataset-name is not 'hf'.", stacklevel=2, ) elif args.dataset_name == "hf": if args.dataset_path in ( VisionArenaDataset.SUPPORTED_DATASET_PATHS.keys() | MultiModalConversationDataset.SUPPORTED_DATASET_PATHS | ConversationDataset.SUPPORTED_DATASET_PATHS ): assert args.backend == "vllm-chat", ( f"{args.dataset_path} needs to use vllm-chat as the backend." ) elif args.dataset_path in ( InstructCoderDataset.SUPPORTED_DATASET_PATHS | AIMODataset.SUPPORTED_DATASET_PATHS ): assert args.backend == "vllm", ( f"{args.dataset_path} needs to use vllm as the backend." ) else: raise ValueError(f"{args.dataset_path} is not supported by hf dataset.") # --random-range-ratio: only used when dataset_name is 'random', # 'random-mm', or 'random-rerank' if ( args.dataset_name not in {"random", "random-mm", "random-rerank"} and args.random_range_ratio is not None ): warnings.warn( "--random-range-ratio will be ignored since \ --dataset-name is not 'random', 'random-mm', or 'random-rerank'.", stacklevel=2, ) # --random-batch-size: only used when dataset_name is 'random-rerank' if ( args.dataset_name != "random-rerank" and getattr(args, "random_batch_size", None) is not None ) and args.random_batch_size != 1: warnings.warn( "--random-batch-size will be ignored since \ --dataset-name is not 'random-rerank'.", stacklevel=2, ) # --no-reranker: only used when dataset_name is 'random-rerank' if args.dataset_name != "random-rerank" and getattr(args, "no_reranker", False): warnings.warn( "--no-reranker will be ignored since \ --dataset-name is not 'random-rerank'.", stacklevel=2, ) # --prefix-len: only used when dataset_name is 'random', 'random-mm', # 'sonnet', or not set. if ( args.dataset_name not in {"random", "random-mm", "sonnet", None} and args.prefix_len is not None ): warnings.warn( "--prefix-len will be ignored since --dataset-name\ is not 'random', 'random-mm', 'sonnet', or not set.", stacklevel=2, ) # === Random Dataset Argument Conflict Detection === # Check for conflicts between regular and random arguments when using # random datasets if args.dataset_name in {"random", "random-mm", "random-rerank"}: random_input_len = getattr(args, "random_input_len", None) random_output_len = getattr(args, "random_output_len", None) random_prefix_len = getattr(args, "random_prefix_len", None) if args.input_len is not None and random_input_len is not None: warnings.warn( "Both --input-len and --random-input-len are specified. " "The random version (--random-input-len) will be preferred " "in this run.", stacklevel=2, ) if args.output_len is not None and random_output_len is not None: warnings.warn( "Both --output-len and --random-output-len are specified. " "The random version (--random-output-len) will be preferred " "in this run.", stacklevel=2, ) if args.prefix_len is not None and random_prefix_len is not None: warnings.warn( "Both --prefix-len and --random-prefix-len are specified. " "The random version (--random-prefix-len) will be preferred " "in this run.", stacklevel=2, ) # === LoRA Settings === if getattr(args, "enable_lora", False) and args.backend != "vllm": raise ValueError("LoRA benchmarking is only supported for vLLM backend") if getattr(args, "enable_lora", False) and args.lora_path is None: raise ValueError("LoRA path must be provided when enable_lora is True") # === Backend-specific Validations === if args.backend == "hf" and args.hf_max_batch_size is None: raise ValueError("HF max batch size is required for HF backend") if args.backend != "hf" and args.hf_max_batch_size is not None: raise ValueError("HF max batch size is only for HF backend.") if ( args.backend in {"hf", "mii"} and getattr(args, "quantization", None) is not None ): raise ValueError("Quantization is only for vLLM backend.") if args.backend == "mii" and args.dtype != "auto": raise ValueError("dtype must be auto for MII backend.") if args.backend == "mii" and args.n != 1: raise ValueError("n must be 1 for MII backend.") if args.backend == "mii" and args.tokenizer != args.model: raise ValueError("Tokenizer must be the same as the model for MII backend.") if args.data_parallel_size > 1 and ( args.distributed_executor_backend != "external_launcher" or args.async_engine ): # --data-parallel is not supported fully. # Old issue: https://github.com/vllm-project/vllm/issues/16222 # Currently we only support data parallel with external launcher # mode (i.e., launch with toruchrun). raise ValueError( "Data parallel is only supported with external launcher mode " "with synchronous engine in offline benchmark, " "please use benchmark serving instead" ) def add_cli_args(parser: argparse.ArgumentParser): parser.add_argument( "--backend", type=str, choices=["vllm", "hf", "mii", "vllm-chat"], default="vllm", ) parser.add_argument( "--dataset-name", type=str, choices=[ "sharegpt", "random", "sonnet", "burstgpt", "hf", "prefix_repetition", "random-mm", "random-rerank", ], help="Name of the dataset to benchmark on.", default="sharegpt", ) parser.add_argument( "--dataset", type=str, default=None, help="Path to the ShareGPT dataset, will be deprecated in\ the next release. The dataset is expected to " "be a json in form of list[dict[..., conversations: " "list[dict[..., value: <prompt_or_response>]]]]", ) parser.add_argument( "--dataset-path", type=str, default=None, help="Path to the dataset" ) parser.add_argument( "--input-len", type=int, default=None, help="Input prompt length for each request", ) parser.add_argument( "--output-len", type=int, default=None, help="Output length for each request. Overrides the " "output length from the dataset.", ) parser.add_argument( "--n", type=int, default=1, help="Number of generated sequences per prompt." ) parser.add_argument( "--num-prompts", type=int, default=1000, help="Number of prompts to process." ) parser.add_argument( "--hf-max-batch-size", type=int, default=None, help="Maximum batch size for HF backend.", ) parser.add_argument( "--output-json", type=str, default=None, help="Path to save the throughput results in JSON format.", ) parser.add_argument( "--async-engine", action="store_true", default=False, help="Use vLLM async engine rather than LLM class.", ) parser.add_argument( "--disable-frontend-multiprocessing", action="store_true", default=False, help="Disable decoupled async engine frontend.", ) parser.add_argument( "--disable-detokenize", action="store_true", help=( "Do not detokenize the response (i.e. do not include " "detokenization time in the measurement)" ), ) # LoRA parser.add_argument( "--lora-path", type=str, default=None, help="Path to the lora adapters to use. This can be an absolute path, " "a relative path, or a Hugging Face model identifier.", ) parser.add_argument( "--prefix-len", type=int, default=0, help="Number of fixed prefix tokens before the random " "context in a request (default: 0).", ) # hf dtaset parser.add_argument( "--hf-subset", type=str, default=None, help="Subset of the HF dataset.", ) parser.add_argument( "--hf-split", type=str, default=None, help="Split of the HF dataset.", ) parser.add_argument( "--profile", action="store_true", default=False, help="Use vLLM Profiling. --profiler-config must be provided on the server.", ) # prefix repetition dataset parser.add_argument( "--prefix-repetition-prefix-len", type=int, default=None, help="Number of prefix tokens per request, used only for prefix " "repetition dataset.", ) parser.add_argument( "--prefix-repetition-suffix-len", type=int, default=None, help="Number of suffix tokens per request, used only for prefix " "repetition dataset. Total input length is prefix_len + suffix_len.", ) parser.add_argument( "--prefix-repetition-num-prefixes", type=int, default=None, help="Number of prefixes to generate, used only for prefix repetition " "dataset. Prompts per prefix is num_requests // num_prefixes.", ) parser.add_argument( "--prefix-repetition-output-len", type=int, default=None, help="Number of output tokens per request, used only for prefix " "repetition dataset.", ) # (random, random-mm, random-rerank) add_random_dataset_base_args(parser) add_random_multimodal_dataset_args(parser) parser = AsyncEngineArgs.add_cli_args(parser) def main(args: argparse.Namespace): validate_args(args) if args.seed is None: args.seed = 0 random.seed(args.seed) # Sample the requests. if ( args.backend == "hf" or args.backend == "mii" ) and args.tokenizer_mode == "auto": # mistral_common tokenizer is only supported on vllm and vllm-chat backends; # for hf and mii backends, we use hf tokenizer args.tokenizer_mode = "hf" tokenizer = get_tokenizer( args.tokenizer, tokenizer_mode=args.tokenizer_mode, trust_remote_code=args.trust_remote_code, ) requests = get_requests(args, tokenizer) is_multi_modal = any(request.multi_modal_data is not None for request in requests) request_outputs: list[RequestOutput] | None = None if args.backend == "vllm": if args.async_engine: elapsed_time = uvloop.run( run_vllm_async( requests, args.n, AsyncEngineArgs.from_cli_args(args), disable_frontend_multiprocessing=args.disable_frontend_multiprocessing, disable_detokenize=args.disable_detokenize, do_profile=args.profile, ) ) else: elapsed_time, request_outputs = run_vllm( requests, args.n, EngineArgs.from_cli_args(args), disable_detokenize=args.disable_detokenize, do_profile=args.profile, ) elif args.backend == "hf": assert args.tensor_parallel_size == 1 if args.profile: raise NotImplementedError("Profiling not implemented yet for backend='hf'.") elapsed_time = run_hf( requests, args.model, tokenizer, args.n, args.hf_max_batch_size, args.trust_remote_code, args.disable_detokenize, ) elif args.backend == "vllm-chat": elapsed_time, request_outputs = run_vllm_chat( requests, args.n, EngineArgs.from_cli_args(args), disable_detokenize=args.disable_detokenize, do_profile=args.profile, ) else:
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/sweep/cli.py
vllm/benchmarks/sweep/cli.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse from vllm.entrypoints.utils import VLLM_SUBCMD_PARSER_EPILOG from .plot import SweepPlotArgs from .plot import main as plot_main from .plot_pareto import SweepPlotParetoArgs from .plot_pareto import main as plot_pareto_main from .serve import SweepServeArgs from .serve import main as serve_main from .serve_sla import SweepServeSLAArgs from .serve_sla import main as serve_sla_main SUBCOMMANDS = ( (SweepServeArgs, serve_main), (SweepServeSLAArgs, serve_sla_main), (SweepPlotArgs, plot_main), (SweepPlotParetoArgs, plot_pareto_main), ) def add_cli_args(parser: argparse.ArgumentParser): subparsers = parser.add_subparsers(required=True, dest="sweep_type") for cmd, entrypoint in SUBCOMMANDS: cmd_subparser = subparsers.add_parser( cmd.parser_name, description=cmd.parser_help, usage=f"vllm bench sweep {cmd.parser_name} [options]", ) cmd_subparser.set_defaults(dispatch_function=entrypoint) cmd.add_cli_args(cmd_subparser) cmd_subparser.epilog = VLLM_SUBCMD_PARSER_EPILOG.format( subcmd=f"sweep {cmd.parser_name}" ) def main(args: argparse.Namespace): args.dispatch_function(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/sweep/plot.py
vllm/benchmarks/sweep/plot.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import json from abc import ABC, abstractmethod from concurrent.futures import ProcessPoolExecutor from dataclasses import dataclass from functools import partial from pathlib import Path from types import TracebackType from typing import ClassVar from typing_extensions import Self, override from vllm.utils.collection_utils import full_groupby from vllm.utils.import_utils import PlaceholderModule from .utils import sanitize_filename try: import matplotlib.pyplot as plt import pandas as pd import seaborn as sns except ImportError: plt = PlaceholderModule("matplotlib").placeholder_attr("pyplot") pd = PlaceholderModule("pandas") seaborn = PlaceholderModule("seaborn") @dataclass class PlotFilterBase(ABC): var: str target: str @classmethod def parse_str(cls, s: str): for op_key in PLOT_FILTERS: if op_key in s: key, value = s.split(op_key) return PLOT_FILTERS[op_key]( key, value.removeprefix(op_key).strip("'").strip('"'), ) else: raise ValueError( f"Invalid operator for plot filter '{s}'. " f"Valid operators are: {sorted(PLOT_FILTERS)}", ) @abstractmethod def apply(self, df: "pd.DataFrame") -> "pd.DataFrame": """Applies this filter to a DataFrame.""" raise NotImplementedError @dataclass class PlotEqualTo(PlotFilterBase): @override def apply(self, df: "pd.DataFrame") -> "pd.DataFrame": try: target = float(self.target) except ValueError: target = self.target return df[df[self.var] == target] @dataclass class PlotNotEqualTo(PlotFilterBase): @override def apply(self, df: "pd.DataFrame") -> "pd.DataFrame": try: target = float(self.target) except ValueError: target = self.target return df[df[self.var] != target] @dataclass class PlotLessThan(PlotFilterBase): @override def apply(self, df: "pd.DataFrame") -> "pd.DataFrame": return df[df[self.var] < float(self.target)] @dataclass class PlotLessThanOrEqualTo(PlotFilterBase): @override def apply(self, df: "pd.DataFrame") -> "pd.DataFrame": return df[df[self.var] <= float(self.target)] @dataclass class PlotGreaterThan(PlotFilterBase): @override def apply(self, df: "pd.DataFrame") -> "pd.DataFrame": return df[df[self.var] > float(self.target)] @dataclass class PlotGreaterThanOrEqualTo(PlotFilterBase): @override def apply(self, df: "pd.DataFrame") -> "pd.DataFrame": return df[df[self.var] >= float(self.target)] # NOTE: The ordering is important! Match longer op_keys first PLOT_FILTERS: dict[str, type[PlotFilterBase]] = { "==": PlotEqualTo, "!=": PlotNotEqualTo, "<=": PlotLessThanOrEqualTo, ">=": PlotGreaterThanOrEqualTo, "<": PlotLessThan, ">": PlotGreaterThan, } class PlotFilters(list[PlotFilterBase]): @classmethod def parse_str(cls, s: str): if not s: return cls() return cls(PlotFilterBase.parse_str(e) for e in s.split(",")) def apply(self, df: "pd.DataFrame") -> "pd.DataFrame": for item in self: df = item.apply(df) return df @dataclass class PlotBinner: var: str bin_size: float @classmethod def parse_str(cls, s: str): for op_key in PLOT_BINNERS: if op_key in s: key, value = s.split(op_key) return PLOT_BINNERS[op_key](key, float(value.removeprefix(op_key))) else: raise ValueError( f"Invalid operator for plot binner '{s}'. " f"Valid operators are: {sorted(PLOT_BINNERS)}", ) def apply(self, df: "pd.DataFrame") -> "pd.DataFrame": """Applies this binner to a DataFrame.""" df = df.copy() df[self.var] = df[self.var] // self.bin_size * self.bin_size return df PLOT_BINNERS: dict[str, type[PlotBinner]] = { "%": PlotBinner, } class PlotBinners(list[PlotBinner]): @classmethod def parse_str(cls, s: str): if not s: return cls() return cls(PlotBinner.parse_str(e) for e in s.split(",")) def apply(self, df: "pd.DataFrame") -> "pd.DataFrame": for item in self: df = item.apply(df) return df def _json_load_bytes(path: Path) -> list[dict[str, object]]: with path.open("rb") as f: return json.load(f) def _convert_inf_nan_strings(data: list[dict[str, object]]) -> list[dict[str, object]]: """ Convert string values "inf", "-inf", and "nan" to their float equivalents. This handles the case where JSON serialization represents inf/nan as strings. """ converted_data = [] for record in data: converted_record = {} for key, value in record.items(): if isinstance(value, str): if value in ["inf", "-inf", "nan"]: converted_record[key] = float(value) else: converted_record[key] = value else: converted_record[key] = value converted_data.append(converted_record) return converted_data def _get_metric(run_data: dict[str, object], metric_key: str): try: return run_data[metric_key] except KeyError as exc: raise ValueError(f"Cannot find metric {metric_key!r} in {run_data=}") from exc def _get_group(run_data: dict[str, object], group_keys: list[str]): return tuple((k, str(_get_metric(run_data, k))) for k in group_keys) def _get_fig_path(fig_dir: Path, group: tuple[tuple[str, str], ...], fig_name: str): parts = list[str]() # Start with figure name (always provided, defaults to "FIGURE") parts.append(fig_name) # Always append group data if present if group: parts.extend(f"{k}={v}" for k, v in group) return fig_dir / sanitize_filename("-".join(parts) + ".png") class DummyExecutor: map = map def __enter__(self) -> Self: return self def __exit__( self, exc_type: type[BaseException] | None, exc_value: BaseException | None, exc_traceback: TracebackType | None, ) -> None: return None def _plot_fig( fig_dir: Path, fig_group_data: tuple[tuple[tuple[str, str], ...], list[dict[str, object]]], row_by: list[str], col_by: list[str], curve_by: list[str], *, var_x: str, var_y: str, filter_by: PlotFilters, bin_by: PlotBinners, scale_x: str | None, scale_y: str | None, dry_run: bool, fig_name: str, error_bars: bool, fig_height: float, fig_dpi: int, ): fig_group, fig_data = fig_group_data row_groups = full_groupby( fig_data, key=lambda item: _get_group(item, row_by), ) num_rows = len(row_groups) num_cols = max( len(full_groupby(row_data, key=lambda item: _get_group(item, col_by))) for _, row_data in row_groups ) fig_path = _get_fig_path(fig_dir, fig_group, fig_name) print("[BEGIN FIGURE]") print(f"Group: {dict(fig_group)}") print(f"Grid: {num_rows} rows x {num_cols} cols") print(f"Output file: {fig_path}") if dry_run: print("[END FIGURE]") return # Convert string "inf", "-inf", and "nan" to their float equivalents fig_data = _convert_inf_nan_strings(fig_data) df = pd.DataFrame.from_records(fig_data) if var_x not in df.columns: raise ValueError( f"Cannot find {var_x=!r} in parameter sweep results. " f"Available variables: {df.columns.tolist()}" ) if var_y not in df.columns: raise ValueError( f"Cannot find {var_y=!r} in parameter sweep results. " f"Available variables: {df.columns.tolist()}" ) for k in row_by: if k not in df.columns: raise ValueError( f"Cannot find row_by={k!r} in parameter sweep results. " f"Available variables: {df.columns.tolist()}" ) for k in col_by: if k not in df.columns: raise ValueError( f"Cannot find col_by={k!r} in parameter sweep results. " f"Available variables: {df.columns.tolist()}" ) for k in curve_by: if k not in df.columns: raise ValueError( f"Cannot find curve_by={k!r} in parameter sweep results. " f"Available variables: {df.columns.tolist()}" ) df = filter_by.apply(df) df = bin_by.apply(df) # Sort by curve_by columns alphabetically for consistent legend ordering if curve_by: df = df.sort_values(by=curve_by) df["row_group"] = ( pd.concat( [k + "=" + df[k].astype(str) for k in row_by], axis=1, ).agg("\n".join, axis=1) if row_by else "(All)" ) df["col_group"] = ( pd.concat( [k + "=" + df[k].astype(str) for k in col_by], axis=1, ).agg("\n".join, axis=1) if col_by else "(All)" ) g = sns.FacetGrid(df, row="row_group", col="col_group", height=fig_height) if row_by and col_by: g.set_titles("{row_name}\n{col_name}") elif row_by: g.set_titles("{row_name}") elif col_by: g.set_titles("{col_name}") else: g.set_titles("") if scale_x: g.set(xscale=scale_x) if scale_y: g.set(yscale=scale_y) if len(curve_by) <= 3: hue, style, size, *_ = (*curve_by, None, None, None) g.map_dataframe( sns.lineplot, x=var_x, y=var_y, hue=hue, style=style, size=size, markers=True, errorbar="sd" if error_bars else None, ) g.add_legend(title=hue) else: df["curve_group"] = ( pd.concat( [k + "=" + df[k].astype(str) for k in curve_by], axis=1, ).agg("\n".join, axis=1) if curve_by else "(All)" ) g.map_dataframe( sns.lineplot, x=var_x, y=var_y, hue="curve_group", markers=True, errorbar="sd" if error_bars else None, ) g.add_legend() g.savefig(fig_path, dpi=fig_dpi) plt.close(g.figure) print("[END FIGURE]") def plot( output_dir: Path, fig_dir: Path, fig_by: list[str], row_by: list[str], col_by: list[str], curve_by: list[str], *, var_x: str, var_y: str, filter_by: PlotFilters, bin_by: PlotBinners, scale_x: str | None, scale_y: str | None, dry_run: bool, fig_name: str = "FIGURE", error_bars: bool = True, fig_height: float = 6.4, fig_dpi: int = 300, ): all_data = [ run_data for path in output_dir.rglob("**/summary.json") for run_data in _json_load_bytes(path) ] if not all_data: raise ValueError(f"Did not find any parameter sweep results under {output_dir}") fig_dir.mkdir(parents=True, exist_ok=True) fig_groups = full_groupby( all_data, key=lambda item: _get_group(item, fig_by), ) with DummyExecutor() if len(fig_groups) <= 1 else ProcessPoolExecutor() as executor: # Resolve the iterable to ensure that the workers are run all( executor.map( partial( _plot_fig, fig_dir, row_by=row_by, col_by=col_by, curve_by=curve_by, var_x=var_x, var_y=var_y, filter_by=filter_by, bin_by=bin_by, scale_x=scale_x, scale_y=scale_y, dry_run=dry_run, fig_name=fig_name, error_bars=error_bars, fig_height=fig_height, fig_dpi=fig_dpi, ), fig_groups, ) ) @dataclass class SweepPlotArgs: output_dir: Path fig_dir: Path fig_by: list[str] row_by: list[str] col_by: list[str] curve_by: list[str] var_x: str var_y: str filter_by: PlotFilters bin_by: PlotBinners scale_x: str | None scale_y: str | None dry_run: bool fig_name: str = "FIGURE" error_bars: bool = True fig_height: float = 6.4 fig_dpi: int = 300 parser_name: ClassVar[str] = "plot" parser_help: ClassVar[str] = "Plot performance curves from parameter sweep results." @classmethod def from_cli_args(cls, args: argparse.Namespace): output_dir = Path(args.OUTPUT_DIR) if not output_dir.exists(): raise ValueError(f"No parameter sweep results under {output_dir}") curve_by = [] if not args.curve_by else args.curve_by.split(",") row_by = [] if not args.row_by else args.row_by.split(",") col_by = [] if not args.col_by else args.col_by.split(",") fig_by = [] if not args.fig_by else args.fig_by.split(",") return cls( output_dir=output_dir, fig_dir=output_dir / args.fig_dir, fig_by=fig_by, row_by=row_by, col_by=col_by, curve_by=curve_by, var_x=args.var_x, var_y=args.var_y, filter_by=PlotFilters.parse_str(args.filter_by), bin_by=PlotBinners.parse_str(args.bin_by), scale_x=args.scale_x, scale_y=args.scale_y, dry_run=args.dry_run, fig_name=args.fig_name, error_bars=not args.no_error_bars, fig_height=args.fig_height, fig_dpi=args.fig_dpi, ) @classmethod def add_cli_args(cls, parser: argparse.ArgumentParser) -> argparse.ArgumentParser: parser.add_argument( "OUTPUT_DIR", type=str, default="results", help="The directory containing the results to plot, " "i.e., the `--output-dir` argument to the parameter sweep script.", ) parser.add_argument( "--fig-dir", type=str, default="", help="The directory to save the figures, relative to `OUTPUT_DIR`. " "By default, the same directory is used.", ) parser.add_argument( "--fig-by", type=str, default="", help="A comma-separated list of variables, such that a separate figure " "is created for each combination of these variables.", ) parser.add_argument( "--row-by", type=str, default="", help="A comma-separated list of variables, such that a separate row " "is created for each combination of these variables.", ) parser.add_argument( "--col-by", type=str, default="", help="A comma-separated list of variables, such that a separate column " "is created for each combination of these variables.", ) parser.add_argument( "--curve-by", type=str, default=None, help="A comma-separated list of variables, such that a separate curve " "is created for each combination of these variables.", ) parser.add_argument( "--var-x", type=str, default="request_throughput", help="The variable for the x-axis.", ) parser.add_argument( "--var-y", type=str, default="p99_e2el_ms", help="The variable for the y-axis", ) parser.add_argument( "--filter-by", type=str, default="", help="A comma-separated list of statements indicating values to filter by. " "This is useful to remove outliers. " "Example: `max_concurrency<1000,max_num_batched_tokens<=4096` means " "plot only the points where `max_concurrency` is less than 1000 and " "`max_num_batched_tokens` is no greater than 4096.", ) parser.add_argument( "--bin-by", type=str, default="", help="A comma-separated list of statements indicating values to bin by. " "This is useful to avoid plotting points that are too close together. " "Example: `request_throughput%%1` means " "use a bin size of 1 for the `request_throughput` variable.", ) parser.add_argument( "--scale-x", type=str, default=None, help="The scale to use for the x-axis. " "Currently only accepts string values such as 'log' and 'sqrt'. " "See also: https://seaborn.pydata.org/generated/seaborn.objects.Plot.scale.html", ) parser.add_argument( "--scale-y", type=str, default=None, help="The scale to use for the y-axis. " "Currently only accepts string values such as 'log' and 'sqrt'. " "See also: https://seaborn.pydata.org/generated/seaborn.objects.Plot.scale.html", ) parser.add_argument( "--fig-name", type=str, default="FIGURE", help="Name prefix for the output figure file. " "Group data is always appended when present. " "Default: 'FIGURE'. Example: --fig-name my_performance_plot", ) parser.add_argument( "--no-error-bars", action="store_true", help="If set, disables error bars on the plot. " "By default, error bars are shown.", ) parser.add_argument( "--fig-height", type=float, default=6.4, help="Height of each subplot in inches. Default: 6.4", ) parser.add_argument( "--fig-dpi", type=int, default=300, help="Resolution of the output figure in dots per inch. Default: 300", ) parser.add_argument( "--dry-run", action="store_true", help="If set, prints the information about each figure to plot, " "then exits without drawing them.", ) return parser def run_main(args: SweepPlotArgs): return plot( output_dir=args.output_dir, fig_dir=args.fig_dir, fig_by=args.fig_by, row_by=args.row_by, col_by=args.col_by, curve_by=args.curve_by, var_x=args.var_x, var_y=args.var_y, filter_by=args.filter_by, bin_by=args.bin_by, scale_x=args.scale_x, scale_y=args.scale_y, dry_run=args.dry_run, fig_name=args.fig_name, error_bars=args.error_bars, fig_height=args.fig_height, fig_dpi=args.fig_dpi, ) def main(args: argparse.Namespace): run_main(SweepPlotArgs.from_cli_args(args)) if __name__ == "__main__": parser = argparse.ArgumentParser(description=SweepPlotArgs.parser_help) SweepPlotArgs.add_cli_args(parser) main(parser.parse_args())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/sweep/plot_pareto.py
vllm/benchmarks/sweep/plot_pareto.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import math from concurrent.futures import ProcessPoolExecutor from dataclasses import dataclass from functools import partial from pathlib import Path from typing import ClassVar from vllm.utils.collection_utils import full_groupby from vllm.utils.import_utils import PlaceholderModule from .plot import DummyExecutor, _json_load_bytes from .utils import sanitize_filename try: import matplotlib.pyplot as plt import pandas as pd import seaborn as sns except ImportError: plt = PlaceholderModule("matplotlib").placeholder_attr("pyplot") pd = PlaceholderModule("pandas") sns = PlaceholderModule("seaborn") def _first_present(run_data: dict[str, object], keys: list[str]): for key in keys: for candidate in {key, key.replace("_", "-"), key.replace("-", "_")}: if candidate in run_data: return run_data[candidate] return None def _get_numeric( run_data: dict[str, object], keys: list[str], *, allow_zero: bool = True, ) -> float | None: value = _first_present(run_data, keys) if value is None: return None try: numeric = float(value) except (TypeError, ValueError) as exc: raise ValueError( f"Expected numeric value for one of {keys}, " f"but found {value!r} in {run_data=}" ) from exc if not allow_zero and numeric == 0: return None return numeric def _infer_user_count( run_data: dict[str, object], user_count_var: str | None, ) -> float | None: candidates = [user_count_var] if user_count_var else [] candidates.extend(["request_rate"]) user_count = _get_numeric(run_data, candidates, allow_zero=False) if user_count is not None: return user_count # Fallback to the observed peak if configured value is missing. return _get_numeric(run_data, ["max_concurrent_requests"], allow_zero=False) def _infer_gpu_count( run_data: dict[str, object], gpu_count_var: str | None, ) -> float: direct_candidates = [gpu_count_var] if gpu_count_var else [] direct_gpu_count = _get_numeric(run_data, direct_candidates, allow_zero=False) if direct_gpu_count: return direct_gpu_count tp_size = _get_numeric(run_data, ["tensor_parallel_size", "tp"]) pp_size = _get_numeric(run_data, ["pipeline_parallel_size", "pp"]) dp_size = _get_numeric(run_data, ["data_parallel_size", "dp"]) world_size = 1.0 if tp_size: world_size *= tp_size if pp_size: world_size *= pp_size if dp_size: world_size *= dp_size return world_size def _get_throughput( run_data: dict[str, object], throughput_var: str, ) -> float: throughput = _get_numeric(run_data, [throughput_var]) if throughput is None: raise ValueError( f"Cannot find throughput metric {throughput_var!r} in run data. " f"Available keys: {sorted(run_data)}" ) return throughput def _prepare_records( all_data: list[dict[str, object]], *, user_count_var: str | None, gpu_count_var: str | None, ) -> tuple[list[dict[str, object]], int]: prepared = [] skipped_missing_users = 0 for record in all_data: throughput = _get_throughput(record, "output_throughput") user_count = _infer_user_count(record, user_count_var) if user_count is None: skipped_missing_users += 1 continue gpu_count = _infer_gpu_count(record, gpu_count_var) tokens_per_user = throughput / user_count tokens_per_gpu = throughput / gpu_count prepared.append( { **record, "tokens_per_user": tokens_per_user, "tokens_per_gpu": tokens_per_gpu, "user_count_estimate": user_count, "gpu_count": gpu_count, } ) return prepared, skipped_missing_users def _pareto_frontier( df: "pd.DataFrame", x_col: str, y_col: str, *, epsilon: float = 1e-9, ) -> "pd.DataFrame": sorted_df = df.sort_values([x_col, y_col], ascending=[False, False]) frontier_indices = [] best_y = -math.inf for idx, row in sorted_df.iterrows(): y_val = row[y_col] if y_val >= best_y - epsilon: frontier_indices.append(idx) best_y = max(best_y, y_val) return df.loc[frontier_indices] def _get_fig_path( fig_dir: Path, fig_group: tuple[tuple[str, str], ...], ) -> Path: parts = ["PARETO"] if fig_group: parts.extend(f"{k}={v}" for k, v in fig_group) filename = sanitize_filename("-".join(parts) + ".png") return fig_dir / filename def _plot_fig( fig_dir: Path, fig_group_data: tuple[tuple[tuple[str, str], ...], list[dict[str, object]]], label_by: list[str], *, dry_run: bool, ): fig_group, fig_data = fig_group_data fig_path = _get_fig_path(fig_dir, fig_group) print("[BEGIN FIGURE]") print(f"Group: {dict(fig_group)}") print(f"Output file: {fig_path}") if dry_run: print("[END FIGURE]") return df = pd.DataFrame.from_records(fig_data) df = df.dropna(subset=["tokens_per_user", "tokens_per_gpu"]) if df.empty: print("No data points available after filtering; skipping.") print("[END FIGURE]") return frontier = _pareto_frontier(df, "tokens_per_user", "tokens_per_gpu") frontier = frontier.sort_values("tokens_per_user") fig, ax = plt.subplots() sns.scatterplot( data=df, x="tokens_per_user", y="tokens_per_gpu", color="0.5", alpha=0.6, ax=ax, label="All runs", ) sns.lineplot( data=frontier, x="tokens_per_user", y="tokens_per_gpu", marker="o", ax=ax, label="Pareto frontier", ) if label_by: for _, row in frontier.iterrows(): label_parts = [] for key in label_by: if key in row: label_parts.append(f"{key}={row[key]}") if label_parts: ax.text( row["tokens_per_user"], row["tokens_per_gpu"], "\n".join(label_parts), fontsize=8, ) ax.set_xlabel("Tokens/s/user") ax.set_ylabel("Tokens/s/GPU") ax.grid(True, linestyle="--", linewidth=0.5, alpha=0.6) fig.tight_layout() fig.savefig(fig_path) plt.close(fig) print( f"Plotted {len(df)} points; Pareto frontier size: {len(frontier)}.", ) print("[END FIGURE]") def plot_pareto( output_dir: Path, user_count_var: str | None, gpu_count_var: str | None, label_by: list[str], *, dry_run: bool, ): fig_dir = output_dir / "pareto" raw_data = [ run_data for path in output_dir.rglob("**/summary.json") for run_data in _json_load_bytes(path) ] if not raw_data: raise ValueError(f"Did not find any parameter sweep results under {output_dir}") fig_dir.mkdir(parents=True, exist_ok=True) prepared_data, skipped_missing_users = _prepare_records( raw_data, user_count_var=user_count_var, gpu_count_var=gpu_count_var, ) if skipped_missing_users: print( f"Skipped {skipped_missing_users} runs without a user count " "(`max_concurrency` or `max_concurrent_requests`).", ) if not prepared_data: raise ValueError( "No data points with both throughput and user count available " "to plot Pareto frontier.", ) fig_groups = full_groupby( prepared_data, key=lambda item: tuple(), ) with DummyExecutor() if len(fig_groups) <= 1 else ProcessPoolExecutor() as executor: all( executor.map( partial( _plot_fig, fig_dir, label_by=label_by, dry_run=dry_run, ), fig_groups, ) ) @dataclass class SweepPlotParetoArgs: output_dir: Path user_count_var: str | None gpu_count_var: str | None label_by: list[str] dry_run: bool parser_name: ClassVar[str] = "plot_pareto" parser_help: ClassVar[str] = ( "Plot Pareto frontier between tokens/s/user and tokens/s/GPU " "from parameter sweep results." ) @classmethod def from_cli_args(cls, args: argparse.Namespace): output_dir = Path(args.OUTPUT_DIR) if not output_dir.exists(): raise ValueError(f"No parameter sweep results under {output_dir}") label_by = [] if not args.label_by else args.label_by.split(",") return cls( output_dir=output_dir, user_count_var=args.user_count_var, gpu_count_var=args.gpu_count_var, label_by=label_by, dry_run=args.dry_run, ) @classmethod def add_cli_args(cls, parser: argparse.ArgumentParser): parser.add_argument( "OUTPUT_DIR", type=str, default="results", help="The directory containing the sweep results to plot.", ) parser.add_argument( "--user-count-var", type=str, default="max_concurrency", help="Result key that stores concurrent user count. " "Falls back to max_concurrent_requests if missing.", ) parser.add_argument( "--gpu-count-var", type=str, default=None, help="Result key that stores GPU count. " "If not provided, falls back to num_gpus/gpu_count " "or tensor_parallel_size * pipeline_parallel_size.", ) parser.add_argument( "--label-by", type=str, default="max_concurrency,gpu_count", help="Comma-separated list of fields to annotate on Pareto frontier " "points.", ) parser.add_argument( "--dry-run", action="store_true", help="If set, prints the figures to plot without drawing them.", ) return parser def run_main(args: SweepPlotParetoArgs): return plot_pareto( output_dir=args.output_dir, user_count_var=args.user_count_var, gpu_count_var=args.gpu_count_var, label_by=args.label_by, dry_run=args.dry_run, ) def main(args: argparse.Namespace): run_main(SweepPlotParetoArgs.from_cli_args(args)) if __name__ == "__main__": parser = argparse.ArgumentParser(description=SweepPlotParetoArgs.parser_help) SweepPlotParetoArgs.add_cli_args(parser) main(parser.parse_args())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/sweep/utils.py
vllm/benchmarks/sweep/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project def sanitize_filename(filename: str) -> str: return filename.replace("/", "_").replace("..", "__").strip("'").strip('"')
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/sweep/param_sweep.py
vllm/benchmarks/sweep/param_sweep.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json import os from typing import Any class ParameterSweep(list["ParameterSweepItem"]): @classmethod def read_json(cls, filepath: os.PathLike): with open(filepath, "rb") as f: data = json.load(f) # Support both list and dict formats if isinstance(data, dict): return cls.read_from_dict(data) return cls.from_records(data) @classmethod def read_from_dict(cls, data: dict[str, dict[str, object]]): """ Read parameter sweep from a dict format where keys are names. Example: { "experiment1": {"max_tokens": 100, "temperature": 0.7}, "experiment2": {"max_tokens": 200, "temperature": 0.9} } """ records = [{"_benchmark_name": name, **params} for name, params in data.items()] return cls.from_records(records) @classmethod def from_records(cls, records: list[dict[str, object]]): if not isinstance(records, list): raise TypeError( f"The parameter sweep should be a list of dictionaries, " f"but found type: {type(records)}" ) # Validate that all _benchmark_name values are unique if provided names = [r["_benchmark_name"] for r in records if "_benchmark_name" in r] if names and len(names) != len(set(names)): duplicates = [name for name in names if names.count(name) > 1] raise ValueError( f"Duplicate _benchmark_name values found: {set(duplicates)}. " f"All _benchmark_name values must be unique." ) return cls(ParameterSweepItem.from_record(record) for record in records) class ParameterSweepItem(dict[str, object]): @classmethod def from_record(cls, record: dict[str, object]): if not isinstance(record, dict): raise TypeError( f"Each item in the parameter sweep should be a dictionary, " f"but found type: {type(record)}" ) return cls(record) def __or__(self, other: dict[str, Any]): return type(self)(super().__or__(other)) @property def name(self) -> str: """ Get the name for this parameter sweep item. Returns the '_benchmark_name' field if present, otherwise returns a text representation of all parameters. """ if "_benchmark_name" in self: return self["_benchmark_name"] return self.as_text(sep="-") # In JSON, we prefer "_" def _iter_param_key_candidates(self, param_key: str): # Inner config arguments are not converted by the CLI if "." in param_key: prefix, rest = param_key.split(".", 1) for prefix_candidate in self._iter_param_key_candidates(prefix): yield prefix_candidate + "." + rest return yield param_key yield param_key.replace("-", "_") yield param_key.replace("_", "-") # In CLI, we prefer "-" def _iter_cmd_key_candidates(self, param_key: str): for k in reversed(tuple(self._iter_param_key_candidates(param_key))): yield "--" + k def _normalize_cmd_key(self, param_key: str): return next(self._iter_cmd_key_candidates(param_key)) def has_param(self, param_key: str) -> bool: return any(k in self for k in self._iter_param_key_candidates(param_key)) def _normalize_cmd_kv_pair(self, k: str, v: object) -> list[str]: """ Normalize a key-value pair into command-line arguments. Returns a list containing either: - A single element for boolean flags (e.g., ['--flag'] or ['--flag=true']) - Two elements for key-value pairs (e.g., ['--key', 'value']) """ if isinstance(v, bool): # For nested params (containing "."), use =true/false syntax if "." in k: return [f"{self._normalize_cmd_key(k)}={'true' if v else 'false'}"] else: return [self._normalize_cmd_key(k if v else "no-" + k)] else: return [self._normalize_cmd_key(k), str(v)] def apply_to_cmd(self, cmd: list[str]) -> list[str]: cmd = list(cmd) for k, v in self.items(): # Skip the '_benchmark_name' field, not a parameter if k == "_benchmark_name": continue # Serialize dict values as JSON if isinstance(v, dict): v = json.dumps(v) for k_candidate in self._iter_cmd_key_candidates(k): try: k_idx = cmd.index(k_candidate) # Replace existing parameter normalized = self._normalize_cmd_kv_pair(k, v) if len(normalized) == 1: # Boolean flag cmd[k_idx] = normalized[0] else: # Key-value pair cmd[k_idx] = normalized[0] cmd[k_idx + 1] = normalized[1] break except ValueError: continue else: # Add new parameter cmd.extend(self._normalize_cmd_kv_pair(k, v)) return cmd def as_text(self, sep: str = ", ") -> str: return sep.join(f"{k}={v}" for k, v in self.items() if k != "_benchmark_name")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/sweep/sla_sweep.py
vllm/benchmarks/sweep/sla_sweep.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json import os from abc import ABC, abstractmethod from dataclasses import dataclass from typing_extensions import override @dataclass class SLACriterionBase(ABC): target: float @abstractmethod def validate(self, actual: float) -> bool: """Return `True` if this criterion is met; otherwise `False`.""" raise NotImplementedError @abstractmethod def format_cond(self, lhs: str) -> str: raise NotImplementedError def print_and_validate( self, metrics: dict[str, float], metrics_key: str, ) -> bool: metric = metrics[metrics_key] result = self.validate(metric) cond = self.format_cond(f"{metrics_key} = {metric:.2f}") print(f"Validating SLA: {cond} | " + ("PASSED" if result else "FAILED")) return result @dataclass class SLALessThan(SLACriterionBase): @override def validate(self, actual: float) -> bool: return actual < self.target @override def format_cond(self, lhs: str) -> str: return f"{lhs}<{self.target:.2f}" @dataclass class SLALessThanOrEqualTo(SLACriterionBase): @override def validate(self, actual: float) -> bool: return actual <= self.target @override def format_cond(self, lhs: str) -> str: return f"{lhs}<={self.target:.2f}" @dataclass class SLAGreaterThan(SLACriterionBase): @override def validate(self, actual: float) -> bool: return actual > self.target @override def format_cond(self, lhs: str) -> str: return f"{lhs}>{self.target:.2f}" @dataclass class SLAGreaterThanOrEqualTo(SLACriterionBase): @override def validate(self, actual: float) -> bool: return actual >= self.target @override def format_cond(self, lhs: str) -> str: return f"{lhs}>={self.target:.2f}" # NOTE: The ordering is important! Match longer op_keys first SLA_CRITERIA: dict[str, type[SLACriterionBase]] = { "<=": SLALessThanOrEqualTo, ">=": SLAGreaterThanOrEqualTo, "<": SLALessThan, ">": SLAGreaterThan, } class SLASweep(list["SLASweepItem"]): @classmethod def read_json(cls, filepath: os.PathLike): with open(filepath, "rb") as f: records = json.load(f) return cls.from_records(records) @classmethod def from_records(cls, records: list[dict[str, str]]): if not isinstance(records, list): raise TypeError( f"The SLA sweep should be a list of dictionaries, " f"but found type: {type(records)}" ) return cls(SLASweepItem.from_record(record) for record in records) class SLASweepItem(dict[str, SLACriterionBase]): @classmethod def from_record(cls, record: dict[str, str]): sla_criteria: dict[str, SLACriterionBase] = {} for metric_key, metric_value in record.items(): for op_key in SLA_CRITERIA: if metric_value.startswith(op_key): sla_criteria[metric_key] = SLA_CRITERIA[op_key]( float(metric_value.removeprefix(op_key)) ) break else: raise ValueError( f"Invalid operator for " f"SLA constraint '{metric_key}={metric_value}'. " f"Valid operators are: {sorted(SLA_CRITERIA)}", ) return cls(sla_criteria) def as_text(self, sep: str = ", ") -> str: return sep.join(v.format_cond(k) for k, v in self.items())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/sweep/__init__.py
vllm/benchmarks/sweep/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/sweep/serve.py
vllm/benchmarks/sweep/serve.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import contextlib import json import shlex from dataclasses import dataclass from datetime import datetime from pathlib import Path from typing import ClassVar from vllm.utils.import_utils import PlaceholderModule from .param_sweep import ParameterSweep, ParameterSweepItem from .server import ServerProcess from .utils import sanitize_filename try: import pandas as pd except ImportError: pd = PlaceholderModule("pandas") @contextlib.contextmanager def run_server( serve_cmd: list[str], after_bench_cmd: list[str], *, show_stdout: bool, serve_overrides: ParameterSweepItem, dry_run: bool, ): server_cmd = serve_overrides.apply_to_cmd(serve_cmd) print("[BEGIN SERVER]") print(f"Server overrides: {serve_overrides}") print(f"Server command: {server_cmd}") if dry_run: yield None print("[END SERVER]") return with ServerProcess(server_cmd, after_bench_cmd, show_stdout=show_stdout) as server: yield server print("[END SERVER]") def _update_run_data( run_data: dict[str, object], serve_overrides: ParameterSweepItem, bench_overrides: ParameterSweepItem, run_number: int, ): run_data["run_number"] = run_number run_data.update(serve_overrides) run_data.update(bench_overrides) return run_data def run_benchmark( server: ServerProcess | None, bench_cmd: list[str], *, serve_overrides: ParameterSweepItem, bench_overrides: ParameterSweepItem, run_number: int, output_path: Path, dry_run: bool, ): benchmark_cmd = [ *bench_overrides.apply_to_cmd(bench_cmd), "--percentile-metrics", "ttft,tpot,itl,e2el", "--save-result", "--result-dir", str(output_path.parent), "--result-filename", output_path.name, ] print("[BEGIN BENCHMARK]") print(f"Benchmark overrides: {bench_overrides}") print(f"Run Number: {run_number}") print(f"Benchmark command: {benchmark_cmd}") print(f"Output file: {output_path}") run_data: dict[str, object] if output_path.exists(): print("Found existing results. Skipping.") with output_path.open("rb") as f: run_data = json.load(f) return _update_run_data( run_data, serve_overrides, bench_overrides, run_number, ) if server is None: if not dry_run: raise ValueError(f"Cannot find results at {output_path}") print("[END BENCHMARK]") return None output_path.parent.mkdir(parents=True, exist_ok=True) server.run_subcommand(benchmark_cmd) server.after_bench() with output_path.open("rb") as f: run_data = json.load(f) run_data = _update_run_data( run_data, serve_overrides, bench_overrides, run_number, ) with output_path.open("w") as f: json.dump(run_data, f, indent=4) print("[END BENCHMARK]") return run_data def _get_comb_base_path( output_dir: Path, serve_comb: ParameterSweepItem, bench_comb: ParameterSweepItem, ): parts = list[str]() if serve_comb: parts.extend(("SERVE-", serve_comb.name)) if bench_comb: parts.extend(("BENCH-", bench_comb.name)) return output_dir / sanitize_filename("-".join(parts)) def _get_comb_run_path(base_path: Path, run_number: int | None): if run_number is None: return base_path / "summary.json" return base_path / f"run={run_number}.json" def _comb_needs_server( serve_comb: ParameterSweepItem, bench_combs: ParameterSweep, output_dir: Path, ): for bench_comb in bench_combs: base_path = _get_comb_base_path(output_dir, serve_comb, bench_comb) if not _get_comb_run_path(base_path, run_number=None).exists(): return True return False def run_comb( server: ServerProcess | None, bench_cmd: list[str], *, serve_comb: ParameterSweepItem, bench_comb: ParameterSweepItem, base_path: Path, num_runs: int, dry_run: bool, ): comb_data = list[dict[str, object]]() for run_number in range(num_runs): run_data = run_benchmark( server, bench_cmd, serve_overrides=serve_comb, bench_overrides=bench_comb, run_number=run_number, output_path=_get_comb_run_path(base_path, run_number), dry_run=dry_run, ) if run_data is not None: comb_data.append(run_data) if dry_run: return None with _get_comb_run_path(base_path, run_number=None).open("w") as f: json.dump(comb_data, f, indent=4) return comb_data def run_combs( serve_cmd: list[str], bench_cmd: list[str], after_bench_cmd: list[str], *, show_stdout: bool, serve_params: ParameterSweep, bench_params: ParameterSweep, output_dir: Path, num_runs: int, dry_run: bool, links: list[tuple[str, str]], ): all_data = list[dict[str, object]]() for serve_comb in serve_params: with ( run_server( serve_cmd, after_bench_cmd, show_stdout=show_stdout, serve_overrides=serve_comb, dry_run=dry_run, ) if _comb_needs_server(serve_comb, bench_params, output_dir) else contextlib.nullcontext() ) as server: for bench_comb in bench_params: should_run = all( serve_key in serve_comb and bench_key in bench_comb and serve_comb[serve_key] == bench_comb[bench_key] for serve_key, bench_key in links ) if not should_run: continue base_path = _get_comb_base_path(output_dir, serve_comb, bench_comb) comb_data = run_comb( server, bench_cmd, serve_comb=serve_comb, bench_comb=bench_comb, base_path=base_path, num_runs=num_runs, dry_run=dry_run, ) if comb_data is not None: all_data.extend(comb_data) if dry_run: return None combined_df = pd.DataFrame.from_records(all_data) combined_df.to_csv(output_dir / "summary.csv") return combined_df @dataclass class SweepServeArgs: serve_cmd: list[str] bench_cmd: list[str] after_bench_cmd: list[str] show_stdout: bool serve_params: ParameterSweep bench_params: ParameterSweep output_dir: Path num_runs: int dry_run: bool resume: str | None link_vars: list[tuple[str, str]] | None parser_name: ClassVar[str] = "serve" parser_help: ClassVar[str] = "Run vLLM server benchmark under multiple settings." @classmethod def from_cli_args(cls, args: argparse.Namespace): serve_cmd = shlex.split(args.serve_cmd) bench_cmd = shlex.split(args.bench_cmd) after_bench_cmd = ( [] if args.after_bench_cmd is None else shlex.split(args.after_bench_cmd) ) if args.serve_params: serve_params = ParameterSweep.read_json(args.serve_params) else: # i.e.: run serve_cmd without any modification serve_params = ParameterSweep.from_records([{}]) if args.bench_params: bench_params = ParameterSweep.read_json(args.bench_params) else: # i.e.: run bench_cmd without any modification bench_params = ParameterSweep.from_records([{}]) link_vars = cls.parse_link_vars(args.link_vars) num_runs = args.num_runs if num_runs < 1: raise ValueError("`num_runs` should be at least 1.") return cls( serve_cmd=serve_cmd, bench_cmd=bench_cmd, after_bench_cmd=after_bench_cmd, show_stdout=args.show_stdout, serve_params=serve_params, bench_params=bench_params, output_dir=Path(args.output_dir), num_runs=num_runs, dry_run=args.dry_run, resume=args.resume, link_vars=link_vars, ) @classmethod def add_cli_args(cls, parser: argparse.ArgumentParser) -> argparse.ArgumentParser: parser.add_argument( "--serve-cmd", type=str, required=True, help="The command used to run the server: `vllm serve ...`", ) parser.add_argument( "--bench-cmd", type=str, required=True, help="The command used to run the benchmark: `vllm bench serve ...`", ) parser.add_argument( "--after-bench-cmd", type=str, default=None, help="After a benchmark run is complete, invoke this command instead of " "the default `ServerWrapper.clear_cache()`.", ) parser.add_argument( "--show-stdout", action="store_true", help="If set, logs the standard output of subcommands. " "Useful for debugging but can be quite spammy.", ) parser.add_argument( "--serve-params", type=str, default=None, help="Path to JSON file containing parameter combinations " "for the `vllm serve` command. Can be either a list of dicts or a dict " "where keys are benchmark names. " "If both `serve_params` and `bench_params` are given, " "this script will iterate over their Cartesian product.", ) parser.add_argument( "--bench-params", type=str, default=None, help="Path to JSON file containing parameter combinations " "for the `vllm bench serve` command. Can be either a list of dicts or " "a dict where keys are benchmark names. " "If both `serve_params` and `bench_params` are given, " "this script will iterate over their Cartesian product.", ) parser.add_argument( "-o", "--output-dir", type=str, default="results", help="The directory to which results are written.", ) parser.add_argument( "--num-runs", type=int, default=3, help="Number of runs per parameter combination.", ) parser.add_argument( "--dry-run", action="store_true", help="If set, prints the commands to run, " "then exits without executing them.", ) parser.add_argument( "--resume", type=str, default=None, help="Set this to the name of a directory under `output_dir` (which is a " "timestamp) to resume a previous execution of this script, i.e., only run " "parameter combinations for which there are still no output files.", ) parser.add_argument( "--link-vars", type=str, default="", help=( "Comma-separated list of linked variables between serve and bench, " "e.g. max_num_seqs=max_concurrency,max_model_len=random_input_len" ), ) return parser @staticmethod def parse_link_vars(s: str) -> list[tuple[str, str]]: if not s: return [] pairs = [] for item in s.split(","): a, b = item.split("=") pairs.append((a.strip(), b.strip())) return pairs def run_main(args: SweepServeArgs): timestamp = args.resume or datetime.now().strftime("%Y%m%d_%H%M%S") output_dir = args.output_dir / timestamp if args.resume and not output_dir.exists(): raise ValueError(f"Cannot resume from non-existent directory ({output_dir})") try: return run_combs( serve_cmd=args.serve_cmd, bench_cmd=args.bench_cmd, after_bench_cmd=args.after_bench_cmd, show_stdout=args.show_stdout, serve_params=args.serve_params, bench_params=args.bench_params, output_dir=output_dir, num_runs=args.num_runs, dry_run=args.dry_run, links=args.link_vars, ) except BaseException as exc: raise RuntimeError( f"The script was terminated early. Use `--resume {timestamp}` " f"to continue the script from its last checkpoint." ) from exc def main(args: argparse.Namespace): run_main(SweepServeArgs.from_cli_args(args)) if __name__ == "__main__": parser = argparse.ArgumentParser(description=SweepServeArgs.parser_help) SweepServeArgs.add_cli_args(parser) main(parser.parse_args())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/sweep/serve_sla.py
vllm/benchmarks/sweep/serve_sla.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import contextlib import json import math from dataclasses import asdict, dataclass from datetime import datetime from pathlib import Path from typing import ClassVar, Literal, get_args from typing_extensions import assert_never from vllm.utils.import_utils import PlaceholderModule from .param_sweep import ParameterSweep, ParameterSweepItem from .serve import SweepServeArgs, run_benchmark, run_server from .server import ServerProcess from .sla_sweep import SLASweep, SLASweepItem from .utils import sanitize_filename try: import pandas as pd except ImportError: pd = PlaceholderModule("pandas") def _get_sla_base_path( output_dir: Path, serve_comb: ParameterSweepItem, bench_comb: ParameterSweepItem, ): parts = list[str]() if serve_comb: parts.extend(("SERVE-", serve_comb.as_text(sep="-"))) if bench_comb: parts.extend(("BENCH-", bench_comb.as_text(sep="-"))) return output_dir / sanitize_filename("-".join(parts)) def _get_sla_iter_path( base_path: Path, sla_comb: SLASweepItem, sla_variable: str, sla_value: int | None, ): if sla_value is None: prefix = sla_comb.as_text(sep="-") return base_path / f"SLA--{prefix}.json" return base_path / f"{sla_variable}={sla_value}" def _get_sla_run_path(iter_path: Path, run_number: int | None): if run_number is None: return iter_path / "summary.json" return iter_path / f"run={run_number}.json" def _sla_needs_server( serve_comb: ParameterSweepItem, bench_combs: ParameterSweep, sla_combs: SLASweep, sla_variable: str, output_dir: Path, ): for bench_comb in bench_combs: base_path = _get_sla_base_path(output_dir, serve_comb, bench_comb) for sla_comb in sla_combs: if not _get_sla_iter_path( base_path, sla_comb, sla_variable, sla_value=None, ).exists(): return True return False def run_sla( server: ServerProcess | None, bench_cmd: list[str], *, serve_comb: ParameterSweepItem, bench_comb: ParameterSweepItem, iter_path: Path, num_runs: int, dry_run: bool, ): iter_data = list[dict[str, object]]() for run_number in range(num_runs): run_data = run_benchmark( server, bench_cmd, serve_overrides=serve_comb, bench_overrides=bench_comb, run_number=run_number, output_path=_get_sla_run_path(iter_path, run_number), dry_run=dry_run, ) if run_data is not None: iter_data.append(run_data) if dry_run: return None with _get_sla_run_path(iter_path, run_number=None).open("w") as f: json.dump(iter_data, f, indent=4) return iter_data SLAVariable = Literal["request_rate", "max_concurrency"] def _estimate_sla_value(run_data: dict[str, object], sla_variable: SLAVariable): request_throughput = float(run_data["request_throughput"]) # type: ignore if sla_variable == "request_rate": return request_throughput if sla_variable == "max_concurrency": mean_latency_ms = float(run_data["mean_e2el_ms"]) # type: ignore return request_throughput * mean_latency_ms / 1000 assert_never(sla_variable) def _estimate_sla_bounds( server: ServerProcess | None, bench_cmd: list[str], *, serve_comb: ParameterSweepItem, bench_comb: ParameterSweepItem, sla_comb: SLASweepItem, base_path: Path, num_runs: int, dry_run: bool, sla_variable: SLAVariable, init_value: int, max_value: int, ): sla_data = list[dict[str, object]]() max_passing: int = 0 min_failing: int = 0 val: int = init_value assert val > 0 while True: print(f"Testing {sla_variable}: {val} req/s") iter_data = run_sla( server, bench_cmd, serve_comb=serve_comb, bench_comb=bench_comb | {sla_variable: val}, iter_path=_get_sla_iter_path(base_path, sla_comb, sla_variable, val), num_runs=num_runs, dry_run=dry_run, ) assert iter_data is not None sla_data.extend(iter_data) iter_data_mean = { k: sum(float(run_data[k]) for run_data in iter_data) / len(iter_data) # type: ignore for k in sla_comb } sla_results = [ criterion.print_and_validate(iter_data_mean, k) for k, criterion in sla_comb.items() ] if all(sla_results): print("SLA criteria are met.") max_passing = val val *= 2 else: print("SLA criteria are not met.") min_failing = val break if val >= max_value: break return sla_data, (max_passing, min_failing) def _find_sla_value( server: ServerProcess | None, bench_cmd: list[str], *, serve_comb: ParameterSweepItem, bench_comb: ParameterSweepItem, sla_comb: SLASweepItem, base_path: Path, num_runs: int, dry_run: bool, sla_variable: SLAVariable, min_value: int, max_value: int, ): sla_data = list[dict[str, object]]() left: int = min_value right: int = max_value while True: val = (left + right) // 2 print(f"Testing {sla_variable}: {val} req/s") iter_data = run_sla( server, bench_cmd, serve_comb=serve_comb, bench_comb=bench_comb | {sla_variable: val}, iter_path=_get_sla_iter_path(base_path, sla_comb, sla_variable, val), num_runs=num_runs, dry_run=dry_run, ) assert iter_data is not None sla_data.extend(iter_data) iter_data_mean = { k: sum(float(run_data[k]) for run_data in iter_data) / len(iter_data) # type: ignore for k in sla_comb } sla_results = [ criterion.print_and_validate(iter_data_mean, k) for k, criterion in sla_comb.items() ] if all(sla_results): print("SLA criteria are met.") left = val else: print("SLA criteria are not met.") right = val if right - left <= 1: break return sla_data, left def search_sla( server: ServerProcess | None, bench_cmd: list[str], *, serve_comb: ParameterSweepItem, bench_comb: ParameterSweepItem, sla_comb: SLASweepItem, sla_variable: SLAVariable, sla_inf_value: int = 65536, # The value that represents infinite QPS base_path: Path, num_runs: int, dry_run: bool, ): print("[SLA START]") print(f"SLA criteria: {sla_comb.as_text()}") sla_data_0 = run_sla( server, bench_cmd, serve_comb=serve_comb, bench_comb=bench_comb | {sla_variable: sla_inf_value}, iter_path=_get_sla_iter_path(base_path, sla_comb, sla_variable, sla_inf_value), num_runs=num_runs, dry_run=dry_run, ) if sla_data_0 is None: assert dry_run print("Omitting SLA search.") print("[SLA END]") return None sla_init_value = math.ceil( sum(_estimate_sla_value(item, sla_variable) for item in sla_data_0) / len(sla_data_0) ) print(f"Initial {sla_variable} to search: {sla_init_value} req/s.") sla_data_1, (sla_min, sla_max) = _estimate_sla_bounds( server, bench_cmd, serve_comb=serve_comb, bench_comb=bench_comb, sla_comb=sla_comb, base_path=base_path, num_runs=num_runs, dry_run=dry_run, sla_variable=sla_variable, init_value=sla_init_value, max_value=sla_inf_value, ) print(f"Range of {sla_variable} to search: [{sla_min}, {sla_max}] req/s.") sla_data_2, sla_value = _find_sla_value( server, bench_cmd, serve_comb=serve_comb, bench_comb=bench_comb, sla_comb=sla_comb, base_path=base_path, num_runs=num_runs, dry_run=dry_run, sla_variable=sla_variable, min_value=sla_min, max_value=sla_max, ) sla_data = sla_data_0 + sla_data_1 + sla_data_2 print(f"Maximum {sla_variable} for SLA: {sla_value} req/s.") with _get_sla_iter_path( base_path, sla_comb, sla_variable, sla_value=None, ).open("w") as f: json.dump(sla_data, f, indent=4) print("[SLA END]") return sla_data def run_slas( serve_cmd: list[str], bench_cmd: list[str], after_bench_cmd: list[str], *, show_stdout: bool, serve_params: ParameterSweep, bench_params: ParameterSweep, sla_params: SLASweep, sla_variable: SLAVariable, output_dir: Path, num_runs: int, dry_run: bool, ): if any(bench_comb.has_param(sla_variable) for bench_comb in bench_params): raise ValueError( f"You should not override `{sla_variable}` in `bench_params` in SLA mode, " "since it is supposed to be determined automatically." ) all_data = list[dict[str, object]]() for serve_comb in serve_params: with ( run_server( serve_cmd, after_bench_cmd, show_stdout=show_stdout, serve_overrides=serve_comb, dry_run=dry_run, ) if _sla_needs_server( serve_comb, bench_params, sla_params, sla_variable, output_dir, ) else contextlib.nullcontext() ) as server: for bench_comb in bench_params: for sla_comb in sla_params: base_path = _get_sla_base_path(output_dir, serve_comb, bench_comb) comb_data = search_sla( server, bench_cmd, serve_comb=serve_comb, bench_comb=bench_comb, sla_comb=sla_comb, sla_variable=sla_variable, base_path=base_path, num_runs=num_runs, dry_run=dry_run, ) if comb_data is not None: all_data.extend(comb_data) if dry_run: return None combined_df = pd.DataFrame.from_records(all_data) combined_df.to_csv(output_dir / "summary.csv") return combined_df @dataclass class SweepServeSLAArgs(SweepServeArgs): sla_params: SLASweep sla_variable: SLAVariable parser_name: ClassVar[str] = "serve_sla" parser_help: ClassVar[str] = "Tune a variable to meet SLAs under multiple settings." @classmethod def from_cli_args(cls, args: argparse.Namespace): # NOTE: Don't use super() as `from_cli_args` calls `cls()` base_args = SweepServeArgs.from_cli_args(args) if args.sla_params: sla_params = SLASweep.read_json(args.sla_params) else: sla_params = SLASweep.from_records([]) return cls( **asdict(base_args), sla_params=sla_params, sla_variable=args.sla_variable, ) @classmethod def add_cli_args(cls, parser: argparse.ArgumentParser) -> argparse.ArgumentParser: parser = super().add_cli_args(parser) sla_group = parser.add_argument_group("sla options") sla_group.add_argument( "--sla-params", type=str, required=True, help="Path to JSON file containing a list of SLA constraints to satisfy. " 'Each constraint is expressed in `{"<KEY>": "<OP><VALUE>"}` format, ' 'e.g.: `{"p99_e2el_ms": "<=500"}` means that ' "the E2E latency should be less than 500ms 99%% of the time. " "Setting this option runs this script in SLA mode, which searches for " "the maximum `sla_variable` that satisfies the constraints for " "each combination of `serve_params`, `bench_params`, and `sla_params`.", ) sla_group.add_argument( "--sla-variable", type=str, choices=get_args(SLAVariable), default="request_rate", help="Whether to tune request rate or maximum concurrency to satisfy " "the SLA constraints.", ) return parser def run_main(args: SweepServeSLAArgs): timestamp = args.resume or datetime.now().strftime("%Y%m%d_%H%M%S") output_dir = args.output_dir / timestamp if args.resume and not output_dir.exists(): raise ValueError(f"Cannot resume from non-existent directory ({output_dir})") try: return run_slas( serve_cmd=args.serve_cmd, bench_cmd=args.bench_cmd, after_bench_cmd=args.after_bench_cmd, show_stdout=args.show_stdout, serve_params=args.serve_params, bench_params=args.bench_params, sla_params=args.sla_params, sla_variable=args.sla_variable, output_dir=output_dir, num_runs=args.num_runs, dry_run=args.dry_run, ) except BaseException as exc: raise RuntimeError( f"The script was terminated early. Use `--resume {timestamp}` " f"to continue the script from its last checkpoint." ) from exc def main(args: argparse.Namespace): run_main(SweepServeSLAArgs.from_cli_args(args)) if __name__ == "__main__": parser = argparse.ArgumentParser(description=SweepServeSLAArgs.parser_help) SweepServeSLAArgs.add_cli_args(parser) main(parser.parse_args())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/sweep/server.py
vllm/benchmarks/sweep/server.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import contextlib import os import signal import subprocess from types import TracebackType import requests from typing_extensions import Self class ServerProcess: def __init__( self, server_cmd: list[str], after_bench_cmd: list[str], *, show_stdout: bool, ) -> None: super().__init__() self.server_cmd = server_cmd self.after_bench_cmd = after_bench_cmd self.show_stdout = show_stdout def __enter__(self) -> Self: self.start() return self def __exit__( self, exc_type: type[BaseException] | None, exc_value: BaseException | None, exc_traceback: TracebackType | None, ) -> None: self.stop() def start(self): # Create new process for clean termination self._server_process = subprocess.Popen( self.server_cmd, start_new_session=True, stdout=None if self.show_stdout else subprocess.DEVNULL, # Need `VLLM_SERVER_DEV_MODE=1` for `_reset_caches` env=os.environ | {"VLLM_SERVER_DEV_MODE": "1"}, ) def stop(self): server_process = self._server_process if server_process.poll() is None: # In case only some processes have been terminated with contextlib.suppress(ProcessLookupError): # We need to kill both API Server and Engine processes os.killpg(os.getpgid(server_process.pid), signal.SIGKILL) def run_subcommand(self, cmd: list[str]): return subprocess.run( cmd, stdout=None if self.show_stdout else subprocess.DEVNULL, check=True, ) def after_bench(self) -> None: if not self.after_bench_cmd: self.reset_caches() return self.run_subcommand(self.after_bench_cmd) def _get_vllm_server_address(self) -> str: server_cmd = self.server_cmd for host_key in ("--host",): if host_key in server_cmd: host = server_cmd[server_cmd.index(host_key) + 1] break else: host = "localhost" for port_key in ("-p", "--port"): if port_key in server_cmd: port = int(server_cmd[server_cmd.index(port_key) + 1]) break else: port = 8000 # The default value in vllm serve return f"http://{host}:{port}" def reset_caches(self) -> None: server_cmd = self.server_cmd # Use `.endswith()` to match `/bin/...` if server_cmd[0].endswith("vllm"): server_address = self._get_vllm_server_address() print(f"Resetting caches at {server_address}") res = requests.post(f"{server_address}/reset_prefix_cache") res.raise_for_status() res = requests.post(f"{server_address}/reset_mm_cache") res.raise_for_status() elif server_cmd[0].endswith("infinity_emb"): if "--vector-disk-cache" in server_cmd: raise NotImplementedError( "Infinity server uses caching but does not expose a method " "to reset the cache" ) else: raise NotImplementedError( f"No implementation of `reset_caches` for `{server_cmd[0]}` server. " "Please specify a custom command via `--after-bench-cmd`." )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/lib/ready_checker.py
vllm/benchmarks/lib/ready_checker.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Utilities for checking endpoint readiness.""" import asyncio import time import aiohttp from tqdm.asyncio import tqdm from .endpoint_request_func import RequestFunc, RequestFuncInput, RequestFuncOutput async def wait_for_endpoint( request_func: RequestFunc, test_input: RequestFuncInput, session: aiohttp.ClientSession, timeout_seconds: int = 600, retry_interval: int = 5, ) -> RequestFuncOutput: """ Wait for an endpoint to become available before starting benchmarks. Args: request_func: The async request function to call test_input: The RequestFuncInput to test with timeout_seconds: Maximum time to wait in seconds (default: 10 minutes) retry_interval: Time between retries in seconds (default: 5 seconds) Returns: RequestFuncOutput: The successful response Raises: ValueError: If the endpoint doesn't become available within the timeout """ deadline = time.perf_counter() + timeout_seconds output = RequestFuncOutput(success=False) print(f"Waiting for endpoint to become up in {timeout_seconds} seconds") with tqdm( total=timeout_seconds, bar_format="{desc} |{bar}| {elapsed} elapsed, {remaining} remaining", unit="s", ) as pbar: while True: # update progress bar remaining = deadline - time.perf_counter() elapsed = timeout_seconds - remaining update_amount = min(elapsed - pbar.n, timeout_seconds - pbar.n) pbar.update(update_amount) pbar.refresh() if remaining <= 0: pbar.close() break # ping the endpoint using request_func try: output = await request_func( request_func_input=test_input, session=session ) if output.success: pbar.close() return output except aiohttp.ClientConnectorError: pass # retry after a delay sleep_duration = min(retry_interval, remaining) if sleep_duration > 0: await asyncio.sleep(sleep_duration) return output
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/lib/utils.py
vllm/benchmarks/lib/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import json import math import os from typing import Any def convert_to_pytorch_benchmark_format( args: argparse.Namespace, metrics: dict[str, list], extra_info: dict[str, Any] ) -> list: """ Save the benchmark results in the format used by PyTorch OSS benchmark with on metric per record https://github.com/pytorch/pytorch/wiki/How-to-integrate-with-PyTorch-OSS-benchmark-database """ records = [] if not os.environ.get("SAVE_TO_PYTORCH_BENCHMARK_FORMAT", False): return records for name, benchmark_values in metrics.items(): record = { "benchmark": { "name": "vLLM benchmark", "extra_info": { "args": vars(args), }, }, "model": { "name": args.model, }, "metric": { "name": name, "benchmark_values": benchmark_values, "extra_info": extra_info, }, } tp = record["benchmark"]["extra_info"]["args"].get("tensor_parallel_size") # Save tensor_parallel_size parameter if it's part of the metadata if not tp and "tensor_parallel_size" in extra_info: record["benchmark"]["extra_info"]["args"]["tensor_parallel_size"] = ( extra_info["tensor_parallel_size"] ) records.append(record) return records class InfEncoder(json.JSONEncoder): def clear_inf(self, o: Any): if isinstance(o, dict): return { str(k) if not isinstance(k, (str, int, float, bool, type(None))) else k: self.clear_inf(v) for k, v in o.items() } elif isinstance(o, list): return [self.clear_inf(v) for v in o] elif isinstance(o, float) and math.isinf(o): return "inf" return o def iterencode(self, o: Any, *args, **kwargs) -> Any: return super().iterencode(self.clear_inf(o), *args, **kwargs) def write_to_json(filename: str, records: list) -> None: with open(filename, "w") as f: json.dump( records, f, cls=InfEncoder, default=lambda o: f"<{type(o).__name__} is not JSON serializable>", )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/lib/__init__.py
vllm/benchmarks/lib/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Benchmark library utilities."""
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/benchmarks/lib/endpoint_request_func.py
vllm/benchmarks/lib/endpoint_request_func.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """The request function for API endpoints.""" import io import json import os import sys import time import traceback from collections.abc import Awaitable from dataclasses import dataclass, field from typing import Any, Literal, Protocol import aiohttp import regex as re from tqdm.asyncio import tqdm AIOHTTP_TIMEOUT = aiohttp.ClientTimeout(total=6 * 60 * 60) class StreamedResponseHandler: """Handles streaming HTTP responses by accumulating chunks until complete messages are available.""" def __init__(self): self.buffer = "" def add_chunk(self, chunk_bytes: bytes) -> list[str]: """Add a chunk of bytes to the buffer and return any complete messages.""" chunk_str = chunk_bytes.decode("utf-8") self.buffer += chunk_str messages = [] # Split by double newlines (SSE message separator) while "\n\n" in self.buffer: message, self.buffer = self.buffer.split("\n\n", 1) message = message.strip() if message: messages.append(message) # if self.buffer is not empty, check if it is a complete message # by removing data: prefix and check if it is a valid JSON if self.buffer.startswith("data: "): message_content = self.buffer.removeprefix("data: ").strip() if message_content == "[DONE]": messages.append(self.buffer.strip()) self.buffer = "" elif message_content: try: json.loads(message_content) messages.append(self.buffer.strip()) self.buffer = "" except json.JSONDecodeError: # Incomplete JSON, wait for more chunks. pass return messages @dataclass class RequestFuncInput: """The input for the request function.""" prompt: str | list[str] api_url: str prompt_len: int output_len: int model: str model_name: str | None = None logprobs: int | None = None extra_headers: dict | None = None extra_body: dict | None = None multi_modal_content: dict | list[dict] | None = None ignore_eos: bool = False language: str | None = None request_id: str | None = None @dataclass class RequestFuncOutput: """The output of the request function including metrics.""" generated_text: str = "" success: bool = False latency: float = 0.0 output_tokens: int = 0 ttft: float = 0.0 # Time to first token itl: list[float] = field(default_factory=list) # list of inter-token latencies tpot: float = 0.0 # avg next-token latencies prompt_len: int = 0 error: str = "" start_time: float = 0.0 class RequestFunc(Protocol): def __call__( self, request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, ) -> Awaitable[RequestFuncOutput]: ... def _validate_api_url( api_url: str, api_name: str, expected_suffixes: str | set[str], ) -> None: if isinstance(expected_suffixes, str): expected_suffixes = {expected_suffixes} expected_suffixes = {*expected_suffixes, "profile"} if not api_url.endswith(tuple(expected_suffixes)): raise ValueError(f"{api_name} URL must end with one of: {expected_suffixes}.") def _update_payload_common( payload: dict[str, Any], request_func_input: RequestFuncInput, ) -> None: if request_func_input.ignore_eos: payload["ignore_eos"] = request_func_input.ignore_eos if request_func_input.extra_body: payload.update(request_func_input.extra_body) def _update_headers_common( headers: dict[str, Any], request_func_input: RequestFuncInput, ) -> None: if request_func_input.extra_headers: headers |= request_func_input.extra_headers if request_func_input.request_id: headers["x-request-id"] = request_func_input.request_id async def async_request_openai_completions( request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, ) -> RequestFuncOutput: """The async request function for the OpenAI Completions API. Args: request_func_input: The input for the request function. pbar: The progress bar to display the progress. Returns: The output of the request function. """ api_url = request_func_input.api_url _validate_api_url(api_url, "OpenAI Completions API", "completions") payload = { "model": request_func_input.model_name if request_func_input.model_name else request_func_input.model, "prompt": request_func_input.prompt, "temperature": 0.0, "repetition_penalty": 1.0, "max_tokens": request_func_input.output_len, "logprobs": request_func_input.logprobs, "stream": True, "stream_options": { "include_usage": True, }, } _update_payload_common(payload, request_func_input) headers = { "Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}", } _update_headers_common(headers, request_func_input) output = RequestFuncOutput() output.prompt_len = request_func_input.prompt_len generated_text = "" st = time.perf_counter() output.start_time = st most_recent_timestamp = st try: async with session.post(url=api_url, json=payload, headers=headers) as response: if response.status == 200: first_chunk_received = False handler = StreamedResponseHandler() async for chunk_bytes in response.content.iter_any(): chunk_bytes = chunk_bytes.strip() if not chunk_bytes: continue messages = handler.add_chunk(chunk_bytes) for message in messages: # NOTE: SSE comments (often used as pings) start with # a colon. These are not JSON data payload and should # be skipped. if message.startswith(":"): continue chunk = message.removeprefix("data: ") if chunk != "[DONE]": data = json.loads(chunk) # NOTE: Some completion API might have a last # usage summary response without a token so we # want to check a token was generated if choices := data.get("choices"): # Note that text could be empty here # e.g. for special tokens text = choices[0].get("text") timestamp = time.perf_counter() # First token if not first_chunk_received: first_chunk_received = True ttft = time.perf_counter() - st output.ttft = ttft # Decoding phase else: output.itl.append(timestamp - most_recent_timestamp) most_recent_timestamp = timestamp generated_text += text or "" elif usage := data.get("usage"): output.output_tokens = usage.get("completion_tokens") if first_chunk_received: output.success = True else: output.success = False output.error = ( "Never received a valid chunk to calculate TTFT." "This response will be marked as failed!" ) output.generated_text = generated_text output.latency = most_recent_timestamp - st else: output.error = response.reason or "" output.success = False except Exception: output.success = False exc_info = sys.exc_info() output.error = "".join(traceback.format_exception(*exc_info)) if pbar: pbar.update(1) return output def _get_chat_content( request_func_input: RequestFuncInput, mm_position: Literal["first", "last"] = "last", ) -> list[dict[str, Any]]: text_contents = [{"type": "text", "text": request_func_input.prompt}] mm_contents = [] if request_func_input.multi_modal_content: mm_content = request_func_input.multi_modal_content if isinstance(mm_content, list): mm_contents.extend(request_func_input.multi_modal_content) elif isinstance(mm_content, dict): mm_contents.append(request_func_input.multi_modal_content) else: raise TypeError( "multi_modal_content must be a dict or list[dict] for openai-chat" ) if mm_position == "first": return mm_contents + text_contents return text_contents + mm_contents async def async_request_openai_chat_completions( request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, mm_position: Literal["first", "last"] = "last", ) -> RequestFuncOutput: api_url = request_func_input.api_url _validate_api_url(api_url, "OpenAI Chat Completions API", "chat/completions") content = _get_chat_content(request_func_input, mm_position=mm_position) payload = { "model": request_func_input.model_name if request_func_input.model_name else request_func_input.model, "messages": [ {"role": "user", "content": content}, ], "temperature": 0.0, "max_completion_tokens": request_func_input.output_len, "stream": True, "stream_options": { "include_usage": True, }, } _update_payload_common(payload, request_func_input) headers = { "Content-Type": "application/json", "Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}", } _update_headers_common(headers, request_func_input) output = RequestFuncOutput() output.prompt_len = request_func_input.prompt_len generated_text = "" ttft = 0.0 st = time.perf_counter() output.start_time = st most_recent_timestamp = st try: async with session.post(url=api_url, json=payload, headers=headers) as response: if response.status == 200: handler = StreamedResponseHandler() async for chunk_bytes in response.content.iter_any(): chunk_bytes = chunk_bytes.strip() if not chunk_bytes: continue messages = handler.add_chunk(chunk_bytes) for message in messages: # NOTE: SSE comments (often used as pings) start with # a colon. These are not JSON data payload and should # be skipped. if message.startswith(":"): continue chunk = message.removeprefix("data: ") if chunk != "[DONE]": timestamp = time.perf_counter() data = json.loads(chunk) if choices := data.get("choices"): content = choices[0]["delta"].get("content") # First token if ttft == 0.0: ttft = timestamp - st output.ttft = ttft # Decoding phase else: output.itl.append(timestamp - most_recent_timestamp) generated_text += content or "" elif usage := data.get("usage"): output.output_tokens = usage.get("completion_tokens") most_recent_timestamp = timestamp output.generated_text = generated_text output.success = True output.latency = most_recent_timestamp - st else: output.error = response.reason or "" output.success = False except Exception: output.success = False exc_info = sys.exc_info() output.error = "".join(traceback.format_exception(*exc_info)) if pbar: pbar.update(1) return output async def async_request_openai_audio( request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, ) -> RequestFuncOutput: # Lazy import without PlaceholderModule to avoid vllm dep. import soundfile api_url = request_func_input.api_url _validate_api_url(api_url, "OpenAI Audio API", {"transcriptions", "translations"}) content = [{"type": "text", "text": request_func_input.prompt}] payload = { "model": request_func_input.model_name if request_func_input.model_name else request_func_input.model, "temperature": 0.0, "max_completion_tokens": request_func_input.output_len, "stream": True, "language": "en", # Flattened due to multipart/form-data "stream_include_usage": True, "stream_continuous_usage_stats": True, } _update_payload_common(payload, request_func_input) headers = { "Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}", } _update_headers_common(headers, request_func_input) # Send audio file def to_bytes(y, sr): buffer = io.BytesIO() soundfile.write(buffer, y, sr, format="WAV") buffer.seek(0) return buffer mm_audio = request_func_input.multi_modal_content if not isinstance(mm_audio, dict) or "audio" not in mm_audio: raise TypeError("multi_modal_content must be a dict containing 'audio'") with to_bytes(*mm_audio["audio"]) as f: form = aiohttp.FormData() form.add_field("file", f, content_type="audio/wav") for key, value in payload.items(): form.add_field(key, str(value)) output = RequestFuncOutput() output.prompt_len = request_func_input.prompt_len generated_text = "" ttft = 0.0 st = time.perf_counter() output.start_time = st most_recent_timestamp = st try: async with session.post( url=api_url, data=form, headers=headers ) as response: if response.status == 200: handler = StreamedResponseHandler() async for chunk_bytes in response.content.iter_any(): chunk_bytes = chunk_bytes.strip() if not chunk_bytes: continue messages = handler.add_chunk(chunk_bytes) for message in messages: chunk = message.decode("utf-8").removeprefix("data: ") if chunk != "[DONE]": timestamp = time.perf_counter() data = json.loads(chunk) if choices := data.get("choices"): content = choices[0]["delta"].get("content") # First token if ttft == 0.0: ttft = timestamp - st output.ttft = ttft # Decoding phase else: output.itl.append( timestamp - most_recent_timestamp ) generated_text += content or "" elif usage := data.get("usage"): output.output_tokens = usage.get( "completion_tokens" ) most_recent_timestamp = timestamp output.generated_text = generated_text output.success = True output.latency = most_recent_timestamp - st else: output.error = response.reason or "" output.success = False except Exception: output.success = False exc_info = sys.exc_info() output.error = "".join(traceback.format_exception(*exc_info)) if pbar: pbar.update(1) return output async def _run_pooling_request( session: aiohttp.ClientSession, api_url: str, payload: dict[str, Any], headers: dict[str, Any], pbar: tqdm | None = None, ) -> RequestFuncOutput: output = RequestFuncOutput() st = time.perf_counter() output.start_time = st try: async with session.post(url=api_url, headers=headers, json=payload) as response: if response.status == 200: output.ttft = output.latency = time.perf_counter() - st if payload.get("encoding_format", "float") == "bytes": metadata = json.loads(response.headers["metadata"]) usage = metadata.get("usage", {}) else: data = await response.json() usage = data.get("usage", {}) output.success = True output.generated_text = "" output.prompt_len = usage.get("prompt_tokens", 0) else: output.success = False output.error = response.reason or "" except Exception as e: output.success = False output.error = str(e) if pbar: pbar.update(1) return output async def async_request_openai_embeddings( request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, ) -> RequestFuncOutput: api_url = request_func_input.api_url _validate_api_url(api_url, "OpenAI Embeddings API", "embeddings") payload = { "model": request_func_input.model_name if request_func_input.model_name else request_func_input.model, "input": request_func_input.prompt, # Many embedding models have short context length, # this is to avoid dropping some of the requests. "truncate_prompt_tokens": -1, } _update_payload_common(payload, request_func_input) headers = { "Content-Type": "application/json", "Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}", } _update_headers_common(headers, request_func_input) return await _run_pooling_request( session, api_url, payload=payload, headers=headers, pbar=pbar, ) async def async_request_vllm_rerank( request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, ) -> RequestFuncOutput: api_url = request_func_input.api_url _validate_api_url(api_url, "vLLM score API", "rerank") assert ( isinstance(request_func_input.prompt, list) and len(request_func_input.prompt) > 1 ) payload = { "model": request_func_input.model_name if request_func_input.model_name else request_func_input.model, "query": request_func_input.prompt[0], "documents": request_func_input.prompt[1:], # Many reranker models have short context length, # this is to avoid dropping some of the requests. "truncate_prompt_tokens": -1, } headers = { "Content-Type": "application/json", "Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}", } _update_headers_common(headers, request_func_input) return await _run_pooling_request( session, api_url, payload=payload, headers=headers, pbar=pbar, ) async def async_request_openai_embeddings_chat( request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, mm_position: Literal["first", "last"] = "last", ) -> RequestFuncOutput: api_url = request_func_input.api_url _validate_api_url(api_url, "OpenAI Embeddings API", "embeddings") content = _get_chat_content(request_func_input, mm_position=mm_position) payload = { "model": request_func_input.model_name if request_func_input.model_name else request_func_input.model, "messages": [ {"role": "user", "content": content}, ], # Many embedding models have short context length, # this is to avoid dropping some of the requests. "truncate_prompt_tokens": -1, } _update_payload_common(payload, request_func_input) headers = { "Content-Type": "application/json", "Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}", } _update_headers_common(headers, request_func_input) return await _run_pooling_request( session, api_url, payload=payload, headers=headers, pbar=pbar, ) def _try_extract_request_idx(request_func_input: RequestFuncInput): if request_func_input.request_id: match = re.search(r"(\d+)$", request_func_input.request_id) if match: try: return int(match.group(1)) except ValueError: pass return None def _preprocess_clip(request_func_input: RequestFuncInput): if request_func_input.multi_modal_content: # Image input request_func_input.prompt = "" def _preprocess_vlm2vec(request_func_input: RequestFuncInput): if request_func_input.multi_modal_content: request_idx = _try_extract_request_idx(request_func_input) # Adjust the ratio manually if needed. use_image_only_prompt = request_idx is None or request_idx % 2 == 0 if use_image_only_prompt: # Image input request_func_input.prompt = "Represent the given image." else: # Text+Image input request_func_input.prompt = ( f"Represent the given image with the following question: " f"{request_func_input.prompt}" ) async def async_request_openai_embeddings_clip( request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, ) -> RequestFuncOutput: _preprocess_clip(request_func_input) return await async_request_openai_embeddings_chat( request_func_input, session, pbar=pbar, ) async def async_request_openai_embeddings_vlm2vec( request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, ) -> RequestFuncOutput: _preprocess_vlm2vec(request_func_input) return await async_request_openai_embeddings_chat( request_func_input, session, pbar=pbar, mm_position="first", ) async def async_request_infinity_embeddings( request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, ) -> RequestFuncOutput: api_url = request_func_input.api_url _validate_api_url(api_url, "Infinity Embeddings API", "embeddings") payload = { "model": request_func_input.model_name if request_func_input.model_name else request_func_input.model, } if request_func_input.prompt: payload["input"] = request_func_input.prompt else: mm_content = request_func_input.multi_modal_content assert isinstance(mm_content, dict) mm_type = mm_content["type"] payload["input"] = mm_content[mm_type]["url"] payload["modality"] = mm_type.split("_", 1)[0] _update_payload_common(payload, request_func_input) headers = { "Content-Type": "application/json", "Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}", } _update_headers_common(headers, request_func_input) return await _run_pooling_request( session, api_url, payload=payload, headers=headers, pbar=pbar, ) async def async_request_infinity_embeddings_clip( request_func_input: RequestFuncInput, session: aiohttp.ClientSession, pbar: tqdm | None = None, ) -> RequestFuncOutput: _preprocess_clip(request_func_input) return await async_request_infinity_embeddings( request_func_input, session, pbar=pbar, ) # TODO: Add more request functions for different API protocols. ASYNC_REQUEST_FUNCS: dict[str, RequestFunc] = { "vllm": async_request_openai_completions, "openai": async_request_openai_completions, "openai-chat": async_request_openai_chat_completions, "openai-audio": async_request_openai_audio, "openai-embeddings": async_request_openai_embeddings, "openai-embeddings-chat": async_request_openai_embeddings_chat, "openai-embeddings-clip": async_request_openai_embeddings_clip, "openai-embeddings-vlm2vec": async_request_openai_embeddings_vlm2vec, # Infinity embedding server: https://github.com/michaelfeil/infinity "infinity-embeddings": async_request_infinity_embeddings, "infinity-embeddings-clip": async_request_infinity_embeddings_clip, # (Infinity embedding server does not support vlm2vec) "vllm-rerank": async_request_vllm_rerank, } OPENAI_COMPATIBLE_BACKENDS = [ k for k, v in ASYNC_REQUEST_FUNCS.items() if v in (async_request_openai_completions, async_request_openai_chat_completions) ]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/fusion.py
vllm/compilation/fusion.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any, NamedTuple import torch import torch._inductor.pattern_matcher as pm from torch import fx from torch._higher_order_ops.auto_functionalize import auto_functionalized from torch._inductor.pattern_matcher import PatternMatcherPass from torch._ops import OpOverload from vllm.config import VllmConfig, get_current_vllm_config from vllm.logger import init_logger from vllm.model_executor.layers.quantization.utils.quant_utils import ( GroupShape, QuantKey, ScaleDesc, kFp8Dynamic64Sym, kFp8Dynamic128Sym, kFp8DynamicTensorSym, kFp8DynamicTokenSym, kFp8StaticTensorSym, kNvfp4Quant, kStaticTensorScale, ) from vllm.platforms import current_platform from .inductor_pass import enable_fake_mode from .matcher_utils import ( MatcherFusedAddRMSNorm, MatcherQuantFP8, MatcherRMSNorm, ) from .vllm_inductor_pass import VllmInductorPass, VllmPatternMatcherPass logger = init_logger(__name__) FP8_DTYPE = current_platform.fp8_dtype() FP4_DTYPE = torch.uint8 def empty_bf16(*args, **kwargs): return torch.empty(*args, **kwargs, dtype=torch.bfloat16, device="cuda") def empty_fp32(*args, **kwargs): return torch.empty(*args, **kwargs, dtype=torch.float32, device="cuda") def empty_i32(*args, **kwargs): return torch.empty(*args, **kwargs, dtype=torch.int32, device="cuda") def empty_i64(*args, **kwargs): return torch.empty(*args, **kwargs, dtype=torch.int64, device="cuda") RMS_OP = torch.ops._C.rms_norm.default RMS_ADD_OP = torch.ops._C.fused_add_rms_norm.default QUANT_OPS: dict[QuantKey, OpOverload] = { kFp8StaticTensorSym: torch.ops._C.static_scaled_fp8_quant.default, # noqa: E501 kFp8DynamicTensorSym: torch.ops._C.dynamic_scaled_fp8_quant.default, # noqa: E501 kFp8DynamicTokenSym: torch.ops._C.dynamic_per_token_scaled_fp8_quant.default, # noqa: E501 } if current_platform.is_cuda() and hasattr(torch.ops._C, "scaled_fp4_quant"): QUANT_OPS[kNvfp4Quant] = torch.ops._C.scaled_fp4_quant.default if current_platform.is_cuda(): QUANT_OPS[kFp8Dynamic128Sym] = torch.ops._C.per_token_group_fp8_quant.default # noqa: E501 QUANT_OPS[kFp8Dynamic64Sym] = torch.ops._C.per_token_group_fp8_quant.default # noqa: E501 class FusedRMSQuantKey(NamedTuple): """ Named tuple for identifying the type of RMSNorm + quant fusion. quant: type of quantization fused_add: does the op also perform the residual add """ quant: QuantKey fused_add: bool def __str__(self): return ( f"FusedQuantKey({self.quant}, with" f"{'' if self.fused_add else 'out'} residual)" ) FUSED_OPS: dict[FusedRMSQuantKey, OpOverload] = { FusedRMSQuantKey( kFp8StaticTensorSym, False ): torch.ops._C.rms_norm_static_fp8_quant.default, # noqa: E501 FusedRMSQuantKey( kFp8StaticTensorSym, True ): torch.ops._C.fused_add_rms_norm_static_fp8_quant.default, # noqa: E501 FusedRMSQuantKey( kFp8DynamicTokenSym, False ): torch.ops._C.rms_norm_dynamic_per_token_quant.default, # noqa: E501 FusedRMSQuantKey( kFp8DynamicTokenSym, True ): torch.ops._C.rms_norm_dynamic_per_token_quant.default, # noqa: E501 FusedRMSQuantKey( kFp8Dynamic128Sym, False ): torch.ops._C.rms_norm_per_block_quant.default, # noqa: E501 FusedRMSQuantKey( kFp8Dynamic128Sym, True ): torch.ops._C.rms_norm_per_block_quant.default, # noqa: E501 FusedRMSQuantKey( kFp8Dynamic64Sym, False ): torch.ops._C.rms_norm_per_block_quant.default, # noqa: E501 FusedRMSQuantKey( kFp8Dynamic64Sym, True ): torch.ops._C.rms_norm_per_block_quant.default, # noqa: E501 } class RMSNormQuantPattern: def __init__( self, epsilon: float, key: FusedRMSQuantKey, has_col_major_scales: bool = False, is_e8m0: bool = False, ): self.epsilon = epsilon self.quant_dtype = key.quant.dtype config = get_current_vllm_config() self.model_dtype = config.model_config.dtype if config.model_config else None assert key in FUSED_OPS, f"unsupported fused rmsnorm+quant op for {key}" self.FUSED_OP = FUSED_OPS[key] self.rmsnorm_matcher = ( MatcherRMSNorm(epsilon) if not key.fused_add else MatcherFusedAddRMSNorm(epsilon) ) self.quant_matcher = MatcherQuantFP8( key.quant, has_col_major_scales=has_col_major_scales, is_e8m0=is_e8m0 ) class RMSNormStaticQuantPattern(RMSNormQuantPattern): def __init__(self, epsilon: float, quant_dtype: torch.dtype, symmetric=True): fused_key = FusedRMSQuantKey( fused_add=False, quant=QuantKey( dtype=quant_dtype, scale=kStaticTensorScale, symmetric=symmetric ), ) super().__init__(epsilon, fused_key) def register(self, pm_pass: PatternMatcherPass): # Cannot use methods, as the self argument affects tracing def pattern(input: torch.Tensor, weight: torch.Tensor, scale: torch.Tensor): result_rms = self.rmsnorm_matcher(input, weight) return self.quant_matcher(result_rms, scale)[0] def replacement(input: torch.Tensor, weight: torch.Tensor, scale: torch.Tensor): # In case we're matching native rms-norm, conversions might be # optimized out. We convert here just to be safe. input = input.to(dtype=self.model_dtype) result = torch.empty( input.shape, device=input.device, dtype=self.quant_dtype ) at = auto_functionalized( self.FUSED_OP, result=result, input=input, weight=weight, scale=scale, epsilon=self.epsilon, ) # result return at[1] inputs = [ # input, weight *self.rmsnorm_matcher.inputs(), self.quant_matcher.inputs()[1], # scale ] pattern(*inputs) pm.register_replacement(pattern, replacement, inputs, pm.fwd_only, pm_pass) class FusedAddRMSNormStaticQuantPattern(RMSNormQuantPattern): def __init__(self, epsilon: float, quant_dtype: torch.dtype, symmetric=True): key = FusedRMSQuantKey( fused_add=True, quant=QuantKey( dtype=quant_dtype, scale=kStaticTensorScale, symmetric=symmetric ), ) super().__init__(epsilon, key) def register(self, pm_pass: PatternMatcherPass): def pattern( input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor, scale: torch.Tensor, ): result_rms, residual = self.rmsnorm_matcher(input, weight, residual) result, _ = self.quant_matcher(result_rms, scale) return result, residual def replacement( input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor, scale: torch.Tensor, ): # In case we're matching native rms-norm, conversions might be # optimized out. We convert here just to be safe. input = input.to(dtype=self.model_dtype) result = torch.empty_like(input, dtype=self.quant_dtype) at = auto_functionalized( self.FUSED_OP, result=result, input=input, residual=residual, weight=weight, scale=scale, epsilon=self.epsilon, ) # result, residual return at[1], at[2] inputs = [ # input, weight, residual *self.rmsnorm_matcher.inputs(), self.quant_matcher.inputs()[1], # scale ] pm.register_replacement( pattern, replacement, inputs, pm.fwd_only, pm_pass, ) class FusedAddRMSNormGroupQuantPattern(RMSNormQuantPattern): def __init__( self, epsilon: float, quant_dtype: torch.dtype, group_shape: GroupShape, symmetric=True, has_col_major_scales: bool = False, is_e8m0: bool = False, ): scale = ScaleDesc(torch.float32, False, group_shape) key = FusedRMSQuantKey( fused_add=True, quant=QuantKey(dtype=quant_dtype, scale=scale, symmetric=symmetric), ) self.group_shape = group_shape self.has_col_major_scales = has_col_major_scales self.is_e8m0 = is_e8m0 super().__init__( epsilon, key, has_col_major_scales=has_col_major_scales, is_e8m0=is_e8m0 ) def register(self, pm_pass: PatternMatcherPass): def pattern(input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor): result_rms, residual = self.rmsnorm_matcher(input, weight, residual) result, scale = self.quant_matcher(result_rms) return result, residual, scale def replacement( input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor ): # In case we're matching native rms-norm, conversions might be # optimized out. We convert here just to be safe. input = input.to(dtype=self.model_dtype) result = torch.empty_like(input, dtype=self.quant_dtype) scale = self.quant_matcher.make_scale(input, self.has_col_major_scales) at = auto_functionalized( self.FUSED_OP, result=result, input=input, weight=weight, scale=scale, epsilon=self.epsilon, scale_ub=None, residual=residual, group_size=self.group_shape[1], is_scale_transposed=self.has_col_major_scales, ) # result, residual, scale return at[1], at[3], at[2] pm.register_replacement( pattern, replacement, self.rmsnorm_matcher.inputs(), pm.fwd_only, pm_pass, ) class RMSNormGroupQuantPattern(RMSNormQuantPattern): def __init__( self, epsilon: float, quant_dtype: torch.dtype, group_shape: GroupShape, symmetric=True, has_col_major_scales: bool = False, is_e8m0: bool = False, ): scale = ScaleDesc(torch.float32, False, group_shape) key = FusedRMSQuantKey( fused_add=False, quant=QuantKey(dtype=quant_dtype, scale=scale, symmetric=symmetric), ) self.group_shape = group_shape super().__init__( epsilon, key, has_col_major_scales=has_col_major_scales, is_e8m0=is_e8m0 ) def register(self, pm_pass: PatternMatcherPass): def pattern(input: torch.Tensor, weight: torch.Tensor): result_rms = self.rmsnorm_matcher(input, weight) result, scale = self.quant_matcher(result_rms) return result, scale def replacement(input: torch.Tensor, weight: torch.Tensor): # In case we're matching native rms-norm, conversions might be # optimized out. We convert here just to be safe. input = input.to(dtype=self.model_dtype) result = torch.empty_like(input, dtype=self.quant_dtype) scale = self.quant_matcher.make_scale( input, transposed=self.quant_matcher.has_col_major_scales ) at = auto_functionalized( self.FUSED_OP, result=result, input=input, weight=weight, scale=scale, epsilon=self.epsilon, scale_ub=None, residual=None, group_size=self.group_shape[1], is_scale_transposed=self.quant_matcher.has_col_major_scales, ) # result, scale return at[1], at[2] pm.register_replacement( pattern, replacement, self.rmsnorm_matcher.inputs(), pm.fwd_only, pm_pass, ) class RMSNormDynamicQuantPattern(RMSNormQuantPattern): def __init__( self, epsilon: float, quant_dtype: torch.dtype, group_shape: GroupShape = GroupShape.PER_TOKEN, symmetric=True, ): scale = ScaleDesc(torch.float32, False, group_shape) key = FusedRMSQuantKey( fused_add=False, quant=QuantKey(dtype=quant_dtype, scale=scale, symmetric=symmetric), ) super().__init__(epsilon, key) def register(self, pm_pass: PatternMatcherPass): def pattern(input: torch.Tensor, weight: torch.Tensor): result_rms = self.rmsnorm_matcher(input, weight) # result, scale return self.quant_matcher(result_rms) def replacement(input: torch.Tensor, weight: torch.Tensor): # In case we're matching native rms-norm, conversions might be # optimized out. We convert here just to be safe. input = input.to(dtype=self.model_dtype) result = torch.empty_like(input, dtype=self.quant_dtype) scale = self.quant_matcher.make_scale(input) at = auto_functionalized( self.FUSED_OP, result=result, input=input, weight=weight, scale=scale, epsilon=self.epsilon, scale_ub=None, residual=None, ) # result, scale return at[1], at[2] pm.register_replacement( pattern, replacement, self.rmsnorm_matcher.inputs(), pm.fwd_only, pm_pass, ) class FusedAddRMSNormDynamicQuantPattern(RMSNormQuantPattern): def __init__( self, epsilon: float, quant_dtype: torch.dtype, group_shape: GroupShape = GroupShape.PER_TOKEN, symmetric=True, ): scale = ScaleDesc(torch.float32, False, group_shape) key = FusedRMSQuantKey( fused_add=True, quant=QuantKey(dtype=quant_dtype, scale=scale, symmetric=symmetric), ) super().__init__(epsilon, key) def register(self, pm_pass: PatternMatcherPass): def pattern(input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor): result_rms, residual = self.rmsnorm_matcher(input, weight, residual) result, scale = self.quant_matcher(result_rms) return result, residual, scale def replacement( input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor ): # In case we're matching native rms-norm, conversions might be # optimized out. We convert here just to be safe. input = input.to(dtype=self.model_dtype) result = torch.empty_like(input, dtype=self.quant_dtype) scale = self.quant_matcher.make_scale(input) at = auto_functionalized( self.FUSED_OP, result=result, input=input, weight=weight, scale=scale, epsilon=self.epsilon, scale_ub=None, residual=residual, ) # result, residual, scale return at[1], at[3], at[2] pm.register_replacement( pattern, replacement, self.rmsnorm_matcher.inputs(), pm.fwd_only, pm_pass, ) class RMSNormQuantFusionPass(VllmPatternMatcherPass): """ This pass fuses rms_norm & quant custom ops into a fused rms_norm_quant op. It also supports fused_add_rms_norm. """ @enable_fake_mode def __init__(self, config: VllmConfig): super().__init__(config) self.patterns: PatternMatcherPass = PatternMatcherPass( pass_name="rmsnorm_quant_fusion_pass" ) # Make sure fused add patterns are before simple rms norm, # as the latter is a subset of the former in torch ops for epsilon in [1e-5, 1e-6]: # Fuse fused_add_rms_norm + static fp8 quant FusedAddRMSNormStaticQuantPattern(epsilon, FP8_DTYPE).register( self.patterns ) # Fuse rms_norm + static fp8 quant RMSNormStaticQuantPattern(epsilon, FP8_DTYPE).register(self.patterns) # Fuse fused_add_rms_norm + dynamic per-token fp8 quant FusedAddRMSNormDynamicQuantPattern(epsilon, FP8_DTYPE).register( self.patterns ) # Fuse rms_norm + dynamic per-token fp8 quant RMSNormDynamicQuantPattern(epsilon, FP8_DTYPE).register(self.patterns) # Only register group quant patterns on CUDA where the C++ op exists if current_platform.is_cuda(): for group_shape in [GroupShape(1, 128), GroupShape(1, 64)]: for has_col_major_scales in [True, False]: for is_e8m0 in [True, False]: # Fuse fused_add_rms_norm + fp8 group quant FusedAddRMSNormGroupQuantPattern( epsilon, FP8_DTYPE, group_shape=group_shape, has_col_major_scales=has_col_major_scales, is_e8m0=is_e8m0, ).register(self.patterns) # Fuse rms_norm + fp8 group quant RMSNormGroupQuantPattern( epsilon, FP8_DTYPE, group_shape=group_shape, has_col_major_scales=has_col_major_scales, is_e8m0=is_e8m0, ).register(self.patterns) self.dump_patterns(config, self.patterns) @VllmInductorPass.time_and_log def __call__(self, graph: fx.Graph): self.matched_count = self.patterns.apply(graph) logger.debug("Replaced %s patterns", self.matched_count) def uuid(self) -> Any: return self.hash_source( self, RMSNormGroupQuantPattern, RMSNormQuantPattern, RMSNormStaticQuantPattern, RMSNormDynamicQuantPattern, FusedAddRMSNormStaticQuantPattern, FusedAddRMSNormDynamicQuantPattern, FusedAddRMSNormGroupQuantPattern, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/partition_rules.py
vllm/compilation/partition_rules.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import contextlib import torch from vllm.logger import init_logger logger = init_logger(__name__) def should_split(node: torch.fx.Node, splitting_ops: list[str]) -> bool: """ Check if a node should be split for dynamo graph partition. It operates on dynamo graph, so the node.target can be anything. We need to check and split only on OpOverload and OpOverloadPacket. """ if node.op != "call_function": return False target = node.target if isinstance(target, torch._ops.OpOverloadPacket): # Example: "aten::add" return target._qualified_op_name in splitting_ops if isinstance(target, torch._ops.OpOverload): # Example: "aten::add" packet_name = target.name() # Example: "aten::add.default" op_overload_name = f"{packet_name}.{target._overloadname}" return op_overload_name in splitting_ops or packet_name in splitting_ops return False @contextlib.contextmanager def inductor_partition_rule_context(splitting_ops: list[str]): """Context manager to temporarily register Inductor partition rules. Registers custom partition rules for specified operators, forcing the Inductor scheduler to partition the graph at these operators. The rules are automatically restored to their previous state on exit. Args: splitting_ops: List of operator names to partition on. """ if not splitting_ops: logger.debug("No partition ops provided; skipping rule registration.") yield return # Save current state before registering saved_splitting_ops: list[str] = list( torch._inductor.config.custom_should_partition_ops ) torch._inductor.config.custom_should_partition_ops = splitting_ops logger.debug( "Registered inductor partition rules for %d operators", len(splitting_ops) ) try: yield finally: # Clear and restore previous state torch._inductor.config.custom_should_partition_ops = saved_splitting_ops logger.debug("Restored previous partition rules state.")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/compiler_interface.py
vllm/compilation/compiler_interface.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import contextlib import copy import os from collections.abc import Callable from contextlib import ExitStack from typing import Any, Literal from unittest.mock import patch import torch import torch._inductor.compile_fx import torch.fx as fx import vllm.envs as envs from vllm.compilation.counter import compilation_counter from vllm.config import VllmConfig from vllm.config.utils import Range from vllm.utils.hashing import safe_hash from vllm.utils.torch_utils import is_torch_equal_or_newer class CompilerInterface: """ The interface for a compiler that can be used by vLLM. """ # The name of the compiler, e.g. inductor. # This is a class-level attribute. name: str def initialize_cache( self, cache_dir: str, disable_cache: bool = False, prefix: str = "" ): """ when the vLLM process uses `cache_dir` as the cache directory, the compiler should initialize itself with the cache directory, e.g. by re-directing its own cache directory to a sub-directory. prefix can be used in combination with cache_dir to figure out the base cache directory, e.g. there're multiple parts of model being compiled, but we want to share the same cache directory for all of them. e.g. cache_dir = "/path/to/dir/backbone", prefix = "backbone" cache_dir = "/path/to/dir/eagle_head", prefix = "eagle_head" """ pass def compute_hash(self, vllm_config: VllmConfig) -> str: """ Gather all the relevant information from the vLLM config, to compute a hash so that we can cache the compiled model. See [`VllmConfig.compute_hash`][vllm.config.VllmConfig.compute_hash] to check what information is already considered by default. This function should only consider the information that is specific to the compiler. """ return "" def compile( self, graph: fx.GraphModule, example_inputs: list[Any], compiler_config: dict[str, Any], compile_range: Range, key: str | None = None, ) -> tuple[Callable | None, Any | None]: """ Compile the graph with the given example inputs and compiler config, with a range. The `compile_range` specifies the range of the inputs, it could be concrete size (if compile_sizes is provided), e.g. [4, 4] or a range [5, 8]. Right now we only support one variable in ranges for all inputs, which is the batchsize (number of tokens) during inference. Dynamo will make sure `graph(*example_inputs)` is valid. The function should return a compiled callable function, as well as a handle that can be used to directly load the compiled function. The handle should be a plain Python object, preferably a string or a file path for readability. If the compiler doesn't support caching, it should return None for the handle. If the compiler fails to compile the graph, it should return None for the compiled function as well. `key` is required for StandaloneInductorAdapter, it specifies where to save the compiled artifact. The compiled artifact gets saved to `cache_dir/key`. """ return None, None def load( self, handle: Any, graph: fx.GraphModule, example_inputs: list[Any], graph_index: int, compile_range: Range, ) -> Callable: """ Load the compiled function from the handle. Raises an error if the handle is invalid. The handle is the second return value of the `compile` function. """ raise NotImplementedError("caching is not supported") class AlwaysHitShapeEnv: """ Why do we need this class: For normal `torch.compile` usage, every compilation will have one Dynamo bytecode compilation and one Inductor compilation. The Inductor compilation happens under the context of the Dynamo bytecode compilation, and that context is used to determine the dynamic shape information, etc. For our use case, we only run Dynamo bytecode compilation once, and run Inductor compilation multiple times with different shapes plus a general shape. The compilation for specific shapes happens outside of the context of the Dynamo bytecode compilation. At that time, we don't have shape environment to provide to Inductor, and it will fail the Inductor code cache lookup. By providing a dummy shape environment that always hits, we can make the Inductor code cache lookup always hit, and we can compile the graph for different shapes as needed. The following dummy methods are obtained by trial-and-error until it works. """ def __init__(self) -> None: self.guards: list[Any] = [] def evaluate_guards_expression(self, *args, **kwargs): return True def get_pruned_guards(self, *args, **kwargs): return [] def produce_guards_expression(self, *args, **kwargs): return "" def get_inductor_factors() -> list[Any]: factors: list[Any] = [] # summarize system state from torch._inductor.codecache import CacheBase system_factors = CacheBase.get_system() factors.append(system_factors) # summarize pytorch state from torch._inductor.codecache import torch_key torch_factors = torch_key() factors.append(torch_factors) return factors def is_compile_cache_enabled( vllm_additional_inductor_config: dict[str, Any], ) -> bool: vllm_inductor_config_disable_cache = vllm_additional_inductor_config.get( "force_disable_caches", False ) # TODO(gmagogsfm): Replace torch._inductor.config.force_disable_caches # with torch.compiler.config.force_disable_caches when minimum PyTorch # version reaches 2.10 return ( not envs.VLLM_DISABLE_COMPILE_CACHE and not torch._inductor.config.force_disable_caches and not vllm_inductor_config_disable_cache ) class InductorStandaloneAdaptor(CompilerInterface): """ The adaptor for the Inductor compiler. Requires PyTorch 2.8+. This is not on by default yet, but we plan to turn it on by default for PyTorch 2.8. Use VLLM_USE_STANDALONE_COMPILE to toggle this on or off. """ name = "inductor_standalone" def __init__(self, save_format: Literal["binary", "unpacked"]): self.save_format = save_format def compute_hash(self, vllm_config: VllmConfig) -> str: factors = get_inductor_factors() hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest()[ :10 ] return hash_str def initialize_cache( self, cache_dir: str, disable_cache: bool = False, prefix: str = "" ): self.cache_dir = cache_dir def compile( self, graph: fx.GraphModule, example_inputs: list[Any], compiler_config: dict[str, Any], compile_range: Range, key: str | None = None, ) -> tuple[Callable | None, Any | None]: compilation_counter.num_inductor_compiles += 1 current_config = {} if compiler_config is not None: current_config.update(compiler_config) set_inductor_config(current_config, compile_range) set_functorch_config() if compile_range.is_single_size(): dynamic_shapes = "from_example_inputs" else: dynamic_shapes = "from_graph" from torch._inductor import standalone_compile compiled_graph = standalone_compile( graph, example_inputs, dynamic_shapes=dynamic_shapes, options={"config_patches": current_config}, ) # Save the compiled artifact to disk in the specified path assert key is not None path = os.path.join(self.cache_dir, key) if is_compile_cache_enabled(compiler_config): compiled_graph.save(path=path, format=self.save_format) compilation_counter.num_compiled_artifacts_saved += 1 return compiled_graph, (key, path) def load( self, handle: Any, graph: fx.GraphModule, example_inputs: list[Any], graph_index: int, compile_range: Range, ) -> Callable: assert isinstance(handle, tuple) assert isinstance(handle[0], str) assert isinstance(handle[1], str) path = handle[1] inductor_compiled_graph = torch._inductor.CompiledArtifact.load( path=path, format=self.save_format ) from torch._inductor.compile_fx import graph_returns_tuple returns_tuple = graph_returns_tuple(graph) def compiled_graph_wrapper(*args): graph_output = inductor_compiled_graph(*args) # unpack the tuple if needed # TODO(rzou): the implication is that we're not # reading the python bytecode correctly in vLLM? if returns_tuple: return graph_output else: return graph_output[0] return compiled_graph_wrapper class InductorAdaptor(CompilerInterface): """ The adaptor for the Inductor compiler, version 2.5, 2.6, 2.7. """ name = "inductor" def compute_hash(self, vllm_config: VllmConfig) -> str: factors = get_inductor_factors() hash_str = safe_hash(str(factors).encode(), usedforsecurity=False).hexdigest()[ :10 ] return hash_str def initialize_cache( self, cache_dir: str, disable_cache: bool = False, prefix: str = "" ): self.cache_dir = cache_dir self.prefix = prefix self.base_cache_dir = cache_dir[: -len(prefix)] if prefix else cache_dir if disable_cache: return # redirect the cache directory to a subdirectory # set flags so that Inductor and Triton store their cache # in the cache_dir, then users only need to copy the cache_dir # to another machine to reuse the cache. inductor_cache = os.path.join(self.base_cache_dir, "inductor_cache") os.makedirs(inductor_cache, exist_ok=True) os.environ["TORCHINDUCTOR_CACHE_DIR"] = inductor_cache triton_cache = os.path.join(self.base_cache_dir, "triton_cache") os.makedirs(triton_cache, exist_ok=True) os.environ["TRITON_CACHE_DIR"] = triton_cache def compile( self, graph: fx.GraphModule, example_inputs: list[Any], compiler_config: dict[str, Any], compile_range: Range, key: str | None = None, ) -> tuple[Callable | None, Any | None]: compilation_counter.num_inductor_compiles += 1 from torch._inductor.compile_fx import compile_fx current_config = {} if compiler_config is not None: current_config.update(compiler_config) # disable remote cache current_config["fx_graph_cache"] = True current_config["fx_graph_remote_cache"] = False set_inductor_config(current_config, compile_range) set_functorch_config() # inductor can inplace modify the graph, so we need to copy it # see https://github.com/pytorch/pytorch/issues/138980 graph = copy.deepcopy(graph) # it's the first time we compile this graph # the assumption is that we don't have nested Inductor compilation. # compiled_fx_graph_hash will only be called once, and we can hook # it to get the hash of the compiled graph directly. hash_str, file_path = None, None from torch._inductor.codecache import FxGraphCache, compiled_fx_graph_hash if torch.__version__.startswith("2.5"): original_load = FxGraphCache.load original_load_name = "torch._inductor.codecache.FxGraphCache.load" def hijack_load(*args, **kwargs): inductor_compiled_graph = original_load(*args, **kwargs) nonlocal file_path compiled_fn = inductor_compiled_graph.current_callable file_path = compiled_fn.__code__.co_filename # noqa if ( not file_path.startswith(self.base_cache_dir) and compiled_fn.__closure__ is not None ): # hooked in the align_inputs_from_check_idxs function # in torch/_inductor/utils.py for cell in compiled_fn.__closure__: if not callable(cell.cell_contents): continue if cell.cell_contents.__code__.co_filename.startswith( self.base_cache_dir ): # this is the real file path compiled from Inductor file_path = cell.cell_contents.__code__.co_filename break return inductor_compiled_graph hijacked_compile_fx_inner = torch._inductor.compile_fx.compile_fx_inner # noqa elif torch.__version__ >= "2.6": # function renamed in 2.6 original_load_name = None def hijacked_compile_fx_inner(*args, **kwargs): output = torch._inductor.compile_fx.compile_fx_inner(*args, **kwargs) nonlocal hash_str inductor_compiled_graph = output if inductor_compiled_graph is not None: nonlocal file_path compiled_fn = inductor_compiled_graph.current_callable file_path = compiled_fn.__code__.co_filename # noqa if ( not file_path.startswith(self.base_cache_dir) and compiled_fn.__closure__ is not None ): # hooked in the align_inputs_from_check_idxs function # in torch/_inductor/utils.py for cell in compiled_fn.__closure__: if not callable(cell.cell_contents): continue code = cell.cell_contents.__code__ if code.co_filename.startswith(self.base_cache_dir): # this is the real file path # compiled from Inductor file_path = code.co_filename break hash_str = inductor_compiled_graph._fx_graph_cache_key return output def hijack_compiled_fx_graph_hash(*args, **kwargs): out = compiled_fx_graph_hash(*args, **kwargs) nonlocal hash_str hash_str = out[0] return out def _check_can_cache(*args, **kwargs): # no error means it can be cached. # Inductor refuses to cache the graph outside of Dynamo # tracing context, and also disables caching for graphs # with high-order ops. # For vLLM, in either case, we want to cache the graph. # see https://github.com/pytorch/pytorch/blob/9f5ebf3fc609105a74eab4ccc24932d6353ff566/torch/_inductor/codecache.py#L1221 # noqa return def _get_shape_env() -> AlwaysHitShapeEnv: return AlwaysHitShapeEnv() with ExitStack() as stack: # hijack to get the compiled graph itself if original_load_name is not None: stack.enter_context(patch(original_load_name, hijack_load)) # for hijacking the hash of the compiled graph stack.enter_context( patch( "torch._inductor.codecache.compiled_fx_graph_hash", hijack_compiled_fx_graph_hash, ) ) # for providing a dummy shape environment stack.enter_context( patch( "torch._inductor.codecache.FxGraphCache._get_shape_env", _get_shape_env, ) ) from torch._functorch._aot_autograd.autograd_cache import AOTAutogradCache # torch 2.8+ on main uses _get_shape_env in AOTAutogradCache if hasattr(AOTAutogradCache, "_get_shape_env"): stack.enter_context( patch( "torch._functorch._aot_autograd.autograd_cache.AOTAutogradCache._get_shape_env", _get_shape_env, ) ) # for forcing the graph to be cached stack.enter_context( patch( "torch._inductor.codecache.FxGraphCache._check_can_cache", _check_can_cache, ) ) # Dynamo metrics context, see method for more details. stack.enter_context(self.metrics_context()) # Disable remote caching. When these are on, on remote cache-hit, # the monkey-patched functions never actually get called. # vLLM today assumes and requires the monkey-patched functions to # get hit. # TODO(zou3519): we're going to replace this all with # standalone_compile sometime. if is_torch_equal_or_newer("2.6"): stack.enter_context( torch._inductor.config.patch(fx_graph_remote_cache=False) ) # InductorAdaptor (unfortunately) requires AOTAutogradCache # to be turned off to run. It will fail to acquire the hash_str # and error if not. # StandaloneInductorAdaptor (PyTorch 2.8+) fixes this problem. stack.enter_context( torch._functorch.config.patch(enable_autograd_cache=False) ) stack.enter_context( torch._functorch.config.patch(enable_remote_autograd_cache=False) ) compiled_graph = compile_fx( graph, example_inputs, inner_compile=hijacked_compile_fx_inner, config_patches=current_config, ) # Turn off the checks if we disable the compilation cache. if is_compile_cache_enabled(compiler_config): if hash_str is None: raise RuntimeError( "vLLM failed to compile the model. The most " "likely reason for this is that a previous compilation " "failed, leading to a corrupted compilation artifact. " "We recommend trying to " "remove ~/.cache/vllm/torch_compile_cache and try again " "to see the real issue. " ) assert file_path is not None, ( "failed to get the file path of the compiled graph" ) return compiled_graph, (hash_str, file_path) def load( self, handle: Any, graph: fx.GraphModule, example_inputs: list[Any], graph_index: int, compile_range: Range, ) -> Callable: assert isinstance(handle, tuple) assert isinstance(handle[0], str) assert isinstance(handle[1], str) hash_str = handle[0] from torch._functorch._aot_autograd.autograd_cache import AOTAutogradCache from torch._inductor.codecache import FxGraphCache with ExitStack() as exit_stack: exit_stack.enter_context( patch( "torch._inductor.codecache.FxGraphCache._get_shape_env", lambda *args, **kwargs: AlwaysHitShapeEnv(), ) ) # torch 2.8+ on main uses _get_shape_env in AOTAutogradCache if hasattr(AOTAutogradCache, "_get_shape_env"): exit_stack.enter_context( patch( "torch._functorch._aot_autograd.autograd_cache.AOTAutogradCache._get_shape_env", lambda *args, **kwargs: AlwaysHitShapeEnv(), ) ) # Dynamo metrics context, see method for more details. exit_stack.enter_context(self.metrics_context()) if torch.__version__.startswith("2.5"): inductor_compiled_graph = FxGraphCache._lookup_graph( hash_str, example_inputs, True, False ) assert inductor_compiled_graph is not None, ( "Inductor cache lookup failed. Please remove" f"the cache directory and try again." # noqa ) elif torch.__version__ >= "2.6": from torch._inductor.output_code import CompiledFxGraphConstantsWithGm constants = CompiledFxGraphConstantsWithGm(graph) inductor_compiled_graph, _ = FxGraphCache._lookup_graph( hash_str, example_inputs, True, None, constants ) assert inductor_compiled_graph is not None, ( "Inductor cache lookup failed. Please remove" f"the cache directory and try again." # noqa ) # Inductor calling convention (function signature): # f(list) -> tuple # Dynamo calling convention (function signature): # f(*args) -> Any # need to know if the graph returns a tuple from torch._inductor.compile_fx import graph_returns_tuple returns_tuple = graph_returns_tuple(graph) # this is the callable we return to Dynamo to run def compiled_graph(*args): # convert args to list list_args = list(args) graph_output = inductor_compiled_graph(list_args) # unpack the tuple if needed if returns_tuple: return graph_output else: return graph_output[0] return compiled_graph def metrics_context(self) -> contextlib.AbstractContextManager: """ This method returns the Dynamo metrics context (if it exists, otherwise a null context). It is used by various compile components. Present in torch>=2.6, it's used inside FxGraphCache in torch==2.6 (but not after). It might also be used in various other torch.compile internal functions. Because it is re-entrant, we always set it (even if entering via Dynamo and the context was already entered). We might want to revisit if it should be set at a different mode of compilation. This is likely a bug in PyTorch: public APIs should not rely on manually setting up internal contexts. But we also rely on non-public APIs which might not provide these guarantees. """ if is_torch_equal_or_newer("2.6"): import torch._dynamo.utils return torch._dynamo.utils.get_metrics_context() else: return contextlib.nullcontext() def set_inductor_config(config, compile_range: Range): if compile_range.is_single_size(): # for a specific batch size, tuning triton kernel parameters # can be beneficial config["max_autotune"] = envs.VLLM_ENABLE_INDUCTOR_MAX_AUTOTUNE config["coordinate_descent_tuning"] = ( envs.VLLM_ENABLE_INDUCTOR_COORDINATE_DESCENT_TUNING ) def set_functorch_config(): torch._functorch.config.bundled_autograd_cache = False class EagerAdaptor(CompilerInterface): name = "eager" def compile( self, graph: fx.GraphModule, example_inputs: list[Any], compiler_config: dict[str, Any], compile_range: Range, key: str | None = None, ) -> tuple[Callable | None, Any | None]: compilation_counter.num_eager_compiles += 1 # we don't need to compile the graph, just return the graph itself. # It does not support caching, return None for the handle. return graph, None
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/fusion_attn.py
vllm/compilation/fusion_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import ABC, abstractmethod from collections.abc import Callable import torch import torch._inductor.pattern_matcher as pm from torch import fx from torch._higher_order_ops.auto_functionalize import auto_functionalized from torch._inductor.pattern_matcher import PatternMatcherPass from vllm.attention.layer import Attention from vllm.config import VllmConfig, get_layers_from_vllm_config from vllm.logger import init_logger from vllm.model_executor.layers.quantization.utils.quant_utils import ( QuantKey, kNvfp4Quant, kStaticTensorScale, ) from vllm.platforms import current_platform from vllm.utils.math_utils import round_up from .fusion import QUANT_OPS, empty_bf16, empty_fp32, empty_i32 from .fx_utils import is_func from .inductor_pass import enable_fake_mode from .matcher_utils import MatcherQuantFP8 from .vllm_inductor_pass import VllmInductorPass, VllmPatternMatcherPass logger = init_logger(__name__) FP8_DTYPE = current_platform.fp8_dtype() FP4_DTYPE = torch.uint8 ATTN_OP = torch.ops.vllm.unified_attention_with_output.default RESHAPE_OP = torch.ops.aten.reshape.default class AttentionQuantPattern(ABC): """ The base class for Attn+Quant fusions. Should not be used directly. """ def __init__( self, layer: Attention, quant_key: QuantKey, dtype: torch.dtype, ): self.layer = layer self.layer_name = layer.layer_name self.num_heads = layer.num_heads self.head_size = layer.head_size self.quant_key = quant_key self.quant_dtype = quant_key.dtype self.dtype = dtype assert self.quant_key in QUANT_OPS, ( f"unsupported quantization scheme {self.quant_key}" ) self.QUANT_OP = QUANT_OPS[self.quant_key] def empty(self, *args, **kwargs): kwargs = {"dtype": self.dtype, "device": "cuda", **kwargs} return torch.empty(*args, **kwargs) def empty_quant(self, *args, **kwargs): kwargs = {"dtype": self.quant_dtype, "device": "cuda", **kwargs} return torch.empty(*args, **kwargs) @staticmethod def wrap_trace_fn(trace_fn, *process_fx_fns: Callable[[fx.GraphModule], None]): def wrapped(*args, **kwargs): gm = trace_fn(*args, **kwargs) for process_fx in process_fx_fns: process_fx(gm) return gm return wrapped @staticmethod def fx_view_to_reshape(gm: torch.fx.GraphModule): from torch._inductor.fx_passes.post_grad import view_to_reshape view_to_reshape(gm) @staticmethod def remove_noop_permutes(gm: torch.fx.GraphModule): for node in gm.graph.nodes: if not is_func(node, torch.ops.aten.permute.default): continue dims = node.args[1] if any(dim != i for i, dim in enumerate(dims)): continue # this is now an identity op, remove node.replace_all_uses_with(node.args[0]) gm.graph.erase_node(node) def register_if_supported(self, pm_pass: PatternMatcherPass): if self.layer.impl.fused_output_quant_supported(self.quant_key): self._register(pm_pass) @abstractmethod def _register(self, pm_pass: PatternMatcherPass): raise NotImplementedError class AttentionFp8StaticQuantPattern(AttentionQuantPattern): """ Fusion for Attention+Fp8StaticQuant. Only triggers when the attention implementation returns True in `fused_output_quant_supported()`. If the pattern is found, the Fp8StaticQuant op will be removed from the graph, and its scale will be passed into Attention op as the `output_scale` argument. """ def __init__( self, layer: Attention, dtype: torch.dtype, symmetric: bool = True, ): quant_key = QuantKey( dtype=FP8_DTYPE, scale=kStaticTensorScale, symmetric=symmetric ) super().__init__(layer, quant_key, dtype) self.quant_matcher = MatcherQuantFP8(quant_key) def _register(self, pm_pass: PatternMatcherPass): def pattern( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, output_attn: torch.Tensor, scale: torch.Tensor, ): at1 = auto_functionalized( ATTN_OP, query=q, key=k, value=v, output=output_attn, layer_name=self.layer_name, output_scale=None, output_block_scale=None, ) attn_out_view = RESHAPE_OP( at1[1], [q.shape[0], self.num_heads * self.head_size] ) return self.quant_matcher(attn_out_view, scale)[0] def replacement( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, output_attn: torch.Tensor, scale: torch.Tensor, ): # attn output in quant_dtype output_attn = torch.ops.aten.full.default( [q.shape[0], self.num_heads, self.head_size], 0.0, dtype=self.quant_dtype, device=q.device, ) at1 = auto_functionalized( ATTN_OP, query=q, key=k, value=v, output=output_attn, layer_name=self.layer_name, output_scale=scale, output_block_scale=None, ) return RESHAPE_OP(at1[1], [-1, self.num_heads * self.head_size]) inputs = [ self.empty(5, self.num_heads, self.head_size), # q self.empty(5, self.num_heads, self.head_size), # k self.empty(5, self.num_heads, self.head_size), # v self.empty(5, self.num_heads, self.head_size), # attn_output empty_fp32(1, 1), # scale ] pm.register_replacement( pattern, replacement, inputs, AttentionQuantPattern.wrap_trace_fn( pm.fwd_only, AttentionQuantPattern.fx_view_to_reshape, AttentionQuantPattern.remove_noop_permutes, ), pm_pass, ) class AttentionNvfp4QuantPattern(AttentionQuantPattern): """ Fusion for Attention+Nvfp4Quant. Only triggers when the attention implementation returns True in `fused_output_quant_supported()`. If the pattern is found, the Nvfp4Quant op will be removed from the graph, and its scale will be passed into Attention op as the `output_scale` argument. """ def __init__(self, layer: Attention, dtype: torch.dtype): super().__init__(layer, kNvfp4Quant, dtype) def _register(self, pm_pass: PatternMatcherPass): def pattern( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, output_attn: torch.Tensor, output_quant: torch.Tensor, output_scale: torch.Tensor, input_scale: torch.Tensor, ): at1 = auto_functionalized( ATTN_OP, query=q, key=k, value=v, output=output_attn, layer_name=self.layer_name, output_scale=None, output_block_scale=None, ) attn_out_view = RESHAPE_OP( at1[1], [q.shape[0], self.num_heads * self.head_size] ) at2 = auto_functionalized( self.QUANT_OP, output=output_quant, input=attn_out_view, output_scale=output_scale, input_scale=input_scale, ) output_scale_view = torch.ops.aten.view.dtype(at2[2], FP8_DTYPE) return at2[1], output_scale_view def replacement( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, output_attn: torch.Tensor, output_quant: torch.Tensor, output_scale: torch.Tensor, input_scale: torch.Tensor, ): # attention output in quant_dtype output_attn = torch.ops.aten.full.default( [q.shape[0], self.num_heads, self.head_size // 2], 0.0, dtype=self.quant_dtype, device=q.device, ) # attention output block scale output_scale_view = torch.ops.aten.view.dtype(output_scale, FP8_DTYPE) at2 = auto_functionalized( ATTN_OP, query=q, key=k, value=v, output=output_attn, layer_name=self.layer_name, output_scale=input_scale, output_block_scale=output_scale_view, ) output = RESHAPE_OP(at2[1], [-1, self.num_heads * self.head_size // 2]) return output, at2[2] inputs = [ empty_bf16(5, self.num_heads, self.head_size), # q empty_bf16(5, self.num_heads, self.head_size), # k empty_bf16(5, self.num_heads, self.head_size), # v empty_bf16(5, self.num_heads, self.head_size), # output_attn self.empty_quant(5, self.num_heads * self.head_size // 2), # output_quant empty_i32( 128, round_up(self.num_heads * self.head_size // 16, 4) ), # output_scale empty_fp32(1, 1), # input_scale ] pm.register_replacement( pattern, replacement, inputs, AttentionQuantPattern.wrap_trace_fn( pm.fwd_only, AttentionQuantPattern.fx_view_to_reshape, AttentionQuantPattern.remove_noop_permutes, ), pm_pass, ) class AttnFusionPass(VllmPatternMatcherPass): """ This pass fuses post-attention quantization onto attention if supported. It uses the pattern matcher and matches each layer manually, as strings cannot be wildcarded. This also lets us check support on attention layers upon registration instead of during pattern matching. Currently, only static fp8 quant is supported, but patterns could easily be added for other quant schemes and dtypes. The bigger hurdle for wider support are attention kernels, which need to support fusing output quant. """ @enable_fake_mode def __init__(self, config: VllmConfig): super().__init__(config) self.patterns = PatternMatcherPass(pass_name="attn_fusion_pass") attn_layers = get_layers_from_vllm_config(config, Attention) for layer_name, layer in attn_layers.items(): pattern_fp8 = AttentionFp8StaticQuantPattern( layer, config.model_config.dtype ) pattern_fp8.register_if_supported(self.patterns) if current_platform.is_cuda() and hasattr(torch.ops._C, "scaled_fp4_quant"): pattern_nvfp4 = AttentionNvfp4QuantPattern( layer, config.model_config.dtype ) pattern_nvfp4.register_if_supported(self.patterns) if len(attn_layers) == 0: logger.warning( "Attention + quant fusion is enabled, but no attention layers " "were found in CompilationConfig.static_forward_context " "so no fusion patterns were registered." ) self.dump_patterns(config, self.patterns) @VllmInductorPass.time_and_log def __call__(self, graph: torch.fx.graph.Graph) -> None: self.matched_count = self.patterns.apply(graph) logger.debug("Fused quant onto %s attention nodes", self.matched_count) def uuid(self): return VllmInductorPass.hash_source( self, AttentionQuantPattern, AttentionFp8StaticQuantPattern, AttentionNvfp4QuantPattern, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/backends.py
vllm/compilation/backends.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import ast import dataclasses import hashlib import json import operator import os import pprint import time from collections.abc import Callable, Sequence from contextlib import contextmanager from copy import deepcopy from functools import partial from typing import Any import torch import torch.fx as fx from torch._dispatch.python import enable_python_dispatcher import vllm.envs as envs from vllm.compilation.inductor_pass import pass_context from vllm.compilation.partition_rules import ( inductor_partition_rule_context, should_split, ) from vllm.config import CompilationConfig, CUDAGraphMode, VllmConfig from vllm.config.compilation import DynamicShapesType from vllm.config.utils import Range, hash_factors from vllm.logger import init_logger from vllm.logging_utils import lazy from vllm.platforms import current_platform from vllm.utils.import_utils import resolve_obj_by_qualname from vllm.utils.torch_utils import is_torch_equal_or_newer from .caching import VllmSerializableFunction from .compiler_interface import ( CompilerInterface, EagerAdaptor, InductorAdaptor, InductorStandaloneAdaptor, is_compile_cache_enabled, ) from .counter import compilation_counter from .inductor_pass import InductorPass from .pass_manager import PostGradPassManager logger = init_logger(__name__) def make_compiler(compilation_config: CompilationConfig) -> CompilerInterface: if compilation_config.backend == "inductor": # Use standalone compile only if requested, version is new enough, # and the symbol actually exists in this PyTorch build. if ( envs.VLLM_USE_STANDALONE_COMPILE and is_torch_equal_or_newer("2.8.0.dev") and hasattr(torch._inductor, "standalone_compile") ): logger.debug("Using InductorStandaloneAdaptor") return InductorStandaloneAdaptor( compilation_config.compile_cache_save_format ) else: logger.debug("Using InductorAdaptor") return InductorAdaptor() elif compilation_config.backend == "eager": logger.debug("Using EagerAdaptor") return EagerAdaptor() else: logger.debug("Using custom backend: %s", compilation_config.backend) compiler = resolve_obj_by_qualname(current_platform.get_compile_backend())() assert isinstance(compiler, CompilerInterface) return compiler class CompilerManager: """ A manager to manage the compilation process, including caching the compiled graph, loading the compiled graph, and compiling the graph. The cache is a dict mapping `(runtime_shape, graph_index, backend_name)` to `any_data` returned from the compiler. When serializing the cache, we save it to a Python file for readability. We don't use json here because json doesn't support int as key. """ def __init__(self, compilation_config: CompilationConfig): self.cache: dict[tuple[Range, int, str], Any] = dict() self.is_cache_updated = False self.compilation_config = compilation_config self.compiler = make_compiler(compilation_config) def compute_hash(self, vllm_config: VllmConfig) -> str: return self.compiler.compute_hash(vllm_config) @contextmanager def compile_context(self, compile_range: Range): """Provide compilation context for the duration of compilation to set any torch global properties we want to scope to a single Inductor compilation (e.g. partition rules, pass context).""" with pass_context(compile_range): if self.compilation_config.use_inductor_graph_partition: with inductor_partition_rule_context( self.compilation_config.splitting_ops ): yield else: yield def initialize_cache( self, cache_dir: str, disable_cache: bool = False, prefix: str = "" ): """ Initialize the cache directory for the compiler. The organization of the cache directory is as follows: cache_dir=/path/to/hash_str/rank_i_j/prefix/ inside cache_dir, there will be: - vllm_compile_cache.py - computation_graph.py - transformed_code.py for multiple prefixes, they can share the same base cache dir of /path/to/hash_str/rank_i_j/ , to store some common compilation artifacts. """ self.disable_cache = disable_cache self.cache_dir = cache_dir self.cache_file_path = os.path.join(cache_dir, "vllm_compile_cache.py") if not disable_cache and os.path.exists(self.cache_file_path): # load the cache from the file with open(self.cache_file_path) as f: # we use ast.literal_eval to parse the data # because it is a safe way to parse Python literals. # do not use eval(), it is unsafe. cache = ast.literal_eval(f.read()) def check_type(value, ty): if not isinstance(value, ty): raise TypeError(f"Expected {ty} but got {type(value)} for {value}") def parse_key(key: Any) -> tuple[Range, int, str]: range_tuple, graph_index, compiler_name = key check_type(graph_index, int) check_type(compiler_name, str) if isinstance(range_tuple, tuple): start, end = range_tuple check_type(start, int) check_type(end, int) range_tuple = Range(start=start, end=end) check_type(range_tuple, Range) return range_tuple, graph_index, compiler_name self.cache = {parse_key(key): value for key, value in cache.items()} self.compiler.initialize_cache( cache_dir=cache_dir, disable_cache=disable_cache, prefix=prefix ) def save_to_file(self): if self.disable_cache or not self.is_cache_updated: return printer = pprint.PrettyPrinter(indent=4) data = printer.pformat(self.cache) with open(self.cache_file_path, "w") as f: f.write(data) def load( self, graph: fx.GraphModule, example_inputs: list[Any], graph_index: int, compile_range: Range, ) -> Callable | None: if (compile_range, graph_index, self.compiler.name) not in self.cache: return None handle = self.cache[(compile_range, graph_index, self.compiler.name)] compiled_graph = self.compiler.load( handle, graph, example_inputs, graph_index, compile_range ) logger.debug( "Directly load the %s-th graph for compile range %sfrom %s via handle %s", graph_index, str(compile_range), self.compiler.name, handle, ) return compiled_graph def compile( self, graph: fx.GraphModule, example_inputs, additional_inductor_config, compilation_config: CompilationConfig, compile_range: Range, graph_index: int = 0, num_graphs: int = 1, ) -> Any: if graph_index == 0: # before compiling the first graph, record the start time global compilation_start_time compilation_start_time = time.time() compilation_counter.num_backend_compilations += 1 compiled_graph = None # try to load from the cache compiled_graph = self.load(graph, example_inputs, graph_index, compile_range) if compiled_graph is not None: if graph_index == num_graphs - 1: # after loading the last graph for this shape, record the time. # there can be multiple graphs due to piecewise compilation. now = time.time() elapsed = now - compilation_start_time compilation_config.compilation_time += elapsed logger.info( "Directly load the compiled graph(s) for compile range %s " "from the cache, took %.3f s", str(compile_range), elapsed, ) return compiled_graph # no compiler cached the graph, or the cache is disabled, # we need to compile it if isinstance(self.compiler, InductorAdaptor): # Let compile_fx generate a key for us maybe_key = None else: maybe_key = "artifact_compile_range_" maybe_key += f"{compile_range.start}_{compile_range.end}" maybe_key += f"_subgraph_{graph_index}" with self.compile_context(compile_range): compiled_graph, handle = self.compiler.compile( graph, example_inputs, additional_inductor_config, compile_range, maybe_key, ) assert compiled_graph is not None, "Failed to compile the graph" # store the artifact in the cache if is_compile_cache_enabled(additional_inductor_config) and handle is not None: self.cache[(compile_range, graph_index, self.compiler.name)] = handle compilation_counter.num_cache_entries_updated += 1 self.is_cache_updated = True if graph_index == 0: # adds some info logging for the first graph logger.info_once( "Cache the graph of compile range %s for later use", str(compile_range), ) logger.debug( "Store the %s-th graph for compile range%s from %s via handle %s", graph_index, str(compile_range), self.compiler.name, handle, ) # after compiling the last graph, record the end time if graph_index == num_graphs - 1: now = time.time() elapsed = now - compilation_start_time compilation_config.compilation_time += elapsed logger.info_once( "Compiling a graph for compile range %s takes %.2f s", str(compile_range), elapsed, scope="local", ) return compiled_graph @dataclasses.dataclass class SplitItem: submod_name: str graph_id: int is_splitting_graph: bool graph: fx.GraphModule def split_graph( graph: fx.GraphModule, splitting_ops: list[str] ) -> tuple[fx.GraphModule, list[SplitItem]]: # split graph by ops subgraph_id = 0 node_to_subgraph_id: dict[fx.Node, int] = {} split_op_graphs: list[int] = [] for node in graph.graph.nodes: if node.op in ("output", "placeholder"): continue # Check if this is a getitem operation on a node from an earlier subgraph. # If so, assign it to the same subgraph as its input to avoid passing entire # tuple as input to submodules, which is against standalone_compile and # AoTAutograd input requirement. if node.op == "call_function" and node.target == operator.getitem: # Assign this getitem to the same subgraph as its input input_node = node.args[0] if input_node.op != "placeholder": assert input_node in node_to_subgraph_id node_to_subgraph_id[node] = node_to_subgraph_id[input_node] continue if should_split(node, splitting_ops): subgraph_id += 1 node_to_subgraph_id[node] = subgraph_id split_op_graphs.append(subgraph_id) subgraph_id += 1 else: node_to_subgraph_id[node] = subgraph_id # `keep_original_order` is important! # otherwise pytorch might reorder the nodes and # the semantics of the graph will change when we # have mutations in the graph split_gm = torch.fx.passes.split_module.split_module( graph, None, lambda node: node_to_subgraph_id[node], keep_original_order=True ) outputs = [] names = [name for (name, module) in split_gm.named_modules()] for name in names: if "." in name or name == "": # recursive child module or the root module continue module = getattr(split_gm, name) graph_id = int(name.replace("submod_", "")) outputs.append(SplitItem(name, graph_id, (graph_id in split_op_graphs), module)) # sort by integer graph_id, rather than string name outputs.sort(key=lambda x: x.graph_id) return split_gm, outputs compilation_start_time = 0.0 class PiecewiseCompileInterpreter(torch.fx.Interpreter): """Code adapted from `torch.fx.passes.shape_prop.ShapeProp`. It runs the given graph with fake inputs, and compile some submodules specified by `compile_submod_names` with the given compilation configs. NOTE: the order in `compile_submod_names` matters, because it will be used to determine the order of the compiled piecewise graphs. The first graph will handle logging, and the last graph has some special cudagraph output handling. """ def __init__( self, module: torch.fx.GraphModule, compile_submod_names: list[str], vllm_config: VllmConfig, vllm_backend: "VllmBackend", ): super().__init__(module) from torch._guards import detect_fake_mode self.fake_mode = detect_fake_mode() self.compile_submod_names = compile_submod_names self.compilation_config = vllm_config.compilation_config self.vllm_config = vllm_config self.vllm_backend = vllm_backend # When True, it annoyingly dumps the torch.fx.Graph on errors. self.extra_traceback = False def run(self, *args): # maybe instead just assert inputs are fake? fake_args = [ self.fake_mode.from_tensor(t) if isinstance(t, torch.Tensor) else t for t in args ] with self.fake_mode, enable_python_dispatcher(): return super().run(*fake_args) def call_module( self, target: torch.fx.node.Target, args: tuple[torch.fx.node.Argument, ...], kwargs: dict[str, Any], ) -> Any: assert isinstance(target, str) output = super().call_module(target, args, kwargs) if target in self.compile_submod_names: index = self.compile_submod_names.index(target) submod = self.fetch_attr(target) sym_shape_indices = [ i for i, x in enumerate(args) if isinstance(x, torch.SymInt) ] # Lazy import here to avoid circular import from .piecewise_backend import PiecewiseBackend piecewise_backend = PiecewiseBackend( submod, self.vllm_config, index, len(self.compile_submod_names), sym_shape_indices, self.vllm_backend, ) if ( self.compilation_config.cudagraph_mode.has_piecewise_cudagraphs() and not self.compilation_config.use_inductor_graph_partition ): # We're using Dynamo-based piecewise splitting, so we wrap # the whole subgraph with a static graph wrapper. from .cuda_graph import CUDAGraphOptions # resolve the static graph wrapper class (e.g. CUDAGraphWrapper # class) as platform dependent. static_graph_wrapper_class = resolve_obj_by_qualname( current_platform.get_static_graph_wrapper_cls() ) # Always assign PIECEWISE runtime mode to the # CUDAGraphWrapper for piecewise_backend, to distinguish # it from the FULL cudagraph runtime mode, no matter it # is wrapped on a full or piecewise fx graph. self.module.__dict__[target] = static_graph_wrapper_class( runnable=piecewise_backend, vllm_config=self.vllm_config, runtime_mode=CUDAGraphMode.PIECEWISE, cudagraph_options=CUDAGraphOptions( debug_log_enable=piecewise_backend.is_first_graph, gc_disable=not piecewise_backend.is_first_graph, weak_ref_output=piecewise_backend.is_last_graph, ), ) else: self.module.__dict__[target] = piecewise_backend compilation_counter.num_piecewise_capturable_graphs_seen += 1 return output # the tag for the part of model being compiled, # e.g. backbone/eagle_head model_tag: str = "backbone" model_is_encoder: bool = False @contextmanager def set_model_tag(tag: str, is_encoder: bool = False): """Context manager to set the model tag.""" global model_tag global model_is_encoder assert tag != model_tag, ( f"Model tag {tag} is the same as the current tag {model_tag}." ) old_tag = model_tag old_is_encoder = model_is_encoder model_tag = tag model_is_encoder = is_encoder try: yield finally: model_tag = old_tag model_is_encoder = old_is_encoder class VllmBackend: """The compilation backend for `torch.compile` with vLLM. It is used for compilation mode of `CompilationMode.VLLM_COMPILE`, where we customize the compilation. The major work of this backend is to split the graph into piecewise graphs, and pass them to the piecewise backend. This backend also adds the PostGradPassManager to Inductor config, which handles the post-grad passes. """ vllm_config: VllmConfig compilation_config: CompilationConfig _called: bool = False # the graph we compiled graph: fx.GraphModule # the stiching graph module for all the piecewise graphs split_gm: fx.GraphModule piecewise_graphs: list[SplitItem] returned_callable: Callable # Inductor passes to run on the graph pre-defunctionalization post_grad_passes: Sequence[Callable] sym_tensor_indices: list[int] input_buffers: list[torch.Tensor] compiler_manager: CompilerManager # Copy of CompilationConfig.inductor_compile_config + # an entry for PostGradPassManager inductor_config: dict[str, Any] def __init__( self, vllm_config: VllmConfig, prefix: str = "", is_encoder: bool = False, ): # if the model is initialized with a non-empty prefix, # then usually it's enough to use that prefix, # e.g. language_model, vision_model, etc. # when multiple parts are initialized as independent # models, we need to use the model_tag to distinguish # them, e.g. backbone (default), eagle_head, etc. self.prefix = prefix or model_tag # Mark compilation for encoder. self.is_encoder = is_encoder or model_is_encoder # Passes to run on the graph post-grad. self.pass_manager = resolve_obj_by_qualname( current_platform.get_pass_manager_cls() )() self.pass_key = current_platform.pass_key self.sym_tensor_indices = [] self.input_buffers = [] self.vllm_config = vllm_config self.compilation_config = vllm_config.compilation_config self.compiler_manager: CompilerManager = CompilerManager( self.compilation_config ) # Deepcopy the inductor config to detach the post-grad custom pass # from CompilationConfig. # We want to avoid PostGradPassManager in CompilationConfig because # in future we need PostGradPassManager.uuid() to be executed # only at compile time. self.inductor_config = deepcopy(self.compilation_config.inductor_compile_config) # `torch.compile` is JIT compiled, so we don't need to # do anything here def configure_post_pass(self): self.pass_manager.configure(self.vllm_config) # Post-grad custom passes are run using the post_grad_custom_post_pass # hook. If a pass for that hook exists, add it to the pass manager. if self.pass_key in self.inductor_config: if isinstance(self.inductor_config[self.pass_key], PostGradPassManager): raise ValueError( "PostGradPassManager can not be kept in CompilationConfig." ) else: # Config should automatically wrap all inductor passes assert isinstance( self.compilation_config.inductor_compile_config[self.pass_key], InductorPass, ) self.pass_manager.add( self.compilation_config.inductor_compile_config[self.pass_key] ) self.inductor_config[self.pass_key] = self.pass_manager def __call__( self, graph: fx.GraphModule, example_inputs ) -> VllmSerializableFunction: vllm_config = self.vllm_config # Minimal hashing here with existing utilities, reused below. env_factors = envs.compile_factors() env_hash = hash_factors(env_factors) # Compute config/compiler/code hashes once and reuse config_hash = vllm_config.compute_hash() compiler_hash = self.compiler_manager.compute_hash(vllm_config) forward_code_files = list(sorted(self.compilation_config.traced_files)) logger.debug( "Traced files (to be considered for compilation cache):\n%s", lazy(lambda: "\n".join(forward_code_files)), ) hash_content = [] for filepath in forward_code_files: hash_content.append(filepath) if filepath == "<string>": # This means the function was dynamically generated, with # e.g. exec(). We can't actually check these. continue try: with open(filepath) as f: hash_content.append(f.read()) except Exception: logger.warning("Failed to read file %s", filepath) continue code_hash = hashlib.sha256("\n".join(hash_content).encode()).hexdigest() # Clear after consumption self.compilation_config.traced_files.clear() if not self.compilation_config.cache_dir: # no provided cache dir, generate one based on the known factors # that affects the compilation. if none of the factors change, # the cache dir will be the same so that we can reuse the compiled # graph. factors = [env_hash, config_hash, code_hash, compiler_hash] # Use SHA-256 for cache key hashing to be consistent across # compute_hash functions. Truncate for a short cache dir name. hash_key = hashlib.sha256(str(factors).encode()).hexdigest()[:10] cache_dir = os.path.join( envs.VLLM_CACHE_ROOT, "torch_compile_cache", hash_key ) self.compilation_config.cache_dir = cache_dir cache_dir = self.compilation_config.cache_dir os.makedirs(cache_dir, exist_ok=True) self.compilation_config.cache_dir = cache_dir rank = vllm_config.parallel_config.rank dp_rank = vllm_config.parallel_config.data_parallel_index local_cache_dir = os.path.join(cache_dir, f"rank_{rank}_{dp_rank}", self.prefix) os.makedirs(local_cache_dir, exist_ok=True) self.compilation_config.local_cache_dir = local_cache_dir # Honors opt-outs such as CompilationMode.NONE or VLLM_DISABLE_COMPILE_CACHE. disable_cache = not is_compile_cache_enabled(self.inductor_config) if disable_cache: logger.info_once("vLLM's torch.compile cache is disabled.", scope="local") else: logger.info_once( "Using cache directory: %s for vLLM's torch.compile", local_cache_dir, scope="local", ) self.compiler_manager.initialize_cache( local_cache_dir, disable_cache, self.prefix ) # Reuses existing cache key logger.debug( "torch.compile cache factors: env=%s cfg=%s comp=%s code=%s dir=%s", env_hash, config_hash, compiler_hash, code_hash, local_cache_dir, ) # Persist and log only hash-relevant factors together. try: logger.debug( "Compile env factors (raw):\n%s\nVllm config hash: %s", lazy(partial(pprint.pformat, env_factors, width=120)), config_hash, ) meta_path = os.path.join(local_cache_dir, "cache_key_factors.json") if not os.path.exists(meta_path): with open(meta_path, "w") as f: json.dump( { "env": env_factors, # raw factors used for env_hash "config_hash": config_hash, "code_hash": code_hash, "compiler_hash": compiler_hash, }, f, indent=2, sort_keys=True, ) except Exception: # Best-effort only; metadata write failures are non-fatal. logger.warning( ( "Could not write compile cache metadata at %s; continuing without " "metadata. Compiled cache remains valid; diagnostics may be " "limited." ), local_cache_dir, exc_info=True, ) # when dynamo calls the backend, it means the bytecode # transform and analysis are done compilation_counter.num_graphs_seen += 1 from .monitor import torch_compile_start_time dynamo_time = time.time() - torch_compile_start_time logger.info_once( "Dynamo bytecode transform time: %.2f s", dynamo_time, scope="local" ) self.compilation_config.compilation_time += dynamo_time # we control the compilation process, each instance can only be # called once assert not self._called, "VllmBackend can only be called once" self.graph = graph self.configure_post_pass() if self.compilation_config.use_inductor_graph_partition: # Let Inductor decide partitioning; avoid FX-level pre-splitting. fx_split_ops: list[str] = [] else: fx_split_ops = self.compilation_config.splitting_ops or [] self.split_gm, self.piecewise_graphs = split_graph(graph, fx_split_ops) from torch._dynamo.utils import lazy_format_graph_code # depyf will hook lazy_format_graph_code and dump the graph # for debugging, no need to print the graph here lazy_format_graph_code("before split", self.graph) lazy_format_graph_code("after split", self.split_gm) compilation_counter.num_piecewise_graphs_seen += len(self.piecewise_graphs) submod_names_to_compile = [ item.submod_name for item in self.piecewise_graphs if not item.is_splitting_graph ] # Extract fake values from the graph to use them when needed. all_fake_values = [] for i in graph.graph.find_nodes(op="placeholder"): all_fake_values.append(i.meta["example_value"]) fake_args = [ all_fake_values[i] if isinstance(t, torch.Tensor) else t for i, t in enumerate(example_inputs) ] # propagate the split graph to the piecewise backend, # compile submodules with symbolic shapes PiecewiseCompileInterpreter( self.split_gm, submod_names_to_compile, self.vllm_config, self ).run(*fake_args) from torch._guards import detect_fake_mode fake_mode = detect_fake_mode() if ( self.compilation_config.dynamic_shapes_config.evaluate_guards and self.compilation_config.dynamic_shapes_config.type == DynamicShapesType.BACKED ): from torch.utils._sympy.value_ranges import ValueRanges # Drop counter-0/1 specializations guards; for backed dynamic shapes, # torch.compile will specialize for 0/1 inputs or otherwise guards that # shape is >= 2. This is because it's really hard not to hit a check # against 0/1. When we evaluate shape guards, we exclude checking those # guards (We would fail always otherwise). # We avoid that by updating the ranges of backed sizes when the min is # 2 for any, we assume it's 0. for s, r in fake_mode.shape_env.var_to_range.items(): if r.lower == 2: fake_mode.shape_env.var_to_range[s] = ValueRanges(0, r.upper) graph_path = os.path.join(local_cache_dir, "computation_graph.py") if not os.path.exists(graph_path): # code adapted from # https://github.com/thuml/depyf/blob/dab831108a752d1facc00acdd6d4243891845c37/depyf/explain/patched_lazy_format_graph_code.py#L30 # use `print_readable` because it can include submodules src = ( "from __future__ import annotations\nimport torch\n" + self.split_gm.print_readable(print_output=False) ) src = src.replace("<lambda>", "GraphModule") with open(graph_path, "w") as f: f.write(src) logger.debug_once( "Computation graph saved to %s", graph_path, scope="local" ) self._called = True if ( self.compilation_config.cudagraph_mode == CUDAGraphMode.NONE or not self.compilation_config.cudagraph_copy_inputs ): return VllmSerializableFunction( graph, example_inputs, self.prefix, self.split_gm, self.is_encoder ) # index of tensors that have symbolic shapes (batch size) # for weights and static buffers, they will have concrete shapes. # symbolic shape only happens for input tensors. from torch.fx.experimental.symbolic_shapes import is_symbolic self.sym_tensor_indices = [ i for i, x in enumerate(fake_args) if isinstance(x, torch._subclasses.fake_tensor.FakeTensor) and any(is_symbolic(d) for d in x.size()) ] # compiler managed cudagraph input buffers # we assume the first run with symbolic shapes # has the maximum size among all the tensors self.input_buffers = [ example_inputs[x].clone() for x in self.sym_tensor_indices ] # this is the callable we return to Dynamo to run def copy_and_call(*args): list_args = list(args) for i, index in enumerate(self.sym_tensor_indices): runtime_tensor = list_args[index] runtime_shape = runtime_tensor.shape[0] static_tensor = self.input_buffers[i][:runtime_shape] # copy the tensor to the static buffer static_tensor.copy_(runtime_tensor) # replace the tensor in the list_args to the static buffer list_args[index] = static_tensor return self.split_gm(*list_args) return VllmSerializableFunction( graph, example_inputs, self.prefix, copy_and_call, self.is_encoder )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/counter.py
vllm/compilation/counter.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import copy import dataclasses from contextlib import contextmanager @dataclasses.dataclass class CompilationCounter: num_models_seen: int = 0 num_graphs_seen: int = 0 # including the splitting ops num_piecewise_graphs_seen: int = 0 # not including the splitting ops num_piecewise_capturable_graphs_seen: int = 0 num_backend_compilations: int = 0 # Number of gpu_model_runner attempts to trigger CUDAGraphs capture num_gpu_runner_capture_triggers: int = 0 # Number of CUDAGraphs captured num_cudagraph_captured: int = 0 # InductorAdapter.compile calls num_inductor_compiles: int = 0 # EagerAdapter.compile calls num_eager_compiles: int = 0 # The number of time vLLM's compiler cache entry was updated num_cache_entries_updated: int = 0 # The number of standalone_compile compiled artifacts saved num_compiled_artifacts_saved: int = 0 # Number of times a model was loaded with CompilationMode.STOCK_TORCH_COMPILE stock_torch_compile_count: int = 0 def clone(self) -> "CompilationCounter": return copy.deepcopy(self) @contextmanager def expect(self, **kwargs): old = self.clone() yield for k, v in kwargs.items(): assert getattr(self, k) - getattr(old, k) == v, ( f"{k} not as expected, before it is {getattr(old, k)}" f", after it is {getattr(self, k)}, " f"expected diff is {v}" ) compilation_counter = CompilationCounter()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/sequence_parallelism.py
vllm/compilation/sequence_parallelism.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools import torch import torch._inductor.pattern_matcher as pm import torch.fx as fx from torch._inductor.pattern_matcher import PatternMatcherPass from vllm.config import VllmConfig from vllm.config.compilation import Range from vllm.distributed import get_tp_group, tensor_model_parallel_all_reduce from vllm.distributed.parallel_state import get_tensor_model_parallel_world_size from vllm.logger import init_logger from vllm.model_executor.layers.quantization.utils.quant_utils import ( kFp8StaticTensorSym, ) from vllm.platforms import current_platform from .inductor_pass import enable_fake_mode from .matcher_utils import MatcherFusedAddRMSNorm, MatcherQuantFP8, MatcherRMSNorm from .noop_elimination import NoOpEliminationPass from .vllm_inductor_pass import VllmInductorPass, VllmPatternMatcherPass logger = init_logger(__name__) def get_first_out_wrapper(fn): @functools.wraps(fn) def wrapper(*args): return fn(*args)[0] return wrapper class _SequenceParallelPatternHelper: """Helper for sequence parallelism patterns.""" def __init__( self, epsilon: float, dtype: torch.dtype, device: str, ): self.epsilon = epsilon self.dtype = dtype self.device = device self.tp_group = get_tp_group() self.tp_size = get_tensor_model_parallel_world_size() def _all_reduce(self, x: torch.Tensor) -> torch.Tensor: return tensor_model_parallel_all_reduce(x) def _reduce_scatter(self, x: torch.Tensor) -> torch.Tensor: return torch.ops.vllm.reduce_scatter.default( x, dim=0, world_size=self.tp_size, group_name=self.tp_group.unique_name ) def _all_gather(self, x: torch.Tensor) -> torch.Tensor: return torch.ops.vllm.all_gather.default( x, dim=0, world_size=self.tp_size, group_name=self.tp_group.unique_name ) class FirstAllReduceRMSNormPattern(_SequenceParallelPatternHelper): def __init__(self, epsilon: float, dtype: torch.dtype, device: str): super().__init__(epsilon, dtype, device) self.rmsnorm_matcher = MatcherRMSNorm(epsilon) def get_inputs(self): input = torch.empty([1, 8, 4], device=self.device, dtype=self.dtype) arg3_1 = torch.empty([4], device=self.device, dtype=self.dtype) return [input, arg3_1] def register(self, pm_pass: PatternMatcherPass): def pattern( input: torch.Tensor, arg3_1: torch.Tensor, ): all_reduce = self._all_reduce(input) rmsnorm = self.rmsnorm_matcher(all_reduce, arg3_1) return rmsnorm, all_reduce def replacement( input: torch.Tensor, arg3_1: torch.Tensor, ): reduce_scatter = self._reduce_scatter(input) rmsnorm = self.rmsnorm_matcher(reduce_scatter, arg3_1) all_gather = self._all_gather(rmsnorm) return all_gather, reduce_scatter pm.register_replacement( pattern, replacement, self.get_inputs(), pm.fwd_only, pm_pass ) class MiddleAllReduceRMSNormPattern(_SequenceParallelPatternHelper): def __init__(self, epsilon: float, dtype: torch.dtype, device: str): super().__init__(epsilon, dtype, device) self.rmsnorm_matcher = MatcherFusedAddRMSNorm(epsilon) def get_inputs(self): mm_1 = torch.empty([4, 4], device=self.device, dtype=self.dtype) residual = torch.empty([4, 4], device=self.device, dtype=self.dtype) rms_norm_weights = torch.empty([4, 4], device=self.device, dtype=self.dtype) return [ residual, mm_1, rms_norm_weights, ] def register(self, pm_pass: PatternMatcherPass): def pattern( residual: torch.Tensor, mm_1: torch.Tensor, rms_norm_weights: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: all_reduce = self._all_reduce(mm_1) rmsnorm = self.rmsnorm_matcher(all_reduce, rms_norm_weights, residual) return rmsnorm[0], rmsnorm[1] def replacement( residual: torch.Tensor, mm_1: torch.Tensor, rms_norm_weights: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: # pattern matcher replaces from top-to-bottom, # so residual is still the full size here. # once the seqpar pattern with the previous rmsnorm is replaced reduce_scatter = self._reduce_scatter(mm_1) residual = residual[0 : reduce_scatter.size(0), ...] rmsnorm = self.rmsnorm_matcher(reduce_scatter, rms_norm_weights, residual) all_gather = self._all_gather(rmsnorm[0]) # shape of residual changes but that's fine, # next node is already slicing it, now becomes a noop return all_gather, rmsnorm[1] pm.register_replacement( pattern, replacement, self.get_inputs(), pm.fwd_only, pm_pass ) pm.register_replacement( get_first_out_wrapper(pattern), get_first_out_wrapper(replacement), self.get_inputs(), pm.fwd_only, pm_pass, ) FP8_DTYPE = current_platform.fp8_dtype() class FirstAllReduceRMSNormStaticFP8Pattern(_SequenceParallelPatternHelper): def __init__( self, epsilon: float, dtype: torch.dtype, device: str, ): super().__init__(epsilon, dtype, device) self.rmsnorm_matcher = MatcherRMSNorm(epsilon) self.quant_matcher = MatcherQuantFP8(kFp8StaticTensorSym) def get_inputs(self): input = torch.zeros([1, 8, 4], device=self.device, dtype=self.dtype) weight = torch.empty([4], device=self.device, dtype=self.dtype) scale = torch.tensor(1.0, device=self.device, dtype=torch.float32) return [input, weight, scale] def register(self, pm_pass: PatternMatcherPass): def pattern( input: torch.Tensor, weight: torch.Tensor, scale: torch.Tensor, ): all_reduce = self._all_reduce(input) rms = self.rmsnorm_matcher(all_reduce, weight) quant, _ = self.quant_matcher(rms, scale) return quant, all_reduce def replacement( input: torch.Tensor, weight: torch.Tensor, scale: torch.Tensor, ): reduce_scatter = self._reduce_scatter(input) rms = self.rmsnorm_matcher(reduce_scatter, weight) quant, _ = self.quant_matcher(rms, scale) all_gather = self._all_gather(quant) return all_gather, reduce_scatter pm.register_replacement( pattern, replacement, self.get_inputs(), pm.fwd_only, pm_pass ) class MiddleAllReduceRMSNormStaticFP8Pattern(_SequenceParallelPatternHelper): def __init__(self, epsilon: float, dtype: torch.dtype, device: str): super().__init__(epsilon, dtype, device) self.rmsnorm_matcher = MatcherFusedAddRMSNorm(epsilon) self.quant_matcher = MatcherQuantFP8(kFp8StaticTensorSym) def get_inputs(self): mm_1 = torch.empty([4, 4], device=self.device, dtype=self.dtype) residual = torch.empty([4, 4], device=self.device, dtype=self.dtype) rms_norm_weights = torch.empty([4, 4], device=self.device, dtype=self.dtype) scale = torch.empty([1, 1], device=self.device, dtype=torch.float32) return [residual, mm_1, rms_norm_weights, scale] def register(self, pm_pass: PatternMatcherPass): def pattern( residual: torch.Tensor, mm_1: torch.Tensor, rms_norm_weights: torch.Tensor, scale: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: all_reduce = self._all_reduce(mm_1) rms, residual_out = self.rmsnorm_matcher( all_reduce, rms_norm_weights, residual ) quant, _ = self.quant_matcher(rms, scale) return quant, residual_out def replacement( residual: torch.Tensor, mm_1: torch.Tensor, rms_norm_weights: torch.Tensor, scale: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: # pattern matcher replaces from top-to-bottom, # so residual is still the full size here. # add a temporary slice which will become a noop # once the seqpar pattern with the previous rmsnorm is replaced reduce_scatter = self._reduce_scatter(mm_1) residual = residual[0 : reduce_scatter.size(0), ...] rms, residual_out = self.rmsnorm_matcher( reduce_scatter, rms_norm_weights, residual ) quant, _ = self.quant_matcher(rms, scale) all_gather = self._all_gather(quant) # shape of residual changes but that's fine, # next node is already slicing it, now becomes a noop return all_gather, residual_out pm.register_replacement( pattern, replacement, self.get_inputs(), pm.fwd_only, pm_pass ) pm.register_replacement( get_first_out_wrapper(pattern), get_first_out_wrapper(replacement), self.get_inputs(), pm.fwd_only, pm_pass, ) class SequenceParallelismPass(VllmPatternMatcherPass): """ This pass enables sequence parallelism for models. It identifies patterns where an AllReduce operation is followed by an RMSNorm (or RMSNorm and then Quantization) operation. These patterns are replaced with a ReduceScatter operation, followed by a local RMSNorm/Quantization, and then an AllGather operation. The general transformation is: Input -> AllReduce -> RMSNorm -> Output becomes Input -> ReduceScatter -> RMSNorm -> AllGather -> Output While this pass itself does not directly yield performance improvements, it lays the groundwork for subsequent fusion passes, such as GEMM + ReduceScatter and AllGather + GEMM fusions. These fusions can significantly reduce communication overhead and improve overall model performance. This pass splits up the residual tensor across TP ranks and hence divides its size. Because the pattern matcher starts at the end of the graph, the replacement contains a slice that temporarily conforms the input residual to the correct size. After all patterns have been matched, we use a NoOpEliminationPass to clean up what have now become no-op slices. Note that an older version of the pass did not need this as it operated only on custom rms_norm and fused_rms_norm_add custom ops which did not complain about mismatched shapes during replacement. So this approach has the same assumption that correctness is only maintained if all rms_norm operations are split across ranks. Correctness-wise, this is approach strictly better than before - before, the graph was incorrect semantically and shape-wise during the pass. With this approach there's only semantic incorrectness during the pass. Both approaches restore a correct graph once all patterns are matched. """ @enable_fake_mode def __init__(self, config: VllmConfig): super().__init__(config) # Used to clean up redundant views created temporarily # to circumvent residual shape change issues self.noop_cleanup = NoOpEliminationPass(config) self.noop_cleanup.pass_name = f"{self.pass_name}.{self.noop_cleanup.pass_name}" self.patterns: PatternMatcherPass = PatternMatcherPass( pass_name="sequence_parallelism_pass" ) for epsilon in [1e-5, 1e-6]: # RMSNorm + Static FP8 quantization patterns FirstAllReduceRMSNormStaticFP8Pattern( epsilon, self.model_dtype, self.device ).register(self.patterns) MiddleAllReduceRMSNormStaticFP8Pattern( epsilon, self.model_dtype, self.device ).register(self.patterns) # Normal RMSNorm patterns FirstAllReduceRMSNormPattern( epsilon, self.model_dtype, self.device ).register(self.patterns) MiddleAllReduceRMSNormPattern( epsilon, self.model_dtype, self.device ).register(self.patterns) self.dump_patterns(config, self.patterns) def is_applicable_for_range(self, compile_range: Range) -> bool: # When sequence parallelism is enabled, the residual tensor from RMSNorm # needs to be split along the sequence dimension. However, this dimension # is symbolic during piecewise compilation, and splitting symbolic shapes # is not supported. # # This pass is therefore only applied when the sequence dimension is # concrete: # 1. In full-graph compilation mode (no Dynamo splitting ops are used). # For this case we always pad num_tokens to be a multiple of # tensor_parallel_size, so there's no need to check shape % tp_size == 0. # 2. For specific shape provided during compilation (e.g., from # `compile_sizes`), which must be divisible by the tensor-parallel # size. if ( not self.compilation_config.splitting_ops or self.compilation_config.use_inductor_graph_partition ): return True tp_size = get_tensor_model_parallel_world_size() return (compile_range.is_single_size()) and (compile_range.end % tp_size == 0) @VllmInductorPass.time_and_log def __call__(self, graph: fx.Graph): self.matched_count = self.patterns.apply(graph) logger.debug("Replaced %s patterns", self.matched_count) # Clean up reshape nodes self.noop_cleanup(graph)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/fix_functionalization.py
vllm/compilation/fix_functionalization.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import operator from collections.abc import Iterable import torch from torch._higher_order_ops.auto_functionalize import auto_functionalized from vllm.logger import init_logger from vllm.platforms import current_platform from .fx_utils import is_func from .vllm_inductor_pass import VllmInductorPass logger = init_logger(__name__) class FixFunctionalizationPass(VllmInductorPass): """ This pass defunctionalizes certain nodes to avoid redundant tensor copies. After this pass, DCE (dead-code elimination) should never be run, as de-functionalized nodes may appear as dead code. To add new nodes to defunctionalize, add to the if-elif chain in __call__. """ @VllmInductorPass.time_and_log def __call__(self, graph: torch.fx.Graph): # XPU does not support auto-functionalization yet. # Will enable this when switch to vllm-xpu-kernels. if current_platform.is_xpu(): logger.debug( "XPU platform does not support fix functionalizationpass currently." ) return self.nodes_to_remove: list[torch.fx.Node] = [] count = 0 for node in graph.nodes: if not is_func(node, auto_functionalized): continue # Avoid deep if-elif nesting kwargs = node.kwargs at_target = node.args[0] if at_target == torch.ops._C.rotary_embedding.default: query = kwargs["query"] key = kwargs["key"] getitem_nodes = self.getitem_users(node) if ( is_func(query, operator.getitem) and is_func(key, operator.getitem) and query.args[0] == key.args[0] and is_func(query.args[0], torch.ops.aten.split_with_sizes.default) and all( is_func(user, torch.ops.aten.slice_scatter.default) for getitem_node in getitem_nodes.values() for user in getitem_node.users ) ): # Pattern where query and key are slices of an mm_node. # While functionalized, results at [1] and [2] are scattered # back into mm_node. So after de-functionalization, we can # just use mm_node directly. mm_node = query.args[0].args[0] for user in getitem_nodes.values(): for user_of_getitem in user.users: if is_func( user_of_getitem, torch.ops.aten.slice_scatter.default ): user_of_getitem.replace_all_uses_with(mm_node) self._remove(user_of_getitem) self._remove(user) self.insert_defunctionalized(graph, node) self._remove(node) else: # Directly replace the auto_functionalize(rotary_embedding) # with the inplace rotary_embedding. In theory, we shouldn't # do this blindly, but in practice in vLLM it's ok. The best # solution is to use auto_functionalization_v2 and then use # inductor's builtin defunctionalization (reinplacing) pass. mutated_args = {1: "query", 2: "key"} self.defunctionalize(graph, node, mutated_args) # rms_norm replacements avoid the most copies for LLaMa. elif at_target == torch.ops._C.fused_add_rms_norm.default: mutated_args = {1: "input", 2: "residual"} self.defunctionalize(graph, node, mutated_args) elif at_target == torch.ops._C.fused_add_rms_norm_static_fp8_quant.default: # noqa: E501 mutated_args = {1: "result", 2: "residual"} self.defunctionalize(graph, node, mutated_args) elif at_target == torch.ops._C.rms_norm_dynamic_per_token_quant.default: # noqa: E501 mutated_args = {1: "result", 2: "scale", 3: "residual"} self.defunctionalize(graph, node, mutated_args) elif at_target in [ torch.ops._C.rms_norm.default, torch.ops._C.rms_norm_static_fp8_quant.default, ]: mutated_args = {1: "result"} self.defunctionalize(graph, node, mutated_args) elif ( hasattr(torch.ops.vllm, "flashinfer_trtllm_fused_allreduce_norm") and at_target == torch.ops.vllm.flashinfer_trtllm_fused_allreduce_norm.default ): mutated_args = { 1: "allreduce_in", 2: "residual", 3: "norm_out", 4: "quant_out", 5: "scale_out", } self.defunctionalize(graph, node, mutated_args) # For some reason we need to specify the args for both # silu_and_mul and silu_and_mul_quant. The kwargs # pathway gets the wrong answer. elif at_target == torch.ops._C.silu_and_mul.default: mutated_args = {1: "result"} self.defunctionalize( graph, node, mutated_args, args=("result", "input") ) elif at_target == torch.ops._C.silu_and_mul_quant.default: mutated_args = {1: "result"} self.defunctionalize( graph, node, mutated_args, args=("result", "input", "scale") ) elif ( hasattr(torch.ops._C, "silu_and_mul_nvfp4_quant") and at_target == torch.ops._C.silu_and_mul_nvfp4_quant.default ): mutated_args = {1: "result", 2: "result_block_scale"} self.defunctionalize( graph, node, mutated_args, args=( "result", "result_block_scale", "input", "input_global_scale", ), ) # Defunctionalize fused_qk_norm_rope to remove higher-order wrapper. elif at_target == torch.ops._C.fused_qk_norm_rope.default: mutated_args = {1: "qkv"} args = ( "qkv", "num_heads_q", "num_heads_k", "num_heads_v", "head_dim", "eps", "q_weight", "k_weight", "cos_sin_cache", "is_neox", "position_ids", ) self.defunctionalize(graph, node, mutated_args=mutated_args, args=args) else: continue # skip the count count += 1 self.dump_graph(graph, "before_cleanup") # Remove the nodes all at once count_removed = len(self.nodes_to_remove) for node in self.nodes_to_remove: graph.erase_node(node) logger.debug( "De-functionalized %s nodes, removed %s nodes", count, count_removed ) self.nodes_to_remove.clear() def _remove(self, node_or_nodes: torch.fx.Node | Iterable[torch.fx.Node]): """ Stage a node (or nodes) for removal at the end of the pass. """ if isinstance(node_or_nodes, torch.fx.Node): self.nodes_to_remove.append(node_or_nodes) else: self.nodes_to_remove.extend(node_or_nodes) def defunctionalize( self, graph: torch.fx.Graph, node: torch.fx.Node, mutated_args: dict[int, torch.fx.Node | str], args: tuple[torch.fx.Node | str, ...] | None = None, ): """ De-functionalize a node by replacing it with a call to the original. It also replaces the getitem users with the mutated arguments. See replace_users_with_mutated_args and insert_defunctionalized. """ self.replace_users_with_mutated_args(node, mutated_args) self.insert_defunctionalized(graph, node, args=args) self._remove(node) def replace_users_with_mutated_args( self, node: torch.fx.Node, mutated_args: dict[int, torch.fx.Node | str] ): """ Replace all getitem users of the auto-functionalized node with the mutated arguments. :param node: The auto-functionalized node :param mutated_args: The mutated arguments, indexed by getitem index. If the value of an arg is a string, `node.kwargs[arg]` is used. """ for idx, user in self.getitem_users(node).items(): arg = mutated_args[idx] arg = node.kwargs[arg] if isinstance(arg, str) else arg user.replace_all_uses_with(arg) self._remove(user) def getitem_users(self, node: torch.fx.Node) -> dict[int, torch.fx.Node]: """ Returns the operator.getitem users of the auto-functionalized node, indexed by the index they are getting. """ users = {} for user in node.users: if is_func(user, operator.getitem): idx = user.args[1] users[idx] = user return users def insert_defunctionalized( self, graph: torch.fx.Graph, node: torch.fx.Node, args: tuple[torch.fx.Node | str, ...] | None = None, ): """ Insert a new defunctionalized node into the graph before node. If one of the kwargs is 'out', provide args directly, as node.kwargs cannot be used. See https://github.com/pytorch/pytorch/blob/a00faf440888ffb724bad413f329a49e2b6388e7/torch/_inductor/lowering.py#L351 :param graph: Graph to insert the defunctionalized node into :param node: The auto-functionalized node to defunctionalize :param args: If we cannot use kwargs, specify args directly. If an arg is a string, `node.kwargs[arg]` is used. """ # noqa: E501 assert is_func(node, auto_functionalized), ( f"node must be auto-functionalized, is {node} instead" ) # Create a new call to the original function with graph.inserting_before(node): function = node.args[0] if args is None: graph.call_function(function, kwargs=node.kwargs) else: # Args passed as strings refer to items in node.kwargs args = tuple( node.kwargs[arg] if isinstance(arg, str) else arg for arg in args ) graph.call_function(function, args=args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/fx_utils.py
vllm/compilation/fx_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import operator from collections.abc import Iterable, Iterator from torch import fx from torch._higher_order_ops.auto_functionalize import auto_functionalized from torch._ops import OpOverload, OpOverloadPacket def is_func(node: fx.Node, target) -> bool: return node.op == "call_function" and node.target == target def is_auto_func(node: fx.Node, op: OpOverload) -> bool: return is_func(node, auto_functionalized) and node.args[0] == op # Returns the first specified node with the given op (if it exists) def find_specified_fn_maybe(nodes: Iterable[fx.Node], op: OpOverload) -> fx.Node | None: for node in nodes: if node.target == op: return node return None # Returns the first specified node with the given op def find_specified_fn(nodes: Iterable[fx.Node], op: OpOverload) -> fx.Node: node = find_specified_fn_maybe(nodes, op) assert node is not None, f"Could not find {op} in nodes {nodes}" return node # Returns the first auto_functionalized node with the given op (if it exists) def find_auto_fn_maybe(nodes: Iterable[fx.Node], op: OpOverload) -> fx.Node | None: for node in nodes: if is_func(node, auto_functionalized) and node.args[0] == op: # noqa return node return None # Returns the first auto_functionalized node with the given op def find_auto_fn(nodes: Iterable[fx.Node], op: OpOverload) -> fx.Node: node = find_auto_fn_maybe(nodes, op) assert node is not None, f"Could not find {op} in nodes {nodes}" return node # Returns the getitem node that extracts the idx-th element from node # (if it exists) def find_getitem_maybe(node: fx.Node, idx: int) -> fx.Node | None: for user in node.users: if is_func(user, operator.getitem) and user.args[1] == idx: return user return None # Returns the getitem node that extracts the idx-th element from node def find_getitem(node: fx.Node, idx: int) -> fx.Node: ret = find_getitem_maybe(node, idx) assert ret is not None, f"Could not find getitem {idx} in node {node}" return ret # An auto-functionalization-aware utility for finding nodes with a specific op # Also handles op overload packets and finds all overloads def find_op_nodes( op: OpOverload | OpOverloadPacket, graph: fx.Graph ) -> Iterator[fx.Node]: if isinstance(op, OpOverloadPacket): for overload in op.overloads(): overload_op = getattr(op, overload) yield from find_op_nodes(overload_op, graph) return assert isinstance(op, OpOverload) yield from graph.find_nodes(op="call_function", target=op) for n in graph.find_nodes(op="call_function", target=auto_functionalized): if n.args[0] == op: yield n # Asserts that the node only has one user and returns it # Even if a node has only 1 user, it might share storage with another node, # which might need to be taken into account. def get_only_user(node: fx.Node) -> fx.Node: assert len(node.users) == 1 return next(iter(node.users))
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/cuda_graph.py
vllm/compilation/cuda_graph.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import dataclasses from collections import Counter from collections.abc import Callable from contextlib import ExitStack from typing import Any from unittest.mock import patch import torch import vllm.envs as envs from vllm.compilation.counter import compilation_counter from vllm.compilation.monitor import validate_cudagraph_capturing_enabled from vllm.config import CUDAGraphMode, VllmConfig from vllm.distributed.device_communicators.pynccl_allocator import set_graph_pool_id from vllm.forward_context import BatchDescriptor, get_forward_context from vllm.logger import init_logger from vllm.platforms import current_platform from vllm.utils.torch_utils import current_stream, weak_ref_tensors logger = init_logger(__name__) @dataclasses.dataclass(frozen=True) class CUDAGraphStat: num_unpadded_tokens: int num_padded_tokens: int num_paddings: int runtime_mode: str class CUDAGraphLogging: """Aggregate and log cudagraph metrics""" COLUMN_HEADERS = [ "Unpadded Tokens", "Padded Tokens", "Num Paddings", "Runtime Mode", "Count", ] def __init__(self, cg_mode: CUDAGraphMode, cg_capture_sizes: list[int] | None): self.reset() self.cg_mode = str(cg_mode) self.cg_capture_sizes = str(cg_capture_sizes or []) self.settings_header = ( "**CUDAGraph Config Settings:**\n\n" f"- Mode: {self.cg_mode}\n" f"- Capture sizes: {self.cg_capture_sizes}\n\n" "**CUDAGraph Stats:**\n\n" ) def reset(self): self.stats = [] def observe(self, cudagraph_stat: CUDAGraphStat): self.stats.append(cudagraph_stat) def generate_metric_table(self) -> str: stats_counts = Counter(self.stats) # Convert stats to rows of strings, in descending order of observed frequencies rows = [] for stat, count in sorted( stats_counts.items(), key=lambda item: item[1], reverse=True ): rows.append( [ str(stat.num_unpadded_tokens), str(stat.num_padded_tokens), str(stat.num_paddings), stat.runtime_mode, str(count), ] ) # Calculate column widths (max of header and data) col_widths = [] for i, header_text in enumerate(self.COLUMN_HEADERS): max_width = len(header_text) for row in rows: max_width = max(max_width, len(row[i])) col_widths.append(max_width) table_header_list = [ h.ljust(w) for h, w in zip(self.COLUMN_HEADERS, col_widths) ] table_header = "| " + " | ".join(table_header_list) + " |\n" table_separator = "|" + "|".join("-" * (w + 2) for w in col_widths) + "|\n" # Create data rows with proper alignment data_rows = [] for row in rows: formatted_row = [ str(val).ljust(width) for val, width in zip(row, col_widths) ] data_rows.append("| " + " | ".join(formatted_row) + " |") return ( self.settings_header + table_header + table_separator + "\n".join(data_rows) + "\n" ) def log(self, log_fn=logger.info): if not self.stats: return log_fn(self.generate_metric_table()) self.reset() @dataclasses.dataclass class CUDAGraphEntry: batch_descriptor: BatchDescriptor cudagraph: torch.cuda.CUDAGraph | None = None output: Any | None = None # for cudagraph debugging, track the input addresses # during capture, and check if they are the same during replay input_addresses: list[int] | None = None @dataclasses.dataclass class CUDAGraphOptions: debug_log_enable: bool = True gc_disable: bool = False weak_ref_output: bool = True class CUDAGraphWrapper: """Wraps a runnable to add CUDA graph capturing and replaying ability. And provide attribute access to the underlying `runnable` via `__getattr__`. The workflow of this wrapper in the cudagraph dispatching is as follows: 1. At initialization, a runtime mode is assigned to the wrapper (FULL or PIECEWISE). 2. At runtime, the wrapper receives a runtime_mode and a batch_descriptor(key) from the forward context and blindly trust them for cudagraph dispatching. 3. If runtime_mode is NONE or runtime_mode does not match the mode of the wrapper, just call the runnable directly. 4. Otherwise, i.e., the runtime_mode matches the mode of the wrapper, the wrapper will perform cudagraph capture(if key does not exist, create a new entry and cache it) or replay (if key exists in the cache). Note: CUDAGraphWrapper does not store persistent buffers or copy any runtime inputs into that buffers for replay. We assume implementing them is done outside of the wrapper. That is because we do not make any assumption on the dynamic shape (batch size) of the runtime inputs, as a trade-off for staying orthogonal to compilation logic. Nevertheless, tracing and checking the input addresses to be consistent during replay is guaranteed when VLLM_LOGGING_LEVEL == "DEBUG". """ def __init__( self, runnable: Callable, vllm_config: VllmConfig, runtime_mode: CUDAGraphMode, cudagraph_options: CUDAGraphOptions | None = None, ): self.runnable = runnable self.vllm_config = vllm_config self.runtime_mode = runtime_mode self.compilation_config = vllm_config.compilation_config self.first_run_finished = False self.is_debugging_mode = envs.VLLM_LOGGING_LEVEL == "DEBUG" # assert runtime_mode is not NONE(no cudagraph), otherwise, we don't # need to initialize a CUDAGraphWrapper. assert self.runtime_mode != CUDAGraphMode.NONE # TODO: in the future, if we want to use multiple # streams, it might not be safe to share a global pool. # only investigate this when we use multiple streams self.graph_pool = current_platform.get_global_graph_pool() if cudagraph_options is None: cudagraph_options = CUDAGraphOptions() self.cudagraph_options = cudagraph_options # the entries for different batch descriptors that we need to capture # cudagraphs for. self.concrete_cudagraph_entries: dict[BatchDescriptor, CUDAGraphEntry] = {} def __getattr__(self, key: str): # allow accessing the attributes of the runnable. if hasattr(self.runnable, key): return getattr(self.runnable, key) raise AttributeError( f"Attribute {key} not exists in the runnable of " f"cudagraph wrapper: {self.runnable}" ) def unwrap(self) -> Callable: # in case we need to access the original runnable. return self.runnable def __call__(self, *args, **kwargs): forward_context = get_forward_context() batch_descriptor = forward_context.batch_descriptor cudagraph_runtime_mode = forward_context.cudagraph_runtime_mode if ( cudagraph_runtime_mode == CUDAGraphMode.NONE or cudagraph_runtime_mode != self.runtime_mode ): # CUDAGraphMode.NONE could mean the profile run, a warmup run, or # running without cudagraphs. # We do not trigger capture/replay if the runtime mode is not # matches. This enables properly dispatching to the correct # CUDAGraphWrapper when nesting multiple instances with different # runtime modes. return self.runnable(*args, **kwargs) if batch_descriptor not in self.concrete_cudagraph_entries: # create a new entry for this batch descriptor self.concrete_cudagraph_entries[batch_descriptor] = CUDAGraphEntry( batch_descriptor=batch_descriptor ) entry = self.concrete_cudagraph_entries[batch_descriptor] if entry.cudagraph is None: if self.cudagraph_options.debug_log_enable: # Since we capture cudagraph for many different shapes and # capturing is fast, we don't need to log it for every # shape. E.g. we only log it for the first subgraph in # piecewise mode. logger.debug( "Capturing a cudagraph on (%s,%s)", self.runtime_mode.name, entry.batch_descriptor, ) # validate that cudagraph capturing is legal at this point. validate_cudagraph_capturing_enabled() input_addresses = [ x.data_ptr() for x in args if isinstance(x, torch.Tensor) ] entry.input_addresses = input_addresses cudagraph = torch.cuda.CUDAGraph() with ExitStack() as stack: if self.cudagraph_options.gc_disable: # during every model forward for piecewise cudagraph # mode, we will capture many pieces of cudagraphs # (roughly one per layer). running gc again and again # across layers will make the cudagraph capture very slow. # therefore, we only run gc for the first graph, # and disable gc for the rest of the graphs. stack.enter_context(patch("gc.collect", lambda: None)) stack.enter_context(patch("torch.cuda.empty_cache", lambda: None)) if self.graph_pool is not None: set_graph_pool_id(self.graph_pool) else: set_graph_pool_id(current_platform.graph_pool_handle()) # mind-exploding: carefully manage the reference and memory. with torch.cuda.graph( cudagraph, pool=self.graph_pool, stream=current_stream(), ): # `output` is managed by pytorch's cudagraph pool output = self.runnable(*args, **kwargs) if self.cudagraph_options.weak_ref_output: # by converting it to weak ref, # the original `output` will immediately be released # to save memory. It is only safe to do this for # the last graph in piecewise cuadgraph mode, because # the output of the last graph will not be used by # any other cuda graph. output = weak_ref_tensors(output) # here we always use weak ref for the output # to save memory entry.output = weak_ref_tensors(output) entry.cudagraph = cudagraph compilation_counter.num_cudagraph_captured += 1 # important: we need to return the output, rather than # the weak ref of the output, so that pytorch can correctly # manage the memory during cuda graph capture return output if self.is_debugging_mode: # check if the input addresses are the same new_input_addresses = [ x.data_ptr() for x in args if isinstance(x, torch.Tensor) ] assert new_input_addresses == entry.input_addresses, ( f"Input addresses for cudagraphs are different " f"during replay. Expected {entry.input_addresses}, " f"got {new_input_addresses}" ) entry.cudagraph.replay() return entry.output
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/post_cleanup.py
vllm/compilation/post_cleanup.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from torch import fx from vllm.compilation.vllm_inductor_pass import VllmInductorPass class PostCleanupPass(VllmInductorPass): """ This pass performs cleanup after custom passes. It topologically sorts the graph and removes unused nodes. This is needed because the pattern matcher does not guarantee producing a topologically sorted graph, and there may be unused nodes left around. """ @VllmInductorPass.time_and_log def __call__(self, graph: fx.Graph) -> None: from torch._inductor.pattern_matcher import stable_topological_sort stable_topological_sort(graph) graph.eliminate_dead_code()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/base_static_graph.py
vllm/compilation/base_static_graph.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from collections.abc import Callable from typing import Any, Protocol from vllm.config import CUDAGraphMode, VllmConfig class AbstractStaticGraphWrapper(Protocol): """ StaticGraphWrapper interface that allows platforms to wrap a callable to be captured as a static graph. """ def __init__( self, runnable: Callable[..., Any], vllm_config: VllmConfig, runtime_mode: CUDAGraphMode, **kwargs: Any, ) -> None: """ Initializes the StaticGraphWrapper class with graph capturing and execution-related configurations. Args: runnable (Callable): The callable to be wrapped and captured. vllm_config (VllmConfig): Global configuration for vLLM. runtime_mode (CUDAGraphMode): The style of the static graph runtime. See CUDAGraphMode in vllm/config.py. Note that only the subset enum `NONE`, `PIECEWISE` and `FULL` are used as concrete runtime mode for cudagraph dispatching. Keyword Args: kwargs: Additional keyword arguments for platform-specific configurations. """ raise NotImplementedError def __call__(self, *args: Any, **kwargs: Any) -> Any: """ Executes the wrapped callable. If the current runtime mode in the ForwardContext matches the runtime mode of this instance, it replays the CUDAGraph or captures it using the callable if it hasn't been captured yet. Otherwise, it calls the original callable directly. Args: *args: Variable length input arguments to be passed into the callable. **kwargs: Keyword arguments to be passed into the callable. Returns: Any: Output of the executed callable. """ raise NotImplementedError
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/activation_quant_fusion.py
vllm/compilation/activation_quant_fusion.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import ABC, abstractmethod import torch from torch._higher_order_ops.auto_functionalize import auto_functionalized from torch._inductor.pattern_matcher import ( PatternMatcherPass, fwd_only, register_replacement, ) from torch._ops import OpOverload from vllm.config import VllmConfig from vllm.logger import init_logger from vllm.model_executor.layers.quantization.utils.quant_utils import ( QuantKey, kFp8StaticTensorSym, kNvfp4Quant, ) from vllm.platforms import current_platform from .fusion import QUANT_OPS, empty_bf16, empty_fp32, empty_i32 from .inductor_pass import enable_fake_mode from .matcher_utils import MatcherQuantFP8, MatcherSiluAndMul from .vllm_inductor_pass import VllmInductorPass, VllmPatternMatcherPass logger = init_logger(__name__) FP8_DTYPE = current_platform.fp8_dtype() FP4_DTYPE = torch.uint8 SILU_MUL_OP = torch.ops._C.silu_and_mul.default FUSED_OPS: dict[QuantKey, OpOverload] = { kFp8StaticTensorSym: torch.ops._C.silu_and_mul_quant.default, # noqa: E501 } silu_and_mul_nvfp4_quant_supported = current_platform.is_cuda() and hasattr( torch.ops._C, "silu_and_mul_nvfp4_quant" ) if silu_and_mul_nvfp4_quant_supported: FUSED_OPS[kNvfp4Quant] = torch.ops._C.silu_and_mul_nvfp4_quant.default # noqa: E501 class ActivationQuantPattern(ABC): """ The base class for Activation+Quant fusions. Should not be used directly. """ def __init__( self, quant_key: QuantKey, ): self.quant_key = quant_key self.quant_dtype = quant_key.dtype assert self.quant_key in QUANT_OPS, ( f"unsupported quantization scheme {self.quant_key}" ) self.QUANT_OP = QUANT_OPS[self.quant_key] assert self.quant_key in FUSED_OPS, ( f"unsupported fusion scheme {self.quant_key}" ) self.FUSED_OP = FUSED_OPS[self.quant_key] self.silu_and_mul_matcher = MatcherSiluAndMul() def empty_quant(self, *args, **kwargs): kwargs = {"dtype": self.quant_dtype, "device": "cuda", **kwargs} return torch.empty(*args, **kwargs) @abstractmethod def register(self, pm_pass: PatternMatcherPass): raise NotImplementedError class SiluMulFp8StaticQuantPattern(ActivationQuantPattern): """ Fusion for SiluMul+Fp8StaticQuant Pattern """ def __init__(self): super().__init__(kFp8StaticTensorSym) self.quant_matcher = MatcherQuantFP8(kFp8StaticTensorSym) def register(self, pm_pass: PatternMatcherPass): def pattern( input: torch.Tensor, scale: torch.Tensor, ): result_silu_mul = self.silu_and_mul_matcher(input) result_quant = self.quant_matcher(result_silu_mul, scale) return result_quant[0] def replacement( input: torch.Tensor, scale: torch.Tensor, ): d = input.shape[-1] // 2 output_shape = input.shape[:-1] + (d,) result = torch.empty( output_shape, device=input.device, dtype=self.quant_dtype ) at = auto_functionalized( self.FUSED_OP, result=result, input=input, scale=scale ) return at[1] inputs = [ *self.silu_and_mul_matcher.inputs(), # input self.quant_matcher.inputs()[1], # scale ] pattern(*inputs) register_replacement(pattern, replacement, inputs, fwd_only, pm_pass) class SiluMulNvfp4QuantPattern(ActivationQuantPattern): """ Fusion for SiluMul+Nvfp4Quant Pattern """ def __init__(self): super().__init__(kNvfp4Quant) def register(self, pm_pass: PatternMatcherPass): def pattern( result: torch.Tensor, output_scale: torch.Tensor, input: torch.Tensor, scale: torch.Tensor, ): result_silu_mul = self.silu_and_mul_matcher(input) at = auto_functionalized( self.QUANT_OP, output=result, input=result_silu_mul, output_scale=output_scale, input_scale=scale, ) return at[1], at[2] def replacement( result: torch.Tensor, output_scale: torch.Tensor, input: torch.Tensor, scale: torch.Tensor, ): at = auto_functionalized( self.FUSED_OP, result=result, result_block_scale=output_scale, input=input, input_global_scale=scale, ) return at[1], at[2] inputs = [ self.empty_quant(5, 32), # result empty_i32(128, 4), # output_scale empty_bf16(5, 64), # input empty_fp32(1, 1), # scale ] register_replacement(pattern, replacement, inputs, fwd_only, pm_pass) class ActivationQuantFusionPass(VllmPatternMatcherPass): """ This pass fuses a pre-defined set of custom ops into fused ops. It uses the torch pattern matcher to find the patterns and replace them. Because patterns can only be registered once, the pass is a singleton. This will be addressed in a future version of PyTorch: https://github.com/pytorch/pytorch/pull/139321#issuecomment-2452354980 """ @enable_fake_mode def __init__(self, config: VllmConfig): super().__init__(config) self.patterns: PatternMatcherPass = PatternMatcherPass( pass_name="activation_quant_fusion_pass" ) pattern_silu_mul_fp8 = SiluMulFp8StaticQuantPattern() pattern_silu_mul_fp8.register(self.patterns) if silu_and_mul_nvfp4_quant_supported: pattern_silu_mul_nvfp4 = SiluMulNvfp4QuantPattern() pattern_silu_mul_nvfp4.register(self.patterns) self.dump_patterns(config, self.patterns) @VllmInductorPass.time_and_log def __call__(self, graph: torch.fx.Graph): self.matched_count = self.patterns.apply(graph) logger.debug("Replaced %s patterns", self.matched_count) def uuid(self): return VllmInductorPass.hash_source( self, ActivationQuantPattern, SiluMulFp8StaticQuantPattern, SiluMulNvfp4QuantPattern, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/rocm_aiter_fusion.py
vllm/compilation/rocm_aiter_fusion.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any import torch import torch._inductor.pattern_matcher as pm from torch import fx from torch._inductor.pattern_matcher import PatternMatcherPass from torch._ops import OpOverload import vllm.model_executor.layers.quantization.utils.fp8_utils # noqa: F401 from vllm._aiter_ops import rocm_aiter_ops from vllm.compilation.activation_quant_fusion import ActivationQuantPattern from vllm.config import VllmConfig from vllm.logger import init_logger from vllm.model_executor.layers.quantization.utils.quant_utils import ( GroupShape, QuantKey, ScaleDesc, ) from vllm.platforms import current_platform from .fusion import ( FusedRMSQuantKey, ) from .inductor_pass import enable_fake_mode from .matcher_utils import ( MatcherFusedAddRMSNorm, MatcherQuantFP8, MatcherRMSNorm, MatcherSiluAndMul, ) from .vllm_inductor_pass import VllmInductorPass, VllmPatternMatcherPass logger = init_logger(__name__) FP8_DTYPE = current_platform.fp8_dtype() class AiterRMSNormQuantPattern: def __init__( self, epsilon: float, key: FusedRMSQuantKey, match_aiter_quant: bool = True ): self.epsilon = epsilon self.quant_dtype = key.quant.dtype self.rmsnorm_matcher = ( MatcherRMSNorm(epsilon, match_rocm_aiter=True) if not key.fused_add else MatcherFusedAddRMSNorm(epsilon, match_rocm_aiter=True) ) self.quant_matcher = MatcherQuantFP8( key.quant, match_rocm_aiter=match_aiter_quant, ) class AiterRMSNormDynamicQuantPattern(AiterRMSNormQuantPattern): """AITER RMSNorm + Dynamic Quantization pattern.""" FUSED_OP = rocm_aiter_ops.get_rmsnorm_fused_dynamic_quant_op() def __init__( self, epsilon: float, quant_dtype: torch.dtype, match_aiter_quant: bool = True, group_shape: GroupShape = GroupShape.PER_TOKEN, symmetric=True, ): scale = ScaleDesc(torch.float32, False, group_shape) key = FusedRMSQuantKey( fused_add=False, quant=QuantKey(dtype=quant_dtype, scale=scale, symmetric=symmetric), ) super().__init__(epsilon, key, match_aiter_quant) def register(self, pm_pass): def pattern( input: torch.Tensor, weight: torch.Tensor, ): result_rms = self.rmsnorm_matcher(input, weight) result, scale = self.quant_matcher(result_rms) return result, scale def replacement( input: torch.Tensor, weight: torch.Tensor, ): result = self.FUSED_OP( x=input, weight=weight, epsilon=self.epsilon, quant_dtype=self.quant_dtype, ) return result[0], result[1] pm.register_replacement( pattern, replacement, self.rmsnorm_matcher.inputs(), pm.fwd_only, pm_pass, ) class AiterFusedAddRMSNormDynamicQuantPattern(AiterRMSNormQuantPattern): """AITER RMSNorm Fused Add + Dynamic Quantization pattern.""" FUSED_OP = rocm_aiter_ops.get_rmsnorm_fused_add_dynamic_quant_op() def __init__( self, epsilon: float, quant_dtype: torch.dtype, match_aiter_quant: bool = True, group_shape: GroupShape = GroupShape.PER_TOKEN, symmetric=True, ): scale = ScaleDesc(torch.float32, False, group_shape) key = FusedRMSQuantKey( fused_add=True, quant=QuantKey(dtype=quant_dtype, scale=scale, symmetric=symmetric), ) super().__init__(epsilon, key, match_aiter_quant) def register(self, pm_pass): def pattern( input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor, ): result_rms, residual_out = self.rmsnorm_matcher(input, weight, residual) result, scale = self.quant_matcher(result_rms) return result, residual_out, scale def replacement( input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor ): result = self.FUSED_OP( x=input, residual=residual, weight=weight, epsilon=self.epsilon, quant_dtype=self.quant_dtype, ) return result[0], result[1], result[2] pm.register_replacement( pattern, replacement, self.rmsnorm_matcher.inputs(), pm.fwd_only, pm_pass, ) class AiterRMSFp8GroupQuantPattern(AiterRMSNormQuantPattern): """ This pattern fuses aiter rms_norm & group fp8 quant custom ops into an aiter rms_norm_group_fp8_quant op. """ FUSED_OP = rocm_aiter_ops.get_rmsnorm_group_fused_quant_op() def __init__( self, epsilon: float, quant_dtype: torch.dtype, group_shape: GroupShape, match_aiter_quant: bool = True, symmetric=True, ): scale = ScaleDesc(torch.float32, False, group_shape) key = FusedRMSQuantKey( fused_add=False, quant=QuantKey(dtype=quant_dtype, scale=scale, symmetric=symmetric), ) super().__init__(epsilon, key, match_aiter_quant) def register(self, pm_pass: PatternMatcherPass): def pattern( input: torch.Tensor, weight: torch.Tensor, ): result_rms = self.rmsnorm_matcher(input, weight) result, scale = self.quant_matcher(result_rms) return result, scale def replacement( input: torch.Tensor, weight: torch.Tensor, ): at = self.FUSED_OP( x=input, weight=weight, variance_epsilon=self.epsilon, group_size=128, ) return at[0], at[1] pm.register_replacement( pattern, replacement, self.rmsnorm_matcher.inputs(), pm.fwd_only, pm_pass ) class AiterFusedAddRMSFp8GroupQuantPattern(AiterRMSNormQuantPattern): """ This pattern fuses aiter rms_norm_with_add & group fp8 quant custom ops into a aiter rms_norm_with_add_group_fp8_quant op. """ FUSED_OP = rocm_aiter_ops.get_rmsnorm_group_add_fused_quant_op() def __init__( self, epsilon: float, quant_dtype: torch.dtype, group_shape: GroupShape, match_aiter_quant: bool = True, symmetric=True, ): scale = ScaleDesc(torch.float32, False, group_shape) key = FusedRMSQuantKey( fused_add=True, quant=QuantKey(dtype=quant_dtype, scale=scale, symmetric=symmetric), ) super().__init__(epsilon, key, match_aiter_quant) def register(self, pm_pass: PatternMatcherPass): def pattern( input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor, ): result_rms, residual_out = self.rmsnorm_matcher(input, weight, residual) result, scale = self.quant_matcher(result_rms) return result, residual_out, scale def replacement( input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor, ): at = self.FUSED_OP( x=input, residual=residual, weight=weight, variance_epsilon=self.epsilon, group_size=128, ) # result, scale, residual return at[0], at[1], at[2] pm.register_replacement( pattern, replacement, self.rmsnorm_matcher.inputs(), pm.fwd_only, pm_pass ) class RocmAiterRMSNormFusionPass(VllmPatternMatcherPass): """ This pass fuses aiter rms_norm & vllm/aiter quant custom ops into a fused rms_norm_quant op. It also supports fused_add_rms_norm. """ @enable_fake_mode def __init__(self, config: VllmConfig): super().__init__(config) self.patterns: PatternMatcherPass = PatternMatcherPass( pass_name="rocm_aiter_rms_norm_quant_fusion_pass" ) # Make sure fused add patterns are before simple rms norm, # as the latter is a subset of the former in torch ops for epsilon in [1e-5, 1e-6]: # Fuse aiter rms_norm + aiter dynamic group fp8 quant AiterRMSFp8GroupQuantPattern( epsilon, FP8_DTYPE, GroupShape(1, 128) ).register(self.patterns) # Fuse aiter fused_add_rms_norm + aiter dynamic group fp8 quant AiterFusedAddRMSFp8GroupQuantPattern( epsilon, FP8_DTYPE, GroupShape(1, 128) ).register(self.patterns) for match_aiter_quant in [True, False]: # Fuse aiter rms_norm + (aiter / vllm built-in) # dynamic per-token fp8 quant AiterRMSNormDynamicQuantPattern( epsilon, FP8_DTYPE, match_aiter_quant=match_aiter_quant ).register(self.patterns) # Fuse aiter fused_add_rms_norm + (aiter / vllm built-in) # dynamic per-token fp8 quant AiterFusedAddRMSNormDynamicQuantPattern( epsilon, FP8_DTYPE, match_aiter_quant=match_aiter_quant ).register(self.patterns) self.dump_patterns(config, self.patterns) @VllmInductorPass.time_and_log def __call__(self, graph: fx.Graph): self.matched_count = self.patterns.apply(graph) logger.debug("Replaced %s patterns", self.matched_count) def uuid(self) -> Any: fusion_patterns = [ AiterRMSNormDynamicQuantPattern, AiterFusedAddRMSNormDynamicQuantPattern, AiterRMSFp8GroupQuantPattern, AiterFusedAddRMSFp8GroupQuantPattern, ] return self.hash_source(self, *fusion_patterns) class AiterSiluMulFp8GroupQuantPattern(ActivationQuantPattern): """ This pattern fuses aiter silu_and_mul & group fp8 quant custom ops into an aiter silu_and_mul_group_fp8_quant op. """ FUSED_SILU_MUL_QUANT_OP = rocm_aiter_ops.get_act_mul_fused_fp8_group_quant_op() def __init__(self, quant_op: OpOverload): self.silu_and_mul_matcher = MatcherSiluAndMul() self.quant_op = quant_op def register(self, pm_pass: PatternMatcherPass): def pattern( input: torch.Tensor, ): at1 = self.silu_and_mul_matcher(input) at2 = self.quant_op(at1, 128) return at2[0], at2[1] def replacement( input: torch.Tensor, ): at = self.FUSED_SILU_MUL_QUANT_OP(x=input, group_size=128) return at[0], at[1] inputs = [ self.silu_and_mul_matcher.inputs()[0], ] pm.register_replacement(pattern, replacement, inputs, pm.fwd_only, pm_pass) class RocmAiterSiluMulFp8GroupQuantFusionPass(VllmPatternMatcherPass): """ This pass fuses a pre-defined set of custom ops into fused ops. It uses the torch pattern matcher to find the patterns and replace them. Because patterns can only be registered once, the pass is a singleton. This will be addressed in a future version of PyTorch: https://github.com/pytorch/pytorch/pull/139321#issuecomment-2452354980 """ AITER_GROUP_FP8_QUANT_OP = rocm_aiter_ops.get_group_quant_op() TRITON_GROUP_FP8_QUANT_OP = torch.ops.vllm.triton_per_token_group_quant_fp8.default QUANT_OPS = [AITER_GROUP_FP8_QUANT_OP, TRITON_GROUP_FP8_QUANT_OP] @enable_fake_mode def __init__(self, config: VllmConfig): super().__init__(config) self.patterns: PatternMatcherPass = PatternMatcherPass( pass_name="rocm_aiter_silu_mul_fp8_group_quant_fusion_pass" ) for quant_op in self.QUANT_OPS: AiterSiluMulFp8GroupQuantPattern(quant_op).register(self.patterns) self.dump_patterns(config, self.patterns) @VllmInductorPass.time_and_log def __call__(self, graph: torch.fx.Graph): self.matched_count = self.patterns.apply(graph) logger.debug("Replaced %s patterns", self.matched_count) def uuid(self): fusion_patterns = [ ActivationQuantPattern, AiterSiluMulFp8GroupQuantPattern, ] return VllmInductorPass.hash_source(self, *fusion_patterns)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/matcher_utils.py
vllm/compilation/matcher_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import ABC, abstractmethod import torch from torch._higher_order_ops import auto_functionalized from torch._ops import OpOverload from vllm._aiter_ops import rocm_aiter_ops from vllm.config import get_current_vllm_config from vllm.model_executor.layers.activation import SiluAndMul from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.quantization.input_quant_fp8 import QuantFP8 from vllm.model_executor.layers.quantization.utils.quant_utils import ( GroupShape, QuantKey, _normalize_quant_group_shape, kFp8Dynamic64Sym, kFp8Dynamic128Sym, kFp8DynamicTensorSym, kFp8DynamicTokenSym, kFp8StaticTensorSym, kNvfp4Quant, ) from vllm.model_executor.layers.rotary_embedding import RotaryEmbedding from vllm.platforms import current_platform RMS_OP = torch.ops._C.rms_norm.default RMS_ADD_OP = torch.ops._C.fused_add_rms_norm.default ROTARY_OP = torch.ops._C.rotary_embedding.default FLASHINFER_ROTARY_OP = torch.ops.vllm.flashinfer_rotary_embedding.default QUANT_OPS: dict[QuantKey, OpOverload] = { kFp8StaticTensorSym: torch.ops._C.static_scaled_fp8_quant.default, # noqa: E501 kFp8DynamicTensorSym: torch.ops._C.dynamic_scaled_fp8_quant.default, # noqa: E501 kFp8DynamicTokenSym: torch.ops._C.dynamic_per_token_scaled_fp8_quant.default, # noqa: E501 } if current_platform.is_cuda() and hasattr(torch.ops._C, "scaled_fp4_quant"): QUANT_OPS[kNvfp4Quant] = torch.ops._C.scaled_fp4_quant.default # noqa: E501 if current_platform.is_cuda(): QUANT_OPS[kFp8Dynamic128Sym] = torch.ops._C.per_token_group_fp8_quant.default # noqa: E501 QUANT_OPS[kFp8Dynamic64Sym] = torch.ops._C.per_token_group_fp8_quant.default # noqa: E501 SILU_MUL_OP = torch.ops._C.silu_and_mul.default class MatcherCustomOp(ABC): def __init__(self, enabled: bool): config = get_current_vllm_config() self.model_dtype = config.model_config.dtype if config.model_config else None self.device = config.device_config.device if config.device_config else None self.enabled = enabled self.forward = self.forward_custom if enabled else self.forward_native @abstractmethod def forward_custom(self, *args, **kws): pass @abstractmethod def forward_native(self, *args, **kws): pass def __call__(self, *args, **kws): return self.forward(*args, **kws) def empty(self, *args, **kws): return torch.empty(*args, dtype=self.model_dtype, device=self.device, **kws) def empty_int64(self, *args, **kws): return torch.empty(*args, dtype=torch.int64, device=self.device, **kws) def empty_f32(self, *args, **kws): return torch.empty(*args, dtype=torch.float32, device=self.device, **kws) def inputs(self) -> list[torch.Tensor]: """Utility for inputs to the pattern""" raise NotImplementedError class MatcherRotaryEmbedding(MatcherCustomOp): def __init__( self, is_neox: bool, head_size: int, num_heads: int, num_kv_heads: int, use_flashinfer: bool = False, enabled: bool | None = None, ) -> None: if enabled is None: enabled = RotaryEmbedding.enabled() super().__init__(enabled) self.is_neox = is_neox self.head_size = head_size self.num_heads = num_heads self.num_kv_heads = num_kv_heads self.q_size = self.num_heads * self.head_size self.kv_size = self.num_kv_heads * self.head_size self.rotary_dim = head_size if use_flashinfer: self.rotary_op = FLASHINFER_ROTARY_OP else: self.rotary_op = ROTARY_OP def inputs(self) -> list[torch.Tensor]: positions = self.empty_int64(5) query = self.empty(5, self.q_size) key = self.empty(5, self.kv_size) cos_sin_cache = self.empty(4096, self.rotary_dim) return [positions, query, key, cos_sin_cache] def forward_custom( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None, cos_sin_cache: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor | None]: result = auto_functionalized( self.rotary_op, positions=positions, query=query, key=key, head_size=self.head_size, cos_sin_cache=cos_sin_cache, is_neox=self.is_neox, ) query_out = result[1] key_out = result[2] if len(result) > 2 else None return query_out, key_out def forward_native( self, positions: torch.Tensor, query: torch.Tensor, key: torch.Tensor | None, cos_sin_cache: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor | None]: return RotaryEmbedding.forward_static( positions, query, key, self.head_size, self.rotary_dim, cos_sin_cache, self.is_neox, ) class MatcherRMSNorm(MatcherCustomOp): def __init__( self, epsilon: float, enabled: bool | None = None, match_rocm_aiter: bool = False, ): if enabled is None: enabled = RMSNorm.enabled() super().__init__(enabled) self.epsilon = epsilon self._rmsnorm_op = RMS_OP self.match_rocm_aiter = match_rocm_aiter if match_rocm_aiter: self._rmsnorm_op = rocm_aiter_ops.get_rmsnorm_op() def inputs(self): input = self.empty(5, 16) if self.enabled else self.empty_f32(5, 16) weight = self.empty(16) return [input, weight] def forward_rocm_aiter( self, input: torch.Tensor, weight: torch.Tensor, ) -> torch.Tensor: return self._rmsnorm_op( x=input, weight=weight, variance_epsilon=self.epsilon, ) def forward_custom( self, input: torch.Tensor, weight: torch.Tensor, ) -> torch.Tensor: if self.match_rocm_aiter: return self.forward_rocm_aiter(input, weight) result = torch.empty_like(input) _, result = auto_functionalized( self._rmsnorm_op, result=result, input=input, weight=weight, epsilon=self.epsilon, ) return result def forward_native( self, input: torch.Tensor, weight: torch.Tensor, ) -> torch.Tensor: return RMSNorm.forward_static( input, self.epsilon, input.size(-1), self.model_dtype, weight ) class MatcherFusedAddRMSNorm(MatcherCustomOp): def __init__( self, epsilon: float, enabled: bool | None = None, match_rocm_aiter: bool = False, ): if enabled is None: enabled = RMSNorm.enabled() super().__init__(enabled) self.epsilon = epsilon self.match_rocm_aiter = match_rocm_aiter self._rmsnorm_op = RMS_ADD_OP if match_rocm_aiter: self._rmsnorm_op = rocm_aiter_ops.get_rmsnorm_fused_add_op() def inputs(self): input = self.empty(5, 16) if self.enabled else self.empty_f32(5, 16) weight = self.empty(16) residual = self.empty(5, 16) return [input, weight, residual] def forward_rocm_aiter( self, input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: return self._rmsnorm_op( x=input, residual=residual, weight=weight, variance_epsilon=self.epsilon ) def forward_custom( self, input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: if self.match_rocm_aiter: return self.forward_rocm_aiter(input, weight, residual) _, result, residual = auto_functionalized( self._rmsnorm_op, input=input, residual=residual, weight=weight, epsilon=self.epsilon, ) return result, residual def forward_native( self, input: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor, ) -> tuple[torch.Tensor, torch.Tensor]: return RMSNorm.forward_static( input, self.epsilon, input.size(-1), self.model_dtype, weight, residual ) class MatcherQuantFP8(MatcherCustomOp): def __init__( self, quant_key: QuantKey, enabled: bool | None = None, has_col_major_scales: bool = False, is_e8m0: bool = False, match_rocm_aiter: bool = False, ): if enabled is None: enabled = QuantFP8.enabled() super().__init__(enabled) self.quant_key = quant_key self.has_col_major_scales = has_col_major_scales self.is_e8m0 = is_e8m0 self.match_rocm_aiter = match_rocm_aiter if match_rocm_aiter: assert not quant_key.scale.group_shape.is_per_tensor(), ( "ROCm aiter fusion pass does not support per tensor quantization" ) if quant_key.scale.group_shape.is_per_token(): self.QUANT_OP = rocm_aiter_ops.get_per_token_quant_op() else: assert quant_key.scale.group_shape.col == 128, ( "ROCm aiter fusion pass currently supports " "quantization operation with group_size 128" ) if current_platform.is_fp8_fnuz(): self.QUANT_OP = rocm_aiter_ops.get_group_quant_op() else: self.QUANT_OP = ( torch.ops.vllm.triton_per_token_group_quant_fp8.default ) else: assert quant_key in QUANT_OPS, ( f"unsupported quantization scheme {quant_key}" ) self.QUANT_OP = QUANT_OPS[quant_key] assert quant_key.dtype == current_platform.fp8_dtype(), ( "Only QuantFP8 supported by" ) assert quant_key.scale2 is None self.quant_fp8 = QuantFP8( quant_key.scale.static, quant_key.scale.group_shape, column_major_scales=has_col_major_scales, use_ue8m0=is_e8m0, ) def forward_rocm_aiter( self, input: torch.Tensor, scale: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: quant_key_group_shape = self.quant_key.scale.group_shape if quant_key_group_shape == GroupShape.PER_TOKEN: return self.QUANT_OP( x=input, quant_dtype=self.quant_key.dtype, scale=scale, ) else: return self.QUANT_OP(input, quant_key_group_shape.col) def forward_custom( self, input: torch.Tensor, scale: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: if self.match_rocm_aiter: return self.forward_rocm_aiter(input, scale) result = torch.empty( input.shape, device=input.device, dtype=self.quant_key.dtype ) if self.quant_key.scale.group_shape.is_per_group(): assert scale is None scale = self.make_scale(input, transposed=self.has_col_major_scales) finfo = torch.finfo(self.quant_key.dtype) fp8_min = finfo.min fp8_max = finfo.max _, result, scale = auto_functionalized( self.QUANT_OP, input=input, output_q=result, output_s=scale, group_size=self.quant_key.scale.group_shape[1], eps=1e-10, fp8_min=fp8_min, fp8_max=fp8_max, scale_ue8m0=self.is_e8m0, ) return result, scale if self.quant_key.scale.static: assert scale is not None _, result = auto_functionalized( self.QUANT_OP, result=result, input=input, scale=scale ) return result, scale else: assert scale is None scale = self.make_scale(input) _, result, scale = auto_functionalized( self.QUANT_OP, result=result, input=input, scale=scale, scale_ub=None ) return result, scale def forward_native( self, input: torch.Tensor, scale: torch.Tensor | None = None, ) -> tuple[torch.Tensor, torch.Tensor]: return self.quant_fp8(input, scale) def make_scale(self, input: torch.Tensor, transposed: bool = False): normalized_group_shape = _normalize_quant_group_shape( input, self.quant_key.scale.group_shape ) scale_shape = ( input.shape[0] // normalized_group_shape[0], input.shape[1] // normalized_group_shape[1], ) if transposed: scale_shape = tuple(reversed(scale_shape)) return torch.empty( scale_shape, device=input.device, dtype=torch.float32 ).permute(-1, -2) return torch.empty(scale_shape, device=input.device, dtype=torch.float32) def inputs(self) -> list[torch.Tensor]: input = self.empty(5, 16) if self.quant_key.scale.static: return [input, self.empty_f32(1, 1)] return [input] class MatcherSiluAndMul(MatcherCustomOp): def __init__(self, enabled: bool | None = None): if enabled is None: enabled = SiluAndMul.enabled() super().__init__(enabled) def inputs(self) -> list[torch.Tensor]: input = self.empty(5, 4) return [input] def forward_custom( self, x: torch.Tensor, ) -> torch.Tensor: d = x.shape[-1] // 2 output_shape = x.shape[:-1] + (d,) out = torch.empty(output_shape, dtype=x.dtype, device=x.device) result = auto_functionalized(SILU_MUL_OP, result=out, input=x) return result[1] def forward_native( self, x: torch.Tensor, ) -> torch.Tensor: return SiluAndMul.forward_native(x)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/vllm/compilation/decorators.py
vllm/compilation/decorators.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import contextlib import hashlib import inspect import os import sys from collections.abc import Callable from typing import TypeVar, overload from unittest.mock import patch import torch import torch.nn as nn from packaging import version from torch._dynamo.symbolic_convert import InliningInstructionTranslator import vllm.envs as envs from vllm.compilation.counter import compilation_counter from vllm.compilation.wrapper import TorchCompileWithNoGuardsWrapper from vllm.config import ( CompilationMode, VllmConfig, get_current_vllm_config, set_current_vllm_config, ) from vllm.config.compilation import DynamicShapesType from vllm.logger import init_logger from vllm.sequence import IntermediateTensors from vllm.utils.import_utils import resolve_obj_by_qualname from vllm.utils.torch_utils import is_torch_equal_or_newer, supports_dynamo from .monitor import start_monitoring_torch_compile logger = init_logger(__name__) IGNORE_COMPILE_KEY = "_ignore_compile_vllm" _T = TypeVar("_T", bound=type[nn.Module]) def ignore_torch_compile(cls: _T) -> _T: """ A decorator to ignore support_torch_compile decorator on the class. This is useful when a parent class has a support_torch_compile decorator, but we don't want to compile the class `cls` that inherits the parent class. This only ignores compiling the forward of the class the decorator is applied to. If the parent has ignore_torch_compile but the child has support_torch_compile, the child will still be compiled. If the class has one or more submodules that have support_torch_compile decorator applied, compile will not be ignored for those submodules. """ setattr(cls, IGNORE_COMPILE_KEY, True) return cls def _should_ignore_torch_compile(cls) -> bool: """ Check if the class should be ignored for torch.compile. """ return getattr(cls, IGNORE_COMPILE_KEY, False) @overload def support_torch_compile( *, enable_if: Callable[[VllmConfig], bool] | None = None, ) -> Callable[[_T], _T]: ... @overload def support_torch_compile( *, dynamic_arg_dims: dict[str, int | list[int]] | None, ) -> Callable[[_T], _T]: ... @overload def support_torch_compile( *, mark_unbacked_dims: dict[str, int | list[int]] | None, ) -> Callable[[_T], _T]: ... @overload def support_torch_compile( *, dynamic_arg_dims: dict[str, int | list[int]] | None, mark_unbacked_dims: dict[str, int | list[int]] | None, ) -> Callable[[_T], _T]: ... @overload def support_torch_compile(cls: _T) -> _T: ... def support_torch_compile( cls: _T | None = None, *, dynamic_arg_dims: dict[str, int | list[int]] | None = None, mark_unbacked_dims: dict[str, int | list[int]] | None = None, enable_if: Callable[[VllmConfig], bool] | None = None, shape_invariants: Callable[..., None] = lambda *args, **kwargs: None, ) -> Callable[[_T], _T] | _T: """ A decorator to add support for compiling the forward method of a class. Usage 1: use directly as a decorator without arguments: ```python @support_torch_compile class MyModel(nn.Module): def forward(self, x: torch.Tensor, y: Optional[torch.Tensor]): ... ``` Usage 2: use as a decorator with arguments: ```python @support_torch_compile(dynamic_arg_dims={"x": 0, "y": 0}) class MyModel(nn.Module): def forward(self, x: torch.Tensor, y: Optional[torch.Tensor]): ... ``` `dynamic_arg_dims` is a dictionary that maps argument names to the dynamic dimensions of the argument. The dynamic dimensions can be either a single integer or a list of integers. if `dynamic_arg_dims` is `None`, it is inferred from the type annotation of the `forward` method, based on the following default rules: - if the argument is annotated as `torch.Tensor` or `Optional[torch.Tensor]`, the first dimension will be marked as dynamic. - if the argument is annotated as `IntermediateTensors`, the first dimension of all the tensors in the intermediate tensors will be marked as dynamic. During runtime, when we actually mark dimensions of tensors, it depends on the value of arguments: - if it is a single integer (can be negative), the corresponding dimension of the argument will be marked as dynamic. - if it is `None`, ignored. - if it is `IntermediateTensors`, all the tensors in the intermediate tensors will be marked as dynamic. - otherwise, it will raise an error. NOTE: if an argument is `None`, it should always be passed as `None` during the lifetime of the model, otherwise, it cannot be captured as a single computation graph. `enable_if` is a function that takes a `VllmConfig` object as input and returns a boolean value indicating whether to compile the model or not. This is useful if you want to compile the model only when certain conditions are met. `mark_unbacked_dims` is a dictionary that maps argument names with a dynamic dim to be decorated with `mark_unbacked`. This is useful if we would like to enforce that dynamo does not specialize on 0/1 values in the case of dummy input such as for vision model compilation `shape_invariants` is a function that gets compiled right before forward. The function should have the torch._check calls that are needed to set the relationships between different input sizes. For example: torch._check(input_ids.size()[0] == inputs_embeds.size()[0]) This enforces constraints on the symbolic shapes without hardcoding specific values. It is needed for some models to avoid data dependent errors. """ def cls_decorator_helper(cls: _T) -> _T: # helper to pass `dynamic_arg_dims` to `_support_torch_compile` # to avoid too much indentation for `_support_torch_compile` if not hasattr(cls, "forward"): raise TypeError("decorated class should have a forward method.") sig = inspect.signature(cls.forward) inferred_dynamic_arg_dims = dynamic_arg_dims if inferred_dynamic_arg_dims is None: inferred_dynamic_arg_dims = {} for k, v in sig.parameters.items(): if v.annotation in [ torch.Tensor, torch.Tensor | None, IntermediateTensors, IntermediateTensors | None, ]: inferred_dynamic_arg_dims[k] = 0 logger.debug( ("Inferred dynamic dimensions for forward method of %s: %s"), cls, list(inferred_dynamic_arg_dims.keys()), ) if len(inferred_dynamic_arg_dims) == 0: raise ValueError( "No dynamic dimensions found in the forward method of " f"{cls}. Please provide dynamic_arg_dims explicitly." ) for k in inferred_dynamic_arg_dims: if k not in sig.parameters: raise ValueError( f"Argument {k} not found in the forward method of {cls}" ) return _support_torch_compile( cls, inferred_dynamic_arg_dims, mark_unbacked_dims, enable_if, shape_invariants, ) if cls is not None: # use `support_torch_compile` as a decorator without arguments assert isinstance(cls, type) return cls_decorator_helper(cls) return cls_decorator_helper def _model_hash_key(fn) -> str: import vllm sha256_hash = hashlib.sha256() sha256_hash.update(vllm.__version__.encode()) sha256_hash.update(fn.__qualname__.encode()) sha256_hash.update(str(fn.__code__.co_firstlineno).encode()) return sha256_hash.hexdigest() def _verify_source_unchanged(source_info, vllm_config) -> None: from .caching import _compute_code_hash, _compute_code_hash_with_content file_contents = {} for source in source_info.inlined_sources: module = sys.modules[source.module] file = inspect.getfile(module) vllm_config.compilation_config.traced_files.add(file) file_contents[file] = source.content expected_checksum = _compute_code_hash_with_content(file_contents) actual_checksum = _compute_code_hash(set(file_contents.keys())) if expected_checksum != actual_checksum: raise RuntimeError( "Source code has changed since the last compilation. Recompiling the model." ) def _support_torch_compile( cls: _T, dynamic_arg_dims: dict[str, int | list[int]], mark_unbacked_dims: dict[str, int | list[int]] | None = None, enable_if: Callable[[VllmConfig], bool] | None = None, shape_invariants: Callable[..., None] = lambda *args, **kwargs: None, ) -> _T: """ A decorator to add support for compiling the forward method of a class. """ if TorchCompileWithNoGuardsWrapper in cls.__bases__: # support decorating multiple times return cls # take care of method resolution order # make sure super().__init__ is called on the base class # other than TorchCompileWithNoGuardsWrapper cls.__bases__ = cls.__bases__ + (TorchCompileWithNoGuardsWrapper,) old_init = cls.__init__ setattr(cls, IGNORE_COMPILE_KEY, False) def __init__( self, *, vllm_config: VllmConfig | None = None, prefix: str = "", **kwargs ): if vllm_config is None: vllm_config = get_current_vllm_config() # NOTE: to support multimodal models (such as encoder), # we may not have vllm_config so we may need to patch # it sig = inspect.signature(old_init) if "vllm_config" in sig.parameters: kwargs["vllm_config"] = vllm_config if "prefix" in sig.parameters: kwargs["prefix"] = prefix old_init(self, **kwargs) self.vllm_config = vllm_config self.compilation_config = self.vllm_config.compilation_config enable_compile = enable_if is None or enable_if(vllm_config) # for CompilationMode.STOCK_TORCH_COMPILE , the upper level model runner # will handle the compilation, so we don't need to do anything here. self.do_not_compile = ( self.compilation_config.mode in [CompilationMode.NONE, CompilationMode.STOCK_TORCH_COMPILE] or not supports_dynamo() or _should_ignore_torch_compile(self.__class__) or not enable_compile ) if self.do_not_compile: return self._check_shape_invariants = shape_invariants compilation_counter.num_models_seen += 1 self.compiled = False TorchCompileWithNoGuardsWrapper.__init__(self) cls.__init__ = __init__ def _mark_dynamic_inputs(mod, type, *args, **kwargs): def mark_dynamic(arg, dims): if type == DynamicShapesType.UNBACKED: if is_torch_equal_or_newer("2.10.0.dev"): for dim in dims: torch._dynamo.decorators.mark_unbacked( arg, dim, hint_override=arg.size()[dim] ) else: torch._dynamo.decorators.mark_unbacked(arg, dims) else: torch._dynamo.mark_dynamic(arg, dims) sig = inspect.signature(mod.__class__.forward) bound_args = sig.bind(mod, *args, **kwargs) bound_args.apply_defaults() for k, dims in dynamic_arg_dims.items(): arg = bound_args.arguments.get(k) if arg is not None: dims = [dims] if isinstance(dims, int) else dims if isinstance(arg, torch.Tensor): # In case dims is specified with negative indexing dims = [arg.ndim + dim if dim < 0 else dim for dim in dims] mark_dynamic(arg, dims) elif isinstance(arg, IntermediateTensors): for tensor in arg.tensors.values(): # In case dims is specified with negative indexing dims = [tensor.ndim + dim if dim < 0 else dim for dim in dims] mark_dynamic(tensor, dims) else: raise ValueError( "Unsupported dynamic dimensions" f" {dims} for argument {k} with type {type(arg)}." ) if mark_unbacked_dims: for k, dims in mark_unbacked_dims.items(): arg = bound_args.arguments.get(k) if arg is not None: dims = [dims] if isinstance(dims, int) else dims if isinstance(arg, torch.Tensor): # In case dims is specified with negative indexing dims = [arg.ndim + dim if dim < 0 else dim for dim in dims] if is_torch_equal_or_newer("2.10.0.dev"): for dim in dims: torch._dynamo.decorators.mark_unbacked( arg, dim, hint_override=arg.size()[dim] ) else: torch._dynamo.decorators.mark_unbacked(arg, dims) def __call__(self, *args, **kwargs): # torch.compiler.is_compiling() means we are inside the compilation # e.g. TPU has the compilation logic in model runner, so we don't # need to compile the model inside. if self.do_not_compile or torch.compiler.is_compiling(): return self.forward(*args, **kwargs) # if aot_compiled_fn is set, just call it. if getattr(self, "aot_compiled_fn", None) is not None: return self.aot_compiled_fn(self, *args, **kwargs) ds_type = self.compilation_config.dynamic_shapes_config.type cache_dir = None aot_compilation_path = None if envs.VLLM_USE_AOT_COMPILE: """ When using torch.compile in AOT mode, we store the cache artifacts under VLLM_CACHE_ROOT/torch_aot_compile/{hash}/rank_i_j. The {hash} contains all of the factors except for the source files being traced through, because we don't actually know which source files to check at this point (before dynamo runs). On loading we will actually look at the source files being traced through. If any source file have changed (compared with the serialized backend artifacts), then we need to generate a new AOT compile artifact from scratch. """ from .caching import compilation_config_hash_factors factors: list[str] = compilation_config_hash_factors(self.vllm_config) factors.append(_model_hash_key(self.forward)) hash_key = hashlib.sha256(str(factors).encode()).hexdigest() cache_dir = os.path.join( envs.VLLM_CACHE_ROOT, "torch_aot_compile", hash_key, ) rank = self.vllm_config.parallel_config.rank dp_rank = self.vllm_config.parallel_config.data_parallel_index cache_dir = os.path.join(cache_dir, f"rank_{rank}_{dp_rank}") aot_compilation_path = os.path.join(cache_dir, "model") try: with ( set_current_vllm_config(self.vllm_config), open(aot_compilation_path, "rb") as f, ): start_monitoring_torch_compile(self.vllm_config) loaded_fn = torch.compiler.load_compiled_function( f, f_globals=self.forward.__globals__ ) _verify_source_unchanged(loaded_fn.source_info(), self.vllm_config) if not self.compilation_config.dynamic_shapes_config.evaluate_guards: loaded_fn.disable_guard_check() self.aot_compiled_fn = loaded_fn except Exception as e: if os.path.exists(aot_compilation_path): logger.warning( "Cannot load aot compilation from path %s, error: %s", aot_compilation_path, str(e), ) if envs.VLLM_FORCE_AOT_LOAD: raise e if getattr(self, "aot_compiled_fn", None) is not None: logger.info( "Directly load AOT compilation from path %s", aot_compilation_path ) return self.aot_compiled_fn(self, *args, **kwargs) if self.compiled: assert ( not envs.VLLM_USE_AOT_COMPILE or self.vllm_config.compilation_config.backend == "eager" ) return TorchCompileWithNoGuardsWrapper.__call__(self, *args, **kwargs) # This is the path for the first compilation. # the first compilation needs to have dynamic shapes marked _mark_dynamic_inputs( self, ds_type, *args, **kwargs, ) # here, it is the starting point of the `torch.compile` process start_monitoring_torch_compile(self.vllm_config) original_code_object = self.original_code_object() logger.debug("Start compiling function %s", original_code_object) # we do not want tp delete the original code object entries since # we depend on them now to look up cached compiled functions. # torch._dynamo.eval_frame.remove_from_cache(original_code_object) # collect all relevant files traced by Dynamo, # so that the compilation cache can trigger re-compilation # properly when any of these files change. # 1. the file containing the top-level forward function self.compilation_config.traced_files.add(original_code_object.co_filename) # 2. every time Dynamo sees a function call, it will inline # the function by calling InliningInstructionTranslator.inline_call_ # we hijack this function to know all the functions called # during Dynamo tracing, and their corresponding files inline_call = InliningInstructionTranslator.inline_call_ def patched_inline_call(self_): code = self_.f_code self.compilation_config.traced_files.add(code.co_filename) return inline_call(self_) # Disable the C++ compilation of symbolic shape guards. C++-fication # of symbolic shape guards can improve guard overhead. But, since # vllm skip guards anyways, setting this flag to False can improve # compile time. dynamo_config_patches = {} try: _ = torch._dynamo.config.enable_cpp_symbolic_shape_guards dynamo_config_patches["enable_cpp_symbolic_shape_guards"] = False except AttributeError: # Note: this config is not available in torch 2.6, we can skip # if the config doesn't exist logger.debug("enable_cpp_symbolic_shape_guards config not available") # Prepare backed_size_oblivious config patch if needed fx_config_patches = {} if ds_type == DynamicShapesType.BACKED_SIZE_OBLIVIOUS: fx_config_patches["backed_size_oblivious"] = True # Prepare inductor config patches # assume_32bit_indexing is only available in torch 2.10.0.dev+ inductor_config_patches = {} if is_torch_equal_or_newer("2.10.0.dev"): inductor_config_patches["assume_32bit_indexing"] = True with ( patch.object( InliningInstructionTranslator, "inline_call_", patched_inline_call ), torch._dynamo.config.patch(**dynamo_config_patches), maybe_use_cudagraph_partition_wrapper(self.vllm_config), torch.fx.experimental._config.patch(**fx_config_patches), _torch27_patch_tensor_subclasses(), torch._inductor.config.patch(**inductor_config_patches), ): use_aot_compile = envs.VLLM_USE_AOT_COMPILE if self.vllm_config.compilation_config.backend == "eager": logger.warning("Detected eager backend, disabling AOT compile.") use_aot_compile = False if use_aot_compile: self.aot_compiled_fn = self.aot_compile(*args, **kwargs) output = self.aot_compiled_fn(self, *args, **kwargs) assert aot_compilation_path is not None assert cache_dir is not None try: os.makedirs(cache_dir, exist_ok=True) self.aot_compiled_fn.save_compiled_function(aot_compilation_path) except Exception as e: logger.warning( "Cannot save aot compilation to path %s, error: %s", aot_compilation_path, str(e), ) else: output = TorchCompileWithNoGuardsWrapper.__call__(self, *args, **kwargs) self.compiled = True return output cls.__call__ = __call__ return cls @contextlib.contextmanager def maybe_use_cudagraph_partition_wrapper(vllm_config: VllmConfig): """ Context manager to set/unset customized cudagraph partition wrappers. If we're using Inductor-based graph partitioning, we currently have the whole `fx.Graph` before Inductor lowering and the piecewise splitting happens after all graph passes and fusions. Here, we add a custom hook for Inductor to wrap each partition with our static graph wrapper class to maintain more control over static graph capture and replay. """ from vllm.config import CUDAGraphMode compilation_config = vllm_config.compilation_config if ( compilation_config.cudagraph_mode.has_piecewise_cudagraphs() and compilation_config.use_inductor_graph_partition ): from torch._inductor.utils import CUDAGraphWrapperMetadata from vllm.compilation.cuda_graph import CUDAGraphOptions from vllm.platforms import current_platform static_graph_wrapper_class = resolve_obj_by_qualname( current_platform.get_static_graph_wrapper_cls() ) def customized_cudagraph_wrapper(f, metadata: CUDAGraphWrapperMetadata): partition_id = metadata.partition_index num_partitions = metadata.num_partitions return static_graph_wrapper_class( runnable=f, vllm_config=vllm_config, runtime_mode=CUDAGraphMode.PIECEWISE, cudagraph_options=CUDAGraphOptions( debug_log_enable=partition_id == 0, gc_disable=partition_id != 0, weak_ref_output=partition_id == num_partitions - 1, ), ) torch._inductor.utils.set_customized_partition_wrappers( customized_cudagraph_wrapper ) yield if ( compilation_config.cudagraph_mode.has_piecewise_cudagraphs() and compilation_config.use_inductor_graph_partition ): torch._inductor.utils.set_customized_partition_wrappers(None) @contextlib.contextmanager def _torch27_patch_tensor_subclasses(): """ Add support for using tensor subclasses (ie `BasevLLMParameter`, ect) when using torch 2.7.0. This enables using weight_loader_v2 and the use of `BasevLLMParameters` without having to replace them with regular tensors before `torch.compile`-time. """ from vllm.model_executor.parameter import ( BasevLLMParameter, ModelWeightParameter, RowvLLMParameter, _ColumnvLLMParameter, ) def return_false(*args, **kwargs): return False if version.parse("2.7") <= version.parse(torch.__version__) < version.parse("2.8"): yield return with ( torch._dynamo.config.patch( "traceable_tensor_subclasses", [ BasevLLMParameter, ModelWeightParameter, _ColumnvLLMParameter, RowvLLMParameter, ], ), patch( "torch._dynamo.variables.torch.can_dispatch_torch_function", return_false ), ): yield
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false