repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/pre_commit/mypy.py
tools/pre_commit/mypy.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Run mypy on changed files. This script is designed to be used as a pre-commit hook. It runs mypy on files that have been changed. It groups files into different mypy calls based on their directory to avoid import following issues. Usage: python tools/pre_commit/mypy.py <ci> <python_version> <changed_files...> Args: ci: "1" if running in CI, "0" otherwise. In CI, follow_imports is set to "silent" for the main group of files. python_version: Python version to use (e.g., "3.10") or "local" to use the local Python version. changed_files: List of changed files to check. """ import subprocess import sys import regex as re FILES = [ "vllm/*.py", "vllm/assets", "vllm/distributed", "vllm/engine", "vllm/entrypoints", "vllm/executor", "vllm/inputs", "vllm/logging_utils", "vllm/multimodal", "vllm/platforms", "vllm/plugins", "vllm/tokenizers", "vllm/transformers_utils", "vllm/triton_utils", "vllm/usage", "vllm/utils", "vllm/worker", "vllm/v1/core", "vllm/v1/engine", "vllm/v1/executor", "vllm/v1/metrics", "vllm/v1/pool", "vllm/v1/sample", "vllm/v1/worker", ] # After fixing errors resulting from changing follow_imports # from "skip" to "silent", move the following directories to FILES SEPARATE_GROUPS = [ "tests", # v0 related "vllm/attention", "vllm/compilation", "vllm/lora", "vllm/model_executor", # v1 related "vllm/v1/attention", "vllm/v1/kv_offload", "vllm/v1/spec_decode", "vllm/v1/structured_output", ] # TODO(woosuk): Include the code from Megatron and HuggingFace. EXCLUDE = [ "vllm/engine/arg_utils.py", "vllm/model_executor/parallel_utils", "vllm/model_executor/models", "vllm/model_executor/layers/fla/ops", # Ignore triton kernels in ops. "vllm/attention/ops", ] def group_files(changed_files: list[str]) -> dict[str, list[str]]: """ Group changed files into different mypy calls. Args: changed_files: List of changed files. Returns: A dictionary mapping file group names to lists of changed files. """ exclude_pattern = re.compile(f"^{'|'.join(EXCLUDE)}.*") files_pattern = re.compile(f"^({'|'.join(FILES)}).*") file_groups = {"": []} file_groups.update({k: [] for k in SEPARATE_GROUPS}) for changed_file in changed_files: # Skip files which should be ignored completely if exclude_pattern.match(changed_file): continue # Group files by mypy call if files_pattern.match(changed_file): file_groups[""].append(changed_file) continue else: for directory in SEPARATE_GROUPS: if re.match(f"^{directory}.*", changed_file): file_groups[directory].append(changed_file) break return file_groups def mypy( targets: list[str], python_version: str | None, follow_imports: str | None, file_group: str, ) -> int: """ Run mypy on the given targets. Args: targets: List of files or directories to check. python_version: Python version to use (e.g., "3.10") or None to use the default mypy version. follow_imports: Value for the --follow-imports option or None to use the default mypy behavior. file_group: The file group name for logging purposes. Returns: The return code from mypy. """ args = ["mypy"] if python_version is not None: args += ["--python-version", python_version] if follow_imports is not None: args += ["--follow-imports", follow_imports] print(f"$ {' '.join(args)} {file_group}") return subprocess.run(args + targets, check=False).returncode def main(): ci = sys.argv[1] == "1" python_version = sys.argv[2] file_groups = group_files(sys.argv[3:]) if python_version == "local": python_version = f"{sys.version_info.major}.{sys.version_info.minor}" returncode = 0 for file_group, changed_files in file_groups.items(): follow_imports = None if ci and file_group == "" else "skip" if changed_files: returncode |= mypy( changed_files, python_version, follow_imports, file_group ) return returncode if __name__ == "__main__": sys.exit(main())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/pre_commit/enforce_regex_import.py
tools/pre_commit/enforce_regex_import.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import subprocess from pathlib import Path import regex as re FORBIDDEN_PATTERNS = re.compile(r"^\s*(?:import\s+re(?:$|\s|,)|from\s+re\s+import)") ALLOWED_PATTERNS = [ re.compile(r"^\s*import\s+regex\s+as\s+re\s*$"), re.compile(r"^\s*import\s+regex\s*$"), ] def get_staged_python_files() -> list[str]: try: result = subprocess.run( ["git", "diff", "--cached", "--name-only", "--diff-filter=AM"], capture_output=True, text=True, check=True, ) files = result.stdout.strip().split("\n") if result.stdout.strip() else [] return [f for f in files if f.endswith(".py")] except subprocess.CalledProcessError: return [] def is_forbidden_import(line: str) -> bool: line = line.strip() return bool( FORBIDDEN_PATTERNS.match(line) and not any(pattern.match(line) for pattern in ALLOWED_PATTERNS) ) def check_file(filepath: str) -> list[tuple[int, str]]: violations = [] try: with open(filepath, encoding="utf-8") as f: for line_num, line in enumerate(f, 1): if is_forbidden_import(line): violations.append((line_num, line.strip())) except (OSError, UnicodeDecodeError): pass return violations def main() -> int: files = get_staged_python_files() if not files: return 0 total_violations = 0 for filepath in files: if not Path(filepath).exists(): continue if filepath == "setup.py": continue violations = check_file(filepath) if violations: print(f"\n❌ {filepath}:") for line_num, line in violations: print(f" Line {line_num}: {line}") total_violations += 1 if total_violations > 0: print(f"\n💡 Found {total_violations} violation(s).") print("❌ Please replace 'import re' with 'import regex as re'") print(" Also replace 'from re import ...' with 'from regex import ...'") # noqa: E501 print("✅ Allowed imports:") print(" - import regex as re") print(" - import regex") # noqa: E501 return 1 return 0 if __name__ == "__main__": raise SystemExit(main())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_outputs.py
tests/test_outputs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from vllm.outputs import RequestOutput pytestmark = pytest.mark.cpu_test def test_request_output_forward_compatible(): output = RequestOutput( request_id="test_request_id", prompt="test prompt", prompt_token_ids=[1, 2, 3], prompt_logprobs=None, outputs=[], finished=False, example_arg_added_in_new_version="some_value", ) assert output is not None
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_sequence.py
tests/test_sequence.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch from vllm.sequence import IntermediateTensors def test_sequence_intermediate_tensors_equal(): class AnotherIntermediateTensors(IntermediateTensors): pass intermediate_tensors = IntermediateTensors({}) another_intermediate_tensors = AnotherIntermediateTensors({}) assert intermediate_tensors != another_intermediate_tensors empty_intermediate_tensors_1 = IntermediateTensors({}) empty_intermediate_tensors_2 = IntermediateTensors({}) assert empty_intermediate_tensors_1 == empty_intermediate_tensors_2 different_key_intermediate_tensors_1 = IntermediateTensors( {"1": torch.zeros([2, 4], dtype=torch.int32)} ) difference_key_intermediate_tensors_2 = IntermediateTensors( {"2": torch.zeros([2, 4], dtype=torch.int32)} ) assert different_key_intermediate_tensors_1 != difference_key_intermediate_tensors_2 same_key_different_value_intermediate_tensors_1 = IntermediateTensors( {"1": torch.zeros([2, 4], dtype=torch.int32)} ) same_key_different_value_intermediate_tensors_2 = IntermediateTensors( {"1": torch.zeros([2, 5], dtype=torch.int32)} ) assert ( same_key_different_value_intermediate_tensors_1 != same_key_different_value_intermediate_tensors_2 ) same_key_same_value_intermediate_tensors_1 = IntermediateTensors( {"1": torch.zeros([2, 4], dtype=torch.int32)} ) same_key_same_value_intermediate_tensors_2 = IntermediateTensors( {"1": torch.zeros([2, 4], dtype=torch.int32)} ) assert ( same_key_same_value_intermediate_tensors_1 == same_key_same_value_intermediate_tensors_2 )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_logger.py
tests/test_logger.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import enum import json import logging import os import sys import tempfile from dataclasses import dataclass from json.decoder import JSONDecodeError from tempfile import NamedTemporaryFile from typing import Any from unittest.mock import MagicMock, patch from uuid import uuid4 import pytest from vllm.entrypoints.logger import RequestLogger from vllm.logger import ( _DATE_FORMAT, _FORMAT, _configure_vllm_root_logger, enable_trace_function_call, init_logger, ) from vllm.logging_utils import NewLineFormatter from vllm.logging_utils.dump_input import prepare_object_to_dump def f1(x): return f2(x) def f2(x): return x def test_trace_function_call(): fd, path = tempfile.mkstemp() cur_dir = os.path.dirname(__file__) enable_trace_function_call(path, cur_dir) f1(1) with open(path) as f: content = f.read() assert "f1" in content assert "f2" in content sys.settrace(None) os.remove(path) def test_default_vllm_root_logger_configuration(monkeypatch): """This test presumes that VLLM_CONFIGURE_LOGGING (default: True) and VLLM_LOGGING_CONFIG_PATH (default: None) are not configured and default behavior is activated.""" monkeypatch.setenv("VLLM_LOGGING_COLOR", "0") _configure_vllm_root_logger() logger = logging.getLogger("vllm") assert logger.level == logging.INFO assert not logger.propagate handler = logger.handlers[0] assert isinstance(handler, logging.StreamHandler) assert handler.stream == sys.stdout # we use DEBUG level for testing by default # assert handler.level == logging.INFO formatter = handler.formatter assert formatter is not None assert isinstance(formatter, NewLineFormatter) assert formatter._fmt == _FORMAT assert formatter.datefmt == _DATE_FORMAT def test_descendent_loggers_depend_on_and_propagate_logs_to_root_logger(monkeypatch): """This test presumes that VLLM_CONFIGURE_LOGGING (default: True) and VLLM_LOGGING_CONFIG_PATH (default: None) are not configured and default behavior is activated.""" monkeypatch.setenv("VLLM_CONFIGURE_LOGGING", "1") monkeypatch.delenv("VLLM_LOGGING_CONFIG_PATH", raising=False) root_logger = logging.getLogger("vllm") root_handler = root_logger.handlers[0] unique_name = f"vllm.{uuid4()}" logger = init_logger(unique_name) assert logger.name == unique_name assert logger.level == logging.NOTSET assert not logger.handlers assert logger.propagate message = "Hello, world!" with patch.object(root_handler, "emit") as root_handle_mock: logger.info(message) root_handle_mock.assert_called_once() _, call_args, _ = root_handle_mock.mock_calls[0] log_record = call_args[0] assert unique_name == log_record.name assert message == log_record.msg assert message == log_record.msg assert log_record.levelno == logging.INFO def test_logger_configuring_can_be_disabled(monkeypatch): """This test calls _configure_vllm_root_logger again to test custom logging config behavior, however mocks are used to ensure no changes in behavior or configuration occur.""" monkeypatch.setenv("VLLM_CONFIGURE_LOGGING", "0") monkeypatch.delenv("VLLM_LOGGING_CONFIG_PATH", raising=False) with patch("vllm.logger.dictConfig") as dict_config_mock: _configure_vllm_root_logger() dict_config_mock.assert_not_called() def test_an_error_is_raised_when_custom_logging_config_file_does_not_exist(monkeypatch): """This test calls _configure_vllm_root_logger again to test custom logging config behavior, however it fails before any change in behavior or configuration occurs.""" monkeypatch.setenv("VLLM_CONFIGURE_LOGGING", "1") monkeypatch.setenv( "VLLM_LOGGING_CONFIG_PATH", "/if/there/is/a/file/here/then/you/did/this/to/yourself.json", ) with pytest.raises(RuntimeError) as ex_info: _configure_vllm_root_logger() assert ex_info.type == RuntimeError # noqa: E721 assert "File does not exist" in str(ex_info) def test_an_error_is_raised_when_custom_logging_config_is_invalid_json(monkeypatch): """This test calls _configure_vllm_root_logger again to test custom logging config behavior, however it fails before any change in behavior or configuration occurs.""" monkeypatch.setenv("VLLM_CONFIGURE_LOGGING", "1") with NamedTemporaryFile(encoding="utf-8", mode="w") as logging_config_file: logging_config_file.write("---\nloggers: []\nversion: 1") logging_config_file.flush() monkeypatch.setenv("VLLM_LOGGING_CONFIG_PATH", logging_config_file.name) with pytest.raises(JSONDecodeError) as ex_info: _configure_vllm_root_logger() assert ex_info.type == JSONDecodeError assert "Expecting value" in str(ex_info) @pytest.mark.parametrize( "unexpected_config", ( "Invalid string", [{"version": 1, "loggers": []}], 0, ), ) def test_an_error_is_raised_when_custom_logging_config_is_unexpected_json( monkeypatch, unexpected_config: Any, ): """This test calls _configure_vllm_root_logger again to test custom logging config behavior, however it fails before any change in behavior or configuration occurs.""" monkeypatch.setenv("VLLM_CONFIGURE_LOGGING", "1") with NamedTemporaryFile(encoding="utf-8", mode="w") as logging_config_file: logging_config_file.write(json.dumps(unexpected_config)) logging_config_file.flush() monkeypatch.setenv("VLLM_LOGGING_CONFIG_PATH", logging_config_file.name) with pytest.raises(ValueError) as ex_info: _configure_vllm_root_logger() assert ex_info.type == ValueError # noqa: E721 assert "Invalid logging config. Expected dict, got" in str(ex_info) def test_custom_logging_config_is_parsed_and_used_when_provided(monkeypatch): """This test calls _configure_vllm_root_logger again to test custom logging config behavior, however mocks are used to ensure no changes in behavior or configuration occur.""" monkeypatch.setenv("VLLM_CONFIGURE_LOGGING", "1") valid_logging_config = { "loggers": { "vllm.test_logger.logger": { "handlers": [], "propagate": False, } }, "version": 1, } with NamedTemporaryFile(encoding="utf-8", mode="w") as logging_config_file: logging_config_file.write(json.dumps(valid_logging_config)) logging_config_file.flush() monkeypatch.setenv("VLLM_LOGGING_CONFIG_PATH", logging_config_file.name) with patch("vllm.logger.dictConfig") as dict_config_mock: _configure_vllm_root_logger() dict_config_mock.assert_called_with(valid_logging_config) def test_custom_logging_config_causes_an_error_if_configure_logging_is_off(monkeypatch): """This test calls _configure_vllm_root_logger again to test custom logging config behavior, however mocks are used to ensure no changes in behavior or configuration occur.""" monkeypatch.setenv("VLLM_CONFIGURE_LOGGING", "0") valid_logging_config = { "loggers": { "vllm.test_logger.logger": { "handlers": [], } }, "version": 1, } with NamedTemporaryFile(encoding="utf-8", mode="w") as logging_config_file: logging_config_file.write(json.dumps(valid_logging_config)) logging_config_file.flush() monkeypatch.setenv("VLLM_LOGGING_CONFIG_PATH", logging_config_file.name) with pytest.raises(RuntimeError) as ex_info: _configure_vllm_root_logger() assert ex_info.type is RuntimeError expected_message_snippet = ( "VLLM_CONFIGURE_LOGGING evaluated to false, but " "VLLM_LOGGING_CONFIG_PATH was given." ) assert expected_message_snippet in str(ex_info) # Remember! The root logger is assumed to have been configured as # though VLLM_CONFIGURE_LOGGING=1 and VLLM_LOGGING_CONFIG_PATH=None. root_logger = logging.getLogger("vllm") other_logger_name = f"vllm.test_logger.{uuid4()}" other_logger = init_logger(other_logger_name) assert other_logger.handlers != root_logger.handlers assert other_logger.level != root_logger.level assert other_logger.propagate def test_prepare_object_to_dump(): str_obj = "str" assert prepare_object_to_dump(str_obj) == "'str'" list_obj = [1, 2, 3] assert prepare_object_to_dump(list_obj) == "[1, 2, 3]" dict_obj = {"a": 1, "b": "b"} assert prepare_object_to_dump(dict_obj) in [ "{a: 1, b: 'b'}", "{b: 'b', a: 1}", ] set_obj = {1, 2, 3} assert prepare_object_to_dump(set_obj) == "[1, 2, 3]" tuple_obj = ("a", "b", "c") assert prepare_object_to_dump(tuple_obj) == "['a', 'b', 'c']" class CustomEnum(enum.Enum): A = enum.auto() B = enum.auto() C = enum.auto() assert prepare_object_to_dump(CustomEnum.A) == repr(CustomEnum.A) @dataclass class CustomClass: a: int b: str assert prepare_object_to_dump(CustomClass(1, "b")) == "CustomClass(a=1, b='b')" def test_request_logger_log_outputs(): """Test the new log_outputs functionality.""" # Create a mock logger to capture log calls mock_logger = MagicMock() with patch("vllm.entrypoints.logger.logger", mock_logger): request_logger = RequestLogger(max_log_len=None) # Test basic output logging request_logger.log_outputs( request_id="test-123", outputs="Hello, world!", output_token_ids=[1, 2, 3, 4], finish_reason="stop", is_streaming=False, delta=False, ) mock_logger.info.assert_called_once() call_args = mock_logger.info.call_args.args assert "Generated response %s%s" in call_args[0] assert call_args[1] == "test-123" assert call_args[3] == "Hello, world!" assert call_args[4] == [1, 2, 3, 4] assert call_args[5] == "stop" def test_request_logger_log_outputs_streaming_delta(): """Test log_outputs with streaming delta mode.""" mock_logger = MagicMock() with patch("vllm.entrypoints.logger.logger", mock_logger): request_logger = RequestLogger(max_log_len=None) # Test streaming delta logging request_logger.log_outputs( request_id="test-456", outputs="Hello", output_token_ids=[1], finish_reason=None, is_streaming=True, delta=True, ) mock_logger.info.assert_called_once() call_args = mock_logger.info.call_args.args assert "Generated response %s%s" in call_args[0] assert call_args[1] == "test-456" assert call_args[2] == " (streaming delta)" assert call_args[3] == "Hello" assert call_args[4] == [1] assert call_args[5] is None def test_request_logger_log_outputs_streaming_complete(): """Test log_outputs with streaming complete mode.""" mock_logger = MagicMock() with patch("vllm.entrypoints.logger.logger", mock_logger): request_logger = RequestLogger(max_log_len=None) # Test streaming complete logging request_logger.log_outputs( request_id="test-789", outputs="Complete response", output_token_ids=[1, 2, 3], finish_reason="length", is_streaming=True, delta=False, ) mock_logger.info.assert_called_once() call_args = mock_logger.info.call_args.args assert "Generated response %s%s" in call_args[0] assert call_args[1] == "test-789" assert call_args[2] == " (streaming complete)" assert call_args[3] == "Complete response" assert call_args[4] == [1, 2, 3] assert call_args[5] == "length" def test_request_logger_log_outputs_with_truncation(): """Test log_outputs respects max_log_len setting.""" mock_logger = MagicMock() with patch("vllm.entrypoints.logger.logger", mock_logger): # Set max_log_len to 10 request_logger = RequestLogger(max_log_len=10) # Test output truncation long_output = "This is a very long output that should be truncated" long_token_ids = list(range(20)) # 20 tokens request_logger.log_outputs( request_id="test-truncate", outputs=long_output, output_token_ids=long_token_ids, finish_reason="stop", is_streaming=False, delta=False, ) mock_logger.info.assert_called_once() call_args = mock_logger.info.call_args # Check that output was truncated to first 10 characters logged_output = call_args[0][3] assert logged_output == "This is a " assert len(logged_output) == 10 # Check that token IDs were truncated to first 10 tokens logged_token_ids = call_args[0][4] assert logged_token_ids == list(range(10)) assert len(logged_token_ids) == 10 def test_request_logger_log_outputs_none_values(): """Test log_outputs handles None values correctly.""" mock_logger = MagicMock() with patch("vllm.entrypoints.logger.logger", mock_logger): request_logger = RequestLogger(max_log_len=None) # Test with None output_token_ids request_logger.log_outputs( request_id="test-none", outputs="Test output", output_token_ids=None, finish_reason="stop", is_streaming=False, delta=False, ) mock_logger.info.assert_called_once() call_args = mock_logger.info.call_args.args assert "Generated response %s%s" in call_args[0] assert call_args[1] == "test-none" assert call_args[3] == "Test output" assert call_args[4] is None assert call_args[5] == "stop" def test_request_logger_log_outputs_empty_output(): """Test log_outputs handles empty output correctly.""" mock_logger = MagicMock() with patch("vllm.entrypoints.logger.logger", mock_logger): request_logger = RequestLogger(max_log_len=5) # Test with empty output request_logger.log_outputs( request_id="test-empty", outputs="", output_token_ids=[], finish_reason="stop", is_streaming=False, delta=False, ) mock_logger.info.assert_called_once() call_args = mock_logger.info.call_args.args assert "Generated response %s%s" in call_args[0] assert call_args[1] == "test-empty" assert call_args[3] == "" assert call_args[4] == [] assert call_args[5] == "stop" def test_request_logger_log_outputs_integration(): """Test that log_outputs can be called alongside log_inputs.""" mock_logger = MagicMock() with patch("vllm.entrypoints.logger.logger", mock_logger): request_logger = RequestLogger(max_log_len=None) # Test that both methods can be called without interference request_logger.log_inputs( request_id="test-integration", prompt="Test prompt", prompt_token_ids=[1, 2, 3], prompt_embeds=None, params=None, lora_request=None, ) request_logger.log_outputs( request_id="test-integration", outputs="Test output", output_token_ids=[4, 5, 6], finish_reason="stop", is_streaming=False, delta=False, ) # Should have been called twice - once for inputs, once for outputs assert mock_logger.info.call_count == 2 # Check that the calls were made with correct patterns input_call = mock_logger.info.call_args_list[0][0] output_call = mock_logger.info.call_args_list[1][0] assert "Received request %s" in input_call[0] assert input_call[1] == "test-integration" assert "Generated response %s%s" in output_call[0] assert output_call[1] == "test-integration" def test_streaming_complete_logs_full_text_content(): """Test that streaming complete logging includes full accumulated text, not just token count.""" mock_logger = MagicMock() with patch("vllm.entrypoints.logger.logger", mock_logger): request_logger = RequestLogger(max_log_len=None) # Test with actual content instead of token count format full_response = "This is a complete response from streaming" request_logger.log_outputs( request_id="test-streaming-full-text", outputs=full_response, output_token_ids=None, finish_reason="streaming_complete", is_streaming=True, delta=False, ) mock_logger.info.assert_called_once() call_args = mock_logger.info.call_args.args # Verify the logged output is the full text, not a token count format logged_output = call_args[3] assert logged_output == full_response assert "tokens>" not in logged_output assert "streaming_complete" not in logged_output # Verify other parameters assert call_args[1] == "test-streaming-full-text" assert call_args[2] == " (streaming complete)" assert call_args[5] == "streaming_complete" # Add vllm prefix to make sure logs go through the vllm logger test_logger = init_logger("vllm.test_logger") def mp_function(**kwargs): # This function runs in a subprocess test_logger.warning("This is a subprocess: %s", kwargs.get("a")) test_logger.error("This is a subprocess error.") test_logger.debug("This is a subprocess debug message: %s.", kwargs.get("b")) def test_caplog_mp_fork(caplog_vllm, caplog_mp_fork): with caplog_vllm.at_level(logging.DEBUG, logger="vllm"), caplog_mp_fork(): import multiprocessing ctx = multiprocessing.get_context("fork") p = ctx.Process( target=mp_function, name=f"SubProcess{1}", kwargs={"a": "AAAA", "b": "BBBBB"}, ) p.start() p.join() assert "AAAA" in caplog_vllm.text assert "BBBBB" in caplog_vllm.text def test_caplog_mp_spawn(caplog_mp_spawn): with caplog_mp_spawn(logging.DEBUG) as log_holder: import multiprocessing ctx = multiprocessing.get_context("spawn") p = ctx.Process( target=mp_function, name=f"SubProcess{1}", kwargs={"a": "AAAA", "b": "BBBBB"}, ) p.start() p.join() assert "AAAA" in log_holder.text assert "BBBBB" in log_holder.text
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_vllm_port.py
tests/test_vllm_port.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os from unittest.mock import patch import pytest from vllm.envs import get_vllm_port def test_get_vllm_port_not_set(): """Test when VLLM_PORT is not set.""" with patch.dict(os.environ, {}, clear=True): assert get_vllm_port() is None def test_get_vllm_port_valid(): """Test when VLLM_PORT is set to a valid integer.""" with patch.dict(os.environ, {"VLLM_PORT": "5678"}, clear=True): assert get_vllm_port() == 5678 def test_get_vllm_port_invalid(): """Test when VLLM_PORT is set to a non-integer value.""" with ( patch.dict(os.environ, {"VLLM_PORT": "abc"}, clear=True), pytest.raises(ValueError, match="must be a valid integer"), ): get_vllm_port() def test_get_vllm_port_uri(): """Test when VLLM_PORT is set to a URI.""" with ( patch.dict(os.environ, {"VLLM_PORT": "tcp://localhost:5678"}, clear=True), pytest.raises(ValueError, match="appears to be a URI"), ): get_vllm_port()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_seed_behavior.py
tests/test_seed_behavior.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import random import numpy as np import torch from vllm.platforms.interface import Platform def test_seed_behavior(): # Test with a specific seed Platform.seed_everything(42) random_value_1 = random.randint(0, 100) np_random_value_1 = np.random.randint(0, 100) torch_random_value_1 = torch.randint(0, 100, (1,)).item() Platform.seed_everything(42) random_value_2 = random.randint(0, 100) np_random_value_2 = np.random.randint(0, 100) torch_random_value_2 = torch.randint(0, 100, (1,)).item() assert random_value_1 == random_value_2 assert np_random_value_1 == np_random_value_2 assert torch_random_value_1 == torch_random_value_2
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_config.py
tests/test_config.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import logging import os from dataclasses import MISSING, Field, asdict, dataclass, field from unittest.mock import patch import pytest from pydantic import ValidationError from vllm.compilation.backends import VllmBackend from vllm.config import ( CompilationConfig, ModelConfig, PoolerConfig, SchedulerConfig, VllmConfig, update_config, ) from vllm.config.compilation import CompilationMode, CUDAGraphMode from vllm.config.load import LoadConfig from vllm.config.utils import get_field from vllm.config.vllm import ( OPTIMIZATION_LEVEL_TO_CONFIG, OptimizationLevel, ) from vllm.model_executor.layers.pooler import PoolingType from vllm.platforms import current_platform def test_compile_config_repr_succeeds(): # setup: VllmBackend mutates the config object config = VllmConfig() backend = VllmBackend(config) backend.configure_post_pass() # test that repr(config) succeeds val = repr(config) assert "VllmConfig" in val assert "inductor_passes" in val @dataclass class _TestConfigFields: a: int b: dict = field(default_factory=dict) c: str = "default" def test_get_field(): with pytest.raises(ValueError): get_field(_TestConfigFields, "a") b = get_field(_TestConfigFields, "b") assert isinstance(b, Field) assert b.default is MISSING assert b.default_factory is dict c = get_field(_TestConfigFields, "c") assert isinstance(c, Field) assert c.default == "default" assert c.default_factory is MISSING @dataclass class _TestNestedConfig: a: _TestConfigFields = field(default_factory=lambda: _TestConfigFields(a=0)) def test_update_config(): # Simple update config1 = _TestConfigFields(a=0) new_config1 = update_config(config1, {"a": 42}) assert new_config1.a == 42 # Nonexistent field with pytest.raises(AssertionError): new_config1 = update_config(config1, {"nonexistent": 1}) # Nested update with dataclass config2 = _TestNestedConfig() new_inner_config = _TestConfigFields(a=1, c="new_value") new_config2 = update_config(config2, {"a": new_inner_config}) assert new_config2.a == new_inner_config # Nested update with dict config3 = _TestNestedConfig() new_config3 = update_config(config3, {"a": {"c": "new_value"}}) assert new_config3.a.c == "new_value" # Nested update with invalid type with pytest.raises(AssertionError): new_config3 = update_config(config3, {"a": "new_value"}) @pytest.mark.parametrize( ("model_id", "expected_runner_type", "expected_convert_type"), [ ("distilbert/distilgpt2", "generate", "none"), ("intfloat/multilingual-e5-small", "pooling", "none"), ("jason9693/Qwen2.5-1.5B-apeach", "pooling", "classify"), ("cross-encoder/ms-marco-MiniLM-L-6-v2", "pooling", "none"), ("Qwen/Qwen2.5-Math-RM-72B", "pooling", "none"), ("openai/whisper-small", "generate", "none"), ], ) def test_auto_runner(model_id, expected_runner_type, expected_convert_type): config = ModelConfig(model_id, runner="auto") assert config.runner_type == expected_runner_type assert config.convert_type == expected_convert_type @pytest.mark.parametrize( ("model_id", "expected_runner_type", "expected_convert_type"), [ ("distilbert/distilgpt2", "pooling", "embed"), ("intfloat/multilingual-e5-small", "pooling", "none"), ("jason9693/Qwen2.5-1.5B-apeach", "pooling", "classify"), ("cross-encoder/ms-marco-MiniLM-L-6-v2", "pooling", "none"), ("Qwen/Qwen2.5-Math-RM-72B", "pooling", "none"), ("openai/whisper-small", "pooling", "embed"), ], ) def test_pooling_runner(model_id, expected_runner_type, expected_convert_type): config = ModelConfig(model_id, runner="pooling") assert config.runner_type == expected_runner_type assert config.convert_type == expected_convert_type @pytest.mark.parametrize( ("model_id", "expected_runner_type", "expected_convert_type"), [ ("Qwen/Qwen2.5-1.5B-Instruct", "draft", "none"), ], ) def test_draft_runner(model_id, expected_runner_type, expected_convert_type): config = ModelConfig(model_id, runner="draft") assert config.runner_type == expected_runner_type assert config.convert_type == expected_convert_type MODEL_IDS_EXPECTED = [ ("Qwen/Qwen1.5-7B", 32768), ("mistralai/Mistral-7B-v0.1", 4096), ("mistralai/Mistral-7B-Instruct-v0.2", 32768), ] @pytest.mark.parametrize("model_id_expected", MODEL_IDS_EXPECTED) def test_disable_sliding_window(model_id_expected): model_id, expected = model_id_expected model_config = ModelConfig(model_id, disable_sliding_window=True) assert model_config.max_model_len == expected @pytest.mark.skipif( current_platform.is_rocm(), reason="Xformers backend is not supported on ROCm." ) def test_get_pooling_config(): model_id = "sentence-transformers/all-MiniLM-L12-v2" model_config = ModelConfig(model_id) assert model_config.pooler_config is not None assert model_config.pooler_config.normalize assert model_config.pooler_config.pooling_type == PoolingType.MEAN.name @pytest.mark.skipif( current_platform.is_rocm(), reason="Xformers backend is not supported on ROCm." ) def test_get_pooling_config_from_args(): model_id = "sentence-transformers/all-MiniLM-L12-v2" pooler_config = PoolerConfig(pooling_type="CLS", normalize=True) model_config = ModelConfig(model_id, pooler_config=pooler_config) assert asdict(model_config.pooler_config) == asdict(pooler_config) @pytest.mark.parametrize( ("model_id", "default_pooling_type", "pooling_type"), [ ("tomaarsen/Qwen3-Reranker-0.6B-seq-cls", "LAST", "LAST"), # LLM ("intfloat/e5-small", "CLS", "MEAN"), # BertModel ("Qwen/Qwen2.5-Math-RM-72B", "ALL", "ALL"), # reward ("Qwen/Qwen2.5-Math-PRM-7B", "STEP", "STEP"), # step reward ], ) def test_default_pooling_type(model_id, default_pooling_type, pooling_type): model_config = ModelConfig(model_id) assert model_config._model_info.default_pooling_type == default_pooling_type assert model_config.pooler_config.pooling_type == pooling_type @pytest.mark.parametrize( ("model_id", "expected_is_moe_model"), [ ("RedHatAI/Qwen3-8B-speculator.eagle3", False), ("RedHatAI/Llama-3.1-8B-Instruct-NVFP4", False), ("RedHatAI/Llama-3.2-1B-FP8", False), ("RedHatAI/Mistral-Small-24B-Instruct-2501-quantized.w8a8", False), ("RedHatAI/gpt-oss-20b", True), ("RedHatAI/DeepSeek-V2.5-1210-FP8", True), ("RedHatAI/Llama-4-Scout-17B-16E-Instruct", True), ("RedHatAI/Mixtral-8x7B-Instruct-v0.1", True), ], ) def test_moe_model_detection(model_id, expected_is_moe_model): model_config = ModelConfig(model_id) # Just check that is_moe field exists and is a boolean assert model_config.is_moe == expected_is_moe_model @pytest.mark.parametrize( ("model_id", "quantized"), [ ("RedHatAI/Qwen3-8B-speculator.eagle3", False), ("RedHatAI/Llama-3.1-8B-Instruct-NVFP4", True), ("RedHatAI/Llama-3.2-1B-FP8", True), ("RedHatAI/Mistral-Small-24B-Instruct-2501-quantized.w8a8", True), ("RedHatAI/gpt-oss-20b", True), ("RedHatAI/DeepSeek-V2.5-1210-FP8", True), ("RedHatAI/Mixtral-8x7B-Instruct-v0.1", False), ], ) def test_is_quantized(model_id, quantized): model_config = ModelConfig(model_id) # Just check that quantized field exists and is a boolean assert model_config.is_quantized == quantized @pytest.mark.skipif( current_platform.is_rocm(), reason="Xformers backend is not supported on ROCm." ) def test_get_bert_tokenization_sentence_transformer_config(): model_id = "BAAI/bge-base-en-v1.5" bge_model_config = ModelConfig(model_id) bert_bge_model_config = bge_model_config._get_encoder_config() assert bert_bge_model_config["max_seq_length"] == 512 assert bert_bge_model_config["do_lower_case"] def test_rope_customization(): TEST_ROPE_PARAMETERS = { "rope_theta": 16_000_000.0, "rope_type": "dynamic", "factor": 2.0, } LLAMA_ROPE_PARAMETERS = {"rope_theta": 500000.0, "rope_type": "default"} LONGCHAT_ROPE_PARAMETERS = {"rope_type": "linear", "factor": 8.0} llama_model_config = ModelConfig("meta-llama/Meta-Llama-3-8B-Instruct") assert ( getattr(llama_model_config.hf_config, "rope_parameters", None) == LLAMA_ROPE_PARAMETERS ) assert llama_model_config.max_model_len == 8192 llama_model_config = ModelConfig( "meta-llama/Meta-Llama-3-8B-Instruct", hf_overrides={"rope_parameters": TEST_ROPE_PARAMETERS}, ) assert ( getattr(llama_model_config.hf_config, "rope_parameters", None) == TEST_ROPE_PARAMETERS ) assert llama_model_config.max_model_len == 16384 longchat_model_config = ModelConfig("lmsys/longchat-13b-16k") # Check if LONGCHAT_ROPE_PARAMETERS entries are in longchat_model_config assert all( longchat_model_config.hf_config.rope_parameters.get(key) == value for key, value in LONGCHAT_ROPE_PARAMETERS.items() ) assert longchat_model_config.max_model_len == 16384 longchat_model_config = ModelConfig( "lmsys/longchat-13b-16k", hf_overrides={ "rope_parameters": TEST_ROPE_PARAMETERS, }, ) assert ( getattr(longchat_model_config.hf_config, "rope_parameters", None) == TEST_ROPE_PARAMETERS ) assert longchat_model_config.max_model_len == 4096 def test_nested_hf_overrides(): """Test that nested hf_overrides work correctly.""" # Test with a model that has text_config model_config = ModelConfig( "Qwen/Qwen2-VL-2B-Instruct", hf_overrides={ "text_config": { "hidden_size": 1024, }, }, ) assert model_config.hf_config.text_config.hidden_size == 1024 # Test with deeply nested overrides model_config = ModelConfig( "Qwen/Qwen2-VL-2B-Instruct", hf_overrides={ "text_config": { "hidden_size": 2048, "num_attention_heads": 16, }, "vision_config": { "hidden_size": 512, }, }, ) assert model_config.hf_config.text_config.hidden_size == 2048 assert model_config.hf_config.text_config.num_attention_heads == 16 assert model_config.hf_config.vision_config.hidden_size == 512 @pytest.mark.skipif( current_platform.is_rocm(), reason="Encoder Decoder models not supported on ROCm." ) @pytest.mark.parametrize( ("model_id", "is_encoder_decoder"), [ ("facebook/opt-125m", False), ("openai/whisper-tiny", True), ("meta-llama/Llama-3.2-1B-Instruct", False), ], ) def test_is_encoder_decoder(model_id, is_encoder_decoder): config = ModelConfig(model_id) assert config.is_encoder_decoder == is_encoder_decoder @pytest.mark.parametrize( ("model_id", "uses_mrope"), [ ("facebook/opt-125m", False), ("Qwen/Qwen2-VL-2B-Instruct", True), ], ) def test_uses_mrope(model_id, uses_mrope): config = ModelConfig(model_id) assert config.uses_mrope == uses_mrope def test_generation_config_loading(): model_id = "Qwen/Qwen2.5-1.5B-Instruct" # When set generation_config to "vllm", the default generation config # will not be loaded. model_config = ModelConfig(model_id, generation_config="vllm") assert model_config.get_diff_sampling_param() == {} # When set generation_config to "auto", the default generation config # should be loaded. model_config = ModelConfig(model_id, generation_config="auto") correct_generation_config = { "repetition_penalty": 1.1, "temperature": 0.7, "top_p": 0.8, "top_k": 20, } assert model_config.get_diff_sampling_param() == correct_generation_config # The generation config could be overridden by the user. override_generation_config = {"temperature": 0.5, "top_k": 5} model_config = ModelConfig( model_id, generation_config="auto", override_generation_config=override_generation_config, ) override_result = correct_generation_config.copy() override_result.update(override_generation_config) assert model_config.get_diff_sampling_param() == override_result # When generation_config is set to "vllm" and override_generation_config # is set, the override_generation_config should be used directly. model_config = ModelConfig( model_id, generation_config="vllm", override_generation_config=override_generation_config, ) assert model_config.get_diff_sampling_param() == override_generation_config @pytest.mark.parametrize( "pt_load_map_location", [ "cuda", {"": "cuda"}, ], ) def test_load_config_pt_load_map_location(pt_load_map_location): load_config = LoadConfig(pt_load_map_location=pt_load_map_location) config = VllmConfig(load_config=load_config) assert config.load_config.pt_load_map_location == pt_load_map_location @pytest.mark.parametrize( ("model_id", "max_model_len", "expected_max_len", "should_raise"), [ ("BAAI/bge-reranker-base", None, 512, False), ("BAAI/bge-reranker-base", 256, 256, False), ("BAAI/bge-reranker-base", 513, 512, True), ("deepseek-ai/DeepSeek-R1-Distill-Qwen-7B", None, 131072, False), ("deepseek-ai/DeepSeek-R1-Distill-Qwen-7B", 131073, 131072, True), ], ) def test_get_and_verify_max_len( model_id, max_model_len, expected_max_len, should_raise ): """Test get_and_verify_max_len with different configurations.""" model_config = ModelConfig(model_id) if should_raise: with pytest.raises(ValueError): model_config.get_and_verify_max_len(max_model_len) else: actual_max_len = model_config.get_and_verify_max_len(max_model_len) assert actual_max_len == expected_max_len class MockConfig: """Simple mock object for testing maybe_pull_model_tokenizer_for_runai""" def __init__(self, model: str, tokenizer: str): self.model = model self.tokenizer = tokenizer self.model_weights = None @pytest.mark.parametrize( "s3_url", [ "s3://example-bucket-1/model/", "s3://example-bucket-2/model/", ], ) @patch("vllm.transformers_utils.runai_utils.ObjectStorageModel.pull_files") def test_s3_url_model_tokenizer_paths(mock_pull_files, s3_url): """Test that S3 URLs create deterministic local directories for model and tokenizer.""" # Mock pull_files to avoid actually downloading files during tests mock_pull_files.return_value = None # Create first mock and run the method config1 = MockConfig(model=s3_url, tokenizer=s3_url) ModelConfig.maybe_pull_model_tokenizer_for_runai(config1, s3_url, s3_url) # Check that model and tokenizer point to existing directories assert os.path.exists(config1.model), ( f"Model directory does not exist: {config1.model}" ) assert os.path.isdir(config1.model), ( f"Model path is not a directory: {config1.model}" ) assert os.path.exists(config1.tokenizer), ( f"Tokenizer directory does not exist: {config1.tokenizer}" ) assert os.path.isdir(config1.tokenizer), ( f"Tokenizer path is not a directory: {config1.tokenizer}" ) # Verify that the paths are different from the original S3 URL assert config1.model != s3_url, "Model path should be converted to local directory" assert config1.tokenizer != s3_url, ( "Tokenizer path should be converted to local directory" ) # Store the original paths created_model_dir = config1.model create_tokenizer_dir = config1.tokenizer # Create a new mock and run the method with the same S3 URL config2 = MockConfig(model=s3_url, tokenizer=s3_url) ModelConfig.maybe_pull_model_tokenizer_for_runai(config2, s3_url, s3_url) # Check that the new directories exist assert os.path.exists(config2.model), ( f"Model directory does not exist: {config2.model}" ) assert os.path.isdir(config2.model), ( f"Model path is not a directory: {config2.model}" ) assert os.path.exists(config2.tokenizer), ( f"Tokenizer directory does not exist: {config2.tokenizer}" ) assert os.path.isdir(config2.tokenizer), ( f"Tokenizer path is not a directory: {config2.tokenizer}" ) # Verify that the paths are deterministic (same as before) assert config2.model == created_model_dir, ( f"Model paths are not deterministic. " f"Original: {created_model_dir}, New: {config2.model}" ) assert config2.tokenizer == create_tokenizer_dir, ( f"Tokenizer paths are not deterministic. " f"Original: {create_tokenizer_dir}, New: {config2.tokenizer}" ) @patch("vllm.transformers_utils.runai_utils.ObjectStorageModel.pull_files") def test_s3_url_different_models_create_different_directories(mock_pull_files): """Test that different S3 URLs create different local directories.""" # Mock pull_files to avoid actually downloading files during tests mock_pull_files.return_value = None s3_url1 = "s3://example-bucket-1/model/" s3_url2 = "s3://example-bucket-2/model/" # Create mocks with different S3 URLs and run the method config1 = MockConfig(model=s3_url1, tokenizer=s3_url1) ModelConfig.maybe_pull_model_tokenizer_for_runai(config1, s3_url1, s3_url1) config2 = MockConfig(model=s3_url2, tokenizer=s3_url2) ModelConfig.maybe_pull_model_tokenizer_for_runai(config2, s3_url2, s3_url2) # Verify that different URLs produce different directories assert config1.model != config2.model, ( f"Different S3 URLs should create different model directories. " f"URL1 model: {config1.model}, URL2 model: {config2.model}" ) assert config1.tokenizer != config2.tokenizer, ( f"Different S3 URLs should create different tokenizer directories. " f"URL1 tokenizer: {config1.tokenizer}, " f"URL2 tokenizer: {config2.tokenizer}" ) # Verify that both sets of directories exist assert os.path.exists(config1.model) and os.path.isdir(config1.model) assert os.path.exists(config1.tokenizer) and os.path.isdir(config1.tokenizer) assert os.path.exists(config2.model) and os.path.isdir(config2.model) assert os.path.exists(config2.tokenizer) and os.path.isdir(config2.tokenizer) @pytest.mark.parametrize( ("model_id", "expected_attn_type", "expected_result", "reason"), [ # pooling models ( "jason9693/Qwen2.5-1.5B-apeach", "decoder", True, "Pooling models with causal attn and last pooling support chunked prefill.", ), ( "Qwen/Qwen3-Embedding-0.6B", "decoder", True, "Pooling models with causal attn and last pooling support chunked prefill.", ), ( "Qwen/Qwen2.5-Math-PRM-7B", "decoder", False, "Pooling models with step pooling does not support chunked prefill.", ), ( "internlm/internlm2-1_8b-reward", "decoder", True, "Pooling models with causal attn and all pooling support chunked prefill.", ), ( "BAAI/bge-base-en", "encoder_only", False, "Pooling models with bidirectional attn does not support chunked prefill.", ), ( "boltuix/NeuroBERT-NER", "encoder_only", False, "Pooling models with bidirectional attn does not support chunked prefill.", ), ( "papluca/xlm-roberta-base-language-detection", "encoder_only", False, "Pooling models with bidirectional attn does not support chunked prefill.", ), ( "Alibaba-NLP/gte-Qwen2-1.5B-instruct", "encoder_only", False, "Pooling models with bidirectional attn does not support chunked prefill.", ), ( "intfloat/e5-small", "encoder_only", False, "Pooling models with bidirectional attn does not support chunked prefill.", ), # multimodal models ( "openai/clip-vit-base-patch32", "decoder", True, "Pooling models with causal attn and last pooling support chunked prefill.", ), ( "google/siglip-base-patch16-224", "encoder_only", False, "Pooling models with bidirectional attn does not support chunked prefill.", ), # generate models ( "Qwen/Qwen3-0.6B", "decoder", True, "Generative models support chunked prefill.", ), ( "Qwen/Qwen3-Next-80B-A3B-Instruct", "hybrid", True, "Generative models support chunked prefill.", ), ( "ibm-granite/granite-4.0-h-small", "hybrid", True, "Generative models support chunked prefill.", ), ( "state-spaces/mamba-130m-hf", "attention_free", True, "Generative models support chunked prefill.", ), # encoder_decoder models ( "openai/whisper-small", "encoder_decoder", False, "Encoder decoder models does not support chunked prefill.", ), ], ) def test_is_chunked_prefill_supported( model_id: str, expected_attn_type: str, expected_result: bool, reason: str, caplog_vllm, ): model_config = ModelConfig(model_id, trust_remote_code=True) assert model_config.attn_type == expected_attn_type with caplog_vllm.at_level(level=logging.DEBUG, logger="vllm"): assert model_config.is_chunked_prefill_supported == expected_result assert reason in caplog_vllm.text @pytest.mark.parametrize( ("model_id", "expected_attn_type", "expected_result", "reason"), [ # pooling models ( "jason9693/Qwen2.5-1.5B-apeach", "decoder", True, "Pooling models with causal attn and last pooling support prefix caching.", ), ( "Qwen/Qwen3-Embedding-0.6B", "decoder", True, "Pooling models with causal attn and last pooling support prefix caching.", ), ( "Qwen/Qwen2.5-Math-PRM-7B", "decoder", False, "Pooling models with step pooling does not support prefix caching.", ), ( "internlm/internlm2-1_8b-reward", "decoder", True, "Pooling models with causal attn and all pooling support prefix caching.", ), ( "BAAI/bge-base-en", "encoder_only", False, "Pooling models with bidirectional attn does not support prefix caching.", ), ( "boltuix/NeuroBERT-NER", "encoder_only", False, "Pooling models with bidirectional attn does not support prefix caching.", ), ( "papluca/xlm-roberta-base-language-detection", "encoder_only", False, "Pooling models with bidirectional attn does not support prefix caching.", ), ( "Alibaba-NLP/gte-Qwen2-1.5B-instruct", "encoder_only", False, "Pooling models with bidirectional attn does not support prefix caching.", ), ( "intfloat/e5-small", "encoder_only", False, "Pooling models with bidirectional attn does not support prefix caching.", ), # multimodal models ( "openai/clip-vit-base-patch32", "decoder", True, "Pooling models with causal attn and last pooling support prefix caching.", ), ( "google/siglip-base-patch16-224", "encoder_only", False, "Pooling models with bidirectional attn does not support prefix caching.", ), # generate models ( "Qwen/Qwen3-0.6B", "decoder", True, "Generative models support prefix caching.", ), ( "Qwen/Qwen3-Next-80B-A3B-Instruct", "hybrid", False, "Hybrid models does not support prefix caching since the feature is still experimental.", # noqa: E501 ), ( "ibm-granite/granite-4.0-h-small", "hybrid", False, "Hybrid models does not support prefix caching since the feature is still experimental.", # noqa: E501 ), ( "state-spaces/mamba-130m-hf", "attention_free", False, "Attention free models does not support prefix caching since the feature is still experimental.", # noqa: E501 ), # encoder_decoder models ( "openai/whisper-small", "encoder_decoder", False, "Encoder decoder models does not support prefix caching.", ), ], ) def test_is_prefix_caching_supported( model_id: str, expected_attn_type: str, expected_result: bool, reason: str, caplog_vllm, ): model_config = ModelConfig(model_id, trust_remote_code=True) assert model_config.attn_type == expected_attn_type with caplog_vllm.at_level(level=logging.DEBUG, logger="vllm"): assert model_config.is_prefix_caching_supported == expected_result assert reason in caplog_vllm.text @pytest.mark.parametrize( ("backend", "custom_ops", "expected"), [ ("eager", [], True), ("eager", ["+fused_layernorm"], True), ("eager", ["all", "-fused_layernorm"], False), ("inductor", [], False), ("inductor", ["none", "+fused_layernorm"], True), ("inductor", ["none", "-fused_layernorm"], False), ], ) def test_is_custom_op_enabled(backend: str, custom_ops: list[str], expected: bool): """Test that is_custom_op_enabled works correctly.""" config = VllmConfig( compilation_config=CompilationConfig(backend=backend, custom_ops=custom_ops) ) assert config.compilation_config.is_custom_op_enabled("fused_layernorm") is expected def test_vllm_config_defaults_are_none(): """Verify that optimization-level defaults are None when not set by user.""" # Test all optimization levels to ensure defaults work correctly for opt_level in OptimizationLevel: config = object.__new__(VllmConfig) config.compilation_config = CompilationConfig() config.optimization_level = opt_level config.model_config = None # Use the global optimization level defaults default_config = OPTIMIZATION_LEVEL_TO_CONFIG[opt_level] # Verify that all pass_config values are None before defaults are applied for pass_k in default_config["compilation_config"]["pass_config"]: assert getattr(config.compilation_config.pass_config, pass_k) is None # Verify that other config values are None before defaults are applied for k in default_config["compilation_config"]: if k != "pass_config": assert getattr(config.compilation_config, k) is None @pytest.mark.parametrize( ("model_id", "compiliation_config", "optimization_level"), [ ( None, CompilationConfig(backend="eager", custom_ops=["+quant_fp8"]), OptimizationLevel.O0, ), (None, CompilationConfig(), OptimizationLevel.O0), (None, CompilationConfig(), OptimizationLevel.O1), (None, CompilationConfig(), OptimizationLevel.O2), (None, CompilationConfig(), OptimizationLevel.O3), ( "RedHatAI/Qwen3-8B-speculator.eagle3", CompilationConfig(backend="inductor", custom_ops=["+quant_fp8"]), OptimizationLevel.O2, ), ( "RedHatAI/Qwen3-8B-speculator.eagle3", CompilationConfig(), OptimizationLevel.O0, ), ( "RedHatAI/Qwen3-8B-speculator.eagle3", CompilationConfig(), OptimizationLevel.O1, ), ( "RedHatAI/Qwen3-8B-speculator.eagle3", CompilationConfig(), OptimizationLevel.O2, ), ( "RedHatAI/Qwen3-8B-speculator.eagle3", CompilationConfig(), OptimizationLevel.O3, ), ("RedHatAI/DeepSeek-V2.5-1210-FP8", CompilationConfig(), OptimizationLevel.O0), ("RedHatAI/DeepSeek-V2.5-1210-FP8", CompilationConfig(), OptimizationLevel.O1), ("RedHatAI/DeepSeek-V2.5-1210-FP8", CompilationConfig(), OptimizationLevel.O2), ("RedHatAI/DeepSeek-V2.5-1210-FP8", CompilationConfig(), OptimizationLevel.O3), ], ) def test_vllm_config_defaults(model_id, compiliation_config, optimization_level): """Test that optimization-level defaults are correctly applied.""" model_config = None if model_id is not None: model_config = ModelConfig(model_id) vllm_config = VllmConfig( model_config=model_config, compilation_config=compiliation_config, optimization_level=optimization_level, ) else: vllm_config = VllmConfig( compilation_config=compiliation_config, optimization_level=optimization_level, ) # Use the global optimization level defaults default_config = OPTIMIZATION_LEVEL_TO_CONFIG[optimization_level] # Verify pass_config defaults (nested under compilation_config) pass_config_dict = default_config["compilation_config"]["pass_config"] for pass_k, pass_v in pass_config_dict.items(): actual = getattr(vllm_config.compilation_config.pass_config, pass_k) expected = pass_v(vllm_config) if callable(pass_v) else pass_v assert actual == expected, ( f"pass_config.{pass_k}: expected {expected}, got {actual}" ) # Verify other compilation_config defaults compilation_config_dict = default_config["compilation_config"] for k, v in compilation_config_dict.items(): if k != "pass_config": actual = getattr(vllm_config.compilation_config, k) expected = v(vllm_config) if callable(v) else v assert actual == expected, ( f"compilation_config.{k}: expected {expected}, got {actual}" ) def test_vllm_config_callable_defaults(): """Test that callable defaults work in the config system. Verifies that lambdas in default configs can inspect VllmConfig properties (e.g., is_quantized, is_model_moe) to conditionally set optimization flags. """ config_no_model = VllmConfig(optimization_level=OptimizationLevel.O2) # Callable that checks if model exists has_model = lambda cfg: cfg.model_config is not None assert has_model(config_no_model) is False # Test with quantized model quantized_model = ModelConfig("RedHatAI/Llama-3.2-1B-FP8") config_quantized = VllmConfig( model_config=quantized_model, optimization_level=OptimizationLevel.O2 ) enable_if_quantized = lambda cfg: ( cfg.model_config is not None and cfg.model_config.is_quantized ) assert enable_if_quantized(config_quantized) is True assert enable_if_quantized(config_no_model) is False # Test with MoE model moe_model = ModelConfig("deepseek-ai/DeepSeek-V2-Lite") config_moe = VllmConfig( model_config=moe_model, optimization_level=OptimizationLevel.O2 ) enable_if_sequential = lambda cfg: ( cfg.model_config is not None and not cfg.model_config.is_moe ) assert enable_if_sequential(config_moe) is False assert enable_if_sequential(config_quantized) is True def test_vllm_config_explicit_overrides(): """Test that explicit property overrides work correctly with callable defaults.
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_pooling_params.py
tests/test_pooling_params.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass import pytest from tests.models.utils import EmbedModelInfo from vllm import PoolingParams from vllm.config import ModelConfig, PoolerConfig EMBEDDING_MODELS = [ EmbedModelInfo("intfloat/multilingual-e5-small", is_matryoshka=False), EmbedModelInfo( "Snowflake/snowflake-arctic-embed-m-v1.5", is_matryoshka=True, matryoshka_dimensions=[256], ), ] classify_parameters = ["use_activation"] embed_parameters = ["dimensions", "normalize"] step_pooling_parameters = ["step_tag_id", "returned_token_ids"] @dataclass() class MockModelConfig: pooler_config: PoolerConfig def test_task(): pooling_params = PoolingParams() pooling_params.verify(task="score") pooling_params = PoolingParams(task="score") pooling_params.verify(task="score") with pytest.raises(ValueError): pooling_params.verify(task="classify") def test_embed(): task = "embed" model_config = MockModelConfig(pooler_config=PoolerConfig(pooling_type="CLS")) pooling_params = PoolingParams(normalize=None) pooling_params.verify(task=task, model_config=model_config) pooling_params = PoolingParams(normalize=True) pooling_params.verify(task=task, model_config=model_config) pooling_params = PoolingParams(normalize=False) pooling_params.verify(task=task, model_config=model_config) invalid_parameters = classify_parameters + step_pooling_parameters for p in invalid_parameters: with pytest.raises(ValueError): pooling_params = PoolingParams(**{p: True}) pooling_params.verify(task=task, model_config=model_config) @pytest.mark.parametrize("model_info", EMBEDDING_MODELS) def test_embed_dimensions(model_info: EmbedModelInfo): task = "embed" model_config = ModelConfig( model_info.name, task="auto", tokenizer=model_info.name, tokenizer_mode="auto", trust_remote_code=False, seed=0, dtype="float16", ) pooling_params = PoolingParams(dimensions=None) pooling_params.verify(task=task, model_config=model_config) with pytest.raises(ValueError): pooling_params = PoolingParams(dimensions=1) pooling_params.verify(task=task, model_config=model_config) if model_info.is_matryoshka: assert model_info.matryoshka_dimensions is not None pooling_params = PoolingParams(dimensions=model_info.matryoshka_dimensions[0]) pooling_params.verify(task=task, model_config=model_config) @pytest.mark.parametrize("task", ["score", "classify"]) def test_classify(task): model_config = MockModelConfig(pooler_config=PoolerConfig(pooling_type="CLS")) pooling_params = PoolingParams(use_activation=None) pooling_params.verify(task=task, model_config=model_config) pooling_params = PoolingParams(use_activation=True) pooling_params.verify(task=task, model_config=model_config) pooling_params = PoolingParams(use_activation=False) pooling_params.verify(task=task, model_config=model_config) invalid_parameters = embed_parameters + step_pooling_parameters for p in invalid_parameters: with pytest.raises(ValueError): pooling_params = PoolingParams(**{p: True}) pooling_params.verify(task=task, model_config=model_config) @pytest.mark.parametrize("pooling_type", ["ALL", "STEP"]) def test_token_embed(pooling_type: str): task = "token_embed" model_config = MockModelConfig( pooler_config=PoolerConfig(pooling_type=pooling_type) ) pooling_params = PoolingParams(normalize=None) pooling_params.verify(task=task, model_config=model_config) pooling_params = PoolingParams(normalize=True) pooling_params.verify(task=task, model_config=model_config) pooling_params = PoolingParams(normalize=False) pooling_params.verify(task=task, model_config=model_config) invalid_parameters = classify_parameters if pooling_type != "STEP": invalid_parameters = classify_parameters + step_pooling_parameters for p in invalid_parameters: with pytest.raises(ValueError): pooling_params = PoolingParams(**{p: True}) pooling_params.verify(task=task, model_config=model_config) @pytest.mark.parametrize("pooling_type", ["ALL", "STEP"]) def test_token_classify(pooling_type: str): task = "token_classify" model_config = MockModelConfig( pooler_config=PoolerConfig(pooling_type=pooling_type) ) pooling_params = PoolingParams(use_activation=None) pooling_params.verify(task=task, model_config=model_config) pooling_params = PoolingParams(use_activation=True) pooling_params.verify(task=task, model_config=model_config) pooling_params = PoolingParams(use_activation=False) pooling_params.verify(task=task, model_config=model_config) invalid_parameters = embed_parameters if pooling_type != "STEP": invalid_parameters = embed_parameters + step_pooling_parameters for p in invalid_parameters: with pytest.raises(ValueError): pooling_params = PoolingParams(**{p: True}) pooling_params.verify(task=task, model_config=model_config)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_envs.py
tests/test_envs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os from unittest.mock import patch import pytest import vllm.envs as envs from vllm.envs import ( disable_envs_cache, enable_envs_cache, env_list_with_choices, env_set_with_choices, env_with_choices, environment_variables, ) def test_getattr_without_cache(monkeypatch: pytest.MonkeyPatch): assert envs.VLLM_HOST_IP == "" assert envs.VLLM_PORT is None monkeypatch.setenv("VLLM_HOST_IP", "1.1.1.1") monkeypatch.setenv("VLLM_PORT", "1234") assert envs.VLLM_HOST_IP == "1.1.1.1" assert envs.VLLM_PORT == 1234 # __getattr__ is not decorated with functools.cache assert not hasattr(envs.__getattr__, "cache_info") def test_getattr_with_cache(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_HOST_IP", "1.1.1.1") monkeypatch.setenv("VLLM_PORT", "1234") # __getattr__ is not decorated with functools.cache assert not hasattr(envs.__getattr__, "cache_info") # Enable envs cache and ignore ongoing environment changes enable_envs_cache() # __getattr__ is decorated with functools.cache assert hasattr(envs.__getattr__, "cache_info") start_hits = envs.__getattr__.cache_info().hits # 2 more hits due to VLLM_HOST_IP and VLLM_PORT accesses assert envs.VLLM_HOST_IP == "1.1.1.1" assert envs.VLLM_PORT == 1234 assert envs.__getattr__.cache_info().hits == start_hits + 2 # All environment variables are cached for environment_variable in environment_variables: envs.__getattr__(environment_variable) assert envs.__getattr__.cache_info().hits == start_hits + 2 + len( environment_variables ) # Reset envs.__getattr__ back to none-cached version to # avoid affecting other tests envs.__getattr__ = envs.__getattr__.__wrapped__ def test_getattr_with_reset(monkeypatch: pytest.MonkeyPatch) -> None: monkeypatch.setenv("VLLM_HOST_IP", "1.1.1.1") # __getattr__ is not decorated with functools.cache assert not hasattr(envs.__getattr__, "cache_info") # Enable envs cache and ignore ongoing environment changes enable_envs_cache() assert envs.VLLM_HOST_IP == "1.1.1.1" # With cache enabled, the environment variable value is cached and unchanged monkeypatch.setenv("VLLM_HOST_IP", "2.2.2.2") assert envs.VLLM_HOST_IP == "1.1.1.1" disable_envs_cache() assert envs.VLLM_HOST_IP == "2.2.2.2" # After cache disabled, the environment variable value would be synced # with os.environ monkeypatch.setenv("VLLM_HOST_IP", "3.3.3.3") assert envs.VLLM_HOST_IP == "3.3.3.3" def test_is_envs_cache_enabled() -> None: assert not envs._is_envs_cache_enabled() enable_envs_cache() assert envs._is_envs_cache_enabled() # Only wrap one-layer of cache, so we only need to # call disable once to reset. enable_envs_cache() enable_envs_cache() enable_envs_cache() disable_envs_cache() assert not envs._is_envs_cache_enabled() disable_envs_cache() assert not envs._is_envs_cache_enabled() class TestEnvWithChoices: """Test cases for env_with_choices function.""" def test_default_value_returned_when_env_not_set(self): """Test default is returned when env var is not set.""" env_func = env_with_choices( "NONEXISTENT_ENV", "default", ["option1", "option2"] ) assert env_func() == "default" def test_none_default_returned_when_env_not_set(self): """Test that None is returned when env not set and default is None.""" env_func = env_with_choices("NONEXISTENT_ENV", None, ["option1", "option2"]) assert env_func() is None def test_valid_value_returned_case_sensitive(self): """Test that valid value is returned in case sensitive mode.""" with patch.dict(os.environ, {"TEST_ENV": "option1"}): env_func = env_with_choices( "TEST_ENV", "default", ["option1", "option2"], case_sensitive=True ) assert env_func() == "option1" def test_valid_lowercase_value_returned_case_insensitive(self): """Test that lowercase value is accepted in case insensitive mode.""" with patch.dict(os.environ, {"TEST_ENV": "option1"}): env_func = env_with_choices( "TEST_ENV", "default", ["OPTION1", "OPTION2"], case_sensitive=False ) assert env_func() == "option1" def test_valid_uppercase_value_returned_case_insensitive(self): """Test that uppercase value is accepted in case insensitive mode.""" with patch.dict(os.environ, {"TEST_ENV": "OPTION1"}): env_func = env_with_choices( "TEST_ENV", "default", ["option1", "option2"], case_sensitive=False ) assert env_func() == "OPTION1" def test_invalid_value_raises_error_case_sensitive(self): """Test that invalid value raises ValueError in case sensitive mode.""" with patch.dict(os.environ, {"TEST_ENV": "invalid"}): env_func = env_with_choices( "TEST_ENV", "default", ["option1", "option2"], case_sensitive=True ) with pytest.raises( ValueError, match="Invalid value 'invalid' for TEST_ENV" ): env_func() def test_case_mismatch_raises_error_case_sensitive(self): """Test that case mismatch raises ValueError in case sensitive mode.""" with patch.dict(os.environ, {"TEST_ENV": "OPTION1"}): env_func = env_with_choices( "TEST_ENV", "default", ["option1", "option2"], case_sensitive=True ) with pytest.raises( ValueError, match="Invalid value 'OPTION1' for TEST_ENV" ): env_func() def test_invalid_value_raises_error_case_insensitive(self): """Test that invalid value raises ValueError when case insensitive.""" with patch.dict(os.environ, {"TEST_ENV": "invalid"}): env_func = env_with_choices( "TEST_ENV", "default", ["option1", "option2"], case_sensitive=False ) with pytest.raises( ValueError, match="Invalid value 'invalid' for TEST_ENV" ): env_func() def test_callable_choices_resolved_correctly(self): """Test that callable choices are resolved correctly.""" def get_choices(): return ["dynamic1", "dynamic2"] with patch.dict(os.environ, {"TEST_ENV": "dynamic1"}): env_func = env_with_choices("TEST_ENV", "default", get_choices) assert env_func() == "dynamic1" def test_callable_choices_with_invalid_value(self): """Test that callable choices raise error for invalid values.""" def get_choices(): return ["dynamic1", "dynamic2"] with patch.dict(os.environ, {"TEST_ENV": "invalid"}): env_func = env_with_choices("TEST_ENV", "default", get_choices) with pytest.raises( ValueError, match="Invalid value 'invalid' for TEST_ENV" ): env_func() class TestEnvListWithChoices: """Test cases for env_list_with_choices function.""" def test_default_list_returned_when_env_not_set(self): """Test that default list is returned when env var is not set.""" env_func = env_list_with_choices( "NONEXISTENT_ENV", ["default1", "default2"], ["option1", "option2"] ) assert env_func() == ["default1", "default2"] def test_empty_default_list_returned_when_env_not_set(self): """Test that empty default list is returned when env not set.""" env_func = env_list_with_choices("NONEXISTENT_ENV", [], ["option1", "option2"]) assert env_func() == [] def test_single_valid_value_parsed_correctly(self): """Test that single valid value is parsed correctly.""" with patch.dict(os.environ, {"TEST_ENV": "option1"}): env_func = env_list_with_choices("TEST_ENV", [], ["option1", "option2"]) assert env_func() == ["option1"] def test_multiple_valid_values_parsed_correctly(self): """Test that multiple valid values are parsed correctly.""" with patch.dict(os.environ, {"TEST_ENV": "option1,option2"}): env_func = env_list_with_choices("TEST_ENV", [], ["option1", "option2"]) assert env_func() == ["option1", "option2"] def test_values_with_whitespace_trimmed(self): """Test that values with whitespace are trimmed correctly.""" with patch.dict(os.environ, {"TEST_ENV": " option1 , option2 "}): env_func = env_list_with_choices("TEST_ENV", [], ["option1", "option2"]) assert env_func() == ["option1", "option2"] def test_empty_values_filtered_out(self): """Test that empty values are filtered out.""" with patch.dict(os.environ, {"TEST_ENV": "option1,,option2,"}): env_func = env_list_with_choices("TEST_ENV", [], ["option1", "option2"]) assert env_func() == ["option1", "option2"] def test_empty_string_returns_default(self): """Test that empty string returns default.""" with patch.dict(os.environ, {"TEST_ENV": ""}): env_func = env_list_with_choices( "TEST_ENV", ["default"], ["option1", "option2"] ) assert env_func() == ["default"] def test_only_commas_returns_default(self): """Test that string with only commas returns default.""" with patch.dict(os.environ, {"TEST_ENV": ",,,"}): env_func = env_list_with_choices( "TEST_ENV", ["default"], ["option1", "option2"] ) assert env_func() == ["default"] def test_case_sensitive_validation(self): """Test case sensitive validation.""" with patch.dict(os.environ, {"TEST_ENV": "option1,OPTION2"}): env_func = env_list_with_choices( "TEST_ENV", [], ["option1", "option2"], case_sensitive=True ) with pytest.raises(ValueError, match="Invalid value 'OPTION2' in TEST_ENV"): env_func() def test_case_insensitive_validation(self): """Test case insensitive validation.""" with patch.dict(os.environ, {"TEST_ENV": "OPTION1,option2"}): env_func = env_list_with_choices( "TEST_ENV", [], ["option1", "option2"], case_sensitive=False ) assert env_func() == ["OPTION1", "option2"] def test_invalid_value_in_list_raises_error(self): """Test that invalid value in list raises ValueError.""" with patch.dict(os.environ, {"TEST_ENV": "option1,invalid,option2"}): env_func = env_list_with_choices("TEST_ENV", [], ["option1", "option2"]) with pytest.raises(ValueError, match="Invalid value 'invalid' in TEST_ENV"): env_func() def test_callable_choices_resolved_correctly(self): """Test that callable choices are resolved correctly.""" def get_choices(): return ["dynamic1", "dynamic2"] with patch.dict(os.environ, {"TEST_ENV": "dynamic1,dynamic2"}): env_func = env_list_with_choices("TEST_ENV", [], get_choices) assert env_func() == ["dynamic1", "dynamic2"] def test_callable_choices_with_invalid_value(self): """Test that callable choices raise error for invalid values.""" def get_choices(): return ["dynamic1", "dynamic2"] with patch.dict(os.environ, {"TEST_ENV": "dynamic1,invalid"}): env_func = env_list_with_choices("TEST_ENV", [], get_choices) with pytest.raises(ValueError, match="Invalid value 'invalid' in TEST_ENV"): env_func() def test_duplicate_values_preserved(self): """Test that duplicate values in the list are preserved.""" with patch.dict(os.environ, {"TEST_ENV": "option1,option1,option2"}): env_func = env_list_with_choices("TEST_ENV", [], ["option1", "option2"]) assert env_func() == ["option1", "option1", "option2"] class TestEnvSetWithChoices: """Test cases for env_set_with_choices function.""" def test_default_list_returned_when_env_not_set(self): """Test that default list is returned when env var is not set.""" env_func = env_set_with_choices( "NONEXISTENT_ENV", ["default1", "default2"], ["option1", "option2"] ) assert env_func() == {"default1", "default2"} def test_empty_default_list_returned_when_env_not_set(self): """Test that empty default list is returned when env not set.""" env_func = env_set_with_choices("NONEXISTENT_ENV", [], ["option1", "option2"]) assert env_func() == set() def test_single_valid_value_parsed_correctly(self): """Test that single valid value is parsed correctly.""" with patch.dict(os.environ, {"TEST_ENV": "option1"}): env_func = env_set_with_choices("TEST_ENV", [], ["option1", "option2"]) assert env_func() == {"option1"} def test_multiple_valid_values_parsed_correctly(self): """Test that multiple valid values are parsed correctly.""" with patch.dict(os.environ, {"TEST_ENV": "option1,option2"}): env_func = env_set_with_choices("TEST_ENV", [], ["option1", "option2"]) assert env_func() == {"option1", "option2"} def test_values_with_whitespace_trimmed(self): """Test that values with whitespace are trimmed correctly.""" with patch.dict(os.environ, {"TEST_ENV": " option1 , option2 "}): env_func = env_set_with_choices("TEST_ENV", [], ["option1", "option2"]) assert env_func() == {"option1", "option2"} def test_empty_values_filtered_out(self): """Test that empty values are filtered out.""" with patch.dict(os.environ, {"TEST_ENV": "option1,,option2,"}): env_func = env_set_with_choices("TEST_ENV", [], ["option1", "option2"]) assert env_func() == {"option1", "option2"} def test_empty_string_returns_default(self): """Test that empty string returns default.""" with patch.dict(os.environ, {"TEST_ENV": ""}): env_func = env_set_with_choices( "TEST_ENV", ["default"], ["option1", "option2"] ) assert env_func() == {"default"} def test_only_commas_returns_default(self): """Test that string with only commas returns default.""" with patch.dict(os.environ, {"TEST_ENV": ",,,"}): env_func = env_set_with_choices( "TEST_ENV", ["default"], ["option1", "option2"] ) assert env_func() == {"default"} def test_case_sensitive_validation(self): """Test case sensitive validation.""" with patch.dict(os.environ, {"TEST_ENV": "option1,OPTION2"}): env_func = env_set_with_choices( "TEST_ENV", [], ["option1", "option2"], case_sensitive=True ) with pytest.raises(ValueError, match="Invalid value 'OPTION2' in TEST_ENV"): env_func() def test_case_insensitive_validation(self): """Test case insensitive validation.""" with patch.dict(os.environ, {"TEST_ENV": "OPTION1,option2"}): env_func = env_set_with_choices( "TEST_ENV", [], ["option1", "option2"], case_sensitive=False ) assert env_func() == {"OPTION1", "option2"} def test_invalid_value_in_list_raises_error(self): """Test that invalid value in list raises ValueError.""" with patch.dict(os.environ, {"TEST_ENV": "option1,invalid,option2"}): env_func = env_set_with_choices("TEST_ENV", [], ["option1", "option2"]) with pytest.raises(ValueError, match="Invalid value 'invalid' in TEST_ENV"): env_func() def test_callable_choices_resolved_correctly(self): """Test that callable choices are resolved correctly.""" def get_choices(): return ["dynamic1", "dynamic2"] with patch.dict(os.environ, {"TEST_ENV": "dynamic1,dynamic2"}): env_func = env_set_with_choices("TEST_ENV", [], get_choices) assert env_func() == {"dynamic1", "dynamic2"} def test_callable_choices_with_invalid_value(self): """Test that callable choices raise error for invalid values.""" def get_choices(): return ["dynamic1", "dynamic2"] with patch.dict(os.environ, {"TEST_ENV": "dynamic1,invalid"}): env_func = env_set_with_choices("TEST_ENV", [], get_choices) with pytest.raises(ValueError, match="Invalid value 'invalid' in TEST_ENV"): env_func() def test_duplicate_values_deduped(self): """Test that duplicate values in the list are deduped.""" with patch.dict(os.environ, {"TEST_ENV": "option1,option1,option2"}): env_func = env_set_with_choices("TEST_ENV", [], ["option1", "option2"]) assert env_func() == {"option1", "option2"} class TestVllmConfigureLogging: """Test cases for VLLM_CONFIGURE_LOGGING environment variable.""" def test_configure_logging_defaults_to_true(self): """Test that VLLM_CONFIGURE_LOGGING defaults to True when not set.""" # Ensure the env var is not set with patch.dict(os.environ, {}, clear=False): if "VLLM_CONFIGURE_LOGGING" in os.environ: del os.environ["VLLM_CONFIGURE_LOGGING"] # Clear cache if it exists if hasattr(envs.__getattr__, "cache_clear"): envs.__getattr__.cache_clear() result = envs.VLLM_CONFIGURE_LOGGING assert result is True assert isinstance(result, bool) def test_configure_logging_with_zero_string(self): """Test that VLLM_CONFIGURE_LOGGING='0' evaluates to False.""" with patch.dict(os.environ, {"VLLM_CONFIGURE_LOGGING": "0"}): # Clear cache if it exists if hasattr(envs.__getattr__, "cache_clear"): envs.__getattr__.cache_clear() result = envs.VLLM_CONFIGURE_LOGGING assert result is False assert isinstance(result, bool) def test_configure_logging_with_one_string(self): """Test that VLLM_CONFIGURE_LOGGING='1' evaluates to True.""" with patch.dict(os.environ, {"VLLM_CONFIGURE_LOGGING": "1"}): # Clear cache if it exists if hasattr(envs.__getattr__, "cache_clear"): envs.__getattr__.cache_clear() result = envs.VLLM_CONFIGURE_LOGGING assert result is True assert isinstance(result, bool) def test_configure_logging_with_invalid_value_raises_error(self): """Test that invalid VLLM_CONFIGURE_LOGGING value raises ValueError.""" with patch.dict(os.environ, {"VLLM_CONFIGURE_LOGGING": "invalid"}): # Clear cache if it exists if hasattr(envs.__getattr__, "cache_clear"): envs.__getattr__.cache_clear() with pytest.raises(ValueError, match="invalid literal for int"): _ = envs.VLLM_CONFIGURE_LOGGING
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_triton_utils.py
tests/test_triton_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import sys import types from unittest import mock from vllm.triton_utils.importing import TritonLanguagePlaceholder, TritonPlaceholder def test_triton_placeholder_is_module(): triton = TritonPlaceholder() assert isinstance(triton, types.ModuleType) assert triton.__name__ == "triton" def test_triton_language_placeholder_is_module(): triton_language = TritonLanguagePlaceholder() assert isinstance(triton_language, types.ModuleType) assert triton_language.__name__ == "triton.language" def test_triton_placeholder_decorators(): triton = TritonPlaceholder() @triton.jit def foo(x): return x @triton.autotune def bar(x): return x @triton.heuristics def baz(x): return x assert foo(1) == 1 assert bar(2) == 2 assert baz(3) == 3 def test_triton_placeholder_decorators_with_args(): triton = TritonPlaceholder() @triton.jit(debug=True) def foo(x): return x @triton.autotune(configs=[], key="x") def bar(x): return x @triton.heuristics({"BLOCK_SIZE": lambda args: 128 if args["x"] > 1024 else 64}) def baz(x): return x assert foo(1) == 1 assert bar(2) == 2 assert baz(3) == 3 def test_triton_placeholder_language(): lang = TritonLanguagePlaceholder() assert isinstance(lang, types.ModuleType) assert lang.__name__ == "triton.language" assert lang.constexpr is None assert lang.dtype is None assert lang.int64 is None assert lang.int32 is None assert lang.tensor is None def test_triton_placeholder_language_from_parent(): triton = TritonPlaceholder() lang = triton.language assert isinstance(lang, TritonLanguagePlaceholder) def test_no_triton_fallback(): # clear existing triton modules sys.modules.pop("triton", None) sys.modules.pop("triton.language", None) sys.modules.pop("vllm.triton_utils", None) sys.modules.pop("vllm.triton_utils.importing", None) # mock triton not being installed with mock.patch.dict(sys.modules, {"triton": None}): from vllm.triton_utils import HAS_TRITON, tl, triton assert HAS_TRITON is False assert triton.__class__.__name__ == "TritonPlaceholder" assert triton.language.__class__.__name__ == "TritonLanguagePlaceholder" assert tl.__class__.__name__ == "TritonLanguagePlaceholder"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_embedded_commit.py
tests/test_embedded_commit.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import vllm def test_embedded_commit_defined(): assert hasattr(vllm, "__version__") assert hasattr(vllm, "__version_tuple__") assert vllm.__version__ != "dev" assert vllm.__version_tuple__ != (0, 0, "dev")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_inputs.py
tests/test_inputs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from vllm.config import ModelConfig from vllm.inputs import zip_enc_dec_prompts from vllm.inputs.parse import parse_raw_prompts from vllm.inputs.preprocess import InputPreprocessor from vllm.tokenizers import cached_tokenizer_from_config pytestmark = pytest.mark.cpu_test STRING_INPUTS = [ "", "foo", "foo bar", "foo baz bar", "foo bar qux baz", ] TOKEN_INPUTS = [ [-1], [1], [1, 2], [1, 3, 4], [1, 2, 4, 3], ] INPUTS_SLICES = [ slice(None, None, -1), slice(None, None, 2), slice(None, None, -2), ] # Test that a nested mixed-type list of lists raises a TypeError. @pytest.mark.parametrize("invalid_input", [[[1, 2], ["foo", "bar"]]]) def test_invalid_input_raise_type_error(invalid_input): with pytest.raises(TypeError): parse_raw_prompts(invalid_input) def test_parse_raw_single_batch_empty(): with pytest.raises(ValueError, match="at least one prompt"): parse_raw_prompts([]) with pytest.raises(ValueError, match="at least one prompt"): parse_raw_prompts([[]]) @pytest.mark.parametrize("string_input", STRING_INPUTS) def test_parse_raw_single_batch_string_consistent(string_input: str): assert parse_raw_prompts(string_input) == parse_raw_prompts([string_input]) @pytest.mark.parametrize("token_input", TOKEN_INPUTS) def test_parse_raw_single_batch_token_consistent(token_input: list[int]): assert parse_raw_prompts(token_input) == parse_raw_prompts([token_input]) @pytest.mark.parametrize("inputs_slice", INPUTS_SLICES) def test_parse_raw_single_batch_string_slice(inputs_slice: slice): assert parse_raw_prompts(STRING_INPUTS)[inputs_slice] == parse_raw_prompts( STRING_INPUTS[inputs_slice] ) @pytest.mark.parametrize( "mm_processor_kwargs,expected_mm_kwargs", [ (None, [{}, {}]), ({}, [{}, {}]), ({"foo": 100}, [{"foo": 100}, {"foo": 100}]), ([{"foo": 100}, {"bar": 200}], [{"foo": 100}, {"bar": 200}]), ], ) def test_zip_enc_dec_prompts(mm_processor_kwargs, expected_mm_kwargs): """Test mm_processor_kwargs init for zipping enc/dec prompts.""" encoder_prompts = ["An encoder prompt", "Another encoder prompt"] decoder_prompts = ["A decoder prompt", "Another decoder prompt"] zipped_prompts = zip_enc_dec_prompts( encoder_prompts, decoder_prompts, mm_processor_kwargs ) assert len(zipped_prompts) == len(encoder_prompts) == len(decoder_prompts) for enc, dec, exp_kwargs, zipped in zip( encoder_prompts, decoder_prompts, expected_mm_kwargs, zipped_prompts ): assert isinstance(zipped, dict) assert len(zipped.keys()) == 3 assert zipped["encoder_prompt"] == enc assert zipped["decoder_prompt"] == dec assert zipped["mm_processor_kwargs"] == exp_kwargs @pytest.mark.parametrize( "model_id", [ "facebook/chameleon-7b", ], ) @pytest.mark.parametrize( "prompt", [ "", {"prompt_token_ids": []}, ], ) @pytest.mark.skip( reason=( "Applying huggingface processor on text inputs results in " "significant performance regression for multimodal models. " "See https://github.com/vllm-project/vllm/issues/26320" ) ) def test_preprocessor_always_mm_code_path(model_id, prompt): model_config = ModelConfig(model=model_id) tokenizer = cached_tokenizer_from_config(model_config) input_preprocessor = InputPreprocessor(model_config, tokenizer) # HF processor adds sep token sep_token_id = tokenizer.vocab[tokenizer.sep_token] processed_inputs = input_preprocessor.preprocess(prompt) assert sep_token_id in processed_inputs["prompt_token_ids"]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/conftest.py
tests/conftest.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import contextlib import pathlib from copy import deepcopy from tblib import pickling_support # Import fixture from tests.v1.entrypoints.conftest import sample_json_schema # noqa # ruff: noqa # Install support for pickling exceptions so that we can nicely propagate # failures from tests running in a subprocess. # This should be run before any custom exception subclasses are defined. pickling_support.install() import http.server import json import math import mimetypes import os import socket import tempfile import threading from collections.abc import Generator from contextlib import nullcontext from enum import Enum from typing import Any, Callable, TypedDict, TypeVar, cast, TYPE_CHECKING import numpy as np import pytest import torch import torch.nn as nn import torch.nn.functional as F from huggingface_hub import snapshot_download from PIL import Image from transformers import ( AutoConfig, AutoModelForCausalLM, AutoTokenizer, BatchEncoding, BatchFeature, ) from transformers.models.auto.auto_factory import _BaseAutoModelClass from tests.models.utils import TokensTextLogprobs, TokensTextLogprobsPromptLogprobs from vllm import LLM, SamplingParams, envs from vllm.assets.audio import AudioAsset from vllm.assets.image import ImageAsset from vllm.assets.video import VideoAsset from vllm.config.model import ConvertOption, RunnerOption, _get_and_verify_dtype from vllm.connections import global_http_connection from vllm.distributed import ( cleanup_dist_env_and_memory, init_distributed_environment, initialize_model_parallel, ) from vllm.logger import init_logger from vllm.logprobs import Logprob from vllm.multimodal.base import MediaWithBytes from vllm.multimodal.utils import fetch_image from vllm.outputs import RequestOutput from vllm.sampling_params import BeamSearchParams from vllm.transformers_utils.utils import maybe_model_redirect from vllm.utils.collection_utils import is_list_of from vllm.utils.torch_utils import set_default_torch_num_threads from torch._inductor.utils import fresh_cache if TYPE_CHECKING: from transformers import PreTrainedTokenizer, PreTrainedTokenizerFast from transformers.generation.utils import GenerateOutput logger = init_logger(__name__) _TEST_DIR = os.path.dirname(__file__) _TEST_PROMPTS = [os.path.join(_TEST_DIR, "prompts", "example.txt")] _LONG_PROMPTS = [os.path.join(_TEST_DIR, "prompts", "summary.txt")] _SYS_MSG = os.path.join(_TEST_DIR, "system_messages", "sonnet3.5_nov2024.txt") _M = TypeVar("_M") _PromptMultiModalInput = list[_M] | list[list[_M]] PromptImageInput = _PromptMultiModalInput[Image.Image] PromptAudioInput = _PromptMultiModalInput[tuple[np.ndarray, int]] PromptVideoInput = _PromptMultiModalInput[np.ndarray] def _read_prompts(filename: str) -> list[str]: with open(filename) as f: prompts = f.readlines() return prompts class ImageAssetPrompts(TypedDict): stop_sign: str cherry_blossom: str class ImageTestAssets(list[ImageAsset]): def __init__(self) -> None: super().__init__( [ ImageAsset("stop_sign"), ImageAsset("cherry_blossom"), ] ) def prompts(self, prompts: ImageAssetPrompts) -> list[str]: """ Convenience method to define the prompt for each test image. The order of the returned prompts matches the order of the assets when iterating through this object. """ return [prompts["stop_sign"], prompts["cherry_blossom"]] class VideoAssetPrompts(TypedDict): baby_reading: str class VideoTestAssets(list[VideoAsset]): def __init__(self) -> None: super().__init__( [ VideoAsset("baby_reading"), ] ) def prompts(self, prompts: VideoAssetPrompts) -> list[str]: return [prompts["baby_reading"]] class AudioAssetPrompts(TypedDict): mary_had_lamb: str winning_call: str class AudioTestAssets(list[AudioAsset]): def __init__(self) -> None: super().__init__( [ AudioAsset("mary_had_lamb"), AudioAsset("winning_call"), ] ) def prompts(self, prompts: AudioAssetPrompts) -> list[str]: return [prompts["mary_had_lamb"], prompts["winning_call"]] IMAGE_ASSETS = ImageTestAssets() """Singleton instance of {class}`ImageTestAssets`.""" VIDEO_ASSETS = VideoTestAssets() """Singleton instance of {class}`VideoTestAssets`.""" AUDIO_ASSETS = AudioTestAssets() """Singleton instance of {class}`AudioTestAssets`.""" @pytest.fixture(autouse=True) def init_test_http_connection(): # pytest_asyncio may use a different event loop per test # so we need to make sure the async client is created anew global_http_connection.reuse_client = False @pytest.fixture def dist_init(): temp_file = tempfile.mkstemp()[1] init_distributed_environment( world_size=1, rank=0, distributed_init_method=f"file://{temp_file}", local_rank=0, backend="nccl", ) initialize_model_parallel(1, 1) yield cleanup_dist_env_and_memory() @pytest.fixture() def should_do_global_cleanup_after_test(request) -> bool: """Allow subdirectories to skip global cleanup by overriding this fixture. This can provide a ~10x speedup for non-GPU unit tests since they don't need to initialize torch. """ return not request.node.get_closest_marker("skip_global_cleanup") @pytest.fixture(autouse=True) def cleanup_fixture(should_do_global_cleanup_after_test: bool): yield if should_do_global_cleanup_after_test: cleanup_dist_env_and_memory() @pytest.fixture def workspace_init(): """Initialize the workspace manager for tests that need it. This fixture initializes the workspace manager with a CUDA device if available, and resets it after the test completes. Tests that create a full vLLM engine should NOT use this fixture as the engine will initialize the workspace manager itself. """ from vllm.v1.worker.workspace import ( init_workspace_manager, reset_workspace_manager, ) if torch.cuda.is_available(): device = torch.device("cuda:0") init_workspace_manager(device) yield reset_workspace_manager() @pytest.fixture(autouse=True) def dynamo_reset(): yield torch._dynamo.reset() @pytest.fixture def example_prompts() -> list[str]: return [prompt for filename in _TEST_PROMPTS for prompt in _read_prompts(filename)] @pytest.fixture def example_system_message() -> str: with open(_SYS_MSG) as f: return f.read() class DecoderPromptType(Enum): """For encoder/decoder models only.""" CUSTOM = 1 NONE = 2 EMPTY_STR = 3 @pytest.fixture def example_long_prompts() -> list[str]: return [prompt for filename in _LONG_PROMPTS for prompt in _read_prompts(filename)] @pytest.fixture(scope="session") def image_assets() -> ImageTestAssets: return IMAGE_ASSETS @pytest.fixture(scope="session") def video_assets() -> VideoTestAssets: return VIDEO_ASSETS @pytest.fixture(scope="session") def audio_assets() -> AudioTestAssets: return AUDIO_ASSETS _T = TypeVar("_T", nn.Module, torch.Tensor, BatchEncoding, BatchFeature, dict) _R = TypeVar("_R") class HfRunner: def get_default_device(self): from vllm.platforms import current_platform return "cpu" if current_platform.is_cpu() else current_platform.device_type def wrap_device(self, x: _T, device: str | None = None) -> _T: if x is None or isinstance(x, (bool,)): return x if device is None: device = self.device if isinstance(x, dict): return {k: self.wrap_device(v, device) for k, v in x.items()} if hasattr(x, "device") and x.device.type == device: return x return x.to(device) def __init__( self, model_name: str, dtype: str = "auto", *, model_kwargs: dict[str, Any] | None = None, trust_remote_code: bool = True, is_sentence_transformer: bool = False, is_cross_encoder: bool = False, skip_tokenizer_init: bool = False, auto_cls: type[_BaseAutoModelClass] = AutoModelForCausalLM, # Set this to avoid hanging issue default_torch_num_threads: int | None = None, ) -> None: init_ctx = ( nullcontext() if default_torch_num_threads is None else set_default_torch_num_threads(default_torch_num_threads) ) with init_ctx: self._init( model_name=model_name, dtype=dtype, model_kwargs=model_kwargs, trust_remote_code=trust_remote_code, is_sentence_transformer=is_sentence_transformer, is_cross_encoder=is_cross_encoder, skip_tokenizer_init=skip_tokenizer_init, auto_cls=auto_cls, ) def _init( self, model_name: str, dtype: str = "auto", *, model_kwargs: dict[str, Any] | None = None, trust_remote_code: bool = True, is_sentence_transformer: bool = False, is_cross_encoder: bool = False, skip_tokenizer_init: bool = False, auto_cls: type[_BaseAutoModelClass] = AutoModelForCausalLM, ) -> None: model_name = maybe_model_redirect(model_name) self.model_name = model_name self.config = AutoConfig.from_pretrained( model_name, trust_remote_code=trust_remote_code, ) self.device = self.get_default_device() self.dtype = dtype = _get_and_verify_dtype( self.model_name, self.config, dtype=dtype, is_pooling_model=is_sentence_transformer or is_cross_encoder, ) model_kwargs = model_kwargs if model_kwargs is not None else {} model_kwargs.setdefault("dtype", dtype) if is_sentence_transformer: # Lazy init required for AMD CI from sentence_transformers import SentenceTransformer self.model = SentenceTransformer( model_name, device=self.device, model_kwargs=model_kwargs, trust_remote_code=trust_remote_code, ) elif is_cross_encoder: # Lazy init required for AMD CI from sentence_transformers import CrossEncoder self.model = CrossEncoder( model_name, device=self.device, automodel_args=model_kwargs, trust_remote_code=trust_remote_code, ) else: model = cast( nn.Module, auto_cls.from_pretrained( model_name, trust_remote_code=trust_remote_code, **model_kwargs, ), ) # in case some unquantized custom models are not in same dtype if getattr(model, "quantization_method", None) is None and any( p.dtype != self.dtype for p in model.parameters() ): model = model.to(dtype=self.dtype) if ( getattr(model, "quantization_method", None) != "bitsandbytes" and len({p.device for p in model.parameters()}) < 2 ): model = model.to(device=self.device) self.model = model if not skip_tokenizer_init: self.tokenizer: "PreTrainedTokenizer | PreTrainedTokenizerFast" = ( AutoTokenizer.from_pretrained( model_name, dtype=dtype, trust_remote_code=trust_remote_code, ) ) # don't put this import at the top level # it will call torch.cuda.device_count() from transformers import AutoProcessor self.processor = AutoProcessor.from_pretrained( model_name, dtype=dtype, trust_remote_code=trust_remote_code, ) if skip_tokenizer_init: self.tokenizer = self.processor.tokenizer def get_inputs( self, prompts: list[str] | list[list[int]], images: PromptImageInput | None = None, videos: PromptVideoInput | None = None, audios: PromptAudioInput | None = None, tokenization_kwargs: dict[str, Any] | None = None, ) -> list[BatchFeature | BatchEncoding | dict[str, torch.Tensor]]: if images is not None: assert len(prompts) == len(images) if videos is not None: assert len(prompts) == len(videos) if audios is not None: assert len(prompts) == len(audios) all_inputs: list[BatchFeature | BatchEncoding | dict[str, torch.Tensor]] = [] for i, prompt in enumerate(prompts): if isinstance(prompt, str): # Create a copy to avoid modifying the original dict processor_kwargs = ( tokenization_kwargs.copy() if tokenization_kwargs is not None else {} ) processor_kwargs.update( { "text": prompt, "return_tensors": "pt", } ) if images is not None and (image := images[i]) is not None: processor_kwargs["images"] = image if videos is not None and (video := videos[i]) is not None: processor_kwargs["videos"] = video if audios is not None and (audio_inputs := audios[i]) is not None: # HACK - not all processors take sampling_rate; we should # clean this up in the future. if len(audio_inputs) == 2: audio, sr = audio_inputs processor_kwargs["audio"] = audio processor_kwargs["sampling_rate"] = sr else: processor_kwargs["audio"] = audio_inputs inputs = self.processor(**processor_kwargs) if isinstance(inputs, BatchFeature): inputs = inputs.to(dtype=self.dtype) all_inputs.append(inputs) else: # check that prompt is (batched) list of integers (token ids) if not is_list_of(prompt, typ=int, check="all"): raise ValueError( "Prompt must be a list of ints corresponding to the prompt token ids." ) # check that no multimodal input is provided if images or videos or audios: raise ValueError( "When providing prompt token ids multimodal inputs are not supported." ) input_dict = { "input_ids": torch.tensor(prompt, dtype=torch.long).unsqueeze(0), } all_inputs.append(input_dict) return all_inputs def get_prompt_embeddings(self, prompts: list[str]) -> list[torch.Tensor]: all_inputs = self.get_inputs(prompts) embeddings = [] for inputs in all_inputs: input_ids = self.wrap_device(inputs)["input_ids"] embedding = self.model.get_input_embeddings()(input_ids).squeeze(0) embeddings.append(embedding) return embeddings def classify(self, prompts: list[str]) -> list[list[float]]: # output is final logits all_inputs = self.get_inputs(prompts) outputs: list[list[float]] = [] problem_type = getattr(self.config, "problem_type", "") for inputs in all_inputs: output = self.model(**self.wrap_device(inputs)) assert isinstance(output.logits, torch.Tensor) if problem_type == "regression": logits = output.logits[0].tolist() elif problem_type == "multi_label_classification": logits = output.logits.sigmoid()[0].tolist() else: logits = output.logits.softmax(dim=-1)[0].tolist() outputs.append(logits) return outputs def generate( self, prompts: list[str] | list[list[int]], images: PromptImageInput | None = None, videos: PromptVideoInput | None = None, audios: PromptAudioInput | None = None, **kwargs: Any, ) -> list[tuple[list[list[int]], list[str]]]: all_inputs = self.get_inputs( prompts, images=images, videos=videos, audios=audios ) outputs: list[tuple[list[list[int]], list[str]]] = [] for inputs in all_inputs: output_ids: torch.Tensor = self.model.generate( **self.wrap_device(inputs), use_cache=True, **kwargs, ) output_str = self.processor.batch_decode( output_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False, ) outputs.append((output_ids.cpu().tolist(), output_str)) return outputs def generate_greedy( self, prompts: list[str] | list[list[int]], max_tokens: int, images: PromptImageInput | None = None, videos: PromptVideoInput | None = None, audios: PromptAudioInput | None = None, **kwargs: Any, ) -> list[tuple[list[int], str]]: outputs = self.generate( prompts, do_sample=False, max_new_tokens=max_tokens, images=images, videos=videos, audios=audios, **kwargs, ) return [(output_ids[0], output_str[0]) for output_ids, output_str in outputs] def generate_beam_search( self, prompts: list[str], beam_width: int, max_tokens: int, images: PromptImageInput | None = None, videos: PromptVideoInput | None = None, audios: PromptAudioInput | None = None, ) -> list[tuple[list[list[int]], list[str]]]: outputs = self.generate( prompts, do_sample=False, max_new_tokens=max_tokens, num_beams=beam_width, num_return_sequences=beam_width, images=images, videos=videos, audios=audios, ) for i in range(len(outputs)): output_ids, output_str = outputs[i] for j in range(len(output_ids)): output_ids[j] = [ x for x in output_ids[j] if x != self.tokenizer.pad_token_id ] outputs[i] = (output_ids, output_str) return outputs def generate_greedy_logprobs( self, prompts: list[str], max_tokens: int, images: PromptImageInput | None = None, videos: PromptVideoInput | None = None, audios: PromptAudioInput | None = None, **kwargs: Any, ) -> list[list[torch.Tensor]]: all_inputs = self.get_inputs( prompts, images=images, videos=videos, audios=audios ) all_logprobs: list[list[torch.Tensor]] = [] for inputs in all_inputs: output: "GenerateOutput" = self.model.generate( **self.wrap_device(inputs), use_cache=True, do_sample=False, max_new_tokens=max_tokens, output_hidden_states=True, return_dict_in_generate=True, **kwargs, ) seq_logprobs = self._hidden_states_to_seq_logprobs(output.hidden_states) all_logprobs.append(seq_logprobs) return all_logprobs def _hidden_states_to_seq_logprobs( self, hidden_states: tuple[tuple[torch.Tensor, ...], ...], ) -> list[torch.Tensor]: output_embeddings = self.model.get_output_embeddings() seq_logprobs: list[torch.Tensor] = [] for _, hidden_state in enumerate(hidden_states): last_hidden_states = hidden_state[-1][0] logits = torch.matmul( last_hidden_states.to( device=output_embeddings.weight.device, dtype=output_embeddings.weight.dtype, ), output_embeddings.weight.t(), ) if getattr(output_embeddings, "bias", None) is not None: logits += output_embeddings.bias.unsqueeze(0) logprobs = F.log_softmax(logits, dim=-1, dtype=torch.float32) seq_logprobs.append(logprobs) return seq_logprobs def _hidden_states_to_logprobs( self, hidden_states: tuple[tuple[torch.Tensor, ...], ...], num_logprobs: int | None, ) -> tuple[list[dict[int, float]], int]: seq_logprobs = self._hidden_states_to_seq_logprobs(hidden_states) output_len = len(hidden_states) # convert to dict seq_logprobs_lst: list[dict[int, float]] = [] for tok_idx, tok_logprobs in enumerate(seq_logprobs): # drop prompt logprobs if tok_idx == 0: tok_logprobs = tok_logprobs[-1, :].reshape(1, -1) topk = tok_logprobs.topk(num_logprobs) tok_logprobs_dct = {} for token_id, logprob in zip(topk.indices[0], topk.values[0]): tok_logprobs_dct[token_id.item()] = logprob.item() seq_logprobs_lst.append(tok_logprobs_dct) return ( seq_logprobs_lst, output_len, ) def generate_greedy_logprobs_limit( self, prompts: list[str], max_tokens: int, num_logprobs: int | None, images: PromptImageInput | None = None, audios: PromptAudioInput | None = None, videos: PromptVideoInput | None = None, **kwargs: Any, ) -> list[TokensTextLogprobs]: all_inputs = self.get_inputs( prompts, images=images, videos=videos, audios=audios ) all_logprobs: list[list[dict[int, float]]] = [] all_output_ids: list[list[int]] = [] all_output_strs: list[str] = [] for inputs in all_inputs: output: "GenerateOutput" = self.model.generate( **self.wrap_device(inputs), use_cache=True, do_sample=False, max_new_tokens=max_tokens, output_hidden_states=True, return_dict_in_generate=True, **kwargs, ) # Encoder-decoder models return decoder_hidden_states instead of # hidden_states hidden_states = ( getattr(output, "hidden_states", None) or output.decoder_hidden_states ) ( seq_logprobs_lst, output_len, ) = self._hidden_states_to_logprobs(hidden_states, num_logprobs) all_logprobs.append(seq_logprobs_lst) seq_ids = output.sequences[0] output_len = len(seq_logprobs_lst) output_ids = seq_ids[-output_len:] all_output_ids.append(output_ids.tolist()) all_output_strs.append(self.tokenizer.decode(output_ids)) outputs = zip(all_output_ids, all_output_strs, all_logprobs) return [ (output_ids, output_str, output_logprobs) for output_ids, output_str, output_logprobs in outputs ] def encode(self, prompts: list[str], *args, **kwargs) -> list[list[torch.Tensor]]: return self.model.encode(prompts, *args, **kwargs) def predict(self, prompts: list[list[str]], *args, **kwargs) -> torch.Tensor: return self.model.predict(prompts, *args, convert_to_tensor=True, **kwargs) def __enter__(self): return self def __exit__(self, exc_type, exc_value, traceback): del self.model cleanup_dist_env_and_memory() @pytest.fixture(scope="session") def hf_runner(): return HfRunner class VllmRunner: """ The default value of some arguments have been modified from {class}`~vllm.LLM` as follows: - `trust_remote_code`: Set to `True` instead of `False` for convenience. - `seed`: Set to `0` instead of `None` for test reproducibility. - `max_model_len`: Set to `1024` instead of `None` to reduce memory usage. - `block_size`: To reduce memory usage, set default to `64` if on XPU devices, otherwise default to `16`. - `enable_chunked_prefill`: Set to `False` instead of `None` for test reproducibility. - `enforce_eager`: Set to `False` to test CUDA graph. """ def __init__( self, model_name: str, runner: RunnerOption = "auto", convert: ConvertOption = "auto", tokenizer_name: str | None = None, tokenizer_mode: str = "auto", trust_remote_code: bool = True, seed: int = 0, max_model_len: int | None = 1024, dtype: str = "auto", disable_log_stats: bool = True, tensor_parallel_size: int = 1, block_size: int = 16 if not torch.xpu.is_available() else 64, enable_chunked_prefill: bool | None = False, swap_space: int = 4, enforce_eager: bool | None = False, # Set this to avoid hanging issue default_torch_num_threads: int | None = None, **kwargs, ) -> None: init_ctx = ( nullcontext() if default_torch_num_threads is None else set_default_torch_num_threads(default_torch_num_threads) ) if not kwargs.get("compilation_config", None): # Note(@tdoublep): This is set to 4 because some tests (e.g., hybrid # model tests) may set max_num_seqs=4. If min cudagraph_capture_size is # set to larger than max_num_seqs, then it will lead to *no* graphs # being captured which can trigger edge cases that we don't handle yet. kwargs["compilation_config"] = {"cudagraph_capture_sizes": [4]} # Make sure we have atleast one cudagraph large enough for a single decode. if (speculative_config := kwargs.get("speculative_config")) and ( num_speculative_tokens := speculative_config["num_speculative_tokens"] ): kwargs["compilation_config"]["cudagraph_capture_sizes"].append( num_speculative_tokens + 1 ) with init_ctx: self.llm = LLM( model=model_name, runner=runner, convert=convert, tokenizer=tokenizer_name, tokenizer_mode=tokenizer_mode, trust_remote_code=trust_remote_code, dtype=dtype, seed=seed, swap_space=swap_space, enforce_eager=enforce_eager, disable_log_stats=disable_log_stats, tensor_parallel_size=tensor_parallel_size, max_model_len=max_model_len, block_size=block_size, enable_chunked_prefill=enable_chunked_prefill, **kwargs, ) def get_inputs( self, prompts: list[str] | list[torch.Tensor] | list[list[int]], images: PromptImageInput | None = None, videos: PromptVideoInput | None = None, audios: PromptAudioInput | None = None, ) -> list[dict[str, Any]]: if any( x is not None and len(x) != len(prompts) for x in [images, videos, audios] ): raise ValueError( "All non-None multimodal inputs must have the same length as prompts" ) inputs = list[dict[str, Any]]() for i, prompt in enumerate(prompts): prompt_dict = dict[str, Any]() if isinstance(prompt, str): prompt_dict["prompt"] = prompt elif isinstance(prompt, list): prompt_dict["prompt_token_ids"] = prompt else: prompt_dict["prompt_embeds"] = prompt multi_modal_data = dict[str, Any]() if images is not None and (image := images[i]) is not None: multi_modal_data["image"] = image if videos is not None and (video := videos[i]) is not None: multi_modal_data["video"] = video if audios is not None and (audio := audios[i]) is not None: multi_modal_data["audio"] = audio if multi_modal_data: prompt_dict["multi_modal_data"] = multi_modal_data inputs.append(prompt_dict) return inputs def generate( self, prompts: list[str] | list[torch.Tensor] | list[list[int]], sampling_params: SamplingParams, images: PromptImageInput | None = None, videos: PromptVideoInput | None = None, audios: PromptAudioInput | None = None, return_logprobs: bool = False, **kwargs: Any, ) -> list[tuple[list[list[int]], list[str]]] | tuple[list, list]: inputs = self.get_inputs(prompts, images=images, videos=videos, audios=audios) req_outputs = self.llm.generate( inputs, sampling_params=sampling_params, **kwargs ) outputs: list[tuple[list[list[int]], list[str]]] = [] logprobs = [] for req_output in req_outputs: prompt_str = req_output.prompt prompt_ids = req_output.prompt_token_ids req_sample_output_ids: list[list[int]] = [] req_sample_output_strs: list[str] = [] req_logprobs = [] for sample in req_output.outputs: output_str = sample.text output_ids = list(sample.token_ids) req_sample_output_ids.append(prompt_ids + output_ids) req_sample_output_strs.append((prompt_str or "") + output_str) if sample.logprobs: req_logprobs.extend(sample.logprobs) outputs.append((req_sample_output_ids, req_sample_output_strs)) logprobs.append(req_logprobs) return outputs if not return_logprobs else (outputs, logprobs) @staticmethod def _final_steps_generate_w_logprobs( req_outputs: list[RequestOutput], include_prompt_token_ids: bool = False, ) -> list[TokensTextLogprobsPromptLogprobs]: outputs: list[TokensTextLogprobsPromptLogprobs] = [] for req_output in req_outputs: assert len(req_output.outputs) > 0 for sample in req_output.outputs: output_str = sample.text output_ids = list(sample.token_ids) output_logprobs = sample.logprobs if include_prompt_token_ids: outputs.append( ( # type: ignore[arg-type] output_ids, output_str, output_logprobs, req_output.prompt_token_ids, req_output.prompt_logprobs, ) ) else: outputs.append( ( output_ids, output_str, output_logprobs, req_output.prompt_logprobs, ) ) return outputs def generate_w_logprobs( self, prompts: list[str], sampling_params: SamplingParams, images: PromptImageInput | None = None, audios: PromptAudioInput | None = None, videos: PromptVideoInput | None = None, include_prompt_token_ids: bool = False, **kwargs: Any, ) -> list[TokensTextLogprobs] | list[TokensTextLogprobsPromptLogprobs]: inputs = self.get_inputs(prompts, images=images, videos=videos, audios=audios) req_outputs = self.llm.generate( inputs, sampling_params=sampling_params, **kwargs ) toks_str_logsprobs_prompt_logprobs = self._final_steps_generate_w_logprobs( req_outputs, include_prompt_token_ids ) # Omit prompt logprobs if not required by sampling params return ( [x[0:-1] for x in toks_str_logsprobs_prompt_logprobs]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/utils.py
tests/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import asyncio import contextlib import copy import functools import importlib import itertools import json import os import random import signal import subprocess import sys import tempfile import time import warnings from collections.abc import Callable, Iterable from contextlib import ExitStack, contextmanager, suppress from multiprocessing import Process from pathlib import Path from typing import Any, Literal from unittest.mock import patch import anthropic import cloudpickle import httpx import openai import pytest import requests import torch import torch.nn.functional as F from openai.types.completion import Completion from typing_extensions import ParamSpec import vllm.envs as envs from tests.models.utils import TextTextLogprobs from vllm.distributed import ( ensure_model_parallel_initialized, init_distributed_environment, ) from vllm.engine.arg_utils import AsyncEngineArgs from vllm.entrypoints.cli.serve import ServeSubcommand from vllm.model_executor.model_loader import get_model_loader from vllm.platforms import current_platform from vllm.tokenizers import get_tokenizer from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.mem_constants import GB_bytes from vllm.utils.network_utils import get_open_port from vllm.utils.torch_utils import cuda_device_count_stateless if current_platform.is_rocm(): from amdsmi import ( amdsmi_get_gpu_vram_usage, amdsmi_get_processor_handles, amdsmi_init, amdsmi_shut_down, ) @contextmanager def _nvml(): try: amdsmi_init() yield finally: amdsmi_shut_down() elif current_platform.is_cuda(): from vllm.third_party.pynvml import ( nvmlDeviceGetHandleByIndex, nvmlDeviceGetMemoryInfo, nvmlInit, nvmlShutdown, ) @contextmanager def _nvml(): try: nvmlInit() yield finally: nvmlShutdown() else: @contextmanager def _nvml(): yield VLLM_PATH = Path(__file__).parent.parent """Path to root of the vLLM repository.""" class RemoteOpenAIServer: DUMMY_API_KEY = "token-abc123" # vLLM's OpenAI server does not need API key def _start_server( self, model: str, vllm_serve_args: list[str], env_dict: dict[str, str] | None ) -> None: """Subclasses override this method to customize server process launch""" env = os.environ.copy() # the current process might initialize cuda, # to be safe, we should use spawn method env["VLLM_WORKER_MULTIPROC_METHOD"] = "spawn" if env_dict is not None: env.update(env_dict) serve_cmd = ["vllm", "serve", model, *vllm_serve_args] print(f"Launching RemoteOpenAIServer with: {' '.join(serve_cmd)}") print(f"Environment variables: {env}") self.proc: subprocess.Popen = subprocess.Popen( serve_cmd, env=env, stdout=sys.stdout, stderr=sys.stderr, ) def __init__( self, model: str, vllm_serve_args: list[str], *, env_dict: dict[str, str] | None = None, seed: int = 0, auto_port: bool = True, max_wait_seconds: float | None = None, override_hf_configs: dict[str, Any] | None = None, ) -> None: if auto_port: if "-p" in vllm_serve_args or "--port" in vllm_serve_args: raise ValueError( "You have manually specified the port when `auto_port=True`." ) # No need for a port if using unix sockets if "--uds" not in vllm_serve_args: # Don't mutate the input args vllm_serve_args = vllm_serve_args + ["--port", str(get_open_port())] if seed is not None: if "--seed" in vllm_serve_args: raise ValueError( f"You have manually specified the seed when `seed={seed}`." ) vllm_serve_args = vllm_serve_args + ["--seed", str(seed)] if override_hf_configs is not None: vllm_serve_args = vllm_serve_args + [ "--hf-overrides", json.dumps(override_hf_configs), ] parser = FlexibleArgumentParser(description="vLLM's remote OpenAI server.") subparsers = parser.add_subparsers(required=False, dest="subparser") parser = ServeSubcommand().subparser_init(subparsers) args = parser.parse_args(["--model", model, *vllm_serve_args]) self.uds = args.uds if args.uds: self.host = None self.port = None else: self.host = str(args.host or "127.0.0.1") self.port = int(args.port) self.show_hidden_metrics = args.show_hidden_metrics_for_version is not None # download the model before starting the server to avoid timeout is_local = os.path.isdir(model) if not is_local: engine_args = AsyncEngineArgs.from_cli_args(args) model_config = engine_args.create_model_config() load_config = engine_args.create_load_config() model_loader = get_model_loader(load_config) model_loader.download_model(model_config) self._start_server(model, vllm_serve_args, env_dict) max_wait_seconds = max_wait_seconds or 240 self._wait_for_server(url=self.url_for("health"), timeout=max_wait_seconds) def __enter__(self): return self def __exit__(self, exc_type, exc_value, traceback): self.proc.terminate() try: self.proc.wait(8) except subprocess.TimeoutExpired: # force kill if needed self.proc.kill() def _poll(self) -> int | None: """Subclasses override this method to customize process polling""" return self.proc.poll() def _wait_for_server(self, *, url: str, timeout: float): # run health check start = time.time() client = ( httpx.Client(transport=httpx.HTTPTransport(uds=self.uds)) if self.uds else requests ) while True: try: if client.get(url).status_code == 200: break except Exception: # this exception can only be raised by requests.get, # which means the server is not ready yet. # the stack trace is not useful, so we suppress it # by using `raise from None`. result = self._poll() if result is not None and result != 0: raise RuntimeError("Server exited unexpectedly.") from None time.sleep(0.5) if time.time() - start > timeout: raise RuntimeError("Server failed to start in time.") from None @property def url_root(self) -> str: return ( f"http://{self.uds.split('/')[-1]}" if self.uds else f"http://{self.host}:{self.port}" ) def url_for(self, *parts: str) -> str: return self.url_root + "/" + "/".join(parts) def get_client(self, **kwargs): if "timeout" not in kwargs: kwargs["timeout"] = 600 return openai.OpenAI( base_url=self.url_for("v1"), api_key=self.DUMMY_API_KEY, max_retries=0, **kwargs, ) def get_async_client(self, **kwargs): if "timeout" not in kwargs: kwargs["timeout"] = 600 return openai.AsyncOpenAI( base_url=self.url_for("v1"), api_key=self.DUMMY_API_KEY, max_retries=0, **kwargs, ) def get_client_anthropic(self, **kwargs): if "timeout" not in kwargs: kwargs["timeout"] = 600 return anthropic.Anthropic( base_url=self.url_for(), api_key=self.DUMMY_API_KEY, max_retries=0, **kwargs, ) def get_async_client_anthropic(self, **kwargs): if "timeout" not in kwargs: kwargs["timeout"] = 600 return anthropic.AsyncAnthropic( base_url=self.url_for(), api_key=self.DUMMY_API_KEY, max_retries=0, **kwargs ) class RemoteOpenAIServerCustom(RemoteOpenAIServer): """Launch test server with custom child process""" def _start_server( self, model: str, vllm_serve_args: list[str], env_dict: dict[str, str] | None ) -> None: self.proc: Process = Process( target=self.child_process_fxn, args=(env_dict, model, vllm_serve_args) ) # type: ignore[assignment] self.proc.start() def __init__( self, model: str, vllm_serve_args: list[str], child_process_fxn: Callable[[dict[str, str] | None, str, list[str]], None], *, env_dict: dict[str, str] | None = None, seed: int = 0, auto_port: bool = True, max_wait_seconds: float | None = None, ) -> None: """Store custom child process function then invoke superclass constructor which will indirectly launch it.""" self.child_process_fxn = child_process_fxn super().__init__( model=model, vllm_serve_args=vllm_serve_args, env_dict=env_dict, seed=seed, auto_port=auto_port, max_wait_seconds=max_wait_seconds, ) def _poll(self) -> int | None: return self.proc.exitcode def __exit__(self, exc_type, exc_value, traceback): self.proc.terminate() self.proc.join(8) if self.proc.is_alive(): # force kill if needed self.proc.kill() def _test_completion( client: openai.OpenAI, model: str, prompt: str, token_ids: list[int], ): results = [] # test with text prompt completion = client.completions.create( model=model, prompt=prompt, max_tokens=5, temperature=0.0 ) results.append( { "test": "single_completion", "text": completion.choices[0].text, "finish_reason": completion.choices[0].finish_reason, "usage": completion.usage, } ) # test using token IDs completion = client.completions.create( model=model, prompt=token_ids, max_tokens=5, temperature=0.0, ) results.append( { "test": "token_ids", "text": completion.choices[0].text, "finish_reason": completion.choices[0].finish_reason, "usage": completion.usage, } ) # test seeded random sampling completion = client.completions.create( model=model, prompt=prompt, max_tokens=5, seed=33, temperature=1.0 ) results.append( { "test": "seeded_sampling", "text": completion.choices[0].text, "finish_reason": completion.choices[0].finish_reason, "usage": completion.usage, } ) # test seeded random sampling with multiple prompts completion = client.completions.create( model=model, prompt=[prompt, prompt], max_tokens=5, seed=33, temperature=1.0 ) results.append( { "test": "seeded_sampling", "text": [choice.text for choice in completion.choices], "finish_reason": [choice.finish_reason for choice in completion.choices], "usage": completion.usage, } ) # test simple list batch = client.completions.create( model=model, prompt=[prompt, prompt], max_tokens=5, temperature=0.0, ) results.append( { "test": "simple_list", "text0": batch.choices[0].text, "text1": batch.choices[1].text, } ) # test streaming batch = client.completions.create( model=model, prompt=[prompt, prompt], max_tokens=5, temperature=0.0, stream=True, ) texts = [""] * 2 for chunk in batch: assert len(chunk.choices) == 1 choice = chunk.choices[0] texts[choice.index] += choice.text results.append( { "test": "streaming", "texts": texts, } ) return results def _test_completion_close( client: openai.OpenAI, model: str, prompt: str, ): results = [] # test with text prompt completion = client.completions.create( model=model, prompt=prompt, max_tokens=1, logprobs=5, temperature=0.0 ) logprobs = completion.choices[0].logprobs.top_logprobs[0] logprobs = {k: round(v, 2) for k, v in logprobs.items()} results.append( { "test": "completion_close", "logprobs": logprobs, } ) return results def _test_chat( client: openai.OpenAI, model: str, prompt: str, ): results = [] messages = [{"role": "user", "content": [{"type": "text", "text": prompt}]}] # test with text prompt chat_response = client.chat.completions.create( model=model, messages=messages, max_tokens=5, temperature=0.0 ) results.append( { "test": "completion_close", "text": chat_response.choices[0].message.content, "finish_reason": chat_response.choices[0].finish_reason, "usage": chat_response.usage, } ) return results def _test_embeddings( client: openai.OpenAI, model: str, text: str, ): results = [] # test with text input embeddings = client.embeddings.create( model=model, input=text, encoding_format="float", ) results.append( { "test": "single_embedding", "embedding": embeddings.data[0].embedding, "usage": embeddings.usage, } ) return results def _test_image_text( client: openai.OpenAI, model_name: str, image_url: str, ): results = [] # test pure text input messages = [ { "role": "user", "content": [ {"type": "text", "text": "How do you feel today?"}, ], } ] chat_completion = client.chat.completions.create( model=model_name, messages=messages, temperature=0.0, max_tokens=1, logprobs=True, top_logprobs=5, ) top_logprobs = chat_completion.choices[0].logprobs.content[0].top_logprobs for x in top_logprobs: x.logprob = round(x.logprob, 2) results.append( { "test": "pure_text", "logprobs": top_logprobs, } ) messages = [ { "role": "user", "content": [ {"type": "image_url", "image_url": {"url": image_url}}, {"type": "text", "text": "What's in this image?"}, ], } ] chat_completion = client.chat.completions.create( model=model_name, messages=messages, temperature=0.0, max_tokens=1, logprobs=True, top_logprobs=5, ) top_logprobs = chat_completion.choices[0].logprobs.content[0].top_logprobs results.append( { "test": "text_image", "logprobs": top_logprobs, } ) return results def compare_two_settings( model: str, arg1: list[str], arg2: list[str], env1: dict[str, str] | None = None, env2: dict[str, str] | None = None, *, method: str = "generate", max_wait_seconds: float | None = None, ) -> None: """ Launch API server with two different sets of arguments/environments and compare the results of the API calls. Args: model: The model to test. arg1: The first set of arguments to pass to the API server. arg2: The second set of arguments to pass to the API server. env1: The first set of environment variables to pass to the API server. env2: The second set of environment variables to pass to the API server. """ compare_all_settings( model, [arg1, arg2], [env1, env2], method=method, max_wait_seconds=max_wait_seconds, ) def compare_all_settings( model: str, all_args: list[list[str]], all_envs: list[dict[str, str] | None], *, method: str = "generate", max_wait_seconds: float | None = None, ) -> None: """ Launch API server with several different sets of arguments/environments and compare the results of the API calls with the first set of arguments. Args: model: The model to test. all_args: A list of argument lists to pass to the API server. all_envs: A list of environment dictionaries to pass to the API server. """ trust_remote_code = False for args in all_args: if "--trust-remote-code" in args: trust_remote_code = True break tokenizer_mode = "auto" for args in all_args: if "--tokenizer-mode" in args: tokenizer_mode = args[args.index("--tokenizer-mode") + 1] break tokenizer = get_tokenizer( model, trust_remote_code=trust_remote_code, tokenizer_mode=tokenizer_mode, ) can_force_load_format = True for args in all_args: if "--load-format" in args: can_force_load_format = False break prompt = "Hello, my name is" token_ids = tokenizer(prompt).input_ids ref_results: list = [] for i, (args, env) in enumerate(zip(all_args, all_envs)): if can_force_load_format: # we are comparing the results and # usually we don't need real weights. # we force to use dummy weights by default, # and it should work for most of the cases. # if not, we can use VLLM_TEST_FORCE_LOAD_FORMAT # environment variable to force the load format, # e.g. in quantization tests. args = args + ["--load-format", envs.VLLM_TEST_FORCE_LOAD_FORMAT] compare_results: list = [] results = ref_results if i == 0 else compare_results with RemoteOpenAIServer( model, args, env_dict=env, max_wait_seconds=max_wait_seconds ) as server: client = server.get_client() # test models list models = client.models.list() models = models.data served_model = models[0] results.append( { "test": "models_list", "id": served_model.id, "root": served_model.root, } ) if method == "generate": results += _test_completion(client, model, prompt, token_ids) elif method == "generate_close": results += _test_completion_close(client, model, prompt) elif method == "generate_chat": results += _test_chat(client, model, prompt) elif method == "generate_with_image": results += _test_image_text( client, model, "https://vllm-public-assets.s3.us-west-2.amazonaws.com/vision_model_images/RGBA_comp.png", ) elif method == "encode": results += _test_embeddings(client, model, prompt) else: raise ValueError(f"Unknown method: {method}") if i > 0: # if any setting fails, raise an error early ref_args = all_args[0] ref_envs = all_envs[0] compare_args = all_args[i] compare_envs = all_envs[i] for ref_result, compare_result in zip(ref_results, compare_results): ref_result = copy.deepcopy(ref_result) compare_result = copy.deepcopy(compare_result) if "embedding" in ref_result and method == "encode": sim = F.cosine_similarity( torch.tensor(ref_result["embedding"]), torch.tensor(compare_result["embedding"]), dim=0, ) assert sim >= 0.999, ( f"Embedding for {model=} are not the same.\n" f"cosine_similarity={sim}\n" ) del ref_result["embedding"] del compare_result["embedding"] assert ref_result == compare_result, ( f"Results for {model=} are not the same.\n" f"{ref_args=} {ref_envs=}\n" f"{compare_args=} {compare_envs=}\n" f"{ref_result=}\n" f"{compare_result=}\n" ) def init_test_distributed_environment( tp_size: int, pp_size: int, rank: int, distributed_init_port: str, local_rank: int = -1, ) -> None: distributed_init_method = f"tcp://localhost:{distributed_init_port}" init_distributed_environment( world_size=pp_size * tp_size, rank=rank, distributed_init_method=distributed_init_method, local_rank=local_rank, ) ensure_model_parallel_initialized(tp_size, pp_size) def multi_process_parallel( monkeypatch: pytest.MonkeyPatch, tp_size: int, pp_size: int, test_target: Any, ) -> None: import ray # Using ray helps debugging the error when it failed # as compared to multiprocessing. # NOTE: We need to set working_dir for distributed tests, # otherwise we may get import errors on ray workers # NOTE: Force ray not to use gitignore file as excluding, otherwise # it will not move .so files to working dir. # So we have to manually add some of large directories os.environ["RAY_RUNTIME_ENV_IGNORE_GITIGNORE"] = "1" ray.init( runtime_env={ "working_dir": VLLM_PATH, "excludes": [ "build", ".git", "cmake-build-*", "shellcheck", "dist", "ep_kernels_workspace", ], } ) distributed_init_port = get_open_port() refs = [] for rank in range(tp_size * pp_size): refs.append( test_target.remote( monkeypatch, tp_size, pp_size, rank, distributed_init_port, ), ) ray.get(refs) ray.shutdown() @contextmanager def error_on_warning(category: type[Warning] = Warning): """ Within the scope of this context manager, tests will fail if any warning of the given category is emitted. """ with warnings.catch_warnings(): warnings.filterwarnings("error", category=category) yield def get_physical_device_indices(devices): visible_devices = os.environ.get("CUDA_VISIBLE_DEVICES") if visible_devices is None: return devices visible_indices = [int(x) for x in visible_devices.split(",")] index_mapping = {i: physical for i, physical in enumerate(visible_indices)} return [index_mapping[i] for i in devices if i in index_mapping] @_nvml() def wait_for_gpu_memory_to_clear( *, devices: list[int], threshold_bytes: int | None = None, threshold_ratio: float | None = None, timeout_s: float = 120, ) -> None: assert threshold_bytes is not None or threshold_ratio is not None # Use nvml instead of pytorch to reduce measurement error from torch cuda # context. devices = get_physical_device_indices(devices) start_time = time.time() while True: output: dict[int, str] = {} output_raw: dict[int, tuple[float, float]] = {} for device in devices: if current_platform.is_rocm(): dev_handle = amdsmi_get_processor_handles()[device] mem_info = amdsmi_get_gpu_vram_usage(dev_handle) gb_used = mem_info["vram_used"] / 2**10 gb_total = mem_info["vram_total"] / 2**10 else: dev_handle = nvmlDeviceGetHandleByIndex(device) mem_info = nvmlDeviceGetMemoryInfo(dev_handle) gb_used = mem_info.used / 2**30 gb_total = mem_info.total / 2**30 output_raw[device] = (gb_used, gb_total) output[device] = f"{gb_used:.02f}/{gb_total:.02f}" print("gpu memory used/total (GiB): ", end="") for k, v in output.items(): print(f"{k}={v}; ", end="") print("") if threshold_bytes is not None: is_free = lambda used, total: used <= threshold_bytes / 2**30 threshold = f"{threshold_bytes / 2**30} GiB" else: is_free = lambda used, total: used / total <= threshold_ratio threshold = f"{threshold_ratio:.2f}" dur_s = time.time() - start_time if all(is_free(used, total) for used, total in output_raw.values()): print( f"Done waiting for free GPU memory on devices {devices=} " f"({threshold=}) {dur_s=:.02f}" ) break if dur_s >= timeout_s: raise ValueError( f"Memory of devices {devices=} not free after " f"{dur_s=:.02f} ({threshold=})" ) time.sleep(5) _P = ParamSpec("_P") def fork_new_process_for_each_test(func: Callable[_P, None]) -> Callable[_P, None]: """Decorator to fork a new process for each test function. See https://github.com/vllm-project/vllm/issues/7053 for more details. """ @functools.wraps(func) def wrapper(*args: _P.args, **kwargs: _P.kwargs) -> None: # Make the process the leader of its own process group # to avoid sending SIGTERM to the parent process os.setpgrp() from _pytest.outcomes import Skipped # Create a unique temporary file to store exception info from child # process. Use test function name and process ID to avoid collisions. with ( tempfile.NamedTemporaryFile( delete=False, mode="w+b", prefix=f"vllm_test_{func.__name__}_{os.getpid()}_", suffix=".exc", ) as exc_file, ExitStack() as delete_after, ): exc_file_path = exc_file.name delete_after.callback(os.remove, exc_file_path) pid = os.fork() print(f"Fork a new process to run a test {pid}") if pid == 0: # Parent process responsible for deleting, don't delete # in child. delete_after.pop_all() try: func(*args, **kwargs) except Skipped as e: # convert Skipped to exit code 0 print(str(e)) os._exit(0) except Exception as e: import traceback tb_string = traceback.format_exc() # Try to serialize the exception object first exc_to_serialize: dict[str, Any] try: # First, try to pickle the actual exception with # its traceback. exc_to_serialize = {"pickled_exception": e} # Test if it can be pickled cloudpickle.dumps(exc_to_serialize) except (Exception, KeyboardInterrupt): # Fall back to string-based approach. exc_to_serialize = { "exception_type": type(e).__name__, "exception_msg": str(e), "traceback": tb_string, } try: with open(exc_file_path, "wb") as f: cloudpickle.dump(exc_to_serialize, f) except Exception: # Fallback: just print the traceback. print(tb_string) os._exit(1) else: os._exit(0) else: pgid = os.getpgid(pid) _pid, _exitcode = os.waitpid(pid, 0) # ignore SIGTERM signal itself old_signal_handler = signal.signal(signal.SIGTERM, signal.SIG_IGN) # kill all child processes os.killpg(pgid, signal.SIGTERM) # restore the signal handler signal.signal(signal.SIGTERM, old_signal_handler) if _exitcode != 0: # Try to read the exception from the child process exc_info = {} if os.path.exists(exc_file_path): with ( contextlib.suppress(Exception), open(exc_file_path, "rb") as f, ): exc_info = cloudpickle.load(f) if ( original_exception := exc_info.get("pickled_exception") ) is not None: # Re-raise the actual exception object if it was # successfully pickled. assert isinstance(original_exception, Exception) raise original_exception if (original_tb := exc_info.get("traceback")) is not None: # Use string-based traceback for fallback case raise AssertionError( f"Test {func.__name__} failed when called with" f" args {args} and kwargs {kwargs}" f" (exit code: {_exitcode}):\n{original_tb}" ) from None # Fallback to the original generic error raise AssertionError( f"function {func.__name__} failed when called with" f" args {args} and kwargs {kwargs}" f" (exit code: {_exitcode})" ) from None return wrapper def spawn_new_process_for_each_test(f: Callable[_P, None]) -> Callable[_P, None]: """Decorator to spawn a new process for each test function.""" @functools.wraps(f) def wrapper(*args: _P.args, **kwargs: _P.kwargs) -> None: # Check if we're already in a subprocess if os.environ.get("RUNNING_IN_SUBPROCESS") == "1": # If we are, just run the function directly return f(*args, **kwargs) import torch.multiprocessing as mp with suppress(RuntimeError): mp.set_start_method("spawn") # Get the module module_name = f.__module__ # Create a process with environment variable set env = os.environ.copy() env["RUNNING_IN_SUBPROCESS"] = "1" with tempfile.TemporaryDirectory() as tempdir: output_filepath = os.path.join(tempdir, "new_process.tmp") # `cloudpickle` allows pickling complex functions directly input_bytes = cloudpickle.dumps((f, output_filepath)) repo_root = str(VLLM_PATH.resolve()) env = dict(env or os.environ) env["PYTHONPATH"] = repo_root + os.pathsep + env.get("PYTHONPATH", "") cmd = [sys.executable, "-m", f"{module_name}"] returned = subprocess.run( cmd, input=input_bytes, capture_output=True, env=env ) # check if the subprocess is successful try: returned.check_returncode() except Exception as e: # wrap raised exception to provide more information raise RuntimeError( f"Error raised in subprocess:\n{returned.stderr.decode()}" ) from e return wrapper def create_new_process_for_each_test( method: Literal["spawn", "fork"] | None = None, ) -> Callable[[Callable[_P, None]], Callable[_P, None]]: """Creates a decorator that runs each test function in a new process. Args:
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_routing_simulator.py
tests/test_routing_simulator.py
#!/usr/bin/env python3 # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Test script for the token-to-expert routing simulator. This script demonstrates how to use the routing simulator to test different routing strategies and analyze their performance, including integration tests with FusedMoE layer. """ import tempfile import pytest import torch from vllm.config import VllmConfig, set_current_vllm_config from vllm.distributed import ( init_distributed_environment, initialize_model_parallel, ) from vllm.model_executor.layers.fused_moe.routing_simulator import ( DistributionBasedRouting, RoutingSimulator, ) @pytest.fixture def device(): """Fixture to provide the appropriate device for testing.""" return torch.device("cuda" if torch.cuda.is_available() else "cpu") @pytest.mark.parametrize("num_tokens", [1, 16, 256]) @pytest.mark.parametrize("hidden_size", [64, 1024]) @pytest.mark.parametrize("num_experts", [16, 128]) @pytest.mark.parametrize("top_k", [1, 4]) def test_basic_functionality( num_tokens: int, hidden_size: int, num_experts: int, top_k: int, device, ): """Test basic functionality of the routing simulator.""" # Test each routing strategy strategies = RoutingSimulator.get_available_strategies() hidden_states = torch.randn(num_tokens, hidden_size, device=device) router_logits = torch.randn(num_tokens, num_experts, device=device) for strategy in strategies: # Simulate routing topk_weights, topk_ids = RoutingSimulator.simulate_routing( hidden_states=hidden_states, router_logits=router_logits, strategy_name=strategy, top_k=top_k, ) # Check output shapes assert topk_weights.shape == ( num_tokens, top_k, ), f"Wrong weights shape for {strategy}" assert topk_ids.shape == ( num_tokens, top_k, ), f"Wrong ids shape for {strategy}" # Check that expert IDs are valid assert topk_ids.min() >= 0, f"Invalid expert ID (negative) for {strategy}" assert topk_ids.max() < num_experts, ( f"Invalid expert ID (too large) for {strategy}" ) def test_routing_strategy_integration(monkeypatch, device): """Test that the routing strategy environment variable works with FusedMoE.""" pytest.importorskip("vllm.model_executor.layers.fused_moe.layer") import vllm.envs as envs from vllm.model_executor.layers.fused_moe.layer import FusedMoE # Test parameters num_tokens = 32 hidden_size = 16 num_experts = 4 top_k = 2 # Create test data hidden_states = torch.randn(num_tokens, hidden_size, device=device) router_logits = torch.randn(num_tokens, num_experts, device=device) # Test different routing strategies strategies = RoutingSimulator.get_available_strategies() vllm_config = VllmConfig() with set_current_vllm_config(vllm_config): temp_file = tempfile.mkstemp()[1] init_distributed_environment( world_size=1, rank=0, local_rank=0, distributed_init_method=f"file://{temp_file}", ) initialize_model_parallel( tensor_model_parallel_size=1, pipeline_model_parallel_size=1, ) fused_moe = FusedMoE( num_experts=num_experts, top_k=top_k, hidden_size=hidden_size, intermediate_size=0, use_grouped_topk=False, renormalize=True, ) for strategy in strategies: # Set environment variable env_name = "VLLM_MOE_ROUTING_SIMULATION_STRATEGY" monkeypatch.setenv(env_name, strategy) # Force reload of environment variable envs.environment_variables[env_name] = lambda s=strategy: s # Test the select_experts method topk_weights, topk_ids = fused_moe.select_experts( hidden_states=hidden_states, router_logits=router_logits, ) # Verify output shapes assert topk_weights.shape == (num_tokens, top_k), ( f"Wrong weights shape for {strategy}" ) assert topk_ids.shape == (num_tokens, top_k), f"Wrong ids shape for {strategy}" # Verify expert IDs are valid assert topk_ids.min() >= 0, f"Invalid expert ID (negative) for {strategy}" assert topk_ids.max() < num_experts, ( f"Invalid expert ID (too large) for {strategy}" ) def test_distribution_based_routing_with_custom_strategy(): """Test registering and using DistributionBasedRouting with custom parameters.""" device = torch.device("cuda" if torch.cuda.is_available() else "cpu") # Register custom distribution-based strategy custom_strategy = DistributionBasedRouting(distribution="normal", mean=2.0, std=0.5) RoutingSimulator.register_strategy("custom_normal", custom_strategy) # Test data num_tokens = 60 hidden_size = 48 num_experts = 6 top_k = 3 hidden_states = torch.randn(num_tokens, hidden_size, device=device) router_logits = torch.randn(num_tokens, num_experts, device=device) # Use the custom strategy topk_weights, topk_ids = RoutingSimulator.simulate_routing( hidden_states=hidden_states, router_logits=router_logits, strategy_name="custom_normal", top_k=top_k, ) # Check output shapes assert topk_weights.shape == (num_tokens, top_k) assert topk_ids.shape == (num_tokens, top_k) # Check that expert IDs are valid assert topk_ids.min() >= 0 assert topk_ids.max() < num_experts def test_instance_compatibility(): """Test that static methods work correctly.""" device = torch.device("cuda" if torch.cuda.is_available() else "cpu") # Test static method directly hidden_states = torch.randn(10, 8, device=device) router_logits = torch.randn(10, 4, device=device) topk_weights, topk_ids = RoutingSimulator.simulate_routing( hidden_states=hidden_states, router_logits=router_logits, strategy_name="uniform_random", top_k=2, ) assert topk_weights.shape == (10, 2) assert topk_ids.shape == (10, 2)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_version.py
tests/test_version.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from unittest.mock import patch import pytest from vllm import version def test_version_is_defined(): assert version.__version__ is not None def test_version_tuple(): assert len(version.__version_tuple__) in (3, 4, 5) @pytest.mark.parametrize( "version_tuple, version_str, expected", [ ((0, 0, "dev"), "0.0", True), ((0, 0, "dev"), "foobar", True), ((0, 7, 4), "0.6", True), ((0, 7, 4), "0.5", False), ((0, 7, 4), "0.7", False), ((1, 2, 3), "1.1", True), ((1, 2, 3), "1.0", False), ((1, 2, 3), "1.2", False), # This won't work as expected ((1, 0, 0), "1.-1", True), ((1, 0, 0), "0.9", False), ((1, 0, 0), "0.17", False), ], ) def test_prev_minor_version_was(version_tuple, version_str, expected): with patch("vllm.version.__version_tuple__", version_tuple): assert version._prev_minor_version_was(version_str) == expected
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_logprobs.py
tests/test_logprobs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from vllm.logprobs import ( FlatLogprobs, Logprob, LogprobsOnePosition, append_logprobs_for_next_position, create_prompt_logprobs, create_sample_logprobs, ) def test_create_logprobs_non_flat() -> None: prompt_logprobs = create_prompt_logprobs(flat_logprobs=False) assert isinstance(prompt_logprobs, list) # Ensure first prompt position logprobs is None assert len(prompt_logprobs) == 1 assert prompt_logprobs[0] is None sample_logprobs = create_sample_logprobs(flat_logprobs=False) assert isinstance(sample_logprobs, list) assert len(sample_logprobs) == 0 def test_create_logprobs_flat() -> None: prompt_logprobs = create_prompt_logprobs(flat_logprobs=True) assert isinstance(prompt_logprobs, FlatLogprobs) assert prompt_logprobs.start_indices == [0] assert prompt_logprobs.end_indices == [0] assert len(prompt_logprobs.token_ids) == 0 assert len(prompt_logprobs.logprobs) == 0 assert len(prompt_logprobs.ranks) == 0 assert len(prompt_logprobs.decoded_tokens) == 0 # Ensure first prompt position logprobs is empty assert len(prompt_logprobs) == 1 assert prompt_logprobs[0] == dict() sample_logprobs = create_sample_logprobs(flat_logprobs=True) assert isinstance(sample_logprobs, FlatLogprobs) assert len(sample_logprobs.start_indices) == 0 assert len(sample_logprobs.end_indices) == 0 assert len(sample_logprobs.token_ids) == 0 assert len(sample_logprobs.logprobs) == 0 assert len(sample_logprobs.ranks) == 0 assert len(sample_logprobs.decoded_tokens) == 0 assert len(sample_logprobs) == 0 def test_append_logprobs_for_next_position_none_flat() -> None: logprobs = create_sample_logprobs(flat_logprobs=False) append_logprobs_for_next_position( logprobs, token_ids=[1], logprobs=[0.1], decoded_tokens=["1"], rank=10, num_logprobs=-1, ) append_logprobs_for_next_position( logprobs, token_ids=[2, 3], logprobs=[0.2, 0.3], decoded_tokens=["2", "3"], rank=11, num_logprobs=-1, ) assert isinstance(logprobs, list) assert logprobs == [ {1: Logprob(logprob=0.1, rank=10, decoded_token="1")}, { 2: Logprob(logprob=0.2, rank=11, decoded_token="2"), 3: Logprob(logprob=0.3, rank=1, decoded_token="3"), }, ] def test_append_logprobs_for_next_position_flat() -> None: logprobs = create_sample_logprobs(flat_logprobs=True) append_logprobs_for_next_position( logprobs, token_ids=[1], logprobs=[0.1], decoded_tokens=["1"], rank=10, num_logprobs=-1, ) append_logprobs_for_next_position( logprobs, token_ids=[2, 3], logprobs=[0.2, 0.3], decoded_tokens=["2", "3"], rank=11, num_logprobs=-1, ) assert isinstance(logprobs, FlatLogprobs) assert logprobs.start_indices == [0, 1] assert logprobs.end_indices == [1, 3] assert logprobs.token_ids == [1, 2, 3] assert logprobs.logprobs == [0.1, 0.2, 0.3] assert logprobs.ranks == [10, 11, 1] assert logprobs.decoded_tokens == ["1", "2", "3"] LOGPROBS_ONE_POSITION_0: LogprobsOnePosition = { 1: Logprob(logprob=0.1, rank=10, decoded_token="10") } LOGPROBS_ONE_POSITION_1: LogprobsOnePosition = { 2: Logprob(logprob=0.2, rank=20, decoded_token="20"), 3: Logprob(logprob=0.3, rank=30, decoded_token="30"), } LOGPROBS_ONE_POSITION_2: LogprobsOnePosition = { 4: Logprob(logprob=0.4, rank=40, decoded_token="40"), 5: Logprob(logprob=0.5, rank=50, decoded_token="50"), 6: Logprob(logprob=0.6, rank=60, decoded_token="60"), } def test_flat_logprobs_append() -> None: logprobs = FlatLogprobs() logprobs.append(LOGPROBS_ONE_POSITION_0) logprobs.append(LOGPROBS_ONE_POSITION_1) assert logprobs.start_indices == [0, 1] assert logprobs.end_indices == [1, 3] assert logprobs.token_ids == [1, 2, 3] assert logprobs.logprobs == [0.1, 0.2, 0.3] assert logprobs.ranks == [10, 20, 30] assert logprobs.decoded_tokens == ["10", "20", "30"] logprobs.append(LOGPROBS_ONE_POSITION_2) assert logprobs.start_indices == [0, 1, 3] assert logprobs.end_indices == [1, 3, 6] assert logprobs.token_ids == [1, 2, 3, 4, 5, 6] assert logprobs.logprobs == [0.1, 0.2, 0.3, 0.4, 0.5, 0.6] assert logprobs.ranks == [10, 20, 30, 40, 50, 60] assert logprobs.decoded_tokens == ["10", "20", "30", "40", "50", "60"] def test_flat_logprobs_extend() -> None: logprobs = FlatLogprobs() # Extend with list[LogprobsOnePosition] logprobs.extend([LOGPROBS_ONE_POSITION_2, LOGPROBS_ONE_POSITION_0]) assert logprobs.start_indices == [0, 3] assert logprobs.end_indices == [3, 4] assert logprobs.token_ids == [4, 5, 6, 1] assert logprobs.logprobs == [0.4, 0.5, 0.6, 0.1] assert logprobs.ranks == [40, 50, 60, 10] assert logprobs.decoded_tokens == ["40", "50", "60", "10"] other_logprobs = FlatLogprobs() other_logprobs.extend([LOGPROBS_ONE_POSITION_1, LOGPROBS_ONE_POSITION_0]) # Extend with another FlatLogprobs logprobs.extend(other_logprobs) assert logprobs.start_indices == [0, 3, 4, 6] assert logprobs.end_indices == [3, 4, 6, 7] assert logprobs.token_ids == [4, 5, 6, 1, 2, 3, 1] assert logprobs.logprobs == [0.4, 0.5, 0.6, 0.1, 0.2, 0.3, 0.1] assert logprobs.ranks == [40, 50, 60, 10, 20, 30, 10] assert logprobs.decoded_tokens == ["40", "50", "60", "10", "20", "30", "10"] def test_flat_logprobs_access() -> None: logprobs = FlatLogprobs() logprobs.extend( [LOGPROBS_ONE_POSITION_1, LOGPROBS_ONE_POSITION_2, LOGPROBS_ONE_POSITION_0] ) assert logprobs.start_indices == [0, 2, 5] assert logprobs.end_indices == [2, 5, 6] assert logprobs.token_ids == [2, 3, 4, 5, 6, 1] assert logprobs.logprobs == [0.2, 0.3, 0.4, 0.5, 0.6, 0.1] assert logprobs.ranks == [20, 30, 40, 50, 60, 10] assert logprobs.decoded_tokens == ["20", "30", "40", "50", "60", "10"] # Test __len__ assert len(logprobs) == 3 # Test __iter__ for actual_logprobs, expected_logprobs in zip( logprobs, [LOGPROBS_ONE_POSITION_1, LOGPROBS_ONE_POSITION_2, LOGPROBS_ONE_POSITION_0], ): assert actual_logprobs == expected_logprobs # Test __getitem__ : single item assert logprobs[0] == LOGPROBS_ONE_POSITION_1 assert logprobs[1] == LOGPROBS_ONE_POSITION_2 assert logprobs[2] == LOGPROBS_ONE_POSITION_0 # Test __getitem__ : slice logprobs02 = logprobs[:2] assert len(logprobs02) == 2 assert logprobs02[0] == LOGPROBS_ONE_POSITION_1 assert logprobs02[1] == LOGPROBS_ONE_POSITION_2 assert logprobs02.start_indices == [0, 2] assert logprobs02.end_indices == [2, 5] assert logprobs02.token_ids == [2, 3, 4, 5, 6] assert logprobs02.logprobs == [0.2, 0.3, 0.4, 0.5, 0.6] assert logprobs02.ranks == [20, 30, 40, 50, 60] assert logprobs02.decoded_tokens == ["20", "30", "40", "50", "60"] logprobs_last2 = logprobs[-2:] assert len(logprobs_last2) == 2 assert logprobs_last2[0] == LOGPROBS_ONE_POSITION_2 assert logprobs_last2[1] == LOGPROBS_ONE_POSITION_0 assert logprobs_last2.start_indices == [0, 3] assert logprobs_last2.end_indices == [3, 4] assert logprobs_last2.token_ids == [4, 5, 6, 1] assert logprobs_last2.logprobs == [0.4, 0.5, 0.6, 0.1] assert logprobs_last2.ranks == [40, 50, 60, 10] assert logprobs_last2.decoded_tokens == ["40", "50", "60", "10"]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/__init__.py
tests/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/ci_envs.py
tests/ci_envs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ These envs only work for a small part of the tests, fix what you need! """ import os from collections.abc import Callable from typing import TYPE_CHECKING, Any from vllm.envs import maybe_convert_bool if TYPE_CHECKING: VLLM_CI_NO_SKIP: bool = False VLLM_CI_DTYPE: str | None = None VLLM_CI_HEAD_DTYPE: str | None = None VLLM_CI_HF_DTYPE: str | None = None environment_variables: dict[str, Callable[[], Any]] = { # A model family has many models with the same architecture. # By default, a model family tests only one model. # Through this flag, all models can be tested. "VLLM_CI_NO_SKIP": lambda: bool(int(os.getenv("VLLM_CI_NO_SKIP", "0"))), # Allow changing the dtype used by vllm in tests "VLLM_CI_DTYPE": lambda: os.getenv("VLLM_CI_DTYPE", None), # Allow changing the head dtype used by vllm in tests "VLLM_CI_HEAD_DTYPE": lambda: os.getenv("VLLM_CI_HEAD_DTYPE", None), # Allow changing the head dtype used by transformers in tests "VLLM_CI_HF_DTYPE": lambda: os.getenv("VLLM_CI_HF_DTYPE", None), # Allow control over whether tests use enforce_eager "VLLM_CI_ENFORCE_EAGER": lambda: maybe_convert_bool( os.getenv("VLLM_CI_ENFORCE_EAGER", None) ), } def __getattr__(name: str): # lazy evaluation of environment variables if name in environment_variables: return environment_variables[name]() raise AttributeError(f"module {__name__!r} has no attribute {name!r}") def __dir__(): return list(environment_variables.keys()) def is_set(name: str): """Check if an environment variable is explicitly set.""" if name in environment_variables: return name in os.environ raise AttributeError(f"module {__name__!r} has no attribute {name!r}")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_regression.py
tests/test_regression.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Containing tests that check for regressions in vLLM's behavior. It should include tests that are reported by users and making sure they will never happen again. """ import gc import pytest import torch from vllm import LLM, SamplingParams @pytest.mark.skip(reason="In V1, we reject tokens > max_seq_len") def test_duplicated_ignored_sequence_group(): """https://github.com/vllm-project/vllm/issues/1655""" sampling_params = SamplingParams(temperature=0.01, top_p=0.1, max_tokens=256) llm = LLM( model="distilbert/distilgpt2", max_num_batched_tokens=4096, tensor_parallel_size=1, ) prompts = ["This is a short prompt", "This is a very long prompt " * 1000] outputs = llm.generate(prompts, sampling_params=sampling_params) assert len(prompts) == len(outputs) def test_max_tokens_none(): sampling_params = SamplingParams(temperature=0.01, top_p=0.1, max_tokens=None) llm = LLM( model="distilbert/distilgpt2", max_num_batched_tokens=4096, tensor_parallel_size=1, ) prompts = ["Just say hello!"] outputs = llm.generate(prompts, sampling_params=sampling_params) assert len(prompts) == len(outputs) def test_gc(): llm = LLM(model="distilbert/distilgpt2", enforce_eager=True) del llm gc.collect() torch.cuda.empty_cache() # The memory allocated for model and KV cache should be released. # The memory allocated for PyTorch and others should be less than 50MB. # Usually, it's around 10MB. allocated = torch.cuda.memory_allocated() assert allocated < 50 * 1024 * 1024 def test_model_from_modelscope(monkeypatch: pytest.MonkeyPatch): # model: https://modelscope.cn/models/qwen/Qwen1.5-0.5B-Chat/summary with monkeypatch.context() as m: m.setenv("VLLM_USE_MODELSCOPE", "True") # Don't use HF_TOKEN for ModelScope repos, otherwise it will fail # with 400 Client Error: Bad Request. m.setenv("HF_TOKEN", "") llm = LLM(model="qwen/Qwen1.5-0.5B-Chat") prompts = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] sampling_params = SamplingParams(temperature=0.8, top_p=0.95) outputs = llm.generate(prompts, sampling_params) assert len(outputs) == 4
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_scalartype.py
tests/test_scalartype.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from vllm.scalar_type import scalar_types @pytest.mark.parametrize( "type_tuple", ( (-8, 7, scalar_types.int4), (0, 15, scalar_types.uint4), (-8, 7, scalar_types.uint4b8), (-128, 127, scalar_types.uint8b128), (-6.0, 6.0, scalar_types.float4_e2m1f), (-28.0, 28.0, scalar_types.float6_e3m2f), (torch.int8, scalar_types.int8), (torch.uint8, scalar_types.uint8), (torch.float8_e5m2, scalar_types.float8_e5m2), (torch.float8_e4m3fn, scalar_types.float8_e4m3fn), (torch.bfloat16, scalar_types.float16_e8m7), (torch.float16, scalar_types.float16_e5m10), ), ids=lambda x: str(x), ) def test_scalar_type_min_max(type_tuple): print(type_tuple) if len(type_tuple) == 3: min, max, t = type_tuple else: torch_type, t = type_tuple if torch_type.is_floating_point: min = torch.finfo(torch_type).min max = torch.finfo(torch_type).max else: min = torch.iinfo(torch_type).min max = torch.iinfo(torch_type).max print(t, min, max, t.min(), t.max()) assert min == t.min(), f"min: {min} != {t.min()}" assert max == t.max(), f"max: {max} != {t.max()}"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/test_attention_backend_registry.py
tests/test_attention_backend_registry.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from vllm.attention.backends.abstract import ( AttentionBackend, AttentionImpl, ) from vllm.attention.backends.registry import ( AttentionBackendEnum, MambaAttentionBackendEnum, register_backend, ) class CustomAttentionImpl(AttentionImpl): """Mock custom attention implementation for testing.""" def __init__(self, *args, **kwargs): super().__init__() def forward(self, *args, **kwargs): """Mock forward pass.""" pass class CustomAttentionBackend(AttentionBackend): """Mock custom attention backend for testing.""" @staticmethod def get_name(): return "CUSTOM" @staticmethod def get_impl_cls(): return CustomAttentionImpl @staticmethod def get_builder_cls(): """Mock builder class.""" return None @staticmethod def get_required_kv_cache_layout(): """Mock KV cache layout.""" return None class CustomMambaAttentionImpl(AttentionImpl): """Mock custom mamba attention implementation for testing.""" def __init__(self, *args, **kwargs): super().__init__() def forward(self, *args, **kwargs): """Mock forward pass.""" pass class CustomMambaAttentionBackend(AttentionBackend): """Mock custom mamba attention backend for testing.""" @staticmethod def get_name(): return "CUSTOM_MAMBA" @staticmethod def get_impl_cls(): return CustomMambaAttentionImpl @staticmethod def get_builder_cls(): """Mock builder class.""" return None @staticmethod def get_required_kv_cache_layout(): """Mock KV cache layout.""" return None def test_custom_is_not_alias_of_any_backend(): # Get all members of AttentionBackendEnum all_backends = list(AttentionBackendEnum) # Find any aliases of CUSTOM aliases = [] for backend in all_backends: if backend.name != "CUSTOM" and backend is AttentionBackendEnum.CUSTOM: aliases.append(backend.name) # CUSTOM should not be an alias of any other backend assert len(aliases) == 0, ( f"BUG! CUSTOM is an alias of: {', '.join(aliases)}!\n" f"CUSTOM.value = {repr(AttentionBackendEnum.CUSTOM.value)}\n" f"This happens when CUSTOM has the same value as another backend.\n" f"When you register to CUSTOM, you're actually registering to {aliases[0]}!\n" f"All backend values:\n" + "\n".join(f" {b.name}: {repr(b.value)}" for b in all_backends) ) # Verify CUSTOM has its own unique identity assert AttentionBackendEnum.CUSTOM.name == "CUSTOM", ( f"CUSTOM.name should be 'CUSTOM', but got '{AttentionBackendEnum.CUSTOM.name}'" ) def test_register_custom_backend_with_class_path(): # Register with explicit class path register_backend( backend=AttentionBackendEnum.CUSTOM, class_path="tests.test_attention_backend_registry.CustomAttentionBackend", is_mamba=False, ) # Check that CUSTOM backend is registered assert AttentionBackendEnum.CUSTOM.is_overridden(), ( "CUSTOM should be overridden after registration" ) # Get the registered class path class_path = AttentionBackendEnum.CUSTOM.get_path() assert class_path == "tests.test_attention_backend_registry.CustomAttentionBackend" # Get the backend class backend_cls = AttentionBackendEnum.CUSTOM.get_class() assert backend_cls.get_name() == "CUSTOM" assert backend_cls.get_impl_cls() == CustomAttentionImpl def test_mamba_custom_is_not_alias_of_any_backend(): # Get all mamba backends all_backends = list(MambaAttentionBackendEnum) # Find any aliases of CUSTOM aliases = [] for backend in all_backends: if backend.name != "CUSTOM" and backend is MambaAttentionBackendEnum.CUSTOM: aliases.append(backend.name) # CUSTOM should not be an alias of any other backend assert len(aliases) == 0, ( f"BUG! MambaAttentionBackendEnum.CUSTOM is an alias of: {', '.join(aliases)}!\n" f"CUSTOM.value = {repr(MambaAttentionBackendEnum.CUSTOM.value)}\n" f"All mamba backend values:\n" + "\n".join(f" {b.name}: {repr(b.value)}" for b in all_backends) ) def test_register_custom_mamba_backend_with_class_path(): # Register with explicit class path register_backend( backend=MambaAttentionBackendEnum.CUSTOM, class_path="tests.test_attention_backend_registry.CustomMambaAttentionBackend", is_mamba=True, ) # Check that the backend is registered assert MambaAttentionBackendEnum.CUSTOM.is_overridden() # Get the registered class path class_path = MambaAttentionBackendEnum.CUSTOM.get_path() assert ( class_path == "tests.test_attention_backend_registry.CustomMambaAttentionBackend" ) # Get the backend class backend_cls = MambaAttentionBackendEnum.CUSTOM.get_class() assert backend_cls.get_name() == "CUSTOM_MAMBA" assert backend_cls.get_impl_cls() == CustomMambaAttentionImpl
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/vllm_test_utils/setup.py
tests/vllm_test_utils/setup.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from setuptools import setup setup( name="vllm_test_utils", version="0.1", packages=["vllm_test_utils"], )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/vllm_test_utils/vllm_test_utils/__init__.py
tests/vllm_test_utils/vllm_test_utils/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ vllm_utils is a package for vLLM testing utilities. It does not import any vLLM modules. """ from .blame import BlameResult, blame from .monitor import MonitoredValues, monitor __all__ = ["blame", "BlameResult", "monitor", "MonitoredValues"]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/vllm_test_utils/vllm_test_utils/blame.py
tests/vllm_test_utils/vllm_test_utils/blame.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import contextlib import dataclasses import sys import traceback from collections.abc import Callable, Generator @dataclasses.dataclass class BlameResult: found: bool = False trace_stack: str = "" @contextlib.contextmanager def blame(func: Callable) -> Generator[BlameResult, None, None]: """ Trace the function calls to find the first function that satisfies the condition. The trace stack will be stored in the result. Usage: ```python with blame(lambda: some_condition()) as result: # do something if result.found: print(result.trace_stack) """ result = BlameResult() def _trace_calls(frame, event, arg=None): nonlocal result if event in ["call", "return"]: # for every function call or return try: # Temporarily disable the trace function sys.settrace(None) # check condition here if not result.found and func(): result.found = True result.trace_stack = "".join(traceback.format_stack()) # Re-enable the trace function sys.settrace(_trace_calls) except NameError: # modules are deleted during shutdown pass return _trace_calls try: sys.settrace(_trace_calls) yield result finally: sys.settrace(None)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/vllm_test_utils/vllm_test_utils/monitor.py
tests/vllm_test_utils/vllm_test_utils/monitor.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import contextlib import dataclasses import sys import traceback from collections.abc import Callable, Generator from typing import Generic, TypeVar _T = TypeVar("_T") @dataclasses.dataclass class MonitoredValues(Generic[_T]): values: list[_T] = dataclasses.field(default_factory=list) trace_stacks: list[str] = dataclasses.field(default_factory=list) @contextlib.contextmanager def monitor( measure_func: Callable[[], _T], ) -> Generator[MonitoredValues[_T], None, None]: """ Trace the function calls to continuously monitor the change of a value. Usage: ```python def measure_func(): ... # measure the current value return current_value with monitor(measure_func) as monitored_values: # do something monitored_values.values # all changes of the values monitored_values.trace_stacks # trace stacks of every change ``` """ monitored_values = MonitoredValues[_T]() def _trace_calls(frame, event, arg=None): nonlocal monitored_values if event in ["line"]: # triggered by every line of Python code. # only Python functions will trigger it, # c/cpp functions will not trigger it. try: # Temporarily disable the trace function sys.settrace(None) # do a measurement current_value = measure_func() if ( len(monitored_values.values) == 0 or current_value != monitored_values.values[-1] ): monitored_values.values.append(current_value) monitored_values.trace_stacks.append( "".join(traceback.format_stack()) ) # Re-enable the trace function sys.settrace(_trace_calls) except NameError: # modules are deleted during shutdown pass return _trace_calls try: sys.settrace(_trace_calls) yield monitored_values finally: sys.settrace(None)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/benchmarks/test_serve_cli.py
tests/benchmarks/test_serve_cli.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import subprocess import pytest from ..utils import RemoteOpenAIServer MODEL_NAME = "meta-llama/Llama-3.2-1B-Instruct" @pytest.fixture(scope="module") def server(): args = ["--max-model-len", "1024", "--enforce-eager", "--load-format", "dummy"] with RemoteOpenAIServer(MODEL_NAME, args) as remote_server: yield remote_server @pytest.mark.benchmark def test_bench_serve(server): # Test default model detection and input/output len command = [ "vllm", "bench", "serve", "--host", server.host, "--port", str(server.port), "--input-len", "32", "--output-len", "4", "--num-prompts", "5", ] result = subprocess.run(command, capture_output=True, text=True) print(result.stdout) print(result.stderr) assert result.returncode == 0, f"Benchmark failed: {result.stderr}" @pytest.mark.benchmark def test_bench_serve_chat(server): command = [ "vllm", "bench", "serve", "--model", MODEL_NAME, "--host", server.host, "--port", str(server.port), "--dataset-name", "random", "--random-input-len", "32", "--random-output-len", "4", "--num-prompts", "5", "--endpoint", "/v1/chat/completions", "--backend", "openai-chat", ] result = subprocess.run(command, capture_output=True, text=True) print(result.stdout) print(result.stderr) assert result.returncode == 0, f"Benchmark failed: {result.stderr}"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/benchmarks/test_random_dataset.py
tests/benchmarks/test_random_dataset.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import random from typing import Any, NamedTuple, cast import numpy as np import pytest from transformers import AutoTokenizer, PreTrainedTokenizerBase from vllm.benchmarks.datasets import ( RandomDataset, RandomMultiModalDataset, SampleRequest, ) @pytest.fixture(scope="session") def hf_tokenizer() -> PreTrainedTokenizerBase: # Use a small, commonly available tokenizer return AutoTokenizer.from_pretrained("gpt2") class Params(NamedTuple): num_requests: int prefix_len: int range_ratio: float input_len: int output_len: int @pytest.fixture(scope="session") def random_dataset_params() -> Params: return Params( num_requests=16, prefix_len=7, range_ratio=0.3, input_len=50, output_len=20 ) def _fingerprint_sample(req: SampleRequest) -> tuple[str, int, int]: """Project a SampleRequest into a comparable tuple.""" return (req.prompt, req.prompt_len, req.expected_output_len) def _collect_samples( dataset: RandomDataset, tokenizer: PreTrainedTokenizerBase, num_requests: int = 16, prefix_len: int = 7, range_ratio: float = 0.3, input_len: int = 50, output_len: int = 20, ) -> list[tuple[str, int, int]]: samples = dataset.sample( tokenizer=tokenizer, num_requests=num_requests, prefix_len=prefix_len, range_ratio=range_ratio, input_len=input_len, output_len=output_len, ) return [_fingerprint_sample(s) for s in samples] @pytest.mark.benchmark def test_random_dataset_same_seed( hf_tokenizer: PreTrainedTokenizerBase, random_dataset_params: Params ) -> None: """Same seed should yield identical outputs, even if global RNGs change. This guards against accidental reliance on Python's random or np.random in RandomDataset after moving to numpy.default_rng. """ p = random_dataset_params common_seed = 123 dataset_a = RandomDataset(random_seed=common_seed) dataset_b = RandomDataset(random_seed=common_seed) a = _collect_samples( dataset_a, hf_tokenizer, num_requests=p.num_requests, prefix_len=p.prefix_len, range_ratio=p.range_ratio, input_len=p.input_len, output_len=p.output_len, ) # Perturb global RNG state to ensure isolation random.seed(999) _ = [random.random() for _ in range(100)] np.random.seed(888) _ = [np.random.random() for _ in range(100)] b = _collect_samples( dataset_b, hf_tokenizer, num_requests=p.num_requests, prefix_len=p.prefix_len, range_ratio=p.range_ratio, input_len=p.input_len, output_len=p.output_len, ) assert a == b @pytest.mark.benchmark def test_random_dataset_different_seeds( hf_tokenizer: PreTrainedTokenizerBase, random_dataset_params: Params ) -> None: """Different seeds should change outputs with overwhelming likelihood.""" p = random_dataset_params seed_a = 0 dataset_a = RandomDataset(random_seed=seed_a) a = _collect_samples( dataset_a, hf_tokenizer, num_requests=p.num_requests, prefix_len=p.prefix_len, range_ratio=p.range_ratio, input_len=p.input_len, output_len=p.output_len, ) seed_b = 999 dataset_b = RandomDataset(random_seed=seed_b) # Perturb global RNG with same seed as dataset_a to ensure isolation random.seed(seed_a) np.random.seed(seed_a) b = _collect_samples( dataset_b, hf_tokenizer, num_requests=p.num_requests, prefix_len=p.prefix_len, range_ratio=p.range_ratio, input_len=p.input_len, output_len=p.output_len, ) assert a != b # ----------------------------- # RandomMultiModalDataset tests # ----------------------------- def _mm_fingerprint_sample( req: SampleRequest, ) -> tuple[str, int, int, int, list[str]]: """Create a compact fingerprint for multimodal samples. Includes: - prompt string - prompt_len - expected_output_len - count of multimodal items - per-item type and URL prefix (e.g., 'data:image/jpeg;base64,') """ items = req.multi_modal_data or [] item_prefixes: list[str] = [] for it in items: if isinstance(it, dict) and it.get("type") == "image_url": url = it.get("image_url", {}).get("url", "") # Only keep a short identifying prefix to avoid huge strings item_prefixes.append(f"image:{url[:22]}") elif isinstance(it, dict) and it.get("type") == "video_url": url = it.get("video_url", {}).get("url", "") item_prefixes.append(f"video:{url[:22]}") else: item_prefixes.append("unknown:") return ( req.prompt, req.prompt_len, req.expected_output_len, len(items), item_prefixes, ) def _collect_mm_samples( dataset: RandomMultiModalDataset, tokenizer: PreTrainedTokenizerBase, *, num_requests: int = 8, prefix_len: int = 3, range_ratio: float = 0.0, input_len: int = 20, output_len: int = 5, base_items_per_request: int = 2, num_mm_items_range_ratio: float = 0.0, limit_mm_per_prompt: dict[str, int] | None = None, bucket_config: dict[tuple[int, int, int], float] | None = None, enable_multimodal_chat: bool = False, ) -> list[SampleRequest]: if limit_mm_per_prompt is None: limit_mm_per_prompt = {"image": 5, "video": 0} if bucket_config is None: bucket_config = {(32, 32, 1): 0.5, (52, 64, 1): 0.5} return dataset.sample( tokenizer=tokenizer, num_requests=num_requests, prefix_len=prefix_len, range_ratio=range_ratio, input_len=input_len, output_len=output_len, base_items_per_request=base_items_per_request, num_mm_items_range_ratio=num_mm_items_range_ratio, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, enable_multimodal_chat=enable_multimodal_chat, ) @pytest.mark.benchmark def test_random_mm_same_seed(hf_tokenizer: PreTrainedTokenizerBase) -> None: seed = 42 ds_a = RandomMultiModalDataset(random_seed=seed) ds_b = RandomMultiModalDataset(random_seed=seed) a = _collect_mm_samples(ds_a, hf_tokenizer) b = _collect_mm_samples(ds_b, hf_tokenizer) fa = [_mm_fingerprint_sample(s) for s in a] fb = [_mm_fingerprint_sample(s) for s in b] assert fa == fb @pytest.mark.benchmark def test_random_mm_different_seeds( hf_tokenizer: PreTrainedTokenizerBase, ) -> None: ds_a = RandomMultiModalDataset(random_seed=0) ds_b = RandomMultiModalDataset(random_seed=999) a = _collect_mm_samples(ds_a, hf_tokenizer) b = _collect_mm_samples(ds_b, hf_tokenizer) fa = [_mm_fingerprint_sample(s) for s in a] fb = [_mm_fingerprint_sample(s) for s in b] assert fa != fb @pytest.mark.benchmark def test_random_mm_respects_limits( hf_tokenizer: PreTrainedTokenizerBase, ) -> None: ds = RandomMultiModalDataset(random_seed=0) # Requesting 3 items with a per-prompt limit of 1 should error per current # design (dataset refuses to silently clamp below the requested baseline). with pytest.raises(ValueError): _collect_mm_samples( ds, hf_tokenizer, num_requests=12, base_items_per_request=3, num_mm_items_range_ratio=0.0, limit_mm_per_prompt={"image": 1, "video": 0}, bucket_config={(32, 32, 1): 1.0}, ) @pytest.mark.benchmark def test_random_mm_zero_prob_entries_are_removed( hf_tokenizer: PreTrainedTokenizerBase, ) -> None: ds = RandomMultiModalDataset(random_seed=0) # Second bucket has zero probability and should be ignored after # normalization samples = _collect_mm_samples( ds, hf_tokenizer, num_requests=6, base_items_per_request=2, num_mm_items_range_ratio=0.0, limit_mm_per_prompt={"image": 10, "video": 0}, bucket_config={(32, 32, 1): 1.0, (52, 64, 1): 0.0}, ) for s in samples: assert isinstance(s.multi_modal_data, list) typed_mm = cast(list[dict[str, Any]], s.multi_modal_data) for it in typed_mm: assert it.get("type") == "image_url" @pytest.mark.benchmark def test_random_mm_zero_items(hf_tokenizer: PreTrainedTokenizerBase) -> None: ds = RandomMultiModalDataset(random_seed=0) samples = _collect_mm_samples( ds, hf_tokenizer, num_requests=5, base_items_per_request=0, num_mm_items_range_ratio=0.0, limit_mm_per_prompt={"image": 5, "video": 0}, bucket_config={(32, 32, 1): 1.0}, ) for s in samples: assert s.multi_modal_data == [] @pytest.mark.benchmark def test_random_mm_num_items_per_prompt(hf_tokenizer: PreTrainedTokenizerBase) -> None: ds = RandomMultiModalDataset(random_seed=0) # Fixed number of images per prompt # set num_mm_items_range_ratio to 0.0 # TODO: modify video values when video sampling is implemented samples_fixed_items = _collect_mm_samples( ds, hf_tokenizer, num_requests=5, base_items_per_request=3, num_mm_items_range_ratio=0.0, limit_mm_per_prompt={"image": 3, "video": 0}, bucket_config={(32, 32, 1): 1.0}, ) # Must have 5 requests each with 3 mm items per prompt assert len(samples_fixed_items) == 5 for s in samples_fixed_items: mm_data = cast(list[dict[str, Any]], s.multi_modal_data) assert len(mm_data) == 3 for it in mm_data: assert it.get("type") == "image_url" @pytest.mark.benchmark def test_random_mm_bucket_config_not_mutated( hf_tokenizer: PreTrainedTokenizerBase, ) -> None: ds = RandomMultiModalDataset(random_seed=0) # This bucket config is not normalized to sum to 1 # and has more buckets than requested images original = {(32, 32, 1): 0.2, (52, 64, 1): 6, (25, 64, 1): 3} # Keep a snapshot to compare after sampling snapshot = dict(original) _ = _collect_mm_samples( ds, hf_tokenizer, num_requests=4, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt={"image": 1, "video": 0}, bucket_config=original, ) # Ensure the original dict content is unchanged assert original == snapshot # Vary number of mm items per prompt # set num_mm_items_range_ratio to 0.5 samples_varying_items = _collect_mm_samples( ds, hf_tokenizer, num_requests=5, base_items_per_request=2, num_mm_items_range_ratio=0.5, limit_mm_per_prompt={"image": 4, "video": 0}, bucket_config={(32, 32, 1): 1.0}, ) # Must have 5 requests each with less than 4 mm items per prompt # but at least 1 mm item per prompt assert len(samples_varying_items) == 5 for s in samples_varying_items: mm_data = cast(list[dict[str, Any]], s.multi_modal_data) assert len(mm_data) <= 4 assert len(mm_data) >= 1 for it in mm_data: assert it.get("type") == "image_url" @pytest.mark.benchmark def test_random_mm_video_sampling(hf_tokenizer: PreTrainedTokenizerBase) -> None: """Test video sampling functionality in RandomMultiModalDataset.""" ds = RandomMultiModalDataset(random_seed=42) # Test with video bucket configuration bucket_config = { (64, 64, 1): 0.3, # Images (64, 64, 8): 0.7, # Videos } limit_mm_per_prompt = {"image": 2, "video": 2} samples = _collect_mm_samples( ds, hf_tokenizer, num_requests=5, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, ) assert len(samples) == 5 # Check that we have both images and videos video_count = 0 image_count = 0 for s in samples: mm_data = cast(list[dict[str, Any]], s.multi_modal_data) assert len(mm_data) == 1 item = mm_data[0] if item.get("type") == "video_url": video_count += 1 # Verify video URL format url = item.get("video_url", {}).get("url", "") assert url.startswith("data:video/mp4;base64,") elif item.get("type") == "image_url": image_count += 1 # Verify image URL format url = item.get("image_url", {}).get("url", "") assert url.startswith("data:image/jpeg;base64,") # Should have some videos due to 0.7 probability assert video_count > 0 assert image_count > 0 @pytest.mark.benchmark def test_random_mm_video_only_sampling(hf_tokenizer: PreTrainedTokenizerBase) -> None: """Test sampling with only video buckets.""" ds = RandomMultiModalDataset(random_seed=42) bucket_config = { (64, 64, 8): 1.0, # Only videos } limit_mm_per_prompt = {"image": 0, "video": 1} samples = _collect_mm_samples( ds, hf_tokenizer, num_requests=3, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, ) assert len(samples) == 3 for s in samples: mm_data = cast(list[dict[str, Any]], s.multi_modal_data) assert len(mm_data) == 1 item = mm_data[0] assert item.get("type") == "video_url" url = item.get("video_url", {}).get("url", "") assert url.startswith("data:video/mp4;base64,") @pytest.mark.benchmark def test_random_mm_video_deterministic_sampling( hf_tokenizer: PreTrainedTokenizerBase, ) -> None: """Test that video sampling is deterministic with same seed.""" seed = 123 ds_a = RandomMultiModalDataset(random_seed=seed) ds_b = RandomMultiModalDataset(random_seed=seed) bucket_config = { (64, 64, 8): 1.0, # Only videos } limit_mm_per_prompt = {"image": 0, "video": 1} a = _collect_mm_samples( ds_a, hf_tokenizer, num_requests=3, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, ) b = _collect_mm_samples( ds_b, hf_tokenizer, num_requests=3, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, ) fa = [_mm_fingerprint_sample(s) for s in a] fb = [_mm_fingerprint_sample(s) for s in b] assert fa == fb
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/benchmarks/test_throughput_cli.py
tests/benchmarks/test_throughput_cli.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import subprocess import pytest MODEL_NAME = "meta-llama/Llama-3.2-1B-Instruct" @pytest.mark.benchmark def test_bench_throughput(): command = [ "vllm", "bench", "throughput", "--model", MODEL_NAME, "--input-len", "32", "--output-len", "1", "--enforce-eager", "--load-format", "dummy", ] result = subprocess.run(command, capture_output=True, text=True) print(result.stdout) print(result.stderr) assert result.returncode == 0, f"Benchmark failed: {result.stderr}"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/benchmarks/__init__.py
tests/benchmarks/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/benchmarks/test_param_sweep.py
tests/benchmarks/test_param_sweep.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json import tempfile from pathlib import Path import pytest from vllm.benchmarks.sweep.param_sweep import ParameterSweep, ParameterSweepItem class TestParameterSweepItem: """Test ParameterSweepItem functionality.""" @pytest.mark.parametrize( "input_dict,expected", [ ( {"compilation_config.use_inductor_graph_partition": False}, "--compilation-config.use_inductor_graph_partition=false", ), ( {"compilation_config.use_inductor_graph_partition": True}, "--compilation-config.use_inductor_graph_partition=true", ), ], ) def test_nested_boolean_params(self, input_dict, expected): """Test that nested boolean params use =true/false syntax.""" item = ParameterSweepItem.from_record(input_dict) cmd = item.apply_to_cmd(["vllm", "serve", "model"]) assert expected in cmd @pytest.mark.parametrize( "input_dict,expected", [ ({"enable_prefix_caching": False}, "--no-enable-prefix-caching"), ({"enable_prefix_caching": True}, "--enable-prefix-caching"), ({"disable_log_stats": False}, "--no-disable-log-stats"), ({"disable_log_stats": True}, "--disable-log-stats"), ], ) def test_non_nested_boolean_params(self, input_dict, expected): """Test that non-nested boolean params use --no- prefix.""" item = ParameterSweepItem.from_record(input_dict) cmd = item.apply_to_cmd(["vllm", "serve", "model"]) assert expected in cmd @pytest.mark.parametrize( "compilation_config", [ {"cudagraph_mode": "full", "mode": 2, "use_inductor_graph_partition": True}, { "cudagraph_mode": "piecewise", "mode": 3, "use_inductor_graph_partition": False, }, ], ) def test_nested_dict_value(self, compilation_config): """Test that nested dict values are serialized as JSON.""" item = ParameterSweepItem.from_record( {"compilation_config": compilation_config} ) cmd = item.apply_to_cmd(["vllm", "serve", "model"]) assert "--compilation-config" in cmd # The dict should be JSON serialized idx = cmd.index("--compilation-config") assert json.loads(cmd[idx + 1]) == compilation_config @pytest.mark.parametrize( "input_dict,expected_key,expected_value", [ ({"model": "test-model"}, "--model", "test-model"), ({"max_tokens": 100}, "--max-tokens", "100"), ({"temperature": 0.7}, "--temperature", "0.7"), ], ) def test_string_and_numeric_values(self, input_dict, expected_key, expected_value): """Test that string and numeric values are handled correctly.""" item = ParameterSweepItem.from_record(input_dict) cmd = item.apply_to_cmd(["vllm", "serve"]) assert expected_key in cmd assert expected_value in cmd @pytest.mark.parametrize( "input_dict,expected_key,key_idx_offset", [ ({"max_tokens": 200}, "--max-tokens", 1), ({"enable_prefix_caching": False}, "--no-enable-prefix-caching", 0), ], ) def test_replace_existing_parameter(self, input_dict, expected_key, key_idx_offset): """Test that existing parameters in cmd are replaced.""" item = ParameterSweepItem.from_record(input_dict) if key_idx_offset == 1: # Key-value pair cmd = item.apply_to_cmd(["vllm", "serve", "--max-tokens", "100", "model"]) assert expected_key in cmd idx = cmd.index(expected_key) assert cmd[idx + 1] == "200" assert "100" not in cmd else: # Boolean flag cmd = item.apply_to_cmd( ["vllm", "serve", "--enable-prefix-caching", "model"] ) assert expected_key in cmd assert "--enable-prefix-caching" not in cmd class TestParameterSweep: """Test ParameterSweep functionality.""" def test_from_records_list(self): """Test creating ParameterSweep from a list of records.""" records = [ {"max_tokens": 100, "temperature": 0.7}, {"max_tokens": 200, "temperature": 0.9}, ] sweep = ParameterSweep.from_records(records) assert len(sweep) == 2 assert sweep[0]["max_tokens"] == 100 assert sweep[1]["max_tokens"] == 200 def test_read_from_dict(self): """Test creating ParameterSweep from a dict format.""" data = { "experiment1": {"max_tokens": 100, "temperature": 0.7}, "experiment2": {"max_tokens": 200, "temperature": 0.9}, } sweep = ParameterSweep.read_from_dict(data) assert len(sweep) == 2 # Check that items have the _benchmark_name field names = {item["_benchmark_name"] for item in sweep} assert names == {"experiment1", "experiment2"} # Check that parameters are preserved for item in sweep: if item["_benchmark_name"] == "experiment1": assert item["max_tokens"] == 100 assert item["temperature"] == 0.7 elif item["_benchmark_name"] == "experiment2": assert item["max_tokens"] == 200 assert item["temperature"] == 0.9 def test_read_json_list_format(self): """Test reading JSON file with list format.""" records = [ {"max_tokens": 100, "temperature": 0.7}, {"max_tokens": 200, "temperature": 0.9}, ] with tempfile.NamedTemporaryFile(mode="w", suffix=".json", delete=False) as f: json.dump(records, f) temp_path = Path(f.name) try: sweep = ParameterSweep.read_json(temp_path) assert len(sweep) == 2 assert sweep[0]["max_tokens"] == 100 assert sweep[1]["max_tokens"] == 200 finally: temp_path.unlink() def test_read_json_dict_format(self): """Test reading JSON file with dict format.""" data = { "experiment1": {"max_tokens": 100, "temperature": 0.7}, "experiment2": {"max_tokens": 200, "temperature": 0.9}, } with tempfile.NamedTemporaryFile(mode="w", suffix=".json", delete=False) as f: json.dump(data, f) temp_path = Path(f.name) try: sweep = ParameterSweep.read_json(temp_path) assert len(sweep) == 2 # Check that items have the _benchmark_name field names = {item["_benchmark_name"] for item in sweep} assert names == {"experiment1", "experiment2"} finally: temp_path.unlink() def test_unique_benchmark_names_validation(self): """Test that duplicate _benchmark_name values raise an error.""" # Test with duplicate names in list format records = [ {"_benchmark_name": "exp1", "max_tokens": 100}, {"_benchmark_name": "exp1", "max_tokens": 200}, ] with pytest.raises(ValueError, match="Duplicate _benchmark_name values"): ParameterSweep.from_records(records) def test_unique_benchmark_names_multiple_duplicates(self): """Test validation with multiple duplicate names.""" records = [ {"_benchmark_name": "exp1", "max_tokens": 100}, {"_benchmark_name": "exp1", "max_tokens": 200}, {"_benchmark_name": "exp2", "max_tokens": 300}, {"_benchmark_name": "exp2", "max_tokens": 400}, ] with pytest.raises(ValueError, match="Duplicate _benchmark_name values"): ParameterSweep.from_records(records) def test_no_benchmark_names_allowed(self): """Test that records without _benchmark_name are allowed.""" records = [ {"max_tokens": 100, "temperature": 0.7}, {"max_tokens": 200, "temperature": 0.9}, ] sweep = ParameterSweep.from_records(records) assert len(sweep) == 2 def test_mixed_benchmark_names_allowed(self): """Test that mixing records with and without _benchmark_name is allowed.""" records = [ {"_benchmark_name": "exp1", "max_tokens": 100}, {"max_tokens": 200, "temperature": 0.9}, ] sweep = ParameterSweep.from_records(records) assert len(sweep) == 2 class TestParameterSweepItemKeyNormalization: """Test key normalization in ParameterSweepItem.""" def test_underscore_to_hyphen_conversion(self): """Test that underscores are converted to hyphens in CLI.""" item = ParameterSweepItem.from_record({"max_tokens": 100}) cmd = item.apply_to_cmd(["vllm", "serve"]) assert "--max-tokens" in cmd def test_nested_key_preserves_suffix(self): """Test that nested keys preserve the suffix format.""" # The suffix after the dot should preserve underscores item = ParameterSweepItem.from_record( {"compilation_config.some_nested_param": "value"} ) cmd = item.apply_to_cmd(["vllm", "serve"]) # The prefix (compilation_config) gets converted to hyphens, # but the suffix (some_nested_param) is preserved assert any("compilation-config.some_nested_param" in arg for arg in cmd)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/benchmarks/test_random_multimodal_dataset_video.py
tests/benchmarks/test_random_multimodal_dataset_video.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import base64 import os from tempfile import NamedTemporaryFile from typing import Any, cast import cv2 import pytest from transformers import AutoTokenizer, PreTrainedTokenizerBase from vllm.benchmarks.datasets import RandomMultiModalDataset, SampleRequest @pytest.fixture(scope="session") def hf_tokenizer() -> PreTrainedTokenizerBase: """Use a small, commonly available tokenizer.""" return AutoTokenizer.from_pretrained("gpt2") @pytest.fixture def video_dataset() -> RandomMultiModalDataset: """Create a RandomMultiModalDataset instance for testing.""" return RandomMultiModalDataset(random_seed=42) @pytest.mark.benchmark def test_generate_synthetic_video_different_seeds(): """Test that different seeds produce different videos.""" dataset1 = RandomMultiModalDataset(random_seed=123) dataset2 = RandomMultiModalDataset(random_seed=456) width, height, num_frames = 64, 48, 8 video1 = dataset1.generate_synthetic_video(width, height, num_frames) video2 = dataset2.generate_synthetic_video(width, height, num_frames) # Videos should be different due to different seeds assert video1["bytes"] != video2["bytes"] @pytest.mark.benchmark def test_map_config_to_modality(video_dataset: RandomMultiModalDataset): """Test modality mapping for different configurations.""" # Test image configuration (num_frames = 1) assert video_dataset.map_config_to_modality((256, 256, 1)) == "image" assert video_dataset.map_config_to_modality((720, 1280, 1)) == "image" # Test video configurations (num_frames > 1) assert video_dataset.map_config_to_modality((256, 256, 8)) == "video" assert video_dataset.map_config_to_modality((720, 1280, 16)) == "video" assert video_dataset.map_config_to_modality((64, 64, 32)) == "video" # Test invalid configurations with pytest.raises(ValueError, match="Invalid multimodal item configuration"): video_dataset.map_config_to_modality((256, 256, 0)) with pytest.raises(ValueError, match="Invalid multimodal item configuration"): video_dataset.map_config_to_modality((256, 256, -1)) @pytest.mark.benchmark def test_generate_mm_item_video(video_dataset: RandomMultiModalDataset): """Test generating multimodal items for video configurations.""" # Test video item generation video_config = (64, 48, 8) # height, width, num_frames result = video_dataset.generate_mm_item(video_config) # Check the result structure matches OpenAI API format assert isinstance(result, dict) assert result["type"] == "video_url" assert "video_url" in result assert "url" in result["video_url"] # Check that the URL is a data URL with base64 encoded video url = result["video_url"]["url"] assert url.startswith("data:video/mp4;base64,") # Decode and verify the video content base64_data = url.split(",")[1] video_bytes = base64.b64decode(base64_data) assert len(video_bytes) > 0 # Verify the video can be decoded with NamedTemporaryFile(suffix=".mp4", delete=False) as temp_file: temp_path = temp_file.name temp_file.write(video_bytes) try: cap = cv2.VideoCapture(temp_path) assert cap.isOpened() frame_count = int(cap.get(cv2.CAP_PROP_FRAME_COUNT)) frame_width = int(cap.get(cv2.CAP_PROP_FRAME_WIDTH)) frame_height = int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT)) assert frame_count == 8 assert frame_width == 48 assert frame_height == 64 cap.release() finally: if os.path.exists(temp_path): os.unlink(temp_path) @pytest.mark.benchmark def test_generate_mm_item_image(video_dataset: RandomMultiModalDataset): """Test generating multimodal items for image configurations.""" # Test image item generation image_config = (64, 48, 1) # height, width, num_frames=1 result = video_dataset.generate_mm_item(image_config) # Check the result structure matches OpenAI API format assert isinstance(result, dict) assert result["type"] == "image_url" assert "image_url" in result assert "url" in result["image_url"] # Check that the URL is a data URL with base64 encoded image url = result["image_url"]["url"] assert url.startswith("data:image/jpeg;base64,") @pytest.mark.benchmark def test_generate_mm_item_invalid_config(video_dataset: RandomMultiModalDataset): """Test error handling for invalid configurations.""" with pytest.raises(ValueError, match="Invalid multimodal item configuration"): video_dataset.generate_mm_item((256, 256, 0)) @pytest.mark.benchmark def test_sample_with_video_buckets( video_dataset: RandomMultiModalDataset, hf_tokenizer: PreTrainedTokenizerBase ): """Test sampling with video bucket configurations.""" # Configure bucket with video probability > 0 bucket_config = { (64, 64, 1): 0.3, # Images (64, 64, 8): 0.7, # Videos } limit_mm_per_prompt = {"image": 5, "video": 3} samples = video_dataset.sample( tokenizer=hf_tokenizer, num_requests=5, base_items_per_request=2, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, input_len=20, output_len=5, ) assert len(samples) == 5 # Check that samples contain both images and videos video_count = 0 image_count = 0 for sample in samples: assert isinstance(sample, SampleRequest) assert sample.multi_modal_data is not None assert isinstance(sample.multi_modal_data, list) mm_data = cast(list[dict[str, Any]], sample.multi_modal_data) assert len(mm_data) == 2 # base_items_per_request for item in mm_data: if item["type"] == "video_url": video_count += 1 # Verify video URL format url = item["video_url"]["url"] assert url.startswith("data:video/mp4;base64,") elif item["type"] == "image_url": image_count += 1 # Verify image URL format url = item["image_url"]["url"] assert url.startswith("data:image/jpeg;base64,") # Should have some videos due to 0.7 probability assert video_count > 0 assert image_count > 0 @pytest.mark.benchmark def test_sample_video_only_buckets( video_dataset: RandomMultiModalDataset, hf_tokenizer: PreTrainedTokenizerBase ): """Test sampling with only video buckets.""" bucket_config = { (64, 64, 8): 1.0, # Only videos } limit_mm_per_prompt = {"image": 0, "video": 2} samples = video_dataset.sample( tokenizer=hf_tokenizer, num_requests=3, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, input_len=20, output_len=5, ) assert len(samples) == 3 for sample in samples: assert isinstance(sample, SampleRequest) assert sample.multi_modal_data is not None assert isinstance(sample.multi_modal_data, list) mm_data = cast(list[dict[str, Any]], sample.multi_modal_data) assert len(mm_data) == 1 item = mm_data[0] assert item["type"] == "video_url" url = item["video_url"]["url"] assert url.startswith("data:video/mp4;base64,") @pytest.mark.benchmark def test_sample_respects_video_limits( video_dataset: RandomMultiModalDataset, hf_tokenizer: PreTrainedTokenizerBase ): """Test that sampling respects video limits per prompt.""" bucket_config = { (64, 64, 8): 1.0, # Only videos } # Set very low video limit limit_mm_per_prompt = {"image": 0, "video": 1} samples = video_dataset.sample( tokenizer=hf_tokenizer, num_requests=3, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, input_len=20, output_len=5, ) assert len(samples) == 3 for sample in samples: mm_data = cast(list[dict[str, Any]], sample.multi_modal_data) assert len(mm_data) <= 1 # Should respect video limit @pytest.mark.benchmark def test_sample_mixed_buckets_with_zero_probability( video_dataset: RandomMultiModalDataset, hf_tokenizer: PreTrainedTokenizerBase ): """Test sampling with mixed buckets including zero probability entries.""" bucket_config = { (64, 64, 1): 0.5, # Images (64, 64, 8): 0.5, # Videos (128, 128, 16): 0.0, # Zero probability videos (should be ignored) } limit_mm_per_prompt = {"image": 2, "video": 2} samples = video_dataset.sample( tokenizer=hf_tokenizer, num_requests=4, base_items_per_request=2, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, input_len=20, output_len=5, ) assert len(samples) == 4 # Should only see 64x64 videos, not 128x128 videos for sample in samples: mm_data = cast(list[dict[str, Any]], sample.multi_modal_data) for item in mm_data: if item["type"] == "video_url": # Decode video to verify dimensions url = item["video_url"]["url"] base64_data = url.split(",")[1] video_bytes = base64.b64decode(base64_data) with NamedTemporaryFile(suffix=".mp4", delete=False) as temp_file: # noqa temp_path = temp_file.name temp_file.write(video_bytes) try: cap = cv2.VideoCapture(temp_path) frame_width = int(cap.get(cv2.CAP_PROP_FRAME_WIDTH)) frame_height = int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT)) cap.release() # Should be 64x64, not 128x128 assert frame_width == 64 assert frame_height == 64 finally: if os.path.exists(temp_path): os.unlink(temp_path) @pytest.mark.benchmark def test_sample_deterministic_with_videos(hf_tokenizer: PreTrainedTokenizerBase): """Test that sampling with videos is deterministic with same seed.""" dataset1 = RandomMultiModalDataset(random_seed=123) dataset2 = RandomMultiModalDataset(random_seed=123) bucket_config = { (64, 64, 1): 0.3, # Images (64, 64, 8): 0.7, # Videos } limit_mm_per_prompt = {"image": 2, "video": 2} samples1 = dataset1.sample( tokenizer=hf_tokenizer, num_requests=3, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, input_len=20, output_len=5, ) samples2 = dataset2.sample( tokenizer=hf_tokenizer, num_requests=3, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, input_len=20, output_len=5, ) assert len(samples1) == len(samples2) # Compare multimodal data for s1, s2 in zip(samples1, samples2): assert s1.multi_modal_data == s2.multi_modal_data @pytest.mark.benchmark def test_sample_different_seeds_produce_different_videos( hf_tokenizer: PreTrainedTokenizerBase, ): """Test that different seeds produce different video content.""" dataset1 = RandomMultiModalDataset(random_seed=123) dataset2 = RandomMultiModalDataset(random_seed=456) bucket_config = { (64, 64, 8): 1.0, # Only videos } limit_mm_per_prompt = {"image": 0, "video": 1} samples1 = dataset1.sample( tokenizer=hf_tokenizer, num_requests=2, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, input_len=20, output_len=5, ) samples2 = dataset2.sample( tokenizer=hf_tokenizer, num_requests=2, base_items_per_request=1, num_mm_items_range_ratio=0.0, limit_mm_per_prompt=limit_mm_per_prompt, bucket_config=bucket_config, input_len=20, output_len=5, ) # Video content should be different for s1, s2 in zip(samples1, samples2): mm_data1 = cast(list[dict[str, Any]], s1.multi_modal_data) mm_data2 = cast(list[dict[str, Any]], s2.multi_modal_data) assert len(mm_data1) == len(mm_data2) == 1 url1 = mm_data1[0]["video_url"]["url"] url2 = mm_data2[0]["video_url"]["url"] assert url1 != url2 # Different video content
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/benchmarks/test_plot_filters.py
tests/benchmarks/test_plot_filters.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pandas as pd import pytest from vllm.benchmarks.sweep.plot import ( PlotEqualTo, PlotFilterBase, PlotFilters, PlotGreaterThan, PlotGreaterThanOrEqualTo, PlotLessThan, PlotLessThanOrEqualTo, PlotNotEqualTo, ) class TestPlotFilters: """Test PlotFilter functionality including 'inf' edge case.""" def setup_method(self): """Create sample DataFrames for testing.""" # DataFrame with numeric values self.df_numeric = pd.DataFrame( { "request_rate": [1.0, 5.0, 10.0, 50.0, 100.0], "value": [10, 20, 30, 40, 50], } ) # DataFrame with float('inf') - note: string "inf" values are coerced # to float when loading data, so we only test with float('inf') self.df_inf_float = pd.DataFrame( { "request_rate": [1.0, 5.0, 10.0, float("inf"), float("inf")], "value": [10, 20, 30, 40, 50], } ) @pytest.mark.parametrize( "target,expected_count", [ ("5.0", 1), ("10.0", 1), ("1.0", 1), ], ) def test_equal_to_numeric(self, target, expected_count): """Test PlotEqualTo with numeric values.""" filter_obj = PlotEqualTo("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count def test_equal_to_inf_float(self): """Test PlotEqualTo with float('inf').""" filter_obj = PlotEqualTo("request_rate", "inf") result = filter_obj.apply(self.df_inf_float) # Should match both float('inf') entries because float('inf') == float('inf') assert len(result) == 2 @pytest.mark.parametrize( "target,expected_count", [ ("5.0", 4), # All except 5.0 ("1.0", 4), # All except 1.0 ], ) def test_not_equal_to_numeric(self, target, expected_count): """Test PlotNotEqualTo with numeric values.""" filter_obj = PlotNotEqualTo("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count def test_not_equal_to_inf_float(self): """Test PlotNotEqualTo with float('inf').""" filter_obj = PlotNotEqualTo("request_rate", "inf") result = filter_obj.apply(self.df_inf_float) # Should exclude float('inf') entries assert len(result) == 3 @pytest.mark.parametrize( "target,expected_count", [ ("10.0", 2), # 1.0, 5.0 ("50.0", 3), # 1.0, 5.0, 10.0 ("5.0", 1), # 1.0 ], ) def test_less_than(self, target, expected_count): """Test PlotLessThan with numeric values.""" filter_obj = PlotLessThan("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count @pytest.mark.parametrize( "target,expected_count", [ ("10.0", 3), # 1.0, 5.0, 10.0 ("5.0", 2), # 1.0, 5.0 ], ) def test_less_than_or_equal_to(self, target, expected_count): """Test PlotLessThanOrEqualTo with numeric values.""" filter_obj = PlotLessThanOrEqualTo("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count @pytest.mark.parametrize( "target,expected_count", [ ("10.0", 2), # 50.0, 100.0 ("5.0", 3), # 10.0, 50.0, 100.0 ], ) def test_greater_than(self, target, expected_count): """Test PlotGreaterThan with numeric values.""" filter_obj = PlotGreaterThan("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count @pytest.mark.parametrize( "target,expected_count", [ ("10.0", 3), # 10.0, 50.0, 100.0 ("5.0", 4), # 5.0, 10.0, 50.0, 100.0 ], ) def test_greater_than_or_equal_to(self, target, expected_count): """Test PlotGreaterThanOrEqualTo with numeric values.""" filter_obj = PlotGreaterThanOrEqualTo("request_rate", target) result = filter_obj.apply(self.df_numeric) assert len(result) == expected_count @pytest.mark.parametrize( "filter_str,expected_var,expected_target,expected_type", [ ("request_rate==5.0", "request_rate", "5.0", PlotEqualTo), ("request_rate!=10.0", "request_rate", "10.0", PlotNotEqualTo), ("request_rate<50.0", "request_rate", "50.0", PlotLessThan), ("request_rate<=50.0", "request_rate", "50.0", PlotLessThanOrEqualTo), ("request_rate>10.0", "request_rate", "10.0", PlotGreaterThan), ("request_rate>=10.0", "request_rate", "10.0", PlotGreaterThanOrEqualTo), ("request_rate==inf", "request_rate", "inf", PlotEqualTo), ("request_rate!='inf'", "request_rate", "inf", PlotNotEqualTo), ], ) def test_parse_str(self, filter_str, expected_var, expected_target, expected_type): """Test parsing filter strings.""" filter_obj = PlotFilterBase.parse_str(filter_str) assert isinstance(filter_obj, expected_type) assert filter_obj.var == expected_var assert filter_obj.target == expected_target def test_parse_str_inf_edge_case(self): """Test parsing 'inf' string in filter.""" filter_obj = PlotFilterBase.parse_str("request_rate==inf") assert isinstance(filter_obj, PlotEqualTo) assert filter_obj.var == "request_rate" assert filter_obj.target == "inf" def test_parse_multiple_filters(self): """Test parsing multiple filters.""" filters = PlotFilters.parse_str("request_rate>5.0,value<=40") assert len(filters) == 2 assert isinstance(filters[0], PlotGreaterThan) assert isinstance(filters[1], PlotLessThanOrEqualTo) def test_parse_empty_filter(self): """Test parsing empty filter string.""" filters = PlotFilters.parse_str("") assert len(filters) == 0
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/benchmarks/test_latency_cli.py
tests/benchmarks/test_latency_cli.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import subprocess import pytest MODEL_NAME = "meta-llama/Llama-3.2-1B-Instruct" @pytest.mark.benchmark def test_bench_latency(): command = [ "vllm", "bench", "latency", "--model", MODEL_NAME, "--input-len", "32", "--output-len", "1", "--enforce-eager", "--load-format", "dummy", ] result = subprocess.run(command, capture_output=True, text=True) print(result.stdout) print(result.stderr) assert result.returncode == 0, f"Benchmark failed: {result.stderr}"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/tools/test_config_validator.py
tests/tools/test_config_validator.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import ast import pytest from tools.pre_commit.validate_config import validate_ast _TestConfig1 = """ @config class _TestConfig1: pass """ _TestConfig2 = ''' @config @dataclass class _TestConfig2: a: int """docstring""" ''' _TestConfig3 = """ @config @dataclass class _TestConfig3: a: int = 1 """ _TestConfig4 = ''' @config @dataclass class _TestConfig4: a: Union[Literal[1], Literal[2]] = 1 """docstring""" ''' @pytest.mark.parametrize( ("test_config", "expected_error"), [ (_TestConfig1, "must be a dataclass"), (_TestConfig2, "must have a default"), (_TestConfig3, "must have a docstring"), (_TestConfig4, "must use a single Literal"), ], ) def test_config(test_config, expected_error): tree = ast.parse(test_config) with pytest.raises(Exception, match=expected_error): validate_ast(tree)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/tools/__init__.py
tests/tools/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/reference_mxfp4.py
tests/quantization/reference_mxfp4.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch BFLOAT16_EXP_BIAS = 127 BFLOAT16_MANTISSA_BITS = 7 BFLOAT16_EXP_BITS = 8 FLOAT16_EXP_BIAS = 15 FLOAT16_MANTISSA_BITS = 10 FLOAT16_EXP_BITS = 5 FLOAT8_E8M0_MAX_EXP = 127 FLOAT4_EXP_BIAS = 1 FLOAT4_MANTISSA_BITS = 1 FLOAT16_VAL_TO_ADD = 1 << (FLOAT16_MANTISSA_BITS - FLOAT4_MANTISSA_BITS - 1) FLOAT16_SIGN_EXPONENT_MASK = ( (1 << (FLOAT16_EXP_BITS + 1)) - 1 ) << FLOAT16_MANTISSA_BITS BFLOAT16_VAL_TO_ADD = 1 << (BFLOAT16_MANTISSA_BITS - FLOAT4_MANTISSA_BITS - 1) BFLOAT16_SIGN_EXPONENT_MASK = ( (1 << (BFLOAT16_EXP_BITS + 1)) - 1 ) << BFLOAT16_MANTISSA_BITS def e8m0_to_half(scale, half_dtype: torch.dtype): assert scale.dtype == torch.uint8 scale_exp = scale.to(torch.int16) - 127 # This can be implemented with bitwise operations in a proper kernel. scale_half = 2.0 ** (scale_exp.to(torch.float)) return scale_half.to(half_dtype) def upcast_fp4_to_fp16_or_bf16( val, float_dtype: torch.dtype, half_exp_bias: int, half_mantissa_bits: int ): assert val.dtype == torch.uint8 unpacked = torch.zeros( *val.shape[:-1], val.shape[-1] * 2, dtype=torch.uint8, device=val.device ) unpacked[..., 1::2] = (val >> 4) & 0x0F # Extract high 4 bits. unpacked[..., ::2] = val & 0x0F # Extract low 4 bits. # Takes one float4 values represented as b0000xxxx, # and converts it to the corresponding float16 value. sign = unpacked >> 3 exp = (unpacked >> 1) & 3 new_mantissa = unpacked & 1 # if exp == 0 and new_mantissa == 0: # new_exp = 0 # else: # new_exp = exp - FLOAT4_EXP_BIAS + FLOAT16_EXP_BIAS # int8_t works with float16, but may overflow with bfloat16. new_exp = exp - FLOAT4_EXP_BIAS + half_exp_bias # Cast b0000 to 0. in fp16/bf16. new_exp = new_exp * torch.logical_or(exp > 0, new_mantissa > 0) # Cast b0001 to 0.5 in fp16/bf16. new_mantissa = torch.logical_and(new_mantissa, exp > 0) new_mantissa = new_mantissa.to(torch.int32) new_exp = new_exp.to(torch.int32) sign = sign.to(torch.int32) qdq_val = ( (sign << 15) + (new_exp << half_mantissa_bits) + (new_mantissa << (half_mantissa_bits - 1)) ) assert qdq_val.max() <= 65535 assert qdq_val.min() >= 0 qdq_val = qdq_val.to(torch.uint16) result = qdq_val.view(float_dtype) return result def dq_mxfp4_torch( x: torch.Tensor, scale: torch.Tensor, float_dtype: torch.dtype ) -> torch.Tensor: assert x.dtype == torch.uint8 assert scale.dtype == torch.uint8 if float_dtype == torch.float16: half_exp_bias = FLOAT16_EXP_BIAS half_mantissa_bits = FLOAT16_MANTISSA_BITS elif float_dtype == torch.bfloat16: half_exp_bias = BFLOAT16_EXP_BIAS half_mantissa_bits = BFLOAT16_MANTISSA_BITS scale_half = e8m0_to_half(scale, half_dtype=float_dtype) x_half = upcast_fp4_to_fp16_or_bf16( x, float_dtype=float_dtype, half_exp_bias=half_exp_bias, half_mantissa_bits=half_mantissa_bits, ) x_half = x_half.reshape(*x_half.shape[:-1], -1, 32) x_half = x_half * scale_half[..., None] x_half = x_half.reshape(*x_half.shape[:-2], -1) return x_half def fp16_to_fp4_simulate( val, half_mantissa_bits: int, half_exp_bits: int, half_exp_bias: int ): # Casts an fp16/bf16 input to the restricted values of float4_e2m1, # that is to say [0., 0.5, 1.0, 1.5, 2.0, 3.0, 4.0, 6.0, -0.0, # -0.5, -1.0, -1.5, -2.0, -3.0, -4.0, -6.0]. float_type = val.dtype # "rshift_cuda" not implemented for 'UInt16' val_view = val.view(torch.int16) # .to(torch.int32) exp = val_view >> half_mantissa_bits exp = exp & ((1 << half_exp_bits) - 1) exp = exp.view(torch.uint16).to(torch.int32) sign = (val_view >> (half_mantissa_bits + half_exp_bits)) & 1 mantissa_last = (val_view >> (half_mantissa_bits - 1)) & 1 exp_unbias = exp - half_exp_bias new_exp = exp_unbias + FLOAT4_EXP_BIAS exp_shift = (new_exp <= 0) * (1 - new_exp) # Typically 9. # Take the min to prevent overflow on `uint16_t half`. This is the case for # very small values, correctly mapped to `round_close`. tail_bits = half_mantissa_bits - FLOAT4_MANTISSA_BITS + exp_shift tail_bits[tail_bits >= 16] = 16 mantissa_plus_one = val_view & ((1 << (half_mantissa_bits + 1)) - 1) half = 1 << (tail_bits - 1) tail = mantissa_plus_one & ((1 << tail_bits) - 1) round_close = tail < half # round towards 0 round_away = tail > half # round away from 0 tie = tail == half new_mantissa_close = torch.zeros(val.shape, device=val.device, dtype=torch.bool) new_exp_close = torch.zeros(val.shape, device=val.device, dtype=torch.uint16) new_mantissa_away = torch.zeros(val.shape, device=val.device, dtype=torch.bool) new_exp_away = torch.zeros(val.shape, device=val.device, dtype=torch.uint16) new_exp_tie = torch.zeros(val.shape, device=val.device, dtype=torch.uint16) # 1. round down # if new_exp == 0: # case [0.5, 0.749999] # new_mantissa = 0 # elif new_exp < 0: # case [0, 0.24999] # new_mantissa = 0 # else: # new_mantissa = mantissa_last new_mantissa_close = (new_exp > 0) * mantissa_last new_exp_close = exp # # 2. round up # if new_exp <= 0: # case [0.250001, 0.499999] and [0.75001, 0.99999] # new_mantissa = 0 # new_exp += 1 # elif mantissa_last == 0: # new_mantissa = 1 # else: # new_mantissa = 0 # new_exp += 1 new_mantissa_away = torch.logical_and(new_exp > 0, mantissa_last == 0) new_exp_away = exp + torch.logical_or(new_exp <= 0, mantissa_last == 1) # # 3. tie # 0.25 -> 0. (handled by `exp > (half_exp_bias - 2)`) # 0.75 -> 1. # 1.25 -> 1. # 1.75 -> 2. # 2.5 -> 2. # 3.5 -> 4. # 5. -> 4. new_exp_tie = (exp > (half_exp_bias - 2)) * (exp + (mantissa_last == 1)) # Gather round up, round down and tie. new_exp = ( round_away * new_exp_away + round_close * new_exp_close + tie * new_exp_tie ) new_mantissa = round_away * new_mantissa_away + round_close * new_mantissa_close # if new_exp > 3: # new_mantissa = 1 new_mantissa = new_mantissa + (new_exp > (2 + half_exp_bias)) * (new_mantissa == 0) # Clamp the exponent to acceptable values. new_exp = (new_exp >= (half_exp_bias - 2)) * torch.clamp( new_exp, half_exp_bias - 2, half_exp_bias + 2 ) sign = sign.to(torch.int32) new_mantissa = new_mantissa.to(torch.int32) qdq_val = ( (sign << 15) + (new_exp << half_mantissa_bits) + (new_mantissa << (half_mantissa_bits - 1)) ) assert qdq_val.max() <= 65535 assert qdq_val.min() >= 0 assert qdq_val.dtype == torch.int32 qdq_val = qdq_val.to(torch.uint16) result = qdq_val.view(float_type) return result def qdq_mxfp4_torch( x: torch.Tensor, scale_calculation_mode: str = "even" ) -> torch.Tensor: half_dtype = x.dtype if half_dtype == torch.float16: half_mantissa_bits = FLOAT16_MANTISSA_BITS half_exp_bits = FLOAT16_EXP_BITS half_exp_bias = FLOAT16_EXP_BIAS val_to_add = FLOAT16_VAL_TO_ADD sign_exponent_mask = FLOAT16_SIGN_EXPONENT_MASK elif half_dtype == torch.bfloat16: half_mantissa_bits = BFLOAT16_MANTISSA_BITS half_exp_bits = BFLOAT16_EXP_BITS half_exp_bias = BFLOAT16_EXP_BIAS val_to_add = BFLOAT16_VAL_TO_ADD sign_exponent_mask = BFLOAT16_SIGN_EXPONENT_MASK else: raise ValueError("not implemented") x = x.reshape(*x.shape[:-1], -1, 32) block_max = torch.max(torch.abs(x), dim=-1).values block_max = block_max.view(torch.uint16).to(torch.int32) block_max_uint = torch.bitwise_and(block_max + val_to_add, sign_exponent_mask) assert block_max_uint.max() <= 65535 assert block_max_uint.min() >= 0 assert block_max_uint.dtype == torch.int32 block_max_uint = block_max_uint.to(torch.uint16) block_max = block_max_uint.view(half_dtype) scale_exp = ( FLOAT8_E8M0_MAX_EXP + torch.floor(torch.log2(block_max)).to(torch.int32) - 2 ) scale_exp = torch.clamp(scale_exp, 0, 2 * FLOAT8_E8M0_MAX_EXP) scale = 2.0 ** (scale_exp - FLOAT8_E8M0_MAX_EXP) scale = scale.to(half_dtype) x = x / scale[..., None] x_fp4 = fp16_to_fp4_simulate( x, half_exp_bits=half_exp_bits, half_mantissa_bits=half_mantissa_bits, half_exp_bias=half_exp_bias, ) x_fp4 = x_fp4 * scale[..., None] return x_fp4.reshape(*x_fp4.shape[:-2], -1)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_compressed_tensors.py
tests/quantization/test_compressed_tensors.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Test model set-up and weight loading for llmcompressor-quantized models. Run `pytest tests/quantization/test_compressed_tensors.py`. """ import pytest import torch from compressed_tensors.quantization import QuantizationType from tests.models.utils import check_logprobs_close from vllm.model_executor.layers.fused_moe import UnquantizedFusedMoEMethod from vllm.model_executor.layers.quantization.compressed_tensors.compressed_tensors import ( # noqa: E501 CompressedTensors24, CompressedTensorsLinearMethod, CompressedTensorsW4A4Fp4, CompressedTensorsW4A8Fp8, CompressedTensorsW4A16Fp4, CompressedTensorsW4A16Sparse24, CompressedTensorsW8A8Fp8, CompressedTensorsW8A8Int8, CompressedTensorsW8A16Fp8, CompressedTensorsWNA16, ) from vllm.model_executor.layers.quantization.input_quant_fp8 import QuantFP8 from vllm.model_executor.layers.quantization.utils.fp8_utils import W8A8BlockFp8LinearOp from vllm.model_executor.layers.quantization.utils.quant_utils import ( cutlass_fp4_supported, ) from vllm.model_executor.layers.quantization.utils.w8a8_utils import ( sparse_cutlass_supported, ) from vllm.platforms import current_platform # AITER only supports per-channel-per-channel INT8 gemm # and per-tensor-per-tensor INT8 GEMM. # It does not support mix precision MM and mix quantization scheme. ROCM_AITER_SUPPORTED_INT8_MODEL = [ "neuralmagic/Llama-3.2-1B-quantized.w8a8", "nm-testing/tinyllama-oneshot-w8a8-channel-dynamic-token-v2", ] # TritonScaledMMLinearKernel only supports symmetric quantization. ROCM_TRITON_SCALED_MM_SUPPORTED_INT8_MODEL = [ "nm-testing/tinyllama-oneshot-w8w8-test-static-shape-change", "nm-testing/tinyllama-oneshot-w8-channel-a8-tensor", "neuralmagic/Llama-3.2-1B-quantized.w8a8", "nm-testing/tinyllama-oneshot-w8a8-dynamic-token-v2", "nm-testing/tinyllama-oneshot-w8a8-channel-dynamic-token-v2", ] @pytest.fixture(scope="function", autouse=True) def enable_pickle(monkeypatch): """`LLM.apply_model` requires pickling a function.""" monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") @pytest.mark.parametrize( "model_args", [ ( "nm-testing/tinyllama-oneshot-w8w8-test-static-shape-change", "tensor", QuantizationType.INT, 2560, True, ), ( "nm-testing/asym-w8w8-int8-static-per-tensor-tiny-llama", "tensor", QuantizationType.INT, 2560, False, ), ], ) def test_compressed_tensors_w8a8_static_setup(vllm_runner, model_args): model_path, strategy, quant_type, shape_0, is_symmetric = model_args if ( current_platform.is_rocm() and model_path not in ROCM_TRITON_SCALED_MM_SUPPORTED_INT8_MODEL ): pytest.skip(f"Skip model {model_path} as it is not supported on ROCm.") with vllm_runner(model_path, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj o_proj = layer.self_attn.o_proj gate_up_proj = layer.mlp.gate_up_proj down_proj = layer.mlp.down_proj # assert zp for symmetric and asymmetric cases def zp_valid(zp: torch.Tensor | None): if is_symmetric: return zp is None return zp is not None and zp.dtype is torch.int32 assert zp_valid(qkv_proj.input_zero_point) assert zp_valid(o_proj.input_zero_point) assert zp_valid(gate_up_proj.input_zero_point) assert zp_valid(down_proj.input_zero_point) assert isinstance(qkv_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(o_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(gate_up_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(down_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(qkv_proj.scheme, CompressedTensorsW8A8Int8) assert qkv_proj.scheme.strategy == strategy assert qkv_proj.scheme.is_static_input_scheme expected_type = torch.int8 assert qkv_proj.weight.dtype is expected_type assert o_proj.weight.dtype is expected_type assert gate_up_proj.weight.dtype is expected_type if qkv_proj.scheme.strategy == "tensor": # Make sure it is a channelwise buffer # After running process_weights_after_loading assert len(qkv_proj.weight_scale.shape) == 2 assert qkv_proj.weight_scale.shape[0] == shape_0 assert qkv_proj.weight_scale.shape[1] == 1 assert qkv_proj.weight_scale.dtype is torch.float32 assert qkv_proj.input_scale.dtype is torch.float32 llm.apply_model(check_model) output = llm.generate_greedy(["Hello my name is"], max_tokens=4) assert output @pytest.mark.parametrize( "model_path", [ "neuralmagic/Llama-3.2-1B-quantized.w8a8", ], ) @pytest.mark.parametrize("max_tokens", [4]) @pytest.mark.parametrize("num_logprobs", [10]) @pytest.mark.parametrize( "use_aiter", [True, False] if current_platform.is_rocm() else [False] ) def test_compressed_tensors_w8a8_logprobs( hf_runner, vllm_runner, example_prompts, model_path, max_tokens, num_logprobs, use_aiter, monkeypatch, ): if ( current_platform.is_rocm() and model_path not in ROCM_TRITON_SCALED_MM_SUPPORTED_INT8_MODEL ): pytest.skip(f"Skip model {model_path} as it is not supported on ROCm.") if use_aiter: if model_path not in ROCM_AITER_SUPPORTED_INT8_MODEL: pytest.skip(f"Skip model {model_path} as it is not support by aiter.") # this will enable VLLM_ROCM_USE_AITER_LINEAR monkeypatch.setenv("VLLM_ROCM_USE_AITER", "1") dtype = "bfloat16" # skip language translation prompt for the static per tensor models if model_path in ( "nm-testing/Meta-Llama-3-8B-Instruct-W8A8-Static-Per-Tensor-Sym", "nm-testing/Meta-Llama-3-8B-Instruct-W8A8-Static-Per-Tensor-Asym", ): example_prompts = example_prompts[0:-1] with hf_runner(model_path, dtype=dtype) as hf_model: hf_outputs = hf_model.generate_greedy_logprobs_limit( example_prompts, max_tokens, num_logprobs ) with vllm_runner(model_path, dtype=dtype, enforce_eager=True) as vllm_model: vllm_outputs = vllm_model.generate_greedy_logprobs( example_prompts, max_tokens, num_logprobs ) check_logprobs_close( outputs_0_lst=hf_outputs, outputs_1_lst=vllm_outputs, name_0="hf", name_1="vllm", ) if current_platform.is_rocm(): torch.cuda.synchronize() def test_compressed_tensors_no_enforce_eager(vllm_runner): model_path = "nm-testing/tinyllama-oneshot-w8w8-test-static-shape-change" with vllm_runner(model_path) as llm: output = llm.generate_greedy("Hello my name is", max_tokens=4) assert output @pytest.mark.parametrize( "model_args", [ ("nm-testing/tinyllama-oneshot-w8a8-dynamic-token-v2", "tensor"), ( "nm-testing/tinyllama-oneshot-w8a8-channel-dynamic-token-v2", "channel", ), ], ) @pytest.mark.parametrize( "use_aiter", [True, False] if current_platform.is_rocm() else [False] ) def test_compressed_tensors_w8a8_dynamic_per_token( vllm_runner, model_args, use_aiter, monkeypatch, ): model_path, strategy = model_args if ( current_platform.is_rocm() and model_path not in ROCM_TRITON_SCALED_MM_SUPPORTED_INT8_MODEL ): pytest.skip(f"Skip model {model_path} as it is not supported on ROCm.") if use_aiter: if model_path not in ROCM_AITER_SUPPORTED_INT8_MODEL: pytest.skip(f"Skip model {model_path} as it is not support by aiter.") # this will enable VLLM_ROCM_USE_AITER_LINEAR monkeypatch.setenv("VLLM_ROCM_USE_AITER", "1") with vllm_runner(model_path, enforce_eager=True, dtype=torch.float16) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(qkv_proj.scheme, CompressedTensorsW8A8Int8) assert not qkv_proj.scheme.is_static_input_scheme assert qkv_proj.scheme.strategy == strategy assert qkv_proj.weight.dtype is torch.int8 llm.apply_model(check_model) output = llm.generate_greedy(["Hello my name is"], max_tokens=4) assert output @pytest.mark.parametrize( "wNa16_args", [ ( "nm-testing/tinyllama-oneshot-w4a16-channel-v2", "channel", None, 8, True, False, ), ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-W4A16-G128-Asym-Updated-ActOrder", "group", 128, 8, False, True, ), ], ) @pytest.mark.skipif( not current_platform.is_cuda(), reason="The tests are skipped on non-CUDA platform." ) def test_compressed_tensors_wNa16(vllm_runner, wNa16_args): model, strategy, group, pack_factor, symmetric, has_g_idx = wNa16_args with vllm_runner(model, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(qkv_proj.scheme, CompressedTensorsWNA16) assert qkv_proj.scheme.strategy == strategy assert qkv_proj.scheme.group_size == (-1 if group is None else group) assert qkv_proj.scheme.pack_factor == pack_factor assert qkv_proj.scheme.symmetric == symmetric assert qkv_proj.scheme.has_g_idx == has_g_idx llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) assert output @pytest.mark.skipif( not current_platform.is_cuda(), reason="This test is skipped on non-CUDA platform." ) def test_compressed_tensors_w4a16_marlin24(vllm_runner): model_path = "nm-testing/llama7b-one-shot-2_4-w4a16-marlin24-t" with vllm_runner(model_path, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(qkv_proj.scheme, CompressedTensorsW4A16Sparse24) assert qkv_proj.weight_packed.dtype is torch.int32 llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) assert output def test_compressed_tensors_fp8(vllm_runner): model_path = "nm-testing/Meta-Llama-3-8B-FP8-compressed-tensors-test" with vllm_runner(model_path, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance( qkv_proj.scheme, (CompressedTensorsW8A8Fp8, CompressedTensorsW8A16Fp8), ) assert qkv_proj.input_scale.dtype is torch.float32 if isinstance(qkv_proj.scheme, CompressedTensorsW8A8Fp8): assert len(qkv_proj.input_scale.shape) == 0 assert qkv_proj.weight.dtype is current_platform.fp8_dtype() assert qkv_proj.weight_scale.dtype is torch.float32 assert len(qkv_proj.weight_scale.shape) == 0 llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) assert output @pytest.mark.skipif( not current_platform.is_cuda(), reason="This test is skipped on non-CUDA platform." ) def test_compressed_tensors_kv_cache(vllm_runner): model_path = "nm-testing/TinyLlama-1.1B-compressed-tensors-kv-cache-scheme" with vllm_runner(model_path, enforce_eager=True, kv_cache_dtype="fp8") as llm: output = llm.generate_greedy("Hello world!", max_tokens=4) assert output @pytest.mark.skipif( not sparse_cutlass_supported(), reason="Sparse FP8 is not yet supported on this GPU type.", ) def _test_2of4_quant_models(qkv_proj, weight_strategy, input_strategy, format="dense"): assert isinstance(qkv_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(qkv_proj.scheme, CompressedTensors24) assert qkv_proj.scheme.weight_quant.strategy == weight_strategy assert qkv_proj.scheme.input_quant.strategy == input_strategy assert qkv_proj.scheme.quantized assert qkv_proj.quant_method.quantization_config.sparsity_scheme_map sparsity_map = qkv_proj.quant_method.quantization_config.sparsity_scheme_map # noqa: E501 assert sparsity_map.get("Linear").format == format assert sparsity_map.get("Linear").sparsity_structure == "2:4" @pytest.mark.skipif( not current_platform.is_cuda() or not current_platform.has_device_capability(90), reason="Sparse FP8 is not yet supported on this GPU type.", ) @pytest.mark.parametrize( "args_2of4", [ ( "nm-testing/Meta-Llama-3-8B-Instruct-FP8-Dynamic-2of4-testing", "channel", "token", ), ( "nm-testing/Meta-Llama-3-8B-Instruct-FP8-Static-Per-Tensor-testing", "channel", "tensor", ), ( "nm-testing/Meta-Llama-3-8B-Instruct-FP8-Static-testing", "tensor", "tensor", ), ( "nm-testing/Meta-Llama-3-8B-Instruct-FP8-Dynamic-IA-Per-Tensor-Weight-testing", "tensor", "token", ), ], ) def test_compressed_tensors_2of4_quant_fp8(vllm_runner, args_2of4): model, weight_strategy, input_strategy = args_2of4 with vllm_runner(model, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert qkv_proj.scheme.weights_dtype == torch.float8_e4m3fn _test_2of4_quant_models(qkv_proj, weight_strategy, input_strategy) llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) print(output) assert output @pytest.mark.skipif( not current_platform.is_cuda() or not current_platform.has_device_capability(90), reason="Sparse FP8 is not yet supported on this GPU type.", ) @pytest.mark.parametrize( "args_2of4", [ ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-chnl_wts_per_tok_dyn_act_fp8-BitM", "channel", "token", ), ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-chnl_wts_tensor_act_fp8-BitM", "channel", "tensor", ), ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-tensor_wts_per_tok_dyn_act_fp8-BitM", "tensor", "token", ), ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-tensor_wts_tensor_act_fp8-BitM", "tensor", "tensor", ), ], ) def test_compressed_tensors_2of4_quant_fp8_compressed(vllm_runner, args_2of4): model, weight_strategy, input_strategy = args_2of4 with vllm_runner(model, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert qkv_proj.scheme.weights_dtype == torch.float8_e4m3fn _test_2of4_quant_models( qkv_proj, weight_strategy, input_strategy, format="sparse-24-bitmask", ) llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) print(output) assert output @pytest.mark.skipif( not sparse_cutlass_supported(), reason="cutlass is not yet supported on this GPU type.", ) @pytest.mark.parametrize( "args_2of4", [ ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-chnl_wts_per_tok_dyn_act_int8-BitM", "channel", "token", ), ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-chnl_wts_tensor_act_int8-BitM", "channel", "tensor", ), ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-tensor_wts_per_tok_dyn_act_int8-BitM", "tensor", "token", ), ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-tensor_wts_tensor_act_int8-BitM", "tensor", "tensor", ), ], ) def test_compressed_tensors_2of4_quant_int8_compressed(vllm_runner, args_2of4): model, weight_strategy, input_strategy = args_2of4 with vllm_runner(model, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert qkv_proj.scheme.weights_dtype == torch.int8 _test_2of4_quant_models( qkv_proj, weight_strategy, input_strategy, format="sparse-24-bitmask", ) llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) print(output) assert output @pytest.mark.skipif( not sparse_cutlass_supported(), reason="Sparse FP8 is not yet supported on this GPU type.", ) @pytest.mark.parametrize( "args_2of4", [ ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-INT8-Dynamic-IA-Per-Channel-Weight-testing", "channel", "token", ), ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-INT8-Static-testing", "tensor", "tensor", ), ( "nm-testing/TinyLlama-1.1B-Chat-v1.0-INT8-Dynamic-IA-Per-Tensor-Weight-testing", "tensor", "token", ), ], ) def test_compressed_tensors_2of4_quant_int8(vllm_runner, args_2of4): model, weight_strategy, input_strategy = args_2of4 with vllm_runner(model, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert qkv_proj.scheme.weights_dtype == torch.int8 _test_2of4_quant_models(qkv_proj, weight_strategy, input_strategy) llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) print(output) assert output @pytest.mark.skipif( not sparse_cutlass_supported(), reason="2of4 Sparse is not yet supported on this GPU type.", ) @pytest.mark.parametrize( "args_2of4", [("nm-testing/TinyLlama-1.1B-Chat-v1.0-2of4-Sparse-Dense-Compressor")], ) def test_compressed_tensors_2of4_sparse(vllm_runner, args_2of4): model = args_2of4 with vllm_runner(model, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(qkv_proj.scheme, CompressedTensors24) assert qkv_proj.scheme.weight_quant is None assert qkv_proj.scheme.input_quant is None assert not qkv_proj.scheme.quantized assert qkv_proj.quant_method.quantization_config.sparsity_scheme_map sparsity_map = qkv_proj.quant_method.quantization_config.sparsity_scheme_map # noqa: E501 assert sparsity_map.get("Linear").format == "dense" assert sparsity_map.get("Linear").sparsity_structure == "2:4" llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) print(output) assert output @pytest.mark.skipif( not sparse_cutlass_supported(), reason="Cutlass is not yet supported on this GPU type.", ) @pytest.mark.parametrize( "args_2of4", [("nm-testing/llama2.c-stories42M-pruned2.4-compressed")] ) def test_compressed_tensors_2of4_sparse_compressed(vllm_runner, args_2of4): model = args_2of4 with vllm_runner(model, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(qkv_proj.scheme, CompressedTensors24) assert qkv_proj.scheme.weight_quant is None assert qkv_proj.scheme.input_quant is None assert not qkv_proj.scheme.quantized assert qkv_proj.quant_method.quantization_config.sparsity_scheme_map sparsity_map = qkv_proj.quant_method.quantization_config.sparsity_scheme_map # noqa: E501 assert sparsity_map.get("Linear").format == "sparse-24-bitmask" assert sparsity_map.get("Linear").sparsity_structure == "2:4" llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) print(output) assert output @pytest.mark.parametrize( "args", [ # TODO: Enable once model is available again # ("nm-testing/TinyLlama-1.1B-Chat-v1.0-NVFP4A16", CompressedTensorsW4A16Fp4), ("nm-testing/TinyLlama-1.1B-Chat-v1.0-NVFP4", CompressedTensorsW4A4Fp4), ], ) def test_compressed_tensors_nvfp4(vllm_runner, args): model, scheme = args with vllm_runner(model, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, CompressedTensorsLinearMethod) if ( isinstance(qkv_proj.scheme, scheme) or isinstance(qkv_proj.scheme, CompressedTensorsW4A16Fp4) and not cutlass_fp4_supported() ): assert True else: raise AssertionError("FP4 Scheme Mismatch") assert qkv_proj.scheme.group_size == 16 llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) print(output) assert output @pytest.mark.skipif( not current_platform.is_cuda() or not current_platform.has_device_capability(90), reason="W4A8 FP8 is not yet supported on this GPU type.", ) @pytest.mark.parametrize( "args", [("czhu-cohere/TinyLlama-1.1B-Chat-v1.0-W4A8-e2e", CompressedTensorsW4A8Fp8)], ) def test_compressed_tensors_w4a8_fp8(vllm_runner, args): model, scheme = args with vllm_runner(model, enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj o_proj = layer.self_attn.o_proj gate_up_proj = layer.mlp.gate_up_proj down_proj = layer.mlp.down_proj for proj in (qkv_proj, o_proj, gate_up_proj, down_proj): assert isinstance(proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(proj.scheme, scheme) assert proj.weight_packed.dtype is torch.int32 assert proj.weight_scale.dtype is torch.float8_e4m3fn assert proj.weight_chan_scale.dtype is torch.float32 assert proj.scheme.group_size == 128 llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) print(output) assert output @pytest.mark.skipif( not current_platform.is_cuda(), reason="This test is skipped on non-CUDA platform." ) @pytest.mark.parametrize( "model,prompt,exp_perplexity", [ ( "nm-testing/Llama-3.2-1B-Instruct-spinquantR1R2R4-w4a16", "Flat is better than nested.\nSparse is better than dense.", 150.0, ), ( "nm-testing/Llama-3.2-1B-Instruct-quip-w4a16", "Flat is better than nested.\nSparse is better than dense.", 150.0, ), ], ) def test_compressed_tensors_transforms_perplexity( vllm_runner, model, prompt, exp_perplexity ): with vllm_runner(model, enforce_eager=True) as llm: perplexity = llm.generate_prompt_perplexity([prompt])[0] print(perplexity) assert perplexity <= exp_perplexity def test_compressed_tensors_fp8_block_enabled(vllm_runner): model_path = "RedHatAI/Qwen3-0.6B-FP8-BLOCK" with vllm_runner(model_path, enforce_eager=True) as llm: fp8_dtype = current_platform.fp8_dtype() def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, CompressedTensorsLinearMethod) assert isinstance(qkv_proj.scheme, CompressedTensorsW8A8Fp8) assert isinstance( qkv_proj.scheme.w8a8_block_fp8_linear, W8A8BlockFp8LinearOp ) assert qkv_proj.weight.dtype is fp8_dtype assert qkv_proj.weight_scale.dtype is torch.float32 assert len(qkv_proj.weight.shape) == 2 assert len(qkv_proj.weight_scale.shape) == 2 input_quant_op = qkv_proj.scheme.w8a8_block_fp8_linear.input_quant_op assert isinstance(input_quant_op, QuantFP8) assert input_quant_op._forward_method == input_quant_op.forward_cuda llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) assert output @pytest.mark.skipif( not current_platform.is_cuda(), reason="This test is not for non-CUDA platforms", ) def test_compressed_tensors_moe_ignore_with_model(vllm_runner): """ Integration test for MoE layer ignore functionality with a real model. This test would verify that when loading a compressed-tensors quantized MoE model where some MoE layers are in the ignore list, those layers use UnquantizedFusedMoEMethod while non-ignored layers use the quantized method. Expected model structure: - Compressed-tensors quantized MoE model (e.g., Mixtral-based) - Config with ignore list containing specific MoE layers - Multiple MoE layers where some are quantized and some are not """ # model_path = "nm-testing/tinysmokeqwen3moe-W4A16-first-only" # CT 12.3 model_path = "nm-testing/tinysmokeqwen3moe-W4A16-first-only-CTstable" # CT 12.2 with vllm_runner(model_path, enforce_eager=True) as llm: def check_model(model): from vllm.model_executor.layers.fused_moe import FusedMoE from vllm.model_executor.layers.quantization.compressed_tensors.compressed_tensors_moe import ( # noqa: E501 CompressedTensorsMoEMethod, ) # Check layer 0 MoE (should be quantized) layer_quantized = model.model.layers[0].mlp.experts assert isinstance(layer_quantized, FusedMoE) assert isinstance(layer_quantized.quant_method, CompressedTensorsMoEMethod) # Check layer 10 MoE (should be unquantized + ignored) layer_unquantized = model.model.layers[3].mlp.experts assert isinstance(layer_unquantized, FusedMoE) assert isinstance(layer_unquantized.quant_method, UnquantizedFusedMoEMethod) llm.apply_model(check_model) # Verify the model can generate output output = llm.generate_greedy("Hello, my name is", max_tokens=4) assert output
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_lm_head.py
tests/quantization/test_lm_head.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests whether gptq models with quantized lm_head can be loaded. Run `pytest tests/quantization/test_quant_lm_head_true.py --forked`. """ import pytest import torch from vllm.model_executor.layers.quantization.gptq import GPTQLinearMethod from vllm.model_executor.layers.quantization.gptq_marlin import GPTQMarlinLinearMethod from vllm.model_executor.layers.vocab_parallel_embedding import ( UnquantizedEmbeddingMethod, ) PROMPT = "On the surface of Mars, we found" MODELS_QUANT = [ ("ModelCloud/Qwen1.5-1.8B-Chat-GPTQ-4bits-dynamic-cfg-with-lm_head", True), ("TheBloke/TinyLlama-1.1B-Chat-v1.0-GPTQ", False), ] @pytest.mark.parametrize("model_id, lm_head_quantized", MODELS_QUANT) def test_lm_head( vllm_runner, model_id: str, lm_head_quantized: bool, monkeypatch, ) -> None: # `LLM.apply_model` requires pickling a function. monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") with vllm_runner( model_id, dtype=torch.float16, max_model_len=2048, enforce_eager=True ) as vllm_model: def check_model(model): lm_head_layer = model.lm_head if lm_head_quantized: assert isinstance( lm_head_layer.quant_method, (GPTQLinearMethod, GPTQMarlinLinearMethod), ) else: assert isinstance( lm_head_layer.quant_method, UnquantizedEmbeddingMethod ) vllm_model.apply_model(check_model) print(vllm_model.generate_greedy(["Hello my name is"], max_tokens=4)[0][1])
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_auto_round.py
tests/quantization/test_auto_round.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Test model set-up and inference for quantized HF models supported on the AutoRound. Validating the configuration and printing results for manual checking. Run `pytest tests/quantization/test_auto_round.py`. """ import pytest from vllm.platforms import current_platform MODELS = [ "OPEA/Qwen2.5-0.5B-Instruct-int4-sym-inc", ##auto_round:auto_gptq "Intel/Qwen2-0.5B-Instruct-int4-sym-AutoRound", ##auto_round:auto_awq ] @pytest.mark.skipif( not current_platform.is_cpu() and not current_platform.is_xpu() and not current_platform.is_cuda(), reason="only supports CPU/XPU/CUDA backend.", ) @pytest.mark.parametrize("model", MODELS) def test_auto_round(vllm_runner, model): with vllm_runner(model, enforce_eager=True) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=8) assert output print(f"{output[0][1]}")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_register_quantization_config.py
tests/quantization/test_register_quantization_config.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests register custom quantization config. See https://github.com/vllm-project/vllm/issues/11926 for more details. Run `pytest tests/quantization/test_register_quantization_config.py`. """ import logging from typing import Any import pytest import torch import torch.nn.functional as F from vllm.model_executor.layers.linear import ( LinearBase, # noqa: E501 UnquantizedLinearMethod, ) from vllm.model_executor.layers.quantization import ( QuantizationMethods, get_quantization_config, register_quantization_config, ) from vllm.model_executor.layers.quantization.base_config import ( QuantizationConfig, # noqa: E501 ) class FakeQuantLinearMethod(UnquantizedLinearMethod): """Fake quantization linear method for per-token dynamic quantization.""" def __init__(self, num_bits: int = 8) -> None: """Initialize the quantization method.""" super().__init__() self.num_bits = num_bits def apply( self, layer: torch.nn.Module, x: torch.Tensor, bias: torch.Tensor | None = None, ) -> torch.Tensor: """Perform fake quantization before the linear layer.""" # Calculate the scales dynamically max_val = torch.amax(x, dim=(0, -1), keepdims=True) min_val = torch.amin(x, dim=(0, -1), keepdims=True) scales = (max_val - min_val) / (2**self.num_bits - 1) # Fake quantize the input quant_x = torch.clamp( torch.round(x / scales), -(2 ** (self.num_bits - 1)), 2 ** (self.num_bits - 1) - 1, ) dequant_x = quant_x * scales return F.linear(dequant_x, layer.weight, bias) @register_quantization_config("custom_quant") class CustomQuantConfig(QuantizationConfig): """Custom quantization config for per-token dynamic fake quantization.""" def __init__(self, num_bits: int = 8) -> None: """Initialize the quantization config.""" super().__init__() self.num_bits = num_bits def get_name(self) -> QuantizationMethods: """Name of the quantization method.""" return "custom_quant" def get_supported_act_dtypes(self) -> list[torch.dtype]: """List of supported activation dtypes.""" return [torch.float16, torch.bfloat16] @classmethod def get_min_capability(cls) -> int: """Minimum GPU capability to support the quantization method.""" return -1 @staticmethod def get_config_filenames() -> list[str]: """List of filenames to search for in the model directory.""" return [] @classmethod def from_config(cls, config: dict[str, Any]) -> "CustomQuantConfig": """Create a config class from the model's quantization config.""" return CustomQuantConfig(num_bits=config.get("num_bits", 8)) def get_quant_method( self, layer: torch.nn.Module, prefix: str ) -> FakeQuantLinearMethod | None: """Get the quantize method to use for the quantized layer.""" if isinstance(layer, LinearBase): return FakeQuantLinearMethod(num_bits=self.num_bits) return None def test_register_quantization_config(caplog_vllm): """Test register custom quantization config.""" # The quantization method `custom_quant` should be registered. assert get_quantization_config("custom_quant") == CustomQuantConfig # The quantization method `custom_quant` is already exists, # should raise a warning when re-registering it. with caplog_vllm.at_level(logging.WARNING): register_quantization_config("custom_quant")(CustomQuantConfig) assert any( "The quantization method 'custom_quant' already exists" in message for message in caplog_vllm.messages ), "Expected a warning when re-registering custom_quant" @pytest.mark.parametrize( argnames="model", argvalues=[ "meta-llama/Llama-3.2-1B-Instruct", ], ) def test_custom_quant(vllm_runner, model, monkeypatch): """Test infer with the custom quantization method.""" # `LLM.apply_model` requires pickling a function. monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") with vllm_runner( model_name=model, quantization="custom_quant", enforce_eager=True ) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj # Check the quantization method is FakeQuantLinearMethod assert isinstance(qkv_proj.quant_method, FakeQuantLinearMethod) llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=1) assert output
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_mixed_precision.py
tests/quantization/test_mixed_precision.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Test quark-quantized {MXFP4, FP8} mixed precision models. Run `pytest tests/quantization/test_mixed_precision.py`. """ import importlib import importlib.metadata from dataclasses import dataclass import lm_eval import pytest from packaging import version QUARK_MXFP4_AVAILABLE = importlib.util.find_spec("quark") is not None and version.parse( importlib.metadata.version("amd-quark") ) >= version.parse("0.8.99") @dataclass class ModelCase: model_id: str tp: int @dataclass class EvaluationConfig: model_name: str def get_model_args(self) -> str: return ( f"pretrained={self.model_name}," "tensor_parallel_size=4,dtype=auto,gpu_memory_utilization=0.8,trust_remote_code=False" ) TEST_CONFIGS = { # Mixed-precision (AMP) model # - Demonstrates end-to-end pipeline functionality "amd/Qwen3-8B-WMXFP4FP8-AMXFP4FP8-AMP-KVFP8": {"arc_challenge": 0.52, "mmlu": 0.72}, # Non-mixed-precision (PTQ) model # - Reference for pipeline compatibility verification -> No conflicts or breakings "amd/Llama-2-70b-chat-hf-FP8-MLPerf-fp8_attn_quark_format": { "arc_challenge": 0.53, "mmlu": 0.61, }, } @pytest.mark.parametrize("model_name, accuracy_numbers", TEST_CONFIGS.items()) @pytest.mark.skipif(not QUARK_MXFP4_AVAILABLE, reason="amd-quark>=0.9 is not available") def test_mixed_precision_model_accuracies(model_name: str, accuracy_numbers: dict): results = lm_eval.simple_evaluate( model="vllm", model_args=EvaluationConfig(model_name).get_model_args(), tasks=list(accuracy_numbers.keys()), batch_size=8, ) rtol = 0.05 for task, expect_accuracy in accuracy_numbers.items(): measured_accuracy = results["results"][task]["acc,none"] assert ( measured_accuracy - rtol < expect_accuracy and measured_accuracy + rtol > expect_accuracy ), f"Expected: {expect_accuracy} | Measured: {measured_accuracy}"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_quark.py
tests/quantization/test_quark.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Test model set-up and weight loading for quark-quantized models. Run `pytest tests/quantization/test_quark.py`. See also `tests/kernels/moe/test_ocp_mx_moe.py`. """ import importlib.metadata from dataclasses import dataclass from importlib.util import find_spec import huggingface_hub import lm_eval import pytest import torch from packaging import version from vllm.model_executor.layers.quantization.quark.quark import ( # noqa: E501 QuarkLinearMethod, QuarkW8A8Fp8, QuarkW8A8Int8, ) from vllm.platforms import current_platform from .reference_mxfp4 import dq_mxfp4_torch, qdq_mxfp4_torch QUARK_MXFP4_AVAILABLE = find_spec("quark") is not None and version.parse( importlib.metadata.version("amd-quark") ) >= version.parse("0.8.99") if QUARK_MXFP4_AVAILABLE: from quark.torch.export.nn.modules.realquantizer import StaticScaledRealQuantizer from quark.torch.kernel import mx as mx_kernel from quark.torch.quantization.config.config import FP4PerGroupSpec try: huggingface_hub.list_repo_refs( "amd/Llama-3.3-70B-Instruct-WMXFP4-AMXFP4-KVFP8-Scale-UINT8-SQ" ) HF_HUB_AMD_ORG_ACCESS = True except huggingface_hub.errors.RepositoryNotFoundError: HF_HUB_AMD_ORG_ACCESS = False @pytest.fixture(scope="function", autouse=True) def enable_pickle(monkeypatch): """`LLM.apply_model` requires pickling a function.""" monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") @pytest.mark.parametrize("kv_cache_dtype", ["auto", "fp8"]) @pytest.mark.parametrize("tp", [1]) def test_quark_fp8_w_per_tensor_a_per_tensor(vllm_runner, kv_cache_dtype, tp): model_path = "amd/Llama-3.1-8B-Instruct-FP8-KV-Quark-test" with vllm_runner( model_path, enforce_eager=True, kv_cache_dtype=kv_cache_dtype, tensor_parallel_size=tp, ) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, QuarkLinearMethod) assert isinstance(qkv_proj.scheme, QuarkW8A8Fp8) if isinstance(qkv_proj.scheme, QuarkW8A8Fp8): assert len(qkv_proj.input_scale.shape) == 0 assert qkv_proj.weight.dtype is current_platform.fp8_dtype() assert len(qkv_proj.weight_scale.shape) == 0 llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) assert output @pytest.mark.parametrize("tp", [1]) def test_quark_fp8_w_per_channel_a_per_token(vllm_runner, tp): model_path = "amd/Qwen2.5-1.5B-Instruct-ptpc-Quark-ts" with vllm_runner(model_path, enforce_eager=True, tensor_parallel_size=tp) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, QuarkLinearMethod) assert isinstance(qkv_proj.scheme, QuarkW8A8Fp8) if isinstance(qkv_proj.scheme, QuarkW8A8Fp8): assert qkv_proj.weight.dtype is current_platform.fp8_dtype() assert qkv_proj.weight_scale.shape[0] == qkv_proj.weight.shape[1] assert qkv_proj.weight_scale.shape[1] == 1 llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) assert output @pytest.mark.parametrize("tp", [1]) def test_quark_int8_w_per_tensor_a_per_tensor(vllm_runner, tp): model_path = "amd/Llama-3.1-8B-Instruct-w-int8-a-int8-sym-test" with vllm_runner(model_path, enforce_eager=True, tensor_parallel_size=tp) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj assert isinstance(qkv_proj.quant_method, QuarkLinearMethod) assert isinstance(qkv_proj.scheme, QuarkW8A8Int8) llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) assert output def test_quark_fp8_parity(vllm_runner): quark_model_id = "amd-quark/llama-tiny-fp8-quark-quant-method" fp8_model_id = "amd-quark/llama-tiny-fp8-quant-method" llm_kwargs = { "tensor_parallel_size": 1, "enforce_eager": True, "gpu_memory_utilization": 0.1, } with ( vllm_runner(quark_model_id, **llm_kwargs) as quark_handle, vllm_runner(fp8_model_id, **llm_kwargs) as fp8_handle, ): def get_state_dict(model): return {k: v.cpu() for k, v in model.state_dict().items()} (quark_state_dict,) = quark_handle.apply_model(get_state_dict) (fp8_state_dict,) = fp8_handle.apply_model(get_state_dict) assert fp8_state_dict.keys() == quark_state_dict.keys() for key in fp8_state_dict: assert torch.equal(fp8_state_dict[key], quark_state_dict[key]) @dataclass class AccuracyTestConfig: model_name: str excepted_value: float def get_model_args( self, tp_size: int, model_max_len: int | None = None, kwargs: dict | None = None, ) -> dict: if kwargs is None: kwargs = {} model_args = { "pretrained": self.model_name, "dtype": "auto", "add_bos_token": True, "tensor_parallel_size": tp_size, "gpu_memory_utilization": 0.7, **kwargs, } if model_max_len is not None: model_args["max_model_len"] = model_max_len return model_args GSM8K_ACCURACY_CONFIGS = [ # Private model. AccuracyTestConfig( model_name="amd/DeepSeek-R1-WMXFP4-AMXFP4-Scale-UINT8-MoE-Quant", excepted_value=0.96, ), ] WIKITEXT_ACCURACY_CONFIGS = [ AccuracyTestConfig( model_name="fxmarty/qwen1.5_moe_a2.7b_chat_w_fp4_a_fp6_e2m3", excepted_value=11.3, ), AccuracyTestConfig( model_name="fxmarty/qwen1.5_moe_a2.7b_chat_w_fp6_e3m2_a_fp6_e3m2", excepted_value=10.6, ), AccuracyTestConfig( model_name="fxmarty/qwen_1.5-moe-a2.7b-mxfp4", excepted_value=12.4 ), ] @pytest.mark.skipif(not QUARK_MXFP4_AVAILABLE, reason="amd-quark>=0.9 is not available") @pytest.mark.parametrize("config", WIKITEXT_ACCURACY_CONFIGS) @pytest.mark.parametrize("tp_size", [1, 2]) def test_ocp_mx_wikitext_correctness(config: AccuracyTestConfig, tp_size: int): if torch.cuda.device_count() < tp_size: pytest.skip( f"This test requires >={tp_size} gpus, got only {torch.cuda.device_count()}" ) task = "wikitext" rtol = 0.1 # Smaller cudagraph_capture_sizes to speed up the test. results = lm_eval.simple_evaluate( model="vllm", model_args=config.get_model_args( tp_size=tp_size, kwargs={"cudagraph_capture_sizes": [16]} ), tasks=task, batch_size=64, ) EXPECTED_VALUE = config.excepted_value measured_value = results["results"][task]["word_perplexity,none"] assert ( measured_value < EXPECTED_VALUE + rtol and measured_value > EXPECTED_VALUE - rtol ), f"Expected: {EXPECTED_VALUE} | Measured: {measured_value}" @pytest.mark.parametrize("config", GSM8K_ACCURACY_CONFIGS) @pytest.mark.skipif(not QUARK_MXFP4_AVAILABLE, reason="amd-quark>=0.9 is not available") @pytest.mark.skipif( not HF_HUB_AMD_ORG_ACCESS, reason="Read access to huggingface.co/amd is required for this test.", ) def test_mxfp4_gsm8k_correctness(config: AccuracyTestConfig): if torch.cuda.device_count() < 8: pytest.skip( f"This test requires >=8 gpus, got only {torch.cuda.device_count()}" ) task = "gsm8k" rtol = 0.03 results = lm_eval.simple_evaluate( model="vllm", model_args=config.get_model_args(tp_size=8, model_max_len=38768), tasks=task, batch_size=64, num_fewshot=8, ) EXPECTED_VALUE = config.excepted_value measured_value = results["results"][task]["exact_match,strict-match"] assert ( measured_value - rtol < EXPECTED_VALUE and measured_value + rtol > EXPECTED_VALUE ), f"Expected: {EXPECTED_VALUE} | Measured: {measured_value}" @pytest.mark.skipif(not QUARK_MXFP4_AVAILABLE, reason="amd-quark>=0.9 is not available") @pytest.mark.parametrize("float_dtype", [torch.bfloat16, torch.float16]) @pytest.mark.parametrize("scalings", [[2.3, 0.03, 7.3, 0.1, 0.004, 17.3, 1e4, 1e-4]]) def test_mxfp4_fused_qdq_match_quark(float_dtype: torch.dtype, scalings: list[int]): torch.manual_seed(0) hidden_size = 64 * 32 inp = (torch.rand(1, hidden_size, dtype=float_dtype, device="cuda") - 0.5) * 2 for i in range(hidden_size // 32): inp[:, i * 32 : (i + 1) * 32] = ( inp[:, i * 32 : (i + 1) * 32] * scalings[i % len(scalings)] ) inp_kernel = inp.clone() inp_kernel_clone = inp_kernel.clone() res_hip = mx_kernel.qdq_mxfp4_hip(inp_kernel_clone, "even") res_torch = qdq_mxfp4_torch(inp_kernel, "even") for i in range(hidden_size // 32): assert torch.all(torch.isfinite(res_hip[:, i * 32 : (i + 1) * 32])) assert torch.all(torch.isfinite(res_torch[:, i * 32 : (i + 1) * 32])) torch.testing.assert_close( res_hip[:, i * 32 : (i + 1) * 32], res_torch[:, i * 32 : (i + 1) * 32] ) @pytest.mark.skipif(not QUARK_MXFP4_AVAILABLE, reason="amd-quark>=0.9 is not available") @pytest.mark.parametrize("float_dtype", [torch.bfloat16, torch.float16]) @pytest.mark.parametrize("scalings", [[2.3, 0.03, 7.3, 0.1, 0.004, 17.3, 1e4, 1e-4]]) def test_mxfp4_dequant_kernel_match_quark( float_dtype: torch.dtype, scalings: list[int] ): qspec = FP4PerGroupSpec( ch_axis=-1, group_size=32, scale_format="e8m0", scale_calculation_mode="even", is_dynamic=False, ).to_quantization_spec() weight_quantizer = StaticScaledRealQuantizer( qspec=qspec, quantizer=None, reorder=False, real_quantized=True, float_dtype=float_dtype, device="cuda", ) observer = qspec.observer_cls(qspec, device="cuda") hidden_size = 512 shape = (11008, hidden_size) w = (torch.rand(shape, device="cuda", dtype=float_dtype) - 0.5) * 2 # Make it so that different groups have different scales. for i in range(hidden_size // 32): w[:, i * 32 : (i + 1) * 32] = ( w[:, i * 32 : (i + 1) * 32] * scalings[i % len(scalings)] ) observer(w) scale, _ = observer._calculate_qparams() weight_quantizer.scale = scale w_mxfp4 = weight_quantizer.to_real_quantize_params(w).to("cuda") weight_quantizer.maybe_convert_and_transpose_scale() scale = weight_quantizer.scale out_hip = mx_kernel.dq_mxfp4_hip(w_mxfp4, scale, float_dtype) out_torch = dq_mxfp4_torch(w_mxfp4, scale, float_dtype) assert torch.equal(out_hip, out_torch)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_modelopt.py
tests/quantization/test_modelopt.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Test ModelOpt quantization method setup and weight loading. Run `pytest tests/quantization/test_modelopt.py`. """ import os from typing import NoReturn import pytest import torch from tests.quantization.utils import is_quant_method_supported @pytest.fixture(scope="function", autouse=True) def enable_pickle(monkeypatch): """`LLM.apply_model` requires pickling a function.""" monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") def _skip(msg: str) -> NoReturn: pytest.skip(msg) raise RuntimeError(msg) def _snapshot_download_or_skip(model_id: str) -> str: try: from huggingface_hub import snapshot_download except Exception as e: # pragma: no cover _skip(f"huggingface_hub is required to download {model_id}: {e}") try: return snapshot_download( repo_id=model_id, repo_type="model", # These checkpoints are already small; download full repo for simplicity. allow_patterns=["*"], ) except Exception as e: _skip(f"Failed to download {model_id} from the HF Hub: {e}") @pytest.mark.skipif( not is_quant_method_supported("modelopt"), reason="ModelOpt FP8 is not supported on this GPU type.", ) def test_modelopt_fp8_checkpoint_setup(vllm_runner): """Test ModelOpt FP8 checkpoint loading and structure validation.""" # TODO: provide a small publicly available test checkpoint model_path = ( "/home/scratch.omniml_data_1/zhiyu/ckpts/test_ckpts/" "TinyLlama-1.1B-Chat-v1.0-fp8-0710" ) # Skip test if checkpoint doesn't exist if not os.path.exists(model_path): pytest.skip( f"Test checkpoint not found at {model_path}. " "This test requires a local ModelOpt FP8 checkpoint." ) with vllm_runner(model_path, quantization="modelopt", enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj o_proj = layer.self_attn.o_proj gate_up_proj = layer.mlp.gate_up_proj down_proj = layer.mlp.down_proj # Check that ModelOpt quantization method is properly applied from vllm.model_executor.layers.quantization.modelopt import ( ModelOptFp8LinearMethod, ) assert isinstance(qkv_proj.quant_method, ModelOptFp8LinearMethod) assert isinstance(o_proj.quant_method, ModelOptFp8LinearMethod) assert isinstance(gate_up_proj.quant_method, ModelOptFp8LinearMethod) assert isinstance(down_proj.quant_method, ModelOptFp8LinearMethod) # Check weight dtype is FP8 assert qkv_proj.weight.dtype == torch.float8_e4m3fn assert o_proj.weight.dtype == torch.float8_e4m3fn assert gate_up_proj.weight.dtype == torch.float8_e4m3fn assert down_proj.weight.dtype == torch.float8_e4m3fn # Check scales are present and have correct dtype assert hasattr(qkv_proj, "weight_scale") assert hasattr(qkv_proj, "input_scale") assert qkv_proj.weight_scale.dtype == torch.float32 assert qkv_proj.input_scale.dtype == torch.float32 assert hasattr(o_proj, "weight_scale") assert hasattr(o_proj, "input_scale") assert o_proj.weight_scale.dtype == torch.float32 assert o_proj.input_scale.dtype == torch.float32 assert hasattr(gate_up_proj, "weight_scale") assert hasattr(gate_up_proj, "input_scale") assert gate_up_proj.weight_scale.dtype == torch.float32 assert gate_up_proj.input_scale.dtype == torch.float32 assert hasattr(down_proj, "weight_scale") assert hasattr(down_proj, "input_scale") assert down_proj.weight_scale.dtype == torch.float32 assert down_proj.input_scale.dtype == torch.float32 llm.apply_model(check_model) # Run a simple generation test to ensure the model works output = llm.generate_greedy(["Hello my name is"], max_tokens=4) assert output print(f"ModelOpt FP8 output: {output}") @pytest.mark.skipif( not is_quant_method_supported("modelopt"), reason="ModelOpt FP8 is not supported on this GPU type.", ) def test_modelopt_fp8_pc_pt_checkpoint_setup(vllm_runner): """Test ModelOpt FP8_PER_CHANNEL_PER_TOKEN checkpoint setup.""" model_id = "CedricHwang/qwen2.5-0.5b-modelopt-fp8-pc-pt" model_path = _snapshot_download_or_skip(model_id) with vllm_runner(model_path, quantization="modelopt", enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj o_proj = layer.self_attn.o_proj gate_up_proj = layer.mlp.gate_up_proj down_proj = layer.mlp.down_proj from vllm.model_executor.layers.quantization.modelopt import ( ModelOptFp8PcPtLinearMethod, ) assert isinstance(qkv_proj.quant_method, ModelOptFp8PcPtLinearMethod) assert isinstance(o_proj.quant_method, ModelOptFp8PcPtLinearMethod) assert isinstance(gate_up_proj.quant_method, ModelOptFp8PcPtLinearMethod) assert isinstance(down_proj.quant_method, ModelOptFp8PcPtLinearMethod) assert qkv_proj.weight.dtype == torch.float8_e4m3fn assert o_proj.weight.dtype == torch.float8_e4m3fn assert gate_up_proj.weight.dtype == torch.float8_e4m3fn assert down_proj.weight.dtype == torch.float8_e4m3fn # Per-channel scales; activations are dynamically scaled per token. assert hasattr(qkv_proj, "weight_scale") assert qkv_proj.weight_scale.dtype == torch.float32 assert qkv_proj.weight_scale.dim() == 1 assert not hasattr(qkv_proj, "input_scale") assert hasattr(o_proj, "weight_scale") assert o_proj.weight_scale.dtype == torch.float32 assert o_proj.weight_scale.dim() == 1 assert not hasattr(o_proj, "input_scale") assert hasattr(gate_up_proj, "weight_scale") assert gate_up_proj.weight_scale.dtype == torch.float32 assert gate_up_proj.weight_scale.dim() == 1 assert not hasattr(gate_up_proj, "input_scale") assert hasattr(down_proj, "weight_scale") assert down_proj.weight_scale.dtype == torch.float32 assert down_proj.weight_scale.dim() == 1 assert not hasattr(down_proj, "input_scale") llm.apply_model(check_model) output = llm.generate_greedy(["Hello my name is"], max_tokens=4) assert output print(f"ModelOpt FP8_PER_CHANNEL_PER_TOKEN output: {output}") @pytest.mark.skipif( not is_quant_method_supported("modelopt"), reason="ModelOpt FP8 is not supported on this GPU type.", ) def test_modelopt_fp8_pb_wo_checkpoint_setup(vllm_runner): """Test ModelOpt FP8_PB_WO checkpoint setup.""" model_id = "CedricHwang/qwen2.5-0.5b-modelopt-fp8-pb-wo" model_path = _snapshot_download_or_skip(model_id) with vllm_runner(model_path, quantization="modelopt", enforce_eager=True) as llm: def check_model(model): layer = model.model.layers[0] qkv_proj = layer.self_attn.qkv_proj o_proj = layer.self_attn.o_proj gate_up_proj = layer.mlp.gate_up_proj down_proj = layer.mlp.down_proj from vllm.model_executor.layers.quantization.modelopt import ( ModelOptFp8PbWoLinearMethod, ) assert isinstance(qkv_proj.quant_method, ModelOptFp8PbWoLinearMethod) assert isinstance(o_proj.quant_method, ModelOptFp8PbWoLinearMethod) assert isinstance(gate_up_proj.quant_method, ModelOptFp8PbWoLinearMethod) assert isinstance(down_proj.quant_method, ModelOptFp8PbWoLinearMethod) assert qkv_proj.weight.dtype == torch.float8_e4m3fn assert o_proj.weight.dtype == torch.float8_e4m3fn assert gate_up_proj.weight.dtype == torch.float8_e4m3fn assert down_proj.weight.dtype == torch.float8_e4m3fn # Block scales; should be materialized as a 2D [out_blk, in_blk] tensor. assert hasattr(qkv_proj, "weight_scale") assert qkv_proj.weight_scale.dtype == torch.float32 assert qkv_proj.weight_scale.dim() == 2 assert hasattr(o_proj, "weight_scale") assert o_proj.weight_scale.dtype == torch.float32 assert o_proj.weight_scale.dim() == 2 assert hasattr(gate_up_proj, "weight_scale") assert gate_up_proj.weight_scale.dtype == torch.float32 assert gate_up_proj.weight_scale.dim() == 2 assert hasattr(down_proj, "weight_scale") assert down_proj.weight_scale.dtype == torch.float32 assert down_proj.weight_scale.dim() == 2 llm.apply_model(check_model) output = llm.generate_greedy(["Hello my name is"], max_tokens=4) assert output print(f"ModelOpt FP8_PB_WO output: {output}")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_torchao.py
tests/quantization/test_torchao.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import importlib.metadata import importlib.util import pytest import torch DTYPE = ["bfloat16"] TORCHAO_AVAILABLE = importlib.util.find_spec("torchao") is not None @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") def test_pre_quantized_model(vllm_runner): with vllm_runner( "drisspg/fp8-opt-125m", quantization="torchao", dtype="bfloat16", enforce_eager=True, ) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") @pytest.mark.parametrize( "pt_load_map_location", [ "cuda:0", # {"": "cuda"}, ], ) def test_opt_125m_int8wo_model_loading_with_params(vllm_runner, pt_load_map_location): torch._dynamo.reset() model_name = "jerryzh168/opt-125m-int8wo-partial-quant" with vllm_runner( model_name=model_name, quantization="torchao", dtype="bfloat16", pt_load_map_location=pt_load_map_location, enforce_eager=True, ) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") def test_opt_125m_int4wo_model_per_module_quant(vllm_runner): torch._dynamo.reset() model_name = "jerryzh168/opt-125m-int4wo-per-module" with vllm_runner( model_name=model_name, quantization="torchao", dtype="bfloat16", pt_load_map_location="cuda:0", enforce_eager=True, ) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") def test_qwenvl_int8wo_model_loading_with_params(vllm_runner): torch._dynamo.reset() model_name = "mobicham/Qwen2.5-VL-3B-Instruct_int8wo_ao" with vllm_runner( model_name=model_name, quantization="torchao", dtype="bfloat16", pt_load_map_location="cuda:0", enforce_eager=True, ) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") @pytest.mark.skip( reason="since torchao nightly is only compatible with torch nightly" "currently https://github.com/pytorch/ao/issues/2919, we'll have to skip " "torchao tests that requires newer versions (0.14.0.dev+) for now" ) def test_opt_125m_awq_int4wo_model_loading_with_params(vllm_runner): torch._dynamo.reset() model_name = "torchao-testing/opt-125m-AWQConfig-Int4WeightOnlyConfig-v2-0.14.0.dev" with vllm_runner( model_name=model_name, quantization="torchao", dtype="bfloat16", pt_load_map_location="cuda:0", ) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") def test_online_quant_config_dict_json(vllm_runner): """Testing on the fly quantization, load_weights integration point, with config dict serialized to json string """ torch._dynamo.reset() model_name = "facebook/opt-125m" import json from torchao.core.config import config_to_dict from torchao.quantization import Float8DynamicActivationFloat8WeightConfig, PerRow torchao_quant_config = Float8DynamicActivationFloat8WeightConfig( granularity=PerRow() ) hf_overrides = { "quantization_config_dict_json": json.dumps( config_to_dict(torchao_quant_config) ) } with vllm_runner( model_name=model_name, dtype="bfloat16", pt_load_map_location="cuda:0", quantization="torchao", hf_overrides=hf_overrides, enforce_eager=True, ) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") def test_online_quant_config_file(vllm_runner): """Testing on the fly quantization, load_weights integration point, with config file """ torch._dynamo.reset() model_name = "facebook/opt-125m" import json from tempfile import NamedTemporaryFile from torchao.core.config import config_to_dict from torchao.quantization import Float8DynamicActivationFloat8WeightConfig, PerRow config = Float8DynamicActivationFloat8WeightConfig(granularity=PerRow()) with NamedTemporaryFile(mode="w", delete=False) as f: f.write(json.dumps(config_to_dict(config))) # close the file to save it f.close() config_file_name = str(f.name) hf_overrides = {"quantization_config_file": config_file_name} with vllm_runner( model_name=model_name, dtype="bfloat16", pt_load_map_location="cuda:0", quantization="torchao", hf_overrides=hf_overrides, enforce_eager=True, ) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") def test_reload_weights(): import json from torchao.core.config import config_to_dict from torchao.quantization import Float8DynamicActivationFloat8WeightConfig, PerRow from vllm import LLM, SamplingParams torchao_quant_config = Float8DynamicActivationFloat8WeightConfig( granularity=PerRow() ) hf_overrides = { "quantization_config_dict_json": json.dumps( config_to_dict(torchao_quant_config) ) } llm = LLM( model="Qwen/Qwen3-0.6B", dtype="bfloat16", load_format="dummy", enforce_eager=True, quantization="torchao", hf_overrides=hf_overrides, ) # Update load format from `dummy` to `auto` llm.collective_rpc( "update_config", args=({"load_config": {"load_format": "auto"}},) ) # Now reload real weights inplace llm.collective_rpc("reload_weights") prompts = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] # Create a sampling params object. sampling_params = SamplingParams(temperature=0, top_p=0.95) outputs = llm.generate(prompts, sampling_params) # make sure it runs for output in outputs: generated_text = output.outputs[0].text assert generated_text # can also uncomment locally to make sure the generated # output makes sense # prompt = output.prompt # print(f"Prompt: {prompt!r}") # print(f"Output: {generated_text!r}") # print("-" * 60) @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") @pytest.mark.skip( reason="since torchao nightly is only compatible with torch nightly" "currently https://github.com/pytorch/ao/issues/2919, we'll have to skip " "torchao tests that requires newer versions (0.15.0.dev+) for now" ) def test_safetensors_model_loading_with_params(vllm_runner): torch._dynamo.reset() # using this model to test safetensors loading with file sharding model_name = "torchao-testing/Qwen3-8B-INT4-0.15.0dev-safetensors" with vllm_runner(model_name=model_name, dtype="bfloat16") as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") @pytest.mark.skip( reason="since torchao nightly is only compatible with torch nightly" "currently https://github.com/pytorch/ao/issues/2919, we'll have to skip " "torchao tests that requires newer versions (0.14.0.dev+) for now" ) def test_opt_125m_module_fqn_to_config_regex_model(vllm_runner): torch._dynamo.reset() model_name = "torchao-testing/opt-125m-ModuleFqnToConfig-v1-regex-0.14.0.dev" with vllm_runner( model_name=model_name, dtype="bfloat16", pt_load_map_location="cuda:0" ) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") @pytest.mark.skip( reason="since torchao nightly is only compatible with torch nightly" "currently https://github.com/pytorch/ao/issues/2919, we'll have to skip " "torchao tests that requires newer versions (0.14.0.dev+) for now" ) def test_opt_125m_int4wo_model_running_preshuffled_kernel(vllm_runner, monkeypatch): """We load a model with Int4Tensor (plain format) linear weights and verify that the weight is updated to Int4PreshuffledTensor after loading in vllm """ from torchao.quantization import Int4PreshuffledTensor from torchao.utils import _is_fbgemm_gpu_genai_available, is_sm_at_least_90 torch._dynamo.reset() monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") model_name = "torchao-testing/opt-125m-Int4WeightOnlyConfig-v2-0.14.0.dev" # Note: using enforce_eager=True because the `bf16i4bf16_shuffled` doesn't # have meta kernel implemented yet, can remove this flag after that is implemented with vllm_runner( model_name=model_name, quantization="torchao", dtype="bfloat16", pt_load_map_location="cuda:0", enforce_eager=True, ) as llm: def has_int4_preshuffled_tensor_weight(model): return isinstance( model.model.decoder.layers[0].self_attn.qkv_proj.weight, Int4PreshuffledTensor, ) def get_weight_attrs(model): weight = model.model.decoder.layers[0].self_attn.qkv_proj.weight return [ weight.requires_grad, weight.input_dim, weight.output_dim, hasattr(weight, "weight_loader"), ] llm_engine = llm.get_llm().llm_engine has_int4_preshuffled_tensor = any( llm_engine.apply_model(has_int4_preshuffled_tensor_weight) ) weight_attrs = llm_engine.apply_model(get_weight_attrs)[0] # making sure we are using Int4PreshuffledTensor on H100 GPU, when # fbgemm_gpu_genai # library is installed, otherwise it should be using Int4Tensor if _is_fbgemm_gpu_genai_available() and is_sm_at_least_90(): assert has_int4_preshuffled_tensor else: assert not has_int4_preshuffled_tensor assert weight_attrs == [False, 1, 0, True] output = llm.generate_greedy(["The capital of France is"], max_tokens=32) assert output @pytest.mark.skipif(not TORCHAO_AVAILABLE, reason="torchao is not available") @pytest.mark.skip( reason="since torchao nightly is only compatible with torch nightly" "currently https://github.com/pytorch/ao/issues/2919, we'll have to skip " "torchao tests that requires newer versions (0.14.0.dev+) for now" ) def test_opt_125m_int4wo_model_running_preshuffled_kernel_online_quant( vllm_runner, monkeypatch ): """We load a bf16 model and online quantize the model to int4, then verify that the weights are updated to Int4PreshuffledTensor after online quantization """ from torchao.quantization import Int4PreshuffledTensor from torchao.utils import _is_fbgemm_gpu_genai_available, is_sm_at_least_90 torch._dynamo.reset() model_name = "facebook/opt-125m" monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") import json from torchao.core.config import config_to_dict from torchao.quantization import Int4WeightOnlyConfig torchao_quant_config = Int4WeightOnlyConfig( group_size=128, int4_packing_format="plain" ) hf_overrides = { "quantization_config_dict_json": json.dumps( config_to_dict(torchao_quant_config) ) } # Note: using enforce_eager=True because the `bf16i4bf16_shuffled` doesn't # have meta kernel implemented yet, can remove this flag after that is implemented with vllm_runner( model_name=model_name, quantization="torchao", dtype="bfloat16", pt_load_map_location="cuda:0", hf_overrides=hf_overrides, enforce_eager=True, ) as llm: def has_int4_preshuffled_tensor_weight(model): return isinstance( model.model.decoder.layers[0].self_attn.qkv_proj.weight, Int4PreshuffledTensor, ) def get_weight_attrs(model): weight = model.model.decoder.layers[0].self_attn.qkv_proj.weight return [ weight.requires_grad, weight.input_dim, weight.output_dim, hasattr(weight, "weight_loader"), ] llm_engine = llm.get_llm().llm_engine has_int4_preshuffled_tensor = any( llm_engine.apply_model(has_int4_preshuffled_tensor_weight) ) weight_attrs = llm_engine.apply_model(get_weight_attrs)[0] # making sure we are using Int4PreshuffledTensor on H100 GPU, when # fbgemm_gpu_genai # library is installed, otherwise it should be using Int4Tensor if _is_fbgemm_gpu_genai_available() and is_sm_at_least_90(): assert has_int4_preshuffled_tensor else: assert not has_int4_preshuffled_tensor assert weight_attrs == [False, 1, 0, True] output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output if __name__ == "__main__": pytest.main([__file__])
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/fp_quant.py
tests/quantization/fp_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Test model set-up and inference for quantized HF models supported on the GPU backend using FPQuant. Validating the configuration and printing results for manual checking. Run `pytest tests/quantization/test_fp_quant.py`. """ import pytest from tests.quantization.utils import is_quant_method_supported MODELS = [ "ISTA-DASLab/Qwen3-0.6B-RTN-NVFP4", "ISTA-DASLab/Qwen3-0.6B-RTN-MXFP4", ] DTYPE = ["bfloat16"] EAGER = [True, False] @pytest.mark.skipif( not is_quant_method_supported("fp_quant"), reason="FPQuant is not supported on this GPU type.", ) @pytest.mark.parametrize("model", MODELS) @pytest.mark.parametrize("eager", EAGER) def test_fpquant(vllm_runner, model, eager): with vllm_runner(model, enforce_eager=eager) as llm: output = llm.generate_greedy(["1 2 3 4 5"], max_tokens=2) assert output[0][1] == "1 2 3 4 5 6"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_cpu_offload.py
tests/quantization/test_cpu_offload.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Expanded quantized model tests for CPU offloading # Base tests: tests/basic_correctness/test_cpu_offload.py import pytest from tests.quantization.utils import is_quant_method_supported from ..utils import compare_two_settings @pytest.mark.skipif( not is_quant_method_supported("fp8"), reason="fp8 is not supported on this GPU type.", ) def test_cpu_offload_fp8(): # Test loading a quantized checkpoint compare_two_settings( "neuralmagic/Qwen2-1.5B-Instruct-FP8", ["--enforce_eager"], ["--enforce_eager", "--cpu-offload-gb", "1"], max_wait_seconds=480, ) @pytest.mark.skipif( not is_quant_method_supported("gptq_marlin"), reason="gptq_marlin is not supported on this GPU type.", ) def test_cpu_offload_gptq(monkeypatch): # This quant method is sensitive to dummy weights, so we force real weights monkeypatch.setenv("VLLM_TEST_FORCE_LOAD_FORMAT", "auto") # Test GPTQ Marlin compare_two_settings( "Qwen/Qwen2-1.5B-Instruct-GPTQ-Int4", ["--enforce_eager"], ["--enforce_eager", "--cpu-offload-gb", "1"], max_wait_seconds=480, ) @pytest.mark.skipif( not is_quant_method_supported("awq_marlin"), reason="awq_marlin is not supported on this GPU type.", ) def test_cpu_offload_awq(monkeypatch): # This quant method is sensitive to dummy weights, so we force real weights monkeypatch.setenv("VLLM_TEST_FORCE_LOAD_FORMAT", "auto") # Test AWQ Marlin compare_two_settings( "Qwen/Qwen2-1.5B-Instruct-AWQ", ["--enforce_eager"], ["--enforce_eager", "--cpu-offload-gb", "1"], max_wait_seconds=480, ) @pytest.mark.skipif( not is_quant_method_supported("gptq_marlin"), reason="gptq_marlin is not supported on this GPU type.", ) def test_cpu_offload_compressed_tensors(monkeypatch): # This quant method is sensitive to dummy weights, so we force real weights monkeypatch.setenv("VLLM_TEST_FORCE_LOAD_FORMAT", "auto") # Test wNa16 compare_two_settings( "nm-testing/tinyllama-oneshot-w4a16-channel-v2", ["--enforce_eager"], ["--enforce_eager", "--cpu-offload-gb", "1"], max_wait_seconds=480, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_fp8.py
tests/quantization/test_fp8.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests whether FP8 computation is enabled correctly. Run `pytest tests/quantization/test_fp8.py --forked`. """ import pytest import torch from tests.quantization.utils import is_quant_method_supported from vllm import _custom_ops as ops from vllm.model_executor.layers.fused_moe import FusedMoE from vllm.model_executor.layers.quantization.fp8 import ( Fp8Config, Fp8KVCacheMethod, Fp8LinearMethod, Fp8MoeBackend, Fp8MoEMethod, ) from vllm.model_executor.model_loader.weight_utils import default_weight_loader from vllm.platforms import current_platform MODELS = [ "neuralmagic/Meta-Llama-3-8B-Instruct-FP8-KV", # The checkpoint below was removed from the HF. # TODO: add a small replacement checkpoint. pytest.param( "nm-testing/Qwen2-0.5B-Instruct-FP8-SkipQKV", marks=pytest.mark.skip(reason="Checkpoint removed from HF."), ), ] @pytest.mark.skipif( not is_quant_method_supported("fp8"), reason="FP8 is not supported on this GPU type.", ) @pytest.mark.parametrize("model_id", MODELS) @pytest.mark.parametrize("force_marlin", [False, True]) @pytest.mark.parametrize( "use_rocm_aiter", [True, False] if current_platform.is_rocm() else [False] ) def test_model_load_and_run( vllm_runner, model_id: str, force_marlin: bool, use_rocm_aiter: bool, monkeypatch ) -> None: if use_rocm_aiter: monkeypatch.setenv("VLLM_ROCM_USE_AITER", "1") if force_marlin: monkeypatch.setenv("VLLM_TEST_FORCE_FP8_MARLIN", "1") with vllm_runner(model_id, enforce_eager=True) as llm: # note: this does not test accuracy, just that we can run through # see lm-eval tests for accuracy outputs = llm.generate_greedy(["Hello my name is"], max_tokens=4) print(outputs[0][1]) KV_CACHE_MODELS = [ # AutoFP8 format using separate .k_scale and .v_scale # The original checkpoint below was removed from the Hub. To unblock CI and # until a small replacement with split K/V scales is found, skip this case. # See PR #27717 for context. pytest.param( "nm-testing/Qwen2-1.5B-Instruct-FP8-K-V", marks=pytest.mark.skip( reason=( "Checkpoint removed from HF; temporarily disabling this " "AutoFP8 split K/V case (PR #27717)." ) ), ), ] @pytest.mark.skipif( not is_quant_method_supported("fp8"), reason="FP8 is not supported on this GPU type.", ) @pytest.mark.parametrize("model_id", KV_CACHE_MODELS) @pytest.mark.parametrize( "use_rocm_aiter", [True, False] if current_platform.is_rocm() else [False] ) def test_kv_cache_model_load_and_run( vllm_runner, model_id: str, use_rocm_aiter: bool, monkeypatch ): if use_rocm_aiter: monkeypatch.setenv("VLLM_ROCM_USE_AITER", "1") # `LLM.apply_model` requires pickling a function. monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") with vllm_runner(model_id, kv_cache_dtype="fp8", enforce_eager=True) as llm: def check_model(model): attn = model.model.layers[0].self_attn.attn assert isinstance(attn.quant_method, Fp8KVCacheMethod) if not current_platform.is_rocm(): # NOTE: This code path requires validation on Non-CUDA platform # NOTE: it is valid for scales to be 1.0 (default value), but # we know these checkpoints have scales < 1.0 assert 0.0 < attn._k_scale < 1.0 assert 0.0 < attn._v_scale < 1.0 else: # NOTE: This code path is for ROCm platform # NOTE: it is valid for scales to be 1.0 (default value), but # we know these checkpoints have scales < 1.0 # However on ROCm platform, the _k_scale and _v_scale will be # scaled by a factor of 2 as described in # vllm/model_executor/layers/quantization/kv_cache.py assert 0.0 < attn._k_scale < (1.0 * 2.0) assert 0.0 < attn._v_scale < (1.0 * 2.0) llm.apply_model(check_model) # note: this does not test accuracy, just that we can run through # see lm-eval tests for accuracy outputs = llm.generate_greedy(["Hello my name is"], max_tokens=4) print(outputs[0][1]) @pytest.mark.skipif( not is_quant_method_supported("fp8"), reason="FP8 is not supported on this GPU type.", ) @pytest.mark.parametrize("kv_cache_dtype", ["auto", "fp8"]) @pytest.mark.parametrize("force_marlin", [False, True]) @pytest.mark.parametrize( "use_rocm_aiter", [True, False] if current_platform.is_rocm() else [False] ) def test_load_fp16_model( vllm_runner, kv_cache_dtype: str, force_marlin: bool, use_rocm_aiter: bool, monkeypatch, ) -> None: if use_rocm_aiter: monkeypatch.setenv("VLLM_ROCM_USE_AITER", "1") # `LLM.apply_model` requires pickling a function. monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") if force_marlin: monkeypatch.setenv("VLLM_TEST_FORCE_FP8_MARLIN", "1") with vllm_runner( "facebook/opt-125m", quantization="fp8", enforce_eager=True, kv_cache_dtype=kv_cache_dtype, ) as llm: def check_model(model): fc1 = model.model.decoder.layers[0].fc1 assert isinstance(fc1.quant_method, Fp8LinearMethod) if kv_cache_dtype == "fp8": attn = model.model.decoder.layers[0].self_attn.attn assert isinstance(attn.quant_method, Fp8KVCacheMethod) assert attn._k_scale == 1.0 assert attn._v_scale == 1.0 if current_platform.is_cuda(): if current_platform.supports_fp8() and not force_marlin: # For GPUs with hardware support, we keep weights in fp8 assert fc1.weight.dtype == torch.float8_e4m3fn else: # For GPUs without hardware support, we pack the fp8 weights # for weight-only quantization using Marlin kernels assert fc1.weight.dtype == torch.int32 elif current_platform.is_rocm(): if current_platform.supports_fp8() and not force_marlin: # For GPUs with hardware support, we keep weights in fp8 assert fc1.weight.dtype == current_platform.fp8_dtype() else: # unsupported ROCm platform pytest.skip( "Skip `test_load_fp16_model`. " "It only runs on ROCm platform with FP8 compute." " e.g. MI300X and above." ) else: # unsupported platform pytest.skip( "Skip `test_load_fp16_model`. " "It only runs on CUDA and ROCm platform." ) llm.apply_model(check_model) @pytest.mark.skipif( not is_quant_method_supported("fp8"), reason="FP8 is not supported on this GPU type.", ) @pytest.mark.parametrize("dtype", [torch.float16, torch.bfloat16]) def test_scaled_fp8_quant(dtype) -> None: def quantize_ref(tensor, inv_scale): # The reference implementation that fully aligns to # the kernel being tested. finfo = torch.finfo(torch.float8_e4m3fn) scale = inv_scale.reciprocal() qweight = (tensor.to(torch.float32) * scale).clamp(min=finfo.min, max=finfo.max) qweight = qweight.to(torch.float8_e4m3fn) return qweight def per_tensor_dequantize(tensor, inv_scale, dtype): fake_qweight = tensor.to(dtype) dq_weight = fake_qweight * inv_scale return dq_weight # Note that we use a shape % 4 != 0 to cover edge cases, # because scaled_fp8_quant is vectorized by 4. x = (torch.randn(size=(11, 11), device="cuda") * 13).to(dtype) # Dynamic quantization ref_y, inv_scale = ops.scaled_fp8_quant(x, None) ref_y = per_tensor_dequantize(ref_y, inv_scale, dtype) # Reference dynamic quantization y = quantize_ref(x, inv_scale) torch.testing.assert_close(ref_y, per_tensor_dequantize(y, inv_scale, dtype)) # Static quantization y, _ = ops.scaled_fp8_quant(x, inv_scale) torch.testing.assert_close(ref_y, per_tensor_dequantize(y, inv_scale, dtype)) # Padding y, _ = ops.scaled_fp8_quant(x, inv_scale, num_token_padding=17) assert y.shape[0] == 17 torch.testing.assert_close( ref_y, per_tensor_dequantize(torch.narrow(y, 0, 0, x.shape[0]), inv_scale, dtype), ) # non-contiguous input with padding m, n, padded_stride = 975, 512, 576 padded_tensor = (torch.randn(size=(m, padded_stride), device="cuda") * 13).to(dtype) x_nc = padded_tensor[:, :n] # shape (m, n) with stride (padded_stride, 1) assert not x_nc.is_contiguous() assert x_nc.stride(0) == padded_stride # dynamic quantization ref_y_nc, inv_scale_nc = ops.scaled_fp8_quant(x_nc, None) ref_y_nc = per_tensor_dequantize(ref_y_nc, inv_scale_nc, dtype) # reference dynamic quantization y_nc = quantize_ref(x_nc, inv_scale_nc) torch.testing.assert_close( ref_y_nc, per_tensor_dequantize(y_nc, inv_scale_nc, dtype) ) # static quantization y_nc, _ = ops.scaled_fp8_quant(x_nc, inv_scale_nc) torch.testing.assert_close( ref_y_nc, per_tensor_dequantize(y_nc, inv_scale_nc, dtype) ) # padding after non-contiguous input quantization y_nc_pad, _ = ops.scaled_fp8_quant(x_nc, inv_scale_nc, num_token_padding=m + 10) assert y_nc_pad.shape[0] == m + 10 torch.testing.assert_close( ref_y_nc, per_tensor_dequantize( torch.narrow(y_nc_pad, 0, 0, x_nc.shape[0]), inv_scale_nc, dtype ), ) @pytest.mark.parametrize("method_cls", [Fp8LinearMethod, Fp8MoEMethod]) # FP8 weight reloading does not support online quantization @pytest.mark.parametrize("is_checkpoint_fp8_serialized", [True]) # skip False @pytest.mark.parametrize("weight_block_size", [None, [1, 1]]) # any postprocessing that is applied to the weights such as padding and repacking # (excluding device sharding) must also be applied to the reloaded weights # # this is the case for marlin as well as per-tensor Fp8MoEMethod @pytest.mark.parametrize("use_marlin", [False]) # skip True def test_fp8_reloading( method_cls, is_checkpoint_fp8_serialized, weight_block_size, use_marlin, dist_init ): if is_checkpoint_fp8_serialized is False: pytest.skip("FP8 weight reloading does not support online quantization") if method_cls is Fp8MoEMethod and weight_block_size is None: pytest.skip( "FP8 Tensor weight reloading does not support fusing w13_weight_scale. " "If this is your use case, consider using a restore function like #26327" ) with torch.device("cuda:0"): config = Fp8Config( is_checkpoint_fp8_serialized=is_checkpoint_fp8_serialized, weight_block_size=weight_block_size, ) if method_cls is Fp8LinearMethod: layer = torch.nn.Linear(1, 1) method = method_cls(config) method.create_weights( layer=layer, input_size_per_partition=1, output_partition_sizes=[1], input_size=1, output_size=1, params_dtype=torch.bfloat16, weight_loader=default_weight_loader, ) else: layer = FusedMoE( num_experts=1, top_k=1, hidden_size=1, intermediate_size=1, ) method = method_cls(config, layer) method.create_weights( layer=layer, num_experts=1, hidden_size=1, intermediate_size_per_partition=1, params_dtype=torch.bfloat16, weight_loader=default_weight_loader, ) # Fp8LinearMethod uses use_marlin # Fp8MoEMethod uses fp8_backend method.use_marlin = use_marlin method.fp8_backend = Fp8MoeBackend.MARLIN if use_marlin else None # capture weights format during loading original_metadata = [ (name, param.shape, getattr(param, "weight_loader", default_weight_loader)) for name, param in layer.named_parameters() ] # test loading for name, shape, _ in original_metadata: param = getattr(layer, name) weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, torch.zeros(shape)) # cannot use empty method.process_weights_after_loading(layer) # test reloading works after loading # assuming that no reshaping occurred for name, shape, original_weight_loader in original_metadata: param = getattr(layer, name) weight_loader = getattr(param, "weight_loader", default_weight_loader) assert weight_loader is original_weight_loader weight_loader(param, torch.zeros(shape)) # cannot use empty method.process_weights_after_loading(layer)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_ptpc_fp8.py
tests/quantization/test_ptpc_fp8.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests whether PTPC w8a8 FP8 computation is enabled correctly. Run `pytest tests/quantization/test_ptpc_fp8.py --forked`. """ import pytest import torch from tests.quantization.utils import is_quant_method_supported from vllm.model_executor.layers.quantization.fp8 import Fp8KVCacheMethod from vllm.model_executor.layers.quantization.ptpc_fp8 import PTPCFp8LinearMethod from vllm.platforms import current_platform UNSUPPORTED_STR = ( "Currently torch._scaled_mm (hipBLASLt) rowwise gemm only " "support output dtype of bfloat16. torch.float16 is specified." ) @pytest.fixture(scope="function", autouse=True) def enable_pickle(monkeypatch): """`LLM.apply_model` requires pickling a function.""" monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") @pytest.mark.skipif( not is_quant_method_supported("ptpc_fp8"), reason="PTPC FP8 is not supported on this GPU type.", ) @pytest.mark.skipif(not current_platform.is_rocm(), reason="This test is for ROCm GPU.") @pytest.mark.parametrize("dtype", ["auto", "bfloat16", "float16"]) @pytest.mark.parametrize("kv_cache_dtype", ["auto", "fp8", "fp8_e4m3"]) def test_ptpc_fp8_rocm(vllm_runner, dtype: str, kv_cache_dtype: str) -> None: try: llm = vllm_runner( "facebook/opt-125m", dtype=dtype, quantization="ptpc_fp8", enforce_eager=True, kv_cache_dtype=kv_cache_dtype, ) except AssertionError as e: if str(e) == UNSUPPORTED_STR: # If the error message matches, the test passes return else: # If the error message does not match, re-raise the exception raise with llm: def check_model(model): fc1 = model.model.decoder.layers[0].fc1 assert isinstance(fc1.quant_method, PTPCFp8LinearMethod) if kv_cache_dtype == "ptpc_fp8": attn = model.model.decoder.layers[0].self_attn.attn assert isinstance(attn.quant_method, Fp8KVCacheMethod) assert attn._k_scale == 1.0 assert attn._v_scale == 1.0 if current_platform.has_device_capability(94): # For GPUs with hardware support, we keep weights in fp8 assert fc1.weight.dtype == torch.float8_e4m3fnuz llm.apply_model(check_model) output = llm.generate_greedy("Hello my name is", max_tokens=4) assert output
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/utils.py
tests/quantization/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from vllm.model_executor.layers.quantization import get_quantization_config from vllm.platforms import current_platform def is_quant_method_supported(quant_method: str) -> bool: # Currently, all quantization methods require Nvidia or AMD GPUs if not (current_platform.is_cuda() or current_platform.is_rocm()): return False capability = current_platform.get_device_capability() assert capability is not None min_capability = get_quantization_config(quant_method).get_min_capability() return capability.to_int() >= min_capability
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_experts_int8.py
tests/quantization/test_experts_int8.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # flake8: noqa """Tests experts_int8 quantization startup and generation, doesn't test correctness """ import pytest from tests.quantization.utils import is_quant_method_supported from ..models.registry import HF_EXAMPLE_MODELS MODELS = ["ai21labs/Jamba-tiny-random", "pfnet/plamo-2-1b"] @pytest.mark.skipif( not is_quant_method_supported("experts_int8"), reason="ExpertsInt8 is not supported on this GPU type.", ) @pytest.mark.parametrize("model", MODELS) @pytest.mark.parametrize("dtype", ["bfloat16"]) @pytest.mark.parametrize("max_tokens", [4]) def test_model_experts_int8_startup( hf_runner, vllm_runner, example_prompts, model: str, dtype: str, max_tokens: int, ) -> None: model_info = HF_EXAMPLE_MODELS.find_hf_info(model) model_info.check_transformers_version(on_fail="skip") with vllm_runner( model, dtype=dtype, enforce_eager=True, quantization="experts_int8" ) as vllm_model: vllm_model.generate_greedy(example_prompts, max_tokens)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/__init__.py
tests/quantization/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_configs.py
tests/quantization/test_configs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests whether Marlin models can be loaded from the autogptq config. Run `pytest tests/quantization/test_configs.py --forked`. """ from dataclasses import dataclass import pytest from vllm.config import ModelConfig @dataclass class ModelPair: model_marlin: str model_gptq: str # Model Id // Quantization Arg // Expected Type MODEL_ARG_EXPTYPES = [ # AUTOGPTQ # compat: autogptq <=0.7.1 is_marlin_format: bool # Model Serialized in Exllama Format. ("TheBloke/Llama-2-7B-Chat-GPTQ", None, "gptq_marlin"), ("TheBloke/Llama-2-7B-Chat-GPTQ", "marlin", "gptq_marlin"), ("TheBloke/Llama-2-7B-Chat-GPTQ", "gptq", "gptq"), ("TheBloke/Llama-2-7B-Chat-GPTQ", "awq", "ERROR"), # compat: autogptq >=0.8.0 use checkpoint_format: str # Model Serialized in Exllama Format. ("LnL-AI/TinyLlama-1.1B-Chat-v1.0-GPTQ-4bit", None, "gptq_marlin"), ("LnL-AI/TinyLlama-1.1B-Chat-v1.0-GPTQ-4bit", "marlin", "gptq_marlin"), ("LnL-AI/TinyLlama-1.1B-Chat-v1.0-GPTQ-4bit", "gptq", "gptq"), ("LnL-AI/TinyLlama-1.1B-Chat-v1.0-GPTQ-4bit", "awq", "ERROR"), # AUTOAWQ ("TheBloke/OpenHermes-2.5-Mistral-7B-AWQ", None, "awq_marlin"), ("TheBloke/OpenHermes-2.5-Mistral-7B-AWQ", "awq", "awq"), ("TheBloke/OpenHermes-2.5-Mistral-7B-AWQ", "marlin", "awq_marlin"), ("TheBloke/OpenHermes-2.5-Mistral-7B-AWQ", "gptq", "ERROR"), ] @pytest.mark.parametrize("model_arg_exptype", MODEL_ARG_EXPTYPES) def test_auto_gptq(model_arg_exptype: tuple[str, None, str]) -> None: model_path, quantization_arg, expected_type = model_arg_exptype try: model_config = ModelConfig(model_path, quantization=quantization_arg) found_quantization_type = model_config.quantization except ValueError: found_quantization_type = "ERROR" assert found_quantization_type == expected_type, ( f"Expected quant_type == {expected_type} for {model_path}, " f"but found {found_quantization_type} " f"for no --quantization {quantization_arg} case" )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_gptq_dynamic.py
tests/quantization/test_gptq_dynamic.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests whether gptq models with dynamic quantized can be loaded. Run `pytest tests/quantization/test_gptq_dynamic.py --forked`. """ import pytest import torch from vllm.model_executor.layers.linear import UnquantizedLinearMethod from vllm.model_executor.layers.quantization.gptq import GPTQLinearMethod from vllm.model_executor.layers.quantization.gptq_marlin import GPTQMarlinLinearMethod from vllm.model_executor.layers.quantization.utils.gptq_utils import ( get_dynamic_override, ) PROMPT = "On the surface of Mars, we found" # The first layer is quantized using bits=4, group_size=128 # The second layer is quantized using bits=8, group_size=32 # All other layers (layer index >= 2) are not quantized MODEL_QUANT = [ ("ModelCloud/Qwen1.5-1.8B-Chat-GPTQ-4bits-dynamic-cfg-with-lm_head-symTrue", True), ( "ModelCloud/Qwen1.5-1.8B-Chat-GPTQ-4bits-dynamic-cfg-with-lm_head-symFalse", False, ), ] @pytest.mark.parametrize("model_id, use_marlin_kernel", MODEL_QUANT) def test_gptq_with_dynamic( vllm_runner, model_id: str, use_marlin_kernel: bool, monkeypatch ): # `LLM.apply_model` requires pickling a function. monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") linear_method_cls = ( GPTQMarlinLinearMethod if use_marlin_kernel else (GPTQLinearMethod) ) with vllm_runner( model_id, dtype=torch.float16, max_model_len=2048, enforce_eager=True ) as llm: def check_model(model): for name, submodule in model.named_modules(): if name == "lm_head": assert isinstance(submodule.quant_method, linear_method_cls) elif name == "model.layers.0.self_attn.qkv_proj": # The first layer is quantized using bits=4, group_size=128 # desc_act=True assert isinstance(submodule.quant_method, linear_method_cls) config = submodule.quant_method.quant_config assert config.weight_bits == 4 assert config.group_size == 128 assert config.desc_act elif name == "model.layers.1.self_attn.qkv_proj": # The second layer is quantized using bits=8, group_size=32 # desc_act=False assert isinstance(submodule.quant_method, linear_method_cls) config = submodule.quant_method.quant_config assert ( get_dynamic_override(config, layer_name=name, key="bits") == 8 ) assert ( get_dynamic_override(config, layer_name=name, key="group_size") == 32 ) assert not get_dynamic_override( config, layer_name=name, key="desc_act" ) elif ( name == "model.layers.2.self_attn.qkv_proj" or name == "model.layers.2.mlp.gate_up_proj" ): # All other layers (layer index >= 2) are not quantized assert isinstance(submodule.quant_method, UnquantizedLinearMethod) llm.apply_model(check_model)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_ipex_quant.py
tests/quantization/test_ipex_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Test model set-up and inference for quantized HF models supported on the CPU/GPU backend using IPEX (including AWQ/GPTQ). Validating the configuration and printing results for manual checking. Run `pytest tests/quantization/test_ipex_quant.py`. """ import pytest from vllm.platforms import current_platform MODELS = [ "AMead10/Llama-3.2-1B-Instruct-AWQ", "shuyuej/Llama-3.2-1B-Instruct-GPTQ", # with g_idx ] DTYPE = ["bfloat16"] @pytest.mark.skipif( not current_platform.is_cpu() and not current_platform.is_xpu(), reason="only supports Intel CPU/XPU backend.", ) @pytest.mark.parametrize("model", MODELS) @pytest.mark.parametrize("dtype", DTYPE) def test_ipex_quant(vllm_runner, model, dtype): with vllm_runner(model, dtype=dtype, enforce_eager=True) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=4) assert output print(output)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_blackwell_moe.py
tests/quantization/test_blackwell_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json import os from typing import Any import pytest from tests.utils import RemoteOpenAIServer from vllm.platforms import current_platform if not current_platform.is_device_capability_family(100): pytest.skip( "This test only runs on Blackwell GPUs (SM10x).", allow_module_level=True ) @pytest.fixture(scope="module", autouse=True) def set_test_environment(): """Sets environment variables required for this test module.""" # Make sure TRTLLM attention is available os.environ["VLLM_HAS_FLASHINFER_CUBIN"] = "1" # Set compilation threads to 16 to speed up startup os.environ["FLASHINFER_NVCC_THREADS"] = "16" # Overide the backbone layers to 4 for faster startup HF_OVERRIDE_TEXT = { "num_layers": 4, "num_hidden_layers": 4, } HF_OVERRIDE_MM = { "text_config": {"num_layers": 4, "num_hidden_layers": 4}, } def can_initialize( model: str, hf_overrides: dict[str, Any] | None = None, extra_args: list[str] | None = None, ): # Server arguments extra_args = extra_args if extra_args is not None else [] server_args = [ "--max-model-len", "2048", "--max-num-batched-tokens", "256", "--load-format", "dummy", "--trust-remote-code", "--limit-mm-per-prompt", json.dumps({"image": 0}), *extra_args, ] # Launch server and make a simple request with RemoteOpenAIServer( model, server_args, max_wait_seconds=1500, # Due to FlashInfer compile override_hf_configs=hf_overrides, ) as server: client = server.get_client() # Make a simple request to verify the server works completion = client.completions.create( model=model, prompt=["Hello, World!"], temperature=0, max_tokens=2, ) print(completion) assert completion.choices[0].text is not None ## Llama4 ## @pytest.mark.skip( reason=( "RuntimeError: run_moe() Expected a value of type " "'Optional[List[Tensor]]' for argument '_9' but instead found type " "'list'." ) ) def test_llama4_fp8_tensor_moe_flashinfer_cutlass(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_FP8", "1") monkeypatch.setenv("VLLM_FLASHINFER_MOE_BACKEND", "throughput") can_initialize( "nvidia/Llama-4-Scout-17B-16E-Instruct-FP8", hf_overrides=HF_OVERRIDE_MM ) def test_llama4_fp8_tensor_moe_flashinfer_trtllm(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_FP8", "1") monkeypatch.setenv("VLLM_FLASHINFER_MOE_BACKEND", "latency") can_initialize( "nvidia/Llama-4-Scout-17B-16E-Instruct-FP8", hf_overrides=HF_OVERRIDE_MM ) def test_llama4_nvfp4_moe_flashinfer_cutlass(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_FP4", "1") monkeypatch.setenv("VLLM_FLASHINFER_MOE_BACKEND", "throughput") can_initialize( "nvidia/Llama-4-Scout-17B-16E-Instruct-FP4", hf_overrides=HF_OVERRIDE_MM ) def test_llama4_nvfp4_moe_flashinfer_trtllm(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_FP4", "1") monkeypatch.setenv("VLLM_FLASHINFER_MOE_BACKEND", "latency") can_initialize( "nvidia/Llama-4-Scout-17B-16E-Instruct-FP4", hf_overrides=HF_OVERRIDE_MM ) ## DeepSeekV3 ## def test_deepseek_fp8_block_moe_deep_gemm(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_DEEP_GEMM", "1") can_initialize("deepseek-ai/DeepSeek-V3.1", hf_overrides=HF_OVERRIDE_TEXT) @pytest.mark.skip( reason=( "Known issue: lack of kernel support. " "Expected failure: assert self.block_quant is None" ) ) def test_deepseek_fp8_block_moe_flashinfer_cutlass(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_FP8", "1") monkeypatch.setenv("VLLM_FLASHINFER_MOE_BACKEND", "throughput") can_initialize("deepseek-ai/DeepSeek-V3.1", hf_overrides=HF_OVERRIDE_TEXT) def test_deepseek_fp8_block_moe_flashinfer_trtllm(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_FP8", "1") monkeypatch.setenv("VLLM_FLASHINFER_MOE_BACKEND", "latency") can_initialize("deepseek-ai/DeepSeek-V3.1", hf_overrides=HF_OVERRIDE_TEXT) def test_deepseek_nvfp4_moe_flashinfer_cutlass(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_FP4", "1") monkeypatch.setenv("VLLM_FLASHINFER_MOE_BACKEND", "throughput") can_initialize("nvidia/DeepSeek-R1-0528-FP4-v2", hf_overrides=HF_OVERRIDE_TEXT) def test_deepseek_nvfp4_moe_flashinfer_trtllm(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_FP4", "1") monkeypatch.setenv("VLLM_FLASHINFER_MOE_BACKEND", "latency") can_initialize("nvidia/DeepSeek-R1-0528-FP4-v2", hf_overrides=HF_OVERRIDE_TEXT) ## GPT-OSS ## def test_gptoss_mxfp4bf16_moe_flashinfer(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_MXFP4_BF16", "1") can_initialize("openai/gpt-oss-20b", hf_overrides=HF_OVERRIDE_TEXT) def test_gptoss_mxfp4mxfp8_moe_flashinfer_cutlass(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_MXFP4_MXFP8_CUTLASS", "1") can_initialize("openai/gpt-oss-20b", hf_overrides=HF_OVERRIDE_TEXT) def test_gptoss_mxfp4mxfp8_moe_flashinfer_trtllm(monkeypatch: pytest.MonkeyPatch): monkeypatch.setenv("VLLM_USE_FLASHINFER_MOE_MXFP4_MXFP8", "1") can_initialize("openai/gpt-oss-20b", hf_overrides=HF_OVERRIDE_TEXT) def test_gptoss_eager(monkeypatch: pytest.MonkeyPatch): can_initialize( "openai/gpt-oss-20b", hf_overrides=HF_OVERRIDE_TEXT, extra_args=["--enforce-eager"], )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_rtn.py
tests/quantization/test_rtn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright © 2025, Oracle and/or its affiliates. """Tests RTN quantization startup and generation, doesn't test correctness """ import pytest from tests.quantization.utils import is_quant_method_supported MODELS = [ "ai21labs/Jamba-tiny-dev", # MoE model ] @pytest.mark.skipif( not is_quant_method_supported("rtn"), reason="RTN is not supported on this GPU type.", ) @pytest.mark.parametrize("model", MODELS) @pytest.mark.parametrize("dtype", ["bfloat16"]) @pytest.mark.parametrize("max_tokens", [10]) def test_model_rtn_startup( hf_runner, vllm_runner, example_prompts, model: str, dtype: str, max_tokens: int, ) -> None: with vllm_runner( model, enforce_eager=True, dtype=dtype, quantization="rtn" ) as vllm_model: vllm_model.generate_greedy(example_prompts, max_tokens)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_cpu_wna16.py
tests/quantization/test_cpu_wna16.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from vllm.platforms import current_platform if not current_platform.is_cpu(): pytest.skip("skipping CPU-only tests", allow_module_level=True) MODELS = [ "TheBloke/TinyLlama-1.1B-Chat-v1.0-AWQ", "TheBloke/TinyLlama-1.1B-Chat-v1.0-GPTQ", # with g_idx ] DTYPE = ["bfloat16"] @pytest.mark.parametrize("model", MODELS) @pytest.mark.parametrize("dtype", DTYPE) def test_ipex_quant(vllm_runner, model, dtype): with vllm_runner(model, dtype=dtype) as llm: output = llm.generate_greedy(["The capital of France is"], max_tokens=32) assert output print(output)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/quantization/test_gptq_v2.py
tests/quantization/test_gptq_v2.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests whether vllm correctly load and run gptq_v2 format checkpoints. Run `pytest tests/quantization/test_gptq_v2.py --forked`. """ import pytest import torch from transformers import AutoTokenizer from vllm import SamplingParams from vllm.model_executor.layers.quantization.gptq import GPTQLinearMethod # A dummy small model quantized by GPTQModel, stored in GPTQ v2 format MODELS = ["XXXXyu/Qwen3-1.7B-w2g64-gptq_v2"] # Generate multiple sequences for testing, because an 1.7B 2-bit model # cannot always generate normal texts. N_SEQ = 5 @pytest.mark.parametrize("model_id", MODELS) def test_model_load(vllm_runner, model_id, monkeypatch): # `LLM.apply_model` requires pickling a function. monkeypatch.setenv("VLLM_ALLOW_INSECURE_SERIALIZATION", "1") # Only check the default GPTQ linear method (used for 2/3-bit models). # 4/8-bit linear methods like Marlin already support gptq_v2. linear_method_cls = GPTQLinearMethod with vllm_runner(model_id, dtype=torch.float16, max_model_len=512) as llm: def check_model(model_id): for name, submodule in model_id.named_modules(): # Could check more modules if necessary if name == "model_id.layers.0.self_attn.qkv_proj": assert isinstance(submodule.quant_method, linear_method_cls) config = submodule.quant_method.quant_config assert config.checkpoint_format == "gptq_v2" assert submodule.quant_method.use_v2_format # Just break since currently we only check 1 module break # Check if gptq_v2 format is correctly loaded llm.apply_model(check_model) @pytest.mark.parametrize("model_id", MODELS) def test_model_inference(vllm_runner, model_id): # Prepare prompt to test the model's generation result. prompt = "What is the meaning of life?" messages = [ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": prompt}, ] tokenizer = AutoTokenizer.from_pretrained(model_id) text = tokenizer.apply_chat_template( messages, tokenize=False, add_generation_prompt=True, enable_thinking=False, # If thinking model, set it to false ) sampling_params = SamplingParams( n=N_SEQ, max_tokens=128, temperature=0.7, top_p=0.8, top_k=20, min_p=0, presence_penalty=2, ) with vllm_runner(model_id, dtype=torch.float16, max_model_len=512) as llm: # Generate a response to verify inference correctness output = llm.generate(text, sampling_params) # Make sure the output exists assert output assert output[0][1] assert len(output[0][1]) == N_SEQ def has_normal_char_distribution(texts, min_len): for text in texts: # Response too short if len(text) < min_len: return False # Basic ratio checks letters = sum(c.isalpha() for c in text) spaces = sum(c.isspace() for c in text) total = len(text) letter_ratio = letters / total space_ratio = spaces / total # At least 1 normal text should exist within output sequences # Normal text should be mostly letters with reasonable spacing # Some magic numbers, could be adjusted if 0.5 <= letter_ratio <= 0.9 and 0.01 <= space_ratio <= 0.3: return True # No sequence contains normal text, output might be broken return False # Apply some simple checks for giberish output # Print the output sequences if failed assert has_normal_char_distribution(output[0][1], 5), output[0][1]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/tpu/test_moe_pallas.py
tests/tpu/test_moe_pallas.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Tests for the Pallas MOE implementation. Run `pytest tests/kernels/moe/test_moe_pallas.py`. """ import pytest import torch import torch_xla from vllm.model_executor.layers.fused_moe.moe_pallas import fused_moe as pallas_moe from vllm.model_executor.layers.fused_moe.moe_torch_iterative import ( fused_moe as torch_moe, ) from vllm.platforms import current_platform if not current_platform.is_tpu(): pytest.skip("This test needs a TPU.", allow_module_level=True) NUM_EXPERTS = [8, 64] EP_SIZE = [1] TOP_KS = [2, 6] # The Pallas GMM kernel requires num_tokens * topk to be a multiple of 16 @pytest.mark.parametrize("m", [8, 16, 64, 2048]) @pytest.mark.parametrize("n", [128, 1024, 2048]) @pytest.mark.parametrize("k", [128, 511, 1024]) @pytest.mark.parametrize("e", NUM_EXPERTS) @pytest.mark.parametrize("topk", TOP_KS) @pytest.mark.parametrize("ep_size", EP_SIZE) @pytest.mark.parametrize("dtype", [torch.bfloat16]) def test_pallas_moe( m: int, n: int, k: int, e: int, topk: int, ep_size: int, dtype: torch.dtype, ): import torch_xla.core.xla_model as xm with torch.device(xm.xla_device()): a = torch.randn((m, k), dtype=dtype) / 10 w1 = torch.randn((e, 2 * n, k), dtype=dtype) / 10 w2 = torch.randn((e, k, n), dtype=dtype) / 10 score = torch.randn((m, e), dtype=dtype) # TODO: Support ep if ep_size > 1: pytest.skip("No support for ep_size > 1 yet") else: e_map = None # Run both implementations torch_output = torch_moe( hidden_states=a, w1=w1, w2=w2, gating_output=score, topk=topk, global_num_experts=e, expert_map=e_map, renormalize=False, ) pallas_output = pallas_moe( hidden_states=a, w1=w1, w2=w2, gating_output=score, topk=topk, global_num_experts=e, expert_map=e_map, renormalize=False, ) torch_xla.sync(wait=False) # Compare outputs torch.testing.assert_close( pallas_output.cpu(), torch_output.cpu(), atol=2e-2, rtol=0, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/tpu/test_custom_dispatcher.py
tests/tpu/test_custom_dispatcher.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from vllm.config import CompilationMode from ..utils import compare_two_settings # --enforce-eager on TPU causes graph compilation # this times out default Health Check in the MQLLMEngine, # so we set the timeout here to 30s def test_custom_dispatcher(monkeypatch: pytest.MonkeyPatch): with monkeypatch.context() as m: m.setenv("VLLM_RPC_TIMEOUT", "30000") compare_two_settings( "Qwen/Qwen2.5-1.5B-Instruct", arg1=[ "--max-model-len=256", "--max-num-seqs=32", "--enforce-eager", f"-O{CompilationMode.DYNAMO_TRACE_ONCE}", ], arg2=[ "--max-model-len=256", "--max-num-seqs=32", "--enforce-eager", f"-O{CompilationMode.STOCK_TORCH_COMPILE}", ], env1={}, env2={}, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/tpu/test_compilation.py
tests/tpu/test_compilation.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import glob import os import tempfile import depyf def test_tpu_compilation(): temp_dir = tempfile.mkdtemp() with depyf.prepare_debug(temp_dir): from vllm import LLM, SamplingParams prompts = [ "A robot may not injure a human being", "It is only with the heart that one can see rightly;", "The greatest glory in living lies not in never falling,", ] answers = [ " or, through inaction", " what is essential ", " but in rising ", ] # Currently, top-p sampling is disabled. `top_p` should be 1.0. N = 1 sampling_params = SamplingParams(temperature=0.7, top_p=1.0, n=N, max_tokens=16) llm = LLM( model="Qwen/Qwen2-1.5B-Instruct", max_num_batched_tokens=256, max_model_len=256, max_num_seqs=32, enforce_eager=False, ) outputs = llm.generate(prompts, sampling_params) for output, answer in zip(outputs, answers): prompt = output.prompt generated_text = output.outputs[0].text print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}") assert generated_text.startswith(answer) compiled_codes = sorted( glob.glob(os.path.join(temp_dir, "__transformed_code*for_forward.py")) ) for i, compiled_code in enumerate(compiled_codes): print("{} file: {}".format(i + 1, compiled_code)) # We should only trigger Dynamo compilation 2 times: # 1. Forward pass without kv_caches # 2. Forward pass with kv_caches # Check we have 2 compiled codes assert len(compiled_codes) == 2 kv_cache_prefix = "kv_cache" attn_prefix = "ragged_paged_attention" def extract_compiled_index(s): parts = s.replace(".", "_").split("_") numbers = [int(part) for part in parts if part.isdigit()] return numbers[0] # Check all the compilations are as expected. The dump files include the # captured graph for the forward function of the nn.Module. compiled_fns = sorted( glob.glob(os.path.join(temp_dir, "__compiled_fn*Forward_graph*.py")), key=lambda s: extract_compiled_index(s), ) for i, compiled_fn in enumerate(compiled_fns): print("{} file: {}".format(i + 1, compiled_fn)) # The first compilation should not have any kv_caches with open(compiled_fns[0]) as f: content = f.read() assert kv_cache_prefix not in content # The second compilation should have kv_caches and the # ragged_paged_attention with open(compiled_fns[1]) as f: content = f.read() assert kv_cache_prefix in content and attn_prefix in content
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/tpu/test_quantization_accuracy.py
tests/tpu/test_quantization_accuracy.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass import lm_eval import pytest TASK = "gsm8k" FILTER = "exact_match,strict-match" RTOL = 0.03 @dataclass class GSM8KAccuracyTestConfig: model_name: str expected_value: float def get_model_args(self) -> str: return f"pretrained={self.model_name},max_model_len=4096,max_num_seqs=32" # NOTE: Accuracy scores measured on GPUs. ACCURACY_CONFIGS = [ GSM8KAccuracyTestConfig( model_name="neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8", expected_value=0.76, ), # no bias # NOTE(rob): We cannot re-initialize vLLM in the same process for TPU, # so only one of these tests can run in a single call to pytest. As # a follow-up, move this into the LM-EVAL section of the CI. # GSM8KAccuracyTestConfig( # model_name="neuralmagic/Qwen2-7B-Instruct-quantized.w8a8", # expected_value=0.66), # bias in QKV layers ] @pytest.mark.parametrize("config", ACCURACY_CONFIGS) def test_gsm8k_correctness(config: GSM8KAccuracyTestConfig): results = lm_eval.simple_evaluate( model="vllm", model_args=config.get_model_args(), tasks="gsm8k", batch_size="auto", ) EXPECTED_VALUE = config.expected_value measured_value = results["results"][TASK][FILTER] assert ( measured_value - RTOL < EXPECTED_VALUE and measured_value + RTOL > EXPECTED_VALUE ), f"Expected: {EXPECTED_VALUE} | Measured: {measured_value}"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/tpu/__init__.py
tests/tpu/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/tpu/lora/__init__.py
tests/tpu/lora/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/tpu/lora/test_lora.py
tests/tpu/lora/test_lora.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest from torch_xla._internal import tpu import vllm from vllm.lora.request import LoRARequest # This file contains tests to ensure that LoRA works correctly on the TPU # backend. We use a series of custom trained adapters for Qwen2.5-3B-Instruct # for this. The adapters are: # Username6568/Qwen2.5-3B-Instruct-1_plus_1_equals_x_adapter, where x ranges # from 1 to 4. # These adapters are trained using a standard huggingface peft training script, # where all the inputs are "What is 1+1? \n" and all the outputs are "x". We run # 100 training iterations with a training batch size of 100. def setup_vllm(num_loras: int, tp: int) -> vllm.LLM: return vllm.LLM( model="Qwen/Qwen2.5-3B-Instruct", max_model_len=256, max_num_seqs=8, tensor_parallel_size=tp, enable_lora=True, max_loras=num_loras, max_lora_rank=8, ) TPU_TENSOR_PARALLEL_SIZES = ( [1, tpu.num_available_chips()] if tpu.num_available_chips() > 1 else [1] ) @pytest.mark.parametrize("tp", TPU_TENSOR_PARALLEL_SIZES) def test_single_lora(tp: int): """ This test ensures we can run a single LoRA adapter on the TPU backend. We run "Username6568/Qwen2.5-3B-Instruct-1_plus_1_equals_1_adapter" which will force Qwen2.5-3B-Instruct to claim 1+1=1. """ llm = setup_vllm(1, tp) prompt = "What is 1+1? \n" lora_request = LoRARequest( "lora_adapter_1", 1, "Username6568/Qwen2.5-3B-Instruct-1_plus_1_equals_1_adapter", ) output = ( llm.generate( prompt, sampling_params=vllm.SamplingParams(max_tokens=256, temperature=0), lora_request=lora_request, )[0] .outputs[0] .text ) answer = output.strip()[0] assert answer.isdigit() assert int(answer) == 1 @pytest.mark.parametrize("tp", TPU_TENSOR_PARALLEL_SIZES) def test_lora_hotswapping(tp: int): """ This test ensures we can run multiple LoRA adapters on the TPU backend, even if we only have space to store 1. We run "Username6568/Qwen2.5-3B-Instruct-1_plus_1_equals_x_adapter" which will force Qwen2.5-3B-Instruct to claim 1+1=x, for a range of x. """ lora_name_template = "Username6568/Qwen2.5-3B-Instruct-1_plus_1_equals_{}_adapter" lora_requests = [ LoRARequest(f"lora_adapter_{i}", i, lora_name_template.format(i)) for i in range(1, 5) ] llm = setup_vllm(1, tp) prompt = "What is 1+1? \n" for i, req in enumerate(lora_requests): output = ( llm.generate( prompt, sampling_params=vllm.SamplingParams(max_tokens=256, temperature=0), lora_request=req, )[0] .outputs[0] .text ) answer = output.strip()[0] assert answer.isdigit() assert int(answer) == i + 1 @pytest.mark.parametrize("tp", TPU_TENSOR_PARALLEL_SIZES) def test_multi_lora(tp: int): """ This test ensures we can run multiple LoRA adapters on the TPU backend, when we have enough space to store all of them. We run "Username6568/Qwen2.5-3B-Instruct-1_plus_1_equals_x_adapter" which will force Qwen2.5-3B-Instruct to claim 1+1=x, for a range of x. """ lora_name_template = "Username6568/Qwen2.5-3B-Instruct-1_plus_1_equals_{}_adapter" lora_requests = [ LoRARequest(f"lora_adapter_{i}", i, lora_name_template.format(i)) for i in range(1, 5) ] llm = setup_vllm(4, tp) prompt = "What is 1+1? \n" for i, req in enumerate(lora_requests): output = ( llm.generate( prompt, sampling_params=vllm.SamplingParams(max_tokens=256, temperature=0), lora_request=req, )[0] .outputs[0] .text ) answer = output.strip()[0] assert answer.isdigit() assert int(output.strip()[0]) == i + 1
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/standalone_tests/lazy_imports.py
tests/standalone_tests/lazy_imports.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Description: Test the lazy import module # The utility function cannot be placed in `vllm.utils` # this needs to be a standalone script import sys # List of modules that should not be imported too early. # Lazy import `torch._inductor.async_compile` to avoid creating # too many processes before we set the number of compiler threads. # Lazy import `cv2` to avoid bothering users who only use text models. # `cv2` can easily mess up the environment. module_names = ["torch._inductor.async_compile", "cv2"] # set all modules in `module_names` to be None. # if we import any modules during `import vllm`, there would be a # hard error and nice stacktrace on the first import. for module_name in module_names: sys.modules[module_name] = None # type: ignore[assignment] import vllm # noqa
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/evals/gsm8k/gsm8k_eval.py
tests/evals/gsm8k/gsm8k_eval.py
#!/usr/bin/env python3 # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Isolated GSM8K evaluation script for vLLM serve endpoint. """ import argparse import ast import asyncio import json import os import time from collections.abc import Generator import aiohttp import numpy as np import regex as re import requests from tqdm.asyncio import tqdm INVALID = -9999999 def download_and_cache_file(url: str, filename: str | None = None) -> str: """Download and cache a file from a URL.""" if filename is None: filename = os.path.join("/tmp", url.split("/")[-1]) if os.path.exists(filename): return filename print(f"Downloading from {url} to {filename}") response = requests.get(url, stream=True) response.raise_for_status() with open(filename, "wb") as f: for chunk in response.iter_content(chunk_size=1024): f.write(chunk) return filename def load_gsm8k_data() -> tuple[list[dict], list[dict]]: """Load GSM8K train and test data""" train_url = "https://raw.githubusercontent.com/openai/grade-school-math/master/grade_school_math/data/train.jsonl" test_url = "https://raw.githubusercontent.com/openai/grade-school-math/master/grade_school_math/data/test.jsonl" train_file = download_and_cache_file(train_url) test_file = download_and_cache_file(test_url) train_data = list(read_jsonl(train_file)) test_data = list(read_jsonl(test_file)) return train_data, test_data def read_jsonl(filename: str) -> Generator[dict, None, None]: """Read a JSONL file.""" with open(filename) as fin: for line in fin: if not line.startswith("#"): yield json.loads(line) def get_answer_value(answer_str: str) -> int: """Extract the numerical answer from the response.""" answer_str = answer_str.replace(",", "") numbers = re.findall(r"\d+", answer_str) if len(numbers) < 1: return INVALID try: return ast.literal_eval(numbers[-1]) except SyntaxError: return INVALID async def call_vllm_api( session: aiohttp.ClientSession, prompt: str, temperature: float, max_tokens: int, stop: list[str] | None = None, url: str | None = None, seed: int | None = None, ) -> tuple[str, int]: """Call vLLM's OpenAI-compatible completions endpoint. Returns: Tuple of (response_text, completion_tokens) """ data = { "prompt": prompt, "temperature": temperature, "max_tokens": max_tokens, "stop": stop, } if seed is not None: data["seed"] = seed try: async with session.post(f"{url}/v1/completions", json=data) as response: response.raise_for_status() result = await response.json() text = result["choices"][0]["text"] completion_tokens = result.get("usage", {}).get("completion_tokens", 0) return text, completion_tokens except Exception as e: print(f"Error calling vLLM API: {e}") return "", 0 def evaluate_gsm8k( num_questions: int = 1319, num_shots: int = 5, max_tokens: int = 256, host: str = "http://127.0.0.1", port: int = 8000, temperature: float = 0.0, seed: int | None = 42, ) -> dict[str, float | int]: """ Evaluate GSM8K accuracy using vLLM serve endpoint. Returns dict with accuracy, invalid_rate, latency, etc. """ base_url = f"{host}:{port}" # Load GSM8K train and test data train_data, test_data = load_gsm8k_data() # Limit to available test questions num_questions = min(num_questions, len(test_data)) # Build few-shot examples from train split (like lm-eval does) few_shot_examples = "" for i in range(num_shots): few_shot_examples += ( f"Question: {train_data[i]['question']}\n" f"Answer: {train_data[i]['answer']}\n\n" ) # Prepare test questions and labels from test split questions = [] labels = [] for i in range(num_questions): questions.append(f"Question: {test_data[i]['question']}\nAnswer:") labels.append(get_answer_value(test_data[i]["answer"])) assert all(label != INVALID for label in labels), "Some labels are invalid" # Run evaluation async def run_async_evaluation(): states: list[str] = [""] * num_questions output_tokens: list[int] = [0] * num_questions async def get_answer(session: aiohttp.ClientSession, i: int) -> tuple[str, int]: prompt = few_shot_examples + questions[i] answer, tokens = await call_vllm_api( session=session, prompt=prompt, temperature=temperature, max_tokens=max_tokens, stop=["Question", "Assistant:", "<|separator|>"], url=base_url, seed=seed, ) states[i] = answer output_tokens[i] = tokens return answer, tokens async with aiohttp.ClientSession( timeout=aiohttp.ClientTimeout(total=600) ) as session: tasks = [get_answer(session, i) for i in range(num_questions)] await tqdm.gather(*tasks, desc="Evaluating") return states, output_tokens print(f"Running GSM8K evaluation: {num_questions} questions, {num_shots}-shot") tic = time.perf_counter() states, output_tokens = asyncio.run(run_async_evaluation()) latency = time.perf_counter() - tic # Compute metrics preds = [get_answer_value(state) for state in states] accuracy = np.mean(np.array(preds) == np.array(labels)) invalid_rate = np.mean(np.array(preds) == INVALID) total_output_tokens = sum(output_tokens) tokens_per_second = total_output_tokens / latency if latency > 0 else 0.0 result = { "accuracy": accuracy, "invalid_rate": invalid_rate, "latency": latency, "questions_per_second": num_questions / latency, "total_output_tokens": total_output_tokens, "tokens_per_second": tokens_per_second, "num_questions": num_questions, "num_shots": num_shots, "max_tokens": max_tokens, "timestamp": time.time(), } return result def main() -> None: parser = argparse.ArgumentParser(description="GSM8K evaluation for vLLM serve") parser.add_argument( "--num-shots", type=int, default=5, help="Number of few-shot examples" ) parser.add_argument( "--num-questions", type=int, default=1319, help="Number of questions to evaluate", ) parser.add_argument( "--max-tokens", type=int, default=256, help="Max tokens for generation" ) parser.add_argument("--host", type=str, default="http://127.0.0.1", help="Host URL") parser.add_argument("--port", type=int, default=8000, help="Port number") parser.add_argument( "--temperature", type=float, default=0.0, help="Temperature for generation" ) parser.add_argument( "--seed", type=int, default=42, help="Random seed for reproducibility" ) parser.add_argument("--save-results", type=str, help="Save results to JSON file") args = parser.parse_args() result = evaluate_gsm8k( num_questions=args.num_questions, num_shots=args.num_shots, max_tokens=args.max_tokens, host=args.host, port=args.port, temperature=args.temperature, seed=args.seed, ) # Print results to terminal print("\nResults:") print(f"Accuracy: {result['accuracy']:.3f}") print(f"Invalid responses: {result['invalid_rate']:.3f}") print(f"Total latency: {result['latency']:.3f} s") print(f"Questions per second: {result['questions_per_second']:.3f}") print(f"Total output tokens: {result['total_output_tokens']}") print(f"Output tokens per second: {result['tokens_per_second']:.3f}") # Optional file saving if args.save_results: with open(args.save_results, "w") as f: json.dump(result, f, indent=2) print(f"Results saved to {args.save_results}") if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/evals/gsm8k/conftest.py
tests/evals/gsm8k/conftest.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from pathlib import Path def pytest_addoption(parser): """Add custom command line options.""" parser.addoption( "--config-list-file", default="configs/models-small.txt", help="File containing list of config files to test", ) def pytest_generate_tests(metafunc): """Generate test parameters from config files.""" if "config_filename" in metafunc.fixturenames: config_list_file = metafunc.config.getoption("--config-list-file") # Handle both relative and absolute paths config_list_path = Path(config_list_file) if not config_list_path.is_absolute(): # If relative, try relative to test directory first test_dir_path = Path(__file__).parent / config_list_file if test_dir_path.exists(): config_list_path = test_dir_path else: # Try relative to current working directory config_list_path = Path.cwd() / config_list_file print(f"Looking for config list at: {config_list_path}") config_files = [] if config_list_path.exists(): # Determine config directory (same directory as the list file) config_dir = config_list_path.parent with open(config_list_path) as f: for line in f: line = line.strip() if line and not line.startswith("#"): config_path = config_dir / line print(f"Checking config file: {config_path}") if config_path.exists(): config_files.append(config_path) print(f" ✓ Found: {config_path}") else: print(f" ✗ Missing: {config_path}") else: print(f"Config list file not found: {config_list_path}") # Generate test parameters if config_files: metafunc.parametrize( "config_filename", config_files, ids=[config_file.stem for config_file in config_files], ) else: print("No config files found, test will be skipped")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/evals/gsm8k/test_gsm8k_correctness.py
tests/evals/gsm8k/test_gsm8k_correctness.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ GSM8K evaluation using vLLM server and isolated GSM8K script. Replacement for lm-eval-harness with better performance and control. Usage: pytest -s -v tests/evals/gsm8k/test_gsm8k_correctness.py \ --config-list-file=configs/models-small.txt """ import shlex import yaml from tests.utils import RemoteOpenAIServer from .gsm8k_eval import evaluate_gsm8k TOL = 0.08 # Absolute tolerance for accuracy comparison def run_gsm8k_eval(eval_config: dict, server_url: str) -> dict: """Run GSM8K evaluation using our isolated script.""" # Extract host and port from server URL if "://" in server_url: server_url = server_url.split("://")[1] host_port = server_url.split("/")[0] # Remove path if present if ":" in host_port: host, p = host_port.split(":") port = int(p) else: host = host_port port = 8000 # Add http:// prefix if not present if not host.startswith("http"): host = f"http://{host}" # Run GSM8K evaluation results = evaluate_gsm8k( num_questions=eval_config["num_questions"], num_shots=eval_config["num_fewshot"], host=host, port=port, ) return results def test_gsm8k_correctness(config_filename): """Test GSM8K correctness for a given model configuration.""" eval_config = yaml.safe_load(config_filename.read_text(encoding="utf-8")) # Parse server arguments from config (use shlex to handle quoted strings) server_args_str = eval_config.get("server_args", "") server_args = shlex.split(server_args_str) if server_args_str else [] # Add standard server arguments server_args.extend( [ "--trust-remote-code", ] ) env_dict = eval_config.get("env", None) print(f"Starting GSM8K evaluation for model: {eval_config['model_name']}") print(f"Expected metric threshold: {eval_config['accuracy_threshold']}") print(f"Number of questions: {eval_config['num_questions']}") print(f"Number of few-shot examples: {eval_config['num_fewshot']}") print(f"Server args: {' '.join(server_args)}") print(f"Environment variables: {env_dict}") # Launch server and run evaluation with RemoteOpenAIServer( eval_config["model_name"], server_args, env_dict=env_dict, max_wait_seconds=600, ) as remote_server: server_url = remote_server.url_for("v1") print(f"Server started at: {server_url}") results = run_gsm8k_eval(eval_config, server_url) measured_metric = results["accuracy"] expected_metric = eval_config["accuracy_threshold"] print(f"GSM8K Results for {eval_config['model_name']}:") print(f" Measured metric: {measured_metric:.4f}") print(f" Expected metric: {expected_metric:.4f}") print(f" Tolerance: {TOL:.4f}") print(f" Questions: {results['num_questions']}") print(f" Invalid rate: {results['invalid_rate']:.3f}") print(f" Latency: {results['latency']:.1f}s") print(f" QPS: {results['questions_per_second']:.1f}") # Verify metric is within tolerance assert measured_metric >= expected_metric - TOL, ( f"GSM8K metric too low: {measured_metric:.4f} < " f"{expected_metric:.4f} - {TOL:.4f} = {expected_metric - TOL:.4f}" ) print(f"✅ GSM8K test passed for {eval_config['model_name']}")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/evals/gsm8k/__init__.py
tests/evals/gsm8k/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/evals/gpt_oss/conftest.py
tests/evals/gpt_oss/conftest.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Pytest configuration for GPT-OSS evaluation tests. """ def pytest_addoption(parser): """Add command line options for pytest.""" parser.addoption("--model", action="store", help="Model name to evaluate") parser.addoption( "--metric", action="store", type=float, help="Expected metric threshold" ) parser.addoption( "--server-args", action="store", default="", help="Additional server arguments" )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/evals/gpt_oss/__init__.py
tests/evals/gpt_oss/__init__.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/evals/gpt_oss/test_gpqa_correctness.py
tests/evals/gpt_oss/test_gpqa_correctness.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ GPQA evaluation using vLLM server and GPT-OSS evaluation package. Usage: pytest -s -v tests/evals/gpt_oss/test_gpqa_correctness.py \ --model openai/gpt-oss-20b \ --metric 0.58 \ --server-args "--tensor-parallel-size 2" """ import subprocess import sys import regex as re from tests.utils import RemoteOpenAIServer TOL = 0.05 # Absolute tolerance for accuracy comparison def run_gpqa_eval(model_name: str, base_url: str) -> float: """Run GPQA evaluation using the gpt-oss evaluation package.""" # Build the command to run the evaluation cmd = [ sys.executable, "-m", "gpt_oss.evals", "--eval", "gpqa", "--model", model_name, "--reasoning-effort", "low", "--base-url", base_url, "--n-threads", "200", ] try: # Run the evaluation result = subprocess.run( cmd, text=True, capture_output=True, timeout=1800, # 30 minute timeout env={"OPENAI_API_KEY": "dummy"}, ) print("Evaluation process output:\n", result.stdout) # Parse the output to extract the score match = re.search(r"'metric':\s*([\d.]+)", result.stdout) if match: return float(match.group(1)) # If we still can't find it, raise an error raise ValueError( f"Could not parse score from evaluation output:\n{result.stdout}" ) except subprocess.TimeoutExpired as e: raise RuntimeError("Evaluation timed out") from e except subprocess.CalledProcessError as e: raise RuntimeError( f"Evaluation failed with exit code {e.returncode}:\n" f"stdout: {e.stdout}\nstderr: {e.stderr}" ) from e def test_gpqa_correctness(request): """Test GPQA correctness for GPT-OSS model.""" # Get command line arguments model_name = request.config.getoption("--model") expected_metric = request.config.getoption("--metric") server_args_str = request.config.getoption("--server-args") # Parse server arguments server_args = [] if server_args_str: server_args = server_args_str.split() # Add standard server arguments server_args.extend( [ "--trust-remote-code", ] ) print(f"Starting GPQA evaluation for model: {model_name}") print(f"Expected metric threshold: {expected_metric}") print(f"Server args: {' '.join(server_args)}") # Launch server and run evaluation with RemoteOpenAIServer( model_name, server_args, max_wait_seconds=1800 ) as remote_server: base_url = remote_server.url_for("v1") print(f"Server started at: {base_url}") measured_metric = run_gpqa_eval(model_name, base_url) print(f"GPQA Results for {model_name}:") print(f" Measured metric: {measured_metric:.4f}") print(f" Expected metric: {expected_metric:.4f}") print(f" Tolerance: {TOL:.4f}") # Verify metric is within tolerance assert measured_metric >= expected_metric - TOL, ( f"GPQA metric too low: {measured_metric:.4f} < " f"{expected_metric:.4f} - {TOL:.4f} = {expected_metric - TOL:.4f}" ) print(f"✅ GPQA test passed for {model_name}")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_graph_partition.py
tests/compile/test_graph_partition.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import operator import pytest import torch from torch.fx.experimental.proxy_tensor import make_fx from vllm.compilation.backends import split_graph def test_getitem_moved_to_producer_subgraph(): """ Test that getitem operations are moved to the same subgraph as their input, preventing tuple inputs to submodules. """ def model_fn(x: torch.Tensor) -> torch.Tensor: # torch.split returns a tuple, creating real getitem operations # Should become first submodule that produces tuple chunks = torch.split(x, x.shape[0] // 2, dim=0) # Following ops should become second submodule that consumes tuple result_0 = torch.relu(chunks[0]) result_1 = torch.relu(chunks[1]) return torch.cat([result_0, result_1], dim=0) x = torch.randn(4, 3) gm = make_fx(model_fn)(x) has_getitem = any( node.op == "call_function" and node.target == operator.getitem for node in gm.graph.nodes ) assert has_getitem, "Test setup failed: graph should contain getitem operations" # Split on tuple producer aten::split split_ops = ["aten::split.Tensor"] split_gm, split_items = split_graph(gm, split_ops) assert len(split_items) == 2, "Graph should be split into 2 submodules" for split_item in split_items: submodule = split_item.graph getitem_on_placeholder = [] for node in submodule.graph.nodes: if ( node.op == "call_function" and node.target == operator.getitem and node.args[0].op == "placeholder" ): getitem_on_placeholder.append(node) assert len(getitem_on_placeholder) == 0, ( f"Submodule {split_item.submod_name} has getitem operations on " f"placeholder nodes: {[n.name for n in getitem_on_placeholder]}. " "This means tuple inputs were not properly eliminated." ) new_x = torch.randn(4, 3) output_original = gm(new_x) output_split = split_gm(new_x) assert torch.allclose(output_original, output_split), "Output mismatch" def test_no_tuple_inputs_with_multiple_consumers(): """ Test that when a tuple is consumed by multiple split operations, getitem operations are properly moved to avoid tuple inputs. """ def model_fn(x: torch.Tensor) -> torch.Tensor: # torch.split returns a tuple, creating real getitem operations # Should become first submodule that produces tuple chunks = torch.split(x, x.shape[0] // 2, dim=0) # These should become second submodule consuming tuple result_1 = torch.relu(chunks[0]) result_2 = torch.relu(chunks[1]) # Artificial graph splitting point to create another # independent submodule that consumes tuple later # This would become the third submodule result_1 = torch.sigmoid(result_1) # Fourth submodule that consumes tuple result = torch.cat([chunks[0], chunks[1], result_1, result_2]) return result x = torch.randn(4, 3) gm = make_fx(model_fn)(x) has_getitem = any( node.op == "call_function" and node.target == operator.getitem for node in gm.graph.nodes ) assert has_getitem, "Test setup failed: graph should contain getitem operations" split_ops = ["aten::split.Tensor", "aten::sigmoid"] split_gm, split_items = split_graph(gm, split_ops) assert len(split_items) == 4, "Graph should be split into 4 submodules" for split_item in split_items: submodule = split_item.graph for node in submodule.graph.nodes: if ( node.op == "call_function" and node.target == operator.getitem and node.args[0].op == "placeholder" ): pytest.fail( f"Submodule {split_item.submod_name} has getitem on " f"placeholder {node.args[0].name}, indicating it receives " "a tuple input" ) new_x = torch.randn(4, 3) output_original = gm(new_x) output_split = split_gm(new_x) assert torch.allclose(output_original, output_split), "Output mismatch after split"
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_config.py
tests/compile/test_config.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import copy from contextlib import nullcontext from unittest.mock import patch import pytest from pydantic import ValidationError from vllm.compilation.counter import compilation_counter from vllm.compilation.fix_functionalization import FixFunctionalizationPass from vllm.config import CompilationConfig, CUDAGraphMode, ParallelConfig, VllmConfig from vllm.config.compilation import CompilationMode, PassConfig from vllm.engine.arg_utils import EngineArgs from vllm.platforms import current_platform from vllm.utils.torch_utils import ( _is_torch_equal_or_newer, is_torch_equal, ) # This import automatically registers `torch.ops.silly.attention` from . import silly_attention # noqa: F401 def test_version(): # Test the version comparison logic using the private function assert _is_torch_equal_or_newer("2.8.0.dev20250624+cu128", "2.8.0.dev") assert _is_torch_equal_or_newer("2.8.0a0+gitc82a174", "2.8.0.dev") assert _is_torch_equal_or_newer("2.8.0", "2.8.0.dev") assert _is_torch_equal_or_newer("2.8.1", "2.8.0.dev") assert not _is_torch_equal_or_newer("2.7.1", "2.8.0.dev") def test_get_raw_stream_patch(): """Test that get_raw_stream patch is applied only for torch 2.9.0 or 2.9.1.""" import builtins # Check if get_raw_stream exists in builtins has_patch = hasattr(builtins, "get_raw_stream") # Import torch to get actual version is_torch_2_9 = is_torch_equal("2.9.0") or is_torch_equal("2.9.1") if is_torch_2_9: # For torch 2.9.x, the patch should be applied assert has_patch, "get_raw_stream should be patched for torch 2.9.x" # Verify it's callable (it should be the _cuda_getCurrentRawStream function) get_raw_stream = builtins.get_raw_stream # type: ignore[attr-defined] assert callable(get_raw_stream) # Verify it's the correct function from torch._C from torch._C import _cuda_getCurrentRawStream assert get_raw_stream is _cuda_getCurrentRawStream def test_copy_pass(): vllm_config = VllmConfig() inductor_pass = FixFunctionalizationPass(vllm_config) copied_inductor_pass = copy.deepcopy(inductor_pass) assert ( copied_inductor_pass.compilation_config.use_inductor_graph_partition == vllm_config.compilation_config.use_inductor_graph_partition ) assert ( copied_inductor_pass.compilation_config.splitting_ops == vllm_config.compilation_config.splitting_ops ) def test_custom_op(): # proper syntax _ = CompilationConfig(custom_ops=["+quant_fp8", "-silu_and_mul"]) with pytest.raises(ValueError, match="Invalid syntax '"): _ = CompilationConfig(custom_ops=["quant_fp8"]) # forked needed to workaround https://github.com/vllm-project/vllm/issues/21073 @pytest.mark.forked # NB: We don't test VLLM_DISABLE_COMPILE_CACHE=0 because that depends # on the state of the cache directory on the current machine, which # may be influenced by other tests. @pytest.mark.parametrize("val", ["1"]) def test_VLLM_DISABLE_COMPILE_CACHE(vllm_runner, monkeypatch, val): # Disable multiprocessing so that the counter is in the same process monkeypatch.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "0") monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", val) compilation_config = { "cudagraph_mode": CUDAGraphMode.NONE, # speed things up a bit } with ( compilation_counter.expect( num_cache_entries_updated=0, num_compiled_artifacts_saved=0 ), # loading the model causes compilation (if enabled) to happen vllm_runner( "facebook/opt-125m", compilation_config=compilation_config, gpu_memory_utilization=0.4, ) as _, ): pass # forked needed to workaround https://github.com/vllm-project/vllm/issues/21073 @pytest.mark.forked @pytest.mark.parametrize( "cudagraph_mode,num_cudagraph_captured", [ (CUDAGraphMode.NONE, 0), (CUDAGraphMode.FULL_DECODE_ONLY, 1), (CUDAGraphMode.PIECEWISE, 13), (CUDAGraphMode.FULL_AND_PIECEWISE, 14), ], ) def test_use_cudagraphs( vllm_runner, monkeypatch, cudagraph_mode, num_cudagraph_captured ): # Disable multiprocessing so that the counter is in the same process monkeypatch.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "0") compilation_config = { "cudagraph_capture_sizes": [100], "cudagraph_mode": cudagraph_mode, } num_gpu_runner_capture_triggers = 1 if cudagraph_mode != CUDAGraphMode.NONE else 0 with ( compilation_counter.expect( num_graphs_seen=1, num_gpu_runner_capture_triggers=num_gpu_runner_capture_triggers, num_cudagraph_captured=num_cudagraph_captured, ), # loading the model causes compilation (if enabled) to happen vllm_runner( "facebook/opt-125m", compilation_config=compilation_config, gpu_memory_utilization=0.4, ) as _, ): pass # forked needed to workaround https://github.com/vllm-project/vllm/issues/21073 @pytest.mark.forked def test_stock_torch_compile(vllm_runner, monkeypatch): # Disable multiprocessing so that the counter is in the same process monkeypatch.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "0") with ( compilation_counter.expect(stock_torch_compile_count=1), # loading the model causes compilation (if enabled) to happen vllm_runner( "facebook/opt-125m", compilation_config={"mode": CompilationMode.STOCK_TORCH_COMPILE}, gpu_memory_utilization=0.4, ) as _, ): pass # forked needed to workaround https://github.com/vllm-project/vllm/issues/21073 @pytest.mark.forked def test_no_compilation(vllm_runner, monkeypatch): # Disable multiprocessing so that the counter is in the same process monkeypatch.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "0") with ( compilation_counter.expect(num_graphs_seen=0, stock_torch_compile_count=0), # loading the model causes compilation (if enabled) to happen vllm_runner( "facebook/opt-125m", compilation_config={"mode": CompilationMode.NONE}, gpu_memory_utilization=0.4, ) as _, ): pass # forked needed to workaround https://github.com/vllm-project/vllm/issues/21073 @pytest.mark.forked def test_enforce_eager(vllm_runner, monkeypatch): # Disable multiprocessing so that the counter is in the same process monkeypatch.setenv("VLLM_ENABLE_V1_MULTIPROCESSING", "0") with ( compilation_counter.expect(num_graphs_seen=0, stock_torch_compile_count=0), # loading the model causes compilation (if enabled) to happen vllm_runner( "facebook/opt-125m", enforce_eager=True, gpu_memory_utilization=0.4 ) as _, ): pass def test_splitting_ops_dynamic(): # Default config config = VllmConfig() # Default V1 config leaves cudagraph mode unset; splitting ops are only # populated when the engine decides to use piecewise compilation. assert config.compilation_config.cudagraph_mode == CUDAGraphMode.FULL_AND_PIECEWISE assert config.compilation_config.splitting_ops_contain_attention() # When use_inductor_graph_partition=True config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, use_inductor_graph_partition=True, splitting_ops=["vllm::unified_attention"], ) ) # with inductor partition we use splitting_ops directly for # partition rules assert config.compilation_config.splitting_ops == ["vllm::unified_attention"] # When attn_fusion pass enabled. config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, pass_config=PassConfig(fuse_attn_quant=True, eliminate_noops=True), custom_ops=["+quant_fp8"], cudagraph_mode=CUDAGraphMode.PIECEWISE, ) ) assert config.compilation_config.splitting_ops == [] # cudagraph mode also fall back to FULL assert config.compilation_config.cudagraph_mode == CUDAGraphMode.FULL # splitting_ops can not contain attention ops when attn_fusion # pass enabled. with pytest.raises(ValidationError): config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, pass_config=PassConfig(fuse_attn_quant=True, eliminate_noops=True), custom_ops=["+quant_fp8"], cudagraph_mode=CUDAGraphMode.PIECEWISE, # work around for accessing all attntion ops splitting_ops=CompilationConfig()._attention_ops, ) ) # When both use_inductor_graph_partition and attn_fusion pass enabled. config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, use_inductor_graph_partition=True, pass_config=PassConfig(fuse_attn_quant=True, eliminate_noops=True), custom_ops=["+quant_fp8"], cudagraph_mode=CUDAGraphMode.PIECEWISE, ) ) # With inductor graph partition, attn_fusion and splitting_ops # work together. Default splitting_ops include attention ops. assert config.compilation_config.splitting_ops_contain_attention() # fuse_attn_quant is directly supported under # use_inductor_graph_partition=True, and cudagraph_mode # is unchanged. assert config.compilation_config.cudagraph_mode == CUDAGraphMode.PIECEWISE def test_moe_splitting_ops_deepep_ht_inductor_partition(): # Inductor partition case: user-provided splitting_ops should be # preserved and MoE ops should be appended for DeepEP HT with dp>1. config = VllmConfig( parallel_config=ParallelConfig( all2all_backend="deepep_high_throughput", data_parallel_size=8, ), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, use_inductor_graph_partition=True, splitting_ops=[ "vllm::unified_attention", "vllm::moe_forward", "vllm::moe_forward_shared", ], ), ) splitting_ops = config.compilation_config.splitting_ops assert splitting_ops == [ "vllm::unified_attention", "vllm::moe_forward", "vllm::moe_forward_shared", ] def test_should_split(): import torch from vllm.compilation.partition_rules import should_split graph = torch.fx.Graph() node = torch.fx.Node( graph=graph, name="dummy_node", op="call_function", target=torch.ops.aten.add.default, args=(), kwargs={}, ) # supports OpOverloadPacket splitting_ops = ["aten::add"] assert should_split(node, splitting_ops) # supports OpOverload splitting_ops = ["aten::add.default"] assert should_split(node, splitting_ops) # supports OpOverload splitting_ops = ["aten::add.Tensor"] assert not should_split(node, splitting_ops) q, k, v, out = [torch.randn(1)] * 4 # supports custom ops as OpOverloadPacket node = torch.fx.Node( graph=graph, name="dummy_node", op="call_function", target=torch.ops.silly.attention, args=(q, k, v, out), kwargs={}, ) splitting_ops = ["silly::attention"] assert should_split(node, splitting_ops) # supports custom ops as OpOverload node = torch.fx.Node( graph=graph, name="dummy_node", op="call_function", target=torch.ops.silly.attention.default, args=(q, k, v, out), kwargs={}, ) splitting_ops = ["silly::attention"] assert should_split(node, splitting_ops) splitting_ops = ["silly::attention.default"] assert should_split(node, splitting_ops) @pytest.mark.skipif( not current_platform.support_static_graph_mode(), reason="Skip if not cudagraph mode supported", ) @pytest.mark.parametrize( ( "cudagraph_capture_sizes", "max_cudagraph_capture_size", "tp_size", "enable_sp", "max_num_batched_tokens", "cudagraph_mode", "expected_max_size", ), [ (None, None, 1, False, 2048, CUDAGraphMode.FULL_AND_PIECEWISE, 256), ([1, 2, 4], 4, 1, False, 2048, CUDAGraphMode.FULL_AND_PIECEWISE, 4), ( [1, 2, 4], 8, 1, False, 2048, CUDAGraphMode.FULL_AND_PIECEWISE, ValidationError, ), ([1, 256], None, 1, False, 2048, CUDAGraphMode.FULL_AND_PIECEWISE, 256), ([], None, 1, False, 2048, CUDAGraphMode.NONE, 0), (None, 0, 1, False, 2048, CUDAGraphMode.NONE, 0), # truncated to nearest multiple of 8 or 16 (None, 257, 1, False, 2048, CUDAGraphMode.FULL_AND_PIECEWISE, 256), # max from list ([1, 2, 4, 15], None, 1, False, 2048, CUDAGraphMode.FULL_AND_PIECEWISE, 15), # filtered out 15 due to SP ([1, 2, 4, 15], None, 2, True, 2048, CUDAGraphMode.FULL_AND_PIECEWISE, 4), # limited by the max_tokens ([1, 2, 4, 15], None, 1, False, 8, CUDAGraphMode.FULL_AND_PIECEWISE, 4), # the list should contain at least 1 element when use cudagraph ([], None, 1, False, 2048, CUDAGraphMode.FULL_AND_PIECEWISE, ValidationError), # the max capturing size should be >= 1 when use cudagraph (None, 0, 1, False, 2048, CUDAGraphMode.FULL_AND_PIECEWISE, ValidationError), ], ) def test_cudagraph_sizes_post_init( cudagraph_capture_sizes, max_cudagraph_capture_size, tp_size, enable_sp, max_num_batched_tokens, cudagraph_mode, expected_max_size, ): ctx = nullcontext() if expected_max_size == ValidationError: ctx = pytest.raises(expected_max_size) with ( ctx, patch("vllm.config.parallel.cuda_device_count_stateless", return_value=tp_size), ): compilation_config = CompilationConfig( cudagraph_capture_sizes=cudagraph_capture_sizes, max_cudagraph_capture_size=max_cudagraph_capture_size, pass_config=PassConfig( enable_sp=enable_sp, fuse_norm_quant=True, fuse_act_quant=True, eliminate_noops=True, ), cudagraph_mode=cudagraph_mode, ) engine_args = EngineArgs( model="facebook/opt-125m", tensor_parallel_size=tp_size, max_num_seqs=min(max_num_batched_tokens, 128), max_num_batched_tokens=max_num_batched_tokens, compilation_config=compilation_config, ) vllm_config = engine_args.create_engine_config() assert ( vllm_config.compilation_config.max_cudagraph_capture_size == expected_max_size ) def test_cached_compilation_config(): import torch from torch._inductor.utils import run_and_get_code from vllm.config import get_cached_compilation_config, set_current_vllm_config from vllm.model_executor.layers.quantization.input_quant_fp8 import QuantFP8 from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape dtype = torch.bfloat16 device = torch.device("cuda:0") batch_size, num_qo_heads, head_size = 8, 16, 128 # access and cache default compilation config # default compilation config does not contain +quant_fp8 custom op. If this is # used, the generated code would use inductor-generated triton kernel instead # of the custom op `torch.ops._C.static_scaled_fp8_quant`. get_cached_compilation_config() vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, custom_ops=["+quant_fp8"], ) ) # set_current_vllm_config should clear cached compilation config and # use the new compilation_config in vllm_config with set_current_vllm_config(vllm_config): query_quant = QuantFP8(static=True, group_shape=GroupShape.PER_TENSOR) query_quant = torch.compile(query_quant) _q_scale = torch.tensor(1.0, dtype=torch.float32, device="cuda") query = torch.randn( batch_size, num_qo_heads * head_size, dtype=dtype, device=device ) _, code = run_and_get_code(query_quant, query, _q_scale) code = " ".join(code) assert "torch.ops._C.static_scaled_fp8_quant.default(" in code
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_fusion.py
tests/compile/test_fusion.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch import vllm.plugins from vllm._aiter_ops import IS_AITER_FOUND, rocm_aiter_ops from vllm.compilation.fusion import FUSED_OPS, FusedRMSQuantKey, RMSNormQuantFusionPass from vllm.compilation.fx_utils import find_op_nodes from vllm.compilation.matcher_utils import QUANT_OPS from vllm.compilation.noop_elimination import NoOpEliminationPass from vllm.compilation.post_cleanup import PostCleanupPass from vllm.config import ( CompilationConfig, CompilationMode, ModelConfig, PassConfig, VllmConfig, ) from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.quantization.utils.fp8_utils import ( W8A8BlockFp8LinearOp, ) from vllm.model_executor.layers.quantization.utils.quant_utils import ( GroupShape, QuantKey, ScaleDesc, ) from vllm.model_executor.layers.quantization.utils.w8a8_utils import ( Fp8LinearOp, cutlass_block_fp8_supported, cutlass_fp8_supported, maybe_create_device_identity, ) from vllm.platforms import current_platform from vllm.utils.deep_gemm import is_deep_gemm_supported from ..utils import override_cutlass_fp8_supported from .backend import TestBackend FP8_DTYPE = current_platform.fp8_dtype() RMS_OP = torch.ops._C.rms_norm.default RMS_ADD_OP = torch.ops._C.fused_add_rms_norm.default class TestModel(torch.nn.Module): def __init__( self, hidden_size: int, eps: float, group_shape: GroupShape, use_aiter: bool = False, cuda_force_torch: bool = False, use_aiter_quant_op: bool = True, *args, **kwargs, ): super().__init__(*args, **kwargs) self.use_aiter = use_aiter self.use_aiter_quant_op = use_aiter_quant_op self.cuda_force_torch = cuda_force_torch self.group_shape = group_shape self.enable_quant_fp8_custom_op = None # Will be set later if applicable self.norm = [RMSNorm(hidden_size, eps) for _ in range(4)] # Setup quantization scale descriptor static = group_shape == GroupShape.PER_TENSOR and not use_aiter quant_scale = ScaleDesc(torch.float32, static, group_shape) self.quant_key = QuantKey(dtype=FP8_DTYPE, scale=quant_scale, symmetric=True) # Setup scales if static: self.scale = [torch.rand(1, dtype=torch.float32) for _ in range(3)] else: self.scale = [None for _ in range(3)] # Setup weights self.w = [ torch.rand(hidden_size, hidden_size).to(dtype=FP8_DTYPE) for _ in range(3) ] if not group_shape.is_per_group() or use_aiter: self.w = [self.w[0].t() for _ in range(3)] # Setup weight scales if group_shape.is_per_group(): scale_size = ( (hidden_size + 128 - 1) // 128 if use_aiter else hidden_size // group_shape[1] ) wscale_shape: tuple[int, ...] = (scale_size, scale_size) else: wscale_shape = (1,) self.wscale = [torch.rand(wscale_shape, dtype=torch.float32) for _ in range(3)] # Setup FP8 linear operation is_per_group = group_shape.is_per_group() if is_per_group and use_aiter: self.fp8_linear = W8A8BlockFp8LinearOp( weight_group_shape=GroupShape(128, 128), act_quant_group_shape=group_shape, use_aiter_and_is_supported=use_aiter_quant_op, ) # AITER blockwise doesn't use enable_quant_fp8_custom_op elif is_per_group: self.fp8_linear = W8A8BlockFp8LinearOp( weight_group_shape=GroupShape(group_shape[1], group_shape[1]), act_quant_group_shape=group_shape, cutlass_block_fp8_supported=cutlass_block_fp8_supported(), use_aiter_and_is_supported=False, ) self.enable_quant_fp8_custom_op = self.fp8_linear.input_quant_op.enabled() elif use_aiter: self.fp8_linear = Fp8LinearOp( act_quant_static=False, act_quant_group_shape=group_shape, ) self.fp8_linear.quant_fp8.use_aiter = use_aiter_quant_op self.enable_quant_fp8_custom_op = self.fp8_linear.quant_fp8.enabled() else: with override_cutlass_fp8_supported(not cuda_force_torch): self.fp8_linear = Fp8LinearOp( act_quant_static=static, act_quant_group_shape=group_shape, ) self.enable_quant_fp8_custom_op = self.fp8_linear.quant_fp8.enabled() self.enable_rms_norm_custom_op = self.norm[0].enabled() def forward(self, x): # avoid having graph input be an arg to a pattern directly x = resid = torch.relu(x) y = self.norm[0](x) x2 = self.fp8_linear.apply( y, self.w[0], self.wscale[0], input_scale=self.scale[0] ) # make sure resid is used for replacement to work y2, resid = self.norm[1](x2, resid) x3 = self.fp8_linear.apply( y2, self.w[1], self.wscale[1], input_scale=self.scale[1] ) y3, resid = self.norm[2](x3, resid) # use resid here x4 = self.fp8_linear.apply( y3, self.w[2], self.wscale[2], input_scale=self.scale[2] ) y4, resid = self.norm[3](x4, resid) # use resid here return y4 def ops_in_model_before(self): if ( self.use_aiter and self.group_shape.is_per_group() and current_platform.is_fp8_fnuz() ): return [rocm_aiter_ops.get_group_quant_op()] if self.use_aiter and self.group_shape.is_per_group(): return [torch.ops.vllm.triton_per_token_group_quant_fp8.default] if self.use_aiter and self.use_aiter_quant_op: return [rocm_aiter_ops.get_per_token_quant_op()] if self.use_aiter: return [QUANT_OPS[self.quant_key]] if self.enable_quant_fp8_custom_op: return [QUANT_OPS[self.quant_key]] return [torch.ops.aten.reciprocal] def ops_in_model_after(self): if self.use_aiter and self.group_shape.is_per_group(): from vllm.compilation.rocm_aiter_fusion import ( AiterFusedAddRMSFp8GroupQuantPattern, AiterRMSFp8GroupQuantPattern, ) return [ AiterFusedAddRMSFp8GroupQuantPattern.FUSED_OP, AiterRMSFp8GroupQuantPattern.FUSED_OP, ] if self.use_aiter: from vllm.compilation.rocm_aiter_fusion import ( AiterFusedAddRMSNormDynamicQuantPattern, AiterRMSNormDynamicQuantPattern, ) return [ AiterFusedAddRMSNormDynamicQuantPattern.FUSED_OP, AiterRMSNormDynamicQuantPattern.FUSED_OP, ] return [ FUSED_OPS[FusedRMSQuantKey(self.quant_key, True)], FUSED_OPS[FusedRMSQuantKey(self.quant_key, False)], ] def ops_in_model_before_partial(self): return ( [RMS_OP, RMS_ADD_OP] if self.enable_rms_norm_custom_op else [torch.ops.aten.rsqrt] ) GROUP_SHAPES = [ GroupShape.PER_TOKEN, GroupShape.PER_TENSOR, GroupShape(1, 128), GroupShape(1, 64), ] def _run_fusion_test( model, fusion_pass, vllm_config, dtype, hidden_size, num_tokens, ): """Helper function for common fusion test logic. Must be called within vllm_config context. """ noop_pass = NoOpEliminationPass(vllm_config) cleanup_pass = PostCleanupPass(vllm_config) backend = TestBackend(noop_pass, fusion_pass, cleanup_pass) backend2 = TestBackend(noop_pass, cleanup_pass) x = torch.rand(num_tokens, hidden_size) torch._dynamo.mark_dynamic(x, 0) model_fused = torch.compile(model, backend=backend) result_fused = model_fused(x) model_unfused = torch.compile(model, backend=backend2) result_unfused = model_unfused(x) if dtype == torch.float16: ATOL, RTOL = (2e-3, 2e-3) else: ATOL, RTOL = (1e-2, 1e-2) torch.testing.assert_close(result_fused, result_unfused, atol=ATOL, rtol=RTOL) assert fusion_pass.matched_count == 3 backend.check_before_ops(model.ops_in_model_before()) backend.check_after_ops(model.ops_in_model_after()) return backend, backend2 @pytest.mark.parametrize("dtype", [torch.float16, torch.bfloat16]) @pytest.mark.parametrize("hidden_size", [256]) @pytest.mark.parametrize("num_tokens", [257]) @pytest.mark.parametrize("eps", [1e-5, 1e-6]) @pytest.mark.parametrize("group_shape", GROUP_SHAPES) @pytest.mark.parametrize("enable_rms_norm_custom_op", [True, False]) @pytest.mark.parametrize("enable_quant_fp8_custom_op", [True, False]) # cuda_force_torch used to test torch code path on platforms that # cutlass_fp8_supported() == True. @pytest.mark.parametrize( "cuda_force_torch", [True, False] if cutlass_fp8_supported() else [True] ) @pytest.mark.skipif( not current_platform.is_cuda_alike(), reason="Only test on CUDA and ROCm" ) def test_fusion_rmsnorm_quant( dtype, hidden_size, num_tokens, eps, group_shape, enable_rms_norm_custom_op, enable_quant_fp8_custom_op, cuda_force_torch, ): if not enable_quant_fp8_custom_op and group_shape.is_per_group(): pytest.skip("Unsupported unwrapped quant fp8 op for blockwise quantization") if group_shape == GroupShape(1, 64) and ( cutlass_block_fp8_supported() or is_deep_gemm_supported() ): pytest.skip("Unsupported group shape 64 for CUTLASS/DeepGemm") custom_ops = [] if enable_rms_norm_custom_op: custom_ops.append("+rms_norm") if enable_quant_fp8_custom_op: custom_ops.append("+quant_fp8") vllm_config = VllmConfig( model_config=ModelConfig(dtype=dtype), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, custom_ops=custom_ops, pass_config=PassConfig( fuse_norm_quant=True, fuse_act_quant=True, eliminate_noops=True ), ), ) with vllm.config.set_current_vllm_config(vllm_config): # Setup device before model creation torch.set_default_device("cuda") torch.set_default_dtype(dtype) torch.manual_seed(1) maybe_create_device_identity() fusion_pass = RMSNormQuantFusionPass(vllm_config) model = TestModel( hidden_size=hidden_size, eps=eps, group_shape=group_shape, use_aiter=False, cuda_force_torch=cuda_force_torch, ) backend, _ = _run_fusion_test( model, fusion_pass, vllm_config, dtype, hidden_size, num_tokens ) backend.check_before_ops( model.ops_in_model_before_partial(), fully_replaced=False ) # If RMSNorm custom op is disabled (native/torch impl used), # there's a risk that the fused add doesn't get included in the # replacement and only the rms part gets fused with quant. # Hence, we check only 2 add nodes are left (final fused rmsnorm add). if not enable_rms_norm_custom_op: n_add_nodes = lambda g: sum(1 for _ in find_op_nodes(torch.ops.aten.add, g)) # 7 = 1 (RMS) + 3x2 (3xRMS_ADD, 2 each) assert n_add_nodes(backend.graph_pre_pass) == 7 assert n_add_nodes(backend.graph_post_pass) == 2 GROUP_SHAPE_QUANT_OPS_MATCHS = [ (GroupShape.PER_TOKEN, True), (GroupShape.PER_TOKEN, False), (GroupShape(1, 128), True), ] @pytest.mark.parametrize("dtype", [torch.bfloat16]) @pytest.mark.parametrize("hidden_size", [256]) @pytest.mark.parametrize("num_tokens", [257]) @pytest.mark.parametrize("eps", [1e-5, 1e-6]) @pytest.mark.parametrize( "group_shape, use_aiter_quant_op", GROUP_SHAPE_QUANT_OPS_MATCHS ) @pytest.mark.skipif( (not current_platform.is_rocm() or not IS_AITER_FOUND), reason="Only test on ROCm with aiter package installed", ) def test_aiter_fusion_rmsnorm_quant( dtype: torch.dtype, hidden_size: int, num_tokens: int, eps: float, group_shape: GroupShape, use_aiter_quant_op: bool, monkeypatch: pytest.MonkeyPatch, ): vllm_config = VllmConfig( model_config=ModelConfig(dtype=dtype), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, custom_ops=["+rms_norm", "+quant_fp8"], pass_config=PassConfig(fuse_norm_quant=True, eliminate_noops=True), ), ) with vllm.config.set_current_vllm_config(vllm_config), monkeypatch.context() as m: from vllm.compilation.rocm_aiter_fusion import RocmAiterRMSNormFusionPass m.setenv("VLLM_ROCM_USE_AITER", "1") rocm_aiter_ops.refresh_env_variables() torch.set_default_device("cuda") torch.set_default_dtype(dtype) torch.manual_seed(1) maybe_create_device_identity() fusion_pass = RocmAiterRMSNormFusionPass(vllm_config) model = TestModel( hidden_size=hidden_size, eps=eps, group_shape=group_shape, use_aiter=True, use_aiter_quant_op=use_aiter_quant_op, ) _run_fusion_test( model, fusion_pass, vllm_config, dtype, hidden_size, num_tokens )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_compile_ranges.py
tests/compile/test_compile_ranges.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from typing import Any import torch from torch import fx as fx from torch import nn # This import automatically registers `torch.ops.silly.attention` import tests.compile.silly_attention # noqa from vllm.compilation.counter import compilation_counter from vllm.compilation.decorators import support_torch_compile from vllm.compilation.inductor_pass import ( InductorPass, get_pass_context, ) from vllm.config import ( VllmConfig, set_current_vllm_config, ) from vllm.config.compilation import CompilationConfig, CompilationMode from vllm.config.scheduler import SchedulerConfig from vllm.config.utils import Range from vllm.forward_context import set_forward_context BATCH_SIZE = 64 MLP_SIZE = 128 @support_torch_compile class TestModel(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = "", **kwargs) -> None: super().__init__() def forward(self, x: torch.Tensor) -> torch.Tensor: x = x + x attn_output = torch.empty_like(x) torch.ops.silly.attention(x, x, x, attn_output) x = attn_output x = x * 3 return x @torch.inference_mode def run_model(vllm_config: VllmConfig, model: nn.Module, batch_sizes: list[int]): with set_forward_context({}, vllm_config=vllm_config): model(torch.randn(BATCH_SIZE, MLP_SIZE)) for batch_size in batch_sizes: model(torch.randn(batch_size, MLP_SIZE)) class PostGradRangeChecker(InductorPass): def __init__(self, ranges: list[Range]): self.ranges = ranges self.num_calls = 0 def __call__(self, graph: fx.Graph): compile_range = get_pass_context().compile_range assert compile_range in self.ranges, ( f"Compile range {compile_range} not in {self.ranges}" ) self.num_calls += 1 def uuid(self) -> str: state: dict[str, Any] = {} return InductorPass.hash_dict(state) def test_compile_ranges(use_fresh_inductor_cache): post_grad_range_checker = PostGradRangeChecker( [ Range(start=1, end=8), Range(start=16, end=16), Range(start=9, end=32), Range(start=64, end=64), Range(start=33, end=8192), ] ) torch.set_default_device("cuda") vllm_config = VllmConfig( scheduler_config=SchedulerConfig( max_num_batched_tokens=8192, max_model_len=8192, is_encoder_decoder=False, ), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, compile_ranges_split_points=[8, 32], compile_sizes=[16, 64, 128], inductor_compile_config={ "post_grad_custom_post_pass": post_grad_range_checker, }, ), ) with set_current_vllm_config(vllm_config): model = TestModel(vllm_config=vllm_config, prefix="").eval() # Number of compilations: 3 for each compile range + 2 compile sizes batch_sizes = [1, 4, 16, 24, 48, 64, 8192] with compilation_counter.expect( num_graphs_seen=1, num_piecewise_graphs_seen=1, num_backend_compilations=5, ): run_model(vllm_config, model, batch_sizes) assert post_grad_range_checker.num_calls == 5 def test_compile_config_get_compile_ranges(): compilation_config = CompilationConfig( compile_ranges_split_points=[8, 32], ) VllmConfig( scheduler_config=SchedulerConfig( max_num_batched_tokens=8192, max_model_len=8192, is_encoder_decoder=False, ), compilation_config=compilation_config, ) assert compilation_config.get_compile_ranges() == [ Range(start=1, end=8), Range(start=9, end=32), Range(start=33, end=8192), ] def test_inductor_cache_compile_ranges(monkeypatch, use_fresh_inductor_cache): # To force multiple compilations, we disable the compile cache monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", "1") post_grad_range_checker = PostGradRangeChecker( ranges=[ Range(start=1, end=8), Range(start=9, end=8192), ] ) scheduler_config = SchedulerConfig( max_num_batched_tokens=8192, max_model_len=8192, is_encoder_decoder=False, ) torch.set_default_device("cuda") def create_vllm_config(): return VllmConfig( scheduler_config=scheduler_config, compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, compile_ranges_split_points=[8], inductor_compile_config={ "post_grad_custom_post_pass": post_grad_range_checker, }, ), ) vllm_config_1 = create_vllm_config() with set_current_vllm_config(vllm_config_1): model1 = TestModel(vllm_config=vllm_config_1, prefix="").eval() batch_sizes = [1, 16] run_model(vllm_config_1, model1, batch_sizes) assert post_grad_range_checker.num_calls == 2 post_grad_range_checker.num_calls = 0 # Create a new vllm config with the new pass context vllm_config_2 = create_vllm_config() with set_current_vllm_config(vllm_config_2): model2 = TestModel(vllm_config=vllm_config_2, prefix="").eval() batch_sizes = [4, 32] run_model(vllm_config_2, model2, batch_sizes) # Check that cache is used, so the number of calls # should be 0 assert post_grad_range_checker.num_calls == 0
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_aot_compile.py
tests/compile/test_aot_compile.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import functools import multiprocessing import tempfile from contextlib import contextmanager import pytest import torch import vllm.model_executor.layers.activation from vllm.compilation.decorators import support_torch_compile from vllm.config import ( CompilationConfig, CompilationMode, VllmConfig, set_current_vllm_config, ) from vllm.envs import disable_envs_cache from vllm.forward_context import set_forward_context from vllm.utils.torch_utils import is_torch_equal_or_newer from ..utils import create_new_process_for_each_test def reference_fn(x: torch.Tensor): assert x.shape[0] <= 42 assert x.shape[0] % 2 == 0 for _ in range(3000): x = x + x.shape[0] return x @support_torch_compile class CompiledMod(torch.nn.Module): def __init__(self, **kwargs): super().__init__() def forward(self, x: torch.Tensor): return reference_fn(x) def make_vllm_config() -> VllmConfig: return VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, ) ) @contextmanager def use_vllm_config(vllm_config: VllmConfig): with set_forward_context({}, vllm_config), set_current_vllm_config(vllm_config): yield @pytest.mark.skipif( not is_torch_equal_or_newer("2.10.0.dev"), reason="requires torch 2.10" ) def test_no_dynamo_cache_entry(monkeypatch: pytest.MonkeyPatch): with monkeypatch.context() as m: vllm_config = make_vllm_config() args = (torch.randn(10, 10),) expected = reference_fn(*args) with use_vllm_config(vllm_config): m.setenv("VLLM_USE_AOT_COMPILE", "0") with ( pytest.raises(RuntimeError, match="Detected recompile"), torch.compiler.set_stance("fail_on_recompile"), ): CompiledMod(vllm_config=vllm_config)(*args) disable_envs_cache() m.setenv("VLLM_USE_AOT_COMPILE", "1") torch._dynamo.reset() with torch.compiler.set_stance("fail_on_recompile"): actual = CompiledMod(vllm_config=vllm_config)(*args) assert torch.allclose(actual, expected) @pytest.mark.skipif( not is_torch_equal_or_newer("2.10.0.dev"), reason="requires torch 2.10" ) def test_force_aot_load(monkeypatch: pytest.MonkeyPatch): with tempfile.TemporaryDirectory() as tmpdirname, monkeypatch.context() as m: args = (torch.randn(10, 10),) m.setenv("VLLM_USE_AOT_COMPILE", "1") m.setenv("VLLM_FORCE_AOT_LOAD", "1") m.setenv("VLLM_CACHE_ROOT", tmpdirname) vllm_config = make_vllm_config() with use_vllm_config(vllm_config), pytest.raises(FileNotFoundError): CompiledMod(vllm_config=vllm_config)(*args) @pytest.mark.skipif( not is_torch_equal_or_newer("2.10.0.dev"), reason="requires torch 2.10" ) def test_save_and_load(monkeypatch: pytest.MonkeyPatch): with monkeypatch.context() as m: args = (torch.randn(10, 10),) with tempfile.TemporaryDirectory() as tmpdirname: m.setenv("VLLM_CACHE_ROOT", tmpdirname) m.setenv("VLLM_USE_AOT_COMPILE", "1") vllm_config = make_vllm_config() with use_vllm_config(vllm_config): expected = CompiledMod(vllm_config=vllm_config)(*args) disable_envs_cache() m.setenv("VLLM_FORCE_AOT_LOAD", "1") vllm_config = make_vllm_config() with use_vllm_config(vllm_config): ret = CompiledMod(vllm_config=vllm_config)(*args) assert torch.allclose(ret, expected) @pytest.mark.skipif( not is_torch_equal_or_newer("2.10.0.dev"), reason="requires torch 2.10" ) def test_shape_env(monkeypatch: pytest.MonkeyPatch): """ Test that the shape environment is correctly serialized and preserved when loading from cache. """ with monkeypatch.context() as m: args = (torch.randn(10, 10),) with tempfile.TemporaryDirectory() as tmpdirname: m.setenv("VLLM_CACHE_ROOT", tmpdirname) m.setenv("VLLM_USE_AOT_COMPILE", "1") vllm_config = make_vllm_config() with use_vllm_config(vllm_config): compiled_mod = CompiledMod(vllm_config=vllm_config) compiled_mod(*args) artifacts = compiled_mod.aot_compiled_fn._artifacts guards_string = artifacts.compiled_fn.shape_env.format_guards() assert guards_string == " - s77 <= 42\n - Eq(Mod(s77, 2), 0)" disable_envs_cache() m.setenv("VLLM_FORCE_AOT_LOAD", "1") vllm_config = make_vllm_config() with use_vllm_config(vllm_config): compiled_mod = CompiledMod(vllm_config=vllm_config) compiled_mod(*args) artifacts = compiled_mod.aot_compiled_fn._artifacts guards_string = artifacts.compiled_fn.shape_env.format_guards() assert guards_string == " - s77 <= 42\n - Eq(Mod(s77, 2), 0)" @pytest.mark.skipif( not is_torch_equal_or_newer("2.10.0.dev"), reason="requires torch 2.10" ) @create_new_process_for_each_test("spawn") def test_gpt2_cache_hit(monkeypatch: pytest.MonkeyPatch): """ Test that compiling gpt2 twice results in a cache hit and capture torch dynamic symbol creations to ensure make_symbol not called on cache hit. """ import torch.fx.experimental.symbolic_shapes as symbolic_shapes_module from torch.utils._sympy.symbol import make_symbol from vllm import LLM create_symbol_counter = multiprocessing.Value("i", 0) original_make_symbol = make_symbol @functools.wraps(original_make_symbol) def counting_make_symbol(prefix, idx, **kwargs): with create_symbol_counter.get_lock(): create_symbol_counter.value += 1 return original_make_symbol(prefix, idx, **kwargs) symbolic_shapes_module.make_symbol = counting_make_symbol try: with monkeypatch.context() as m, tempfile.TemporaryDirectory() as tmpdirname: m.setenv("VLLM_CACHE_ROOT", tmpdirname) m.setenv("VLLM_USE_AOT_COMPILE", "1") # First compilation - initialize model and generate llm_model = LLM( model="gpt2", compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, ), max_model_len=256, ) llm_model.generate("Hello, my name is") assert create_symbol_counter.value == 2 create_symbol_counter.value = 0 # Clean up first model del llm_model disable_envs_cache() vllm.model_executor.layers.activation._ACTIVATION_REGISTRY._dict.clear() # Second compilation - should hit cache m.setenv("VLLM_FORCE_AOT_LOAD", "1") llm_model = LLM( model="gpt2", compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, ), max_model_len=256, ) llm_model.generate("Hello, my name is") assert create_symbol_counter.value == 0 finally: # Restore original method symbolic_shapes_module.make_symbol = original_make_symbol
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_dynamic_shapes_compilation.py
tests/compile/test_dynamic_shapes_compilation.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import gc import tempfile from contextlib import contextmanager import pytest import torch from vllm import LLM, SamplingParams from vllm.compilation.decorators import support_torch_compile from vllm.config import CompilationConfig, VllmConfig, set_current_vllm_config from vllm.config.compilation import ( CompilationMode, DynamicShapesConfig, DynamicShapesType, ) from vllm.forward_context import set_forward_context from vllm.tokenizers import get_tokenizer from vllm.utils.torch_utils import is_torch_equal_or_newer def get_test_models(): """Get list of models to test based on PyTorch version""" # TODO "Qwen/Qwen3-4B-Instruct-2507" fails Fix issue and support it. return ["gpt2", "Qwen/Qwen2-7B-Instruct", "meta-llama/Llama-3.1-8B"] @pytest.mark.parametrize("model_name", get_test_models()) @pytest.mark.parametrize( "shapes_type", [ DynamicShapesType.BACKED, DynamicShapesType.UNBACKED, DynamicShapesType.BACKED_SIZE_OBLIVIOUS, ], ) @pytest.mark.parametrize("use_aot_compile", ["0", "1"]) @pytest.mark.parametrize("use_bytecode_hook", [True, False]) @pytest.mark.parametrize("evaluate_guards", [False, True]) @pytest.mark.skipif( not is_torch_equal_or_newer("2.10.0.dev"), reason="requires torch 2.10" ) def test_dynamic_shapes_compilation( monkeypatch, model_name, shapes_type, use_aot_compile, use_bytecode_hook, evaluate_guards, ): """Test that all dynamic shapes types compile successfully""" if use_bytecode_hook and shapes_type == DynamicShapesType.UNBACKED: pytest.skip("UNBACKED dynamic shapes require VLLM_USE_BYTECODE_HOOK=0") if evaluate_guards and shapes_type == DynamicShapesType.UNBACKED: pytest.skip("unbacked dynamic shapes do not add guards") if evaluate_guards and use_aot_compile: pytest.skip("evaluate_guards requires use_aot_compile=0") monkeypatch.setenv("VLLM_USE_AOT_COMPILE", use_aot_compile) monkeypatch.setenv("VLLM_USE_BYTECODE_HOOK", "1" if use_bytecode_hook else "0") prompt = "Hello, my name is" print(f"Testing {shapes_type.name} dynamic shapes...") # Initialize the model with specific dynamic shapes configuration model = LLM( model=model_name, compilation_config={ "mode": CompilationMode.VLLM_COMPILE, "dynamic_shapes_config": { "type": shapes_type.value, "evaluate_guards": evaluate_guards, }, }, max_model_len=1024, ) output = model.generate(prompt) result = output[0].outputs[0].text # Example of setting the sampling parameters tokenizer = get_tokenizer(model_name) yes_tokens = tokenizer.encode("yes", add_special_tokens=False) no_tokens = tokenizer.encode("no", add_special_tokens=False) allowed_ids = list(set(yes_tokens + no_tokens)) sampling_params = SamplingParams( max_tokens=1, temperature=0, allowed_token_ids=allowed_ids ) output = model.generate( "answer with yes or no is " + result + " rubbish for prompt " + prompt + "?", sampling_params=sampling_params, ) result = output[0].outputs[0].text assert result == "yes" # Clean up GPU memory del model gc.collect() torch.cuda.empty_cache() torch.cuda.synchronize() print("GPU memory cleared") @pytest.mark.parametrize("use_aot_compile", ["0", "1"]) @pytest.mark.parametrize( "dynamic_shapes_type", [ DynamicShapesType.BACKED, DynamicShapesType.BACKED_SIZE_OBLIVIOUS, ], ) @pytest.mark.parametrize("evaluate_guards", [False, True]) def test_model_specialization_with_evaluate_guards( monkeypatch, use_aot_compile, dynamic_shapes_type, evaluate_guards ): """Test that evaluate_guards correctly detects shape specialization violations. """ if ( use_aot_compile == "1" and dynamic_shapes_type == DynamicShapesType.BACKED and evaluate_guards ): pytest.skip("evaluate_guards for backed does not work with aot_compile=1") @support_torch_compile class ModelWithSizeCheck(torch.nn.Module): def __init__(self, **kwargs): super().__init__() def forward(self, x: torch.Tensor): # This will cause specialization - torch.compile will guard on # sx.shape[0] if x.shape[0] >= 10: return x * 10 else: return x * 10 @support_torch_compile class ModelWithOneSizeCheck(torch.nn.Module): def __init__(self, **kwargs): super().__init__() def forward(self, x: torch.Tensor): # This will cause 0/1 specializations. if x.shape[0] == 0: return x * 10 if x.shape[0] == 1: return x * 10 else: return x * 10 @contextmanager def use_vllm_config(vllm_config: VllmConfig): with set_forward_context({}, vllm_config), set_current_vllm_config(vllm_config): yield monkeypatch.setenv("TOKENIZERS_PARALLELISM", "true") monkeypatch.setenv("VLLM_USE_AOT_COMPILE", use_aot_compile) monkeypatch.setenv("VLLM_USE_BYTECODE_HOOK", "0") # Create vllm config with the desired settings from vllm.config import CompilationMode vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, dynamic_shapes_config=DynamicShapesConfig( type=dynamic_shapes_type, evaluate_guards=evaluate_guards, ), ) ) def test(model_class, input1, input2, is_01_specialization=False): with ( torch.no_grad(), use_vllm_config(vllm_config), tempfile.TemporaryDirectory() as tmpdirname, ): monkeypatch.setenv("VLLM_CACHE_ROOT", tmpdirname) model = model_class(vllm_config=vllm_config).cuda() model(input1) if evaluate_guards and ( not ( is_01_specialization and dynamic_shapes_type == DynamicShapesType.BACKED ) ): # This should fail because guards were added. with pytest.raises(RuntimeError) as excinfo: model(input2) # Expected failure - guard was violated error_msg = str(excinfo.value) assert ( "GuardManager check failed" in error_msg or "Detected recompile when torch.compile stance" in error_msg ), error_msg else: model(input2) test(ModelWithSizeCheck, torch.randn(20, 10).cuda(), torch.randn(5, 10).cuda()) test(ModelWithSizeCheck, torch.randn(5, 10).cuda(), torch.randn(20, 10).cuda()) test( ModelWithOneSizeCheck, torch.randn(20, 10).cuda(), torch.randn(1, 10).cuda(), is_01_specialization=True, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_qk_norm_rope_fusion.py
tests/compile/test_qk_norm_rope_fusion.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from tests.compile.backend import TestBackend from vllm.attention.backends.abstract import AttentionType from vllm.attention.layer import Attention from vllm.compilation.matcher_utils import FLASHINFER_ROTARY_OP, RMS_OP, ROTARY_OP from vllm.compilation.noop_elimination import NoOpEliminationPass from vllm.compilation.post_cleanup import PostCleanupPass from vllm.compilation.qk_norm_rope_fusion import ( FUSED_QK_ROPE_OP, QKNormRoPEFusionPass, ) from vllm.config import ( CompilationConfig, CompilationMode, ModelConfig, PassConfig, VllmConfig, set_current_vllm_config, ) from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.rotary_embedding import RotaryEmbedding from vllm.platforms import current_platform RSQRT_OP = torch.ops.aten.rsqrt.default INDEX_SELECT_OP = torch.ops.aten.index.Tensor class QKNormRoPETestModel(torch.nn.Module): def __init__( self, *, num_heads: int, num_kv_heads: int, head_dim: int, eps: float, is_neox: bool, vllm_config: VllmConfig, dtype: torch.dtype, prefix: str = "model.layers.0.self_attn.attn", ) -> None: super().__init__() self.num_heads = num_heads self.num_kv_heads = num_kv_heads self.head_dim = head_dim self.q_size = num_heads * head_dim self.kv_size = num_kv_heads * head_dim self.rotary_dim = head_dim self.eps = eps self.dtype = dtype # Register layer metadata for the fusion pass via Attention. self.attn = Attention( num_heads=self.num_heads, head_size=self.head_dim, scale=1.0 / self.head_dim**0.5, num_kv_heads=self.num_kv_heads, cache_config=vllm_config.cache_config, prefix=prefix, attn_type=AttentionType.DECODER, ) self.q_norm = RMSNorm(self.head_dim, eps=self.eps) self.k_norm = RMSNorm(self.head_dim, eps=self.eps) self.rotary_emb = RotaryEmbedding( self.head_dim, rotary_dim=self.rotary_dim, max_position_embeddings=4096, base=10000, is_neox_style=is_neox, dtype=self.dtype, ) self.enable_rms_norm_custom_op = self.q_norm.enabled() self.enable_rope_custom_op = self.rotary_emb.enabled() def forward(self, qkv: torch.Tensor, positions: torch.Tensor): q, k, v = qkv.split([self.q_size, self.kv_size, self.kv_size], dim=-1) q_by_head = q.view(*q.shape[:-1], q.shape[-1] // self.head_dim, self.head_dim) q_by_head = self.q_norm(q_by_head) q = q_by_head.view(q.shape) k_by_head = k.view(*k.shape[:-1], k.shape[-1] // self.head_dim, self.head_dim) k_by_head = self.k_norm(k_by_head) k = k_by_head.view(k.shape) q, k = self.rotary_emb(positions, q, k) return q, k, v def ops_in_model_before(self) -> list[torch._ops.OpOverload]: ops = [] if self.enable_rms_norm_custom_op: ops.append(RMS_OP) else: ops.append(RSQRT_OP) if self.enable_rope_custom_op: if self.rotary_emb.use_flashinfer: ops.append(FLASHINFER_ROTARY_OP) else: ops.append(ROTARY_OP) else: ops.append(INDEX_SELECT_OP) return ops def ops_in_model_after(self) -> list[torch._ops.OpOverload]: return [FUSED_QK_ROPE_OP] @pytest.mark.parametrize("eps", [1e-5, 1e-6]) @pytest.mark.parametrize("is_neox", [True, False]) @pytest.mark.parametrize("enable_rms_norm_custom_op", [True, False]) @pytest.mark.parametrize("enable_rope_custom_op", [True]) @pytest.mark.parametrize("dtype", [torch.bfloat16, torch.float16]) @pytest.mark.skipif( not current_platform.is_cuda_alike(), reason="Only test on cuda and rocm platform", ) def test_qk_norm_rope_fusion( eps, is_neox, enable_rms_norm_custom_op, enable_rope_custom_op, dtype ): if not hasattr(torch.ops._C, "fused_qk_norm_rope"): pytest.skip("fused_qk_norm_rope custom op not available") torch.set_default_device("cuda") torch.set_default_dtype(dtype) torch.manual_seed(0) custom_ops: list[str] = [] if enable_rms_norm_custom_op: custom_ops.append("+rms_norm") if enable_rope_custom_op: custom_ops.append("+rotary_embedding") vllm_config = VllmConfig( model_config=ModelConfig(dtype=dtype), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, custom_ops=custom_ops, pass_config=PassConfig( enable_qk_norm_rope_fusion=True, eliminate_noops=True, ), ), ) num_heads, num_kv_heads, head_dim = 16, 4, 128 T = 5 with set_current_vllm_config(vllm_config): model = QKNormRoPETestModel( num_heads=num_heads, num_kv_heads=num_kv_heads, head_dim=head_dim, eps=eps, is_neox=is_neox, vllm_config=vllm_config, dtype=dtype, ) noop_pass = NoOpEliminationPass(vllm_config) fusion_pass = QKNormRoPEFusionPass(vllm_config) cleanup_pass = PostCleanupPass(vllm_config) backend = TestBackend(noop_pass, fusion_pass, cleanup_pass) backend_baseline = TestBackend(noop_pass, cleanup_pass) qkv = torch.randn(T, model.q_size + 2 * model.kv_size) pos = torch.arange(T, dtype=torch.long, device=qkv.device) qkv_unfused = qkv.clone() pos_unfused = pos.clone() torch._dynamo.mark_dynamic(qkv, 0) torch._dynamo.mark_dynamic(pos, 0) model_fused = torch.compile(model, backend=backend) q_fused, k_fused, v_fused = model_fused(qkv, pos) torch._dynamo.mark_dynamic(qkv_unfused, 0) torch._dynamo.mark_dynamic(pos_unfused, 0) model_unfused = torch.compile(model, backend=backend_baseline) q_unfused, k_unfused, v_unfused = model_unfused(qkv_unfused, pos_unfused) if dtype == torch.float16: ATOL, RTOL = (2e-3, 2e-3) else: ATOL, RTOL = (1e-2, 1e-2) torch.testing.assert_close(q_unfused, q_fused, atol=ATOL, rtol=RTOL) torch.testing.assert_close(k_unfused, k_fused, atol=ATOL, rtol=RTOL) torch.testing.assert_close(v_unfused, v_fused, atol=ATOL, rtol=RTOL) assert fusion_pass.matched_count == 1 backend.check_before_ops(model.ops_in_model_before()) backend.check_after_ops(model.ops_in_model_after())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_decorator.py
tests/compile/test_decorator.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch from torch import nn from vllm.compilation.counter import compilation_counter from vllm.compilation.decorators import ignore_torch_compile, support_torch_compile from vllm.config import ( CacheConfig, CompilationConfig, CompilationMode, CUDAGraphMode, VllmConfig, set_current_vllm_config, ) from vllm.forward_context import BatchDescriptor, set_forward_context from vllm.utils.torch_utils import is_torch_equal_or_newer # This import automatically registers `torch.ops.silly.attention` from . import silly_attention # noqa: F401 BATCH_SIZE = 32 MLP_SIZE = 128 @torch.inference_mode def run_model( vllm_config: VllmConfig, model: nn.Module, cudagraph_runtime_mode: CUDAGraphMode ): with set_forward_context({}, vllm_config=vllm_config): # warmup for the model with cudagraph_mode NONE model(torch.randn(BATCH_SIZE, MLP_SIZE).cuda()) # simulate cudagraphs capturing with set_forward_context( {}, vllm_config=vllm_config, cudagraph_runtime_mode=cudagraph_runtime_mode, batch_descriptor=BatchDescriptor( num_tokens=2, ), ): model(torch.randn(2, MLP_SIZE).cuda()) with set_forward_context( {}, vllm_config=vllm_config, cudagraph_runtime_mode=cudagraph_runtime_mode, batch_descriptor=BatchDescriptor( num_tokens=1, ), ): model(torch.randn(1, MLP_SIZE).cuda()) # simulate cudagraphs replay with set_forward_context( {}, vllm_config=vllm_config, cudagraph_runtime_mode=cudagraph_runtime_mode, batch_descriptor=BatchDescriptor( num_tokens=2, ), ): output = model(torch.randn(2, MLP_SIZE).cuda()) output = output.cpu() return output.cpu() @pytest.mark.parametrize("use_inductor_graph_partition", [True, False]) def test_ignore_torch_compile_decorator(use_inductor_graph_partition, monkeypatch): # disable compile cache so that we can count the number of compilations # appropriately monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", "1") if use_inductor_graph_partition and not is_torch_equal_or_newer("2.9.0.dev"): pytest.skip("inductor graph partition is only available in PyTorch 2.9+") # piecewise vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, splitting_ops=["silly::attention"], cudagraph_capture_sizes=[1, 2], use_inductor_graph_partition=use_inductor_graph_partition, ) ) cudagraph_runtime_mode = CUDAGraphMode.PIECEWISE expected_num_graphs_seen = 1 expected_num_cudagraph_captured = ( 4 # num_cudagraph_sizes * num cudagraphs to capture ) if use_inductor_graph_partition: expected_num_piecewise_graphs_seen = 1 expected_num_piecewise_capturable_graphs_seen = 1 expected_num_backend_compilations = 1 else: expected_num_piecewise_graphs_seen = 3 expected_num_piecewise_capturable_graphs_seen = 2 expected_num_backend_compilations = 2 @support_torch_compile class A(nn.Module): def __init__( self, *, vllm_config: VllmConfig, prefix: str = "", **kwargs ) -> None: super().__init__() def forward(self, x: torch.Tensor) -> torch.Tensor: x = x + x attn_output = torch.empty_like(x) torch.ops.silly.attention(x, x, x, attn_output) x = attn_output x = x * 3 return x @ignore_torch_compile class B(A): ... @support_torch_compile class C(B): ... with set_current_vllm_config(vllm_config): mod_A = A(vllm_config=vllm_config, prefix="").eval().cuda() # A has support_torch_compile with compilation_counter.expect( num_graphs_seen=expected_num_graphs_seen, num_piecewise_graphs_seen=expected_num_piecewise_graphs_seen, num_piecewise_capturable_graphs_seen=expected_num_piecewise_capturable_graphs_seen, num_backend_compilations=expected_num_backend_compilations, num_cudagraph_captured=expected_num_cudagraph_captured, ): run_model(vllm_config, mod_A, cudagraph_runtime_mode) with set_current_vllm_config(vllm_config): mod_B = B(vllm_config=vllm_config, prefix="").eval().cuda() # B's ignore_torch_compile should override A's support_torch_compile with compilation_counter.expect( num_graphs_seen=0, num_piecewise_graphs_seen=0, num_piecewise_capturable_graphs_seen=0, num_backend_compilations=0, num_cudagraph_captured=0, ): run_model(vllm_config, mod_B, cudagraph_runtime_mode) with set_current_vllm_config(vllm_config): mod_C = C(vllm_config=vllm_config, prefix="").eval().cuda() # C's support_torch_compile should override B's ignore_torch_compile with compilation_counter.expect( num_graphs_seen=expected_num_graphs_seen, num_piecewise_graphs_seen=expected_num_piecewise_graphs_seen, num_piecewise_capturable_graphs_seen=expected_num_piecewise_capturable_graphs_seen, num_backend_compilations=expected_num_backend_compilations, num_cudagraph_captured=expected_num_cudagraph_captured, ): run_model(vllm_config, mod_C, cudagraph_runtime_mode) # Only enable torch.compile if # vllm_config.cache_config.kv_sharing_fast_prefill=True @support_torch_compile( enable_if=lambda vllm_config: vllm_config.cache_config.kv_sharing_fast_prefill ) class B(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = "", **kwargs) -> None: super().__init__() def forward(self, x: torch.Tensor) -> torch.Tensor: x = x + x attn_output = torch.empty_like(x) torch.ops.silly.attention(x, x, x, attn_output) x = attn_output x = x + x return x # Only enable torch.compile if # vllm_config.cache_config.kv_sharing_fast_prefill=False @support_torch_compile( enable_if=lambda vllm_config: not vllm_config.cache_config.kv_sharing_fast_prefill ) class A(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = "", **kwargs) -> None: super().__init__() self.mod1 = B(vllm_config=vllm_config, prefix=prefix, **kwargs) self.mod2 = B(vllm_config=vllm_config, prefix=prefix, **kwargs) def forward(self, x: torch.Tensor) -> torch.Tensor: x = self.mod1(x) attn_output = torch.empty_like(x) torch.ops.silly.attention(x, x, x, attn_output) x = attn_output x = self.mod2(x) return x @pytest.mark.parametrize("use_inductor_graph_partition", [True, False]) def test_conditional_compile_enable_if(use_inductor_graph_partition, monkeypatch): # disable compile cache so that we can count the number of compilations # appropriately monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", "1") if use_inductor_graph_partition and not is_torch_equal_or_newer("2.9.0.dev"): pytest.skip("inductor graph partition is only available in PyTorch 2.9+") vllm_config = VllmConfig( cache_config=CacheConfig( kv_sharing_fast_prefill=True, ), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, splitting_ops=["silly::attention"], cudagraph_capture_sizes=[1, 2], use_inductor_graph_partition=use_inductor_graph_partition, ), ) cudagraph_runtime_mode = CUDAGraphMode.PIECEWISE with set_current_vllm_config(vllm_config): mod_A = A(vllm_config=vllm_config, prefix="").eval().cuda() if use_inductor_graph_partition: expected_num_piecewise_graphs_seen = 2 expected_num_piecewise_capturable_graphs_seen = 2 expected_num_backend_compilations = 2 else: expected_num_piecewise_graphs_seen = 6 expected_num_piecewise_capturable_graphs_seen = 4 expected_num_backend_compilations = 4 # A has support_torch_compile but enable_if fn returns False # enalbe_if will be True for B, so we expect mod1 and mod2 # to be compiled with compilation_counter.expect( num_graphs_seen=2, num_piecewise_graphs_seen=expected_num_piecewise_graphs_seen, # 3 piecewise graphs per instance of B() num_piecewise_capturable_graphs_seen=expected_num_piecewise_capturable_graphs_seen, num_backend_compilations=expected_num_backend_compilations, num_cudagraph_captured=8, # num_cudagraph_sizes * num cudagraphable graphs to capture ): run_model(vllm_config, mod_A, cudagraph_runtime_mode) # Set kv_sharing_fast_prefill=False # which will cause A to be compiled and B to not be compiled vllm_config = VllmConfig( cache_config=CacheConfig( kv_sharing_fast_prefill=False, ), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, splitting_ops=["silly::attention"], cudagraph_capture_sizes=[1, 2], use_inductor_graph_partition=use_inductor_graph_partition, ), ) with set_current_vllm_config(vllm_config): mod_A = A(vllm_config=vllm_config, prefix="").eval().cuda() if use_inductor_graph_partition: expected_num_piecewise_graphs_seen = 1 expected_num_piecewise_capturable_graphs_seen = 1 expected_num_backend_compilations = 1 else: # 3 attn ops and 4 non-attn ops expected_num_piecewise_graphs_seen = 7 expected_num_piecewise_capturable_graphs_seen = 4 expected_num_backend_compilations = 4 with compilation_counter.expect( num_graphs_seen=1, num_piecewise_graphs_seen=expected_num_piecewise_graphs_seen, # 3 attn ops and 4 non-attn ops num_piecewise_capturable_graphs_seen=expected_num_piecewise_capturable_graphs_seen, num_backend_compilations=expected_num_backend_compilations, num_cudagraph_captured=8, # num_cudagraph_sizes * num cudagraphable graphs to capture ): run_model(vllm_config, mod_A, cudagraph_runtime_mode)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_wrapper.py
tests/compile/test_wrapper.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os import pytest import torch from vllm.compilation.wrapper import TorchCompileWithNoGuardsWrapper from vllm.config import ( CompilationConfig, CompilationMode, VllmConfig, set_current_vllm_config, ) class MyMod(torch.nn.Module): def forward(self, x: torch.Tensor, cache: torch.Tensor | None = None): if x.size()[0] >= 4: return x * 2 else: return x * 100 class MyWrapper(TorchCompileWithNoGuardsWrapper): def __init__(self, model): self.model = model super().__init__() def forward(self, x: torch.Tensor): # type: ignore[override] # this is the function to be compiled return self.model(x) @pytest.mark.parametrize("use_bytecode_hook", [True, False]) def test_torch_compile_wrapper(use_bytecode_hook, monkeypatch): """Test basic functionality of TorchCompileWithNoGuardsWrapper.""" # Set the environment variable for this test monkeypatch.setenv("VLLM_USE_BYTECODE_HOOK", "1" if use_bytecode_hook else "0") # Create a proper vLLM config instead of mocking vllm_config = VllmConfig() vllm_config.compilation_config = CompilationConfig() vllm_config.compilation_config.mode = CompilationMode.DYNAMO_TRACE_ONCE vllm_config.compilation_config.backend = "inductor" # Test DYNAMO_TRACE_ONCE with set_current_vllm_config(vllm_config): torch._dynamo.reset() mod = MyMod() wrapper = MyWrapper(mod) # First call should trigger compilation x = torch.tensor([1, 2, 3, 4]) torch._dynamo.mark_dynamic(x, 0) result1 = wrapper(x) expected1 = torch.tensor([2, 4, 6, 8]) assert torch.allclose(result1, expected1), ( f"Expected {expected1}, got {result1}" ) # Second call should use compiled code x2 = torch.tensor([1, 2, 3]) result2 = wrapper(x2) expected2 = torch.tensor([2, 4, 6]) assert torch.allclose(result2, expected2), ( f"Expected {expected2}, got {result2}" ) # without the wrapper result would be different. result3 = mod(x2) expected3 = torch.tensor([100, 200, 300]) assert torch.allclose(result3, expected3), ( f"Expected {result3}, got {expected3}" ) # with STOCK_TORCH_COMPILE we do not remove guards. vllm_config.compilation_config.mode = CompilationMode.STOCK_TORCH_COMPILE torch._dynamo.reset() with set_current_vllm_config(vllm_config): mod = MyMod() wrapper = MyWrapper(mod) # First call should trigger compilation x = torch.tensor([1, 2, 3, 4]) torch._dynamo.mark_dynamic(x, 0) result1 = wrapper(x) expected1 = torch.tensor([2, 4, 6, 8]) assert torch.allclose(result1, expected1), ( f"Expected {expected1}, got {result1}" ) # Second call should triger another compilation x2 = torch.tensor([1, 2, 3]) result2 = wrapper(x2) expected2 = torch.tensor([100, 200, 300]) assert torch.allclose(result2, expected2), ( f"Expected {expected2}, got {result2}" ) # NO_COMPILATION level not supported. vllm_config.compilation_config.mode = None torch._dynamo.reset() with set_current_vllm_config(vllm_config): torch._dynamo.reset() mod = MyMod() try: wrapper = MyWrapper(mod) except Exception: return raise AssertionError("expected an exception to be raised") if __name__ == "__main__": # Run with both parameter values class MockMonkeypatch: def setenv(self, name, value): os.environ[name] = value mp = MockMonkeypatch() print("Testing with VLLM_USE_BYTECODE_HOOK=False") test_torch_compile_wrapper(False, mp) print("Testing with VLLM_USE_BYTECODE_HOOK=True") test_torch_compile_wrapper(True, mp) print("All tests passed!")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_fusion_attn.py
tests/compile/test_fusion_attn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import copy import pytest import torch._dynamo from tests.compile.backend import LazyInitPass, TestBackend from tests.utils import flat_product from tests.v1.attention.utils import BatchSpec, create_common_attn_metadata from vllm._custom_ops import cutlass_scaled_fp4_mm, scaled_fp4_quant from vllm.attention.backends.abstract import AttentionMetadata from vllm.attention.backends.registry import AttentionBackendEnum from vllm.attention.layer import Attention from vllm.compilation.fusion_attn import ATTN_OP, AttnFusionPass from vllm.compilation.fx_utils import find_op_nodes from vllm.compilation.matcher_utils import QUANT_OPS from vllm.compilation.noop_elimination import NoOpEliminationPass from vllm.compilation.post_cleanup import PostCleanupPass from vllm.config import ( AttentionConfig, CacheConfig, CompilationConfig, CompilationMode, ModelConfig, PassConfig, SchedulerConfig, VllmConfig, set_current_vllm_config, ) from vllm.forward_context import get_forward_context, set_forward_context from vllm.model_executor.layers.quantization.utils.quant_utils import ( QuantKey, kFp8StaticTensorSym, kNvfp4Quant, ) from vllm.model_executor.layers.quantization.utils.w8a8_utils import Fp8LinearOp from vllm.platforms import current_platform from vllm.utils.flashinfer import has_flashinfer from vllm.v1.kv_cache_interface import AttentionSpec FP8_DTYPE = current_platform.fp8_dtype() FP4_DTYPE = torch.uint8 class AttentionQuantPatternModel(torch.nn.Module): """Base model for AttentionQuantPattern fusion.""" def __init__( self, num_qo_heads: int, num_kv_heads: int, head_size: int, kv_cache_dtype: torch.dtype, device: torch.device, vllm_config: VllmConfig, **kwargs, ): super().__init__() self.num_qo_heads = num_qo_heads self.num_kv_heads = num_kv_heads self.head_size = head_size self.kv_cache_dtype = kv_cache_dtype self.device = device self.vllm_config = vllm_config self.attn = Attention( num_heads=self.num_qo_heads, head_size=self.head_size, scale=1.0 / (self.head_size**0.5), num_kv_heads=self.num_kv_heads, cache_config=vllm_config.cache_config, prefix="model.layers.0.self_attn.attn", ) self.attn._k_scale = self.attn._k_scale.to(device) self.attn._v_scale = self.attn._v_scale.to(device) self.block_size = 16 # Initialize attn MetadataBuilder self.builder = self.attn.attn_backend.get_builder_cls()( kv_cache_spec=AttentionSpec( block_size=self.block_size, num_kv_heads=self.num_kv_heads, head_size=self.head_size, dtype=self.kv_cache_dtype, ), layer_names=[self.attn.layer_name], vllm_config=self.vllm_config, device=self.device, ) def build_attn_metadata(self, batch_size: int) -> AttentionMetadata: """Initialize attention metadata.""" # Create common attn metadata batch_spec = BatchSpec(seq_lens=[1] * batch_size, query_lens=[1] * batch_size) common_attn_metadata = create_common_attn_metadata( batch_spec, self.block_size, self.device, arange_block_indices=True ) max_blocks = (max(batch_spec.seq_lens) + self.block_size - 1) // self.block_size num_blocks = batch_size * max_blocks backend = self.attn.backend # TODO(luka) use get_kv_cache_stride_order # Create dummy KV cache for the selected backend if backend == AttentionBackendEnum.ROCM_ATTN: # k/v as 1st dimention # HND: [num_blocks, num_kv_heads, block_size, head_size] kv_cache = torch.zeros( 2, num_blocks, self.num_kv_heads, self.block_size, self.head_size, dtype=self.kv_cache_dtype, device=self.device, ) elif backend == AttentionBackendEnum.ROCM_AITER_UNIFIED_ATTN: # k/v as 1st dimention # NHD: [num_blocks, block_size, num_kv_heads, head_size] kv_cache = torch.zeros( 2, num_blocks, self.block_size, self.num_kv_heads, self.head_size, dtype=self.kv_cache_dtype, device=self.device, ) elif backend == AttentionBackendEnum.TRITON_ATTN: # k/v as 2nd dimention # NHD: [num_blocks, block_size, num_kv_heads, head_size] kv_cache = torch.zeros( num_blocks, 2, self.num_kv_heads, self.block_size, self.head_size, dtype=self.kv_cache_dtype, device=self.device, ) elif backend == AttentionBackendEnum.FLASHINFER: kv_cache = torch.zeros( num_blocks, 2, self.num_kv_heads, self.block_size, self.head_size, dtype=self.kv_cache_dtype, device=self.device, ).permute(0, 1, 3, 2, 4) else: raise ValueError(f"Unsupported backend: {backend}") self.attn.kv_cache = [kv_cache] # Build attn metadata self.attn_metadata = self.builder.build( common_prefix_len=0, common_attn_metadata=common_attn_metadata ) return self.attn_metadata class TestAttentionFp8StaticQuantPatternModel(AttentionQuantPatternModel): """Test model for AttentionFp8StaticQuantPattern fusion.""" quant_key = kFp8StaticTensorSym def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.fp8_linear = Fp8LinearOp( act_quant_static=self.quant_key.scale.static, act_quant_group_shape=self.quant_key.scale.group_shape, ) hidden_size = self.num_qo_heads * self.head_size self.w = kwargs.get( "w", { "weight": torch.randn(hidden_size, hidden_size) .to(dtype=FP8_DTYPE, device=self.device) .t(), "wscale": torch.tensor([1.0], dtype=torch.float32, device=self.device), "scale": torch.tensor([1.0], dtype=torch.float32, device=self.device), }, ) def forward(self, q: torch.Tensor, k: torch.Tensor, v: torch.Tensor): """Forward pass that creates the pattern to be fused.""" attn_output = self.attn(q, k, v) return self.fp8_linear.apply( input=attn_output, weight=self.w["weight"], weight_scale=self.w["wscale"], input_scale=self.w["scale"], ) class TestAttentionNvfp4QuantPatternModel(AttentionQuantPatternModel): """Test model for AttentionNvfp4QuantPattern fusion.""" quant_key = kNvfp4Quant def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) hidden_size = self.num_qo_heads * self.head_size self.w = kwargs.get( "w", { "weight": torch.randint( 256, (hidden_size, hidden_size // 2), dtype=FP4_DTYPE, device=self.device, ), "wscale_swizzled": torch.randn(hidden_size, hidden_size // 16).to( dtype=FP8_DTYPE, device=self.device ), "wscale": torch.tensor([500], dtype=torch.float32, device=self.device), "scale": torch.tensor([0.002], dtype=torch.float32, device=self.device), }, ) def forward(self, q: torch.Tensor, k: torch.Tensor, v: torch.Tensor): """Forward pass that creates the pattern to be fused.""" attn_output = self.attn(q, k, v) quant_output, output_block_scale = scaled_fp4_quant( attn_output, 1 / self.w["scale"] ) return cutlass_scaled_fp4_mm( a=quant_output, b=self.w["weight"], block_scale_a=output_block_scale, block_scale_b=self.w["wscale_swizzled"], alpha=self.w["scale"] * self.w["wscale"], out_dtype=attn_output.dtype, ) MODELS_FP8: list[tuple[str, type]] = [] MODELS_FP4: list[tuple[str, type]] = [] HEADS: list[tuple[int, int]] = [] SPLIT_ATTENTION: list[bool] = [] BACKENDS_FP8: list[AttentionBackendEnum] = [] BACKENDS_FP4: list[AttentionBackendEnum] = [] if current_platform.is_cuda(): HEADS = [(64, 8), (40, 8)] MODELS_FP8 = [ ( "nvidia/Llama-4-Scout-17B-16E-Instruct-FP8", TestAttentionFp8StaticQuantPatternModel, ) ] MODELS_FP4 = [ ( "nvidia/Llama-4-Scout-17B-16E-Instruct-FP4", TestAttentionNvfp4QuantPatternModel, ) ] BACKENDS_FP8 = [AttentionBackendEnum.TRITON_ATTN, AttentionBackendEnum.FLASHINFER] BACKENDS_FP4 = [AttentionBackendEnum.FLASHINFER] elif current_platform.is_rocm(): HEADS = [(32, 8), (40, 8)] MODELS_FP8 = [ ("amd/Llama-3.1-8B-Instruct-FP8-KV", TestAttentionFp8StaticQuantPatternModel) ] BACKENDS = [ AttentionBackendEnum.ROCM_AITER_UNIFIED_ATTN, AttentionBackendEnum.ROCM_ATTN, AttentionBackendEnum.TRITON_ATTN, ] @pytest.mark.parametrize("num_qo_heads, num_kv_heads", HEADS) @pytest.mark.parametrize("head_size", [128]) @pytest.mark.parametrize( "batch_size", [7, 256, 533] if current_platform.is_cuda() else [8] ) @pytest.mark.parametrize("dtype", [torch.bfloat16, torch.float16]) @pytest.mark.parametrize( "backend, model_name, model_class, custom_ops", # Test attention+quant_fp8 fusion with custom and torch impls of QuantFP8 list(flat_product(BACKENDS_FP8, MODELS_FP8, ["+quant_fp8", "-quant_fp8"])) # quant_fp4 only has the custom impl + list(flat_product(BACKENDS_FP4, MODELS_FP4, [""])), ) @pytest.mark.skipif( not current_platform.is_cuda_alike(), reason="Only test ROCm or CUDA" ) @pytest.mark.skipif(not current_platform.supports_fp8(), reason="Need FP8") def test_attention_quant_pattern( num_qo_heads: int, num_kv_heads: int, head_size: int, batch_size: int, dtype: torch.dtype, custom_ops: str, model_name: str, model_class: type[AttentionQuantPatternModel], backend: AttentionBackendEnum, dist_init, ): """Test AttentionStaticQuantPattern fusion pass""" if backend == AttentionBackendEnum.FLASHINFER and ( not current_platform.is_device_capability((10, 0)) or not has_flashinfer() ): pytest.skip("FlashInfer attn fusion requires Blackwell and flashinfer") custom_ops_list = custom_ops.split(",") if custom_ops else [] device = torch.device("cuda:0") torch.set_default_dtype(dtype) torch.manual_seed(42) model_config = ModelConfig( model=model_name, max_model_len=2048, dtype=dtype, ) vllm_config = VllmConfig( model_config=model_config, scheduler_config=SchedulerConfig( max_num_seqs=1024, max_model_len=model_config.max_model_len, is_encoder_decoder=model_config.is_encoder_decoder, ), compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, custom_ops=custom_ops_list, ), cache_config=CacheConfig(cache_dtype="fp8"), attention_config=AttentionConfig(backend=backend), ) # Create test inputs q = torch.randn(batch_size, num_qo_heads * head_size, dtype=dtype, device=device) k = torch.randn(batch_size, num_kv_heads * head_size, dtype=dtype, device=device) v = torch.randn(batch_size, num_kv_heads * head_size, dtype=dtype, device=device) # Mark first dimension as dynamic for realistic testing torch._dynamo.mark_dynamic(q, 0) torch._dynamo.mark_dynamic(k, 0) torch._dynamo.mark_dynamic(v, 0) # Run model directly without compilation and fusion vllm_config_unfused = copy.deepcopy(vllm_config) with ( set_current_vllm_config(vllm_config_unfused), set_forward_context(attn_metadata=None, vllm_config=vllm_config_unfused), ): model_unfused = model_class( num_qo_heads=num_qo_heads, num_kv_heads=num_kv_heads, head_size=head_size, kv_cache_dtype=FP8_DTYPE, device=device, vllm_config=vllm_config_unfused, ) model_unfused = model_unfused.to(device) forward_ctx = get_forward_context() forward_ctx.attn_metadata = model_unfused.build_attn_metadata(batch_size) # Run model directly without fusion # Still compile so query QuantFP8 has closer numerics result_unfused = torch.compile(model_unfused, fullgraph=True)(q, k, v) # Run model with attn fusion enabled vllm_config.compilation_config.pass_config = PassConfig( fuse_attn_quant=True, eliminate_noops=True ) with ( set_current_vllm_config(vllm_config), set_forward_context(attn_metadata=None, vllm_config=vllm_config), ): model_fused = model_class( num_qo_heads=num_qo_heads, num_kv_heads=num_kv_heads, head_size=head_size, kv_cache_dtype=FP8_DTYPE, device=device, vllm_config=vllm_config, w=model_unfused.w, ) model_fused = model_fused.to(device) forward_ctx = get_forward_context() forward_ctx.attn_metadata = model_fused.build_attn_metadata(batch_size) # Create test backend with fusion passes enabled noop_pass = NoOpEliminationPass(vllm_config) attn_pass = LazyInitPass(AttnFusionPass, vllm_config) cleanup_pass = PostCleanupPass(vllm_config) test_backend = TestBackend(noop_pass, attn_pass, cleanup_pass) # Compile model with fusion enabled model_compiled = torch.compile( model_fused, backend=test_backend, fullgraph=True ) assert model_compiled.attn._o_scale_float is None result_fused_1 = model_compiled(q, k, v) if backend == AttentionBackendEnum.FLASHINFER: # With the Flashinfer backend after the 1st round of the forward # pass, output quant scale should be loaded into the attn layer's # _o_scale_float, the 2nd round should reuse the loaded # _o_scale_float assert model_compiled.attn._o_scale_float is not None result_fused_2 = model_compiled(q, k, v) assert model_compiled.attn._o_scale_float is not None torch.testing.assert_close( result_unfused, result_fused_2, atol=1e-2, rtol=1e-2 ) # Check attn fusion support quant_key: QuantKey = model_class.quant_key attn_fusion_supported = [ layer.impl.fused_output_quant_supported(quant_key) for key, layer in vllm_config.compilation_config.static_forward_context.items() ] assert sum(attn_fusion_supported) == len(attn_fusion_supported), ( "All layers should support attention fusion" ) # Check quantization ops in the graph before and after fusion quant_op = ( torch.ops.aten.reciprocal if "-quant_fp8" in custom_ops_list else QUANT_OPS[quant_key] ) # Note: for fp8, fully_replaced=False because query quant ops remain in graph. # Only output quant ops are fused into attention. test_backend.check_before_ops([quant_op], fully_replaced=quant_key is kNvfp4Quant) # access the underlying `AttnFusionPass` on the `LazyInitPass` assert attn_pass.pass_.matched_count == sum(attn_fusion_supported) # Check attention ops in the graph before and after fusion attn_nodes_pre = list(find_op_nodes(ATTN_OP, test_backend.graph_pre_pass)) attn_nodes_post = list(find_op_nodes(ATTN_OP, test_backend.graph_post_pass)) assert len(attn_nodes_pre) > 0, "Should have attention nodes before fusion" assert len(attn_nodes_pre) == len(attn_nodes_post), ( "Should have same number of attention nodes before and after fusion" ) assert attn_nodes_pre[0].kwargs.get("output_scale") is None, ( "Attention should not have output_scale before fusion" ) assert attn_nodes_post[0].kwargs.get("output_scale") is not None, ( "Attention should have output_scale after fusion" ) assert attn_nodes_pre[0].kwargs.get("output_block_scale") is None, ( "Attention should not have output_block_scale before fusion" ) if quant_key.dtype == FP8_DTYPE: assert attn_nodes_post[0].kwargs.get("output_block_scale") is None, ( "Attention should not have output_block_scale after FP8 fusion" ) elif quant_key.dtype == FP4_DTYPE: assert attn_nodes_post[0].kwargs.get("output_block_scale") is not None, ( "Attention should have output_block_scale after FP4 fusion" ) # Check that results are close torch.testing.assert_close(result_unfused, result_fused_1, atol=1e-2, rtol=1e-2)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/silly_attention.py
tests/compile/silly_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Shared PyTorch custom silly attention for compilation tests. Centralizes custom operation definitions to avoid duplicate registrations. """ import torch from torch.library import Library from vllm.utils.torch_utils import direct_register_custom_op # Shared library for all compilation test operations # Using "silly" namespace to match existing test expectations # import this file will automatically register # torch ops for testing (like silly.attention) silly_lib = Library("silly", "FRAGMENT") # Global counter that counts the number of times attention is invoked _global_counter = 0 def get_global_counter(): """Get the current global counter value""" return _global_counter def reset_global_counter(): """Reset the global counter to 0""" global _global_counter _global_counter = 0 def silly_attention( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, out: torch.Tensor ) -> None: """ Unified attention implementation that depends on all inputs and affects the output. Always increments a global counter that tests can use or ignore. """ global _global_counter # Always increment the global counter _global_counter += 1 # Unified implementation that depends on all inputs out.copy_(q + k + v) def silly_attention_fake( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, out: torch.Tensor ) -> None: """Fake implementation for testing""" return # Register the unified attention operation direct_register_custom_op( op_name="attention", op_func=silly_attention, mutates_args=["out"], fake_impl=silly_attention_fake, target_lib=silly_lib, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_functionalization.py
tests/compile/test_functionalization.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch import vllm.envs as envs from vllm.compilation.activation_quant_fusion import ActivationQuantFusionPass from vllm.compilation.fix_functionalization import FixFunctionalizationPass from vllm.compilation.fusion import RMSNormQuantFusionPass from vllm.compilation.fx_utils import find_auto_fn, find_auto_fn_maybe, is_func from vllm.compilation.noop_elimination import NoOpEliminationPass from vllm.compilation.post_cleanup import PostCleanupPass from vllm.config import ( CompilationConfig, ModelConfig, PassConfig, VllmConfig, set_current_vllm_config, ) from vllm.model_executor.layers.activation import SiluAndMul from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape from vllm.model_executor.layers.quantization.utils.w8a8_utils import Fp8LinearOp from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.platforms import current_platform from .backend import TestBackend TEST_FP8 = current_platform.supports_fp8() FP8_DTYPE = current_platform.fp8_dtype() class TestSiluMul(torch.nn.Module): def __init__(self, hidden_size: int = 128): super().__init__() self.silu_and_mul = SiluAndMul() self.wscale = torch.rand(1, dtype=torch.float32) self.scale = torch.rand(1, dtype=torch.float32) if TEST_FP8: self.w = torch.rand(hidden_size, hidden_size).to(dtype=FP8_DTYPE).t() self.fp8_linear = Fp8LinearOp( act_quant_static=True, act_quant_group_shape=GroupShape.PER_TENSOR, ) def forward(self, x): y = self.silu_and_mul(x) if TEST_FP8: x2 = self.fp8_linear.apply(y, self.w, self.wscale, input_scale=self.wscale) return x2 else: return y def example_inputs(self, num_tokens=32, hidden_size=128): return (torch.rand(num_tokens, hidden_size * 2),) def ops_in_model(self, do_fusion): if TEST_FP8 and do_fusion: return [torch.ops._C.silu_and_mul_quant.default] else: return [torch.ops._C.silu_and_mul.default] def ops_not_in_model(self): return [] class TestFusedAddRMSNorm(torch.nn.Module): def __init__(self, hidden_size=16, intermediate_size=32): super().__init__() self.hidden_size = hidden_size self.intermediate_size = intermediate_size self.gate_proj = torch.nn.Parameter( torch.empty((intermediate_size, hidden_size)) ) self.norm = RMSNorm(intermediate_size, 1e-05) self.norm.weight = torch.nn.Parameter(torch.ones(intermediate_size)) torch.nn.init.normal_(self.gate_proj, std=0.02) if TEST_FP8: self.fp8_linear = Fp8LinearOp(act_quant_static=True) self.scale = torch.rand(1, dtype=torch.float32) self.w = torch.rand(hidden_size, intermediate_size).to(dtype=FP8_DTYPE).t() self.wscale = torch.rand(1, dtype=torch.float32) def forward(self, hidden_states, residual): # Reshape input view = hidden_states.reshape(-1, self.hidden_size) # matrix multiplication permute = self.gate_proj.permute(1, 0) mm = torch.mm(view, permute) # layer normalization norm_output, residual_output = self.norm(mm, residual) if TEST_FP8: # scaled_mm with static input quantization fp8_linear_result = self.fp8_linear.apply( norm_output, self.w, self.wscale, input_scale=self.scale.to(norm_output.device), ) return fp8_linear_result, residual_output else: return norm_output, residual_output def example_inputs(self, batch_size=8, hidden_size=16, seq_len=16): hidden_states = torch.randn((batch_size * seq_len, hidden_size)) residual = torch.randn((batch_size * seq_len, hidden_size)) return (hidden_states, residual) def ops_in_model(self, do_fusion): if TEST_FP8 and do_fusion: return [torch.ops._C.fused_add_rms_norm_static_fp8_quant.default] else: return [torch.ops._C.fused_add_rms_norm.default] def ops_not_in_model(self): return [] class TestRotaryEmbedding(torch.nn.Module): def __init__(self, head_dim=64, max_position=2048, base=10000): super().__init__() self.head_dim = head_dim self.rotary_emb = get_rope( self.head_dim, max_position=max_position, rope_parameters={"rope_type": "default", "rope_theta": base}, ) def forward(self, positions, q, k): q_rotated, k_rotated = self.rotary_emb(positions, q, k) return q_rotated, k_rotated def example_inputs(self, num_tokens=32, head_dim=64): positions = torch.arange(num_tokens, dtype=torch.long) q = torch.randn(num_tokens, head_dim) k = torch.randn(num_tokens, head_dim) return (positions, q, k) def ops_in_model(self, do_fusion): return [torch.ops._C.rotary_embedding.default] def ops_not_in_model(self): return [] class TestRotaryEmbeddingSliceScatter(torch.nn.Module): def __init__(self, head_dim=64, num_heads=4, max_position=2048, base=10000): super().__init__() self.head_dim = head_dim self.num_heads = num_heads self.hidden_size = head_dim * num_heads self.qkv_proj = torch.nn.Linear( self.hidden_size, self.hidden_size * 3, bias=False ) self.rotary_emb = get_rope( self.head_dim, max_position=max_position, rope_parameters={"rope_type": "default", "rope_theta": base}, ) def forward(self, positions, hidden_states): # Simulate the pattern: mm -> split_with_sizes -> rotary_embedding # -> slice_scatter -> split_with_sizes qkv = self.qkv_proj(hidden_states) split_sizes = [self.hidden_size, self.hidden_size, self.hidden_size] q, k, v = torch.split(qkv, split_sizes, dim=-1) q_rotated, k_rotated = self.rotary_emb(positions, q, k) qkv_updated = torch.cat([q_rotated, k_rotated, v], dim=-1) return qkv_updated def example_inputs(self, num_tokens=32, head_dim=64, num_heads=4): hidden_size = head_dim * num_heads positions = torch.arange(num_tokens, dtype=torch.long) hidden_states = torch.randn(num_tokens, hidden_size) return (positions, hidden_states) def ops_in_model(self, do_fusion): return [torch.ops._C.rotary_embedding.default] def ops_not_in_model(self): return [torch.ops.aten.slice_scatter.default] MODELS = [ TestSiluMul, TestFusedAddRMSNorm, TestRotaryEmbedding, TestRotaryEmbeddingSliceScatter, ] @pytest.mark.parametrize("dtype", [torch.float16, torch.bfloat16]) @pytest.mark.parametrize("model_class", MODELS) @pytest.mark.parametrize("do_fusion", [True, False]) @pytest.mark.skipif(envs.VLLM_TARGET_DEVICE != "cuda", reason="Only test on CUDA") def test_fix_functionalization( model_class: torch.nn.Module, do_fusion: bool, dtype: torch.dtype ): torch.set_default_device("cuda") torch.set_default_dtype(dtype) vllm_config = VllmConfig( model_config=ModelConfig(dtype=dtype), compilation_config=CompilationConfig( custom_ops=["all"], pass_config=PassConfig( fuse_norm_quant=do_fusion, fuse_act_quant=do_fusion, eliminate_noops=True, ), ), ) with set_current_vllm_config(vllm_config): assert RMSNorm.enabled() noop_pass = NoOpEliminationPass(vllm_config) fusion_pass = RMSNormQuantFusionPass(vllm_config) cleanup_pass = PostCleanupPass(vllm_config) act_quant_fusion_pass = ActivationQuantFusionPass(vllm_config) passes = ( [noop_pass, fusion_pass, act_quant_fusion_pass, cleanup_pass] if do_fusion else [noop_pass, cleanup_pass] ) func_pass = FixFunctionalizationPass(vllm_config) backend_func = TestBackend(*passes, func_pass) backend_no_func = TestBackend(*passes) model = model_class() torch.compile(model, backend=backend_func)(*model.example_inputs()) torch.compile(model, backend=backend_no_func)(*model.example_inputs()) # check if the functionalization pass is applied for op in model.ops_in_model(do_fusion): find_auto_fn(backend_no_func.graph_post_pass.nodes, op) assert find_auto_fn_maybe(backend_func.graph_post_pass.nodes, op) is None # make sure the ops were all de-functionalized found = dict() for node in backend_func.graph_post_pass.nodes: for op in model.ops_in_model(do_fusion): if is_func(node, op): found[op] = True for op in model.ops_not_in_model(): if is_func(node, op): found[op] = True assert all(found[op] for op in model.ops_in_model(do_fusion)) assert all(not found.get(op) for op in model.ops_not_in_model())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/__init__.py
tests/compile/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_silu_mul_quant_fusion.py
tests/compile/test_silu_mul_quant_fusion.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import itertools import pytest import torch import vllm.envs as envs from tests.kernels.quantization.nvfp4_utils import quant_nvfp4_tensor from vllm._aiter_ops import IS_AITER_FOUND from vllm._custom_ops import cutlass_scaled_fp4_mm, scaled_fp4_quant from vllm.compilation.activation_quant_fusion import ( FUSED_OPS, SILU_MUL_OP, ActivationQuantFusionPass, ) from vllm.compilation.fusion import QUANT_OPS from vllm.compilation.noop_elimination import NoOpEliminationPass from vllm.compilation.post_cleanup import PostCleanupPass from vllm.config import ( CompilationConfig, CompilationMode, PassConfig, VllmConfig, set_current_vllm_config, ) from vllm.model_executor.layers.activation import SiluAndMul from vllm.model_executor.layers.quantization.utils.fp8_utils import W8A8BlockFp8LinearOp from vllm.model_executor.layers.quantization.utils.quant_utils import ( GroupShape, kFp8StaticTensorSym, kNvfp4Quant, ) from vllm.model_executor.layers.quantization.utils.w8a8_utils import ( Fp8LinearOp, maybe_create_device_identity, ) from vllm.platforms import current_platform from ..utils import override_cutlass_fp8_supported from .backend import TestBackend FP8_DTYPE = current_platform.fp8_dtype() FP4_DTYPE = torch.uint8 def is_nvfp4_supported(): return current_platform.has_device_capability(100) class TestSiluMulFp8QuantModel(torch.nn.Module): def __init__(self, hidden_size: int, cuda_force_torch: bool, **kwargs): super().__init__() self.silu_and_mul = SiluAndMul() self.wscale = torch.rand(1, dtype=torch.float32) self.scale = torch.rand(1, dtype=torch.float32) self.w = torch.rand(hidden_size, hidden_size).to(dtype=FP8_DTYPE).t() with override_cutlass_fp8_supported(not cuda_force_torch): self.fp8_linear = Fp8LinearOp( act_quant_static=True, act_quant_group_shape=GroupShape.PER_TENSOR, ) self.enable_silu_mul_custom_op = self.silu_and_mul.enabled() self.enable_quant_fp8_custom_op = self.fp8_linear.quant_fp8.enabled() def forward(self, x): y = self.silu_and_mul(x) x2 = self.fp8_linear.apply(y, self.w, self.wscale, input_scale=self.wscale) return x2 def ops_in_model_before(self): return [ SILU_MUL_OP if self.enable_silu_mul_custom_op else torch.ops.aten.mul, ( QUANT_OPS[kFp8StaticTensorSym] if self.enable_quant_fp8_custom_op else torch.ops.aten.reciprocal ), ] def ops_in_model_after(self): return [FUSED_OPS[kFp8StaticTensorSym]] class TestSiluMulNvfp4QuantModel(torch.nn.Module): def __init__(self, hidden_size: int, x: torch.Tensor, **kwargs): super().__init__() from vllm.compilation.activation_quant_fusion import ( silu_and_mul_nvfp4_quant_supported, ) assert silu_and_mul_nvfp4_quant_supported self.silu_and_mul = SiluAndMul() self.enable_silu_mul_custom_op = self.silu_and_mul.enabled() # create nvfp4 weight w = torch.rand((hidden_size, hidden_size)) self.w, self.w_block_scale, self.w_global_scale = quant_nvfp4_tensor(w) # get global scale offline _, _, self.y_global_scale = quant_nvfp4_tensor(self.silu_and_mul(x)) self.alpha = 1.0 / (self.w_global_scale * self.y_global_scale) def forward(self, x): y = self.silu_and_mul(x) y_quant, y_block_scale = scaled_fp4_quant(y, self.y_global_scale) out = cutlass_scaled_fp4_mm( a=y_quant, b=self.w, block_scale_a=y_block_scale, block_scale_b=self.w_block_scale, alpha=self.alpha, out_dtype=y.dtype, ) return out def ops_in_model_before(self): return [ SILU_MUL_OP if self.enable_silu_mul_custom_op else torch.ops.aten.mul, QUANT_OPS[kNvfp4Quant], ] def ops_in_model_after(self): return [FUSED_OPS[kNvfp4Quant]] class TestSiluMulGroupFp8QuantModel(torch.nn.Module): def __init__(self, hidden_size: int, **kwargs): super().__init__() self.silu_and_mul = SiluAndMul() self.w8a8_block_fp8_linear = W8A8BlockFp8LinearOp( weight_group_shape=GroupShape(128, 128), act_quant_group_shape=GroupShape(1, 128), cutlass_block_fp8_supported=False, use_aiter_and_is_supported=True, ) self.w = torch.rand(hidden_size, hidden_size).to(dtype=FP8_DTYPE).t() scale_hidden_size = (hidden_size + 128 - 1) // 128 self.wscale = torch.rand( (scale_hidden_size, scale_hidden_size), dtype=torch.float32 ) self.enable_silu_mul_custom_op = self.silu_and_mul.enabled() def forward(self, x): y = self.silu_and_mul(x) x2 = self.w8a8_block_fp8_linear.apply(y, self.w, self.wscale) return x2 def ops_in_model_before(self): return [ SILU_MUL_OP if self.enable_silu_mul_custom_op else torch.ops.aten.mul, ] def ops_in_model_after(self): return [torch.ops.vllm.rocm_aiter_act_mul_and_fp8_group_quant] @pytest.mark.parametrize("num_tokens", [32, 64]) @pytest.mark.parametrize("hidden_size", [128, 256]) @pytest.mark.parametrize("dtype", [torch.bfloat16, torch.float16]) @pytest.mark.parametrize("enable_silu_mul_custom_op", [True, False]) @pytest.mark.parametrize( "model_class, enable_quant_fp8_custom_op, cuda_force_torch", list(itertools.product([TestSiluMulFp8QuantModel], [True, False], [True, False])) + [ (TestSiluMulNvfp4QuantModel, False, False), (TestSiluMulGroupFp8QuantModel, False, False), ], ) # cuda_force_torch used to test torch code path on platforms that # cutlass_fp8_supported() == True. @pytest.mark.skipif( envs.VLLM_TARGET_DEVICE not in ["cuda", "rocm"], reason="Only test on CUDA and ROCm" ) def test_fusion_silu_and_mul_quant( num_tokens: int, hidden_size: int, dtype: torch.dtype, model_class: type[ TestSiluMulFp8QuantModel | TestSiluMulNvfp4QuantModel | TestSiluMulGroupFp8QuantModel ], enable_silu_mul_custom_op: bool, enable_quant_fp8_custom_op: bool, cuda_force_torch: bool, ): if model_class is TestSiluMulNvfp4QuantModel and not is_nvfp4_supported(): pytest.skip("NVFP4 is not supported on this GPU.") if model_class is TestSiluMulGroupFp8QuantModel and not IS_AITER_FOUND: pytest.skip("AITER is not supported on this GPU.") torch.set_default_device("cuda") torch.set_default_dtype(dtype) maybe_create_device_identity() x = torch.rand(num_tokens, hidden_size * 2) # Reshape pass is needed for the fusion pass to work custom_ops = [] if enable_silu_mul_custom_op: custom_ops.append("+silu_and_mul") if enable_quant_fp8_custom_op: custom_ops.append("+quant_fp8") config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, custom_ops=custom_ops, pass_config=PassConfig(fuse_act_quant=True, eliminate_noops=True), ), ) with set_current_vllm_config(config): fusion_passes = [ActivationQuantFusionPass(config)] if IS_AITER_FOUND: from vllm.compilation.rocm_aiter_fusion import ( RocmAiterSiluMulFp8GroupQuantFusionPass, ) fusion_passes += [RocmAiterSiluMulFp8GroupQuantFusionPass(config)] passes = [NoOpEliminationPass(config), *fusion_passes, PostCleanupPass(config)] backend = TestBackend(*passes) model = model_class( hidden_size=hidden_size, cuda_force_torch=cuda_force_torch, x=x ) # First dimension dynamic torch._dynamo.mark_dynamic(x, 0) result = model(x) model2 = torch.compile(model, backend=backend) result2 = model2(x) # Check that it gives the same answer if model_class == TestSiluMulFp8QuantModel: atol, rtol = 1e-3, 1e-3 elif model_class == TestSiluMulNvfp4QuantModel: atol, rtol = 1e-1, 1e-1 elif model_class == TestSiluMulGroupFp8QuantModel: atol, rtol = 5e-2, 5e-2 torch.testing.assert_close( result[0].to(dtype=dtype), result2[0].to(dtype=dtype), atol=atol, rtol=rtol ) assert sum([p.matched_count for p in fusion_passes]) == 1 # In pre-nodes, quant op should be present and fused kernels should not backend.check_before_ops(model.ops_in_model_before()) # In post-nodes, fused kernels should be present and quant op should not backend.check_after_ops(model.ops_in_model_after())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/backend.py
tests/compile/backend.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import weakref from collections.abc import Callable, Sequence from contextlib import nullcontext from copy import deepcopy import depyf from torch import fx from torch._ops import OpOverload from torch.fx._utils import lazy_format_graph_code from vllm.compilation.fx_utils import find_op_nodes from vllm.compilation.inductor_pass import InductorPass from vllm.compilation.pass_manager import with_pattern_match_debug from vllm.compilation.vllm_inductor_pass import VllmInductorPass from vllm.config import VllmConfig, get_current_vllm_config from vllm.logger import init_logger logger = init_logger("vllm.tests.compile.backend") class LazyInitPass(InductorPass): """ If there's a pass that we want to initialize lazily in a test, we can wrap it in LazyInitPass, which will initialize the pass when invoked and then immediately invoke it. """ def __init__(self, pass_cls: type[VllmInductorPass], vllm_config: VllmConfig): self.pass_cls = pass_cls self.vllm_config = weakref.proxy(vllm_config) # avoid cycle def __call__(self, graph: fx.Graph) -> None: self.pass_ = self.pass_cls(self.vllm_config) self.pass_(graph) class TestBackend: """ This class provides a simple Inductor backend that can be used for testing. It takes a list of custom passes and runs them after Inductor's passes. It also saves the graph before and after the custom passes for inspection. Inductor config can be modified directly by editing the inductor_config property. This can be helpful for adding passes like the 'pre_grad_custom_pass' and the 'post_grad_custom_pre_pass'. Inductor config is default-initialized from VllmConfig.CompilationConfig. """ def __init__(self, *passes: InductorPass | Callable[[fx.Graph], None]): self.custom_passes = list(passes) vllm_config = get_current_vllm_config() compile_config = vllm_config.compilation_config # Deepcopy to allow multiple TestBackend instances to use the same VllmConfig self.inductor_config = deepcopy(compile_config.inductor_compile_config) self.inductor_config["force_disable_caches"] = True self.inductor_config["post_grad_custom_post_pass"] = self.post_pass if debug_dump_path := vllm_config.compile_debug_dump_path(): logger.debug("Dumping depyf output to %s", debug_dump_path) self.debug_ctx = depyf.prepare_debug(debug_dump_path.as_posix()) else: self.debug_ctx = nullcontext() def __call__(self, graph: fx.GraphModule, example_inputs): self.graph_pre_compile = deepcopy(graph) from torch._inductor.compile_fx import compile_fx with self.debug_ctx: return compile_fx( graph, example_inputs, config_patches=self.inductor_config ) @with_pattern_match_debug def post_pass(self, graph: fx.Graph): self.graph_pre_pass = deepcopy(graph) lazy_format_graph_code("graph_pre_pass", graph.owning_module) VllmInductorPass.dump_prefix = 0 for pass_ in self.custom_passes: pass_(graph) VllmInductorPass.dump_prefix += 1 VllmInductorPass.dump_prefix = None self.graph_post_pass = deepcopy(graph) lazy_format_graph_code("graph_post_pass", graph.owning_module) # assign by reference, will reflect the final state of the graph self.final_graph = graph def check_before_ops(self, ops: Sequence[OpOverload], fully_replaced=True): for op in ops: num_pre = len(list(find_op_nodes(op, self.graph_pre_pass))) num_post = len(list(find_op_nodes(op, self.graph_post_pass))) assert num_pre > 0, f"Op {op.name()} not found in pre-pass graph" assert num_pre > num_post, f"All nodes remain for op {op.name()}" if fully_replaced: assert num_post == 0, f"Unexpected op {op.name()} in post-pass graph" def check_after_ops(self, ops: Sequence[OpOverload]): for op in ops: num_pre = len(list(find_op_nodes(op, self.graph_pre_pass))) num_post = len(list(find_op_nodes(op, self.graph_post_pass))) assert num_pre == 0, f"Unexpected op {op.name()} in pre-pass graph" assert num_post > 0, f"Op {op.name()} not found in post-pass graph" def op_count(self, op: OpOverload, before=False) -> int: graph = self.graph_pre_pass if before else self.graph_post_pass return len(list(find_op_nodes(op, graph)))
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_pass_manager.py
tests/compile/test_pass_manager.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import copy import pytest import torch from vllm.compilation.inductor_pass import ( CallableInductorPass, InductorPass, pass_context, ) from vllm.compilation.pass_manager import PostGradPassManager from vllm.config import ModelConfig, VllmConfig from vllm.config.utils import Range # dummy custom pass that doesn't inherit def simple_callable(graph: torch.fx.Graph): pass # Should fail to add directly to the pass manager def test_bad_callable(): config = VllmConfig() pass_manager = PostGradPassManager() pass_manager.configure(config) with pytest.raises(AssertionError): pass_manager.add(simple_callable) # Pass that inherits from InductorPass class ProperPass(InductorPass): def __call__(self, graph: torch.fx.graph.Graph) -> None: pass @pytest.mark.parametrize( "callable", [ ProperPass(), # Can also wrap callables in CallableInductorPass for compliance CallableInductorPass(simple_callable), CallableInductorPass(simple_callable, InductorPass.hash_source(__file__)), ], ) def test_pass_manager_uuid(callable): # Set the pass context as PassManager uuid uses it with pass_context(Range(start=1, end=8)): # Some passes need dtype to be set config = VllmConfig(model_config=ModelConfig(dtype=torch.bfloat16)) pass_manager = PostGradPassManager() pass_manager.configure(config) # Check that UUID is different if the same pass is added 2x pass_manager.add(callable) uuid1 = pass_manager.uuid() pass_manager.add(callable) uuid2 = pass_manager.uuid() assert uuid1 != uuid2 # UUID should be the same as the original one, # as we constructed in the same way. pass_manager2 = PostGradPassManager() pass_manager2.configure(config) pass_manager2.add(callable) assert uuid1 == pass_manager2.uuid() # UUID should be different due to config change config2 = copy.deepcopy(config) config2.compilation_config.pass_config.fuse_norm_quant = ( not config2.compilation_config.pass_config.fuse_norm_quant ) config2.compilation_config.pass_config.fuse_act_quant = ( not config2.compilation_config.pass_config.fuse_act_quant ) pass_manager3 = PostGradPassManager() pass_manager3.configure(config2) pass_manager3.add(callable) assert uuid1 != pass_manager3.uuid()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/test_noop_elimination.py
tests/compile/test_noop_elimination.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch import vllm from vllm.compilation.noop_elimination import NoOpEliminationPass from vllm.config import CompilationConfig, CompilationMode, PassConfig, VllmConfig from .backend import TestBackend @pytest.mark.parametrize("dtype", [torch.float16, torch.bfloat16, torch.float32]) # Important edge case is when `num_tokens == buffer_size` @pytest.mark.parametrize( ("num_tokens", "buffer_size"), [(256, 256), (256, 512), (1024, 1024), (1024, 1025)] ) @pytest.mark.parametrize("hidden_size", [64, 4096]) def test_noop_elimination(dtype, num_tokens, hidden_size, buffer_size): torch.set_default_device("cuda") torch.set_default_dtype(dtype) torch.manual_seed(1) class Model(torch.nn.Module): def __init__(self) -> None: super().__init__() self.pos_embed = torch.empty(buffer_size, hidden_size, dtype=dtype) def forward(self, x): x += self.pos_embed[: x.shape[0]] # Chain of reshapes y = x.reshape(-1, 128, 32) z = y.reshape(-1, 4096) # No-op reshape a = z.reshape(-1, 4096) # Final reshape that should remain b = a.reshape(-1, 128, 32) # No-op slice c = b[0 : b.shape[0]] # The pass should replace the result of this op with `c`. d = torch.slice_scatter( torch.ones_like(c), # Dummy tensor to be scattered into c, # Source tensor 0, # dim 0, # start c.shape[0], # end ) return d vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, pass_config=PassConfig(eliminate_noops=True), ) ) with vllm.config.set_current_vllm_config(vllm_config): noop_pass = NoOpEliminationPass(vllm_config) backend = TestBackend(noop_pass) model = Model() # First dimension dynamic x = torch.rand(num_tokens, hidden_size) torch._dynamo.mark_dynamic(x, 0) result = model(x) model2 = torch.compile(model, backend=backend) result2 = model2(x) ATOL, RTOL = (2e-3, 2e-3) torch.testing.assert_close(result, result2, atol=ATOL, rtol=RTOL) # The no-op reshape and slice should be eliminated. # The initial slice on the positional embedding should remain. # The chain of reshapes should be fused into a single reshape. assert backend.op_count(torch.ops.aten.reshape.default) == 1 assert backend.op_count(torch.ops.aten.slice.Tensor) == 1 assert backend.op_count(torch.ops.aten.slice_scatter.default) == 0 def test_non_noop_slice_preserved(): """Ensure that a slice with end=-1 (dropping last row) is NOT eliminated. Regression test for a bug where end=-1 was treated like an inferred dimension (reshape semantics) leading to incorrect elimination. """ torch.set_default_device("cuda") x = torch.randn(16, 16) class SliceModel(torch.nn.Module): def forward(self, x): base = x.clone() src = torch.ones(15, 16) y = torch.slice_scatter(base, src, dim=0, start=0, end=-1) return x[0:-1, :], y vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, pass_config=PassConfig(eliminate_noops=True), ) ) with vllm.config.set_current_vllm_config(vllm_config): noop_pass = NoOpEliminationPass(vllm_config) backend = TestBackend(noop_pass) model = SliceModel() ref = model(x) compiled = torch.compile(model, backend=backend) out = compiled(x) torch.testing.assert_close(ref, out) # The slice should remain (not a no-op). assert backend.op_count(torch.ops.aten.slice.Tensor) == 1 assert backend.op_count(torch.ops.aten.slice_scatter.default) == 1
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/distributed/test_async_tp.py
tests/compile/distributed/test_async_tp.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json import pytest import torch import vllm.envs as envs from vllm.compilation.collective_fusion import AsyncTPPass from vllm.config import ( CompilationConfig, CompilationMode, DeviceConfig, ModelConfig, PassConfig, VllmConfig, ) from vllm.distributed import ( tensor_model_parallel_all_gather, tensor_model_parallel_reduce_scatter, ) from vllm.distributed.parallel_state import ( init_distributed_environment, initialize_model_parallel, ) from vllm.platforms import current_platform from vllm.utils.system_utils import update_environment_variables from ...models.registry import HF_EXAMPLE_MODELS from ...utils import ( compare_two_settings, create_new_process_for_each_test, multi_gpu_test, ) from ..backend import TestBackend FP8_DTYPE = current_platform.fp8_dtype() prompts = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] class TestMMRSModel(torch.nn.Module): def __init__(self, hidden_size=16, dtype=torch.float16): super().__init__() self.hidden_size = hidden_size self.dtype = dtype self.gate_proj = torch.nn.Parameter( torch.empty((self.hidden_size * 2, hidden_size)), requires_grad=False ) # Initialize weights torch.nn.init.normal_(self.gate_proj, std=0.02) def forward(self, hidden_states): """ Forward pass implementing the mm + reduce scatter in the FX graph """ # Reshape input view = hidden_states.reshape(-1, self.hidden_size) # matrix multiplication permute = self.gate_proj.permute(1, 0) mm = torch.mm(view, permute) reduce_scatter = tensor_model_parallel_reduce_scatter(mm, dim=0) return reduce_scatter def ops_in_model_before(self): return [torch.ops.vllm.reduce_scatter.default] def ops_in_model_after(self): return [torch.ops.symm_mem.fused_matmul_reduce_scatter.default] class TestAGMMModel(torch.nn.Module): def __init__(self, hidden_size=16, dtype=torch.float16): super().__init__() self.hidden_size = hidden_size self.dtype = dtype self.weight = torch.nn.Parameter( torch.empty((hidden_size, hidden_size)), requires_grad=False ) # Initialize weights torch.nn.init.normal_(self.weight, std=0.02) def forward(self, hidden_states): """ Forward pass implementing the mm + all gather in the FX graph """ # Reshape input view = hidden_states.reshape(-1, self.hidden_size) all_gather = tensor_model_parallel_all_gather(view, dim=0) permute = self.weight.permute(1, 0) mm = torch.mm(all_gather, permute) return mm def ops_in_model_before(self): return [torch.ops.vllm.all_gather.default] def ops_in_model_after(self): return [torch.ops.symm_mem.fused_all_gather_matmul.default] class _BaseScaledMMModel(torch.nn.Module): def __init__(self, hidden_size=16, dtype=torch.float16): super().__init__() self.hidden_size = hidden_size self.dtype = dtype self.weight = ( torch.empty([hidden_size, hidden_size], dtype=FP8_DTYPE) .contiguous() .transpose(0, 1) ) # Initialize scale_b for _scaled_mm. self.scale_b = torch.ones(1, self.hidden_size, dtype=torch.float32) class TestScaledMMRSModel(_BaseScaledMMModel): def forward(self, input: torch.Tensor): """ Forward pass implementing the scaled_mm + reduce scatter in the FX graph """ fp8_input = input.to(FP8_DTYPE) scale_a = torch.ones(input.shape[0], 1, dtype=torch.float32) scaled_mm = torch._scaled_mm( fp8_input, self.weight, scale_a=scale_a, scale_b=self.scale_b, out_dtype=self.dtype, ) reduce_scatter = tensor_model_parallel_reduce_scatter(scaled_mm, dim=0) return reduce_scatter def ops_in_model_before(self): return [torch.ops.vllm.reduce_scatter.default] def ops_in_model_after(self): return [torch.ops.vllm.patched_fused_scaled_matmul_reduce_scatter.default] class TestAGScaledMMModel(_BaseScaledMMModel): def forward(self, input: torch.Tensor): """ Forward pass implementing the all gather + scaled_mm in the FX graph """ # Reshape input fp8_input = input.to(FP8_DTYPE) all_gather = tensor_model_parallel_all_gather(fp8_input, dim=0) scale_a = torch.ones(all_gather.shape[0], 1, dtype=torch.float32) scaled_mm = torch._scaled_mm( all_gather, self.weight, scale_a=scale_a, scale_b=self.scale_b, out_dtype=self.dtype, ) return scaled_mm def ops_in_model_before(self): return [torch.ops.vllm.all_gather.default] def ops_in_model_after(self): return [torch.ops.symm_mem.fused_all_gather_scaled_matmul.default] class TestCutlassScaledMMRSModel(_BaseScaledMMModel): def forward(self, input: torch.Tensor): """ Forward pass implementing the cutlass_scaled_mm + reduce scatter in the FX graph """ fp8_input = input.to(FP8_DTYPE) scale_a = torch.ones(input.shape[0], 1, dtype=torch.float32) mm_out = torch.empty( (fp8_input.shape[0], self.weight.shape[1]), dtype=self.dtype, device=input.device, ) torch.ops._C.cutlass_scaled_mm( mm_out, fp8_input, self.weight, scale_a, self.scale_b, None ) reduce_scatter = tensor_model_parallel_reduce_scatter(mm_out, dim=0) return reduce_scatter def ops_in_model_before(self): return [torch.ops.vllm.reduce_scatter.default] def ops_in_model_after(self): return [torch.ops.vllm.patched_fused_scaled_matmul_reduce_scatter.default] class TestAGCutlassScaledMMModel(_BaseScaledMMModel): def forward(self, input: torch.Tensor): """ Forward pass implementing the all gather + cutlass_scaled_mm in the FX graph """ # Reshape input fp8_input = input.to(FP8_DTYPE) all_gather = tensor_model_parallel_all_gather(fp8_input, dim=0) scale_a = torch.ones(all_gather.shape[0], 1, dtype=torch.float32) mm_out = torch.empty( (all_gather.shape[0], self.weight.shape[1]), dtype=self.dtype, device=all_gather.device, ) torch.ops._C.cutlass_scaled_mm( mm_out, all_gather, self.weight, scale_a, self.scale_b, None ) return mm_out def ops_in_model_before(self): return [torch.ops.vllm.all_gather.default] def ops_in_model_after(self): return [torch.ops.symm_mem.fused_all_gather_scaled_matmul.default] @multi_gpu_test(num_gpus=2) @pytest.mark.parametrize( "test_model", [ TestMMRSModel, TestAGMMModel, TestScaledMMRSModel, TestAGScaledMMModel, TestCutlassScaledMMRSModel, TestAGCutlassScaledMMModel, ], ) @pytest.mark.parametrize("batch_size", [8]) @pytest.mark.parametrize("seq_len", [16]) @pytest.mark.parametrize("hidden_size", [16]) @pytest.mark.parametrize("dtype", [torch.float16, torch.bfloat16]) @pytest.mark.parametrize("dynamic", [True, False]) @pytest.mark.skipif(envs.VLLM_TARGET_DEVICE not in ["cuda"], reason="Only test on CUDA") def test_async_tp_pass_replace( test_model: str, batch_size: int, seq_len: int, hidden_size: int, dtype: torch.dtype, dynamic: bool, ): if ( test_model in ( TestScaledMMRSModel, TestAGScaledMMModel, TestCutlassScaledMMRSModel, TestAGCutlassScaledMMModel, ) and dtype == torch.float16 ): pytest.skip( "Only bf16 high precision output types are supported for " "per-token (row-wise) scaling" ) num_processes = 2 def run_torch_spawn(fn, nprocs): # need to use torch.mp.spawn otherwise will have problems with # torch.distributed and cuda torch.multiprocessing.spawn( fn, args=( num_processes, test_model, batch_size, seq_len, hidden_size, dtype, dynamic, ), nprocs=nprocs, ) run_torch_spawn(async_tp_pass_on_test_model, num_processes) def async_tp_pass_on_test_model( local_rank: int, world_size: int, test_model_cls: torch.nn.Module, batch_size: int, seq_len: int, hidden_size: int, dtype: torch.dtype, dynamic: bool, ): current_platform.seed_everything(0) device = torch.device(f"cuda:{local_rank}") torch.cuda.set_device(device) torch.set_default_device(device) torch.set_default_dtype(dtype) update_environment_variables( { "RANK": str(local_rank), "LOCAL_RANK": str(local_rank), "WORLD_SIZE": str(world_size), "MASTER_ADDR": "localhost", "MASTER_PORT": "12345", } ) # initialize distributed init_distributed_environment() initialize_model_parallel(tensor_model_parallel_size=world_size) # configure vllm config for SequenceParallelismPass vllm_config = VllmConfig() vllm_config.compilation_config = CompilationConfig( pass_config=PassConfig( fuse_gemm_comms=True, ), ) vllm_config.device_config = DeviceConfig(device=torch.device("cuda")) # this is a fake model name to construct the model config # in the vllm_config, it's not really used. model_name = "RedHatAI/Llama-3.2-1B-Instruct-FP8" vllm_config.model_config = ModelConfig( model=model_name, trust_remote_code=True, dtype=dtype, seed=42 ) async_tp_pass = AsyncTPPass(vllm_config) backend = TestBackend(async_tp_pass) assert ( async_tp_pass.compilation_config.splitting_ops == vllm_config.compilation_config.splitting_ops ) assert ( async_tp_pass.compilation_config.use_inductor_graph_partition == vllm_config.compilation_config.use_inductor_graph_partition ) model = test_model_cls(hidden_size, dtype) # Pass dtype to model constructor hidden_states = torch.randn( (batch_size * seq_len, hidden_size), dtype=dtype, requires_grad=False ) if dynamic: torch._dynamo.mark_dynamic(hidden_states, 0) compiled_model = torch.compile(model, backend=backend) compiled_model(hidden_states) assert async_tp_pass.matched_count == 1 # In pre-nodes, all gather or reduce scatter should exist, # fused_matmul_reduce_scatter or fused_all_gather_matmul should not backend.check_before_ops(model.ops_in_model_before(), fully_replaced=False) # In post-nodes, fused_matmul_reduce_scatter or \ # fused_all_gather_matmul should exist backend.check_after_ops(model.ops_in_model_after()) @create_new_process_for_each_test() @pytest.mark.parametrize( "model_id", ["meta-llama/Llama-3.2-1B-Instruct", "RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8"], ) @pytest.mark.parametrize("tp_size", [2]) @pytest.mark.parametrize("async_tp_enabled", [True]) @pytest.mark.parametrize("distributed_backend", ["mp"]) @pytest.mark.parametrize("eager_mode", [False, True]) def test_async_tp_pass_correctness( model_id: str, tp_size: int, async_tp_enabled: bool, distributed_backend: str, eager_mode: bool, num_gpus_available: int, ): model_info = HF_EXAMPLE_MODELS.find_hf_info(model_id) model_info.check_transformers_version(on_fail="skip") model_info.check_available_online(on_fail="skip") pp_size = 1 if num_gpus_available < tp_size: pytest.skip(f"Need at least {tp_size} x {pp_size} GPUs") common_args = [ "--dtype", "bfloat16", "--max-model-len", "2048", "--max-num-seqs", "8", ] if eager_mode: common_args.append("--enforce-eager") compilation_config = { "mode": CompilationMode.VLLM_COMPILE, "compile_sizes": [2, 4, 8], "splitting_ops": [], "pass_config": {"fuse_gemm_comms": async_tp_enabled}, } async_tp_args = [ *common_args, "--tensor-parallel-size", str(tp_size), "--distributed-executor-backend", distributed_backend, "--compilation_config", json.dumps(compilation_config), ] tp_args = [ *common_args, "--tensor-parallel-size", str(tp_size), "--distributed-executor-backend", "mp", ] compare_two_settings(model_id, async_tp_args, tp_args, method="generate")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/distributed/__init__.py
tests/compile/distributed/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/distributed/test_sequence_parallelism.py
tests/compile/distributed/test_sequence_parallelism.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pytest import torch import vllm.envs as envs from vllm.compilation.fusion import RMSNormQuantFusionPass from vllm.compilation.fx_utils import find_auto_fn from vllm.compilation.noop_elimination import NoOpEliminationPass from vllm.compilation.post_cleanup import PostCleanupPass from vllm.compilation.sequence_parallelism import SequenceParallelismPass from vllm.compilation.vllm_inductor_pass import VllmInductorPass from vllm.config import ( CompilationConfig, CUDAGraphMode, DeviceConfig, ModelConfig, PassConfig, VllmConfig, get_current_vllm_config, set_current_vllm_config, ) from vllm.distributed import tensor_model_parallel_all_reduce from vllm.distributed.parallel_state import ( init_distributed_environment, initialize_model_parallel, ) from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape from vllm.model_executor.layers.quantization.utils.w8a8_utils import Fp8LinearOp from vllm.platforms import current_platform from vllm.utils.system_utils import update_environment_variables from ...utils import multi_gpu_test from ..backend import TestBackend FP8_DTYPE = current_platform.fp8_dtype() prompts = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] class TestAllReduceRMSNormModel(torch.nn.Module): def __init__(self, hidden_size=16, eps=1e-6): super().__init__() self.hidden_size = hidden_size self.eps = eps self.norm = [RMSNorm(hidden_size, eps) for i in range(4)] self.w = [torch.rand(hidden_size, hidden_size) for _ in range(3)] def forward(self, x): z = torch.relu(x) x = resid = tensor_model_parallel_all_reduce(z) y = self.norm[0](x) z2 = torch.mm(y, self.w[0]) x2 = tensor_model_parallel_all_reduce(z2) y2, resid = self.norm[1](x2, resid) z3 = torch.mm(y2, self.w[1]) x3 = tensor_model_parallel_all_reduce(z3) y3, resid = self.norm[2](x3, resid) z4 = torch.mm(y3, self.w[2]) x4 = tensor_model_parallel_all_reduce(z4) y4, resid = self.norm[3](x4, resid) return y4 def ops_in_model_before(self): return [torch.ops.vllm.all_reduce.default] def ops_in_model_after(self): return [ torch.ops.vllm.all_gather.default, torch.ops.vllm.reduce_scatter.default, ] def ops_in_model(self): if RMSNorm.enabled(): return [ torch.ops._C.rms_norm.default, torch.ops._C.fused_add_rms_norm.default, ] else: return [] class TestAllReduceRMSNormStaticQuantFP8Model(torch.nn.Module): def __init__(self, hidden_size=16, eps=1e-6): super().__init__() self.vllm_config = get_current_vllm_config() self.hidden_size = hidden_size self.eps = eps self.norm = [RMSNorm(hidden_size, eps) for i in range(4)] self.wscale = [torch.rand(1, dtype=torch.float32) for _ in range(3)] self.w = [ torch.rand(hidden_size, hidden_size) .to(dtype=current_platform.fp8_dtype()) .t() for _ in range(3) ] self.fp8_linear = Fp8LinearOp( act_quant_static=True, act_quant_group_shape=GroupShape.PER_TENSOR, ) self.scale = [torch.rand(1, dtype=torch.float32) for _ in range(3)] def forward(self, hidden_states): # avoid having graph input be an arg to a pattern directly z = torch.relu(hidden_states) x = resid = tensor_model_parallel_all_reduce(z) y = self.norm[0](x) z2 = self.fp8_linear.apply( y, self.w[0], self.wscale[0], input_scale=self.scale[0] ) x2 = tensor_model_parallel_all_reduce(z2) y2, resid = self.norm[1](x2, resid) z3 = self.fp8_linear.apply( y2, self.w[1], self.wscale[1], input_scale=self.scale[1] ) x3 = tensor_model_parallel_all_reduce(z3) y3, resid = self.norm[2](x3, resid) # use resid here z4 = self.fp8_linear.apply( y3, self.w[2], self.wscale[2], input_scale=self.scale[2] ) x4 = tensor_model_parallel_all_reduce(z4) y4, resid = self.norm[3](x4, resid) # use resid here return y4 def ops_in_model_after(self): return [ torch.ops.vllm.all_gather.default, torch.ops.vllm.reduce_scatter.default, ] def ops_in_model_before(self): return [ torch.ops.vllm.all_reduce.default, ] def ops_in_model(self): if self.vllm_config.compilation_config.pass_config.fuse_norm_quant: return [torch.ops._C.fused_add_rms_norm_static_fp8_quant.default] elif RMSNorm.enabled(): return [ torch.ops._C.fused_add_rms_norm.default, ] elif self.fp8_linear.quant_fp8.enabled(): return [ torch.ops._C.static_scaled_fp8_quant.default, ] else: return [] @multi_gpu_test(num_gpus=2) @pytest.mark.parametrize( "test_model_cls, custom_ops", [ (TestAllReduceRMSNormModel, "+rms_norm"), (TestAllReduceRMSNormModel, "-rms_norm"), (TestAllReduceRMSNormStaticQuantFP8Model, "+rms_norm,+quant_fp8"), (TestAllReduceRMSNormStaticQuantFP8Model, "+rms_norm,-quant_fp8"), (TestAllReduceRMSNormStaticQuantFP8Model, "-rms_norm,+quant_fp8"), (TestAllReduceRMSNormStaticQuantFP8Model, "-rms_norm,-quant_fp8"), ], ) @pytest.mark.parametrize("batch_size", [8]) @pytest.mark.parametrize("seq_len", [16]) @pytest.mark.parametrize("hidden_size", [16]) @pytest.mark.parametrize("dtype", [torch.float16, torch.bfloat16]) @pytest.mark.parametrize("fuse_norm_quant", [True, False]) @pytest.mark.parametrize("dynamic", [False, True]) @pytest.mark.skipif(envs.VLLM_TARGET_DEVICE not in ["cuda"], reason="Only test on CUDA") def test_sequence_parallelism_pass( test_model_cls: type[torch.nn.Module], custom_ops: str, batch_size: int, seq_len: int, hidden_size: int, dtype: torch.dtype, fuse_norm_quant: bool, dynamic: bool, ): num_processes = 2 def run_torch_spawn(fn, nprocs): # need to use torch.mp.spawn otherwise will have problems with # torch.distributed and cuda torch.multiprocessing.spawn( fn, args=( num_processes, test_model_cls, custom_ops, batch_size, seq_len, hidden_size, dtype, fuse_norm_quant, dynamic, ), nprocs=nprocs, ) run_torch_spawn(sequence_parallelism_pass_on_test_model, num_processes) def sequence_parallelism_pass_on_test_model( local_rank: int, world_size: int, test_model_cls: type[torch.nn.Module], custom_ops: str, batch_size: int, seq_len: int, hidden_size: int, dtype: torch.dtype, fuse_norm_quant: bool, dynamic: bool, ): current_platform.seed_everything(0) device = torch.device(f"cuda:{local_rank}") torch.cuda.set_device(device) torch.set_default_device(device) torch.set_default_dtype(dtype) update_environment_variables( { "RANK": str(local_rank), "LOCAL_RANK": str(local_rank), "WORLD_SIZE": str(world_size), "MASTER_ADDR": "localhost", "MASTER_PORT": "12345", } ) # initialize distributed init_distributed_environment() initialize_model_parallel(tensor_model_parallel_size=world_size) # configure vllm config for SequenceParallelismPass custom_ops_list = custom_ops.split(",") if custom_ops else [] compilation_config = CompilationConfig( splitting_ops=[], # avoid automatic rms_norm enablement cudagraph_mode=CUDAGraphMode.NONE, # avoid piecewise warnings custom_ops=custom_ops_list, pass_config=PassConfig( enable_sp=True, fuse_norm_quant=fuse_norm_quant, eliminate_noops=True, ), ) # NoOp needed for fusion device_config = DeviceConfig(device=torch.device("cuda")) # this is a fake model name to construct the model config # in the vllm_config, it's not really used. model_name = "RedHatAI/Llama-3.2-1B-Instruct-FP8" model_config = ModelConfig( model=model_name, trust_remote_code=True, dtype=dtype, seed=42 ) vllm_config = VllmConfig( model_config=model_config, device_config=device_config, compilation_config=compilation_config, ) with set_current_vllm_config(vllm_config): noop_pass = NoOpEliminationPass(vllm_config) sequence_parallelism_pass = SequenceParallelismPass(vllm_config) cleanup_pass = PostCleanupPass(vllm_config) assert ( sequence_parallelism_pass.compilation_config.splitting_ops == vllm_config.compilation_config.splitting_ops ) assert ( sequence_parallelism_pass.compilation_config.use_inductor_graph_partition == vllm_config.compilation_config.use_inductor_graph_partition ) passes_for_backend: list[VllmInductorPass] = [ noop_pass, sequence_parallelism_pass, ] if fuse_norm_quant: fusion_pass = RMSNormQuantFusionPass(vllm_config) passes_for_backend.append(fusion_pass) passes_for_backend.append(cleanup_pass) backend = TestBackend(*passes_for_backend) model = test_model_cls(hidden_size) hidden_states = torch.randn((batch_size * seq_len, hidden_size), dtype=dtype) if dynamic: torch._dynamo.mark_dynamic(hidden_states, 0) compiled_model = torch.compile(model, backend=backend) compiled_model(hidden_states) assert sequence_parallelism_pass.matched_count == 4 # In pre-nodes, all reduce should be there, # reduce scatter and all gather should not for op in model.ops_in_model_before(): assert backend.op_count(op, before=True) == 4 # In post-nodes, reduce scatter and all gather should be there, # all reduce should not for op in model.ops_in_model_after(): assert backend.op_count(op, before=False) == 4 for op in model.ops_in_model(): find_auto_fn(backend.graph_post_pass.nodes, op)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/distributed/test_fusions_e2e.py
tests/compile/distributed/test_fusions_e2e.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from __future__ import annotations import itertools import logging from collections.abc import Iterable from typing import Any, NamedTuple import pytest import regex as re from tests.v1.attention.utils import AttentionBackendEnum from vllm import LLM, SamplingParams from vllm.config import CompilationConfig, CompilationMode, CUDAGraphMode, PassConfig from vllm.platforms import current_platform from vllm.utils.flashinfer import has_flashinfer from vllm.utils.torch_utils import is_torch_equal_or_newer from ...utils import flat_product, multi_gpu_test is_blackwell = lambda: current_platform.is_device_capability_family(100) """Are we running on Blackwell, a lot of tests depend on it""" class Matches(NamedTuple): attention_fusion: int = 0 allreduce_fusion: int = 0 rms_quant_norm_fusion: int = 0 sequence_parallel: int = 0 async_tp: int = 0 class ModelBackendTestCase(NamedTuple): model_name: str model_kwargs: dict[str, Any] backend: AttentionBackendEnum matches: Matches MODELS_FP8: list[ModelBackendTestCase] = [] MODELS_FP4: list[ModelBackendTestCase] = [] MODELS_GROUP_FP8: list[ModelBackendTestCase] = [] MODELS: list[ModelBackendTestCase] = [] # tp-only if current_platform.is_cuda(): MODELS_FP8 = [ ModelBackendTestCase( # Use smaller model for L40s in CI model_name="RedHatAI/Meta-Llama-3.1-8B-Instruct-FP8", model_kwargs=dict(max_model_len=1024, kv_cache_dtype="fp8"), backend=AttentionBackendEnum.TRITON_ATTN, matches=Matches( attention_fusion=32, allreduce_fusion=65, sequence_parallel=65, async_tp=128, ), ), ModelBackendTestCase( model_name="nvidia/Llama-4-Scout-17B-16E-Instruct-FP8", model_kwargs=dict(max_model_len=1024, kv_cache_dtype="fp8"), # TODO FlashInfer attn broken on Hopper with kvcache=fp8: # https://github.com/vllm-project/vllm/issues/28568 backend=AttentionBackendEnum.FLASHINFER if is_blackwell() else AttentionBackendEnum.TRITON_ATTN, matches=Matches( attention_fusion=48, allreduce_fusion=96, sequence_parallel=96, async_tp=95, # mlp is moe, no fusion there ), ), ] MODELS_FP4 = [ ModelBackendTestCase( model_name="nvidia/Llama-3.1-8B-Instruct-FP4", model_kwargs=dict(max_model_len=1024, kv_cache_dtype="fp8"), backend=AttentionBackendEnum.FLASHINFER, matches=Matches( attention_fusion=32, allreduce_fusion=65, sequence_parallel=65, async_tp=128, ), ), ] # TP only MODELS = [ ModelBackendTestCase( model_name="meta-llama/Llama-3.1-8B-Instruct", model_kwargs=dict(max_model_len=1024), backend=AttentionBackendEnum.TRITON_ATTN, matches=Matches( attention_fusion=0, allreduce_fusion=65, sequence_parallel=65, async_tp=128, ), ), ModelBackendTestCase( model_name="Qwen/Qwen3-30B-A3B", model_kwargs=dict(max_model_len=1024), backend=AttentionBackendEnum.TRITON_ATTN, matches=Matches( attention_fusion=0, allreduce_fusion=97, sequence_parallel=97, async_tp=96, # MLP is MoE, half the fusions of dense ), ), ] elif current_platform.is_rocm(): MODELS_FP8 = [ ModelBackendTestCase( model_name="amd/Llama-3.1-8B-Instruct-FP8-KV", model_kwargs=dict(max_model_len=1024), backend=AttentionBackendEnum.TRITON_ATTN, matches=Matches(attention_fusion=32), ), ModelBackendTestCase( model_name="amd/Llama-3.1-8B-Instruct-FP8-KV", model_kwargs=dict(max_model_len=1024), backend=AttentionBackendEnum.ROCM_ATTN, matches=Matches(attention_fusion=32), ), ModelBackendTestCase( model_name="amd/Llama-3.1-8B-Instruct-FP8-KV", model_kwargs=dict(max_model_len=1024), backend=AttentionBackendEnum.ROCM_AITER_UNIFIED_ATTN, matches=Matches(attention_fusion=32), ), ] CUSTOM_OPS_FP8 = ["-quant_fp8", "+quant_fp8"] def has_cuda_graph_wrapper_metadata() -> bool: from importlib import import_module try: module = import_module("torch._inductor.utils") module.CUDAGraphWrapperMetadata # noqa B018 except AttributeError: return False return True @pytest.mark.parametrize( "model_name, model_kwargs, backend, matches, custom_ops", # Test attention+quant_fp8 fusion with custom and torch impls of QuantFP8 list(flat_product(MODELS_FP8, CUSTOM_OPS_FP8)) # quant_fp4 only has the custom impl + list(flat_product(MODELS_FP4, [""])), ) @pytest.mark.parametrize( "inductor_graph_partition", [ pytest.param( True, marks=pytest.mark.skipif( not has_cuda_graph_wrapper_metadata(), reason="This test requires" "torch._inductor.utils.CUDAGraphWrapperMetadata to run", ), ), False, ], ) def test_attn_quant( model_name: str, model_kwargs: dict[str, Any], backend: AttentionBackendEnum, matches: Matches, custom_ops: str, inductor_graph_partition: bool, caplog_mp_spawn, monkeypatch, ): if backend == AttentionBackendEnum.FLASHINFER and ( not is_blackwell() or not has_flashinfer() ): pytest.skip("FlashInfer attn fusion requires Blackwell and flashinfer") if inductor_graph_partition and not is_torch_equal_or_newer("2.9.0.dev"): pytest.skip("Inductor graph partition requires torch>=2.9") custom_ops_list = custom_ops.split(",") if custom_ops else [] if inductor_graph_partition: mode = CUDAGraphMode.FULL_AND_PIECEWISE splitting_ops: list[str] | None = None else: # FIXME: Llama-4-Scout-17B-16E-Instruct-FP8 + FlashInfer + Blackwell end at # CUDAGraphMode.NONE here because it derives an attention backend that # does not support full cudagraphs mode = CUDAGraphMode.FULL_DECODE_ONLY splitting_ops = [] # Disable, compile cache to make sure custom passes run. # Otherwise, we can't verify fusion happened through the logs. monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", "1") # To capture subprocess logs, we need to know whether spawn or fork is used. # Force spawn as it is more general. monkeypatch.setenv("VLLM_WORKER_MULTIPROC_METHOD", "spawn") model_kwargs["attention_config"] = {"backend": backend.name} compilation_config = CompilationConfig( # Testing properties custom_ops=custom_ops_list, use_inductor_graph_partition=inductor_graph_partition, cudagraph_mode=mode, splitting_ops=splitting_ops, # Common mode=CompilationMode.VLLM_COMPILE, pass_config=PassConfig(fuse_attn_quant=True, eliminate_noops=True), # Inductor caches custom passes by default as well via uuid inductor_compile_config={"force_disable_caches": True}, ) with caplog_mp_spawn(logging.DEBUG) as log_holder: run_model(compilation_config, model_name, **model_kwargs) log_matches = re.findall( r"fusion_attn.py:\d+] Fused quant onto (\d+) attention nodes", log_holder.text, ) assert len(log_matches) == 1, log_holder.text assert int(log_matches[0]) == matches.attention_fusion CUSTOM_OPS_RMS_NORM = ["-rms_norm", "+rms_norm"] def custom_ops_product(*custom_ops_lists: list[str]) -> Iterable[str]: for op_list in itertools.product(*custom_ops_lists): yield ",".join(op_list) @multi_gpu_test(num_gpus=2) @pytest.mark.parametrize( "model_name, model_kwargs, backend, matches, custom_ops", # Toggle RMSNorm and QuantFP8 for FP8 models list( flat_product( MODELS_FP8, custom_ops_product(CUSTOM_OPS_FP8, CUSTOM_OPS_RMS_NORM) ) ) # Toggle RMSNorm for FP4 models and unquant models + list(flat_product(MODELS_FP4 + MODELS, CUSTOM_OPS_RMS_NORM)), ) @pytest.mark.parametrize("inductor_graph_partition", [True, False]) @pytest.mark.skipif( not current_platform.is_cuda() or not has_flashinfer() or not current_platform.has_device_capability(90), reason="allreduce+rmsnorm fusion requires flashinfer", ) def test_tp2_attn_quant_allreduce_rmsnorm( model_name: str, model_kwargs: dict, backend: AttentionBackendEnum, matches: Matches, custom_ops: str, inductor_graph_partition: bool, caplog_mp_spawn, monkeypatch, ): if inductor_graph_partition and not is_torch_equal_or_newer("2.9.0.dev"): pytest.skip("Inductor graph partition requires torch>=2.9") if "fp4" in model_name.lower() and not is_blackwell(): pytest.skip("NVFP4 quant requires Blackwell") if backend == AttentionBackendEnum.FLASHINFER and not is_blackwell(): # FlashInfer attn fusion requires Blackwell matches = matches._replace(attention_fusion=0) custom_ops_list = custom_ops.split(",") if custom_ops else [] if inductor_graph_partition: mode = CUDAGraphMode.FULL_AND_PIECEWISE splitting_ops: list[str] | None = None else: mode = CUDAGraphMode.FULL_DECODE_ONLY splitting_ops = [] # Disable, compile cache to make sure custom passes run. # Otherwise, we can't verify fusion happened through the logs. monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", "1") # To capture subprocess logs, we need to know whether spawn or fork is used. # Force spawn as it is more general. monkeypatch.setenv("VLLM_WORKER_MULTIPROC_METHOD", "spawn") model_kwargs["attention_config"] = {"backend": backend.name} compilation_config = CompilationConfig( # Testing properties use_inductor_graph_partition=inductor_graph_partition, cudagraph_mode=mode, custom_ops=custom_ops_list, splitting_ops=splitting_ops, # Common mode=CompilationMode.VLLM_COMPILE, pass_config=PassConfig( fuse_attn_quant=True, eliminate_noops=True, fuse_allreduce_rms=True, ), # Inductor caches custom passes by default as well via uuid inductor_compile_config={"force_disable_caches": True}, ) with caplog_mp_spawn(logging.DEBUG) as log_holder: run_model( compilation_config, model_name, tensor_parallel_size=2, **model_kwargs ) log_matches = re.findall( r"fusion_attn.py:\d+] Fused quant onto (\d+) attention nodes", log_holder.text, ) # 2 for each compile range # (global compile range can be split due to fuse_allreduce_rmsnorm) num_compile_ranges = len(compilation_config.get_compile_ranges()) assert num_compile_ranges in [1, 2] assert len(log_matches) == 2 * num_compile_ranges, log_holder.text assert all(int(log_match) == matches.attention_fusion for log_match in log_matches) log_matches = re.findall( r"collective_fusion.py:\d+] Replaced (\d+) patterns", log_holder.text, ) assert len(log_matches) == 2, log_holder.text assert int(log_matches[0]) == matches.allreduce_fusion assert int(log_matches[1]) == matches.allreduce_fusion log_matches = re.findall( r"pass_manager.py:\d+] Skipping .*AllReduceFusionPass.* with compile range", log_holder.text, ) assert len(log_matches) == 2 * (num_compile_ranges - 1), log_holder.text @multi_gpu_test(num_gpus=2) @pytest.mark.parametrize( "model_name, model_kwargs, backend, matches, custom_ops", # Toggle RMSNorm and QuantFP8 for FP8 models list( flat_product( MODELS_FP8, custom_ops_product(CUSTOM_OPS_FP8, CUSTOM_OPS_RMS_NORM) ) ) # Toggle RMSNorm for FP4 models and unquant models + list(flat_product(MODELS_FP4 + MODELS, CUSTOM_OPS_RMS_NORM)), ) @pytest.mark.parametrize("inductor_graph_partition", [True, False]) @pytest.mark.skipif( not current_platform.is_cuda(), reason="sequence parallel only tested on CUDA", ) def test_tp2_attn_quant_async_tp( model_name: str, model_kwargs: dict, backend: AttentionBackendEnum, matches: Matches, custom_ops: str, inductor_graph_partition: bool, caplog_mp_spawn, monkeypatch, ): if is_blackwell(): # TODO: https://github.com/vllm-project/vllm/issues/27893 pytest.skip("Blackwell is not supported for AsyncTP pass") if inductor_graph_partition and not is_torch_equal_or_newer("2.9.0.dev"): pytest.skip("Inductor graph partition requires torch>=2.9") if "fp4" in model_name.lower() and not is_blackwell(): pytest.skip("NVFP4 quant requires Blackwell") if backend == AttentionBackendEnum.FLASHINFER: if not has_flashinfer(): pytest.skip("FlashInfer backend requires flashinfer installed") if not is_blackwell(): # FlashInfer attn fusion requires Blackwell matches = matches._replace(attention_fusion=0) custom_ops_list = custom_ops.split(",") if custom_ops else [] if inductor_graph_partition: mode = CUDAGraphMode.FULL_AND_PIECEWISE splitting_ops: list[str] | None = None else: mode = CUDAGraphMode.FULL_DECODE_ONLY splitting_ops = [] # Disable, compile cache to make sure custom passes run. # Otherwise, we can't verify fusion happened through the logs. monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", "1") # To capture subprocess logs, we need to know whether spawn or fork is used. # Force spawn as it is more general. monkeypatch.setenv("VLLM_WORKER_MULTIPROC_METHOD", "spawn") model_kwargs["attention_config"] = {"backend": backend.name} compilation_config = CompilationConfig( # Testing properties use_inductor_graph_partition=inductor_graph_partition, cudagraph_mode=mode, custom_ops=custom_ops_list, splitting_ops=splitting_ops, # Common level=CompilationMode.VLLM_COMPILE, pass_config=PassConfig( fuse_attn_quant=True, eliminate_noops=True, enable_sp=True, fuse_gemm_comms=True, ), # Inductor caches custom passes by default as well via uuid inductor_compile_config={"force_disable_caches": True}, ) with caplog_mp_spawn(logging.DEBUG) as log_holder: run_model( compilation_config, model_name, tensor_parallel_size=2, **model_kwargs ) log_matches = re.findall( r"fusion_attn.py:\d+] Fused quant onto (\d+) attention nodes", log_holder.text, ) assert len(log_matches) == 2, log_holder.text assert int(log_matches[0]) == matches.attention_fusion assert int(log_matches[1]) == matches.attention_fusion log_matches = re.findall( r"sequence_parallelism.py:\d+] Replaced (\d+) patterns", log_holder.text, ) assert len(log_matches) == 2, log_holder.text assert int(log_matches[0]) == matches.sequence_parallel assert int(log_matches[1]) == matches.sequence_parallel log_matches = re.findall( r"collective_fusion.py:\d+] Replaced (\d+) patterns", log_holder.text, ) assert len(log_matches) == 2, log_holder.text assert int(log_matches[0]) == matches.async_tp assert int(log_matches[1]) == matches.async_tp def run_model(compile_config: int | CompilationConfig, model: str, **model_kwargs): compilation_config = ( compile_config if isinstance(compile_config, CompilationConfig) else CompilationConfig(mode=compile_config) ) prompts = [ "Hello, my name is", "The president of the United States is", "The capital of France is", "The future of AI is", ] sampling_params = SamplingParams(temperature=0) # Allow override from model_kwargs model_kwargs = {"tensor_parallel_size": 1, **model_kwargs} model_kwargs = {"disable_custom_all_reduce": True, **model_kwargs} # No cudagraphs by default if compilation_config.cudagraph_mode is None: compilation_config.cudagraph_mode = CUDAGraphMode.NONE llm = LLM( model=model, compilation_config=compilation_config, **model_kwargs, ) outputs = llm.generate(prompts, sampling_params) # Print the outputs. for output in outputs: prompt = output.prompt generated_text = output.outputs[0].text print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}") # Get the compile ranges split points after vllm config post init # in order to compute compile ranges correctly compilation_config.compile_ranges_split_points = ( llm.llm_engine.vllm_config.compilation_config.compile_ranges_split_points ) if current_platform.is_cuda(): MODELS_GROUP_FP8 = [ ModelBackendTestCase( model_name="Qwen/Qwen3-30B-A3B-FP8", model_kwargs=dict(max_model_len=1024, kv_cache_dtype="fp8"), backend=AttentionBackendEnum.TRITON_ATTN, matches=Matches( rms_quant_norm_fusion=48, ), ), ] CUSTOM_OPS_QUANT_RMS_NORM = ["+quant_fp8,+rms_norm"] @pytest.mark.parametrize( "model_name, model_kwargs, backend, matches, custom_ops", # Test rms norm+group quant_fp8 fusion list[tuple[Any, ...]](flat_product(MODELS_GROUP_FP8, CUSTOM_OPS_QUANT_RMS_NORM)), ) @pytest.mark.parametrize("inductor_graph_partition", [True, False]) # TODO: remove skip after we fix the fusion thoroughly @pytest.mark.skipif(is_blackwell(), reason="Temporarily disabled on Blackwell") def test_rms_group_quant( model_name: str, model_kwargs: dict[str, Any], backend: AttentionBackendEnum, matches: Matches, custom_ops: str, inductor_graph_partition: bool, caplog_mp_spawn, monkeypatch, ): if inductor_graph_partition and not is_torch_equal_or_newer("2.9.0.dev"): pytest.skip("Inductor graph partition requires torch>=2.9") custom_ops_list = custom_ops.split(",") if custom_ops else [] if inductor_graph_partition: mode = CUDAGraphMode.FULL_AND_PIECEWISE splitting_ops: list[str] | None = None else: mode = CUDAGraphMode.FULL_DECODE_ONLY splitting_ops = [] # Disable, compile cache to make sure custom passes run. # Otherwise, we can't verify fusion happened through the logs. monkeypatch.setenv("VLLM_DISABLE_COMPILE_CACHE", "1") # To capture subprocess logs, we need to know whether spawn or fork is used. # Force spawn as it is more general. monkeypatch.setenv("VLLM_WORKER_MULTIPROC_METHOD", "spawn") model_kwargs["attention_config"] = {"backend": backend.name} compilation_config = CompilationConfig( # Testing properties custom_ops=custom_ops_list, use_inductor_graph_partition=inductor_graph_partition, cudagraph_mode=mode, splitting_ops=splitting_ops, # Common mode=CompilationMode.VLLM_COMPILE, pass_config=PassConfig( fuse_norm_quant=True, fuse_act_quant=True, eliminate_noops=True ), # Inductor caches custom passes by default as well via uuid inductor_compile_config={"force_disable_caches": True}, ) with caplog_mp_spawn(logging.DEBUG) as log_holder: run_model(compilation_config, model_name, **model_kwargs) log_matches = re.findall( r"\[fusion.py:\d+] Replaced (\d+) patterns", log_holder.text, ) assert len(log_matches) == 1, log_holder.text assert int(log_matches[0]) == matches.rms_quant_norm_fusion
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/distributed/test_fusion_all_reduce.py
tests/compile/distributed/test_fusion_all_reduce.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from importlib.util import find_spec import pytest import torch import vllm.envs as envs from vllm._custom_ops import cutlass_scaled_fp4_mm, scaled_fp4_quant from vllm.compilation.collective_fusion import AllReduceFusionPass from vllm.compilation.fix_functionalization import FixFunctionalizationPass from vllm.compilation.noop_elimination import NoOpEliminationPass from vllm.compilation.post_cleanup import PostCleanupPass from vllm.config import ( CompilationConfig, CompilationMode, DeviceConfig, ModelConfig, PassConfig, VllmConfig, set_current_vllm_config, ) from vllm.distributed import tensor_model_parallel_all_reduce from vllm.distributed.parallel_state import ( init_distributed_environment, initialize_model_parallel, ) from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.quantization.utils.w8a8_utils import ( Fp8LinearOp, GroupShape, ) from vllm.platforms import current_platform from vllm.utils.system_utils import update_environment_variables from ...utils import has_module_attribute, multi_gpu_test from ..backend import TestBackend class TestAllReduceRMSNormModel(torch.nn.Module): def __init__(self, hidden_size=16, token_num=16, eps=1e-6): super().__init__() self.hidden_size = hidden_size self.eps = eps self.norm = [RMSNorm(hidden_size, eps) for i in range(4)] self.w = [torch.rand(hidden_size, hidden_size) for _ in range(3)] def forward(self, x): # avoid having graph input be an arg to a pattern directly z = torch.relu(x) x = resid = tensor_model_parallel_all_reduce(z) y = self.norm[0](x) z2 = torch.mm(y, self.w[0]) x2 = tensor_model_parallel_all_reduce(z2) y2, resid = self.norm[1](x2, resid) z3 = torch.mm(y2, self.w[1]) x3 = tensor_model_parallel_all_reduce(z3) y3, resid = self.norm[2](x3, resid) z4 = torch.mm(y3, self.w[2]) x4 = tensor_model_parallel_all_reduce(z4) y4, resid = self.norm[3](x4, resid) return y4 def ops_in_model_before(self): return [torch.ops.vllm.all_reduce.default] def ops_in_model_after(self): return [torch.ops.vllm.flashinfer_trtllm_fused_allreduce_norm.default] class TestAllReduceRMSNormStaticQuantFP8Model(torch.nn.Module): def __init__(self, hidden_size=16, token_num=16, eps=1e-6): super().__init__() self.hidden_size = hidden_size self.eps = eps self.norm = [RMSNorm(hidden_size, eps) for i in range(4)] self.wscale = [torch.rand(1, dtype=torch.float32) for _ in range(3)] self.w = [ torch.rand(hidden_size, hidden_size) .to(dtype=current_platform.fp8_dtype()) .t() for _ in range(3) ] self.fp8_linear = Fp8LinearOp( act_quant_static=True, act_quant_group_shape=GroupShape.PER_TENSOR, ) self.scale = [torch.rand(1, dtype=torch.float32) for _ in range(3)] def forward(self, hidden_states): # avoid having graph input be an arg to a pattern directly z = torch.relu(hidden_states) x = resid = tensor_model_parallel_all_reduce(z) y = self.norm[0](x) z2 = self.fp8_linear.apply( y, self.w[0], self.wscale[0], input_scale=self.scale[0] ) x2 = tensor_model_parallel_all_reduce(z2) y2, resid = self.norm[1](x2, resid) z3 = self.fp8_linear.apply( y2, self.w[1], self.wscale[1], input_scale=self.scale[1] ) x3 = tensor_model_parallel_all_reduce(z3) y3, resid = self.norm[2](x3, resid) # use resid here z4 = self.fp8_linear.apply( y3, self.w[2], self.wscale[2], input_scale=self.scale[2] ) x4 = tensor_model_parallel_all_reduce(z4) y4, resid = self.norm[3](x4, resid) # use resid here return y4 def ops_in_model_after(self): return [torch.ops.vllm.flashinfer_trtllm_fused_allreduce_norm.default] def ops_in_model_before(self): return [ torch.ops.vllm.all_reduce.default, torch.ops._C.static_scaled_fp8_quant.default if self.fp8_linear.quant_fp8.enabled() else torch.ops.aten.reciprocal.default, ] class TestAllReduceFusedAddRMSNormStaticQuantFP4Model(torch.nn.Module): def __init__(self, hidden_size=16, token_num=16, eps=1e-6): super().__init__() self.hidden_size = hidden_size self.eps = eps self.norm = [RMSNorm(hidden_size, eps) for i in range(4)] self.w = [torch.rand(hidden_size, hidden_size) for _ in range(3)] self.agscale = [torch.rand(1, dtype=torch.float32) for _ in range(3)] wgscale = [torch.rand(1, dtype=torch.float32) for _ in range(3)] self.alpha = [1 / (w * a) for w, a in zip(wgscale, self.agscale)] wq_gen, wscale_gen = zip( *(scaled_fp4_quant(w, wg) for w, wg in zip(self.w, wgscale)) ) self.wq, self.wscale = list(wq_gen), list(wscale_gen) print(f"{self.wq=}, {self.wscale=}") def forward(self, hidden_states): # avoid having graph input be an arg to a pattern directly z = torch.relu(hidden_states) x = resid = tensor_model_parallel_all_reduce(z) y = self.norm[0](x) yq, y_scale = scaled_fp4_quant(y, self.agscale[0]) z2 = cutlass_scaled_fp4_mm( yq, self.wq[0], y_scale, self.wscale[0], self.alpha[0], out_dtype=y.dtype ) x2 = tensor_model_parallel_all_reduce(z2) y2, resid = self.norm[1](x2, resid) yq2, y_scale2 = scaled_fp4_quant(y2, self.agscale[1]) z3 = cutlass_scaled_fp4_mm( yq2, self.wq[1], y_scale2, self.wscale[1], self.alpha[1], out_dtype=y2.dtype ) x3 = tensor_model_parallel_all_reduce(z3) y3, resid = self.norm[2](x3, resid) # use resid here yq3, y_scale3 = scaled_fp4_quant(y3, self.agscale[2]) z4 = cutlass_scaled_fp4_mm( yq3, self.wq[2], y_scale3, self.wscale[2], self.alpha[2], out_dtype=y3.dtype ) x4 = tensor_model_parallel_all_reduce(z4) y4, resid = self.norm[3](x4, resid) # use resid here return y4 def ops_in_model_after(self): return [torch.ops.vllm.flashinfer_trtllm_fused_allreduce_norm.default] def ops_in_model_before(self): return [ torch.ops.vllm.all_reduce.default, torch.ops._C.scaled_fp4_quant.default, ] @multi_gpu_test(num_gpus=2) @pytest.mark.parametrize( "test_model, enable_quant_fp8_custom_op", [ (TestAllReduceRMSNormModel, False), (TestAllReduceRMSNormStaticQuantFP8Model, True), (TestAllReduceRMSNormStaticQuantFP8Model, False), (TestAllReduceFusedAddRMSNormStaticQuantFP4Model, False), ], ) @pytest.mark.parametrize("batch_size", [8]) @pytest.mark.parametrize("seq_len", [8]) @pytest.mark.parametrize("hidden_size", [64]) @pytest.mark.parametrize("dtype", [torch.bfloat16]) @pytest.mark.parametrize("enable_rms_norm_custom_op", [True, False]) @pytest.mark.skipif(envs.VLLM_TARGET_DEVICE not in ["cuda"], reason="Only test on CUDA") @pytest.mark.skipif( not find_spec("flashinfer") or not has_module_attribute("flashinfer.comm", "trtllm_allreduce_fusion"), reason="flashinfer is not found or flashinfer " "is not compiled with trtllm_allreduce_fusion", ) def test_all_reduce_fusion_pass_replace( test_model: torch.nn.Module, batch_size: int, seq_len: int, hidden_size: int, dtype: torch.dtype, enable_rms_norm_custom_op, enable_quant_fp8_custom_op, ): num_processes = 2 if ( test_model == TestAllReduceFusedAddRMSNormStaticQuantFP4Model and not current_platform.has_device_capability(100) ): pytest.skip( "Skip as nvfp4 is only supported on " "devices with compute capability 10.0 (Blackwell)" ) def run_torch_spawn(fn, nprocs): torch.multiprocessing.spawn( fn, args=( num_processes, test_model, batch_size, seq_len, hidden_size, dtype, enable_rms_norm_custom_op, enable_quant_fp8_custom_op, ), nprocs=nprocs, ) run_torch_spawn(all_reduce_fusion_pass_on_test_model, num_processes) def all_reduce_fusion_pass_on_test_model( local_rank: int, world_size: int, test_model_cls: torch.nn.Module, batch_size: int, seq_len: int, hidden_size: int, dtype: torch.dtype, enable_rms_norm_custom_op, enable_quant_fp8_custom_op, ): current_platform.seed_everything(0) device = torch.device(f"cuda:{local_rank}") torch.cuda.set_device(device) torch.set_default_device(device) torch.set_default_dtype(dtype) update_environment_variables( { "RANK": str(local_rank), "LOCAL_RANK": str(local_rank), "WORLD_SIZE": str(world_size), "MASTER_ADDR": "localhost", "MASTER_PORT": "12345", } ) init_distributed_environment() initialize_model_parallel(tensor_model_parallel_size=world_size) custom_ops = [] if enable_rms_norm_custom_op: custom_ops.append("+rms_norm") if enable_quant_fp8_custom_op: custom_ops.append("+quant_fp8") vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, custom_ops=custom_ops ) ) vllm_config.compilation_config.pass_config = PassConfig( fuse_allreduce_rms=True, eliminate_noops=True ) vllm_config.device_config = DeviceConfig(device=torch.device("cuda")) vllm_config.parallel_config.rank = local_rank # Setup rank for debug path # this is a fake model name to construct the model config # in the vllm_config, it's not really used. model_name = "RedHatAI/Llama-3.2-1B-Instruct-FP8" vllm_config.model_config = ModelConfig( model=model_name, trust_remote_code=True, dtype=dtype, seed=42 ) with set_current_vllm_config(vllm_config): all_reduce_fusion_pass = AllReduceFusionPass(vllm_config) noop_pass = NoOpEliminationPass(vllm_config) func_pass = FixFunctionalizationPass(vllm_config) cleanup_pass = PostCleanupPass(vllm_config) backend = TestBackend( noop_pass, all_reduce_fusion_pass, func_pass, cleanup_pass ) token_num = batch_size * seq_len model = test_model_cls(hidden_size, token_num) hidden_states = torch.randn((token_num, hidden_size), requires_grad=False) compiled_model = torch.compile(model, backend=backend) compiled_model(hidden_states) assert all_reduce_fusion_pass.matched_count == 4, ( f"{all_reduce_fusion_pass.matched_count=}" ) backend.check_before_ops(model.ops_in_model_before(), fully_replaced=False) backend.check_after_ops(model.ops_in_model_after()) del all_reduce_fusion_pass
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/fullgraph/test_full_cudagraph.py
tests/compile/fullgraph/test_full_cudagraph.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import contextlib import os import weakref import pytest from tests.utils import wait_for_gpu_memory_to_clear from tests.v1.attention.utils import full_cg_backend_configs as backend_configs from vllm import LLM, SamplingParams from vllm.config import CompilationConfig from vllm.platforms import current_platform from vllm.utils.torch_utils import is_torch_equal_or_newer @contextlib.contextmanager def temporary_environ(env_vars): """ Temporarily set environment variables and restore them afterward. We have to do this vs monkeypatch because monkeypatch doesn't work with "module" scoped fixtures. """ original_env = {k: os.environ.get(k) for k in env_vars} try: os.environ.update(env_vars) yield finally: for k, v in original_env.items(): if v is None: os.environ.pop(k, None) else: os.environ[k] = v model_backends_full_cudagraph = [] # deepseek-ai/DeepSeek-V2-Lite with MLA MLA_backends = ["FlashMLA", "FlashAttentionMLA", "CutlassMLA"] for mla_backend in MLA_backends: model_backends_full_cudagraph.append( ("deepseek-ai/DeepSeek-V2-Lite", backend_configs[mla_backend]) ) # Qwen/Qwen2-1.5B-Instruct with other backends other_backend_configs = [ backend_configs[c] for c in backend_configs if c not in MLA_backends ] for backend_config in other_backend_configs: model_backends_full_cudagraph.append(("Qwen/Qwen2-1.5B-Instruct", backend_config)) @pytest.fixture(scope="class") def llm_pair(request): model, backend_config, use_inductor_graph_partition = request.param backend_config.comp_config["use_inductor_graph_partition"] = ( use_inductor_graph_partition ) if use_inductor_graph_partition and not is_torch_equal_or_newer("2.9.0.dev"): pytest.skip("Inductor graph partition only supported in torch>=2.9") # Dynamically skip test if GPU capability is not met if ( backend_config.specific_gpu_arch and backend_config.specific_gpu_arch != current_platform.get_device_capability() ): if backend_config.specific_gpu_arch == (9, 0): pytest.skip("Only Hopper GPUs support FA3 and FlashMLA") elif backend_config.specific_gpu_arch == (10, 0): pytest.skip("Only Blackwell GPUs support Cutlass MLA") env_vars = { # Force native sampler to avoid potential nondeterminism in FlashInfer # when per-request generators are not used in V1. "VLLM_USE_FLASHINFER_SAMPLER": "0", } with temporary_environ(env_vars): full = LLM( model=model, gpu_memory_utilization=0.43, trust_remote_code=True, max_model_len=1024, max_num_seqs=128, compilation_config=CompilationConfig(**backend_config.comp_config), generation_config="vllm", seed=42, ) piecewise = LLM( model=model, gpu_memory_utilization=0.43, trust_remote_code=True, max_model_len=1024, max_num_seqs=128, compilation_config=CompilationConfig(cudagraph_mode="PIECEWISE"), generation_config="vllm", seed=42, ) # PyTest caches the fixture values so we use weakref.proxy to enable GC yield weakref.proxy(full), weakref.proxy(piecewise) del full del piecewise wait_for_gpu_memory_to_clear( devices=[0], threshold_ratio=0.1, ) @pytest.mark.parametrize( "llm_pair", [ pytest.param((model, backend_config, use_inductor_graph_partition)) for model, backend_config in model_backends_full_cudagraph for use_inductor_graph_partition in [True, False] ], indirect=True, ) class TestFullCUDAGraph: """ Use a class such that an llm pair is constructed once for all batch_size/max_tokens combinations and released immediately after. Module-scope fixtures would stick around the whole time, meaning there would be multiple LLM instances hogging memory simultaneously. """ @pytest.mark.parametrize( ("batch_size", "max_tokens"), [ (1, 10), (7, 10), (16, 10), (25, 10), (32, 10), (45, 10), (64, 10), (123, 10), (8, 5), (8, 30), ], ) def test_full_cudagraph(self, batch_size, max_tokens, llm_pair: tuple[LLM, LLM]): """ Test various batch sizes and max_tokens to ensure that the full cudagraph compilation works for padded cases too. """ full_cudagraph_llm, piecewise_llm = llm_pair prompts = ["the quick brown fox"] * batch_size # Use purely greedy decoding to avoid top-p truncation sensitivity # that can amplify tiny numeric differences across runtimes. sampling_params = SamplingParams( temperature=0.0, max_tokens=max_tokens, top_p=1.0 ) piecewise_responses = piecewise_llm.generate(prompts, sampling_params) full_responses = full_cudagraph_llm.generate(prompts, sampling_params) # Check that all responses are the same for piecewise_res, full_res in zip(piecewise_responses, full_responses): assert ( piecewise_res.outputs[0].text.lower() == full_res.outputs[0].text.lower() ) @pytest.mark.skipif(not current_platform.is_cuda(), reason="Skip if not cuda") def test_full_cudagraph_with_invalid_backend(): # Flex_Attention is not supported with full cuda graph with pytest.raises(RuntimeError): LLM( model="Qwen/Qwen2-1.5B-Instruct", compilation_config=CompilationConfig(cudagraph_mode="FULL"), attention_config={"backend": "FLEX_ATTENTION"}, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/fullgraph/test_multiple_graphs.py
tests/compile/fullgraph/test_multiple_graphs.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Test (piecewise) compilation with a simple model where multiple submodules are compiled and graph captured separately. """ import pytest import torch from torch import nn from vllm.compilation.backends import set_model_tag from vllm.compilation.counter import compilation_counter from vllm.compilation.decorators import ignore_torch_compile, support_torch_compile from vllm.config import ( CompilationConfig, CompilationMode, CUDAGraphMode, VllmConfig, set_current_vllm_config, ) from vllm.forward_context import BatchDescriptor, set_forward_context from vllm.utils.torch_utils import is_torch_equal_or_newer from ...utils import create_new_process_for_each_test # This import automatically registers `torch.ops.silly.attention` from .. import silly_attention # noqa: F401 BATCH_SIZE = 32 MLP_SIZE = 128 HIDDEN_SIZE = 1024 RANDOM_SEED = 0 @support_torch_compile class ParentModel(nn.Module): def __init__(self, *, vllm_config: VllmConfig, prefix: str = "", **kwargs) -> None: super().__init__() def forward(self, x: torch.Tensor) -> torch.Tensor: return x class Attention(nn.Module): def __init__(self, mlp_size: int, hidden_size: int) -> None: super().__init__() self.pre_attn = nn.Linear(mlp_size, hidden_size, bias=False) self.post_attn = nn.Linear(hidden_size, mlp_size, bias=False) self.rms_norm_weight = nn.Parameter(torch.ones(hidden_size)) # Initialize to same weights for testing nn.init.xavier_normal_( self.pre_attn.weight.data, generator=torch.Generator().manual_seed(RANDOM_SEED), gain=0.001, ) nn.init.xavier_normal_( self.post_attn.weight.data, generator=torch.Generator().manual_seed(RANDOM_SEED), gain=0.001, ) def rms_norm_ref(self, x: torch.Tensor) -> torch.Tensor: x_f32 = x.float() return ( x_f32 * torch.rsqrt(torch.mean(x_f32.square(), dim=-1, keepdim=True) + 1e-6) * self.rms_norm_weight ).to(x.dtype) def forward(self, x: torch.Tensor) -> torch.Tensor: x = self.pre_attn(x) x = self.rms_norm_ref(x) attn_output = torch.empty_like(x) torch.ops.silly.attention(x, x, x, attn_output) x = attn_output x = self.rms_norm_ref(x) x = self.post_attn(x) return x @support_torch_compile class CompiledAttention(nn.Module): def __init__( self, *, mlp_size: int, hidden_size: int, vllm_config: VllmConfig, prefix: str = "", **kwargs, ) -> None: super().__init__() self.attn = Attention(mlp_size, hidden_size) def forward(self, x: torch.Tensor) -> torch.Tensor: return self.attn(x) @support_torch_compile class CompiledAttentionTwo(CompiledAttention): def forward(self, x: torch.Tensor) -> torch.Tensor: return self.attn(x) + x @ignore_torch_compile class SimpleModelWithTwoGraphs(ParentModel): def __init__( self, *, mlp_size: int, hidden_size: int, vllm_config: VllmConfig, prefix: str = "", **kwargs, ) -> None: super().__init__(vllm_config=vllm_config, prefix=prefix) # Test will fail without set_model_tag here with error: # "ValueError: too many values to unpack (expected 3)" # This is because CompiledAttention and CompiledAttentionTwo # have different implementations but the same torch.compile # cache dir will be used as default prefix is 'model_tag' with set_model_tag("attn_one"): self.attn_one = CompiledAttention( mlp_size=mlp_size, hidden_size=hidden_size, vllm_config=vllm_config, prefix=f"{prefix}.attn_one", ) with set_model_tag("attn_two"): self.attn_two = CompiledAttentionTwo( mlp_size=mlp_size, hidden_size=hidden_size, vllm_config=vllm_config, prefix=f"{prefix}.attn_two", ) self.hidden_states = torch.zeros((BATCH_SIZE, MLP_SIZE)).cuda() def forward(self, x: torch.Tensor) -> torch.Tensor: bsz = x.shape[0] # CUDAGraph expects same tensor addresses for each run self.hidden_states[:bsz].copy_(x) x = self.attn_one(self.hidden_states[:bsz]) self.hidden_states[:bsz].copy_(x) x = self.attn_two(self.hidden_states[:bsz]) return x @torch.inference_mode def run_model( vllm_config: VllmConfig, model: nn.Module, inputs: torch.Tensor, cudagraph_runtime_mode: CUDAGraphMode, ): with set_forward_context({}, vllm_config=vllm_config): # warmup for the model with cudagraph_mode NONE model(inputs) # simulate cudagraphs capturing with set_forward_context( {}, vllm_config=vllm_config, cudagraph_runtime_mode=cudagraph_runtime_mode, batch_descriptor=BatchDescriptor( num_tokens=2, ), ): model(inputs[:2]) with set_forward_context( {}, vllm_config=vllm_config, cudagraph_runtime_mode=cudagraph_runtime_mode, batch_descriptor=BatchDescriptor( num_tokens=1, ), ): model(inputs[:1]) # simulate cudagraphs replay with set_forward_context( {}, vllm_config=vllm_config, cudagraph_runtime_mode=cudagraph_runtime_mode, batch_descriptor=BatchDescriptor( num_tokens=2, ), ): output = model(inputs[:2]) output = output.cpu() return output.cpu() @pytest.mark.parametrize("use_inductor_graph_partition", [False, True]) @pytest.mark.parametrize("use_bytecode_hook", [True, False]) @create_new_process_for_each_test("spawn") def test_multi_graph_piecewise_compile( use_inductor_graph_partition: bool, use_bytecode_hook: bool, monkeypatch ): # Set the environment variable for this test monkeypatch.setenv("VLLM_USE_BYTECODE_HOOK", "1" if use_bytecode_hook else "0") if use_inductor_graph_partition and not is_torch_equal_or_newer("2.9.0.dev"): pytest.skip("inductor graph partition is only available in PyTorch 2.9+") outputs = [] # vllmcompile compile vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, cudagraph_mode=CUDAGraphMode.PIECEWISE, splitting_ops=["silly::attention"], cudagraph_capture_sizes=[1, 2], use_inductor_graph_partition=use_inductor_graph_partition, ) ) cudagraph_runtime_mode = CUDAGraphMode.PIECEWISE with set_current_vllm_config(vllm_config): model = ( SimpleModelWithTwoGraphs( mlp_size=MLP_SIZE, hidden_size=HIDDEN_SIZE, vllm_config=vllm_config, prefix="", ) .eval() .cuda() ) # Pre-allocate memory for CUDAGraph which expects # static tensor addresses inputs = torch.randn(BATCH_SIZE, MLP_SIZE).cuda() if use_inductor_graph_partition: # Splitting happens at Inductor lowering level, # total piecewise fx graphs is equal to total graphs num_piecewise_fx = 2 num_piecewise_capturable_fx = 2 else: # attn_one, attn_two each has 3 piecewise graphs # (pre attn, post attn, silly_attention) each num_piecewise_fx = 6 # attn_one, attn_two has pre attn and post attn each, total=4 num_piecewise_capturable_fx = 4 with compilation_counter.expect( num_graphs_seen=2, # two graphs for the model num_piecewise_graphs_seen=num_piecewise_fx, num_piecewise_capturable_graphs_seen=num_piecewise_capturable_fx, num_backend_compilations=num_piecewise_capturable_fx, num_cudagraph_captured=8, # num_cudagraph_sizes * num_partitions ): outputs.append(run_model(vllm_config, model, inputs, cudagraph_runtime_mode)) # no compile or cudagraph vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.NONE, ) ) cudagraph_runtime_mode = CUDAGraphMode.NONE with set_current_vllm_config(vllm_config): model = ( SimpleModelWithTwoGraphs( mlp_size=MLP_SIZE, hidden_size=HIDDEN_SIZE, vllm_config=vllm_config, prefix="", ) .eval() .cuda() ) with compilation_counter.expect( num_graphs_seen=0, num_piecewise_graphs_seen=0, num_piecewise_capturable_graphs_seen=0, num_backend_compilations=0, num_cudagraph_captured=0, ): outputs.append(run_model(vllm_config, model, inputs, cudagraph_runtime_mode)) # piecewise compile without CUDA graph vllm_config = VllmConfig( compilation_config=CompilationConfig( mode=CompilationMode.VLLM_COMPILE, cudagraph_mode=CUDAGraphMode.NONE, splitting_ops=["silly::attention"], use_inductor_graph_partition=use_inductor_graph_partition, ) ) cudagraph_runtime_mode = CUDAGraphMode.PIECEWISE with set_current_vllm_config(vllm_config): model = ( SimpleModelWithTwoGraphs( mlp_size=MLP_SIZE, hidden_size=HIDDEN_SIZE, vllm_config=vllm_config, prefix="", ) .eval() .cuda() ) with compilation_counter.expect( num_graphs_seen=2, num_piecewise_graphs_seen=num_piecewise_fx, num_piecewise_capturable_graphs_seen=num_piecewise_capturable_fx, num_backend_compilations=num_piecewise_capturable_fx, num_cudagraph_captured=0, # no cudagraph captured ): outputs.append(run_model(vllm_config, model, inputs, cudagraph_runtime_mode)) # Generally don't expect outputs with and without inductor # to be bitwise equivalent assert torch.allclose(outputs[0], outputs[1]) # Expect bitwise equivalence using inductor w/ and w/o cudagraph assert torch.equal(outputs[0], outputs[2])
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tests/compile/fullgraph/__init__.py
tests/compile/fullgraph/__init__.py
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false