repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/test_http_api/test_file_management_within_dataset/test_stop_parse_documents.py
sdk/python/test/test_http_api/test_file_management_within_dataset/test_stop_parse_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor from time import sleep import pytest from common import INVALID_API_TOKEN, bulk_upload_documents, list_documnets, parse_documnets, stop_parse_documnets from libs.auth import RAGFlowHttpApiAuth from libs.utils import wait_for def validate_document_parse_done(auth, dataset_id, document_ids): for document_id in document_ids: res = list_documnets(auth, dataset_id, params={"id": document_id}) doc = res["data"]["docs"][0] assert doc["run"] == "DONE" assert len(doc["process_begin_at"]) > 0 assert doc["process_duration"] > 0 assert doc["progress"] > 0 assert "Task done" in doc["progress_msg"] def validate_document_parse_cancel(auth, dataset_id, document_ids): for document_id in document_ids: res = list_documnets(auth, dataset_id, params={"id": document_id}) doc = res["data"]["docs"][0] assert doc["run"] == "CANCEL" assert len(doc["process_begin_at"]) > 0 assert doc["progress"] == 0.0 @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, auth, expected_code, expected_message): res = stop_parse_documnets(auth, "dataset_id") assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.skip class TestDocumentsParseStop: @pytest.mark.parametrize( "payload, expected_code, expected_message", [ pytest.param(None, 102, """AttributeError("\'NoneType\' object has no attribute \'get\'")""", marks=pytest.mark.skip), pytest.param({"document_ids": []}, 102, "`document_ids` is required", marks=pytest.mark.p1), pytest.param({"document_ids": ["invalid_id"]}, 102, "You don't own the document invalid_id.", marks=pytest.mark.p3), pytest.param({"document_ids": ["\n!?。;!?\"'"]}, 102, """You don\'t own the document \n!?。;!?"\'.""", marks=pytest.mark.p3), pytest.param("not json", 102, "AttributeError(\"'str' object has no attribute 'get'\")", marks=pytest.mark.skip), pytest.param(lambda r: {"document_ids": r[:1]}, 0, "", marks=pytest.mark.p1), pytest.param(lambda r: {"document_ids": r}, 0, "", marks=pytest.mark.p1), ], ) def test_basic_scenarios(self, get_http_api_auth, add_documents_func, payload, expected_code, expected_message): @wait_for(10, 1, "Document parsing timeout") def condition(_auth, _dataset_id, _document_ids): for _document_id in _document_ids: res = list_documnets(_auth, _dataset_id, {"id": _document_id}) if res["data"]["docs"][0]["run"] != "DONE": return False return True dataset_id, document_ids = add_documents_func parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) if callable(payload): payload = payload(document_ids) res = stop_parse_documnets(get_http_api_auth, dataset_id, payload) assert res["code"] == expected_code if expected_code == 0: completed_document_ids = list(set(document_ids) - set(payload["document_ids"])) condition(get_http_api_auth, dataset_id, completed_document_ids) validate_document_parse_cancel(get_http_api_auth, dataset_id, payload["document_ids"]) validate_document_parse_done(get_http_api_auth, dataset_id, completed_document_ids) else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "invalid_dataset_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id.", ), ], ) def test_invalid_dataset_id( self, get_http_api_auth, add_documents_func, invalid_dataset_id, expected_code, expected_message, ): dataset_id, document_ids = add_documents_func parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) res = stop_parse_documnets(get_http_api_auth, invalid_dataset_id, {"document_ids": document_ids}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.skip @pytest.mark.parametrize( "payload", [ lambda r: {"document_ids": ["invalid_id"] + r}, lambda r: {"document_ids": r[:1] + ["invalid_id"] + r[1:3]}, lambda r: {"document_ids": r + ["invalid_id"]}, ], ) def test_stop_parse_partial_invalid_document_id(self, get_http_api_auth, add_documents_func, payload): dataset_id, document_ids = add_documents_func parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) if callable(payload): payload = payload(document_ids) res = stop_parse_documnets(get_http_api_auth, dataset_id, payload) assert res["code"] == 102 assert res["message"] == "You don't own the document invalid_id." validate_document_parse_cancel(get_http_api_auth, dataset_id, document_ids) @pytest.mark.p3 def test_repeated_stop_parse(self, get_http_api_auth, add_documents_func): dataset_id, document_ids = add_documents_func parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) res = stop_parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 0 res = stop_parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 102 assert res["message"] == "Can't stop parsing document with progress at 0 or 1" @pytest.mark.p3 def test_duplicate_stop_parse(self, get_http_api_auth, add_documents_func): dataset_id, document_ids = add_documents_func parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) res = stop_parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids + document_ids}) assert res["code"] == 0 assert res["data"]["success_count"] == 3 assert f"Duplicate document ids: {document_ids[0]}" in res["data"]["errors"] @pytest.mark.skip(reason="unstable") def test_stop_parse_100_files(get_http_api_auth, add_dataset_func, tmp_path): document_num = 100 dataset_id = add_dataset_func document_ids = bulk_upload_documents(get_http_api_auth, dataset_id, document_num, tmp_path) parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) sleep(1) res = stop_parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 0 validate_document_parse_cancel(get_http_api_auth, dataset_id, document_ids) @pytest.mark.skip(reason="unstable") def test_concurrent_parse(get_http_api_auth, add_dataset_func, tmp_path): document_num = 50 dataset_id = add_dataset_func document_ids = bulk_upload_documents(get_http_api_auth, dataset_id, document_num, tmp_path) parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( stop_parse_documnets, get_http_api_auth, dataset_id, {"document_ids": document_ids[i : i + 1]}, ) for i in range(document_num) ] responses = [f.result() for f in futures] assert all(r["code"] == 0 for r in responses) validate_document_parse_cancel(get_http_api_auth, dataset_id, document_ids)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/test_http_api/test_file_management_within_dataset/test_download_document.py
sdk/python/test/test_http_api/test_file_management_within_dataset/test_download_document.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json from concurrent.futures import ThreadPoolExecutor import pytest from common import INVALID_API_TOKEN, bulk_upload_documents, download_document, upload_documnets from libs.auth import RAGFlowHttpApiAuth from libs.utils import compare_by_hash from requests import codes @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, tmp_path, auth, expected_code, expected_message): res = download_document(auth, "dataset_id", "document_id", tmp_path / "ragflow_tes.txt") assert res.status_code == codes.ok with (tmp_path / "ragflow_tes.txt").open("r") as f: response_json = json.load(f) assert response_json["code"] == expected_code assert response_json["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "generate_test_files", [ "docx", "excel", "ppt", "image", "pdf", "txt", "md", "json", "eml", "html", ], indirect=True, ) def test_file_type_validation(get_http_api_auth, add_dataset, generate_test_files, request): dataset_id = add_dataset fp = generate_test_files[request.node.callspec.params["generate_test_files"]] res = upload_documnets(get_http_api_auth, dataset_id, [fp]) document_id = res["data"][0]["id"] res = download_document( get_http_api_auth, dataset_id, document_id, fp.with_stem("ragflow_test_download"), ) assert res.status_code == codes.ok assert compare_by_hash( fp, fp.with_stem("ragflow_test_download"), ) class TestDocumentDownload: @pytest.mark.p3 @pytest.mark.parametrize( "document_id, expected_code, expected_message", [ ( "invalid_document_id", 102, "The dataset not own the document invalid_document_id.", ), ], ) def test_invalid_document_id(self, get_http_api_auth, add_documents, tmp_path, document_id, expected_code, expected_message): dataset_id, _ = add_documents res = download_document( get_http_api_auth, dataset_id, document_id, tmp_path / "ragflow_test_download_1.txt", ) assert res.status_code == codes.ok with (tmp_path / "ragflow_test_download_1.txt").open("r") as f: response_json = json.load(f) assert response_json["code"] == expected_code assert response_json["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<NotFound '404: Not Found'>"), ( "invalid_dataset_id", 102, "You do not own the dataset invalid_dataset_id.", ), ], ) def test_invalid_dataset_id(self, get_http_api_auth, add_documents, tmp_path, dataset_id, expected_code, expected_message): _, document_ids = add_documents res = download_document( get_http_api_auth, dataset_id, document_ids[0], tmp_path / "ragflow_test_download_1.txt", ) assert res.status_code == codes.ok with (tmp_path / "ragflow_test_download_1.txt").open("r") as f: response_json = json.load(f) assert response_json["code"] == expected_code assert response_json["message"] == expected_message @pytest.mark.p3 def test_same_file_repeat(self, get_http_api_auth, add_documents, tmp_path, ragflow_tmp_dir): num = 5 dataset_id, document_ids = add_documents for i in range(num): res = download_document( get_http_api_auth, dataset_id, document_ids[0], tmp_path / f"ragflow_test_download_{i}.txt", ) assert res.status_code == codes.ok assert compare_by_hash( ragflow_tmp_dir / "ragflow_test_upload_0.txt", tmp_path / f"ragflow_test_download_{i}.txt", ) @pytest.mark.p3 def test_concurrent_download(get_http_api_auth, add_dataset, tmp_path): document_count = 20 dataset_id = add_dataset document_ids = bulk_upload_documents(get_http_api_auth, dataset_id, document_count, tmp_path) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( download_document, get_http_api_auth, dataset_id, document_ids[i], tmp_path / f"ragflow_test_download_{i}.txt", ) for i in range(document_count) ] responses = [f.result() for f in futures] assert all(r.status_code == codes.ok for r in responses) for i in range(document_count): assert compare_by_hash( tmp_path / f"ragflow_test_upload_{i}.txt", tmp_path / f"ragflow_test_download_{i}.txt", )
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/test_http_api/test_file_management_within_dataset/test_upload_documents.py
sdk/python/test/test_http_api/test_file_management_within_dataset/test_upload_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import string from concurrent.futures import ThreadPoolExecutor import pytest import requests from common import DOCUMENT_NAME_LIMIT, FILE_API_URL, HOST_ADDRESS, INVALID_API_TOKEN, list_datasets, upload_documnets from libs.auth import RAGFlowHttpApiAuth from libs.utils.file_utils import create_txt_file from requests_toolbelt import MultipartEncoder @pytest.mark.p1 @pytest.mark.usefixtures("clear_datasets") class TestAuthorization: @pytest.mark.parametrize( "auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, auth, expected_code, expected_message): res = upload_documnets(auth, "dataset_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsUpload: @pytest.mark.p1 def test_valid_single_upload(self, get_http_api_auth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") res = upload_documnets(get_http_api_auth, dataset_id, [fp]) assert res["code"] == 0 assert res["data"][0]["dataset_id"] == dataset_id assert res["data"][0]["name"] == fp.name @pytest.mark.p1 @pytest.mark.parametrize( "generate_test_files", [ "docx", "excel", "ppt", "image", "pdf", "txt", "md", "json", "eml", "html", ], indirect=True, ) def test_file_type_validation(self, get_http_api_auth, add_dataset_func, generate_test_files, request): dataset_id = add_dataset_func fp = generate_test_files[request.node.callspec.params["generate_test_files"]] res = upload_documnets(get_http_api_auth, dataset_id, [fp]) assert res["code"] == 0 assert res["data"][0]["dataset_id"] == dataset_id assert res["data"][0]["name"] == fp.name @pytest.mark.p2 @pytest.mark.parametrize( "file_type", ["exe", "unknown"], ) def test_unsupported_file_type(self, get_http_api_auth, add_dataset_func, tmp_path, file_type): dataset_id = add_dataset_func fp = tmp_path / f"ragflow_test.{file_type}" fp.touch() res = upload_documnets(get_http_api_auth, dataset_id, [fp]) assert res["code"] == 500 assert res["message"] == f"ragflow_test.{file_type}: This type of file has not been supported yet!" @pytest.mark.p2 def test_missing_file(self, get_http_api_auth, add_dataset_func): dataset_id = add_dataset_func res = upload_documnets(get_http_api_auth, dataset_id) assert res["code"] == 101 assert res["message"] == "No file part!" @pytest.mark.p3 def test_empty_file(self, get_http_api_auth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = tmp_path / "empty.txt" fp.touch() res = upload_documnets(get_http_api_auth, dataset_id, [fp]) assert res["code"] == 0 assert res["data"][0]["size"] == 0 @pytest.mark.p3 def test_filename_empty(self, get_http_api_auth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") url = f"{HOST_ADDRESS}{FILE_API_URL}".format(dataset_id=dataset_id) fields = (("file", ("", fp.open("rb"))),) m = MultipartEncoder(fields=fields) res = requests.post( url=url, headers={"Content-Type": m.content_type}, auth=get_http_api_auth, data=m, ) assert res.json()["code"] == 101 assert res.json()["message"] == "No file selected!" @pytest.mark.p2 def test_filename_exceeds_max_length(self, get_http_api_auth, add_dataset_func, tmp_path): dataset_id = add_dataset_func # filename_length = 129 fp = create_txt_file(tmp_path / f"{'a' * (DOCUMENT_NAME_LIMIT - 3)}.txt") res = upload_documnets(get_http_api_auth, dataset_id, [fp]) assert res["code"] == 101 assert res["message"] == "File name should be less than 128 bytes." @pytest.mark.p2 def test_invalid_dataset_id(self, get_http_api_auth, tmp_path): fp = create_txt_file(tmp_path / "ragflow_test.txt") res = upload_documnets(get_http_api_auth, "invalid_dataset_id", [fp]) assert res["code"] == 100 assert res["message"] == """LookupError("Can\'t find the dataset with ID invalid_dataset_id!")""" @pytest.mark.p2 def test_duplicate_files(self, get_http_api_auth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") res = upload_documnets(get_http_api_auth, dataset_id, [fp, fp]) assert res["code"] == 0 assert len(res["data"]) == 2 for i in range(len(res["data"])): assert res["data"][i]["dataset_id"] == dataset_id expected_name = fp.name if i != 0: expected_name = f"{fp.stem}({i}){fp.suffix}" assert res["data"][i]["name"] == expected_name @pytest.mark.p2 def test_same_file_repeat(self, get_http_api_auth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") for i in range(10): res = upload_documnets(get_http_api_auth, dataset_id, [fp]) assert res["code"] == 0 assert len(res["data"]) == 1 assert res["data"][0]["dataset_id"] == dataset_id expected_name = fp.name if i != 0: expected_name = f"{fp.stem}({i}){fp.suffix}" assert res["data"][0]["name"] == expected_name @pytest.mark.p3 def test_filename_special_characters(self, get_http_api_auth, add_dataset_func, tmp_path): dataset_id = add_dataset_func illegal_chars = '<>:"/\\|?*' translation_table = str.maketrans({char: "_" for char in illegal_chars}) safe_filename = string.punctuation.translate(translation_table) fp = tmp_path / f"{safe_filename}.txt" fp.write_text("Sample text content") res = upload_documnets(get_http_api_auth, dataset_id, [fp]) assert res["code"] == 0 assert len(res["data"]) == 1 assert res["data"][0]["dataset_id"] == dataset_id assert res["data"][0]["name"] == fp.name @pytest.mark.p1 def test_multiple_files(self, get_http_api_auth, add_dataset_func, tmp_path): dataset_id = add_dataset_func expected_document_count = 20 fps = [] for i in range(expected_document_count): fp = create_txt_file(tmp_path / f"ragflow_test_{i}.txt") fps.append(fp) res = upload_documnets(get_http_api_auth, dataset_id, fps) assert res["code"] == 0 res = list_datasets(get_http_api_auth, {"id": dataset_id}) assert res["data"][0]["document_count"] == expected_document_count @pytest.mark.p3 def test_concurrent_upload(self, get_http_api_auth, add_dataset_func, tmp_path): dataset_id = add_dataset_func expected_document_count = 20 fps = [] for i in range(expected_document_count): fp = create_txt_file(tmp_path / f"ragflow_test_{i}.txt") fps.append(fp) with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(upload_documnets, get_http_api_auth, dataset_id, fps[i : i + 1]) for i in range(expected_document_count)] responses = [f.result() for f in futures] assert all(r["code"] == 0 for r in responses) res = list_datasets(get_http_api_auth, {"id": dataset_id}) assert res["data"][0]["document_count"] == expected_document_count
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/test_http_api/test_file_management_within_dataset/test_list_documents.py
sdk/python/test/test_http_api/test_file_management_within_dataset/test_list_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor import pytest from common import INVALID_API_TOKEN, list_documnets from libs.auth import RAGFlowHttpApiAuth from libs.utils import is_sorted @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, auth, expected_code, expected_message): res = list_documnets(auth, "dataset_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsList: @pytest.mark.p1 def test_default(self, get_http_api_auth, add_documents): dataset_id, _ = add_documents res = list_documnets(get_http_api_auth, dataset_id) assert res["code"] == 0 assert len(res["data"]["docs"]) == 5 assert res["data"]["total"] == 5 @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id. ", ), ], ) def test_invalid_dataset_id(self, get_http_api_auth, dataset_id, expected_code, expected_message): res = list_documnets(get_http_api_auth, dataset_id) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page": None, "page_size": 2}, 0, 2, ""), ({"page": 0, "page_size": 2}, 0, 2, ""), ({"page": 2, "page_size": 2}, 0, 2, ""), ({"page": 3, "page_size": 2}, 0, 1, ""), ({"page": "3", "page_size": 2}, 0, 1, ""), pytest.param( {"page": -1, "page_size": 2}, 100, 0, "1064", marks=pytest.mark.skip(reason="issues/5851"), ), pytest.param( {"page": "a", "page_size": 2}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip(reason="issues/5851"), ), ], ) def test_page( self, get_http_api_auth, add_documents, params, expected_code, expected_page_size, expected_message, ): dataset_id, _ = add_documents res = list_documnets(get_http_api_auth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["docs"]) == expected_page_size assert res["data"]["total"] == 5 else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page_size": None}, 0, 5, ""), ({"page_size": 0}, 0, 0, ""), ({"page_size": 1}, 0, 1, ""), ({"page_size": 6}, 0, 5, ""), ({"page_size": "1"}, 0, 1, ""), pytest.param( {"page_size": -1}, 100, 0, "1064", marks=pytest.mark.skip(reason="issues/5851"), ), pytest.param( {"page_size": "a"}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip(reason="issues/5851"), ), ], ) def test_page_size( self, get_http_api_auth, add_documents, params, expected_code, expected_page_size, expected_message, ): dataset_id, _ = add_documents res = list_documnets(get_http_api_auth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["docs"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "params, expected_code, assertions, expected_message", [ ({"orderby": None}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"orderby": "create_time"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"orderby": "update_time"}, 0, lambda r: (is_sorted(r["data"]["docs"], "update_time", True)), ""), pytest.param({"orderby": "name", "desc": "False"}, 0, lambda r: (is_sorted(r["data"]["docs"], "name", False)), "", marks=pytest.mark.skip(reason="issues/5851")), pytest.param({"orderby": "unknown"}, 102, 0, "orderby should be create_time or update_time", marks=pytest.mark.skip(reason="issues/5851")), ], ) def test_orderby( self, get_http_api_auth, add_documents, params, expected_code, assertions, expected_message, ): dataset_id, _ = add_documents res = list_documnets(get_http_api_auth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: if callable(assertions): assert assertions(res) else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "params, expected_code, assertions, expected_message", [ ({"desc": None}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"desc": "true"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"desc": "True"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"desc": True}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), pytest.param({"desc": "false"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", False)), "", marks=pytest.mark.skip(reason="issues/5851")), ({"desc": "False"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", False)), ""), ({"desc": False}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", False)), ""), ({"desc": "False", "orderby": "update_time"}, 0, lambda r: (is_sorted(r["data"]["docs"], "update_time", False)), ""), pytest.param({"desc": "unknown"}, 102, 0, "desc should be true or false", marks=pytest.mark.skip(reason="issues/5851")), ], ) def test_desc( self, get_http_api_auth, add_documents, params, expected_code, assertions, expected_message, ): dataset_id, _ = add_documents res = list_documnets(get_http_api_auth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: if callable(assertions): assert assertions(res) else: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "params, expected_num", [ ({"keywords": None}, 5), ({"keywords": ""}, 5), ({"keywords": "0"}, 1), ({"keywords": "ragflow_test_upload"}, 5), ({"keywords": "unknown"}, 0), ], ) def test_keywords(self, get_http_api_auth, add_documents, params, expected_num): dataset_id, _ = add_documents res = list_documnets(get_http_api_auth, dataset_id, params=params) assert res["code"] == 0 assert len(res["data"]["docs"]) == expected_num assert res["data"]["total"] == expected_num @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_num, expected_message", [ ({"name": None}, 0, 5, ""), ({"name": ""}, 0, 5, ""), ({"name": "ragflow_test_upload_0.txt"}, 0, 1, ""), ( {"name": "unknown.txt"}, 102, 0, "You don't own the document unknown.txt.", ), ], ) def test_name( self, get_http_api_auth, add_documents, params, expected_code, expected_num, expected_message, ): dataset_id, _ = add_documents res = list_documnets(get_http_api_auth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: if params["name"] in [None, ""]: assert len(res["data"]["docs"]) == expected_num else: assert res["data"]["docs"][0]["name"] == params["name"] else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "document_id, expected_code, expected_num, expected_message", [ (None, 0, 5, ""), ("", 0, 5, ""), (lambda r: r[0], 0, 1, ""), ("unknown.txt", 102, 0, "You don't own the document unknown.txt."), ], ) def test_id( self, get_http_api_auth, add_documents, document_id, expected_code, expected_num, expected_message, ): dataset_id, document_ids = add_documents if callable(document_id): params = {"id": document_id(document_ids)} else: params = {"id": document_id} res = list_documnets(get_http_api_auth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: if params["id"] in [None, ""]: assert len(res["data"]["docs"]) == expected_num else: assert res["data"]["docs"][0]["id"] == params["id"] else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "document_id, name, expected_code, expected_num, expected_message", [ (lambda r: r[0], "ragflow_test_upload_0.txt", 0, 1, ""), (lambda r: r[0], "ragflow_test_upload_1.txt", 0, 0, ""), (lambda r: r[0], "unknown", 102, 0, "You don't own the document unknown."), ( "id", "ragflow_test_upload_0.txt", 102, 0, "You don't own the document id.", ), ], ) def test_name_and_id( self, get_http_api_auth, add_documents, document_id, name, expected_code, expected_num, expected_message, ): dataset_id, document_ids = add_documents if callable(document_id): params = {"id": document_id(document_ids), "name": name} else: params = {"id": document_id, "name": name} res = list_documnets(get_http_api_auth, dataset_id, params=params) if expected_code == 0: assert len(res["data"]["docs"]) == expected_num else: assert res["message"] == expected_message @pytest.mark.p3 def test_concurrent_list(self, get_http_api_auth, add_documents): dataset_id, _ = add_documents with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(list_documnets, get_http_api_auth, dataset_id) for i in range(100)] responses = [f.result() for f in futures] assert all(r["code"] == 0 for r in responses) @pytest.mark.p3 def test_invalid_params(self, get_http_api_auth, add_documents): dataset_id, _ = add_documents params = {"a": "b"} res = list_documnets(get_http_api_auth, dataset_id, params=params) assert res["code"] == 0 assert len(res["data"]["docs"]) == 5
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/test_http_api/test_file_management_within_dataset/conftest.py
sdk/python/test/test_http_api/test_file_management_within_dataset/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import bulk_upload_documents, delete_documnets @pytest.fixture(scope="function") def add_document_func(request, get_http_api_auth, add_dataset, ragflow_tmp_dir): dataset_id = add_dataset document_ids = bulk_upload_documents(get_http_api_auth, dataset_id, 1, ragflow_tmp_dir) def cleanup(): delete_documnets(get_http_api_auth, dataset_id, {"ids": document_ids}) request.addfinalizer(cleanup) return dataset_id, document_ids[0] @pytest.fixture(scope="class") def add_documents(request, get_http_api_auth, add_dataset, ragflow_tmp_dir): dataset_id = add_dataset document_ids = bulk_upload_documents(get_http_api_auth, dataset_id, 5, ragflow_tmp_dir) def cleanup(): delete_documnets(get_http_api_auth, dataset_id, {"ids": document_ids}) request.addfinalizer(cleanup) return dataset_id, document_ids @pytest.fixture(scope="function") def add_documents_func(get_http_api_auth, add_dataset_func, ragflow_tmp_dir): dataset_id = add_dataset_func document_ids = bulk_upload_documents(get_http_api_auth, dataset_id, 3, ragflow_tmp_dir) return dataset_id, document_ids
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/test_http_api/test_file_management_within_dataset/test_delete_documents.py
sdk/python/test/test_http_api/test_file_management_within_dataset/test_delete_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor import pytest from common import INVALID_API_TOKEN, bulk_upload_documents, delete_documnets, list_documnets from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, auth, expected_code, expected_message): res = delete_documnets(auth, "dataset_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsDeletion: @pytest.mark.p1 @pytest.mark.parametrize( "payload, expected_code, expected_message, remaining", [ (None, 0, "", 0), ({"ids": []}, 0, "", 0), ({"ids": ["invalid_id"]}, 102, "Documents not found: ['invalid_id']", 3), ( {"ids": ["\n!?。;!?\"'"]}, 102, """Documents not found: [\'\\n!?。;!?"\\\'\']""", 3, ), ( "not json", 100, "AttributeError(\"'str' object has no attribute 'get'\")", 3, ), (lambda r: {"ids": r[:1]}, 0, "", 2), (lambda r: {"ids": r}, 0, "", 0), ], ) def test_basic_scenarios( self, get_http_api_auth, add_documents_func, payload, expected_code, expected_message, remaining, ): dataset_id, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = delete_documnets(get_http_api_auth, dataset_id, payload) assert res["code"] == expected_code if res["code"] != 0: assert res["message"] == expected_message res = list_documnets(get_http_api_auth, dataset_id) assert len(res["data"]["docs"]) == remaining assert res["data"]["total"] == remaining @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id. ", ), ], ) def test_invalid_dataset_id(self, get_http_api_auth, add_documents_func, dataset_id, expected_code, expected_message): _, document_ids = add_documents_func res = delete_documnets(get_http_api_auth, dataset_id, {"ids": document_ids[:1]}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "payload", [ lambda r: {"ids": ["invalid_id"] + r}, lambda r: {"ids": r[:1] + ["invalid_id"] + r[1:3]}, lambda r: {"ids": r + ["invalid_id"]}, ], ) def test_delete_partial_invalid_id(self, get_http_api_auth, add_documents_func, payload): dataset_id, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = delete_documnets(get_http_api_auth, dataset_id, payload) assert res["code"] == 102 assert res["message"] == "Documents not found: ['invalid_id']" res = list_documnets(get_http_api_auth, dataset_id) assert len(res["data"]["docs"]) == 0 assert res["data"]["total"] == 0 @pytest.mark.p2 def test_repeated_deletion(self, get_http_api_auth, add_documents_func): dataset_id, document_ids = add_documents_func res = delete_documnets(get_http_api_auth, dataset_id, {"ids": document_ids}) assert res["code"] == 0 res = delete_documnets(get_http_api_auth, dataset_id, {"ids": document_ids}) assert res["code"] == 102 assert "Documents not found" in res["message"] @pytest.mark.p2 def test_duplicate_deletion(self, get_http_api_auth, add_documents_func): dataset_id, document_ids = add_documents_func res = delete_documnets(get_http_api_auth, dataset_id, {"ids": document_ids + document_ids}) assert res["code"] == 0 assert "Duplicate document ids" in res["data"]["errors"][0] assert res["data"]["success_count"] == 3 res = list_documnets(get_http_api_auth, dataset_id) assert len(res["data"]["docs"]) == 0 assert res["data"]["total"] == 0 @pytest.mark.p3 def test_concurrent_deletion(get_http_api_auth, add_dataset, tmp_path): documnets_num = 100 dataset_id = add_dataset document_ids = bulk_upload_documents(get_http_api_auth, dataset_id, documnets_num, tmp_path) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( delete_documnets, get_http_api_auth, dataset_id, {"ids": document_ids[i : i + 1]}, ) for i in range(documnets_num) ] responses = [f.result() for f in futures] assert all(r["code"] == 0 for r in responses) @pytest.mark.p3 def test_delete_1k(get_http_api_auth, add_dataset, tmp_path): documnets_num = 1_000 dataset_id = add_dataset document_ids = bulk_upload_documents(get_http_api_auth, dataset_id, documnets_num, tmp_path) res = list_documnets(get_http_api_auth, dataset_id) assert res["data"]["total"] == documnets_num res = delete_documnets(get_http_api_auth, dataset_id, {"ids": document_ids}) assert res["code"] == 0 res = list_documnets(get_http_api_auth, dataset_id) assert res["data"]["total"] == 0
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/test_http_api/test_file_management_within_dataset/test_update_document.py
sdk/python/test/test_http_api/test_file_management_within_dataset/test_update_document.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import DOCUMENT_NAME_LIMIT, INVALID_API_TOKEN, list_documnets, update_documnet from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, auth, expected_code, expected_message): res = update_documnet(auth, "dataset_id", "document_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsUpdated: @pytest.mark.p1 @pytest.mark.parametrize( "name, expected_code, expected_message", [ ("new_name.txt", 0, ""), ( f"{'a' * (DOCUMENT_NAME_LIMIT - 3)}.txt", 101, "The name should be less than 128 bytes.", ), ( 0, 100, """AttributeError("\'int\' object has no attribute \'encode\'")""", ), ( None, 100, """AttributeError("\'NoneType\' object has no attribute \'encode\'")""", ), ( "", 101, "The extension of file can't be changed", ), ( "ragflow_test_upload_0", 101, "The extension of file can't be changed", ), ( "ragflow_test_upload_1.txt", 102, "Duplicated document name in the same dataset.", ), ( "RAGFLOW_TEST_UPLOAD_1.TXT", 0, "", ), ], ) def test_name(self, get_http_api_auth, add_documents, name, expected_code, expected_message): dataset_id, document_ids = add_documents res = update_documnet(get_http_api_auth, dataset_id, document_ids[0], {"name": name}) assert res["code"] == expected_code if expected_code == 0: res = list_documnets(get_http_api_auth, dataset_id, {"id": document_ids[0]}) assert res["data"]["docs"][0]["name"] == name else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "document_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_document_id", 102, "The dataset doesn't own the document.", ), ], ) def test_invalid_document_id(self, get_http_api_auth, add_documents, document_id, expected_code, expected_message): dataset_id, _ = add_documents res = update_documnet(get_http_api_auth, dataset_id, document_id, {"name": "new_name.txt"}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<NotFound '404: Not Found'>"), ( "invalid_dataset_id", 102, "You don't own the dataset.", ), ], ) def test_invalid_dataset_id(self, get_http_api_auth, add_documents, dataset_id, expected_code, expected_message): _, document_ids = add_documents res = update_documnet(get_http_api_auth, dataset_id, document_ids[0], {"name": "new_name.txt"}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "meta_fields, expected_code, expected_message", [({"test": "test"}, 0, ""), ("test", 102, "meta_fields must be a dictionary")], ) def test_meta_fields(self, get_http_api_auth, add_documents, meta_fields, expected_code, expected_message): dataset_id, document_ids = add_documents res = update_documnet(get_http_api_auth, dataset_id, document_ids[0], {"meta_fields": meta_fields}) if expected_code == 0: res = list_documnets(get_http_api_auth, dataset_id, {"id": document_ids[0]}) assert res["data"]["docs"][0]["meta_fields"] == meta_fields else: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "chunk_method, expected_code, expected_message", [ ("naive", 0, ""), ("manual", 0, ""), ("qa", 0, ""), ("table", 0, ""), ("paper", 0, ""), ("book", 0, ""), ("laws", 0, ""), ("presentation", 0, ""), ("picture", 0, ""), ("one", 0, ""), ("knowledge_graph", 0, ""), ("email", 0, ""), ("tag", 0, ""), ("", 102, "`chunk_method` doesn't exist"), ( "other_chunk_method", 102, "`chunk_method` other_chunk_method doesn't exist", ), ], ) def test_chunk_method(self, get_http_api_auth, add_documents, chunk_method, expected_code, expected_message): dataset_id, document_ids = add_documents res = update_documnet(get_http_api_auth, dataset_id, document_ids[0], {"chunk_method": chunk_method}) assert res["code"] == expected_code if expected_code == 0: res = list_documnets(get_http_api_auth, dataset_id, {"id": document_ids[0]}) if chunk_method == "": assert res["data"]["docs"][0]["chunk_method"] == "naive" else: assert res["data"]["docs"][0]["chunk_method"] == chunk_method else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"chunk_count": 1}, 102, "Can't change `chunk_count`."), pytest.param( {"create_date": "Fri, 14 Mar 2025 16:53:42 GMT"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"create_time": 1}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"created_by": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"dataset_id": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"id": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"location": "ragflow_test.txt"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"process_begin_at": 1}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"process_duration": 1.0}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param({"progress": 1.0}, 102, "Can't change `progress`."), pytest.param( {"progress_msg": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"run": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"size": 1}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"source_type": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"thumbnail": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), ({"token_count": 1}, 102, "Can't change `token_count`."), pytest.param( {"type": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"update_date": "Fri, 14 Mar 2025 16:33:17 GMT"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"update_time": 1}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), ], ) def test_invalid_field( self, get_http_api_auth, add_documents, payload, expected_code, expected_message, ): dataset_id, document_ids = add_documents res = update_documnet(get_http_api_auth, dataset_id, document_ids[0], payload) assert res["code"] == expected_code assert res["message"] == expected_message class TestUpdateDocumentParserConfig: @pytest.mark.p2 @pytest.mark.parametrize( "chunk_method, parser_config, expected_code, expected_message", [ ("naive", {}, 0, ""), ( "naive", { "chunk_token_num": 128, "layout_recognize": "DeepDOC", "html4excel": False, "delimiter": r"\n", "task_page_size": 12, "raptor": {"use_raptor": False}, }, 0, "", ), pytest.param( "naive", {"chunk_token_num": -1}, 100, "AssertionError('chunk_token_num should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"chunk_token_num": 0}, 100, "AssertionError('chunk_token_num should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"chunk_token_num": 100000000}, 100, "AssertionError('chunk_token_num should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"chunk_token_num": 3.14}, 102, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"chunk_token_num": "1024"}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), ( "naive", {"layout_recognize": "DeepDOC"}, 0, "", ), ( "naive", {"layout_recognize": "Naive"}, 0, "", ), ("naive", {"html4excel": True}, 0, ""), ("naive", {"html4excel": False}, 0, ""), pytest.param( "naive", {"html4excel": 1}, 100, "AssertionError('html4excel should be True or False')", marks=pytest.mark.skip(reason="issues/6098"), ), ("naive", {"delimiter": ""}, 0, ""), ("naive", {"delimiter": "`##`"}, 0, ""), pytest.param( "naive", {"delimiter": 1}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"task_page_size": -1}, 100, "AssertionError('task_page_size should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"task_page_size": 0}, 100, "AssertionError('task_page_size should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"task_page_size": 100000000}, 100, "AssertionError('task_page_size should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"task_page_size": 3.14}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"task_page_size": "1024"}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), ("naive", {"raptor": {"use_raptor": True}}, 0, ""), ("naive", {"raptor": {"use_raptor": False}}, 0, ""), pytest.param( "naive", {"invalid_key": "invalid_value"}, 100, """AssertionError("Abnormal \'parser_config\'. Invalid key: invalid_key")""", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_keywords": -1}, 100, "AssertionError('auto_keywords should be in range from 0 to 32')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_keywords": 32}, 100, "AssertionError('auto_keywords should be in range from 0 to 32')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_questions": 3.14}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_keywords": "1024"}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_questions": -1}, 100, "AssertionError('auto_questions should be in range from 0 to 10')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_questions": 10}, 100, "AssertionError('auto_questions should be in range from 0 to 10')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_questions": 3.14}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_questions": "1024"}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"topn_tags": -1}, 100, "AssertionError('topn_tags should be in range from 0 to 10')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"topn_tags": 10}, 100, "AssertionError('topn_tags should be in range from 0 to 10')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"topn_tags": 3.14}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"topn_tags": "1024"}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), ], ) def test_parser_config( self, get_http_api_auth, add_documents, chunk_method, parser_config, expected_code, expected_message, ): dataset_id, document_ids = add_documents res = update_documnet( get_http_api_auth, dataset_id, document_ids[0], {"chunk_method": chunk_method, "parser_config": parser_config}, ) assert res["code"] == expected_code if expected_code == 0: res = list_documnets(get_http_api_auth, dataset_id, {"id": document_ids[0]}) if parser_config == {}: assert res["data"]["docs"][0]["parser_config"] == { "chunk_token_num": 128, "delimiter": r"\n", "html4excel": False, "layout_recognize": "DeepDOC", "raptor": {"use_raptor": False}, } else: for k, v in parser_config.items(): assert res["data"]["docs"][0]["parser_config"][k] == v if expected_code != 0 or expected_message: assert res["message"] == expected_message
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/test_http_api/test_file_management_within_dataset/test_parse_documents.py
sdk/python/test/test_http_api/test_file_management_within_dataset/test_parse_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor import pytest from common import INVALID_API_TOKEN, bulk_upload_documents, list_documnets, parse_documnets from libs.auth import RAGFlowHttpApiAuth from libs.utils import wait_for @wait_for(30, 1, "Document parsing timeout") def condition(_auth, _dataset_id, _document_ids=None): res = list_documnets(_auth, _dataset_id) target_docs = res["data"]["docs"] if _document_ids is None: for doc in target_docs: if doc["run"] != "DONE": return False return True target_ids = set(_document_ids) for doc in target_docs: if doc["id"] in target_ids: if doc.get("run") != "DONE": return False return True def validate_document_details(auth, dataset_id, document_ids): for document_id in document_ids: res = list_documnets(auth, dataset_id, params={"id": document_id}) doc = res["data"]["docs"][0] assert doc["run"] == "DONE" assert len(doc["process_begin_at"]) > 0 assert doc["process_duration"] > 0 assert doc["progress"] > 0 assert "Task done" in doc["progress_msg"] @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, auth, expected_code, expected_message): res = parse_documnets(auth, "dataset_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsParse: @pytest.mark.parametrize( "payload, expected_code, expected_message", [ pytest.param(None, 102, """AttributeError("\'NoneType\' object has no attribute \'get\'")""", marks=pytest.mark.skip), pytest.param({"document_ids": []}, 102, "`document_ids` is required", marks=pytest.mark.p1), pytest.param({"document_ids": ["invalid_id"]}, 102, "Documents not found: ['invalid_id']", marks=pytest.mark.p3), pytest.param({"document_ids": ["\n!?。;!?\"'"]}, 102, """Documents not found: [\'\\n!?。;!?"\\\'\']""", marks=pytest.mark.p3), pytest.param("not json", 102, "AttributeError(\"'str' object has no attribute 'get'\")", marks=pytest.mark.skip), pytest.param(lambda r: {"document_ids": r[:1]}, 0, "", marks=pytest.mark.p1), pytest.param(lambda r: {"document_ids": r}, 0, "", marks=pytest.mark.p1), ], ) def test_basic_scenarios(self, get_http_api_auth, add_documents_func, payload, expected_code, expected_message): dataset_id, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = parse_documnets(get_http_api_auth, dataset_id, payload) assert res["code"] == expected_code if expected_code != 0: assert res["message"] == expected_message if expected_code == 0: condition(get_http_api_auth, dataset_id, payload["document_ids"]) validate_document_details(get_http_api_auth, dataset_id, payload["document_ids"]) @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id.", ), ], ) def test_invalid_dataset_id( self, get_http_api_auth, add_documents_func, dataset_id, expected_code, expected_message, ): _, document_ids = add_documents_func res = parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.parametrize( "payload", [ pytest.param(lambda r: {"document_ids": ["invalid_id"] + r}, marks=pytest.mark.p3), pytest.param(lambda r: {"document_ids": r[:1] + ["invalid_id"] + r[1:3]}, marks=pytest.mark.p1), pytest.param(lambda r: {"document_ids": r + ["invalid_id"]}, marks=pytest.mark.p3), ], ) def test_parse_partial_invalid_document_id(self, get_http_api_auth, add_documents_func, payload): dataset_id, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = parse_documnets(get_http_api_auth, dataset_id, payload) assert res["code"] == 102 assert res["message"] == "Documents not found: ['invalid_id']" condition(get_http_api_auth, dataset_id) validate_document_details(get_http_api_auth, dataset_id, document_ids) @pytest.mark.p3 def test_repeated_parse(self, get_http_api_auth, add_documents_func): dataset_id, document_ids = add_documents_func res = parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 0 condition(get_http_api_auth, dataset_id) res = parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 0 @pytest.mark.p3 def test_duplicate_parse(self, get_http_api_auth, add_documents_func): dataset_id, document_ids = add_documents_func res = parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids + document_ids}) assert res["code"] == 0 assert "Duplicate document ids" in res["data"]["errors"][0] assert res["data"]["success_count"] == 3 condition(get_http_api_auth, dataset_id) validate_document_details(get_http_api_auth, dataset_id, document_ids) @pytest.mark.p3 def test_parse_100_files(get_http_api_auth, add_dataset_func, tmp_path): @wait_for(100, 1, "Document parsing timeout") def condition(_auth, _dataset_id, _document_num): res = list_documnets(_auth, _dataset_id, {"page_size": _document_num}) for doc in res["data"]["docs"]: if doc["run"] != "DONE": return False return True document_num = 100 dataset_id = add_dataset_func document_ids = bulk_upload_documents(get_http_api_auth, dataset_id, document_num, tmp_path) res = parse_documnets(get_http_api_auth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 0 condition(get_http_api_auth, dataset_id, document_num) validate_document_details(get_http_api_auth, dataset_id, document_ids) @pytest.mark.p3 def test_concurrent_parse(get_http_api_auth, add_dataset_func, tmp_path): @wait_for(120, 1, "Document parsing timeout") def condition(_auth, _dataset_id, _document_num): res = list_documnets(_auth, _dataset_id, {"page_size": _document_num}) for doc in res["data"]["docs"]: if doc["run"] != "DONE": return False return True document_num = 100 dataset_id = add_dataset_func document_ids = bulk_upload_documents(get_http_api_auth, dataset_id, document_num, tmp_path) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( parse_documnets, get_http_api_auth, dataset_id, {"document_ids": document_ids[i : i + 1]}, ) for i in range(document_num) ] responses = [f.result() for f in futures] assert all(r["code"] == 0 for r in responses) condition(get_http_api_auth, dataset_id, document_num) validate_document_details(get_http_api_auth, dataset_id, document_ids)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/libs/__init__.py
sdk/python/test/libs/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/libs/auth.py
sdk/python/test/libs/auth.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from requests.auth import AuthBase class RAGFlowHttpApiAuth(AuthBase): def __init__(self, token): self._token = token def __call__(self, r): r.headers["Authorization"] = f'Bearer {self._token}' return r
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/libs/utils/file_utils.py
sdk/python/test/libs/utils/file_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json from docx import Document # pip install python-docx from openpyxl import Workbook # pip install openpyxl from PIL import Image, ImageDraw # pip install Pillow from pptx import Presentation # pip install python-pptx from reportlab.pdfgen import canvas # pip install reportlab def create_docx_file(path): doc = Document() doc.add_paragraph("这是一个测试 DOCX 文件。") doc.save(path) return path def create_excel_file(path): wb = Workbook() ws = wb.active ws["A1"] = "测试 Excel 文件" wb.save(path) return path def create_ppt_file(path): prs = Presentation() slide = prs.slides.add_slide(prs.slide_layouts[0]) slide.shapes.title.text = "测试 PPT 文件" prs.save(path) return path def create_image_file(path): img = Image.new("RGB", (100, 100), color="blue") draw = ImageDraw.Draw(img) draw.text((10, 40), "Test", fill="white") img.save(path) return path def create_pdf_file(path): if not isinstance(path, str): path = str(path) c = canvas.Canvas(path) c.drawString(100, 750, "测试 PDF 文件") c.save() return path def create_txt_file(path): with open(path, "w", encoding="utf-8") as f: f.write("这是测试 TXT 文件的内容。") return path def create_md_file(path): md_content = "# 测试 MD 文件\n\n这是一份 Markdown 格式的测试文件。" with open(path, "w", encoding="utf-8") as f: f.write(md_content) return path def create_json_file(path): data = {"message": "这是测试 JSON 文件", "value": 123} with open(path, "w", encoding="utf-8") as f: json.dump(data, f, indent=2) return path def create_eml_file(path): eml_content = ( "From: sender@example.com\n" "To: receiver@example.com\n" "Subject: 测试 EML 文件\n\n" "这是一封测试邮件的内容。\n" ) with open(path, "w", encoding="utf-8") as f: f.write(eml_content) return path def create_html_file(path): html_content = ( "<html>\n" "<head><title>测试 HTML 文件</title></head>\n" "<body><h1>这是一个测试 HTML 文件</h1></body>\n" "</html>" ) with open(path, "w", encoding="utf-8") as f: f.write(html_content) return path
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/libs/utils/__init__.py
sdk/python/test/libs/utils/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import base64 import functools import hashlib import time from pathlib import Path def encode_avatar(image_path): with Path.open(image_path, "rb") as file: binary_data = file.read() base64_encoded = base64.b64encode(binary_data).decode("utf-8") return base64_encoded def compare_by_hash(file1, file2, algorithm="sha256"): def _calc_hash(file_path): hash_func = hashlib.new(algorithm) with open(file_path, "rb") as f: while chunk := f.read(8192): hash_func.update(chunk) return hash_func.hexdigest() return _calc_hash(file1) == _calc_hash(file2) def wait_for(timeout=10, interval=1, error_msg="Timeout"): def decorator(func): @functools.wraps(func) def wrapper(*args, **kwargs): start_time = time.time() while True: result = func(*args, **kwargs) if result is True: return result elapsed = time.time() - start_time if elapsed > timeout: assert False, error_msg time.sleep(interval) return wrapper return decorator def is_sorted(data, field, descending=True): timestamps = [ds[field] for ds in data] return all(a >= b for a, b in zip(timestamps, timestamps[1:])) if descending else all(a <= b for a, b in zip(timestamps, timestamps[1:]))
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/sdk/python/test/libs/utils/hypothesis_utils.py
sdk/python/test/libs/utils/hypothesis_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import hypothesis.strategies as st @st.composite def valid_names(draw): base_chars = "abcdefghijklmnopqrstuvwxyz_" first_char = draw(st.sampled_from([c for c in base_chars if c.isalpha() or c == "_"])) remaining = draw(st.text(alphabet=st.sampled_from(base_chars), min_size=0, max_size=128 - 2)) name = (first_char + remaining)[:128] return name.encode("utf-8").decode("utf-8")
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/mcp/server/server.py
mcp/server/server.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json import logging import random import time from collections import OrderedDict from collections.abc import AsyncIterator from contextlib import asynccontextmanager from functools import wraps import click import requests from starlette.applications import Starlette from starlette.middleware import Middleware from starlette.responses import JSONResponse, Response from starlette.routing import Mount, Route from strenum import StrEnum import mcp.types as types from mcp.server.lowlevel import Server class LaunchMode(StrEnum): SELF_HOST = "self-host" HOST = "host" class Transport(StrEnum): SSE = "sse" STEAMABLE_HTTP = "streamable-http" BASE_URL = "http://127.0.0.1:9380" HOST = "127.0.0.1" PORT = "9382" HOST_API_KEY = "" MODE = "" TRANSPORT_SSE_ENABLED = True TRANSPORT_STREAMABLE_HTTP_ENABLED = True JSON_RESPONSE = True class RAGFlowConnector: _MAX_DATASET_CACHE = 32 _CACHE_TTL = 300 _dataset_metadata_cache: OrderedDict[str, tuple[dict, float | int]] = OrderedDict() # "dataset_id" -> (metadata, expiry_ts) _document_metadata_cache: OrderedDict[str, tuple[list[tuple[str, dict]], float | int]] = OrderedDict() # "dataset_id" -> ([(document_id, doc_metadata)], expiry_ts) def __init__(self, base_url: str, version="v1"): self.base_url = base_url self.version = version self.api_url = f"{self.base_url}/api/{self.version}" def bind_api_key(self, api_key: str): self.api_key = api_key self.authorization_header = {"Authorization": "{} {}".format("Bearer", self.api_key)} def _post(self, path, json=None, stream=False, files=None): if not self.api_key: return None res = requests.post(url=self.api_url + path, json=json, headers=self.authorization_header, stream=stream, files=files) return res def _get(self, path, params=None, json=None): res = requests.get(url=self.api_url + path, params=params, headers=self.authorization_header, json=json) return res def _is_cache_valid(self, ts): return time.time() < ts def _get_expiry_timestamp(self): offset = random.randint(-30, 30) return time.time() + self._CACHE_TTL + offset def _get_cached_dataset_metadata(self, dataset_id): entry = self._dataset_metadata_cache.get(dataset_id) if entry: data, ts = entry if self._is_cache_valid(ts): self._dataset_metadata_cache.move_to_end(dataset_id) return data return None def _set_cached_dataset_metadata(self, dataset_id, metadata): self._dataset_metadata_cache[dataset_id] = (metadata, self._get_expiry_timestamp()) self._dataset_metadata_cache.move_to_end(dataset_id) if len(self._dataset_metadata_cache) > self._MAX_DATASET_CACHE: self._dataset_metadata_cache.popitem(last=False) def _get_cached_document_metadata_by_dataset(self, dataset_id): entry = self._document_metadata_cache.get(dataset_id) if entry: data_list, ts = entry if self._is_cache_valid(ts): self._document_metadata_cache.move_to_end(dataset_id) return {doc_id: doc_meta for doc_id, doc_meta in data_list} return None def _set_cached_document_metadata_by_dataset(self, dataset_id, doc_id_meta_list): self._document_metadata_cache[dataset_id] = (doc_id_meta_list, self._get_expiry_timestamp()) self._document_metadata_cache.move_to_end(dataset_id) def list_datasets(self, page: int = 1, page_size: int = 1000, orderby: str = "create_time", desc: bool = True, id: str | None = None, name: str | None = None): res = self._get("/datasets", {"page": page, "page_size": page_size, "orderby": orderby, "desc": desc, "id": id, "name": name}) if not res: raise Exception([types.TextContent(type="text", text=res.get("Cannot process this operation."))]) res = res.json() if res.get("code") == 0: result_list = [] for data in res["data"]: d = {"description": data["description"], "id": data["id"]} result_list.append(json.dumps(d, ensure_ascii=False)) return "\n".join(result_list) return "" def retrieval( self, dataset_ids, document_ids=None, question="", page=1, page_size=30, similarity_threshold=0.2, vector_similarity_weight=0.3, top_k=1024, rerank_id: str | None = None, keyword: bool = False, force_refresh: bool = False, ): if document_ids is None: document_ids = [] # If no dataset_ids provided or empty list, get all available dataset IDs if not dataset_ids: dataset_list_str = self.list_datasets() dataset_ids = [] # Parse the dataset list to extract IDs if dataset_list_str: for line in dataset_list_str.strip().split('\n'): if line.strip(): try: dataset_info = json.loads(line.strip()) dataset_ids.append(dataset_info["id"]) except (json.JSONDecodeError, KeyError): # Skip malformed lines continue data_json = { "page": page, "page_size": page_size, "similarity_threshold": similarity_threshold, "vector_similarity_weight": vector_similarity_weight, "top_k": top_k, "rerank_id": rerank_id, "keyword": keyword, "question": question, "dataset_ids": dataset_ids, "document_ids": document_ids, } # Send a POST request to the backend service (using requests library as an example, actual implementation may vary) res = self._post("/retrieval", json=data_json) if not res: raise Exception([types.TextContent(type="text", text=res.get("Cannot process this operation."))]) res = res.json() if res.get("code") == 0: data = res["data"] chunks = [] # Cache document metadata and dataset information document_cache, dataset_cache = self._get_document_metadata_cache(dataset_ids, force_refresh=force_refresh) # Process chunks with enhanced field mapping including per-chunk metadata for chunk_data in data.get("chunks", []): enhanced_chunk = self._map_chunk_fields(chunk_data, dataset_cache, document_cache) chunks.append(enhanced_chunk) # Build structured response (no longer need response-level document_metadata) response = { "chunks": chunks, "pagination": { "page": data.get("page", page), "page_size": data.get("page_size", page_size), "total_chunks": data.get("total", len(chunks)), "total_pages": (data.get("total", len(chunks)) + page_size - 1) // page_size, }, "query_info": { "question": question, "similarity_threshold": similarity_threshold, "vector_weight": vector_similarity_weight, "keyword_search": keyword, "dataset_count": len(dataset_ids), }, } return [types.TextContent(type="text", text=json.dumps(response, ensure_ascii=False))] raise Exception([types.TextContent(type="text", text=res.get("message"))]) def _get_document_metadata_cache(self, dataset_ids, force_refresh=False): """Cache document metadata for all documents in the specified datasets""" document_cache = {} dataset_cache = {} try: for dataset_id in dataset_ids: dataset_meta = None if force_refresh else self._get_cached_dataset_metadata(dataset_id) if not dataset_meta: # First get dataset info for name dataset_res = self._get("/datasets", {"id": dataset_id, "page_size": 1}) if dataset_res and dataset_res.status_code == 200: dataset_data = dataset_res.json() if dataset_data.get("code") == 0 and dataset_data.get("data"): dataset_info = dataset_data["data"][0] dataset_meta = {"name": dataset_info.get("name", "Unknown"), "description": dataset_info.get("description", "")} self._set_cached_dataset_metadata(dataset_id, dataset_meta) if dataset_meta: dataset_cache[dataset_id] = dataset_meta docs = None if force_refresh else self._get_cached_document_metadata_by_dataset(dataset_id) if docs is None: page = 1 page_size = 30 doc_id_meta_list = [] docs = {} while page: docs_res = self._get(f"/datasets/{dataset_id}/documents?page={page}") docs_data = docs_res.json() if docs_data.get("code") == 0 and docs_data.get("data", {}).get("docs"): for doc in docs_data["data"]["docs"]: doc_id = doc.get("id") if not doc_id: continue doc_meta = { "document_id": doc_id, "name": doc.get("name", ""), "location": doc.get("location", ""), "type": doc.get("type", ""), "size": doc.get("size"), "chunk_count": doc.get("chunk_count"), "create_date": doc.get("create_date", ""), "update_date": doc.get("update_date", ""), "token_count": doc.get("token_count"), "thumbnail": doc.get("thumbnail", ""), "dataset_id": doc.get("dataset_id", dataset_id), "meta_fields": doc.get("meta_fields", {}), } doc_id_meta_list.append((doc_id, doc_meta)) docs[doc_id] = doc_meta page += 1 if docs_data.get("data", {}).get("total", 0) - page * page_size <= 0: page = None self._set_cached_document_metadata_by_dataset(dataset_id, doc_id_meta_list) if docs: document_cache.update(docs) except Exception as e: # Gracefully handle metadata cache failures logging.error(f"Problem building the document metadata cache: {str(e)}") pass return document_cache, dataset_cache def _map_chunk_fields(self, chunk_data, dataset_cache, document_cache): """Preserve all original API fields and add per-chunk document metadata""" # Start with ALL raw data from API (preserve everything like original version) mapped = dict(chunk_data) # Add dataset name enhancement dataset_id = chunk_data.get("dataset_id") or chunk_data.get("kb_id") if dataset_id and dataset_id in dataset_cache: mapped["dataset_name"] = dataset_cache[dataset_id]["name"] else: mapped["dataset_name"] = "Unknown" # Add document name convenience field mapped["document_name"] = chunk_data.get("document_keyword", "") # Add per-chunk document metadata document_id = chunk_data.get("document_id") if document_id and document_id in document_cache: mapped["document_metadata"] = document_cache[document_id] return mapped class RAGFlowCtx: def __init__(self, connector: RAGFlowConnector): self.conn = connector @asynccontextmanager async def sse_lifespan(server: Server) -> AsyncIterator[dict]: ctx = RAGFlowCtx(RAGFlowConnector(base_url=BASE_URL)) logging.info("Legacy SSE application started with StreamableHTTP session manager!") try: yield {"ragflow_ctx": ctx} finally: logging.info("Legacy SSE application shutting down...") app = Server("ragflow-mcp-server", lifespan=sse_lifespan) def with_api_key(required=True): def decorator(func): @wraps(func) async def wrapper(*args, **kwargs): ctx = app.request_context ragflow_ctx = ctx.lifespan_context.get("ragflow_ctx") if not ragflow_ctx: raise ValueError("Get RAGFlow Context failed") connector = ragflow_ctx.conn if MODE == LaunchMode.HOST: headers = ctx.session._init_options.capabilities.experimental.get("headers", {}) token = None # lower case here, because of Starlette conversion auth = headers.get("authorization", "") if auth.startswith("Bearer "): token = auth.removeprefix("Bearer ").strip() elif "api_key" in headers: token = headers["api_key"] if required and not token: raise ValueError("RAGFlow API key or Bearer token is required.") connector.bind_api_key(token) else: connector.bind_api_key(HOST_API_KEY) return await func(*args, connector=connector, **kwargs) return wrapper return decorator @app.list_tools() @with_api_key(required=True) async def list_tools(*, connector) -> list[types.Tool]: dataset_description = connector.list_datasets() return [ types.Tool( name="ragflow_retrieval", description="Retrieve relevant chunks from the RAGFlow retrieve interface based on the question. You can optionally specify dataset_ids to search only specific datasets, or omit dataset_ids entirely to search across ALL available datasets. You can also optionally specify document_ids to search within specific documents. When dataset_ids is not provided or is empty, the system will automatically search across all available datasets. Below is the list of all available datasets, including their descriptions and IDs:" + dataset_description, inputSchema={ "type": "object", "properties": { "dataset_ids": { "type": "array", "items": {"type": "string"}, "description": "Optional array of dataset IDs to search. If not provided or empty, all datasets will be searched." }, "document_ids": { "type": "array", "items": {"type": "string"}, "description": "Optional array of document IDs to search within." }, "question": { "type": "string", "description": "The question or query to search for." }, "page": { "type": "integer", "description": "Page number for pagination", "default": 1, "minimum": 1, }, "page_size": { "type": "integer", "description": "Number of results to return per page (default: 10, max recommended: 50 to avoid token limits)", "default": 10, "minimum": 1, "maximum": 100, }, "similarity_threshold": { "type": "number", "description": "Minimum similarity threshold for results", "default": 0.2, "minimum": 0.0, "maximum": 1.0, }, "vector_similarity_weight": { "type": "number", "description": "Weight for vector similarity vs term similarity", "default": 0.3, "minimum": 0.0, "maximum": 1.0, }, "keyword": { "type": "boolean", "description": "Enable keyword-based search", "default": False, }, "top_k": { "type": "integer", "description": "Maximum results to consider before ranking", "default": 1024, "minimum": 1, "maximum": 1024, }, "rerank_id": { "type": "string", "description": "Optional reranking model identifier", }, "force_refresh": { "type": "boolean", "description": "Set to true only if fresh dataset and document metadata is explicitly required. Otherwise, cached metadata is used (default: false).", "default": False, }, }, "required": ["question"], }, ), ] @app.call_tool() @with_api_key(required=True) async def call_tool(name: str, arguments: dict, *, connector) -> list[types.TextContent | types.ImageContent | types.EmbeddedResource]: if name == "ragflow_retrieval": document_ids = arguments.get("document_ids", []) dataset_ids = arguments.get("dataset_ids", []) question = arguments.get("question", "") page = arguments.get("page", 1) page_size = arguments.get("page_size", 10) similarity_threshold = arguments.get("similarity_threshold", 0.2) vector_similarity_weight = arguments.get("vector_similarity_weight", 0.3) keyword = arguments.get("keyword", False) top_k = arguments.get("top_k", 1024) rerank_id = arguments.get("rerank_id") force_refresh = arguments.get("force_refresh", False) # If no dataset_ids provided or empty list, get all available dataset IDs if not dataset_ids: dataset_list_str = connector.list_datasets() dataset_ids = [] # Parse the dataset list to extract IDs if dataset_list_str: for line in dataset_list_str.strip().split('\n'): if line.strip(): try: dataset_info = json.loads(line.strip()) dataset_ids.append(dataset_info["id"]) except (json.JSONDecodeError, KeyError): # Skip malformed lines continue return connector.retrieval( dataset_ids=dataset_ids, document_ids=document_ids, question=question, page=page, page_size=page_size, similarity_threshold=similarity_threshold, vector_similarity_weight=vector_similarity_weight, keyword=keyword, top_k=top_k, rerank_id=rerank_id, force_refresh=force_refresh, ) raise ValueError(f"Tool not found: {name}") def create_starlette_app(): routes = [] middleware = None if MODE == LaunchMode.HOST: from starlette.types import ASGIApp, Receive, Scope, Send class AuthMiddleware: def __init__(self, app: ASGIApp): self.app = app async def __call__(self, scope: Scope, receive: Receive, send: Send): if scope["type"] != "http": await self.app(scope, receive, send) return path = scope["path"] if path.startswith("/messages/") or path.startswith("/sse") or path.startswith("/mcp"): headers = dict(scope["headers"]) token = None auth_header = headers.get(b"authorization") if auth_header and auth_header.startswith(b"Bearer "): token = auth_header.removeprefix(b"Bearer ").strip() elif b"api_key" in headers: token = headers[b"api_key"] if not token: response = JSONResponse({"error": "Missing or invalid authorization header"}, status_code=401) await response(scope, receive, send) return await self.app(scope, receive, send) middleware = [Middleware(AuthMiddleware)] # Add SSE routes if enabled if TRANSPORT_SSE_ENABLED: from mcp.server.sse import SseServerTransport sse = SseServerTransport("/messages/") async def handle_sse(request): async with sse.connect_sse(request.scope, request.receive, request._send) as streams: await app.run(streams[0], streams[1], app.create_initialization_options(experimental_capabilities={"headers": dict(request.headers)})) return Response() routes.extend( [ Route("/sse", endpoint=handle_sse, methods=["GET"]), Mount("/messages/", app=sse.handle_post_message), ] ) # Add streamable HTTP route if enabled streamablehttp_lifespan = None if TRANSPORT_STREAMABLE_HTTP_ENABLED: from starlette.types import Receive, Scope, Send from mcp.server.streamable_http_manager import StreamableHTTPSessionManager session_manager = StreamableHTTPSessionManager( app=app, event_store=None, json_response=JSON_RESPONSE, stateless=True, ) async def handle_streamable_http(scope: Scope, receive: Receive, send: Send) -> None: await session_manager.handle_request(scope, receive, send) @asynccontextmanager async def streamablehttp_lifespan(app: Starlette) -> AsyncIterator[None]: async with session_manager.run(): logging.info("StreamableHTTP application started with StreamableHTTP session manager!") try: yield finally: logging.info("StreamableHTTP application shutting down...") routes.append(Mount("/mcp", app=handle_streamable_http)) return Starlette( debug=True, routes=routes, middleware=middleware, lifespan=streamablehttp_lifespan, ) @click.command() @click.option("--base-url", type=str, default="http://127.0.0.1:9380", help="API base URL for RAGFlow backend") @click.option("--host", type=str, default="127.0.0.1", help="Host to bind the RAGFlow MCP server") @click.option("--port", type=int, default=9382, help="Port to bind the RAGFlow MCP server") @click.option( "--mode", type=click.Choice(["self-host", "host"]), default="self-host", help=("Launch mode:\n self-host: run MCP for a single tenant (requires --api-key)\n host: multi-tenant mode, users must provide Authorization headers"), ) @click.option("--api-key", type=str, default="", help="API key to use when in self-host mode") @click.option( "--transport-sse-enabled/--no-transport-sse-enabled", default=True, help="Enable or disable legacy SSE transport mode (default: enabled)", ) @click.option( "--transport-streamable-http-enabled/--no-transport-streamable-http-enabled", default=True, help="Enable or disable streamable-http transport mode (default: enabled)", ) @click.option( "--json-response/--no-json-response", default=True, help="Enable or disable JSON response mode for streamable-http (default: enabled)", ) def main(base_url, host, port, mode, api_key, transport_sse_enabled, transport_streamable_http_enabled, json_response): import os import uvicorn from dotenv import load_dotenv load_dotenv() def parse_bool_flag(key: str, default: bool) -> bool: val = os.environ.get(key, str(default)) return str(val).strip().lower() in ("1", "true", "yes", "on") global BASE_URL, HOST, PORT, MODE, HOST_API_KEY, TRANSPORT_SSE_ENABLED, TRANSPORT_STREAMABLE_HTTP_ENABLED, JSON_RESPONSE BASE_URL = os.environ.get("RAGFLOW_MCP_BASE_URL", base_url) HOST = os.environ.get("RAGFLOW_MCP_HOST", host) PORT = os.environ.get("RAGFLOW_MCP_PORT", str(port)) MODE = os.environ.get("RAGFLOW_MCP_LAUNCH_MODE", mode) HOST_API_KEY = os.environ.get("RAGFLOW_MCP_HOST_API_KEY", api_key) TRANSPORT_SSE_ENABLED = parse_bool_flag("RAGFLOW_MCP_TRANSPORT_SSE_ENABLED", transport_sse_enabled) TRANSPORT_STREAMABLE_HTTP_ENABLED = parse_bool_flag("RAGFLOW_MCP_TRANSPORT_STREAMABLE_ENABLED", transport_streamable_http_enabled) JSON_RESPONSE = parse_bool_flag("RAGFLOW_MCP_JSON_RESPONSE", json_response) if MODE == LaunchMode.SELF_HOST and not HOST_API_KEY: raise click.UsageError("--api-key is required when --mode is 'self-host'") if TRANSPORT_STREAMABLE_HTTP_ENABLED and MODE == LaunchMode.HOST: raise click.UsageError("The --host mode is not supported with streamable-http transport yet.") if not TRANSPORT_STREAMABLE_HTTP_ENABLED and JSON_RESPONSE: JSON_RESPONSE = False print( r""" __ __ ____ ____ ____ _____ ______ _______ ____ | \/ |/ ___| _ \ / ___|| ____| _ \ \ / / ____| _ \ | |\/| | | | |_) | \___ \| _| | |_) \ \ / /| _| | |_) | | | | | |___| __/ ___) | |___| _ < \ V / | |___| _ < |_| |_|\____|_| |____/|_____|_| \_\ \_/ |_____|_| \_\ """, flush=True, ) print(f"MCP launch mode: {MODE}", flush=True) print(f"MCP host: {HOST}", flush=True) print(f"MCP port: {PORT}", flush=True) print(f"MCP base_url: {BASE_URL}", flush=True) if not any([TRANSPORT_SSE_ENABLED, TRANSPORT_STREAMABLE_HTTP_ENABLED]): print("At least one transport should be enabled, enable streamable-http automatically", flush=True) TRANSPORT_STREAMABLE_HTTP_ENABLED = True if TRANSPORT_SSE_ENABLED: print("SSE transport enabled: yes", flush=True) print("SSE endpoint available at /sse", flush=True) else: print("SSE transport enabled: no", flush=True) if TRANSPORT_STREAMABLE_HTTP_ENABLED: print("Streamable HTTP transport enabled: yes", flush=True) print("Streamable HTTP endpoint available at /mcp", flush=True) if JSON_RESPONSE: print("Streamable HTTP mode: JSON response enabled", flush=True) else: print("Streamable HTTP mode: SSE over HTTP enabled", flush=True) else: print("Streamable HTTP transport enabled: no", flush=True) if JSON_RESPONSE: print("Warning: --json-response ignored because streamable transport is disabled.", flush=True) uvicorn.run( create_starlette_app(), host=HOST, port=int(PORT), ) if __name__ == "__main__": """ Launch examples: 1. Self-host mode with both SSE and Streamable HTTP (in JSON response mode) enabled (default): uv run mcp/server/server.py --host=127.0.0.1 --port=9382 \ --base-url=http://127.0.0.1:9380 \ --mode=self-host --api-key=ragflow-xxxxx 2. Host mode (multi-tenant, self-host only, clients must provide Authorization headers): uv run mcp/server/server.py --host=127.0.0.1 --port=9382 \ --base-url=http://127.0.0.1:9380 \ --mode=host 3. Disable legacy SSE (only streamable HTTP will be active): uv run mcp/server/server.py --no-transport-sse-enabled \ --mode=self-host --api-key=ragflow-xxxxx 4. Disable streamable HTTP (only legacy SSE will be active): uv run mcp/server/server.py --no-transport-streamable-http-enabled \ --mode=self-host --api-key=ragflow-xxxxx 5. Use streamable HTTP with SSE-style events (disable JSON response): uv run mcp/server/server.py --transport-streamable-http-enabled --no-json-response \ --mode=self-host --api-key=ragflow-xxxxx 6. Disable both transports (for testing): uv run mcp/server/server.py --no-transport-sse-enabled --no-transport-streamable-http-enabled \ --mode=self-host --api-key=ragflow-xxxxx """ main()
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/mcp/client/client.py
mcp/client/client.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from mcp.client.session import ClientSession from mcp.client.sse import sse_client async def main(): try: # To access RAGFlow server in `host` mode, you need to attach `api_key` for each request to indicate identification. # async with sse_client("http://localhost:9382/sse", headers={"api_key": "ragflow-IyMGI1ZDhjMTA2ZTExZjBiYTMyMGQ4Zm"}) as streams: # Or follow the requirements of OAuth 2.1 Section 5 with Authorization header # async with sse_client("http://localhost:9382/sse", headers={"Authorization": "Bearer ragflow-IyMGI1ZDhjMTA2ZTExZjBiYTMyMGQ4Zm"}) as streams: async with sse_client("http://localhost:9382/sse") as streams: async with ClientSession( streams[0], streams[1], ) as session: await session.initialize() tools = await session.list_tools() print(f"{tools.tools=}") response = await session.call_tool(name="ragflow_retrieval", arguments={"dataset_ids": ["ce3bb17cf27a11efa69751e139332ced"], "document_ids": [], "question": "How to install neovim?"}) print(f"Tool response: {response.model_dump()}") except Exception as e: print(e) if __name__ == "__main__": from anyio import run run(main)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/mcp/client/streamable_http_client.py
mcp/client/streamable_http_client.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from mcp import ClientSession from mcp.client.streamable_http import streamablehttp_client async def main(): try: async with streamablehttp_client("http://localhost:9382/mcp/") as (read_stream, write_stream, _): async with ClientSession(read_stream, write_stream) as session: await session.initialize() tools = await session.list_tools() print(f"{tools.tools=}") response = await session.call_tool(name="ragflow_retrieval", arguments={"dataset_ids": ["bc4177924a7a11f09eff238aa5c10c94"], "document_ids": [], "question": "How to install neovim?"}) print(f"Tool response: {response.model_dump()}") except Exception as e: print(e) if __name__ == "__main__": from anyio import run run(main)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/use_existing_torch.py
use_existing_torch.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import glob for file in (*glob.glob("requirements/*.txt"), "pyproject.toml"): print(f">>> cleaning {file}") with open(file) as f: lines = f.readlines() if "torch" in "".join(lines).lower(): print("removed:") with open(file, "w") as f: for line in lines: if "torch" not in line.lower(): f.write(line) else: print(line.strip()) print(f"<<< done cleaning {file}\n")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/setup.py
setup.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import ctypes import importlib.util import json import logging import os import re import shutil import subprocess import sys import sysconfig from pathlib import Path from shutil import which import torch from packaging.version import Version, parse from setuptools import Extension, setup from setuptools.command.build_ext import build_ext from setuptools_scm import get_version from torch.utils.cpp_extension import CUDA_HOME, ROCM_HOME def load_module_from_path(module_name, path): spec = importlib.util.spec_from_file_location(module_name, path) module = importlib.util.module_from_spec(spec) sys.modules[module_name] = module spec.loader.exec_module(module) return module ROOT_DIR = Path(__file__).parent logger = logging.getLogger(__name__) # cannot import envs directly because it depends on vllm, # which is not installed yet envs = load_module_from_path("envs", os.path.join(ROOT_DIR, "vllm", "envs.py")) VLLM_TARGET_DEVICE = envs.VLLM_TARGET_DEVICE if sys.platform.startswith("darwin") and VLLM_TARGET_DEVICE != "cpu": logger.warning("VLLM_TARGET_DEVICE automatically set to `cpu` due to macOS") VLLM_TARGET_DEVICE = "cpu" elif not (sys.platform.startswith("linux") or sys.platform.startswith("darwin")): logger.warning( "vLLM only supports Linux platform (including WSL) and MacOS." "Building on %s, " "so vLLM may not be able to run correctly", sys.platform, ) VLLM_TARGET_DEVICE = "empty" elif sys.platform.startswith("linux") and os.getenv("VLLM_TARGET_DEVICE") is None: if torch.version.hip is not None: VLLM_TARGET_DEVICE = "rocm" logger.info("Auto-detected ROCm") elif torch.version.cuda is not None: VLLM_TARGET_DEVICE = "cuda" logger.info("Auto-detected CUDA") else: VLLM_TARGET_DEVICE = "cpu" def is_sccache_available() -> bool: return which("sccache") is not None and not bool( int(os.getenv("VLLM_DISABLE_SCCACHE", "0")) ) def is_ccache_available() -> bool: return which("ccache") is not None def is_ninja_available() -> bool: return which("ninja") is not None def is_freethreaded(): return bool(sysconfig.get_config_var("Py_GIL_DISABLED")) class CMakeExtension(Extension): def __init__(self, name: str, cmake_lists_dir: str = ".", **kwa) -> None: super().__init__(name, sources=[], py_limited_api=not is_freethreaded(), **kwa) self.cmake_lists_dir = os.path.abspath(cmake_lists_dir) class cmake_build_ext(build_ext): # A dict of extension directories that have been configured. did_config: dict[str, bool] = {} # # Determine number of compilation jobs and optionally nvcc compile threads. # def compute_num_jobs(self): # `num_jobs` is either the value of the MAX_JOBS environment variable # (if defined) or the number of CPUs available. num_jobs = envs.MAX_JOBS if num_jobs is not None: num_jobs = int(num_jobs) logger.info("Using MAX_JOBS=%d as the number of jobs.", num_jobs) else: try: # os.sched_getaffinity() isn't universally available, so fall # back to os.cpu_count() if we get an error here. num_jobs = len(os.sched_getaffinity(0)) except AttributeError: num_jobs = os.cpu_count() nvcc_threads = None if _is_cuda() and CUDA_HOME is not None: try: nvcc_version = get_nvcc_cuda_version() if nvcc_version >= Version("11.2"): # `nvcc_threads` is either the value of the NVCC_THREADS # environment variable (if defined) or 1. # when it is set, we reduce `num_jobs` to avoid # overloading the system. nvcc_threads = envs.NVCC_THREADS if nvcc_threads is not None: nvcc_threads = int(nvcc_threads) logger.info( "Using NVCC_THREADS=%d as the number of nvcc threads.", nvcc_threads, ) else: nvcc_threads = 1 num_jobs = max(1, num_jobs // nvcc_threads) except Exception as e: logger.warning("Failed to get NVCC version: %s", e) return num_jobs, nvcc_threads # # Perform cmake configuration for a single extension. # def configure(self, ext: CMakeExtension) -> None: # If we've already configured using the CMakeLists.txt for # this extension, exit early. if ext.cmake_lists_dir in cmake_build_ext.did_config: return cmake_build_ext.did_config[ext.cmake_lists_dir] = True # Select the build type. # Note: optimization level + debug info are set by the build type default_cfg = "Debug" if self.debug else "RelWithDebInfo" cfg = envs.CMAKE_BUILD_TYPE or default_cfg cmake_args = [ "-DCMAKE_BUILD_TYPE={}".format(cfg), "-DVLLM_TARGET_DEVICE={}".format(VLLM_TARGET_DEVICE), ] verbose = envs.VERBOSE if verbose: cmake_args += ["-DCMAKE_VERBOSE_MAKEFILE=ON"] if is_sccache_available(): cmake_args += [ "-DCMAKE_C_COMPILER_LAUNCHER=sccache", "-DCMAKE_CXX_COMPILER_LAUNCHER=sccache", "-DCMAKE_CUDA_COMPILER_LAUNCHER=sccache", "-DCMAKE_HIP_COMPILER_LAUNCHER=sccache", ] elif is_ccache_available(): cmake_args += [ "-DCMAKE_C_COMPILER_LAUNCHER=ccache", "-DCMAKE_CXX_COMPILER_LAUNCHER=ccache", "-DCMAKE_CUDA_COMPILER_LAUNCHER=ccache", "-DCMAKE_HIP_COMPILER_LAUNCHER=ccache", ] # Pass the python executable to cmake so it can find an exact # match. cmake_args += ["-DVLLM_PYTHON_EXECUTABLE={}".format(sys.executable)] # Pass the python path to cmake so it can reuse the build dependencies # on subsequent calls to python. cmake_args += ["-DVLLM_PYTHON_PATH={}".format(":".join(sys.path))] # Override the base directory for FetchContent downloads to $ROOT/.deps # This allows sharing dependencies between profiles, # and plays more nicely with sccache. # To override this, set the FETCHCONTENT_BASE_DIR environment variable. fc_base_dir = os.path.join(ROOT_DIR, ".deps") fc_base_dir = os.environ.get("FETCHCONTENT_BASE_DIR", fc_base_dir) cmake_args += ["-DFETCHCONTENT_BASE_DIR={}".format(fc_base_dir)] # # Setup parallelism and build tool # num_jobs, nvcc_threads = self.compute_num_jobs() if nvcc_threads: cmake_args += ["-DNVCC_THREADS={}".format(nvcc_threads)] if is_ninja_available(): build_tool = ["-G", "Ninja"] cmake_args += [ "-DCMAKE_JOB_POOL_COMPILE:STRING=compile", "-DCMAKE_JOB_POOLS:STRING=compile={}".format(num_jobs), ] else: # Default build tool to whatever cmake picks. build_tool = [] # Make sure we use the nvcc from CUDA_HOME if _is_cuda() and CUDA_HOME is not None: cmake_args += [f"-DCMAKE_CUDA_COMPILER={CUDA_HOME}/bin/nvcc"] elif _is_hip() and ROCM_HOME is not None: cmake_args += [f"-DROCM_PATH={ROCM_HOME}"] other_cmake_args = os.environ.get("CMAKE_ARGS") if other_cmake_args: cmake_args += other_cmake_args.split() subprocess.check_call( ["cmake", ext.cmake_lists_dir, *build_tool, *cmake_args], cwd=self.build_temp, ) def build_extensions(self) -> None: # Ensure that CMake is present and working try: subprocess.check_output(["cmake", "--version"]) except OSError as e: raise RuntimeError("Cannot find CMake executable") from e # Create build directory if it does not exist. if not os.path.exists(self.build_temp): os.makedirs(self.build_temp) targets = [] def target_name(s: str) -> str: return s.removeprefix("vllm.").removeprefix("vllm_flash_attn.") # Build all the extensions for ext in self.extensions: self.configure(ext) targets.append(target_name(ext.name)) num_jobs, _ = self.compute_num_jobs() build_args = [ "--build", ".", f"-j={num_jobs}", *[f"--target={name}" for name in targets], ] subprocess.check_call(["cmake", *build_args], cwd=self.build_temp) # Install the libraries for ext in self.extensions: # Install the extension into the proper location outdir = Path(self.get_ext_fullpath(ext.name)).parent.absolute() # Skip if the install directory is the same as the build directory if outdir == self.build_temp: continue # CMake appends the extension prefix to the install path, # and outdir already contains that prefix, so we need to remove it. prefix = outdir for _ in range(ext.name.count(".")): prefix = prefix.parent # prefix here should actually be the same for all components install_args = [ "cmake", "--install", ".", "--prefix", prefix, "--component", target_name(ext.name), ] subprocess.check_call(install_args, cwd=self.build_temp) def run(self): # First, run the standard build_ext command to compile the extensions super().run() # copy vllm/vllm_flash_attn/**/*.py from self.build_lib to current # directory so that they can be included in the editable build import glob files = glob.glob( os.path.join(self.build_lib, "vllm", "vllm_flash_attn", "**", "*.py"), recursive=True, ) for file in files: dst_file = os.path.join( "vllm/vllm_flash_attn", file.split("vllm/vllm_flash_attn/")[-1] ) print(f"Copying {file} to {dst_file}") os.makedirs(os.path.dirname(dst_file), exist_ok=True) self.copy_file(file, dst_file) if _is_cuda() or _is_hip(): # copy vllm/third_party/triton_kernels/**/*.py from self.build_lib # to current directory so that they can be included in the editable # build print( f"Copying {self.build_lib}/vllm/third_party/triton_kernels " "to vllm/third_party/triton_kernels" ) shutil.copytree( f"{self.build_lib}/vllm/third_party/triton_kernels", "vllm/third_party/triton_kernels", dirs_exist_ok=True, ) class precompiled_build_ext(build_ext): """Disables extension building when using precompiled binaries.""" def run(self) -> None: return def build_extensions(self) -> None: print("Skipping build_ext: using precompiled extensions.") return class precompiled_wheel_utils: """Extracts libraries and other files from an existing wheel.""" @staticmethod def fetch_metadata_for_variant( commit: str, variant: str | None ) -> tuple[list[dict], str]: """ Fetches metadata for a specific variant of the precompiled wheel. """ variant_dir = f"{variant}/" if variant is not None else "" repo_url = f"https://wheels.vllm.ai/{commit}/{variant_dir}vllm/" meta_url = repo_url + "metadata.json" print(f"Trying to fetch nightly build metadata from {meta_url}") from urllib.request import urlopen with urlopen(meta_url) as resp: # urlopen raises HTTPError on unexpected status code wheels = json.loads(resp.read().decode("utf-8")) return wheels, repo_url @staticmethod def is_rocm_system() -> bool: """Detect ROCm without relying on torch (for build environment).""" if os.getenv("ROCM_PATH"): return True if os.path.isdir("/opt/rocm"): return True if which("rocminfo") is not None: return True try: import torch return torch.version.hip is not None except ImportError: return False @staticmethod def find_local_rocm_wheel() -> str | None: """Search for a local vllm wheel in common locations.""" import glob for pattern in ["/vllm-workspace/dist/vllm-*.whl", "./dist/vllm-*.whl"]: wheels = glob.glob(pattern) if wheels: return sorted(wheels)[-1] return None @staticmethod def fetch_wheel_from_pypi_index(index_url: str, package: str = "vllm") -> str: """Fetch the latest wheel URL from a PyPI-style simple index.""" import platform from html.parser import HTMLParser from urllib.parse import urljoin from urllib.request import urlopen arch = platform.machine() class WheelLinkParser(HTMLParser): def __init__(self): super().__init__() self.wheels = [] def handle_starttag(self, tag, attrs): if tag == "a": for name, value in attrs: if name == "href" and value.endswith(".whl"): self.wheels.append(value) simple_url = f"{index_url.rstrip('/')}/{package}/" print(f"Fetching wheel list from {simple_url}") with urlopen(simple_url) as resp: html = resp.read().decode("utf-8") parser = WheelLinkParser() parser.feed(html) for wheel in reversed(parser.wheels): if arch in wheel: if wheel.startswith("http"): return wheel return urljoin(simple_url, wheel) raise ValueError(f"No compatible wheel found for {arch} at {simple_url}") @staticmethod def determine_wheel_url_rocm() -> tuple[str, str | None]: """Determine the precompiled wheel for ROCm.""" # Search for local wheel first local_wheel = precompiled_wheel_utils.find_local_rocm_wheel() if local_wheel is not None: print(f"Found local ROCm wheel: {local_wheel}") return local_wheel, None # Fall back to AMD's PyPI index index_url = os.getenv( "VLLM_ROCM_WHEEL_INDEX", "https://pypi.amd.com/vllm-rocm/simple" ) print(f"Fetching ROCm precompiled wheel from {index_url}") wheel_url = precompiled_wheel_utils.fetch_wheel_from_pypi_index(index_url) download_filename = wheel_url.split("/")[-1].split("#")[0] print(f"Using ROCm precompiled wheel: {wheel_url}") return wheel_url, download_filename @staticmethod def determine_wheel_url() -> tuple[str, str | None]: """ Try to determine the precompiled wheel URL or path to use. The order of preference is: 1. user-specified wheel location (can be either local or remote, via VLLM_PRECOMPILED_WHEEL_LOCATION) 2. user-specified variant (VLLM_PRECOMPILED_WHEEL_VARIANT) from nightly repo 3. the variant corresponding to VLLM_MAIN_CUDA_VERSION from nightly repo 4. the default variant from nightly repo If downloading from the nightly repo, the commit can be specified via VLLM_PRECOMPILED_WHEEL_COMMIT; otherwise, the head commit in the main branch is used. """ wheel_location = os.getenv("VLLM_PRECOMPILED_WHEEL_LOCATION", None) if wheel_location is not None: print(f"Using user-specified precompiled wheel location: {wheel_location}") return wheel_location, None else: # ROCm: use local wheel or AMD's PyPI index # TODO: When we have ROCm nightly wheels, we can update this logic. if precompiled_wheel_utils.is_rocm_system(): return precompiled_wheel_utils.determine_wheel_url_rocm() import platform arch = platform.machine() # try to fetch the wheel metadata from the nightly wheel repo main_variant = "cu" + envs.VLLM_MAIN_CUDA_VERSION.replace(".", "") variant = os.getenv("VLLM_PRECOMPILED_WHEEL_VARIANT", main_variant) commit = os.getenv("VLLM_PRECOMPILED_WHEEL_COMMIT", "").lower() if not commit or len(commit) != 40: print( f"VLLM_PRECOMPILED_WHEEL_COMMIT not valid: {commit}" ", trying to fetch base commit in main branch" ) commit = precompiled_wheel_utils.get_base_commit_in_main_branch() print(f"Using precompiled wheel commit {commit} with variant {variant}") try_default = False wheels, repo_url, download_filename = None, None, None try: wheels, repo_url = precompiled_wheel_utils.fetch_metadata_for_variant( commit, variant ) except Exception as e: logger.warning( "Failed to fetch precompiled wheel metadata for variant %s: %s", variant, e, ) try_default = True # try outside handler to keep the stacktrace simple if try_default: print("Trying the default variant from remote") wheels, repo_url = precompiled_wheel_utils.fetch_metadata_for_variant( commit, None ) # if this also fails, then we have nothing more to try / cache assert wheels is not None and repo_url is not None, ( "Failed to fetch precompiled wheel metadata" ) # The metadata.json has the following format: # see .buildkite/scripts/generate-nightly-index.py for details """[{ "package_name": "vllm", "version": "0.11.2.dev278+gdbc3d9991", "build_tag": null, "python_tag": "cp38", "abi_tag": "abi3", "platform_tag": "manylinux1_x86_64", "variant": null, "filename": "vllm-0.11.2.dev278+gdbc3d9991-cp38-abi3-manylinux1_x86_64.whl", "path": "../vllm-0.11.2.dev278%2Bgdbc3d9991-cp38-abi3-manylinux1_x86_64.whl" }, ...]""" from urllib.parse import urljoin for wheel in wheels: # TODO: maybe check more compatibility later? (python_tag, abi_tag, etc) if wheel.get("package_name") == "vllm" and arch in wheel.get( "platform_tag", "" ): print(f"Found precompiled wheel metadata: {wheel}") if "path" not in wheel: raise ValueError(f"Wheel metadata missing path: {wheel}") wheel_url = urljoin(repo_url, wheel["path"]) download_filename = wheel.get("filename") print(f"Using precompiled wheel URL: {wheel_url}") break else: raise ValueError( f"No precompiled vllm wheel found for architecture {arch} " f"from repo {repo_url}. All available wheels: {wheels}" ) return wheel_url, download_filename @staticmethod def extract_precompiled_and_patch_package( wheel_url_or_path: str, download_filename: str | None ) -> dict: import tempfile import zipfile temp_dir = None try: if not os.path.isfile(wheel_url_or_path): # use provided filename first, then derive from URL wheel_filename = download_filename or wheel_url_or_path.split("/")[-1] temp_dir = tempfile.mkdtemp(prefix="vllm-wheels") wheel_path = os.path.join(temp_dir, wheel_filename) print(f"Downloading wheel from {wheel_url_or_path} to {wheel_path}") from urllib.request import urlretrieve urlretrieve(wheel_url_or_path, filename=wheel_path) else: wheel_path = wheel_url_or_path print(f"Using existing wheel at {wheel_path}") package_data_patch = {} with zipfile.ZipFile(wheel_path) as wheel: files_to_copy = [ "vllm/_C.abi3.so", "vllm/_moe_C.abi3.so", "vllm/_flashmla_C.abi3.so", "vllm/_flashmla_extension_C.abi3.so", "vllm/_sparse_flashmla_C.abi3.so", "vllm/vllm_flash_attn/_vllm_fa2_C.abi3.so", "vllm/vllm_flash_attn/_vllm_fa3_C.abi3.so", "vllm/cumem_allocator.abi3.so", # ROCm-specific libraries "vllm/_rocm_C.abi3.so", ] flash_attn_regex = re.compile( r"vllm/vllm_flash_attn/(?:[^/.][^/]*/)*(?!\.)[^/]*\.py" ) triton_kernels_regex = re.compile( r"vllm/third_party/triton_kernels/(?:[^/.][^/]*/)*(?!\.)[^/]*\.py" ) file_members = list( filter(lambda x: x.filename in files_to_copy, wheel.filelist) ) file_members += list( filter(lambda x: flash_attn_regex.match(x.filename), wheel.filelist) ) file_members += list( filter( lambda x: triton_kernels_regex.match(x.filename), wheel.filelist ) ) for file in file_members: print(f"[extract] {file.filename}") target_path = os.path.join(".", file.filename) os.makedirs(os.path.dirname(target_path), exist_ok=True) with ( wheel.open(file.filename) as src, open(target_path, "wb") as dst, ): shutil.copyfileobj(src, dst) pkg = os.path.dirname(file.filename).replace("/", ".") package_data_patch.setdefault(pkg, []).append( os.path.basename(file.filename) ) return package_data_patch finally: if temp_dir is not None: print(f"Removing temporary directory {temp_dir}") shutil.rmtree(temp_dir) @staticmethod def get_base_commit_in_main_branch() -> str: try: # Get the latest commit hash of the upstream main branch. resp_json = subprocess.check_output( [ "curl", "-s", "https://api.github.com/repos/vllm-project/vllm/commits/main", ] ).decode("utf-8") upstream_main_commit = json.loads(resp_json)["sha"] print(f"Upstream main branch latest commit: {upstream_main_commit}") # In Docker build context, .git may be immutable or missing. if envs.VLLM_DOCKER_BUILD_CONTEXT: return upstream_main_commit # Check if the upstream_main_commit exists in the local repo try: subprocess.check_output( ["git", "cat-file", "-e", f"{upstream_main_commit}"] ) except subprocess.CalledProcessError: # If not present, fetch it from the remote repository. # Note that this does not update any local branches, # but ensures that this commit ref and its history are # available in our local repo. subprocess.check_call( ["git", "fetch", "https://github.com/vllm-project/vllm", "main"] ) # Then get the commit hash of the current branch that is the same as # the upstream main commit. current_branch = ( subprocess.check_output(["git", "branch", "--show-current"]) .decode("utf-8") .strip() ) base_commit = ( subprocess.check_output( ["git", "merge-base", f"{upstream_main_commit}", current_branch] ) .decode("utf-8") .strip() ) return base_commit except ValueError as err: raise ValueError(err) from None except Exception as err: logger.warning( "Failed to get the base commit in the main branch. " "Using the nightly wheel. The libraries in this " "wheel may not be compatible with your dev branch: %s", err, ) return "nightly" def _no_device() -> bool: return VLLM_TARGET_DEVICE == "empty" def _is_cuda() -> bool: has_cuda = torch.version.cuda is not None return VLLM_TARGET_DEVICE == "cuda" and has_cuda and not _is_tpu() def _is_hip() -> bool: return ( VLLM_TARGET_DEVICE == "cuda" or VLLM_TARGET_DEVICE == "rocm" ) and torch.version.hip is not None def _is_tpu() -> bool: return VLLM_TARGET_DEVICE == "tpu" def _is_cpu() -> bool: return VLLM_TARGET_DEVICE == "cpu" def _is_xpu() -> bool: return VLLM_TARGET_DEVICE == "xpu" def _build_custom_ops() -> bool: return _is_cuda() or _is_hip() or _is_cpu() def get_rocm_version(): # Get the Rocm version from the ROCM_HOME/bin/librocm-core.so # see https://github.com/ROCm/rocm-core/blob/d11f5c20d500f729c393680a01fa902ebf92094b/rocm_version.cpp#L21 try: if ROCM_HOME is None: return None librocm_core_file = Path(ROCM_HOME) / "lib" / "librocm-core.so" if not librocm_core_file.is_file(): return None librocm_core = ctypes.CDLL(librocm_core_file) VerErrors = ctypes.c_uint32 get_rocm_core_version = librocm_core.getROCmVersion get_rocm_core_version.restype = VerErrors get_rocm_core_version.argtypes = [ ctypes.POINTER(ctypes.c_uint32), ctypes.POINTER(ctypes.c_uint32), ctypes.POINTER(ctypes.c_uint32), ] major = ctypes.c_uint32() minor = ctypes.c_uint32() patch = ctypes.c_uint32() if ( get_rocm_core_version( ctypes.byref(major), ctypes.byref(minor), ctypes.byref(patch) ) == 0 ): return f"{major.value}.{minor.value}.{patch.value}" return None except Exception: return None def get_nvcc_cuda_version() -> Version: """Get the CUDA version from nvcc. Adapted from https://github.com/NVIDIA/apex/blob/8b7a1ff183741dd8f9b87e7bafd04cfde99cea28/setup.py """ assert CUDA_HOME is not None, "CUDA_HOME is not set" nvcc_output = subprocess.check_output( [CUDA_HOME + "/bin/nvcc", "-V"], universal_newlines=True ) output = nvcc_output.split() release_idx = output.index("release") + 1 nvcc_cuda_version = parse(output[release_idx].split(",")[0]) return nvcc_cuda_version def get_vllm_version() -> str: # Allow overriding the version. This is useful to build platform-specific # wheels (e.g. CPU, TPU) without modifying the source. if env_version := os.getenv("VLLM_VERSION_OVERRIDE"): print(f"Overriding VLLM version with {env_version} from VLLM_VERSION_OVERRIDE") os.environ["SETUPTOOLS_SCM_PRETEND_VERSION"] = env_version return get_version(write_to="vllm/_version.py") version = get_version(write_to="vllm/_version.py") sep = "+" if "+" not in version else "." # dev versions might contain + if _no_device(): if envs.VLLM_TARGET_DEVICE == "empty": version += f"{sep}empty" elif _is_cuda(): if envs.VLLM_USE_PRECOMPILED and not envs.VLLM_SKIP_PRECOMPILED_VERSION_SUFFIX: version += f"{sep}precompiled" else: cuda_version = str(get_nvcc_cuda_version()) if cuda_version != envs.VLLM_MAIN_CUDA_VERSION: cuda_version_str = cuda_version.replace(".", "")[:3] # skip this for source tarball, required for pypi if "sdist" not in sys.argv: version += f"{sep}cu{cuda_version_str}" elif _is_hip(): # Get the Rocm Version rocm_version = get_rocm_version() or torch.version.hip if rocm_version and rocm_version != envs.VLLM_MAIN_CUDA_VERSION: version += f"{sep}rocm{rocm_version.replace('.', '')[:3]}" elif _is_tpu(): version += f"{sep}tpu" elif _is_cpu(): if envs.VLLM_TARGET_DEVICE == "cpu": version += f"{sep}cpu" elif _is_xpu(): version += f"{sep}xpu" else: raise RuntimeError("Unknown runtime environment") return version def get_requirements() -> list[str]: """Get Python package dependencies from requirements.txt.""" requirements_dir = ROOT_DIR / "requirements" def _read_requirements(filename: str) -> list[str]: with open(requirements_dir / filename) as f: requirements = f.read().strip().split("\n") resolved_requirements = [] for line in requirements: if line.startswith("-r "): resolved_requirements += _read_requirements(line.split()[1]) elif ( not line.startswith("--") and not line.startswith("#") and line.strip() != "" ): resolved_requirements.append(line) return resolved_requirements if _no_device(): requirements = _read_requirements("common.txt") elif _is_cuda(): requirements = _read_requirements("cuda.txt") cuda_major, cuda_minor = torch.version.cuda.split(".") modified_requirements = [] for req in requirements: if "vllm-flash-attn" in req and cuda_major != "12": # vllm-flash-attn is built only for CUDA 12.x. # Skip for other versions. continue modified_requirements.append(req) requirements = modified_requirements elif _is_hip(): requirements = _read_requirements("rocm.txt") elif _is_tpu(): requirements = _read_requirements("tpu.txt") elif _is_cpu(): requirements = _read_requirements("cpu.txt") elif _is_xpu(): requirements = _read_requirements("xpu.txt") else: raise ValueError("Unsupported platform, please use CUDA, ROCm, or CPU.") return requirements ext_modules = [] if _is_cuda() or _is_hip(): ext_modules.append(CMakeExtension(name="vllm._moe_C")) ext_modules.append(CMakeExtension(name="vllm.cumem_allocator")) # Optional since this doesn't get built (produce an .so file). This is just # copying the relevant .py files from the source repository. ext_modules.append(CMakeExtension(name="vllm.triton_kernels", optional=True)) if _is_hip(): ext_modules.append(CMakeExtension(name="vllm._rocm_C")) if _is_cuda(): ext_modules.append(CMakeExtension(name="vllm.vllm_flash_attn._vllm_fa2_C")) if envs.VLLM_USE_PRECOMPILED or ( CUDA_HOME and get_nvcc_cuda_version() >= Version("12.3") ): # FA3 requires CUDA 12.3 or later ext_modules.append(CMakeExtension(name="vllm.vllm_flash_attn._vllm_fa3_C")) # Optional since this doesn't get built (produce an .so file) when # not targeting a hopper system ext_modules.append(CMakeExtension(name="vllm._flashmla_C", optional=True)) ext_modules.append( CMakeExtension(name="vllm._flashmla_extension_C", optional=True) ) if _build_custom_ops(): ext_modules.append(CMakeExtension(name="vllm._C")) package_data = { "vllm": [ "py.typed", "model_executor/layers/fused_moe/configs/*.json",
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_long_document_qa_throughput.py
benchmarks/benchmark_long_document_qa_throughput.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Offline benchmark to test the long document QA throughput. Example usage: # This workload samples 8 different prompts with a default input # length of 20000 tokens, then replicates each prompt 2 times # in random order. python benchmark_long_document_qa_throughput.py \ --model meta-llama/Llama-2-7b-chat-hf \ --enable-prefix-caching \ --num-documents 8 \ --repeat-count 2 Commandline arguments: --num-documents: The number of documents to sample prompts from. --document-length: The length of each document in tokens. (Optional, default: 20000) --output-len: The number of tokens to generate for each prompt. (Optional, default: 10) --repeat-count: The number of times to repeat each prompt. (Optional, default: 2) --repeat-mode: The mode to repeat prompts. The supported modes are: - 'random': shuffle the prompts randomly. (Default) - 'tile': the entire prompt list is repeated in sequence. (Potentially lowest cache hit) - 'interleave': each prompt is repeated consecutively before moving to the next element. (Highest cache hit) --shuffle-seed: Random seed when the repeat mode is "random". (Optional, default: 0) In the meantime, it also supports all the vLLM engine args to initialize the LLM engine. You can refer to the `vllm.engine.arg_utils.EngineArgs` for more details. """ import dataclasses import random import time from vllm import LLM, SamplingParams from vllm.engine.arg_utils import EngineArgs from vllm.utils.argparse_utils import FlexibleArgumentParser def test_long_document_qa(llm=None, sampling_params=None, prompts=None): """ Test long document QA with the given prompts and sampling parameters. Print the time spent in processing all the prompts. Args: llm: The language model used for generating responses. sampling_params: Sampling parameter used to generate the response. prompts: A list of prompt strings to be processed by the LLM. """ start_time = time.time() llm.generate(prompts, sampling_params=sampling_params) end_time = time.time() print(f"Time to execute all requests: {end_time - start_time:.4f} secs") def repeat_prompts(prompts, repeat_count, mode: str): """ Repeat each prompt in the list for a specified number of times. The order of prompts in the output list depends on the mode. Args: prompts: A list of prompts to be repeated. repeat_count: The number of times each prompt is repeated. mode: The mode of repetition. Supported modes are: - 'random': Shuffle the prompts randomly after repetition. - 'tile': Repeat the entire prompt list in sequence. Example: [1, 2, 3] -> [1, 2, 3, 1, 2, 3]. - 'interleave': Repeat each prompt consecutively before moving to the next. Example: [1, 2, 3] -> [1, 1, 2, 2, 3, 3]. Returns: A list of repeated prompts in the specified order. Raises: ValueError: If an invalid mode is provided. """ print("Repeat mode: ", mode) if mode == "random": repeated_prompts = prompts * repeat_count random.shuffle(repeated_prompts) return repeated_prompts elif mode == "tile": return prompts * repeat_count elif mode == "interleave": repeated_prompts = [] for prompt in prompts: repeated_prompts.extend([prompt] * repeat_count) return repeated_prompts else: raise ValueError( f"Invalid mode: {mode}, only support 'random', 'tile', 'interleave'" ) def main(args): random.seed(args.shuffle_seed) # Prepare the prompts: # we append the document id at the beginning to avoid any of the document # being the prefix of other documents prompts = [ str(i) + " ".join(["hi"] * args.document_length) for i in range(args.num_documents) ] prompts = repeat_prompts(prompts, args.repeat_count, mode=args.repeat_mode) warmup_prompts = [ "This is warm up request " + str(i) + " ".join(["hi"] * args.document_length) for i in range(args.num_documents) ] # Create the LLM engine engine_args = EngineArgs.from_cli_args(args) llm = LLM(**dataclasses.asdict(engine_args)) sampling_params = SamplingParams(temperature=0, max_tokens=args.output_len) print("------warm up------") test_long_document_qa( llm=llm, prompts=warmup_prompts, sampling_params=sampling_params, ) print("------start generating------") test_long_document_qa( llm=llm, prompts=prompts, sampling_params=sampling_params, ) def create_argument_parser(): parser = FlexibleArgumentParser( description="Benchmark the performance with or " "without automatic prefix caching." ) parser.add_argument( "--document-length", type=int, # Roughly the number of tokens for a system paper, # excluding images default=20000, help="Range of input lengths for sampling prompts, " 'specified as "min:max" (e.g., "128:256").', ) parser.add_argument( "--num-documents", type=int, default=8, help="Range of input lengths for sampling prompts, " 'specified as "min:max" (e.g., "128:256").', ) parser.add_argument("--output-len", type=int, default=10) parser.add_argument( "--repeat-count", type=int, default=2, help="Number of times to repeat each prompt", ) parser.add_argument( "--repeat-mode", type=str, default="random", help="The mode to repeat prompts. The supported " 'modes are "random", "tile", and "interleave". ' "See repeat_prompts() in the source code for details.", ) parser.add_argument( "--shuffle-seed", type=int, default=0, help='Random seed when the repeat mode is "random"', ) parser = EngineArgs.add_cli_args(parser) return parser if __name__ == "__main__": parser = create_argument_parser() args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_latency.py
benchmarks/benchmark_latency.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import sys if __name__ == "__main__": print("""DEPRECATED: This script has been moved to the vLLM CLI. Please use the following command instead: vllm bench latency For help with the new command, run: vllm bench latency --help Alternatively, you can run the new command directly with: python -m vllm.entrypoints.cli.main bench latency --help """) sys.exit(1)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_batch_invariance.py
benchmarks/benchmark_batch_invariance.py
#!/usr/bin/env python3 # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Benchmark to measure the performance overhead of VLLM_BATCH_INVARIANT mode. This benchmark runs the same workload twice: 1. With VLLM_BATCH_INVARIANT=0 (baseline) 2. With VLLM_BATCH_INVARIANT=1 (batch invariant mode) And reports the timing and throughput metrics for comparison. Environment variables: VLLM_BENCH_MODEL: Model to benchmark (default: "Qwen/Qwen3-1.7B") VLLM_BENCH_TP_SIZE: Tensor parallel size (default: 1, use 8 for deepseek) VLLM_BENCH_BATCH_SIZE: Max batch size (default: 128) VLLM_BENCH_NUM_TRIALS: Number of trials to run (default: 5) VLLM_BENCH_MIN_PROMPT: Min prompt length in words (default: 1024) VLLM_BENCH_MAX_PROMPT: Max prompt length in words (default: 2048) VLLM_BENCH_MAX_TOKENS: Max tokens to generate (default: 128) VLLM_BENCH_TEMPERATURE: Temperature for sampling (default: 0.0) VLLM_BENCH_GPU_MEMORY_UTILIZATION: GPU memory utilization (default: 0.4) VLLM_BENCH_MAX_MODEL_LEN: Max model length (default: 5120) VLLM_BENCH_BACKEND: Attention backend (default: FLASH_ATTN) Example usage: # Benchmark qwen3 (default) python benchmarks/benchmark_batch_invariance.py # Benchmark deepseek with 8 GPUs VLLM_BENCH_MODEL="deepseek-ai/DeepSeek-V3" VLLM_BENCH_TP_SIZE=8 \\ python benchmarks/benchmark_batch_invariance.py # Quick test with fewer trials VLLM_BENCH_NUM_TRIALS=2 VLLM_BENCH_BATCH_SIZE=32 \\ python benchmarks/benchmark_batch_invariance.py """ import contextlib import os import random import time from vllm import LLM, SamplingParams from vllm.platforms import current_platform def _random_prompt(min_words: int = 1024, max_words: int = 1024 * 2) -> str: """Generate a random prompt for benchmarking.""" prompt_templates = [ "Question: What is the capital of France?\nAnswer: The capital of France is", "Q: How does photosynthesis work?\nA: Photosynthesis is the process by which", "User: Can you explain quantum mechanics?\nAssistant: Quantum mechanics is", "Once upon a time in a distant galaxy, there lived", "The old man walked slowly down the street, remembering", "In the year 2157, humanity finally discovered", "To implement a binary search tree in Python, first we need to", "The algorithm works by iterating through the array and", "Here's how to optimize database queries using indexing:", "The Renaissance was a period in European history that", "Climate change is caused by several factors including", "The human brain contains approximately 86 billion neurons which", "I've been thinking about getting a new laptop because", "Yesterday I went to the store and bought", "My favorite thing about summer is definitely", ] base_prompt = random.choice(prompt_templates) if max_words < min_words: max_words = min_words target_words = random.randint(min_words, max_words) if target_words > 50: padding_text = ( " This is an interesting topic that deserves more explanation. " * (target_words // 50) ) base_prompt = base_prompt + padding_text return base_prompt def run_benchmark_with_batch_invariant( model: str, tp_size: int, max_batch_size: int, num_trials: int, min_prompt: int, max_prompt: int, max_tokens: int, temperature: float, gpu_mem_util: float, max_model_len: int, backend: str, batch_invariant: bool, seed: int = 12345, ) -> dict: """ Run the benchmark with the specified configuration. Returns a dict with timing and throughput metrics. """ random.seed(seed) # Set environment variables if batch_invariant: os.environ["VLLM_BATCH_INVARIANT"] = "1" else: os.environ["VLLM_BATCH_INVARIANT"] = "0" print(f"\n{'=' * 80}") print(f"BENCHMARK: VLLM_BATCH_INVARIANT={int(batch_invariant)}") print(f" Model: {model}") print(f" TP Size: {tp_size}") print(f" Backend: {backend}") print(f" Max Batch Size: {max_batch_size}") print(f" Trials: {num_trials}") print(f" Max Tokens: {max_tokens}") print(f"{'=' * 80}\n") sampling = SamplingParams( temperature=temperature, top_p=0.95, max_tokens=max_tokens, seed=20240919, ) needle_prompt = "There once was a " llm = None try: # Create LLM engine start_init = time.perf_counter() llm = LLM( model=model, max_num_seqs=max_batch_size, gpu_memory_utilization=gpu_mem_util, max_model_len=max_model_len, dtype="bfloat16", tensor_parallel_size=tp_size, attention_config={"backend": backend}, enable_prefix_caching=False, ) init_time = time.perf_counter() - start_init print(f"Engine initialization time: {init_time:.2f}s\n") # Generate baseline print("Generating baseline (warmup)...") baseline_out = llm.generate([needle_prompt], sampling) assert len(baseline_out) == 1 baseline_text = baseline_out[0].outputs[0].text print(f"Baseline output: '{baseline_text[:50]}...'\n") # Run trials and measure timing trial_times: list[float] = [] total_tokens = 0 total_prompts = 0 for trial in range(num_trials): # Create a batch prompts: list[str] = [] batch_size = random.randint(max_batch_size // 2, max_batch_size) needle_pos = random.randint(0, batch_size - 1) for i in range(batch_size): if i == needle_pos: prompts.append(needle_prompt) else: prompts.append(_random_prompt(min_prompt, max_prompt)) # Measure time for this trial start_time = time.perf_counter() outputs = llm.generate(prompts, sampling) trial_time = time.perf_counter() - start_time trial_times.append(trial_time) total_prompts += len(prompts) # Count tokens for output in outputs: if output.outputs: total_tokens += len(output.outputs[0].token_ids) print( f"Trial {trial + 1}/{num_trials}: " f"batch_size={batch_size}, " f"time={trial_time:.2f}s" ) # Verify needle output still matches needle_output = outputs[needle_pos] assert needle_output.prompt == needle_prompt # Compute statistics avg_time = sum(trial_times) / len(trial_times) min_time = min(trial_times) max_time = max(trial_times) throughput = total_tokens / sum(trial_times) prompts_per_sec = total_prompts / sum(trial_times) print(f"\n{'=' * 80}") print("RESULTS:") print(f" Average time per trial: {avg_time:.2f}s") print(f" Min time: {min_time:.2f}s") print(f" Max time: {max_time:.2f}s") print(f" Total tokens generated: {total_tokens}") print(f" Total prompts processed: {total_prompts}") print(f" Throughput: {throughput:.2f} tokens/s") print(f" Prompts/s: {prompts_per_sec:.2f}") print(f"{'=' * 80}\n") return { "init_time": init_time, "avg_time": avg_time, "min_time": min_time, "max_time": max_time, "total_tokens": total_tokens, "total_prompts": total_prompts, "throughput": throughput, "prompts_per_sec": prompts_per_sec, "trial_times": trial_times, } finally: # Cleanup if llm is not None: with contextlib.suppress(Exception): llm.shutdown() def main(): # Check platform support if not (current_platform.is_cuda() and current_platform.has_device_capability(90)): print("ERROR: Requires CUDA and >= Hopper (SM90)") print(f"Current platform: {current_platform.device_type}") if current_platform.is_cuda(): print(f"Device capability: {current_platform.get_device_capability()}") return 1 # Read configuration from environment model = os.getenv("VLLM_BENCH_MODEL", "Qwen/Qwen3-1.7B") tp_size = int(os.getenv("VLLM_BENCH_TP_SIZE", "1")) max_batch_size = int(os.getenv("VLLM_BENCH_BATCH_SIZE", "128")) num_trials = int(os.getenv("VLLM_BENCH_NUM_TRIALS", "5")) min_prompt = int(os.getenv("VLLM_BENCH_MIN_PROMPT", "1024")) max_prompt = int(os.getenv("VLLM_BENCH_MAX_PROMPT", "2048")) max_tokens = int(os.getenv("VLLM_BENCH_MAX_TOKENS", "128")) temperature = float(os.getenv("VLLM_BENCH_TEMPERATURE", "0.0")) gpu_mem_util = float(os.getenv("VLLM_BENCH_GPU_MEMORY_UTILIZATION", "0.4")) max_model_len = int(os.getenv("VLLM_BENCH_MAX_MODEL_LEN", "5120")) backend = os.getenv("VLLM_BENCH_BACKEND", "FLASH_ATTN") print("\n" + "=" * 80) print("VLLM BATCH INVARIANCE BENCHMARK") print("=" * 80) print("\nConfiguration:") print(f" Model: {model}") print(f" Tensor Parallel Size: {tp_size}") print(f" Attention Backend: {backend}") print(f" Max Batch Size: {max_batch_size}") print(f" Number of Trials: {num_trials}") print(f" Prompt Length Range: {min_prompt}-{max_prompt} words") print(f" Max Tokens to Generate: {max_tokens}") print(f" Temperature: {temperature}") print(f" GPU Memory Utilization: {gpu_mem_util}") print(f" Max Model Length: {max_model_len}") print("=" * 80) # Run benchmark WITHOUT batch invariance (baseline) print("\n" + "=" * 80) print("PHASE 1: Running WITHOUT batch invariance (baseline)") print("=" * 80) baseline_results = run_benchmark_with_batch_invariant( model=model, tp_size=tp_size, max_batch_size=max_batch_size, num_trials=num_trials, min_prompt=min_prompt, max_prompt=max_prompt, max_tokens=max_tokens, temperature=temperature, gpu_mem_util=gpu_mem_util, max_model_len=max_model_len, backend=backend, batch_invariant=False, ) # Run benchmark WITH batch invariance print("\n" + "=" * 80) print("PHASE 2: Running WITH batch invariance") print("=" * 80) batch_inv_results = run_benchmark_with_batch_invariant( model=model, tp_size=tp_size, max_batch_size=max_batch_size, num_trials=num_trials, min_prompt=min_prompt, max_prompt=max_prompt, max_tokens=max_tokens, temperature=temperature, gpu_mem_util=gpu_mem_util, max_model_len=max_model_len, backend=backend, batch_invariant=True, ) # Compare results print("\n" + "=" * 80) print("COMPARISON: Batch Invariance vs Baseline") print("=" * 80) init_overhead_pct = ( (batch_inv_results["init_time"] - baseline_results["init_time"]) / baseline_results["init_time"] * 100 ) time_overhead_pct = ( (batch_inv_results["avg_time"] - baseline_results["avg_time"]) / baseline_results["avg_time"] * 100 ) throughput_change_pct = ( (batch_inv_results["throughput"] - baseline_results["throughput"]) / baseline_results["throughput"] * 100 ) print("\nInitialization Time:") print(f" Baseline: {baseline_results['init_time']:.2f}s") print(f" Batch Invariant: {batch_inv_results['init_time']:.2f}s") print(f" Overhead: {init_overhead_pct:+.2f}%") print("\nAverage Trial Time:") print(f" Baseline: {baseline_results['avg_time']:.2f}s") print(f" Batch Invariant: {batch_inv_results['avg_time']:.2f}s") print(f" Overhead: {time_overhead_pct:+.2f}%") print("\nThroughput (tokens/s):") print(f" Baseline: {baseline_results['throughput']:.2f}") print(f" Batch Invariant: {batch_inv_results['throughput']:.2f}") print(f" Change: {throughput_change_pct:+.2f}%") print("\nPrompts/s:") print(f" Baseline: {baseline_results['prompts_per_sec']:.2f}") print(f" Batch Invariant: {batch_inv_results['prompts_per_sec']:.2f}") print("\n" + "=" * 80) print("SUMMARY") print("=" * 80) if time_overhead_pct > 0: print( f"Batch invariance mode adds approximately {time_overhead_pct:.1f}% " "overhead" ) else: print( f"Batch invariance mode is approximately {-time_overhead_pct:.1f}% " "faster (unexpected!)" ) if abs(throughput_change_pct) < 1.0: print("Throughput difference is negligible (< 1%)") elif throughput_change_pct < 0: print( f"Throughput decreased by {-throughput_change_pct:.1f}% " "with batch invariance" ) else: print( f"Throughput increased by {throughput_change_pct:.1f}% " "with batch invariance (unexpected!)" ) print("=" * 80 + "\n") return 0 if __name__ == "__main__": exit(main())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_throughput.py
benchmarks/benchmark_throughput.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import sys if __name__ == "__main__": print("""DEPRECATED: This script has been moved to the vLLM CLI. Please use the following command instead: vllm bench throughput For help with the new command, run: vllm bench throughput --help Alternatively, you can run the new command directly with: python -m vllm.entrypoints.cli.main bench throughput --help """) sys.exit(1)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/backend_request_func.py
benchmarks/backend_request_func.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import io import json import os import sys import time import traceback from dataclasses import dataclass, field import aiohttp import huggingface_hub.constants from tqdm.asyncio import tqdm from transformers import AutoTokenizer, PreTrainedTokenizer, PreTrainedTokenizerFast # NOTE(simon): do not import vLLM here so the benchmark script # can run without vLLM installed. AIOHTTP_TIMEOUT = aiohttp.ClientTimeout(total=6 * 60 * 60) @dataclass class RequestFuncInput: prompt: str api_url: str prompt_len: int output_len: int model: str model_name: str | None = None logprobs: int | None = None extra_body: dict | None = None multi_modal_content: dict | list[dict] | None = None ignore_eos: bool = False language: str | None = None request_id: str | None = None @dataclass class RequestFuncOutput: generated_text: str = "" success: bool = False latency: float = 0.0 output_tokens: int = 0 ttft: float = 0.0 # Time to first token itl: list[float] = field(default_factory=list) # list of inter-token latencies tpot: float = 0.0 # avg next-token latencies prompt_len: int = 0 error: str = "" async def async_request_tgi( request_func_input: RequestFuncInput, pbar: tqdm | None = None, ) -> RequestFuncOutput: api_url = request_func_input.api_url assert api_url.endswith("generate_stream") async with aiohttp.ClientSession( trust_env=True, timeout=AIOHTTP_TIMEOUT ) as session: params = { "max_new_tokens": request_func_input.output_len, "do_sample": True, "temperature": 0.01, # TGI does not accept 0.0 temperature. "top_p": 0.99, # TGI does not accept 1.0 top_p. "truncate": request_func_input.prompt_len, "ignore_eos_token": request_func_input.ignore_eos, } payload = { "inputs": request_func_input.prompt, "parameters": params, } headers = None if request_func_input.request_id: headers = {"x-request-id": request_func_input.request_id} output = RequestFuncOutput() output.prompt_len = request_func_input.prompt_len if request_func_input.ignore_eos: output.output_tokens = request_func_input.output_len else: output.output_tokens = None ttft = 0.0 st = time.perf_counter() most_recent_timestamp = st try: async with session.post( url=api_url, json=payload, headers=headers ) as response: if response.status == 200: async for chunk_bytes in response.content: chunk_bytes = chunk_bytes.strip() if not chunk_bytes: continue chunk_bytes = chunk_bytes.decode("utf-8") # NOTE: Sometimes TGI returns a ping response without # any data, we should skip it. if chunk_bytes.startswith(":"): continue chunk = chunk_bytes.removeprefix("data:") data = json.loads(chunk) timestamp = time.perf_counter() # First token if ttft == 0.0: ttft = time.perf_counter() - st output.ttft = ttft # Decoding phase else: output.itl.append(timestamp - most_recent_timestamp) most_recent_timestamp = timestamp output.latency = most_recent_timestamp - st output.success = True output.generated_text = data["generated_text"] else: output.error = response.reason or "" output.success = False except Exception: output.success = False exc_info = sys.exc_info() output.error = "".join(traceback.format_exception(*exc_info)) if pbar: pbar.update(1) return output async def async_request_trt_llm( request_func_input: RequestFuncInput, pbar: tqdm | None = None, ) -> RequestFuncOutput: api_url = request_func_input.api_url assert api_url.endswith("generate_stream") async with aiohttp.ClientSession( trust_env=True, timeout=AIOHTTP_TIMEOUT ) as session: payload = { "accumulate_tokens": True, "text_input": request_func_input.prompt, "temperature": 0.0, "top_p": 1.0, "max_tokens": request_func_input.output_len, "stream": True, } if request_func_input.ignore_eos: payload["min_length"] = request_func_input.output_len headers = None if request_func_input.request_id: headers = {"x-request-id": request_func_input.request_id} output = RequestFuncOutput() output.prompt_len = request_func_input.prompt_len ttft = 0.0 st = time.perf_counter() most_recent_timestamp = st try: async with session.post( url=api_url, json=payload, headers=headers ) as response: if response.status == 200: async for chunk_bytes in response.content: chunk_bytes = chunk_bytes.strip() if not chunk_bytes: continue chunk = chunk_bytes.decode("utf-8").removeprefix("data:") data = json.loads(chunk) output.generated_text += data["text_output"] timestamp = time.perf_counter() # First token if ttft == 0.0: ttft = timestamp - st output.ttft = ttft # Decoding phase else: output.itl.append(timestamp - most_recent_timestamp) most_recent_timestamp = timestamp output.latency = most_recent_timestamp - st output.success = True else: output.error = response.reason or "" output.success = False except Exception: output.success = False exc_info = sys.exc_info() output.error = "".join(traceback.format_exception(*exc_info)) if pbar: pbar.update(1) return output async def async_request_deepspeed_mii( request_func_input: RequestFuncInput, pbar: tqdm | None = None, ) -> RequestFuncOutput: api_url = request_func_input.api_url assert api_url.endswith(("completions", "profile")), ( "OpenAI Completions API URL must end with 'completions' or 'profile'." ) async with aiohttp.ClientSession( trust_env=True, timeout=AIOHTTP_TIMEOUT ) as session: payload = { "model": request_func_input.model, "prompt": request_func_input.prompt, "max_tokens": request_func_input.output_len, "temperature": 0.01, # deepspeed-mii does not accept 0.0 temp. "top_p": 1.0, } headers = {"Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}"} if request_func_input.request_id: headers["x-request-id"] = request_func_input.request_id output = RequestFuncOutput() output.prompt_len = request_func_input.prompt_len # NOTE: DeepSpeed-MII doesn't support streaming as of Jan 28 2024, # will use 0 as placeholder. # See https://github.com/microsoft/DeepSpeed-MII/pull/311 output.ttft = 0 st = time.perf_counter() try: async with session.post( url=api_url, json=payload, headers=headers ) as response: if response.status == 200: parsed_resp = await response.json() output.latency = time.perf_counter() - st if "choices" in parsed_resp: output.generated_text = parsed_resp["choices"][0]["text"] elif "text" in parsed_resp: output.generated_text = parsed_resp["text"][0] else: output.error = ( "Unexpected response format: " "neither 'choices' nor 'text' found" ) output.success = False output.success = True else: output.error = response.reason or "" output.success = False except Exception: output.success = False exc_info = sys.exc_info() output.error = "".join(traceback.format_exception(*exc_info)) if pbar: pbar.update(1) return output async def async_request_openai_completions( request_func_input: RequestFuncInput, pbar: tqdm | None = None, ) -> RequestFuncOutput: api_url = request_func_input.api_url assert api_url.endswith(("completions", "profile")), ( "OpenAI Completions API URL must end with 'completions' or 'profile'." ) async with aiohttp.ClientSession( trust_env=True, timeout=AIOHTTP_TIMEOUT ) as session: payload = { "model": request_func_input.model_name if request_func_input.model_name else request_func_input.model, "prompt": request_func_input.prompt, "temperature": 0.0, "repetition_penalty": 1.0, "max_tokens": request_func_input.output_len, "logprobs": request_func_input.logprobs, "stream": True, "stream_options": { "include_usage": True, }, } if request_func_input.ignore_eos: payload["ignore_eos"] = request_func_input.ignore_eos if request_func_input.extra_body: payload.update(request_func_input.extra_body) headers = {"Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}"} if request_func_input.request_id: headers["x-request-id"] = request_func_input.request_id output = RequestFuncOutput() output.prompt_len = request_func_input.prompt_len generated_text = "" st = time.perf_counter() most_recent_timestamp = st try: async with session.post( url=api_url, json=payload, headers=headers ) as response: if response.status == 200: first_chunk_received = False async for chunk_bytes in response.content: chunk_bytes = chunk_bytes.strip() if not chunk_bytes: continue chunk = chunk_bytes.decode("utf-8").removeprefix("data: ") if chunk != "[DONE]": data = json.loads(chunk) # NOTE: Some completion API might have a last # usage summary response without a token so we # want to check a token was generated if choices := data.get("choices"): # Note that text could be empty here # e.g. for special tokens text = choices[0].get("text") timestamp = time.perf_counter() # First token if not first_chunk_received: first_chunk_received = True ttft = time.perf_counter() - st output.ttft = ttft # Decoding phase else: output.itl.append(timestamp - most_recent_timestamp) most_recent_timestamp = timestamp generated_text += text or "" if usage := data.get("usage"): output.output_tokens = usage.get("completion_tokens") if first_chunk_received: output.success = True else: output.success = False output.error = ( "Never received a valid chunk to calculate TTFT." "This response will be marked as failed!" ) output.generated_text = generated_text output.latency = most_recent_timestamp - st else: output.error = response.reason or "" output.success = False except Exception: output.success = False exc_info = sys.exc_info() output.error = "".join(traceback.format_exception(*exc_info)) if pbar: pbar.update(1) return output async def async_request_openai_chat_completions( request_func_input: RequestFuncInput, pbar: tqdm | None = None, ) -> RequestFuncOutput: api_url = request_func_input.api_url assert api_url.endswith(("chat/completions", "profile")), ( "OpenAI Chat Completions API URL must end with 'chat/completions'." ) async with aiohttp.ClientSession( trust_env=True, timeout=AIOHTTP_TIMEOUT ) as session: content = [{"type": "text", "text": request_func_input.prompt}] if request_func_input.multi_modal_content: mm_content = request_func_input.multi_modal_content if isinstance(mm_content, list): content.extend(mm_content) elif isinstance(mm_content, dict): content.append(mm_content) else: raise TypeError( "multi_modal_content must be a dict or list[dict] for openai-chat" ) payload = { "model": request_func_input.model_name if request_func_input.model_name else request_func_input.model, "messages": [ {"role": "user", "content": content}, ], "temperature": 0.0, "max_completion_tokens": request_func_input.output_len, "stream": True, "stream_options": { "include_usage": True, }, } if request_func_input.ignore_eos: payload["ignore_eos"] = request_func_input.ignore_eos if request_func_input.extra_body: payload.update(request_func_input.extra_body) headers = { "Content-Type": "application/json", "Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}", } if request_func_input.request_id: headers["x-request-id"] = request_func_input.request_id output = RequestFuncOutput() output.prompt_len = request_func_input.prompt_len generated_text = "" ttft = 0.0 st = time.perf_counter() most_recent_timestamp = st try: async with session.post( url=api_url, json=payload, headers=headers ) as response: if response.status == 200: async for chunk_bytes in response.content: chunk_bytes = chunk_bytes.strip() if not chunk_bytes: continue chunk_bytes = chunk_bytes.decode("utf-8") # NOTE: SSE comments (often used as pings) start with a colon. # These are not JSON data payload and should be skipped. if chunk_bytes.startswith(":"): continue chunk = chunk_bytes.removeprefix("data: ") if chunk != "[DONE]": timestamp = time.perf_counter() data = json.loads(chunk) if choices := data.get("choices"): content = choices[0]["delta"].get("content") # First token if ttft == 0.0: ttft = timestamp - st output.ttft = ttft # Decoding phase else: output.itl.append(timestamp - most_recent_timestamp) generated_text += content or "" elif usage := data.get("usage"): output.output_tokens = usage.get("completion_tokens") most_recent_timestamp = timestamp output.generated_text = generated_text output.success = True output.latency = most_recent_timestamp - st else: output.error = response.reason or "" output.success = False except Exception: output.success = False exc_info = sys.exc_info() output.error = "".join(traceback.format_exception(*exc_info)) if pbar: pbar.update(1) return output async def async_request_openai_audio( request_func_input: RequestFuncInput, pbar: tqdm | None = None, ) -> RequestFuncOutput: # Lazy import without PlaceholderModule to avoid vllm dep. import soundfile api_url = request_func_input.api_url assert api_url.endswith(("transcriptions", "translations")), ( "OpenAI Chat Completions API URL must end with 'transcriptions' " ) "or `translations`." async with aiohttp.ClientSession( trust_env=True, timeout=AIOHTTP_TIMEOUT ) as session: content = [{"type": "text", "text": request_func_input.prompt}] payload = { "model": request_func_input.model_name if request_func_input.model_name else request_func_input.model, "temperature": 0.0, "max_completion_tokens": request_func_input.output_len, "stream": True, "language": "en", # Flattened due to multipart/form-data "stream_include_usage": True, "stream_continuous_usage_stats": True, } if request_func_input.extra_body: payload.update(request_func_input.extra_body) headers = { "Authorization": f"Bearer {os.environ.get('OPENAI_API_KEY')}", } if request_func_input.request_id: headers["x-request-id"] = request_func_input.request_id # Send audio file def to_bytes(y, sr): buffer = io.BytesIO() soundfile.write(buffer, y, sr, format="WAV") buffer.seek(0) return buffer mm_audio = request_func_input.multi_modal_content if not isinstance(mm_audio, dict) or "audio" not in mm_audio: raise TypeError("multi_modal_content must be a dict containing 'audio'") with to_bytes(*mm_audio["audio"]) as f: form = aiohttp.FormData() form.add_field("file", f, content_type="audio/wav") for key, value in payload.items(): form.add_field(key, str(value)) output = RequestFuncOutput() output.prompt_len = request_func_input.prompt_len generated_text = "" ttft = 0.0 st = time.perf_counter() most_recent_timestamp = st try: async with session.post( url=api_url, data=form, headers=headers ) as response: if response.status == 200: async for chunk_bytes in response.content: chunk_bytes = chunk_bytes.strip() if not chunk_bytes: continue chunk = chunk_bytes.decode("utf-8").removeprefix("data: ") if chunk != "[DONE]": timestamp = time.perf_counter() data = json.loads(chunk) if choices := data.get("choices"): content = choices[0]["delta"].get("content") # First token if ttft == 0.0: ttft = timestamp - st output.ttft = ttft # Decoding phase else: output.itl.append( timestamp - most_recent_timestamp ) generated_text += content or "" elif usage := data.get("usage"): output.output_tokens = usage.get( "completion_tokens" ) most_recent_timestamp = timestamp output.generated_text = generated_text output.success = True output.latency = most_recent_timestamp - st else: output.error = response.reason or "" output.success = False except Exception: output.success = False exc_info = sys.exc_info() output.error = "".join(traceback.format_exception(*exc_info)) if pbar: pbar.update(1) return output def get_model(pretrained_model_name_or_path: str) -> str: if os.getenv("VLLM_USE_MODELSCOPE", "False").lower() == "true": from modelscope import snapshot_download from vllm.model_executor.model_loader.weight_utils import get_lock # Use file lock to prevent multiple processes from # downloading the same model weights at the same time. with get_lock(pretrained_model_name_or_path): model_path = snapshot_download( model_id=pretrained_model_name_or_path, local_files_only=huggingface_hub.constants.HF_HUB_OFFLINE, ignore_file_pattern=[".*.pt", ".*.safetensors", ".*.bin"], ) return model_path return pretrained_model_name_or_path def get_tokenizer( pretrained_model_name_or_path: str, tokenizer_mode: str = "auto", trust_remote_code: bool = False, **kwargs, ) -> PreTrainedTokenizer | PreTrainedTokenizerFast: if pretrained_model_name_or_path is not None and not os.path.exists( pretrained_model_name_or_path ): pretrained_model_name_or_path = get_model(pretrained_model_name_or_path) if tokenizer_mode == "slow": if kwargs.get("use_fast", False): raise ValueError("Cannot use the fast tokenizer in slow tokenizer mode.") kwargs["use_fast"] = False if tokenizer_mode == "mistral": try: from vllm.tokenizers.mistral import MistralTokenizer except ImportError as e: raise ImportError( "MistralTokenizer requires vllm package.\n" "Please install it with `pip install vllm` " "to use mistral tokenizer mode." ) from e return MistralTokenizer.from_pretrained(str(pretrained_model_name_or_path)) else: return AutoTokenizer.from_pretrained( pretrained_model_name_or_path, trust_remote_code=trust_remote_code, **kwargs, ) ASYNC_REQUEST_FUNCS = { "tgi": async_request_tgi, "vllm": async_request_openai_completions, "lmdeploy": async_request_openai_completions, "deepspeed-mii": async_request_deepspeed_mii, "openai": async_request_openai_completions, "openai-chat": async_request_openai_chat_completions, "openai-audio": async_request_openai_audio, "tensorrt-llm": async_request_trt_llm, "scalellm": async_request_openai_completions, "sglang": async_request_openai_completions, "llama.cpp": async_request_openai_completions, } OPENAI_COMPATIBLE_BACKENDS = [ k for k, v in ASYNC_REQUEST_FUNCS.items() if v in (async_request_openai_completions, async_request_openai_chat_completions) ]
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_prioritization.py
benchmarks/benchmark_prioritization.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Benchmark offline prioritization.""" import argparse import dataclasses import json import random import time from transformers import AutoTokenizer, PreTrainedTokenizerBase from vllm.engine.arg_utils import EngineArgs from vllm.utils.argparse_utils import FlexibleArgumentParser # Select a equi-probable random priority def get_random_flag(): return 0 if random.random() < 0.5 else 1 def sample_requests( dataset_path: str, num_requests: int, tokenizer: PreTrainedTokenizerBase, fixed_output_len: int | None, ) -> list[tuple[str, int, int, int]]: if fixed_output_len is not None and fixed_output_len < 4: raise ValueError("output_len too small") # Load the dataset. with open(dataset_path) as f: dataset = json.load(f) # Filter out the conversations with less than 2 turns. dataset = [data for data in dataset if len(data["conversations"]) >= 2] # Only keep the first two turns of each conversation. dataset = [ (data["conversations"][0]["value"], data["conversations"][1]["value"]) for data in dataset ] # Shuffle the dataset. random.shuffle(dataset) # Filter out sequences that are too long or too short filtered_dataset: list[tuple[str, int, int]] = [] for i in range(len(dataset)): if len(filtered_dataset) == num_requests: break # Tokenize the prompts and completions. prompt = dataset[i][0] prompt_token_ids = tokenizer(prompt).input_ids completion = dataset[i][1] completion_token_ids = tokenizer(completion).input_ids prompt_len = len(prompt_token_ids) output_len = ( len(completion_token_ids) if fixed_output_len is None else fixed_output_len ) if prompt_len < 4 or output_len < 4: # Prune too short sequences. continue if prompt_len > 1024 or prompt_len + output_len > 2048: # Prune too long sequences. continue priority = get_random_flag() filtered_dataset.append((prompt, prompt_len, output_len, priority)) return filtered_dataset def run_vllm( requests: list[tuple[str, int, int]], n: int, engine_args: EngineArgs, disable_detokenize: bool = False, ) -> float: from vllm import LLM, SamplingParams llm = LLM(**dataclasses.asdict(engine_args)) assert all( llm.llm_engine.model_config.max_model_len >= (request[1] + request[2]) for request in requests ), ( "Please ensure that max_model_len is greater than the sum of" " input_len and output_len for all requests." ) # Add the requests to the engine. prompts = [] sampling_params = [] priority = [] for prompt, _, output_len, _priority in requests: prompts.append(prompt) priority.append(_priority) sampling_params.append( SamplingParams( n=n, temperature=1.0, top_p=1.0, ignore_eos=True, max_tokens=output_len, detokenize=not disable_detokenize, ) ) start = time.perf_counter() llm.generate(prompts, sampling_params, priority=priority, use_tqdm=True) end = time.perf_counter() return end - start def main(args: argparse.Namespace): print(args) random.seed(args.seed) # Sample the requests. tokenizer = AutoTokenizer.from_pretrained( args.tokenizer, trust_remote_code=args.trust_remote_code ) if args.dataset is None: # Synthesize a prompt with the given input length. prompt = "hi" * (args.input_len - 1) requests = [ (prompt, args.input_len, args.output_len, get_random_flag()) for _ in range(args.num_prompts) ] else: requests = sample_requests( args.dataset, args.num_prompts, tokenizer, args.output_len ) if args.backend == "vllm": elapsed_time = run_vllm( requests, args.n, EngineArgs.from_cli_args(args), args.disable_detokenize ) else: raise ValueError(f"Unknown backend: {args.backend}") total_num_tokens = sum( prompt_len + output_len for _, prompt_len, output_len, priority in requests ) print( f"Throughput: {len(requests) / elapsed_time:.2f} requests/s, " f"{total_num_tokens / elapsed_time:.2f} tokens/s" ) # Output JSON results if specified if args.output_json: results = { "elapsed_time": elapsed_time, "num_requests": len(requests), "total_num_tokens": total_num_tokens, "requests_per_second": len(requests) / elapsed_time, "tokens_per_second": total_num_tokens / elapsed_time, } with open(args.output_json, "w") as f: json.dump(results, f, indent=4) def create_argument_parser(): parser = FlexibleArgumentParser(description="Benchmark the throughput.") parser.add_argument( "--backend", type=str, choices=["vllm", "hf", "mii"], default="vllm" ) parser.add_argument( "--dataset", type=str, default=None, help="Path to the dataset." ) parser.add_argument( "--input-len", type=int, default=None, help="Input prompt length for each request", ) parser.add_argument( "--output-len", type=int, default=None, help="Output length for each request. Overrides the " "output length from the dataset.", ) parser.add_argument( "--n", type=int, default=1, help="Number of generated sequences per prompt." ) parser.add_argument( "--num-prompts", type=int, default=200, help="Number of prompts to process." ) parser.add_argument( "--output-json", type=str, default=None, help="Path to save the throughput results in JSON format.", ) parser.add_argument( "--disable-detokenize", action="store_true", help=( "Do not detokenize responses (i.e. do not include " "detokenization time in the latency measurement)" ), ) parser = EngineArgs.add_cli_args(parser) return parser if __name__ == "__main__": parser = create_argument_parser() args = parser.parse_args() if args.tokenizer is None: args.tokenizer = args.model if args.dataset is None: assert args.input_len is not None assert args.output_len is not None else: assert args.input_len is None main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_block_pool.py
benchmarks/benchmark_block_pool.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import gc from benchmark_utils import TimeCollector from tabulate import tabulate from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.v1.core.block_pool import BlockPool def main(args): rows = [] for allocate_block in args.allocate_blocks: # Enforce a GC collect ahead to minimize the impact among runs gc.collect() block_pool = BlockPool(num_gpu_blocks=args.num_gpu_blocks, enable_caching=True) get_blocks_times = TimeCollector(TimeCollector.US) free_blocks_times = TimeCollector(TimeCollector.US) for _ in range(args.num_iteration): with get_blocks_times: blocks = block_pool.get_new_blocks(allocate_block) with free_blocks_times: block_pool.free_blocks(blocks) rows.append( [get_blocks_times.cnt, args.num_gpu_blocks, allocate_block] + get_blocks_times.dump_avg_max() + free_blocks_times.dump_avg_max() ) print( tabulate( rows, headers=[ "Iterations", "Total\nBlocks", "Allocated\nBlocks", "Get Blocks\nAvg (us)", "Get Blocks\nMax (us)", "Free Blocks\nAvg (us)", "Free Blocks\nMax (us)", ], tablefmt="grid", floatfmt=".3f", ) ) def invoke_main() -> None: parser = FlexibleArgumentParser( description="Benchmark the performance of BlockPool for KV Cache." ) parser.add_argument("--num-gpu-blocks", type=int, default=100000) parser.add_argument( "--num-iteration", type=int, default=1000, help="Number of iterations to run to stabilize final data readings", ) parser.add_argument( "--allocate-blocks", type=int, nargs="*", default=[10, 50, 100, 500, 1000], help="Number of blocks to allocate", ) args = parser.parse_args() main(args) if __name__ == "__main__": invoke_main() # pragma: no cover
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_prefix_caching.py
benchmarks/benchmark_prefix_caching.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Benchmark the efficiency of prefix caching. This script allows you to benchmark the performance of a model with and without prefix caching using either fixed prompts or prompts sampled from the ShareGPT dataset. Fixed example usage: python benchmark_prefix_caching.py \ --model meta-llama/Llama-2-7b-chat-hf \ --enable-prefix-caching \ --num-prompts 1 \ --repeat-count 100 \ --input-length-range 128:256 ShareGPT example usage: # This command samples 20 prompts with input lengths # between 128 and 256 tokens from the ShareGPT dataset, # then replicates each prompt 5 times. python benchmark_prefix_caching.py \ --model meta-llama/Llama-2-7b-chat-hf \ --dataset-path /path/to/ShareGPT_V3_unfiltered_cleaned_split.json \ --enable-prefix-caching \ --num-prompts 20 \ --repeat-count 5 \ --input-length-range 128:256 """ import dataclasses import json import random import time from transformers import PreTrainedTokenizerBase from vllm import LLM, SamplingParams from vllm.engine.arg_utils import EngineArgs from vllm.utils.argparse_utils import FlexibleArgumentParser try: from vllm.tokenizers import get_tokenizer except ImportError: from backend_request_func import get_tokenizer PROMPT = "You are a helpful assistant in recognizes the content of tables in markdown format. Here is a table as fellows. You need to answer my question about the table.\n# Table\n|Opening|Opening|Sl. No.|Film|Cast|Director|Music Director|Notes|\n|----|----|----|----|----|----|----|----|\n|J A N|9|1|Agni Pushpam|Jayabharathi, Kamalahasan|Jeassy|M. K. Arjunan||\n|J A N|16|2|Priyamvada|Mohan Sharma, Lakshmi, KPAC Lalitha|K. S. Sethumadhavan|V. Dakshinamoorthy||\n|J A N|23|3|Yakshagaanam|Madhu, Sheela|Sheela|M. S. Viswanathan||\n|J A N|30|4|Paalkkadal|Sheela, Sharada|T. K. Prasad|A. T. Ummer||\n|F E B|5|5|Amma|Madhu, Srividya|M. Krishnan Nair|M. K. Arjunan||\n|F E B|13|6|Appooppan|Thikkurissi Sukumaran Nair, Kamal Haasan|P. Bhaskaran|M. S. Baburaj||\n|F E B|20|7|Srishti|Chowalloor Krishnankutty, Ravi Alummoodu|K. T. Muhammad|M. S. Baburaj||\n|F E B|20|8|Vanadevatha|Prem Nazir, Madhubala|Yusufali Kechery|G. Devarajan||\n|F E B|27|9|Samasya|Madhu, Kamalahaasan|K. Thankappan|Shyam||\n|F E B|27|10|Yudhabhoomi|K. P. Ummer, Vidhubala|Crossbelt Mani|R. K. Shekhar||\n|M A R|5|11|Seemantha Puthran|Prem Nazir, Jayabharathi|A. B. Raj|M. K. Arjunan||\n|M A R|12|12|Swapnadanam|Rani Chandra, Dr. Mohandas|K. G. George|Bhaskar Chandavarkar||\n|M A R|19|13|Thulavarsham|Prem Nazir, sreedevi, Sudheer|N. Sankaran Nair|V. Dakshinamoorthy||\n|M A R|20|14|Aruthu|Kaviyoor Ponnamma, Kamalahasan|Ravi|G. Devarajan||\n|M A R|26|15|Swimming Pool|Kamal Haasan, M. G. Soman|J. Sasikumar|M. K. Arjunan||\n\n# Question\nWhat' s the content in the (1,1) cells\n" # noqa: E501 def test_prefix(llm=None, sampling_params=None, prompts=None): start_time = time.time() llm.generate(prompts, sampling_params=sampling_params) end_time = time.time() print(f"cost time {end_time - start_time}") @dataclasses.dataclass class Request: prompt: str prompt_len: int output_len: int def sample_tokens(tokenizer: PreTrainedTokenizerBase, length: int) -> list[int]: vocab = tokenizer.get_vocab() all_special_ids = set(tokenizer.all_special_ids) # Remove the special tokens. return random.choices( [v for v in vocab.values() if v not in all_special_ids], k=length, ) def sample_requests_from_dataset( dataset_path: str, num_requests: int, tokenizer: PreTrainedTokenizerBase, input_length_range: tuple[int, int], fixed_output_len: int | None, ) -> list[Request]: if fixed_output_len is not None and fixed_output_len < 4: raise ValueError("output_len too small") # Load the dataset. with open(dataset_path) as f: dataset = json.load(f) # Filter out the conversations with less than 2 turns. dataset = [data for data in dataset if len(data["conversations"]) >= 2] # Only keep the first two turns of each conversation. dataset = [ (data["conversations"][0]["value"], data["conversations"][1]["value"]) for data in dataset ] # Shuffle the dataset. random.shuffle(dataset) min_len, max_len = input_length_range assert min_len >= 0 and max_len >= min_len, "input_length_range too small" # Filter out sequences that are too long or too short filtered_requests: list[Request] = [] for i in range(len(dataset)): if len(filtered_requests) == num_requests: break # Tokenize the prompts and completions. prompt_token_ids = tokenizer(dataset[i][0]).input_ids prompt = tokenizer.decode(prompt_token_ids) completion = dataset[i][1] completion_token_ids = tokenizer(completion).input_ids prompt_len = len(prompt_token_ids) output_len = ( len(completion_token_ids) if fixed_output_len is None else fixed_output_len ) if min_len <= prompt_len <= max_len: filtered_requests.append(Request(prompt, prompt_len, output_len)) return filtered_requests def sample_requests_from_random( num_requests: int, tokenizer: PreTrainedTokenizerBase, input_length_range: tuple[int, int], fixed_output_len: int | None, prefix_len: int, ) -> list[Request]: requests = [] prefix_token_ids = sample_tokens(tokenizer, prefix_len) min_len, max_len = input_length_range for i in range(num_requests): unique_part_token_ids = sample_tokens( tokenizer, random.randint(min_len - prefix_len, max_len - prefix_len) ) prompt_token_ids = prefix_token_ids + unique_part_token_ids prompt = tokenizer.decode(prompt_token_ids) prompt_len = len(prompt_token_ids) assert min_len <= prompt_len <= max_len, ( f"prompt_len {prompt_len} out of range {min_len}:{max_len}" ) requests.append(Request(prompt, prompt_len, fixed_output_len)) return requests def repeat_and_sort_requests( requests: list[Request], repeat_count: int, sort: bool = False ) -> list[str]: repeated_requests = requests * repeat_count if sort: repeated_requests.sort(key=lambda x: x[1]) else: random.shuffle(repeated_requests) return [req.prompt for req in repeated_requests] def main(args): tokenizer = get_tokenizer(args.model, trust_remote_code=True) input_length_range = tuple(map(int, args.input_length_range.split(":"))) random.seed(args.seed) if args.dataset_path is not None: if args.prefix_len > 0: raise ValueError( "prefix-len is not supported when dataset-path is provided." ) print(f"Start to sample {args.num_prompts} prompts from {args.dataset_path}") filtered_requests = sample_requests_from_dataset( dataset_path=args.dataset_path, num_requests=args.num_prompts, tokenizer=tokenizer, input_length_range=input_length_range, fixed_output_len=args.output_len, ) else: print(f"Start to sample {args.num_prompts} prompts from random") filtered_requests = sample_requests_from_random( num_requests=args.num_prompts, tokenizer=tokenizer, input_length_range=input_length_range, fixed_output_len=args.output_len, prefix_len=args.prefix_len, ) # Print some helpful stats of the requests. print(f"Sampled {len(filtered_requests)} requests.") prompt_lens = [req.prompt_len for req in filtered_requests] print(f"Average input length: {sum(prompt_lens) / len(prompt_lens)}") print(f"P50 input length: {sorted(prompt_lens)[len(prompt_lens) // 2]}") print(f"Min Prompt Length: {min(prompt_lens)}") print(f"Max Prompt Length: {max(prompt_lens)}") engine_args = EngineArgs.from_cli_args(args) llm = LLM(**dataclasses.asdict(engine_args)) sampling_params = SamplingParams( temperature=0, max_tokens=args.output_len, detokenize=not args.disable_detokenize, ) print("Testing filtered requests") prompts = repeat_and_sort_requests( filtered_requests, repeat_count=args.repeat_count, sort=args.sort ) print("------start generating------") test_prefix( llm=llm, prompts=prompts, sampling_params=sampling_params, ) def create_argument_parser(): parser = FlexibleArgumentParser( description="Benchmark the performance with or without " "automatic prefix caching." ) parser.add_argument( "--dataset-path", type=str, default=None, help="Path to the dataset." ) parser.add_argument("--output-len", type=int, default=10) parser.add_argument( "--num-prompts", type=int, required=True, help="Number of the prompts sampled from dataset", ) parser.add_argument( "--repeat-count", type=int, default=1, help="Number of times to repeat each prompt", ) parser.add_argument( "--sort", action="store_true", help="Sort prompts by input length" ) parser.add_argument( "--input-length-range", type=str, required=True, help="Range of input lengths for sampling prompts," 'specified as "min:max" (e.g., "128:256").', ) parser.add_argument( "--prefix-len", type=int, default=0, help="Specifies the length of a common prefix to be " "added to the input prompt. The input-length-range will " "subtract this length when filtering prompts. Only used " "when dataset-path is not provided.", ) parser.add_argument( "--disable-detokenize", action="store_true", help=( "Do not detokenize responses (i.e. do not include " "detokenization time in the latency measurement)" ), ) parser = EngineArgs.add_cli_args(parser) return parser if __name__ == "__main__": parser = create_argument_parser() args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_prefix_block_hash.py
benchmarks/benchmark_prefix_block_hash.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Simple benchmark to compare prefix-cache block hashing algorithms. Example: python benchmark_prefix_block_hash.py --num-blocks 20000 --block-size 32 """ from __future__ import annotations import argparse import random import statistics import sys import time from collections.abc import Callable, Iterable, Sequence from vllm.utils.hashing import get_hash_fn_by_name from vllm.v1.core.kv_cache_utils import BlockHash, hash_block_tokens, init_none_hash SUPPORTED_ALGOS = ("sha256", "sha256_cbor", "xxhash", "xxhash_cbor") def _generate_blocks( num_blocks: int, block_size: int, vocab_size: int, seed: int ) -> list[list[int]]: rng = random.Random(seed) return [ [rng.randrange(vocab_size) for _ in range(block_size)] for _ in range(num_blocks) ] def _hash_all_blocks( hash_fn: Callable[[object], bytes], blocks: Iterable[Sequence[int]], ) -> float: parent_hash: BlockHash | None = None start = time.perf_counter() for block in blocks: parent_hash = hash_block_tokens(hash_fn, parent_hash, block, extra_keys=None) end = time.perf_counter() return end - start def _benchmark( hash_algo: str, blocks: list[list[int]], trials: int, ) -> tuple[float, float, float] | None: try: hash_fn = get_hash_fn_by_name(hash_algo) init_none_hash(hash_fn) timings = [_hash_all_blocks(hash_fn, blocks) for _ in range(trials)] except ModuleNotFoundError as exc: print(f"Skipping {hash_algo}: {exc}", file=sys.stderr) return None avg = statistics.mean(timings) best = min(timings) # throughput: tokens / second tokens_hashed = len(blocks) * len(blocks[0]) throughput = tokens_hashed / best return avg, best, throughput def main() -> None: parser = argparse.ArgumentParser(description=__doc__) parser.add_argument("--num-blocks", type=int, default=10000, help="Block count.") parser.add_argument("--block-size", type=int, default=32, help="Tokens per block.") parser.add_argument( "--vocab-size", type=int, default=32000, help="Token id range [0, vocab_size)." ) parser.add_argument("--seed", type=int, default=0, help="Random seed.") parser.add_argument( "--trials", type=int, default=5, help="Number of timed trials per algorithm." ) parser.add_argument( "--algorithms", nargs="+", default=SUPPORTED_ALGOS, choices=SUPPORTED_ALGOS, help="Hash algorithms to benchmark.", ) args = parser.parse_args() blocks = _generate_blocks( args.num_blocks, args.block_size, args.vocab_size, args.seed ) print( f"Benchmarking {len(args.algorithms)} algorithms on " f"{args.num_blocks} blocks (block size={args.block_size})." ) for algo in args.algorithms: result = _benchmark(algo, blocks, args.trials) if result is None: continue avg, best, throughput = result print( f"{algo:14s} avg: {avg:.6f}s best: {best:.6f}s " f"throughput: {throughput / 1e6:.2f}M tokens/s" ) if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_serving.py
benchmarks/benchmark_serving.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import sys if __name__ == "__main__": print("""DEPRECATED: This script has been moved to the vLLM CLI. Please use the following command instead: vllm bench serve For help with the new command, run: vllm bench serve --help Alternatively, you can run the new command directly with: python -m vllm.entrypoints.cli.main bench serve --help """) sys.exit(1)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_serving_structured_output.py
benchmarks/benchmark_serving_structured_output.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project r"""Benchmark online serving throughput with structured outputs. On the server side, run one of the following commands: (vLLM OpenAI API server) vllm serve <your_model> On the client side, run: python benchmarks/benchmark_serving_structured_output.py \ --backend <backend> \ --model <your_model> \ --dataset json \ --structured-output-ratio 1.0 \ --request-rate 10 \ --num-prompts 1000 when using tgi backend, add --endpoint /generate_stream to the end of the command above. """ import argparse import asyncio import copy import dataclasses import json import os import random import time import uuid import warnings from collections.abc import AsyncGenerator from contextlib import nullcontext from dataclasses import dataclass import datasets import numpy as np import pandas as pd from backend_request_func import ( ASYNC_REQUEST_FUNCS, RequestFuncInput, RequestFuncOutput, ) from tqdm.asyncio import tqdm from transformers import PreTrainedTokenizerBase try: from vllm.tokenizers import get_tokenizer except ImportError: from backend_request_func import get_tokenizer try: from vllm.utils.argparse_utils import FlexibleArgumentParser except ImportError: from argparse import ArgumentParser as FlexibleArgumentParser from vllm.v1.structured_output.backend_xgrammar import ( has_xgrammar_unsupported_json_features, ) MILLISECONDS_TO_SECONDS_CONVERSION = 1000 @dataclass class BenchmarkMetrics: completed: int total_input: int total_output: int request_throughput: float request_goodput: float output_throughput: float total_token_throughput: float mean_ttft_ms: float median_ttft_ms: float std_ttft_ms: float percentiles_ttft_ms: list[tuple[float, float]] mean_tpot_ms: float median_tpot_ms: float std_tpot_ms: float percentiles_tpot_ms: list[tuple[float, float]] mean_itl_ms: float median_itl_ms: float std_itl_ms: float percentiles_itl_ms: list[tuple[float, float]] # E2EL stands for end-to-end latency per request. # It is the time taken on the client side from sending # a request to receiving a complete response. mean_e2el_ms: float median_e2el_ms: float std_e2el_ms: float percentiles_e2el_ms: list[tuple[float, float]] @dataclasses.dataclass class SampleRequest: """A class representing a single inference request for benchmarking. Attributes: prompt: The input text prompt for the model. multi_modal_data: Optional dictionary containing multi-modal data (e.g. images). prompt_len: The length of the prompt in tokens. expected_output_len: The expected length of the output in tokens. """ prompt: str prompt_len: int expected_output_len: int schema: dict structure_type: str completion: str = None def sample_requests( tokenizer: PreTrainedTokenizerBase, args: argparse.Namespace ) -> list[SampleRequest]: if args.dataset == "json" or args.dataset == "json-unique": if args.json_schema_path is None: dir_path = os.path.dirname(os.path.realpath(__file__)) args.json_schema_path = os.path.join( dir_path, "structured_schemas", "structured_schema_1.json" ) json_schemas = [] with open(args.json_schema_path) as f: schema = json.load(f) if args.dataset == "json-unique": json_schemas = [copy.deepcopy(schema) for _ in range(args.num_prompts)] for i in range(len(json_schemas)): if "properties" not in json_schemas[i]: json_schemas[i]["properties"] = {} json_schemas[i]["properties"][f"__optional_field_{uuid.uuid4()}"] = { "type": "string", "description": "An unique optional field to avoid cached schemas", } else: json_schemas = [schema] * args.num_prompts def gen_prompt(index: int): return f"Generate an example of a brief user profile given the following schema: {json.dumps(get_schema(index))}" # noqa: E501 def get_schema(index: int): return json_schemas[index % len(json_schemas)] requests = [ SampleRequest( prompt=gen_prompt(i), prompt_len=len(tokenizer(gen_prompt(i)).input_ids), expected_output_len=args.output_len, schema=get_schema(i), structure_type=args.structure_type, ) for i in range(args.num_prompts) ] elif args.dataset == "grammar": schema = """ root ::= select_statement select_statement ::= "SELECT " column " from " table " where " condition column ::= "col_1 " | "col_2 " table ::= "table_1 " | "table_2 " condition ::= column "= " number number ::= "1 " | "2 " """ prompt = "Generate an SQL query to show the 'username' \ and 'email' from the 'users' table." input_len = len(tokenizer(prompt).input_ids) print(f"Input length of the prompt: {input_len} tokens") requests = [ SampleRequest( prompt=prompt, prompt_len=input_len, expected_output_len=args.output_len, schema=schema, structure_type=args.structure_type, ) for _ in range(args.num_prompts) ] elif args.dataset == "regex": regex = r"\w+@\w+\.com\n" args.regex = regex prompt = "Generate an email address for Alan Turing, \ who works in Enigma. End in .com and new line. \ Example result: alan.turing@enigma.com\n" input_len = len(tokenizer(prompt).input_ids) print(f"Input length of the prompt: {input_len} tokens") requests = [ SampleRequest( prompt=prompt, prompt_len=input_len, expected_output_len=args.output_len, schema=regex, structure_type=args.structure_type, ) for _ in range(args.num_prompts) ] elif args.dataset == "choice": choice = ["Positive", "Negative"] args.choice = choice prompt = "Classify this sentiment: vLLM is wonderful!" input_len = len(tokenizer(prompt).input_ids) print(f"Input length of the prompt: {input_len} tokens") requests = [ SampleRequest( prompt=prompt, prompt_len=input_len, expected_output_len=args.output_len, schema=choice, structure_type=args.structure_type, ) for _ in range(args.num_prompts) ] elif args.dataset == "xgrammar_bench": requests: list[SampleRequest] = [] dataset = datasets.load_dataset("NousResearch/json-mode-eval", split="train") full_dataset_len = len(dataset) def _filter_func(item): import json schema = json.loads(item["schema"]) return not has_xgrammar_unsupported_json_features(schema) dataset = dataset.filter(_filter_func) num_filtered_out = full_dataset_len - len(dataset) print( f"dataset has {len(dataset)} entries after filtering " f"out {num_filtered_out} entries with unsupported features" ) len_dataset = len(dataset) for data_point_idx in range(args.num_prompts): idx = data_point_idx while idx >= len_dataset: idx -= len_dataset schema = dataset["schema"][idx] prompt = tokenizer.apply_chat_template( dataset["prompt"][idx], tokenize=False, add_generation_prompt=True ) input_len = len(tokenizer(prompt).input_ids) completion = dataset["completion"][idx] requests.append( SampleRequest( prompt=prompt, prompt_len=input_len, expected_output_len=args.output_len, schema=schema, structure_type=args.structure_type, completion=completion, ) ) return requests async def get_request( input_requests: list[SampleRequest], request_rate: float, burstiness: float = 1.0, ) -> AsyncGenerator[tuple[int, SampleRequest], None]: """ Asynchronously generates requests at a specified rate with OPTIONAL burstiness. Args: input_requests: A list of input requests, each represented as a tuple. request_rate: The rate at which requests are generated (requests/s). burstiness (optional): The burstiness factor of the request generation. Only takes effect when request_rate is not inf. Default value is 1, which follows a Poisson process. Otherwise, the request intervals follow a gamma distribution. A lower burstiness value (0 < burstiness < 1) results in more bursty requests, while a higher burstiness value (burstiness > 1) results in a more uniform arrival of requests. """ input_requests = iter(input_requests) # Calculate scale parameter theta to maintain the desired request_rate. assert burstiness > 0, ( f"A positive burstiness factor is expected, but given {burstiness}." ) theta = 1.0 / (request_rate * burstiness) for i, request in enumerate(input_requests): yield i, request if request_rate == float("inf"): # If the request rate is infinity, then we don't need to wait. continue # Sample the request interval from the gamma distribution. # If burstiness is 1, it follows exponential distribution. interval = np.random.gamma(shape=burstiness, scale=theta) # The next request will be sent after the interval. await asyncio.sleep(interval) def calculate_metrics( input_requests: list[tuple[str, int, int]], outputs: list[RequestFuncOutput], dur_s: float, tokenizer: PreTrainedTokenizerBase, selected_percentile_metrics: list[str], selected_percentiles: list[float], goodput_config_dict: dict[str, float] | None = None, ) -> tuple[BenchmarkMetrics, list[int]]: actual_output_lens: list[int] = [] total_input = 0 completed = 0 good_completed = 0 itls: list[float] = [] tpots: list[float] = [] all_tpots: list[float] = [] ttfts: list[float] = [] e2els: list[float] = [] for i in range(len(outputs)): if outputs[i].success: # We use the tokenizer to count the number of output tokens for all # serving backends instead of looking at len(outputs[i].itl) since # multiple output tokens may be bundled together # Note : this may inflate the output token count slightly output_len = len( tokenizer(outputs[i].generated_text, add_special_tokens=False).input_ids ) actual_output_lens.append(output_len) total_input += input_requests[i].prompt_len tpot = 0 if output_len > 1: latency_minus_ttft = outputs[i].latency - outputs[i].ttft tpot = latency_minus_ttft / (output_len - 1) tpots.append(tpot) outputs[i].tpot = tpot # Note: if output_len <= 1, we regard tpot as 0 for goodput all_tpots.append(tpot) itls += outputs[i].itl ttfts.append(outputs[i].ttft) e2els.append(outputs[i].latency) completed += 1 else: actual_output_lens.append(0) if goodput_config_dict: valid_metrics = [] slo_values = [] if "ttft" in goodput_config_dict: valid_metrics.append(ttfts) slo_values.append( goodput_config_dict["ttft"] / MILLISECONDS_TO_SECONDS_CONVERSION ) if "tpot" in goodput_config_dict: valid_metrics.append(all_tpots) slo_values.append( goodput_config_dict["tpot"] / MILLISECONDS_TO_SECONDS_CONVERSION ) if "e2el" in goodput_config_dict: valid_metrics.append(e2els) slo_values.append( goodput_config_dict["e2el"] / MILLISECONDS_TO_SECONDS_CONVERSION ) for req_metric in zip(*valid_metrics): is_good_req = all([s >= r for s, r in zip(slo_values, req_metric)]) if is_good_req: good_completed += 1 if completed == 0: warnings.warn( "All requests failed. This is likely due to a misconfiguration " "on the benchmark arguments.", stacklevel=2, ) metrics = BenchmarkMetrics( completed=completed, total_input=total_input, total_output=sum(actual_output_lens), request_throughput=completed / dur_s, request_goodput=good_completed / dur_s, output_throughput=sum(actual_output_lens) / dur_s, total_token_throughput=(total_input + sum(actual_output_lens)) / dur_s, mean_ttft_ms=np.mean(ttfts or 0) * 1000, # ttfts is empty if streaming is not supported by backend std_ttft_ms=np.std(ttfts or 0) * 1000, median_ttft_ms=np.median(ttfts or 0) * 1000, percentiles_ttft_ms=[ (p, np.percentile(ttfts or 0, p) * 1000) for p in selected_percentiles ], mean_tpot_ms=np.mean(tpots or 0) * 1000, std_tpot_ms=np.std(tpots or 0) * 1000, median_tpot_ms=np.median(tpots or 0) * 1000, percentiles_tpot_ms=[ (p, np.percentile(tpots or 0, p) * 1000) for p in selected_percentiles ], mean_itl_ms=np.mean(itls or 0) * 1000, std_itl_ms=np.std(itls or 0) * 1000, median_itl_ms=np.median(itls or 0) * 1000, percentiles_itl_ms=[ (p, np.percentile(itls or 0, p) * 1000) for p in selected_percentiles ], mean_e2el_ms=np.mean(e2els or 0) * 1000, std_e2el_ms=np.std(e2els or 0) * 1000, median_e2el_ms=np.median(e2els or 0) * 1000, percentiles_e2el_ms=[ (p, np.percentile(e2els or 0, p) * 1000) for p in selected_percentiles ], ) return metrics, actual_output_lens async def benchmark( backend: str, api_url: str, base_url: str, model_id: str, tokenizer: PreTrainedTokenizerBase, input_requests: list[SampleRequest], request_rate: float, burstiness: float, disable_tqdm: bool, profile: bool, selected_percentile_metrics: list[str], selected_percentiles: list[str], ignore_eos: bool, max_concurrency: int | None, structured_output_ratio: float, goodput_config_dict: dict[str, float] | None = None, ): if backend in ASYNC_REQUEST_FUNCS: request_func = ASYNC_REQUEST_FUNCS[backend] else: raise ValueError(f"Unknown backend: {backend}") def prepare_extra_body(request) -> dict: extra_body = {} # Add the schema to the extra_body extra_body["structured_outputs"] = {} extra_body["structured_outputs"][request.structure_type] = request.schema return extra_body print("Starting initial single prompt test run...") structured_output_req_idx = random.sample( range(len(input_requests)), int(len(input_requests) * structured_output_ratio) ) test_request = input_requests[0] test_req_extra_body = ( prepare_extra_body(test_request) if 0 in structured_output_req_idx else None ) test_input = RequestFuncInput( model=model_id, prompt=test_request.prompt, api_url=api_url, prompt_len=test_request.prompt_len, output_len=test_request.expected_output_len, ignore_eos=ignore_eos, extra_body=test_req_extra_body, ) test_output = await request_func(request_func_input=test_input) if not test_output.success: raise ValueError( "Initial test run failed - Please make sure benchmark arguments " f"are correctly specified. Error: {test_output.error}" ) else: print("Initial test run completed. Starting main benchmark run...") if profile: print("Starting profiler...") profile_input = RequestFuncInput( model=model_id, prompt=test_request.prompt, api_url=base_url + "/start_profile", prompt_len=test_request.prompt_len, output_len=test_request.expected_output_len, ignore_eos=ignore_eos, extra_body=test_req_extra_body, ) profile_output = await request_func(request_func_input=profile_input) if profile_output.success: print("Profiler started") distribution = "Poisson process" if burstiness == 1.0 else "Gamma distribution" print(f"Traffic request rate: {request_rate}") print(f"Burstiness factor: {burstiness} ({distribution})") print(f"Maximum request concurrency: {max_concurrency}") pbar = None if disable_tqdm else tqdm(total=len(input_requests)) semaphore = asyncio.Semaphore(max_concurrency) if max_concurrency else nullcontext() async def limited_request_func(request_func_input, pbar): async with semaphore: return await request_func(request_func_input=request_func_input, pbar=pbar) benchmark_start_time = time.perf_counter() tasks: list[asyncio.Task] = [] expected: list[str] = [] async for i, request in get_request(input_requests, request_rate, burstiness): extra_body = ( prepare_extra_body(request) if i in structured_output_req_idx else None ) request_func_input = RequestFuncInput( model=model_id, prompt=request.prompt, api_url=api_url, prompt_len=request.prompt_len, output_len=request.expected_output_len, ignore_eos=ignore_eos, extra_body=extra_body, ) expected.append(request.completion) tasks.append( asyncio.create_task( limited_request_func(request_func_input=request_func_input, pbar=pbar) ) ) outputs: list[RequestFuncOutput] = await asyncio.gather(*tasks) if pbar is not None: pbar.close() benchmark_duration = time.perf_counter() - benchmark_start_time metrics, actual_output_lens = calculate_metrics( input_requests=input_requests, outputs=outputs, dur_s=benchmark_duration, tokenizer=tokenizer, selected_percentile_metrics=selected_percentile_metrics, selected_percentiles=selected_percentiles, goodput_config_dict=goodput_config_dict, ) print("{s:{c}^{n}}".format(s=" Serving Benchmark Result ", n=50, c="=")) print("{:<40} {:<10}".format("Successful requests:", metrics.completed)) if max_concurrency is not None: print("{:<40} {:<10}".format("Maximum request concurrency:", max_concurrency)) if request_rate != float("inf"): print("{:<40} {:<10.2f}".format("Request rate configured (RPS):", request_rate)) print("{:<40} {:<10.2f}".format("Benchmark duration (s):", benchmark_duration)) print("{:<40} {:<10}".format("Total input tokens:", metrics.total_input)) print("{:<40} {:<10}".format("Total generated tokens:", metrics.total_output)) print( "{:<40} {:<10.2f}".format( "Request throughput (req/s):", metrics.request_throughput ) ) if goodput_config_dict: print( "{:<40} {:<10.2f}".format( "Request goodput (req/s):", metrics.request_goodput ) ) print( "{:<40} {:<10.2f}".format( "Output token throughput (tok/s):", metrics.output_throughput ) ) print( "{:<40} {:<10.2f}".format( "Total token throughput (tok/s):", metrics.total_token_throughput ) ) result = { "duration": benchmark_duration, "completed": metrics.completed, "total_input_tokens": metrics.total_input, "total_output_tokens": metrics.total_output, "request_throughput": metrics.request_throughput, "output_throughput": metrics.output_throughput, "total_token_throughput": metrics.total_token_throughput, "ttft_description": pd.Series([output.ttft for output in outputs]) .describe() .to_dict(), "tpot_description": pd.Series([output.tpot for output in outputs]) .describe() .to_dict(), "input_lens": [output.prompt_len for output in outputs], "output_lens": actual_output_lens, "ttfts": [output.ttft for output in outputs], "itls": [output.itl for output in outputs], "errors": [output.error for output in outputs], } ret = [ {"generated": output.generated_text, "expected": gt} for output, gt in zip(outputs, expected) ] def process_one_metric( # E.g., "ttft" metric_attribute_name: str, # E.g., "TTFT" metric_name: str, # E.g., "Time to First Token" metric_header: str, ): # This function prints and adds statistics of the specified # metric. if metric_attribute_name not in selected_percentile_metrics: return print("{s:{c}^{n}}".format(s=metric_header, n=50, c="-")) print( "{:<40} {:<10.2f}".format( f"Mean {metric_name} (ms):", getattr(metrics, f"mean_{metric_attribute_name}_ms"), ) ) print( "{:<40} {:<10.2f}".format( f"Median {metric_name} (ms):", getattr(metrics, f"median_{metric_attribute_name}_ms"), ) ) result[f"mean_{metric_attribute_name}_ms"] = getattr( metrics, f"mean_{metric_attribute_name}_ms" ) result[f"median_{metric_attribute_name}_ms"] = getattr( metrics, f"median_{metric_attribute_name}_ms" ) result[f"std_{metric_attribute_name}_ms"] = getattr( metrics, f"std_{metric_attribute_name}_ms" ) for p, value in getattr(metrics, f"percentiles_{metric_attribute_name}_ms"): p_word = str(int(p)) if int(p) == p else str(p) print("{:<40} {:<10.2f}".format(f"P{p_word} {metric_name} (ms):", value)) result[f"p{p_word}_{metric_attribute_name}_ms"] = value process_one_metric("ttft", "TTFT", "Time to First Token") process_one_metric("tpot", "TPOT", "Time per Output Token (excl. 1st token)") process_one_metric("itl", "ITL", "Inter-token Latency") process_one_metric("e2el", "E2EL", "End-to-end Latency") print("=" * 50) if profile: print("Stopping profiler...") profile_input = RequestFuncInput( model=model_id, prompt=test_request.prompt, api_url=base_url + "/stop_profile", prompt_len=test_request.prompt_len, output_len=test_request.expected_output_len, extra_body={test_request.structure_type: test_request.schema}, ) profile_output = await request_func(request_func_input=profile_input) if profile_output.success: print("Profiler stopped") return result, ret def evaluate(ret, args): def _eval_correctness_json(expected, actual): # extract json string from string using regex import regex as re actual = actual.replace("\n", "").replace(" ", "").strip() try: actual = re.search(r"\{.*\}", actual).group() actual = json.loads(actual) except Exception: return False return True def _eval_correctness_choice(expected, actual): return actual in args.choice def _eval_correctness_regex(expected, actual): import regex as re return re.match(args.regex, actual) is not None def _eval_correctness(expected, actual): if args.structure_type == "json": return _eval_correctness_json(expected, actual) elif args.structure_type == "regex": return _eval_correctness_regex(expected, actual) elif args.structure_type == "choice": return _eval_correctness_choice(expected, actual) else: return None scores = [] for res in ret: score = _eval_correctness(res["expected"], res["generated"]) res["correctness"] = score scores.append(score) not_none_scores = [score for score in scores if score is not None] return ( (sum(not_none_scores) / len(not_none_scores) * 100) if len(not_none_scores) > 0 else None ) def parse_goodput(slo_pairs): goodput_config_dict = {} try: for slo_pair in slo_pairs: slo_name, slo_val = slo_pair.split(":") goodput_config_dict[slo_name] = float(slo_val) except ValueError as err: raise argparse.ArgumentTypeError( "Invalid format found for service level objectives. " 'Specify service level objectives for goodput as "KEY:VALUE" ' "pairs, where the key is a metric name, and the value is a " "number in milliseconds." ) from err return goodput_config_dict def check_goodput_args(args): goodput_config_dict = {} VALID_NAMES = ["ttft", "tpot", "e2el"] if args.goodput: goodput_config_dict = parse_goodput(args.goodput) for slo_name, slo_val in goodput_config_dict.items(): if slo_name not in VALID_NAMES: raise ValueError( f"Invalid metric name found, {slo_name}: {slo_val}. " "The service level objective name should be one of " f"{str(VALID_NAMES)}. " ) if slo_val < 0: raise ValueError( f"Invalid value found, {slo_name}: {slo_val}. " "The service level objective value should be " "non-negative." ) return goodput_config_dict def main(args: argparse.Namespace): print(args) random.seed(args.seed) np.random.seed(args.seed) backend = args.backend model_id = args.model tokenizer_id = args.tokenizer if args.tokenizer is not None else args.model if args.base_url is not None: api_url = f"{args.base_url}{args.endpoint}" base_url = f"{args.base_url}" else: api_url = f"http://{args.host}:{args.port}{args.endpoint}" base_url = f"http://{args.host}:{args.port}" tokenizer = get_tokenizer( tokenizer_id, trust_remote_code=args.trust_remote_code, tokenizer_mode=args.tokenizer_mode, ) if args.dataset == "grammar": args.structure_type = "grammar" elif args.dataset == "regex": args.structure_type = "regex" elif args.dataset == "choice": args.structure_type = "choice" else: args.structure_type = "json" if args.no_structured_output: args.structured_output_ratio = 0 if args.save_results: result_file_name = f"{args.structured_output_ratio}so" result_file_name += f"_{backend}" result_file_name += f"_{args.request_rate}qps" result_file_name += f"_{args.model.split('/')[-1]}" result_file_name += f"_{args.dataset}" result_file_name += f"_{args.num_prompts}" result_file_name += f"_out{args.output_len}" result_file_name += ".txt" else: result_file_name = None input_requests = sample_requests(tokenizer, args) goodput_config_dict = check_goodput_args(args) benchmark_result, ret = asyncio.run( benchmark( backend=backend, api_url=api_url, base_url=base_url, model_id=model_id, tokenizer=tokenizer, input_requests=input_requests, request_rate=args.request_rate, burstiness=args.burstiness, disable_tqdm=args.disable_tqdm, profile=args.profile, selected_percentile_metrics=args.percentile_metrics.split(","), selected_percentiles=[float(p) for p in args.metric_percentiles.split(",")], ignore_eos=args.ignore_eos, max_concurrency=args.max_concurrency, structured_output_ratio=args.structured_output_ratio, goodput_config_dict=goodput_config_dict, ) ) # Save config and results to json score = evaluate(ret, args) print("correct_rate(%)", score, "\n") if args.save_results: results = { "backend": backend, "model_id": model_id, "tokenizer_id": tokenizer_id, "num_prompts": args.num_prompts, "request_rate": args.request_rate if args.request_rate < float("inf") else "inf", "burstiness": args.burstiness, "max_concurrency": args.max_concurrency, "correct_rate(%)": score, } results = {"outputs": ret, **results, **benchmark_result} # Save to file if args.result_filename: result_file_name = args.result_filename if args.result_dir: result_file_name = os.path.join(args.result_dir, result_file_name) with open(result_file_name, "w", encoding="utf-8") as outfile: json.dump(results, outfile, indent=4) def create_argument_parser(): parser = FlexibleArgumentParser( description="Benchmark the online serving throughput." ) parser.add_argument( "--backend", type=str, default="vllm", choices=list(ASYNC_REQUEST_FUNCS.keys()), ) parser.add_argument( "--base-url", type=str, default=None, help="Server or API base url if not using http host and port.", ) # Use 127.0.0.1 here instead of localhost to force the use of ipv4 parser.add_argument("--host", type=str, default="127.0.0.1") parser.add_argument("--port", type=int, default=8000) parser.add_argument( "--endpoint", type=str, default="/v1/completions", help="API endpoint.", ) parser.add_argument( "--dataset", default="json", choices=["json", "json-unique", "grammar", "regex", "choice", "xgrammar_bench"], ) parser.add_argument( "--json-schema-path", type=str, default=None, help="Path to json schema." ) parser.add_argument( "--max-concurrency", type=int, default=None, help="Maximum number of concurrent requests. This can be used " "to help simulate an environment where a higher level component " "is enforcing a maximum number of concurrent requests. While the " "--request-rate argument controls the rate at which requests are " "initiated, this argument will control how many are actually allowed " "to execute at a time. This means that when used in combination, the " "actual request rate may be lower than specified with --request-rate, " "if the server is not processing requests fast enough to keep up.", ) parser.add_argument( "--model", type=str, required=True, help="Name of the model.", ) parser.add_argument( "--tokenizer", type=str, help="Name or path of the tokenizer, if not using the default tokenizer.", ) parser.add_argument( "--tokenizer-mode", type=str, default="auto", help="Name or path of the tokenizer, if not using the default tokenizer.", ) parser.add_argument( "--num-prompts", type=int, default=1000, help="Number of prompts to process.", ) parser.add_argument( "--output-len", type=int, default=128, help="Number of output tokens.", ) parser.add_argument( "--request-rate", type=float, default=float("inf"), help="Number of requests per second. If this is inf, " "then all the requests are sent at time 0. "
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_utils.py
benchmarks/benchmark_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import json import math import os import time from types import TracebackType from typing import Any def convert_to_pytorch_benchmark_format( args: argparse.Namespace, metrics: dict[str, list], extra_info: dict[str, Any] ) -> list: """ Save the benchmark results in the format used by PyTorch OSS benchmark with on metric per record https://github.com/pytorch/pytorch/wiki/How-to-integrate-with-PyTorch-OSS-benchmark-database """ records = [] if not os.environ.get("SAVE_TO_PYTORCH_BENCHMARK_FORMAT", False): return records for name, benchmark_values in metrics.items(): record = { "benchmark": { "name": "vLLM benchmark", "extra_info": { "args": vars(args), }, }, "model": { "name": args.model, }, "metric": { "name": name, "benchmark_values": benchmark_values, "extra_info": extra_info, }, } tp = record["benchmark"]["extra_info"]["args"].get("tensor_parallel_size") # Save tensor_parallel_size parameter if it's part of the metadata if not tp and "tensor_parallel_size" in extra_info: record["benchmark"]["extra_info"]["args"]["tensor_parallel_size"] = ( extra_info["tensor_parallel_size"] ) records.append(record) return records class InfEncoder(json.JSONEncoder): def clear_inf(self, o: Any): if isinstance(o, dict): return {k: self.clear_inf(v) for k, v in o.items()} elif isinstance(o, list): return [self.clear_inf(v) for v in o] elif isinstance(o, float) and math.isinf(o): return "inf" return o def iterencode(self, o: Any, *args, **kwargs) -> Any: return super().iterencode(self.clear_inf(o), *args, **kwargs) def write_to_json(filename: str, records: list) -> None: with open(filename, "w") as f: json.dump( records, f, cls=InfEncoder, default=lambda o: f"<{type(o).__name__} object is not JSON serializable>", ) # Collect time and generate time metrics # # Example Usage: # collector = TimeCollector(TimeCollector.US) # for _ in range(total_iteration): # with collector: # ... # collector.dump_avg_max() class TimeCollector: NS: int = 1 US: int = NS * 1000 MS: int = US * 1000 S: int = MS * 1000 def __init__(self, scale: int) -> None: self.cnt: int = 0 self._sum: int = 0 self._max: int | None = None self.scale = scale self.start_time: int = time.monotonic_ns() def collect(self, v: int) -> None: self.cnt += 1 self._sum += v if self._max is None: self._max = v else: self._max = max(self._max, v) def avg(self) -> float | str: return self._sum * 1.0 / self.cnt / self.scale if self.cnt > 0 else "N/A" def max(self) -> float | str: return self._max / self.scale if self._max else "N/A" def dump_avg_max(self) -> list[float | str]: return [self.avg(), self.max()] def __enter__(self) -> None: self.start_time = time.monotonic_ns() def __exit__( self, exc_type: type[BaseException] | None, exc_value: BaseException | None, exc_traceback: TracebackType | None, ) -> None: self.collect(time.monotonic_ns() - self.start_time)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_hash.py
benchmarks/benchmark_hash.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Micro benchmark comparing built-in hash(), SHA-256, and xxHash. This focuses on a single test payload shaped like the prefix-cache hash input: (32-byte bytes object, 32-int tuple) Usage: python benchmarks/hash_micro_benchmark.py --iterations 20000 """ from __future__ import annotations import argparse import random import statistics import time from collections.abc import Callable, Iterable from vllm.utils.hashing import sha256, xxhash def _generate_test_data(seed: int) -> tuple[bytes, tuple[int, ...]]: """Generate a deterministic test payload.""" random.seed(seed) bytes_data = bytes(random.getrandbits(8) for _ in range(32)) int_tuple = tuple(random.randint(1, 1_000_000) for _ in range(32)) return (bytes_data, int_tuple) def _benchmark_func(func: Callable[[tuple], object], data: tuple, iterations: int): """Return (avg_seconds, std_seconds) for hashing `data` `iterations` times.""" times: list[float] = [] # Warm-up to avoid first-run noise. for _ in range(200): func(data) for _ in range(iterations): start = time.perf_counter() func(data) end = time.perf_counter() times.append(end - start) avg = statistics.mean(times) std = statistics.stdev(times) if len(times) > 1 else 0.0 return avg, std def _run_benchmarks( benchmarks: Iterable[tuple[str, Callable[[tuple], object]]], data: tuple, iterations: int, ): """Yield (name, avg, std) for each benchmark, skipping unavailable ones.""" for name, func in benchmarks: try: avg, std = _benchmark_func(func, data, iterations) except ModuleNotFoundError as exc: print(f"Skipping {name}: {exc}") continue yield name, avg, std def builtin_hash(data: tuple) -> int: """Wrapper for Python's built-in hash().""" return hash(data) def main() -> None: parser = argparse.ArgumentParser(description=__doc__) parser.add_argument( "--iterations", type=int, default=10_000, help="Number of measured iterations per hash function.", ) parser.add_argument( "--seed", type=int, default=42, help="Random seed for test payload." ) args = parser.parse_args() data = _generate_test_data(args.seed) benchmarks = ( ("SHA256 (pickle)", sha256), ("xxHash (pickle)", xxhash), ("built-in hash()", builtin_hash), ) print("=" * 60) print("HASH FUNCTION MICRO BENCHMARK") print("=" * 60) print("Test data: (32-byte bytes object, 32-int tuple)") print(f"Iterations: {args.iterations:,}") print("=" * 60) results = list(_run_benchmarks(benchmarks, data, args.iterations)) builtin_entry = next((r for r in results if r[0] == "built-in hash()"), None) print("\nResults:") for name, avg, std in results: print(f" {name:16s}: {avg * 1e6:8.2f} ± {std * 1e6:6.2f} μs") if builtin_entry: _, builtin_avg, _ = builtin_entry print("\n" + "=" * 60) print("SUMMARY (relative to built-in hash())") print("=" * 60) for name, avg, _ in results: if name == "built-in hash()": continue speed_ratio = avg / builtin_avg print(f"• {name} is {speed_ratio:.1f}x slower than built-in hash()") else: print("\nBuilt-in hash() result missing; cannot compute speed ratios.") if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/benchmark_ngram_proposer.py
benchmarks/benchmark_ngram_proposer.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import gc import time from unittest import mock import numpy as np from benchmark_utils import TimeCollector from tabulate import tabulate from vllm.config import ( CacheConfig, DeviceConfig, LoadConfig, ModelConfig, ParallelConfig, SchedulerConfig, SpeculativeConfig, VllmConfig, ) from vllm.platforms import current_platform from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.v1.spec_decode.ngram_proposer import NgramProposer from vllm.v1.worker.gpu_input_batch import InputBatch from vllm.v1.worker.gpu_model_runner import GPUModelRunner def benchmark_propose(args): rows = [] for max_ngram in args.max_ngram: collector = TimeCollector(TimeCollector.US) model_config = ModelConfig( model="facebook/opt-125m", max_model_len=args.num_token + args.num_spec_token, tokenizer="facebook/opt-125m", tokenizer_mode="auto", dtype="auto", seed=0, trust_remote_code=False, ) proposer = NgramProposer( vllm_config=VllmConfig( model_config=model_config, speculative_config=SpeculativeConfig( prompt_lookup_min=args.min_ngram, prompt_lookup_max=max_ngram, num_speculative_tokens=args.num_spec_token, method="ngram", ), ) ) # Warm up proposer.propose(np.random.randint(0, 20, (args.num_token,))) gc.collect() for _ in range(args.num_iteration): tokens = np.random.randint(0, 20, (args.num_req, args.num_token)) with collector: for i in range(args.num_req): proposer.propose(tokens[i, :]) rows.append( [args.num_req, args.num_token, args.min_ngram, max_ngram] + collector.dump_avg_max() ) print( tabulate( rows, headers=[ "# Request", "# Token", "Min Ngram", "Max Ngram", "Avg (us)", "Max (us)", ], tablefmt="grid", floatfmt=".3f", ) ) def benchmark_batched_propose(args): NUM_SPECULATIVE_TOKENS_NGRAM = 10 PROMPT_LOOKUP_MIN = 5 PROMPT_LOOKUP_MAX = 15 MAX_MODEL_LEN = int(1e7) DEVICE = current_platform.device_type model_config = ModelConfig(model="facebook/opt-125m", runner="generate") speculative_config = SpeculativeConfig( target_model_config=model_config, target_parallel_config=ParallelConfig(), method="ngram", num_speculative_tokens=NUM_SPECULATIVE_TOKENS_NGRAM, prompt_lookup_max=PROMPT_LOOKUP_MAX, prompt_lookup_min=PROMPT_LOOKUP_MIN, ) vllm_config = VllmConfig( model_config=model_config, cache_config=CacheConfig(), speculative_config=speculative_config, device_config=DeviceConfig(device=current_platform.device_type), parallel_config=ParallelConfig(), load_config=LoadConfig(), scheduler_config=SchedulerConfig( max_model_len=model_config.max_model_len, is_encoder_decoder=model_config.is_encoder_decoder, ), ) # monkey patch vllm.v1.worker.gpu_model_runner.get_pp_group mock_pp_group = mock.MagicMock() mock_pp_group.world_size = 1 with mock.patch( "vllm.v1.worker.gpu_model_runner.get_pp_group", return_value=mock_pp_group ): runner = GPUModelRunner(vllm_config, DEVICE) # hack max model len runner.max_model_len = MAX_MODEL_LEN runner.drafter.max_model_len = MAX_MODEL_LEN dummy_input_batch = InputBatch( max_num_reqs=args.num_req, max_model_len=MAX_MODEL_LEN, max_num_batched_tokens=args.num_req * args.num_token, device=DEVICE, pin_memory=False, vocab_size=256000, block_sizes=[16], ) dummy_input_batch._req_ids = list(str(id) for id in range(args.num_req)) dummy_input_batch.spec_decode_unsupported_reqs = () dummy_input_batch.num_tokens_no_spec = [args.num_token] * args.num_req dummy_input_batch.token_ids_cpu = np.random.randint( 0, 20, (args.num_req, args.num_token) ) runner.input_batch = dummy_input_batch sampled_token_ids = [[0]] * args.num_req print("Starting benchmark") # first run is warmup so ignore it for _ in range(args.num_iteration): start = time.time() runner.drafter.propose( sampled_token_ids, dummy_input_batch.req_ids, dummy_input_batch.num_tokens_no_spec, dummy_input_batch.token_ids_cpu, dummy_input_batch.spec_decode_unsupported_reqs, ) end = time.time() print(f"Iteration time (s): {end - start}") def invoke_main() -> None: parser = FlexibleArgumentParser( description="Benchmark the performance of N-gram speculative decode drafting" ) parser.add_argument( "--batched", action="store_true", help="consider time to prepare batch" ) parser.add_argument( "--num-iteration", type=int, default=100, help="Number of iterations to run to stabilize final data readings", ) parser.add_argument( "--num-req", type=int, default=128, help="Number of requests in the batch" ) parser.add_argument( "--num-token", type=int, default=1500, help="Number of tokens for each request" ) parser.add_argument( "--min-ngram", type=int, default=3, help="Minimum n-gram to match", ) parser.add_argument( "--max-ngram", type=int, nargs="*", default=[5, 7, 10, 15, 20], help="Maximum n-gram to match", ) parser.add_argument( "--num-spec-token", type=int, default=3, help="Number of speculative tokens to generate", ) args = parser.parse_args() if not args.batched: benchmark_propose(args) else: benchmark_batched_propose(args) """ # Example command lines: # time python3 benchmarks/benchmark_ngram_proposer.py # time python3 benchmarks/benchmark_ngram_proposer.py --batched --num-iteration 4 --num-token 1000000 --num-req 128 """ # noqa: E501 if __name__ == "__main__": invoke_main() # pragma: no cover
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/overheads/benchmark_hashing.py
benchmarks/overheads/benchmark_hashing.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import cProfile import pstats from vllm import LLM, SamplingParams from vllm.utils.argparse_utils import FlexibleArgumentParser # A very long prompt, total number of tokens is about 15k. LONG_PROMPT = ["You are an expert in large language models, aren't you?"] * 1000 LONG_PROMPT = " ".join(LONG_PROMPT) def main(args): llm = LLM( model=args.model, enforce_eager=True, enable_prefix_caching=True, tensor_parallel_size=args.tensor_parallel_size, ) sampling_params = SamplingParams(temperature=0, max_tokens=args.output_len) profiler = cProfile.Profile() print("------warm up------") for i in range(3): output = llm.generate(LONG_PROMPT, sampling_params) print(output[0].outputs[0].text) print("------start generating------") for i in range(3): profiler.runctx( "llm.generate(LONG_PROMPT, sampling_params)", globals(), locals() ) # analyze the runtime of hashing function stats = pstats.Stats(profiler) stats.sort_stats("cumulative") total_time = 0 total_calls = 0 for func in stats.stats: if "hash_of_block" in func[2]: total_time = stats.stats[func][3] total_calls = stats.stats[func][0] percentage = (total_time / stats.total_tt) * 100 print( f"Hashing took {total_time:.2f} seconds,{percentage:.2f}% of the total runtime." ) if __name__ == "__main__": parser = FlexibleArgumentParser( description="Benchmark the performance of hashing function in" "automatic prefix caching." ) parser.add_argument("--model", type=str, default="lmsys/longchat-7b-16k") parser.add_argument("--tensor-parallel-size", "-tp", type=int, default=1) parser.add_argument("--output-len", type=int, default=10) parser.add_argument( "--enable-prefix-caching", action="store_true", help="enable prefix caching" ) args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/disagg_benchmarks/rate_limiter.py
benchmarks/disagg_benchmarks/rate_limiter.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import asyncio import time class RateLimiter: """Token bucket rate limiter implementation""" def __init__(self, rate_limit): self.rate_limit = rate_limit # Requests per second self.num_available_tokens = rate_limit # Available tokens self.last_refill = time.monotonic() # Last token refill time self.lock = asyncio.Lock() # Synchronization lock async def acquire(self): """Acquire a token from the rate limiter""" while True: async with self.lock: current_time = time.monotonic() elapsed = current_time - self.last_refill # Refill num_available_tokens if more than 1 second has passed if elapsed > 1.0: self.num_available_tokens = self.rate_limit self.last_refill = current_time # Check if num_available_tokens are available if self.num_available_tokens > 0: self.num_available_tokens -= 1 return True # Calculate wait time if no num_available_tokens available wait_time = 1.0 - elapsed await asyncio.sleep(wait_time) async def __aenter__(self): """Enter async context manager - acquire token""" await self.acquire() return self async def __aexit__(self, exc_type, exc_value, traceback): """Exit async context manager - no cleanup needed""" pass
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/disagg_benchmarks/visualize_benchmark_results.py
benchmarks/disagg_benchmarks/visualize_benchmark_results.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import json import matplotlib.pyplot as plt import pandas as pd if __name__ == "__main__": data = [] for name in ["disagg_prefill", "chunked_prefill"]: for qps in [2, 4, 6, 8]: with open(f"results/{name}-qps-{qps}.json") as f: x = json.load(f) x["name"] = name x["qps"] = qps data.append(x) df = pd.DataFrame.from_dict(data) dis_df = df[df["name"] == "disagg_prefill"] chu_df = df[df["name"] == "chunked_prefill"] plt.style.use("bmh") plt.rcParams["font.size"] = 20 for key in [ "mean_ttft_ms", "median_ttft_ms", "p99_ttft_ms", "mean_itl_ms", "median_itl_ms", "p99_itl_ms", ]: fig, ax = plt.subplots(figsize=(11, 7)) plt.plot( dis_df["qps"], dis_df[key], label="disagg_prefill", marker="o", linewidth=4 ) plt.plot( chu_df["qps"], chu_df[key], label="chunked_prefill", marker="o", linewidth=4 ) ax.legend() ax.set_xlabel("QPS") ax.set_ylabel(key) ax.set_ylim(bottom=0) fig.savefig(f"results/{key}.png") plt.close(fig)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/disagg_benchmarks/request_queue.py
benchmarks/disagg_benchmarks/request_queue.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import asyncio from collections import deque class RequestQueue: """Request queue manager with concurrency control""" def __init__(self, max_concurrent, max_queue_size): # Maximum concurrent requests self.max_concurrent = max_concurrent self.max_queue_size = max_queue_size # Maximum queue size # Concurrency control self.semaphore = asyncio.Semaphore(max_concurrent) self.queue = deque() # Request queue self.queue_size = 0 # Current queue size self.lock = asyncio.Lock() # Sync queue Lock async def enqueue(self, task): """Add a request task to the queue""" async with self.lock: if self.queue_size >= self.max_queue_size: return False self.queue.append(task) self.queue_size += 1 return True async def process(self): """Process queued requests using semaphore for concurrency control""" while True: if self.queue: async with self.semaphore, self.lock: task = self.queue.popleft() self.queue_size -= 1 await task await asyncio.sleep(0.01) # Yield control to event loop
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/disagg_benchmarks/disagg_prefill_proxy_server.py
benchmarks/disagg_benchmarks/disagg_prefill_proxy_server.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import asyncio import logging import os import time import uuid from urllib.parse import urlparse import aiohttp from quart import Quart, Response, make_response, request # Configure logging logging.basicConfig(level=logging.INFO) logger = logging.getLogger(__name__) def parse_args(): """parse command line arguments""" parser = argparse.ArgumentParser(description="vLLM P/D disaggregation proxy server") # Add args parser.add_argument( "--timeout", type=float, default=6 * 60 * 60, help="Timeout for backend service requests in seconds (default: 21600)", ) parser.add_argument( "--port", type=int, default=8000, help="Port to run the server on (default: 8000)", ) parser.add_argument( "--prefill-url", type=str, default="http://localhost:8100", help="Prefill service base URL (protocol + host[:port])", ) parser.add_argument( "--decode-url", type=str, default="http://localhost:8200", help="Decode service base URL (protocol + host[:port])", ) parser.add_argument( "--kv-host", type=str, default="localhost", help="Hostname or IP used by KV transfer (default: localhost)", ) parser.add_argument( "--prefill-kv-port", type=int, default=14579, help="Prefill KV port (default: 14579)", ) parser.add_argument( "--decode-kv-port", type=int, default=14580, help="Decode KV port (default: 14580)", ) return parser.parse_args() def main(): """parse command line arguments""" args = parse_args() # Initialize configuration using command line parameters AIOHTTP_TIMEOUT = aiohttp.ClientTimeout(total=args.timeout) PREFILL_SERVICE_URL = args.prefill_url DECODE_SERVICE_URL = args.decode_url PORT = args.port PREFILL_KV_ADDR = f"{args.kv_host}:{args.prefill_kv_port}" DECODE_KV_ADDR = f"{args.kv_host}:{args.decode_kv_port}" logger.info( "Proxy resolved KV addresses -> prefill: %s, decode: %s", PREFILL_KV_ADDR, DECODE_KV_ADDR, ) app = Quart(__name__) # Attach the configuration object to the application instance so helper # coroutines can read the resolved backend URLs and timeouts without using # globals. app.config.update( { "AIOHTTP_TIMEOUT": AIOHTTP_TIMEOUT, "PREFILL_SERVICE_URL": PREFILL_SERVICE_URL, "DECODE_SERVICE_URL": DECODE_SERVICE_URL, "PREFILL_KV_ADDR": PREFILL_KV_ADDR, "DECODE_KV_ADDR": DECODE_KV_ADDR, } ) def _normalize_base_url(url: str) -> str: """Remove any trailing slash so path joins behave predictably.""" return url.rstrip("/") def _get_host_port(url: str) -> str: """Return the hostname:port portion for logging and KV headers.""" parsed = urlparse(url) host = parsed.hostname or "localhost" port = parsed.port if port is None: port = 80 if parsed.scheme == "http" else 443 return f"{host}:{port}" PREFILL_BASE = _normalize_base_url(PREFILL_SERVICE_URL) DECODE_BASE = _normalize_base_url(DECODE_SERVICE_URL) KV_TARGET = _get_host_port(DECODE_SERVICE_URL) def _build_headers(request_id: str) -> dict[str, str]: """Construct the headers expected by vLLM's P2P disagg connector.""" headers: dict[str, str] = {"X-Request-Id": request_id, "X-KV-Target": KV_TARGET} api_key = os.environ.get("OPENAI_API_KEY") if api_key: headers["Authorization"] = f"Bearer {api_key}" return headers async def _run_prefill( request_path: str, payload: dict, headers: dict[str, str], request_id: str, ): url = f"{PREFILL_BASE}{request_path}" start_ts = time.perf_counter() logger.info("[prefill] start request_id=%s url=%s", request_id, url) try: async with ( aiohttp.ClientSession(timeout=AIOHTTP_TIMEOUT) as session, session.post(url=url, json=payload, headers=headers) as resp, ): if resp.status != 200: error_text = await resp.text() raise RuntimeError( f"Prefill backend error {resp.status}: {error_text}" ) await resp.read() logger.info( "[prefill] done request_id=%s status=%s elapsed=%.2fs", request_id, resp.status, time.perf_counter() - start_ts, ) except asyncio.TimeoutError as exc: raise RuntimeError(f"Prefill service timeout at {url}") from exc except aiohttp.ClientError as exc: raise RuntimeError(f"Prefill service unavailable at {url}") from exc async def _stream_decode( request_path: str, payload: dict, headers: dict[str, str], request_id: str, ): url = f"{DECODE_BASE}{request_path}" # Stream tokens from the decode service once the prefill stage has # materialized KV caches on the target workers. logger.info("[decode] start request_id=%s url=%s", request_id, url) try: async with ( aiohttp.ClientSession(timeout=AIOHTTP_TIMEOUT) as session, session.post(url=url, json=payload, headers=headers) as resp, ): if resp.status != 200: error_text = await resp.text() logger.error( "Decode backend error %s - %s", resp.status, error_text ) err_msg = ( '{"error": "Decode backend error ' + str(resp.status) + '"}' ) yield err_msg.encode() return logger.info( "[decode] streaming response request_id=%s status=%s", request_id, resp.status, ) async for chunk_bytes in resp.content.iter_chunked(1024): yield chunk_bytes logger.info("[decode] finished streaming request_id=%s", request_id) except asyncio.TimeoutError: logger.error("Decode service timeout at %s", url) yield b'{"error": "Decode service timeout"}' except aiohttp.ClientError as exc: logger.error("Decode service error at %s: %s", url, exc) yield b'{"error": "Decode service unavailable"}' async def process_request(): """Process a single request through prefill and decode stages""" try: original_request_data = await request.get_json() # Create prefill request (max_tokens=1) prefill_request = original_request_data.copy() prefill_request["max_tokens"] = 1 if "max_completion_tokens" in prefill_request: prefill_request["max_completion_tokens"] = 1 # Execute prefill stage # The request id encodes both KV socket addresses so the backend can # shuttle tensors directly via NCCL once the prefill response # completes. request_id = ( f"___prefill_addr_{PREFILL_KV_ADDR}___decode_addr_" f"{DECODE_KV_ADDR}_{uuid.uuid4().hex}" ) headers = _build_headers(request_id) await _run_prefill(request.path, prefill_request, headers, request_id) # Execute decode stage and stream response # Pass the unmodified user request so the decode phase can continue # sampling with the already-populated KV cache. generator = _stream_decode( request.path, original_request_data, headers, request_id ) response = await make_response(generator) response.timeout = None # Disable timeout for streaming response return response except Exception: logger.exception("Error processing request") return Response( response=b'{"error": "Internal server error"}', status=500, content_type="application/json", ) @app.route("/v1/completions", methods=["POST"]) async def handle_request(): """Handle incoming API requests with concurrency and rate limiting""" try: return await process_request() except asyncio.CancelledError: logger.warning("Request cancelled") return Response( response=b'{"error": "Request cancelled"}', status=503, content_type="application/json", ) # Start the Quart server with host can be set to 0.0.0.0 app.run(port=PORT) if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/disagg_benchmarks/round_robin_proxy.py
benchmarks/disagg_benchmarks/round_robin_proxy.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import asyncio import itertools import aiohttp from aiohttp import web class RoundRobinProxy: def __init__(self, target_ports): self.target_ports = target_ports self.port_cycle = itertools.cycle(self.target_ports) async def handle_request(self, request): target_port = next(self.port_cycle) target_url = f"http://localhost:{target_port}{request.path_qs}" async with aiohttp.ClientSession() as session: try: # Forward the request async with session.request( method=request.method, url=target_url, headers=request.headers, data=request.content, ) as response: # Start sending the response resp = web.StreamResponse( status=response.status, headers=response.headers ) await resp.prepare(request) # Stream the response content async for chunk in response.content.iter_any(): await resp.write(chunk) await resp.write_eof() return resp except Exception as e: return web.Response(text=f"Error: {str(e)}", status=500) async def main(): proxy = RoundRobinProxy([8100, 8200]) app = web.Application() app.router.add_route("*", "/{path:.*}", proxy.handle_request) runner = web.AppRunner(app) await runner.setup() site = web.TCPSite(runner, "localhost", 8000) await site.start() print("Proxy server started on http://localhost:8000") # Keep the server running await asyncio.Event().wait() if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/fused_kernels/layernorm_rms_benchmarks.py
benchmarks/fused_kernels/layernorm_rms_benchmarks.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import pickle as pkl import time from collections.abc import Callable, Iterable from dataclasses import dataclass from itertools import product import torch import torch.utils.benchmark as TBenchmark from torch.utils.benchmark import Measurement as TMeasurement from tqdm import tqdm import vllm._custom_ops as ops from vllm.model_executor.layers.layernorm import RMSNorm from vllm.model_executor.layers.quantization.utils.fp8_utils import ( per_token_group_quant_fp8, ) @dataclass class bench_params_t: num_tokens: int hidden_size: int add_residual: bool dtype: torch.dtype group_size: list[int] def description(self): return ( f"N {self.num_tokens} " f"x D {self.hidden_size} " f"x R {self.add_residual} " f"x DT {self.dtype}" f"x GS {self.group_size}" ) def get_bench_params() -> list[bench_params_t]: ## Test Fixtures NUM_TOKENS = [2**x for x in range(11)] HIDDEN_SIZES = list(range(1024, 8129, 1024)) ADD_RESIDUAL = [True, False] DTYPES = [torch.bfloat16, torch.float] GROUP_SIZES = [[1, 64], [1, 128]] combinations = product(NUM_TOKENS, HIDDEN_SIZES, ADD_RESIDUAL, DTYPES, GROUP_SIZES) bench_params = list( map(lambda x: bench_params_t(x[0], x[1], x[2], x[3], x[4]), combinations) ) return bench_params # Reference impls def unfused_int8_impl( rms_norm_layer: RMSNorm, x: torch.Tensor, residual: torch.Tensor | None, quant_dtype: torch.dtype, group_size: list[int], ): # Norm torch_out = None if residual is None: torch_out = rms_norm_layer.forward_cuda(x, residual) else: torch_out, _ = rms_norm_layer.forward_cuda(x, residual) # Quant torch_out, _, _ = ops.scaled_int8_quant(torch_out) def unfused_fp8_impl( rms_norm_layer: RMSNorm, x: torch.Tensor, residual: torch.Tensor | None, quant_dtype: torch.dtype, group_size: list[int], ): # Norm torch_out = None if residual is None: torch_out = rms_norm_layer.forward_cuda(x, residual) else: torch_out, _ = rms_norm_layer.forward_cuda(x, residual) # Quant torch_out, _ = ops.scaled_fp8_quant(torch_out) def unfused_groupwise_fp8_impl( rms_norm_layer: RMSNorm, x: torch.Tensor, residual: torch.Tensor | None, quant_dtype: torch.dtype, group_size: list[int], ): # Norm torch_out = None if residual is None: torch_out = rms_norm_layer.forward_cuda(x, residual) else: torch_out, _ = rms_norm_layer.forward_cuda(x, residual) # Quant torch_out, _ = per_token_group_quant_fp8( torch_out, group_size=group_size[1], use_ue8m0=False ) def fused_impl( rms_norm_layer: RMSNorm, # this stores the weights x: torch.Tensor, residual: torch.Tensor | None, quant_dtype: torch.dtype, group_size: list[int], ): out, _ = ops.rms_norm_dynamic_per_token_quant( x, rms_norm_layer.weight, 1e-6, quant_dtype, residual=residual ) def fused_groupwise_impl( rms_norm_layer: RMSNorm, # this stores the weights x: torch.Tensor, residual: torch.Tensor | None, quant_dtype: torch.dtype, group_size: list[int], ): out, _ = ops.rms_norm_per_block_quant( x, rms_norm_layer.weight, 1e-6, quant_dtype, group_size, residual=residual, is_scale_transposed=True, ) # Bench functions def bench_fn( rms_norm_layer: RMSNorm, x: torch.Tensor, residual: torch.Tensor, quant_dtype: torch.dtype, group_size: list[int], label: str, sub_label: str, fn: Callable, description: str, ) -> TMeasurement: min_run_time = 1 globals = { "rms_norm_layer": rms_norm_layer, "x": x, "residual": residual, "quant_dtype": quant_dtype, "group_size": group_size, "fn": fn, } return TBenchmark.Timer( stmt="fn(rms_norm_layer, x, residual, quant_dtype, group_size)", globals=globals, label=label, sub_label=sub_label, description=description, ).blocked_autorange(min_run_time=min_run_time) def bench(params: bench_params_t, label: str, sub_label: str) -> Iterable[TMeasurement]: # Make inputs layer = RMSNorm(params.hidden_size, 1e-6).to(dtype=params.dtype) # Make weights layer.weight.data.normal_(mean=1.0, std=0.1) # Make inputs scale = 1 / params.hidden_size x = ( torch.randn( params.num_tokens, params.hidden_size, dtype=params.dtype, device="cuda" ) * scale ) residual = ( (torch.randn_like(x) * scale).to(device="cuda") if params.add_residual else None ) timers = [] # unfused int8 impl. timers.append( bench_fn( layer, x, residual, torch.int8, params.group_size, label, sub_label, unfused_int8_impl, "unfused_int8_impl", ) ) # unfused fp8 impl. timers.append( bench_fn( layer, x, residual, torch.float8_e4m3fn, params.group_size, label, sub_label, unfused_fp8_impl, "unfused_fp8_impl", ) ) # fused int8 impl. timers.append( bench_fn( layer, x, residual, torch.int8, params.group_size, label, sub_label, fused_impl, "fused_int8_impl", ) ) # fused fp8 impl. timers.append( bench_fn( layer, x, residual, torch.float8_e4m3fn, params.group_size, label, sub_label, fused_impl, "fused_fp8_impl", ) ) # unfused groupwise fp8 impl. timers.append( bench_fn( layer, x, residual, torch.float8_e4m3fn, params.group_size, label, sub_label, unfused_groupwise_fp8_impl, "unfused_groupwise_fp8_impl", ) ) # fused groupwise fp8 impl. timers.append( bench_fn( layer, x, residual, torch.float8_e4m3fn, params.group_size, label, sub_label, fused_groupwise_impl, "fused_groupwise_fp8_impl", ) ) print_timers(timers) return timers # launch bench # runner def print_timers(timers: Iterable[TMeasurement]): compare = TBenchmark.Compare(timers) compare.print() def main(): torch.set_default_device("cuda") bench_params = get_bench_params() timers = [] for bp in tqdm(bench_params): timers.extend(bench(bp, "rms-norm-dynamic-per-token-quant", bp.description())) print_timers(timers) # pickle all the results timestamp = int(time.time()) with open(f"rms_norm_dpt_quant-{timestamp}.pkl", "wb") as f: pkl.dump(timers, f) if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/multi_turn/convert_sharegpt_to_openai.py
benchmarks/multi_turn/convert_sharegpt_to_openai.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Download dataset from: https://huggingface.co/datasets/philschmid/sharegpt-raw/blob/main/sharegpt_20230401_clean_lang_split.json Convert to OpenAI API: export INPUT_FILE=sharegpt_20230401_clean_lang_split.json python convert_sharegpt_to_openai.py $INPUT_FILE sharegpt_conv_128.json --max-items=128 """ import argparse import json import random from statistics import mean from typing import Any import pandas as pd # type: ignore import tqdm # type: ignore from transformers import AutoTokenizer # type: ignore def has_non_english_chars(text: str) -> bool: return not text.isascii() def content_is_valid( content: str, min_content_len: int | None, max_content_len: int | None ) -> bool: if min_content_len and len(content) < min_content_len: return False if max_content_len and len(content) > max_content_len: return False return has_non_english_chars(content) def print_stats( conversations: "list[dict[Any, Any]]", tokenizer: AutoTokenizer | None = None ) -> None: # Collect statistics stats = [] print("\nCollecting statistics...") for item in tqdm.tqdm(conversations): # item has "id" and "messages" messages = item["messages"] user_turns = 0 assistant_turns = 0 user_words = 0 assistant_words = 0 conv_chars = 0 user_tokens: list[int] = [] assistant_tokens: list[int] = [] for m in messages: content = m["content"] conv_chars += len(content) content_num_words = content.count(" ") + 1 num_tokens = 0 if tokenizer: num_tokens = len(tokenizer(m["content"]).input_ids) if m["role"] == "user": user_turns += 1 user_words += content_num_words if tokenizer: user_tokens.append(num_tokens) elif m["role"] == "assistant": assistant_turns += 1 assistant_words += content_num_words if tokenizer: assistant_tokens.append(num_tokens) # assert user_turns == assistant_turns, \ # f"Invalid conversation ID {item['id']}" conv_words = user_words + assistant_words item_stats = { "user_turns": user_turns, "assistant_turns": assistant_turns, "user_words": user_words, "assistant_words": assistant_words, "conv_turns": len(messages), "conv_words": conv_words, "conv_characters": conv_chars, } if len(user_tokens) > 0: item_stats["user_tokens"] = int(mean(user_tokens)) if len(assistant_tokens) > 0: item_stats["assistant_tokens"] = int(mean(assistant_tokens)) stats.append(item_stats) print("\nStatistics:") percentiles = [0.25, 0.5, 0.75, 0.9, 0.99, 0.999, 0.9999] df = pd.DataFrame(stats) print(df.describe(percentiles=percentiles).transpose()) def convert_sharegpt_to_openai( seed: int, input_file: str, output_file: str, max_items: int | None, min_content_len: int | None = None, max_content_len: int | None = None, min_turns: int | None = None, max_turns: int | None = None, model: str | None = None, ) -> None: if min_turns and max_turns: assert min_turns <= max_turns if min_content_len and max_content_len: # Verify that min is not larger than max if both were given assert min_content_len <= max_content_len print( f"Input parameters:\n{seed=}, {max_items=}, {min_content_len=}," f" {max_content_len=}, {min_turns=}, {max_turns=}\n" ) random.seed(seed) tokenizer = None if model is not None: print(f"Loading tokenizer from: {model}") tokenizer = AutoTokenizer.from_pretrained(model) # Read the ShareGPT JSON file print(f"Reading file: {input_file}") with open(input_file, encoding="utf-8") as f: # Should be a list of dicts # Each dict should have "id" (string) and "conversations" (list of dicts) sharegpt_data = json.load(f) assert isinstance(sharegpt_data, list), "Input file should contain a list of dicts" print(f"Total items in input file: {len(sharegpt_data):,}") print(f"Shuffling dataset with seed {seed}") random.shuffle(sharegpt_data) # Map conversation ID to the all the messages conversation_parts: dict[str, list[Any]] = {} for item in tqdm.tqdm(sharegpt_data): assert "id" in item, "Missing key 'id'" assert "conversations" in item, "Missing key 'conversations'" # Conversation ID (e.g: "hiWPlMD") and part/session (0, 1, 2, etc.) conv_id, _ = item["id"].split("_") new_turns = item["conversations"] if conv_id not in conversation_parts: # Start new conversation conversation_parts[conv_id] = [] elif len(conversation_parts[conv_id]) > 0 and len(new_turns) > 0: prev_turns = conversation_parts[conv_id][-1] if prev_turns[-1]["from"] == new_turns[0]["from"]: new_turns = new_turns[1:] if len(new_turns) > 0: # We assume that parts are in order in the ShareGPT dataset conversation_parts[conv_id].append(new_turns) dataset: list[dict[str, Any]] = [] for conv_id, conv_parts in conversation_parts.items(): new_item = {"id": conv_id} conversations: list[dict[str, str]] = [] # Merge all parts for conv_part in conv_parts: conversations.extend(conv_part) if len(conversations) > 0: new_item["conversations"] = conversations dataset.append(new_item) print(f"Total unique conversations (IDs) in input file: {len(dataset):,}") # Final output data final_openai_dataset: list[dict] = [] # Filter conversations from the ShareGPT dataset and convert to OpenAI format for item in tqdm.tqdm(dataset): messages: list[dict] = [] assert "id" in item, "Missing key 'id'" assert "conversations" in item, "Missing key 'conversations'" conv_id = item["id"] conversations = item["conversations"] if min_turns is not None and len(conversations) < min_turns: # Skip short conversations continue # Convert each message in the conversation, up to max_turns if specified for i, turn in enumerate(conversations): assert "from" in turn and "value" in turn, ( f"Invalid conversation ID {conv_id} - missing 'from' or 'value'" ) role = None turn_from = turn["from"] if turn_from in {"human", "user"}: role = "user" elif turn_from in {"gpt", "bing", "chatgpt", "bard"}: role = "assistant" elif turn_from == "system": role = "system" assert role is not None, ( f"Invalid conversation ID {conv_id} - 'from'='{turn_from}' is invalid" ) if i == 0 and role != "user": # If the first message is from assistant (gpt), skip it. # this happens when the conversation is a follow-up # to a previous conversation (from the same user). continue if max_turns is not None and i >= max_turns: break # Convert message to OpenAI format (with "role" and "content") content = turn["value"] messages.append({"role": role, "content": content}) # Add the converted conversation to the OpenAI format if len(messages) > 0: valid_messages = True # First turn should always be from the user user_turn = True for m in messages: # Make sure that turns alternate between user and assistant if (user_turn and m["role"] != "user") or ( not user_turn and m["role"] != "assistant" ): valid_messages = False break user_turn = not user_turn content = m["content"] valid_messages = content_is_valid( content, min_content_len, max_content_len ) if not valid_messages: break if valid_messages is True: final_openai_dataset.append({"id": conv_id, "messages": messages}) assert len(final_openai_dataset) > 0, "Final number of conversations is zero" print_stats(final_openai_dataset) print_stats_again = False if max_items is not None and len(final_openai_dataset) > max_items: print(f"\n\nSampling {max_items} items from the dataset...") print_stats_again = True final_openai_dataset = random.sample(final_openai_dataset, max_items) if print_stats_again: # Print stats after the dataset changed print_stats(final_openai_dataset, tokenizer) # Write the converted data to a new JSON file final_size = len(final_openai_dataset) print(f"\nTotal conversations converted (after filtering): {final_size:,}") print(f"\nWriting file: {output_file}") with open(output_file, "w", encoding="utf-8") as f: json.dump(final_openai_dataset, f, ensure_ascii=False, indent=2) def main() -> None: parser = argparse.ArgumentParser( description="Convert ShareGPT dataset to OpenAI API format" ) parser.add_argument("input_file", help="Path to the input ShareGPT JSON file") parser.add_argument( "output_file", help="Path to the output OpenAI format JSON file" ) parser.add_argument( "--seed", type=int, default=0, help="Seed for random number generators" ) parser.add_argument( "--max-items", type=int, default=None, help="Maximum number of items in the output file", ) parser.add_argument( "--min-turns", type=int, default=None, help="Minimum number of turns per conversation", ) parser.add_argument( "--max-turns", type=int, default=None, help="Maximum number of turns per conversation", ) parser.add_argument( "--min-content-len", type=int, default=None, help="Min number of characters in the messages' content", ) parser.add_argument( "--max-content-len", type=int, default=None, help="Max number of characters in the messages' content", ) parser.add_argument( "--model", type=str, default=None, help="LLM model, only the tokenizer will be used", ) args = parser.parse_args() convert_sharegpt_to_openai( args.seed, args.input_file, args.output_file, args.max_items, args.min_content_len, args.max_content_len, args.min_turns, args.max_turns, args.model, ) if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/multi_turn/bench_utils.py
benchmarks/multi_turn/bench_utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import logging from enum import Enum class Color(Enum): RED = "\033[91m" GREEN = "\033[92m" BLUE = "\033[94m" PURPLE = "\033[95m" CYAN = "\033[96m" YELLOW = "\033[93m" RESET = "\033[0m" def __str__(self): return self.value TEXT_SEPARATOR = "-" * 100 # Configure the logger logging.basicConfig( level=logging.INFO, format="%(asctime)s [%(levelname)s] - %(message)s", datefmt="%d-%m-%Y %H:%M:%S", ) logger = logging.getLogger(__name__)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/multi_turn/bench_dataset.py
benchmarks/multi_turn/bench_dataset.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from abc import ABC, abstractmethod from statistics import mean from typing import Any, NamedTuple import numpy as np # type: ignore import pandas as pd # type: ignore from bench_utils import ( TEXT_SEPARATOR, Color, logger, ) from tqdm import tqdm from transformers import AutoTokenizer # type: ignore # Conversation ID is a string (e.g: "UzTK34D") ConvId = str # A list of dicts (dicts with keys "id" and "messages") ShareGptConversations = list[dict[str, Any]] # A list of dicts (dicts with keys "role" and "content") MessagesList = list[dict[str, str]] # Map conversation ID to conversation messages ConversationsMap = list[ConvId, MessagesList] class Distribution(ABC): @abstractmethod def sample(self, size: int = 1) -> np.ndarray: pass class UniformDistribution(Distribution): def __init__( self, min_val: int | float, max_val: int | float, is_integer: bool = True, ) -> None: self.min_val = min_val self.max_val = max_val self.is_integer = is_integer def sample(self, size: int = 1) -> np.ndarray: if self.is_integer: return np.random.randint( int(self.min_val), int(self.max_val + 1), size=size ) else: return np.random.uniform(self.min_val, self.max_val, size=size) def __repr__(self) -> str: return f"UniformDistribution[{self.min_val}, {self.max_val}]" class ConstantDistribution(Distribution): def __init__(self, value: int | float) -> None: self.value = value self.max_val = value def sample(self, size: int = 1) -> np.ndarray: return np.full(shape=size, fill_value=self.value) def __repr__(self) -> str: return f"Constant[{self.value}]" class ZipfDistribution(Distribution): def __init__(self, alpha: float, max_val: int | None = None) -> None: self.alpha = alpha self.max_val = max_val def sample(self, size: int = 1) -> np.ndarray: samples = np.random.zipf(self.alpha, size=size) if self.max_val: samples = np.minimum(samples, self.max_val) return samples def __repr__(self) -> str: return f"ZipfDistribution[{self.alpha}]" class PoissonDistribution(Distribution): def __init__(self, alpha: float, max_val: int | None = None) -> None: self.alpha = alpha self.max_val = max_val def sample(self, size: int = 1) -> np.ndarray: samples = np.random.poisson(self.alpha, size=size) if self.max_val: samples = np.minimum(samples, self.max_val) return samples def __repr__(self) -> str: return f"PoissonDistribution[{self.alpha}]" class LognormalDistribution(Distribution): def __init__( self, mean: float | None = None, sigma: float | None = None, average: int | None = None, median_ratio: float | None = None, max_val: int | None = None, ) -> None: self.average = average self.median_ratio = median_ratio self.max_val = max_val if average is not None: if average < 1: raise ValueError("Lognormal average must be positive") if mean or sigma: raise ValueError( "When using lognormal average, you can't provide mean/sigma" ) if self.median_ratio is None: # Default value that provides relatively wide range of values self.median_ratio = 0.85 # Calculate mean/sigma of np.random.lognormal based on the average mean, sigma = self._generate_lognormal_by_median( target_average=self.average, median_ratio=self.median_ratio ) else: if mean is None or sigma is None: raise ValueError( "Must provide both mean and sigma if average is not used" ) if mean <= 0 or sigma < 0: raise ValueError( "Lognormal mean must be positive and sigma must be non-negative" ) # Mean and standard deviation of the underlying normal distribution # Based on numpy.random.lognormal self.mean = mean self.sigma = sigma @staticmethod def _generate_lognormal_by_median( target_average: int, median_ratio: float ) -> tuple[float, float]: """ Compute (mu, sigma) for a lognormal distribution given: - a target average (mean of the distribution) - a ratio of median / mean (controls skewness), assume mean > median Background: If Z ~ Normal(mu, sigma^2), then X = exp(Z) ~ LogNormal(mu, sigma). * mean(X) = exp(mu + sigma^2 / 2) * median(X) = exp(mu) So: median / mean = exp(mu) / exp(mu + sigma^2 / 2) = exp(-sigma^2 / 2) Rearranging: sigma^2 = 2 * ln(mean / median) mu = ln(median) This gives a unique (mu, sigma) for any valid mean and median. """ # Check input validity: median must be smaller than mean if median_ratio <= 0 or median_ratio >= 1: raise ValueError("median_ratio must be in range (0, 1)") target_median = target_average * median_ratio # Solve sigma^2 = 2 * ln(mean / median) sigma = np.sqrt(2 * np.log(target_average / target_median)) mu = np.log(target_median) return mu, sigma def sample(self, size: int = 1) -> np.ndarray: samples = np.random.lognormal(mean=self.mean, sigma=self.sigma, size=size) if self.average is not None: # Scale to average samples *= self.average / samples.mean() if self.max_val: samples = np.minimum(samples, self.max_val) return np.round(samples).astype(int) def __repr__(self) -> str: if self.average: return ( f"LognormalDistribution[{self.average}, " f"{self.median_ratio}, {self.max_val}]" ) return f"LognormalDistribution[{self.mean}, {self.sigma}, {self.max_val}]" class GenConvArgs(NamedTuple): num_conversations: int text_files: list[str] input_num_turns: Distribution input_common_prefix_num_tokens: Distribution input_prefix_num_tokens: Distribution input_num_tokens: Distribution output_num_tokens: Distribution print_stats: bool def verify_field_exists( conf: dict, field_name: str, section: str, subsection: str ) -> None: if field_name not in conf: raise ValueError( f"Missing field '{field_name}' in {section=} and {subsection=}" ) def get_random_distribution( conf: dict, section: str, subsection: str, optional: bool = False ) -> Distribution: # section can be "prompt_input" or "prompt_output" (both required) conf = conf[section] if optional and subsection not in conf: # Optional subsection, if not found assume the value is always 0 return ConstantDistribution(0) # subsection can be "num_turns", "num_tokens" or "prefix_num_tokens" if subsection not in conf: raise ValueError(f"Missing subsection {subsection} in section {section}") conf = conf[subsection] distribution = conf.get("distribution") if distribution is None: raise ValueError( f"Missing field 'distribution' in {section=} and {subsection=}" ) if distribution == "constant": verify_field_exists(conf, "value", section, subsection) return ConstantDistribution(conf["value"]) elif distribution == "zipf": verify_field_exists(conf, "alpha", section, subsection) max_val = conf.get("max", None) return ZipfDistribution(conf["alpha"], max_val=max_val) elif distribution == "poisson": verify_field_exists(conf, "alpha", section, subsection) max_val = conf.get("max", None) return PoissonDistribution(conf["alpha"], max_val=max_val) elif distribution == "lognormal": max_val = conf.get("max", None) if "average" in conf: # Infer lognormal mean/sigma (numpy) from input average median_ratio = conf.get("median_ratio", None) return LognormalDistribution( average=conf["average"], median_ratio=median_ratio, max_val=max_val ) # Use mean/sigma directly (for full control over the distribution) verify_field_exists(conf, "mean", section, subsection) verify_field_exists(conf, "sigma", section, subsection) return LognormalDistribution( mean=conf["mean"], sigma=conf["sigma"], max_val=max_val ) elif distribution == "uniform": verify_field_exists(conf, "min", section, subsection) verify_field_exists(conf, "max", section, subsection) min_value = conf["min"] max_value = conf["max"] assert min_value > 0 assert min_value <= max_value is_integer = isinstance(min_value, int) and isinstance(max_value, int) return UniformDistribution(min_value, max_value, is_integer) else: raise ValueError(f"Unknown distribution: {distribution}") def parse_input_json_file(conf: dict) -> GenConvArgs: # Validate the input file assert isinstance(conf, dict) required_fields = [ "filetype", "num_conversations", "text_files", "prompt_input", "prompt_output", ] for field in required_fields: assert field in conf, f"Missing field {field} in input {conf}" assert conf["filetype"] == "generate_conversations" assert conf["num_conversations"] > 0, "num_conversations should be larger than zero" text_files = conf["text_files"] assert isinstance(text_files, list), "Field 'text_files' should be a list" assert len(text_files) > 0, ( "Field 'text_files' should be a list with at least one file" ) # Parse the parameters for the prompt input/output workload input_num_turns = get_random_distribution(conf, "prompt_input", "num_turns") input_num_tokens = get_random_distribution(conf, "prompt_input", "num_tokens") input_common_prefix_num_tokens = get_random_distribution( conf, "prompt_input", "common_prefix_num_tokens", optional=True ) input_prefix_num_tokens = get_random_distribution( conf, "prompt_input", "prefix_num_tokens" ) output_num_tokens = get_random_distribution(conf, "prompt_output", "num_tokens") print_stats: bool = conf.get("print_stats", False) assert isinstance(print_stats, bool), ( "Field 'print_stats' should be either 'true' or 'false'" ) args = GenConvArgs( num_conversations=conf["num_conversations"], text_files=text_files, input_num_turns=input_num_turns, input_common_prefix_num_tokens=input_common_prefix_num_tokens, input_prefix_num_tokens=input_prefix_num_tokens, input_num_tokens=input_num_tokens, output_num_tokens=output_num_tokens, print_stats=print_stats, ) return args def print_conv_stats(conversations: ConversationsMap, tokenizer: AutoTokenizer) -> None: # Collect statistics conv_stats: list[dict[Any, Any]] = [] req_stats: list[int] = [] print("\nCollecting statistics...") for messages in conversations.values(): # messages is a list of dicts user_tokens: list[int] = [] assistant_tokens: list[int] = [] request_tokens: list[int] = [] req_tokens = 0 for m in messages: content = m["content"] num_tokens = len(tokenizer(content).input_ids) if m["role"] == "user": user_tokens.append(num_tokens) # New user prompt including all chat history req_tokens += num_tokens request_tokens.append(req_tokens) elif m["role"] == "assistant": assistant_tokens.append(num_tokens) # Update assistant answer # (will be part of chat history for the next user prompt) req_tokens += num_tokens item_stats = { "conversation_turns": len(messages), "user_tokens": mean(user_tokens), "assistant_tokens": mean(assistant_tokens), } conv_stats.append(item_stats) req_stats.extend(request_tokens) # Print statistics percentiles = [0.25, 0.5, 0.75, 0.9, 0.99] print(TEXT_SEPARATOR) print(f"{Color.YELLOW}Conversations statistics:{Color.RESET}") print(TEXT_SEPARATOR) df = pd.DataFrame(conv_stats) print(df.describe(percentiles=percentiles).transpose()) print(TEXT_SEPARATOR) print(f"{Color.YELLOW}Request statistics:{Color.RESET}") print(TEXT_SEPARATOR) df = pd.DataFrame(req_stats, columns=["request_tokens"]) print(df.describe(percentiles=percentiles).transpose()) print(TEXT_SEPARATOR) def generate_conversations( args: GenConvArgs, tokenizer: AutoTokenizer ) -> ConversationsMap: # Text for all user prompts # (text from the input text files will be appended to this line) base_prompt_text = "Please rewrite the following text and add more content: " base_prompt_token_count = len( tokenizer.encode(base_prompt_text, add_special_tokens=False) ) logger.info(f"{Color.PURPLE}Generating conversations...{Color.RESET}") logger.info(args) list_of_tokens = [] for filename in args.text_files: # Load text file that will be used to generate prompts with open(filename) as file: data = file.read() tokens_in_file = tokenizer.encode(data, add_special_tokens=False) list_of_tokens.extend(tokens_in_file) logger.info( f"Loaded {len(tokens_in_file)} tokens from file {filename}, " f"total tokens so far: {len(list_of_tokens)}" ) conversations: ConversationsMap = {} conv_id = 0 # Generate number of turns for every conversation turn_count: np.ndarray = args.input_num_turns.sample(args.num_conversations) # Turn count should be at least 2 (one user prompt and one assistant answer) turn_count = np.maximum(turn_count, 2) # Round up to an even number (every user prompt should have an answer) turn_count = turn_count + (turn_count % 2) # Generate number of prefix tokens for every conversation conv_prefix_tokens: np.ndarray = args.input_prefix_num_tokens.sample( args.num_conversations ) # Used to reduce shared text between conversations # (jump/skip over text sections between conversations) base_offset = 0 # Common prefix size for all conversations (only 1 sample required) common_prefix_text = "" common_prefix_tokens: int = args.input_common_prefix_num_tokens.sample(1)[0] if common_prefix_tokens > 0: # Using "." at the end to separate sentences common_prefix_text = ( tokenizer.decode(list_of_tokens[: common_prefix_tokens - 2]) + "." ) base_offset += common_prefix_tokens for conv_id in tqdm( range(args.num_conversations), total=args.num_conversations, desc="Generating conversations", unit="conv", ): # Generate a single conversation messages: MessagesList = [] nturns = turn_count[conv_id] # User prompt token count per turn (with lower limit) input_token_count: np.ndarray = args.input_num_tokens.sample(nturns).astype(int) input_token_count = np.maximum(input_token_count, base_prompt_token_count) # Assistant answer token count per turn (with lower limit) output_token_count: np.ndarray = args.output_num_tokens.sample(nturns).astype( int ) output_token_count = np.maximum(output_token_count, 1) user_turn = True for turn_id in range(nturns): if user_turn: role = "user" num_tokens = input_token_count[turn_id] # Generate the user prompt, # use a unique prefix (the conv_id) for each conversation # (to avoid shared prefix between conversations) content = f"{conv_id} is a nice number... " if len(common_prefix_text) > 0 and turn_id == 0: content = common_prefix_text + content # Update the number of tokens left for the content num_tokens -= len(tokenizer.encode(content, add_special_tokens=False)) if turn_id == 0: prefix_num_tokens = conv_prefix_tokens[conv_id] if prefix_num_tokens > 0: # Add prefix text (context) to the first turn start_offset = base_offset end_offset = start_offset + prefix_num_tokens assert len(list_of_tokens) > end_offset, ( "Not enough input text to generate " f"{prefix_num_tokens} tokens for the " f"prefix text ({start_offset=}, {end_offset=})" ) content += f"{conv_id}, " + tokenizer.decode( list_of_tokens[start_offset:end_offset] ) base_offset += prefix_num_tokens # Add the actual user prompt/question after the prefix text content += base_prompt_text num_tokens -= base_prompt_token_count if num_tokens > 0: # Add text from the input file (to reach the desired token count) start_offset = base_offset + turn_id * input_token_count.max() end_offset = start_offset + num_tokens assert len(list_of_tokens) > end_offset, ( f"Not enough input text to generate {num_tokens} tokens " f"for the prompt ({start_offset=}, {end_offset=})" ) # Convert tokens back to text content += tokenizer.decode(list_of_tokens[start_offset:end_offset]) else: role = "assistant" # This content will not be used as input to the LLM server # (actual answers will be used instead). # Content is only required to determine the min_tokens/max_tokens # (inputs to the LLM server). num_tokens = output_token_count[turn_id] assert len(list_of_tokens) > num_tokens, ( f"Not enough input text to generate {num_tokens} " "tokens for assistant content" ) content = tokenizer.decode(list_of_tokens[:num_tokens]) # Append the user/assistant message to the list of messages messages.append({"role": role, "content": content}) user_turn = not user_turn # Add the new conversation conversations[f"CONV_ID_{conv_id}"] = messages # Increase base offset for the next conversation base_offset += nturns if args.print_stats: print_conv_stats(conversations, tokenizer) return conversations def conversations_list_to_dict(input_list: ShareGptConversations) -> ConversationsMap: conversations: ConversationsMap = {} for item in input_list: conv_id: str = item["id"] assert isinstance(conv_id, str) assert conv_id not in conversations, ( f"Conversation ID {conv_id} found more than once in the input" ) messages: MessagesList = item["messages"] assert isinstance(messages, list), ( f"Conversation messages should be a list (ID: {conv_id})" ) assert len(messages) > 0, f"Conversation with no messages (ID: {conv_id})" conversations[conv_id] = messages logger.info(f"Using {len(conversations)} unique conversations (IDs)") assert len(conversations) == len(input_list) # Print statistics about the selected conversations stats: list[dict[str, Any]] = [] for conv_data in conversations.values(): stats.append({"num_turns": len(conv_data)}) print(TEXT_SEPARATOR) print(f"{Color.YELLOW}Conversations statistics:{Color.RESET}") print(TEXT_SEPARATOR) percentiles = [0.25, 0.5, 0.75, 0.9, 0.99, 0.999, 0.9999] conv_stats = pd.DataFrame(stats).describe(percentiles=percentiles) print(conv_stats.transpose()) print(TEXT_SEPARATOR) return conversations def conversations_dict_to_list(input_dict: ConversationsMap) -> ShareGptConversations: output: ShareGptConversations = [] for conv_id, conv_data in input_dict.items(): new_item = {"id": conv_id, "messages": conv_data} output.append(new_item) return output
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/multi_turn/benchmark_serving_multi_turn.py
benchmarks/multi_turn/benchmark_serving_multi_turn.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import asyncio import json import logging import multiprocessing as mp import os import random import time from collections import Counter, deque from datetime import datetime from enum import Enum from http import HTTPStatus from statistics import mean from typing import NamedTuple import aiohttp # type: ignore import numpy as np # type: ignore import pandas as pd # type: ignore from bench_dataset import ( ConversationsMap, ConvId, GenConvArgs, MessagesList, ShareGptConversations, conversations_dict_to_list, conversations_list_to_dict, generate_conversations, parse_input_json_file, ) from bench_utils import TEXT_SEPARATOR, Color, logger from transformers import AutoTokenizer # type: ignore NUM_TOKENS_FROM_DATASET = 0 TERM_SIGNAL = None class ConversationSampling(str, Enum): ROUND_ROBIN = "round_robin" RANDOM = "random" def __str__(self): return self.value class ClientArgs(NamedTuple): seed: int max_num_requests: int | None skip_first_turn: bool max_turns: int | None max_active_conversations: int verbose: bool print_content: bool verify_output: bool conversation_sampling: ConversationSampling request_rate: float max_retries: int class RequestArgs(NamedTuple): chat_url: str model: str stream: bool limit_min_tokens: int # Use negative value for no limit limit_max_tokens: int # Use negative value for no limit timeout_sec: int class BenchmarkArgs(NamedTuple): url: str num_clients: int early_stop: bool class ServerResponse(NamedTuple): valid: bool ttft_ms: float # time to first chunk tpot_ms: float # time per output chunk (one or more tokens) latency_ms: float start_time_ms: float first_chunk: str # first chunk of the content content: str # includes the first_chunk num_chunks: int def __str__(self) -> str: return f"ttft_ms {self.ttft_ms:.2f}, tpot_ms {self.tpot_ms:.2f}, latency_ms {self.latency_ms:.2f}" # noqa: E501 class RequestStats(NamedTuple): ttft_ms: float tpot_ms: float latency_ms: float start_time_ms: float input_num_turns: int input_num_tokens: int output_num_tokens: int output_num_chunks: int output_num_first_chunk_tokens: int approx_cached_percent: float conversation_id: str client_id: int def __str__(self) -> str: return ( f"ttft_ms {self.ttft_ms:.2f}, tpot_ms {self.tpot_ms:.2f}, latency_ms {self.latency_ms:.2f}, input_num_tokens {self.input_num_tokens}, " # noqa: E501 f"output_num_tokens {self.output_num_tokens} ({self.output_num_chunks} chunks, {self.output_num_first_chunk_tokens} tokens in first chunk), " # noqa: E501 f"approx_cached_percent {self.approx_cached_percent:.2f}%" ) class MetricStats: def __init__(self) -> None: self.min: float | None = None self.max: float | None = None self.avg: float | None = None self.sum = 0.0 self.count = 0 def update(self, value: float) -> None: if self.min is None: self.min = value else: self.min = min(self.min, value) if self.max is None: self.max = value else: self.max = max(self.max, value) self.sum += value self.count += 1 self.avg = self.sum / self.count def __repr__(self) -> str: if self.count == 0: return "no data" return f"avg: {self.avg:>10.3f}, min: {self.min:>10.3f}, max: {self.max:>10.3f}" class MovingAverage: def __init__(self, window_size: int) -> None: self.window_size = window_size self.window = np.zeros(window_size) self.index = 0 self.sum = 0.0 self.count = 0 self.avg: float | None = None def update(self, new_value: float) -> None: if self.count < self.window_size: # Filling up the window self.sum += new_value self.window[self.count] = new_value self.count += 1 else: # Window is full, start replacing old values old_value = self.window[self.index] self.sum = self.sum - old_value + new_value self.window[self.index] = new_value self.index = (self.index + 1) % self.window_size self.avg = self.sum / self.count def __repr__(self) -> str: if self.count == 0: return "no data" return f"avg: {self.avg:>10.3f} ({self.count} samples)" class DebugStats: def __init__(self, logger: logging.Logger, window_size: int) -> None: self.logger = logger self.metrics: dict[str, MovingAverage | MetricStats] = { "moving_avg_ttft_ms": MovingAverage(window_size), "moving_avg_tpot_ms": MovingAverage(window_size), "ttft_ms": MetricStats(), "tpot_ms": MetricStats(), "latency_ms": MetricStats(), "input_num_turns": MetricStats(), "input_num_tokens": MetricStats(), "output_num_tokens": MetricStats(), } def update(self, data: RequestStats) -> None: self.metrics["ttft_ms"].update(data.ttft_ms) self.metrics["moving_avg_ttft_ms"].update(data.ttft_ms) self.metrics["tpot_ms"].update(data.tpot_ms) self.metrics["moving_avg_tpot_ms"].update(data.tpot_ms) self.metrics["latency_ms"].update(data.latency_ms) self.metrics["input_num_turns"].update(data.input_num_turns) self.metrics["input_num_tokens"].update(data.input_num_tokens) self.metrics["output_num_tokens"].update(data.output_num_tokens) def print(self) -> None: self.logger.info("-" * 50) for k, v in self.metrics.items(): kv_info = f"[{k:25}] {v}" self.logger.info(kv_info) self.logger.info("-" * 50) def nanosec_to_millisec(value: float) -> float: return value / 1000000.0 def nanosec_to_sec(value: float) -> float: return value / 1000000000.0 async def send_request( session: aiohttp.ClientSession, messages: list[dict[str, str]], chat_url: str, model: str, stream: bool = True, min_tokens: int | None = None, max_tokens: int | None = None, timeout_sec: int = 120, ) -> ServerResponse: payload = { "model": model, "messages": messages, "seed": 0, "temperature": 0.0, } if stream: payload["stream"] = True payload["stream_options"] = {"include_usage": False} if min_tokens is not None: payload["min_tokens"] = min_tokens if max_tokens is not None: payload["max_tokens"] = max_tokens headers = {"Content-Type": "application/json"} # Calculate the timeout for the request if max_tokens is not None: # Assume TPOT of 200ms and use max_tokens to determine timeout token_based_timeout = int(max_tokens * 0.2) if token_based_timeout > timeout_sec: timeout_sec = token_based_timeout logger.info( "Using timeout of %ds based on max_tokens %d", timeout_sec, max_tokens, ) timeout = aiohttp.ClientTimeout(total=timeout_sec) valid_response = True ttft: float | None = None chunk_delay: list[int] = [] latency: float | None = None first_chunk = "" generated_text = "" start_time: int = time.perf_counter_ns() most_recent_timestamp: int = start_time async with session.post( url=chat_url, json=payload, headers=headers, timeout=timeout ) as response: http_status = HTTPStatus(response.status) if http_status == HTTPStatus.OK: async for chunk_bytes in response.content: chunk_bytes = chunk_bytes.strip() if not chunk_bytes: continue chunk = chunk_bytes.decode("utf-8").removeprefix("data: ") if chunk == "[DONE]": # End of stream latency = time.perf_counter_ns() - start_time elif stream is False: data = json.loads(chunk) message = data["choices"][0]["message"] assert message["role"] == "assistant" generated_text += message["content"] else: timestamp: int = time.perf_counter_ns() data = json.loads(chunk) # Delta is the new content/text/data delta = data["choices"][0]["delta"] if delta.get("content", None): if ttft is None: # First token first_token_time = time.perf_counter_ns() ttft = first_token_time - start_time first_chunk = delta["content"] else: # Decoding phase chunk_delay.append(timestamp - most_recent_timestamp) generated_text += delta["content"] most_recent_timestamp = timestamp else: valid_response = False content = await response.text() logger.warning( f"{Color.YELLOW}Received HTTP status {http_status.value} " f"({http_status.phrase}): {content}{Color.RESET}" ) if latency is None: latency = -1.0 if valid_response: # Streaming is disabled, latency was not set latency = time.perf_counter_ns() - start_time if ttft is None: # The response was a single chunk ttft = latency # Each chunk may include more than one token tpot: float = mean(chunk_delay) if len(chunk_delay) > 0 else 0.0 num_chunks: int = len(chunk_delay) sr = ServerResponse( valid=valid_response, ttft_ms=nanosec_to_millisec(ttft) if ttft > 0.0 else -1.0, tpot_ms=nanosec_to_millisec(tpot), latency_ms=nanosec_to_millisec(latency), start_time_ms=nanosec_to_millisec(start_time), first_chunk=first_chunk, content=generated_text, num_chunks=num_chunks, ) return sr def get_short_string(input: str) -> str: n = 20 if len(input) < 400: return input return f"{input[:n]}...{input[-n:]}" def get_token_count(tokenizer: AutoTokenizer, text: str) -> int: return len(tokenizer(text, add_special_tokens=False).input_ids) def get_messages_token_count( tokenizer: AutoTokenizer, messages: list[dict[str, str]] ) -> int: token_count = 0 for m in messages: token_count += get_token_count(tokenizer, m["content"]) return token_count async def send_turn( session: aiohttp.ClientSession, client_id: int, conv_id: str, conversation_messages: MessagesList, messages_to_use: int, tokenizer: AutoTokenizer, req_args: RequestArgs, verbose: bool, verify_output: bool, ) -> RequestStats | None: assert messages_to_use > 0 assert messages_to_use <= len(conversation_messages) messages = conversation_messages[:messages_to_use] # Index of the next message (the role should be "user") index = messages_to_use - 1 # Verify that the message has only two keys, "role" and "content" assert len(messages[index].keys()) == 2 assert "role" in messages[index] and "content" in messages[index] assert messages[index]["role"] == "user", ( f"Failed on conversation ID {conv_id}, message role should be user" ) if verbose: print( f"{Color.CYAN}Messages (conversation ID {conv_id}," f" {len(messages)} turns):{Color.RESET}", messages, ) # None means that there is no upper/lower limit for the output token count min_tokens = None if req_args.limit_min_tokens < 0 else req_args.limit_min_tokens max_tokens = None if req_args.limit_max_tokens < 0 else req_args.limit_max_tokens if len(conversation_messages) > messages_to_use: # The conversation contains an assistant answer for the next user prompt if ( min_tokens == NUM_TOKENS_FROM_DATASET or max_tokens == NUM_TOKENS_FROM_DATASET ): # Compute number of tokens in the answer (from the input conversation) assistant_answer = conversation_messages[messages_to_use] answer_num_tokens = get_token_count(tokenizer, assistant_answer["content"]) assert assistant_answer["role"] == "assistant" if min_tokens == NUM_TOKENS_FROM_DATASET: min_tokens = max(1, answer_num_tokens) if max_tokens == NUM_TOKENS_FROM_DATASET: max_tokens = max(1, answer_num_tokens) # Send the current conversation to LLM and get a response response: ServerResponse = await send_request( session, messages, req_args.chat_url, req_args.model, req_args.stream, min_tokens, max_tokens, req_args.timeout_sec, ) if response.valid is False: # Request failed return None # Compute number of tokens in input / output input_num_tokens = get_messages_token_count(tokenizer, messages) # Num tokens in the user's last question question_num_tokens = get_token_count(tokenizer, messages[index]["content"]) # Num tokens in the history/context of the question assert input_num_tokens >= question_num_tokens history_num_tokens = input_num_tokens - question_num_tokens # Num tokens in the LLM's answer (first chunk and full answer) first_chunk_tokens = get_token_count(tokenizer, response.first_chunk) output_content = response.content output_num_tokens = get_token_count(tokenizer, output_content) # Prefix caching approximated cached percent approx_cached_percent = ( 100.0 * (history_num_tokens / input_num_tokens) if input_num_tokens > 0 else 0.0 ) # Compute the correct TTFT and TPOT (based on tokens and not chunks). # Required because multiple output tokens may be bundled in a single chunk. if output_num_tokens > 1 and output_num_tokens > first_chunk_tokens: # More than one token and more than one chunk in the output decode_ms = response.latency_ms - response.ttft_ms decode_num_tokens = output_num_tokens - first_chunk_tokens tpot_ms = decode_ms / decode_num_tokens else: # In this case: output_num_tokens == first_chunk_tokens # Output was a single chunk (output_num_tokens > 1) # or even a single token (output_num_tokens == 1) tpot_ms = 0.0 if first_chunk_tokens > 1: # First chunk had multiple tokens, adjust TTFT for a single token delta_ms = (first_chunk_tokens - 1) * tpot_ms ttft_ms = max(0.1, response.ttft_ms - delta_ms) else: # First chunk had only one token ttft_ms = response.ttft_ms rs = RequestStats( ttft_ms=ttft_ms, tpot_ms=tpot_ms, latency_ms=response.latency_ms, start_time_ms=response.start_time_ms, input_num_turns=len(messages), input_num_tokens=input_num_tokens, output_num_tokens=output_num_tokens, output_num_chunks=response.num_chunks, output_num_first_chunk_tokens=first_chunk_tokens, approx_cached_percent=approx_cached_percent, conversation_id=conv_id, client_id=client_id, ) if verbose: print( f"\n{Color.YELLOW}Response ({output_num_tokens} tokens):{Color.RESET}", output_content, ) print(f"{Color.YELLOW}Response metrics: {rs}{Color.RESET}") print("-" * 70) # Save the LLM's answer (will be used as part of the context for the next user turn) answer_index = messages_to_use if len(conversation_messages) > answer_index: assert conversation_messages[answer_index]["role"] == "assistant", ( f"Failed on conversation ID {conv_id}, message role should be assistant" ) orig_content = conversation_messages[answer_index]["content"] if verify_output: # Compare the new answer to the answer from the input file debug_info = ( f"LLM/dataset answers do not match ({conv_id}):" f"\n'{get_short_string(output_content)}' (len: {len(output_content)})," f"\n'{get_short_string(orig_content)}' (len: {len(orig_content)})" ) if orig_content != output_content: raise ValueError(debug_info) # Update the answer conversation_messages[answer_index]["content"] = output_content else: # A user prompt that has no answer, add the answer as a new message new_answer = {"role": "assistant", "content": output_content} conversation_messages.append(new_answer) return rs async def poisson_sleep(request_rate: float, verbose: bool = False) -> None: # Generate a random time interval from the Poisson distribution assert request_rate > 0 interval = np.random.exponential(1.0 / request_rate) if verbose: logger.info(f"Sleeping for {interval:.3f} seconds...") await asyncio.sleep(interval) async def exponential_backoff_sleep( attempt_cnt: int, base_rate: float = 1.0, backoff_factor: float = 2.0, jitter_fraction: float = 0.10, verbose: bool = False, ) -> None: # Sleep with exponential backoff and jitter after a failed request. backoff_delay = base_rate * (backoff_factor**attempt_cnt) jittered_delay = backoff_delay * ( 1 + np.random.uniform(-jitter_fraction, jitter_fraction) ) if verbose: logger.info(f"Backoff for {jittered_delay:.3f} seconds...") await asyncio.sleep(jittered_delay) async def client_main( args: ClientArgs, req_args: RequestArgs, client_id: int, tokenizer: AutoTokenizer, stop_event: mp.Event, # type: ignore task_queue: mp.Queue, result_queue: mp.Queue, conv_queue: mp.Queue, ) -> None: logger.info( f"{Color.CYAN}Started client {client_id}: max_num_requests={args.max_num_requests}, max_active_conversations={args.max_active_conversations}{Color.RESET}" # noqa: E501 ) # Set unique seed per client (each client runs in its own process) # Add 1 to ensure no client uses the same seed as the main process client_seed = args.seed + client_id + 1 random.seed(client_seed) np.random.seed(client_seed) # Active conversations active_convs: ConversationsMap = {} conv_id_queue: deque = deque(maxlen=args.max_active_conversations) # Keep track of how many messages have been used for each conversation turns_count: Counter = Counter() num_successes = 0 num_failures = 0 # Track the timestamp (time.perf_counter()) # of the last turn per conversation (only for debug) time_of_last_turn: dict[ConvId, float] = {} # Flag that indicates that there are no new tasks (conversations) for the client task_queue_empty = False async with aiohttp.ClientSession() as session: # Print progress while task_queue_empty is False: result = None if ( args.max_num_requests and num_successes + num_failures == args.max_num_requests ): logger.info( f"{Color.YELLOW}Client {client_id} reached " f"request limit{Color.RESET}" ) break if stop_event.is_set(): # type: ignore logger.info( f"{Color.YELLOW}Client {client_id} received " f"a termination signal{Color.RESET}" ) break while ( len(active_convs) < args.max_active_conversations and task_queue_empty is False ): # Get a new conversation from the task queue conv_id, messages = task_queue.get() if conv_id is TERM_SIGNAL: task_queue_empty = True break if args.skip_first_turn: # Skip the first turn (both user and assistant), # relevant if warmup was enabled. # Default turns_count[conv_id] will be zero if conv_id # was never inserted/updated in turns_count. turns_count[conv_id] += 2 if turns_count[conv_id] < len(messages): # Add new conversation active_convs[conv_id] = messages conv_id_queue.append(conv_id) if args.verbose: logger.info( f"{Color.GREEN}Client {client_id} will use conversation ID {conv_id} (active conversations {len(active_convs)}){Color.RESET}" # noqa: E501 ) elif args.verbose: # No more messages (conversation finished during the warmup) logger.info( f"{Color.YELLOW}Client {client_id} will not use conversation ID {conv_id} (all {len(messages)} messages already sent){Color.RESET}" # noqa: E501 ) if len(active_convs) == 0 or task_queue_empty: logger.info( f"{Color.YELLOW}Client {client_id} has no more work{Color.RESET}" ) break # Pick an active conversation for the next request if args.conversation_sampling == ConversationSampling.ROUND_ROBIN: conv_id = conv_id_queue.pop() else: # ConversationSampling.RANDOM active_ids = list(active_convs.keys()) conv_id = random.choice(active_ids) messages = active_convs[conv_id] assert isinstance(messages, list) and len(messages) > 0 # Update the amount of messages to use turns_count[conv_id] += 1 current_turn = turns_count[conv_id] assert current_turn < len(messages), ( f"Turn number {current_turn} is invalid for conversation ID {conv_id}" f" that has only {len(messages)} messages" ) if args.verbose: curr_time_sec: float = time.perf_counter() time_since_last_turn: str | float = "N/A" if conv_id in time_of_last_turn: time_since_last_turn = round( curr_time_sec - time_of_last_turn[conv_id], 3 ) logger.info( f"Client {client_id} using conversation ID {conv_id} (turn: {current_turn}, time since last turn [sec]: {time_since_last_turn})" # noqa: E501 ) time_of_last_turn[conv_id] = curr_time_sec success = False for attempt_cnt in range(args.max_retries + 1): try: exception = False result = await send_turn( session, client_id, conv_id, messages, current_turn, tokenizer, req_args, args.print_content, args.verify_output, ) if result is not None: result_queue.put(result) success = True break else: logger.warning( f"{Color.YELLOW}Client {client_id} - Request rejected during conversation ID {conv_id} (turn: {current_turn}){Color.RESET}" # noqa: E501 ) except asyncio.exceptions.TimeoutError: exception = True logger.error( "%sClient %d - Timeout during conversation ID %s (turn: %d). " "Base timeout is %ss (set with --request-timeout-sec), but the " "effective timeout may be longer based on max_tokens. If this " "is unexpected, consider increasing the timeout or checking " "model performance.%s", Color.RED, client_id, conv_id, current_turn, req_args.timeout_sec, Color.RESET, ) except Exception: exception = True logger.exception( f"{Color.RED}Client {client_id} - Exception during conversation ID {conv_id} (turn: {current_turn}){Color.RESET}" # noqa: E501 ) # Sleep before retry if not last attempt if not success and attempt_cnt < args.max_retries: await exponential_backoff_sleep(attempt_cnt, verbose=args.verbose) if not success: num_failures += 1 # Remove the conversation (should not be used again) active_convs.pop(conv_id) if exception: break # Exit gracefully instead of raising an error else: num_successes += 1 # Update the turns counter to include the LLM response # The LLM response will be used as context for the next user turn turns_count[conv_id] += 1 max_turns = len(messages) if args.max_turns is not None: # Limit the number of turns in the conversation max_turns = min(args.max_turns, max_turns) if turns_count[conv_id] >= max_turns: # Conversation has no more turns (no longer active) # save the updated conversation (with the LLM server's answer) conv_queue.put((conv_id, active_convs.pop(conv_id))) if args.verbose: logger.info( f"{Color.GREEN}Client {client_id} finished " f"conversation ID {conv_id}{Color.RESET}" ) else: # Conversation is not finished, insert it at the back of the queue conv_id_queue.appendleft(conv_id) # Sleep between requests (if lambda is positive) if args.request_rate > 0: await poisson_sleep(args.request_rate, args.verbose) # Send indication that the client is done conv_queue.put((TERM_SIGNAL, TERM_SIGNAL)) logger.info( f"{Color.CYAN}Client {client_id} is done " f"({num_successes=}, {num_failures=}){Color.RESET}" ) def worker_function( client_id: int, tokenizer: AutoTokenizer, client_args: ClientArgs, req_args: RequestArgs, stop_event: mp.Event, # type: ignore task_queue: mp.Queue, result_queue: mp.Queue, conv_queue: mp.Queue, ) -> None: asyncio.run( client_main( client_args, req_args, client_id, tokenizer, stop_event, task_queue, result_queue, conv_queue, ) ) def get_client_config( args: argparse.Namespace, input_conv: ConversationsMap ) -> tuple[ClientArgs, RequestArgs]: if args.num_clients < 1: raise ValueError("Number of clients must be a positive number") if len(input_conv) < args.num_clients: raise ValueError( "Number of conversations must be equal or larger than the number of clients" ) max_req_per_client: int | None = None if args.max_num_requests is not None: # Max number of requests per client req_per_client = args.max_num_requests // args.num_clients if req_per_client < 1: raise ValueError("Number of requests should be at least one per client") max_req_per_client = req_per_client max_active_conversations = args.max_active_conversations if max_active_conversations is None: # Each client will have only one active conversation at a time max_active_conversations = args.num_clients if max_active_conversations > len(input_conv): raise ValueError( f"Max active conversations {max_active_conversations} " "must be equal or less than the total number of conversations" ) # Max number of active conversations per client max_active_conv_per_client = max_active_conversations // args.num_clients if max_active_conv_per_client < 1: raise ValueError( f"Max active conversations {max_active_conversations} " "must be equal or greater than the number of clients" ) # Skip the first user turn (as part of the warmup) skip_first_turn = args.warmup_step # Common arguments for all clients client_args = ClientArgs( seed=args.seed, max_num_requests=max_req_per_client, skip_first_turn=skip_first_turn, max_turns=args.max_turns, max_active_conversations=max_active_conv_per_client, verbose=args.verbose, print_content=args.print_content, verify_output=args.verify_output, conversation_sampling=args.conversation_sampling, request_rate=args.request_rate, max_retries=args.max_retries, ) if args.limit_min_tokens > 0 or args.limit_max_tokens > 0: if args.limit_min_tokens < 1 or args.limit_max_tokens < 1: raise ValueError( "Invalid min/max tokens limits (both limits should be provided)" ) if args.limit_min_tokens > args.limit_max_tokens: raise ValueError( "Invalid min/max tokens limits (min should not be larger than max)" ) if args.request_timeout_sec <= 0: raise ValueError("Request timeout must be a positive number") # Arguments for API requests chat_url = f"{args.url}/v1/chat/completions" model_name = args.served_model_name if args.served_model_name else args.model req_args = RequestArgs( chat_url=chat_url, model=model_name, stream=not args.no_stream, limit_min_tokens=args.limit_min_tokens, limit_max_tokens=args.limit_max_tokens, timeout_sec=args.request_timeout_sec, ) return client_args, req_args async def main_mp( client_args: ClientArgs, req_args: RequestArgs, bench_args: BenchmarkArgs, tokenizer: AutoTokenizer, input_conv: ConversationsMap, ) -> tuple[ConversationsMap, list[RequestStats]]: # An event that will trigger graceful termination of all the clients stop_event = mp.Event() # Queue for input conversations (from the input file/dataset) task_queue: mp.Queue = mp.Queue() # Queue for client measurements (TTFT, TPOT, etc. for each request) result_queue: mp.Queue = mp.Queue() # Queue for output conversations (with the LLM answers, sent by the server) conv_queue: mp.Queue = mp.Queue() output_conv: ConversationsMap = {} client_metrics: list[RequestStats] = [] # Start all clients start_time = time.perf_counter_ns() logger.info(f"{Color.GREEN}Starting {bench_args.num_clients} clients{Color.RESET}") clients = [] for client_id in range(bench_args.num_clients): client = mp.Process( name=f"client_{client_id}", target=worker_function, args=( client_id, tokenizer, client_args, req_args, stop_event, task_queue, result_queue, conv_queue, ), ) clients.append(client) client.start() # Submit all the input conversations as tasks for the clients for conv_id, messages in input_conv.items(): task_queue.put((conv_id, messages)) # Add termination signals for clients for _ in range(bench_args.num_clients): task_queue.put((TERM_SIGNAL, TERM_SIGNAL))
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/cutlass_benchmarks/sparse_benchmarks.py
benchmarks/cutlass_benchmarks/sparse_benchmarks.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import copy import itertools import pickle as pkl import time from collections.abc import Callable, Iterable import torch import torch.utils.benchmark as TBenchmark from torch.utils.benchmark import Measurement as TMeasurement from utils import make_rand_sparse_tensors from weight_shapes import WEIGHT_SHAPES from vllm import _custom_ops as ops from vllm.utils.argparse_utils import FlexibleArgumentParser DEFAULT_MODELS = list(WEIGHT_SHAPES.keys()) DEFAULT_BATCH_SIZES = [1, 16, 32, 64, 128, 256, 512] DEFAULT_TP_SIZES = [1] # bench def bench_fn( label: str, sub_label: str, description: str, fn: Callable, *args, **kwargs ) -> TMeasurement: min_run_time = 1 globals = { "args": args, "kwargs": kwargs, "fn": fn, } return TBenchmark.Timer( stmt="fn(*args, **kwargs)", globals=globals, label=label, sub_label=sub_label, description=description, ).blocked_autorange(min_run_time=min_run_time) def bench_int8( dtype: torch.dtype, m: int, k: int, n: int, label: str, sub_label: str ) -> Iterable[TMeasurement]: assert dtype == torch.int8 b_compressed, e, a, b = make_rand_sparse_tensors(torch.int8, m, n, k) scale_a = torch.tensor(1.0, device="cuda", dtype=torch.float32) scale_b = torch.tensor(1.0, device="cuda", dtype=torch.float32) bias = torch.zeros((n,), device="cuda", dtype=torch.bfloat16) out = ops.cutlass_scaled_sparse_mm( a, b_compressed, e, scale_a, scale_b, torch.bfloat16 ) out_ref = ops.cutlass_scaled_mm(a, b, scale_a, scale_b, torch.bfloat16) if not torch.allclose(out, out_ref): print("Incorrect results") print(out) print(out_ref) else: print("Correct results") timers = [] # pytorch impl - bfloat16 timers.append( bench_fn( label, sub_label, "pytorch_bf16_bf16_bf16_matmul-no-scales", torch.mm, a.to(dtype=torch.bfloat16), b.to(dtype=torch.bfloat16), ) ) # pytorch impl - float16 timers.append( bench_fn( label, sub_label, "pytorch_fp16_fp16_fp16_matmul-no-scales", torch.mm, a.to(dtype=torch.float16), b.to(dtype=torch.float16), ) ) # cutlass impl timers.append( bench_fn( label, sub_label, "cutlass_i8_i8_bf16_scaled_mm", ops.cutlass_scaled_mm, a, b, scale_a, scale_b, torch.bfloat16, ) ) # cutlass with bias timers.append( bench_fn( label, sub_label, "cutlass_i8_i8_bf16_scaled_mm_bias", ops.cutlass_scaled_mm, a, b, scale_a, scale_b, torch.bfloat16, bias, ) ) # cutlass sparse impl timers.append( bench_fn( label, sub_label, "cutlass_i8_i8_bf16_scaled_sparse_mm", ops.cutlass_scaled_sparse_mm, a, b_compressed, e, scale_a, scale_b, torch.bfloat16, ) ) # cutlass sparse with bias timers.append( bench_fn( label, sub_label, "cutlass_i8_i8_bf16_scaled_sparse_mm_bias", ops.cutlass_scaled_sparse_mm, a, b_compressed, e, scale_a, scale_b, torch.bfloat16, bias, ) ) return timers def bench_fp8( dtype: torch.dtype, m: int, k: int, n: int, label: str, sub_label: str ) -> Iterable[TMeasurement]: assert dtype == torch.float8_e4m3fn b_compressed, e, a, b = make_rand_sparse_tensors(torch.float8_e4m3fn, m, n, k) scale_a = torch.tensor(1.0, device="cuda", dtype=torch.float32) scale_b = torch.tensor(1.0, device="cuda", dtype=torch.float32) bias = torch.zeros((n,), device="cuda", dtype=torch.bfloat16) out = ops.cutlass_scaled_sparse_mm( a, b_compressed, e, scale_a, scale_b, torch.bfloat16 ) out_ref = ops.cutlass_scaled_mm(a, b, scale_a, scale_b, torch.bfloat16) if not torch.allclose(out, out_ref): print("Incorrect results") print(out) print(out_ref) else: print("Correct results") timers = [] # pytorch impl w. bf16 timers.append( bench_fn( label, sub_label, "pytorch_bf16_bf16_bf16_matmul-no-scales", torch.mm, a.to(dtype=torch.bfloat16, device="cuda"), b.to(dtype=torch.bfloat16, device="cuda"), ) ) # pytorch impl: bf16 output, without fp8 fast accum timers.append( bench_fn( label, sub_label, "pytorch_fp8_fp8_bf16_scaled_mm", torch._scaled_mm, a, b, scale_a=scale_a, scale_b=scale_b, out_dtype=torch.bfloat16, ) ) # pytorch impl: bf16 output, with fp8 fast accum timers.append( bench_fn( label, sub_label, "pytorch_fp8_fp8_bf16_scaled_mm_fast_accum", torch._scaled_mm, a, b, scale_a=scale_a, scale_b=scale_b, out_dtype=torch.bfloat16, use_fast_accum=True, ) ) # pytorch impl: fp16 output, without fp8 fast accum timers.append( bench_fn( label, sub_label, "pytorch_fp8_fp8_fp16_scaled_mm", torch._scaled_mm, a, b, scale_a=scale_a, scale_b=scale_b, out_dtype=torch.float16, ) ) # pytorch impl: fp16 output, with fp8 fast accum timers.append( bench_fn( label, sub_label, "pytorch_fp8_fp8_fp16_scaled_mm_fast_accum", torch._scaled_mm, a, b, scale_a=scale_a, scale_b=scale_b, out_dtype=torch.float16, use_fast_accum=True, ) ) # cutlass impl: bf16 output timers.append( bench_fn( label, sub_label, "cutlass_fp8_fp8_bf16_scaled_mm", ops.cutlass_scaled_mm, a, b, scale_a, scale_b, torch.bfloat16, ) ) # cutlass impl: bf16 output timers.append( bench_fn( label, sub_label, "cutlass_fp8_fp8_bf16_scaled_sparse_mm", ops.cutlass_scaled_sparse_mm, a, b_compressed, e, scale_a, scale_b, torch.bfloat16, ) ) # cutlass impl: fp16 output timers.append( bench_fn( label, sub_label, "cutlass_fp8_fp8_fp16_scaled_sparse_mm", ops.cutlass_scaled_sparse_mm, a, b_compressed, e, scale_a, scale_b, torch.float16, ) ) # cutlass impl: bf16 output, with bias timers.append( bench_fn( label, sub_label, "cutlass_fp8_fp8_bf16_scaled_sparse_mm_bias", ops.cutlass_scaled_sparse_mm, a, b_compressed, e, scale_a, scale_b, torch.bfloat16, bias, ) ) # cutlass impl: fp16 output, with bias timers.append( bench_fn( label, sub_label, "cutlass_fp8_fp8_fp16_scaled_sparse_mm_bias", ops.cutlass_scaled_sparse_mm, a, b_compressed, e, scale_a, scale_b, torch.float16, bias.to(dtype=torch.float16), ) ) return timers def bench( dtype: torch.dtype, m: int, k: int, n: int, label: str, sub_label: str ) -> Iterable[TMeasurement]: if dtype == torch.int8: return bench_int8(dtype, m, k, n, label, sub_label) if dtype == torch.float8_e4m3fn: return bench_fp8(dtype, m, k, n, label, sub_label) raise ValueError("unsupported type") # runner def print_timers(timers: Iterable[TMeasurement]): compare = TBenchmark.Compare(timers) compare.print() def run( dtype: torch.dtype, MKNs: Iterable[tuple[int, int, int]] ) -> Iterable[TMeasurement]: results = [] for m, k, n in MKNs: timers = bench(dtype, m, k, n, f"scaled-{dtype}-gemm", f"MKN=({m}x{k}x{n})") print_timers(timers) results.extend(timers) return results # output makers def make_output( data: Iterable[TMeasurement], MKNs: Iterable[tuple[int, int, int]], base_description: str, timestamp=None, ): print(f"== All Results {base_description} ====") print_timers(data) # pickle all the results timestamp = int(time.time()) if timestamp is None else timestamp with open(f"{base_description}-{timestamp}.pkl", "wb") as f: pkl.dump(data, f) # argparse runners def run_square_bench(args): dim_sizes = list(range(args.dim_start, args.dim_end + 1, args.dim_increment)) MKNs = list(zip(dim_sizes, dim_sizes, dim_sizes)) data = run(args.dtype, MKNs) make_output(data, MKNs, f"square_bench-{args.dtype}") def run_range_bench(args): dim_sizes = list(range(args.dim_start, args.dim_end, args.dim_increment)) n = len(dim_sizes) Ms = [args.m_constant] * n if args.m_constant is not None else dim_sizes Ks = [args.k_constant] * n if args.k_constant is not None else dim_sizes Ns = [args.n_constant] * n if args.n_constant is not None else dim_sizes MKNs = list(zip(Ms, Ks, Ns)) data = run(args.dtype, MKNs) make_output(data, MKNs, f"range_bench-{args.dtype}") def run_model_bench(args): print("Benchmarking models:") for i, model in enumerate(args.models): print(f"[{i}] {model}") def model_shapes(model_name: str, tp_size: int) -> list[tuple[int, int]]: KNs = [] for KN, tp_split_dim in copy.deepcopy(WEIGHT_SHAPES[model_name]): KN[tp_split_dim] = KN[tp_split_dim] // tp_size KNs.append(KN) return KNs model_bench_data = [] models_tps = list(itertools.product(args.models, args.tp_sizes)) for model, tp_size in models_tps: Ms = args.batch_sizes KNs = model_shapes(model, tp_size) MKNs = [] for m in Ms: for k, n in KNs: MKNs.append((m, k, n)) data = run(args.dtype, MKNs) model_bench_data.append(data) # Print all results for data, model_tp in zip(model_bench_data, models_tps): model, tp_size = model_tp print(f"== Results {args.dtype} {model}-TP{tp_size} ====") print_timers(data) timestamp = int(time.time()) all_data = [] for d in model_bench_data: all_data.extend(d) # pickle all data with open(f"model_bench-{args.dtype}-{timestamp}.pkl", "wb") as f: pkl.dump(all_data, f) if __name__ == "__main__": def to_torch_dtype(dt): if dt == "int8": return torch.int8 if dt == "fp8": return torch.float8_e4m3fn raise ValueError("unsupported dtype") parser = FlexibleArgumentParser( description=""" Benchmark Cutlass GEMM. To run square GEMMs: python3 ./benchmarks/cutlass_benchmarks/sparse_benchmarks.py --dtype fp8 square_bench --dim-start 128 --dim-end 512 --dim-increment 64 To run constant N and K and sweep M: python3 ./benchmarks/cutlass_benchmarks/sparse_benchmarks.py --dtype fp8 range_bench --dim-start 128 --dim-end 512 --dim-increment 64 --n-constant 16384 --k-constant 16384 To run dimensions from a model: python3 ./benchmarks/cutlass_benchmarks/sparse_benchmarks.py --dtype fp8 model_bench --models meta-llama/Llama-2-7b-hf --batch-sizes 16 --tp-sizes 1 Output: - a .pkl file, that is a list of raw torch.benchmark.utils.Measurements for the pytorch and cutlass implementations for the various GEMMs. """, # noqa: E501 formatter_class=argparse.RawTextHelpFormatter, ) parser.add_argument( "--dtype", type=to_torch_dtype, required=True, help="Available options are ['int8', 'fp8']", ) subparsers = parser.add_subparsers(dest="cmd") square_parser = subparsers.add_parser("square_bench") square_parser.add_argument("--dim-start", type=int, required=True) square_parser.add_argument("--dim-end", type=int, required=True) square_parser.add_argument("--dim-increment", type=int, required=True) square_parser.set_defaults(func=run_square_bench) range_parser = subparsers.add_parser("range_bench") range_parser.add_argument("--dim-start", type=int, required=True) range_parser.add_argument("--dim-end", type=int, required=True) range_parser.add_argument("--dim-increment", type=int, required=True) range_parser.add_argument("--m-constant", type=int, default=None) range_parser.add_argument("--n-constant", type=int, default=None) range_parser.add_argument("--k-constant", type=int, default=None) range_parser.set_defaults(func=run_range_bench) model_parser = subparsers.add_parser("model_bench") model_parser.add_argument( "--models", nargs="+", type=str, default=DEFAULT_MODELS, choices=WEIGHT_SHAPES.keys(), ) model_parser.add_argument( "--tp-sizes", nargs="+", type=int, default=DEFAULT_TP_SIZES ) model_parser.add_argument( "--batch-sizes", nargs="+", type=int, default=DEFAULT_BATCH_SIZES ) model_parser.set_defaults(func=run_model_bench) args = parser.parse_args() args.func(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/cutlass_benchmarks/utils.py
benchmarks/cutlass_benchmarks/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Cutlass bench utils from collections.abc import Iterable import torch import vllm._custom_ops as ops def to_fp8(tensor: torch.Tensor) -> torch.Tensor: finfo = torch.finfo(torch.float8_e4m3fn) return torch.round(tensor.clamp(min=finfo.min, max=finfo.max)).to( dtype=torch.float8_e4m3fn ) def to_int8(tensor: torch.Tensor) -> torch.Tensor: return torch.round(tensor.clamp(min=-128, max=127)).to(dtype=torch.int8) def to_bf16(tensor: torch.Tensor) -> torch.Tensor: return tensor.to(dtype=torch.bfloat16) def to_fp16(tensor: torch.Tensor) -> torch.Tensor: return tensor.to(dtype=torch.float16) def make_rand_tensors( dtype: torch.dtype, m: int, n: int, k: int ) -> tuple[torch.Tensor, torch.Tensor]: a = torch.randn((m, k), device="cuda") * 5 b = torch.randn((n, k), device="cuda").t() * 5 if dtype == torch.int8: return to_int8(a), to_int8(b) if dtype == torch.float8_e4m3fn: return to_fp8(a), to_fp8(b) raise ValueError("unsupported dtype") def prune_to_2_4(tensor): # Reshape tensor to [N, 4] where N is number of groups of 4 original_shape = tensor.shape reshaped = tensor.reshape(-1, 4) # Get indices of top 2 absolute values in each group of 4 _, indices = torch.topk(torch.abs(reshaped), k=2, dim=1) # Create binary mask mask = torch.zeros_like(reshaped) mask.scatter_(dim=1, index=indices, src=torch.ones_like(indices, dtype=mask.dtype)) # Apply mask and reshape back pruned = reshaped * mask # Turn all -0.0 to 0.0 pruned[pruned == -0.0] = 0.0 return pruned.reshape(original_shape) def make_rand_sparse_tensors( dtype: torch.dtype, m: int, n: int, k: int ) -> tuple[torch.Tensor, torch.Tensor]: a = torch.randn((m, k), device="cuda") * 5 b = torch.randn((n, k), device="cuda").t() * 5 b = prune_to_2_4(b.t()).t() if dtype == torch.int8: a, b = to_int8(a), to_int8(b) elif dtype == torch.float8_e4m3fn: a, b = to_fp8(a), to_fp8(b) elif dtype == torch.float16: a, b = to_fp16(a), to_fp16(b) elif dtype == torch.bfloat16: a, b = to_bf16(a), to_bf16(b) else: raise ValueError("unsupported dtype") b_compressed, e = ops.cutlass_sparse_compress(b.t()) # Compressed B, Metadata, Original A, B return b_compressed, e, a, b def make_n_rand_sparse_tensors( num_tensors: int, dtype: torch.dtype, m: int, n: int, k: int ) -> tuple[Iterable[torch.Tensor], Iterable[torch.Tensor]]: ABs = [] for _ in range(num_tensors): b_comp, e, a, b = make_rand_sparse_tensors(dtype, m, n, k) if b_comp is not None: ABs.append(make_rand_sparse_tensors(dtype, m, n, k)) BComps, Es, As, Bs = zip(*ABs) return list(BComps), list(Es), list(As), list(Bs)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/cutlass_benchmarks/weight_shapes.py
benchmarks/cutlass_benchmarks/weight_shapes.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Weight Shapes are in the format # ([K, N], TP_SPLIT_DIM) # Example: # A shape of ([14336, 4096], 0) indicates the following GEMM shape, # - TP1 : K = 14336, N = 4096 # - TP2 : K = 7168, N = 4096 # A shape of ([4096, 6144], 1) indicates the following GEMM shape, # - TP1 : K = 4096, N = 6144 # - TP4 : K = 4096, N = 1536 # TP1 shapes WEIGHT_SHAPES = { "mistralai/Mistral-7B-v0.1": [ ([4096, 6144], 1), ([4096, 4096], 0), ([4096, 28672], 1), ([14336, 4096], 0), ], "meta-llama/Llama-2-7b-hf": [ ([4096, 12288], 1), ([4096, 4096], 0), ([4096, 22016], 1), ([11008, 4096], 0), ], "meta-llama/Llama-3-8b": [ ([4096, 6144], 1), ([4096, 4096], 0), ([4096, 28672], 1), ([14336, 4096], 0), ], "meta-llama/Llama-2-13b-hf": [ ([5120, 15360], 1), ([5120, 5120], 0), ([5120, 27648], 1), ([13824, 5120], 0), ], "meta-llama/Llama-2-70b-hf": [ ([8192, 10240], 1), ([8192, 8192], 0), ([8192, 57344], 1), ([28672, 8192], 0), ], }
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/cutlass_benchmarks/w8a8_benchmarks.py
benchmarks/cutlass_benchmarks/w8a8_benchmarks.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import copy import itertools import pickle as pkl import time from collections.abc import Callable, Iterable import torch import torch.utils.benchmark as TBenchmark from torch.utils.benchmark import Measurement as TMeasurement from utils import make_rand_tensors from weight_shapes import WEIGHT_SHAPES from vllm import _custom_ops as ops from vllm.model_executor.layers.quantization.utils.fp8_utils import ( w8a8_triton_block_scaled_mm, ) from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.math_utils import cdiv DEFAULT_MODELS = list(WEIGHT_SHAPES.keys()) DEFAULT_BATCH_SIZES = [1, 16, 32, 64, 128, 256, 512] DEFAULT_TP_SIZES = [1] # bench def bench_fn( label: str, sub_label: str, description: str, fn: Callable, *args, **kwargs ) -> TMeasurement: min_run_time = 1 globals = { "args": args, "kwargs": kwargs, "fn": fn, } return TBenchmark.Timer( stmt="fn(*args, **kwargs)", globals=globals, label=label, sub_label=sub_label, description=description, ).blocked_autorange(min_run_time=min_run_time) def bench_int8( dtype: torch.dtype, m: int, k: int, n: int, label: str, sub_label: str, bench_kernels: list[str] | None = None, ) -> Iterable[TMeasurement]: """Benchmark INT8-based kernels.""" assert dtype == torch.int8 a, b = make_rand_tensors(torch.int8, m, n, k) scale_a = torch.tensor(1.0, device="cuda", dtype=torch.float32) scale_b = torch.tensor(1.0, device="cuda", dtype=torch.float32) bias = torch.zeros((n,), device="cuda", dtype=torch.bfloat16) azp = torch.zeros((m,), device="cuda", dtype=torch.int32) azp_adj = torch.zeros((n,), device="cuda", dtype=torch.int32) bench_fns = { "pytorch_bf16_bf16_bf16_matmul-no-scales": lambda: torch.mm( a.to(dtype=torch.bfloat16), b.to(dtype=torch.bfloat16) ), "pytorch_fp16_fp16_fp16_matmul-no-scales": lambda: torch.mm( a.to(dtype=torch.float16), b.to(dtype=torch.float16) ), "cutlass_i8_i8_bf16_scaled_mm": lambda: ops.cutlass_scaled_mm( a, b, scale_a, scale_b, torch.bfloat16 ), "cutlass_i8_i8_bf16_scaled_mm_bias": lambda: ops.cutlass_scaled_mm( a, b, scale_a, scale_b, torch.bfloat16, bias ), "cutlass_i8_i8_bf16_scaled_mm_azp": lambda: ops.cutlass_scaled_mm_azp( a, b, scale_a, scale_b, torch.bfloat16, azp_adj ), "cutlass_i8_i8_bf16_scaled_mm_azp_bias": lambda: ops.cutlass_scaled_mm_azp( a, b, scale_a, scale_b, torch.bfloat16, azp_adj, None, bias ), "cutlass_i8_i8_bf16_scaled_mm_azp_pt": lambda: ops.cutlass_scaled_mm_azp( a, b, scale_a, scale_b, torch.bfloat16, azp_adj, azp ), "cutlass_i8_i8_bf16_scaled_mm_azp_pt_bias": lambda: ops.cutlass_scaled_mm_azp( a, b, scale_a, scale_b, torch.bfloat16, azp_adj, azp, bias ), } timers = [] for name, fn in bench_fns.items(): # If bench_kernels is None, run all. Otherwise, run only exact matches. if bench_kernels is None or name in bench_kernels: print(f"Running {name}") timers.append(bench_fn(label, sub_label, name, fn)) return timers def bench_fp8( dtype: torch.dtype, m: int, k: int, n: int, label: str, sub_label: str, bench_kernels: list[str] | None = None, ) -> Iterable[TMeasurement]: """Benchmark FP8-based kernels.""" assert dtype == torch.float8_e4m3fn a, b = make_rand_tensors(torch.float8_e4m3fn, m, n, k) a_cont = a.contiguous() scale_a = torch.tensor(1.0, device="cuda", dtype=torch.float32) scale_b = torch.tensor(1.0, device="cuda", dtype=torch.float32) block_scale_a = torch.rand((m, cdiv(k, 128)), device="cuda", dtype=torch.float32) block_scale_b = torch.rand( cdiv(k, 128), cdiv(n, 128), device="cuda", dtype=torch.float32 ) block_scale_a_M_major = block_scale_a.t().contiguous().t() block_scale_b_K_major = block_scale_b.t().contiguous().t() bias = torch.zeros((n,), device="cuda", dtype=torch.bfloat16) print(m, k, n) bench_fns = { "pytorch_bf16_bf16_bf16_matmul-no-scales": lambda: torch.mm( a.to(dtype=torch.bfloat16), b.to(dtype=torch.bfloat16) ), "pytorch_fp16_fp16_fp16_matmul-no-scales": lambda: torch.mm( a.to(dtype=torch.float16), b.to(dtype=torch.float16) ), "pytorch_fp8_fp8_fp16_scaled_mm": lambda: torch._scaled_mm( a, b, scale_a, scale_b, out_dtype=torch.float16 ), "pytorch_fp8_fp8_fp16_scaled_mm_fast_accum": lambda: torch._scaled_mm( a, b, scale_a, scale_b, out_dtype=torch.float16, use_fast_accum=True ), "pytorch_fp8_fp8_bf16_scaled_mm": lambda: torch._scaled_mm( a, b, scale_a, scale_b, out_dtype=torch.bfloat16 ), "pytorch_fp8_fp8_bf16_scaled_mm_fast_accum": lambda: torch._scaled_mm( a, b, scale_a, scale_b, out_dtype=torch.bfloat16, use_fast_accum=True ), "cutlass_fp8_fp8_bf16_scaled_mm": lambda: ops.cutlass_scaled_mm( a, b, scale_a, scale_b, torch.bfloat16 ), "cutlass_fp8_fp8_fp16_scaled_mm": lambda: ops.cutlass_scaled_mm( a, b, scale_a, scale_b, torch.float16 ), "cutlass_fp8_fp8_bf16_scaled_mm_bias": lambda: ops.cutlass_scaled_mm( a, b, scale_a, scale_b, torch.bfloat16, bias ), "cutlass_fp8_fp8_fp16_scaled_mm_bias": lambda: ops.cutlass_scaled_mm( a, b, scale_a, scale_b, torch.float16, bias.to(dtype=torch.float16) ), "triton_fp8_fp8_fp16_scaled_mm_blockwise": lambda: w8a8_triton_block_scaled_mm( a_cont, b.t(), block_scale_a, block_scale_b.t(), (128, 128) ), "cutlass_fp8_fp8_fp16_scaled_mm_blockwise": lambda: ops.cutlass_scaled_mm( a, b, block_scale_a_M_major, block_scale_b_K_major, torch.float16 ), } timers = [] for name, fn in bench_fns.items(): # If bench_kernels is None, run all. Otherwise, run only exact matches. if bench_kernels is None or name in bench_kernels: print(f"Running {name}") timers.append(bench_fn(label, sub_label, name, fn)) return timers def bench( dtype: torch.dtype, m: int, k: int, n: int, label: str, sub_label: str, bench_kernels: list[str] | None = None, ) -> Iterable[TMeasurement]: if dtype == torch.int8: return bench_int8(dtype, m, k, n, label, sub_label, bench_kernels) if dtype == torch.float8_e4m3fn: return bench_fp8(dtype, m, k, n, label, sub_label, bench_kernels) raise ValueError("unsupported type") # runner def print_timers(timers: Iterable[TMeasurement]): compare = TBenchmark.Compare(timers) compare.print() def run( dtype: torch.dtype, MKNs: Iterable[tuple[int, int, int]], bench_kernels: list[str] | None = None, ) -> Iterable[TMeasurement]: results = [] for m, k, n in MKNs: timers = bench( dtype, m, k, n, f"scaled-{dtype}-gemm", f"MKN=({m}x{k}x{n})", bench_kernels=bench_kernels, ) print_timers(timers) results.extend(timers) return results def make_output( data: Iterable[TMeasurement], MKNs: Iterable[tuple[int, int, int]], base_description: str, timestamp=None, ): print(f"== All Results {base_description} ====") print_timers(data) # pickle all the results timestamp = int(time.time()) if timestamp is None else timestamp with open(f"{base_description}-{timestamp}.pkl", "wb") as f: pkl.dump(data, f) def run_square_bench(args): dim_sizes = list(range(args.dim_start, args.dim_end + 1, args.dim_increment)) MKNs = list(zip(dim_sizes, dim_sizes, dim_sizes)) data = run(args.dtype, MKNs, bench_kernels=args.kernels) make_output(data, MKNs, f"square_bench-{args.dtype}") def run_range_bench(args): dim_sizes = list(range(args.dim_start, args.dim_end, args.dim_increment)) n = len(dim_sizes) Ms = [args.m_constant] * n if args.m_constant is not None else dim_sizes Ks = [args.k_constant] * n if args.k_constant is not None else dim_sizes Ns = [args.n_constant] * n if args.n_constant is not None else dim_sizes MKNs = list(zip(Ms, Ks, Ns)) data = run(args.dtype, MKNs, bench_kernels=args.kernels) make_output(data, MKNs, f"range_bench-{args.dtype}") def run_model_bench(args): print("Benchmarking models:") for i, model in enumerate(args.models): print(f"[{i}] {model}") def model_shapes(model_name: str, tp_size: int) -> list[tuple[int, int]]: KNs = [] for KN, tp_split_dim in copy.deepcopy(WEIGHT_SHAPES[model_name]): KN[tp_split_dim] = KN[tp_split_dim] // tp_size KNs.append(KN) return KNs model_bench_data = [] models_tps = list(itertools.product(args.models, args.tp_sizes)) for model, tp_size in models_tps: Ms = args.batch_sizes KNs = model_shapes(model, tp_size) MKNs = [] for m in Ms: for k, n in KNs: MKNs.append((m, k, n)) data = run(args.dtype, MKNs, bench_kernels=args.kernels) model_bench_data.append(data) # Print all results for data, model_tp in zip(model_bench_data, models_tps): model, tp_size = model_tp print(f"== Results {args.dtype} {model}-TP{tp_size} ====") print_timers(data) timestamp = int(time.time()) all_data = [] for d in model_bench_data: all_data.extend(d) # pickle all data with open(f"model_bench-{args.dtype}-{timestamp}.pkl", "wb") as f: pkl.dump(all_data, f) if __name__ == "__main__": def to_torch_dtype(dt): if dt == "int8": return torch.int8 if dt == "fp8": return torch.float8_e4m3fn raise ValueError("unsupported dtype") parser = FlexibleArgumentParser( description=""" Benchmark Cutlass GEMM. To run square GEMMs: python3 ./benchmarks/cutlass_benchmarks/w8a8_benchmarks.py --dtype fp8 square_bench --dim-start 128 --dim-end 512 --dim-increment 64 To run constant N and K and sweep M: python3 ./benchmarks/cutlass_benchmarks/w8a8_benchmarks.py --dtype fp8 range_bench --dim-start 128 --dim-end 512 --dim-increment 64 --n-constant 16384 --k-constant 16384 To run dimensions from a model: python3 ./benchmarks/cutlass_benchmarks/w8a8_benchmarks.py --dtype fp8 model_bench --models meta-llama/Llama-2-7b-hf --batch-sizes 16 --tp-sizes 1 Output: - a .pkl file, that is a list of raw torch.benchmark.utils.Measurements for the pytorch and cutlass implementations for the various GEMMs. """, # noqa: E501 formatter_class=argparse.RawTextHelpFormatter, ) parser.add_argument( "--dtype", type=to_torch_dtype, required=True, help="Available options are ['int8', 'fp8']", ) parser.add_argument( "--kernels", nargs="+", type=str, default=None, help="Exact names of the kernels to benchmark. If not set, runs all kernels.", ) subparsers = parser.add_subparsers(dest="cmd") square_parser = subparsers.add_parser("square_bench") square_parser.add_argument("--dim-start", type=int, required=True) square_parser.add_argument("--dim-end", type=int, required=True) square_parser.add_argument("--dim-increment", type=int, required=True) square_parser.set_defaults(func=run_square_bench) range_parser = subparsers.add_parser("range_bench") range_parser.add_argument("--dim-start", type=int, required=True) range_parser.add_argument("--dim-end", type=int, required=True) range_parser.add_argument("--dim-increment", type=int, required=True) range_parser.add_argument("--m-constant", type=int, default=None) range_parser.add_argument("--n-constant", type=int, default=None) range_parser.add_argument("--k-constant", type=int, default=None) range_parser.set_defaults(func=run_range_bench) model_parser = subparsers.add_parser("model_bench") model_parser.add_argument( "--models", nargs="+", type=str, default=DEFAULT_MODELS, choices=WEIGHT_SHAPES.keys(), ) model_parser.add_argument( "--tp-sizes", nargs="+", type=int, default=DEFAULT_TP_SIZES ) model_parser.add_argument( "--batch-sizes", nargs="+", type=int, default=DEFAULT_BATCH_SIZES ) model_parser.set_defaults(func=run_model_bench) args = parser.parse_args() args.func(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_layernorm.py
benchmarks/kernels/benchmark_layernorm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import time import torch from vllm.model_executor.layers.layernorm import RMSNorm from vllm.platforms import current_platform from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.torch_utils import STR_DTYPE_TO_TORCH_DTYPE @torch.inference_mode() def main( num_tokens: int, hidden_size: int, add_residual: bool, dtype: torch.dtype, seed: int = 0, do_profile: bool = False, num_warmup_iters: int = 5, num_iters: int = 100, ) -> None: current_platform.seed_everything(seed) torch.set_default_device("cuda") layer = RMSNorm(hidden_size).to(dtype=dtype) layer.weight.data.normal_(mean=1.0, std=0.1) scale = 1 / (2 * hidden_size) x = torch.randn(num_tokens, hidden_size, dtype=dtype) x *= scale residual = torch.randn_like(x) * scale if add_residual else None def run_cuda_benchmark(num_iters: int, profile: bool = False) -> float: torch.cuda.synchronize() if profile: torch.cuda.cudart().cudaProfilerStart() start_time = time.perf_counter() for _ in range(num_iters): layer(x, residual) torch.cuda.synchronize() end_time = time.perf_counter() if profile: torch.cuda.cudart().cudaProfilerStop() return (end_time - start_time) / num_iters # Warmup. print("Warming up...") run_benchmark = run_cuda_benchmark run_benchmark(num_iters=num_warmup_iters, profile=False) # Benchmark. if do_profile: latency = run_benchmark(num_iters=1, profile=True) else: latency = run_benchmark(num_iters=num_iters, profile=False) print(f"Kernel running time: {latency * 1000000:.3f} us") if __name__ == "__main__": parser = FlexibleArgumentParser(description="Benchmark the layernorm kernel.") parser.add_argument("--num-tokens", type=int, default=4096) parser.add_argument("--hidden-size", type=int, default=8192) parser.add_argument("--add-residual", action="store_true") parser.add_argument( "--dtype", type=str, choices=["half", "bfloat16", "float"], default="half" ) parser.add_argument("--seed", type=int, default=0) parser.add_argument("--profile", action="store_true") parser.add_argument("--num-warmup-iters", type=int, default=5) parser.add_argument( "--num-iters", type=int, default=100, help="Number of benchmark iterations. " "If --profile is set, this number is ignored", ) args = parser.parse_args() print(args) main( num_tokens=args.num_tokens, hidden_size=args.hidden_size, add_residual=args.add_residual, dtype=STR_DTYPE_TO_TORCH_DTYPE[args.dtype], seed=args.seed, do_profile=args.profile, num_warmup_iters=args.num_warmup_iters, num_iters=args.num_iters, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/bench_nvfp4_quant.py
benchmarks/kernels/bench_nvfp4_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import copy import itertools import torch from weight_shapes import WEIGHT_SHAPES from vllm import _custom_ops as ops from vllm.platforms import current_platform from vllm.scalar_type import scalar_types from vllm.triton_utils import triton from vllm.utils.flashinfer import flashinfer_fp4_quantize if not current_platform.has_device_capability(100): raise RuntimeError("NVFP4 requires compute capability of 10.0 (Blackwell)") FLOAT4_E2M1_MAX = scalar_types.float4_e2m1f.max() FLOAT8_E4M3_MAX = torch.finfo(torch.float8_e4m3fn).max PROVIDER_CFGS = { "vllm": dict(backend="vllm", enabled=True), "flashinfer": dict(backend="flashinfer", enabled=True), } _enabled = [k for k, v in PROVIDER_CFGS.items() if v["enabled"]] def compute_global_scale(tensor: torch.Tensor) -> torch.Tensor: """Compute global scale for FP4 quantization.""" amax = torch.abs(tensor).max().to(torch.float32) return FLOAT8_E4M3_MAX * FLOAT4_E2M1_MAX / amax @triton.testing.perf_report( triton.testing.Benchmark( x_names=["batch_size"], x_vals=[1, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096], x_log=False, line_arg="provider", line_vals=_enabled, line_names=_enabled, ylabel="us (lower is better)", plot_name="NVFP4 Input Quantization Latency (us)", args={}, ) ) def benchmark(batch_size, provider, N, K): M = batch_size device = "cuda" dtype = torch.bfloat16 # Create input tensor a = torch.randn((M, K), device=device, dtype=dtype) # Compute global scale for activation a_global_scale = compute_global_scale(a) quantiles = [0.5, 0.2, 0.8] cfg = PROVIDER_CFGS[provider] if cfg["backend"] == "vllm": # vLLM's FP4 quantization ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: ops.scaled_fp4_quant(a, a_global_scale), quantiles=quantiles, ) elif cfg["backend"] == "flashinfer": # FlashInfer's FP4 quantization # Use is_sf_swizzled_layout=True to match vLLM's output format ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: flashinfer_fp4_quantize( a, a_global_scale, is_sf_swizzled_layout=True ), quantiles=quantiles, ) # Convert ms to us for better readability at small batch sizes to_us = lambda t_ms: t_ms * 1000 return to_us(ms), to_us(max_ms), to_us(min_ms) def prepare_shapes(args): out = [] for model, tp_size in itertools.product(args.models, args.tp_sizes): for KN, tp_dim in copy.deepcopy(WEIGHT_SHAPES[model]): KN[tp_dim] //= tp_size KN.append(model) out.append(KN) return out def _test_accuracy_once(M: int, K: int, dtype: torch.dtype, device: str): """Test accuracy between vLLM and FlashInfer FP4 quantization.""" # Create input tensor a = torch.randn((M, K), device=device, dtype=dtype) # Compute global scale a_global_scale = compute_global_scale(a) # vLLM quantization vllm_fp4, vllm_scale = ops.scaled_fp4_quant(a, a_global_scale) # FlashInfer quantization (with swizzled layout to match vLLM's output) flashinfer_fp4, flashinfer_scale = flashinfer_fp4_quantize( a, a_global_scale, is_sf_swizzled_layout=True ) flashinfer_scale = flashinfer_scale.view(torch.float8_e4m3fn) # Compare outputs torch.testing.assert_close( vllm_fp4, flashinfer_fp4, ) print(f"M={M}, K={K}, dtype={dtype}: PASSED") def test_accuracy(): """Run accuracy tests across various shapes.""" print("\n" + "=" * 60) print("Running accuracy tests: vLLM vs FlashInfer") print("=" * 60) device = "cuda" dtype = torch.bfloat16 # Test various batch sizes and hidden dimensions Ms = [1, 1024] Ks = [4096] for M in Ms: for K in Ks: _test_accuracy_once(M, K, dtype, device) print("\nAll accuracy tests passed!") if __name__ == "__main__": parser = argparse.ArgumentParser( description="Benchmark NVFP4 quantization: vLLM vs FlashInfer" ) parser.add_argument( "--models", nargs="+", type=str, default=["meta-llama/Llama-3.1-8B-Instruct"], choices=list(WEIGHT_SHAPES.keys()), ) parser.add_argument("--tp-sizes", nargs="+", type=int, default=[1]) parser.add_argument( "--save-path", type=str, default=None, help="Path to save benchmark results", ) parser.add_argument( "--accuracy", action="store_true", help="Run accuracy tests", ) args = parser.parse_args() if args.accuracy: test_accuracy() for K, N, model in prepare_shapes(args): print(f"\n{model}, N={N} K={K}") benchmark.run( print_data=True, save_path=args.save_path, N=N, K=K, ) print("\nBenchmark finished!")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_moe_permute_unpermute.py
benchmarks/kernels/benchmark_moe_permute_unpermute.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse from typing import Any, TypedDict import ray import torch from transformers import AutoConfig from vllm.model_executor.layers.fused_moe.fused_moe import * from vllm.model_executor.layers.fused_moe.moe_permute_unpermute import ( _moe_permute, _moe_unpermute_and_reduce, moe_permute, moe_unpermute, ) from vllm.model_executor.layers.fused_moe.utils import _fp8_quantize from vllm.platforms import current_platform from vllm.utils.argparse_utils import FlexibleArgumentParser FP8_DTYPE = current_platform.fp8_dtype() class BenchmarkConfig(TypedDict): BLOCK_SIZE_M: int BLOCK_SIZE_N: int BLOCK_SIZE_K: int GROUP_SIZE_M: int num_warps: int num_stages: int def benchmark_permute( num_tokens: int, num_experts: int, hidden_size: int, topk: int, dtype: torch.dtype, use_fp8_w8a8: bool, use_int8_w8a16: bool, num_iters: int = 100, use_customized_permute: bool = False, ) -> float: # init_dtype = torch.float16 if use_fp8_w8a8 else dtype hidden_states = torch.randn(num_tokens, hidden_size, dtype=dtype) # output_hidden_states = torch.empty_like(hidden_states) if use_fp8_w8a8: align_block_size = 128 # deepgemm needs 128 m aligned block qhidden_states, scale = _fp8_quantize(hidden_states, None, None) else: align_block_size = None qhidden_states = hidden_states gating_output = torch.randn(num_iters, num_tokens, num_experts, dtype=torch.float32) input_gating = torch.randn(num_tokens, num_experts, dtype=torch.float32) topk_weights, topk_ids, token_expert_indices = fused_topk( qhidden_states, input_gating, topk, False ) def prepare(i: int): input_gating.copy_(gating_output[i]) def run(): if use_customized_permute: ( permuted_hidden_states, a1q_scale, first_token_off, inv_perm_idx, m_indices, ) = moe_permute( qhidden_states, a1q_scale=None, topk_ids=topk_ids, n_expert=num_experts, expert_map=None, align_block_size=align_block_size, ) else: ( permuted_hidden_states, a1q_scale, sorted_token_ids, expert_ids, inv_perm, ) = _moe_permute( qhidden_states, None, topk_ids, num_experts, None, align_block_size ) # JIT compilation & warmup run() torch.cuda.synchronize() # Capture 10 invocations with CUDA graph graph = torch.cuda.CUDAGraph() with torch.cuda.graph(graph): for _ in range(10): run() torch.cuda.synchronize() # Warmup for _ in range(5): graph.replay() torch.cuda.synchronize() start_event = torch.Event(enable_timing=True) end_event = torch.Event(enable_timing=True) latencies: list[float] = [] for i in range(num_iters): prepare(i) torch.cuda.synchronize() start_event.record() graph.replay() end_event.record() end_event.synchronize() latencies.append(start_event.elapsed_time(end_event)) avg = sum(latencies) / (num_iters * 10) * 1000 # us graph.reset() return avg def benchmark_unpermute( num_tokens: int, num_experts: int, hidden_size: int, topk: int, dtype: torch.dtype, use_fp8_w8a8: bool, use_int8_w8a16: bool, num_iters: int = 100, use_customized_permute: bool = False, ) -> float: # init_dtype = torch.float16 if use_fp8_w8a8 else dtype hidden_states = torch.randn(num_tokens, hidden_size, dtype=dtype) output_hidden_states = torch.empty_like(hidden_states) if use_fp8_w8a8: align_block_size = 128 # deepgemm needs 128 m aligned block qhidden_states, scale = _fp8_quantize(hidden_states, None, None) else: align_block_size = None qhidden_states = hidden_states input_gating = torch.randn(num_tokens, num_experts, dtype=torch.float32) topk_weights, topk_ids, token_expert_indices = fused_topk( qhidden_states, input_gating, topk, False ) def prepare(): if use_customized_permute: ( permuted_hidden_states, a1q_scale, first_token_off, inv_perm_idx, m_indices, ) = moe_permute( qhidden_states, a1q_scale=None, topk_ids=topk_ids, n_expert=num_experts, expert_map=None, align_block_size=align_block_size, ) # convert to fp16/bf16 as gemm output return ( permuted_hidden_states.to(dtype), first_token_off, inv_perm_idx, m_indices, ) else: ( permuted_qhidden_states, a1q_scale, sorted_token_ids, expert_ids, inv_perm, ) = _moe_permute( qhidden_states, None, topk_ids, num_experts, None, align_block_size ) # convert to fp16/bf16 as gemm output return ( permuted_qhidden_states.to(dtype), a1q_scale, sorted_token_ids, expert_ids, inv_perm, ) def run(input: tuple): if use_customized_permute: ( permuted_hidden_states, first_token_off, inv_perm_idx, m_indices, ) = input output = torch.empty_like(hidden_states) moe_unpermute( output, permuted_hidden_states, topk_weights, inv_perm_idx, first_token_off, ) else: ( permuted_hidden_states, a1q_scale, sorted_token_ids, expert_ids, inv_perm, ) = input _moe_unpermute_and_reduce( output_hidden_states, permuted_hidden_states, inv_perm, topk_weights, True, ) # JIT compilation & warmup input = prepare() run(input) torch.cuda.synchronize() # Capture 10 invocations with CUDA graph graph = torch.cuda.CUDAGraph() with torch.cuda.graph(graph): for _ in range(10): run(input) torch.cuda.synchronize() # Warmup for _ in range(5): graph.replay() torch.cuda.synchronize() start_event = torch.Event(enable_timing=True) end_event = torch.Event(enable_timing=True) latencies: list[float] = [] for i in range(num_iters): torch.cuda.synchronize() start_event.record() graph.replay() end_event.record() end_event.synchronize() latencies.append(start_event.elapsed_time(end_event)) avg = sum(latencies) / (num_iters * 10) * 1000 # us graph.reset() return avg @ray.remote(num_gpus=1) class BenchmarkWorker: def __init__(self, seed: int) -> None: torch.set_default_device("cuda") current_platform.seed_everything(seed) self.seed = seed # Get the device ID to allocate tensors and kernels # on the respective GPU. This is required for Ray to work # correctly with multi-GPU tuning on the ROCm platform. self.device_id = int(ray.get_gpu_ids()[0]) def benchmark( self, num_tokens: int, num_experts: int, hidden_size: int, topk: int, dtype: torch.dtype, use_fp8_w8a8: bool, use_int8_w8a16: bool, use_customized_permute: bool = False, ) -> tuple[dict[str, int], float]: current_platform.seed_everything(self.seed) permute_time = benchmark_permute( num_tokens, num_experts, hidden_size, topk, dtype, use_fp8_w8a8, use_int8_w8a16, num_iters=100, use_customized_permute=use_customized_permute, ) unpermute_time = benchmark_unpermute( num_tokens, num_experts, hidden_size, topk, dtype, use_fp8_w8a8, use_int8_w8a16, num_iters=100, use_customized_permute=use_customized_permute, ) return permute_time, unpermute_time def get_weight_block_size_safety(config, default_value=None): quantization_config = getattr(config, "quantization_config", {}) if isinstance(quantization_config, dict): return quantization_config.get("weight_block_size", default_value) return default_value def main(args: argparse.Namespace): print(args) config = AutoConfig.from_pretrained( args.model, trust_remote_code=args.trust_remote_code ) if config.architectures[0] == "DbrxForCausalLM": E = config.ffn_config.moe_num_experts topk = config.ffn_config.moe_top_k elif config.architectures[0] == "JambaForCausalLM": E = config.num_experts topk = config.num_experts_per_tok elif ( config.architectures[0] == "DeepseekV3ForCausalLM" or config.architectures[0] == "DeepseekV2ForCausalLM" or config.architectures[0] == "Glm4MoeForCausalLM" ): E = config.n_routed_experts topk = config.num_experts_per_tok elif config.architectures[0] in ["Qwen2MoeForCausalLM", "Qwen3MoeForCausalLM"]: E = config.num_experts topk = config.num_experts_per_tok else: # Support for llama4 config = config.get_text_config() # Default: Mixtral. E = config.num_local_experts topk = config.num_experts_per_tok hidden_size = config.hidden_size dtype = torch.float16 if current_platform.is_rocm() else config.dtype use_fp8_w8a8 = args.dtype == "fp8_w8a8" use_int8_w8a16 = args.dtype == "int8_w8a16" use_customized_permute = args.use_customized_permute if args.batch_size is None: batch_sizes = [ 1, 2, 4, 8, 16, 24, 32, 48, 64, 96, 128, 256, 512, 1024, 1536, 2048, 3072, 4096, ] else: batch_sizes = [args.batch_size] ray.init() num_gpus = int(ray.available_resources()["GPU"]) workers = [BenchmarkWorker.remote(args.seed) for _ in range(num_gpus)] def _distribute(method: str, inputs: list[Any]) -> list[Any]: outputs = [] worker_idx = 0 for input_args in inputs: worker = workers[worker_idx] worker_method = getattr(worker, method) output = worker_method.remote(*input_args) outputs.append(output) worker_idx = (worker_idx + 1) % num_gpus return ray.get(outputs) outputs = _distribute( "benchmark", [ ( batch_size, E, hidden_size, topk, dtype, use_fp8_w8a8, use_int8_w8a16, use_customized_permute, ) for batch_size in batch_sizes ], ) for batch_size, (permute, unpermute) in zip(batch_sizes, outputs): print(f"Batch size: {batch_size}") print(f"Permute time: {permute:.2f} us") print(f"Unpermute time: {unpermute:.2f} us") if __name__ == "__main__": parser = FlexibleArgumentParser() parser.add_argument( "--model", type=str, default="mistralai/Mixtral-8x7B-Instruct-v0.1" ) parser.add_argument( "--dtype", type=str, choices=["auto", "fp8_w8a8", "int8_w8a16"], default="auto" ) parser.add_argument("--use-customized-permute", action="store_true") parser.add_argument("--seed", type=int, default=0) parser.add_argument("--batch-size", type=int, required=False) parser.add_argument("--trust-remote-code", action="store_true") args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_machete.py
benchmarks/kernels/benchmark_machete.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import copy import itertools import math import os import pickle as pkl import time from collections.abc import Callable, Iterable from dataclasses import dataclass from itertools import product import pandas as pd import torch import torch.utils.benchmark as TBenchmark from torch.utils.benchmark import Measurement as TMeasurement from weight_shapes import WEIGHT_SHAPES from vllm import _custom_ops as ops from vllm.model_executor.layers.quantization.utils.marlin_utils import ( GPTQ_MARLIN_MAX_PARALLEL, GPTQ_MARLIN_MIN_THREAD_N, marlin_permute_scales, marlin_zero_points, ) from vllm.model_executor.layers.quantization.utils.marlin_utils_test import ( MarlinWorkspace, ) from vllm.model_executor.layers.quantization.utils.quant_utils import ( pack_rows, quantize_weights, ) from vllm.scalar_type import ScalarType, scalar_types from vllm.utils.argparse_utils import FlexibleArgumentParser DEFAULT_MODELS = ["meta-llama/Llama-3-8b", "meta-llama/Llama-2-70b-hf"] DEFAULT_BATCH_SIZES = [1, 16, 32, 64, 128, 256, 512, 1024] DEFAULT_TP_SIZES = [1] NVTX_PROFILE = os.environ.get("NVTX_PROFILE", False) if NVTX_PROFILE: import nvtx def terse_type_name(dt): return { torch.bfloat16: "bf16", torch.float16: "fp16", torch.int8: "int8", torch.float8_e4m3fn: "fp8", torch.float: "float", torch.int: "int", }[dt] @dataclass class BenchmarkTensors: w_ref: torch.Tensor a: torch.Tensor w_q: torch.Tensor group_size: int | None wtype: ScalarType w_g_s: torch.Tensor w_g_zp: torch.Tensor | None w_ch_s: torch.Tensor | None w_tok_s: torch.Tensor | None @dataclass class TypeConfig: act_type: torch.dtype weight_type: ScalarType output_type: torch.dtype | None group_scale_type: torch.dtype | None group_zero_type: torch.dtype | None channel_scale_type: torch.dtype | None token_scale_type: torch.dtype | None def rand_data(shape, dtype=torch.float16, scale=1): if dtype.is_floating_point: return (scale * torch.rand(shape, device="cuda") - 0.3).to(dtype) else: return torch.randint(-15, 15, shape, dtype=dtype, device="cuda") def quantize_and_pack( atype: torch.dtype, w: torch.Tensor, wtype: ScalarType, stype: torch.dtype | None, group_size: int | None, zero_points: bool = False, ): assert wtype.is_integer(), "TODO: support floating point weights" w_ref, w_q, w_s, w_zp = quantize_weights( w, wtype, group_size=group_size, zero_points=zero_points, # to match how the kernel applies zps ref_zero_points_after_scales=True, ) w_q = pack_rows(w_q, wtype.size_bits, *w_q.shape) return w_ref, w_q, w_s, w_zp def create_bench_tensors( shape: tuple[int, int, int], types: TypeConfig, group_size: int | None ) -> list[BenchmarkTensors]: m, n, k = shape # we want to make sure that weights don't fit into L2 cache between runs so # we construct enough weights to exceed L2 cache, which is 50mb on a H100 # so we target total weight size > 2*50mb num_weights = math.ceil( 2 * 50 * 1024**2 * 8 / (k * n * types.weight_type.size_bits) ) a = rand_data((m, k), types.act_type, scale=5) benchmark_tensors: list[BenchmarkTensors] = [] for _ in range(num_weights): w = rand_data((k, n), types.act_type, scale=5) if types.group_scale_type is not None: w = w.to(types.group_scale_type) if w.dtype.itemsize == 1: w = w.to(torch.float16) w_ref, w_q_packed, w_s, w_zp = quantize_and_pack( a.dtype, w, types.weight_type, types.group_scale_type, group_size, types.group_zero_type is not None, ) if not a.dtype.is_floating_point: aiinfo = torch.iinfo(a.dtype) w_ref = w_ref.round().clamp(aiinfo.min, aiinfo.max) w_ref = w_ref.to(torch.float32) w_ch_s = ( None if types.channel_scale_type is None else rand_data((n,), types.channel_scale_type) ) w_tok_s = ( None if types.token_scale_type is None else rand_data((m,), types.token_scale_type) ) benchmark_tensors.append( BenchmarkTensors( w_ref=w_ref, a=a, w_q=w_q_packed, wtype=types.weight_type, w_g_s=w_s, w_g_zp=w_zp, group_size=group_size, w_ch_s=w_ch_s, w_tok_s=w_tok_s, ) ) return benchmark_tensors def torch_matmul_f16_create_bench_fn(bt: BenchmarkTensors) -> Callable: a = bt.a w = bt.w_ref.to(bt.a.dtype) # use float reference tensor if a.dtype not in [torch.float16, torch.bfloat16]: a = a.to(torch.float16) w = w.to(torch.float16) return lambda: torch.matmul(a, w) def cutlass_scaled_mm_create_bench_fn(bt: BenchmarkTensors) -> Callable: if bt.w_ch_s is not None and bt.w_tok_s is not None: scale_a = bt.w_tok_s.to(torch.float32) scale_b = bt.w_ch_s.to(torch.float32) else: scale_a = torch.tensor(1.0, dtype=torch.float32, device=bt.a.device) scale_b = torch.tensor(1.0, dtype=torch.float32, device=bt.a.device) w_col_major = bt.w_ref.to(bt.a.dtype).t().contiguous().t() return lambda: ops.cutlass_scaled_mm( bt.a, w_col_major, scale_a, scale_b, out_dtype=torch.float16 ) def marlin_create_bench_fn(bt: BenchmarkTensors) -> Callable: device = bt.a.device workspace = MarlinWorkspace( bt.w_ref.shape[1], GPTQ_MARLIN_MIN_THREAD_N, GPTQ_MARLIN_MAX_PARALLEL ) if bt.w_g_zp is None: w_zp = torch.empty(0, dtype=torch.int, device=device) else: w_zp = marlin_zero_points( bt.w_g_zp, bt.w_ref.shape[0], bt.w_ref.shape[1], bt.wtype.size_bits ) if bt.group_size is None: w_s = torch.tensor([], device="cuda", dtype=torch.half) else: w_s = marlin_permute_scales( bt.w_g_s, bt.w_ref.shape[0], bt.w_ref.shape[1], bt.group_size ) sort_indices = torch.empty(0, dtype=torch.int, device=device) g_idx = torch.empty(0, dtype=torch.int, device=device) w_q = ops.gptq_marlin_repack( bt.w_q, sort_indices, bt.w_ref.shape[0], bt.w_ref.shape[1], bt.wtype.size_bits ) if bt.a.dtype.is_floating_point: assert bt.w_ch_s is None assert bt.w_tok_s is None assert bt.group_size is not None fn = lambda: ops.gptq_marlin_gemm( a=bt.a, c=None, b_q_weight=w_q, b_bias=None, b_scales=w_s, a_scales=None, global_scale=None, b_zeros=w_zp, g_idx=g_idx, perm=sort_indices, workspace=workspace.scratch, b_q_type=bt.wtype, size_m=bt.a.shape[0], size_n=bt.w_ref.shape[1], size_k=bt.w_ref.shape[0], is_k_full=True, is_zp_float=False, ) else: assert bt.a.dtype == torch.int8 assert bt.wtype == scalar_types.uint4b8 raise NotImplementedError("QQQ is not supported anymore") return fn def machete_create_bench_fn( bt: BenchmarkTensors, out_type=torch.dtype, schedule=None ) -> Callable: w_q = bt.w_q.t().contiguous().t() # make col major w_q = ops.machete_prepack_B( w_q, bt.a.dtype, bt.wtype, None if bt.w_g_s is None else bt.w_g_s.dtype ) w_g_zp = bt.w_g_zp if w_g_zp is not None: w_g_zp = -1 * bt.w_g_s * (w_g_zp.to(bt.w_g_s.dtype)) return lambda: ops.machete_mm( a=bt.a, b_q=w_q, b_type=bt.wtype, b_group_scales=bt.w_g_s, b_group_zeros=w_g_zp, b_group_size=bt.group_size, b_channel_scales=bt.w_ch_s, a_token_scales=bt.w_tok_s, out_type=out_type, schedule=schedule, ) def cutlass_w4a8_create_bench_fn( bt: BenchmarkTensors, out_type=torch.dtype, schedule=None ) -> Callable: w_q = bt.w_q.t().contiguous().t() # make col major w_q = ops.cutlass_encode_and_reorder_int4b(w_q) # expects fp8 scales w_s = ops.cutlass_pack_scale_fp8(bt.w_g_s.to(torch.float8_e4m3fn)) return lambda: ops.cutlass_w4a8_mm( a=bt.a, b_q=w_q, b_group_scales=w_s, b_group_size=bt.group_size, b_channel_scales=bt.w_ch_s, a_token_scales=bt.w_tok_s, maybe_schedule=schedule, ) # impl # bench def bench_fns(label: str, sub_label: str, description: str, fns: list[Callable]): min_run_time = 1 if not NVTX_PROFILE else 0.1 res = TBenchmark.Timer( stmt=""" for fn in fns: fn() """, globals={"fns": fns}, label=label, sub_label=sub_label, description=description, ).blocked_autorange(min_run_time=min_run_time) if NVTX_PROFILE: with ( nvtx.annotate("mm-bench"), nvtx.annotate(f"{label}|{sub_label}|{description}"), ): fns[0]() return res _SWEEP_SCHEDULES_RESULTS: pd.DataFrame | None = None _SWEEP_SCHEDULES_RESULTS_CSV: str | None = None def bench( types: TypeConfig, group_size: int, m: int, k: int, n: int, label: str, sub_label: str, sweep_schedules: bool = True, ) -> list[TMeasurement]: benchmark_tensors = create_bench_tensors((m, n, k), types, group_size) sub_label += f", L={len(benchmark_tensors)}" name_type_string = f"W{types.weight_type}" + f"-A{terse_type_name(types.act_type)}" if types.group_scale_type is not None: name_type_string += f"-GS{terse_type_name(types.group_scale_type)}" if types.group_zero_type is not None: name_type_string += f"-GZ{terse_type_name(types.group_zero_type)}" if group_size is not None: name_type_string += f"-G{group_size}" if types.channel_scale_type is not None: name_type_string += f"-CS{terse_type_name(types.channel_scale_type)}" if types.token_scale_type is not None: name_type_string += f"-TS{terse_type_name(types.token_scale_type)}" timers = [] # pytorch impl timers.append( bench_fns( label, sub_label, "torch.matmul (fp16)", [torch_matmul_f16_create_bench_fn(bt) for bt in benchmark_tensors], ) ) if types.act_type == torch.int8 or types.act_type == torch.float8_e4m3fn: timers.append( bench_fns( label, sub_label, f"cutlass_scaled_mm ({terse_type_name(types.act_type)})", [cutlass_scaled_mm_create_bench_fn(bt) for bt in benchmark_tensors], ) ) if types.act_type != torch.float8_e4m3fn: timers.append( bench_fns( label, sub_label, f"marlin ({name_type_string})", [marlin_create_bench_fn(bt) for bt in benchmark_tensors], ) ) # machete timers.append( bench_fns( label, sub_label, f"machete ({name_type_string})", [ machete_create_bench_fn(bt, out_type=types.output_type) for bt in benchmark_tensors ], ) ) # cutlass w4a8 if types.act_type == torch.float8_e4m3fn and group_size == 128: timers.append( bench_fns( label, sub_label, f"cutlass w4a8 ({name_type_string})", [ cutlass_w4a8_create_bench_fn(bt, out_type=types.output_type) for bt in benchmark_tensors ], ) ) if sweep_schedules: global _SWEEP_SCHEDULES_RESULTS print("Finding best schedule for machete") best = None best_schedule = None schedules = ops.machete_supported_schedules( a_type=types.act_type, b_type=types.weight_type, group_scales_type=types.group_scale_type, group_zeros_type=types.group_zero_type, token_scales_type=types.token_scale_type, channel_scales_type=types.channel_scale_type, out_type=types.output_type, ) if schedules is None or len(schedules) == 0: raise ValueError("No schedules found to sweep") for schedule in reversed(schedules): schedule_M = int(schedule.split("_")[0].split("x")[1]) # Prune known bad schedules if schedule_M >= 2 * max(m, 16) or schedule_M < m // 4: continue res = bench_fns( label, sub_label, "machete_best", [ machete_create_bench_fn( bt, out_type=types.output_type, schedule=schedule ) for bt in benchmark_tensors ], ) results_row = { "M": m, "K": k, "N": n, "group_size": group_size, "schedule": schedule, "median": res.median, } if _SWEEP_SCHEDULES_RESULTS is None: _SWEEP_SCHEDULES_RESULTS = pd.DataFrame(columns=results_row.keys()) _SWEEP_SCHEDULES_RESULTS.loc[len(_SWEEP_SCHEDULES_RESULTS)] = results_row print(f" {res.median:5.5} ", schedule) if not best or res.median < best.median: best = res best_schedule = schedule print("Best schedule:", best_schedule) timers.append(best) return timers # runner def print_timers(timers: list[TMeasurement]): compare = TBenchmark.Compare(timers) compare.print() def run(args, MKNs: Iterable[tuple[int, int, int]]) -> Iterable[TMeasurement]: types = TypeConfig( act_type=args.act_type, weight_type=scalar_types.uint4b8 if args.group_zero_type is None else scalar_types.uint4, output_type=args.out_type, group_scale_type=args.group_scale_type, group_zero_type=args.group_zero_type, channel_scale_type=args.channel_scale_type, token_scale_type=args.token_scale_type, ) results: list[TMeasurement] = [] for m, k, n in MKNs: timers = bench( types, args.group_size, m, k, n, f"{args.act_type}-gemm", f"MKN=({m}x{k}x{n})", sweep_schedules=args.sweep_schedules, ) print_timers(timers) results.extend(timers) return results # output makers def make_output( data: list[TMeasurement], MKNs: Iterable[tuple[int, int, int]], base_description: str, timestamp=None, ): print(f"== All Results {base_description} ====") print_timers(data) # pickle all the results timestamp = int(time.time()) if timestamp is None else timestamp with open(f"{base_description}-{timestamp}.pkl", "wb") as f: pkl.dump(data, f) # argparse runners def run_square_bench(args): dim_sizes = list(range(args.dim_start, args.dim_end + 1, args.dim_increment)) MKNs = list(zip(dim_sizes, dim_sizes, dim_sizes)) data = run(args.dtype, args.sweep_schedules, MKNs) make_output(data, MKNs, f"square_bench-{args.dtype}") def run_range_bench(args): m_start, k_start, n_start = (int(x) for x in args.dim_start.split(",")) m_end, k_end, n_end = (int(x) for x in args.dim_end.split(",")) m_increment, k_increment, n_increment = ( int(x) for x in args.dim_increment.split(",") ) Ms = list(range(m_start, m_end + 1, m_increment)) Ks = list(range(k_start, k_end + 1, k_increment)) Ns = list(range(n_start, n_end + 1, n_increment)) MKNs = list(product(Ms, Ks, Ns)) data = run(args.dtype, args.sweep_schedules, MKNs) make_output(data, MKNs, f"range_bench-{args.dtype}") def run_model_bench(args): print("Benchmarking models:") for i, model in enumerate(args.models): print(f"[{i}] {model}") def model_shapes(model_name: str, tp_size: int) -> list[tuple[int, int]]: KNs = [] for KN, tp_split_dim in copy.deepcopy(WEIGHT_SHAPES[model_name]): KN[tp_split_dim] = KN[tp_split_dim] // tp_size KNs.append(KN) return KNs model_bench_data = [] models_tps = list(itertools.product(args.models, args.tp_sizes)) for model, tp_size in models_tps: Ms = args.batch_sizes KNs = model_shapes(model, tp_size) MKNs = [] for m in Ms: for k, n in KNs: MKNs.append((m, k, n)) data = run(args, MKNs) model_bench_data.append(data) type_string = f"{args.act_type}" # Print all results for data, model_tp in zip(model_bench_data, models_tps): model, tp_size = model_tp print(f"== Results {type_string} {model}-TP{tp_size} ====") print_timers(data) timestr = time.strftime("%Y%m%d-%H%M%S") all_results = [] for d in model_bench_data: all_results.extend(d) # pickle all data with open(f"model_bench-{type_string}-{timestr}.pkl", "wb") as f: args_dict = vars(args) args_dict.pop("func") pkl.dump( { "args": args_dict, "results": all_results, }, f, ) if __name__ == "__main__": def to_torch_dtype(dt): return { "bfloat16": torch.bfloat16, "float16": torch.float16, "int8": torch.int8, "float8_e4m3fn": torch.float8_e4m3fn, "int": torch.int, "float": torch.float, }[dt] class ToTorchDtype(argparse.Action): def __call__(self, parser, namespace, values, option_string=None): setattr(namespace, self.dest, to_torch_dtype(values)) parser = FlexibleArgumentParser( description=""" Benchmark Machete GEMM. To run square GEMMs: python3 ./benchmarks/kernels/benchmark_machete.py --dtype float16 square_bench --dim-start 128 --dim-end 512 --dim-increment 64 To run constant N and K and sweep M: python3 ./benchmarks/kernels/benchmark_machete.py --dtype float16 range_bench --dim-start 128 --dim-end 512 --dim-increment 64 --n-constant 16384 --k-constant 16384 To run dimensions from a model: python3 ./benchmarks/kernels/benchmark_machete.py --dtype float16 model_bench --models meta-llama/Llama-2-7b-hf --batch-sizes 16 --tp-sizes 1 Output: - a .pkl file, that is a list of raw torch.benchmark.utils.Measurements for the pytorch and cutlass implementations for the various GEMMs. """, # noqa: E501 formatter_class=argparse.RawTextHelpFormatter, ) parser.add_argument( "--act-type", action=ToTorchDtype, required=True, choices=["bfloat16", "float16", "int8", "float8_e4m3fn"], ) parser.add_argument( "--group-scale-type", action=ToTorchDtype, choices=["bfloat16", "float16"], ) parser.add_argument( "--group-zero-type", type=to_torch_dtype, choices=["bfloat16", "float16"], ) parser.add_argument( "--channel-scale-type", action=ToTorchDtype, choices=["float"], ) parser.add_argument( "--token-scale-type", action=ToTorchDtype, choices=["float"], ) parser.add_argument( "--out-type", action=ToTorchDtype, choices=["bfloat16", "float16"], ) parser.add_argument( "--group-size", type=int, help="Available options are ['None', '-1', '128'], default=128", default=128, ) parser.add_argument( "--sweep-schedules", action="store_true", help="Run a sweep over all supported schedules", ) parser.add_argument( "--sweep-csv-out", help="CSV to store sweep results", default="sch_sweep_results.csv", ) subparsers = parser.add_subparsers(dest="cmd", required=True) square_parser = subparsers.add_parser("square_bench") square_parser.add_argument("--dim-start", type=int, required=True) square_parser.add_argument("--dim-end", type=int, required=True) square_parser.add_argument("--dim-increment", type=int, required=True) square_parser.set_defaults(func=run_square_bench) range_parser = subparsers.add_parser("range_bench") range_parser.add_argument( "--dim-start", type=str, required=True, help="Start value for M,K,N as common separated list", ) range_parser.add_argument( "--dim-end", type=str, required=True, help="End value (inclusive) for M,K,N as common separated list", ) range_parser.add_argument( "--dim-increment", type=str, required=True, help="Increment value for M,K,N as common separated list", ) range_parser.set_defaults(func=run_range_bench) model_parser = subparsers.add_parser("model_bench") model_parser.add_argument( "--models", nargs="+", type=str, default=DEFAULT_MODELS, choices=WEIGHT_SHAPES.keys(), ) model_parser.add_argument( "--tp-sizes", nargs="+", type=int, default=DEFAULT_TP_SIZES ) model_parser.add_argument( "--batch-sizes", nargs="+", type=int, default=DEFAULT_BATCH_SIZES ) model_parser.set_defaults(func=run_model_bench) args = parser.parse_args() _SWEEP_SCHEDULES_RESULTS_CSV = args.sweep_csv_out args.func(args) if _SWEEP_SCHEDULES_RESULTS is not None: _SWEEP_SCHEDULES_RESULTS.to_csv(_SWEEP_SCHEDULES_RESULTS_CSV)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/graph_machete_bench.py
benchmarks/kernels/graph_machete_bench.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import math import pickle from collections import defaultdict import matplotlib.pyplot as plt import pandas as pd import regex as re import seaborn as sns from torch.utils.benchmark import Measurement as TMeasurement from vllm.utils.argparse_utils import FlexibleArgumentParser if __name__ == "__main__": parser = FlexibleArgumentParser( description="Benchmark the latency of processing a single batch of " "requests till completion." ) parser.add_argument("filename", type=str) args = parser.parse_args() with open(args.filename, "rb") as f: data = pickle.load(f) raw_results: list[TMeasurement] = data["results"] results = defaultdict(lambda: list()) for v in raw_results: result = re.search(r"MKN=\(\d+x(\d+x\d+)\)", v.task_spec.sub_label) if result is not None: KN = result.group(1) else: raise Exception("MKN not found") result = re.search(r"MKN=\((\d+)x\d+x\d+\)", v.task_spec.sub_label) if result is not None: M = result.group(1) else: raise Exception("MKN not found") kernel = v.task_spec.description results[KN].append({"kernel": kernel, "batch_size": M, "median": v.median}) rows = int(math.ceil(len(results) / 2)) fig, axs = plt.subplots(rows, 2, figsize=(12, 5 * rows)) axs = axs.flatten() for axs_idx, (shape, data) in enumerate(results.items()): plt.sca(axs[axs_idx]) df = pd.DataFrame(data) sns.lineplot( data=df, x="batch_size", y="median", hue="kernel", style="kernel", markers=True, dashes=False, palette="Dark2", ) plt.title(f"Shape: {shape}") plt.ylabel("time (median, s)") plt.tight_layout() plt.savefig("graph_machete_bench.pdf")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_marlin.py
benchmarks/kernels/benchmark_marlin.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import torch.utils.benchmark as benchmark from benchmark_shapes import WEIGHT_SHAPES from vllm import _custom_ops as ops from vllm.model_executor.layers.quantization.gptq_marlin_24 import ( GPTQ_MARLIN_24_MAX_PARALLEL, GPTQ_MARLIN_24_MIN_THREAD_N, GPTQ_MARLIN_24_SUPPORTED_GROUP_SIZES, GPTQ_MARLIN_24_SUPPORTED_QUANT_TYPES, ) from vllm.model_executor.layers.quantization.utils.allspark_utils import ( ALLSPARK_AMPERE_M_CUBLAS_THRESHOLD, ALLSPARK_SUPPORTED_QUANT_TYPES, ) from vllm.model_executor.layers.quantization.utils.marlin_utils import ( GPTQ_MARLIN_MAX_PARALLEL, GPTQ_MARLIN_MIN_THREAD_N, MARLIN_SUPPORTED_GROUP_SIZES, query_marlin_supported_quant_types, ) from vllm.model_executor.layers.quantization.utils.marlin_utils_fp4 import ( FP4_MARLIN_SUPPORTED_GROUP_SIZES, rand_marlin_weight_fp4_like, ) from vllm.model_executor.layers.quantization.utils.marlin_utils_fp8 import ( marlin_quant_fp8_torch, ) from vllm.model_executor.layers.quantization.utils.marlin_utils_test import ( MarlinWorkspace, awq_marlin_quantize, marlin_quantize, ) from vllm.model_executor.layers.quantization.utils.marlin_utils_test_24 import ( marlin_24_quantize, ) from vllm.model_executor.layers.quantization.utils.quant_utils import ( gptq_pack, gptq_quantize_weights, quantize_weights, sort_weights, ) from vllm.scalar_type import ScalarType, scalar_types from vllm.utils.argparse_utils import FlexibleArgumentParser DEFAULT_MODELS = ["meta-llama/Llama-2-7b-hf/TP1"] DEFAULT_BATCH_SIZES = [1, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192] ACT_ORDER_OPTS = [False, True] K_FULL_OPTS = [False, True] def bench_run( results: list[benchmark.Measurement], model: str, act_order: bool, is_k_full: bool, quant_type: ScalarType, group_size: int, size_m: int, size_k: int, size_n: int, ): label = "Quant Matmul" sub_label = "{}, act={} k_full={}, q={}, g={}, MKN=({}x{}x{})".format( model, act_order, is_k_full, str(quant_type), group_size, size_m, size_k, size_n ) print(f"Testing: {sub_label}") a = torch.randn(size_m, size_k).to(torch.half).cuda() b = torch.rand(size_k, size_n).to(torch.half).cuda() has_zp = quant_type in [scalar_types.uint4, scalar_types.uint8] if act_order and (group_size == -1 or group_size == size_k or has_zp): return if size_k % group_size != 0: return marlin_24_supported = ( quant_type in GPTQ_MARLIN_24_SUPPORTED_QUANT_TYPES and group_size in GPTQ_MARLIN_24_SUPPORTED_GROUP_SIZES ) repack_supported = ( quant_type in GPTQ_MARLIN_24_SUPPORTED_QUANT_TYPES and group_size in MARLIN_SUPPORTED_GROUP_SIZES ) allspark_supported = ( quant_type in ALLSPARK_SUPPORTED_QUANT_TYPES and group_size == -1 and not act_order and is_k_full ) def gen_marlin_params(): # Marlin quant marlin_g_idx = marlin_sort_indices = marlin_zp = marlin_s2 = None if quant_type == scalar_types.float4_e2m1f: if group_size != 16 or act_order: return marlin_w_ref, marlin_q_w, marlin_s, marlin_s2 = rand_marlin_weight_fp4_like( b.T, group_size ) elif quant_type == scalar_types.float8_e4m3fn: if group_size not in [-1, 128] or act_order: return marlin_w_ref, marlin_q_w, marlin_s = marlin_quant_fp8_torch(b.T, group_size) elif group_size == 16: return elif has_zp: marlin_w_ref, marlin_q_w, marlin_s, marlin_zp = awq_marlin_quantize( b, quant_type, group_size ) else: marlin_w_ref, marlin_q_w, marlin_s, marlin_g_idx, marlin_sort_indices, _ = ( marlin_quantize(b, quant_type, group_size, act_order) ) return ( marlin_w_ref, marlin_q_w, marlin_s, marlin_s2, marlin_zp, marlin_g_idx, marlin_sort_indices, ) def gen_marlin_24_params(): marlin_24_w_ref = marlin_24_q_w_comp = marlin_24_meta = marlin_24_s = None if marlin_24_supported: (marlin_24_w_ref, marlin_24_q_w_comp, marlin_24_meta, marlin_24_s) = ( marlin_24_quantize(b, quant_type, group_size) ) return (marlin_24_w_ref, marlin_24_q_w_comp, marlin_24_meta, marlin_24_s) def gen_repack_params(): q_w_gptq = None repack_sort_indices = None if repack_supported: (w_ref, q_w, s, g_idx, rand_perm) = gptq_quantize_weights( b, quant_type, group_size, act_order ) q_w_gptq = gptq_pack(q_w, quant_type.size_bits, size_k, size_n) # For act_order, sort the "weights" and "g_idx" # so that group ids are increasing repack_sort_indices = torch.empty(0, dtype=torch.int, device=b.device) if act_order: (q_w, g_idx, repack_sort_indices) = sort_weights(q_w, g_idx) return q_w_gptq, repack_sort_indices def gen_allspark_params(): qw_reorder = s_reorder = zp_reorder = sm_count = sm_version = ( CUBLAS_M_THRESHOLD ) = None nonlocal allspark_supported if allspark_supported: properties = torch.cuda.get_device_properties(b.device.index) sm_count = properties.multi_processor_count sm_version = properties.major * 10 + properties.minor supported_arch = sm_version >= 80 and sm_version < 90 allspark_supported = allspark_supported and supported_arch if supported_arch: w_ref, qw, s, zp = quantize_weights(b, quant_type, group_size, has_zp) qw = qw.to(torch.uint8) qw_reorder, s_reorder, zp_reorder = ops.allspark_repack_weight( qw, s, zp, has_zp ) CUBLAS_M_THRESHOLD = ALLSPARK_AMPERE_M_CUBLAS_THRESHOLD return ( qw_reorder, s_reorder, zp_reorder, sm_count, sm_version, CUBLAS_M_THRESHOLD, ) ( marlin_w_ref, marlin_q_w, marlin_s, marlin_s2, marlin_zp, marlin_g_idx, marlin_sort_indices, ) = gen_marlin_params() marlin_24_w_ref, marlin_24_q_w_comp, marlin_24_meta, marlin_24_s = ( gen_marlin_24_params() ) q_w_gptq, repack_sort_indices = gen_repack_params() qw_reorder, s_reorder, zp_reorder, sm_count, sm_version, CUBLAS_M_THRESHOLD = ( gen_allspark_params() ) # Prepare marlin_workspace = MarlinWorkspace( size_n, GPTQ_MARLIN_MIN_THREAD_N, GPTQ_MARLIN_MAX_PARALLEL ) marlin_24_workspace = MarlinWorkspace( size_n, GPTQ_MARLIN_24_MIN_THREAD_N, GPTQ_MARLIN_24_MAX_PARALLEL ) globals = { # Gen params "quant_type": quant_type, "group_size": group_size, "size_m": size_m, "size_n": size_n, "size_k": size_k, "a": a, # Marlin params "marlin_w_ref": marlin_w_ref, "marlin_q_w": marlin_q_w, "marlin_s": marlin_s, "marlin_s2": marlin_s2, "marlin_zp": marlin_zp, "marlin_g_idx": marlin_g_idx, "marlin_sort_indices": marlin_sort_indices, "marlin_workspace": marlin_workspace, "is_k_full": is_k_full, # Marlin_24 params "marlin_24_w_ref": marlin_24_w_ref, "marlin_24_q_w_comp": marlin_24_q_w_comp, "marlin_24_meta": marlin_24_meta, "marlin_24_s": marlin_24_s, "marlin_24_workspace": marlin_24_workspace, # GPTQ params "q_w_gptq": q_w_gptq, "repack_sort_indices": repack_sort_indices, # AllSpark W8A16 params "qw_reorder": qw_reorder, "s_reorder": s_reorder, "zp_reorder": zp_reorder, "sm_count": sm_count, "sm_version": sm_version, "CUBLAS_M_THRESHOLD": CUBLAS_M_THRESHOLD, # Kernels "gptq_marlin_gemm": ops.gptq_marlin_gemm, "gptq_marlin_24_gemm": ops.gptq_marlin_24_gemm, "gptq_marlin_repack": ops.gptq_marlin_repack, "allspark_w8a16_gemm": ops.allspark_w8a16_gemm, } min_run_time = 1 # Warmup pytorch for _ in range(5): torch.matmul(a, marlin_w_ref) results.append( benchmark.Timer( stmt="torch.matmul(a, marlin_w_ref)", globals=globals, label=label, sub_label=sub_label, description="pytorch_gemm", ).blocked_autorange(min_run_time=min_run_time) ) results.append( benchmark.Timer( stmt="output = gptq_marlin_gemm(a, None, marlin_q_w, marlin_s, None, marlin_s2, marlin_zp, marlin_g_idx, marlin_sort_indices, marlin_workspace.scratch, quant_type, size_m, size_n, size_k, is_k_full, False, False, False)", # noqa: E501 globals=globals, label=label, sub_label=sub_label, description="gptq_marlin_gemm", ).blocked_autorange(min_run_time=min_run_time) ) results.append( benchmark.Timer( stmt="output = gptq_marlin_gemm(a, None, marlin_q_w, marlin_s, None, marlin_s2, marlin_zp, marlin_g_idx, marlin_sort_indices, marlin_workspace.scratch, quant_type, size_m, size_n, size_k, is_k_full, False, True, False)", # noqa: E501 globals=globals, label=label, sub_label=sub_label, description="gptq_marlin_gemm_fp32", ).blocked_autorange(min_run_time=min_run_time) ) if marlin_24_supported: results.append( benchmark.Timer( stmt="output = gptq_marlin_24_gemm(a, marlin_24_q_w_comp, marlin_24_meta, marlin_24_s, marlin_24_workspace.scratch, quant_type, size_m, size_n, size_k)", # noqa: E501 globals=globals, label=label, sub_label=sub_label, description="gptq_marlin_24_gemm", ).blocked_autorange(min_run_time=min_run_time) ) if repack_supported: results.append( benchmark.Timer( stmt="q_res = gptq_marlin_repack(q_w_gptq, repack_sort_indices, size_k, size_n, quant_type.size_bits)", # noqa: E501 globals=globals, label=label, sub_label=sub_label, description="gptq_marlin_repack", ).blocked_autorange(min_run_time=min_run_time) ) if allspark_supported: results.append( benchmark.Timer( stmt="output = allspark_w8a16_gemm(a, qw_reorder, s_reorder, zp_reorder, size_n, group_size, sm_count, sm_version, CUBLAS_M_THRESHOLD, False, True)", # noqa: E501 globals=globals, label=label, sub_label=sub_label, description="allspark_w8a16_gemm_fp32", ).blocked_autorange(min_run_time=min_run_time) ) def main(args): print("Benchmarking models:") for i, model in enumerate(args.models): print(f"[{i}] {model}") results: list[benchmark.Measurement] = [] for model in args.models: for layer in WEIGHT_SHAPES[model]: size_k = layer[0] size_n = layer[1] if len(args.limit_k) > 0 and size_k not in args.limit_k: continue if len(args.limit_n) > 0 and size_n not in args.limit_n: continue for act_order in ACT_ORDER_OPTS: if ( len(args.limit_act_order) > 0 and act_order not in args.limit_act_order ): continue for is_k_full in K_FULL_OPTS: if ( len(args.limit_k_full) > 0 and is_k_full not in args.limit_k_full ): continue for quant_type in query_marlin_supported_quant_types(): if ( len(args.limit_num_bits) > 0 and quant_type.size_bits not in args.limit_num_bits ): continue for group_size in ( MARLIN_SUPPORTED_GROUP_SIZES + FP4_MARLIN_SUPPORTED_GROUP_SIZES ): if ( len(args.limit_group_size) > 0 and group_size not in args.limit_group_size ): continue # For act_order, the group_size must be less than # size_k if act_order and (group_size == size_k or group_size == -1): continue for size_m in args.batch_sizes: bench_run( results, model, act_order, is_k_full, quant_type, group_size, size_m, size_k, size_n, ) compare = benchmark.Compare(results) compare.print() # For quick benchmarking use: # python benchmark_marlin.py --batch-sizes 1 16 32 --limit-k 4096 --limit-n 4096 --limit-group-size 128 --limit-num-bits 4 --limit-act-order 0 --limit-k-full 1 # noqa E501 # if __name__ == "__main__": parser = FlexibleArgumentParser( description="Benchmark Marlin across specified models/shapes/batches" ) parser.add_argument( "--models", nargs="+", type=str, default=DEFAULT_MODELS, choices=WEIGHT_SHAPES.keys(), ) parser.add_argument( "--batch-sizes", nargs="+", type=int, default=DEFAULT_BATCH_SIZES ) parser.add_argument("--limit-k", nargs="+", type=int, default=[]) parser.add_argument("--limit-n", nargs="+", type=int, default=[]) parser.add_argument("--limit-group-size", nargs="+", type=int, default=[]) parser.add_argument("--limit-num-bits", nargs="+", type=int, default=[]) parser.add_argument("--limit-act-order", nargs="+", type=int, default=[]) parser.add_argument("--limit-k-full", nargs="+", type=int, default=[]) args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_per_token_group_quant.py
benchmarks/kernels/benchmark_per_token_group_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import math from collections.abc import Callable from contextlib import contextmanager from unittest.mock import patch import torch from vllm.model_executor.layers.quantization.utils import fp8_utils, int8_utils from vllm.platforms import current_platform @contextmanager def _triton_mode(): """Temporarily force the Triton fallback path""" with patch("vllm.platforms.current_platform.is_cuda", return_value=False): yield def _time_cuda( fn: Callable[[], tuple[torch.Tensor, torch.Tensor]], warmup_iters: int, bench_iters: int, ) -> float: # warmup for _ in range(warmup_iters): fn() torch.cuda.synchronize() start = torch.Event(enable_timing=True) end = torch.Event(enable_timing=True) start.record() for _ in range(bench_iters): fn() end.record() torch.cuda.synchronize() return start.elapsed_time(end) / bench_iters # ms/iter def _run_single( shape: tuple[int, int], group_size: int, dtype: str, *, column_major: bool = False, scale_ue8m0: bool = False, warmup_iters: int, bench_iters: int, ) -> None: num_tokens, hidden_dim = shape device = torch.device("cuda") torch.manual_seed(42) x = torch.randn(num_tokens, hidden_dim, device=device, dtype=torch.bfloat16) * 8 if dtype == "fp8": def cuda_impl(): return fp8_utils.per_token_group_quant_fp8( x, group_size, column_major_scales=column_major, use_ue8m0=scale_ue8m0, ) def triton_impl(): with _triton_mode(): return fp8_utils.per_token_group_quant_fp8( x, group_size, column_major_scales=column_major, use_ue8m0=scale_ue8m0, ) elif dtype == "int8": def cuda_impl(): return int8_utils.per_token_group_quant_int8(x, group_size) def triton_impl(): with _triton_mode(): return int8_utils.per_token_group_quant_int8(x, group_size) else: raise ValueError("dtype must be 'fp8' or 'int8'") cuda_ms = _time_cuda(cuda_impl, warmup_iters, bench_iters) triton_ms = _time_cuda(triton_impl, warmup_iters, bench_iters) speedup = triton_ms / cuda_ms if cuda_ms else math.inf cfg_desc = ( f"shape={shape} gs={group_size:<3} col_major={column_major:<5} " f"ue8m0={scale_ue8m0:<5} dtype={dtype}" ) print( f"{cfg_desc:55} | CUDA {cuda_ms:7.3f} ms | Triton {triton_ms:7.3f} ms | " f"speed-up ×{speedup:5.2f}" ) def parse_args(): parser = argparse.ArgumentParser() parser.add_argument("--warmup-iters", type=int, default=10) parser.add_argument("--bench-iters", type=int, default=100) parser.add_argument("--dtype", choices=["fp8", "int8", "both"], default="both") return parser.parse_args() if __name__ == "__main__": if not current_platform.is_cuda(): raise RuntimeError("CUDA device is required to run this benchmark.") args = parse_args() warmup_iters, bench_iters = args.warmup_iters, args.bench_iters shapes = [(32, 128), (64, 256), (16, 512)] group_sizes = [64, 128] dtypes = ["fp8", "int8"] if args.dtype == "both" else [args.dtype] header = ( "Configuration".ljust(55) + " | " + "CUDA (ms)".center(12) + " | " + "Triton (ms)".center(13) + " | " + "Speed-up" ) print(header) print("-" * len(header)) for dtype in dtypes: for shape in shapes: for gs in group_sizes: if dtype == "fp8": for col_major in (False, True): for ue8m0 in (False, True): _run_single( shape, gs, dtype, column_major=col_major, scale_ue8m0=ue8m0, warmup_iters=warmup_iters, bench_iters=bench_iters, ) else: # INT8 has no col-major / ue8m0 switches _run_single( shape, gs, dtype, warmup_iters=warmup_iters, bench_iters=bench_iters, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_quant.py
benchmarks/kernels/benchmark_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import time import torch from vllm import _custom_ops as ops from vllm.platforms import current_platform from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.torch_utils import STR_DTYPE_TO_TORCH_DTYPE @torch.inference_mode() def main( num_tokens: int, hidden_size: int, static_scale: bool, quant_dtype: torch.dtype, dtype: torch.dtype, seed: int = 0, do_profile: bool = False, num_warmup_iters: int = 5, num_iters: int = 100, ) -> None: current_platform.seed_everything(seed) torch.set_default_device("cuda") x = torch.randn(num_tokens, hidden_size, dtype=dtype) scale = torch.randn(1, 1, dtype=torch.float32) if static_scale else None def run_cuda_benchmark(num_iters: int, profile: bool = False) -> float: torch.cuda.synchronize() if profile: torch.cuda.cudart().cudaProfilerStart() start_time = time.perf_counter() for _ in range(num_iters): if quant_dtype == torch.int8: ops.scaled_int8_quant(x, scale) else: ops.scaled_fp8_quant(x, scale) torch.cuda.synchronize() end_time = time.perf_counter() if profile: torch.cuda.cudart().cudaProfilerStop() return (end_time - start_time) / num_iters # Warmup. print("Warming up...") run_benchmark = run_cuda_benchmark run_benchmark(num_iters=num_warmup_iters, profile=False) # Benchmark. if do_profile: latency = run_benchmark(num_iters=1, profile=True) else: latency = run_benchmark(num_iters=num_iters, profile=False) print(f"Kernel running time: {latency * 1000000:.3f} us") if __name__ == "__main__": def to_torch_dtype(dt): if dt == "int8": return torch.int8 if dt == "fp8": return torch.float8_e4m3fn raise ValueError(f"Unsupported dtype: {dt}") parser = FlexibleArgumentParser( description="Benchmark the quantization (fp8 or int8) kernel." ) parser.add_argument("--num-tokens", type=int, default=4096) parser.add_argument("--hidden-size", type=int, default=8192) parser.add_argument("--static-scale", action="store_true") parser.add_argument( "--quant-dtype", type=str, choices=["fp8", "int8"], default="int8" ) parser.add_argument( "--dtype", type=str, choices=["half", "bfloat16", "float"], default="half" ) parser.add_argument("--seed", type=int, default=0) parser.add_argument("--profile", action="store_true") parser.add_argument("--num-warmup-iters", type=int, default=5) parser.add_argument( "--num-iters", type=int, default=100, help="Number of benchmark iterations. " "If --profile is set, this number is ignored", ) args = parser.parse_args() print(args) main( num_tokens=args.num_tokens, hidden_size=args.hidden_size, static_scale=args.static_scale, quant_dtype=to_torch_dtype(args.quant_dtype), dtype=STR_DTYPE_TO_TORCH_DTYPE[args.dtype], seed=args.seed, do_profile=args.profile, num_warmup_iters=args.num_warmup_iters, num_iters=args.num_iters, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_mrope.py
benchmarks/kernels/benchmark_mrope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # This script benchmarks the mrope kernel (mainly for Qwen2VL and Qwen2.5VL models). # It generates test data, runs benchmarks, and saves results to a CSV file. # # The CSV file (named with current date/time) contains these columns: # model_name, tp_size, num_tokens, num_heads, num_kv_heads, head_dim, max_position, # is_neox_style, rope_parameters, dtype, torch_mean, torch_median, torch_p99, # torch_min, torch_max, triton_mean, triton_median, triton_p99, triton_min, triton_max, # speedup # # == Usage Examples == # # Single model benchmark: # python3 benchmark_mrope.py --model-name Qwen/Qwen2-VL-7B-Instruct --tp-size 1 \ # --warmup-iter 10 --benchmark-iter 100 --dtype bfloat16 --seed 0 --num-tokens 1024 # # All models benchmark: # python3 benchmark_mrope.py --model-name "" --tp-size 1 --warmup-iter 10 \ # --benchmark-iter 100 --dtype bfloat16 --seed 0 --num-tokens 1024 # # All models with different TP sizes: # python3 benchmark_mrope.py --model-name "" --tp-size 1 2 4 8 --warmup-iter 10 \ # --benchmark-iter 100 --dtype bfloat16 --seed 0 --num-tokens 1024 # # All models with different token counts: # python3 benchmark_mrope.py --model-name "" --tp-size 1 --warmup-iter 10 \ # --benchmark-iter 100 --dtype bfloat16 --seed 0 --num-tokens 1024 4096 16384 import csv import os import time from datetime import datetime from typing import Any import numpy as np import torch from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.platforms import current_platform from vllm.transformers_utils.config import get_config from vllm.utils.argparse_utils import FlexibleArgumentParser device = torch.device("cuda" if torch.cuda.is_available() else "cpu") def generate_test_data( num_tokens: int, num_q_heads: int, num_kv_heads: int, head_size: int, max_position_embeddings: int, dtype: torch.dtype, device: torch.device, ): """Generate test data for given configuration.""" # Create 2D positions (3, num_tokens) for multimodal case positions = torch.randint( 0, max_position_embeddings // 4, (3, num_tokens), device=device ) # Create query and key tensors query = torch.randn(num_tokens, num_q_heads * head_size, dtype=dtype, device=device) key = torch.randn(num_tokens, num_kv_heads * head_size, dtype=dtype, device=device) return positions, query, key def calculate_stats(times: list[float]) -> dict[str, float]: """Calculate statistics from a list of times.""" times_array = np.array(times) return { "mean": np.mean(times_array), "median": np.median(times_array), "p99": np.percentile(times_array, 99), "min": np.min(times_array), "max": np.max(times_array), } def benchmark_mrope( model_name: str, num_tokens: int, head_dim: int, tp_size: int, num_heads: int, num_kv_heads: int, max_position: int = 8192, is_neox_style: bool = True, rope_parameters: dict[str, Any] | None = None, dtype: torch.dtype = torch.bfloat16, seed: int = 0, warmup_iter: int = 10, benchmark_iter: int = 100, csv_writer=None, ): current_platform.seed_everything(seed) torch.set_default_device(device) # the parameters to compute the q k v size based on tp_size mrope_helper_class = get_rope( head_size=head_dim, max_position=max_position, is_neox_style=is_neox_style, rope_parameters=rope_parameters, dtype=dtype, ).to(device=device) print(80 * "=") print( f"Evaluating model: {model_name} " f"with tp_size: {tp_size} " f"and num_tokens: {num_tokens}, " f"dtype: {dtype}" ) # create q k v input tensors # create rotary pos emb input tensors positions, query, key = generate_test_data( num_tokens, num_heads, num_kv_heads, head_dim, max_position, dtype, device ) # Warm up for _ in range(warmup_iter): mrope_helper_class.forward_native( positions, query.clone(), key.clone(), ) mrope_helper_class.forward_cuda( positions, query.clone(), key.clone(), ) torch.cuda.synchronize() # Time reference implementation torch_times = [] for _ in range(benchmark_iter): query_clone = query.clone() key_clone = key.clone() torch.cuda.synchronize() start_time = time.time() mrope_helper_class.forward_native( positions, query_clone, key_clone, ) torch.cuda.synchronize() torch_times.append(time.time() - start_time) # Time triton kernel implementation triton_times = [] for _ in range(benchmark_iter): query_clone = query.clone() key_clone = key.clone() torch.cuda.synchronize() start_time = time.time() mrope_helper_class.forward_cuda( positions, query_clone, key_clone, ) torch.cuda.synchronize() triton_times.append(time.time() - start_time) # Calculate statistics torch_stats = calculate_stats(torch_times) triton_stats = calculate_stats(triton_times) print(f"\nPerformance for config ({num_tokens}, {num_heads}, {num_kv_heads}):") print( f"Torch implementation: " f"mean={torch_stats['mean']:.8f}s, " f"median={torch_stats['median']:.8f}s, " f"p99={torch_stats['p99']:.8f}s" ) print( f"Triton implementation: " f"mean={triton_stats['mean']:.8f}s, " f"median={triton_stats['median']:.8f}s, " f"p99={triton_stats['p99']:.8f}s" ) print( f"Triton Speedup over Torch: {torch_stats['mean'] / triton_stats['mean']:.8f}x" ) # Write to CSV if csv_writer: row = [ model_name, tp_size, num_tokens, num_heads, num_kv_heads, head_dim, max_position, is_neox_style, str(rope_parameters), str(dtype).split(".")[-1], torch_stats["mean"], torch_stats["median"], torch_stats["p99"], torch_stats["min"], torch_stats["max"], triton_stats["mean"], triton_stats["median"], triton_stats["p99"], triton_stats["min"], triton_stats["max"], torch_stats["mean"] / triton_stats["mean"], # speedup ] csv_writer.writerow(row) return torch_stats, triton_stats if __name__ == "__main__": parser = FlexibleArgumentParser( description="Benchmark the rotary embedding kernels." ) parser.add_argument("--model-name", type=str, default="") parser.add_argument("--tp-size", type=int, default=1) parser.add_argument("--warmup-iter", type=int, default=10) parser.add_argument("--benchmark-iter", type=int, default=100) parser.add_argument("--dtype", type=str, choices=["bfloat16"], default="bfloat16") parser.add_argument("--seed", type=int, default=0) parser.add_argument("--num-tokens", type=int, nargs="+", required=False) parser.add_argument("--trust-remote-code", action="store_true") parser.add_argument("--output-csv", type=str, default="mrope_benchmark_results.csv") args = parser.parse_args() print(args) # Create CSV file for results timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") csv_filename = f"{os.path.splitext(args.output_csv)[0]}_{timestamp}.csv" with open(csv_filename, "w", newline="") as csvfile: csv_writer = csv.writer(csvfile) # Write header header = [ "model_name", "tp_size", "num_tokens", "num_heads", "num_kv_heads", "head_dim", "max_position", "is_neox_style", "rope_parameters", "dtype", "torch_mean", "torch_median", "torch_p99", "torch_min", "torch_max", "triton_mean", "triton_median", "triton_p99", "triton_min", "triton_max", "speedup", ] csv_writer.writerow(header) model_tp_dict = {} if args.model_name == "": model_tp_dict = { "Qwen/Qwen2-VL-2B-Instruct": [1], "Qwen/Qwen2-VL-7B-Instruct": [1], "Qwen/Qwen2-VL-72B-Instruct": [2, 4, 8], "Qwen/Qwen2.5-VL-3B-Instruct": [1, 2, 4, 8], "Qwen/Qwen2.5-VL-7B-Instruct": [1, 2, 4, 8], "Qwen/Qwen2.5-VL-72B-Instruct": [2, 4, 8], } else: model_tp_dict[args.model_name] = [args.tp_size] if args.num_tokens is None: num_tokens_list = [2**i for i in range(0, 18)] else: num_tokens_list = args.num_tokens for model_name, tp_list in model_tp_dict.items(): config = get_config(model_name, trust_remote_code=args.trust_remote_code) for tp_size in tp_list: # get the model config total_num_kv_heads = config.num_key_value_heads total_num_heads = config.num_attention_heads num_heads = total_num_heads // tp_size num_kv_heads = max(1, total_num_kv_heads // tp_size) head_dim = config.hidden_size // total_num_heads q_size = num_heads * head_dim kv_size = num_kv_heads * head_dim is_neox_style = True rope_parameters = config.rope_parameters max_position = config.max_position_embeddings for num_tokens in num_tokens_list: benchmark_mrope( model_name=model_name, num_tokens=num_tokens, head_dim=head_dim, tp_size=tp_size, num_heads=num_heads, num_kv_heads=num_kv_heads, max_position=max_position, is_neox_style=is_neox_style, rope_parameters=rope_parameters, dtype=getattr(torch, args.dtype), seed=args.seed, warmup_iter=args.warmup_iter, benchmark_iter=args.benchmark_iter, csv_writer=csv_writer, ) print(f"Benchmark results saved to {csv_filename}")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_trtllm_prefill_attention.py
benchmarks/kernels/benchmark_trtllm_prefill_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import csv import os from datetime import datetime import flashinfer import torch from vllm.utils.math_utils import round_up FLOAT32_BYTES = torch.finfo(torch.float).bits // 8 FP8_DTYPE = torch.float8_e4m3fn FP4_DTYPE = torch.uint8 def to_float8(x, dtype=torch.float8_e4m3fn): finfo = torch.finfo(dtype) min_val, max_val = x.aminmax() amax = torch.maximum(min_val.abs(), max_val.abs()).clamp(min=1e-12) scale = finfo.max / amax * 0.1 x_scl_sat = (x * scale).clamp(min=finfo.min, max=finfo.max) return x_scl_sat.to(dtype), scale.float().reciprocal() @torch.no_grad() def benchmark_prefill( dtype: torch.dtype, quant_dtypes: tuple[torch.dtype | None, torch.dtype | None, torch.dtype | None], batch_size: int, max_seq_len: int, num_heads: tuple[int, int] = (64, 8), head_size: int = 128, kv_layout: str = "HND", block_size: int = 16, warmup: int = 10, trials: int = 20, ): torch.set_default_device("cuda") torch.manual_seed(0) q_quant_dtype, kv_quant_dtype, o_quant_dtype = quant_dtypes q_quant_dtype = q_quant_dtype or dtype kv_quant_dtype = kv_quant_dtype or dtype o_quant_dtype = o_quant_dtype or dtype max_q_len = max_kv_len = max_seq_len num_qo_heads, num_kv_heads = num_heads assert num_qo_heads % num_kv_heads == 0 sm_scale = float(1.0 / (head_size**0.5)) # large number to reduce kv_cache reuse NUM_BLOCKS = int(256000 / block_size) kv_cache_shape = None if kv_layout == "NHD": kv_cache_shape = (NUM_BLOCKS, 2, block_size, num_kv_heads, head_size) elif kv_layout == "HND": kv_cache_shape = (NUM_BLOCKS, 2, num_kv_heads, block_size, head_size) else: raise ValueError(f"Invalid kv_layout: {kv_layout}") q_lens = torch.randint(1, max_q_len, (batch_size,), dtype=torch.int32) q_lens[-1] = max_q_len q_indptr = torch.cat( [ torch.tensor([0], dtype=torch.int32), torch.cumsum(q_lens, dim=0, dtype=torch.int32), ] ) # Always using 1.0 scale to reflect the real perf in benchmarking q_scale = 1.0 ref_query = torch.randn( torch.sum(q_lens).item(), num_qo_heads, head_size, dtype=dtype ) if q_quant_dtype == FP8_DTYPE: query, _ = to_float8(ref_query) else: query = ref_query kv_lens = torch.randint(0, max_kv_len, (batch_size,), dtype=torch.int32) kv_lens[-1] = max_kv_len seq_lens = kv_lens + q_lens max_seq_len = torch.max(seq_lens).item() # Always using 1.0 scale to reflect the real perf in benchmarking k_scale = v_scale = 1.0 ref_kv_cache = torch.randn(kv_cache_shape, dtype=dtype) if kv_quant_dtype == FP8_DTYPE: kv_cache, _ = to_float8(ref_kv_cache) else: kv_cache = ref_kv_cache max_num_blocks_per_seq = (max_seq_len + block_size - 1) // block_size block_tables = torch.randint( 0, NUM_BLOCKS, (batch_size, max_num_blocks_per_seq), dtype=torch.int32 ) kv_indptr = [0] kv_indices = [] kv_last_page_lens = [] for i in range(batch_size): seq_len = seq_lens[i] assert seq_len > 0 num_blocks = (seq_len + block_size - 1) // block_size kv_indices.extend(block_tables[i, :num_blocks]) kv_indptr.append(kv_indptr[-1] + num_blocks) kv_last_page_len = seq_len % block_size if kv_last_page_len == 0: kv_last_page_len = block_size kv_last_page_lens.append(kv_last_page_len) kv_indptr = torch.tensor(kv_indptr, dtype=torch.int32) kv_indices = torch.tensor(kv_indices, dtype=torch.int32) kv_last_page_lens = torch.tensor(kv_last_page_lens, dtype=torch.int32) workspace_buffer = torch.zeros(1024 * 1024 * 1024, dtype=torch.int8) wrapper = flashinfer.BatchPrefillWithPagedKVCacheWrapper( workspace_buffer, kv_layout ) wrapper.plan( q_indptr, kv_indptr, kv_indices, kv_last_page_lens, num_qo_heads, num_kv_heads, head_size, block_size, causal=True, sm_scale=sm_scale, q_data_type=dtype, kv_data_type=dtype, ) def time_fn(fn, warmup=10, trials=20): torch.cuda.synchronize() start = torch.Event(enable_timing=True) end = torch.Event(enable_timing=True) times = [] for i in range(warmup): fn() for i in range(trials): start.record() fn() end.record() torch.cuda.synchronize() times.append(start.elapsed_time(end)) # ms return sum(times) / len(times), torch.std(torch.tensor(times)) o_scale = 1.0 o_sf_scale = None output_baseline = torch.empty(ref_query.shape, dtype=dtype) if o_quant_dtype == FP4_DTYPE: o_sf_scale = 500.0 output_trtllm = flashinfer.utils.FP4Tensor( torch.empty(query.shape[:-1] + (query.shape[-1] // 2,), dtype=torch.uint8), torch.empty( ( round_up(query.shape[0], 128), round_up(query.shape[1] * query.shape[2] // 16, 4), ), dtype=torch.float8_e4m3fn, ), ) else: output_trtllm = torch.empty(query.shape, dtype=o_quant_dtype) def baseline_prefill(): return wrapper.run( ref_query, ref_kv_cache, k_scale=k_scale, v_scale=v_scale, out=output_baseline, ) def trtllm_prefill(): return flashinfer.prefill.trtllm_batch_context_with_kv_cache( query=query, kv_cache=kv_cache, workspace_buffer=workspace_buffer, block_tables=block_tables, seq_lens=seq_lens, max_q_len=max_q_len, max_kv_len=max_seq_len, bmm1_scale=q_scale * k_scale * sm_scale, bmm2_scale=v_scale / o_scale, batch_size=batch_size, cum_seq_lens_q=q_indptr, cum_seq_lens_kv=kv_indptr, o_sf_scale=o_sf_scale, out=output_trtllm, ) baseline_mean, baseline_std = time_fn(baseline_prefill) trtllm_mean, trtllm_std = time_fn(trtllm_prefill) # Calculate percentage speedup (positive means TRT is faster) speedup_percent = (baseline_mean - trtllm_mean) / baseline_mean print( f"\t{batch_size}\t{max_seq_len}\t{trtllm_mean:8.3f}\t{trtllm_std.item():8.3f}" f"\t{baseline_mean:8.3f}\t{baseline_std.item():8.3f}\t{speedup_percent:8.3f}" ) # Return results for CSV writing return { "batch_size": batch_size, "trtllm_mean": trtllm_mean, "trtllm_std": trtllm_std.item(), "baseline_mean": baseline_mean, "baseline_std": baseline_std.item(), "speedup_percent": speedup_percent, "q_dtype": str(q_quant_dtype), "kv_cache_dtype": str(kv_quant_dtype), "output_dtype": str(o_quant_dtype), "block_size": block_size, "num_kv_heads": num_kv_heads, "head_size": head_size, "max_seq_len": max_seq_len, } def write_results_to_csv(results, filename=None): """Write benchmark results to CSV file.""" if filename is None: timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") filename = f"flashinfer_trtllm_benchmark_{timestamp}.csv" fieldnames = [ "batch_size", "trtllm_mean", "trtllm_std", "baseline_mean", "baseline_std", "speedup_percent", "q_dtype", "kv_cache_dtype", "output_dtype", "block_size", "num_kv_heads", "head_size", "max_seq_len", ] file_exists = os.path.exists(filename) with open(filename, "a", newline="") as csvfile: writer = csv.DictWriter(csvfile, fieldnames=fieldnames) if not file_exists: writer.writeheader() for result in results: writer.writerow(result) print(f"Results written to {filename}") if __name__ == "__main__": batch_sizes = [1, 4, 8, 16, 32, 64, 128, 256] max_seq_lens = [1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072] all_results = [] dtype = torch.bfloat16 quant_dtypes = [ # (q_quant_dtype, kv_quant_dtype, o_quant_dtype) (None, None, None), (FP8_DTYPE, FP8_DTYPE, None), (FP8_DTYPE, FP8_DTYPE, FP8_DTYPE), (FP8_DTYPE, FP8_DTYPE, FP4_DTYPE), ] for quant_dtype in quant_dtypes: q_quant_dtype, kv_quant_dtype, o_quant_dtype = quant_dtype q_quant_dtype = q_quant_dtype or dtype kv_quant_dtype = kv_quant_dtype or dtype o_quant_dtype = o_quant_dtype or dtype print( f"Running benchmark for q_dtype = {q_quant_dtype}, " f"kv_cache_dtype: {kv_quant_dtype}, " f"output_dtype: {o_quant_dtype}" ) print( "\tbatch_size\tmax_seq_len\ttrtllm_mean\ttrtllm_std\tbaseline_mean\t" "baseline_std\tspeedup_percent" ) for max_seq_len in max_seq_lens: for bs in batch_sizes: result = benchmark_prefill( dtype=dtype, quant_dtypes=quant_dtype, batch_size=bs, max_seq_len=max_seq_len, ) all_results.append(result) # Write all results to CSV write_results_to_csv(all_results)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_bitblas.py
benchmarks/kernels/benchmark_bitblas.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) Microsoft Corporation. # Licensed under the MIT License. from packaging import version from vllm.model_executor.layers.quantization.utils.bitblas_utils import ( MINIMUM_BITBLAS_VERSION, ) try: import bitblas if version.parse(bitblas.__version__) < version.parse(MINIMUM_BITBLAS_VERSION): raise ImportError( "bitblas version is wrong. Please " f"install bitblas>={MINIMUM_BITBLAS_VERSION}" ) except ImportError as e: bitblas_import_exception = e raise ValueError( "Trying to use the bitblas backend, but could not import" f"with the following error: {bitblas_import_exception}. " "Please install bitblas through the following command: " f"`pip install bitblas>={MINIMUM_BITBLAS_VERSION}`" ) from bitblas_import_exception from bitblas import Matmul, MatmulConfig, auto_detect_nvidia_target from vllm.utils.argparse_utils import FlexibleArgumentParser parser = FlexibleArgumentParser( description="Benchmark BitBLAS int4 on a specific target." ) # Add arguments to the parser parser.add_argument( "--target", type=str, default=auto_detect_nvidia_target(), help="Specify the target device for benchmarking.", ) parser.add_argument( "--group_size", type=int, default=None, help="Group size for grouped quantization." ) parser.add_argument( "--A_dtype", type=str, default="float16", choices=["float16", "float32", "float64", "int32", "int8"], help="Data type of activation A.", ) parser.add_argument( "--W_dtype", type=str, default="int4", choices=[ "float16", "float32", "float64", "int32", "int8", "int4", "int2", "int1", "nf4", "fp4_e2m1", ], help="Data type of weight W.", ) parser.add_argument( "--accum_dtype", type=str, default="float16", choices=["float16", "int32"], help="Data type for accumulation.", ) parser.add_argument( "--out_dtype", type=str, default="float16", choices=["float16", "float32", "int32", "int8"], help="Data type for output.", ) parser.add_argument( "--layout", type=str, default="nt", choices=["nt", "nn"], help="Matrix layout, 'nt' for non-transpose A and transpose W.", ) parser.add_argument( "--with_bias", action="store_true", help="Include bias in the benchmark." ) parser.add_argument( "--with_scaling", action="store_true", help="Include scaling factor in the quantization.", ) parser.add_argument( "--with_zeros", action="store_true", help="Include zeros in the quantization." ) parser.add_argument( "--zeros_mode", type=str, default=None, choices=["original", "rescale", "quantized"], help="Specify the mode for calculating zeros.", ) # Parse the arguments args = parser.parse_args() # Assign arguments to variables target = args.target A_dtype = args.A_dtype W_dtype = args.W_dtype accum_dtype = args.accum_dtype out_dtype = args.out_dtype layout = args.layout with_bias = args.with_bias group_size = args.group_size with_scaling = args.with_scaling with_zeros = args.with_zeros zeros_mode = args.zeros_mode # Define a list of shared arguments that repeat in every config shared_args = [ A_dtype, W_dtype, out_dtype, accum_dtype, layout, with_bias, group_size, with_scaling, with_zeros, zeros_mode, ] # Define just the (M, K, N) shapes in a more compact list shapes = [ # square test (1, 16384, 16384), # BLOOM-176B (1, 43008, 14336), (1, 14336, 14336), (1, 57344, 14336), (1, 14336, 57344), # OPT-65B (1, 9216, 9216), (1, 36864, 9216), (1, 9216, 36864), (1, 22016, 8192), # LLAMA-70B/65B (1, 8192, 22016), (1, 8192, 8192), (1, 28672, 8192), (1, 8192, 28672), # square test (16384, 16384, 16384), # BLOOM-176B (8192, 43008, 14336), (8192, 14336, 14336), (8192, 57344, 14336), (8192, 14336, 57344), # OPT-65B (8192, 9216, 9216), (8192, 36864, 9216), (8192, 9216, 36864), (8192, 22016, 8192), # LLAMA-70B/65B (8192, 8192, 22016), (8192, 8192, 8192), (8192, 28672, 8192), (8192, 8192, 28672), ] # Build test shapes with all the shared arguments test_shapes = [(MatmulConfig, Matmul, (*shape, *shared_args)) for shape in shapes] benchmark_sets = [] benchmark_sets.extend(test_shapes) benchmark_results = {} for config_class, operator, input_args in benchmark_sets: config = config_class(*input_args) matmul = operator(config, target=target, enable_tuning=True) kernel_latency = matmul.profile_latency() print("Time cost is: {:.3f} ms".format(kernel_latency)) profile_config = { f"{operator.__name__}-{'-'.join([str(i) for i in input_args])}": { "BitBLAS_top20_latency": kernel_latency, } } benchmark_results.update(profile_config) # Define headers for the table headers = [ "PrimFunc", "Input Arguments", "BitBLAS Top20 Latency", ] # Calculate column widths for pretty printing col_widths = [0, 0, 0] for config_key, values in benchmark_results.items(): args_split = config_key.split("-") func_name = args_split[0] input_args_str = "-".join(args_split[1:]) col_widths[0] = max(col_widths[0], len(func_name) + 2, len(headers[0]) + 2) col_widths[1] = max(col_widths[1], len(input_args_str) + 2, len(headers[1]) + 2) col_widths[2] = max( col_widths[2], len(f"{values['BitBLAS_top20_latency']:.3f} ms") + 2, len(headers[2]) + 2, ) # break only if you want to measure widths from a single example; # otherwise, let it loop over all items. # Print header for i, header in enumerate(headers): headers[i] = header.ljust(col_widths[i]) print("".join(headers)) print("-" * sum(col_widths)) # Print rows for config_key, values in benchmark_results.items(): args_split = config_key.split("-") func_name = args_split[0] input_args_str = "-".join(args_split[1:]) row = [ func_name, input_args_str, f"{values['BitBLAS_top20_latency']:.3f} ms", ] row_str = "".join( [str(cell).ljust(col_widths[idx]) for idx, cell in enumerate(row)] ) print(row_str)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_trtllm_decode_attention.py
benchmarks/kernels/benchmark_trtllm_decode_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import csv import os from datetime import datetime import flashinfer import torch from vllm.utils.math_utils import round_up FLOAT32_BYTES = torch.finfo(torch.float).bits // 8 FP8_DTYPE = torch.float8_e4m3fn FP4_DTYPE = torch.uint8 def to_float8(x, dtype=torch.float8_e4m3fn): finfo = torch.finfo(dtype) min_val, max_val = x.aminmax() amax = torch.maximum(min_val.abs(), max_val.abs()).clamp(min=1e-12) scale = finfo.max / amax * 0.1 x_scl_sat = (x * scale).clamp(min=finfo.min, max=finfo.max) return x_scl_sat.to(dtype), scale.float().reciprocal() @torch.no_grad() def benchmark_decode( dtype: torch.dtype, quant_dtypes: tuple[torch.dtype | None, torch.dtype | None, torch.dtype | None], batch_size: int, max_seq_len: int, num_heads: tuple[int, int] = (64, 8), head_size: int = 128, kv_layout: str = "HND", block_size: int = 16, warmup: int = 10, trials: int = 20, ): torch.set_default_device("cuda") torch.manual_seed(0) q_quant_dtype, kv_quant_dtype, o_quant_dtype = quant_dtypes q_quant_dtype = q_quant_dtype or dtype kv_quant_dtype = kv_quant_dtype or dtype o_quant_dtype = o_quant_dtype or dtype num_qo_heads, num_kv_heads = num_heads assert num_qo_heads % num_kv_heads == 0 sm_scale = float(1.0 / (head_size**0.5)) # large number to reduce kv_cache reuse NUM_BLOCKS = int(256000 / block_size) kv_cache_shape = None if kv_layout == "NHD": kv_cache_shape = (NUM_BLOCKS, 2, block_size, num_kv_heads, head_size) elif kv_layout == "HND": kv_cache_shape = (NUM_BLOCKS, 2, num_kv_heads, block_size, head_size) else: raise ValueError(f"Invalid kv_layout: {kv_layout}") # Always using 1.0 scale to reflect the real perf in benchmarking q_scale = 1.0 ref_query = torch.randn(batch_size, num_qo_heads, head_size, dtype=dtype) if q_quant_dtype == FP8_DTYPE: query, _ = to_float8(ref_query) else: query = ref_query kv_lens = torch.randint(1, max_seq_len, (batch_size,), dtype=torch.int32) kv_lens[-1] = max_seq_len seq_lens = kv_lens max_seq_len = torch.max(seq_lens).item() # Always using 1.0 scale to reflect the real perf in benchmarking k_scale = v_scale = 1.0 ref_kv_cache = torch.randn(kv_cache_shape, dtype=dtype) if kv_quant_dtype == FP8_DTYPE: kv_cache, _ = to_float8(ref_kv_cache) else: kv_cache = ref_kv_cache max_num_blocks_per_seq = (max_seq_len + block_size - 1) // block_size block_tables = torch.randint( 0, NUM_BLOCKS, (batch_size, max_num_blocks_per_seq), dtype=torch.int32 ) kv_indptr = [0] kv_indices = [] kv_last_page_lens = [] for i in range(batch_size): seq_len = seq_lens[i] assert seq_len > 0 num_blocks = (seq_len + block_size - 1) // block_size kv_indices.extend(block_tables[i, :num_blocks]) kv_indptr.append(kv_indptr[-1] + num_blocks) kv_last_page_len = seq_len % block_size if kv_last_page_len == 0: kv_last_page_len = block_size kv_last_page_lens.append(kv_last_page_len) kv_indptr = torch.tensor(kv_indptr, dtype=torch.int32) kv_indices = torch.tensor(kv_indices, dtype=torch.int32) kv_last_page_lens = torch.tensor(kv_last_page_lens, dtype=torch.int32) workspace_buffer = torch.zeros(1024 * 1024 * 1024, dtype=torch.int8) wrapper = flashinfer.BatchDecodeWithPagedKVCacheWrapper( workspace_buffer, kv_layout, use_tensor_cores=True, ) wrapper.plan( kv_indptr, kv_indices, kv_last_page_lens, num_qo_heads, num_kv_heads, head_size, block_size, "NONE", sm_scale=sm_scale, q_data_type=dtype, kv_data_type=dtype, ) def time_fn(fn, warmup=10, trials=20): torch.cuda.synchronize() start = torch.Event(enable_timing=True) end = torch.Event(enable_timing=True) times = [] for i in range(warmup): fn() for i in range(trials): start.record() fn() end.record() torch.cuda.synchronize() times.append(start.elapsed_time(end)) # ms return sum(times) / len(times), torch.std(torch.tensor(times)) o_scale = 1.0 o_sf_scale = None output_baseline = torch.empty(ref_query.shape, dtype=dtype) if o_quant_dtype == FP4_DTYPE: o_sf_scale = 500.0 output_trtllm = flashinfer.utils.FP4Tensor( torch.empty(query.shape[:-1] + (query.shape[-1] // 2,), dtype=torch.uint8), torch.empty( ( round_up(query.shape[0], 128), round_up(query.shape[1] * query.shape[2] // 16, 4), ), dtype=torch.float8_e4m3fn, ), ) else: output_trtllm = torch.empty(query.shape, dtype=o_quant_dtype) def baseline_decode(): return wrapper.run( ref_query, ref_kv_cache, k_scale=k_scale, v_scale=v_scale, out=output_baseline, ) def trtllm_decode(): return flashinfer.decode.trtllm_batch_decode_with_kv_cache( query=query, kv_cache=kv_cache, workspace_buffer=workspace_buffer, block_tables=block_tables, seq_lens=seq_lens, max_seq_len=max_seq_len, bmm1_scale=q_scale * k_scale * sm_scale, bmm2_scale=v_scale / o_scale, o_sf_scale=o_sf_scale, out=output_trtllm, ) baseline_mean, baseline_std = time_fn(baseline_decode) trtllm_mean, trtllm_std = time_fn(trtllm_decode) # Calculate percentage speedup (positive means TRT is faster) speedup_percent = (baseline_mean - trtllm_mean) / baseline_mean print( f"\t{batch_size}\t{max_seq_len}\t{trtllm_mean:.3f}\t{trtllm_std.item():.3f}" f"\t{baseline_mean:.3f}\t{baseline_std.item():.3f}\t{speedup_percent:.3f}" ) # Return results for CSV writing return { "batch_size": batch_size, "trtllm_mean": trtllm_mean, "trtllm_std": trtllm_std.item(), "baseline_mean": baseline_mean, "baseline_std": baseline_std.item(), "speedup_percent": speedup_percent, "q_dtype": str(q_quant_dtype), "kv_cache_dtype": str(kv_quant_dtype), "output_dtype": str(o_quant_dtype), "block_size": block_size, "num_kv_heads": num_kv_heads, "head_size": head_size, "max_seq_len": max_seq_len, } def write_results_to_csv(results, filename=None): """Write benchmark results to CSV file.""" if filename is None: timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") filename = f"flashinfer_trtllm_benchmark_{timestamp}.csv" fieldnames = [ "batch_size", "trtllm_mean", "trtllm_std", "baseline_mean", "baseline_std", "speedup_percent", "q_dtype", "kv_cache_dtype", "output_dtype", "block_size", "num_kv_heads", "head_size", "max_seq_len", ] file_exists = os.path.exists(filename) with open(filename, "a", newline="") as csvfile: writer = csv.DictWriter(csvfile, fieldnames=fieldnames) if not file_exists: writer.writeheader() for result in results: writer.writerow(result) print(f"Results written to {filename}") if __name__ == "__main__": batch_sizes = [1, 4, 8, 16, 32, 64, 128, 256] max_seq_lens = [1024, 2048, 4096, 8192, 16384, 32768, 65536, 131072] all_results = [] dtype = torch.bfloat16 quant_dtypes = [ # (q_quant_dtype, kv_quant_dtype, o_quant_dtype) (None, None, None), (None, FP8_DTYPE, None), (FP8_DTYPE, FP8_DTYPE, None), (FP8_DTYPE, FP8_DTYPE, FP8_DTYPE), (FP8_DTYPE, FP8_DTYPE, FP4_DTYPE), ] for quant_dtype in quant_dtypes: q_quant_dtype, kv_quant_dtype, o_quant_dtype = quant_dtype q_quant_dtype = q_quant_dtype or dtype kv_quant_dtype = kv_quant_dtype or dtype o_quant_dtype = o_quant_dtype or dtype print( f"Running benchmark for q_dtype = {q_quant_dtype}, " f"kv_cache_dtype: {kv_quant_dtype}, " f"output_dtype: {o_quant_dtype}" ) print( "\tbatch_size\tmax_seq_len\ttrtllm_mean\ttrtllm_std\tbaseline_mean\t" "baseline_std\tspeedup_percent" ) for max_seq_len in max_seq_lens: for bs in batch_sizes: result = benchmark_decode( dtype=dtype, quant_dtypes=quant_dtype, batch_size=bs, max_seq_len=max_seq_len, ) all_results.append(result) # Write all results to CSV write_results_to_csv(all_results)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_paged_attention.py
benchmarks/kernels/benchmark_paged_attention.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import random import time import torch from vllm import _custom_ops as ops from vllm.logger import init_logger from vllm.platforms import current_platform from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.torch_utils import ( STR_DTYPE_TO_TORCH_DTYPE, create_kv_caches_with_random, ) logger = init_logger(__name__) NUM_BLOCKS = 128 * 1024 PARTITION_SIZE = 512 PARTITION_SIZE_ROCM = 256 @torch.inference_mode() def main( version: str, num_seqs: int, seq_len: int, num_query_heads: int, num_kv_heads: int, head_size: int, use_alibi: bool, block_size: int, dtype: torch.dtype, seed: int, do_profile: bool, device: str = "cuda", kv_cache_dtype: str | None = None, ) -> None: current_platform.seed_everything(seed) scale = float(1.0 / (head_size**0.5)) query = torch.empty( num_seqs, num_query_heads, head_size, dtype=dtype, device=device ) query.uniform_(-scale, scale) assert num_query_heads % num_kv_heads == 0 alibi_slopes = None if use_alibi: alibi_slopes = torch.randn(num_query_heads, dtype=torch.float, device=device) seq_lens = [seq_len for _ in range(num_seqs)] max_seq_len = max(seq_lens) seq_lens = torch.tensor(seq_lens, dtype=torch.int, device=device) # Create the block tables. max_num_blocks_per_seq = (max_seq_len + block_size - 1) // block_size block_tables_lst: list[list[int]] = [] for _ in range(num_seqs): block_table = [ random.randint(0, NUM_BLOCKS - 1) for _ in range(max_num_blocks_per_seq) ] block_tables_lst.append(block_table) block_tables = torch.tensor(block_tables_lst, dtype=torch.int, device=device) # Create the KV cache. key_caches, value_caches = create_kv_caches_with_random( NUM_BLOCKS, block_size, 1, num_kv_heads, head_size, kv_cache_dtype, dtype, device=device, ) key_cache, value_cache = key_caches[0], value_caches[0] # Prepare for the paged attention kernel. output = torch.empty_like(query) if version == "v2": if current_platform.is_rocm(): global PARTITION_SIZE if not args.custom_paged_attn and not current_platform.is_navi(): PARTITION_SIZE = 1024 else: PARTITION_SIZE = PARTITION_SIZE_ROCM num_partitions = (max_seq_len + PARTITION_SIZE - 1) // PARTITION_SIZE tmp_output = torch.empty( size=(num_seqs, num_query_heads, num_partitions, head_size), dtype=output.dtype, device=output.device, ) exp_sums = torch.empty( size=(num_seqs, num_query_heads, num_partitions), dtype=torch.float32, device=output.device, ) max_logits = torch.empty_like(exp_sums) def run_cuda_benchmark(num_iters: int, profile: bool = False) -> float: torch.cuda.synchronize() if profile: torch.cuda.cudart().cudaProfilerStart() start_time = time.perf_counter() # Using default kv_scale k_scale = v_scale = torch.tensor(1.0, dtype=torch.float32, device=device) for _ in range(num_iters): if version == "v1": ops.paged_attention_v1( output, query, key_cache, value_cache, num_kv_heads, scale, block_tables, seq_lens, block_size, max_seq_len, alibi_slopes, kv_cache_dtype, k_scale, v_scale, ) elif version == "v2": if not args.custom_paged_attn: ops.paged_attention_v2( output, exp_sums, max_logits, tmp_output, query, key_cache, value_cache, num_kv_heads, scale, block_tables, seq_lens, block_size, max_seq_len, alibi_slopes, kv_cache_dtype, k_scale, v_scale, ) else: ops.paged_attention_rocm( output, exp_sums, max_logits, tmp_output, query, key_cache, value_cache, num_kv_heads, scale, block_tables, seq_lens, None, block_size, max_seq_len, alibi_slopes, kv_cache_dtype, k_scale, v_scale, ) else: raise ValueError(f"Invalid version: {version}") torch.cuda.synchronize() end_time = time.perf_counter() if profile: torch.cuda.cudart().cudaProfilerStop() return (end_time - start_time) / num_iters # Warmup. print("Warming up...") run_benchmark = run_cuda_benchmark run_benchmark(num_iters=3, profile=False) # Benchmark. if do_profile: latency = run_benchmark(num_iters=1, profile=True) else: latency = run_benchmark(num_iters=100, profile=False) print(f"Kernel running time: {latency * 1000000:.3f} us") if __name__ == "__main__": logger.warning( "This script benchmarks the paged attention kernel. " "By default this is no longer used in vLLM inference." ) parser = FlexibleArgumentParser(description="Benchmark the paged attention kernel.") parser.add_argument("--version", type=str, choices=["v1", "v2"], default="v2") parser.add_argument("--batch-size", type=int, default=8) parser.add_argument("--seq-len", type=int, default=4096) parser.add_argument("--num-query-heads", type=int, default=64) parser.add_argument("--num-kv-heads", type=int, default=8) parser.add_argument( "--head-size", type=int, choices=[64, 80, 96, 112, 120, 128, 192, 256], default=128, ) parser.add_argument("--block-size", type=int, choices=[16, 32], default=16) parser.add_argument("--use-alibi", action="store_true") parser.add_argument( "--dtype", type=str, choices=["half", "bfloat16", "float"], default="half" ) parser.add_argument("--seed", type=int, default=0) parser.add_argument("--profile", action="store_true") parser.add_argument( "--kv-cache-dtype", type=str, choices=["auto", "fp8", "fp8_e5m2", "fp8_e4m3"], default="auto", help="Data type for kv cache storage. If 'auto', will use model " "data type. CUDA 11.8+ supports fp8 (=fp8_e4m3) and fp8_e5m2. " "ROCm (AMD GPU) supports fp8 (=fp8_e4m3)", ) parser.add_argument( "--custom-paged-attn", action="store_true", help="Use custom paged attention" ) args = parser.parse_args() print(args) if args.num_query_heads % args.num_kv_heads != 0: raise ValueError("num_query_heads must be divisible by num_kv_heads") main( version=args.version, num_seqs=args.batch_size, seq_len=args.seq_len, num_query_heads=args.num_query_heads, num_kv_heads=args.num_kv_heads, head_size=args.head_size, block_size=args.block_size, use_alibi=args.use_alibi, dtype=STR_DTYPE_TO_TORCH_DTYPE[args.dtype], seed=args.seed, do_profile=args.profile, kv_cache_dtype=args.kv_cache_dtype, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_mla_k_concat.py
benchmarks/kernels/benchmark_mla_k_concat.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Benchmark script comparing torch.cat vs direct copy for k_nope/k_pe concatenation in MLA (Multi-head Latent Attention) prefill. This validates that the optimization from commit 8d4142bd is beneficial across various batch sizes, not just the originally tested batch size of 32768. """ import time from collections.abc import Callable import torch # DeepSeek-V3 MLA dimensions NUM_HEADS = 128 QK_NOPE_HEAD_DIM = 128 PE_DIM = 64 def cat_method(k_nope: torch.Tensor, k_pe: torch.Tensor) -> torch.Tensor: """Original torch.cat approach with expand.""" return torch.cat((k_nope, k_pe.expand((*k_nope.shape[:-1], -1))), dim=-1) def direct_copy_method(k_nope: torch.Tensor, k_pe: torch.Tensor) -> torch.Tensor: """Optimized direct copy approach (avoids expand + cat overhead).""" k = torch.empty( (*k_nope.shape[:-1], k_nope.shape[-1] + k_pe.shape[-1]), dtype=k_nope.dtype, device=k_nope.device, ) k[..., : k_nope.shape[-1]] = k_nope k[..., k_nope.shape[-1] :] = k_pe return k def benchmark_method( method: Callable, k_nope: torch.Tensor, k_pe: torch.Tensor, num_warmup: int = 10, num_iters: int = 100, ) -> float: """Benchmark a concatenation method and return mean latency in ms.""" # Warmup for _ in range(num_warmup): _ = method(k_nope, k_pe) torch.cuda.synchronize() # Benchmark start = time.perf_counter() for _ in range(num_iters): _ = method(k_nope, k_pe) torch.cuda.synchronize() end = time.perf_counter() return (end - start) / num_iters * 1000 # Convert to ms @torch.inference_mode() def run_benchmark(dtype: torch.dtype, dtype_name: str): """Run benchmark for a specific dtype.""" torch.set_default_device("cuda") # Batch sizes to test (powers of 2 from 32 to 65536) batch_sizes = [32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536] print("=" * 80) print("Benchmark: torch.cat vs direct copy for MLA k_nope/k_pe concatenation") print("=" * 80) print( f"Tensor shapes: k_nope=[B, {NUM_HEADS}, {QK_NOPE_HEAD_DIM}], " f"k_pe=[B, 1, {PE_DIM}]" ) print(f"dtype: {dtype_name}") print() print( f"{'Batch Size':>12} | {'cat (ms)':>10} | {'direct (ms)':>12} | " f"{'Speedup':>8} | {'Reduction':>10}" ) print("-" * 70) results = [] for batch_size in batch_sizes: # Create input tensors (generate in float32 then convert for FP8 compatibility) k_nope = torch.randn( batch_size, NUM_HEADS, QK_NOPE_HEAD_DIM, dtype=torch.float32, device="cuda" ).to(dtype) k_pe = torch.randn( batch_size, 1, PE_DIM, dtype=torch.float32, device="cuda" ).to(dtype) # Benchmark both methods cat_time = benchmark_method(cat_method, k_nope, k_pe) direct_time = benchmark_method(direct_copy_method, k_nope, k_pe) speedup = cat_time / direct_time reduction = (1 - direct_time / cat_time) * 100 results.append((batch_size, cat_time, direct_time, speedup, reduction)) print( f"{batch_size:>12} | {cat_time:>10.3f} | {direct_time:>12.3f} | " f"{speedup:>7.2f}x | {reduction:>9.1f}%" ) print("=" * 80) # Summary statistics speedups = [r[3] for r in results] print("\nSpeedup summary:") print(f" Min: {min(speedups):.2f}x") print(f" Max: {max(speedups):.2f}x") print(f" Mean: {sum(speedups) / len(speedups):.2f}x") # Find crossover point crossover_batch = None for batch_size, _, _, speedup, _ in results: if speedup >= 1.0: crossover_batch = batch_size break print("\nConclusion:") if crossover_batch: print(f" - Direct copy becomes beneficial at batch size >= {crossover_batch}") # Filter for large batches (>= 512 which is typical for prefill) large_batch_speedups = [r[3] for r in results if r[0] >= 512] if large_batch_speedups: avg_large = sum(large_batch_speedups) / len(large_batch_speedups) print(f" - For batch sizes >= 512: avg speedup = {avg_large:.2f}x") print(" - MLA prefill typically uses large batches, so optimization is effective") return results @torch.inference_mode() def main(): # Test bfloat16 print("\n") run_benchmark(torch.bfloat16, "bfloat16") # Test float8_e4m3fn print("\n") run_benchmark(torch.float8_e4m3fn, "float8_e4m3fn") if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_moe_align_block_size.py
benchmarks/kernels/benchmark_moe_align_block_size.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import itertools import torch from vllm.model_executor.layers.fused_moe.moe_align_block_size import ( moe_align_block_size, ) from vllm.triton_utils import triton def get_topk_ids(num_tokens: int, num_experts: int, topk: int) -> torch.Tensor: return torch.stack( [ torch.randperm(num_experts, dtype=torch.int32, device="cuda")[:topk] for _ in range(num_tokens) ] ) # test configurations num_tokens_range = [1, 16, 256, 4096] num_experts_range = [16, 64, 224, 256, 280, 512] topk_range = [1, 2, 8] ep_size_range = [1, 8] configs = list( itertools.product(num_tokens_range, num_experts_range, topk_range, ep_size_range) ) @triton.testing.perf_report( triton.testing.Benchmark( x_names=["num_tokens", "num_experts", "topk", "ep_size"], x_vals=configs, line_arg="provider", line_vals=["vllm"], line_names=["vLLM"], plot_name="moe-align-block-size-performance", args={}, ) ) def benchmark(num_tokens, num_experts, topk, ep_size, provider): """Benchmark function for Triton.""" block_size = 256 torch.cuda.manual_seed_all(0) topk_ids = get_topk_ids(num_tokens, num_experts, topk) e_map = None if ep_size != 1: local_e = num_experts // ep_size e_ids = torch.randperm(num_experts, device="cuda", dtype=torch.int32)[:local_e] e_map = torch.full((num_experts,), -1, device="cuda", dtype=torch.int32) e_map[e_ids] = torch.arange(local_e, device="cuda", dtype=torch.int32) quantiles = [0.5, 0.2, 0.8] if provider == "vllm": ms, min_ms, max_ms = triton.testing.do_bench( lambda: moe_align_block_size( topk_ids, block_size, num_experts, e_map, ignore_invalid_experts=True ), quantiles=quantiles, ) return 1000 * ms, 1000 * max_ms, 1000 * min_ms if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--num_experts", type=int, default=64, choices=[8, 16, 32, 64, 128, 256], ) parser.add_argument( "--topk", type=int, default=8, choices=[2, 4, 8], help="Top-k value for correctness check.", ) args = parser.parse_args() benchmark.run(print_data=True, show_plots=True)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_cutlass_fp4_moe.py
benchmarks/kernels/benchmark_cutlass_fp4_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Benchmark the performance of the cutlass_moe_fp4 kernel vs the triton_moe kernel. The cutlass_moe_fp4 kernel takes in fp4 quantized weights and 16-bit activations. The triton_moe kernel takes in fp8 weights(tensor scaled to fp8) and 16-bit activations. """ import nvtx import torch import torch.utils.benchmark as benchmark from vllm import _custom_ops as ops from vllm.config import ParallelConfig, VllmConfig, set_current_vllm_config from vllm.model_executor.layers.fused_moe.config import ( fp8_w8a8_moe_quant_config, nvfp4_moe_quant_config, ) from vllm.model_executor.layers.fused_moe.cutlass_moe import cutlass_moe_fp4 from vllm.model_executor.layers.fused_moe.fused_moe import fused_experts, fused_topk from vllm.scalar_type import scalar_types from vllm.utils.argparse_utils import FlexibleArgumentParser WEIGHT_SHAPES_MOE = { "nvidia/DeepSeek-R1-FP4": [ [256, 8, 2048, 7168], ], } DEFAULT_MODELS = [ "nvidia/DeepSeek-R1-FP4", ] DEFAULT_BATCH_SIZES = [4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048] DEFAULT_TP_SIZES = [1] PER_ACT_TOKEN_OPTS = [False] PER_OUT_CH_OPTS = [False] FLOAT4_E2M1_MAX = scalar_types.float4_e2m1f.max() FLOAT8_E4M3_MAX = torch.finfo(torch.float8_e4m3fn).max def to_fp8(tensor: torch.Tensor): finfo = torch.finfo(torch.float8_e4m3fn) return torch.round(tensor.clamp(min=finfo.min, max=finfo.max)).to( dtype=torch.float8_e4m3fn ) def bench_run( results: list[benchmark.Measurement], model: str, num_experts: int, topk: int, per_act_token: bool, per_out_ch: bool, mkn: tuple[int, int, int], ): label = "NVFP4 Blockscaled CUTLASS MOE vs FP8 Tensor Scaled Triton" sub_label = ( "{}, num_experts={}, topk={}, per_act_token={} per_out_ch={}, MKN=({})".format( model, num_experts, topk, per_act_token, per_out_ch, mkn ) ) print(f"Testing: {sub_label}") (m, k, n) = mkn dtype = torch.half device = "cuda" a = torch.randn((m, k), device=device, dtype=dtype) / 10 w1 = torch.randn((num_experts, 2 * n, k), device=device, dtype=dtype) / 10 w2 = torch.randn((num_experts, k, n), device=device, dtype=dtype) / 10 _, a_fp8_scale = ops.scaled_fp8_quant(a) w1_fp8q = torch.empty( (num_experts, 2 * n, k), device=device, dtype=torch.float8_e4m3fn ) w2_fp8q = torch.empty((num_experts, k, n), device=device, dtype=torch.float8_e4m3fn) w1_fp8scale = torch.empty((num_experts, 1, 1), device=device, dtype=torch.float32) w2_fp8scale = torch.empty((num_experts, 1, 1), device=device, dtype=torch.float32) for expert in range(num_experts): w1_fp8q[expert], w1_fp8scale[expert] = ops.scaled_fp8_quant(w1[expert]) w2_fp8q[expert], w2_fp8scale[expert] = ops.scaled_fp8_quant(w2[expert]) w1_fp8q_notransp = w1_fp8q.clone() w2_fp8q_notransp = w2_fp8q.clone() w1_fp8q = w1_fp8q.transpose(1, 2) w2_fp8q = w2_fp8q.transpose(1, 2) score = torch.randn((m, num_experts), device=device, dtype=dtype) topk_weights, topk_ids, _ = fused_topk(a, score, topk, renormalize=False) quant_blocksize = 16 w1_blockscale = torch.empty( (num_experts, 2 * n, k // quant_blocksize), device=device, dtype=torch.float8_e4m3fn, ) w2_blockscale = torch.empty( (num_experts, k, n // quant_blocksize), device=device, dtype=torch.float8_e4m3fn ) # n_b_scales = 2 * n if per_out_ch else 1 # k_b_scales = k if per_out_ch else 1 w1_fp4 = torch.empty((num_experts, 2 * n, k // 2), device=device, dtype=torch.uint8) w2_fp4 = torch.empty((num_experts, k, n // 2), device=device, dtype=torch.uint8) w1_gs = torch.empty((num_experts,), device=device, dtype=torch.float32) w2_gs = torch.empty((num_experts,), device=device, dtype=torch.float32) a1_gs = torch.ones((num_experts,), device=device, dtype=torch.float32) a2_gs = torch.ones((num_experts,), device=device, dtype=torch.float32) for expert in range(num_experts): w1_e = w1[expert] w2_e = w2[expert] w1_amax = torch.abs(w1_e).max().to(torch.float32) w2_amax = torch.abs(w2_e).max().to(torch.float32) w1_gs[expert] = FLOAT8_E4M3_MAX * FLOAT4_E2M1_MAX / w1_amax w2_gs[expert] = FLOAT8_E4M3_MAX * FLOAT4_E2M1_MAX / w2_amax w1_fp4[expert], w1_blockscale[expert] = ops.scaled_fp4_quant( w1_e, w1_gs[expert] ) w2_fp4[expert], w2_blockscale[expert] = ops.scaled_fp4_quant( w2_e, w2_gs[expert] ) def run_triton_moe( a: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, w1_scale: torch.Tensor, w2_scale: torch.Tensor, a_fp8_scale: torch.Tensor, num_repeats: int, ): quant_config = fp8_w8a8_moe_quant_config( w1_scale=w1_scale, w2_scale=w2_scale, a1_scale=a_fp8_scale, ) for _ in range(num_repeats): fused_experts( a, w1, w2, topk_weights, topk_ids, quant_config=quant_config, ) def run_cutlass_moe_fp4( a: torch.Tensor, w1_fp4: torch.Tensor, w2_fp4: torch.Tensor, w1_blockscale: torch.Tensor, w2_blockscale: torch.Tensor, w1_gs: torch.Tensor, w2_gs: torch.Tensor, a1_gs: torch.Tensor, a2_gs: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, m: int, n: int, k: int, e: int, device: torch.device, num_repeats: int, ): quant_config = nvfp4_moe_quant_config( a1_gscale=a1_gs, a2_gscale=a2_gs, w1_scale=w1_blockscale, w2_scale=w2_blockscale, g1_alphas=w1_gs, g2_alphas=w2_gs, ) for _ in range(num_repeats): with nvtx.annotate("cutlass_moe_fp4", color="green"): cutlass_moe_fp4( a=a, w1_fp4=w1_fp4, w2_fp4=w2_fp4, topk_weights=topk_weights, topk_ids=topk_ids, m=m, n=n, k=k, e=num_experts, quant_config=quant_config, ) def run_cutlass_from_graph( a: torch.Tensor, a1_gscale: torch.Tensor, w1_fp4: torch.Tensor, w1_blockscale: torch.Tensor, w1_alphas: torch.Tensor, a2_gscale: torch.Tensor, w2_fp4: torch.Tensor, w2_blockscale: torch.Tensor, w2_alphas: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, m: int, n: int, k: int, e: int, device: torch.device, ): quant_config = nvfp4_moe_quant_config( a1_gscale=a1_gs, a2_gscale=a2_gs, w1_scale=w1_blockscale, w2_scale=w2_blockscale, g1_alphas=w1_gs, g2_alphas=w2_gs, ) with set_current_vllm_config( VllmConfig(parallel_config=ParallelConfig(pipeline_parallel_size=1)) ): return cutlass_moe_fp4( a=a, w1_fp4=w1_fp4, w2_fp4=w2_fp4, topk_weights=topk_weights, topk_ids=topk_ids, m=m, n=n, k=k, e=num_experts, quant_config=quant_config, ) def run_triton_from_graph( a: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, w1_scale: torch.Tensor, w2_scale: torch.Tensor, a_fp8_scale: torch.Tensor, ): with set_current_vllm_config( VllmConfig(parallel_config=ParallelConfig(pipeline_parallel_size=1)) ): quant_config = fp8_w8a8_moe_quant_config( w1_scale=w1_scale, w2_scale=w2_scale, a1_scale=a_fp8_scale, ) return fused_experts( a, w1, w2, topk_weights, topk_ids, quant_config=quant_config, ) def replay_graph(graph, num_repeats): for _ in range(num_repeats): graph.replay() torch.cuda.synchronize() cutlass_stream = torch.cuda.Stream() cutlass_graph = torch.cuda.CUDAGraph() with torch.cuda.graph(cutlass_graph, stream=cutlass_stream): run_cutlass_from_graph( a=a, a1_gscale=a1_gs, w1_fp4=w1_fp4, w1_blockscale=w1_blockscale, w1_alphas=w1_gs, a2_gscale=a2_gs, w2_fp4=w2_fp4, w2_blockscale=w2_blockscale, w2_alphas=w2_gs, topk_weights=topk_weights, topk_ids=topk_ids, m=m, n=n, k=k, e=num_experts, device=device, ) torch.cuda.synchronize() triton_stream = torch.cuda.Stream() triton_graph = torch.cuda.CUDAGraph() with torch.cuda.graph(triton_graph, stream=triton_stream): run_triton_from_graph( a, w1_fp8q_notransp, w2_fp8q_notransp, topk_weights, topk_ids, w1_fp8scale, w2_fp8scale, a_fp8_scale, ) torch.cuda.synchronize() min_run_time = 5 num_warmup = 5 num_runs = 25 globals = { # Baseline params "w1": w1, "w2": w2, "score": score, "topk": topk, "w1_fp8q_notransp": w1_fp8q_notransp, "w2_fp8q_notransp": w2_fp8q_notransp, "w1_fp8scale": w1_fp8scale, "w2_fp8scale": w2_fp8scale, "a_fp8_scale": a_fp8_scale, # Cutlass params "a": a, "a1_gscale": a1_gs, "w1_fp4": w1_fp4, "w1_blockscale": w1_blockscale, "w1_alphas": w1_gs, "a2_gscale": a2_gs, "w2_fp4": w2_fp4, "w2_blockscale": w2_blockscale, "w2_alphas": w2_gs, "topk_weights": topk_weights, "topk_ids": topk_ids, "m": m, "n": n, "k": k, "e": num_experts, "device": device, # cuda graph params "cutlass_graph": cutlass_graph, "triton_graph": triton_graph, # Gen params "num_runs": num_runs, # Kernels "run_triton_moe": run_triton_moe, "run_cutlass_moe_fp4": run_cutlass_moe_fp4, "replay_graph": replay_graph, } # Warmup run_triton_moe( a, w1_fp8q_notransp, w2_fp8q_notransp, topk_weights, topk_ids, w1_fp8scale, w2_fp8scale, a_fp8_scale, num_warmup, ) results.append( benchmark.Timer( stmt="run_triton_moe(a, w1_fp8q_notransp, w2_fp8q_notransp, topk_weights, topk_ids, w1_fp8scale, w2_fp8scale, a_fp8_scale, num_runs)", # noqa: E501 globals=globals, label=label, sub_label=sub_label, description="triton_moe", ).blocked_autorange(min_run_time=min_run_time) ) # Warmup replay_graph(triton_graph, num_warmup) results.append( benchmark.Timer( stmt="replay_graph(triton_graph, num_runs)", globals=globals, label=label, sub_label=sub_label, description="triton_moe_cuda_graphs", ).blocked_autorange(min_run_time=min_run_time) ) # Warmup run_cutlass_moe_fp4( a, w1_fp4, w2_fp4, w1_blockscale, w2_blockscale, w1_gs, w2_gs, a1_gs, a2_gs, topk_weights, topk_ids, m, n, k, num_experts, device, num_warmup, ) results.append( benchmark.Timer( stmt="run_cutlass_moe_fp4(a, w1_fp4, w2_fp4, w1_blockscale, w2_blockscale, w1_alphas, w2_alphas, a1_gscale, a2_gscale, topk_weights, topk_ids, m, n, k, e, device, num_runs)", # noqa: E501 globals=globals, label=label, sub_label=sub_label, description="cutlass_moe_fp4", ).blocked_autorange(min_run_time=min_run_time) ) # Warmup replay_graph(cutlass_graph, num_warmup) results.append( benchmark.Timer( stmt="replay_graph(cutlass_graph, num_runs)", globals=globals, label=label, sub_label=sub_label, description="cutlass_moe_fp4_cuda_graphs", ).blocked_autorange(min_run_time=min_run_time) ) def main(args): print("Benchmarking models:") for i, model in enumerate(args.models): print(f"[{i}] {model}") results: list[benchmark.Measurement] = [] for model in args.models: for tp in args.tp_sizes: for layer in WEIGHT_SHAPES_MOE[model]: num_experts = layer[0] topk = layer[1] size_k = layer[2] size_n = layer[3] // tp if len(args.limit_k) > 0 and size_k not in args.limit_k: continue if len(args.limit_n) > 0 and size_n not in args.limit_n: continue for per_act_token in PER_ACT_TOKEN_OPTS: for per_out_ch in PER_OUT_CH_OPTS: for size_m in args.batch_sizes: mkn = (size_m, size_k, size_n) bench_run( results, model, num_experts, topk, per_act_token, per_out_ch, mkn, ) compare = benchmark.Compare(results) compare.print() if __name__ == "__main__": parser = FlexibleArgumentParser( description="Benchmark NVFP4 CUTLASS MOE across specified models/shapes/batches" ) parser.add_argument( "--models", nargs="+", type=str, default=DEFAULT_MODELS, choices=WEIGHT_SHAPES_MOE.keys(), ) parser.add_argument("--tp-sizes", nargs="+", type=int, default=DEFAULT_TP_SIZES) parser.add_argument( "--batch-sizes", nargs="+", type=int, default=DEFAULT_BATCH_SIZES ) parser.add_argument("--limit-k", nargs="+", type=int, default=[]) parser.add_argument("--limit-n", nargs="+", type=int, default=[]) parser.add_argument("--limit-num-groups", nargs="+", type=int, default=[]) parser.add_argument("--limit-per-act-token", nargs="+", type=int, default=[]) parser.add_argument("--limit-per-out-ch", nargs="+", type=int, default=[]) args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_rmsnorm.py
benchmarks/kernels/benchmark_rmsnorm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import itertools import torch from flashinfer.norm import fused_add_rmsnorm, rmsnorm from torch import nn from vllm import _custom_ops as vllm_ops from vllm.triton_utils import triton class HuggingFaceRMSNorm(nn.Module): def __init__(self, hidden_size: int, eps: float = 1e-6) -> None: super().__init__() self.weight = nn.Parameter(torch.ones(hidden_size)) self.variance_epsilon = eps def forward( self, x: torch.Tensor, residual: torch.Tensor | None = None, ) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]: orig_dtype = x.dtype x = x.to(torch.float32) if residual is not None: x = x + residual.to(torch.float32) residual = x.to(orig_dtype) variance = x.pow(2).mean(dim=-1, keepdim=True) x = x * torch.rsqrt(variance + self.variance_epsilon) x = x.to(orig_dtype) * self.weight if residual is None: return x else: return x, residual def rmsnorm_naive( x: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor | None = None, eps: float = 1e-6, ): naive_norm = HuggingFaceRMSNorm(x.shape[-1], eps=eps) naive_norm.weight = nn.Parameter(weight) naive_norm = naive_norm.to(x.device) orig_shape = x.shape x = x.view(-1, x.shape[-1]) if residual is not None: residual = residual.view(-1, residual.shape[-1]) output = naive_norm(x, residual) if isinstance(output, tuple): output = (output[0].view(orig_shape), output[1].view(orig_shape)) else: output = output.view(orig_shape) return output def rmsnorm_flashinfer( x: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor | None = None, eps: float = 1e-6, ): orig_shape = x.shape x = x.view(-1, x.shape[-1]) if residual is not None: residual = residual.view(-1, residual.shape[-1]) if residual is not None: fused_add_rmsnorm(x, residual, weight, eps) output = (x, residual) else: output = rmsnorm(x, weight, eps) if isinstance(output, tuple): output = (output[0].view(orig_shape), output[1].view(orig_shape)) else: output = output.view(orig_shape) return output def rmsnorm_vllm( x: torch.Tensor, weight: torch.Tensor, residual: torch.Tensor | None = None, eps: float = 1e-6, ): orig_shape = x.shape x = x.view(-1, x.shape[-1]) if residual is not None: residual = residual.view(-1, residual.shape[-1]) if residual is not None: vllm_ops.fused_add_rms_norm(x, residual, weight, eps) output = (x, residual) else: out = torch.empty_like(x) vllm_ops.rms_norm(out, x, weight, eps) output = out if isinstance(output, tuple): output = (output[0].view(orig_shape), output[1].view(orig_shape)) else: output = output.view(orig_shape) return output def calculate_diff(batch_size, seq_len, hidden_size, use_residual=True): dtype = torch.bfloat16 x = torch.randn(batch_size, seq_len, hidden_size, dtype=dtype, device="cuda") weight = torch.ones(hidden_size, dtype=dtype, device="cuda") residual = torch.randn_like(x) if use_residual else None output_naive = rmsnorm_naive( x.clone(), weight, residual.clone() if residual is not None else None ) output_flashinfer = rmsnorm_flashinfer( x.clone(), weight, residual.clone() if residual is not None else None ) output_vllm = rmsnorm_vllm( x.clone(), weight, residual.clone() if residual is not None else None ) if use_residual: output_naive = output_naive[0] output_flashinfer = output_flashinfer[0] output_vllm = output_vllm[0] print(f"Naive output={output_naive}") print(f"FlashInfer output={output_flashinfer}") print(f"vLLM output={output_vllm}") if torch.allclose( output_naive, output_flashinfer, atol=1e-2, rtol=1e-2 ) and torch.allclose(output_naive, output_vllm, atol=1e-2, rtol=1e-2): print("✅ All implementations match") else: print("❌ Implementations differ") batch_size_range = [2**i for i in range(0, 7, 2)] seq_length_range = [2**i for i in range(6, 11, 1)] head_num_range = [32, 48] configs = list(itertools.product(head_num_range, batch_size_range, seq_length_range)) def get_benchmark(use_residual): @triton.testing.perf_report( triton.testing.Benchmark( x_names=["head_num", "batch_size", "seq_len"], x_vals=[list(_) for _ in configs], line_arg="provider", line_vals=["huggingface", "flashinfer", "vllm"], line_names=["HuggingFace", "FlashInfer", "vLLM"], styles=[("blue", "-"), ("green", "-"), ("red", "-")], ylabel="us", plot_name=f"rmsnorm-perf-{'with' if use_residual else 'without'}-residual", args={}, ) ) def benchmark(head_num, batch_size, seq_len, provider): dtype = torch.bfloat16 hidden_size = head_num * 128 # assuming head_dim = 128 x = torch.randn(batch_size, seq_len, hidden_size, dtype=dtype, device="cuda") weight = torch.ones(hidden_size, dtype=dtype, device="cuda") residual = torch.randn_like(x) if use_residual else None quantiles = [0.5, 0.2, 0.8] if provider == "huggingface": ms, min_ms, max_ms = triton.testing.do_bench( lambda: rmsnorm_naive( x.clone(), weight, residual.clone() if residual is not None else None, ), quantiles=quantiles, ) elif provider == "flashinfer": ms, min_ms, max_ms = triton.testing.do_bench( lambda: rmsnorm_flashinfer( x.clone(), weight, residual.clone() if residual is not None else None, ), quantiles=quantiles, ) else: ms, min_ms, max_ms = triton.testing.do_bench( lambda: rmsnorm_vllm( x.clone(), weight, residual.clone() if residual is not None else None, ), quantiles=quantiles, ) return 1000 * ms, 1000 * max_ms, 1000 * min_ms return benchmark if __name__ == "__main__": import argparse parser = argparse.ArgumentParser() parser.add_argument( "--batch-size", type=int, default=4, help="Batch size", ) parser.add_argument( "--seq-len", type=int, default=128, help="Sequence length", ) parser.add_argument( "--hidden-size", type=int, default=4096, help="Hidden size (2nd dimension) of the sequence", ) parser.add_argument( "--use-residual", action="store_true", help="Whether to use residual connection" ) parser.add_argument( "--save-path", type=str, default="./configs/rmsnorm/", help="Path to save rmsnorm benchmark results", ) args = parser.parse_args() # Run correctness test calculate_diff( batch_size=args.batch_size, seq_len=args.seq_len, hidden_size=args.hidden_size, use_residual=args.use_residual, ) # Get the benchmark function with proper use_residual setting benchmark = get_benchmark(args.use_residual) # Run performance benchmark benchmark.run(print_data=True, save_path=args.save_path)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/bench_per_token_quant_fp8.py
benchmarks/kernels/bench_per_token_quant_fp8.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import itertools from collections.abc import Callable from unittest.mock import patch import pandas as pd import torch from vllm.model_executor.layers.quantization.input_quant_fp8 import QuantFP8 from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape from vllm.triton_utils import triton from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.torch_utils import STR_DTYPE_TO_TORCH_DTYPE def with_triton_mode(fn): """Temporarily force the Triton fallback path""" def wrapped(*args, **kwargs): with patch("vllm.platforms.current_platform.is_cuda", return_value=False): return fn(*args, **kwargs) return wrapped # TODO(luka): use standalone_compile utility def with_dyn_arg(fn: Callable, arg_index: int, dim_index: int): def inner(*args): torch._dynamo.mark_dynamic(args[arg_index], dim_index) return fn(*args) return inner def bench_compile(fn: Callable): # recompile for different shapes fwd = torch.compile(fn, fullgraph=True, dynamic=False) # First dim is explicitly dynamic to simulate vLLM usage return with_dyn_arg(fwd, 0, 0) torch._dynamo.config.recompile_limit = 8888 def calculate_diff( batch_size: int, hidden_size: int, group_shape: GroupShape, dtype: torch.dtype, ): """Calculate the difference between Inductor and CUDA implementations.""" device = torch.device("cuda") x = torch.randn((batch_size, hidden_size), dtype=dtype, device=device) quant_fp8 = QuantFP8(False, group_shape, column_major_scales=False) torch_out, torch_scale = bench_compile(quant_fp8.forward_native)(x) torch_eager_out, torch_eager_scale = quant_fp8.forward_native(x) cuda_out, cuda_scale = quant_fp8.forward_cuda(x) try: torch.testing.assert_close( cuda_out.to(torch.float32), torch_out.to(torch.float32), rtol=1e-3, atol=1e-5, ) torch.testing.assert_close(cuda_scale, torch_scale, rtol=1e-3, atol=1e-5) torch.testing.assert_close( cuda_out.to(torch.float32), torch_eager_out.to(torch.float32), rtol=1e-3, atol=1e-5, ) torch.testing.assert_close(cuda_scale, torch_eager_scale, rtol=1e-3, atol=1e-5) print("✅ All implementations match") except AssertionError as e: print("❌ Implementations differ") print(e) configs = [] def benchmark_quantization( batch_size, hidden_size, provider, group_shape: GroupShape, col_major: bool, dtype: torch.dtype, ): device = torch.device("cuda") x = torch.randn(batch_size, hidden_size, device=device, dtype=dtype) quantiles = [0.5, 0.2, 0.8] quant_fp8 = QuantFP8(False, group_shape, column_major_scales=col_major) if provider == "torch": fn = lambda: bench_compile(quant_fp8.forward_native)(x.clone()) elif provider == "cuda": fn = lambda: quant_fp8.forward_cuda(x.clone()) elif provider == "triton": if not group_shape.is_per_group(): # Triton only supported for per-group return 0, 0, 0 fn = lambda: with_triton_mode(quant_fp8.forward_cuda)(x.clone()) ms, min_ms, max_ms = triton.testing.do_bench_cudagraph(fn, quantiles=quantiles) return 1000 * ms, 1000 * max_ms, 1000 * min_ms # TODO(luka) extract to utils def compute_geomean_speedups( df: pd.DataFrame, baseline_col: str, speedup_cols: list[str], groupby_cols: list[str] | None = None, ) -> pd.DataFrame: """ Compute geometric mean speedups over a baseline column. Args: df: Input dataframe baseline_col: Column to use as baseline speedup_cols: Columns to compute speedups for groupby_cols: Columns to group by. If None, compute over entire df. Returns: pd.DataFrame with geometric mean speedups """ from scipy.stats import gmean def geo_speedup(group: pd.DataFrame) -> pd.Series: ratios = { col: (group[baseline_col] / group[col]).values for col in speedup_cols } return pd.Series({col: gmean(vals) for col, vals in ratios.items()}) if groupby_cols is None: result = geo_speedup(df).to_frame().T else: result = ( df.groupby(groupby_cols) .apply(geo_speedup, include_groups=False) .reset_index() ) return result if __name__ == "__main__": parser = FlexibleArgumentParser( description="Benchmark the various implementations of QuantFP8 (dynamic-only)" ) parser.add_argument("-c", "--check", action="store_true") parser.add_argument( "--dtype", type=str, choices=["half", "bfloat16", "float"], default="bfloat16" ) parser.add_argument( "--hidden-sizes", type=int, nargs="+", default=[896, 1024, 2048, 4096, 7168], help="Hidden sizes to benchmark", ) parser.add_argument( "--batch-sizes", type=int, nargs="+", default=[1, 16, 128, 512, 1024], help="Batch sizes to benchmark", ) parser.add_argument( "--group-sizes", type=int, nargs="+", default=None, help="Group sizes for GroupShape(1,N) to benchmark. " "Use 0 for PER_TENSOR, -1 for PER_TOKEN (default: 0,-1,64,128)", ) parser.add_argument( "--no-column-major", action="store_true", help="Disable column-major scales testing", ) args = parser.parse_args() assert args dtype = STR_DTYPE_TO_TORCH_DTYPE[args.dtype] hidden_sizes = args.hidden_sizes batch_sizes = args.batch_sizes if args.group_sizes is not None: group_shapes = [] for size in args.group_sizes: if size == 0: group_shapes.append(GroupShape.PER_TENSOR) elif size == -1: group_shapes.append(GroupShape.PER_TOKEN) else: group_shapes.append(GroupShape(1, size)) else: group_shapes = [ GroupShape.PER_TENSOR, GroupShape.PER_TOKEN, GroupShape(1, 64), GroupShape(1, 128), ] column_major_scales = [False] if args.no_column_major else [True, False] config_gen = itertools.product( group_shapes, column_major_scales, batch_sizes, hidden_sizes, ) # filter out column-major scales for non-group, reverse order configs.extend(c[::-1] for c in config_gen if (c[0].is_per_group() or not c[1])) print(f"Running {len(configs)} configurations:") print(f" Hidden sizes: {hidden_sizes}") print(f" Batch sizes: {batch_sizes}") print(f" Group shapes: {[str(g) for g in group_shapes]}") print(f" Column major scales: {column_major_scales}") print() if args.check: for group_shape in group_shapes: group_size = group_shape[1] print(f"{group_size=}") calculate_diff( batch_size=4, hidden_size=4096, group_shape=group_shape, dtype=dtype ) benchmark = triton.testing.perf_report( triton.testing.Benchmark( x_names=["hidden_size", "batch_size", "col_major", "group_shape"], x_vals=configs, line_arg="provider", line_vals=["torch", "cuda", "triton"], line_names=["Torch (Compiled)", "CUDA", "Triton"], styles=[("blue", "-"), ("green", "-"), ("black", "-")], ylabel="us", plot_name="QuantFP8 performance", args={}, ) )(benchmark_quantization) df = benchmark.run(print_data=True, dtype=dtype, return_df=True) # Print geomean speedups geo_table_grouped = compute_geomean_speedups( df, baseline_col="Torch (Compiled)", speedup_cols=["CUDA", "Triton"], groupby_cols=["col_major", "group_shape"], ) print("Speedup over Torch (Compiled)") print(geo_table_grouped.to_string(index=False))
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/bench_fp8_gemm.py
benchmarks/kernels/bench_fp8_gemm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import copy import itertools import torch from weight_shapes import WEIGHT_SHAPES from vllm._custom_ops import cutlass_scaled_mm as vllm_scaled_mm from vllm._custom_ops import scaled_fp8_quant as vllm_scaled_fp8_quant from vllm.triton_utils import triton PROVIDER_CFGS = { "torch-bf16": dict(enabled=True), "fp8-tensor-w-token-a": dict( w="tensor", a="token", no_a_quant=False, enabled=False ), "fp8-tensor-w-tensor-a": dict( w="tensor", a="tensor", no_a_quant=False, enabled=True ), "fp8-channel-w-token-a": dict( w="channel", a="token", no_a_quant=False, enabled=True ), "fp8-channel-w-tensor-a": dict( w="channel", a="tensor", no_a_quant=False, enabled=False ), "fp8-tensor-w-token-a-noquant": dict( w="tensor", a="token", no_a_quant=True, enabled=False ), "fp8-tensor-w-tensor-a-noquant": dict( w="tensor", a="tensor", no_a_quant=True, enabled=True ), "fp8-channel-w-token-a-noquant": dict( w="channel", a="token", no_a_quant=True, enabled=True ), "fp8-channel-w-tensor-a-noquant": dict( w="channel", a="tensor", no_a_quant=True, enabled=False ), } _enabled = [k for k, v in PROVIDER_CFGS.items() if v["enabled"]] def _quant_weight_fp8(b: torch.Tensor, w_type: str, device: str): if w_type == "tensor": scale_b = torch.ones(1, device=device, dtype=torch.float32) b_fp8, scale_b_fp8 = vllm_scaled_fp8_quant(b, scale_b) else: b_fp8, scale_b_fp8 = vllm_scaled_fp8_quant(b, use_per_token_if_dynamic=True) return b_fp8.t(), scale_b_fp8 def build_fp8_runner(cfg, a, b, dtype, device): b_fp8, scale_b_fp8 = _quant_weight_fp8(b, cfg["w"], device) scale_a_const = ( torch.ones(1, device=device, dtype=torch.float32) if cfg["a"] == "tensor" else None ) if cfg["no_a_quant"]: if cfg["a"] == "tensor": a_fp8, scale_a_fp8 = vllm_scaled_fp8_quant(a, scale_a_const) else: a_fp8, scale_a_fp8 = vllm_scaled_fp8_quant(a, use_per_token_if_dynamic=True) def run(): return vllm_scaled_mm(a_fp8, b_fp8, scale_a_fp8, scale_b_fp8, dtype) return run if cfg["a"] == "tensor": def run(): a_fp8, scale_a_fp8 = vllm_scaled_fp8_quant(a, scale_a_const) return vllm_scaled_mm(a_fp8, b_fp8, scale_a_fp8, scale_b_fp8, dtype) else: def run(): a_fp8, scale_a_fp8 = vllm_scaled_fp8_quant(a, use_per_token_if_dynamic=True) return vllm_scaled_mm(a_fp8, b_fp8, scale_a_fp8, scale_b_fp8, dtype) return run @triton.testing.perf_report( triton.testing.Benchmark( x_names=["batch_size"], x_vals=[1, 16, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384], x_log=False, line_arg="provider", line_vals=_enabled, line_names=_enabled, ylabel="TFLOP/s (larger is better)", plot_name="BF16 vs FP8 GEMMs", args={}, ) ) def benchmark(batch_size, provider, N, K): M = batch_size device = "cuda" dtype = torch.bfloat16 a = torch.randn((M, K), device=device, dtype=dtype) b = torch.randn((N, K), device=device, dtype=dtype) quantiles = [0.5, 0.2, 0.8] if provider == "torch-bf16": ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: torch.nn.functional.linear(a, b), quantiles=quantiles ) else: cfg = PROVIDER_CFGS[provider] run_quant = build_fp8_runner(cfg, a, b, dtype, device) ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: run_quant(), quantiles=quantiles ) to_tflops = lambda t_ms: (2 * M * N * K) * 1e-12 / (t_ms * 1e-3) return to_tflops(ms), to_tflops(max_ms), to_tflops(min_ms) def prepare_shapes(args): out = [] for model, tp_size in itertools.product(args.models, args.tp_sizes): for KN, tp_dim in copy.deepcopy(WEIGHT_SHAPES[model]): KN[tp_dim] //= tp_size KN.append(model) out.append(KN) return out if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--models", nargs="+", type=str, default=["meta-llama/Llama-3.1-8B-Instruct"], choices=list(WEIGHT_SHAPES.keys()), ) parser.add_argument("--tp-sizes", nargs="+", type=int, default=[1]) args = parser.parse_args() for K, N, model in prepare_shapes(args): print(f"{model}, N={N} K={K}, BF16 vs FP8 GEMMs TFLOP/s:") benchmark.run( print_data=True, show_plots=True, save_path=f"bench_fp8_res_n{N}_k{K}", N=N, K=K, ) print("Benchmark finished!")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_moe.py
benchmarks/kernels/benchmark_moe.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import gc import json import os import time from contextlib import nullcontext from datetime import datetime from itertools import product from typing import Any, TypedDict import ray import torch from ray.experimental.tqdm_ray import tqdm from vllm.model_executor.layers.fused_moe.config import ( FusedMoEQuantConfig, _get_config_dtype_str, ) from vllm.model_executor.layers.fused_moe.fused_moe import * from vllm.platforms import current_platform from vllm.transformers_utils.config import get_config from vllm.triton_utils import triton from vllm.utils.argparse_utils import FlexibleArgumentParser FP8_DTYPE = current_platform.fp8_dtype() # Default interval for clearing Triton JIT cache during tuning # Set to 0 to disable automatic cache clearing _CACHE_CLEAR_INTERVAL_ENV = "VLLM_MOE_TUNE_CACHE_CLEAR_INTERVAL" TRITON_CACHE_CLEAR_INTERVAL = int(os.environ.get(_CACHE_CLEAR_INTERVAL_ENV, "50")) def clear_triton_cache(): """Clear Triton JIT compilation cache and Python/CUDA memory. This helps prevent OOM during tuning with large models (many experts). """ # Force Python garbage collection gc.collect() # Clear CUDA memory cache if torch.cuda.is_available(): torch.cuda.empty_cache() # Try to clear Triton's runtime cache try: import triton if ( hasattr(triton, "runtime") and hasattr(triton.runtime, "cache") and hasattr(triton.runtime.cache, "clear") ): triton.runtime.cache.clear() except ImportError: # Triton not installed, skip cache clearing pass except AttributeError: # Triton version doesn't have expected cache API pass except Exception as e: print(f"Warning: Failed to clear Triton cache: {e}") # Additional garbage collection after clearing caches gc.collect() def ensure_divisibility(numerator, denominator, text): """Ensure that numerator is divisible by the denominator.""" assert numerator % denominator == 0, "{} {} is not divisible by tp {}.".format( text, numerator, denominator ) class BenchmarkConfig(TypedDict): BLOCK_SIZE_M: int BLOCK_SIZE_N: int BLOCK_SIZE_K: int GROUP_SIZE_M: int num_warps: int num_stages: int def benchmark_config( config: BenchmarkConfig, num_tokens: int, num_experts: int, shard_intermediate_size: int, hidden_size: int, topk: int, dtype: torch.dtype, use_fp8_w8a8: bool, use_int8_w8a16: bool, num_iters: int = 100, block_quant_shape: list[int] = None, use_deep_gemm: bool = False, ) -> float: init_dtype = torch.float16 if use_fp8_w8a8 else dtype x = torch.randn(num_tokens, hidden_size, dtype=dtype) if use_int8_w8a16: w1 = torch.randint( -127, 127, ( num_experts, shard_intermediate_size, hidden_size, ), dtype=torch.int8, ) w2 = torch.randint( -127, 127, ( num_experts, hidden_size, shard_intermediate_size // 2, ), dtype=torch.int8, ) else: w1 = torch.randn( num_experts, shard_intermediate_size, hidden_size, dtype=init_dtype ) w2 = torch.randn( num_experts, hidden_size, shard_intermediate_size // 2, dtype=init_dtype ) gating_output = torch.randn(num_iters, num_tokens, num_experts, dtype=torch.float32) w1_scale = None w2_scale = None a1_scale = None a2_scale = None if use_int8_w8a16: w1_scale = torch.randn( (num_experts, 2 * shard_intermediate_size), dtype=torch.float32 ) w2_scale = torch.randn((hidden_size, num_experts), dtype=torch.float32) if use_deep_gemm: # we use the default block shape for deepgemm block_quant_shape = [128, 128] if use_fp8_w8a8: if block_quant_shape: block_n, block_k = block_quant_shape[0], block_quant_shape[1] E = num_experts N = shard_intermediate_size // 2 K = hidden_size factor_for_scale = 1e-2 n_tiles_w1 = (2 * N + block_n - 1) // block_n n_tiles_w2 = (K + block_n - 1) // block_n k_tiles_w1 = (K + block_k - 1) // block_k k_tiles_w2 = (N + block_k - 1) // block_k w1_scale = ( torch.rand((E, n_tiles_w1, k_tiles_w1), dtype=torch.float32) * factor_for_scale ) w2_scale = ( torch.rand((E, n_tiles_w2, k_tiles_w2), dtype=torch.float32) * factor_for_scale ) else: w1_scale = torch.randn(num_experts, dtype=torch.float32) w2_scale = torch.randn(num_experts, dtype=torch.float32) a1_scale = torch.randn(1, dtype=torch.float32) a2_scale = torch.randn(1, dtype=torch.float32) w1 = w1.to(FP8_DTYPE) w2 = w2.to(FP8_DTYPE) input_gating = torch.empty(num_tokens, num_experts, dtype=torch.float32) def prepare(i: int): input_gating.copy_(gating_output[i]) def run(): from vllm.model_executor.layers.fused_moe import override_config if use_fp8_w8a8: quant_dtype = torch.float8_e4m3fn elif use_int8_w8a16: quant_dtype = torch.int8 else: quant_dtype = None quant_config = FusedMoEQuantConfig.make( quant_dtype=quant_dtype, w1_scale=w1_scale, w2_scale=w2_scale, a1_scale=a1_scale, a2_scale=a2_scale, block_shape=block_quant_shape, ) with override_config(config): topk_weights, topk_ids, token_expert_indices = fused_topk( x, input_gating, topk, renormalize=not use_deep_gemm ) return fused_experts( x, w1, w2, topk_weights, topk_ids, inplace=True, quant_config=quant_config, allow_deep_gemm=use_deep_gemm, ) # JIT compilation & warmup run() torch.cuda.synchronize() # Capture 10 invocations with CUDA graph graph = torch.cuda.CUDAGraph() with torch.cuda.graph(graph): for _ in range(10): run() torch.cuda.synchronize() # Warmup for _ in range(5): graph.replay() torch.cuda.synchronize() start_event = torch.Event(enable_timing=True) end_event = torch.Event(enable_timing=True) latencies: list[float] = [] for i in range(num_iters): prepare(i) torch.cuda.synchronize() start_event.record() graph.replay() end_event.record() end_event.synchronize() latencies.append(start_event.elapsed_time(end_event)) avg = sum(latencies) / (num_iters * 10) * 1000 # us graph.reset() return avg def get_rocm_tuning_space(use_fp16): block_mn_range = [16, 32, 64, 128, 256] block_k_range = [16, 32, 64, 128, 256] if not use_fp16: block_k_range.remove(16) # BLOCK_K=16 not supported for fp8 num_warps_range = [1, 2, 4, 8] group_m_range = [1, 4, 8, 16, 32] num_stage_range = [2] waves_per_eu_range = [0, 1, 2, 4] matrix_instr_nonkdim_range = [16, 32] if use_fp16 else [] kpack_range = [1, 2] if use_fp16 else [] param_ranges = { "BLOCK_SIZE_M": block_mn_range, "BLOCK_SIZE_N": block_mn_range, "BLOCK_SIZE_K": block_k_range, "GROUP_SIZE_M": group_m_range, "num_warps": num_warps_range, "num_stages": num_stage_range, "waves_per_eu": waves_per_eu_range, } if use_fp16: param_ranges["matrix_instr_nonkdim"] = matrix_instr_nonkdim_range param_ranges["kpack"] = kpack_range return param_ranges def get_configs_compute_bound(use_fp16, block_quant_shape) -> list[dict[str, int]]: configs: list[BenchmarkConfig] = [] if current_platform.is_rocm(): param_ranges = get_rocm_tuning_space(use_fp16) else: # Reduced search space for faster tuning. # TODO(woosuk): Increase the search space and use a performance model to # prune the search space. block_m_range = [16, 32, 64, 128, 256] block_n_range = [32, 64, 128, 256] block_k_range = [64, 128, 256] num_warps_range = [4, 8] group_m_range = [1, 16, 32, 64] num_stage_range = [2, 3, 4, 5] param_ranges = { "BLOCK_SIZE_M": block_m_range, "BLOCK_SIZE_N": block_n_range, "BLOCK_SIZE_K": block_k_range, "GROUP_SIZE_M": group_m_range, "num_warps": num_warps_range, "num_stages": num_stage_range, } keys, values = zip(*param_ranges.items()) for config_values in product(*values): config = dict(zip(keys, config_values)) configs.append(config) # Remove configs that are not compatible with fp8 block quantization # BLOCK_SIZE_K must be a multiple of block_k # BLOCK_SIZE_N must be a multiple of block_n if block_quant_shape is not None and not use_fp16: block_n, block_k = block_quant_shape[0], block_quant_shape[1] for config in configs[:]: if ( config["BLOCK_SIZE_K"] % block_k != 0 or config["BLOCK_SIZE_N"] % block_n != 0 ): configs.remove(config) return configs def prune_rocm_search_space( num_tokens, shard_intermediate_size, hidden_size, search_space, is_fp16, topk ): N1, K1 = shard_intermediate_size, hidden_size N2, K2 = hidden_size, shard_intermediate_size // 2 pruned_space_1 = prune_rocm_configs( num_tokens * topk, N1, K1, search_space, is_fp16 ) pruned_space_2 = prune_rocm_configs( num_tokens * topk, N2, K2, search_space, is_fp16 ) search_space = merge_unique_dicts(pruned_space_1, pruned_space_2) return search_space # The following code is inspired by ROCm/Triton GEMM tuning script: # https://github.com/ROCm/triton/blob/triton-mlir/scripts/amd/gemm/tune_gemm.py#L89 def prune_rocm_configs(M, N, K, configs, is_fp16=True): pruned_configs = [] elemBytes_a = 2 if is_fp16 else 1 elemBytes_b = 2 if is_fp16 else 1 mfma = 16 if M < 32 or N < 32 else 32 # TODO (zhanglx): figure out the boundary between large and small gemms large_gemm = False if M >= 2048 and N >= 2048: large_gemm = True for config in configs: BLOCK_SIZE_M = config.get("BLOCK_SIZE_M") BLOCK_SIZE_N = config.get("BLOCK_SIZE_N") BLOCK_SIZE_K = config.get("BLOCK_SIZE_K") num_warps = config.get("num_warps") if is_fp16: matrix_instr_nonkdim = config.get("matrix_instr_nonkdim") if matrix_instr_nonkdim > mfma: continue if mfma == 4 and BLOCK_SIZE_K < 64: continue # some layouts could not work properly in case # number elements per thread is less 1 if BLOCK_SIZE_M * BLOCK_SIZE_N < 64: continue SPLIT_K = config.get("SPLIT_K", 1) GROUP_M = config.get("GROUP_SIZE_M") if is_fp16: if ( matrix_instr_nonkdim > BLOCK_SIZE_M or matrix_instr_nonkdim > BLOCK_SIZE_N ): continue if matrix_instr_nonkdim >= M and matrix_instr_nonkdim != BLOCK_SIZE_M: continue if matrix_instr_nonkdim >= N and matrix_instr_nonkdim != BLOCK_SIZE_N: continue # Skip BLOCK_SIZE that is too large compare to M/N # unless BLOCK_SIZE is already small enough if M * 2 < BLOCK_SIZE_M and BLOCK_SIZE_M != 16: continue if N * 2 < BLOCK_SIZE_N and BLOCK_SIZE_N != 16: continue # skip large split_k when not necessary if SPLIT_K != 1 and not need_split_k(M, N, K): continue # skip split_k that leads to EVEN_K = false leap = SPLIT_K * BLOCK_SIZE_K modv = K % leap if modv != 0: continue # skip large GROUP_M if GROUP_M * BLOCK_SIZE_M > M and GROUP_M != 1: continue # out of shared memory resource # TODO (zhanglx): This does not consider the LDS usage in the epilogue LDS = ( BLOCK_SIZE_K * BLOCK_SIZE_M * elemBytes_a + BLOCK_SIZE_K * BLOCK_SIZE_N * elemBytes_b ) if LDS > 65536: continue # Skip small block sizes and num_warps for large gemm # For fp16 and f8, we want to only use BLOCK_SIZE >= 64 if large_gemm: if BLOCK_SIZE_M < 64 or BLOCK_SIZE_N < 64: continue if BLOCK_SIZE_K < 64: continue if num_warps < 4: continue pruned_configs.append(config) return pruned_configs def need_split_k(SIZE_M, SIZE_N, SIZE_K): return (SIZE_M < 64 or SIZE_N < 64) and SIZE_K > 1024 def merge_unique_dicts(list1, list2): result = [] combined_list = list1.copy() combined_list.extend(list2) for dictionary in combined_list: if dictionary not in result: result.append(dictionary) return result @ray.remote(num_gpus=1) class BenchmarkWorker: def __init__(self, seed: int) -> None: torch.set_default_device("cuda") current_platform.seed_everything(seed) self.seed = seed # Get the device ID to allocate tensors and kernels # on the respective GPU. This is required for Ray to work # correctly with multi-GPU tuning on the ROCm platform. self.device_id = int(ray.get_gpu_ids()[0]) def benchmark( self, num_tokens: int, num_experts: int, shard_intermediate_size: int, hidden_size: int, topk: int, dtype: torch.dtype, use_fp8_w8a8: bool, use_int8_w8a16: bool, block_quant_shape: list[int] = None, use_deep_gemm: bool = False, ) -> tuple[dict[str, int], float]: current_platform.seed_everything(self.seed) dtype_str = _get_config_dtype_str( dtype, use_int8_w8a16=use_int8_w8a16, use_fp8_w8a8=use_fp8_w8a8 ) # NOTE(woosuk): The current naming convention uses w2.shape[2], which # is the intermediate size after silu_and_mul. block_n = block_quant_shape[0] if block_quant_shape else None block_k = block_quant_shape[1] if block_quant_shape else None op_config = get_moe_configs( num_experts, shard_intermediate_size // 2, dtype_str, block_n, block_k ) if op_config is None: config = get_default_config( num_tokens, num_experts, shard_intermediate_size, hidden_size, topk, dtype_str, block_quant_shape, ) else: config = op_config[min(op_config.keys(), key=lambda x: abs(x - num_tokens))] kernel_time = benchmark_config( config, num_tokens, num_experts, shard_intermediate_size, hidden_size, topk, dtype, use_fp8_w8a8, use_int8_w8a16, num_iters=100, block_quant_shape=block_quant_shape, use_deep_gemm=use_deep_gemm, ) return config, kernel_time def tune( self, num_tokens: int, num_experts: int, shard_intermediate_size: int, hidden_size: int, topk: int, dtype: torch.dtype, use_fp8_w8a8: bool, use_int8_w8a16: bool, search_space: list[dict[str, int]], block_quant_shape: list[int], use_deep_gemm: bool, ) -> dict[str, int]: best_config = None best_time = float("inf") if current_platform.is_rocm(): is_fp16 = not (use_fp8_w8a8 or use_int8_w8a16) search_space = prune_rocm_search_space( num_tokens, shard_intermediate_size, hidden_size, search_space, is_fp16, topk, ) need_device_guard = False if current_platform.is_rocm(): visible_device = os.environ.get("ROCR_VISIBLE_DEVICES", None) if visible_device != f"{self.device_id}": need_device_guard = True with torch.cuda.device(self.device_id) if need_device_guard else nullcontext(): for idx, config in enumerate(tqdm(search_space)): try: kernel_time = benchmark_config( config, num_tokens, num_experts, shard_intermediate_size, hidden_size, topk, dtype, use_fp8_w8a8, use_int8_w8a16, num_iters=20, block_quant_shape=block_quant_shape, use_deep_gemm=use_deep_gemm, ) except triton.runtime.autotuner.OutOfResources: # Some configurations may be invalid and fail to compile. continue if kernel_time < best_time: best_time = kernel_time best_config = config # Periodically clear Triton JIT cache to prevent OOM # This is especially important for large models with many experts if ( TRITON_CACHE_CLEAR_INTERVAL > 0 and idx > 0 and idx % TRITON_CACHE_CLEAR_INTERVAL == 0 ): clear_triton_cache() # Final cleanup after tuning completes clear_triton_cache() now = datetime.now() print(f"{now.ctime()}] Completed tuning for batch_size={num_tokens}") assert best_config is not None return best_config def sort_config(config: BenchmarkConfig) -> BenchmarkConfig: return { "BLOCK_SIZE_M": config["BLOCK_SIZE_M"], "BLOCK_SIZE_N": config["BLOCK_SIZE_N"], "BLOCK_SIZE_K": config["BLOCK_SIZE_K"], "GROUP_SIZE_M": config["GROUP_SIZE_M"], "num_warps": config["num_warps"], "num_stages": config["num_stages"], **( {"waves_per_eu": config["waves_per_eu"]} if "waves_per_eu" in config else {} ), **( {"matrix_instr_nonkdim": config["matrix_instr_nonkdim"]} if "matrix_instr_nonkdim" in config else {} ), **({"kpack": config["kpack"]} if "kpack" in config else {}), } def save_configs( configs: dict[int, BenchmarkConfig], num_experts: int, shard_intermediate_size: int, hidden_size: int, topk: int, dtype: torch.dtype, use_fp8_w8a8: bool, use_int8_w8a16: bool, block_quant_shape: list[int], save_dir: str, ) -> None: dtype_str = _get_config_dtype_str( dtype, use_int8_w8a16=use_int8_w8a16, use_fp8_w8a8=use_fp8_w8a8 ) # NOTE(woosuk): The current naming convention uses w2.shape[2], which # is the intermediate size after silu_and_mul. filename = get_config_file_name( num_experts, shard_intermediate_size // 2, dtype_str, block_quant_shape ) os.makedirs(save_dir, exist_ok=True) filename = os.path.join(save_dir, filename) print(f"Writing best config to {filename}...") with open(filename, "w") as f: json.dump({"triton_version": triton.__version__, **configs}, f, indent=4) f.write("\n") def get_weight_block_size_safety(config, default_value=None): quantization_config = getattr(config, "quantization_config", {}) if isinstance(quantization_config, dict): return quantization_config.get("weight_block_size", default_value) return default_value def main(args: argparse.Namespace): print(args) config = get_config(model=args.model, trust_remote_code=args.trust_remote_code) if args.model_prefix: config = getattr(config, args.model_prefix) if config.architectures[0] == "DbrxForCausalLM": E = config.ffn_config.moe_num_experts topk = config.ffn_config.moe_top_k intermediate_size = config.ffn_config.ffn_hidden_size hidden_size = config.hidden_size elif config.architectures[0] == "JambaForCausalLM": E = config.num_experts topk = config.num_experts_per_tok intermediate_size = config.intermediate_size hidden_size = config.hidden_size elif config.architectures[0] in ( "DeepseekV2ForCausalLM", "DeepseekV3ForCausalLM", "DeepseekV32ForCausalLM", "Glm4MoeForCausalLM", "NemotronHForCausalLM", ): E = config.n_routed_experts topk = config.num_experts_per_tok intermediate_size = config.moe_intermediate_size hidden_size = config.hidden_size elif config.architectures[0] in ( "Qwen2MoeForCausalLM", "Qwen3MoeForCausalLM", "Qwen3NextForCausalLM", ): E = config.num_experts topk = config.num_experts_per_tok intermediate_size = config.moe_intermediate_size hidden_size = config.hidden_size elif config.architectures[0] == "Qwen3VLMoeForConditionalGeneration": text_config = config.get_text_config() E = text_config.num_experts topk = text_config.num_experts_per_tok intermediate_size = text_config.moe_intermediate_size hidden_size = text_config.hidden_size elif config.architectures[0] in ("HunYuanMoEV1ForCausalLM"): E = config.num_experts topk = config.moe_topk[0] intermediate_size = config.moe_intermediate_size[0] hidden_size = config.hidden_size elif config.architectures[0] in ["Qwen3OmniMoeForConditionalGeneration"]: E = config.thinker_config.text_config.num_experts topk = config.thinker_config.text_config.num_experts_per_tok intermediate_size = config.thinker_config.text_config.moe_intermediate_size hidden_size = config.thinker_config.text_config.hidden_size else: # Support for llama4 config = config.get_text_config() # Default: Mixtral. E = config.num_local_experts topk = config.num_experts_per_tok intermediate_size = config.intermediate_size hidden_size = config.hidden_size enable_ep = bool(args.enable_expert_parallel) if enable_ep: ensure_divisibility(E, args.tp_size, "Number of experts") E = E // args.tp_size shard_intermediate_size = 2 * intermediate_size else: ensure_divisibility(intermediate_size, args.tp_size, "intermediate_size") shard_intermediate_size = 2 * intermediate_size // args.tp_size dtype = torch.float16 if current_platform.is_rocm() else config.dtype use_fp8_w8a8 = args.dtype == "fp8_w8a8" use_int8_w8a16 = args.dtype == "int8_w8a16" block_quant_shape = get_weight_block_size_safety(config) if args.batch_size is None: batch_sizes = [ 1, 2, 4, 8, 16, 24, 32, 48, 64, 96, 128, 256, 512, 1024, 1536, 2048, 3072, 4096, ] else: batch_sizes = args.batch_size use_deep_gemm = bool(args.use_deep_gemm) if current_platform.is_rocm() and "HIP_VISIBLE_DEVICES" in os.environ: # Ray will set ROCR_VISIBLE_DEVICES for device visibility logger.warning( "Ray uses ROCR_VISIBLE_DEVICES to control device accessibility." "Replacing HIP_VISIBLE_DEVICES with ROCR_VISIBLE_DEVICES." ) val = os.environ["HIP_VISIBLE_DEVICES"] os.environ["ROCR_VISIBLE_DEVICES"] = val del os.environ["HIP_VISIBLE_DEVICES"] ray.init() num_gpus = int(ray.available_resources()["GPU"]) workers = [BenchmarkWorker.remote(args.seed) for _ in range(num_gpus)] def _distribute(method: str, inputs: list[Any]) -> list[Any]: outputs = [] worker_idx = 0 for input_args in inputs: worker = workers[worker_idx] worker_method = getattr(worker, method) output = worker_method.remote(*input_args) outputs.append(output) worker_idx = (worker_idx + 1) % num_gpus return ray.get(outputs) if args.tune: is_fp16 = not (use_fp8_w8a8 or use_int8_w8a16) search_space = get_configs_compute_bound(is_fp16, block_quant_shape) print(f"Start tuning over {len(search_space)} configurations...") if use_deep_gemm: raise ValueError( "Tuning with --use-deep-gemm is not supported as it only tunes Triton " "kernels. Please remove the flag." ) start = time.time() configs = _distribute( "tune", [ ( batch_size, E, shard_intermediate_size, hidden_size, topk, dtype, use_fp8_w8a8, use_int8_w8a16, search_space, block_quant_shape, use_deep_gemm, ) for batch_size in batch_sizes ], ) best_configs = { M: sort_config(config) for M, config in zip(batch_sizes, configs) } save_configs( best_configs, E, shard_intermediate_size, hidden_size, topk, dtype, use_fp8_w8a8, use_int8_w8a16, block_quant_shape, args.save_dir, ) end = time.time() print(f"Tuning took {end - start:.2f} seconds") else: outputs = _distribute( "benchmark", [ ( batch_size, E, shard_intermediate_size, hidden_size, topk, dtype, use_fp8_w8a8, use_int8_w8a16, block_quant_shape, use_deep_gemm, ) for batch_size in batch_sizes ], ) for batch_size, (config, kernel_time) in zip(batch_sizes, outputs): print(f"Batch size: {batch_size}, config: {config}") print(f"Kernel time: {kernel_time:.2f} us") if __name__ == "__main__": parser = FlexibleArgumentParser() parser.add_argument( "--model", type=str, default="mistralai/Mixtral-8x7B-Instruct-v0.1" ) parser.add_argument( "--tp-size", "-tp", "--tensor-parallel-size", type=int, default=2 ) parser.add_argument("--enable-expert-parallel", "-enable-ep", action="store_true") parser.add_argument( "--dtype", type=str, choices=["auto", "fp8_w8a8", "int8_w8a16"], default="auto" ) parser.add_argument("--use-deep-gemm", action="store_true") parser.add_argument( "--save-dir", type=str, default="./", help="Directory to save tuned results" ) parser.add_argument("--seed", type=int, default=0) parser.add_argument("--batch-size", type=int, nargs="+", required=False) parser.add_argument("--tune", action="store_true") parser.add_argument("--trust-remote-code", action="store_true") parser.add_argument("--model-prefix", type=str, required=False) args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_cutlass_moe_fp8.py
benchmarks/kernels/benchmark_cutlass_moe_fp8.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Benchmark the performance of the cutlass_moe_fp8 kernel vs the triton_moe kernel. Both kernels take in fp8 quantized weights and 16-bit activations, but use different quantization strategies and backends. """ import nvtx import torch from vllm import _custom_ops as ops from vllm.model_executor.layers.fused_moe.config import fp8_w8a8_moe_quant_config from vllm.model_executor.layers.fused_moe.cutlass_moe import cutlass_moe_fp8 from vllm.model_executor.layers.fused_moe.fused_moe import fused_experts, fused_topk from vllm.platforms import current_platform from vllm.utils.argparse_utils import FlexibleArgumentParser # Weight shapes for different models: [num_experts, topk, hidden_size, # intermediate_size] WEIGHT_SHAPES_MOE = { "mixtral-8x7b": [ [8, 2, 4096, 14336], ], "deepseek-v2": [ [160, 6, 5120, 12288], ], "custom-small": [ [8, 2, 2048, 7168], ], "glm45-fp8": [ [128, 8, 4096, 1408], ], "Llama-4-Maverick-17B-128E-Instruct-FP8": [ [128, 1, 5120, 8192], ], } DEFAULT_MODELS = [ "mixtral-8x7b", ] DEFAULT_BATCH_SIZES = [4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048] DEFAULT_TP_SIZES = [1] PER_ACT_TOKEN_OPTS = [False, True] PER_OUT_CH_OPTS = [False, True] FP8_DTYPE = current_platform.fp8_dtype() def bench_run( results: list, model: str, num_experts: int, topk: int, per_act_token: bool, per_out_ch: bool, mkn: tuple[int, int, int], ): (m, k, n) = mkn dtype = torch.half device = "cuda" # Create input activations a = torch.randn((m, k), device=device, dtype=dtype) / 10 # Create weights w1 = torch.randn((num_experts, 2 * n, k), device=device, dtype=dtype) / 10 w2 = torch.randn((num_experts, k, n), device=device, dtype=dtype) / 10 # Create FP8 quantized weights and scales for both kernels w1_fp8q = torch.empty((num_experts, 2 * n, k), device=device, dtype=FP8_DTYPE) w2_fp8q = torch.empty((num_experts, k, n), device=device, dtype=FP8_DTYPE) # Create scales based on quantization strategy if per_out_ch: # Per-channel quantization w1_scale = torch.empty( (num_experts, 2 * n, 1), device=device, dtype=torch.float32 ) w2_scale = torch.empty((num_experts, k, 1), device=device, dtype=torch.float32) else: # Per-tensor quantization w1_scale = torch.empty((num_experts, 1, 1), device=device, dtype=torch.float32) w2_scale = torch.empty((num_experts, 1, 1), device=device, dtype=torch.float32) # Quantize weights for expert in range(num_experts): if per_out_ch: # Per-channel quantization - not yet implemented properly # For now, fall back to per-tensor quantization w1_fp8q[expert], w1_scale_temp = ops.scaled_fp8_quant(w1[expert]) w2_fp8q[expert], w2_scale_temp = ops.scaled_fp8_quant(w2[expert]) # Expand scalar scales to the expected per-channel shape w1_scale[expert] = w1_scale_temp.expand(2 * n, 1) w2_scale[expert] = w2_scale_temp.expand(k, 1) else: # Per-tensor quantization w1_fp8q[expert], w1_scale_temp = ops.scaled_fp8_quant(w1[expert]) w2_fp8q[expert], w2_scale_temp = ops.scaled_fp8_quant(w2[expert]) # Store scalar scales in [1, 1] tensors w1_scale[expert, 0, 0] = w1_scale_temp w2_scale[expert, 0, 0] = w2_scale_temp # Prepare weights for CUTLASS (no transpose needed) w1_fp8q_cutlass = w1_fp8q # Keep original [E, 2N, K] w2_fp8q_cutlass = w2_fp8q # Keep original [E, K, N] # Create router scores and get topk score = torch.randn((m, num_experts), device=device, dtype=dtype) topk_weights, topk_ids, _ = fused_topk(a, score, topk, renormalize=False) # WORKAROUND: CUTLASS MoE FP8 has issues with per-token quantization # Force per-tensor quantization for all cases to match working e2e setup a1_scale = torch.full((), 1e-2, device=device, dtype=torch.float32) a2_scale = torch.full((), 1e-2, device=device, dtype=torch.float32) # Force per-tensor quantization for all cases per_act_token = False # Create stride tensors for CUTLASS ab_strides1 = torch.full((num_experts,), k, dtype=torch.int64, device=device) ab_strides2 = torch.full((num_experts,), n, dtype=torch.int64, device=device) c_strides1 = torch.full((num_experts,), 2 * n, dtype=torch.int64, device=device) c_strides2 = torch.full((num_experts,), k, dtype=torch.int64, device=device) def run_triton_moe( a: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, w1_scale: torch.Tensor, w2_scale: torch.Tensor, a1_scale: torch.Tensor, a2_scale: torch.Tensor, num_repeats: int, ): quant_config = fp8_w8a8_moe_quant_config( w1_scale=w1_scale, w2_scale=w2_scale, a1_scale=a1_scale, a2_scale=a2_scale, per_act_token_quant=per_act_token, per_out_ch_quant=per_out_ch, ) for _ in range(num_repeats): fused_experts( a, w1, w2, topk_weights, topk_ids, quant_config=quant_config, ) def run_cutlass_moe_fp8( a: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, ab_strides1: torch.Tensor, ab_strides2: torch.Tensor, c_strides1: torch.Tensor, c_strides2: torch.Tensor, w1_scale: torch.Tensor, w2_scale: torch.Tensor, a1_scale: torch.Tensor, a2_scale: torch.Tensor, num_repeats: int, ): quant_config = fp8_w8a8_moe_quant_config( w1_scale=w1_scale, w2_scale=w2_scale, a1_scale=a1_scale, a2_scale=a2_scale, per_act_token_quant=per_act_token, per_out_ch_quant=per_out_ch, ) for _ in range(num_repeats): with nvtx.annotate("cutlass_moe_fp8", color="blue"): cutlass_moe_fp8( a=a, w1_q=w1, w2_q=w2, topk_weights=topk_weights, topk_ids=topk_ids, ab_strides1=ab_strides1, ab_strides2=ab_strides2, c_strides1=c_strides1, c_strides2=c_strides2, quant_config=quant_config, activation="silu", global_num_experts=num_experts, ) # Pre-create quantization config to avoid creating it inside CUDA graph quant_config = fp8_w8a8_moe_quant_config( w1_scale=w1_scale, w2_scale=w2_scale, a1_scale=a1_scale, a2_scale=a2_scale, per_act_token_quant=per_act_token, per_out_ch_quant=per_out_ch, ) # Create CUDA graphs for CUTLASS (match benchmark_moe.py pattern exactly) cutlass_stream = torch.cuda.Stream() cutlass_graph = torch.cuda.CUDAGraph() with torch.cuda.graph(cutlass_graph, stream=cutlass_stream): # Capture 10 invocations like benchmark_moe.py for _ in range(10): cutlass_moe_fp8( a=a, w1_q=w1_fp8q_cutlass, w2_q=w2_fp8q_cutlass, topk_weights=topk_weights, topk_ids=topk_ids, ab_strides1=ab_strides1, ab_strides2=ab_strides2, c_strides1=c_strides1, c_strides2=c_strides2, quant_config=quant_config, activation="silu", global_num_experts=num_experts, ) torch.cuda.synchronize() # Create CUDA graphs for Triton (match benchmark_moe.py pattern exactly) triton_stream = torch.cuda.Stream() triton_graph = torch.cuda.CUDAGraph() with torch.cuda.graph(triton_graph, stream=triton_stream): # Capture 10 invocations like benchmark_moe.py for _ in range(10): fused_experts( a, w1_fp8q, w2_fp8q, topk_weights, topk_ids, quant_config=quant_config, ) torch.cuda.synchronize() def bench_cuda_graph(graph, num_warmup=5, num_iters=100): """Benchmark CUDA graph using events like benchmark_moe.py""" # Warmup for _ in range(num_warmup): graph.replay() torch.cuda.synchronize() # Timing start_event = torch.Event(enable_timing=True) end_event = torch.Event(enable_timing=True) latencies = [] for _ in range(num_iters): torch.cuda.synchronize() start_event.record() graph.replay() end_event.record() end_event.synchronize() latencies.append(start_event.elapsed_time(end_event)) # Divide by 10 since graph contains 10 calls return sum(latencies) / (num_iters * 10) # Benchmark parameters num_warmup = 5 num_iters = 100 # Benchmark only CUDA graphs (more reliable and faster) # Benchmark Triton MoE with CUDA graphs triton_graph_time = bench_cuda_graph( triton_graph, num_warmup=num_warmup, num_iters=num_iters ) # Benchmark CUTLASS MoE with CUDA graphs cutlass_graph_time = bench_cuda_graph( cutlass_graph, num_warmup=num_warmup, num_iters=num_iters ) # Convert ms to us and return results triton_time_us = triton_graph_time * 1000 cutlass_time_us = cutlass_graph_time * 1000 return { "batch_size": m, "triton_time_us": triton_time_us, "cutlass_time_us": cutlass_time_us, } def main(args): print("Benchmarking models:") for i, model in enumerate(args.models): print(f"[{i}] {model}") all_results = [] for model in args.models: for tp in args.tp_sizes: for layer in WEIGHT_SHAPES_MOE[model]: num_experts = layer[0] topk = layer[1] size_k = layer[2] size_n = layer[3] // tp if len(args.limit_k) > 0 and size_k not in args.limit_k: continue if len(args.limit_n) > 0 and size_n not in args.limit_n: continue for per_act_token in args.per_act_token_opts: for per_out_ch in args.per_out_ch_opts: print( f"\n=== {model}, experts={num_experts}, topk={topk}," f"per_act={per_act_token}, per_out_ch={per_out_ch} ===" ) config_results = [] for size_m in args.batch_sizes: mkn = (size_m, size_k, size_n) result = bench_run( [], # Not used anymore model, num_experts, topk, per_act_token, per_out_ch, mkn, ) if result: config_results.append(result) # Print results table for this configuration if config_results: print( f"\n{'Batch Size':<12}" f"{'Triton (us)':<15}" f"{'CUTLASS (us)':<15}" ) print("-" * 45) for result in config_results: print( f"{result['batch_size']:<12}" f"{result['triton_time_us']:<15.2f}" f"{result['cutlass_time_us']:<15.2f}" ) all_results.extend(config_results) print(f"\nTotal benchmarks completed: {len(all_results)}") if __name__ == "__main__": parser = FlexibleArgumentParser( description="""Benchmark CUTLASS FP8 MOE vs Triton FP8 FUSED MOE across specified models/shapes/batches Example usage: python benchmark_cutlass_moe_fp8.py \ --model "Llama-4-Maverick-17B-128E-Instruct-FP8" \ --tp-sizes 8 \ --batch-size 2 4 8 \ --per-act-token-opts false \ --per-out-ch-opts false """ ) parser.add_argument( "--models", nargs="+", type=str, default=DEFAULT_MODELS, choices=WEIGHT_SHAPES_MOE.keys(), ) parser.add_argument("--tp-sizes", nargs="+", type=int, default=DEFAULT_TP_SIZES) parser.add_argument( "--batch-sizes", nargs="+", type=int, default=DEFAULT_BATCH_SIZES ) parser.add_argument("--limit-k", nargs="+", type=int, default=[]) parser.add_argument("--limit-n", nargs="+", type=int, default=[]) parser.add_argument( "--per-act-token-opts", nargs="+", type=lambda x: x.lower() == "true", default=[False, True], help="Per-activation token quantization options (true/false)", ) parser.add_argument( "--per-out-ch-opts", nargs="+", type=lambda x: x.lower() == "true", default=[False, True], help="Per-output channel quantization options (true/false)", ) args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/bench_int8_gemm.py
benchmarks/kernels/bench_int8_gemm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import copy import itertools import torch from weight_shapes import WEIGHT_SHAPES from vllm._custom_ops import cutlass_scaled_mm as vllm_scaled_mm from vllm._custom_ops import scaled_int8_quant as vllm_scaled_int8_quant from vllm.triton_utils import triton PROVIDER_CFGS = { "torch-bf16": dict(enabled=True), "int8-tensor-w-token-a": dict( w="tensor", a="token", no_a_quant=False, enabled=False ), "int8-tensor-w-tensor-a": dict( w="tensor", a="tensor", no_a_quant=False, enabled=True ), "int8-channel-w-token-a": dict( w="channel", a="token", no_a_quant=False, enabled=True ), "int8-channel-w-tensor-a": dict( w="channel", a="tensor", no_a_quant=False, enabled=False ), "int8-tensor-w-token-a-noquant": dict( w="tensor", a="token", no_a_quant=True, enabled=False ), "int8-tensor-w-tensor-a-noquant": dict( w="tensor", a="tensor", no_a_quant=True, enabled=True ), "int8-channel-w-token-a-noquant": dict( w="channel", a="token", no_a_quant=True, enabled=True ), "int8-channel-w-tensor-a-noquant": dict( w="channel", a="tensor", no_a_quant=True, enabled=False ), } def _quant_weight(b, w_type, device): if w_type == "tensor": scale_b = torch.ones(1, device=device, dtype=torch.float32) b_int8, scale_b_int8, _ = vllm_scaled_int8_quant(b, scale_b) assert scale_b_int8.numel() == 1 else: # channel b_int8, scale_b_int8, _ = vllm_scaled_int8_quant(b) assert scale_b_int8.numel() == b.shape[0] return b_int8.t(), scale_b_int8 def build_int8_runner(cfg, a, b, dtype, device): # quant before running the kernel b_int8, scale_b_int8 = _quant_weight(b, cfg["w"], device) scale_a_const = None if cfg["a"] == "tensor": scale_a_const = torch.ones(1, device=device, dtype=torch.float32) # no quant, create activation ahead if cfg["no_a_quant"]: if cfg["a"] == "tensor": a_int8, scale_a_int8, _ = vllm_scaled_int8_quant(a, scale_a_const) else: # token a_int8, scale_a_int8, _ = vllm_scaled_int8_quant(a) def run_quant(): return vllm_scaled_mm(a_int8, b_int8, scale_a_int8, scale_b_int8, dtype) return run_quant # dynamic quant, create activation inside if cfg["a"] == "tensor": def run_quant(): a_int8, scale_a_int8, _ = vllm_scaled_int8_quant(a, scale_a_const) return vllm_scaled_mm(a_int8, b_int8, scale_a_int8, scale_b_int8, dtype) else: # token def run_quant(): a_int8, scale_a_int8, _ = vllm_scaled_int8_quant(a) return vllm_scaled_mm(a_int8, b_int8, scale_a_int8, scale_b_int8, dtype) return run_quant _enabled = [k for k, v in PROVIDER_CFGS.items() if v.get("enabled")] @triton.testing.perf_report( triton.testing.Benchmark( x_names=["batch_size"], x_vals=[1, 16, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384], x_log=False, line_arg="provider", line_vals=_enabled, line_names=[k for k in _enabled], ylabel="TFLOP/s (larger is better)", plot_name="BF16 vs INT8 GEMMs", args={}, ) ) def benchmark(batch_size, provider, N, K): M = batch_size device = "cuda" dtype = torch.bfloat16 a = torch.randn((M, K), device=device, dtype=dtype) b = torch.randn((N, K), device=device, dtype=dtype) quantiles = [0.5, 0.2, 0.8] if provider == "torch-bf16": ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: torch.nn.functional.linear(a, b), quantiles=quantiles ) else: cfg = PROVIDER_CFGS[provider] run_quant = build_int8_runner(cfg, a, b, dtype, device) ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: run_quant(), quantiles=quantiles ) to_tflops = lambda t_ms: (2 * M * N * K) * 1e-12 / (t_ms * 1e-3) return to_tflops(ms), to_tflops(max_ms), to_tflops(min_ms) def prepare_shapes(args): KN_model_names = [] for model, tp_size in itertools.product(args.models, args.tp_sizes): for KN, tp_dim in copy.deepcopy(WEIGHT_SHAPES[model]): KN[tp_dim] //= tp_size KN.append(model) KN_model_names.append(KN) return KN_model_names if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--models", nargs="+", type=str, default=["meta-llama/Llama-3.1-8B-Instruct"], choices=list(WEIGHT_SHAPES.keys()), help="List of models to benchmark", ) parser.add_argument( "--tp-sizes", nargs="+", type=int, default=[1], help="List of tensor parallel sizes", ) args = parser.parse_args() for K, N, model in prepare_shapes(args): print(f"{model}, N={N} K={K}, BF16 vs INT8 GEMMs TFLOP/s:") benchmark.run( print_data=True, show_plots=True, save_path=f"bench_int8_res_n{N}_k{K}", N=N, K=K, ) print("Benchmark finished!")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/utils.py
benchmarks/kernels/utils.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import dataclasses from collections.abc import Callable, Iterable from typing import Any import torch import torch.utils.benchmark as TBenchmark from torch.utils.benchmark import Measurement as TMeasurement @dataclasses.dataclass class CudaGraphBenchParams: num_ops_in_cuda_graph: int @dataclasses.dataclass class ArgPool: """ When some argument of the benchmarking function is annotated with this type, the benchmarking class (BenchMM) will collapse the argument to a pick a single value from the given list of values, during function invocation. For every invocation during a benchmarking run, it will choose a different value from the list. """ values: Iterable[Any] def __getitem__(self, index): return self.values[index] class Bench: class ArgsIterator: def __init__(self, args_list, kwargs_list): assert len(args_list) == len(kwargs_list) self.args_list = args_list self.kwargs_list = kwargs_list self.n = len(self.args_list) self.idx = 0 def __next__(self): while True: yield (self.args_list[self.idx], self.kwargs_list[self.idx]) self.idx += 1 self.idx = self.idx % self.n def reset(self): self.idx = 0 @property def n_args(self): return self.n def __init__( self, cuda_graph_params: CudaGraphBenchParams | None, label: str, sub_label: str, description: str, fn: Callable, *args, **kwargs, ): self.cuda_graph_params = cuda_graph_params self.use_cuda_graph = self.cuda_graph_params is not None self.label = label self.sub_label = sub_label self.description = description self.fn = fn # Process args self._args = args self._kwargs = kwargs self.args_list, self.kwargs_list = self.collapse_argpool(*args, **kwargs) self.args_iterator = self.ArgsIterator(self.args_list, self.kwargs_list) # Cudagraph runner self.g = None if self.use_cuda_graph: self.g = self.get_cuda_graph_runner() # benchmark run params self.min_run_time = 1 def collapse_argpool(self, *args, **kwargs): argpool_args = [arg for arg in args if isinstance(arg, ArgPool)] + [ arg for arg in kwargs.values() if isinstance(arg, ArgPool) ] if len(argpool_args) == 0: return [args], [kwargs] # Make sure all argpools are of the same size argpool_size = len(argpool_args[0].values) assert all([argpool_size == len(arg.values) for arg in argpool_args]) # create copies of the args args_list = [] kwargs_list = [] for _ in range(argpool_size): args_list.append(args) kwargs_list.append(kwargs.copy()) for i in range(argpool_size): # collapse args; Just pick the ith value args_list[i] = tuple( [arg[i] if isinstance(arg, ArgPool) else arg for arg in args_list[i]] ) # collapse kwargs kwargs_i = kwargs_list[i] arg_pool_keys = [k for k, v in kwargs_i.items() if isinstance(v, ArgPool)] for k in arg_pool_keys: # again just pick the ith value kwargs_i[k] = kwargs_i[k][i] kwargs_list[i] = kwargs_i return args_list, kwargs_list def get_cuda_graph_runner(self): assert self.use_cuda_graph assert self.args_iterator is not None num_graph_ops = self.cuda_graph_params.num_ops_in_cuda_graph # warmup args_it = self.args_iterator.__next__() for _ in range(2): args, kwargs = next(args_it) self.fn(*args, **kwargs) self.args_iterator.reset() args_it = self.args_iterator.__next__() stream = torch.cuda.Stream() with torch.cuda.stream(stream): g = torch.cuda.CUDAGraph() with torch.cuda.graph(g): for _ in range(num_graph_ops): args, kwargs = next(args_it) self.fn(*args, **kwargs) return g def run_cudagrah(self) -> TMeasurement: assert self.use_cuda_graph globals = {"g": self.g} return TBenchmark.Timer( stmt="g.replay()", globals=globals, label=( f"{self.label}" f" | cugraph {self.cuda_graph_params.num_ops_in_cuda_graph} ops" ), sub_label=self.sub_label, description=self.description, ).blocked_autorange(min_run_time=self.min_run_time) def run_eager(self) -> TMeasurement: setup = None stmt = None globals = None has_arg_pool = self.args_iterator.n_args > 1 if has_arg_pool: setup = """ args_iterator.reset() args_it = args_iterator.__next__() """ stmt = """ args, kwargs = next(args_it) fn(*args, **kwargs) """ globals = {"fn": self.fn, "args_iterator": self.args_iterator} else: # no arg pool. Just use the args and kwargs directly self.args_iterator.reset() args_it = self.args_iterator.__next__() args, kwargs = next(args_it) setup = "" stmt = """ fn(*args, **kwargs) """ globals = {"fn": self.fn, "args": args, "kwargs": kwargs} return TBenchmark.Timer( stmt=stmt, setup=setup, globals=globals, label=self.label, sub_label=self.sub_label, description=self.description, ).blocked_autorange(min_run_time=self.min_run_time) def run(self) -> TMeasurement: timer = None if self.use_cuda_graph: # noqa SIM108 timer = self.run_cudagrah() else: timer = self.run_eager() if not timer.meets_confidence() or timer.has_warnings: print("Doesn't meet confidence - re-running bench ...") return self.run() return timer def __enter__(self): return self def __exit__(self, exc_type, exc_value, traceback): if exc_type: print(f"exc type {exc_type}") print(f"exc value {exc_value}") print(f"exc traceback {traceback}")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_2d_silu_mul_fp8_quant.py
benchmarks/kernels/benchmark_2d_silu_mul_fp8_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from dataclasses import dataclass from enum import Enum from itertools import product from typing import Any import torch import torch.utils.benchmark as TBenchmark from torch.utils.benchmark import Measurement as TMeasurement from vllm.model_executor.layers.quantization.utils.fp8_utils import ( _per_token_group_quant_fp8_colmajor, silu_mul_per_token_group_quant_fp8_colmajor, ) from vllm.triton_utils import triton from vllm.utils.deep_gemm import is_deep_gemm_e8m0_used from .utils import ArgPool, Bench, CudaGraphBenchParams GROUP_SIZE = 128 FLOAT8_T = torch.float8_e4m3fn def print_timers(timers: list[TMeasurement], cuda_graph_nops: int): print( f"Note : The timings reported above is for {cuda_graph_nops} " "consecutive invocations of the benchmarking functions. " f"Please divide by {cuda_graph_nops} for single invocation " "timings." ) compare = TBenchmark.Compare(timers) compare.print() class ImplType(Enum): SILU_MUL_PER_TOKEN_GROUP_QUANT_FP8_COLMAJOR = 1 REFERENCE = 2 def get_impl(self): if self == ImplType.SILU_MUL_PER_TOKEN_GROUP_QUANT_FP8_COLMAJOR: return silu_mul_per_token_group_quant_fp8_colmajor elif self == ImplType.REFERENCE: return reference raise ValueError(f"Unrecognized ImplType {self}") @dataclass class BenchmarkTensors: input: torch.Tensor output: torch.Tensor # Reference act output tensor ref_act_out: torch.Tensor ref_quant_out: torch.Tensor @staticmethod def make(T: int, N: int) -> "BenchmarkTensors": assert T % GROUP_SIZE == 0 assert N % (GROUP_SIZE * 2) == 0 input = torch.rand((T, N), dtype=torch.bfloat16, device="cuda") # silu_mul_per_token_group_quant_fp8_colmajor output. output = torch.rand((T, N // 2), dtype=torch.bfloat16, device="cuda").to( FLOAT8_T ) # reference output. ref_act_out = torch.empty((T, N // 2), dtype=torch.bfloat16, device="cuda") ref_quant_out = torch.empty( (T, N // 2), dtype=torch.bfloat16, device="cuda" ).to(FLOAT8_T) return BenchmarkTensors( input=input, output=output, ref_act_out=ref_act_out, ref_quant_out=ref_quant_out, ) @property def T(self): return self.input.size(0) @property def N(self): return self.input.size(1) def make_impl_kwargs(self, impl_type: ImplType) -> dict[str, Any]: if impl_type == ImplType.SILU_MUL_PER_TOKEN_GROUP_QUANT_FP8_COLMAJOR: return { "input": self.input, "output": self.output, "use_ue8m0": is_deep_gemm_e8m0_used(), } elif impl_type == ImplType.REFERENCE: return { "input": self.input, "act_out": self.ref_act_out, "quant_out": self.ref_quant_out, "use_ue8m0": is_deep_gemm_e8m0_used(), } raise ValueError(f"Unrecognized impl_type {impl_type}") def reference_quant(x: torch.Tensor, quant_out: torch.Tensor, use_ue8m0: bool): """ Reference triton quant kernel from, vllm.model_executor.layers.quantization.utils.fp8_utils """ assert quant_out.size() == x.size() # Allocate the scale tensor column-major format. shape = (x.shape[-1] // GROUP_SIZE,) + x.shape[:-1] x_q = quant_out x_s = torch.empty(shape, device=x.device, dtype=torch.float32).permute(-1, -2) M = x.numel() // GROUP_SIZE N = GROUP_SIZE BLOCK = triton.next_power_of_2(N) # heuristics for number of warps num_warps = min(max(BLOCK // 256, 1), 8) num_stages = 1 finfo = torch.finfo(FLOAT8_T) fp8_min = finfo.min fp8_max = finfo.max _per_token_group_quant_fp8_colmajor[(M,)]( x, x_q, x_s, GROUP_SIZE, x.shape[1], x.stride(0), x_s.stride(1), eps=1e-10, fp8_min=fp8_min, fp8_max=fp8_max, use_ue8m0=use_ue8m0, BLOCK=BLOCK, num_warps=num_warps, num_stages=num_stages, ) return x_q, x_s def reference( input: torch.Tensor, act_out: torch.Tensor, quant_out: torch.Tensor, use_ue8m0: bool, ) -> tuple[torch.Tensor, torch.Tensor]: torch.ops._C.silu_and_mul(act_out, input) return reference_quant(act_out, quant_out, use_ue8m0) def bench_impl( bench_tensors: list[BenchmarkTensors], impl_type: ImplType ) -> TMeasurement: T = bench_tensors[0].T N = bench_tensors[0].N arg_pool_size = len(bench_tensors) kwargs_list = [bt.make_impl_kwargs(impl_type) for bt in bench_tensors] # warmup for kwargs in kwargs_list: impl_type.get_impl()(**kwargs) torch.cuda.synchronize() # Merge into a single kwargs and qualify arguments as ArgPool kwargs = {k: ArgPool([]) for k in kwargs_list[0]} for _kwargs in kwargs_list: for k, v in _kwargs.items(): kwargs[k].values.append(v) cuda_graph_params = None cuda_graph_params = CudaGraphBenchParams(arg_pool_size) timer = None with Bench( cuda_graph_params, "silu-mul-quant", f"num_tokens={T}, N={N}", impl_type.name, impl_type.get_impl(), **kwargs, ) as bench: timer = bench.run() return timer def test_correctness(T: int, N: int): print(f"Testing num_tokens={T}, N={N} ...") bench_tensor = BenchmarkTensors.make(T, N) def output_from_impl(impl: ImplType) -> tuple[torch.Tensor, torch.Tensor]: return impl.get_impl()(**bench_tensor.make_impl_kwargs(impl)) # reference output ref_out_q, ref_out_s = output_from_impl(ImplType.REFERENCE) # test ouptut out_q, out_s = output_from_impl( ImplType.SILU_MUL_PER_TOKEN_GROUP_QUANT_FP8_COLMAJOR ) torch.testing.assert_close(ref_out_q.to(torch.float32), out_q.to(torch.float32)) torch.testing.assert_close(ref_out_s, out_s) def run(Ts: list[int], Ns: list[int], arg_pool_size: int) -> list[TMeasurement]: timers = [] for N, T in product(Ns, Ts): test_correctness(T, N) bench_tensors: list[BenchmarkTensors] = [ BenchmarkTensors.make(T, N) for _ in range(arg_pool_size) ] silu_mul_quant_timer = bench_impl( bench_tensors, ImplType.SILU_MUL_PER_TOKEN_GROUP_QUANT_FP8_COLMAJOR ) timers.append(silu_mul_quant_timer) reference_timer = bench_impl(bench_tensors, ImplType.REFERENCE) timers.append(reference_timer) print_timers( [silu_mul_quant_timer, reference_timer], cuda_graph_nops=arg_pool_size ) print_timers(timers, cuda_graph_nops=arg_pool_size) return timers if __name__ == "__main__": T = [128 * i for i in range(1, 16)] + [2048 * i for i in range(1, 65)] N = [2048, 4096, 8192] print(f"T = {T}, N = {N}") run(T, N, arg_pool_size=8)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_silu_mul_fp8_quant.py
benchmarks/kernels/benchmark_silu_mul_fp8_quant.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Comprehensive 3-way SiLU Benchmark Suite This benchmark compares three SiLU implementations: 1. SiLU V2 (CUDA) - Optimized CUDA kernel implementation 2. Triton Kernel - Triton-based implementation The suite generates detailed performance comparisons including: - Memory bandwidth utilization - Speedup ratios (baseline vs optimized implementations) - Performance across different expert configurations and token distributions """ from collections.abc import Callable import matplotlib.pyplot as plt import numpy as np import torch from vllm.model_executor.layers.fused_moe.batched_deep_gemm_moe import ( persistent_masked_m_silu_mul_quant, ) from vllm.platforms import current_platform from vllm.triton_utils import tl, triton from vllm.utils.deep_gemm import is_deep_gemm_e8m0_used @triton.jit def _silu_mul_fp8_quant_deep_gemm( # Pointers ------------------------------------------------------------ input_ptr, # 16-bit activations (E, T, 2*H) y_q_ptr, # fp8 quantized activations (E, T, H) y_s_ptr, # 16-bit scales (E, T, G) counts_ptr, # int32 num tokens per expert (E) # Sizes --------------------------------------------------------------- H: tl.constexpr, # hidden dimension (per output) GROUP_SIZE: tl.constexpr, # elements per group (usually 128) # Strides for input (elements) --------------------------------------- stride_i_e, stride_i_t, stride_i_h, # Strides for y_q (elements) ----------------------------------------- stride_yq_e, stride_yq_t, stride_yq_h, # Strides for y_s (elements) ----------------------------------------- stride_ys_e, stride_ys_t, stride_ys_g, # Stride for counts (elements) stride_counts_e, # Numeric params ------------------------------------------------------ eps: tl.constexpr, fp8_min: tl.constexpr, fp8_max: tl.constexpr, use_ue8m0: tl.constexpr, # Meta --------------------------------------------------------------- BLOCK: tl.constexpr, NUM_STAGES: tl.constexpr, ): G = H // GROUP_SIZE # map program id -> (e, g) pid = tl.program_id(0) e = pid // G g = pid % G e = e.to(tl.int64) g = g.to(tl.int64) # number of valid tokens for this expert n_tokens = tl.load(counts_ptr + e * stride_counts_e).to(tl.int64) cols = tl.arange(0, BLOCK).to(tl.int64) mask = cols < BLOCK base_input_offset = e * stride_i_e + g * GROUP_SIZE * stride_i_h base_gate_offset = base_input_offset + cols * stride_i_h base_up_offset = base_input_offset + H * stride_i_h + cols * stride_i_h base_yq_offset = e * stride_yq_e + g * GROUP_SIZE * stride_yq_h + cols * stride_yq_h base_ys_offset = e * stride_ys_e + g * stride_ys_g for t in tl.range(0, n_tokens, num_stages=NUM_STAGES): gate = tl.load( input_ptr + base_gate_offset + t * stride_i_t, mask=mask, other=0.0 ).to(tl.float32) up = tl.load(input_ptr + base_up_offset + t * stride_i_t, mask=mask, other=0.0) gate = gate * (1.0 / (1.0 + tl.exp(-gate))) y = gate * up y_s = tl.maximum(tl.max(tl.abs(y)), eps) / fp8_max if use_ue8m0: y_s = tl.exp2(tl.ceil(tl.log2(y_s))) y_q = tl.clamp(y / y_s, fp8_min, fp8_max).to(y_q_ptr.dtype.element_ty) tl.store(y_q_ptr + base_yq_offset + t * stride_yq_t, y_q, mask=mask) tl.store(y_s_ptr + base_ys_offset + t * stride_ys_t, y_s) def silu_mul_fp8_quant_deep_gemm_triton( y: torch.Tensor, # (E, T, 2*H) tokens_per_expert: torch.Tensor, # (E,) number of valid tokens per expert num_parallel_tokens, group_size: int = 128, eps: float = 1e-10, expert_offsets: torch.Tensor = None, ) -> tuple[torch.Tensor, torch.Tensor]: """Quantize silu(y[..., :H]) * y[..., H:] to FP8 with group per-token scales y has shape (E, T, 2*H). The first half of the last dimension is silu-activated, multiplied by the second half, then quantized into FP8. Returns `(y_q, y_s)` where * `y_q`: FP8 tensor, shape (E, T, H), same layout as y[..., :H] * `y_s`: FP32 tensor, shape (E, T, H // group_size), strides (T*G, 1, T) """ assert y.ndim == 3, "y must be (E, T, 2*H)" E, T, H2 = y.shape assert H2 % 2 == 0, "last dim of y must be even (2*H)" H = H2 // 2 G = (H + group_size - 1) // group_size assert H % group_size == 0, "H must be divisible by group_size" assert tokens_per_expert.ndim == 1 and tokens_per_expert.shape[0] == E, ( "tokens_per_expert must be shape (E,)" ) tokens_per_expert = tokens_per_expert.to(device=y.device, dtype=torch.int32) # allocate outputs fp8_dtype = torch.float8_e4m3fn y_q = torch.empty((E, T, H), dtype=fp8_dtype, device=y.device) # strides (elements) stride_i_e, stride_i_t, stride_i_h = y.stride() stride_yq_e, stride_yq_t, stride_yq_h = y_q.stride() # desired scale strides (elements): (T*G, 1, T) stride_ys_e = T * G stride_ys_t = 1 stride_ys_g = T y_s = torch.empty_strided( (E, T, G), (stride_ys_e, stride_ys_t, stride_ys_g), dtype=torch.float32, device=y.device, ) stride_cnt_e = tokens_per_expert.stride()[0] # Static grid over experts and H-groups. # A loop inside the kernel handles the token dim grid = (E * G,) f_info = torch.finfo(fp8_dtype) fp8_max = f_info.max fp8_min = f_info.min _silu_mul_fp8_quant_deep_gemm[grid]( y, y_q, y_s, tokens_per_expert, H, group_size, stride_i_e, stride_i_t, stride_i_h, stride_yq_e, stride_yq_t, stride_yq_h, stride_ys_e, stride_ys_t, stride_ys_g, stride_cnt_e, eps, fp8_min, fp8_max, is_deep_gemm_e8m0_used(), BLOCK=group_size, NUM_STAGES=4, num_warps=1, ) return y_q, y_s # Parse generation strategies strategies = ["random_imbalanced", "uniform", "max_t"] def benchmark( kernel: Callable, E: int, T: int, H: int, total_tokens: int, num_parallel_tokens: int = 64, G: int = 128, runs: int = 200, num_warmups: int = 20, gen_strategy: str = "default", iterations_per_run: int = 20, ): def generate_data(seed_offset=0): """Generate input data with given seed offset""" current_platform.seed_everything(42 + seed_offset) y = torch.rand((E, T, 2 * H), dtype=torch.bfloat16, device="cuda").contiguous() if gen_strategy == "random_imbalanced": def generate_expert_loads(n_e, total_tokens, ratio, device="cuda"): mean = total_tokens // n_e min_max = mean // ratio e = torch.ones(size=(E,), dtype=torch.int64, device=device) * mean e[0] = min_max r = torch.rand(size=(E - 1,)) r /= r.sum() r *= total_tokens - min_max r = r.round().long() e[1:] = r.to(device=device) return e tokens_per_expert = generate_expert_loads(E, total_tokens, 0.7, "cuda") elif gen_strategy == "uniform": r = torch.rand(size=(E,)) r /= r.sum() r *= total_tokens r = r.round().long() tokens_per_expert = r elif gen_strategy == "max_t": tokens_per_expert = torch.empty(size=(E,), dtype=torch.int32, device="cuda") tokens_per_expert.fill_(total_tokens / E) elif gen_strategy == "first_t": tokens_per_expert = torch.zeros(size=(E,), dtype=torch.int32, device="cuda") tokens_per_expert[0] = min(T, total_tokens) else: raise ValueError(f"Unknown generation strategy: {gen_strategy}") return y, tokens_per_expert dataset_count = 4 # Pre-generate different input matrices for each iteration to avoid cache effects data_sets = [generate_data(i) for i in range(dataset_count)] # Warmup y, tokens_per_expert = data_sets[0] for _ in range(num_warmups): kernel( y, tokens_per_expert, num_parallel_tokens=num_parallel_tokens, group_size=G ) torch.cuda.synchronize() start_event = torch.Event(enable_timing=True) end_event = torch.Event(enable_timing=True) # Benchmark latencies: list[float] = [] for _ in range(runs): torch.cuda.synchronize() start_event.record() for i in range(iterations_per_run): y, tokens_per_expert = data_sets[i % dataset_count] kernel( y, tokens_per_expert, num_parallel_tokens=num_parallel_tokens, group_size=G, ) end_event.record() end_event.synchronize() total_time_ms = start_event.elapsed_time(end_event) per_iter_time_ms = total_time_ms / iterations_per_run latencies.append(per_iter_time_ms) # Use median instead of average for better outlier handling median_time_ms = np.median(latencies) median_time_s = median_time_ms / 1000 # Calculate actual work done (using first dataset for consistency) _, tokens_per_expert = data_sets[0] actual_tokens = tokens_per_expert.sum().item() actual_elements = actual_tokens * H # GFLOPS: operations per element = exp + 3 muls + 1 div + quantization ops ≈ 8 ops ops_per_element = 8 total_ops = actual_elements * ops_per_element gflops = total_ops / median_time_s / 1e9 # Memory bandwidth: bfloat16 inputs (2 bytes), fp8 output (1 byte), scales (4 bytes) input_bytes = actual_tokens * 2 * H * 2 # 2*H bfloat16 inputs output_bytes = actual_tokens * H * 1 # H fp8 outputs scale_bytes = actual_tokens * (H // G) * 4 # scales in float32 total_bytes = input_bytes + output_bytes + scale_bytes memory_bw = total_bytes / median_time_s / 1e9 HOPPER_BANDWIDTH_TBPS = 3.35 return ( median_time_ms, gflops, memory_bw, (memory_bw / (HOPPER_BANDWIDTH_TBPS * 1024)) * 100, ) def create_comparison_plot( ratios, silu_v2_times, triton_times, config_labels, strategy_name, id ): fig, ax = plt.subplots(1, 1, figsize=(18, 6)) # Configure x-axis positions x = np.arange(len(config_labels)) width = 0.25 # Execution Time plot (lower is better) ax.bar(x, silu_v2_times, width, label="SiLU V2 (CUDA)", alpha=0.8, color="blue") ax.bar( x + width, triton_times, width, label="Triton Kernel", alpha=0.8, color="green" ) # Add speedup labels over each bar trio for i in range(len(x)): triton_v2_speedup = ratios[i][1] # triton/v2 max_height = max(silu_v2_times[i], triton_times[i]) # Triton/V2 speedup ax.text( x[i] + width / 2, max_height + max_height * 0.02, f"{triton_v2_speedup:.2f}x", ha="center", va="bottom", fontweight="bold", fontsize=8, ) ax.set_xlabel("Configuration") ax.set_ylabel("% Utilization") ax.set_title( f"Memory Bandwidth Utilization (%) - {strategy_name}\n(Higher is Better)" ) ax.set_xticks(x) ax.set_xticklabels(config_labels, rotation=45, ha="right") ax.legend() ax.grid(True, alpha=0.3) plt.tight_layout() return fig, ax def create_combined_plot(all_results): num_strategies = len(all_results) fig, axes = plt.subplots(num_strategies, 1, figsize=(22, 7 * num_strategies)) if num_strategies == 1: axes = [axes] for idx, ( strategy_name, all_ratios, all_silu_v2_results, all_triton_results, config_labels, config_x_axis, ) in enumerate(all_results): ax = axes[idx] # Flatten the nested results to get bandwidth percentages for plotting silu_v2_bandwidths = [] triton_bandwidths = [] flat_ratios = [] for config_results in all_silu_v2_results: for result in config_results: silu_v2_bandwidths.append(result[3]) # bandwidth percentage for config_results in all_triton_results: for result in config_results: triton_bandwidths.append(result[3]) # bandwidth percentage for config_ratios in all_ratios: for ratio in config_ratios: flat_ratios.append(ratio) # Configure x-axis positions x = np.arange(len(config_labels)) width = 0.25 # Bandwidth utilization plot (higher is better) ax.bar( x, silu_v2_bandwidths, width, label="SiLU V2 (CUDA)", alpha=0.8, color="blue", ) ax.bar( x + width, triton_bandwidths, width, label="Triton Kernel", alpha=0.8, color="green", ) # Add speedup labels over each bar trio for i in range(len(x)): triton_v2_speedup = flat_ratios[i] # triton/v2 max_height = max(silu_v2_bandwidths[i], triton_bandwidths[i]) # Triton/V2 speedup ax.text( x[i] + width / 2, max_height + max_height * 0.02, f"{triton_v2_speedup:.2f}x", ha="center", va="bottom", fontweight="bold", fontsize=8, ) ax.set_xlabel("Configuration") ax.set_ylabel("% Utilization") ax.set_title( f"Memory Bandwidth Utilization (%) - {strategy_name}\n(Higher is Better)" ) ax.set_xticks(x) ax.set_xticklabels(config_labels, rotation=45, ha="right") ax.legend() ax.grid(True, alpha=0.3) plt.tight_layout() filename = "silu_benchmark_combined_3way.png" plt.savefig(filename, dpi=300, bbox_inches="tight") plt.show() return filename outer_dim = 7168 configs = [ # DeepSeekV3 Configs # (1, 56, 7168), (8, 1024, 7168), # (32, 56, 7168), # DeepSeekV3 Configs (32, 1024, 7168), # DeepSeekV3 Configs (256, 1024, 7168), ] runs = 100 num_warmups = 20 strategy_descriptions = { "uniform": "Uniform Random", "random_imbalanced": "Imbalanced Random", "max_t": "Even Assignment", "first_t": "experts[0] = T, experts[1:] = 0", } print(f"GPU: {torch.cuda.get_device_name()}") print(f"Testing strategies: {', '.join(strategies)}") print(f"Configurations: {len(configs)} configs") all_results = [] # Run benchmarks for each strategy for id, strategy in enumerate(strategies): print(f"\n{'=' * 60}") print(f"Testing strategy: {strategy_descriptions[strategy]}") print(f"{'=' * 60}") # Collect benchmark data for all three algorithms config_labels = [] config_x_axis = [] all_silu_v2_results = [] all_triton_results = [] all_ratios = [] for E, T, H in configs: total_tokens_config = [] for i in [8, 16, 32, 64, 128, 256, 512]: if i <= T: total_tokens_config.append(i * E) config_x_axis.append(total_tokens_config) silu_v2_results = [] triton_results = [] ratios = [] for total_tokens in total_tokens_config: config_label = f"E={E},T={T},H={H},TT={total_tokens}" config_labels.append(config_label) # SiLU V2 (CUDA kernel) results time_ms_silu_v2, gflops, gbps, perc = benchmark( persistent_masked_m_silu_mul_quant, E, T, H, total_tokens, runs=runs, num_warmups=num_warmups, gen_strategy=strategy, ) silu_v2_results.append((time_ms_silu_v2, gflops, gbps, perc)) # Triton kernel results time_ms_triton, gflops, gbps, perc = benchmark( silu_mul_fp8_quant_deep_gemm_triton, E, T, H, total_tokens, runs=runs, num_warmups=num_warmups, gen_strategy=strategy, ) triton_results.append((time_ms_triton, gflops, gbps, perc)) # Calculate speedup ratios (triton baseline / implementation) triton_v2_ratio = time_ms_triton / time_ms_silu_v2 ratios.append(triton_v2_ratio) print( f"Completed: {config_label}:" f" V2: {time_ms_silu_v2:.3f}ms," f" Triton: {time_ms_triton:.3f}ms" ) all_silu_v2_results.append(silu_v2_results) all_triton_results.append(triton_results) all_ratios.append(ratios) # Store results for combined plotting all_results.append( ( strategy_descriptions[strategy], all_ratios, all_silu_v2_results, all_triton_results, config_labels, config_x_axis, ) ) # Print summary table for this strategy print(f"\nSummary Table - {strategy_descriptions[strategy]}:") print(f" {'V2 Time(ms)':<12} {'Triton Time(ms)':<14} {'Triton/V2':<10}") print("-" * 90) for i, (E, T, H) in enumerate(configs): # Get the first result for each config (simplifying for summary) v2_time = silu_v2_results[i][0] triton_time = triton_results[i][0] triton_v2_speedup = triton_time / v2_time config_label = f"E={E:3d},T={T:4d},H={H:4d}" print( f"{config_label:<20} {v2_time:8.5f} {triton_time:10.5f} " f"{triton_v2_speedup:8.2f}x" ) def create_total_tokens_plot(all_results): num_strategies = len(all_results) num_configs = len(configs) fig, axs = plt.subplots( num_strategies, num_configs * 2, figsize=(32, 8 * num_strategies) ) # Add main title to the entire figure fig.suptitle( "Performance Analysis: Speedup vs Bandwidth Utilization (SiLU V2, and Triton)", fontsize=18, fontweight="bold", y=0.98, ) # Handle single strategy case if num_strategies == 1: axs = axs.reshape(1, -1) # Handle single config case if num_configs == 1: axs = axs.reshape(-1, 2) for strategy_idx, result in enumerate(all_results): ( strategy_name, all_ratios, all_silu_v2_results, all_triton_results, config_labels, config_x_axis, ) = result for config_idx in range(num_configs): # Speedup plot (left column) ax_speedup = axs[strategy_idx, config_idx * 2] # Bandwidth plot (right column) ax_bandwidth = axs[strategy_idx, config_idx * 2 + 1] E, T, H = configs[config_idx] ratios = all_ratios[config_idx] total_tokens_values = config_x_axis[config_idx] # Extract speedup ratios triton_v2_ratios = [ratio for ratio in ratios] # Extract bandwidth percentages for all implementations v2_bandwidth_percentages = [ result[3] for result in all_silu_v2_results[config_idx] ] triton_bandwidth_percentages = [ result[3] for result in all_triton_results[config_idx] ] # Plot speedup ratios vs total tokens (left plot) ax_speedup.plot( total_tokens_values, triton_v2_ratios, "go-", linewidth=3, markersize=8, label="Triton/V2 Speedup", ) ax_speedup.set_title( f"{strategy_name}\nSpeedup vs Baseline (Triton)\nE={E}, T={T}, H={H}", fontsize=12, fontweight="bold", ) ax_speedup.set_xlabel("Total Tokens", fontweight="bold", fontsize=11) ax_speedup.set_ylabel("Speedup Ratio", fontweight="bold", fontsize=11) ax_speedup.legend(prop={"weight": "bold"}) ax_speedup.grid(True, alpha=0.3) # Plot bandwidth utilization (right plot) ax_bandwidth.plot( total_tokens_values, v2_bandwidth_percentages, "o-", linewidth=3, markersize=8, label="SiLU V2", color="blue", ) ax_bandwidth.plot( total_tokens_values, triton_bandwidth_percentages, "o-", linewidth=3, markersize=8, label="Triton", color="green", ) ax_bandwidth.set_title( f"{strategy_name}\nBandwidth Utilization (Hopper)\nE={E}, T={T}, H={H}", fontsize=12, fontweight="bold", ) ax_bandwidth.set_xlabel("Total Tokens", fontweight="bold", fontsize=11) ax_bandwidth.set_ylabel( "% of Peak Bandwidth", fontweight="bold", fontsize=11 ) ax_bandwidth.legend(prop={"weight": "bold"}) ax_bandwidth.grid(True, alpha=0.3) # Format x-axis labels for both plots for ax in [ax_speedup, ax_bandwidth]: ax.set_xticks(total_tokens_values) ax.set_xticklabels( [ f"{tt // 1000}K" if tt >= 1000 else str(tt) for tt in total_tokens_values ], fontweight="bold", ) # Make tick labels bold for label in ax.get_xticklabels() + ax.get_yticklabels(): label.set_fontweight("bold") # Add value labels on Triton/V2 speedup points for x, y in zip(total_tokens_values, triton_v2_ratios): ax_speedup.annotate( f"{y:.2f}x", (x, y), textcoords="offset points", xytext=(0, -15), ha="center", fontsize=9, fontweight="bold", bbox=dict(boxstyle="round,pad=0.2", facecolor="green", alpha=0.3), ) plt.tight_layout() plt.subplots_adjust(top=0.93) # Make room for main title filename = "silu_benchmark_total_tokens_3way.png" plt.savefig(filename, dpi=300, bbox_inches="tight") plt.show() return filename # Create comprehensive 3-way comparison plots combined_plot_filename = create_combined_plot(all_results) total_tokens_plot_filename = create_total_tokens_plot(all_results) print(f"\n{'=' * 80}") print("3-Way Benchmark Suite Complete!") print(f"Generated combined comparison plot: {combined_plot_filename}") print(f"Generated total tokens analysis plot: {total_tokens_plot_filename}") print("Compared: SiLU V2 (CUDA), and Triton implementations") print(f"{'=' * 80}")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_grouped_gemm_cutlass.py
benchmarks/kernels/benchmark_grouped_gemm_cutlass.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import torch import torch.utils.benchmark as benchmark from benchmark_shapes import WEIGHT_SHAPES_MOE from vllm import _custom_ops as ops from vllm.config import ParallelConfig, VllmConfig, set_current_vllm_config from vllm.model_executor.layers.fused_moe.config import fp8_w8a8_moe_quant_config from vllm.model_executor.layers.fused_moe.cutlass_moe import cutlass_moe_fp8 from vllm.model_executor.layers.fused_moe.fused_moe import ( fused_experts, fused_topk, ) from vllm.utils.argparse_utils import FlexibleArgumentParser DEFAULT_MODELS = [ "mistralai/Mixtral-8x7B-Instruct-v0.1", "deepseek-ai/DeepSeek-V2-Lite", "ibm-granite/granite-3.0-1b-a400m", "ibm-granite/granite-3.0-3b-a800m", ] DEFAULT_BATCH_SIZES = [1, 4, 8, 16, 32, 64, 128, 256, 512] DEFAULT_TP_SIZES = [1] PER_ACT_TOKEN_OPTS = [False] PER_OUT_CH_OPTS = [False] def to_fp8(tensor: torch.Tensor): finfo = torch.finfo(torch.float8_e4m3fn) return torch.round(tensor.clamp(min=finfo.min, max=finfo.max)).to( dtype=torch.float8_e4m3fn ) def bench_run( results: list[benchmark.Measurement], model: str, num_experts: int, topk: int, per_act_token: bool, per_out_ch: bool, mkn: tuple[int, int, int], ): label = "Quant Matmul" sub_label = ( "{}, num_experts={}, topk={}, per_act_token={} per_out_ch={}, MKN=({})".format( model, num_experts, topk, per_act_token, per_out_ch, mkn ) ) print(f"Testing: {sub_label}") (m, k, n) = mkn dtype = torch.half a = torch.randn((m, k), device="cuda", dtype=dtype) / 10 w1 = torch.randn((num_experts, 2 * n, k), device="cuda", dtype=dtype) / 10 w2 = torch.randn((num_experts, k, n), device="cuda", dtype=dtype) / 10 _, a_scale = ops.scaled_fp8_quant(a) w1_q = torch.empty( (num_experts, 2 * n, k), device="cuda", dtype=torch.float8_e4m3fn ) w2_q = torch.empty((num_experts, k, n), device="cuda", dtype=torch.float8_e4m3fn) w1_scale = torch.empty((num_experts, 1, 1), device="cuda", dtype=torch.float32) w2_scale = torch.empty((num_experts, 1, 1), device="cuda", dtype=torch.float32) for expert in range(num_experts): w1_q[expert], w1_scale[expert] = ops.scaled_fp8_quant(w1[expert]) w2_q[expert], w2_scale[expert] = ops.scaled_fp8_quant(w2[expert]) score = torch.randn((m, num_experts), device="cuda", dtype=dtype) topk_weights, topk_ids, token_expert_indices = fused_topk( a, score, topk, renormalize=False ) ab_strides1 = torch.full((num_experts,), k, device="cuda", dtype=torch.int64) ab_strides2 = torch.full((num_experts,), n, device="cuda", dtype=torch.int64) c_strides1 = torch.full((num_experts,), 2 * n, device="cuda", dtype=torch.int64) c_strides2 = torch.full((num_experts,), k, device="cuda", dtype=torch.int64) def run_triton_moe( a: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, w1_scale: torch.Tensor, w2_scale: torch.Tensor, a_scale: torch.Tensor, num_repeats: int, ): quant_config = fp8_w8a8_moe_quant_config( w1_scale=w1_scale, w2_scale=w2_scale, a1_scale=a_scale, ) for _ in range(num_repeats): fused_experts( a, w1, w2, topk_weights, topk_ids, quant_config=quant_config, ) def run_cutlass_moe( a: torch.Tensor, a_scale: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, w1_scale: torch.Tensor, w2_scale: torch.Tensor, ab_strides1: torch.Tensor, ab_strides2: torch.Tensor, c_strides1: torch.Tensor, c_strides2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, per_act_token: bool, num_repeats: int, ): quant_config = fp8_w8a8_moe_quant_config( w1_scale=w1_scale, w2_scale=w2_scale, per_act_token_quant=per_act_token, ) for _ in range(num_repeats): cutlass_moe_fp8( a, w1, w2, topk_weights, topk_ids, ab_strides1, ab_strides2, c_strides1, c_strides2, quant_config=quant_config, ) def run_cutlass_from_graph( a: torch.Tensor, a_scale: torch.Tensor, w1_q: torch.Tensor, w2_q: torch.Tensor, w1_scale: torch.Tensor, w2_scale: torch.Tensor, ab_strides1: torch.Tensor, ab_strides2: torch.Tensor, c_strides1: torch.Tensor, c_strides2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, ): quant_config = fp8_w8a8_moe_quant_config( w1_scale=w1_scale, w2_scale=w2_scale, per_act_token_quant=per_act_token, ) with set_current_vllm_config( VllmConfig(parallel_config=ParallelConfig(pipeline_parallel_size=1)) ): return cutlass_moe_fp8( a, w1_q, w2_q, topk_weights, topk_ids, ab_strides1, ab_strides2, c_strides1, c_strides2, quant_config=quant_config, ) def run_triton_from_graph( a: torch.Tensor, w1: torch.Tensor, w2: torch.Tensor, topk_weights: torch.Tensor, topk_ids: torch.Tensor, w1_scale: torch.Tensor, w2_scale: torch.Tensor, a_scale: torch.Tensor, ): quant_config = fp8_w8a8_moe_quant_config( w1_scale=w1_scale, w2_scale=w2_scale, a1_scale=a_scale, ) with set_current_vllm_config( VllmConfig(parallel_config=ParallelConfig(pipeline_parallel_size=1)) ): return fused_experts( a, w1, w2, topk_weights, topk_ids, quant_config=quant_config, ) def replay_graph(graph, num_repeats): for _ in range(num_repeats): graph.replay() torch.cuda.synchronize() cutlass_stream = torch.cuda.Stream() cutlass_graph = torch.cuda.CUDAGraph() with torch.cuda.graph(cutlass_graph, stream=cutlass_stream): run_cutlass_from_graph( a, a_scale, w1_q, w2_q, w1_scale, w2_scale, ab_strides1, ab_strides2, c_strides1, c_strides2, topk_weights, topk_ids, ) torch.cuda.synchronize() triton_stream = torch.cuda.Stream() triton_graph = torch.cuda.CUDAGraph() with torch.cuda.graph(triton_graph, stream=triton_stream): run_triton_from_graph( a, w1_q, w2_q, topk_weights, topk_ids, w1_scale, w2_scale, a_scale, ) torch.cuda.synchronize() min_run_time = 5 num_warmup = 5 num_runs = 25 globals = { # Baseline params "w1": w1, "w2": w2, "score": score, "topk": topk, # Cutlass params "a_scale": a_scale, "w1_q": w1_q, "w2_q": w2_q, "w1_scale": w1_scale, "w2_scale": w2_scale, "per_act_token": per_act_token, "ab_strides1": ab_strides1, "ab_strides2": ab_strides2, "c_strides1": c_strides1, "c_strides2": c_strides2, # cuda graph params "cutlass_graph": cutlass_graph, "triton_graph": triton_graph, # Gen params "a": a, "topk_weights": topk_weights, "topk_ids": topk_ids, "num_runs": num_runs, # Kernels "run_triton_moe": run_triton_moe, "run_cutlass_moe": run_cutlass_moe, "replay_graph": replay_graph, } # Warmup run_triton_moe( a, w1_q, w2_q, topk_weights, topk_ids, w1_scale, w2_scale, a_scale, num_warmup, ) results.append( benchmark.Timer( stmt="run_triton_moe(a, w1_q, w2_q, topk_weights, topk_ids, w1_scale, w2_scale, a_scale, num_runs)", # noqa: E501 globals=globals, label=label, sub_label=sub_label, description="triton_moe", ).blocked_autorange(min_run_time=min_run_time) ) # Warmup replay_graph(triton_graph, num_warmup) results.append( benchmark.Timer( stmt="replay_graph(triton_graph, num_runs)", globals=globals, label=label, sub_label=sub_label, description="triton_moe_cuda_graphs", ).blocked_autorange(min_run_time=min_run_time) ) # Warmup run_cutlass_moe( a, a_scale, w1_q, w2_q, w1_scale, w2_scale, ab_strides1, ab_strides2, c_strides1, c_strides2, topk_weights, topk_ids, per_act_token, num_warmup, ) results.append( benchmark.Timer( stmt="run_cutlass_moe(a, a_scale, w1_q, w2_q, w1_scale, w2_scale, ab_strides1, ab_strides2, c_strides1, c_strides2, topk_weights, topk_ids, per_act_token, num_runs)", # noqa: E501 globals=globals, label=label, sub_label=sub_label, description="grouped_gemm_moe", ).blocked_autorange(min_run_time=min_run_time) ) # Warmup replay_graph(cutlass_graph, num_warmup) results.append( benchmark.Timer( stmt="replay_graph(cutlass_graph, num_runs)", globals=globals, label=label, sub_label=sub_label, description="grouped_gemm_moe_cuda_graphs", ).blocked_autorange(min_run_time=min_run_time) ) def main(args): print("Benchmarking models:") for i, model in enumerate(args.models): print(f"[{i}] {model}") results: list[benchmark.Measurement] = [] for model in args.models: for tp in args.tp_sizes: for layer in WEIGHT_SHAPES_MOE[model]: num_experts = layer[0] topk = layer[1] size_k = layer[2] size_n = layer[3] // tp if len(args.limit_k) > 0 and size_k not in args.limit_k: continue if len(args.limit_n) > 0 and size_n not in args.limit_n: continue for per_act_token in PER_ACT_TOKEN_OPTS: for per_out_ch in PER_OUT_CH_OPTS: for size_m in DEFAULT_BATCH_SIZES: mkn = (size_m, size_k, size_n) bench_run( results, model, num_experts, topk, per_act_token, per_out_ch, mkn, ) compare = benchmark.Compare(results) compare.print() if __name__ == "__main__": parser = FlexibleArgumentParser( description="Benchmark Marlin across specified models/shapes/batches" ) parser.add_argument( "--models", nargs="+", type=str, default=DEFAULT_MODELS, choices=WEIGHT_SHAPES_MOE.keys(), ) parser.add_argument("--tp-sizes", nargs="+", type=int, default=DEFAULT_TP_SIZES) parser.add_argument( "--batch-sizes", nargs="+", type=int, default=DEFAULT_BATCH_SIZES ) parser.add_argument("--limit-k", nargs="+", type=int, default=[]) parser.add_argument("--limit-n", nargs="+", type=int, default=[]) parser.add_argument("--limit-num-groups", nargs="+", type=int, default=[]) parser.add_argument("--limit-per-act-token", nargs="+", type=int, default=[]) parser.add_argument("--limit-per-out-ch", nargs="+", type=int, default=[]) args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/bench_nvfp4_qutlass.py
benchmarks/kernels/bench_nvfp4_qutlass.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # # Copyright (C) 2025 Roberto L. Castro (Roberto.LopezCastro@ist.ac.at). # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import argparse import copy import itertools import torch from compressed_tensors.transform.utils.hadamard import deterministic_hadamard_matrix from weight_shapes import WEIGHT_SHAPES from vllm import _custom_ops as ops # use existing nvfp4 gemm in vllm from vllm._custom_ops import fusedQuantizeNv from vllm.model_executor.layers.quantization.qutlass_utils import to_blocked from vllm.triton_utils import triton PROVIDER_CFGS = { "torch-bf16": dict(enabled=True), "nvfp4": dict(no_a_quant=False, enabled=True), "nvfp4-noquant": dict(no_a_quant=True, enabled=True), } _enabled = [k for k, v in PROVIDER_CFGS.items() if v["enabled"]] def get_hadamard_matrix(group_size: int, dtype: torch.dtype, device: torch.device): return ( deterministic_hadamard_matrix(group_size, dtype=dtype, device=device) * group_size**-0.5 ) def _quant_weight_nvfp4( b: torch.Tensor, forward_hadamard_matrix: torch.Tensor, global_scale: torch.Tensor, device: str, M: int, N: int, K: int, ): weight_hf_e2m1, weight_hf_e8m0 = fusedQuantizeNv( b, forward_hadamard_matrix, global_scale ) weight_hf_scale_block = to_blocked(weight_hf_e8m0, backend="triton").view( -1, K // 16 ) return weight_hf_e2m1, weight_hf_scale_block def build_nvfp4_runner(cfg, a, b, forward_hadamard_matrix, dtype, device, M, N, K): alpha = torch.tensor([1.0], device="cuda") global_scale = torch.tensor([1.0], device="cuda") weight_hf_e2m1, weight_hf_scale_block = _quant_weight_nvfp4( b, forward_hadamard_matrix, global_scale, device, M, N, K ) if cfg["no_a_quant"]: # Pre-quantize activation input_hf_e2m1, input_hf_e8m0 = fusedQuantizeNv( a, forward_hadamard_matrix, global_scale ) input_hf_scale_block = to_blocked(input_hf_e8m0, backend="triton").view( -1, K // 16 ) def run(): return ops.cutlass_scaled_fp4_mm( input_hf_e2m1, weight_hf_e2m1, input_hf_scale_block, weight_hf_scale_block, alpha, torch.bfloat16, ) return run # Quantize activation on-the-fly def run(): input_hf_e2m1, input_hf_e8m0 = fusedQuantizeNv( a, forward_hadamard_matrix, global_scale ) input_hf_scale_block = to_blocked(input_hf_e8m0, backend="triton").view( -1, K // 16 ) return ops.cutlass_scaled_fp4_mm( input_hf_e2m1, weight_hf_e2m1, input_hf_scale_block, weight_hf_scale_block, alpha, torch.bfloat16, ) return run @triton.testing.perf_report( triton.testing.Benchmark( x_names=["batch_size"], x_vals=[ 1, 4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 24576, 32768, ], x_log=False, line_arg="provider", line_vals=_enabled, line_names=_enabled, ylabel="TFLOP/s (larger is better)", plot_name="BF16 vs NVFP4 GEMMs", args={}, ) ) def benchmark(batch_size, provider, N, K, had_size): M = batch_size device = "cuda" dtype = torch.bfloat16 a = torch.randn((M, K), device=device, dtype=dtype) b = torch.randn((N, K), device=device, dtype=dtype) forward_hadamard_matrix = get_hadamard_matrix(had_size, dtype, device) quantiles = [0.5, 0.2, 0.8] if provider == "torch-bf16": ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: torch.nn.functional.linear(a, b), rep=200, quantiles=quantiles ) else: cfg = PROVIDER_CFGS[provider] run_quant = build_nvfp4_runner( cfg, a, b, forward_hadamard_matrix, dtype, device, M, N, K ) ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: run_quant(), rep=200, quantiles=quantiles ) to_tflops = lambda t_ms: (2 * M * N * K) * 1e-12 / (t_ms * 1e-3) return to_tflops(ms), to_tflops(max_ms), to_tflops(min_ms) def prepare_shapes(args): out = [] for model, tp_size in itertools.product(args.models, args.tp_sizes): for KN, tp_dim in copy.deepcopy(WEIGHT_SHAPES[model]): KN[tp_dim] //= tp_size KN.append(model) out.append(KN) return out if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--models", nargs="+", type=str, default=["meta-llama/Llama-3.3-70B-Instruct"], choices=list(WEIGHT_SHAPES.keys()), ) parser.add_argument("--tp-sizes", nargs="+", type=int, default=[1]) args = parser.parse_args() for K, N, model in prepare_shapes(args): for had_size in [16, 32, 64, 128]: print(f"{model}, N={N} K={K}, HAD={had_size}, BF16 vs NVFP4 GEMMs TFLOP/s:") benchmark.run( print_data=True, show_plots=True, save_path=f"bench_nvfp4_res_n{N}_k{K}", N=N, K=K, had_size=had_size, ) print("Benchmark finished!")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_shapes.py
benchmarks/kernels/benchmark_shapes.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project WEIGHT_SHAPES = { "ideal": [[4 * 256 * 32, 256 * 32]], "mistralai/Mistral-7B-v0.1/TP1": [ [4096, 6144], [4096, 4096], [4096, 28672], [14336, 4096], ], "mistralai/Mistral-7B-v0.1/TP2": [ [4096, 3072], [2048, 4096], [4096, 14336], [7168, 4096], ], "mistralai/Mistral-7B-v0.1/TP4": [ [4096, 1536], [1024, 4096], [4096, 7168], [3584, 4096], ], "meta-llama/Llama-2-7b-hf/TP1": [ [4096, 12288], [4096, 4096], [4096, 22016], [11008, 4096], ], "meta-llama/Llama-2-7b-hf/TP2": [ [4096, 6144], [2048, 4096], [4096, 11008], [5504, 4096], ], "meta-llama/Llama-2-7b-hf/TP4": [ [4096, 3072], [1024, 4096], [4096, 5504], [2752, 4096], ], "meta-llama/Llama-2-13b-hf/TP1": [ [5120, 15360], [5120, 5120], [5120, 27648], [13824, 5120], ], "meta-llama/Llama-2-13b-hf/TP2": [ [5120, 7680], [2560, 5120], [5120, 13824], [6912, 5120], ], "meta-llama/Llama-2-13b-hf/TP4": [ [5120, 3840], [1280, 5120], [5120, 6912], [3456, 5120], ], "meta-llama/Llama-2-70b-hf/TP1": [ [8192, 10240], [8192, 8192], [8192, 57344], [28672, 8192], ], "meta-llama/Llama-2-70b-hf/TP2": [ [8192, 5120], [4096, 8192], [8192, 28672], [14336, 8192], ], "meta-llama/Llama-2-70b-hf/TP4": [ [8192, 2560], [2048, 8192], [8192, 14336], [7168, 8192], ], } WEIGHT_SHAPES_MOE = { "mistralai/Mixtral-8x7B-Instruct-v0.1": [ [8, 2, 4096, 28672], [8, 2, 14336, 4096], ], "deepseek-ai/DeepSeek-V2-Lite": [ [64, 6, 2048, 1408], ], "ibm-granite/granite-3.0-1b-a400m": [ [32, 8, 1024, 1024], ], "ibm-granite/granite-3.0-3b-a800m": [ [40, 8, 1024, 1536], ], }
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/weight_shapes.py
benchmarks/kernels/weight_shapes.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Weight Shapes are in the format # ([K, N], TP_SPLIT_DIM) # Example: # A shape of ([14336, 4096], 0) indicates the following GEMM shape, # - TP1 : K = 14336, N = 4096 # - TP2 : K = 7168, N = 4096 # A shape of ([4096, 6144], 1) indicates the following GEMM shape, # - TP1 : K = 4096, N = 6144 # - TP4 : K = 4096, N = 1536 # TP1 shapes WEIGHT_SHAPES = { "mistralai/Mistral-7B-v0.1": [ ([4096, 6144], 1), ([4096, 4096], 0), ([4096, 28672], 1), ([14336, 4096], 0), ], "meta-llama/Llama-2-7b-hf": [ ([4096, 12288], 1), ([4096, 4096], 0), ([4096, 22016], 1), ([11008, 4096], 0), ], "meta-llama/Llama-3-8b": [ ([4096, 6144], 1), ([4096, 4096], 0), ([4096, 28672], 1), ([14336, 4096], 0), ], "meta-llama/Llama-2-13b-hf": [ ([5120, 15360], 1), ([5120, 5120], 0), ([5120, 27648], 1), ([13824, 5120], 0), ], "meta-llama/Llama-2-70b-hf": [ ([8192, 10240], 1), ([8192, 8192], 0), ([8192, 57344], 1), ([28672, 8192], 0), ], "meta-llama/Llama-3.1-405b-hf": [ ([16384, 18432], 1), ([16384, 16384], 0), ([16384, 106496], 1), ([53248, 16384], 0), ], "meta-llama/Llama-3.1-8B-Instruct": [ ([4096, 6144], 1), ([4096, 4096], 0), ([4096, 28672], 1), ([14336, 4096], 0), ], "meta-llama/Llama-3.3-70B-Instruct": [ ([8192, 10240], 1), ([8192, 8192], 0), ([8192, 57344], 1), ([28672, 8192], 0), ], "mistralai/Mistral-Large-Instruct-2407": [ ([12288, 14336], 1), ([12288, 12288], 0), ([12288, 57344], 1), ([28672, 12288], 0), ], "Qwen/Qwen2.5-7B-Instruct": [ ([3584, 4608], 1), ([3584, 3584], 0), ([3584, 37888], 1), ([18944, 3584], 0), ], "Qwen/Qwen2.5-32B-Instruct": [ ([5120, 7168], 1), ([5120, 5120], 0), ([5120, 55296], 1), ([27648, 5120], 0), ], "Qwen/Qwen2.5-72B-Instruct": [ ([8192, 10240], 1), ([8192, 8192], 0), ([8192, 59136], 1), ([29568, 8192], 0), ], "deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct": [ ([2048, 3072], 1), ([2048, 4096], 1), ([2048, 2048], 0), ([2048, 576], 0), ([2048, 21888], 1), ([10944, 2048], 0), ([2048, 2816], 1), ([1408, 2048], 0), ], "CohereLabs/c4ai-command-a-03-2025": [ ([12288, 14336], 1), ([12288, 12288], 0), ([12288, 73728], 1), ([36864, 12288], 0), ], }
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/bench_block_fp8_gemm.py
benchmarks/kernels/bench_block_fp8_gemm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import os # Disable DeepGEMM for this benchmark to use CUTLASS os.environ["VLLM_USE_DEEP_GEMM"] = "0" import torch from vllm.model_executor.layers.quantization.utils.fp8_utils import ( W8A8BlockFp8LinearOp, ) from vllm.model_executor.layers.quantization.utils.quant_utils import ( GroupShape, ) from vllm.model_executor.layers.quantization.utils.w8a8_utils import ( CUTLASS_BLOCK_FP8_SUPPORTED, ) from vllm.platforms import current_platform from vllm.triton_utils import triton as vllm_triton assert current_platform.is_cuda(), ( "Only support benchmarking w8a8 block fp8 kernel on CUDA device." ) # DeepSeek-V3 weight shapes DEEPSEEK_V3_SHAPES = [ (512 + 64, 7168), (2112, 7168), ((128 + 64) * 128, 7168), (128 * (128 + 128), 512), (7168, 16384), (7168, 18432), (18432 * 2, 7168), (24576, 1536), (12288, 7168), (4096, 7168), (7168, 2048), ] def build_w8a8_block_fp8_runner(M, N, K, block_size, device, use_cutlass): """Build runner function for w8a8 block fp8 matmul.""" factor_for_scale = 1e-2 fp8_info = torch.finfo(torch.float8_e4m3fn) fp8_max, fp8_min = fp8_info.max, fp8_info.min # Create random input tensor (bfloat16, will be quantized by W8A8BlockFp8LinearOp) A_ref = (torch.rand(M, K, dtype=torch.bfloat16, device=device) - 0.5) * 2 * fp8_max # Create quantized weight tensor B_ref = (torch.rand(N, K, dtype=torch.bfloat16, device=device) - 0.5) * 2 * fp8_max B = B_ref.clamp(min=fp8_min, max=fp8_max).to(torch.float8_e4m3fn) # Create weight scales block_n, block_k = block_size[0], block_size[1] n_tiles = (N + block_n - 1) // block_n k_tiles = (K + block_k - 1) // block_k Bs = ( torch.rand(n_tiles, k_tiles, dtype=torch.float32, device=device) * factor_for_scale ) # Create W8A8BlockFp8LinearOp instance weight_group_shape = GroupShape(block_n, block_k) act_quant_group_shape = GroupShape(1, block_k) # Per-token, per-group quantization linear_op = W8A8BlockFp8LinearOp( weight_group_shape=weight_group_shape, act_quant_group_shape=act_quant_group_shape, cutlass_block_fp8_supported=use_cutlass, use_aiter_and_is_supported=False, ) def run(): return linear_op.apply( input=A_ref, weight=B, weight_scale=Bs, input_scale=None, bias=None, ) return run # Determine available providers available_providers = ["torch-bf16", "w8a8-block-fp8-triton"] plot_title = "BF16 vs W8A8 Block FP8 GEMMs" if CUTLASS_BLOCK_FP8_SUPPORTED: available_providers.append("w8a8-block-fp8-cutlass") @vllm_triton.testing.perf_report( vllm_triton.testing.Benchmark( x_names=["batch_size"], x_vals=[1, 16, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384], x_log=False, line_arg="provider", line_vals=available_providers, line_names=available_providers, ylabel="TFLOP/s (larger is better)", plot_name="BF16 vs W8A8 Block FP8 GEMMs", args={}, ) ) def benchmark_tflops(batch_size, provider, N, K, block_size=(128, 128)): M = batch_size device = "cuda" quantiles = [0.5, 0.2, 0.8] if provider == "torch-bf16": a = torch.randn((M, K), device=device, dtype=torch.bfloat16) b = torch.randn((N, K), device=device, dtype=torch.bfloat16) ms, min_ms, max_ms = vllm_triton.testing.do_bench_cudagraph( lambda: torch.nn.functional.linear(a, b), quantiles=quantiles ) elif provider == "w8a8-block-fp8-triton": run_w8a8_triton = build_w8a8_block_fp8_runner( M, N, K, block_size, device, use_cutlass=False ) ms, min_ms, max_ms = vllm_triton.testing.do_bench_cudagraph( lambda: run_w8a8_triton(), quantiles=quantiles ) elif provider == "w8a8-block-fp8-cutlass": run_w8a8_cutlass = build_w8a8_block_fp8_runner( M, N, K, block_size, device, use_cutlass=True ) ms, min_ms, max_ms = vllm_triton.testing.do_bench_cudagraph( lambda: run_w8a8_cutlass(), quantiles=quantiles ) else: raise ValueError(f"Unknown provider: {provider}") to_tflops = lambda t_ms: (2 * M * N * K) * 1e-12 / (t_ms * 1e-3) return to_tflops(ms), to_tflops(max_ms), to_tflops(min_ms) if __name__ == "__main__": block_size = (128, 128) for N, K in DEEPSEEK_V3_SHAPES: print(f"\nBenchmarking DeepSeek-V3, N={N} K={K}") print(f"TFLOP/s comparison (block_size={block_size}):") benchmark_tflops.run( print_data=True, # show_plots=False, # save_path=f"bench_w8a8_block_fp8_tflops_n{N}_k{K}", N=N, K=K, block_size=block_size, ) print("\nBenchmark finished!")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_lora.py
benchmarks/kernels/benchmark_lora.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import copy import json import pickle import time from collections.abc import Callable from dataclasses import dataclass from enum import Enum, auto from itertools import product from pathlib import Path from typing import Any import torch import torch.utils.benchmark as TBenchmark from torch.utils.benchmark import Measurement as TMeasurement from utils import ArgPool, Bench, CudaGraphBenchParams from weight_shapes import WEIGHT_SHAPES from vllm.lora.ops.triton_ops.utils import get_lora_op_configs from vllm.triton_utils import HAS_TRITON, triton if HAS_TRITON: from vllm.lora.ops.triton_ops import ( ## added fused_moe_lora LoRAKernelMeta, fused_moe_lora_expand, fused_moe_lora_shrink, lora_expand, lora_shrink, ) from vllm.lora.ops.triton_ops.fused_moe_lora_op import ( _LORA_PTR_DICT, ## added _LORA_PTR_DICT for fused_moe_lora ) from vllm.lora.ops.triton_ops.utils import _LORA_A_PTR_DICT, _LORA_B_PTR_DICT from vllm import _custom_ops as ops from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.math_utils import round_up DEFAULT_MODELS = list(WEIGHT_SHAPES.keys()) DEFAULT_TP_SIZES = [1] DEFAULT_BATCH_SIZES = [ 1, 16, 32, 64, 128, 192, 256, 320, 384, 448, 512, 640, 768, 896, 1024, 2048, 3072, 4096, 5120, 6144, 7168, 8192, ] DEFAULT_HIDDEN_SIZES = [1024, 2048, 4096, 8192, 16384] DEFAULT_LORA_RANKS = [16] DEFAULT_NUM_LORAS = [1, 2, 3, 4] DEFAULT_SORT_BY_LORA_IDS = [False, True] DEFAULT_SEQ_LENGTHS = [1] DEFAULT_EXPAND_FN_ADD_INPUTS = [True, False] DEFAULT_TOP_K_NUMS = [1] # Added for MoE LoRA top_k DEFAULT_NUM_EXPERTS = [8] # Added for MoE LoRA num_experts # Utilities def dtype_to_str(dtype: torch.dtype): if dtype == torch.float16: return "f16" if dtype == torch.bfloat16: return "bf16" if dtype == torch.float32: return "f32" raise ValueError(f"Unsupported dtype {dtype}") def make_rand_lora_weight_tensor( k: int, n: int, num_loras: int, dtype: torch.dtype, device: str = "cuda" ) -> torch.Tensor: # LoRA weights column major return torch.rand((num_loras, n, k), dtype=dtype).to(device) def make_rand_tensors( a_shape: tuple[int, ...], b_shape: tuple[int, ...], c_shape: tuple[int, ...], a_dtype: torch.dtype, b_dtype: torch.dtype, c_dtype: torch.dtype, num_slices: int, device: str = "cuda", ) -> tuple[torch.Tensor, list[torch.Tensor], torch.Tensor]: """ Make LoRA input/output matrices. """ A = torch.rand(a_shape, dtype=a_dtype).to(device) # LoRA weights column major Bs = [torch.rand(b_shape, dtype=b_dtype).to(device) for _ in range(num_slices)] C = torch.zeros(c_shape, dtype=c_dtype).to(device) return A, Bs, C def make_prompt_lora_mapping( num_prompts: int, num_active_loras: int, sort_by_lora_id: bool, device: str ) -> torch.Tensor: """ All prompts are mapped to a LoRA ID in range [0, num_active_loras). where 0 refers to first lora, 1 refers to second lora and so on. """ assert num_active_loras > 0 if not sort_by_lora_id: return torch.randint(0, num_active_loras, (num_prompts,), dtype=torch.long) # Divide LoRAs equally and in order. part_size = num_prompts // num_active_loras part_size = max(part_size, 1) lora_id = 0 prompt_lora_mapping = [] while len(prompt_lora_mapping) < num_prompts: prompt_lora_mapping.extend([lora_id] * part_size) lora_id = lora_id + 1 if lora_id + 1 < num_active_loras else lora_id return torch.tensor( prompt_lora_mapping[:num_prompts], dtype=torch.long, device=device ) def make_token_lora_mapping( num_tokens: int, num_prompts: int, prompt_lora_mapping: torch.Tensor, seq_len_tensor: torch.Tensor, device: str, ): """ Make token_lora_mapping from prompt_lora_mapping and seq_lens_tensor """ assert prompt_lora_mapping.shape[0] == num_prompts # token to lora index mapping token_lora_mapping = [0] * num_tokens current_offset = 0 for b_id in range(num_prompts): lora_index = prompt_lora_mapping[b_id].item() s = current_offset e = s + seq_len_tensor[b_id].item() token_lora_mapping[s:e] = [lora_index] * (e - s) current_offset += seq_len_tensor[b_id].item() return torch.tensor(token_lora_mapping, dtype=torch.long, device=device) def ref_group_gemm( ref_out: torch.Tensor, input: torch.Tensor, lora_weights: list[torch.Tensor], seq_lens_cpu: torch.Tensor, prompt_lora_mapping_cpu: torch.Tensor, scaling: float, add_inputs: bool | None, ): """ Torch group gemm reference implementation to test correctness of benchmarking operations. """ batches = seq_lens_cpu.size(0) out_list = [] current_offset = 0 for lora_index, b_length in zip(range(batches), seq_lens_cpu): x = input[current_offset : b_length + current_offset, :] current_offset += b_length w = lora_weights[prompt_lora_mapping_cpu[lora_index]] result = torch.nn.functional.linear(x, w) result *= scaling out_list.append(result) cat_result = torch.cat(out_list, dim=0) if add_inputs: ref_out += cat_result else: ref_out.copy_(cat_result) class OpType(Enum): """ LoRA Ops to benchmark and its properties. """ LORA_SHRINK = auto() LORA_EXPAND = auto() ## Adding support for fused moe lora FUSED_MOE_LORA_GATE_UP_SHRINK = auto() ## Gate/Up projection variant with shrink FUSED_MOE_LORA_GATE_UP_EXPAND = auto() ## Gate/Up projection variant with expand FUSED_MOE_LORA_DOWN_SHRINK = auto() ## Down projection variant with shrink FUSED_MOE_LORA_DOWN_EXPAND = auto() ## Down projection variant with expand @staticmethod def from_str(s: str) -> "OpType": if s.lower() == "lora_shrink": return OpType.LORA_SHRINK if s.lower() == "lora_expand": return OpType.LORA_EXPAND # Adding support for fused moe lora, both in gate_up and down if s.lower() == "fused_moe_lora_gate_up_shrink": ## Gate/Up variant with shrink return OpType.FUSED_MOE_LORA_GATE_UP_SHRINK if s.lower() == "fused_moe_lora_gate_up_expand": ## Gate/Up variant with expand return OpType.FUSED_MOE_LORA_GATE_UP_EXPAND if s.lower() == "fused_moe_lora_down_shrink": ## Down variant with shrink return OpType.FUSED_MOE_LORA_DOWN_SHRINK if s.lower() == "fused_moe_lora_down_expand": ## Down variant with expand return OpType.FUSED_MOE_LORA_DOWN_EXPAND raise ValueError(f"Unrecognized str {s} to convert to OpType") def is_shrink_fn(self) -> bool: return self in [OpType.LORA_SHRINK] def is_expand_fn(self) -> bool: return self in [OpType.LORA_EXPAND] def is_fused_moe_lora_fn(self) -> bool: ## adding for fused MoE LoRA return self in [ OpType.FUSED_MOE_LORA_GATE_UP_SHRINK, OpType.FUSED_MOE_LORA_DOWN_SHRINK, OpType.FUSED_MOE_LORA_GATE_UP_EXPAND, OpType.FUSED_MOE_LORA_DOWN_EXPAND, ] def is_fused_moe_lora_gate_up_fn( self, ) -> bool: ## adding for fused MoE LoRA Gate/Up return self in [ OpType.FUSED_MOE_LORA_GATE_UP_SHRINK, OpType.FUSED_MOE_LORA_GATE_UP_EXPAND, ] def is_fused_moe_lora_down_fn(self) -> bool: ## adding for fused MoE LoRA Down return self in [ OpType.FUSED_MOE_LORA_DOWN_SHRINK, OpType.FUSED_MOE_LORA_DOWN_EXPAND, ] def is_fused_moe_lora_shrink_fn(self) -> bool: return self in [ OpType.FUSED_MOE_LORA_GATE_UP_SHRINK, OpType.FUSED_MOE_LORA_DOWN_SHRINK, ] def is_fused_moe_lora_expand_fn(self) -> bool: return self in [ OpType.FUSED_MOE_LORA_GATE_UP_EXPAND, OpType.FUSED_MOE_LORA_DOWN_EXPAND, ] def num_slices(self) -> list[int]: if self.is_fused_moe_lora_gate_up_fn(): return [2] elif self.is_fused_moe_lora_down_fn(): return [1] return [1, 2, 3] def mkn( self, batch_size: int, seq_length: int, hidden_size: int, lora_rank: int ) -> tuple[int, int, int]: num_tokens = batch_size * seq_length if self.is_shrink_fn() or self.is_fused_moe_lora_fn(): m = num_tokens k = hidden_size n = lora_rank elif self.is_expand_fn(): m = num_tokens k = lora_rank n = hidden_size return m, k, n def matmul_dtypes( self, op_dtype: torch.dtype ) -> tuple[torch.dtype, torch.dtype, torch.dtype]: """ return a type, b type and c type for A x B = C """ if self.is_shrink_fn(): return op_dtype, op_dtype, torch.float32 elif self.is_expand_fn(): return torch.float32, op_dtype, op_dtype else: assert self.is_fused_moe_lora_fn() return op_dtype, op_dtype, op_dtype def matmul_shapes_fused_moe_lora( self, m: int, n: int, k: int, num_loras: int, num_slices: int, top_k_num: int, num_experts: int, ) -> tuple[tuple[int], tuple[int], tuple[int], tuple[int]]: if self.is_fused_moe_lora_shrink_fn(): input_shape = ( (m * top_k_num, n) if self in [OpType.FUSED_MOE_LORA_DOWN_SHRINK] else (m, n) ) output_shape = (num_slices, m, top_k_num, k) weight_shape = (num_loras, num_experts, k, n) else: assert self.is_fused_moe_lora_expand_fn() input_shape = (num_slices, m, top_k_num, k) output_shape = (m, top_k_num, n * num_slices) weight_shape = (num_loras, num_experts, n, k) return (input_shape, weight_shape, output_shape) def matmul_shapes( self, batch_size: int, seq_length: int, hidden_size: int, lora_rank: int, num_loras: int, num_slices: int, top_k_num: int | None = None, num_experts: int | None = None, ) -> tuple[tuple[int, ...], tuple[int, ...], tuple[int, ...]]: """ Given num_slices, return the shapes of the A, B, and C matrices in A x B = C, for the op_type """ m, k, n = self.mkn(batch_size, seq_length, hidden_size, lora_rank) b_shape = (num_loras, n, k) # col-major if self in [OpType.LORA_SHRINK]: # LoRA shrink kernels support num_slices inherently in the kernel. return ((m, k), b_shape, (num_slices, m, n)) if self in [OpType.LORA_EXPAND]: # LoRA expand kernels support num_slices inherently in the kernel return ((num_slices, m, k), b_shape, (m, n * num_slices)) if self.is_fused_moe_lora_fn(): return self.matmul_shapes_fused_moe_lora( m, k, n, num_loras, num_slices, top_k_num, num_experts, ) raise ValueError(f"Unrecognized op_type {self}") def bench_fn(self) -> Callable: if self == OpType.LORA_SHRINK: return lora_shrink if self == OpType.LORA_EXPAND: return lora_expand if self in [ OpType.FUSED_MOE_LORA_GATE_UP_SHRINK, OpType.FUSED_MOE_LORA_DOWN_SHRINK, ]: return fused_moe_lora_shrink if self in [ OpType.FUSED_MOE_LORA_GATE_UP_EXPAND, OpType.FUSED_MOE_LORA_DOWN_EXPAND, ]: return fused_moe_lora_expand raise ValueError(f"Unrecognized optype {self}") def run_ref_group_gemm( self, output: torch.Tensor, input: torch.Tensor, lora_weights: list[torch.Tensor], **kwargs, ) -> Callable: """Each benchmark operation expects the input, lora_weights and outputs in a slightly different format. Refer to self.matmul_shapes(). run_ref_group_gemm accounts for those differences in executing a reference group gemm for correctness testing. """ w_dtype = lora_weights[0].dtype num_slices = len(lora_weights) if self in [OpType.LORA_SHRINK]: for slice_idx in range(num_slices): ref_group_gemm( ref_out=output[slice_idx, :], input=input, lora_weights=lora_weights[slice_idx], **kwargs, ) elif self in [OpType.LORA_EXPAND]: hidden_size = lora_weights[0].shape[1] for slice_idx in range(num_slices): slice_offset = slice_idx * hidden_size ref_group_gemm( ref_out=output[:, slice_offset : slice_offset + hidden_size], input=input[slice_idx].clone().to(dtype=w_dtype), lora_weights=lora_weights[slice_idx], **kwargs, ) else: raise ValueError(f"Unrecognized optype {self}") @dataclass class BenchmarkContext: """ LoRA benchmark context """ batch_size: int hidden_size: int num_loras: int num_active_loras: int lora_rank: int sort_by_lora_id: bool dtype: torch.dtype seq_length: int | None = None num_experts: int | None = None # num_experts for MoE based ops top_k_num: int | None = None # top_k for MoE based ops num_slices: int | None = None # num_slices for slice based ops def with_seq_length(self, seq_length: int) -> "BenchmarkContext": ctx = copy.copy(self) ctx.seq_length = seq_length return ctx def with_num_slices(self, num_slices: int) -> "BenchmarkContext": ctx = copy.copy(self) ctx.num_slices = num_slices return ctx def bench_label(self) -> str: return f"lora-{self.dtype}" def bench_sublabel(self, op_type: OpType) -> str: m, k, n = op_type.mkn( self.batch_size, self.seq_length, self.hidden_size, self.lora_rank ) desc = { "bs": self.batch_size, "sl": self.seq_length, "m": m, "k": k, "n": n, "num_loras": self.num_loras, "sort_by_lora": self.sort_by_lora_id, "num_slices": self.num_slices, } return json.dumps(desc) @dataclass class BenchmarkTensors: """ Input/Output tensors used for benchmarks """ # matmul tensors input: torch.Tensor lora_weights_lst: list[torch.Tensor] output: torch.Tensor # LoRA kernel metadata lora_kernel_meta: LoRAKernelMeta # Metadata tensors used in testing correctness seq_lens: torch.Tensor prompt_lora_mapping: torch.Tensor def io_types(self) -> str: return ( f"{dtype_to_str(self.input.dtype)}x" f"{dtype_to_str(self.lora_weights_lst[0].dtype)}=>" f"{dtype_to_str(self.output.dtype)}" ) def get_num_tokens(self, size: int, top_k_num: int, op_type: OpType): return ( size * top_k_num if op_type in [OpType.FUSED_MOE_LORA_DOWN_SHRINK] else size ) @staticmethod def make( ctx: BenchmarkContext, op_type: OpType, device: str = "cuda" ) -> "BenchmarkTensors": # Make input / output matmul tensors. a_shape, b_shape, c_shape = op_type.matmul_shapes( ctx.batch_size, ctx.seq_length, ctx.hidden_size, ctx.lora_rank, ctx.num_loras, ctx.num_slices, ctx.top_k_num, ctx.num_experts, ) a_type, b_type, c_type = op_type.matmul_dtypes(ctx.dtype) input_tensor, lora_weights, output_tensor = make_rand_tensors( a_shape, b_shape, c_shape, a_type, b_type, c_type, num_slices=ctx.num_slices ) # Make metadata tensors. # Keep the metadata tensors in the CPU for further processing if needed. # The tensors get moved to the GPU before benchmarking. assert ctx.num_active_loras <= ctx.num_loras total_tokens = ctx.batch_size * ctx.seq_length # Make metadata tensors involved in correctness testing. # Prepare seq lens tensor seq_len_tensor = torch.randint( ctx.seq_length, ctx.seq_length + 1, (ctx.batch_size,) ) assert total_tokens == seq_len_tensor.sum() # Prepare prompt lora indices tensor prompt_lora_indices_tensor = make_prompt_lora_mapping( ctx.batch_size, ctx.num_active_loras, ctx.sort_by_lora_id, "cpu" ) # Make LoRAKernelMeta token_lora_indices_tensor = make_token_lora_mapping( total_tokens, ctx.batch_size, prompt_lora_indices_tensor, seq_len_tensor, "cpu", ) lora_kernel_meta = LoRAKernelMeta.make( max_loras=ctx.num_loras, max_num_tokens=token_lora_indices_tensor.size(0), device="cpu", ) lora_kernel_meta.prepare_tensors(token_lora_mapping=token_lora_indices_tensor) return BenchmarkTensors( input_tensor, lora_weights, output_tensor, lora_kernel_meta, seq_len_tensor, prompt_lora_indices_tensor, ) def sanity_check(self, ctx: BenchmarkContext, op_type: OpType) -> None: """ Fails asserts when non-conformality is detected. """ num_tokens = ( self.input.shape[1] if op_type.is_fused_moe_lora_expand_fn() else self.input.shape[-2] ) # check metadata tensors ## In down shrink case, each token is repeated top_k_num times assert num_tokens == self.get_num_tokens( torch.sum(self.seq_lens), ctx.top_k_num, op_type ), f"Expected {num_tokens} tokens, but got {torch.sum(self.seq_lens)}" num_seqs = self.seq_lens.shape[0] # assert self.seq_start_loc.shape[0] == num_seqs ## In down shrink case, each prompt corresponds to top_k_num sequences assert self.prompt_lora_mapping.shape[0] == num_seqs assert self.get_num_tokens( self.lora_kernel_meta.token_lora_mapping.shape[0], ctx.top_k_num, op_type ) def to_device(self, device: str): """ Transfer tensors to device if the tensors aren't already on the device """ def to_device(tensor: torch.Tensor): if tensor.device != device: tensor = tensor.to(device=device) return tensor self.input = to_device(self.input) self.output = to_device(self.output) self.seq_lens = to_device(self.seq_lens) self.prompt_lora_mapping = to_device(self.prompt_lora_mapping) for i in range(len(self.lora_weights_lst)): self.lora_weights_lst[i] = to_device(self.lora_weights_lst[i]) # LoRA meta for field_name in LoRAKernelMeta.__dataclass_fields__: field = getattr(self.lora_kernel_meta, field_name) assert isinstance(field, torch.Tensor) setattr( self.lora_kernel_meta, field_name, to_device(field) if field_name != "no_lora_flag_cpu" else field, ) def metadata(self, ctx: BenchmarkContext, op_type: OpType) -> tuple[int, int, int]: """ Return num_seqs, num_tokens and max_seq_len """ num_seqs = self.seq_lens.shape[0] num_tokens = self.get_num_tokens( self.lora_kernel_meta.token_lora_mapping.shape[0], ctx.top_k_num, op_type ) max_seq_len = torch.max(self.seq_lens).item() num_slices = len(self.lora_weights_lst) return num_seqs, num_tokens, max_seq_len, num_slices def fused_moe_lora_data_prepare( self, block_size: int, token_lora_mapping: torch.Tensor, ctx: BenchmarkContext, ): def moe_lora_align_block_size( topk_ids: torch.Tensor, token_lora_mapping: torch.Tensor, block_size: int, num_experts: int, max_loras: int, expert_map: torch.Tensor | None = None, pad_sorted_ids: bool = False, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Aligns tokens and experts into block-sized chunks for LoRA-based mixture-of-experts (MoE) execution. """ max_num_tokens_padded = topk_ids.numel() + num_experts * (block_size - 1) if pad_sorted_ids: max_num_tokens_padded = round_up(max_num_tokens_padded, block_size) sorted_ids = torch.empty( (max_loras * max_num_tokens_padded,), dtype=torch.int32, device=topk_ids.device, ) max_num_m_blocks = triton.cdiv(max_num_tokens_padded, block_size) # Expert ids must be set default to -1 to prevent a blank block expert_ids = torch.empty( (max_loras * max_num_m_blocks,), dtype=torch.int32, device=topk_ids.device, ) num_tokens_post_pad = torch.empty( (max_loras), dtype=torch.int32, device=topk_ids.device ) ops.moe_lora_align_block_size( topk_ids, token_lora_mapping, num_experts, block_size, max_loras, max_num_tokens_padded, max_num_m_blocks, sorted_ids, expert_ids, num_tokens_post_pad, ) if expert_map is not None: expert_ids = expert_map[expert_ids] return sorted_ids, expert_ids, num_tokens_post_pad num_tokens = ctx.batch_size curr_topk_ids = torch.randint( 0, ctx.num_experts, (num_tokens, ctx.top_k_num), device="cuda", dtype=torch.int32, ) topk_weights = torch.randint( 0, ctx.num_experts, (num_tokens, ctx.top_k_num), device="cuda", dtype=torch.int32, ) (sorted_token_ids_lora, expert_ids_lora, num_tokens_post_padded_lora) = ( moe_lora_align_block_size( topk_ids=curr_topk_ids, token_lora_mapping=token_lora_mapping, block_size=block_size, num_experts=ctx.num_experts, max_loras=ctx.num_loras, ) ) sorted_token_ids = sorted_token_ids_lora.view(ctx.num_loras, -1) expert_ids = expert_ids_lora.view(ctx.num_loras, -1) num_tokens_post_padded = num_tokens_post_padded_lora return (topk_weights, sorted_token_ids, expert_ids, num_tokens_post_padded) def as_lora_shrink_kwargs( self, ctx: BenchmarkContext, op_type: OpType ) -> dict[str, Any]: self.sanity_check(ctx, op_type) self.to_device(self.input.device) _, num_tokens, _, num_slices = self.metadata(ctx, op_type) # Sanity check matrix shapes. i_shape, lw_shape, o_shape = ( self.input.shape, self.lora_weights_lst[0].shape, self.output.shape, ) # Expected input shape [num_tokens, hidden_size] assert len(i_shape) == 2 assert i_shape[0] == num_tokens hidden_size = i_shape[1] # Expected lora weight shape [num_loras, lora_rank, hidden_size] assert len(lw_shape) == 3 assert lw_shape[2] == hidden_size lora_rank = lw_shape[1] # Expected output shape [num_slices, num_tokens, lora_rank] assert len(o_shape) == 3 assert o_shape == (num_slices, num_tokens, lora_rank) return { "inputs": self.input, "lora_a_weights": self.lora_weights_lst, "output_tensor": self.output, "token_lora_mapping": self.lora_kernel_meta.token_lora_mapping, "token_indices_sorted_by_lora_ids": ( self.lora_kernel_meta.token_indices_sorted_by_lora_ids ), "num_tokens_per_lora": self.lora_kernel_meta.num_tokens_per_lora, "lora_token_start_loc": self.lora_kernel_meta.lora_token_start_loc, "lora_ids": self.lora_kernel_meta.active_lora_ids, "scaling": 1.0, "no_lora_flag_cpu": self.lora_kernel_meta.no_lora_flag_cpu, } def as_lora_expand_kwargs( self, ctx: BenchmarkContext, op_type: OpType, add_inputs: bool ) -> dict[str, Any]: self.sanity_check(ctx, op_type) self.to_device(self.input.device) _, num_tokens, _, num_slices = self.metadata(ctx, op_type) # Sanity check matrix shapes. i_shape, lw_shape, o_shape = ( self.input.shape, self.lora_weights_lst[0].shape, self.output.shape, ) # Expected input shape : [num_slices, num_tokens, lora_rank] assert len(i_shape) == 3 assert i_shape[0] == num_slices assert i_shape[1] == num_tokens lora_rank = i_shape[2] # Expected lora weight shape : [num_lora, hidden_size, lora_rank] assert len(lw_shape) == 3 assert lw_shape[2] == lora_rank hidden_size = lw_shape[1] # Expected output shape : [num_tokens, hidden_size * num_slices] assert len(o_shape) == 2 assert o_shape == (num_tokens, hidden_size * num_slices) return { "inputs": self.input, "lora_b_weights": self.lora_weights_lst, "output_tensor": self.output, "token_lora_mapping": self.lora_kernel_meta.token_lora_mapping, "token_indices_sorted_by_lora_ids": ( self.lora_kernel_meta.token_indices_sorted_by_lora_ids ), "num_tokens_per_lora": self.lora_kernel_meta.num_tokens_per_lora, "lora_token_start_loc": self.lora_kernel_meta.lora_token_start_loc, "lora_ids": self.lora_kernel_meta.active_lora_ids, "offset_start": 0, "add_inputs": add_inputs, "no_lora_flag_cpu": self.lora_kernel_meta.no_lora_flag_cpu, } def as_fused_moe_lora_shrink_kwargs( self, ctx: BenchmarkContext, op_type: OpType ) -> dict[str, Any]: self.sanity_check(ctx, op_type) self.to_device(self.input.device) _, num_tokens, _, num_slices = self.metadata(ctx, op_type) # Sanity check matrix shapes. i_shape, lw_shape, o_shape = ( self.input.shape, self.lora_weights_lst[0].shape, self.output.shape, ) # Expected input shape : [num_tokens, hidden_size] for gate_up # Expected input shape : [top_k_num * num_tokens, hidden_size] for down assert len(i_shape) == 2 assert i_shape[0] == num_tokens hidden_size = i_shape[1] # Expected lora weight shape [max_lora, num_experts, lora_rank, hidden_size] assert len(lw_shape) == 4 assert lw_shape[-1] == hidden_size lora_rank = lw_shape[-2] # Expected output shape : [num_slices, num_tokens, top_k_num, lora_rank] assert len(o_shape) == 4 assert ( o_shape == (num_slices, num_tokens // ctx.top_k_num, ctx.top_k_num, lora_rank) if op_type in [OpType.FUSED_MOE_LORA_DOWN_SHRINK] else o_shape == (num_slices, num_tokens, ctx.top_k_num, lora_rank) ) kernel_config = get_lora_op_configs( op_type.name.lower(), max_loras=lw_shape[0], batch=num_tokens, hidden_size=hidden_size, rank=lora_rank, num_slices=num_slices, add_inputs=False, ) (topk_weights, sorted_token_ids, expert_ids, num_tokens_post_padded) = ( self.fused_moe_lora_data_prepare( block_size=kernel_config["BLOCK_SIZE_M"], token_lora_mapping=self.lora_kernel_meta.token_lora_mapping, ctx=ctx, ) ) return { "qcurr_hidden_states": self.input, "lora_a_stacked": self.lora_weights_lst, "a_intermediate_cache1": self.output, "topk_weights": topk_weights, "sorted_token_ids": sorted_token_ids, "expert_ids": expert_ids, "num_tokens_post_padded": num_tokens_post_padded, "top_k_num": ctx.top_k_num, "device": self.input.device, "N": lora_rank, "M": topk_weights.shape[0], "EM": sorted_token_ids.shape[1], "K": self.input.shape[1], "num_tokens": num_tokens, "num_experts": ctx.num_experts, "num_slices": num_slices, "shrink_block_size_m": kernel_config["BLOCK_SIZE_M"], "shrink_block_size_n": kernel_config["BLOCK_SIZE_N"], "shrink_block_size_k": kernel_config["BLOCK_SIZE_K"], "shrink_group_size_m": kernel_config["GROUP_SIZE_M"], "shrink_num_warps": kernel_config["NUM_WARPS"], "shrink_num_stages": kernel_config["NUM_STAGES"], "shrink_split_k": kernel_config.get("SPLIT_K", 1), "mul_routed_weight": op_type.is_fused_moe_lora_down_fn(), } def as_fused_moe_lora_expand_kwargs( self, ctx: BenchmarkContext, op_type: OpType ) -> dict[str, Any]: self.sanity_check(ctx, op_type) self.to_device(self.input.device) _, num_tokens, _, num_slices = self.metadata(ctx, op_type) # Sanity check matrix shapes. i_shape, lw_shape, o_shape = ( self.input.shape, self.lora_weights_lst[0].shape, self.output.shape, ) # Expected input shape : [num_slices, num_tokens, top_k_num, lora_rank] assert len(i_shape) == 4 assert i_shape[0] == num_slices assert i_shape[1] == num_tokens lora_rank = i_shape[-1] # Expected lora weight shape : [num_loras, num_experts, hidden_size, lora_rank] assert len(lw_shape) == 4 assert lw_shape[-1] == lora_rank hidden_size = lw_shape[-2] # Expected output shape : [num_tokens, top_k_num, hidden_size * num_slices] assert len(o_shape) == 3 assert o_shape == (num_tokens, ctx.top_k_num, hidden_size * num_slices) kernel_config = get_lora_op_configs( op_type.name.lower(), max_loras=lw_shape[0], batch=num_tokens, hidden_size=hidden_size, rank=lora_rank, num_slices=num_slices, add_inputs=False, ) (topk_weights, sorted_token_ids, expert_ids, num_tokens_post_padded) = ( self.fused_moe_lora_data_prepare( block_size=kernel_config["BLOCK_SIZE_M"], token_lora_mapping=self.lora_kernel_meta.token_lora_mapping, ctx=ctx, ) ) return { "a_intermediate_cache1": self.input, "lora_b_stacked": self.lora_weights_lst, "output": self.output, "topk_weights": topk_weights, "sorted_token_ids": sorted_token_ids, "expert_ids": expert_ids, "num_tokens_post_padded": num_tokens_post_padded, "top_k_num": ctx.top_k_num, "device": self.input.device, "N": lora_rank, "M": topk_weights.shape[0], "EM": sorted_token_ids.shape[1], "K": self.input.shape[1], "num_tokens": num_tokens, "num_experts": ctx.num_experts, "num_slices": num_slices, "max_lora_rank": lora_rank, "w1_output_dim_size": lw_shape[2], "expand_block_size_m": kernel_config["BLOCK_SIZE_M"], "expand_block_size_n": kernel_config["BLOCK_SIZE_N"], "expand_block_size_k": kernel_config["BLOCK_SIZE_K"],
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/bench_mxfp4_qutlass.py
benchmarks/kernels/bench_mxfp4_qutlass.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # # Copyright (C) 2025 Roberto L. Castro (Roberto.LopezCastro@ist.ac.at). # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import argparse import copy import itertools import torch from compressed_tensors.transform.utils.hadamard import deterministic_hadamard_matrix from weight_shapes import WEIGHT_SHAPES from vllm._custom_ops import fusedQuantizeMx, matmul_mxf4_bf16_tn from vllm.model_executor.layers.quantization.qutlass_utils import to_blocked from vllm.triton_utils import triton PROVIDER_CFGS = { "torch-bf16": dict(enabled=True), "mxfp4": dict(no_a_quant=False, enabled=True), "mxfp4-noquant": dict(no_a_quant=True, enabled=True), } _enabled = [k for k, v in PROVIDER_CFGS.items() if v["enabled"]] def get_hadamard_matrix(group_size: int, dtype: torch.dtype, device: torch.device): return ( deterministic_hadamard_matrix(group_size, dtype=dtype, device=device) * group_size**-0.5 ) def _quant_weight_mxfp4( b: torch.Tensor, forward_hadamard_matrix: torch.Tensor, device: str ): weight_hf_e2m1, weight_hf_e8m0 = fusedQuantizeMx( b, forward_hadamard_matrix, method="abs_max" ) weight_hf_scale_block = to_blocked(weight_hf_e8m0, backend="triton") return weight_hf_e2m1, weight_hf_scale_block def build_mxfp4_runner(cfg, a, b, forward_hadamard_matrix, dtype, device): weight_hf_e2m1, weight_hf_scale_block = _quant_weight_mxfp4( b, forward_hadamard_matrix, device ) alpha = torch.tensor([1.0], device="cuda") if cfg["no_a_quant"]: # Pre-quantize activation input_hf_e2m1, input_hf_e8m0 = fusedQuantizeMx( a, forward_hadamard_matrix, method="abs_max" ) input_hf_scale_block = to_blocked(input_hf_e8m0, backend="triton") def run(): return matmul_mxf4_bf16_tn( input_hf_e2m1, weight_hf_e2m1, input_hf_scale_block, weight_hf_scale_block, alpha, ) return run # Quantize activation on-the-fly def run(): input_hf_e2m1, input_hf_e8m0 = fusedQuantizeMx( a, forward_hadamard_matrix, method="abs_max" ) input_hf_scale_block = to_blocked(input_hf_e8m0, backend="triton") return matmul_mxf4_bf16_tn( input_hf_e2m1, weight_hf_e2m1, input_hf_scale_block, weight_hf_scale_block, alpha, ) return run @triton.testing.perf_report( triton.testing.Benchmark( x_names=["batch_size"], x_vals=[ 1, 4, 8, 16, 32, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 24576, 32768, ], x_log=False, line_arg="provider", line_vals=_enabled, line_names=_enabled, ylabel="TFLOP/s (larger is better)", plot_name="BF16 vs MXFP4 GEMMs", args={}, ) ) def benchmark(batch_size, provider, N, K, had_size): M = batch_size device = "cuda" dtype = torch.bfloat16 a = torch.randn((M, K), device=device, dtype=dtype) b = torch.randn((N, K), device=device, dtype=dtype) forward_hadamard_matrix = get_hadamard_matrix(had_size, dtype, device) quantiles = [0.5, 0.2, 0.8] if provider == "torch-bf16": ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: torch.nn.functional.linear(a, b), rep=200, quantiles=quantiles ) else: cfg = PROVIDER_CFGS[provider] run_quant = build_mxfp4_runner( cfg, a, b, forward_hadamard_matrix, dtype, device ) ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: run_quant(), rep=200, quantiles=quantiles ) to_tflops = lambda t_ms: (2 * M * N * K) * 1e-12 / (t_ms * 1e-3) return to_tflops(ms), to_tflops(max_ms), to_tflops(min_ms) def prepare_shapes(args): out = [] for model, tp_size in itertools.product(args.models, args.tp_sizes): for KN, tp_dim in copy.deepcopy(WEIGHT_SHAPES[model]): KN[tp_dim] //= tp_size KN.append(model) out.append(KN) return out if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--models", nargs="+", type=str, default=["meta-llama/Llama-3.3-70B-Instruct"], choices=list(WEIGHT_SHAPES.keys()), ) parser.add_argument("--tp-sizes", nargs="+", type=int, default=[1]) args = parser.parse_args() for K, N, model in prepare_shapes(args): for had_size in [32, 64, 128]: print(f"{model}, N={N} K={K}, HAD={had_size}, BF16 vs MXFP4 GEMMs TFLOP/s:") benchmark.run( print_data=True, show_plots=True, save_path=f"bench_mxfp4_res_n{N}_k{K}", N=N, K=K, had_size=had_size, ) print("Benchmark finished!")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_reshape_and_cache.py
benchmarks/kernels/benchmark_reshape_and_cache.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import random import time import torch from tabulate import tabulate from vllm import _custom_ops as ops from vllm.logger import init_logger from vllm.platforms import current_platform from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.torch_utils import ( STR_DTYPE_TO_TORCH_DTYPE, create_kv_caches_with_random, ) logger = init_logger(__name__) @torch.inference_mode() def run_benchmark( num_tokens: int, num_heads: int, head_size: int, block_size: int, num_blocks: int, dtype: torch.dtype, kv_cache_dtype: str, num_iters: int, benchmark_mode: str, device: str = "cuda", ) -> float: """Return latency (seconds) for given num_tokens.""" if kv_cache_dtype == "fp8" and head_size % 16: raise ValueError("fp8 kv-cache requires head_size to be a multiple of 16.") current_platform.seed_everything(42) torch.set_default_device(device) # create random key / value tensors [T, H, D]. key = torch.randn(num_tokens, num_heads, head_size, dtype=dtype, device=device) value = torch.randn_like(key) # prepare the slot mapping. # each token is assigned a unique slot in the KV-cache. num_slots = block_size * num_blocks if num_tokens > num_slots: raise ValueError("num_tokens cannot exceed the total number of cache slots") slot_mapping_lst = random.sample(range(num_slots), num_tokens) slot_mapping = torch.tensor(slot_mapping_lst, dtype=torch.long, device=device) key_caches, value_caches = create_kv_caches_with_random( num_blocks, block_size, 1, # num_layers num_heads, head_size, kv_cache_dtype, dtype, device=device, ) key_cache, value_cache = key_caches[0], value_caches[0] # to free unused memory del key_caches, value_caches # compute per-kernel scaling factors for fp8 conversion (if used). k_scale = (key.amax() / 64.0).to(torch.float32) v_scale = (value.amax() / 64.0).to(torch.float32) function_under_test = lambda: ops.reshape_and_cache( key, # noqa: F821 value, # noqa: F821 key_cache, # noqa: F821 value_cache, # noqa: F821 slot_mapping, # noqa: F821 kv_cache_dtype, k_scale, v_scale, ) if benchmark_mode == "cudagraph": g = torch.cuda.CUDAGraph() with torch.cuda.graph(g): function_under_test() torch.cuda.synchronize() function_under_test = lambda: g.replay() def run_cuda_benchmark(n_iters: int) -> float: nonlocal key, value, key_cache, value_cache, slot_mapping torch.cuda.synchronize() start = time.perf_counter() for _ in range(n_iters): function_under_test() torch.cuda.synchronize() end = time.perf_counter() return (end - start) / n_iters # warm-up run_cuda_benchmark(3) lat = run_cuda_benchmark(num_iters) # free tensors to mitigate OOM when sweeping del key, value, key_cache, value_cache, slot_mapping torch.cuda.empty_cache() return lat def main(args): rows = [] for exp in range(1, 17): n_tok = 2**exp lat = run_benchmark( num_tokens=n_tok, num_heads=args.num_heads, head_size=args.head_size, block_size=args.block_size, num_blocks=args.num_blocks, dtype=STR_DTYPE_TO_TORCH_DTYPE[args.dtype], kv_cache_dtype=args.kv_cache_dtype, num_iters=args.iters, benchmark_mode=args.mode, device="cuda", ) rows.append([n_tok, lat * 1e6]) # convert to microseconds print(f"Benchmark results for implementation cuda (measuring with {args.mode}):") print(tabulate(rows, headers=["num_tokens", "latency (µs)"], floatfmt=".3f")) if __name__ == "__main__": parser = FlexibleArgumentParser() parser.add_argument("--num-heads", type=int, default=128) parser.add_argument( "--head-size", type=int, choices=[64, 80, 96, 112, 120, 128, 192, 256], default=128, ) parser.add_argument("--block-size", type=int, choices=[16, 32], default=16) parser.add_argument("--num-blocks", type=int, default=128 * 128) parser.add_argument( "--dtype", type=str, choices=["half", "bfloat16", "float"], default="bfloat16", ) parser.add_argument( "--kv-cache-dtype", type=str, choices=["auto", "fp8"], default="auto", ) parser.add_argument("--iters", type=int, default=200) parser.add_argument( "--mode", type=str, choices=["cudagraph", "no_graph"], default="cudagraph", ) args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/bench_nvfp4_gemm.py
benchmarks/kernels/bench_nvfp4_gemm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import copy import itertools import os import torch from weight_shapes import WEIGHT_SHAPES from vllm import _custom_ops as ops from vllm.platforms import current_platform from vllm.scalar_type import scalar_types from vllm.triton_utils import triton if not current_platform.has_device_capability(100): raise RuntimeError("NVFP4 requires compute capability of 10.0 (Blackwell)") FLOAT4_E2M1_MAX = scalar_types.float4_e2m1f.max() FLOAT8_E4M3_MAX = torch.finfo(torch.float8_e4m3fn).max PROVIDER_CFGS = { "torch-bf16": dict(enabled=True), "nvfp4": dict(no_a_quant=False, enabled=True), "nvfp4-noquant": dict(no_a_quant=True, enabled=True), "fbgemm-nvfp4": dict(fbgemm=True, no_a_quant=False, enabled=True), "fbgemm-nvfp4-noquant": dict(fbgemm=True, no_a_quant=True, enabled=True), } _needs_fbgemm = any( v.get("fbgemm", False) for v in PROVIDER_CFGS.values() if v.get("enabled", False) ) if _needs_fbgemm: try: from fbgemm_gpu.experimental.gemm.triton_gemm.fp4_quantize import ( triton_scale_nvfp4_quant, ) except ImportError: print( "WARNING: FBGEMM providers are enabled but fbgemm_gpu is not installed. " "These providers will be skipped. Please install fbgemm_gpu with: " "'pip install fbgemm-gpu-genai' to run them." ) # Disable FBGEMM providers so the benchmark can run. for cfg in PROVIDER_CFGS.values(): if cfg.get("fbgemm"): cfg["enabled"] = False _enabled = [k for k, v in PROVIDER_CFGS.items() if v["enabled"]] def _quant_weight_nvfp4(b: torch.Tensor, device: str, cfg): # Compute global scale for weight b_amax = torch.abs(b).max().to(torch.float32) b_global_scale = FLOAT8_E4M3_MAX * FLOAT4_E2M1_MAX / b_amax if "fbgemm" in cfg and cfg["fbgemm"]: b_fp4, scale_b_fp4 = triton_scale_nvfp4_quant(b, b_global_scale) else: b_fp4, scale_b_fp4 = ops.scaled_fp4_quant(b, b_global_scale) return b_fp4, scale_b_fp4, b_global_scale def build_nvfp4_runner(cfg, a, b, dtype, device): b_fp4, scale_b_fp4, b_global_scale = _quant_weight_nvfp4(b, device, cfg) # Compute global scale for activation # NOTE: This is generally provided ahead-of-time by the model checkpoint. a_amax = torch.abs(a).max().to(torch.float32) a_global_scale = FLOAT8_E4M3_MAX * FLOAT4_E2M1_MAX / a_amax # Alpha for the GEMM operation alpha = 1.0 / (a_global_scale * b_global_scale) if "fbgemm" in cfg and cfg["fbgemm"]: if cfg["no_a_quant"]: a_fp4, scale_a_fp4 = triton_scale_nvfp4_quant(a, a_global_scale) def run(): return torch.ops.fbgemm.f4f4bf16( a_fp4, b_fp4, scale_a_fp4, scale_b_fp4, global_scale=alpha, use_mx=False, ) return run else: def run(): a_fp4, scale_a_fp4 = triton_scale_nvfp4_quant(a, a_global_scale) return torch.ops.fbgemm.f4f4bf16( a_fp4, b_fp4, scale_a_fp4, scale_b_fp4, global_scale=alpha, use_mx=False, ) return run if cfg["no_a_quant"]: # Pre-quantize activation a_fp4, scale_a_fp4 = ops.scaled_fp4_quant(a, a_global_scale) def run(): return ops.cutlass_scaled_fp4_mm( a_fp4, b_fp4, scale_a_fp4, scale_b_fp4, alpha, dtype ) return run # Quantize activation on-the-fly def run(): a_fp4, scale_a_fp4 = ops.scaled_fp4_quant(a, a_global_scale) return ops.cutlass_scaled_fp4_mm( a_fp4, b_fp4, scale_a_fp4, scale_b_fp4, alpha, dtype ) return run @triton.testing.perf_report( triton.testing.Benchmark( x_names=["batch_size"], x_vals=[1, 16, 64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384], x_log=False, line_arg="provider", line_vals=_enabled, line_names=_enabled, ylabel="TFLOP/s (larger is better)", plot_name="BF16 vs NVFP4 GEMMs", args={}, ) ) def benchmark(batch_size, provider, N, K): M = batch_size device = "cuda" dtype = torch.bfloat16 a = torch.randn((M, K), device=device, dtype=dtype) b = torch.randn((N, K), device=device, dtype=dtype) quantiles = [0.5, 0.2, 0.8] if provider == "torch-bf16": ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: torch.nn.functional.linear(a, b), quantiles=quantiles ) else: cfg = PROVIDER_CFGS[provider] run_quant = build_nvfp4_runner(cfg, a, b, dtype, device) ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( lambda: run_quant(), quantiles=quantiles ) to_tflops = lambda t_ms: (2 * M * N * K) * 1e-12 / (t_ms * 1e-3) return to_tflops(ms), to_tflops(max_ms), to_tflops(min_ms) def prepare_shapes(args): out = [] for model, tp_size in itertools.product(args.models, args.tp_sizes): for KN, tp_dim in copy.deepcopy(WEIGHT_SHAPES[model]): KN[tp_dim] //= tp_size KN.append(model) out.append(KN) return out if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--models", nargs="+", type=str, default=["meta-llama/Llama-3.1-8B-Instruct"], choices=list(WEIGHT_SHAPES.keys()), ) parser.add_argument("--tp-sizes", nargs="+", type=int, default=[1]) args = parser.parse_args() for K, N, model in prepare_shapes(args): print(f"{model}, N={N} K={K}, BF16 vs NVFP4 GEMMs TFLOP/s:") save_dir = f"bench_nvfp4_res_n{N}_k{K}" os.makedirs(save_dir, exist_ok=True) benchmark.run( print_data=True, show_plots=True, save_path=save_dir, N=N, K=K, ) print("Benchmark finished!")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_rope.py
benchmarks/kernels/benchmark_rope.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import itertools import torch from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.triton_utils import triton from vllm.utils.argparse_utils import FlexibleArgumentParser batch_size_range = [2**i for i in range(0, 8, 2)] seq_len_range = [2**i for i in range(6, 10, 1)] num_heads_range = [32, 48] configs = list(itertools.product(batch_size_range, seq_len_range, num_heads_range)) def get_benchmark(head_size, rotary_dim, is_neox_style, device): @triton.testing.perf_report( triton.testing.Benchmark( x_names=["batch_size", "seq_len", "num_heads"], x_vals=[list(_) for _ in configs], line_arg="provider", line_vals=["torch", "flashinfer", "vllm"], line_names=["PyTorch", "FlashInfer", "vLLM"], styles=[("blue", "-"), ("green", "-"), ("red", "-")], ylabel="us", plot_name=f"rope-perf{'-neox-style' if is_neox_style else ''}", args={}, ) ) def benchmark(batch_size, seq_len, num_heads, provider): dtype = torch.bfloat16 max_position = 8192 rope_parameters = {"partial_rotary_factor": rotary_dim / head_size} rope = get_rope(head_size, max_position, is_neox_style, rope_parameters) rope = rope.to(dtype=dtype, device=device) cos_sin_cache = rope.cos_sin_cache.to(dtype=torch.float, device=device) positions = torch.randint(0, max_position, (batch_size, seq_len), device=device) query = torch.randn( (batch_size, seq_len, num_heads * head_size), dtype=dtype, device=device ) key = torch.randn_like(query) quantiles = [0.5, 0.2, 0.8] if provider == "torch": ms, min_ms, max_ms = triton.testing.do_bench( lambda: rope.forward_native(positions, query.clone(), key.clone()), quantiles=quantiles, ) elif provider == "flashinfer": ms, min_ms, max_ms = triton.testing.do_bench( lambda: torch.ops.vllm.flashinfer_rotary_embedding( positions, query.clone(), key.clone(), head_size, cos_sin_cache, is_neox_style, ), quantiles=quantiles, ) else: ms, min_ms, max_ms = triton.testing.do_bench( lambda: rope.forward_cuda(positions, query.clone(), key.clone()), quantiles=quantiles, ) return 1000 * ms, 1000 * max_ms, 1000 * min_ms return benchmark if __name__ == "__main__": parser = FlexibleArgumentParser( description="Benchmark the rotary embedding kernels." ) parser.add_argument("--is-neox-style", type=bool, default=True) parser.add_argument("--batch-size", type=int, default=16) parser.add_argument("--seq-len", type=int, default=512) parser.add_argument("--num-heads", type=int, default=8) parser.add_argument( "--head-size", type=int, choices=[64, 80, 96, 112, 120, 128, 192, 256], default=128, ) parser.add_argument("--rotary-dim", type=int, choices=[16, 32], default=32) parser.add_argument( "--dtype", type=str, choices=["bfloat16", "float"], default="float" ) parser.add_argument("--seed", type=int, default=0) parser.add_argument( "--device", type=str, choices=["cuda:0", "cuda:1"], default="cuda:0" ) parser.add_argument("--save-path", type=str, default="./configs/rope/") args = parser.parse_args() # Get the benchmark function benchmark = get_benchmark( args.head_size, args.rotary_dim, args.is_neox_style, args.device ) # Run performance benchmark benchmark.run(print_data=True, save_path=args.save_path)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_device_communicators.py
benchmarks/kernels/benchmark_device_communicators.py
#!/usr/bin/env python3 # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Benchmark script for device communicators: CustomAllreduce (oneshot, twoshot), PyNcclCommunicator, and SymmMemCommunicator (multimem, two-shot). for NCCL symmetric memory you need to set the environment variables NCCL_NVLS_ENABLE=1 NCCL_CUMEM_ENABLE=1 VLLM_USE_NCCL_SYMM_MEM=1, otherwise NCCL does not use fast NVLS implementation for all reduce. Usage: torchrun --nproc_per_node=<N> benchmark_device_communicators.py [options] Example: torchrun --nproc_per_node=2 benchmark_device_communicators.py --sequence-lengths 512 1024 2048 --num-warmup 10 --num-trials 100 """ import json import os import time from collections.abc import Callable from contextlib import nullcontext import torch import torch.distributed as dist from torch.distributed import ProcessGroup from vllm.distributed.device_communicators.custom_all_reduce import CustomAllreduce from vllm.distributed.device_communicators.pynccl import ( PyNcclCommunicator, register_nccl_symmetric_ops, ) from vllm.distributed.device_communicators.pynccl_allocator import ( set_graph_pool_id, ) from vllm.distributed.device_communicators.symm_mem import SymmMemCommunicator from vllm.logger import init_logger from vllm.utils.argparse_utils import FlexibleArgumentParser logger = init_logger(__name__) # Default sequence lengths to benchmark DEFAULT_SEQUENCE_LENGTHS = [128, 512, 1024, 2048, 4096, 8192] # Fixed hidden size and dtype for all benchmarks HIDDEN_SIZE = 8192 BENCHMARK_DTYPE = torch.bfloat16 # CUDA graph settings CUDA_GRAPH_CAPTURE_CYCLES = 10 class CommunicatorBenchmark: """Benchmark class for testing device communicators.""" def __init__( self, rank: int, world_size: int, device: torch.device, cpu_group: ProcessGroup, sequence_lengths: list[int], ): self.rank = rank self.world_size = world_size self.device = device self.cpu_group = cpu_group # Calculate max_size_override based on largest sequence length max_seq_len = max(sequence_lengths) max_tensor_elements = max_seq_len * HIDDEN_SIZE self.max_size_override = max_tensor_elements * BENCHMARK_DTYPE.itemsize + 1 # Initialize communicators self.custom_allreduce = None self.pynccl_comm = None self.symm_mem_comm = None self.symm_mem_comm_multimem = None self.symm_mem_comm_two_shot = None self._init_communicators() def _init_communicators(self): """Initialize all available communicators.""" try: self.custom_allreduce = CustomAllreduce( group=self.cpu_group, device=self.device, max_size=self.max_size_override, ) if not self.custom_allreduce.disabled: logger.info("Rank %s: CustomAllreduce initialized", self.rank) else: logger.info("Rank %s: CustomAllreduce disabled", self.rank) except Exception as e: logger.warning( "Rank %s: Failed to initialize CustomAllreduce: %s", self.rank, e ) self.custom_allreduce = None try: self.pynccl_comm = PyNcclCommunicator( group=self.cpu_group, device=self.device ) if not self.pynccl_comm.disabled: logger.info("Rank %s: PyNcclCommunicator initialized", self.rank) register_nccl_symmetric_ops(self.pynccl_comm) else: logger.info("Rank %s: PyNcclCommunicator disabled", self.rank) self.pynccl_comm = None except Exception as e: logger.warning( "Rank %s: Failed to initialize PyNcclCommunicator: %s", self.rank, e ) self.pynccl_comm = None # Initialize variants for SymmMemCommunicator try: self.symm_mem_comm_multimem = SymmMemCommunicator( group=self.cpu_group, device=self.device, force_multimem=True, max_size_override=self.max_size_override, ) if not self.symm_mem_comm_multimem.disabled: logger.info( "Rank %s: SymmMemCommunicator (multimem) initialized", self.rank ) else: self.symm_mem_comm_multimem = None except Exception as e: logger.warning( "Rank %s: Failed to initialize SymmMemCommunicator (multimem): %s", self.rank, e, ) self.symm_mem_comm_multimem = None try: self.symm_mem_comm_two_shot = SymmMemCommunicator( group=self.cpu_group, device=self.device, force_multimem=False, max_size_override=self.max_size_override, ) if not self.symm_mem_comm_two_shot.disabled: logger.info( "Rank %s: SymmMemCommunicator (two_shot) initialized", self.rank ) else: self.symm_mem_comm_two_shot = None except Exception as e: logger.warning( "Rank %s: Failed to initialize SymmMemCommunicator (two_shot): %s", self.rank, e, ) self.symm_mem_comm_two_shot = None def benchmark_allreduce( self, sequence_length: int, num_warmup: int, num_trials: int ) -> dict[str, float]: """Benchmark allreduce operations for all available communicators.""" results = {} # Define communicators with their benchmark functions communicators = [] if self.custom_allreduce is not None: comm = self.custom_allreduce # CustomAllreduce one-shot communicators.append( ( "ca_1stage", lambda t, c=comm: c.custom_all_reduce(t), lambda t, c=comm: c.should_custom_ar(t), comm.capture(), "1stage", # env variable value ) ) # CustomAllreduce two-shot communicators.append( ( "ca_2stage", lambda t, c=comm: c.custom_all_reduce(t), lambda t, c=comm: c.should_custom_ar(t), comm.capture(), "2stage", # env variable value ) ) if self.pynccl_comm is not None: comm = self.pynccl_comm communicators.append( ( "pynccl", lambda t, c=comm: c.all_reduce(t), lambda t: True, # Always available if initialized nullcontext(), None, # no env variable needed ) ) communicators.append( ( "pynccl-symm", lambda t: torch.ops.vllm.all_reduce_symmetric_with_copy(t), lambda t: True, # Always available if initialized nullcontext(), None, # no env variable needed ) ) if self.symm_mem_comm_multimem is not None: comm = self.symm_mem_comm_multimem communicators.append( ( "symm_mem_multimem", lambda t, c=comm: c.all_reduce(t), lambda t, c=comm: c.should_use_symm_mem(t), nullcontext(), None, # no env variable needed ) ) if self.symm_mem_comm_two_shot is not None: comm = self.symm_mem_comm_two_shot communicators.append( ( "symm_mem_two_shot", lambda t, c=comm: c.all_reduce(t), lambda t, c=comm: c.should_use_symm_mem(t), nullcontext(), None, # no env variable needed ) ) # Benchmark each communicator for name, allreduce_fn, should_use_fn, context, env_var in communicators: # Set environment variable if needed if env_var is not None: os.environ["VLLM_CUSTOM_ALLREDUCE_ALGO"] = env_var else: # Clear the environment variable to avoid interference os.environ.pop("VLLM_CUSTOM_ALLREDUCE_ALGO", None) latency = self.benchmark_allreduce_single( sequence_length, allreduce_fn, should_use_fn, context, num_warmup, num_trials, ) if latency is not None: results[name] = latency return results def benchmark_allreduce_single( self, sequence_length: int, allreduce_fn: Callable[[torch.Tensor], torch.Tensor | None], should_use_fn: Callable[[torch.Tensor], bool], context, num_warmup: int, num_trials: int, ) -> float | None: """Benchmark method with CUDA graph optimization.""" try: # Create test tensor (2D: sequence_length x hidden_size) tensor = torch.randn( sequence_length, HIDDEN_SIZE, dtype=BENCHMARK_DTYPE, device=self.device ) if not should_use_fn(tensor): return None torch.cuda.synchronize() stream = torch.cuda.Stream() with torch.cuda.stream(stream): graph_input = tensor.clone() # Warmup before capture for _ in range(3): allreduce_fn(graph_input) # Capture the graph using context manager with context: graph = torch.cuda.CUDAGraph() graph_pool = torch.cuda.graph_pool_handle() set_graph_pool_id(graph_pool) with torch.cuda.graph(graph, pool=graph_pool, stream=stream): for _ in range(CUDA_GRAPH_CAPTURE_CYCLES): allreduce_fn(graph_input) torch.cuda.synchronize() for _ in range(num_warmup): graph.replay() torch.cuda.synchronize() torch.cuda.synchronize() start_time = time.perf_counter() for _ in range(num_trials): graph.replay() torch.cuda.synchronize() end_time = time.perf_counter() # Convert to ms and divide by CUDA_GRAPH_CAPTURE_CYCLES return ( (end_time - start_time) / num_trials / CUDA_GRAPH_CAPTURE_CYCLES * 1000 ) except Exception as e: logger.error("CUDA graph benchmark failed: %s", e) raise RuntimeError( f"CUDA graph benchmark failed for communicator: {e}" ) from e def _calculate_speedup_info(comm_results: dict[str, float]) -> str: """Calculate speedup information for a single tensor size.""" if not comm_results: return "N/A" # Find the fastest communicator fastest_comm = min(comm_results.keys(), key=lambda k: comm_results[k]) fastest_time = comm_results[fastest_comm] # Calculate speedup vs PyNccl if available if "pynccl" in comm_results: pynccl_time = comm_results["pynccl"] speedup = pynccl_time / fastest_time return f"{fastest_comm} ({speedup:.2f}x)" else: return f"{fastest_comm} (N/A)" def print_results( results: dict[str, dict[str, float]], sequence_lengths: list[int], world_size: int ): """Print benchmark results in a formatted table.""" print(f"\n{'=' * 130}") print("Device Communicator Benchmark Results") print( f"World Size: {world_size}, Data Type: {BENCHMARK_DTYPE}, " f"Hidden Size: {HIDDEN_SIZE}" ) print(f"{'=' * 130}") # Get all communicator names all_comms = set() for size_results in results.values(): all_comms.update(size_results.keys()) all_comms = sorted(list(all_comms)) # Print header header = f"{'Tensor Shape':<20}{'Tensor Size':<15}" for comm in all_comms: header += f"{comm:<20}" header += f"{'Best (Speedup vs PyNccl)':<30}" print(header) print("-" * len(header)) # Print results for each sequence length for seq_len in sequence_lengths: if seq_len in results: # Calculate tensor size in elements and bytes tensor_elements = seq_len * HIDDEN_SIZE tensor_bytes = tensor_elements * BENCHMARK_DTYPE.itemsize # Format tensor size (MB) tensor_size_mb = tensor_bytes / (1024 * 1024) tensor_size_str = f"{tensor_size_mb:.2f} MB" # Format tensor shape tensor_shape = f"({seq_len}, {HIDDEN_SIZE})" row = f"{tensor_shape:<20}{tensor_size_str:<15}" for comm in all_comms: if comm in results[seq_len]: row += f"{results[seq_len][comm]:<20.3f}" else: row += f"{'N/A':<20}" # Calculate speedup information speedup_info = _calculate_speedup_info(results[seq_len]) row += f"{speedup_info:<30}" print(row) print(f"{'=' * 130}") print("All times are in milliseconds (ms) per allreduce operation") print("Speedup column shows: fastest_algorithm (speedup_vs_pynccl)") def main(): parser = FlexibleArgumentParser(description="Benchmark device communicators") parser.add_argument( "--sequence-lengths", type=int, nargs="+", default=DEFAULT_SEQUENCE_LENGTHS, help="Sequence lengths to benchmark (tensor shape: seq_len x hidden_size)", ) parser.add_argument( "--num-warmup", type=int, default=5, help="Number of warmup iterations" ) parser.add_argument( "--num-trials", type=int, default=50, help="Number of benchmark trials" ) parser.add_argument("--output-json", type=str, help="Output results to JSON file") args = parser.parse_args() # Initialize distributed if not dist.is_initialized(): dist.init_process_group(backend="gloo") rank = dist.get_rank() world_size = dist.get_world_size() # Set device device = torch.device(f"cuda:{rank}") torch.cuda.set_device(device) # Get CPU process group cpu_group = dist.new_group(backend="gloo") # Disable USE_SYMM_MEM to avoid affecting the max_sizes # in symm_mem and custom_all_reduce for benchmark os.environ["VLLM_ALLREDUCE_USE_SYMM_MEM"] = "0" # Initialize benchmark benchmark = CommunicatorBenchmark( rank, world_size, device, cpu_group, args.sequence_lengths ) # Run benchmarks all_results = {} for seq_len in args.sequence_lengths: if rank == 0: logger.info( "Benchmarking sequence length: %s (tensor shape: %s x %s)", seq_len, seq_len, HIDDEN_SIZE, ) results = benchmark.benchmark_allreduce( sequence_length=seq_len, num_warmup=args.num_warmup, num_trials=args.num_trials, ) all_results[seq_len] = results # Synchronize between ranks dist.barrier() # Print results (only rank 0) if rank == 0: print_results(all_results, args.sequence_lengths, world_size) # Save to JSON if requested if args.output_json: # Add speedup information to results enhanced_results = {} for seq_len, comm_results in all_results.items(): enhanced_results[seq_len] = { "timings": comm_results, "speedup_info": _calculate_speedup_info(comm_results), } output_data = { "world_size": world_size, "dtype": str(BENCHMARK_DTYPE), "hidden_size": HIDDEN_SIZE, "sequence_lengths": args.sequence_lengths, "num_warmup": args.num_warmup, "num_trials": args.num_trials, "cuda_graph_capture_cycles": CUDA_GRAPH_CAPTURE_CYCLES, "results": enhanced_results, } with open(args.output_json, "w") as f: json.dump(output_data, f, indent=2) logger.info("Results saved to %s", args.output_json) # Cleanup if cpu_group != dist.group.WORLD: dist.destroy_process_group(cpu_group) if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_reshape_and_cache_flash.py
benchmarks/kernels/benchmark_reshape_and_cache_flash.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import random import time import torch from tabulate import tabulate from vllm import _custom_ops as ops from vllm.attention.ops.triton_reshape_and_cache_flash import ( triton_reshape_and_cache_flash, ) from vllm.logger import init_logger from vllm.platforms import current_platform from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.torch_utils import ( STR_DTYPE_TO_TORCH_DTYPE, create_kv_caches_with_random_flash, ) logger = init_logger(__name__) @torch.inference_mode() def run_benchmark( num_tokens: int, num_heads: int, head_size: int, block_size: int, num_blocks: int, dtype: torch.dtype, kv_cache_dtype: str, kv_cache_layout: str, num_iters: int, implementation: str, benchmark_mode: str, device: str = "cuda", ) -> float: """Return latency (seconds) for given num_tokens.""" if kv_cache_dtype == "fp8" and head_size % 16: raise ValueError("fp8 kv-cache requires head_size to be a multiple of 16.") if implementation not in ("cuda", "triton"): raise ValueError( f"Unsupported implementation: {implementation}. " "Only 'cuda' and 'triton' are supported." ) if implementation == "triton" and kv_cache_layout == "HND": return float("nan") # Triton does not support HND layout yet. current_platform.seed_everything(42) torch.set_default_device(device) # create random key / value tensors [T, H, D]. key = torch.randn(num_tokens, num_heads, head_size, dtype=dtype, device=device) value = torch.randn_like(key) # prepare the slot mapping. # each token is assigned a unique slot in the KV-cache. num_slots = block_size * num_blocks if num_tokens > num_slots: raise ValueError("num_tokens cannot exceed the total number of cache slots") slot_mapping_lst = random.sample(range(num_slots), num_tokens) slot_mapping = torch.tensor(slot_mapping_lst, dtype=torch.long, device=device) key_caches, value_caches = create_kv_caches_with_random_flash( num_blocks, block_size, 1, # num_layers num_heads, head_size, kv_cache_dtype, dtype, device=device, cache_layout=kv_cache_layout, ) key_cache, value_cache = key_caches[0], value_caches[0] # to free unused memory del key_caches, value_caches # compute per-kernel scaling factors for fp8 conversion (if used). k_scale = (key.amax() / 64.0).to(torch.float32) v_scale = (value.amax() / 64.0).to(torch.float32) if implementation == "cuda": function_under_test = lambda: ops.reshape_and_cache_flash( key, # noqa: F821 value, # noqa: F821 key_cache, # noqa: F821 value_cache, # noqa: F821 slot_mapping, # noqa: F821 kv_cache_dtype, k_scale, v_scale, ) else: function_under_test = lambda: triton_reshape_and_cache_flash( key, # noqa: F821 value, # noqa: F821 key_cache, # noqa: F821 value_cache, # noqa: F821 slot_mapping, # noqa: F821 kv_cache_dtype, k_scale, v_scale, ) if benchmark_mode == "cudagraph": g = torch.cuda.CUDAGraph() with torch.cuda.graph(g): function_under_test() torch.cuda.synchronize() function_under_test = lambda: g.replay() def run_cuda_benchmark(n_iters: int) -> float: nonlocal key, value, key_cache, value_cache, slot_mapping torch.cuda.synchronize() start = time.perf_counter() for _ in range(n_iters): function_under_test() torch.cuda.synchronize() end = time.perf_counter() return (end - start) / n_iters # warm-up run_cuda_benchmark(3) lat = run_cuda_benchmark(num_iters) # free tensors to mitigate OOM when sweeping del key, value, key_cache, value_cache, slot_mapping torch.cuda.empty_cache() return lat def main(args): rows = [] for layout in ["NHD", "HND"]: for exp in range(1, 17): n_tok = 2**exp lat = run_benchmark( num_tokens=n_tok, num_heads=args.num_heads, head_size=args.head_size, block_size=args.block_size, num_blocks=args.num_blocks, dtype=STR_DTYPE_TO_TORCH_DTYPE[args.dtype], kv_cache_dtype=args.kv_cache_dtype, kv_cache_layout=layout, num_iters=args.iters, implementation=args.implementation, benchmark_mode=args.mode, device="cuda", ) rows.append([n_tok, layout, f"{lat * 1e6:.3f}"]) print( f"Benchmark results for implementation {args.implementation}" f" (measuring with {args.mode}):" ) print(tabulate(rows, headers=["num_tokens", "layout", "latency (µs)"])) if __name__ == "__main__": parser = FlexibleArgumentParser() parser.add_argument("--num-heads", type=int, default=128) parser.add_argument( "--head-size", type=int, choices=[64, 80, 96, 112, 120, 128, 192, 256], default=128, ) parser.add_argument("--block-size", type=int, choices=[16, 32], default=16) parser.add_argument("--num-blocks", type=int, default=128 * 512) parser.add_argument( "--dtype", type=str, choices=["half", "bfloat16", "float"], default="bfloat16", ) parser.add_argument( "--kv-cache-dtype", type=str, choices=["auto", "fp8"], default="auto", ) parser.add_argument("--iters", type=int, default=100) parser.add_argument( "--implementation", type=str, choices=["cuda", "triton"], default="cuda", ) parser.add_argument( "--mode", type=str, choices=["cudagraph", "no_graph"], default="cudagraph", ) args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_w8a8_block_fp8.py
benchmarks/kernels/benchmark_w8a8_block_fp8.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Adapted from sglang quantization/tuning_block_wise_kernel.py import argparse import json import multiprocessing as mp import os import time from datetime import datetime from typing import Any import torch from tqdm import tqdm from vllm.model_executor.layers.quantization.utils.fp8_utils import ( _w8a8_triton_block_scaled_mm, ) from vllm.platforms import current_platform from vllm.triton_utils import triton from vllm.utils.argparse_utils import FlexibleArgumentParser mp.set_start_method("spawn", force=True) assert current_platform.is_cuda(), ( "Only support tune w8a8 block fp8 kernel on CUDA device." ) DTYPE_MAP = { "float32": torch.float32, "float16": torch.float16, "half": torch.half, "bfloat16": torch.bfloat16, } def w8a8_block_matmul( A: torch.Tensor, B: torch.Tensor, As: torch.Tensor, Bs: torch.Tensor, block_size: list[int], config: dict[str, Any], output_dtype: torch.dtype = torch.float16, ) -> torch.Tensor: """This function performs matrix multiplication with block-wise quantization. It takes two input tensors `A` and `B` with scales `As` and `Bs`. The output is returned in the specified `output_dtype`. Args: A: The input tensor, e.g., activation. B: The input tensor, e.g., weight. As: The per-token-group quantization scale for `A`. Bs: The per-block quantization scale for `B`. block_size: The block size for per-block quantization. It should be 2-dim, e.g., [128, 128]. output_dtype: The dtype of the returned tensor. Returns: torch.Tensor: The result of matmul. """ assert len(block_size) == 2 block_n, block_k = block_size[0], block_size[1] assert A.shape[-1] == B.shape[-1] assert A.shape[:-1] == As.shape[:-1] and A.is_contiguous() assert triton.cdiv(A.shape[-1], block_k) == As.shape[-1] M = A.numel() // A.shape[-1] assert B.ndim == 2 and B.is_contiguous() and Bs.ndim == 2 N, K = B.shape assert triton.cdiv(N, block_n) == Bs.shape[0] assert triton.cdiv(K, block_k) == Bs.shape[1] C_shape = A.shape[:-1] + (N,) C = A.new_empty(C_shape, dtype=output_dtype) def grid(META): return ( triton.cdiv(M, META["BLOCK_SIZE_M"]) * triton.cdiv(N, META["BLOCK_SIZE_N"]), ) if A.dtype == torch.float8_e4m3fn: kernel = _w8a8_triton_block_scaled_mm else: raise RuntimeError("Currently, only support tune w8a8 block fp8 kernel.") kernel[grid]( A, B, C, As, Bs, M, N, K, block_n, block_k, A.stride(-2), A.stride(-1), B.stride(1), B.stride(0), C.stride(-2), C.stride(-1), As.stride(-2), As.stride(-1), Bs.stride(1), Bs.stride(0), **config, ) return C def get_configs_compute_bound(): configs = [] for num_stages in [2, 3, 4, 5]: for block_m in [16, 32, 64, 128, 256]: for block_k in [64, 128]: for block_n in [32, 64, 128, 256]: for num_warps in [4, 8]: for group_size in [1, 16, 32, 64]: configs.append( { "BLOCK_SIZE_M": block_m, "BLOCK_SIZE_N": block_n, "BLOCK_SIZE_K": block_k, "GROUP_SIZE_M": group_size, "num_warps": num_warps, "num_stages": num_stages, } ) return configs def get_weight_shapes(tp_size): # NOTE(HandH1998): The weight shapes only works for DeepSeek-V3. # Modify them, if you tune for another different model. # cannot TP total = [ (512 + 64, 7168), (2112, 7168), ((128 + 64) * 128, 7168), (128 * (128 + 128), 512), (7168, 16384), (7168, 18432), ] # N can TP n_tp = [ (18432 * 2, 7168), ((128 + 64) * 128, 7168), (128 * (128 + 128), 512), (24576, 1536), (12288, 7168), (4096, 7168), ] # K can TP k_tp = [(7168, 18432), (7168, 16384), (7168, 2048)] weight_shapes = [] for t in total: weight_shapes.append(t) for n_t in n_tp: new_t = (n_t[0] // tp_size, n_t[1]) weight_shapes.append(new_t) for k_t in k_tp: new_t = (k_t[0], k_t[1] // tp_size) weight_shapes.append(new_t) return weight_shapes def benchmark_config( A, B, As, Bs, block_size, config, out_dtype=torch.float16, num_iters=10 ): def run(): w8a8_block_matmul(A, B, As, Bs, block_size, config, out_dtype) torch.cuda.synchronize() # JIT complication & warmup for _ in range(5): run() torch.cuda.synchronize() start_event = torch.Event(enable_timing=True) end_event = torch.Event(enable_timing=True) latencies: list[float] = [] for i in range(num_iters): torch.cuda.synchronize() start_event.record() run() end_event.record() end_event.synchronize() latencies.append(start_event.elapsed_time(end_event)) avg = sum(latencies) / (num_iters * 10) * 1000 # us return avg def tune(M, N, K, block_size, out_dtype, search_space, input_type): factor_for_scale = 1e-2 if input_type == "fp8": fp8_info = torch.finfo(torch.float8_e4m3fn) fp8_max, fp8_min = fp8_info.max, fp8_info.min A_fp32 = ( (torch.rand(M, K, dtype=torch.float32, device="cuda") - 0.5) * 2 * fp8_max ) A = A_fp32.clamp(min=fp8_min, max=fp8_max).to(torch.float8_e4m3fn) B_fp32 = ( (torch.rand(N, K, dtype=torch.float32, device="cuda") - 0.5) * 2 * fp8_max ) B = B_fp32.clamp(min=fp8_min, max=fp8_max).to(torch.float8_e4m3fn) else: raise RuntimeError("Currently, only support tune w8a8 block fp8 kernel.") block_n, block_k = block_size[0], block_size[1] n_tiles = (N + block_n - 1) // block_n k_tiles = (K + block_k - 1) // block_k As = torch.rand(M, k_tiles, dtype=torch.float32, device="cuda") * factor_for_scale Bs = ( torch.rand(n_tiles, k_tiles, dtype=torch.float32, device="cuda") * factor_for_scale ) best_config = None best_time = float("inf") for config in tqdm(search_space): try: kernel_time = benchmark_config( A, B, As, Bs, block_size, config, out_dtype, num_iters=10, ) except triton.runtime.autotuner.OutOfResources: # Some configurations may be invalid and fail to compile. continue if kernel_time < best_time: best_time = kernel_time best_config = config now = datetime.now() print(f"{now.ctime()}] Completed tuning for batch_size={M}") assert best_config is not None return best_config def save_configs( N, K, block_n, block_k, configs, save_path, input_type="fp8", ) -> None: os.makedirs(save_path, exist_ok=True) device_name = current_platform.get_device_name().replace(" ", "_") json_file_name = ( f"N={N},K={K},device_name={device_name},dtype={input_type}_w8a8," f"block_shape=[{block_n},{block_k}].json" ) config_file_path = os.path.join(save_path, json_file_name) print(f"Writing best config to {config_file_path}...") with open(config_file_path, "w") as f: json.dump(configs, f, indent=4) f.write("\n") def tune_on_gpu(args_dict): """Run tuning on a specific GPU.""" gpu_id = args_dict["gpu_id"] batch_sizes = args_dict["batch_sizes"] weight_shapes = args_dict["weight_shapes"] args = args_dict["args"] torch.cuda.set_device(gpu_id) print(f"Starting tuning on GPU {gpu_id} with batch sizes {batch_sizes}") block_n = args.block_n block_k = args.block_k out_dtype = DTYPE_MAP[args.out_dtype] save_path = args.save_path input_type = args.input_type search_space = get_configs_compute_bound() search_space = [ config for config in search_space if block_k % config["BLOCK_SIZE_K"] == 0 ] start = time.time() for shape in tqdm(weight_shapes, desc=f"GPU {gpu_id} - Shapes"): N, K = shape[0], shape[1] print(f"[GPU {gpu_id}] Tune for weight shape of `N: {N}, K: {K}`") benchmark_results = [ tune( batch_size, N, K, [block_n, block_k], out_dtype, search_space, input_type, ) for batch_size in tqdm(batch_sizes, desc=f"GPU {gpu_id} - Batch sizes") ] best_configs = {M: config for M, config in zip(batch_sizes, benchmark_results)} save_configs(N, K, block_n, block_k, best_configs, save_path, input_type) end = time.time() print(f"Tuning on GPU {gpu_id} took {end - start:.2f} seconds") def distribute_batch_sizes(batch_sizes, num_gpus): """Distribute batch sizes across available GPUs.""" batches_per_gpu = [] for i in range(num_gpus): start_idx = i * len(batch_sizes) // num_gpus end_idx = (i + 1) * len(batch_sizes) // num_gpus batches_per_gpu.append(batch_sizes[start_idx:end_idx]) return batches_per_gpu def main(args): print(args) num_gpus = torch.cuda.device_count() if num_gpus == 0: raise RuntimeError("No GPU available for tuning") print(f"Found {num_gpus} GPUs for parallel tuning") torch.cuda.init() if args.batch_size is None: batch_sizes = [ 1, 2, 4, 8, 16, 24, 32, 48, 64, 96, 128, 256, 512, 1024, 1536, 2048, 3072, 4096, ] else: batch_sizes = [args.batch_size] num_gpus = 1 # If only one batch size, use only one GPU weight_shapes = get_weight_shapes(args.tp_size) batches_per_gpu = distribute_batch_sizes(batch_sizes, num_gpus) process_args = [] for gpu_id in range(num_gpus): process_args.append( { "gpu_id": gpu_id, "batch_sizes": batches_per_gpu[gpu_id], "weight_shapes": weight_shapes, # Each GPU processes all weight shapes "args": args, } ) ctx = mp.get_context("spawn") with ctx.Pool(num_gpus) as pool: pool.map(tune_on_gpu, process_args) print("Multi-GPU tuning completed") if __name__ == "__main__": parser = FlexibleArgumentParser( description=""" Tune triton w8a8 block fp8 for DeepSeek-V3/DeepSeek-R1: python3 benchmark_w8a8_block_fp8.py --tp-size 8 --input-type fp8 Then copy to model_executor/layers/quantization/utils/configs """, formatter_class=argparse.RawTextHelpFormatter, ) parser.add_argument("--tp-size", "-tp", type=int, default=8) parser.add_argument("--input-type", type=str, choices=["fp8"], default="fp8") parser.add_argument( "--out-dtype", type=str, choices=["float32", "float16", "bfloat16", "half"], default="float16", ) parser.add_argument("--block-n", type=int, default=128) parser.add_argument("--block-k", type=int, default=128) parser.add_argument("--batch-size", type=int, required=False) parser.add_argument("--save-path", type=str, default="./") args = parser.parse_args() main(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_activation.py
benchmarks/kernels/benchmark_activation.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # benchmark custom activation op performance import itertools import torch import vllm.model_executor.layers.activation # noqa F401 from vllm.model_executor.custom_op import CustomOp from vllm.platforms import current_platform from vllm.triton_utils import triton from vllm.utils.argparse_utils import FlexibleArgumentParser from vllm.utils.torch_utils import STR_DTYPE_TO_TORCH_DTYPE batch_size_range = [1, 16, 128] seq_len_range = [1, 16, 64, 1024, 4096] intermediate_size = [3072, 9728, 12288] configs = list(itertools.product(batch_size_range, seq_len_range, intermediate_size)) def benchmark_activation( batch_size: int, seq_len: int, intermediate_size: int, provider: str, func_name: str, dtype: torch.dtype, ): device = "cuda" num_tokens = batch_size * seq_len dim = intermediate_size current_platform.seed_everything(42) torch.set_default_device(device) if func_name == "gelu_and_mul": layer = CustomOp.op_registry[func_name](approximate="none") elif func_name == "gelu_and_mul_tanh": layer = CustomOp.op_registry["gelu_and_mul"](approximate="tanh") elif func_name == "fatrelu_and_mul": threshold = 0.5 layer = CustomOp.op_registry[func_name](threshold) else: layer = CustomOp.op_registry[func_name]() x = torch.randn(num_tokens, dim, dtype=dtype, device=device) compiled_layer = torch.compile(layer.forward_native) if provider == "custom": fn = lambda: layer(x) elif provider == "compiled": fn = lambda: compiled_layer(x) ms, min_ms, max_ms = triton.testing.do_bench_cudagraph( fn, quantiles=[0.5, 0.2, 0.8] ) return ms, max_ms, min_ms if __name__ == "__main__": parser = FlexibleArgumentParser(description="Benchmark the custom activation op.") parser.add_argument( "--func-name", type=str, choices=[ "mul_and_silu", "silu_and_mul", "gelu_and_mul", "gelu_and_mul_tanh", "fatrelu_and_mul", "swigluoai_and_mul", "gelu_new", "gelu_fast", "quick_gelu", ], default="silu_and_mul", ) parser.add_argument( "--dtype", type=str, choices=["half", "bfloat16", "float"], default="bfloat16" ) args = parser.parse_args() assert args func_name = args.func_name dtype = STR_DTYPE_TO_TORCH_DTYPE[args.dtype] perf_report = triton.testing.perf_report( triton.testing.Benchmark( x_names=["batch_size", "seq_len", "intermediate_size"], x_vals=configs, line_arg="provider", line_vals=["custom", "compiled"], line_names=["Custom OP", "Compiled"], styles=[("blue", "-"), ("green", "-")], ylabel="ms", plot_name=f"{func_name}-op-performance", args={}, ) ) perf_report( lambda batch_size, seq_len, intermediate_size, provider: benchmark_activation( batch_size, seq_len, intermediate_size, provider, func_name, dtype ) ).run(print_data=True)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/benchmark_fused_collective.py
benchmarks/kernels/benchmark_fused_collective.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Benchmark for FlashInfer fused collective operations vs standard operations. This benchmark compares: 1. FlashInfer's trtllm_allreduce_fusion (fused allreduce + rmsnorm + optional quant) 2. Standard tensor_model_parallel_all_reduce + separate rmsnorm/quant operations Usage with torchrun: torchrun --nproc_per_node=2 benchmark_fused_collective.py """ import argparse import itertools import os import time import pandas as pd import torch # type: ignore import torch.distributed as dist # type: ignore from vllm.config.vllm import CompilationConfig, VllmConfig, set_current_vllm_config from vllm.distributed import ( get_tp_group, tensor_model_parallel_all_reduce, ) from vllm.distributed.parallel_state import ( graph_capture, init_distributed_environment, initialize_model_parallel, ) from vllm.logger import init_logger from vllm.model_executor.layers.layernorm import RMSNorm # noqa from vllm.model_executor.layers.quantization.input_quant_fp8 import QuantFP8 # noqa from vllm.model_executor.layers.quantization.utils.quant_utils import GroupShape # noqa from vllm.platforms import current_platform # noqa RMS_NORM_OP = torch.ops._C.rms_norm FUSED_ADD_RMS_NORM_OP = torch.ops._C.fused_add_rms_norm RMS_NORM_STATIC_FP8_QUANT_OP = torch.ops._C.rms_norm_static_fp8_quant FUSED_ADD_RMS_NORM_STATIC_FP8_QUANT_OP = ( torch.ops._C.fused_add_rms_norm_static_fp8_quant ) SCALED_FP4_QUANT_OP = torch.ops._C.scaled_fp4_quant logger = init_logger(__name__) # Try to import FlashInfer try: import flashinfer.comm as flashinfer_comm # type: ignore if not hasattr(flashinfer_comm, "trtllm_allreduce_fusion"): flashinfer_comm = None logger.warning( "FlashInfer comm module found but missing trtllm_allreduce_fusion" ) except ImportError: flashinfer_comm = None logger.warning("FlashInfer not found, only benchmarking standard operations") # Constants FP8_DTYPE = current_platform.fp8_dtype() MiB = 1024 * 1024 # FlashInfer max sizes per world size # Enable 64MB for 2, 4, 8 world sizes to verify large input sizes # use --disable-oneshot to disable oneshot mode for very large input sizes _FI_MAX_SIZES = { 2: 64 * MiB, # 64MB 4: 64 * MiB, # 64MB 8: 64 * MiB, # 64MB } # Global workspace tensor for FlashInfer _FI_WORKSPACE_TENSOR = None def setup_flashinfer_workspace( world_size: int, rank: int, hidden_dim: int, max_token_num: int, use_fp32_lamport: bool = False, ): """Setup FlashInfer workspace for fused allreduce operations.""" global _FI_WORKSPACE_TENSOR if flashinfer_comm is None: return None, None if world_size not in _FI_MAX_SIZES: logger.warning("FlashInfer not supported for world size %s", world_size) return None, None try: # Create IPC workspace ipc_handles, workspace_tensor = ( flashinfer_comm.trtllm_create_ipc_workspace_for_all_reduce_fusion( tp_rank=rank, tp_size=world_size, max_token_num=max_token_num, hidden_dim=hidden_dim, group=get_tp_group().device_group, use_fp32_lamport=use_fp32_lamport, ) ) _FI_WORKSPACE_TENSOR = workspace_tensor return ipc_handles, workspace_tensor except Exception as e: logger.error("Failed to setup FlashInfer workspace: %s", e) return None, None def cleanup_flashinfer_workspace(ipc_handles): """Cleanup FlashInfer workspace.""" if flashinfer_comm is None or ipc_handles is None: return try: group = get_tp_group().device_group flashinfer_comm.trtllm_destroy_ipc_workspace_for_all_reduce(ipc_handles, group) except Exception as e: logger.error("Failed to cleanup FlashInfer workspace: %s", e) class FlashInferFusedAllReduceParams: """Parameters for FlashInfer fused allreduce operations.""" def __init__( self, rank: int, world_size: int, use_fp32_lamport: bool = False, max_token_num: int = 1024, ): self.rank = rank self.world_size = world_size self.use_fp32_lamport = use_fp32_lamport self.trigger_completion_at_end = True self.launch_with_pdl = True self.fp32_acc = True self.max_token_num = max_token_num def get_trtllm_fused_allreduce_kwargs(self): return { "world_rank": self.rank, "world_size": self.world_size, "launch_with_pdl": self.launch_with_pdl, "trigger_completion_at_end": self.trigger_completion_at_end, "fp32_acc": self.fp32_acc, } def flashinfer_fused_allreduce_rmsnorm( input_tensor: torch.Tensor, residual: torch.Tensor | None, rms_gamma: torch.Tensor, rms_eps: float, allreduce_params: "FlashInferFusedAllReduceParams", use_oneshot: bool, norm_out: torch.Tensor | None = None, ): """FlashInfer fused allreduce + rmsnorm operation.""" if flashinfer_comm is None or _FI_WORKSPACE_TENSOR is None: raise RuntimeError("FlashInfer not available or workspace not initialized") if norm_out is None: norm_out = input_tensor residual_out = residual else: residual_out = input_tensor flashinfer_comm.trtllm_allreduce_fusion( allreduce_in=input_tensor, token_num=input_tensor.shape[0], residual_in=residual, residual_out=residual_out, norm_out=norm_out, rms_gamma=rms_gamma, rms_eps=rms_eps, hidden_dim=input_tensor.shape[-1], workspace_ptrs=_FI_WORKSPACE_TENSOR, pattern_code=flashinfer_comm.AllReduceFusionPattern.kARResidualRMSNorm, allreduce_out=None, quant_out=None, scale_out=None, layout_code=flashinfer_comm.QuantizationSFLayout.SWIZZLED_128x4, scale_factor=None, use_oneshot=use_oneshot, **allreduce_params.get_trtllm_fused_allreduce_kwargs(), ) def flashinfer_fused_allreduce_rmsnorm_fp8_quant( input_tensor: torch.Tensor, residual: torch.Tensor | None, rms_gamma: torch.Tensor, rms_eps: float, scale_factor: torch.Tensor, allreduce_params: FlashInferFusedAllReduceParams, use_oneshot: bool = True, norm_out: torch.Tensor | None = None, quant_out: torch.Tensor | None = None, ): """FlashInfer fused allreduce + rmsnorm + FP8 quantization.""" if flashinfer_comm is None or _FI_WORKSPACE_TENSOR is None: raise RuntimeError("FlashInfer not available or workspace not initialized") if norm_out is None: norm_out = input_tensor residual_out = residual else: residual_out = input_tensor flashinfer_comm.trtllm_allreduce_fusion( allreduce_in=input_tensor, token_num=input_tensor.shape[0], residual_in=residual, residual_out=residual_out, norm_out=norm_out, rms_gamma=rms_gamma, rms_eps=rms_eps, hidden_dim=input_tensor.shape[-1], workspace_ptrs=_FI_WORKSPACE_TENSOR, pattern_code=flashinfer_comm.AllReduceFusionPattern.kARResidualRMSNormFP8Quant, allreduce_out=None, quant_out=quant_out, scale_out=None, layout_code=flashinfer_comm.QuantizationSFLayout.SWIZZLED_128x4, scale_factor=scale_factor, use_oneshot=use_oneshot, **allreduce_params.get_trtllm_fused_allreduce_kwargs(), ) def flashinfer_fused_allreduce_rmsnorm_fp4_quant( input_tensor: torch.Tensor, residual: torch.Tensor | None, rms_gamma: torch.Tensor, rms_eps: float, input_global_scale: torch.Tensor, allreduce_params: FlashInferFusedAllReduceParams, quant_out: torch.Tensor, use_oneshot: bool, output_scale: torch.Tensor, norm_out: torch.Tensor | None = None, ): """FlashInfer fused allreduce + rmsnorm + FP4 quantization.""" if flashinfer_comm is None or _FI_WORKSPACE_TENSOR is None: raise RuntimeError("FlashInfer not available or workspace not initialized") if norm_out is None: norm_out = input_tensor residual_out = residual else: residual_out = input_tensor flashinfer_comm.trtllm_allreduce_fusion( allreduce_in=input_tensor, token_num=input_tensor.shape[0], residual_in=residual, residual_out=residual_out, norm_out=norm_out, rms_gamma=rms_gamma, rms_eps=rms_eps, hidden_dim=input_tensor.shape[-1], workspace_ptrs=_FI_WORKSPACE_TENSOR, pattern_code=flashinfer_comm.AllReduceFusionPattern.kARResidualRMSNormFP4Quant, allreduce_out=None, quant_out=quant_out, scale_out=output_scale, layout_code=flashinfer_comm.QuantizationSFLayout.SWIZZLED_128x4, scale_factor=input_global_scale, use_oneshot=use_oneshot, **allreduce_params.get_trtllm_fused_allreduce_kwargs(), ) class VllmFusedAllreduce: def __init__(self, hidden_dim, dtype): self.rms_eps = 1e-6 self.rms_norm = RMSNorm(hidden_dim, eps=self.rms_eps, dtype=dtype) self.fp8_quant = QuantFP8( static=True, group_shape=GroupShape.PER_TENSOR, ) def allreduce_rmsnorm( self, input_tensor: torch.Tensor, residual: torch.Tensor | None ): allreduce_out = tensor_model_parallel_all_reduce(input_tensor) return self.rms_norm(allreduce_out, residual) def allreduce_rmsnorm_fp8_quant( self, input_tensor: torch.Tensor, residual: torch.Tensor | None, scale_factor: torch.Tensor, ): allreduce_out = tensor_model_parallel_all_reduce(input_tensor) rms_out = self.rms_norm(allreduce_out, residual) if residual is None: quant_out = self.fp8_quant(rms_out, scale_factor) return quant_out else: rms_out, residual_out = rms_out quant_out = self.fp8_quant(rms_out, scale_factor) return quant_out, residual_out def allreduce_rmsnorm_fp4_quant( self, input_tensor: torch.Tensor, residual: torch.Tensor | None, input_global_scale: torch.Tensor, quant_out: torch.Tensor, output_scale: torch.Tensor, ): allreduce_out = tensor_model_parallel_all_reduce(input_tensor) rms_out = self.rms_norm(allreduce_out, residual) if residual is None: SCALED_FP4_QUANT_OP(quant_out, rms_out, output_scale, input_global_scale) return quant_out, output_scale else: rms_out, residual_out = rms_out SCALED_FP4_QUANT_OP(quant_out, rms_out, output_scale, input_global_scale) return quant_out, residual_out, output_scale def create_test_tensors( num_tokens: int, hidden_dim: int, dtype: torch.dtype, use_residual: bool = True ): """Create test tensors for benchmarking.""" input_tensor = torch.randn(num_tokens, hidden_dim, dtype=dtype) residual = ( torch.randn_like(input_tensor) if use_residual else torch.zeros_like(input_tensor) ) rms_gamma = torch.ones(hidden_dim, dtype=dtype) norm_out = None if use_residual else torch.empty_like(input_tensor) # Quantization scales scale_fp8 = torch.tensor(1.0, dtype=torch.float32) scale_fp4 = torch.tensor(1.0, dtype=torch.float32) quant_out_fp8 = torch.empty_like(input_tensor, dtype=FP8_DTYPE) # Pre-allocate FP4 output tensors (to avoid allocation overhead in benchmarks) fp4_quant_out = torch.empty((num_tokens, hidden_dim // 2), dtype=torch.uint8) fp4_output_scale = torch.empty((128, 4), dtype=torch.int32) return ( input_tensor, norm_out, residual, rms_gamma, scale_fp8, quant_out_fp8, scale_fp4, fp4_quant_out, fp4_output_scale, ) def benchmark_operation( operation_func, *args, warmup: int = 5, trials: int = 20, **kwargs ): """Benchmark a single operation using CUDA graphs.""" # Warmup before graph capture for _ in range(warmup): operation_func(*args, **kwargs) torch.cuda.synchronize() # Create CUDA graph graph = torch.cuda.CUDAGraph() num_op_per_cudagraph = 10 # Use vLLM's graph_capture to make tensor_model_parallel_all_reduce graph-safe device = torch.device(f"cuda:{torch.cuda.current_device()}") with graph_capture(device=device), torch.cuda.graph(graph): for _ in range(num_op_per_cudagraph): operation_func(*args, **kwargs) # Graph warmup torch.cuda.synchronize() for _ in range(warmup): graph.replay() # Benchmark with CUDA graph torch.cuda.synchronize() start_time = time.perf_counter() for _ in range(trials // num_op_per_cudagraph): # operation_func(*args, **kwargs) graph.replay() torch.cuda.synchronize() end_time = time.perf_counter() avg_time_ms = ((end_time - start_time) / trials) * 1000 return avg_time_ms def run_benchmarks( num_tokens: int, hidden_dim: int, dtype: torch.dtype, use_residual: bool, allreduce_params: FlashInferFusedAllReduceParams | None, quant_modes: set[str], no_oneshot: bool, ): """Run all benchmarks for given configuration. Args: quant_mode: "none", "fp8_only", "fp4_only", or "all" """ ( input_tensor, norm_out, residual, rms_gamma, scale_fp8, quant_out_fp8, scale_fp4, fp4_quant_out, fp4_output_scale, ) = create_test_tensors(num_tokens, hidden_dim, dtype, use_residual) rms_eps = 1e-6 results = {} vllm_fused_allreduce = VllmFusedAllreduce(hidden_dim, dtype) use_oneshot_options = [False] if no_oneshot else [True, False] # Create RMSNorm and QuantFP8 layers once for native benchmarks if "none" in quant_modes: # Standard AllReduce + RMSNorm for custom_op in ["-rms_norm", "+rms_norm"]: with set_current_vllm_config( VllmConfig(compilation_config=CompilationConfig(custom_ops=[custom_op])) ): try: suffix = ( "_custom_rms_norm" if "+" in custom_op else "_native_rms_norm" ) time_ms = benchmark_operation( vllm_fused_allreduce.allreduce_rmsnorm, input_tensor, residual=residual, ) results[f"standard_allreduce_{suffix}"] = time_ms except Exception as e: logger.error("Standard AllReduce+RMSNorm failed: %s", e) results[f"standard_allreduce_{suffix}"] = float("inf") # Standard AllReduce + RMSNorm Native Compiled with set_current_vllm_config( VllmConfig(compilation_config=CompilationConfig(custom_ops=["-rms_norm"])) ): try: standard_allreduce_rmsnorm_native_compiled = torch.compile( vllm_fused_allreduce.allreduce_rmsnorm, fullgraph=True, dynamic=False, ) time_ms = benchmark_operation( standard_allreduce_rmsnorm_native_compiled, input_tensor, residual=residual, ) results["standard_allreduce_rmsnorm_native_compiled"] = time_ms except Exception as e: logger.error("Standard AllReduce+RMSNorm Native Compiled failed: %s", e) results["standard_allreduce_rmsnorm_native_compiled"] = float("inf") # FlashInfer Fused AllReduce + RMSNorm Oneshot/Twoshot if flashinfer_comm is not None and allreduce_params is not None: for use_oneshot in use_oneshot_options: suffix = "_oneshot" if use_oneshot else "_twoshot" try: time_ms = benchmark_operation( flashinfer_fused_allreduce_rmsnorm, input_tensor, residual=residual, norm_out=norm_out, rms_gamma=rms_gamma, rms_eps=rms_eps, allreduce_params=allreduce_params, use_oneshot=use_oneshot, ) results[f"flashinfer_fused_allreduce_rmsnorm{suffix}"] = time_ms except Exception as e: logger.error("FlashInfer Fused AllReduce+RMSNorm failed: %s", e) results[f"flashinfer_fused_allreduce_rmsnorm{suffix}"] = float( "inf" ) if "fp8" in quant_modes: # Standard AllReduce + RMSNorm + FP8 Quant for rms_norm_custom_op in ["-rms_norm", "+rms_norm"]: suffix = ( "_custom_rms_norm" if "+" in rms_norm_custom_op else "_native_rms_norm" ) for quant_fp8_custom_op in ["-quant_fp8", "+quant_fp8"]: suffix += ( "_custom_quant_fp8" if "+" in quant_fp8_custom_op else "_native_quant_fp8" ) with set_current_vllm_config( VllmConfig( compilation_config=CompilationConfig( custom_ops=[rms_norm_custom_op, quant_fp8_custom_op] ) ) ): try: time_ms = benchmark_operation( vllm_fused_allreduce.allreduce_rmsnorm_fp8_quant, input_tensor, residual=residual, scale_factor=scale_fp8, ) results[f"standard_allreduce{suffix}"] = time_ms except Exception as e: logger.error("Standard AllReduce+RMSNorm+FP8 failed: %s", e) results[f"standard_allreduce{suffix}"] = float("inf") # Standard AllReduce + RMSNorm + FP8 Quant Native Compiled with set_current_vllm_config( VllmConfig( compilation_config=CompilationConfig( custom_ops=["-rms_norm", "-quant_fp8"] ) ) ): try: standard_allreduce_rmsnorm_fp8_quant_native_compiled = torch.compile( vllm_fused_allreduce.allreduce_rmsnorm_fp8_quant, fullgraph=True, dynamic=False, ) time_ms = benchmark_operation( standard_allreduce_rmsnorm_fp8_quant_native_compiled, input_tensor, residual=residual, scale_factor=scale_fp8, ) results["standard_allreduce_rmsnorm_fp8_quant_native_compiled"] = ( time_ms ) except Exception as e: logger.error( "Standard AllReduce+RMSNorm+FP8 Native Compiled failed: %s", e ) results["standard_allreduce_rmsnorm_fp8_quant_native_compiled"] = float( "inf" ) # FlashInfer Fused AllReduce + RMSNorm + FP8 Quant Oneshot if flashinfer_comm is not None and allreduce_params is not None: for use_oneshot in use_oneshot_options: suffix = "_oneshot" if use_oneshot else "_twoshot" try: time_ms = benchmark_operation( flashinfer_fused_allreduce_rmsnorm_fp8_quant, input_tensor, norm_out=norm_out, residual=residual, rms_gamma=rms_gamma, rms_eps=rms_eps, scale_factor=scale_fp8, quant_out=quant_out_fp8, allreduce_params=allreduce_params, use_oneshot=use_oneshot, ) results[f"flashinfer_fused_allreduce_rmsnorm_fp8_quant{suffix}"] = ( time_ms ) except Exception as e: logger.error( "FlashInfer Fused AllReduce+RMSNorm+FP8 Oneshot failed: %s", e, ) results[f"flashinfer_fused_allreduce_rmsnorm_fp8_quant{suffix}"] = ( float("inf") ) if "fp4" in quant_modes and current_platform.has_device_capability(100): # Standard AllReduce + RMSNorm + FP4 Quant for rms_norm_custom_op in ["-rms_norm", "+rms_norm"]: suffix = ( "_custom_rms_norm" if "+" in rms_norm_custom_op else "_native_rms_norm" ) with set_current_vllm_config( VllmConfig( compilation_config=CompilationConfig( custom_ops=[rms_norm_custom_op] ) ) ): try: time_ms = benchmark_operation( vllm_fused_allreduce.allreduce_rmsnorm_fp4_quant, input_tensor, residual=residual, input_global_scale=scale_fp4, quant_out=fp4_quant_out, output_scale=fp4_output_scale, ) results[f"standard_allreduce_{suffix}_fp4_quant"] = time_ms except Exception as e: logger.error("Standard AllReduce+RMSNorm+FP4 failed: %s", e) results[f"standard_allreduce_{suffix}_fp4_quant"] = float("inf") # Standard AllReduce + RMSNorm + FP4 Quant Native Compiled with set_current_vllm_config( VllmConfig(compilation_config=CompilationConfig(custom_ops=["-rms_norm"])) ): try: standard_allreduce_rmsnorm_fp4_quant_native_compiled = torch.compile( vllm_fused_allreduce.allreduce_rmsnorm_fp4_quant, fullgraph=True, dynamic=False, ) time_ms = benchmark_operation( standard_allreduce_rmsnorm_fp4_quant_native_compiled, input_tensor, residual=residual, quant_out=fp4_quant_out, input_global_scale=scale_fp4, output_scale=fp4_output_scale, ) results["standard_allreduce_rmsnorm_fp4_quant_native_compiled"] = ( time_ms ) except Exception as e: logger.error( "Standard AllReduce+RMSNorm+FP4 Native Compiled failed: %s", e ) results["standard_allreduce_rmsnorm_fp4_quant_native_compiled"] = float( "inf" ) # FlashInfer Fused AllReduce + RMSNorm + FP4 Quant Oneshot if flashinfer_comm is not None and allreduce_params is not None: for use_oneshot in use_oneshot_options: suffix = "_oneshot" if use_oneshot else "_twoshot" try: time_ms = benchmark_operation( flashinfer_fused_allreduce_rmsnorm_fp4_quant, input_tensor, residual=residual, norm_out=norm_out, rms_gamma=rms_gamma, rms_eps=rms_eps, input_global_scale=scale_fp4, allreduce_params=allreduce_params, quant_out=fp4_quant_out, output_scale=fp4_output_scale, use_oneshot=use_oneshot, ) results[f"flashinfer_fused_allreduce_rmsnorm_fp4_quant{suffix}"] = ( time_ms ) except Exception as e: logger.error( "FlashInfer Fused AllReduce+RMSNorm+FP4 Oneshot failed: %s", e, ) results[f"flashinfer_fused_allreduce_rmsnorm_fp4_quant{suffix}"] = ( float("inf") ) # FlashInfer Fused AllReduce + RMSNorm + FP4 Quant Two-shot if flashinfer_comm is not None and allreduce_params is not None: try: time_ms = benchmark_operation( flashinfer_fused_allreduce_rmsnorm_fp4_quant, input_tensor, residual=residual, norm_out=norm_out, rms_gamma=rms_gamma, rms_eps=rms_eps, input_global_scale=scale_fp4, allreduce_params=allreduce_params, quant_out=fp4_quant_out, output_scale=fp4_output_scale, use_oneshot=False, ) results["flashinfer_fused_allreduce_rmsnorm_fp4_quant_twoshot"] = ( time_ms ) except Exception as e: logger.error( "FlashInfer Fused AllReduce+RMSNorm+FP4 Two-shot failed: %s", e, ) results["flashinfer_fused_allreduce_rmsnorm_fp4_quant_twoshot"] = float( "inf" ) return results def prepare_results_with_speedups(results_dict): """Prepare results with speedup calculations based on dynamic baseline selection.""" prepared_results = [] # Determine the fastest baseline for each operation type def get_fastest_baseline(op_name, results_dict): """Get the fastest baseline between standard and native_compiled versions.""" if "fp8_quant" in op_name: candidates = [ "standard_allreduce_rmsnorm_fp8_quant", "standard_allreduce_rmsnorm_fp8_quant_native_compiled", ] elif "fp4_quant" in op_name: candidates = [ "standard_allreduce_rmsnorm_fp4_quant", "standard_allreduce_rmsnorm_fp4_quant_native_compiled", ] else: candidates = [ "standard_allreduce_rmsnorm", "standard_allreduce_rmsnorm_native_compiled", ] # Find the fastest among available candidates fastest_time = float("inf") fastest_baseline = None for candidate in candidates: if ( candidate in results_dict and results_dict[candidate] != float("inf") and results_dict[candidate] < fastest_time ): fastest_time = results_dict[candidate] fastest_baseline = candidate return fastest_baseline # Create dynamic baseline mapping dynamic_baseline_mapping = {} for op_name in results_dict: if ( op_name.startswith("flashinfer_") or op_name.startswith("standard_") and not op_name.endswith("_native_compiled") ): dynamic_baseline_mapping[op_name] = get_fastest_baseline( op_name, results_dict ) for op_name, time_ms in results_dict.items(): if time_ms == float("inf"): speedup_str = "FAILED" time_str = "FAILED" else: time_str = f"{time_ms:.3f}" # Find the appropriate baseline for this operation baseline_op = dynamic_baseline_mapping.get(op_name) if baseline_op and baseline_op in results_dict: baseline_time = results_dict[baseline_op] if baseline_time != float("inf") and baseline_time > 0: speedup = baseline_time / time_ms speedup_str = f"{speedup:.2f}x" else: speedup_str = "N/A" else: # For baseline operations, determine if this is the fastest baseline if op_name.endswith("_native_compiled") or ( op_name.startswith("standard_") and not op_name.endswith("_native_compiled") ): fastest_baseline = get_fastest_baseline(op_name, results_dict) if fastest_baseline == op_name: speedup_str = "baseline" else: if fastest_baseline and fastest_baseline in results_dict: baseline_time = results_dict[fastest_baseline] if baseline_time != float("inf") and baseline_time > 0: speedup = baseline_time / time_ms speedup_str = f"{speedup:.2f}x" else: speedup_str = "N/A" else: speedup_str = "N/A" else: speedup_str = "N/A" prepared_results.append( { "operation": op_name, "time_ms": time_ms, "time_str": time_str, "speedup_str": speedup_str, } ) return prepared_results def print_results( results_dict, num_tokens, hidden_dim, dtype, use_residual, quant_modes, input_size_mb, ): """Print benchmark results in a formatted table.""" print(f"\n{'=' * 80}") print( f"Results: num_tokens={num_tokens}, hidden_dim={hidden_dim} " f"(input size: {input_size_mb:.2f} MB)" ) print( f"dtype={dtype}, residual={'yes' if use_residual else 'no'}, " f"quant_modes={','.join(sorted(list(quant_modes)))}" ) print(f"{'=' * 80}") print(f"{'Operation':<50} {'Time (ms)':<12} {'Speedup':<10}") print(f"{'-' * 80}") # Prepare results with speedup calculations prepared_results = prepare_results_with_speedups(results_dict) for result in prepared_results: if result["time_ms"] == float("inf"): time_display = result["time_str"] else: time_display = f"{result['time_ms']:.3f}" print( f"{result['operation']:<50} {time_display:<12} {result['speedup_str']:<10}" ) def format_results_markdown( all_results: list[dict], world_size: int, args: argparse.Namespace ) -> str: """Format all benchmark results as markdown.""" lines: list[str] = [] lines.append("# FlashInfer Fused Collective Operations Benchmark Results") lines.append("") lines.append(f"**World Size:** {world_size} ") lines.append(f"**Hidden Dimension:** {args.hidden_dim} ") lines.append(f"**Warmup Iterations:** {args.warmup} ") lines.append(f"**Benchmark Trials:** {args.trials} ") modes = ",".join(all_results[0]["quant_modes"]) if all_results else "N/A" lines.append(f"**Quantization Modes:** {modes} ") lines.append("") lines.append("---") lines.append("") for entry in all_results: num_tokens = entry["num_tokens"] dtype = entry["dtype"] use_residual = entry["use_residual"] results_dict = entry["results"] input_size_mb = entry["input_size_mb"] residual_str = "with residual" if use_residual else "no residual" lines.append( f"## Configuration: num_tokens={num_tokens}, dtype={dtype}, {residual_str}" ) lines.append(f"**Input Size:** {input_size_mb:.2f} MB") lines.append("") prepared = prepare_results_with_speedups(results_dict) # Build DataFrame for markdown export rows = [ { "Operation": r["operation"].replace("_", " ").title(), "Time (ms)": r["time_str"], "Speedup": r["speedup_str"], } for r in prepared ] df = pd.DataFrame(rows) if df.empty: lines.append("No results.") else: lines.append(df.to_markdown(index=False)) lines.append("")
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
true
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/benchmarks/kernels/deepgemm/benchmark_fp8_block_dense_gemm.py
benchmarks/kernels/deepgemm/benchmark_fp8_block_dense_gemm.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # ruff: noqa: E501 import time import torch from vllm import _custom_ops as ops from vllm.model_executor.layers.quantization.utils.fp8_utils import ( per_token_group_quant_fp8, w8a8_triton_block_scaled_mm, ) from vllm.triton_utils import triton from vllm.utils.deep_gemm import ( calc_diff, fp8_gemm_nt, get_col_major_tma_aligned_tensor, per_block_cast_to_fp8, ) def benchmark_shape( m: int, n: int, k: int, warmup: int = 100, repeat: int = 10000, verbose: bool = False, ) -> dict: """Benchmark all implementations for a specific (m, n, k) shape.""" if verbose: print(f"\n=== Benchmarking shape: m={m}, n={n}, k={k} ===") # Create test tensors A = torch.randn((m, k), device="cuda", dtype=torch.bfloat16) B = torch.randn((n, k), device="cuda", dtype=torch.bfloat16) # Reference result in BF16 torch.cuda.synchronize() C_ref = A @ B.t() # Pre-quantize B for all implementations # (weights can be pre-quantized offline) B_deepgemm, B_scale_deepgemm = per_block_cast_to_fp8(B, [128, 128], use_ue8m0=True) B_vllm, B_scale_vllm = per_block_cast_to_fp8(B, [128, 128], use_ue8m0=True) # Block size configuration block_size = [128, 128] # Pre-quantize A for all implementations A_deepgemm, A_scale_deepgemm = per_token_group_quant_fp8(A, block_size[1]) A_scale_deepgemm = get_col_major_tma_aligned_tensor(A_scale_deepgemm) C_deepgemm = torch.empty((m, n), device="cuda", dtype=torch.bfloat16) A_vllm, A_scale_vllm = per_token_group_quant_fp8(A, block_size[1]) A_vllm_cutlass, A_scale_vllm_cutlass = per_token_group_quant_fp8( A, block_size[1], column_major_scales=True ) # === DeepGEMM Implementation === def deepgemm_gemm(): fp8_gemm_nt( (A_deepgemm, A_scale_deepgemm), (B_deepgemm, B_scale_deepgemm), C_deepgemm ) return C_deepgemm # === vLLM Triton Implementation === def vllm_triton_gemm(): return w8a8_triton_block_scaled_mm( A_vllm, B_vllm, A_scale_vllm, B_scale_vllm, block_size, output_dtype=torch.bfloat16, ) # === vLLM CUTLASS Implementation === def vllm_cutlass_gemm(): return ops.cutlass_scaled_mm( A_vllm_cutlass, B_vllm.T, scale_a=A_scale_vllm_cutlass, scale_b=B_scale_vllm.T, out_dtype=torch.bfloat16, ) # Run correctness check first if verbose: print("Running correctness check...") C_deepgemm = deepgemm_gemm() C_vllm_triton = vllm_triton_gemm() C_vllm_cutlass = vllm_cutlass_gemm() deepgemm_diff = calc_diff(C_deepgemm, C_ref) vllm_triton_diff = calc_diff(C_vllm_triton, C_ref) vllm_cutlass_diff = calc_diff(C_vllm_cutlass, C_ref) if verbose: print(f"DeepGEMM vs Reference difference: {deepgemm_diff:.6f}") print(f"vLLM Triton vs Reference difference: {vllm_triton_diff:.6f}") print(f"vLLM CUTLASS vs Reference difference: {vllm_cutlass_diff:.6f}") print( "vLLM Triton vs DeepGEMM difference: " f"{calc_diff(C_vllm_triton, C_deepgemm):.6f}" ) print( "vLLM CUTLASS vs DeepGEMM difference: " f"{calc_diff(C_vllm_cutlass, C_deepgemm):.6f}" ) # Benchmark implementations implementations = { "DeepGEMM": deepgemm_gemm, "vLLM Triton": vllm_triton_gemm, "vLLM CUTLASS": vllm_cutlass_gemm, } benchmark_results = {"shape": {"m": m, "n": n, "k": k}, "implementations": {}} for name, func in implementations.items(): # Warmup for _ in range(warmup): func() torch.cuda.synchronize() # Timing loop torch.cuda.synchronize() start = time.time() for _ in range(repeat): func() torch.cuda.synchronize() end = time.time() # Calculate timing and TFLOPS avg_time_ms = (end - start) / repeat * 1000 avg_time_us = avg_time_ms * 1000 tflops = 2 * m * n * k / (avg_time_ms * 1e-3) / 1e12 gb_s = (m * k + k * n + m * n * 2) / 1e9 / (avg_time_ms * 1e-3) benchmark_results["implementations"][name] = { "time_ms": avg_time_ms, "time_us": avg_time_us, "tflops": tflops, "gb_s": gb_s, "diff": { "DeepGEMM": 0.0 if name == "DeepGEMM" else calc_diff(func(), C_deepgemm), "Reference": deepgemm_diff if name == "DeepGEMM" else (vllm_triton_diff if name == "vLLM Triton" else vllm_cutlass_diff), }, } if verbose: print(f"{name}: {avg_time_ms:.3f} ms, {tflops:.2f} TFLOPS, {gb_s:.2f} GB/s") # Calculate speedups baseline = benchmark_results["implementations"]["DeepGEMM"]["time_ms"] for name, data in benchmark_results["implementations"].items(): if name != "DeepGEMM": speedup = baseline / data["time_ms"] benchmark_results["implementations"][name]["speedup_vs_deepgemm"] = speedup if verbose: print( f"DeepGEMM is {1 / speedup:.2f}x " f"{'faster' if 1 / speedup > 1 else 'slower'} than {name}" ) vllm_triton_time = benchmark_results["implementations"]["vLLM Triton"]["time_ms"] vllm_cutlass_time = benchmark_results["implementations"]["vLLM CUTLASS"]["time_ms"] cutlass_vs_triton = vllm_triton_time / vllm_cutlass_time benchmark_results["implementations"]["vLLM CUTLASS"]["speedup_vs_triton"] = ( cutlass_vs_triton ) if verbose: print( f"vLLM CUTLASS is {cutlass_vs_triton:.2f}x " f"{'faster' if cutlass_vs_triton > 1 else 'slower'} than vLLM Triton" ) return benchmark_results def format_table_row(values, widths): """Format a row with specified column widths.""" return "| " + " | ".join(f"{val:{w}}" for val, w in zip(values, widths)) + " |" def print_table(headers, rows, title=None): """Print a table with headers and rows.""" if title: print(f"\n{title}") # Calculate column widths based on headers and data widths = [ max(len(str(h)), max(len(str(row[i])) for row in rows)) for i, h in enumerate(headers) ] # Create separator line separator = "+-" + "-+-".join("-" * w for w in widths) + "-+" # Print table print(separator) print(format_table_row(headers, widths)) print(separator) for row in rows: print(format_table_row(row, widths)) print(separator) def format_speedup(value): """Format speedup value with indicator if it's faster or slower.""" return f"{value:.2f}x {'faster' if value > 1.0 else 'slower'}" def run_benchmarks(verbose: bool = False): """Run benchmarks for a set of common shapes.""" print("===== STARTING FP8 GEMM BENCHMARK =====") # Make sure we're using the GPU if not torch.cuda.is_available(): print("CUDA not available! Tests require GPU.") return # Print system information print(f"PyTorch version: {torch.__version__}") print(f"CUDA version: {torch.version.cuda}") print(f"Triton version: {triton.__version__}") print(f"Using device: {torch.cuda.get_device_name()}") # Enable TF32 for better performance torch.backends.cuda.matmul.allow_tf32 = True torch.backends.cudnn.allow_tf32 = True # Set seeds for reproducibility torch.manual_seed(42) torch.cuda.manual_seed(42) # Define benchmark shapes (m, n, k) shapes = [ (8, 4096, 7168), (8, 7168, 18432), (8, 18432, 7168), (64, 4096, 7168), (64, 7168, 18432), (64, 18432, 7168), (64, 24576, 1536), (64, 32768, 512), (64, 7168, 16384), (128, 4096, 7168), (128, 7168, 18432), (128, 18432, 7168), (1024, 4096, 7168), (1024, 18432, 7168), (2048, 4096, 7168), (4096, 4096, 7168), ] shapes = [ # (64, 2112, 7168), (64, 24576, 1536), (64, 32768, 512), (64, 7168, 16384), (64, 4096, 7168), (64, 7168, 2048), # (128, 2112, 7168), (128, 24576, 1536), (128, 32768, 512), (128, 7168, 16384), (128, 4096, 7168), (128, 7168, 2048), # (4096, 2112, 7168), (4096, 24576, 1536), (4096, 32768, 512), (4096, 7168, 16384), (4096, 4096, 7168), (4096, 7168, 2048), ] all_results = [] for m, n, k in shapes: result = benchmark_shape(m, n, k, verbose=verbose) all_results.append(result) # Print results in a nicely formatted table print("\n===== PERFORMANCE COMPARISON =====") # Print DeepGEMM table deepgemm_headers = ["m", "n", "k", "Time (μs)", "TFLOPS", "GB/s"] deepgemm_rows = [] for result in all_results: shape = result["shape"] impl_data = result["implementations"]["DeepGEMM"] deepgemm_rows.append( [ shape["m"], shape["n"], shape["k"], f"{impl_data['time_us']:.1f}", f"{impl_data['tflops']:.1f}", f"{impl_data['gb_s']:.1f}", ] ) print_table(deepgemm_headers, deepgemm_rows, title="DeepGEMM Implementation:") # Print vLLM Triton table triton_headers = ["m", "n", "k", "Time (μs)", "TFLOPS", "GB/s", "vs DeepGEMM"] triton_rows = [] for result in all_results: shape = result["shape"] impl_data = result["implementations"]["vLLM Triton"] speedup = impl_data.get("speedup_vs_deepgemm", 1.0) triton_rows.append( [ shape["m"], shape["n"], shape["k"], f"{impl_data['time_us']:.1f}", f"{impl_data['tflops']:.1f}", f"{impl_data['gb_s']:.1f}", format_speedup(speedup), ] ) print_table(triton_headers, triton_rows, title="vLLM Triton Implementation:") # Print vLLM CUTLASS table cutlass_headers = [ "m", "n", "k", "Time (μs)", "TFLOPS", "GB/s", "vs DeepGEMM", "vs Triton", ] cutlass_rows = [] for result in all_results: shape = result["shape"] impl_data = result["implementations"]["vLLM CUTLASS"] vs_deepgemm = impl_data.get("speedup_vs_deepgemm", 1.0) vs_triton = impl_data.get("speedup_vs_triton", 1.0) cutlass_rows.append( [ shape["m"], shape["n"], shape["k"], f"{impl_data['time_us']:.1f}", f"{impl_data['tflops']:.1f}", f"{impl_data['gb_s']:.1f}", format_speedup(vs_deepgemm), format_speedup(vs_triton), ] ) print_table(cutlass_headers, cutlass_rows, title="vLLM CUTLASS Implementation:") # Calculate and print averages print("\n===== AVERAGE PERFORMANCE =====") implementations = ["DeepGEMM", "vLLM Triton", "vLLM CUTLASS"] avg_metrics = { impl: {"tflops": 0, "gb_s": 0, "time_ms": 0} for impl in implementations } for result in all_results: for impl in implementations: impl_data = result["implementations"][impl] avg_metrics[impl]["tflops"] += impl_data["tflops"] avg_metrics[impl]["gb_s"] += impl_data["gb_s"] avg_metrics[impl]["time_ms"] += impl_data["time_ms"] num_shapes = len(all_results) avg_headers = ["Implementation", "Avg TFLOPS", "Avg GB/s", "Avg Time (ms)"] avg_rows = [] for impl in implementations: avg_tflops = avg_metrics[impl]["tflops"] / num_shapes avg_mem_bw = avg_metrics[impl]["gb_s"] / num_shapes avg_time = avg_metrics[impl]["time_ms"] / num_shapes avg_rows.append( [impl, f"{avg_tflops:.2f}", f"{avg_mem_bw:.2f}", f"{avg_time:.2f}"] ) print_table(avg_headers, avg_rows) # Calculate average speedups avg_speedups = { "DeepGEMM vs vLLM Triton": 0, "DeepGEMM vs vLLM CUTLASS": 0, "vLLM CUTLASS vs vLLM Triton": 0, } for result in all_results: deepgemm_time = result["implementations"]["DeepGEMM"]["time_ms"] vllm_triton_time = result["implementations"]["vLLM Triton"]["time_ms"] vllm_cutlass_time = result["implementations"]["vLLM CUTLASS"]["time_ms"] avg_speedups["DeepGEMM vs vLLM Triton"] += vllm_triton_time / deepgemm_time avg_speedups["DeepGEMM vs vLLM CUTLASS"] += vllm_cutlass_time / deepgemm_time avg_speedups["vLLM CUTLASS vs vLLM Triton"] += ( vllm_triton_time / vllm_cutlass_time ) print("\n===== AVERAGE SPEEDUPS =====") speedup_headers = ["Comparison", "Speedup"] speedup_rows = [] for comparison, total in avg_speedups.items(): avg_speedup = total / num_shapes status = "faster" if avg_speedup > 1 else "slower" speedup_rows.append([comparison, f"{avg_speedup:.2f}x {status}"]) print_table(speedup_headers, speedup_rows) # Average accuracy comparison print("\n===== ACCURACY COMPARISON =====") avg_diff = {impl: 0 for impl in implementations} for result in all_results: for impl in implementations: avg_diff[impl] += result["implementations"][impl]["diff"]["Reference"] diff_headers = ["Implementation", "Avg Diff vs Reference"] diff_rows = [] for impl in implementations: diff_rows.append([impl, f"{avg_diff[impl] / num_shapes:.6f}"]) print_table(diff_headers, diff_rows) if __name__ == "__main__": run_benchmarks(verbose=False)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/install_nixl_from_source_ubuntu.py
tools/install_nixl_from_source_ubuntu.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # install_prerequisites.py import argparse import glob import json import os import subprocess import sys import urllib.request # --- Configuration --- WHEELS_CACHE_HOME = os.environ.get("WHEELS_CACHE_HOME", "/tmp/wheels_cache") ROOT_DIR = os.path.dirname(os.path.abspath(__file__)) UCX_DIR = os.path.join("/tmp", "ucx_source") NIXL_DIR = os.path.join("/tmp", "nixl_source") UCX_INSTALL_DIR = os.path.join("/tmp", "ucx_install") UCX_REPO_URL = "https://github.com/openucx/ucx.git" NIXL_REPO_URL = "https://github.com/ai-dynamo/nixl.git" # --- Helper Functions --- def get_latest_nixl_version(): """Helper function to get latest release version of NIXL""" try: nixl_release_url = "https://api.github.com/repos/ai-dynamo/nixl/releases/latest" with urllib.request.urlopen(nixl_release_url) as response: data = json.load(response) return data.get("tag_name", "0.7.0") except Exception: return "0.7.0" NIXL_VERSION = os.environ.get("NIXL_VERSION", get_latest_nixl_version()) def run_command(command, cwd=".", env=None): """Helper function to run a shell command and check for errors.""" print(f"--> Running command: {' '.join(command)} in '{cwd}'", flush=True) subprocess.check_call(command, cwd=cwd, env=env) def is_pip_package_installed(package_name): """Checks if a package is installed via pip without raising an exception.""" result = subprocess.run( [sys.executable, "-m", "pip", "show", package_name], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, ) return result.returncode == 0 def find_nixl_wheel_in_cache(cache_dir): """Finds a nixl wheel file in the specified cache directory.""" # The repaired wheel will have a 'manylinux' tag, but this glob still works. search_pattern = os.path.join(cache_dir, f"nixl*{NIXL_VERSION}*.whl") wheels = glob.glob(search_pattern) if wheels: # Sort to get the most recent/highest version if multiple exist wheels.sort() return wheels[-1] return None def install_system_dependencies(): """Installs required system packages using apt-get if run as root.""" if os.geteuid() != 0: print("\n---", flush=True) print( "WARNING: Not running as root. \ Skipping system dependency installation.", flush=True, ) print( "Please ensure the listed packages are installed on your system:", flush=True, ) print( " patchelf build-essential git cmake ninja-build \ autotools-dev automake meson libtool libtool-bin", flush=True, ) print("---\n", flush=True) return print("--- Running as root. Installing system dependencies... ---", flush=True) apt_packages = [ "patchelf", # <-- Add patchelf here "build-essential", "git", "cmake", "ninja-build", "autotools-dev", "automake", "meson", "libtool", "libtool-bin", "pkg-config", ] run_command(["apt-get", "update"]) run_command(["apt-get", "install", "-y"] + apt_packages) print("--- System dependencies installed successfully. ---\n", flush=True) def build_and_install_prerequisites(args): """Builds UCX and NIXL from source, creating a self-contained wheel.""" if not args.force_reinstall and is_pip_package_installed("nixl"): print("--> NIXL is already installed. Nothing to do.", flush=True) return cached_wheel = find_nixl_wheel_in_cache(WHEELS_CACHE_HOME) if not args.force_reinstall and cached_wheel: print( f"\n--> Found self-contained wheel: \ {os.path.basename(cached_wheel)}.", flush=True, ) print("--> Installing from cache, skipping all source builds.", flush=True) install_command = [sys.executable, "-m", "pip", "install", cached_wheel] run_command(install_command) print("\n--- Installation from cache complete. ---", flush=True) return print( "\n--> No installed package or cached wheel found. \ Starting full build process...", flush=True, ) print("\n--> Installing auditwheel...", flush=True) run_command([sys.executable, "-m", "pip", "install", "auditwheel"]) install_system_dependencies() ucx_install_path = os.path.abspath(UCX_INSTALL_DIR) print(f"--> Using wheel cache directory: {WHEELS_CACHE_HOME}", flush=True) os.makedirs(WHEELS_CACHE_HOME, exist_ok=True) # -- Step 1: Build UCX from source -- print("\n[1/3] Configuring and building UCX from source...", flush=True) if not os.path.exists(UCX_DIR): run_command(["git", "clone", UCX_REPO_URL, UCX_DIR]) ucx_source_path = os.path.abspath(UCX_DIR) run_command(["git", "checkout", "v1.19.x"], cwd=ucx_source_path) run_command(["./autogen.sh"], cwd=ucx_source_path) configure_command = [ "./configure", f"--prefix={ucx_install_path}", "--enable-shared", "--disable-static", "--disable-doxygen-doc", "--enable-optimizations", "--enable-cma", "--enable-devel-headers", "--with-verbs", "--enable-mt", "--with-ze=no", ] run_command(configure_command, cwd=ucx_source_path) run_command(["make", "-j", str(os.cpu_count() or 1)], cwd=ucx_source_path) run_command(["make", "install"], cwd=ucx_source_path) print("--- UCX build and install complete ---", flush=True) # -- Step 2: Build NIXL wheel from source -- print("\n[2/3] Building NIXL wheel from source...", flush=True) if not os.path.exists(NIXL_DIR): run_command(["git", "clone", NIXL_REPO_URL, NIXL_DIR]) else: run_command(["git", "fetch", "--tags"], cwd=NIXL_DIR) run_command(["git", "checkout", NIXL_VERSION], cwd=NIXL_DIR) print(f"--> Checked out NIXL version: {NIXL_VERSION}", flush=True) build_env = os.environ.copy() build_env["PKG_CONFIG_PATH"] = os.path.join(ucx_install_path, "lib", "pkgconfig") ucx_lib_path = os.path.join(ucx_install_path, "lib") ucx_plugin_path = os.path.join(ucx_lib_path, "ucx") existing_ld_path = os.environ.get("LD_LIBRARY_PATH", "") build_env["LD_LIBRARY_PATH"] = ( f"{ucx_lib_path}:{ucx_plugin_path}:{existing_ld_path}".strip(":") ) build_env["LDFLAGS"] = "-Wl,-rpath,$ORIGIN" print(f"--> Using LD_LIBRARY_PATH: {build_env['LD_LIBRARY_PATH']}", flush=True) temp_wheel_dir = os.path.join(ROOT_DIR, "temp_wheelhouse") run_command( [ sys.executable, "-m", "pip", "wheel", ".", "--no-deps", f"--wheel-dir={temp_wheel_dir}", ], cwd=os.path.abspath(NIXL_DIR), env=build_env, ) # -- Step 3: Repair the wheel by copying UCX libraries -- print("\n[3/3] Repairing NIXL wheel to include UCX libraries...", flush=True) unrepaired_wheel = find_nixl_wheel_in_cache(temp_wheel_dir) if not unrepaired_wheel: raise RuntimeError("Failed to find the NIXL wheel after building it.") # We tell auditwheel to ignore the plugin that mesonpy already handled. auditwheel_command = [ "auditwheel", "repair", "--exclude", "libplugin_UCX.so", # <-- Exclude because mesonpy already includes it unrepaired_wheel, f"--wheel-dir={WHEELS_CACHE_HOME}", ] run_command(auditwheel_command, env=build_env) # --- CLEANUP --- # No more temporary files to remove, just the temp wheelhouse run_command(["rm", "-rf", temp_wheel_dir]) # --- END CLEANUP --- newly_built_wheel = find_nixl_wheel_in_cache(WHEELS_CACHE_HOME) if not newly_built_wheel: raise RuntimeError("Failed to find the repaired NIXL wheel.") print( f"--> Successfully built self-contained wheel: \ {os.path.basename(newly_built_wheel)}. Now installing...", flush=True, ) install_command = [ sys.executable, "-m", "pip", "install", "--no-deps", # w/o "no-deps", it will install cuda-torch newly_built_wheel, ] if args.force_reinstall: install_command.insert(-1, "--force-reinstall") run_command(install_command) print("--- NIXL installation complete ---", flush=True) if __name__ == "__main__": parser = argparse.ArgumentParser( description="Build and install UCX and NIXL dependencies." ) parser.add_argument( "--force-reinstall", action="store_true", help="Force rebuild and reinstall of UCX and NIXL \ even if they are already installed.", ) args = parser.parse_args() build_and_install_prerequisites(args)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/report_build_time_ninja.py
tools/report_build_time_ninja.py
#!/usr/bin/env python3 # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project # Copyright (c) 2018 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. # Modified version of: https://chromium.googlesource.com/chromium/tools/depot_tools.git/+/refs/heads/main/post_build_ninja_summary.py """Summarize the last ninja build, invoked with ninja's -C syntax. > python3 tools/report_build_time_ninja.py -C build/.. Typical output looks like this: ``` Longest build steps for .cpp.o: 1.0 weighted s to build ...torch_bindings.cpp.o (12.4 s elapsed time) 2.0 weighted s to build ..._attn_c.dir/csrc... (23.5 s elapsed time) 2.6 weighted s to build ...torch_bindings.cpp.o (31.5 s elapsed time) 3.2 weighted s to build ...torch_bindings.cpp.o (38.5 s elapsed time) Longest build steps for .so (linking): 0.1 weighted s to build _moe_C.abi3.so (1.0 s elapsed time) 0.5 weighted s to build ...flash_attn_c.abi3.so (1.1 s elapsed time) 6.2 weighted s to build _C.abi3.so (6.2 s elapsed time) Longest build steps for .cu.o: 15.3 weighted s to build ...machete_mm_... (183.5 s elapsed time) 15.3 weighted s to build ...machete_mm_... (183.5 s elapsed time) 15.3 weighted s to build ...machete_mm_... (183.6 s elapsed time) 15.3 weighted s to build ...machete_mm_... (183.7 s elapsed time) 15.5 weighted s to build ...machete_mm_... (185.6 s elapsed time) 15.5 weighted s to build ...machete_mm_... (185.9 s elapsed time) 15.5 weighted s to build ...machete_mm_... (186.2 s elapsed time) 37.4 weighted s to build ...scaled_mm_c3x.cu... (449.0 s elapsed time) 43.9 weighted s to build ...scaled_mm_c2x.cu... (527.4 s elapsed time) 344.8 weighted s to build ...attention_...cu.o (1087.2 s elapsed time) 1110.0 s weighted time (10120.4 s elapsed time sum, 9.1x parallelism) 134 build steps completed, average of 0.12/s ``` """ import argparse import errno import fnmatch import os import sys from collections import defaultdict # The number of long build times to report: long_count = 10 # The number of long times by extension to report long_ext_count = 10 class Target: """Represents a single line read for a .ninja_log file.""" def __init__(self, start, end): """Creates a target object by passing in the start/end times in seconds as a float.""" self.start = start self.end = end # A list of targets, appended to by the owner of this object. self.targets = [] self.weighted_duration = 0.0 def Duration(self): """Returns the task duration in seconds as a float.""" return self.end - self.start def SetWeightedDuration(self, weighted_duration): """Sets the duration, in seconds, passed in as a float.""" self.weighted_duration = weighted_duration def WeightedDuration(self): """Returns the task's weighted duration in seconds as a float. Weighted_duration takes the elapsed time of the task and divides it by how many other tasks were running at the same time. Thus, it represents the approximate impact of this task on the total build time, with serialized or serializing steps typically ending up with much longer weighted durations. weighted_duration should always be the same or shorter than duration. """ # Allow for modest floating-point errors epsilon = 0.000002 if self.weighted_duration > self.Duration() + epsilon: print("{} > {}?".format(self.weighted_duration, self.Duration())) assert self.weighted_duration <= self.Duration() + epsilon return self.weighted_duration def DescribeTargets(self): """Returns a printable string that summarizes the targets.""" # Some build steps generate dozens of outputs - handle them sanely. # The max_length was chosen so that it can fit most of the long # single-target names, while minimizing word wrapping. result = ", ".join(self.targets) max_length = 65 if len(result) > max_length: result = result[:max_length] + "..." return result # Copied with some modifications from ninjatracing def ReadTargets(log, show_all): """Reads all targets from .ninja_log file |log_file|, sorted by duration. The result is a list of Target objects.""" header = log.readline() assert header == "# ninja log v5\n", "unrecognized ninja log version {!r}".format( header ) targets_dict = {} last_end_seen = 0.0 for line in log: parts = line.strip().split("\t") if len(parts) != 5: # If ninja.exe is rudely halted then the .ninja_log file may be # corrupt. Silently continue. continue start, end, _, name, cmdhash = parts # Ignore restart. # Convert from integral milliseconds to float seconds. start = int(start) / 1000.0 end = int(end) / 1000.0 if not show_all and end < last_end_seen: # An earlier time stamp means that this step is the first in a new # build, possibly an incremental build. Throw away the previous # data so that this new build will be displayed independently. # This has to be done by comparing end times because records are # written to the .ninja_log file when commands complete, so end # times are guaranteed to be in order, but start times are not. targets_dict = {} target = None if cmdhash in targets_dict: target = targets_dict[cmdhash] if not show_all and (target.start != start or target.end != end): # If several builds in a row just run one or two build steps # then the end times may not go backwards so the last build may # not be detected as such. However in many cases there will be a # build step repeated in the two builds and the changed # start/stop points for that command, identified by the hash, # can be used to detect and reset the target dictionary. targets_dict = {} target = None if not target: targets_dict[cmdhash] = target = Target(start, end) last_end_seen = end target.targets.append(name) return list(targets_dict.values()) def GetExtension(target, extra_patterns): """Return the file extension that best represents a target. For targets that generate multiple outputs it is important to return a consistent 'canonical' extension. Ultimately the goal is to group build steps by type.""" for output in target.targets: if extra_patterns: for fn_pattern in extra_patterns.split(";"): if fnmatch.fnmatch(output, "*" + fn_pattern + "*"): return fn_pattern # Not a true extension, but a good grouping. if output.endswith("type_mappings"): extension = "type_mappings" break # Capture two extensions if present. For example: file.javac.jar should # be distinguished from file.interface.jar. root, ext1 = os.path.splitext(output) _, ext2 = os.path.splitext(root) extension = ext2 + ext1 # Preserve the order in the file name. if len(extension) == 0: extension = "(no extension found)" if ext1 in [".pdb", ".dll", ".exe"]: extension = "PEFile (linking)" # Make sure that .dll and .exe are grouped together and that the # .dll.lib files don't cause these to be listed as libraries break if ext1 in [".so", ".TOC"]: extension = ".so (linking)" # Attempt to identify linking, avoid identifying as '.TOC' break # Make sure .obj files don't get categorized as mojo files if ext1 in [".obj", ".o"]: break # Jars are the canonical output of java targets. if ext1 == ".jar": break # Normalize all mojo related outputs to 'mojo'. if output.count(".mojom") > 0: extension = "mojo" break return extension def SummarizeEntries(entries, extra_step_types): """Print a summary of the passed in list of Target objects.""" # Create a list that is in order by time stamp and has entries for the # beginning and ending of each build step (one time stamp may have multiple # entries due to multiple steps starting/stopping at exactly the same time). # Iterate through this list, keeping track of which tasks are running at all # times. At each time step calculate a running total for weighted time so # that when each task ends its own weighted time can easily be calculated. task_start_stop_times = [] earliest = -1 latest = 0 total_cpu_time = 0 for target in entries: if earliest < 0 or target.start < earliest: earliest = target.start if target.end > latest: latest = target.end total_cpu_time += target.Duration() task_start_stop_times.append((target.start, "start", target)) task_start_stop_times.append((target.end, "stop", target)) length = latest - earliest weighted_total = 0.0 # Sort by the time/type records and ignore |target| task_start_stop_times.sort(key=lambda times: times[:2]) # Now we have all task start/stop times sorted by when they happen. If a # task starts and stops on the same time stamp then the start will come # first because of the alphabet, which is important for making this work # correctly. # Track the tasks which are currently running. running_tasks = {} # Record the time we have processed up to so we know how to calculate time # deltas. last_time = task_start_stop_times[0][0] # Track the accumulated weighted time so that it can efficiently be added # to individual tasks. last_weighted_time = 0.0 # Scan all start/stop events. for event in task_start_stop_times: time, action_name, target = event # Accumulate weighted time up to now. num_running = len(running_tasks) if num_running > 0: # Update the total weighted time up to this moment. last_weighted_time += (time - last_time) / float(num_running) if action_name == "start": # Record the total weighted task time when this task starts. running_tasks[target] = last_weighted_time if action_name == "stop": # Record the change in the total weighted task time while this task # ran. weighted_duration = last_weighted_time - running_tasks[target] target.SetWeightedDuration(weighted_duration) weighted_total += weighted_duration del running_tasks[target] last_time = time assert len(running_tasks) == 0 # Warn if the sum of weighted times is off by more than half a second. if abs(length - weighted_total) > 500: print( "Warning: Possible corrupt ninja log, results may be " "untrustworthy. Length = {:.3f}, weighted total = {:.3f}".format( length, weighted_total ) ) entries_by_ext = defaultdict(list) for target in entries: extension = GetExtension(target, extra_step_types) entries_by_ext[extension].append(target) for key, values in entries_by_ext.items(): print(" Longest build steps for {}:".format(key)) values.sort(key=lambda x: x.WeightedDuration()) for target in values[-long_count:]: print( " {:8.1f} weighted s to build {} ({:.1f} s elapsed time)".format( target.WeightedDuration(), target.DescribeTargets(), target.Duration(), ) ) print( " {:.1f} s weighted time ({:.1f} s elapsed time sum, {:1.1f}x " "parallelism)".format(length, total_cpu_time, total_cpu_time * 1.0 / length) ) print( " {} build steps completed, average of {:1.2f}/s".format( len(entries), len(entries) / (length) ) ) def main(): log_file = ".ninja_log" parser = argparse.ArgumentParser() parser.add_argument("-C", dest="build_directory", help="Build directory.") parser.add_argument( "-s", "--step-types", help="semicolon separated fnmatch patterns for build-step grouping", ) parser.add_argument("--log-file", help="specific ninja log file to analyze.") args, _extra_args = parser.parse_known_args() if args.build_directory: log_file = os.path.join(args.build_directory, log_file) if args.log_file: log_file = args.log_file if args.step_types: # Make room for the extra build types. global long_ext_count long_ext_count += len(args.step_types.split(";")) try: with open(log_file) as log: entries = ReadTargets(log, False) SummarizeEntries(entries, args.step_types) except OSError: print("Log file {!r} not found, no build summary created.".format(log_file)) return errno.ENOENT if __name__ == "__main__": sys.exit(main())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/generate_cmake_presets.py
tools/generate_cmake_presets.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import json import multiprocessing import os import sys from shutil import which try: # Try to get CUDA_HOME from PyTorch installation, which is the # most reliable source of truth for vLLM's build. from torch.utils.cpp_extension import CUDA_HOME except ImportError: print("Warning: PyTorch not found. Falling back to CUDA_HOME environment variable.") CUDA_HOME = os.environ.get("CUDA_HOME") def get_python_executable(): """Get the current Python executable, which is used to run this script.""" return sys.executable def get_cpu_cores(): """Get the number of CPU cores.""" return multiprocessing.cpu_count() def generate_presets(output_path="CMakeUserPresets.json", force_overwrite=False): """Generates the CMakeUserPresets.json file.""" print("Attempting to detect your system configuration...") # Detect NVCC nvcc_path = None if CUDA_HOME: prospective_path = os.path.join(CUDA_HOME, "bin", "nvcc") if os.path.exists(prospective_path): nvcc_path = prospective_path print(f"Found nvcc via torch.utils.cpp_extension.CUDA_HOME: {nvcc_path}") if not nvcc_path: nvcc_path = which("nvcc") if nvcc_path: print(f"Found nvcc in PATH: {nvcc_path}") if not nvcc_path: nvcc_path_input = input( "Could not automatically find 'nvcc'. Please provide the full " "path to nvcc (e.g., /usr/local/cuda/bin/nvcc): " ) nvcc_path = nvcc_path_input.strip() print(f"Using NVCC path: {nvcc_path}") # Detect Python executable python_executable = get_python_executable() if python_executable: print(f"Found Python via sys.executable: {python_executable}") else: python_executable_prompt = ( "Could not automatically find Python executable. Please provide " "the full path to your Python executable for vLLM development " "(typically from your virtual environment, e.g., " "/home/user/venvs/vllm/bin/python): " ) python_executable = input(python_executable_prompt).strip() if not python_executable: raise ValueError( "Could not determine Python executable. Please provide it manually." ) print(f"Using Python executable: {python_executable}") # Get CPU cores cpu_cores = get_cpu_cores() nvcc_threads = min(4, cpu_cores) cmake_jobs = max(1, cpu_cores // nvcc_threads) print( f"Detected {cpu_cores} CPU cores. " f"Setting NVCC_THREADS={nvcc_threads} and CMake jobs={cmake_jobs}." ) # Get vLLM project root (assuming this script is in vllm/tools/) project_root = os.path.abspath(os.path.join(os.path.dirname(__file__), "..")) print(f"VLLM project root detected as: {project_root}") # Ensure python_executable path is absolute or resolvable if not os.path.isabs(python_executable) and which(python_executable): python_executable = os.path.abspath(which(python_executable)) elif not os.path.isabs(python_executable): print( f"Warning: Python executable '{python_executable}' is not an " "absolute path and not found in PATH. CMake might not find it." ) cache_variables = { "CMAKE_CUDA_COMPILER": nvcc_path, "CMAKE_BUILD_TYPE": "Release", "VLLM_PYTHON_EXECUTABLE": python_executable, "CMAKE_INSTALL_PREFIX": "${sourceDir}", "CMAKE_CUDA_FLAGS": "", "NVCC_THREADS": str(nvcc_threads), } # Detect compiler cache if which("sccache"): print("Using sccache for compiler caching.") for launcher in ("C", "CXX", "CUDA", "HIP"): cache_variables[f"CMAKE_{launcher}_COMPILER_LAUNCHER"] = "sccache" elif which("ccache"): print("Using ccache for compiler caching.") for launcher in ("C", "CXX", "CUDA", "HIP"): cache_variables[f"CMAKE_{launcher}_COMPILER_LAUNCHER"] = "ccache" else: print("No compiler cache ('ccache' or 'sccache') found.") configure_preset = { "name": "release", "binaryDir": "${sourceDir}/cmake-build-release", "cacheVariables": cache_variables, } if which("ninja"): print("Using Ninja generator.") configure_preset["generator"] = "Ninja" cache_variables["CMAKE_JOB_POOLS"] = f"compile={cmake_jobs}" else: print("Ninja not found, using default generator. Build may be slower.") presets = { "version": 6, # Keep in sync with CMakeLists.txt and requirements/build.txt "cmakeMinimumRequired": {"major": 3, "minor": 26, "patch": 1}, "configurePresets": [configure_preset], "buildPresets": [ { "name": "release", "configurePreset": "release", "jobs": cmake_jobs, } ], } output_file_path = os.path.join(project_root, output_path) if os.path.exists(output_file_path): if force_overwrite: print(f"Overwriting existing file '{output_file_path}'") else: overwrite = ( input(f"'{output_file_path}' already exists. Overwrite? (y/N): ") .strip() .lower() ) if overwrite != "y": print("Generation cancelled.") return try: with open(output_file_path, "w") as f: json.dump(presets, f, indent=4) print(f"Successfully generated '{output_file_path}'") print("\nTo use this preset:") print(f"1. Ensure you are in the vLLM root directory: cd {project_root}") print("2. Initialize CMake: cmake --preset release") print("3. Build+install: cmake --build --preset release --target install") except OSError as e: print(f"Error writing file: {e}") if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--force-overwrite", action="store_true", help="Force overwrite existing CMakeUserPresets.json without prompting", ) args = parser.parse_args() generate_presets(force_overwrite=args.force_overwrite)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/profiler/visualize_layerwise_profile.py
tools/profiler/visualize_layerwise_profile.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import copy import json import math import os from pathlib import Path from typing import Any import matplotlib.pyplot as plt import pandas as pd ## JSON parsing utils #### def largest_dist_from_leaf(node: dict, depth: int = 0): if len(node["children"]) == 0: return depth return max( [largest_dist_from_leaf(child, depth=depth + 1) for child in node["children"]] ) def get_entries_at_depth( depth: int, entries_and_traces: list[tuple[Any, Any]], node: dict, curr_depth: int = 0, trace=(), ): # assert that the query is at kernel or module level assert depth == -1 or depth == -2 if curr_depth == 0 and largest_dist_from_leaf(node) <= (abs(depth) - 1): # The tree is not tall enough! entries_and_traces.append((node["entry"], trace)) return if largest_dist_from_leaf(node) == (abs(depth) - 1): entries_and_traces.append((node["entry"], trace)) trace = (node["entry"]["name"],) + trace for child in node["children"]: get_entries_at_depth( depth, entries_and_traces, child, curr_depth=curr_depth + 1, trace=trace ) def fold_nodes(root: dict, nodes_to_fold: list[str]): stack: list[dict] = [root] while len(stack) != 0: node = stack.pop() if node["entry"]["name"] in nodes_to_fold: node["children"] = [] continue for child in node["children"]: stack.append(child) return root ## Operation name cleanup utils #### def trim_string_back(string: str, width: int) -> str: if len(string) > width: offset = len(string) - width + 3 string = string[:-offset] if len(string) > 3: string = string + "..." return string def shorten_plot_legend_strings(legend, max_char_len: int): for t in legend.get_texts(): t.set_text(trim_string_back(abbreviate_known_names(t.get_text()), max_char_len)) def abbreviate_known_names(name: str) -> str: abbreviations = { "MergedColumnParallelLinear": "MCPLinear", "QKVParallelLinear": "QKVPLinear", "RowParallelLinear": "RPLinear", "weight=": "w=", "bfloat16": "bf16", "float16": "f16", } for key, value in abbreviations.items(): name = name.replace(key, value) return name def attempt_to_make_names_unique(entries_and_traces): names, non_unique_names = (set(), set()) def all_the_same(items) -> bool: return all(i == items[0] for i in items) for entry, _ in entries_and_traces: if entry["name"] in names: non_unique_names.add(entry["name"]) else: names.add(entry["name"]) for name in non_unique_names: entries_and_traces_with_name = [ (entry, trace) for entry, trace in entries_and_traces if entry["name"] == name ] zipped_traces = list(zip(*[trace for _, trace in entries_and_traces_with_name])) first_trace_difference = next( ( i for i, trace_eles in enumerate(zipped_traces) if not all_the_same(trace_eles) ), None, ) if first_trace_difference is None: # can't create a unique name, leave the names as they # are they will get aggregated by the pivot_table call continue for entry, trace in entries_and_traces_with_name: entry["name"] = " <- ".join( (entry["name"],) + trace[: first_trace_difference + 1] ) ## Operation grouping utils #### """ Group operations in the given dataframe by some high-level ops like, - gemms - attention - rms_norm etc. """ def group_trace_by_operations(trace_df: "pd.DataFrame") -> "pd.DataFrame": def is_rms_norm(op_name: str): if "rms_norm_kernel" in op_name: return True def is_attention_block(op_name: str): if "flash_fwd" in op_name or "reshape_and_cache_flash_kernel" in op_name: return True def is_quant(op_name: str): if "scaled_fp8_quant" in op_name or "scaled_int8_quant" in op_name: return True # LoRA ops def is_sgmv_shrink(op_name: str): return "sgmv_shrink" in op_name def is_sgmv_expand(op_name: str): return "sgmv_expand" in op_name def is_bgmv_shrink(op_name: str): return "bgmv_shrink" in op_name def is_bgmv_expand(op_name: str): return "bgmv_expand" in op_name def is_cutlass_gemm_op(op_name: str): return ( "void cutlass::Kernel" in op_name or "void cutlass::device_kernel" in op_name ) def is_gemm_op(op_name: str): if is_quant(op_name): return False return ( is_cutlass_gemm_op(op_name) or "xmma_gemm" in op_name or "gemv2T_kernel" in op_name or "splitKreduce" in op_name or "s16816gemm" in op_name ) def is_elementwise_op(op_name: str): return "elementwise_kernel" in op_name def is_mem_op(op_name: str): return "memcpy" in op_name.lower() or "memset" in op_name.lower() def is_vocab_embedding_op(op_name: str): return "vocabparallelembed" in op_name.lower() # nccl ops def is_nccl_op(op_name: str): return "nccl" in op_name.lower() def is_nccl_all_reduce(op_name: str): return is_nccl_op(op_name) and ( "all_reduce" in op_name.lower() or "allreduce" in op_name.lower() ) def is_nccl_gather(op_name: str): return is_nccl_op(op_name) and "gather" in op_name.lower() def is_nccl_broadcast(op_name: str): return is_nccl_op(op_name) and "broadcast" in op_name.lower() # Reduce ops types def is_cross_device_reduce_1stage(op_name: str): return "cross_device_reduce_1stage" in op_name def is_cross_device_reduce_2stage(op_name: str): return "cross_device_reduce_2stage" in op_name def is_custom_ar_all_reduce(op_name: str): return "_C_custom_ar::all_reduce" in op_name def is_reduce_kernel(op_name: str): return "reduce_kernel" in op_name headers = list(trace_df) ops = copy.deepcopy(headers) attention_ops = list(filter(lambda x: is_attention_block(x), ops)) ops = list(filter(lambda x: x not in attention_ops, ops)) quant_ops = list(filter(lambda x: is_quant(x), ops)) ops = list(filter(lambda x: x not in quant_ops, ops)) sgmv_shrink_ops = list(filter(lambda x: is_sgmv_shrink(x), ops)) ops = list(filter(lambda x: x not in sgmv_shrink_ops, ops)) sgmv_expand_ops = list(filter(lambda x: is_sgmv_expand(x), ops)) ops = list(filter(lambda x: x not in sgmv_expand_ops, ops)) bgmv_shrink_ops = list(filter(lambda x: is_bgmv_shrink(x), ops)) ops = list(filter(lambda x: x not in bgmv_shrink_ops, ops)) bgmv_expand_ops = list(filter(lambda x: is_bgmv_expand(x), ops)) ops = list(filter(lambda x: x not in bgmv_expand_ops, ops)) cutlass_gemm_ops = list(filter(lambda x: is_cutlass_gemm_op(x), ops)) ops = list(filter(lambda x: x not in cutlass_gemm_ops, ops)) gemm_ops = list(filter(lambda x: is_gemm_op(x), ops)) ops = list(filter(lambda x: x not in gemm_ops, ops)) rms_norm_ops = list(filter(lambda x: is_rms_norm(x), ops)) ops = list(filter(lambda x: x not in rms_norm_ops, ops)) vocab_embed_ops = list(filter(lambda x: is_vocab_embedding_op(x), ops)) ops = list(filter(lambda x: x not in vocab_embed_ops, ops)) mem_ops = list(filter(lambda x: is_mem_op(x), ops)) ops = list(filter(lambda x: x not in mem_ops, ops)) elementwise_ops = list(filter(lambda x: is_elementwise_op(x), ops)) ops = list(filter(lambda x: x not in elementwise_ops, ops)) nccl_all_reduce_ops = list(filter(lambda x: is_nccl_all_reduce(x), ops)) ops = list(filter(lambda x: x not in nccl_all_reduce_ops, ops)) nccl_gather_ops = list(filter(lambda x: is_nccl_gather(x), ops)) ops = list(filter(lambda x: x not in nccl_gather_ops, ops)) nccl_broadcast_ops = list(filter(lambda x: is_nccl_broadcast(x), ops)) ops = list(filter(lambda x: x not in nccl_broadcast_ops, ops)) nccl_other_ops = list(filter(lambda x: is_nccl_op(x), ops)) ops = list(filter(lambda x: x not in nccl_other_ops, ops)) cross_device_reduce_1stage_ops = list( filter(lambda x: is_cross_device_reduce_1stage(x), ops) ) ops = list(filter(lambda x: x not in cross_device_reduce_1stage_ops, ops)) cross_device_reduce_2stage_ops = list( filter(lambda x: is_cross_device_reduce_2stage(x), ops) ) ops = list(filter(lambda x: x not in cross_device_reduce_2stage_ops, ops)) custom_ar_all_reduce_ops = list(filter(lambda x: is_custom_ar_all_reduce(x), ops)) ops = list(filter(lambda x: x not in custom_ar_all_reduce_ops, ops)) reduce_kernel_ops = list(filter(lambda x: is_reduce_kernel(x), ops)) ops = list(filter(lambda x: x not in reduce_kernel_ops, ops)) if len(attention_ops): trace_df["attention"] = trace_df[attention_ops].agg("sum", axis=1) if len(quant_ops): trace_df["quant_ops"] = trace_df[quant_ops].agg("sum", axis=1) if len(sgmv_shrink_ops): trace_df["sgmv_shrink_ops"] = trace_df[sgmv_shrink_ops].agg("sum", axis=1) if len(sgmv_expand_ops): trace_df["sgmv_expand_ops"] = trace_df[sgmv_expand_ops].agg("sum", axis=1) if len(bgmv_shrink_ops): trace_df["bgmv_shrink_ops"] = trace_df[bgmv_shrink_ops].agg("sum", axis=1) if len(bgmv_expand_ops): trace_df["bgmv_expand_ops"] = trace_df[bgmv_expand_ops].agg("sum", axis=1) if len(cutlass_gemm_ops): trace_df["cutlass_gemm_ops"] = trace_df[cutlass_gemm_ops].agg("sum", axis=1) if len(gemm_ops): trace_df["gemm_ops"] = trace_df[gemm_ops].agg("sum", axis=1) if len(rms_norm_ops): trace_df["rms_norm_ops"] = trace_df[rms_norm_ops].agg("sum", axis=1) if len(vocab_embed_ops): trace_df["vocab_embed_ops"] = trace_df[vocab_embed_ops].agg("sum", axis=1) if len(mem_ops): trace_df["mem_ops"] = trace_df[mem_ops].agg("sum", axis=1) if len(elementwise_ops): trace_df["elementwise_ops"] = trace_df[elementwise_ops].agg("sum", axis=1) if len(nccl_all_reduce_ops): trace_df["nccl_all_reduce_ops"] = trace_df[nccl_all_reduce_ops].agg( "sum", axis=1 ) if len(nccl_gather_ops): trace_df["nccl_gather_ops"] = trace_df[nccl_gather_ops].agg("sum", axis=1) if len(nccl_broadcast_ops): trace_df["nccl_broadcast_ops"] = trace_df[nccl_broadcast_ops].agg("sum", axis=1) if len(nccl_other_ops): trace_df["nccl_other_ops"] = trace_df[nccl_other_ops].agg("sum", axis=1) if len(cross_device_reduce_1stage_ops): trace_df["cross_device_reduce_1stage_ops"] = trace_df[ cross_device_reduce_1stage_ops ].agg("sum", axis=1) if len(cross_device_reduce_2stage_ops): trace_df["cross_device_reduce_2stage_ops"] = trace_df[ cross_device_reduce_2stage_ops ].agg("sum", axis=1) if len(custom_ar_all_reduce_ops): trace_df["custom_ar_all_reduce_ops"] = trace_df[custom_ar_all_reduce_ops].agg( "sum", axis=1 ) if len(reduce_kernel_ops): trace_df["reduce_kernel_ops"] = trace_df[reduce_kernel_ops].agg("sum", axis=1) trace_df.drop( attention_ops + quant_ops + sgmv_shrink_ops + sgmv_expand_ops + bgmv_shrink_ops + bgmv_expand_ops + cutlass_gemm_ops + gemm_ops + rms_norm_ops + vocab_embed_ops + mem_ops + elementwise_ops + nccl_all_reduce_ops + nccl_gather_ops + nccl_broadcast_ops + nccl_other_ops + cross_device_reduce_1stage_ops + cross_device_reduce_2stage_ops + custom_ar_all_reduce_ops + reduce_kernel_ops, axis=1, inplace=True, ) return trace_df ## Data plotting utils #### def plot_trace_df( traces_df: "pd.DataFrame", plot_metric: str, plot_title: str, output: Path | None = None, ): def get_phase_description(traces_df: "pd.DataFrame", phase: str) -> str: phase_df = traces_df.query(f'phase == "{phase}"') descs = phase_df["phase_desc"].to_list() assert all([desc == descs[0] for desc in descs]) return descs[0] phases = traces_df["phase"].unique() phase_descs = [get_phase_description(traces_df, p) for p in phases] traces_df = traces_df.pivot_table( index="phase", columns="name", values=plot_metric, aggfunc="sum" ) traces_df = group_trace_by_operations(traces_df) # Make the figure fig_size_x = max(5, len(phases)) fig, ax = plt.subplots(1, figsize=(fig_size_x, 8), sharex=True) # Draw the stacked bars ops = list(traces_df) bottom = [0] * len(phases) for op in ops: values = [traces_df[op][phase] for phase in phases] values = list(map(lambda x: 0.0 if math.isnan(x) else x, values)) ax.bar(phase_descs, values, label=op, bottom=bottom) bottom = [bottom[j] + values[j] for j in range(len(phases))] # Write the values as text on the bars for bar in ax.patches: if bar.get_height() != 0: ax.text( bar.get_x() + bar.get_width() / 2, bar.get_height() / 2 + bar.get_y(), f"{round(bar.get_height(), 2)}", ha="center", color="w", weight="bold", size=5, ) # Setup legend handles, labels = plt.gca().get_legend_handles_labels() legend = fig.legend(handles, labels, loc="center left", bbox_to_anchor=(1, 1)) shorten_plot_legend_strings(legend, 50) # Setup labels and title plt.setp(ax.get_xticklabels(), rotation=90) ax.set_ylabel(plot_metric) plt.suptitle(plot_title) plt.savefig(output, bbox_inches="tight") print("Created: ", output) def main( json_trace: Path, output_directory: Path, depth: int, # Fetch/Plot operations at this depth of the Json tree plot_metric: str, make_names_unique: bool, top_k: int, json_nodes_to_fold: list[str], ): def prepare_data(profile_json: dict, step_keys: list[str]) -> "pd.DataFrame": def get_entries_and_traces(key: str): entries_and_traces: list[tuple[Any, Any]] = [] for root in profile_json[key]["summary_stats"]: # Fold nodes in the traces as per user request. i.e. simply # make the requested nodes leaf-nodes. root = fold_nodes(root, json_nodes_to_fold) get_entries_at_depth(depth, entries_and_traces, root) return entries_and_traces def keep_only_top_entries( df: "pd.DataFrame", metric: str, top_k: int = 9 ) -> "pd.DataFrame": df.loc[df.nsmallest(len(df) - top_k + 1, metric).index, ["name"]] = "others" return df def get_phase_description(key: str) -> str: num_running_seqs = profile_json[key]["metadata"]["num_running_seqs"] if num_running_seqs is not None: return f"{key}-seqs-{num_running_seqs}" else: return key # Get data for each key traces = list(map(lambda x: get_entries_and_traces(x), step_keys)) # Attempt some cleanup if make_names_unique: for trace in traces: attempt_to_make_names_unique(trace) # To pandas dataframe trace_dfs = list( map(lambda t: pd.DataFrame([entry for entry, _ in t]).fillna(0), traces) ) # Respect top_k if top_k: trace_dfs = list( map( lambda trace_df: keep_only_top_entries( trace_df, "cuda_time_us", top_k ), trace_dfs, ) ) # Fill in information about the step-keys for trace_df, step_key in zip(trace_dfs, step_keys): trace_df["phase"] = step_key trace_df["phase_desc"] = get_phase_description(step_key) # Combine all data frames so they can be put in a single plot traces_df = pd.concat(trace_dfs) # Add a derived metric `cuda_time_ms` traces_df["cuda_time_ms"] = traces_df["cuda_time_us"] / 1000 traces_df = traces_df.fillna(0) return traces_df def make_plot_title_suffix(profile_json: dict) -> str: context = profile_json["context"] sparsity = context.get("sparsity", None) run_type = ( f"Run {context['num_steps']} steps" if context["num_steps"] else ( f"Complete {context['complete_num_requests_per_step']} per " f"step; Run till completion" ) ) return ( f"{context['engine_args']['model']}\n" f"Batch={context['batch_size']}, " f"PromptLen={context['prompt_len']}, " f"NumGpus={context['engine_args']['tensor_parallel_size']}" f"{', Sparsity ' + sparsity if sparsity else ''}\n" f"Run Type: {run_type}" ) profile_json = None with open(json_trace) as f: profile_json = json.load(f) assert profile_json is not None # Get all `llm.generate.step()` profile step_traces = list(profile_json.keys()) assert step_traces[0] == "context" step_traces = step_traces[1:] # have only prefill and decodes prefills = list(filter(lambda x: "prefill" in x, step_traces)) all_decodes = list(filter(lambda x: "decode" in x, step_traces)) assert len(prefills) + len(all_decodes) == len(step_traces) assert len(prefills) == 1 decodes = all_decodes[:: args.step_plot_interval] if decodes[-1] != all_decodes[-1]: # Always have the last decode decodes.append(all_decodes[-1]) prefill_traces = prepare_data(profile_json, prefills) decode_traces = prepare_data(profile_json, decodes) plot_title_suffix = make_plot_title_suffix(profile_json) plot_trace_df( prefill_traces, plot_metric, "prefill " + plot_title_suffix, output_directory / Path("prefill.png"), ) plot_trace_df( decode_traces, plot_metric, "decodes " + plot_title_suffix, output_directory / Path("decode_steps.png"), ) if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--json-trace", type=str, required=True, help="json trace file output by \ examples/offline_inference/profiling.py", ) parser.add_argument( "--output-directory", type=str, required=False, help="Directory to output plots" ) parser.add_argument( "--level", type=str, default="module", choices=["module", "kernel"] ) parser.add_argument( "--top-k", type=int, default=12, help="Only graph the top `top_k` entries by time.", ) parser.add_argument( "--fold-json-node", nargs="+", default=["Sampler", "LogitsProcessor"], help="Do not plot the children of these nodes. Let, \ the node represent the aggregate of all its \ children", ) parser.add_argument( "--plot-metric", type=str, default="cuda_time_ms", help="Metric to plot. some options are cuda_time_ms, \ pct_cuda_time", ) parser.add_argument( "--step-plot-interval", type=int, default=4, help="For every `step_plot_interval` steps, plot 1 step", ) args = parser.parse_args() # Prepare/Extract relevant args make_names_unique = False if args.level == "module": depth = -2 make_names_unique = True elif args.level == "kernel": depth = -1 else: raise Exception(f"Unexpected level value ({args.level})") output_directory = ( args.output_directory if args.output_directory else Path(args.json_trace).parent ) if not os.path.exists(output_directory): os.makedirs(output_directory) main( Path(args.json_trace), output_directory, depth, args.plot_metric, make_names_unique, args.top_k, args.fold_json_node, )
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/profiler/print_layerwise_table.py
tools/profiler/print_layerwise_table.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import argparse import json from vllm.profiler.layerwise_profile import ModelStatsEntry, SummaryStatsEntry from vllm.profiler.utils import TablePrinter, indent_string def flatten_entries(entry_cls, profile_dict: dict): entries_and_depth = [] def get_entries(node, curr_depth=0): entries_and_depth.append((entry_cls(**node["entry"]), curr_depth)) for child in node["children"]: get_entries( child, curr_depth=curr_depth + 1, ) for root in profile_dict: get_entries(root) return entries_and_depth if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--json-trace", type=str, required=True, help="json trace file output by examples/offline_inference/profiling.py", ) parser.add_argument( "--phase", type=str, required=True, help="The phase to print the table for. This is either" "prefill or decode_n, where n is the decode step " "number", ) parser.add_argument( "--table", type=str, choices=["summary", "model"], default="summary", help="Which table to print, the summary table or the layerwise model table", ) args = parser.parse_args() with open(args.json_trace) as f: profile_data = json.load(f) assert args.phase in profile_data, ( f"Cannot find phase {args.phase} in profile data. Choose one among" f"{[x for x in profile_data if 'prefill' in x or 'decode' in x]}" ) # noqa if args.table == "summary": entries_and_depths = flatten_entries( SummaryStatsEntry, profile_data[args.phase]["summary_stats"] ) column_widths = dict(name=80, cuda_time_us=12, pct_cuda_time=12, invocations=15) elif args.table == "model": entries_and_depths = flatten_entries( ModelStatsEntry, profile_data[args.phase]["model_stats"] ) column_widths = dict( name=60, cpu_time_us=12, cuda_time_us=12, pct_cuda_time=12, trace=60 ) # indent entry names based on the depth entries = [] for entry, depth in entries_and_depths: entry.name = indent_string( entry.name, indent=depth, indent_style=lambda indent: "|" + "-" * indent + " ", ) entries.append(entry) TablePrinter(type(entries[0]), column_widths).print_table(entries)
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/profiler/nsys_profile_tools/gputrc2graph.py
tools/profiler/nsys_profile_tools/gputrc2graph.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ This generates gpu kernel analysis output from nsys rep. Will call nsys stats -r cuda_gpu_kern_trace, get non-overlapped gpu cycles, then generate csv and html output for analysis """ import argparse import logging import os import regex as re logger = logging.getLogger(__name__) # helper data class for annotating kernels def load_engine_model(): """returns engine_model built from all json files in the current dir""" import glob import json engine_model = {} json_files = glob.glob(os.path.join(os.path.dirname(__file__) or ".", "*.json")) for fname in json_files: with open(fname, encoding="utf-8") as f: engine_model.update(json.load(f)) return engine_model class GPUTrace2Graph: """ Parses output of nsys report, generates csv and bar chart output """ def __init__(self): import pandas as pd # avoid importing till needed self.pd = pd self.pd.options.mode.copy_on_write = True # helper functions for generating trace->summary csvs def gen_nonoverlapped_sum_from_gputrace(self, in_file, out_file): logger.info("loading %s", in_file) df = self.pd.read_csv( in_file, usecols=["Start (ns)", "Duration (ns)", "Device", "Strm", "Name"] ) df["End (ns)"] = df["Start (ns)"] + df["Duration (ns)"] df = self.sum_non_overlapping_intervals(df) # get ready to print table with elapsed times per kernel df["Instances"] = 1 df_sum = df.groupby("Name", as_index=False).agg( {"Elapsed Time (ns)": "sum", "Duration (ns)": "sum", "Instances": "size"} ) # generate csv df_sum["Total Time (sec)"] = df_sum["Duration (ns)"] / 1e9 df_sum["Elapsed Time (sec)"] = df_sum["Elapsed Time (ns)"] / 1e9 df_sum = df_sum.sort_values(by="Elapsed Time (sec)", ascending=False) df_sum[["Elapsed Time (sec)", "Total Time (sec)", "Instances", "Name"]].to_csv( out_file, index=False ) def sum_non_overlapping_intervals(self, df): """ returns new sorted df with Elapsed Time (ns) column using vectorized operations """ logger.info("sorting %s trace records by start time", str(df.shape)) # Sort by start time and reset index df = df.sort_values(by="Start (ns)").reset_index(drop=True) # Initialize elapsed time as duration df["Elapsed Time (ns)"] = df["Duration (ns)"] # Get numpy arrays for faster operations starts = df["Start (ns)"].values ends = df["End (ns)"].values # Keep track of current interval end current_end = ends[0] display_units = int(len(df) / 100) # Update current_end for overlapping intervals for i in range(1, len(df)): if i % display_units == 0: print(f"processing trace: {int(i / len(df) * 100)} %", end="\r") if starts[i] <= current_end: if ends[i] > current_end: # Partial overlap df.iloc[i, df.columns.get_loc("Elapsed Time (ns)")] = ( ends[i] - current_end ) current_end = ends[i] else: # Complete overlap df.iloc[i, df.columns.get_loc("Elapsed Time (ns)")] = 0 else: # No overlap current_end = ends[i] return df # functions for generating html files def make_html(self, df, output_dir, title): """make html graph from df""" import plotly.express as px if df.empty: return output_name = output_dir + "/result" if not title: title = "Model_Engine" x = "Model_Engine" y = "Elapsed Time (sec)" color = "Category" """ generate kernel mapping table """ # Sort Model_Engine categories by last field after underscore df["Model_Engine"] = self.pd.Categorical( df["Model_Engine"], sorted(df["Model_Engine"].unique(), key=lambda x: x.split("_")[-1]), ) df[["Model_Engine", color, "Instances", "Name", y]].sort_values( by=color ).to_csv(f"{output_name}.csv", index=False) graph = px.histogram( df.round(2), x=x, y=y, title=(f"{y} for {title}"), color=color, text_auto=True, ) # wrap x axis labels graph.update_xaxes(automargin=True) graph.write_html(f"{output_name}.html") """ Generate data table with columns per Model_Engine into result.html """ pivot_df = df.pivot_table( values="Elapsed Time (sec)", index="Category", columns="Model_Engine", aggfunc="sum", observed=False, ).round(2) # Add sum row at bottom pivot_df.loc["total_elapsed_sec"] = pivot_df.sum() pivot_df.fillna("").to_html("temp.html") with ( open(f"{output_name}.html", "a", encoding="utf-8") as outfile, open("temp.html", encoding="utf-8") as infile, ): outfile.write(infile.read()) os.remove("temp.html") print( f"Finished generating: \n" f" {output_name}.html for stack bar chart \n" f" {output_name}.csv for Kernel-Category mapping" ) def anno_gpu_kernname(self, df, mapping): """add "Category" column""" def anno_gpu_kernname_helper(name): for kern_name, val in mapping.items(): if re.search(kern_name, name): return val df["Category"] = df["Name"].apply(anno_gpu_kernname_helper) def make_nongpu_row(self, df, nongpu_sec): """this will append non-gpu time entry at end of df""" nongpu_row = self.pd.DataFrame([df.iloc[-1]]) nongpu_row["Category"] = nongpu_row["Name"] = "CPU(non-GPU)" nongpu_row["Instances"] = 1 nongpu_row["Elapsed Time (sec)"] = nongpu_sec return nongpu_row def is_valid_file(self, base_file): """asserts if base_file is non-existent or is empty""" assert os.path.isfile(base_file) and os.path.getsize(base_file) > 0, ( f"{base_file} doesn't exist or is empty" ) def should_gen_file(self, new_file, base_file): """figure out if new file should be generated from base_file""" self.is_valid_file(base_file) if ( os.path.exists(new_file) and (os.path.getmtime(new_file) > os.path.getmtime(base_file)) and (os.path.getsize(base_file) > 0) ): logger.info("reusing %s", new_file) return False else: logger.info("generating %s", new_file) return True def gen_sum_file(self, file, nsys_cmd): """ generates sum file from nsys trace with times per kernel and returns the name of the sum file """ import subprocess file_dir = os.path.dirname(file) file_name = os.path.basename(file) if not file_dir: file_dir = "." # Walk through trace and get the total non-overlapped time nsys_stats_file = f"{file_dir}/{file_name}_cuda_gpu_trace.csv" sum_file = f"{file_dir}/{file_name}_cuda_gpu_kernel_tracesum.csv" if self.should_gen_file(nsys_stats_file, file): cmd = [ nsys_cmd, "stats", "-r", "cuda_gpu_trace", file, "-o", f"{file_dir}/{file_name}", ] cmd_str = " ".join(cmd) logger.info("+ %s", cmd_str) # estimate time based on calibrated 240M/min file_size_mb = os.path.getsize(file) / 1e6 logger.info( "nsys stats for %.2f MB file expected to take %.2f min", file_size_mb, file_size_mb / 240, ) try: subprocess.run(cmd, check=True) except Exception: logger.error("%s failed; Use --nsys_cmd to specify nsys path", cmd_str) exit(1) logger.info("generating non-overalapped sum %s", sum_file) self.gen_nonoverlapped_sum_from_gputrace(nsys_stats_file, sum_file) self.is_valid_file(sum_file) logger.info("Finished generating %s", sum_file) return sum_file def gen_graph(self, in_file, out_dir, title, nsys_cmd, engine_model): """generates graph and csv file from in_file into out_dir""" # Initialize an empty DataFrame to store combined data combined_df = self.pd.DataFrame() for idx, (file, engine, model, total_sec) in enumerate(in_file): file_dir = os.path.dirname(file) file_name = os.path.basename(file) if not file_dir: file_dir = "." sum_file = self.gen_sum_file(file, nsys_cmd) # read kernel summary file df = self.pd.read_csv(sum_file) # annotate kernel to their categories assert engine_model.get(engine), f"engine {engine} unknown" assert engine_model[engine].get(model), f"model {model} unknown" # remove nsys-rep from file_name for shorter x-label file_name = file_name.replace(".nsys-rep", "") df["Model_Engine"] = f"{model}_{engine}_{file_name}_{idx}" self.anno_gpu_kernname(df, engine_model[engine][model]) # patch in non-gpu time gpu_sec = round(df["Elapsed Time (sec)"].sum(), 1) total_sec = round(float(total_sec), 1) if total_sec < gpu_sec: logger.warning( "Elapsed sec %.2f < GPU sec %.2f resetting Elapsed sec ", total_sec, gpu_sec, ) total_sec = gpu_sec nongpu_row = self.make_nongpu_row(df, total_sec - gpu_sec) df = self.pd.concat([df, nongpu_row], ignore_index=True) combined_df = self.pd.concat([combined_df, df], ignore_index=True) if out_dir is None: out_dir = "." else: os.makedirs(out_dir, exist_ok=True) # generate html file self.make_html(combined_df, out_dir, title) def parse_tuple(s): return tuple(s.split(",")) def main(): logging.basicConfig( format=("%(asctime)s - %(levelname)s - %(message)s"), level=logging.INFO ) parser = argparse.ArgumentParser( description=( "Process nsys rep and generate kernel non-overlapped cycles. \n" "Example:\n" "gputrc2graph.py --in_file d1.nsys-rep,vllm,llama,100 \n" "d2.nsys-rep,vllm,gpt-oss,102 " '--out_dir results/ --title "Model=gpt-oss vLLM chart"' ), formatter_class=argparse.RawDescriptionHelpFormatter, ) # load supported engine_model engine_model_supported = load_engine_model() # Get a string representation of supported engine/model combinations engine_model_supported_str = ", ".join( f"{engine}:[{', '.join(models.keys())}]" for engine, models in engine_model_supported.items() ) parser.add_argument( "--in_file", type=parse_tuple, nargs="+", help=( "list of (nsys-rep, engine, model, elapsed_nonprofiled_sec) " "separated by space. Elapsed_nonprofiled_sec is runtime without " "profiling used to calculate non-gpu time. Specify 0 to use " "elapsed time from nsys-rep but that might inflate non-gpu time. " f"Available engine:[model] are: {engine_model_supported_str} " f"Example: --infile d1.nsys-rep,vllm,llama,100 " "d2.nsys-rep,vllm,gpt-oss,102" ), required=True, ) parser.add_argument("--out_dir", help=("output dir for result.csv/html")) parser.add_argument("--title", help=("title for html chart")) parser.add_argument( "--nsys_cmd", help=("nsys cmd, e.g. /usr/bin/nsys, Default: nsys"), default="nsys", ) args = parser.parse_args() gputrace = GPUTrace2Graph() gputrace.gen_graph( args.in_file, args.out_dir, args.title, args.nsys_cmd, engine_model_supported ) if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/pre_commit/generate_nightly_torch_test.py
tools/pre_commit/generate_nightly_torch_test.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Generates specialized requirements files for nightly PyTorch testing. This script reads the main test requirements input file (`requirements/test.in`) and splits its content into two files: 1. `requirements/nightly_torch_test.txt`: Contains dependencies except PyTorch-related. 2. `torch_nightly_test.txt`: Contains only PyTorch-related packages. """ input_file = "requirements/test.in" output_file = "requirements/nightly_torch_test.txt" # white list of packages that are not compatible with PyTorch nightly directly # with pip install. Please add your package to this list if it is not compatible # or make the dependency test fails. white_list = ["torch", "torchaudio", "torchvision", "mamba_ssm"] with open(input_file) as f: lines = f.readlines() skip_next = False for line in lines: if skip_next: if line.startswith((" ", "\t")) or line.strip() == "": continue skip_next = False if any(k in line.lower() for k in white_list): skip_next = True continue
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/pre_commit/check_spdx_header.py
tools/pre_commit/check_spdx_header.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import sys from enum import Enum class SPDXStatus(Enum): """SPDX header status enumeration""" EMPTY = "empty" # empty __init__.py COMPLETE = "complete" MISSING_LICENSE = "missing_license" # Only has copyright line MISSING_COPYRIGHT = "missing_copyright" # Only has license line MISSING_BOTH = "missing_both" # Completely missing FULL_SPDX_HEADER = ( "# SPDX-License-Identifier: Apache-2.0\n" "# SPDX-FileCopyrightText: Copyright contributors to the vLLM project" ) LICENSE_LINE = "# SPDX-License-Identifier: Apache-2.0" COPYRIGHT_LINE = "# SPDX-FileCopyrightText: Copyright contributors to the vLLM project" # noqa: E501 def check_spdx_header_status(file_path): """Check SPDX header status of the file""" with open(file_path, encoding="UTF-8") as file: lines = file.readlines() if not lines: # Empty file return SPDXStatus.EMPTY # Skip shebang line start_idx = 0 if lines and lines[0].startswith("#!"): start_idx = 1 has_license = False has_copyright = False # Check all lines for SPDX headers (not just the first two) for i in range(start_idx, len(lines)): line = lines[i].strip() if line == LICENSE_LINE: has_license = True elif line == COPYRIGHT_LINE: has_copyright = True # Determine status based on what we found if has_license and has_copyright: return SPDXStatus.COMPLETE elif has_license and not has_copyright: # Only has license line return SPDXStatus.MISSING_COPYRIGHT # Only has copyright line elif not has_license and has_copyright: return SPDXStatus.MISSING_LICENSE else: # Completely missing both lines return SPDXStatus.MISSING_BOTH def add_header(file_path, status): """Add or supplement SPDX header based on status""" with open(file_path, "r+", encoding="UTF-8") as file: lines = file.readlines() file.seek(0, 0) file.truncate() if status == SPDXStatus.MISSING_BOTH: # Completely missing, add complete header if lines and lines[0].startswith("#!"): # Preserve shebang line file.write(lines[0]) file.write(FULL_SPDX_HEADER + "\n") file.writelines(lines[1:]) else: # Add header directly file.write(FULL_SPDX_HEADER + "\n") file.writelines(lines) elif status == SPDXStatus.MISSING_COPYRIGHT: # Only has license line, need to add copyright line # Find the license line and add copyright line after it for i, line in enumerate(lines): if line.strip() == LICENSE_LINE: # Insert copyright line after license line lines.insert( i + 1, f"{COPYRIGHT_LINE}\n", ) break file.writelines(lines) elif status == SPDXStatus.MISSING_LICENSE: # Only has copyright line, need to add license line # Find the copyright line and add license line before it for i, line in enumerate(lines): if line.strip() == COPYRIGHT_LINE: # Insert license line before copyright line lines.insert(i, f"{LICENSE_LINE}\n") break file.writelines(lines) def main(): """Main function""" files_missing_both = [] files_missing_copyright = [] files_missing_license = [] for file_path in sys.argv[1:]: status = check_spdx_header_status(file_path) if status == SPDXStatus.MISSING_BOTH: files_missing_both.append(file_path) elif status == SPDXStatus.MISSING_COPYRIGHT: files_missing_copyright.append(file_path) elif status == SPDXStatus.MISSING_LICENSE: files_missing_license.append(file_path) else: continue # Collect all files that need fixing all_files_to_fix = ( files_missing_both + files_missing_copyright + files_missing_license ) if all_files_to_fix: print("The following files are missing the SPDX header:") if files_missing_both: for file_path in files_missing_both: print(f" {file_path}") add_header(file_path, SPDXStatus.MISSING_BOTH) if files_missing_copyright: for file_path in files_missing_copyright: print(f" {file_path}") add_header(file_path, SPDXStatus.MISSING_COPYRIGHT) if files_missing_license: for file_path in files_missing_license: print(f" {file_path}") add_header(file_path, SPDXStatus.MISSING_LICENSE) sys.exit(1 if all_files_to_fix else 0) if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/pre_commit/check_init_lazy_imports.py
tools/pre_commit/check_init_lazy_imports.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """Ensure we perform lazy loading in vllm/__init__.py. i.e: appears only within the `if typing.TYPE_CHECKING:` guard, **except** for a short whitelist. """ import ast import sys from collections.abc import Iterable from pathlib import Path from typing import Final INIT_PATH: Final = Path("vllm/__init__.py") # If you need to add items to whitelist, do it here. ALLOWED_IMPORTS: Final[frozenset[str]] = frozenset( { "vllm.env_override", } ) ALLOWED_FROM_MODULES: Final[frozenset[str]] = frozenset( { ".version", } ) def _is_internal(name: str | None, *, level: int = 0) -> bool: if level > 0: return True if name is None: return False return name.startswith("vllm.") or name == "vllm" def _fail(violations: Iterable[tuple[int, str]]) -> None: print("ERROR: Disallowed eager imports in vllm/__init__.py:\n", file=sys.stderr) for lineno, msg in violations: print(f" Line {lineno}: {msg}", file=sys.stderr) sys.exit(1) def main() -> None: source = INIT_PATH.read_text(encoding="utf-8") tree = ast.parse(source, filename=str(INIT_PATH)) violations: list[tuple[int, str]] = [] class Visitor(ast.NodeVisitor): def __init__(self) -> None: super().__init__() self._in_type_checking = False def visit_If(self, node: ast.If) -> None: guard_is_type_checking = False test = node.test if isinstance(test, ast.Attribute) and isinstance(test.value, ast.Name): guard_is_type_checking = ( test.value.id == "typing" and test.attr == "TYPE_CHECKING" ) elif isinstance(test, ast.Name): guard_is_type_checking = test.id == "TYPE_CHECKING" if guard_is_type_checking: prev = self._in_type_checking self._in_type_checking = True for child in node.body: self.visit(child) self._in_type_checking = prev for child in node.orelse: self.visit(child) else: self.generic_visit(node) def visit_Import(self, node: ast.Import) -> None: if self._in_type_checking: return for alias in node.names: module_name = alias.name if _is_internal(module_name) and module_name not in ALLOWED_IMPORTS: violations.append( ( node.lineno, f"import '{module_name}' must be inside typing.TYPE_CHECKING", # noqa: E501 ) ) def visit_ImportFrom(self, node: ast.ImportFrom) -> None: if self._in_type_checking: return module_as_written = ("." * node.level) + (node.module or "") if ( _is_internal(node.module, level=node.level) and module_as_written not in ALLOWED_FROM_MODULES ): violations.append( ( node.lineno, f"from '{module_as_written}' import ... must be inside typing.TYPE_CHECKING", # noqa: E501 ) ) Visitor().visit(tree) if violations: _fail(violations) if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/pre_commit/check_pickle_imports.py
tools/pre_commit/check_pickle_imports.py
#!/usr/bin/env python3 # SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import sys import regex as re # List of files (relative to repo root) that are allowed to import pickle or # cloudpickle # # STOP AND READ BEFORE YOU ADD ANYTHING ELSE TO THIS LIST: # The pickle and cloudpickle modules are known to be unsafe when deserializing # data from potentially untrusted parties. They have resulted in multiple CVEs # for vLLM and numerous vulnerabilities in the Python ecosystem more broadly. # Before adding new uses of pickle/cloudpickle, please consider safer # alternatives like msgpack or pydantic that are already in use in vLLM. Only # add to this list if absolutely necessary and after careful security review. ALLOWED_FILES = { # pickle "vllm/multimodal/hasher.py", "vllm/transformers_utils/config.py", "vllm/model_executor/models/registry.py", "vllm/compilation/caching.py", "vllm/distributed/utils.py", "vllm/distributed/parallel_state.py", "vllm/distributed/device_communicators/all_reduce_utils.py", "vllm/distributed/device_communicators/shm_broadcast.py", "vllm/distributed/device_communicators/shm_object_storage.py", "vllm/utils/hashing.py", "tests/tokenizers_/test_hf.py", "tests/utils_/test_hashing.py", "benchmarks/kernels/graph_machete_bench.py", "benchmarks/kernels/benchmark_lora.py", "benchmarks/kernels/benchmark_machete.py", "benchmarks/fused_kernels/layernorm_rms_benchmarks.py", "benchmarks/cutlass_benchmarks/w8a8_benchmarks.py", "benchmarks/cutlass_benchmarks/sparse_benchmarks.py", # cloudpickle "vllm/v1/executor/multiproc_executor.py", "vllm/v1/executor/ray_executor.py", "vllm/entrypoints/llm.py", "tests/utils.py", # pickle and cloudpickle "vllm/v1/serial_utils.py", } PICKLE_RE = re.compile( r"^\s*(import\s+(pickle|cloudpickle)(\s|$|\sas)" r"|from\s+(pickle|cloudpickle)\s+import\b)" ) def scan_file(path: str) -> int: with open(path, encoding="utf-8") as f: for i, line in enumerate(f, 1): if PICKLE_RE.match(line): print( f"{path}:{i}: " "\033[91merror:\033[0m " # red color "Found pickle/cloudpickle import" ) return 1 return 0 def main(): returncode = 0 for filename in sys.argv[1:]: if filename in ALLOWED_FILES: continue returncode |= scan_file(filename) return returncode def test_regex(): test_cases = [ # Should match ("import pickle", True), ("import cloudpickle", True), ("import pickle as pkl", True), ("import cloudpickle as cpkl", True), ("from pickle import *", True), ("from cloudpickle import dumps", True), ("from pickle import dumps, loads", True), ("from cloudpickle import (dumps, loads)", True), (" import pickle", True), ("\timport cloudpickle", True), ("from pickle import loads", True), # Should not match ("import somethingelse", False), ("from somethingelse import pickle", False), ("# import pickle", False), ("print('import pickle')", False), ("import pickleas as asdf", False), ] for i, (line, should_match) in enumerate(test_cases): result = bool(PICKLE_RE.match(line)) assert result == should_match, ( f"Test case {i} failed: '{line}' (expected {should_match}, got {result})" ) print("All regex tests passed.") if __name__ == "__main__": if "--test-regex" in sys.argv: test_regex() else: sys.exit(main())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/pre_commit/validate_config.py
tools/pre_commit/validate_config.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project """ Ensures all fields in a config dataclass have default values and that each field has a docstring. """ import ast import inspect import sys from itertools import pairwise import regex as re def get_attr_docs(cls_node: ast.ClassDef) -> dict[str, str]: """ Get any docstrings placed after attribute assignments in a class body. Adapted from https://davidism.com/attribute-docstrings/ https://davidism.com/mit-license/ """ out = {} # Consider each pair of nodes. for a, b in pairwise(cls_node.body): # Must be an assignment then a constant string. if ( not isinstance(a, (ast.Assign, ast.AnnAssign)) or not isinstance(b, ast.Expr) or not isinstance(b.value, ast.Constant) or not isinstance(b.value.value, str) ): continue doc = inspect.cleandoc(b.value.value) # An assignment can have multiple targets (a = b = v), but an # annotated assignment only has one target. targets = a.targets if isinstance(a, ast.Assign) else [a.target] for target in targets: # Must be assigning to a plain name. if not isinstance(target, ast.Name): continue out[target.id] = doc return out class ConfigValidator(ast.NodeVisitor): def __init__(self): ... def visit_ClassDef(self, node): # Validate class with both @config and @dataclass decorators decorators = [ id for d in node.decorator_list if ( isinstance(d, ast.Name) and ((id := d.id) == "config" or id == "dataclass") ) or ( isinstance(d, ast.Call) and (isinstance(d.func, ast.Name) and (id := d.func.id) == "dataclass") ) ] if set(decorators) == {"config", "dataclass"}: validate_class(node) elif set(decorators) == {"config"}: fail(f"Class {node.name} with config decorator must be a dataclass.", node) self.generic_visit(node) def validate_class(class_node: ast.ClassDef): attr_docs = get_attr_docs(class_node) for stmt in class_node.body: # A field is defined as a class variable that has a type annotation. if isinstance(stmt, ast.AnnAssign): # Skip ClassVar and InitVar # see https://docs.python.org/3/library/dataclasses.html#class-variables # and https://docs.python.org/3/library/dataclasses.html#init-only-variables if ( isinstance(stmt.annotation, ast.Subscript) and isinstance(stmt.annotation.value, ast.Name) and stmt.annotation.value.id in {"ClassVar", "InitVar"} ): continue if isinstance(stmt.target, ast.Name): field_name = stmt.target.id if stmt.value is None: fail( f"Field '{field_name}' in {class_node.name} must have " "a default value.", stmt, ) if field_name not in attr_docs: fail( f"Field '{field_name}' in {class_node.name} must have " "a docstring.", stmt, ) if ( isinstance(stmt.annotation, ast.Subscript) and isinstance(stmt.annotation.value, ast.Name) and stmt.annotation.value.id == "Union" and isinstance(stmt.annotation.slice, ast.Tuple) ): args = stmt.annotation.slice.elts literal_args = [ arg for arg in args if isinstance(arg, ast.Subscript) and isinstance(arg.value, ast.Name) and arg.value.id == "Literal" ] if len(literal_args) > 1: fail( f"Field '{field_name}' in {class_node.name} must " "use a single " "Literal type. Please use 'Literal[Literal1, " "Literal2]' instead of 'Union[Literal1, Literal2]'" ".", stmt, ) def validate_ast(tree: ast.stmt): ConfigValidator().visit(tree) def validate_file(file_path: str): try: print(f"Validating {file_path} config dataclasses ", end="") with open(file_path, encoding="utf-8") as f: source = f.read() tree = ast.parse(source, filename=file_path) validate_ast(tree) except ValueError as e: print(e) raise SystemExit(1) from e else: print("✅") def fail(message: str, node: ast.stmt): raise ValueError(f"❌ line({node.lineno}): {message}") def main(): for filename in sys.argv[1:]: # Only run for Python files in vllm/ or tests/ if not re.match(r"^(vllm|tests)/.*\.py$", filename): continue # Only run if the file contains @config with open(filename, encoding="utf-8") as f: if "@config" in f.read(): validate_file(filename) if __name__ == "__main__": main()
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false
vllm-project/vllm
https://github.com/vllm-project/vllm/blob/0d4044edd85de30d7d4558aeea4d1e95c7c556d6/tools/pre_commit/check_triton_import.py
tools/pre_commit/check_triton_import.py
# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project import subprocess import sys import regex as re FORBIDDEN_IMPORT_RE = re.compile(r"^(from|import)\s+triton(\s|\.|$)") # the way allowed to import triton ALLOWED_LINES = { "from vllm.triton_utils import triton", "from vllm.triton_utils import tl", "from vllm.triton_utils import tl, triton", } ALLOWED_FILES = {"vllm/triton_utils/importing.py"} def is_allowed_file(current_file: str) -> bool: return current_file in ALLOWED_FILES def is_forbidden_import(line: str) -> bool: stripped = line.strip() return bool(FORBIDDEN_IMPORT_RE.match(stripped)) and stripped not in ALLOWED_LINES def parse_diff(diff: str) -> list[str]: violations = [] current_file = None current_lineno = None skip_allowed_file = False for line in diff.splitlines(): if line.startswith("+++ b/"): current_file = line[6:] skip_allowed_file = is_allowed_file(current_file) elif skip_allowed_file: continue elif line.startswith("@@"): match = re.search(r"\+(\d+)", line) if match: current_lineno = int(match.group(1)) - 1 # next "+ line" is here elif line.startswith("+") and not line.startswith("++"): current_lineno += 1 code_line = line[1:] if is_forbidden_import(code_line): violations.append( f"{current_file}:{current_lineno}: {code_line.strip()}" ) return violations def get_diff(diff_type: str) -> str: if diff_type == "staged": return subprocess.check_output( ["git", "diff", "--cached", "--unified=0"], text=True ) elif diff_type == "unstaged": return subprocess.check_output(["git", "diff", "--unified=0"], text=True) else: raise ValueError(f"Unknown diff_type: {diff_type}") def main(): all_violations = [] for diff_type in ["staged", "unstaged"]: try: diff_output = get_diff(diff_type) violations = parse_diff(diff_output) all_violations.extend(violations) except subprocess.CalledProcessError as e: print(f"[{diff_type}] Git diff failed: {e}", file=sys.stderr) if all_violations: print( "❌ Forbidden direct `import triton` detected." " ➤ Use `from vllm.triton_utils import triton` instead.\n" ) for v in all_violations: print(f"❌ {v}") return 1 return 0 if __name__ == "__main__": sys.exit(main())
python
Apache-2.0
0d4044edd85de30d7d4558aeea4d1e95c7c556d6
2026-01-04T14:38:19.902011Z
false