transformers / benchmark_v2 /framework /benchmark_config.py
AbdulElahGwaith's picture
Upload folder using huggingface_hub
a9bd396 verified
import hashlib
import itertools
import json
import logging
from functools import lru_cache
from typing import Any
import torch
from transformers.generation.configuration_utils import CompileConfig
from transformers.utils import is_torch_accelerator_available
from transformers.utils.import_utils import is_flash_attn_2_available, is_kernels_available
KERNELIZATION_AVAILABLE = False
try:
from kernels import Mode, kernelize # noqa: F401
KERNELIZATION_AVAILABLE = True
except ImportError:
pass
logger = logging.getLogger(__name__)
@lru_cache
def is_fa2_or_kernel_available() -> bool:
"""Returns True if the flash_attn_2 or a fallback kernel is available"""
# Early return if flash_attn_2 is available
if is_flash_attn_2_available():
return True
# Early return if kernels is not available
if not is_kernels_available():
logger.warning(
"flash_attention_2 is not available. kernels is not installed. Benchmarking flash_attention_2 will not "
"be possible."
)
return False
# If kernels is available, try to get the flash_attn_2 kernel
try:
from kernels import get_kernel
get_kernel("kernels-community/flash-attn")
except Exception as _:
logger.warning(
"flash_attention_2 is not available. kernels is installed, but the flash_attn kernel is not available."
"Benchmarking flash_attention_2 will not be possible."
)
return False
class BenchmarkConfig:
"""Configuration for a single benchmark scenario."""
all_attn_implementations = ["flash_attention_2", "eager", "sdpa", "flex_attention"]
all_compiled_modes = [None, "default", "reduce-overhead", "max-autotune", "max-autotune-no-cudagraphs"]
def __init__(
self,
warmup_iterations: int = 5,
measurement_iterations: int = 20,
gpu_monitoring: bool = True, # NOTE: you may want to disable this at times as we have obsvered it could heavily slow down benchmarks on AMD
continuous_batching: bool = False,
batch_size: int = 1,
sequence_length: int = 128,
num_tokens_to_generate: int = 128,
attn_implementation: str = "eager",
compile_kwargs: dict[str, Any] | None = None,
kernelize: bool = False,
name: str | None = None,
skip_validity_check: bool = False,
) -> None:
# Benchmark parameters
self.warmup_iterations = warmup_iterations
self.measurement_iterations = measurement_iterations
self.gpu_monitoring = gpu_monitoring
self.continuous_batching = continuous_batching
# Input parameters
self.batch_size = batch_size
self.sequence_length = sequence_length
self.num_tokens_to_generate = num_tokens_to_generate
# Generation parameters
self.attn_implementation = attn_implementation
# Optimization parameters
if compile_kwargs is None:
self.compile_config = None
else:
compile_kwargs["fullgraph"] = compile_kwargs.get("fullgraph", True)
self.compile_config = CompileConfig(**compile_kwargs)
self.kernelize = kernelize
# Constant parameters
self.dtype = "torch.bfloat16"
self.device = torch.accelerator.current_accelerator().type if is_torch_accelerator_available() else "cuda"
self.check_validity(skip_validity_check)
self.name = name if name is not None else self.infer_name()
def check_validity(self, skip_validity_check: bool = False) -> None:
if skip_validity_check:
return
# If flash_attention_2 is selected but not available, default to SDPA
if self.attn_implementation == "flash_attention_2" and not is_fa2_or_kernel_available():
logger.error("Flash attention is not available. Defaulting to SDPA.")
self.attn_implementation = "sdpa"
# The combination of flash_attention_2, compile and generate is not supported # FIXME: support it
if (
not self.continuous_batching
and self.attn_implementation == "flash_attention_2"
and self.compile_config is not None
):
logger.error(
"The combination of flash_attention_2, compile and generate is not supported. Turning off compile."
)
self.compile_config = None
# Continuous batching does not support flex attention as an attention implementation # FIXME: support it
if self.attn_implementation == "flex_attention" and self.continuous_batching:
logger.error(
"Disabling continuous batching because of invalid configuration: flex attention is not supported."
)
self.continuous_batching = False
# Continuous batching supports compile mode "default" or "max-autotune-no-cudagraphs"
if (
self.continuous_batching
and self.compile_config is not None
and self.compile_config.mode not in ["default", "max-autotune-no-cudagraphs"]
):
logger.error(
f"You have continuous batching and compile enabled, but {self.compile_config.mode = } is not supported."
" Supported modes are: default, max-autotune-no-cudagraphs. Changing to default."
)
self.compile_config.mode = "default"
@property
def hash(self) -> str:
return hashlib.sha256(json.dumps(self.to_dict()).encode()).hexdigest()
def infer_name(self, compact: bool = True) -> str:
"""Infer a human-readable name for the benchmark config, either compact or verbose."""
if compact:
iter_str = f"w{self.warmup_iterations}_i{self.measurement_iterations}"
gpu_monitor_str = "monitored" if self.gpu_monitoring else "unmonitored"
dimensions_str = f"b{self.batch_size}_s{self.sequence_length}_n{self.num_tokens_to_generate}"
attn_code = self.attn_implementation
compile_str = f"compiled_{self.compile_config.mode}" if self.compile_config is not None else "uncompiled"
kernelize_str = "kernelized" if self.kernelize else "unkernelized"
continuous_batching_str = "cb" if self.continuous_batching else "generate"
sep = "-"
else:
iter_str = f"{self.warmup_iterations} warmup, {self.measurement_iterations} iterations"
gpu_monitor_str = ("with" if self.gpu_monitoring else "no") + " GPU monitoring"
dimensions_str = f"batch size {self.batch_size}, sequence length {self.sequence_length}, {self.num_tokens_to_generate} generated tokens"
attn_code = f"{self.attn_implementation} attention"
compile_str = "compiled" if self.compile_config is not None else "not compiled"
kernelize_str = "kernelized" if self.kernelize else "not kernelized"
continuous_batching_str = "continuous batching" if self.continuous_batching else "regular generate"
sep = ", "
return sep.join(
[iter_str, gpu_monitor_str, dimensions_str, attn_code, compile_str, kernelize_str, continuous_batching_str]
)
def to_dict(self) -> dict[str, Any]:
return {
"name": self.name,
"warmup_iterations": self.warmup_iterations,
"measurement_iterations": self.measurement_iterations,
"gpu_monitoring": self.gpu_monitoring,
"continuous_batching": self.continuous_batching,
"batch_size": self.batch_size,
"sequence_length": self.sequence_length,
"num_tokens_to_generate": self.num_tokens_to_generate,
"attn_implementation": self.attn_implementation,
"compile_kwargs": self.compile_config.to_dict() if self.compile_config is not None else None,
"kernelize": self.kernelize,
}
@classmethod
def from_dict(cls, data: dict[str, Any], skip_validity_check: bool = False) -> "BenchmarkConfig":
return cls(
warmup_iterations=data.get("warmup_iterations", 5),
measurement_iterations=data.get("measurement_iterations", 20),
gpu_monitoring=data.get("gpu_monitoring", False),
continuous_batching=data.get("continuous_batching", False),
batch_size=data.get("batch_size", 1),
sequence_length=data.get("sequence_length", 128),
num_tokens_to_generate=data.get("num_tokens_to_generate", 128),
attn_implementation=data.get("attn_implementation", "eager"),
compile_kwargs=data.get("compile_kwargs"),
kernelize=data.get("kernelize", False),
name=data.get("name"),
skip_validity_check=skip_validity_check,
)
def adapt_configs(
configs: list[BenchmarkConfig],
warmup_iterations: int | list[int] = 5,
measurement_iterations: int | list[int] = 20,
batch_size: int | list[int] = 1,
sequence_length: int | list[int] = 128,
num_tokens_to_generate: int | list[int] = 128,
gpu_monitoring: bool | list[bool] = True,
) -> list[BenchmarkConfig]:
parameters = (
x if isinstance(x, list) else [x]
for x in [
warmup_iterations,
measurement_iterations,
batch_size,
sequence_length,
num_tokens_to_generate,
gpu_monitoring,
]
)
iterator = itertools.product(*parameters)
adapted_configs = []
for warmup_iters, measurement_iters, bs, seqlen, ntok, monitor in iterator:
for config in configs:
config = config.to_dict()
config["warmup_iterations"] = warmup_iters
config["measurement_iterations"] = measurement_iters
config["batch_size"] = bs
config["sequence_length"] = seqlen
config["num_tokens_to_generate"] = ntok
config["gpu_monitoring"] = monitor
# Remove the old name so it gets re-inferred with the updated values
config.pop("name", None)
adapted_configs.append(BenchmarkConfig.from_dict(config))
return adapted_configs
def get_config_by_level(level: int) -> list[BenchmarkConfig]:
configs = []
# Early return if level is greater than 3: we generate all combinations of configs, maybe even w/ all compile modes
if level >= 3:
for attn_implementation in BenchmarkConfig.all_attn_implementations:
# Usually there is not much to gain by compiling with other modes, but we allow it for level 4
compile_modes = BenchmarkConfig.all_compiled_modes if level >= 4 else [None, "default"]
for cm in compile_modes:
compile_kwargs = {"mode": cm} if cm is not None else None
for kernelize_on in {False, KERNELIZATION_AVAILABLE}:
for cb_on in [False, True]:
configs.append(
BenchmarkConfig(
attn_implementation=attn_implementation,
compile_kwargs=compile_kwargs,
kernelize=kernelize_on,
continuous_batching=cb_on,
)
)
return configs
# Otherwise, we add the configs for the given level
if level >= 0:
configs.append(BenchmarkConfig(attn_implementation="flex_attention", compile_kwargs={}))
if level >= 1:
configs.append(BenchmarkConfig(attn_implementation="flash_attention_2"))
configs.append(BenchmarkConfig(attn_implementation="eager", compile_kwargs={}))
configs.append(BenchmarkConfig(attn_implementation="flash_attention_2", continuous_batching=True))
if level >= 2:
configs.append(BenchmarkConfig(attn_implementation="sdpa", compile_kwargs={}))
configs.append(BenchmarkConfig(attn_implementation="flex_attention", compile_kwargs={}, kernelize=True))
configs.append(BenchmarkConfig(attn_implementation="flash_attention_2", kernelize=True))
configs.append(BenchmarkConfig(attn_implementation="sdpa", continuous_batching=True))
return configs