| | import gc |
| | import inspect |
| | import logging |
| | import os |
| | import queue |
| | import threading |
| | from contextlib import nullcontext |
| | from dataclasses import dataclass |
| | from typing import Any, Callable, Dict, Optional, Union |
| |
|
| | import pandas as pd |
| | import torch |
| | import torch.utils.benchmark as benchmark |
| |
|
| | from diffusers.models.modeling_utils import ModelMixin |
| | from diffusers.utils.testing_utils import require_torch_gpu, torch_device |
| |
|
| |
|
| | logging.basicConfig(level=logging.INFO, format="%(asctime)s %(levelname)s %(name)s: %(message)s") |
| | logger = logging.getLogger(__name__) |
| |
|
| | NUM_WARMUP_ROUNDS = 5 |
| |
|
| |
|
| | def benchmark_fn(f, *args, **kwargs): |
| | t0 = benchmark.Timer( |
| | stmt="f(*args, **kwargs)", |
| | globals={"args": args, "kwargs": kwargs, "f": f}, |
| | num_threads=1, |
| | ) |
| | return float(f"{(t0.blocked_autorange().mean):.3f}") |
| |
|
| |
|
| | def flush(): |
| | gc.collect() |
| | torch.cuda.empty_cache() |
| | torch.cuda.reset_max_memory_allocated() |
| | torch.cuda.reset_peak_memory_stats() |
| |
|
| |
|
| | |
| | def calculate_flops(model, input_dict): |
| | try: |
| | from torchprofile import profile_macs |
| | except ModuleNotFoundError: |
| | raise |
| |
|
| | |
| | sig = inspect.signature(model.forward) |
| | param_names = [ |
| | p.name |
| | for p in sig.parameters.values() |
| | if p.kind |
| | in ( |
| | inspect.Parameter.POSITIONAL_ONLY, |
| | inspect.Parameter.POSITIONAL_OR_KEYWORD, |
| | ) |
| | and p.name != "self" |
| | ] |
| | bound = sig.bind_partial(**input_dict) |
| | bound.apply_defaults() |
| | args = tuple(bound.arguments[name] for name in param_names) |
| |
|
| | model.eval() |
| | with torch.no_grad(): |
| | macs = profile_macs(model, args) |
| | flops = 2 * macs |
| | return flops |
| |
|
| |
|
| | def calculate_params(model): |
| | return sum(p.numel() for p in model.parameters()) |
| |
|
| |
|
| | |
| | |
| | def model_init_fn(model_cls, group_offload_kwargs=None, layerwise_upcasting=False, **init_kwargs): |
| | model = model_cls.from_pretrained(**init_kwargs).eval() |
| | if group_offload_kwargs and isinstance(group_offload_kwargs, dict): |
| | model.enable_group_offload(**group_offload_kwargs) |
| | else: |
| | model.to(torch_device) |
| | if layerwise_upcasting: |
| | model.enable_layerwise_casting( |
| | storage_dtype=torch.float8_e4m3fn, compute_dtype=init_kwargs.get("torch_dtype", torch.bfloat16) |
| | ) |
| | return model |
| |
|
| |
|
| | @dataclass |
| | class BenchmarkScenario: |
| | name: str |
| | model_cls: ModelMixin |
| | model_init_kwargs: Dict[str, Any] |
| | model_init_fn: Callable |
| | get_model_input_dict: Callable |
| | compile_kwargs: Optional[Dict[str, Any]] = None |
| |
|
| |
|
| | @require_torch_gpu |
| | class BenchmarkMixin: |
| | def pre_benchmark(self): |
| | flush() |
| | torch.compiler.reset() |
| |
|
| | def post_benchmark(self, model): |
| | model.cpu() |
| | flush() |
| | torch.compiler.reset() |
| |
|
| | @torch.no_grad() |
| | def run_benchmark(self, scenario: BenchmarkScenario): |
| | |
| | logger.info(f"Running scenario: {scenario.name}.") |
| | try: |
| | model = model_init_fn(scenario.model_cls, **scenario.model_init_kwargs) |
| | num_params = round(calculate_params(model) / 1e9, 2) |
| | try: |
| | flops = round(calculate_flops(model, input_dict=scenario.get_model_input_dict()) / 1e9, 2) |
| | except Exception as e: |
| | logger.info(f"Problem in calculating FLOPs:\n{e}") |
| | flops = None |
| | model.cpu() |
| | del model |
| | except Exception as e: |
| | logger.info(f"Error while initializing the model and calculating FLOPs:\n{e}") |
| | return {} |
| | self.pre_benchmark() |
| |
|
| | |
| | results = {} |
| | plain = None |
| | try: |
| | plain = self._run_phase( |
| | model_cls=scenario.model_cls, |
| | init_fn=scenario.model_init_fn, |
| | init_kwargs=scenario.model_init_kwargs, |
| | get_input_fn=scenario.get_model_input_dict, |
| | compile_kwargs=None, |
| | ) |
| | except Exception as e: |
| | logger.info(f"Benchmark could not be run with the following error:\n{e}") |
| | return results |
| |
|
| | |
| | compiled = {"time": None, "memory": None} |
| | if scenario.compile_kwargs: |
| | try: |
| | compiled = self._run_phase( |
| | model_cls=scenario.model_cls, |
| | init_fn=scenario.model_init_fn, |
| | init_kwargs=scenario.model_init_kwargs, |
| | get_input_fn=scenario.get_model_input_dict, |
| | compile_kwargs=scenario.compile_kwargs, |
| | ) |
| | except Exception as e: |
| | logger.info(f"Compilation benchmark could not be run with the following error\n: {e}") |
| | if plain is None: |
| | return results |
| |
|
| | |
| | result = { |
| | "scenario": scenario.name, |
| | "model_cls": scenario.model_cls.__name__, |
| | "num_params_B": num_params, |
| | "flops_G": flops, |
| | "time_plain_s": plain["time"], |
| | "mem_plain_GB": plain["memory"], |
| | "time_compile_s": compiled["time"], |
| | "mem_compile_GB": compiled["memory"], |
| | } |
| | if scenario.compile_kwargs: |
| | result["fullgraph"] = scenario.compile_kwargs.get("fullgraph", False) |
| | result["mode"] = scenario.compile_kwargs.get("mode", "default") |
| | else: |
| | result["fullgraph"], result["mode"] = None, None |
| | return result |
| |
|
| | def run_bencmarks_and_collate(self, scenarios: Union[BenchmarkScenario, list[BenchmarkScenario]], filename: str): |
| | if not isinstance(scenarios, list): |
| | scenarios = [scenarios] |
| | record_queue = queue.Queue() |
| | stop_signal = object() |
| |
|
| | def _writer_thread(): |
| | while True: |
| | item = record_queue.get() |
| | if item is stop_signal: |
| | break |
| | df_row = pd.DataFrame([item]) |
| | write_header = not os.path.exists(filename) |
| | df_row.to_csv(filename, mode="a", header=write_header, index=False) |
| | record_queue.task_done() |
| |
|
| | record_queue.task_done() |
| |
|
| | writer = threading.Thread(target=_writer_thread, daemon=True) |
| | writer.start() |
| |
|
| | for s in scenarios: |
| | try: |
| | record = self.run_benchmark(s) |
| | if record: |
| | record_queue.put(record) |
| | else: |
| | logger.info(f"Record empty from scenario: {s.name}.") |
| | except Exception as e: |
| | logger.info(f"Running scenario ({s.name}) led to error:\n{e}") |
| | record_queue.put(stop_signal) |
| | logger.info(f"Results serialized to {filename=}.") |
| |
|
| | def _run_phase( |
| | self, |
| | *, |
| | model_cls: ModelMixin, |
| | init_fn: Callable, |
| | init_kwargs: Dict[str, Any], |
| | get_input_fn: Callable, |
| | compile_kwargs: Optional[Dict[str, Any]], |
| | ) -> Dict[str, float]: |
| | |
| | self.pre_benchmark() |
| |
|
| | |
| | model = init_fn(model_cls, **init_kwargs) |
| | if compile_kwargs: |
| | model.compile(**compile_kwargs) |
| |
|
| | |
| | inp = get_input_fn() |
| |
|
| | |
| | run_ctx = torch._inductor.utils.fresh_inductor_cache() if compile_kwargs else nullcontext() |
| | with run_ctx: |
| | for _ in range(NUM_WARMUP_ROUNDS): |
| | _ = model(**inp) |
| | time_s = benchmark_fn(lambda m, d: m(**d), model, inp) |
| | mem_gb = torch.cuda.max_memory_allocated() / (1024**3) |
| | mem_gb = round(mem_gb, 2) |
| |
|
| | |
| | self.post_benchmark(model) |
| | del model |
| | return {"time": time_s, "memory": mem_gb} |
| |
|