File size: 4,768 Bytes
5fed0fc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
import torch
import math
import triton
from typing import Optional

# Ensure CUDA is available and properly initialize device
if not torch.cuda.is_available():
    raise RuntimeError("CUDA is not available. This benchmark requires a CUDA-enabled GPU.")
DEVICE = torch.device("cuda:0")
torch.cuda.set_device(DEVICE)

def alloc_fn(size: int, align: int, stream: Optional[int]):
    assert align == 128
    assert stream == 0
    return torch.empty(size, dtype=torch.int8, device=DEVICE)

triton.set_allocator(alloc_fn)
torch.manual_seed(0)
try:
    torch.cuda.manual_seed_all(0)
except Exception:
    pass
assert triton.runtime.driver.active.get_current_target().backend == "cuda", "This benchmark only supports CUDA backend."

def _bench_ms(fn):
    out = triton.testing.do_bench(fn, quantiles=[0.5])
    if isinstance(out, (tuple, list)):
        return float(out[0])
    return float(out)

def _is_close(x: torch.Tensor, y: torch.Tensor, rtol=1e-2, atol=5e-3):
    return torch.allclose(x, y, rtol=rtol, atol=atol)

def _pt_bmm(A, B):
    # A:[B,M,K], B:[B,K,N] -> [B,M,N]
    return torch.bmm(A.float(), B.float()).to(torch.float16)

def _bench_pair(B, M, N, K, answer_bmm, baseline_bmm=_pt_bmm):
    A = torch.randn(B, M, K, device=DEVICE, dtype=torch.float16)
    Bm = torch.randn(B, K, N, device=DEVICE, dtype=torch.float16)
    
    baseline_ms = _bench_ms(lambda: baseline_bmm(A, Bm))
    answer_ms = _bench_ms(lambda: answer_bmm(A, Bm))
    
    flops = 2.0 * B * M * N * K
    to_tflops = lambda ms: flops * 1e-12 / (ms * 1e-3) if ms is not None else None
    
    ref = baseline_bmm(A, Bm)
    out = answer_bmm(A, Bm)
    passed = _is_close(out, ref, rtol=1e-2, atol=5e-3)
    
    return {
        "B": B, "M": M, "N": N, "K": K,
        "baseline_ms": baseline_ms, "answer_ms": answer_ms,
        "baseline_tflops": to_tflops(baseline_ms),
        "answer_tflops": to_tflops(answer_ms),
        "close_passed": passed,
        "rtol": 1e-2, "atol": 5e-3, "passed": passed,
    }

def _warmup_gpu(iters: int = 10):
    try:
        B, M, K, N = 64, 64, 64, 64
        A = torch.randn(B, M, K, device=DEVICE, dtype=torch.float16)
        Bm = torch.randn(B, K, N, device=DEVICE, dtype=torch.float16)
        for _ in range(max(1, int(iters))):
            _ = torch.bmm(A, Bm)
        torch.cuda.synchronize()
    except Exception:
        pass

def summarize_speedup(answer_bmm, baseline_bmm=_pt_bmm, print_output=False, metadata=None):
    # Warm up GPU to stabilize clocks and caches
    _warmup_gpu(10)
    
    # Get shapes from metadata or use defaults
    if metadata is None:
        metadata = {}
    shapes = metadata.get("shapes", None)
    if shapes is None:
        B_list = metadata.get("B_list", [64, 256, 1024])
        M = metadata.get("M", 64)
        N = metadata.get("N", 64)
        K = metadata.get("K", 64)
        shapes = [(B, M, N, K) for B in B_list]
    
    rows = []
    for (B, M, N, K) in shapes:
        r = _bench_pair(B, M, N, K, answer_bmm, baseline_bmm)
        rows.append(r)
    
    if print_output:
        print("\n=== Answer vs Baseline: Speedup for each shape (based on median time) ===")
    
    speedups = []
    for r in rows:
        tm, cm = r["answer_ms"], r["baseline_ms"]
        if cm is None or tm is None:
            continue
        sp = cm / tm
        speedups.append(sp)
        status = "OK" if r["close_passed"] else "FAIL"
        if print_output:
            print(
                f"B={r['B']:4d} M={r['M']:4d} N={r['N']:4d} K={r['K']:4d}  "
                f"baseline={cm:7.3f} ms  answer={tm:7.3f} ms  speedup={sp:5.2f}x  "
                f"[Passed: {status}  "
                f"rtol={r['rtol']:.1e} atol={r['atol']:.1e}]"
            )
    
    if speedups:
        arith_mean = sum(speedups) / len(speedups)
        geo_mean = math.exp(sum(math.log(s) for s in speedups) / len(speedups))
        median = sorted(speedups)[len(speedups)//2]
        if print_output:
            print("\n--- Summary ---")
            print(f"Sample size: {len(speedups)}")
            print(f"Arithmetic mean speedup: {arith_mean:.3f}x")
            print(f"Geometric mean speedup: {geo_mean:.3f}x")
            print(f"Median speedup: {median:.3f}x")
    else:
        arith_mean = geo_mean = median = 0.0
    
    return rows, arith_mean, geo_mean, median

def run_benchmark(answer_bmm, baseline_bmm=_pt_bmm, print_output=False, metadata=None):
    rows, arith_mean, geo_mean, median = summarize_speedup(answer_bmm, baseline_bmm, print_output=print_output, metadata=metadata)
    return {
        "rows": rows,
        "arithmetic_mean_speedup": arith_mean,
        "geometric_mean_speedup": geo_mean,
        "median_speedup": median,
        "pass_all": all(r["close_passed"] for r in rows),
    }