| """SOTA reference for KDA forward: fla.ops.kda.chunk_kda (Triton). |
| |
| The agent's solution is forbidden from importing this module path (see |
| problem.yaml.forbidden). This file is only used by benchmark.py to draw |
| the SOTA reference line. |
| |
| If FLA's Triton kernel does not run on SM120 (Blackwell consumer-lineage -- |
| some Triton kernels in FLA target Hopper TMA), is_available() returns False |
| and benchmark.py omits the SOTA variant. The H100 reference is documented |
| in problem.yaml for context. |
| """ |
| from __future__ import annotations |
|
|
| import torch |
|
|
|
|
| def _import_fla(): |
| try: |
| from fla.ops.kda import chunk_kda |
| return chunk_kda |
| except Exception: |
| return None |
|
|
|
|
| def sota_forward( |
| q: torch.Tensor, |
| k: torch.Tensor, |
| v: torch.Tensor, |
| g: torch.Tensor, |
| beta: torch.Tensor, |
| scale: float | None = None, |
| ) -> torch.Tensor: |
| """Run FLA's Triton chunk_kda. Returns o (B, T, H, V) in v's dtype.""" |
| chunk_kda = _import_fla() |
| if chunk_kda is None: |
| raise RuntimeError("fla.ops.kda.chunk_kda unavailable") |
| |
| |
| |
| out = chunk_kda( |
| q=q, |
| k=k, |
| v=v, |
| g=g, |
| beta=beta, |
| scale=scale, |
| initial_state=None, |
| output_final_state=False, |
| use_qk_l2norm_in_kernel=False, |
| use_gate_in_kernel=False, |
| ) |
| |
| return out[0] if isinstance(out, tuple) else out |
|
|
|
|
| def is_available() -> bool: |
| if _import_fla() is None: |
| return False |
| |
| try: |
| device = torch.device("cuda:0") |
| B, T, H, K, V = 1, 64, 1, 64, 64 |
| q = torch.randn(B, T, H, K, dtype=torch.bfloat16, device=device) |
| k = torch.randn(B, T, H, K, dtype=torch.bfloat16, device=device) |
| v = torch.randn(B, T, H, V, dtype=torch.bfloat16, device=device) |
| g = torch.randn(B, T, H, K, dtype=torch.float32, device=device) * 0.01 |
| beta = torch.sigmoid(torch.randn(B, T, H, dtype=torch.bfloat16, device=device)) |
| sota_forward(q, k, v, g, beta, scale=K ** -0.5) |
| return True |
| except Exception: |
| return False |
|
|