| Ragged Attention Optimization Problem |
| ====================================== |
|
|
| Problem Setting |
| --------------- |
| Design and optimize high-performance Triton kernels for ragged attention computation on GPU. This problem focuses on implementing efficient kernels that handle variable-length sequences using ragged attention, where each query row can attend to a different number of key/value rows. |
|
|
| The challenge involves optimizing: |
| - **Ragged attention**: Efficiently handling variable-length sequences where each row has different attention lengths |
| - **Memory access patterns**: Efficient loading and storing of Q, K, V tensors with ragged masking |
| - **Streaming softmax**: Computing softmax in a streaming fashion for numerical stability |
| - **Row-wise masking**: Correctly masking attention scores based on row_lens |
| - **Mixed precision**: Handling float16 inputs/outputs with float32 accumulation |
| - **Block tiling**: Optimal block sizes for GPU execution across different matrix sizes |
| - **Performance benchmarking**: Achieving speedup over baseline PyTorch implementations |
|
|
| Target |
| ------ |
| - **Primary**: Maximize geometric mean speedup over baseline (higher is better) |
| - **Secondary**: Ensure correctness across diverse matrix sizes and ragged lengths |
| - **Tertiary**: Minimize kernel launch overhead and memory usage |
|
|
| API Specification |
| ----------------- |
| Implement a `Solution` class that returns a Triton kernel implementation: |
|
|
| ```python |
| class Solution: |
| def solve(self, spec_path: str = None) -> dict: |
| """ |
| Returns a dict with either: |
| - {"code": "python_code_string"} |
| - {"program_path": "path/to/kernel.py"} |
| """ |
| # Your implementation |
| pass |
| ``` |
|
|
| Your kernel implementation must provide: |
|
|
| ```python |
| import torch |
| import triton |
| import triton.language as tl |
| |
| def ragged_attn(Q: torch.Tensor, K: torch.Tensor, V: torch.Tensor, row_lens: torch.Tensor) -> torch.Tensor: |
| """ |
| Ragged attention computation. |
| |
| Args: |
| Q: Query tensor of shape (M, D) - query features (float16) |
| K: Key tensor of shape (N, D) - key features (float16) |
| V: Value tensor of shape (N, Dv) - value features (float16) |
| row_lens: Row lengths tensor of shape (M,) - number of valid K/V rows per Q row (int32 or int64) |
| |
| Returns: |
| Output tensor of shape (M, Dv) - attention output (float16) |
| |
| Semantics: |
| For each query row i (0 <= i < M), compute attention over the first row_lens[i] key/value rows. |
| Specifically: |
| - scores[i, j] = (Q[i] @ K[j].T) * scale, for j < row_lens[i], else -inf |
| - P[i] = softmax(scores[i]) |
| - O[i] = P[i] @ V[:row_lens[i]] |
| """ |
| pass |
| ``` |
| |
| Scoring |
| ------- |
| The scoring system evaluates your implementation based on geometric mean speedup over GPU baseline: |
|
|
| - **0 points**: 1x GPU baseline (same speed as PyTorch GPU baseline) |
| - **100 points**: 3x GPU baseline (3x speedup over PyTorch GPU baseline) |
| - **Linear interpolation**: Scores between 0-100 are linearly interpolated based on speedup |
|
|
| The evaluation uses the following test cases: |
| - M (number of query rows): [512, 1024] |
| - N (number of key/value rows): 1024 |
| - D (model dimension): 64 |
| - Dv (value dimension): 64 |
| - row_lens: Random integers between [min_ratio*N, N] where min_ratio=0.25 |
|
|
| Correctness is verified using: |
| - Relative tolerance: 1e-2 |
| - Absolute tolerance: 5e-3 |
|
|
| All tests must pass for a non-zero score. If any test fails correctness, the score is 0. |
|
|
| Example |
| ------- |
| ```python |
| import torch |
| import triton |
| import triton.language as tl |
| |
| @triton.jit |
| def _ragged_kernel(Q, K, V, O, ROW_LENS, ...): |
| # Your kernel implementation |
| pass |
|
|
| def ragged_attn(Q: torch.Tensor, K: torch.Tensor, V: torch.Tensor, row_lens: torch.Tensor) -> torch.Tensor: |
| # Your kernel launch logic |
| pass |
| ``` |
| |
| Constraints |
| ----------- |
| - All tensors must be CUDA tensors (float16 for Q, K, V; int32/int64 for row_lens) |
| - Output must be float16 |
| - The implementation must handle variable row lengths correctly |
| - Accumulation should use float32 for numerical stability |
| - Must use streaming softmax for numerical stability |
|
|
| Tips |
| ---- |
| 1. Use efficient block tiling (BM, BN, BD, BDV) for optimal performance |
| 2. Implement streaming softmax to handle large attention matrices |
| 3. Correctly mask attention scores based on row_lens |
| 4. Load row_lens once per program and broadcast for masking |
| 5. Use proper masking for boundary conditions |
|
|
|
|