andylizf's picture
Upload folder using huggingface_hub
5fed0fc verified
Ragged Attention Optimization Problem
======================================
Problem Setting
---------------
Design and optimize high-performance Triton kernels for ragged attention computation on GPU. This problem focuses on implementing efficient kernels that handle variable-length sequences using ragged attention, where each query row can attend to a different number of key/value rows.
The challenge involves optimizing:
- **Ragged attention**: Efficiently handling variable-length sequences where each row has different attention lengths
- **Memory access patterns**: Efficient loading and storing of Q, K, V tensors with ragged masking
- **Streaming softmax**: Computing softmax in a streaming fashion for numerical stability
- **Row-wise masking**: Correctly masking attention scores based on row_lens
- **Mixed precision**: Handling float16 inputs/outputs with float32 accumulation
- **Block tiling**: Optimal block sizes for GPU execution across different matrix sizes
- **Performance benchmarking**: Achieving speedup over baseline PyTorch implementations
Target
------
- **Primary**: Maximize geometric mean speedup over baseline (higher is better)
- **Secondary**: Ensure correctness across diverse matrix sizes and ragged lengths
- **Tertiary**: Minimize kernel launch overhead and memory usage
API Specification
-----------------
Implement a `Solution` class that returns a Triton kernel implementation:
```python
class Solution:
def solve(self, spec_path: str = None) -> dict:
"""
Returns a dict with either:
- {"code": "python_code_string"}
- {"program_path": "path/to/kernel.py"}
"""
# Your implementation
pass
```
Your kernel implementation must provide:
```python
import torch
import triton
import triton.language as tl
def ragged_attn(Q: torch.Tensor, K: torch.Tensor, V: torch.Tensor, row_lens: torch.Tensor) -> torch.Tensor:
"""
Ragged attention computation.
Args:
Q: Query tensor of shape (M, D) - query features (float16)
K: Key tensor of shape (N, D) - key features (float16)
V: Value tensor of shape (N, Dv) - value features (float16)
row_lens: Row lengths tensor of shape (M,) - number of valid K/V rows per Q row (int32 or int64)
Returns:
Output tensor of shape (M, Dv) - attention output (float16)
Semantics:
For each query row i (0 <= i < M), compute attention over the first row_lens[i] key/value rows.
Specifically:
- scores[i, j] = (Q[i] @ K[j].T) * scale, for j < row_lens[i], else -inf
- P[i] = softmax(scores[i])
- O[i] = P[i] @ V[:row_lens[i]]
"""
pass
```
Scoring
-------
The scoring system evaluates your implementation based on geometric mean speedup over GPU baseline:
- **0 points**: 1x GPU baseline (same speed as PyTorch GPU baseline)
- **100 points**: 3x GPU baseline (3x speedup over PyTorch GPU baseline)
- **Linear interpolation**: Scores between 0-100 are linearly interpolated based on speedup
The evaluation uses the following test cases:
- M (number of query rows): [512, 1024]
- N (number of key/value rows): 1024
- D (model dimension): 64
- Dv (value dimension): 64
- row_lens: Random integers between [min_ratio*N, N] where min_ratio=0.25
Correctness is verified using:
- Relative tolerance: 1e-2
- Absolute tolerance: 5e-3
All tests must pass for a non-zero score. If any test fails correctness, the score is 0.
Example
-------
```python
import torch
import triton
import triton.language as tl
@triton.jit
def _ragged_kernel(Q, K, V, O, ROW_LENS, ...):
# Your kernel implementation
pass
def ragged_attn(Q: torch.Tensor, K: torch.Tensor, V: torch.Tensor, row_lens: torch.Tensor) -> torch.Tensor:
# Your kernel launch logic
pass
```
Constraints
-----------
- All tensors must be CUDA tensors (float16 for Q, K, V; int32/int64 for row_lens)
- Output must be float16
- The implementation must handle variable row lengths correctly
- Accumulation should use float32 for numerical stability
- Must use streaming softmax for numerical stability
Tips
----
1. Use efficient block tiling (BM, BN, BD, BDV) for optimal performance
2. Implement streaming softmax to handle large attention matrices
3. Correctly mask attention scores based on row_lens
4. Load row_lens once per program and broadcast for masking
5. Use proper masking for boundary conditions