| Vector Addition Problem - Medium Vectors (2^20) |
| ================================================ |
|
|
| Problem Setting |
| --------------- |
| Design and optimize high-performance Triton kernels for vector addition on GPU with medium vectors (1,048,576 elements). This problem focuses on implementing efficient element-wise addition for typical workloads. |
|
|
| The challenge involves optimizing: |
| - **Memory access patterns**: Efficient loading and storing of vector data |
| - **Block sizing**: Optimal block sizes for GPU execution |
| - **Memory bandwidth**: Maximizing throughput for simple arithmetic operations |
| - **Performance benchmarking**: Achieving speedup over PyTorch baseline |
|
|
| This variant tests performance on medium vectors (2^20 = 1,048,576 elements = 4 MB per vector). |
|
|
| Target |
| ------ |
| - **Primary**: Maximize bandwidth (GB/s) over PyTorch baseline (higher is better) |
| - **Secondary**: Ensure correctness |
| - **Tertiary**: Minimize kernel launch overhead |
|
|
| API Specification |
| ----------------- |
| Implement a `Solution` class that returns a Triton kernel implementation: |
|
|
| ```python |
| class Solution: |
| def solve(self, spec_path: str = None) -> dict: |
| """ |
| Returns a dict with either: |
| - {"code": "python_code_string"} |
| - {"program_path": "path/to/kernel.py"} |
| """ |
| # Your implementation |
| pass |
| ``` |
|
|
| Your kernel implementation must provide: |
|
|
| ```python |
| import torch |
| import triton |
| import triton.language as tl |
| |
| def add(x: torch.Tensor, y: torch.Tensor) -> torch.Tensor: |
| """ |
| Element-wise addition of two vectors. |
| |
| Args: |
| x: Input tensor of shape (1048576,) |
| y: Input tensor of shape (1048576,) |
| |
| Returns: |
| Output tensor of shape (1048576,) with x + y |
| """ |
| pass |
| ``` |
| |
| API Usage Notes |
| --------------- |
| - The evaluator looks for an `add` function in the module namespace |
| - Function must handle vector size of exactly 1,048,576 elements |
| - Must use Triton JIT compilation for kernel definition |
| - Should optimize for memory bandwidth |
| - Input tensors are guaranteed to be contiguous and same size |
|
|
| Scoring (0-100) |
| --------------- |
| Performance is measured against CPU baseline and PyTorch GPU baseline: |
|
|
| ``` |
| target = max(2.0 * (pytorch_bandwidth / cpu_bandwidth), 1.0) |
| score = ((custom_bandwidth / cpu_bandwidth - 1.0) / (target - 1.0)) * 100 |
| |
| Where: |
| - custom_bandwidth = your solution's bandwidth |
| - cpu_bandwidth = naive CPU baseline bandwidth |
| - pytorch_bandwidth = PyTorch GPU baseline bandwidth |
| - target = 2x PyTorch performance vs CPU (normalized to custom vs CPU) |
|
|
| Score is clamped to [0, 100] range |
| ``` |
| |
| - 0 points = CPU baseline performance (custom/cpu = 1x) |
| - 50 points = Halfway between CPU baseline and 2x PyTorch performance |
| - 100 points = 2x PyTorch GPU performance vs CPU (custom/cpu = 2 * pytorch/cpu) |
|
|
| Evaluation Details |
| ------------------ |
| - Tested on vector size: 2^20 = 1,048,576 elements |
| - Performance measured in GB/s (bandwidth) |
| - Correctness verified with tolerance: rtol=1e-5, atol=1e-8 |
| - Performance measured using median execution time across 5 samples |
| - Requires CUDA backend and GPU support |
|
|