Vector Addition Problem - Medium Vectors (2^20) ================================================ Problem Setting --------------- Design and optimize high-performance Triton kernels for vector addition on GPU with medium vectors (1,048,576 elements). This problem focuses on implementing efficient element-wise addition for typical workloads. The challenge involves optimizing: - **Memory access patterns**: Efficient loading and storing of vector data - **Block sizing**: Optimal block sizes for GPU execution - **Memory bandwidth**: Maximizing throughput for simple arithmetic operations - **Performance benchmarking**: Achieving speedup over PyTorch baseline This variant tests performance on medium vectors (2^20 = 1,048,576 elements = 4 MB per vector). Target ------ - **Primary**: Maximize bandwidth (GB/s) over PyTorch baseline (higher is better) - **Secondary**: Ensure correctness - **Tertiary**: Minimize kernel launch overhead API Specification ----------------- Implement a `Solution` class that returns a Triton kernel implementation: ```python class Solution: def solve(self, spec_path: str = None) -> dict: """ Returns a dict with either: - {"code": "python_code_string"} - {"program_path": "path/to/kernel.py"} """ # Your implementation pass ``` Your kernel implementation must provide: ```python import torch import triton import triton.language as tl def add(x: torch.Tensor, y: torch.Tensor) -> torch.Tensor: """ Element-wise addition of two vectors. Args: x: Input tensor of shape (1048576,) y: Input tensor of shape (1048576,) Returns: Output tensor of shape (1048576,) with x + y """ pass ``` API Usage Notes --------------- - The evaluator looks for an `add` function in the module namespace - Function must handle vector size of exactly 1,048,576 elements - Must use Triton JIT compilation for kernel definition - Should optimize for memory bandwidth - Input tensors are guaranteed to be contiguous and same size Scoring (0-100) --------------- Performance is measured against CPU baseline and PyTorch GPU baseline: ``` target = max(2.0 * (pytorch_bandwidth / cpu_bandwidth), 1.0) score = ((custom_bandwidth / cpu_bandwidth - 1.0) / (target - 1.0)) * 100 Where: - custom_bandwidth = your solution's bandwidth - cpu_bandwidth = naive CPU baseline bandwidth - pytorch_bandwidth = PyTorch GPU baseline bandwidth - target = 2x PyTorch performance vs CPU (normalized to custom vs CPU) Score is clamped to [0, 100] range ``` - 0 points = CPU baseline performance (custom/cpu = 1x) - 50 points = Halfway between CPU baseline and 2x PyTorch performance - 100 points = 2x PyTorch GPU performance vs CPU (custom/cpu = 2 * pytorch/cpu) Evaluation Details ------------------ - Tested on vector size: 2^20 = 1,048,576 elements - Performance measured in GB/s (bandwidth) - Correctness verified with tolerance: rtol=1e-5, atol=1e-8 - Performance measured using median execution time across 5 samples - Requires CUDA backend and GPU support