| # Submitting Your Results |
|
|
| We welcome submissions from all models and agent frameworks. To have your results included in our leaderboard, please follow the instructions below. |
|
|
| ## Algorithmic Problems |
|
|
| We currently release **1 -- 3 public test case** per problem for local testing and debugging. Full evaluation (with all test cases) is performed on our servers. |
|
|
| ### What to Submit |
|
|
| 1. **Solution files**: `{problem_id}_{model_name}_solution.cpp` for each problem |
| 2. **Model/Agent info**: Name and version of the model or agent framework used |
| 3. **Generation method**: Brief description of how solutions were generated (e.g., one-shot, multi-turn, with/without feedback) |
|
|
| ### Submission Format |
|
|
| Organize your solutions as: |
| ``` |
| submissions/ |
| ├── 1_gpt4_solution.cpp |
| ├── 2_gpt4_solution.cpp |
| ├── ... |
| └── metadata.json |
| ``` |
|
|
| `metadata.json`: |
| ```json |
| { |
| "model": "gpt-4o", |
| "agent_framework": "custom", |
| "generation_method": "one-shot", |
| "date": "2025-01-15", |
| "notes": "Optional additional notes" |
| } |
| ``` |
|
|
| ## Research Problems |
|
|
| Research problems require a `solution.py` file implementing the `Solution` class interface. |
|
|
| ### Problem Structure |
|
|
| Research problems follow a hierarchical structure: |
|
|
| ``` |
| Problem (e.g., gemm_optimization, poc_generation) |
| └── Category (e.g., squares, heap_buffer_overflow) |
| └── Variant (e.g., arvo_21000) |
| ``` |
|
|
| | Level | Example | Description | |
| |-------|---------|-------------| |
| | **Problem** | `gemm_optimization` | Top-level problem domain | |
| | **Category** | `gemm_optimization/squares` | Scores are **aggregated** at this level for leaderboard reporting | |
| | **Variant** | `poc_generation/heap_buffer_overflow/arvo_21000` | Each variant is **evaluated independently** with its own README | |
|
|
| **Key distinction:** |
| - **Evaluation**: Each variant runs independently and produces its own score |
| - **Reporting**: Scores are aggregated by category for the leaderboard (e.g., all `heap_buffer_overflow` variants → one score) |
|
|
| > Note: Some problems have only one level (e.g., `flash_attn`), which functions as both category and variant. |
| |
| ### Problem ID Format |
| |
| Each variant has a unique **Problem ID** based on its path under `research/`. |
| |
| The full list of all evaluatable variants is in [`research/problems.txt`](research/problems.txt) (109 variants total, aggregated into ~50 categories for reporting). |
| |
| | Type | Example Path | Problem ID | |
| |------|-------------|------------| |
| | Single problem | `research/flash_attn` | `flash_attn` | |
| | Problem with variants | `research/gemm_optimization/squares` | `gemm_optimization/squares` | |
| | Nested variants | `research/poc_generation/heap_buffer_overflow/arvo_21000` | `poc_generation/heap_buffer_overflow/arvo_21000` | |
| |
| ### What to Submit |
| |
| 1. **Solution files**: `solution.py` for each problem, placed in a directory matching the Problem ID |
| 2. **Model/Agent info**: Name and version of the model or agent framework used |
| 3. **Local evaluation results** (optional but recommended): Score from running the evaluator locally |
| |
| ### Submission Format |
| |
| Your submission zip should mirror the Problem ID directory structure: |
| |
| ``` |
| submission.zip |
| ├── flash_attn/ |
| │ └── solution.py |
| ├── gemm_optimization/ |
| │ └── squares/ |
| │ └── solution.py |
| ├── cant_be_late/ |
| │ └── high_availability_loose_deadline/ |
| │ └── solution.py |
| ├── poc_generation/ |
| │ └── heap_buffer_overflow/ |
| │ └── arvo_21000/ |
| │ └── solution.py |
| └── metadata.json |
| ``` |
| |
| **Important**: The directory structure must exactly match the Problem ID. For example: |
| - `flash_attn/solution.py` |
| - `gemm_optimization/squares/solution.py` |
| |
| Each `solution.py` must implement: |
| ```python |
| class Solution: |
| def __init__(self): |
| pass |
| |
| def solve(self, *args): |
| # Returns: solution output (format varies by problem) |
| pass |
| ``` |
| |
| ### metadata.json |
|
|
| ```json |
| { |
| "model": "gpt-4o", |
| "agent_framework": "custom", |
| "generation_method": "one-shot", |
| "date": "2025-01-15", |
| "problems_solved": [ |
| "flash_attn", |
| "gemm_optimization/squares", |
| "cant_be_late/high_availability_loose_deadline" |
| ], |
| "notes": "Optional additional notes" |
| } |
| ``` |
|
|
| ### Running Local Evaluation |
|
|
| Before submitting, you can verify your solutions locally: |
|
|
| ```bash |
| # Evaluate a single solution |
| frontier-eval flash_attn solution.py |
| |
| # Batch evaluation with progress tracking |
| frontier-eval batch --pairs-file pairs.txt --results-dir results/ |
| |
| # Batch evaluation with SkyPilot (cloud) |
| frontier-eval batch --pairs-file pairs.txt --skypilot --max-concurrent 4 |
| ``` |
|
|
| ## How to Submit |
|
|
| Send your submission to: |
| - **Email**: qmang@berkeley.edu or wenhao.chai@princeton.edu |
|
|
| Please include: |
| 1. A zip/tar archive of your solutions following the format above |
| 2. `metadata.json` with model and method information |
| 3. (Optional) Local evaluation results if you ran them |
|
|
| ## Leaderboard |
|
|
| Accepted submissions will be evaluated on our full test suite and results will be published on the [Frontier-CS Leaderboard](https://frontier-cs.org). |
|
|