| name: Research Problem Contribution | |
| about: Add a new research problem to Frontier-CS | |
| title: '[Research] Add {problem_name}' | |
| labels: research-problem | |
| ## Problem Overview | |
| **Problem Name**: | |
| **Category**: <!-- os / hpc / ai / db / pl / security --> | |
| **Difficulty**: <!-- easy / medium / hard --> | |
| ### Description | |
| <!-- Brief description of the problem (2-3 sentences) --> | |
| ### Why This Problem? | |
| <!-- What makes this problem interesting/challenging for frontier models? --> | |
| ## Checklist | |
| ### Required Files | |
| - [ ] `config.yaml` - Dependencies, datasets, runtime config | |
| - [ ] `readme` - Problem description with API specification | |
| - [ ] `set_up_env.sh` - Environment setup script | |
| - [ ] `evaluate.sh` - Evaluation entry point | |
| - [ ] `evaluator.py` - Scoring logic (outputs 0-100 score) | |
| - [ ] `resources/` - Problem-specific code/data | |
| ### Problem Structure | |
| ``` | |
| research/{problem_name}/ | |
| βββ config.yaml | |
| βββ readme | |
| βββ set_up_env.sh | |
| βββ evaluate.sh | |
| βββ evaluator.py | |
| βββ resources/ | |
| βββ ... | |
| ``` | |
| ### Testing | |
| - [ ] Verified `set_up_env.sh` runs successfully | |
| - [ ] Verified `evaluate.sh` runs and outputs a numeric score | |
| - [ ] Tested with a baseline/reference solution | |
| **Test Results** (if available): | |
| ``` | |
| Baseline solution score: | |
| Test environment: | |
| ``` | |
| ## Additional Notes | |
| <!-- Any additional context, known issues, or special requirements --> | |