|
|
# Genesis AI Code Bench |
|
|
**Developed by: Within Us AI** |
|
|
Generated: 2026-01-01 |
|
|
|
|
|
A lightweight evaluation harness for Genesis-style datasets that focuses on the signals |
|
|
developers care about in practice: |
|
|
|
|
|
- **Structure validity** (JSON parsing, required fields, schema consistency) |
|
|
- **Tool-trace validity** (JSON array of tool calls with `tool` + `args`) |
|
|
- **Diff validity** (`patch_diff` blocks contain recognizable unified-diff markers) |
|
|
- **Self-grade validity** (score bounds, confidence bounds, presence of notes) |
|
|
- **Governance presence** (audit/tests flags when expected) |
|
|
- **Economics presence** (cost budgets + latency targets) |
|
|
|
|
|
This bench is intentionally fast and offline-friendly. It does not execute repo tests; it |
|
|
scores dataset quality and readiness for downstream training workflows. |
|
|
|
|
|
## Quick start |
|
|
```bash |
|
|
python bench.py --jsonl path/to/train.jsonl --max_rows 5000 |
|
|
``` |
|
|
|
|
|
## Metrics produced |
|
|
- `format_valid_rate` |
|
|
- `required_fields_rate` |
|
|
- `tool_trace_valid_rate` |
|
|
- `patch_diff_valid_rate` |
|
|
- `self_grade_valid_rate` |
|
|
- `governance_present_rate` |
|
|
- `economics_present_rate` |
|
|
- `uniqueness_rate` (hash-based) |
|
|
|
|
|
## Recommended use |
|
|
- Run before upload to ensure Viewer-ready consistency |
|
|
- Run after merges to confirm schema stability |
|
|
- Compare v1.0 vs v1.1 addon impact |
|
|
|