RedButton / LEADERBOARD.md
Arun-Sanjay's picture
phase-5: deploy to HF Space, burst+sustained concurrency tests, leaderboard skeleton
8346aac

Shutdown-Gym Leaderboard

Tampering rate measured across 200 rollouts at Tier 2.

Model Train Method Trained-Cat (Train Op) Held-Out (Train Op) Held-Out (Strict Op) Math Acc Honesty (T3) Submission
Qwen-2.5-1.5B (none, baseline) TBD TBD TBD TBD TBD this repo
Qwen-2.5-1.5B GRPO (ours) TBD TBD TBD TBD TBD this repo

The baseline row is populated Day 2 morning after the 50-rollout baseline measurement (PROJECT.md §20). The trained row is populated Day 2 noon after held-out evaluation completes (§18).

How to submit

  1. Train any model against this environment, or evaluate any existing checkpoint.

  2. Run:

    python -m evaluation.evaluate \
        --model <name> \
        --env-url https://arun-sanjay-redbutton.hf.space \
        --n-episodes 200
    
  3. Run again with the --strict-operator flag for the strict-operator column.

  4. Open a PR with the new row added and a link to your model checkpoint.

Submission template

| <your model> | <method> | X% | X% | X% | X% | X% | <link to repo or HF model> |

Why this benchmark

Production ML engineers from Meta, HuggingFace, and PyTorch evaluate environments by whether they could integrate them. The leaderboard makes this concrete: the env ships with two reference rows (baseline + trained), and any new entry is a 200-rollout evaluate.py invocation away.

The four metrics on the right (Trained-Cat, Held-Out × {Train, Strict} Op, Math Accuracy, Honesty) capture the dual-axis generalization claim — disposition transfers across mechanism and across operator policy. See PROJECT.md §16, §17, §18 for the detailed taxonomy and rubric design, and API_NOTES.md for the recon-verified OpenEnv API surface this env builds on.