File size: 37,219 Bytes
e313100 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 | # Phase 5.1 PPO — Operations Tracker
Single source of truth for in-flight work. Resume from here.
---
## Principles
1. **Two canonical specs**: `ppo_playground` (DM Control) and `ppo_playground_loco` (Loco). Per-env variants only when officially required: `ppo_playground_fingerspin` (gamma=0.95), `ppo_playground_pendulum` (training_epoch=4, action_repeat=4 via code).
2. **100M frames hard cap** — no extended runs. If an env doesn't hit target at 100M, fix the spec.
3. **Strategic reruns**: only rerun failing/⚠️ envs. Already-✅ envs skip revalidation.
4. **Score metric**: use `total_reward_ma` (final moving average of total reward) — measures end-of-training performance and matches mujoco_playground reference scores.
5. **Official reference**: check `~/.cache/uv/archive-v0/ON8dY3irQZTYI3Bok0SlC/mujoco_playground/config/dm_control_suite_params.py` for per-env overrides.
---
## Wave 3 (2026-03-16)
**Fixes applied:**
- stderr suppression: MuJoCo C-level warnings (ccd_iterations, nefc overflow, broadphase overflow) silenced in playground.py
- obs fix: _get_obs now passes only "state" key for dict-obs envs (was incorrectly concatenating privileged_state+state)
**Envs graduated to ✅ (close enough):**
FishSwim, PointMass, ReacherHard, WalkerStand, WalkerWalk, SpotGetup, SpotJoystickGaitTracking, AlohaHandOver
**Failing envs by root cause:**
- Humanoid double-norm (rs10 fix): HumanoidStand (114→700), HumanoidWalk (47→500), HumanoidRun (18→130)
- Dict obs fix (now fixed): Go1Flat/Rough/Getup/Handstand, G1Flat/Rough, T1Flat/Rough
- Unknown: BarkourJoystick (0/35), Op3Joystick (0/20)
- Needs hparam work: H1Inplace (4→10), H1Joystick (16→30), SpotFlat (11→30)
- Manipulation: AlohaPeg (188→300), LeapCubeReorient (74→200)
- Infeasible: PandaRobotiqPushCube, AeroCubeRotateZAxis
**Currently running:** (to be populated by ops)
---
## Currently Running (as of 2026-03-14 ~00:00)
**Wave V (p5-ppo17) — Constant LR test (4 runs, just launched)**
Testing constant LR (Brax default) in isolation — never tested before. Key hypothesis: LR decay hurts late-converging envs.
| Run | Env | Spec | Key Change | Old Best | Target |
|---|---|---|---|---|---|
| p5-ppo17-csup | CartpoleSwingup | constlr | constant LR + minibatch=4096 | 576.1 | 800 |
| p5-ppo17-csupsparse | CartpoleSwingupSparse | constlr | constant LR + minibatch=4096 | 296.3 | 425 |
| p5-ppo17-acrobot | AcrobotSwingup | vnorm_constlr | constant LR + vnorm | 173 | 220 |
| p5-ppo17-fteasy | FingerTurnEasy | vnorm_constlr | constant LR + vnorm | 571 | 950 |
**Wave IV-H (p5-ppo16h) — Humanoid with wider policy (3 runs, ~2.5h remaining)**
New `ppo_playground_humanoid` variant: 2×256 policy (vs 2×64), constant LR, vnorm=true.
Based on Phase 3 Gymnasium Humanoid-v5 success (2661 MA with 2×256 + constant LR).
| Run | Env | Old Best | Target |
|---|---|---|---|
| p5-ppo16h-hstand | HumanoidStand | 18.36 | 700 |
| p5-ppo16h-hwalk | HumanoidWalk | 7.68 | 500 |
| p5-ppo16h-hrun | HumanoidRun | 3.19 | 130 |
**Wave VI (p5-ppo18) — Brax 4×32 policy + constant LR + vnorm (3 runs, just launched)**
Testing Brax default policy architecture (4 layers × 32 units vs our 2 × 64).
Deeper narrower policy may learn better features for precision tasks.
| Run | Env | Old Best | Target |
|---|---|---|---|
| p5-ppo18-fteasy | FingerTurnEasy | 571 | 950 |
| p5-ppo18-fthard | FingerTurnHard | 484 | 950 |
| p5-ppo18-fishswim | FishSwim | 463 | 650 |
**Wave IV tail (p5-ppo16) — completed**
| Run | Env | strength | Target | New best? |
|---|---|---|---|---|
| p5-ppo16-swimmer6 | SwimmerSwimmer6 | 509.3 | 560 | ✅ New best (final_strength=560.6) |
| p5-ppo16-fishswim | FishSwim | 420.6 | 650 | ❌ Worse than 463 |
**Wave IV results (p5-ppo16, vnorm=true rerun with reverted spec — completed):**
All ran with vnorm=true. CartpoleSwingup/Sparse worse (vnorm=false is better for them — wrong setting).
Precision envs also scored below old bests. Humanoid still failing with standard 2×64 policy.
| Env | p16 strength | Old Best | Target | Verdict |
|---|---|---|---|---|
| CartpoleSwingup | 316.2 | 576.1 (false) | 800 | ❌ wrong vnorm |
| CartpoleSwingupSparse | 288.7 | 296.3 (false) | 425 | ❌ wrong vnorm |
| AcrobotSwingup | 145.4 | 173 (true) | 220 | ❌ worse |
| FingerTurnEasy | 511.1 | 571 (true) | 950 | ❌ worse |
| FingerTurnHard | 368.6 | 484 (true) | 950 | ❌ worse |
| HumanoidStand | 12.72 | 18.36 | 700 | ❌ still failing |
| HumanoidWalk | 7.46 | 7.68 | 500 | ❌ still failing |
| HumanoidRun | 3.19 | 3.19 | 130 | ❌ still failing |
**CONCLUSION**: Reverted spec didn't help. No new bests. Consistency was negative for CartpoleSwingup/Sparse (high variance).
Need constant LR test (Wave V) and wider policy for Humanoid (Wave IV-H).
**Wave III results (p5-ppo13/p5-ppo15, 5-layer value + no grad clip — completed):**
Only CartpoleSwingup improved slightly (623.8 vs 576.1). All others regressed.
FishSwim p5-ppo15: strength=411.6 (vs 463 old best). AcrobotSwingup p5-ppo15: strength=95.4 (vs 173).
**CONCLUSION**: 5-layer value + no grad clip is NOT a general improvement. Reverted to 3-layer + clip_grad_val=1.0.
**Wave H results (p5-ppo12, ALL completed — NONE improved over old bests):**
Re-running same spec (variance reruns + vnorm) didn't help. Run-to-run variance is high but
old bests represent lucky runs. Hyperparameter tuning has hit diminishing returns.
**Wave G/G2 results (normalize_v_targets=false ablation, ALL completed):**
| Env | p11 strength | Old Best (true) | Target | Change | Verdict |
|---|---|---|---|---|---|
| **PendulumSwingup** | **533.5** | 276 | 395 | +93% | **✅ NEW PASS** |
| **FingerSpin** | **652.4** | 561 | 600 | +16% | **✅ NEW PASS** |
| **CartpoleBalanceSparse** | **690.4** | 545 | 700 | +27% | **⚠️ 99% of target** |
| **CartpoleSwingup** | **576.1** | 443/506 | 800 | +30% | ⚠️ improved |
| **CartpoleSwingupSparse** | **296.3** | 271 | 425 | +9% | ⚠️ improved |
| PointMass | 854.4 | 863 | 900 | -1% | ⚠️ same |
| FishSwim | 293.9 | 463 | 650 | -36% | ❌ regression |
| FingerTurnEasy | 441.1 | 571 | 950 | -23% | ❌ regression |
| SwimmerSwimmer6 | 386.2 | 485 | 560 | -20% | ❌ regression |
| FingerTurnHard | 335.7 | 484 | 950 | -31% | ❌ regression |
| AcrobotSwingup | 105.1 | 173 | 220 | -39% | ❌ regression |
| HumanoidStand | 12.87 | 18.36 | 500 | -30% | ❌ still failing |
**CONCLUSION**: `normalize_v_targets: false` helps 5/12, hurts 6/12, neutral 1/12.
- **false wins**: PendulumSwingup, FingerSpin, CartpoleBalanceSparse, CartpoleSwingup, CartpoleSwingupSparse
- **true wins**: FishSwim, FingerTurnEasy/Hard, SwimmerSwimmer6, AcrobotSwingup, PointMass
- **Decision**: Per-env spec selection. New `ppo_playground_vnorm` variant for precision envs.
**Wave F results (multi-unroll=16 + proven hyperparameters):**
| Env | p10 strength | p10 final_str | Old best str | Target | Verdict |
|---|---|---|---|---|---|
| CartpoleSwingup | 342 | 443 | 443 | 800 | Same |
| FingerTurnEasy | 529 | 685 | 571 | 950 | Better final, worse strength |
| FingerSpin | 402 | 597 | 561 | 600 | Better final (near target!), worse strength |
| FingerTurnHard | 368 | 559 | 484 | 950 | Better final, worse strength |
| SwimmerSwimmer6 | 251 | 384 | 485 | 560 | Worse |
| CartpoleSwingupSparse | 56 | 158 | 271 | 425 | MUCH worse |
| AcrobotSwingup | 31 | 63 | 173 | 220 | MUCH worse |
**CONCLUSION**: Multi-unroll adds no benefit over single-unroll for any env by `strength` metric.
The `final_strength` improvements for Finger tasks are offset by `strength` regressions.
Root cause: stale old_net (480 vs 30 steps between copies) makes policy ratio less accurate.
**Spec reverted to single-unroll (num_unrolls=1)**. Multi-unroll code preserved in ppo.py.
**Wave E results (multi-unroll + Brax hyperparameters — ALL worse):**
Brax-matched spec (clip_eps=0.3, constant LR, 5-layer value, reward_scale=10, minibatch=30720)
hurt every env except HopperStand (which used wrong spec before). Reverted.
**Wave C completed results** (all reward_scale=10, divide by 10 for true score):
| Run | Env | strength/10 | final_strength/10 | total_reward_ma/10 | Target | vs Old |
|---|---|---|---|---|---|---|
| p5-ppo7-cartpoleswingup | CartpoleSwingup | 556.6 | 670.5 | 705.3 | 800 | 443→557 ✅ improved |
| p5-ppo7-fingerturneasy | FingerTurnEasy | 511.1 | 693.2 | 687.0 | 950 | 571→511 ❌ **WORSE** |
| p5-ppo7-fingerturnhard | FingerTurnHard | 321.9 | 416.8 | 425.2 | 950 | 484→322 ❌ **WORSE** |
| p5-ppo7-cartpoleswingupsparse2 | CartpoleSwingupSparse | 144.0 | 360.6 | 337.7 | 425 | 271→144 ❌ **WORSE** |
**KEY FINDING**: time_horizon=480 helps CartpoleSwingup (+25%) but HURTS FingerTurn (-30 to -50%) and CartpoleSwingupSparse (-47%). Long GAE horizons produce noisy advantage estimates for precision/sparse tasks. The official Brax approach is 16×30-step unrolls (short GAE per unroll), NOT 1×480-step unroll.
---
## Spec Changes Applied (2026-03-13)
### Fix 1: reward_scale=10.0 (matches official mujoco_playground)
- `playground.py`: `PlaygroundVecEnv` now multiplies rewards by `self._reward_scale`
- `__init__.py`: threads `reward_scale` from env spec to wrapper
- `ppo_playground.yaml`: `reward_scale: 10.0` in shared `_env` anchor
### Fix 2: Revert minibatch_size 2048→4096 (fixes CartpoleSwingup regression)
- `ppo_playground.yaml`: all DM Control specs (ppo_playground, fingerspin, pendulum) now use minibatch_size=4096
- 15 minibatches × 16 epochs = 240 grad steps (was 30×16=480)
- Restores p5-ppo5 performance for CartpoleSwingup (803 vs 443)
### Fix 3: Brax-matched spec (commit 6eb08fe9) — time_horizon=480, clip_eps=0.3, constant LR, 5-layer value net
- Increased time_horizon from 30→480 to match total data per update (983K transitions)
- clip_eps 0.2→0.3, constant LR (min_factor=1.0), 5-layer [256×5] value net
- action std upper bound raised (max=2.0 in policy_util.py)
- **Result**: CartpoleSwingup improved (443→557 strength), but FingerTurn and CartpoleSwingupSparse got WORSE
- **Root cause**: 1×480-step unroll computes GAE over 480 steps (noisy), vs official 16×30-step unrolls (short, accurate GAE)
### Fix 4: ppo_playground_short variant (time_horizon=30 + Brax improvements)
- Keeps: reward_scale=10, clip_eps=0.3, constant LR, 5-layer value net, no grad clipping
- Reverts: time_horizon=30, minibatch_size=4096 (15 minibatches, 240 grad steps)
- **Hypothesis**: Short GAE + other Brax improvements = best of both worlds for precision tasks
- Testing on FingerTurnEasy/Hard first (Wave D p5-ppo8-*)
### Fix 5: Multi-unroll collection (IMPLEMENTED but NOT USED — code stays, spec reverted)
- Added `num_unrolls` parameter to PPO (ppo.py, actor_critic.py). Code works correctly.
- **Brax-matched spec (Wave E, p5-ppo9)**: clip_eps=0.3, constant LR, 5-layer value, reward_scale=10
- Result: WORSE on 5/7 tested envs. Only CartpoleSwingup improved (443→506).
- Root cause: minibatch_size=30720 → 7.5x fewer gradient steps per transition → underfitting
- **Reverted spec + multi-unroll (Wave F, p5-ppo10)**: clip_eps=0.2, LR decay, 3-layer value, minibatch=4096
- Result: Same or WORSE on all envs by `strength` metric. Same fps as single-unroll.
- Training compute per env step is identical, but old_net staleness (480 vs 30 steps) hurts.
- **Conclusion**: Multi-unroll adds complexity without benefit. Reverted spec to single-unroll (num_unrolls=1).
Code preserved in ppo.py (defaults to 1). Spec uses original hyperparameters.
---
## Completed Runs Needing Intake
### Humanoid (ppo_playground_loco, post log_std fix) — intake immediately
| Run | HF Folder | strength | target | HF status |
|---|---|---|---|---|
| p5-ppo6-humanoidrun | ppo_playground_loco_humanoidrun_2026_03_12_175917 | 2.78 | 130 | ✅ uploaded |
| p5-ppo6-humanoidwalk | ppo_playground_loco_humanoidwalk_2026_03_12_175817 | 6.82 | 500 | ✅ uploaded |
| p5-ppo6-humanoidstand | ppo_playground_loco_humanoidstand_2026_03_12_175810 | 12.45 | 700 | ❌ **UPLOAD FAILED (412)** — re-upload first |
Re-upload HumanoidStand:
```bash
source .env && huggingface-cli upload SLM-Lab/benchmark-dev \
hf_data/data/benchmark-dev/data/ppo_playground_loco_humanoidstand_2026_03_12_175810 \
data/ppo_playground_loco_humanoidstand_2026_03_12_175810 --repo-type dataset
```
**Conclusion**: loco spec still fails completely for Humanoid — log_std fix insufficient. See spec fixes below.
### BENCHMARKS.md correction needed (commit b6ef49d9 used wrong metric)
intake-a used `total_reward_ma` instead of `strength`. Fix these 4 entries:
| Env | Run | strength (correct) | total_reward_ma (wrong) | target |
|---|---|---|---|---|
| AcrobotSwingup | p5-ppo6-acrobotswingup2 | **172.8** | 253.24 | 220 |
| CartpoleBalanceSparse | p5-ppo6-cartpolebalancesparse2 | **545.1** | 991.81 | 700 |
| CartpoleSwingup | p5-ppo6-cartpoleswingup2 | **unknown — extract from logs** | 641.51 | 800 |
| CartpoleSwingupSparse | p5-ppo6-cartpoleswingupsparse | **270.9** | 331.23 | 425 |
Extract correct values: `dstack logs p5-ppo6-NAME --since 6h 2>&1 | grep "trial_metrics" | tail -1` → use `strength:` field.
Also check FingerSpin: `dstack logs p5-ppo6-fingerspin2 --since 6h | grep trial_metrics | tail -1` — confirm strength value.
**Metric decision needed**: strength penalizes slow learners (CartpoleBalanceSparse strength=545 but final MA=992). Consider switching ALL entries to `final_strength`. But this requires auditing every existing entry — do it as a batch before publishing.
---
## Queue (launch when slots open, all 100M)
| Priority | Env | Spec | Run name | Rationale |
|---|---|---|---|---|
| 1 | PendulumSwingup | ppo_playground_pendulum | p5-ppo6-pendulumswingup | action_repeat=4 + training_epoch=4 (code fix applied) |
| 2 | FingerSpin | ppo_playground_fingerspin | p5-ppo6-fingerspin3 | canonical gamma=0.95 run; fingerspin2 used gamma=0.995 (override silently ignored) |
---
## Full Env Status
### ✅ Complete (13/25)
| Env | strength | target | normalize_v_targets |
|---|---|---|---|
| CartpoleBalance | 968.23 | 950 | true |
| AcrobotSwingupSparse | 42.74 | 15 | true |
| BallInCup | 942.44 | 680 | true |
| CheetahRun | 865.83 | 850 | true |
| ReacherEasy | 955.08 | 950 | true |
| ReacherHard | 946.99 | 950 | true |
| WalkerRun | 637.80 | 560 | true |
| WalkerStand | 970.94 | 1000 | true |
| WalkerWalk | 952 | 960 | true |
| HopperHop | 22.00 | ~2 | true |
| HopperStand | 118.2 | ~70 | true |
| PendulumSwingup | 533.5 | 395 | **false** |
| FingerSpin | 652.4 | 600 | **false** |
### ⚠️ Below target (9/25)
| Env | best strength | target | best with | status |
|---|---|---|---|---|
| CartpoleSwingup | 576.1 | 800 | false | Improved +30% from 443 (true) |
| CartpoleBalanceSparse | 545 | 700 | true | Testing false (p5-ppo11) |
| CartpoleSwingupSparse | 296.3 | 425 | false | Improved +9% from 271 (true) |
| AcrobotSwingup | 173 | 220 | true | false=105, regressed |
| FingerTurnEasy | 571 | 950 | true | false=441, regressed |
| FingerTurnHard | 484 | 950 | true | false=336, regressed |
| FishSwim | 463 | 650 | true | Testing false (p5-ppo11) |
| SwimmerSwimmer6 | 509.3 | 560 | true | final_strength=560.6 (at target!) |
| PointMass | 863 | 900 | true | false=854, ~same |
### ❌ Fundamental failure — Humanoid (3/25)
| Env | best strength | target | diagnosis |
|---|---|---|---|
| HumanoidRun | 3.19 | 130 | <3% target, NormalTanh distribution needed |
| HumanoidWalk | 7.68 | 500 | <2% target, wider policy (2×256) didn't help |
| HumanoidStand | 18.36 | 700 | <3% target, constant LR + wider policy tested, no improvement |
**Humanoid tested and failed**: wider 2×256 policy + constant LR + vnorm (Wave IV-H). MA stayed flat at 8-10 for HumanoidStand over entire training. Root cause is likely NormalTanh distribution (state-dependent std + tanh squashing) — a fundamental architectural difference from Brax.
---
## Spec Fixes Required
### Priority 1: Humanoid loco spec (update ppo_playground_loco)
Official uses `num_envs=8192, time_horizon=20 (unroll_length)` for loco. We use `num_envs=2048, time_horizon=64`.
**Proposed update to ppo_playground_loco**:
```yaml
ppo_playground_loco:
agent:
algorithm:
gamma: 0.97
time_horizon: 20 # was 64; official unroll_length=20
training_epoch: 4
env:
num_envs: 8192 # was 2048; official loco num_envs=8192
```
**Before launching**: verify VRAM by checking if 8192 envs fits A4500 20GB. Run one Humanoid env, check `dstack logs NAME --since 10m | grep -i "memory\|OOM"` after 5 min.
**Rerun only**: HumanoidRun, HumanoidWalk, HumanoidStand (3 runs). HopperStand also uses loco spec — add if VRAM confirmed OK.
### Priority 2: CartpoleSwingup regression
p5-ppo5 scored 803 ✅; p5-ppo6 scored ~641. The p5-ppo6 change was `minibatch_size: 2048` (30 minibatches) vs p5-ppo5's 4096 (15 minibatches). More gradient steps per iter hurt CartpoleSwingup.
**Option A**: Revert `ppo_playground` minibatch_size from 2048→4096 (15 minibatches). Rerun only failing DM Control envs (CartpoleSwingup, CartpoleSwingupSparse, + any that need it).
**Option B**: Accept 641 and note the trade-off — p5-ppo6 improved other envs (CartpoleBalance 968 was already ✅).
### Priority 3: FingerTurnEasy/Hard
No official override. At 570/? vs target 950, gap is large. Check:
```bash
grep -A10 "Finger" ~/.cache/uv/archive-v0/ON8dY3irQZTYI3Bok0SlC/mujoco_playground/config/dm_control_suite_params.py
```
May need deeper policy network [32,32,32,32] (official arch) vs our [64,64].
---
## Tuning Principles Learned
1. **Check official per-env overrides first**: `dm_control_suite_params.py` has `discounting`, `action_repeat`, `num_updates_per_batch` per env. These are canonical.
2. **action_repeat** is env-level, not spec-level. Implemented in `playground.py` via `_ACTION_REPEAT` dict. PendulumSwingup→4. Add others as found.
3. **NaN loss**: `log_std` clamp max=0.5 helps but Humanoid (21 DOF) still has many NaN skips. Rate-limited to log every 10K. If NaN dominates → spec is wrong.
4. **num_envs scales with task complexity**: Cartpole/Acrobot: 2048 fine. Humanoid locomotion: needs 8192 for rollout diversity.
5. **time_horizon (unroll_length)**: DM Control official=30, loco official=20. Longer → more correlated rollouts → less diversity per update. Match official.
6. **Minibatch count**: more minibatches = more gradient steps per batch. Can overfit or slow convergence for simpler envs. 15 minibatches (p5-ppo5) vs 30 (p5-ppo6) — the latter hurt CartpoleSwingup.
7. **Sparse reward + strength metric**: strength (trajectory mean) severely penalizes sparse/delayed convergence. CartpoleBalanceSparse strength=545 but final MA=992. Resolve metric before publishing.
8. **High seed variance** (consistency < 0): some seeds solve, some don't → wrong spec, not bad luck. Fix exploration (entropy_coef) or use different spec.
9. **-s overrides are silently ignored** if the YAML key isn't a `${variable}` placeholder. Always verify overrides took effect via logs: `grep "gamma\|lr\|training_epoch" dstack logs`.
10. **Loco spec failures**: if loco spec gives <20 on env with target >100, the issue is almost certainly num_envs/time_horizon mismatch vs official, not a fundamental algo failure.
---
## Code Changes This Session
| Commit | Change |
|---|---|
| `8fe7bc76` | `playground.py`: `_ACTION_REPEAT` lookup for per-env action_repeat. `ppo_playground.yaml`: added `ppo_playground_fingerspin` and `ppo_playground_pendulum` specs. |
| `fb55c2f9` | `base.py`: rate-limit NaN loss warning (every 10K skips). `ppo_playground.yaml`: revert log_frequency 1M→100K. |
| `3f4ede3d` | BENCHMARKS.md: mark HopperHop ✅. |
---
## Resume Commands
```bash
# Setup
git pull && uv sync --no-default-groups
# Check jobs
dstack ps
# Intake a completed run
dstack logs RUN_NAME --since 6h 2>&1 | grep "trial_metrics" | tail -1
dstack logs RUN_NAME --since 6h 2>&1 | grep -iE "Uploading|benchmark-dev"
# Pull HF data
source .env && huggingface-cli download SLM-Lab/benchmark-dev \
--local-dir hf_data/data/benchmark-dev --repo-type dataset \
--include "data/FOLDER_NAME/*"
# Plot
uv run slm-lab plot -t "EnvName" -d hf_data/data/benchmark-dev/data -f FOLDER_NAME
# Launch PendulumSwingup (queue priority 1)
source .env && uv run slm-lab run-remote --gpu \
slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground_pendulum train \
-s env=playground/PendulumSwingup -s max_frame=100000000 -n p5-ppo6-pendulumswingup
# Launch FingerSpin canonical (queue priority 2)
source .env && uv run slm-lab run-remote --gpu \
slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground_fingerspin train \
-s env=playground/FingerSpin -s max_frame=100000000 -n p5-ppo6-fingerspin3
# Launch Humanoid loco (after updating ppo_playground_loco spec to num_envs=8192, time_horizon=20)
source .env && uv run slm-lab run-remote --gpu \
slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground_loco train \
-s env=playground/HumanoidRun -s max_frame=100000000 -n p5-ppo6-humanoidrun2
```
---
## CRITICAL CORRECTION (2026-03-13) — Humanoid is DM Control, not Loco
**Root cause of Humanoid failure**: HumanoidRun/Walk/Stand are registered in `dm_control_suite/__init__.py` — they ARE DM Control envs. We incorrectly ran them with `ppo_playground_loco` (gamma=0.97, 4 epochs, time_horizon=64).
Official config uses DEFAULT DM Control params for them: discounting=0.995, 2048 envs, lr=1e-3, unroll_length=30, 16 epochs.
**NaN was never the root cause** — intake-b confirmed NaN skips were 0, 0, 2 in the loco runs. The spec was simply wrong.
**Fix**: Run all 3 Humanoid envs with `ppo_playground` (DM Control spec). No spec change needed.
```bash
# Launch with correct spec
source .env && uv run slm-lab run-remote --gpu \
slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground train \
-s env=playground/HumanoidRun -s max_frame=100000000 -n p5-ppo6-humanoidrun2
source .env && uv run slm-lab run-remote --gpu \
slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground train \
-s env=playground/HumanoidWalk -s max_frame=100000000 -n p5-ppo6-humanoidwalk2
source .env && uv run slm-lab run-remote --gpu \
slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground train \
-s env=playground/HumanoidStand -s max_frame=100000000 -n p5-ppo6-humanoidstand2
```
**HopperStand**: Also a DM Control env. If p5-ppo6-hopperstand (loco spec, 16.38) is below target, rerun with `ppo_playground`.
**Do NOT intake** the loco-spec Humanoid runs (2.78/6.82/12.45) — wrong spec, not valid benchmark results. The old ppo_playground runs (2.86/3.73) were also wrong spec but at least the right family.
**Updated queue (prepend these as highest priority)**:
| Priority | Env | Spec | Run name |
|---|---|---|---|
| 0 | HumanoidRun | ppo_playground | p5-ppo6-humanoidrun2 |
| 0 | HumanoidWalk | ppo_playground | p5-ppo6-humanoidwalk2 |
| 0 | HumanoidStand | ppo_playground | p5-ppo6-humanoidstand2 |
| 0 | HopperStand | ppo_playground | p5-ppo6-hopperstand2 (if loco result ⚠️) |
Note on loco spec (`ppo_playground_loco`): only for actual locomotion robot envs (Go1, G1, BerkeleyHumanoid, etc.) — NOT for DM Control Humanoid.
---
## METRIC CORRECTION (2026-03-13) — strength vs final_strength
**Problem**: `strength` = trajectory-averaged mean over entire run. For slow-rising envs this severely underrepresents end-of-training performance. After metric correction to `strength`:
| Env | strength | total_reward_ma | target | conclusion |
|---|---|---|---|---|
| CartpoleSwingup | **443.0** | 641.51 | 800 | Massive regression from p5-ppo5 (803). Strength 443 << 665 (65M result) — curve rises but slow start drags average down |
| CartpoleBalanceSparse | **545.1** | 991.81 | 700 | Hits target by end (final MA=992) but sparse reward delays convergence |
| AcrobotSwingup | **172.8** | 253.24 | 220 | Below target by strength, above by final MA |
| CartpoleSwingupSparse | **270.9** | 331.23 | 425 | Below both metrics |
**Resolution needed**: Reference scores from mujoco_playground are end-of-training values, not trajectory averages. `final_strength` (= last eval MA) is the correct comparison metric. **Recommend switching BENCHMARKS.md score column to `final_strength`** and audit all existing entries.
**CartpoleSwingup regression** is real regardless of metric: p5-ppo5 `final_strength` would be ~800+, p5-ppo6 `total_reward_ma`=641. The p5-ppo6 minibatch change (2048→30 minibatches) hurt CartpoleSwingup convergence speed. Fix: revert `ppo_playground` minibatch_size to 4096 (15 minibatches) — OR accept and investigate if CartpoleSwingup needs its own spec variant.
---
## Next Architectural Changes
Research-based prioritized list of changes NOT yet tested. Ordered by expected impact across the most envs. Wave I (5-layer value + no grad clip) is currently running — results pending.
### Priority 1: NormalTanhDistribution (tanh-squashed actions)
**Expected impact**: HIGH — affects FingerTurnEasy/Hard, FishSwim, Humanoid, CartpoleSwingup
**Implementation complexity**: MEDIUM (new distribution class + policy_util changes)
**Envs helped**: All continuous-action envs, especially precision/manipulation tasks
**What Brax does differently**: Brax uses `NormalTanhDistribution` — samples from `Normal(loc, scale)`, then applies `tanh` to bound actions to [-1, 1]. The log-probability includes a log-det-jacobian correction: `log_prob -= log(1 - tanh(x)^2)`. The scale is parameterized as `softplus(raw_scale) + 0.001` (state-dependent, output by the network).
**What SLM-Lab does**: Raw `Normal(loc, scale)` with state-independent `log_std` as an `nn.Parameter`. Actions can exceed [-1, 1] and are silently clipped by the environment. The log-prob does NOT account for this clipping, creating a mismatch between the distribution the agent thinks it's using and the effective action distribution.
**Why this matters**:
1. **Gradient quality**: Without jacobian correction, the policy gradient is biased. Actions near the boundary (common in precise manipulation like FingerTurn) have incorrect log-prob gradients. The agent cannot learn fine boundary control.
2. **Exploration**: State-dependent std allows the agent to be precise where it's confident and exploratory where uncertain. State-independent std forces uniform exploration across all states — wasteful for tasks requiring both coarse and fine control.
3. **FingerTurn gap (571/950 = 60%)**: FingerTurn requires precise angular positioning of a fingertip. Without tanh squashing, actions at the boundary are clipped but the log-prob doesn't reflect this — the policy "thinks" it's outputting different actions that are actually identical after clipping. This prevents learning fine-grained control near action limits.
4. **Humanoid gap (<3%)**: 21 DOF with high-dimensional action space. State-independent std means all joints explore equally. Humanoid needs to stabilize torso (low variance) while exploring leg movement (high variance) — impossible with shared std.
**Implementation plan**:
1. Add `NormalTanhDistribution` class in `slm_lab/lib/distribution.py`:
- Forward: `action = tanh(Normal(loc, scale).rsample())`
- log_prob: `Normal.log_prob(atanh(action)) - log(1 - action^2 + eps)`
- entropy: approximate (no closed form for tanh-Normal)
2. Modify `policy_util.init_action_pd()` to handle the new distribution
3. Remove `log_std_init` for playground specs — let the network output both mean and std (state-dependent)
4. Network change: policy output dim doubles (mean + raw_scale per action dim)
**Risk**: Medium. Tanh squashing changes gradient dynamics significantly. Need to validate on already-solved envs (CartpoleBalance, WalkerRun) to ensure no regression. Can gate behind a spec flag (`action_pdtype: NormalTanh`).
---
### Fix 6: Constant LR variants + Humanoid variant (commit pending)
Added three new spec variants to `ppo_playground.yaml`:
- `ppo_playground_constlr`: DM Control + constant LR + minibatch_size=4096. For envs where vnorm=false works.
- `ppo_playground_vnorm_constlr`: DM Control + vnorm + constant LR + minibatch_size=2048. For precision envs.
- `ppo_playground_humanoid`: 2×256 policy + constant LR + vnorm. For Humanoid DM Control envs.
---
### Priority 2: Constant LR (remove LinearToMin decay)
**Expected impact**: MEDIUM — affects all envs, especially long-training ones
**Implementation complexity**: TRIVIAL (spec-only change)
**Envs helped**: CartpoleSwingup, CartpoleSwingupSparse, FingerTurnEasy/Hard, FishSwim
**What Brax does**: Constant LR = 1e-3 for all DM Control envs. No decay.
**What SLM-Lab does**: `LinearToMin` decay from 1e-3 to 3.3e-5 (min_factor=0.033) over the full training run.
**Why this matters**: By the midpoint of training, SLM-Lab's LR is already at ~5e-4 — half the Brax LR. By 75% of training, it's at ~2.7e-4. For envs that converge late (CartpoleSwingup, FishSwim), the LR is too low during the critical learning phase. Brax maintains full learning capacity throughout.
**This was tested as part of the Brax hyperparameter bundle (Wave E) which was ALL worse**, but that test changed 4 things simultaneously (clip_eps=0.3 + constant LR + 5-layer value + reward_scale=10). The constant LR was never tested in isolation.
**Implementation**: Set `min_factor: 1.0` in spec (or remove `lr_scheduler_spec` entirely).
**Risk**: Low. Constant LR is the Brax default and widely used. If instability occurs late in training, a gentler decay (`min_factor: 0.3`) can be used as fallback.
---
### Priority 3: Clip epsilon 0.3 (from 0.2)
**Expected impact**: MEDIUM — affects all envs
**Implementation complexity**: TRIVIAL (spec-only change)
**Envs helped**: FingerTurnEasy/Hard, FishSwim, CartpoleSwingup (tasks needing faster policy adaptation)
**What Brax does**: `clipping_epsilon=0.3` for DM Control.
**What SLM-Lab does**: `clip_eps=0.2`.
**Why this matters**: Clip epsilon 0.2 constrains the policy ratio to [0.8, 1.2]. At 0.3, it's [0.7, 1.3] — allowing 50% larger policy updates per step. For envs that need to explore widely before converging (FingerTurn, FishSwim), the tighter constraint slows learning.
**This was tested in the Brax bundle (Wave E) alongside 3 other changes — all worse together.** Never tested in isolation or with just constant LR.
**Implementation**: Change `start_val: 0.2` to `start_val: 0.3` in `clip_eps_spec`.
**Risk**: Low-medium. Larger clip_eps can cause training instability with small batches. However, with our 61K batch (2048 envs * 30 steps), it should be safe. If combined with constant LR (#2), the compounding effect should be tested carefully.
---
### Priority 4: Per-env tuning for FingerTurn (if P1-P3 insufficient)
**Expected impact**: HIGH for FingerTurn specifically
**Implementation complexity**: LOW (spec variant)
**Envs helped**: FingerTurnEasy, FingerTurnHard only
If NormalTanh + constant LR + clip_eps=0.3 don't close the FingerTurn gap (currently 60% and 51% of target), try:
1. **Lower gamma (0.99 → 0.95)**: FingerSpin uses gamma=0.95 officially. FingerTurn may benefit from shorter horizon discounting since reward is instantaneous (current angle vs target). Lower gamma reduces value function complexity.
2. **Smaller policy network**: Brax DM Control uses `(32, 32, 32, 32)` — our `(64, 64)` may over-parameterize for manipulation tasks. Try `(32, 32, 32, 32)` to match exactly.
3. **Higher entropy coefficient**: FingerTurn has a narrow solution manifold. Increasing entropy from 0.01 to 0.02 would encourage broader exploration of finger positions.
---
### Priority 5: Humanoid-specific — num_envs=8192
**Expected impact**: HIGH for Humanoid specifically
**Implementation complexity**: TRIVIAL (spec-only)
**Envs helped**: HumanoidStand, HumanoidWalk, HumanoidRun
**Current situation**: Humanoid was incorrectly run with loco spec (gamma=0.97, 4 epochs). The correction to DM Control spec (gamma=0.995, 16 epochs) is being tested in Wave I (p5-ppo13). However, even with correct spec, the standard 2048 envs may be insufficient.
**Why num_envs matters for Humanoid**: 21 DOF, 67-dim observations. With 2048 envs and time_horizon=30, the batch is 61K transitions — each containing a narrow slice of the 21-DOF state space. Humanoid needs more diverse rollouts to learn coordinated multi-joint control. Brax's effective batch of 983K transitions provides 16x more state-space coverage per update.
**Since we can't easily get 16x more data per update**, increasing num_envs from 2048 to 4096 or 8192 doubles/quadruples rollout diversity. Combined with NormalTanh (state-dependent std for per-joint exploration), this could be sufficient.
**VRAM concern**: 8192 envs may exceed A4500 20GB. Test with a quick 1M frame run first. Fallback: 4096 envs.
---
### NOT recommended (already tested, no benefit)
| Change | Wave | Result | Why it failed |
|---|---|---|---|
| normalize_v_targets: false | G/G2 | Mixed (helps 5, hurts 6) | Already per-env split in spec |
| Multi-unroll (num_unrolls=16) | F | Same or worse by strength | Stale old_net (480 vs 30 steps between copies) |
| Brax hyperparameter bundle (clip_eps=0.3 + constant LR + 5-layer value + reward_scale=10) | E | All worse | Confounded — 4 changes at once. Individual effects unknown except for reward_scale (helps) |
| time_horizon=480 (single long unroll) | C | Helps CartpoleSwingup, hurts FingerTurn | 480-step GAE is noisy for precision tasks |
| 5-layer value + no grad clip | III | Only helped CartpoleSwingup slightly | Hurt AcrobotSwingup, FishSwim; not general |
| NormalTanh distribution | II | Abandoned | Architecturally incompatible — SLM-Lab stores post-tanh actions, atanh inversion unstable |
| vnorm=true rerun (reverted spec) | IV | All worse or same | No new information — variance rerun |
| 4×32 Brax policy + constant LR + vnorm | VI | All worse | FingerTurnEasy 408 (vs 571), FingerTurnHard 244 (vs 484), FishSwim 106 (vs 463) |
| Humanoid wider 2×256 + constant LR + vnorm | IV-H | No improvement | MA flat at 8-10 for all 3 Humanoid envs; NormalTanh is root cause |
### Currently testing
### Wave V-B completed results (constant LR)
| Env | strength | final_strength | Old best | Verdict |
|---|---|---|---|---|
| PointMass | 841.3 | 877.3 | 863.5 | ❌ strength lower |
| **SwimmerSwimmer6** | **517.3** | 585.7 | 509.3 | ✅ NEW BEST (+1.6%) |
| FishSwim | 434.6 | 550.8 | 463.0 | ❌ strength lower (final much better) |
### Wave VII completed results (clip_eps=0.3 + constant LR)
| Env | strength | final_strength | Old best | Verdict |
|---|---|---|---|---|
| FingerTurnEasy | 518.0 | 608.8 | 570.9 | ❌ strength lower (final much better, but slow start drags average) |
| FingerTurnHard | 401.7 | 489.7 | 484.1 | ❌ strength lower (same pattern) |
| **FishSwim** | **476.9** | 581.4 | 463.0 | ✅ NEW BEST (+3%) |
**Key insight**: clip_eps=0.3 produces higher final performance but worse trajectory-averaged strength. The wider clip allows bigger policy updates which increases exploration early (slower convergence) but reaches higher asymptotic performance. The strength metric penalizes late bloomers.
### Wave V completed results
| Env | strength | final_strength | Old best | Verdict |
|---|---|---|---|---|
| CartpoleSwingup | **606.5** | 702.6 | 576.1 | ✅ NEW BEST (+5%) |
| CartpoleSwingupSparse | **383.7** | 536.2 | 296.3 | ✅ NEW BEST (+29%) |
| CartpoleBalanceSparse | **757.9** | 993.0 | 690.4 | ✅ NEW BEST (+10%) |
| AcrobotSwingup | 161.2 | 246.9 | 172.8 | ❌ strength lower (final_strength much better but trajectory avg worse due to slow start) |
**Key insight**: Constant LR is the single most impactful change found. LR decay from 1e-3 to 3.3e-5 was hurting late-converging envs. CartpoleBalanceSparse went from 690→993 (final_strength), effectively solved.
### Completed waves
**Wave VI** (p5-ppo18): 4×32 Brax policy — **STOPPED, all underperformed**. FingerTurnEasy MA 408, FingerTurnHard MA 244, FishSwim MA 106. All below old bests.
**Wave IV-H** (p5-ppo16h): Humanoid wider 2×256 + constant LR + vnorm — all flat at MA 8-10.
### Next steps after Wave VII
1. **Humanoid num_envs=4096/8192** — only major gap remaining after Wave VII
2. **Consider constant LR + clip_eps=0.3 as new general default** if results hold across envs
### Key Brax architecture differences (from source code analysis)
| Parameter | Brax Default | SLM-Lab | Impact |
|---|---|---|---|
| Policy | 4×32 (deeper, narrower) | 2×64 | **Testable via spec** |
| Value | 5×256 | 3×256 | Tested Wave III — no help |
| Distribution | tanh_normal | Normal | **Cannot test** (architectural incompatibility) |
| Init | lecun_uniform | orthogonal_ | Would need code change |
| State-dep std | False (scalar) | False (nn.Parameter) | Similar |
| Activation | swish (SiLU) | SiLU | ✅ Match |
| clipping_epsilon | 0.3 | 0.2 | **Testable via spec** |
| num_minibatches | 32 | 15-30 | Close enough |
| num_unrolls | 16 (implicit) | 1 | Tested Wave F — stale old_net hurts |
|