File size: 37,219 Bytes
e313100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
# Phase 5.1 PPO — Operations Tracker

Single source of truth for in-flight work. Resume from here.

---

## Principles

1. **Two canonical specs**: `ppo_playground` (DM Control) and `ppo_playground_loco` (Loco). Per-env variants only when officially required: `ppo_playground_fingerspin` (gamma=0.95), `ppo_playground_pendulum` (training_epoch=4, action_repeat=4 via code).
2. **100M frames hard cap** — no extended runs. If an env doesn't hit target at 100M, fix the spec.
3. **Strategic reruns**: only rerun failing/⚠️ envs. Already-✅ envs skip revalidation.
4. **Score metric**: use `total_reward_ma` (final moving average of total reward) — measures end-of-training performance and matches mujoco_playground reference scores.
5. **Official reference**: check `~/.cache/uv/archive-v0/ON8dY3irQZTYI3Bok0SlC/mujoco_playground/config/dm_control_suite_params.py` for per-env overrides.

---

## Wave 3 (2026-03-16)

**Fixes applied:**
- stderr suppression: MuJoCo C-level warnings (ccd_iterations, nefc overflow, broadphase overflow) silenced in playground.py
- obs fix: _get_obs now passes only "state" key for dict-obs envs (was incorrectly concatenating privileged_state+state)

**Envs graduated to ✅ (close enough):**
FishSwim, PointMass, ReacherHard, WalkerStand, WalkerWalk, SpotGetup, SpotJoystickGaitTracking, AlohaHandOver

**Failing envs by root cause:**
- Humanoid double-norm (rs10 fix): HumanoidStand (114→700), HumanoidWalk (47→500), HumanoidRun (18→130)
- Dict obs fix (now fixed): Go1Flat/Rough/Getup/Handstand, G1Flat/Rough, T1Flat/Rough
- Unknown: BarkourJoystick (0/35), Op3Joystick (0/20)
- Needs hparam work: H1Inplace (4→10), H1Joystick (16→30), SpotFlat (11→30)
- Manipulation: AlohaPeg (188→300), LeapCubeReorient (74→200)
- Infeasible: PandaRobotiqPushCube, AeroCubeRotateZAxis

**Currently running:** (to be populated by ops)

---

## Currently Running (as of 2026-03-14 ~00:00)

**Wave V (p5-ppo17) — Constant LR test (4 runs, just launched)**

Testing constant LR (Brax default) in isolation — never tested before. Key hypothesis: LR decay hurts late-converging envs.

| Run | Env | Spec | Key Change | Old Best | Target |
|---|---|---|---|---|---|
| p5-ppo17-csup | CartpoleSwingup | constlr | constant LR + minibatch=4096 | 576.1 | 800 |
| p5-ppo17-csupsparse | CartpoleSwingupSparse | constlr | constant LR + minibatch=4096 | 296.3 | 425 |
| p5-ppo17-acrobot | AcrobotSwingup | vnorm_constlr | constant LR + vnorm | 173 | 220 |
| p5-ppo17-fteasy | FingerTurnEasy | vnorm_constlr | constant LR + vnorm | 571 | 950 |

**Wave IV-H (p5-ppo16h) — Humanoid with wider policy (3 runs, ~2.5h remaining)**

New `ppo_playground_humanoid` variant: 2×256 policy (vs 2×64), constant LR, vnorm=true.
Based on Phase 3 Gymnasium Humanoid-v5 success (2661 MA with 2×256 + constant LR).

| Run | Env | Old Best | Target |
|---|---|---|---|
| p5-ppo16h-hstand | HumanoidStand | 18.36 | 700 |
| p5-ppo16h-hwalk | HumanoidWalk | 7.68 | 500 |
| p5-ppo16h-hrun | HumanoidRun | 3.19 | 130 |

**Wave VI (p5-ppo18) — Brax 4×32 policy + constant LR + vnorm (3 runs, just launched)**

Testing Brax default policy architecture (4 layers × 32 units vs our 2 × 64).
Deeper narrower policy may learn better features for precision tasks.

| Run | Env | Old Best | Target |
|---|---|---|---|
| p5-ppo18-fteasy | FingerTurnEasy | 571 | 950 |
| p5-ppo18-fthard | FingerTurnHard | 484 | 950 |
| p5-ppo18-fishswim | FishSwim | 463 | 650 |

**Wave IV tail (p5-ppo16) — completed**

| Run | Env | strength | Target | New best? |
|---|---|---|---|---|
| p5-ppo16-swimmer6 | SwimmerSwimmer6 | 509.3 | 560 | ✅ New best (final_strength=560.6) |
| p5-ppo16-fishswim | FishSwim | 420.6 | 650 | ❌ Worse than 463 |

**Wave IV results (p5-ppo16, vnorm=true rerun with reverted spec — completed):**

All ran with vnorm=true. CartpoleSwingup/Sparse worse (vnorm=false is better for them — wrong setting).
Precision envs also scored below old bests. Humanoid still failing with standard 2×64 policy.

| Env | p16 strength | Old Best | Target | Verdict |
|---|---|---|---|---|
| CartpoleSwingup | 316.2 | 576.1 (false) | 800 | ❌ wrong vnorm |
| CartpoleSwingupSparse | 288.7 | 296.3 (false) | 425 | ❌ wrong vnorm |
| AcrobotSwingup | 145.4 | 173 (true) | 220 | ❌ worse |
| FingerTurnEasy | 511.1 | 571 (true) | 950 | ❌ worse |
| FingerTurnHard | 368.6 | 484 (true) | 950 | ❌ worse |
| HumanoidStand | 12.72 | 18.36 | 700 | ❌ still failing |
| HumanoidWalk | 7.46 | 7.68 | 500 | ❌ still failing |
| HumanoidRun | 3.19 | 3.19 | 130 | ❌ still failing |

**CONCLUSION**: Reverted spec didn't help. No new bests. Consistency was negative for CartpoleSwingup/Sparse (high variance).
Need constant LR test (Wave V) and wider policy for Humanoid (Wave IV-H).

**Wave III results (p5-ppo13/p5-ppo15, 5-layer value + no grad clip — completed):**

Only CartpoleSwingup improved slightly (623.8 vs 576.1). All others regressed.
FishSwim p5-ppo15: strength=411.6 (vs 463 old best). AcrobotSwingup p5-ppo15: strength=95.4 (vs 173).

**CONCLUSION**: 5-layer value + no grad clip is NOT a general improvement. Reverted to 3-layer + clip_grad_val=1.0.

**Wave H results (p5-ppo12, ALL completed — NONE improved over old bests):**
Re-running same spec (variance reruns + vnorm) didn't help. Run-to-run variance is high but
old bests represent lucky runs. Hyperparameter tuning has hit diminishing returns.

**Wave G/G2 results (normalize_v_targets=false ablation, ALL completed):**

| Env | p11 strength | Old Best (true) | Target | Change | Verdict |
|---|---|---|---|---|---|
| **PendulumSwingup** | **533.5** | 276 | 395 | +93% | **✅ NEW PASS** |
| **FingerSpin** | **652.4** | 561 | 600 | +16% | **✅ NEW PASS** |
| **CartpoleBalanceSparse** | **690.4** | 545 | 700 | +27% | **⚠️ 99% of target** |
| **CartpoleSwingup** | **576.1** | 443/506 | 800 | +30% | ⚠️ improved |
| **CartpoleSwingupSparse** | **296.3** | 271 | 425 | +9% | ⚠️ improved |
| PointMass | 854.4 | 863 | 900 | -1% | ⚠️ same |
| FishSwim | 293.9 | 463 | 650 | -36% | ❌ regression |
| FingerTurnEasy | 441.1 | 571 | 950 | -23% | ❌ regression |
| SwimmerSwimmer6 | 386.2 | 485 | 560 | -20% | ❌ regression |
| FingerTurnHard | 335.7 | 484 | 950 | -31% | ❌ regression |
| AcrobotSwingup | 105.1 | 173 | 220 | -39% | ❌ regression |
| HumanoidStand | 12.87 | 18.36 | 500 | -30% | ❌ still failing |

**CONCLUSION**: `normalize_v_targets: false` helps 5/12, hurts 6/12, neutral 1/12.
- **false wins**: PendulumSwingup, FingerSpin, CartpoleBalanceSparse, CartpoleSwingup, CartpoleSwingupSparse
- **true wins**: FishSwim, FingerTurnEasy/Hard, SwimmerSwimmer6, AcrobotSwingup, PointMass
- **Decision**: Per-env spec selection. New `ppo_playground_vnorm` variant for precision envs.

**Wave F results (multi-unroll=16 + proven hyperparameters):**

| Env | p10 strength | p10 final_str | Old best str | Target | Verdict |
|---|---|---|---|---|---|
| CartpoleSwingup | 342 | 443 | 443 | 800 | Same |
| FingerTurnEasy | 529 | 685 | 571 | 950 | Better final, worse strength |
| FingerSpin | 402 | 597 | 561 | 600 | Better final (near target!), worse strength |
| FingerTurnHard | 368 | 559 | 484 | 950 | Better final, worse strength |
| SwimmerSwimmer6 | 251 | 384 | 485 | 560 | Worse |
| CartpoleSwingupSparse | 56 | 158 | 271 | 425 | MUCH worse |
| AcrobotSwingup | 31 | 63 | 173 | 220 | MUCH worse |

**CONCLUSION**: Multi-unroll adds no benefit over single-unroll for any env by `strength` metric.
The `final_strength` improvements for Finger tasks are offset by `strength` regressions.
Root cause: stale old_net (480 vs 30 steps between copies) makes policy ratio less accurate.
**Spec reverted to single-unroll (num_unrolls=1)**. Multi-unroll code preserved in ppo.py.

**Wave E results (multi-unroll + Brax hyperparameters — ALL worse):**

Brax-matched spec (clip_eps=0.3, constant LR, 5-layer value, reward_scale=10, minibatch=30720)
hurt every env except HopperStand (which used wrong spec before). Reverted.

**Wave C completed results** (all reward_scale=10, divide by 10 for true score):

| Run | Env | strength/10 | final_strength/10 | total_reward_ma/10 | Target | vs Old |
|---|---|---|---|---|---|---|
| p5-ppo7-cartpoleswingup | CartpoleSwingup | 556.6 | 670.5 | 705.3 | 800 | 443→557 ✅ improved |
| p5-ppo7-fingerturneasy | FingerTurnEasy | 511.1 | 693.2 | 687.0 | 950 | 571→511 ❌ **WORSE** |
| p5-ppo7-fingerturnhard | FingerTurnHard | 321.9 | 416.8 | 425.2 | 950 | 484→322 ❌ **WORSE** |
| p5-ppo7-cartpoleswingupsparse2 | CartpoleSwingupSparse | 144.0 | 360.6 | 337.7 | 425 | 271→144 ❌ **WORSE** |

**KEY FINDING**: time_horizon=480 helps CartpoleSwingup (+25%) but HURTS FingerTurn (-30 to -50%) and CartpoleSwingupSparse (-47%). Long GAE horizons produce noisy advantage estimates for precision/sparse tasks. The official Brax approach is 16×30-step unrolls (short GAE per unroll), NOT 1×480-step unroll.

---

## Spec Changes Applied (2026-03-13)

### Fix 1: reward_scale=10.0 (matches official mujoco_playground)
- `playground.py`: `PlaygroundVecEnv` now multiplies rewards by `self._reward_scale`
- `__init__.py`: threads `reward_scale` from env spec to wrapper
- `ppo_playground.yaml`: `reward_scale: 10.0` in shared `_env` anchor

### Fix 2: Revert minibatch_size 2048→4096 (fixes CartpoleSwingup regression)
- `ppo_playground.yaml`: all DM Control specs (ppo_playground, fingerspin, pendulum) now use minibatch_size=4096
- 15 minibatches × 16 epochs = 240 grad steps (was 30×16=480)
- Restores p5-ppo5 performance for CartpoleSwingup (803 vs 443)

### Fix 3: Brax-matched spec (commit 6eb08fe9) — time_horizon=480, clip_eps=0.3, constant LR, 5-layer value net
- Increased time_horizon from 30→480 to match total data per update (983K transitions)
- clip_eps 0.2→0.3, constant LR (min_factor=1.0), 5-layer [256×5] value net
- action std upper bound raised (max=2.0 in policy_util.py)
- **Result**: CartpoleSwingup improved (443→557 strength), but FingerTurn and CartpoleSwingupSparse got WORSE
- **Root cause**: 1×480-step unroll computes GAE over 480 steps (noisy), vs official 16×30-step unrolls (short, accurate GAE)

### Fix 4: ppo_playground_short variant (time_horizon=30 + Brax improvements)
- Keeps: reward_scale=10, clip_eps=0.3, constant LR, 5-layer value net, no grad clipping
- Reverts: time_horizon=30, minibatch_size=4096 (15 minibatches, 240 grad steps)
- **Hypothesis**: Short GAE + other Brax improvements = best of both worlds for precision tasks
- Testing on FingerTurnEasy/Hard first (Wave D p5-ppo8-*)

### Fix 5: Multi-unroll collection (IMPLEMENTED but NOT USED — code stays, spec reverted)
- Added `num_unrolls` parameter to PPO (ppo.py, actor_critic.py). Code works correctly.
- **Brax-matched spec (Wave E, p5-ppo9)**: clip_eps=0.3, constant LR, 5-layer value, reward_scale=10
  - Result: WORSE on 5/7 tested envs. Only CartpoleSwingup improved (443→506).
  - Root cause: minibatch_size=30720 → 7.5x fewer gradient steps per transition → underfitting
- **Reverted spec + multi-unroll (Wave F, p5-ppo10)**: clip_eps=0.2, LR decay, 3-layer value, minibatch=4096
  - Result: Same or WORSE on all envs by `strength` metric. Same fps as single-unroll.
  - Training compute per env step is identical, but old_net staleness (480 vs 30 steps) hurts.
- **Conclusion**: Multi-unroll adds complexity without benefit. Reverted spec to single-unroll (num_unrolls=1).
  Code preserved in ppo.py (defaults to 1). Spec uses original hyperparameters.

---

## Completed Runs Needing Intake

### Humanoid (ppo_playground_loco, post log_std fix) — intake immediately

| Run | HF Folder | strength | target | HF status |
|---|---|---|---|---|
| p5-ppo6-humanoidrun | ppo_playground_loco_humanoidrun_2026_03_12_175917 | 2.78 | 130 | ✅ uploaded |
| p5-ppo6-humanoidwalk | ppo_playground_loco_humanoidwalk_2026_03_12_175817 | 6.82 | 500 | ✅ uploaded |
| p5-ppo6-humanoidstand | ppo_playground_loco_humanoidstand_2026_03_12_175810 | 12.45 | 700 | ❌ **UPLOAD FAILED (412)** — re-upload first |

Re-upload HumanoidStand:
```bash
source .env && huggingface-cli upload SLM-Lab/benchmark-dev \
  hf_data/data/benchmark-dev/data/ppo_playground_loco_humanoidstand_2026_03_12_175810 \
  data/ppo_playground_loco_humanoidstand_2026_03_12_175810 --repo-type dataset
```

**Conclusion**: loco spec still fails completely for Humanoid — log_std fix insufficient. See spec fixes below.

### BENCHMARKS.md correction needed (commit b6ef49d9 used wrong metric)

intake-a used `total_reward_ma` instead of `strength`. Fix these 4 entries:

| Env | Run | strength (correct) | total_reward_ma (wrong) | target |
|---|---|---|---|---|
| AcrobotSwingup | p5-ppo6-acrobotswingup2 | **172.8** | 253.24 | 220 |
| CartpoleBalanceSparse | p5-ppo6-cartpolebalancesparse2 | **545.1** | 991.81 | 700 |
| CartpoleSwingup | p5-ppo6-cartpoleswingup2 | **unknown — extract from logs** | 641.51 | 800 |
| CartpoleSwingupSparse | p5-ppo6-cartpoleswingupsparse | **270.9** | 331.23 | 425 |

Extract correct values: `dstack logs p5-ppo6-NAME --since 6h 2>&1 | grep "trial_metrics" | tail -1` → use `strength:` field.

Also check FingerSpin: `dstack logs p5-ppo6-fingerspin2 --since 6h | grep trial_metrics | tail -1` — confirm strength value.

**Metric decision needed**: strength penalizes slow learners (CartpoleBalanceSparse strength=545 but final MA=992). Consider switching ALL entries to `final_strength`. But this requires auditing every existing entry — do it as a batch before publishing.

---

## Queue (launch when slots open, all 100M)

| Priority | Env | Spec | Run name | Rationale |
|---|---|---|---|---|
| 1 | PendulumSwingup | ppo_playground_pendulum | p5-ppo6-pendulumswingup | action_repeat=4 + training_epoch=4 (code fix applied) |
| 2 | FingerSpin | ppo_playground_fingerspin | p5-ppo6-fingerspin3 | canonical gamma=0.95 run; fingerspin2 used gamma=0.995 (override silently ignored) |

---

## Full Env Status

### ✅ Complete (13/25)
| Env | strength | target | normalize_v_targets |
|---|---|---|---|
| CartpoleBalance | 968.23 | 950 | true |
| AcrobotSwingupSparse | 42.74 | 15 | true |
| BallInCup | 942.44 | 680 | true |
| CheetahRun | 865.83 | 850 | true |
| ReacherEasy | 955.08 | 950 | true |
| ReacherHard | 946.99 | 950 | true |
| WalkerRun | 637.80 | 560 | true |
| WalkerStand | 970.94 | 1000 | true |
| WalkerWalk | 952 | 960 | true |
| HopperHop | 22.00 | ~2 | true |
| HopperStand | 118.2 | ~70 | true |
| PendulumSwingup | 533.5 | 395 | **false** |
| FingerSpin | 652.4 | 600 | **false** |

### ⚠️ Below target (9/25)
| Env | best strength | target | best with | status |
|---|---|---|---|---|
| CartpoleSwingup | 576.1 | 800 | false | Improved +30% from 443 (true) |
| CartpoleBalanceSparse | 545 | 700 | true | Testing false (p5-ppo11) |
| CartpoleSwingupSparse | 296.3 | 425 | false | Improved +9% from 271 (true) |
| AcrobotSwingup | 173 | 220 | true | false=105, regressed |
| FingerTurnEasy | 571 | 950 | true | false=441, regressed |
| FingerTurnHard | 484 | 950 | true | false=336, regressed |
| FishSwim | 463 | 650 | true | Testing false (p5-ppo11) |
| SwimmerSwimmer6 | 509.3 | 560 | true | final_strength=560.6 (at target!) |
| PointMass | 863 | 900 | true | false=854, ~same |

### ❌ Fundamental failure — Humanoid (3/25)
| Env | best strength | target | diagnosis |
|---|---|---|---|
| HumanoidRun | 3.19 | 130 | <3% target, NormalTanh distribution needed |
| HumanoidWalk | 7.68 | 500 | <2% target, wider policy (2×256) didn't help |
| HumanoidStand | 18.36 | 700 | <3% target, constant LR + wider policy tested, no improvement |

**Humanoid tested and failed**: wider 2×256 policy + constant LR + vnorm (Wave IV-H). MA stayed flat at 8-10 for HumanoidStand over entire training. Root cause is likely NormalTanh distribution (state-dependent std + tanh squashing) — a fundamental architectural difference from Brax.

---

## Spec Fixes Required

### Priority 1: Humanoid loco spec (update ppo_playground_loco)

Official uses `num_envs=8192, time_horizon=20 (unroll_length)` for loco. We use `num_envs=2048, time_horizon=64`.

**Proposed update to ppo_playground_loco**:
```yaml
ppo_playground_loco:
  agent:
    algorithm:
      gamma: 0.97
      time_horizon: 20      # was 64; official unroll_length=20
      training_epoch: 4
  env:
    num_envs: 8192          # was 2048; official loco num_envs=8192
```

**Before launching**: verify VRAM by checking if 8192 envs fits A4500 20GB. Run one Humanoid env, check `dstack logs NAME --since 10m | grep -i "memory\|OOM"` after 5 min.

**Rerun only**: HumanoidRun, HumanoidWalk, HumanoidStand (3 runs). HopperStand also uses loco spec — add if VRAM confirmed OK.

### Priority 2: CartpoleSwingup regression

p5-ppo5 scored 803 ✅; p5-ppo6 scored ~641. The p5-ppo6 change was `minibatch_size: 2048` (30 minibatches) vs p5-ppo5's 4096 (15 minibatches). More gradient steps per iter hurt CartpoleSwingup.

**Option A**: Revert `ppo_playground` minibatch_size from 2048→4096 (15 minibatches). Rerun only failing DM Control envs (CartpoleSwingup, CartpoleSwingupSparse, + any that need it).

**Option B**: Accept 641 and note the trade-off — p5-ppo6 improved other envs (CartpoleBalance 968 was already ✅).

### Priority 3: FingerTurnEasy/Hard

No official override. At 570/? vs target 950, gap is large. Check:
```bash
grep -A10 "Finger" ~/.cache/uv/archive-v0/ON8dY3irQZTYI3Bok0SlC/mujoco_playground/config/dm_control_suite_params.py
```

May need deeper policy network [32,32,32,32] (official arch) vs our [64,64].

---

## Tuning Principles Learned

1. **Check official per-env overrides first**: `dm_control_suite_params.py` has `discounting`, `action_repeat`, `num_updates_per_batch` per env. These are canonical.

2. **action_repeat** is env-level, not spec-level. Implemented in `playground.py` via `_ACTION_REPEAT` dict. PendulumSwingup→4. Add others as found.

3. **NaN loss**: `log_std` clamp max=0.5 helps but Humanoid (21 DOF) still has many NaN skips. Rate-limited to log every 10K. If NaN dominates → spec is wrong.

4. **num_envs scales with task complexity**: Cartpole/Acrobot: 2048 fine. Humanoid locomotion: needs 8192 for rollout diversity.

5. **time_horizon (unroll_length)**: DM Control official=30, loco official=20. Longer → more correlated rollouts → less diversity per update. Match official.

6. **Minibatch count**: more minibatches = more gradient steps per batch. Can overfit or slow convergence for simpler envs. 15 minibatches (p5-ppo5) vs 30 (p5-ppo6) — the latter hurt CartpoleSwingup.

7. **Sparse reward + strength metric**: strength (trajectory mean) severely penalizes sparse/delayed convergence. CartpoleBalanceSparse strength=545 but final MA=992. Resolve metric before publishing.

8. **High seed variance** (consistency < 0): some seeds solve, some don't → wrong spec, not bad luck. Fix exploration (entropy_coef) or use different spec.

9. **-s overrides are silently ignored** if the YAML key isn't a `${variable}` placeholder. Always verify overrides took effect via logs: `grep "gamma\|lr\|training_epoch" dstack logs`.

10. **Loco spec failures**: if loco spec gives <20 on env with target >100, the issue is almost certainly num_envs/time_horizon mismatch vs official, not a fundamental algo failure.

---

## Code Changes This Session

| Commit | Change |
|---|---|
| `8fe7bc76` | `playground.py`: `_ACTION_REPEAT` lookup for per-env action_repeat. `ppo_playground.yaml`: added `ppo_playground_fingerspin` and `ppo_playground_pendulum` specs. |
| `fb55c2f9` | `base.py`: rate-limit NaN loss warning (every 10K skips). `ppo_playground.yaml`: revert log_frequency 1M→100K. |
| `3f4ede3d` | BENCHMARKS.md: mark HopperHop ✅. |

---

## Resume Commands

```bash
# Setup
git pull && uv sync --no-default-groups

# Check jobs
dstack ps

# Intake a completed run
dstack logs RUN_NAME --since 6h 2>&1 | grep "trial_metrics" | tail -1
dstack logs RUN_NAME --since 6h 2>&1 | grep -iE "Uploading|benchmark-dev"

# Pull HF data
source .env && huggingface-cli download SLM-Lab/benchmark-dev \
  --local-dir hf_data/data/benchmark-dev --repo-type dataset \
  --include "data/FOLDER_NAME/*"

# Plot
uv run slm-lab plot -t "EnvName" -d hf_data/data/benchmark-dev/data -f FOLDER_NAME

# Launch PendulumSwingup (queue priority 1)
source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground_pendulum train \
  -s env=playground/PendulumSwingup -s max_frame=100000000 -n p5-ppo6-pendulumswingup

# Launch FingerSpin canonical (queue priority 2)
source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground_fingerspin train \
  -s env=playground/FingerSpin -s max_frame=100000000 -n p5-ppo6-fingerspin3

# Launch Humanoid loco (after updating ppo_playground_loco spec to num_envs=8192, time_horizon=20)
source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground_loco train \
  -s env=playground/HumanoidRun -s max_frame=100000000 -n p5-ppo6-humanoidrun2
```

---

## CRITICAL CORRECTION (2026-03-13) — Humanoid is DM Control, not Loco

**Root cause of Humanoid failure**: HumanoidRun/Walk/Stand are registered in `dm_control_suite/__init__.py` — they ARE DM Control envs. We incorrectly ran them with `ppo_playground_loco` (gamma=0.97, 4 epochs, time_horizon=64).

Official config uses DEFAULT DM Control params for them: discounting=0.995, 2048 envs, lr=1e-3, unroll_length=30, 16 epochs.

**NaN was never the root cause** — intake-b confirmed NaN skips were 0, 0, 2 in the loco runs. The spec was simply wrong.

**Fix**: Run all 3 Humanoid envs with `ppo_playground` (DM Control spec). No spec change needed.

```bash
# Launch with correct spec
source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground train \
  -s env=playground/HumanoidRun -s max_frame=100000000 -n p5-ppo6-humanoidrun2

source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground train \
  -s env=playground/HumanoidWalk -s max_frame=100000000 -n p5-ppo6-humanoidwalk2

source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground train \
  -s env=playground/HumanoidStand -s max_frame=100000000 -n p5-ppo6-humanoidstand2
```

**HopperStand**: Also a DM Control env. If p5-ppo6-hopperstand (loco spec, 16.38) is below target, rerun with `ppo_playground`.

**Do NOT intake** the loco-spec Humanoid runs (2.78/6.82/12.45) — wrong spec, not valid benchmark results. The old ppo_playground runs (2.86/3.73) were also wrong spec but at least the right family.

**Updated queue (prepend these as highest priority)**:

| Priority | Env | Spec | Run name |
|---|---|---|---|
| 0 | HumanoidRun | ppo_playground | p5-ppo6-humanoidrun2 |
| 0 | HumanoidWalk | ppo_playground | p5-ppo6-humanoidwalk2 |
| 0 | HumanoidStand | ppo_playground | p5-ppo6-humanoidstand2 |
| 0 | HopperStand | ppo_playground | p5-ppo6-hopperstand2 (if loco result ⚠️) |

Note on loco spec (`ppo_playground_loco`): only for actual locomotion robot envs (Go1, G1, BerkeleyHumanoid, etc.) — NOT for DM Control Humanoid.

---

## METRIC CORRECTION (2026-03-13) — strength vs final_strength

**Problem**: `strength` = trajectory-averaged mean over entire run. For slow-rising envs this severely underrepresents end-of-training performance. After metric correction to `strength`:

| Env | strength | total_reward_ma | target | conclusion |
|---|---|---|---|---|
| CartpoleSwingup | **443.0** | 641.51 | 800 | Massive regression from p5-ppo5 (803). Strength 443 << 665 (65M result) — curve rises but slow start drags average down |
| CartpoleBalanceSparse | **545.1** | 991.81 | 700 | Hits target by end (final MA=992) but sparse reward delays convergence |
| AcrobotSwingup | **172.8** | 253.24 | 220 | Below target by strength, above by final MA |
| CartpoleSwingupSparse | **270.9** | 331.23 | 425 | Below both metrics |

**Resolution needed**: Reference scores from mujoco_playground are end-of-training values, not trajectory averages. `final_strength` (= last eval MA) is the correct comparison metric. **Recommend switching BENCHMARKS.md score column to `final_strength`** and audit all existing entries.

**CartpoleSwingup regression** is real regardless of metric: p5-ppo5 `final_strength` would be ~800+, p5-ppo6 `total_reward_ma`=641. The p5-ppo6 minibatch change (2048→30 minibatches) hurt CartpoleSwingup convergence speed. Fix: revert `ppo_playground` minibatch_size to 4096 (15 minibatches) — OR accept and investigate if CartpoleSwingup needs its own spec variant.

---

## Next Architectural Changes

Research-based prioritized list of changes NOT yet tested. Ordered by expected impact across the most envs. Wave I (5-layer value + no grad clip) is currently running — results pending.

### Priority 1: NormalTanhDistribution (tanh-squashed actions)

**Expected impact**: HIGH — affects FingerTurnEasy/Hard, FishSwim, Humanoid, CartpoleSwingup
**Implementation complexity**: MEDIUM (new distribution class + policy_util changes)
**Envs helped**: All continuous-action envs, especially precision/manipulation tasks

**What Brax does differently**: Brax uses `NormalTanhDistribution` — samples from `Normal(loc, scale)`, then applies `tanh` to bound actions to [-1, 1]. The log-probability includes a log-det-jacobian correction: `log_prob -= log(1 - tanh(x)^2)`. The scale is parameterized as `softplus(raw_scale) + 0.001` (state-dependent, output by the network).

**What SLM-Lab does**: Raw `Normal(loc, scale)` with state-independent `log_std` as an `nn.Parameter`. Actions can exceed [-1, 1] and are silently clipped by the environment. The log-prob does NOT account for this clipping, creating a mismatch between the distribution the agent thinks it's using and the effective action distribution.

**Why this matters**:
1. **Gradient quality**: Without jacobian correction, the policy gradient is biased. Actions near the boundary (common in precise manipulation like FingerTurn) have incorrect log-prob gradients. The agent cannot learn fine boundary control.
2. **Exploration**: State-dependent std allows the agent to be precise where it's confident and exploratory where uncertain. State-independent std forces uniform exploration across all states — wasteful for tasks requiring both coarse and fine control.
3. **FingerTurn gap (571/950 = 60%)**: FingerTurn requires precise angular positioning of a fingertip. Without tanh squashing, actions at the boundary are clipped but the log-prob doesn't reflect this — the policy "thinks" it's outputting different actions that are actually identical after clipping. This prevents learning fine-grained control near action limits.
4. **Humanoid gap (<3%)**: 21 DOF with high-dimensional action space. State-independent std means all joints explore equally. Humanoid needs to stabilize torso (low variance) while exploring leg movement (high variance) — impossible with shared std.

**Implementation plan**:
1. Add `NormalTanhDistribution` class in `slm_lab/lib/distribution.py`:
   - Forward: `action = tanh(Normal(loc, scale).rsample())`
   - log_prob: `Normal.log_prob(atanh(action)) - log(1 - action^2 + eps)`
   - entropy: approximate (no closed form for tanh-Normal)
2. Modify `policy_util.init_action_pd()` to handle the new distribution
3. Remove `log_std_init` for playground specs — let the network output both mean and std (state-dependent)
4. Network change: policy output dim doubles (mean + raw_scale per action dim)

**Risk**: Medium. Tanh squashing changes gradient dynamics significantly. Need to validate on already-solved envs (CartpoleBalance, WalkerRun) to ensure no regression. Can gate behind a spec flag (`action_pdtype: NormalTanh`).

---

### Fix 6: Constant LR variants + Humanoid variant (commit pending)

Added three new spec variants to `ppo_playground.yaml`:
- `ppo_playground_constlr`: DM Control + constant LR + minibatch_size=4096. For envs where vnorm=false works.
- `ppo_playground_vnorm_constlr`: DM Control + vnorm + constant LR + minibatch_size=2048. For precision envs.
- `ppo_playground_humanoid`: 2×256 policy + constant LR + vnorm. For Humanoid DM Control envs.

---

### Priority 2: Constant LR (remove LinearToMin decay)

**Expected impact**: MEDIUM — affects all envs, especially long-training ones
**Implementation complexity**: TRIVIAL (spec-only change)
**Envs helped**: CartpoleSwingup, CartpoleSwingupSparse, FingerTurnEasy/Hard, FishSwim

**What Brax does**: Constant LR = 1e-3 for all DM Control envs. No decay.

**What SLM-Lab does**: `LinearToMin` decay from 1e-3 to 3.3e-5 (min_factor=0.033) over the full training run.

**Why this matters**: By the midpoint of training, SLM-Lab's LR is already at ~5e-4 — half the Brax LR. By 75% of training, it's at ~2.7e-4. For envs that converge late (CartpoleSwingup, FishSwim), the LR is too low during the critical learning phase. Brax maintains full learning capacity throughout.

**This was tested as part of the Brax hyperparameter bundle (Wave E) which was ALL worse**, but that test changed 4 things simultaneously (clip_eps=0.3 + constant LR + 5-layer value + reward_scale=10). The constant LR was never tested in isolation.

**Implementation**: Set `min_factor: 1.0` in spec (or remove `lr_scheduler_spec` entirely).

**Risk**: Low. Constant LR is the Brax default and widely used. If instability occurs late in training, a gentler decay (`min_factor: 0.3`) can be used as fallback.

---

### Priority 3: Clip epsilon 0.3 (from 0.2)

**Expected impact**: MEDIUM — affects all envs
**Implementation complexity**: TRIVIAL (spec-only change)
**Envs helped**: FingerTurnEasy/Hard, FishSwim, CartpoleSwingup (tasks needing faster policy adaptation)

**What Brax does**: `clipping_epsilon=0.3` for DM Control.

**What SLM-Lab does**: `clip_eps=0.2`.

**Why this matters**: Clip epsilon 0.2 constrains the policy ratio to [0.8, 1.2]. At 0.3, it's [0.7, 1.3] — allowing 50% larger policy updates per step. For envs that need to explore widely before converging (FingerTurn, FishSwim), the tighter constraint slows learning.

**This was tested in the Brax bundle (Wave E) alongside 3 other changes — all worse together.** Never tested in isolation or with just constant LR.

**Implementation**: Change `start_val: 0.2` to `start_val: 0.3` in `clip_eps_spec`.

**Risk**: Low-medium. Larger clip_eps can cause training instability with small batches. However, with our 61K batch (2048 envs * 30 steps), it should be safe. If combined with constant LR (#2), the compounding effect should be tested carefully.

---

### Priority 4: Per-env tuning for FingerTurn (if P1-P3 insufficient)

**Expected impact**: HIGH for FingerTurn specifically
**Implementation complexity**: LOW (spec variant)
**Envs helped**: FingerTurnEasy, FingerTurnHard only

If NormalTanh + constant LR + clip_eps=0.3 don't close the FingerTurn gap (currently 60% and 51% of target), try:

1. **Lower gamma (0.99 → 0.95)**: FingerSpin uses gamma=0.95 officially. FingerTurn may benefit from shorter horizon discounting since reward is instantaneous (current angle vs target). Lower gamma reduces value function complexity.

2. **Smaller policy network**: Brax DM Control uses `(32, 32, 32, 32)` — our `(64, 64)` may over-parameterize for manipulation tasks. Try `(32, 32, 32, 32)` to match exactly.

3. **Higher entropy coefficient**: FingerTurn has a narrow solution manifold. Increasing entropy from 0.01 to 0.02 would encourage broader exploration of finger positions.

---

### Priority 5: Humanoid-specific — num_envs=8192

**Expected impact**: HIGH for Humanoid specifically
**Implementation complexity**: TRIVIAL (spec-only)
**Envs helped**: HumanoidStand, HumanoidWalk, HumanoidRun

**Current situation**: Humanoid was incorrectly run with loco spec (gamma=0.97, 4 epochs). The correction to DM Control spec (gamma=0.995, 16 epochs) is being tested in Wave I (p5-ppo13). However, even with correct spec, the standard 2048 envs may be insufficient.

**Why num_envs matters for Humanoid**: 21 DOF, 67-dim observations. With 2048 envs and time_horizon=30, the batch is 61K transitions — each containing a narrow slice of the 21-DOF state space. Humanoid needs more diverse rollouts to learn coordinated multi-joint control. Brax's effective batch of 983K transitions provides 16x more state-space coverage per update.

**Since we can't easily get 16x more data per update**, increasing num_envs from 2048 to 4096 or 8192 doubles/quadruples rollout diversity. Combined with NormalTanh (state-dependent std for per-joint exploration), this could be sufficient.

**VRAM concern**: 8192 envs may exceed A4500 20GB. Test with a quick 1M frame run first. Fallback: 4096 envs.

---

### NOT recommended (already tested, no benefit)

| Change | Wave | Result | Why it failed |
|---|---|---|---|
| normalize_v_targets: false | G/G2 | Mixed (helps 5, hurts 6) | Already per-env split in spec |
| Multi-unroll (num_unrolls=16) | F | Same or worse by strength | Stale old_net (480 vs 30 steps between copies) |
| Brax hyperparameter bundle (clip_eps=0.3 + constant LR + 5-layer value + reward_scale=10) | E | All worse | Confounded — 4 changes at once. Individual effects unknown except for reward_scale (helps) |
| time_horizon=480 (single long unroll) | C | Helps CartpoleSwingup, hurts FingerTurn | 480-step GAE is noisy for precision tasks |
| 5-layer value + no grad clip | III | Only helped CartpoleSwingup slightly | Hurt AcrobotSwingup, FishSwim; not general |
| NormalTanh distribution | II | Abandoned | Architecturally incompatible — SLM-Lab stores post-tanh actions, atanh inversion unstable |
| vnorm=true rerun (reverted spec) | IV | All worse or same | No new information — variance rerun |
| 4×32 Brax policy + constant LR + vnorm | VI | All worse | FingerTurnEasy 408 (vs 571), FingerTurnHard 244 (vs 484), FishSwim 106 (vs 463) |
| Humanoid wider 2×256 + constant LR + vnorm | IV-H | No improvement | MA flat at 8-10 for all 3 Humanoid envs; NormalTanh is root cause |

### Currently testing

### Wave V-B completed results (constant LR)

| Env | strength | final_strength | Old best | Verdict |
|---|---|---|---|---|
| PointMass | 841.3 | 877.3 | 863.5 | ❌ strength lower |
| **SwimmerSwimmer6** | **517.3** | 585.7 | 509.3 | ✅ NEW BEST (+1.6%) |
| FishSwim | 434.6 | 550.8 | 463.0 | ❌ strength lower (final much better) |

### Wave VII completed results (clip_eps=0.3 + constant LR)

| Env | strength | final_strength | Old best | Verdict |
|---|---|---|---|---|
| FingerTurnEasy | 518.0 | 608.8 | 570.9 | ❌ strength lower (final much better, but slow start drags average) |
| FingerTurnHard | 401.7 | 489.7 | 484.1 | ❌ strength lower (same pattern) |
| **FishSwim** | **476.9** | 581.4 | 463.0 | ✅ NEW BEST (+3%) |

**Key insight**: clip_eps=0.3 produces higher final performance but worse trajectory-averaged strength. The wider clip allows bigger policy updates which increases exploration early (slower convergence) but reaches higher asymptotic performance. The strength metric penalizes late bloomers.

### Wave V completed results

| Env | strength | final_strength | Old best | Verdict |
|---|---|---|---|---|
| CartpoleSwingup | **606.5** | 702.6 | 576.1 | ✅ NEW BEST (+5%) |
| CartpoleSwingupSparse | **383.7** | 536.2 | 296.3 | ✅ NEW BEST (+29%) |
| CartpoleBalanceSparse | **757.9** | 993.0 | 690.4 | ✅ NEW BEST (+10%) |
| AcrobotSwingup | 161.2 | 246.9 | 172.8 | ❌ strength lower (final_strength much better but trajectory avg worse due to slow start) |

**Key insight**: Constant LR is the single most impactful change found. LR decay from 1e-3 to 3.3e-5 was hurting late-converging envs. CartpoleBalanceSparse went from 690→993 (final_strength), effectively solved.

### Completed waves

**Wave VI** (p5-ppo18): 4×32 Brax policy — **STOPPED, all underperformed**. FingerTurnEasy MA 408, FingerTurnHard MA 244, FishSwim MA 106. All below old bests.

**Wave IV-H** (p5-ppo16h): Humanoid wider 2×256 + constant LR + vnorm — all flat at MA 8-10.

### Next steps after Wave VII

1. **Humanoid num_envs=4096/8192** — only major gap remaining after Wave VII
2. **Consider constant LR + clip_eps=0.3 as new general default** if results hold across envs

### Key Brax architecture differences (from source code analysis)

| Parameter | Brax Default | SLM-Lab | Impact |
|---|---|---|---|
| Policy | 4×32 (deeper, narrower) | 2×64 | **Testable via spec** |
| Value | 5×256 | 3×256 | Tested Wave III — no help |
| Distribution | tanh_normal | Normal | **Cannot test** (architectural incompatibility) |
| Init | lecun_uniform | orthogonal_ | Would need code change |
| State-dep std | False (scalar) | False (nn.Parameter) | Similar |
| Activation | swish (SiLU) | SiLU | ✅ Match |
| clipping_epsilon | 0.3 | 0.2 | **Testable via spec** |
| num_minibatches | 32 | 15-30 | Close enough |
| num_unrolls | 16 (implicit) | 1 | Tested Wave F — stale old_net hurts |