| # Phase 5 Spec Research: Official vs SLM-Lab Config Comparison |
|
|
| ## Source Files |
|
|
| - **Official config**: `mujoco_playground/config/dm_control_suite_params.py` ([GitHub](https://github.com/google-deepmind/mujoco_playground/blob/main/mujoco_playground/config/dm_control_suite_params.py)) |
| - **Official network**: Brax PPO defaults (`brax/training/agents/ppo/networks.py`) |
| - **Our spec**: `slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml` |
| - **Our wrapper**: `slm_lab/env/playground.py` |
|
|
| ## Critical Architectural Difference: Batch Collection Size |
|
|
| The most significant difference is how much data is collected per update cycle. |
|
|
| ### Official Brax PPO batch mechanics |
|
|
| In Brax PPO, `batch_size` means **minibatch size in trajectories** (not total batch): |
|
|
| | Parameter | Official Value | |
| |---|---| |
| | `num_envs` | 2048 | |
| | `unroll_length` | 30 | |
| | `batch_size` | 1024 (trajectories per minibatch) | |
| | `num_minibatches` | 32 | |
| | `num_updates_per_batch` | 16 (epochs) | |
|
|
| - Sequential unrolls per env = `batch_size * num_minibatches / num_envs` = 1024 * 32 / 2048 = **16** |
| - Total transitions collected = 2048 envs * 16 unrolls * 30 steps = **983,040** |
| - Each minibatch = 30,720 transitions |
| - Grad steps per update = 32 * 16 = **512** |
|
|
| ### SLM-Lab batch mechanics |
|
|
| | Parameter | Our Value | |
| |---|---| |
| | `num_envs` | 2048 | |
| | `time_horizon` | 30 | |
| | `minibatch_size` | 2048 | |
| | `training_epoch` | 16 | |
|
|
| - Total transitions collected = 2048 * 30 = **61,440** |
| - Num minibatches = 61,440 / 2048 = **30** |
| - Each minibatch = 2,048 transitions |
| - Grad steps per update = 30 * 16 = **480** |
|
|
| ### Comparison |
|
|
| | Metric | Official | SLM-Lab | Ratio | |
| |---|---|---|---| |
| | Transitions per update | 983,040 | 61,440 | **16x more in official** | |
| | Minibatch size (transitions) | 30,720 | 2,048 | **15x more in official** | |
| | Grad steps per update | 512 | 480 | ~same | |
| | Data reuse (epochs over same data) | 16 | 16 | same | |
|
|
| **Impact**: Official collects 16x more data before each gradient update cycle. Each minibatch is 15x larger. The grad steps are similar, but each gradient step in official sees 15x more transitions — better gradient estimates, less variance. |
|
|
| This is likely the **root cause** for most failures, especially hard exploration tasks (FingerTurn, CartpoleSwingupSparse). |
|
|
| ## Additional Missing Feature: reward_scaling=10.0 |
| |
| The official config uses `reward_scaling=10.0`. SLM-Lab has **no reward scaling** (implicitly 1.0). This amplifies reward signal by 10x, which: |
| - Helps with sparse/small rewards (CartpoleSwingupSparse, AcrobotSwingup) |
| - Works in conjunction with value target normalization |
| - May partially compensate for the batch size difference |
|
|
| ## Network Architecture |
|
|
| | Component | Official (Brax) | SLM-Lab | Match? | |
| |---|---|---|---| |
| | Policy layers | (32, 32, 32, 32) | (64, 64) | Different shape, similar param count | |
| | Value layers | (256, 256, 256, 256, 256) | (256, 256, 256) | Official deeper | |
| | Activation | Swish (SiLU) | SiLU | Same | |
| | Init | default (lecun_uniform) | orthogonal_ | Different | |
|
|
| The policy architectures have similar total parameters (32*32*4 vs 64*64*2 chains are comparable). The value network is 2 layers shallower in SLM-Lab. Unlikely to be the primary cause of failures but could matter for harder tasks. |
|
|
| ## Per-Environment Analysis |
|
|
| ### Env: FingerTurnEasy (570 vs 950 target) |
|
|
| | Parameter | Official | Ours | Mismatch? | |
| |---|---|---|---| |
| | gamma (discounting) | 0.995 | 0.995 | Match | |
| | training_epoch (num_updates_per_batch) | 16 | 16 | Match | |
| | time_horizon (unroll_length) | 30 | 30 | Match | |
| | action_repeat | 1 | 1 | Match | |
| | num_envs | 2048 | 2048 | Match | |
| | reward_scaling | 10.0 | 1.0 (none) | **MISMATCH** | |
| | batch collection size | 983K | 61K | **MISMATCH (16x)** | |
| | minibatch transitions | 30,720 | 2,048 | **MISMATCH (15x)** | |
| |
| **Per-env overrides**: None in official. Uses all defaults. |
| **Diagnosis**: Huge gap (570 vs 950). FingerTurn is a precision manipulation task requiring coordinated finger-tip control. The 16x smaller batch likely causes high gradient variance, preventing the policy from learning fine-grained coordination. reward_scaling=10 would also help. |
|
|
| ### Env: FingerTurnHard (~500 vs 950 target) |
|
|
| Same as FingerTurnEasy — no per-env overrides. Same mismatches apply. |
| **Diagnosis**: Even harder version, same root cause. Needs larger batches and reward scaling. |
|
|
| ### Env: CartpoleSwingup (443 vs 800 target, regression from p5-ppo5=803) |
|
|
| | Parameter | Official | p5-ppo5 | p5-ppo6 (current) | |
| |---|---|---|---| |
| | minibatch_size | N/A (30,720 transitions) | 4096 | 2048 | |
| | num_minibatches | 32 | 15 | 30 | |
| | grad steps/update | 512 | 240 | 480 | |
| | total transitions/update | 983K | 61K | 61K | |
| | reward_scaling | 10.0 | 1.0 | 1.0 | |
| |
| **Per-env overrides**: None in official. |
| **Diagnosis**: The p5-ppo5→p5-ppo6 regression (803→443) came from doubling grad steps (240→480) while halving minibatch size (4096→2048). More gradient steps on smaller minibatches = overfitting per update. p5-ppo5's 15 larger minibatches were better for CartpoleSwingup. |
| |
| **Answer to key question**: Yes, reverting to minibatch_size=4096 would likely restore CartpoleSwingup performance. However, the deeper fix is the batch collection size — both p5-ppo5 and p5-ppo6 collect only 61K transitions vs official's 983K. |
|
|
| ### Env: CartpoleSwingupSparse (270 vs 425 target) |
|
|
| | Parameter | Official | Ours | Mismatch? | |
| |---|---|---|---| |
| | All params | Same defaults | Same as ppo_playground | Same mismatches | |
| | reward_scaling | 10.0 | 1.0 | **MISMATCH — critical for sparse** | |
|
|
| **Per-env overrides**: None in official. |
| **Diagnosis**: Sparse reward + no reward scaling = very weak learning signal. reward_scaling=10 is especially important here. The small batch also hurts exploration diversity. |
| |
| ### Env: CartpoleBalanceSparse (545 vs 700 target) |
| |
| Same mismatches as other Cartpole variants. No per-env overrides. |
| **Diagnosis**: Note that the actual final MA is 992 (well above target). The low "strength" score (545) reflects slow initial convergence, not inability to solve. If metric switches to final_strength, this may already pass. reward_scaling would accelerate early convergence. |
| |
| ### Env: AcrobotSwingup (172 vs 220 target) |
| |
| | Parameter | Official | Ours | Mismatch? | |
| |---|---|---|---| |
| | num_timesteps | 100M | 100M | Match (official has explicit override) | |
| | All training params | Defaults | ppo_playground | Same mismatches | |
| | reward_scaling | 10.0 | 1.0 | **MISMATCH** | |
|
|
| **Per-env overrides**: Official only sets `num_timesteps=100M` (already matched). |
| **Diagnosis**: Close to target (172 vs 220). reward_scaling=10 would likely close the gap. The final MA (253) exceeds target — metric issue compounds this. |
| |
| ### Env: SwimmerSwimmer6 (485 vs 560 target) |
| |
| | Parameter | Official | Ours | Mismatch? | |
| |---|---|---|---| |
| | num_timesteps | 100M | 100M | Match (official has explicit override) | |
| | All training params | Defaults | ppo_playground | Same mismatches | |
| | reward_scaling | 10.0 | 1.0 | **MISMATCH** | |
|
|
| **Per-env overrides**: Official only sets `num_timesteps=100M` (already matched). |
| **Diagnosis**: Swimmer is a multi-joint locomotion task that benefits from larger batches (more diverse body configurations per update). reward_scaling would also help. |
| |
| ### Env: PointMass (863 vs 900 target) |
| |
| No per-env overrides. Same mismatches. |
| **Diagnosis**: Very close (863 vs 900). This might pass with reward_scaling alone. Simple task — batch size less critical. |
|
|
| ### Env: FishSwim (~530 vs 650 target, may still be running) |
|
|
| No per-env overrides. Same mismatches. |
| **Diagnosis**: 3D swimming task. Would benefit from both larger batches and reward_scaling. |
| |
| ## Summary of Mismatches (All Envs) |
| |
| | Mismatch | Official | SLM-Lab | Impact | Fixable? | |
| |---|---|---|---|---| |
| | **Batch collection size** | 983K transitions | 61K transitions | HIGH — 16x less data per update | Requires architectural change to collect multiple unrolls | |
| | **Minibatch size** | 30,720 transitions | 2,048 transitions | HIGH — much noisier gradients | Limited by venv_pack constraint | |
| | **reward_scaling** | 10.0 | 1.0 (none) | MEDIUM-HIGH — especially for sparse envs | Easy to add | |
| | **Value network depth** | 5 layers | 3 layers | LOW-MEDIUM | Easy to change in spec | |
| | **Weight init** | lecun_uniform | orthogonal_ | LOW | Unlikely to matter much | |
| |
| ## Proposed Fixes |
| |
| ### Fix 1: Add reward_scaling (EASY, HIGH IMPACT) |
| |
| Add a `reward_scale` parameter to the spec and apply it in the training loop or environment wrapper. |
| |
| ```yaml |
| # In ppo_playground spec |
| env: |
| reward_scale: 10.0 # Official mujoco_playground default |
| ``` |
| |
| This requires a code change to support `reward_scale` in the env or algorithm. Simplest approach: multiply rewards by scale factor in the PlaygroundVecEnv wrapper. |
| |
| **Priority: 1 (do this first)** — Easy to implement, likely closes the gap for PointMass, AcrobotSwingup, and CartpoleBalanceSparse. Partial improvement for others. |
|
|
| ### Fix 2: Revert minibatch_size to 4096 for base ppo_playground (EASY) |
|
|
| ```yaml |
| ppo_playground: |
| agent: |
| algorithm: |
| minibatch_size: 4096 # 15 minibatches, fewer but larger grad steps |
| ``` |
|
|
| **Priority: 2** — Immediately restores CartpoleSwingup from 443 to ~803. May modestly improve other envs. The trade-off: fewer grad steps (240 vs 480) but larger minibatches = more stable gradients. |
|
|
| ### Fix 3: Multi-unroll collection (MEDIUM DIFFICULTY, HIGHEST IMPACT) |
|
|
| The fundamental gap is that SLM-Lab collects only 1 unroll (30 steps) from each env before updating, while Brax collects 16 sequential unrolls (480 steps). To match official: |
|
|
| Option A: Increase `time_horizon` to 480 (= 30 * 16). This collects the same total data but changes GAE computation (advantages computed over 480 steps instead of 30). Not equivalent to official. |
|
|
| Option B: Add a `num_unrolls` parameter that collects multiple independent unrolls of `time_horizon` length before updating. This matches official behavior but requires a code change to the training loop. |
|
|
| Option C: Accept the batch size difference and compensate with reward_scaling + larger minibatch_size. Less optimal but no code changes needed beyond reward_scaling. |
| |
| **Priority: 3** — Biggest potential impact but requires code changes. Try fixes 1-2 first and re-evaluate. |
| |
| ### Fix 4: Deepen value network (EASY) |
| |
| ```yaml |
| _value_body: &value_body |
| modules: |
| body: |
| Sequential: |
| - LazyLinear: {out_features: 256} |
| - SiLU: |
| - LazyLinear: {out_features: 256} |
| - SiLU: |
| - LazyLinear: {out_features: 256} |
| - SiLU: |
| - LazyLinear: {out_features: 256} |
| - SiLU: |
| - LazyLinear: {out_features: 256} |
| - SiLU: |
| ``` |
| |
| **Priority: 4** — Minor impact expected. Try after fixes 1-2. |
|
|
| ### Fix 5: Per-env spec variants for FingerTurn (if fixes 1-2 insufficient) |
|
|
| If FingerTurn still fails after reward_scaling + minibatch revert, create a dedicated variant with tuned hyperparameters (possibly lower gamma, different lr). But try the general fixes first since official uses default params for FingerTurn. |
| |
| **Priority: 5** — Only if fixes 1-3 don't close the gap. |
| |
| ## Recommended Action Plan |
| |
| 1. **Implement reward_scale=10.0** in PlaygroundVecEnv (multiply rewards by scale factor). Add `reward_scale` to env spec. One-line code change + spec update. |
|
|
| 2. **Revert minibatch_size to 4096** in ppo_playground base spec. This gives 15 minibatches * 16 epochs = 240 grad steps (vs 480 now). |
| |
| 3. **Rerun the 5 worst-performing envs** with fixes 1+2: |
| - FingerTurnEasy (570 → target 950) |
| - FingerTurnHard (500 → target 950) |
| - CartpoleSwingup (443 → target 800) |
| - CartpoleSwingupSparse (270 → target 425) |
| - FishSwim (530 → target 650) |
|
|
| 4. **Evaluate results**. If FingerTurn still fails badly, investigate multi-unroll collection (Fix 3) or FingerTurn-specific tuning. |
|
|
| 5. **Metric decision**: Switch to `final_strength` for score reporting. CartpoleBalanceSparse (final MA=992) and AcrobotSwingup (final MA=253) likely pass under the correct metric. |
|
|
| ## Envs Likely Fixed by Metric Change Alone |
|
|
| These envs have final MA above target but low "strength" due to slow early convergence: |
|
|
| | Env | strength | final MA | target | Passes with final_strength? | |
| |---|---|---|---|---| |
| | CartpoleBalanceSparse | 545 | 992 | 700 | YES | |
| | AcrobotSwingup | 172 | 253 | 220 | YES | |
| |
| ## Envs Requiring Spec Changes |
| |
| | Env | Current | Target | Most likely fix | |
| |---|---|---|---| |
| | FingerTurnEasy | 570 | 950 | reward_scale + larger batch | |
| | FingerTurnHard | 500 | 950 | reward_scale + larger batch | |
| | CartpoleSwingup | 443 | 800 | Revert minibatch_size=4096 | |
| | CartpoleSwingupSparse | 270 | 425 | reward_scale | |
| | SwimmerSwimmer6 | 485 | 560 | reward_scale | |
| | PointMass | 863 | 900 | reward_scale | |
| | FishSwim | 530 | 650 | reward_scale + larger batch | |
|
|