File size: 9,954 Bytes
ebd7dfd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
# SLM Lab Improvements Roadmap

SLM Lab's algorithms (PPO, SAC) are architecturally sound but use 2017-era defaults. This roadmap integrates material advances from the post-PPO RL landscape.

**Source**: [`notes/literature/ai/rl-landscape-2026.md`](../../notes/literature/ai/rl-landscape-2026.md)

**Hardware**: Mac (Apple Silicon) for dev, cloud GPU (A100/H100) for runs.

---

## Status

| Step | What | Status |
|:---:|------|--------|
| **1** | **GPU envs (MuJoCo Playground)** | **NEXT** |
| 2 | Normalization stack (layer norm, percentile) | DONE |
| 3 | CrossQ algorithm (batch norm critics) | DONE |
| 4 | Combine + full benchmark suite | TODO (after Step 1) |
| 5 | High-UTD SAC / RLPD | TODO |
| 6 | Pretrained vision encoders | TODO |

---

## NEXT: Step 1 — GPU Envs (MuJoCo Playground)

**Goal**: Remove env as the bottleneck. Run physics on GPU via [MuJoCo Playground](https://github.com/google-deepmind/mujoco_playground), keep training in PyTorch. Scale to 1000+ parallel envs for large-scale runs.

### The Stack

```
MuJoCo Playground  ← env definitions, registry, wrappers

     Brax           ← EpisodeWrapper, AutoResetWrapper

  MuJoCo MJX        ← JAX reimplementation of MuJoCo physics (GPU/TPU)

   JAX / XLA        ← jit, vmap
```

### API Difference

Playground uses a **stateless functional API**, not Gymnasium OOP:

```python
# Gymnasium (today)                         # Playground
env = gym.make("HalfCheetah-v5")            env = registry.load("CheetahRun")
obs, info = env.reset()                     state = env.reset(rng)        # → State dataclass
obs, rew, term, trunc, info = env.step(a)   state = env.step(state, a)    # → new State
```

Key differences: functional (state passed explicitly), `jax.vmap` for batching (not `VectorEnv`), `jax.jit` for GPU compilation, single `done` flag (no term/trunc split), `observation_size`/`action_size` ints (no `gym.spaces`).

### Environment Catalog

**DM Control Suite (25 envs)** — standard RL benchmarks, but dm_control versions (different obs/reward/termination from Gymnasium MuJoCo):

| Playground | Nearest Gymnasium | Notes |
|-----------|-------------------|-------|
| `CheetahRun` | `HalfCheetah-v5` | Tolerance reward (target speed=10) |
| `HopperHop` / `HopperStand` | `Hopper-v5` | Different reward |
| `WalkerWalk` / `WalkerRun` | `Walker2d-v5` | dm_control version |
| `HumanoidWalk` / `HumanoidRun` | `Humanoid-v5` | CMU humanoid |
| `CartpoleSwingup` | `CartPole-v1` | Swing-up (harder) |
| `ReacherEasy/Hard`, `FingerSpin/Turn*`, `FishSwim`, `PendulumSwingup`, `SwimmerSwimmer6` | — | Various |

No Ant equivalent. Results NOT comparable across env suites.

**Locomotion (19 envs)** — real robots (Unitree Go1/G1/H1, Spot, etc.) with joystick control, gait tracking, recovery.

**Manipulation (10 envs)** — Aloha bimanual, Franka Panda, LEAP hand dexterity.

### Performance

Single-env MJX is ~10x slower than CPU MuJoCo. The win comes from massive parallelism:

| Hardware | Batch Size | Humanoid steps/sec |
|----------|-----------|-------------------|
| M3 Max (CPU) | ~128 | 650K |
| A100 (MJX) | 8,192 | 950K |

Training throughput on single A100: ~720K steps/sec (Cartpole PPO), ~91K steps/sec (Humanoid PPO). SAC 25-50x slower than PPO (off-policy overhead).

**Wall clock (1M frames)**: CPU ~80 min → GPU <5 min (PPO), ~30 min (SAC).

### Integration Design

Adapter at the env boundary. Algorithms unchanged.

```
Spec: env.backend = "playground", env.name = "CheetahRun", env.num_envs = 4096

make_env() routes on backend

PlaygroundVecEnv(VectorEnv)  ← jit+vmap internally, DLPack zero-copy at boundary

VectorClockWrapper → Session.run_rl() (existing, unchanged)
```

Reference implementations: Playground's [`wrapper_torch.py`](https://github.com/google-deepmind/mujoco_playground/blob/main/mujoco_playground/_src/wrapper_torch.py) (`RSLRLBraxWrapper`), [skrl](https://skrl.readthedocs.io/en/develop/api/envs/wrapping.html) Gymnasium-like wrapper.

### Changes

- `slm_lab/env/playground.py`: **New**`PlaygroundVecEnv(VectorEnv)` adapter (JIT, vmap, DLPack, auto-reset, RNG management)
- `slm_lab/env/__init__.py`: `backend` routing in `make_env()`
- `pyproject.toml`: Optional `[playground]` dependency group (`mujoco-playground`, `jax[cuda12]`, `mujoco-mjx`, `brax`)
- Specs: New specs with `backend: playground`, Playground env names, `num_envs: 4096`

No changes to: algorithms, networks, memory, training loop, experiment control.

### Gotchas

1. **JIT startup**: First `reset()`/`step()` triggers XLA compilation (10-60s). One-time.
2. **Static shapes**: `num_envs` fixed at construction. Contacts padded to max possible.
3. **Ampere precision**: RTX 30/40 need `JAX_DEFAULT_MATMUL_PRECISION=highest` or training destabilizes.
4. **No Atari**: Playground is physics-only. Atari stays on CPU Gymnasium.

### Verify

PPO on CheetahRun — same reward as CPU baseline, 100x+ faster wall clock (4096 envs, A100).

### Migration Path

1. **Phase 1** (this step): Adapter + DM Control locomotion (CheetahRun, HopperHop, WalkerWalk, HumanoidWalk/Run)
2. **Phase 2**: Robotics envs (Unitree Go1/G1, Spot, Franka Panda, LEAP hand)
3. **Phase 3**: Isaac Lab (same adapter pattern, PhysX backend, sim-to-real)

---

## TODO: Step 4 — Combine + Full Benchmark Suite

**Goal**: Run PPO v2 and CrossQ+norm on MuJoCo envs. Record wall-clock and final reward (mean ± std, 4 seeds). This is the "before/after" comparison for the roadmap.

**Runs to dispatch** (via dstack, see `docs/BENCHMARKS.md`):

| Algorithm | Env | Spec | Frames |
|-----------|-----|------|--------|
| PPO v2 | HalfCheetah-v5 | `ppo_mujoco_v2_arc.yaml` | 1M |
| PPO v2 | Humanoid-v5 | `ppo_mujoco_v2_arc.yaml` | 2M |
| SAC v2 | HalfCheetah-v5 | `sac_mujoco_v2_arc.yaml` | 1M |
| SAC v2 | Humanoid-v5 | `sac_mujoco_v2_arc.yaml` | 2M |
| CrossQ | HalfCheetah-v5 | `crossq_mujoco.yaml` | 4M |
| CrossQ | Humanoid-v5 | `crossq_mujoco.yaml` | 1M |
| CrossQ | Hopper-v5 | `crossq_mujoco.yaml` | 3M |
| CrossQ | Ant-v5 | `crossq_mujoco.yaml` | 2M |

**Verify**: Both algorithms beat their v1/SAC baselines on at least 2/3 envs.

**Local testing results (200k frames, 4 sessions)**:
- PPO v2 (layer norm + percentile) beats baseline on Humanoid (272.67 vs 246.83, consistency 0.78 vs 0.70)
- Layer norm is the most reliable individual feature — helps on LunarLander (+56%) and Humanoid (+8%)
- CrossQ beats SAC on CartPole (383 vs 238), Humanoid (365 vs 356), with higher consistency
- CrossQ unstable on Ant (loss divergence) — may need tuning for high-dimensional action spaces

---

## Completed: Step 2 — Normalization Stack

**v2 = layer_norm + percentile normalization** (symlog dropped — harms model-free RL).

Changes:
- `net_util.py` / `mlp.py`: `layer_norm` and `batch_norm` params in `build_fc_model()` / `MLPNet`
- `actor_critic.py`: `PercentileNormalizer` (EMA-tracked 5th/95th percentile advantage normalization)
- `math_util.py`: `symlog` / `symexp` (retained but excluded from v2 defaults)
- `ppo.py` / `sac.py`: symlog + percentile normalization integration
- Specs: `ppo_mujoco_v2_arc.yaml`, `sac_mujoco_v2_arc.yaml`

## Completed: Step 3 — CrossQ

CrossQ: SAC variant with no target networks. Uses cross-batch normalization on concatenated (s,a) and (s',a') batches.

Changes:
- `crossq.py`: CrossQ algorithm inheriting from SAC
- `algorithm/__init__.py`: CrossQ import
- Specs: `benchmark/crossq/crossq_mujoco.yaml`, `crossq_classic.yaml`, `crossq_box2d.yaml`, `crossq_atari.yaml`
- Future: actor LayerNorm (TorchArc YAML) — may help underperforming envs (Hopper, InvPendulum, Humanoid)

---

## TODO: Step 5 — High-UTD SAC / RLPD

**Goal**: `utd_ratio` alias for `training_iter`, demo buffer via `ReplayWithDemos` subclass.

Changes:
- `sac.py`: `utd_ratio` alias for `training_iter`
- `replay.py`: `ReplayWithDemos` subclass (50/50 symmetric sampling from demo and online data)
- Spec: `sac_mujoco_highutd_arc.yaml` (UTD=20 + layer norm critic)

**Verify**: High-UTD SAC on Hopper-v5 — converge in ~50% fewer env steps vs standard SAC.

## TODO: Step 6 — Pretrained Vision Encoders

**Goal**: DINOv2 encoder via torcharc, DrQ augmentation wrapper.

Changes:
- `pretrained.py`: `PretrainedEncoder` module (DINOv2, freeze/fine-tune, projection)
- `wrappers.py`: `RandomShiftWrapper` (DrQ-v2 ±4px shift augmentation)
- Spec: `ppo_vision_arc.yaml`

**Verify**: PPO with DINOv2 on DMControl pixel tasks (Walker Walk, Cartpole Swingup). Frozen vs fine-tuned comparison.

---

## Environment Plan (Future)

Three tiers of environment coverage:

| Tier | Platform | Purpose |
|------|----------|---------|
| **Broad/Basic** | [Gymnasium](https://gymnasium.farama.org/) | Standard RL benchmarks (CartPole, MuJoCo, Atari) |
| **Physics-rich** | [DeepMind MuJoCo Playground](https://github.com/google-deepmind/mujoco_playground) | GPU-accelerated locomotion, manipulation, dexterous tasks |
| **Sim-to-real** | [NVIDIA Isaac Lab](https://github.com/isaac-sim/IsaacLab) | GPU-accelerated, sim-to-real transfer, robot learning |

---

## Key Findings

- **Symlog harms model-free RL**: Tested across 6 envs — consistently hurts PPO and SAC. Designed for DreamerV3's world model, not direct value/Q-target compression.
- **Layer norm is the most reliable feature**: Helps on harder envs (LunarLander +56%, Humanoid +8%), neutral on simple envs.
- **CrossQ unstable on some envs**: Loss divergence on Ant and Hopper. Stable on CartPole, Humanoid. May need batch norm tuning for high-dimensional action spaces.
- **Features help more on harder envs**: Simple envs (CartPole, Acrobot) — baseline wins. Complex envs (Humanoid) — v2 and CrossQ win.

## Not in Scope

- World models (DreamerV3/RSSM) — Dasein Agent Phase 3.2c
- Plasticity loss mitigation (CReLU, periodic resets) — future work
- PPO+ principled fixes (ICLR 2025) — evaluate after base normalization stack