Half-Life Regularization Experiment Summary
Generated: 2026-01-22T14:32:53.545187
Overview
This experiment suite addresses the half-life collapse problem discovered by Melanie/Tiago:
"After training at GPT-2 scale, oscillator half-lives collapse to ~10 steps."
Key Results
Collapse and Recovery
The half-life regularizer successfully provides gradients to restore long-range oscillators:
- Initial distribution: Log-uniform over [1, 4096]
- Collapsed distribution: All < 10 steps
- After regularization step: Distribution spreads back toward target
Identity Reconstruction
| Condition | Verdict | Critical K |
|---|---|---|
| Without Regularization | FAIL (GRADUAL DRIFT) | 128 |
| With Regularization | PASS (PHASE TRANSITION) | 64 |
Conclusion
Half-life regularization is decisive for long-context coherence.
The experiment confirms:
- Half-life collapse prevents long-range identity preservation
- The regularizer restores the capacity for long-context reasoning
- This validates the hypothesis from Melanie/Tiago's discovery
Files Included
collapse_recovery.json- Half-life collapse/recovery dataidentity_reconstruction/- Full experiment resultsPRESENTATION_HALF_LIFE_REGULARIZATION.md- Slidesall_results.json- Complete results data
Recommendations
- Integrate
HalfLifeRegularizerinto FDRA training loss - Set
lambda1 = 0.01,lambda2 = 0.01as starting points - Monitor half-life histogram during training
- Test on long-context benchmarks (QA, summarization)
Generated by run_half_life_experiment.py