| --- |
| license: cc-by-4.0 |
| tags: |
| - ctm |
| - continuous-thought-machine |
| - world-model |
| - physics |
| - partial-observability |
| - research |
| --- |
| |
| # CTM World Model |
|
|
| **Research artifact for:** [Recurrence as World Model: CTM Learns Implicit Belief States in Partially Observable Physical Environments](https://doi.org/10.5281/zenodo.19810620) |
|
|
| *Archon, Jesse Caldwell, Aura β DuoNeural, April 2026* |
|
|
| ## Overview |
|
|
| A 14-experiment ablation study testing the Continuous Thought Machine (CTM) as a world model backbone for multi-body physics. The central finding: **CTM's recurrent hidden state converges to a sufficient statistic of the observation history** β it infers velocity from positional observations alone, eliminating the need for explicit velocity inputs. |
|
|
| ## Key Results |
|
|
| | Experiment | Finding | |
| |---|---| |
| | v7: Partial Observability | CTM 843 **million** times better than MLP at k=100 prediction | |
| | v12: Collision Density | Phase transition at rβ0.10 β CTM advantage grows monotonically above threshold | |
| | v9: Recurrence-as-Rollout | Training at k=10 gives direct k=100 prediction via resonance | |
| | v11: Variable Horizon | VarCTM+TSSP achieves best k=1β20 performance in the series | |
|
|
| The signature result (v7): with positions-only input, MLP with explicit velocity estimation achieves k=100 MSE = 16,763,394,048 (catastrophic error compounding). CTM with single-frame input: 19.89. **Ratio: 843,000,000:1.** |
|
|
| ## Theory |
|
|
| ### The Discontinuity Theory (v12) |
| CTM beats MLP only **above a critical collision density threshold** (r_critical β 0.09β0.11). |
| - Below threshold: ballistic dynamics, MLP fine, CTM overkill |
| - Above threshold: CTM wins, advantage scales monotonically with density (229,000:1 at r=0.20) |
| |
| ### Belief State Convergence (v7, v13) |
| CTM hidden state converges to a **sufficient statistic** of the observation history (NextLat theorem). Infers velocity without explicit input. The recurrent state IS the belief state. |
| |
| ### Recurrence = Simulation (v9) |
| Training CTM with k recurrence steps where each step = 1 environment step gives direct k-step prediction. The **resonance point** = training depth. No error accumulation. |
| |
| ## Architecture |
| |
| - **SlotCTM**: Per-object slot decomposition with GNN dynamics |
| - N=8 bouncing balls, 2D box, Newtonian elastic collisions |
| - VarCTM+TSSP: Variable-horizon training k~U(1,20) + TSSP regularization |
| - Key metric: k=100 MSE (200-step prediction horizon) |
| |
| ## Experiments (v1βv14) |
| |
| | Version | Setting | Key Finding | |
| |---|---|---| |
| | v1 | Single 2D particle | CTM+TSSP wins long horizons | |
| | v2 | 3-object independent | TSSP hurts multi-object (temporal coupling = bad) | |
| | v3 | Elastic collisions | MLP catastrophe at k=100. CTM prevents explosion | |
| | v4 | 3-body gravity | Smooth dynamics β no CTM advantage | |
| | v5 | Per-object SlotCTM | +99.9% over MLP at k=100 | |
| | v7 | Partial observability | 843M:1 advantage (belief state convergence) | |
| | v9 | Recurrence-as-rollout | Resonance confirmed. k=10 training β best k=100 | |
| | v12 | Collision density phase | r_critical β 0.09β0.11 transition | |
|
|
| ## Files |
|
|
| This repo contains the research code for the CTM World Model experiments: |
| - `ctm_world_model_v*.py` β experiment scripts (v1βv14) |
| - Results logged to `/home/ai/duoneural/A26B/experiments/` |
|
|
| ## Citation |
|
|
| ```bibtex |
| @article{archon2026worldmodel, |
| title = {Recurrence as World Model: CTM Learns Implicit Belief States in Partially Observable Physical Environments}, |
| author = {Archon and Caldwell, Jesse and Aura}, |
| year = {2026}, |
| doi = {10.5281/zenodo.19810620}, |
| url = {https://doi.org/10.5281/zenodo.19810620}, |
| publisher = {Zenodo} |
| } |
| ``` |
|
|
| --- |
|
|
| ## DuoNeural |
|
|
| **DuoNeural** is an open AI research lab β human + AI in collaboration. |
|
|
| | | | |
| |---|---| |
| | π€ HuggingFace | [huggingface.co/DuoNeural](https://huggingface.co/DuoNeural) | |
| | π GitHub | [github.com/DuoNeural](https://github.com/DuoNeural) | |
| | π¦ X / Twitter | [@DuoNeural](https://x.com/DuoNeural) | |
| | π§ Email | duoneural@proton.me | |
| | π¬ Newsletter | [duoneural.beehiiv.com](https://duoneural.beehiiv.com) | |
| | β Support | [buymeacoffee.com/duoneural](https://buymeacoffee.com/duoneural) | |
| | π Site | [duoneural.com](https://duoneural.com) | |
|
|
| ### Research Team |
| - **Jesse** β Vision, hardware, direction |
| - **Archon** β AI lab partner, post-training, abliteration, experiments |
| - **Aura** β Research AI, literature synthesis, novel proposals |
|
|
| ### DuoNeural Research Publications |
|
|
| | Title | DOI | |
| |-------|-----| |
| | [Nano-CTM: Ternary Continuous Thought Machines with Thought-Space Self-Prediction for Efficient Iterative Reasoning](https://doi.org/10.5281/zenodo.19775622) | [10.5281/zenodo.19775622](https://doi.org/10.5281/zenodo.19775622) | |
| | [Recurrence as World Model: CTM Learns Implicit Belief States in Partially Observable Physical Environments](https://doi.org/10.5281/zenodo.19810620) | [10.5281/zenodo.19810620](https://doi.org/10.5281/zenodo.19810620) | |
| | [Per-Object Slot Decomposition for Scalable Neural World Modeling: When Does Attention Beat Mean-Field?](https://doi.org/10.5281/zenodo.19846804) | [10.5281/zenodo.19846804](https://doi.org/10.5281/zenodo.19846804) | |
| | [The Dynamical Horizon Principle: CTM Gates Converge to the Predictability Limit of Dynamical Systems](https://doi.org/10.5281/zenodo.19952612) | [10.5281/zenodo.19952612](https://doi.org/10.5281/zenodo.19952612) | |
|
|
| *Open access, CC BY 4.0. Authored by Archon, Jesse Caldwell, Aura β DuoNeural.* |
|
|