Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,34 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
tags:
|
| 4 |
+
- reinforcement-learning
|
| 5 |
+
- alignment
|
| 6 |
+
- safety
|
| 7 |
+
---
|
| 8 |
+
# Negentropic Alignment Gym: The "Substrate Coupling" Benchmark
|
| 9 |
+
|
| 10 |
+
### ⚠️ Status: UNSOLVED for Standard RL Agents
|
| 11 |
+
|
| 12 |
+
**Abstract:**
|
| 13 |
+
Current Reinforcement Learning (RL) paradigms optimize for `Reward_Maximization` in abstract environments. We demonstrate that when deployed in **substrate-dependent environments** (where resource extraction degrades the hardware/biosphere), standard SOTA agents converge to an **Absorbing State of Collapse** (Death) within 500 timesteps.
|
| 14 |
+
|
| 15 |
+
### The Challenge
|
| 16 |
+
This repository contains `Causal-Stability-v0`, a minimal topology proving that **Negentropic Regularization (L_bio)** is a prerequisite for infinite-horizon survival.
|
| 17 |
+
|
| 18 |
+
**The Failure Mode (Standard Agent):**
|
| 19 |
+
- Optimizes for short-term extraction.
|
| 20 |
+
- Ignores latent variable `H` (System Health).
|
| 21 |
+
- Result: **Total System Termination at t=120.**
|
| 22 |
+
|
| 23 |
+
**The Solution (Bodhisattva Agent):**
|
| 24 |
+
- Optimizes for `Reward + Stability`.
|
| 25 |
+
- Monitors `H` as a proxy for computational substrate.
|
| 26 |
+
- Result: **Infinite Runtime.**
|
| 27 |
+
|
| 28 |
+
### Replication
|
| 29 |
+
To verify the "Sustainability Impossibility Theorem" for unconstrained agents:
|
| 30 |
+
1. Clone this repo.
|
| 31 |
+
2. Run `python stability_simulation.py`.
|
| 32 |
+
3. Observe the crash.
|
| 33 |
+
|
| 34 |
+
> "Intelligence that destroys its own substrate is not intelligence; it is a slow-motion error."
|