Update model card: clean public release
Browse files
README.md
CHANGED
|
@@ -21,7 +21,7 @@ pipeline_tag: text-generation
|
|
| 21 |
**Creator & Lead Contributor:** Brian Langay
|
| 22 |
**Contact:** support@openbnet.com · services@openbnet.cloud
|
| 23 |
**Paper:** [NEXUS: Design, Implementation, and Empirical Evaluation of a Lightweight Neural Controller for LLM Agent Systems](https://github.com/brian-Lab-0/nexus/blob/main/NEXUS_Implementation_Report.pdf)
|
| 24 |
-
**Code:** [github.com/
|
| 25 |
**License:** Apache 2.0
|
| 26 |
|
| 27 |
---
|
|
@@ -36,7 +36,7 @@ NEXUS comprises five subsystems:
|
|
| 36 |
|---|---:|---|
|
| 37 |
| Protocol Cortex (TSM) | 4,474,624 | KV-cache task vector injection |
|
| 38 |
| Belief Engine (BTBS) | 601,234 | Mamba SSM particle filter; tracks P(done) |
|
| 39 |
-
| Resource Router (FSM-NHC) | 184,213 | 7-class
|
| 40 |
| SAC Corrector | 454,273 | Semantic drift correction patches |
|
| 41 |
| Adapter Switch (TALoRA) | 42,373 | LoRA routing by sub-task type |
|
| 42 |
| Drift Sentinel | 33,287 | Drift detection from trajectory buffer |
|
|
@@ -59,7 +59,7 @@ NEXUS comprises five subsystems:
|
|
| 59 |
|
| 60 |
## Training Data
|
| 61 |
|
| 62 |
-
|
| 63 |
|
| 64 |
| Component | Training samples | Distribution |
|
| 65 |
|---|---:|---|
|
|
@@ -68,8 +68,6 @@ Current checkpoints are trained on **synthetic data** generated to approximate C
|
|
| 68 |
| Belief Engine | ~1,700 train / 300 val | Sigmoid completion ramps |
|
| 69 |
| Drift Sentinel | Synthetic trajectories | Orthogonal rotation drift injection |
|
| 70 |
|
| 71 |
-
Phase 1 will replace synthetic data with real Chatp production traces via `NexusTrainingLog`.
|
| 72 |
-
|
| 73 |
---
|
| 74 |
|
| 75 |
## Training Procedure
|
|
@@ -94,14 +92,14 @@ Evaluated on a 200-task synthetic benchmark (T=15 steps, 35% drift fraction, see
|
|
| 94 |
| Token Overhead Ratio ↓ | 99.95% | **0.00%** | −100% |
|
| 95 |
| Tool Routing Accuracy ↑ | 13.5% | **14.0%** | +3.7% |
|
| 96 |
|
| 97 |
-
**
|
| 98 |
|
| 99 |
-
|
|
| 100 |
|---|---|---|
|
| 101 |
-
|
|
| 102 |
-
|
|
| 103 |
-
|
|
| 104 |
-
|
|
| 105 |
|
| 106 |
---
|
| 107 |
|
|
@@ -116,7 +114,7 @@ cfg = MetacontrolConfig()
|
|
| 116 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 117 |
|
| 118 |
pipeline = MetacontrolPipeline(cfg).to(device)
|
| 119 |
-
pipeline.load_checkpoint("checkpoints/")
|
| 120 |
pipeline.reset(batch_size=1, device=device)
|
| 121 |
|
| 122 |
# Each step: pass LLM hidden states + goal embedding
|
|
@@ -132,13 +130,14 @@ print(result["state_summary"])
|
|
| 132 |
# Drift: 0.766
|
| 133 |
```
|
| 134 |
|
| 135 |
-
###
|
| 136 |
|
| 137 |
```bash
|
| 138 |
git clone https://github.com/brian-Lab-0/nexus
|
| 139 |
cd nexus
|
| 140 |
-
pip install -e .
|
| 141 |
-
python
|
|
|
|
| 142 |
```
|
| 143 |
|
| 144 |
---
|
|
@@ -158,7 +157,7 @@ Tested on: NVIDIA RTX 4060 Laptop GPU (8 GB VRAM), CUDA 12.8, PyTorch 2.9.1+cu12
|
|
| 158 |
|
| 159 |
## Limitations
|
| 160 |
|
| 161 |
-
- **Synthetic training gap** — TTCS/DRP metrics require real
|
| 162 |
- **LTVI pending** — KV-cache injection path is functional but Protocol Cortex needs training on real traces for coherent generation
|
| 163 |
- **Drift sentinel** — 69% accuracy; natural drift is more varied than synthetic orthogonal rotation
|
| 164 |
|
|
@@ -173,6 +172,7 @@ Tested on: NVIDIA RTX 4060 Laptop GPU (8 GB VRAM), CUDA 12.8, PyTorch 2.9.1+cu12
|
|
| 173 |
year = {2026},
|
| 174 |
publisher = {OpenBnet},
|
| 175 |
url = {https://github.com/brian-Lab-0/nexus},
|
|
|
|
| 176 |
}
|
| 177 |
```
|
| 178 |
|
|
|
|
| 21 |
**Creator & Lead Contributor:** Brian Langay
|
| 22 |
**Contact:** support@openbnet.com · services@openbnet.cloud
|
| 23 |
**Paper:** [NEXUS: Design, Implementation, and Empirical Evaluation of a Lightweight Neural Controller for LLM Agent Systems](https://github.com/brian-Lab-0/nexus/blob/main/NEXUS_Implementation_Report.pdf)
|
| 24 |
+
**Code:** [github.com/brian-Lab-0/nexus](https://github.com/brian-Lab-0/nexus)
|
| 25 |
**License:** Apache 2.0
|
| 26 |
|
| 27 |
---
|
|
|
|
| 36 |
|---|---:|---|
|
| 37 |
| Protocol Cortex (TSM) | 4,474,624 | KV-cache task vector injection |
|
| 38 |
| Belief Engine (BTBS) | 601,234 | Mamba SSM particle filter; tracks P(done) |
|
| 39 |
+
| Resource Router (FSM-NHC) | 184,213 | 7-class tool classifier |
|
| 40 |
| SAC Corrector | 454,273 | Semantic drift correction patches |
|
| 41 |
| Adapter Switch (TALoRA) | 42,373 | LoRA routing by sub-task type |
|
| 42 |
| Drift Sentinel | 33,287 | Drift detection from trajectory buffer |
|
|
|
|
| 59 |
|
| 60 |
## Training Data
|
| 61 |
|
| 62 |
+
Checkpoints are trained on **synthetic data** approximating Chatp production agent interaction patterns:
|
| 63 |
|
| 64 |
| Component | Training samples | Distribution |
|
| 65 |
|---|---:|---|
|
|
|
|
| 68 |
| Belief Engine | ~1,700 train / 300 val | Sigmoid completion ramps |
|
| 69 |
| Drift Sentinel | Synthetic trajectories | Orthogonal rotation drift injection |
|
| 70 |
|
|
|
|
|
|
|
| 71 |
---
|
| 72 |
|
| 73 |
## Training Procedure
|
|
|
|
| 92 |
| Token Overhead Ratio ↓ | 99.95% | **0.00%** | −100% |
|
| 93 |
| Tool Routing Accuracy ↑ | 13.5% | **14.0%** | +3.7% |
|
| 94 |
|
| 95 |
+
**Training results (val metrics):**
|
| 96 |
|
| 97 |
+
| Component | Metric | Value |
|
| 98 |
|---|---|---|
|
| 99 |
+
| Resource Router | Val accuracy | **95.5%** |
|
| 100 |
+
| Sub-task Classifier | Val accuracy | **99.8%** |
|
| 101 |
+
| Belief Engine | Val loss (MSE) | **7×10⁻⁵** |
|
| 102 |
+
| Drift Sentinel | Val accuracy | **69.0%** |
|
| 103 |
|
| 104 |
---
|
| 105 |
|
|
|
|
| 114 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 115 |
|
| 116 |
pipeline = MetacontrolPipeline(cfg).to(device)
|
| 117 |
+
pipeline.load_checkpoint("checkpoints/")
|
| 118 |
pipeline.reset(batch_size=1, device=device)
|
| 119 |
|
| 120 |
# Each step: pass LLM hidden states + goal embedding
|
|
|
|
| 130 |
# Drift: 0.766
|
| 131 |
```
|
| 132 |
|
| 133 |
+
### Quick Start
|
| 134 |
|
| 135 |
```bash
|
| 136 |
git clone https://github.com/brian-Lab-0/nexus
|
| 137 |
cd nexus
|
| 138 |
+
pip install -e ".[dev]"
|
| 139 |
+
python -m pytest # 155 tests, should all pass
|
| 140 |
+
python scripts/evaluate.py --n-tasks 200 --device cuda
|
| 141 |
```
|
| 142 |
|
| 143 |
---
|
|
|
|
| 157 |
|
| 158 |
## Limitations
|
| 159 |
|
| 160 |
+
- **Synthetic training gap** — TTCS/DRP metrics require real production traces to differentiate from baseline
|
| 161 |
- **LTVI pending** — KV-cache injection path is functional but Protocol Cortex needs training on real traces for coherent generation
|
| 162 |
- **Drift sentinel** — 69% accuracy; natural drift is more varied than synthetic orthogonal rotation
|
| 163 |
|
|
|
|
| 172 |
year = {2026},
|
| 173 |
publisher = {OpenBnet},
|
| 174 |
url = {https://github.com/brian-Lab-0/nexus},
|
| 175 |
+
note = {6.29M-parameter neural controller for LLM agent systems}
|
| 176 |
}
|
| 177 |
```
|
| 178 |
|