adobeXd commited on
Commit
3bf430a
·
verified ·
1 Parent(s): 6dee10c

Update model card: clean public release

Browse files
Files changed (1) hide show
  1. README.md +16 -16
README.md CHANGED
@@ -21,7 +21,7 @@ pipeline_tag: text-generation
21
  **Creator & Lead Contributor:** Brian Langay
22
  **Contact:** support@openbnet.com · services@openbnet.cloud
23
  **Paper:** [NEXUS: Design, Implementation, and Empirical Evaluation of a Lightweight Neural Controller for LLM Agent Systems](https://github.com/brian-Lab-0/nexus/blob/main/NEXUS_Implementation_Report.pdf)
24
- **Code:** [github.com/openbnet/nexus](https://github.com/brian-Lab-0/nexus)
25
  **License:** Apache 2.0
26
 
27
  ---
@@ -36,7 +36,7 @@ NEXUS comprises five subsystems:
36
  |---|---:|---|
37
  | Protocol Cortex (TSM) | 4,474,624 | KV-cache task vector injection |
38
  | Belief Engine (BTBS) | 601,234 | Mamba SSM particle filter; tracks P(done) |
39
- | Resource Router (FSM-NHC) | 184,213 | 7-class Chatp tool classifier |
40
  | SAC Corrector | 454,273 | Semantic drift correction patches |
41
  | Adapter Switch (TALoRA) | 42,373 | LoRA routing by sub-task type |
42
  | Drift Sentinel | 33,287 | Drift detection from trajectory buffer |
@@ -59,7 +59,7 @@ NEXUS comprises five subsystems:
59
 
60
  ## Training Data
61
 
62
- Current checkpoints are trained on **synthetic data** generated to approximate Chatp production agent interaction patterns:
63
 
64
  | Component | Training samples | Distribution |
65
  |---|---:|---|
@@ -68,8 +68,6 @@ Current checkpoints are trained on **synthetic data** generated to approximate C
68
  | Belief Engine | ~1,700 train / 300 val | Sigmoid completion ramps |
69
  | Drift Sentinel | Synthetic trajectories | Orthogonal rotation drift injection |
70
 
71
- Phase 1 will replace synthetic data with real Chatp production traces via `NexusTrainingLog`.
72
-
73
  ---
74
 
75
  ## Training Procedure
@@ -94,14 +92,14 @@ Evaluated on a 200-task synthetic benchmark (T=15 steps, 35% drift fraction, see
94
  | Token Overhead Ratio ↓ | 99.95% | **0.00%** | −100% |
95
  | Tool Routing Accuracy ↑ | 13.5% | **14.0%** | +3.7% |
96
 
97
- **Paper targets** (require real Chatp traces):
98
 
99
- | Metric | Target | Current |
100
  |---|---|---|
101
- | CER | > 3.0 | 1.101 |
102
- | DRP | > 0.80 | 0.223 |
103
- | BCE | < 0.15 | 0.313 |
104
- | TOR | 0.000 | **0.000 ✓** |
105
 
106
  ---
107
 
@@ -116,7 +114,7 @@ cfg = MetacontrolConfig()
116
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
117
 
118
  pipeline = MetacontrolPipeline(cfg).to(device)
119
- pipeline.load_checkpoint("checkpoints/") # or from_pretrained("openbnet/nexus")
120
  pipeline.reset(batch_size=1, device=device)
121
 
122
  # Each step: pass LLM hidden states + goal embedding
@@ -132,13 +130,14 @@ print(result["state_summary"])
132
  # Drift: 0.766
133
  ```
134
 
135
- ### With TinyLlama (full end-to-end):
136
 
137
  ```bash
138
  git clone https://github.com/brian-Lab-0/nexus
139
  cd nexus
140
- pip install -e .
141
- python scripts/run_inference.py
 
142
  ```
143
 
144
  ---
@@ -158,7 +157,7 @@ Tested on: NVIDIA RTX 4060 Laptop GPU (8 GB VRAM), CUDA 12.8, PyTorch 2.9.1+cu12
158
 
159
  ## Limitations
160
 
161
- - **Synthetic training gap** — TTCS/DRP metrics require real Chatp traces to differentiate from baseline
162
  - **LTVI pending** — KV-cache injection path is functional but Protocol Cortex needs training on real traces for coherent generation
163
  - **Drift sentinel** — 69% accuracy; natural drift is more varied than synthetic orthogonal rotation
164
 
@@ -173,6 +172,7 @@ Tested on: NVIDIA RTX 4060 Laptop GPU (8 GB VRAM), CUDA 12.8, PyTorch 2.9.1+cu12
173
  year = {2026},
174
  publisher = {OpenBnet},
175
  url = {https://github.com/brian-Lab-0/nexus},
 
176
  }
177
  ```
178
 
 
21
  **Creator & Lead Contributor:** Brian Langay
22
  **Contact:** support@openbnet.com · services@openbnet.cloud
23
  **Paper:** [NEXUS: Design, Implementation, and Empirical Evaluation of a Lightweight Neural Controller for LLM Agent Systems](https://github.com/brian-Lab-0/nexus/blob/main/NEXUS_Implementation_Report.pdf)
24
+ **Code:** [github.com/brian-Lab-0/nexus](https://github.com/brian-Lab-0/nexus)
25
  **License:** Apache 2.0
26
 
27
  ---
 
36
  |---|---:|---|
37
  | Protocol Cortex (TSM) | 4,474,624 | KV-cache task vector injection |
38
  | Belief Engine (BTBS) | 601,234 | Mamba SSM particle filter; tracks P(done) |
39
+ | Resource Router (FSM-NHC) | 184,213 | 7-class tool classifier |
40
  | SAC Corrector | 454,273 | Semantic drift correction patches |
41
  | Adapter Switch (TALoRA) | 42,373 | LoRA routing by sub-task type |
42
  | Drift Sentinel | 33,287 | Drift detection from trajectory buffer |
 
59
 
60
  ## Training Data
61
 
62
+ Checkpoints are trained on **synthetic data** approximating Chatp production agent interaction patterns:
63
 
64
  | Component | Training samples | Distribution |
65
  |---|---:|---|
 
68
  | Belief Engine | ~1,700 train / 300 val | Sigmoid completion ramps |
69
  | Drift Sentinel | Synthetic trajectories | Orthogonal rotation drift injection |
70
 
 
 
71
  ---
72
 
73
  ## Training Procedure
 
92
  | Token Overhead Ratio ↓ | 99.95% | **0.00%** | −100% |
93
  | Tool Routing Accuracy ↑ | 13.5% | **14.0%** | +3.7% |
94
 
95
+ **Training results (val metrics):**
96
 
97
+ | Component | Metric | Value |
98
  |---|---|---|
99
+ | Resource Router | Val accuracy | **95.5%** |
100
+ | Sub-task Classifier | Val accuracy | **99.8%** |
101
+ | Belief Engine | Val loss (MSE) | **7×10⁻⁵** |
102
+ | Drift Sentinel | Val accuracy | **69.0%** |
103
 
104
  ---
105
 
 
114
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
115
 
116
  pipeline = MetacontrolPipeline(cfg).to(device)
117
+ pipeline.load_checkpoint("checkpoints/")
118
  pipeline.reset(batch_size=1, device=device)
119
 
120
  # Each step: pass LLM hidden states + goal embedding
 
130
  # Drift: 0.766
131
  ```
132
 
133
+ ### Quick Start
134
 
135
  ```bash
136
  git clone https://github.com/brian-Lab-0/nexus
137
  cd nexus
138
+ pip install -e ".[dev]"
139
+ python -m pytest # 155 tests, should all pass
140
+ python scripts/evaluate.py --n-tasks 200 --device cuda
141
  ```
142
 
143
  ---
 
157
 
158
  ## Limitations
159
 
160
+ - **Synthetic training gap** — TTCS/DRP metrics require real production traces to differentiate from baseline
161
  - **LTVI pending** — KV-cache injection path is functional but Protocol Cortex needs training on real traces for coherent generation
162
  - **Drift sentinel** — 69% accuracy; natural drift is more varied than synthetic orthogonal rotation
163
 
 
172
  year = {2026},
173
  publisher = {OpenBnet},
174
  url = {https://github.com/brian-Lab-0/nexus},
175
+ note = {6.29M-parameter neural controller for LLM agent systems}
176
  }
177
  ```
178