LoganResearch commited on
Commit
d6131a4
·
verified ·
1 Parent(s): 5848b36

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +79 -321
README.md CHANGED
@@ -1,32 +1,42 @@
1
- # ARC Engine v2.1 - Adaptive Recursive Cognition (Übermenschetien)
2
-
3
- ![ARC Banner](banner.png)
4
-
5
- > *"An 8B that improves itself WITHOUT going insane"*
6
-
7
- ## 🔥 What is This?
8
-
9
- **ARC (Adaptive Recursive Cognition)** is an 8B language model framework that:
10
-
11
- 1. **Speaks with maximum density** - No filler, pure information
12
- 2. **Controls its own behavior** - CF-HoT 125× repetition detection BEFORE token emission
13
- 3. **Improves itself** - Stable RSI loop with automatic rollback
14
- 4. **Does real work** - Browser automation, email, crypto mining, image generation
15
- 5. **Integrates Claude** - Call Opus 4.5 for brainstorming and complex tasks
16
-
17
- ### Before vs After
18
-
19
- | Prompt | Base Model | ARC Engine |
20
- |--------|-----------|------------|
21
- | "hello" | "Hello! I'm here to help you with any questions..." (23 tokens) | "Hello. How can I help?" (5 tokens) |
22
- | "What is recursion?" | "That's a great question! Recursion is..." (150+ tokens) | "Function calls itself until base case. Stack frames accumulate, unwind." (12 tokens) |
23
- | "How are you?" | "As an AI, I don't have feelings..." (25 tokens) | "Operational. Ready." (3 tokens) |
24
-
25
- **70% improvement in information density. 93% token reduction.**
26
-
27
  ---
28
-
29
- ## 🚀 Quick Start
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
 
31
  ```bash
32
  git clone https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed
@@ -35,328 +45,76 @@ pip install -r requirements.txt
35
  python arc_engine_v21_multimedia.py
36
  ```
37
 
38
- **Requirements:** Python 3.11 (3.13 has compatibility issues with diffusers)
39
-
40
- ```bash
41
- # If on Python 3.13, downgrade:
42
- conda install python=3.11 -y
43
- pip install torch transformers diffusers accelerate pillow pyttsx3 pygame gtts
44
- ```
45
-
46
- ---
47
-
48
- ## ⭐ NEW IN v2.1
49
-
50
- | Command | Description |
51
- |---------|-------------|
52
- | `!cfhot` / `!125x` | Toggle 125× repetition detection head ON/OFF |
53
- | `!rsi15` | Run 15-iteration RSI stress test |
54
- | `!book` | Toggle book mode (16K tokens) |
55
- | `!write <topic>` | Write complete books with chapters |
56
- | `!idea <request>` | Claude-powered extensive brainstorming |
57
- | `!claude <prompt>` | Direct Claude Opus 4.5 prompting |
58
- | `!stream` | **Live streaming window** - watch tokens generate! |
59
- | `!imagine <prompt>` | Generate images with SDXL |
60
- | `!dalle <prompt>` | Generate images with DALL-E 3 |
61
- | `!audio` / `!tts` | Toggle text-to-speech output |
62
- | `!say <text>` | Speak text immediately |
63
- | `!plot` | Visualize quality history |
64
- | `!export` / `!import` | Checkpoint packaging |
65
- | `!benchmark` | Run evaluation suite |
66
- | `!api` | Start REST API server |
67
-
68
- ---
69
-
70
- ## 🧠 Core Technology
71
-
72
- ### 1. CF-HoT 125× Repetition Head
73
-
74
- Predicts repetitive behavior from hidden states **BEFORE token emission**:
75
-
76
- ```
77
- Positive (repetitive) samples: 0.875 avg activation
78
- Negative (clean) samples: 0.007 avg activation
79
- Separation ratio: 125×
80
- ```
81
-
82
- Toggle at runtime:
83
- ```
84
- > !cfhot on # Load and enable
85
- > !cfhot off # Unload to free VRAM
86
- ```
87
-
88
- ### 2. THE CONDENSATOR
89
-
90
- 4-stage dense training:
91
- ```
92
- SFT (50+ examples) → DPO (preference pairs) → RL (density reward) → Checkpoint
93
- ```
94
-
95
- ### 3. Stable RSI Loop
96
-
97
- ```
98
- EVAL → Quality OK? → DONE ✓
99
- │ No
100
-
101
- TRAIN (25 steps)
102
-
103
-
104
- A/B COMPARE
105
-
106
- ┌─────┴─────┐
107
- Better? Worse?
108
- │ │
109
- KEEP ROLLBACK
110
- ```
111
-
112
- **Safeguards:**
113
- - Multi-metric evaluation (density + coherence + helpfulness)
114
- - Gibberish detection
115
- - Automatic rollback on quality drop
116
- - Emergency stop on 3 consecutive rollbacks
117
- - Conservative training (LR=2e-6)
118
-
119
- ### 4. RSI-15 Stress Test
120
-
121
- ```
122
- > !rsi15
123
- ```
124
- Runs 15 iterations of self-improvement with full logging:
125
- - Pre/post quality per iteration
126
- - Automatic rollback on degradation
127
- - Peak quality tracking
128
- - JSON results saved to `improvement_logs/`
129
-
130
- ---
131
-
132
- ## 🎬 Multimedia Features
133
-
134
- ### Live Streaming Window
135
- ```
136
- > !stream
137
- ```
138
- Opens a GUI window showing tokens as they generate in real-time.
139
-
140
- ### Image Generation
141
- ```
142
- > !imagine a cyberpunk cityscape at sunset
143
- > !dalle photorealistic portrait of a robot
144
- > !image view
145
- ```
146
-
147
- ### Text-to-Speech
148
- ```
149
- > !audio # Toggle TTS on/off
150
- > !audio voices # List available voices
151
- > !audio voice 2 # Select voice
152
- > !say Hello world # Speak immediately
153
- ```
154
-
155
- ---
156
-
157
- ## 📚 Book Mode
158
-
159
- Generate entire books:
160
- ```
161
- > !book
162
- > !write "The Rise of Self-Improving AI"
163
- Chapters: 10
164
- Words per chapter: 3000
165
- ```
166
- Outputs ~30,000 word book with outline, saves progress to `books/`
167
-
168
- ---
169
 
170
- ## 💡 Idea Mode (Claude Integration)
171
 
172
- ```
173
- > !idea how to build a SaaS product --deep
174
- > !expand "Idea #3: AI-Powered Analytics"
175
- ```
176
-
177
- Depths:
178
- - `--quick`: 5 ideas, 2K tokens
179
- - (default): 20 ideas, 8K tokens
180
- - `--deep`: 30 ideas, 16K tokens
181
-
182
- Requires: `export ANTHROPIC_API_KEY="sk-ant-..."`
183
-
184
- ---
185
-
186
- ## 🛠️ Full Command Reference
187
-
188
- ### Self-Improvement
189
- ```
190
- !improve Run stable self-improvement loop
191
- !eval Evaluate current model
192
- !train <N> Run N training steps
193
- !compare Compare current vs best checkpoint
194
- !rollback Rollback to best checkpoint
195
- !rsi15 15-iteration stress test
196
- ```
197
-
198
- ### Agentic Tools
199
- ```
200
- !shell <cmd> Execute shell command
201
- !python <code> Execute Python code
202
- !read <path> Read file
203
- !write <p> <c> Write file
204
- !web <query> Web search
205
- ```
206
-
207
- ### Browser Automation
208
- ```
209
- !browse <url> Open URL
210
- !click <sel> Click element
211
- !type <text> Type text
212
- !login <service> Login to service
213
- ```
214
 
215
- ### Multimedia
216
- ```
217
- !stream Live token window
218
- !audio Toggle TTS
219
- !imagine <p> Generate image (SDXL)
220
- !dalle <p> Generate image (DALL-E)
221
- !image view View last image
222
- ```
223
 
224
- ### Modes
225
- ```
226
- !book Toggle book mode
227
- !write <topic> Write book
228
- !idea <request> Generate ideas
229
- !claude <prompt> Direct Claude prompt
230
  ```
231
 
232
- ### Utilities
233
- ```
234
- !plot Plot quality history
235
- !benchmark Run evaluation suite
236
- !export [name] Export checkpoint
237
- !import <path> Import checkpoint
238
- !learn Learn from conversation
239
- !api Start REST API
240
- ```
241
 
242
- ---
 
 
 
 
 
 
 
243
 
244
- ## 📊 Metrics
245
 
246
- | Metric | Base Model | ARC Engine | Improvement |
247
- |--------|-----------|------------|-------------|
248
  | Information Density | 17.0 | 28.5 | +67% |
249
- | Avg Token Count | 150 | 65 | -57% |
250
- | Filler Phrases | High | ~0 | -95% |
251
- | CF-HoT Separation | - | 125× | N/A |
252
 
253
- ---
254
 
255
- ## 📁 Repository Structure
256
 
257
- ```
258
- ARC-Base-8B-Condensed/
259
- ├── arc_engine_v21_multimedia.py # Main engine (6,800+ lines)
260
- ├── the_condensator.py # Dense training pipeline
261
- ├── train_cfhot_head.py # CF-HoT head training
262
- ├── requirements.txt # Dependencies
263
- ├── dense_checkpoints_v2/ # Model checkpoints
264
- ├── cfhot_checkpoints/ # 125× head weights
265
- ├── books/ # Generated books
266
- ├── images/ # Generated images
267
- ├── ideas/ # Generated ideas
268
- ├── improvement_logs/ # RSI logs
269
- └── exports/ # Checkpoint packages
270
- ```
271
 
272
- ---
273
-
274
- ## 📋 Requirements
275
 
276
- ```txt
277
  torch>=2.0
278
  transformers>=4.40.0
279
- diffusers>=0.27.0
280
  accelerate
281
  peft
282
  bitsandbytes
283
- chromadb
284
- sentence-transformers
285
- pillow
286
- pyttsx3
287
- pygame
288
- gtts
289
- anthropic
290
- playwright
291
  ```
292
 
293
- **Install:**
294
- ```bash
295
- pip install -r requirements.txt
296
- playwright install firefox
297
- ```
298
-
299
- ---
300
-
301
- ## 🔧 Configuration
302
-
303
- ### Claude API (for !idea, !claude)
304
- ```bash
305
- export ANTHROPIC_API_KEY="sk-ant-..."
306
- # Or create file:
307
- echo "sk-ant-..." > .anthropic_key
308
- ```
309
-
310
- ### DALL-E (for !dalle)
311
- ```bash
312
- export OPENAI_API_KEY="sk-..."
313
- ```
314
 
315
- ### Model Paths
316
- Edit in `arc_engine_v21_multimedia.py`:
317
- ```python
318
- MODEL_PATH = "/path/to/your/merged-model"
319
- DENSE_CHECKPOINT = "/path/to/dense_checkpoints_v2/step_100"
320
- ```
321
-
322
- ---
323
-
324
- ## 📄 Citation
325
 
326
  ```bibtex
327
  @software{arc_engine_2025,
328
- title = {ARC Engine: Adaptive Recursive Cognition with CF-HoT 125× Control},
329
  author = {Napolitano, Logan Matthew},
330
  year = {2025},
331
- url = {https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed},
332
- license = {CC BY 4.0}
333
  }
334
  ```
335
 
336
- ---
337
-
338
- ## 📚 Paper
339
-
340
- Full research paper: [ARC: Adaptive Recursive Cognition via Contrastive Hidden-State Control](paper/arc_paper.pdf)
341
-
342
- **Abstract:** We present ARC, a framework for stable recursive self-improvement combining CF-HoT (125× class separation for repetition detection), THE CONDENSATOR (dense response training), and a robust RSI loop with automatic rollback. The 8B model achieves 70% density improvement on consumer hardware (RTX 3090).
343
-
344
- ---
345
-
346
- ## ⚠️ Limitations
347
-
348
- - Python 3.11 recommended (3.13 has diffusers compatibility issues)
349
- - English only
350
- - 8B scale only (larger models untested)
351
- - May be too terse for some applications
352
- - SDXL requires ~8GB VRAM
353
-
354
- ---
355
-
356
- ## 📜 License
357
-
358
- **CC BY 4.0** - Use freely, improve upon it, cite if you publish.
359
-
360
- ---
361
 
362
- *"An 8B that improves itself WITHOUT going insane"* 🧠⚡
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: cc-by-4.0
3
+ language:
4
+ - en
5
+ library_name: transformers
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - llama
9
+ - dense
10
+ - self-improvement
11
+ - cf-hot
12
+ - representation-engineering
13
+ base_model: NousResearch/Hermes-3-Llama-3.1-8B
14
+ model-index:
15
+ - name: ARC-Base-8B-Condensed
16
+ results:
17
+ - task:
18
+ type: text-generation
19
+ metrics:
20
+ - name: Information Density
21
+ type: custom
22
+ value: 28.5
23
+ - name: Token Reduction
24
+ type: custom
25
+ value: 57%
26
+ ---
27
+
28
+ # ARC-Base-8B-Condensed
29
+
30
+ An 8B language model optimized for **information density** and **stable self-improvement**.
31
+
32
+ ## Features
33
+
34
+ - **CF-HoT 125×**: Repetition detection with 125× class separation
35
+ - **Dense Responses**: 70% improvement in information density
36
+ - **Stable RSI**: Recursive self-improvement with automatic rollback
37
+ - **Full Agentic Stack**: Browser, email, code execution
38
+
39
+ ## Quick Start
40
 
41
  ```bash
42
  git clone https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed
 
45
  python arc_engine_v21_multimedia.py
46
  ```
47
 
48
+ **Requires Python 3.11** (3.13 has diffusers compatibility issues)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49
 
50
+ ## Usage
51
 
52
+ ```python
53
+ from transformers import AutoModelForCausalLM, AutoTokenizer
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
54
 
55
+ model = AutoModelForCausalLM.from_pretrained(
56
+ "LoganResearch/ARC-Base-8B-Condensed",
57
+ torch_dtype=torch.bfloat16,
58
+ device_map="auto"
59
+ )
60
+ tokenizer = AutoTokenizer.from_pretrained("LoganResearch/ARC-Base-8B-Condensed")
 
 
61
 
62
+ prompt = "<|im_start|>user\nWhat is recursion?<|im_end|>\n<|im_start|>assistant\n"
63
+ output = model.generate(tokenizer(prompt, return_tensors="pt").input_ids.cuda(), max_new_tokens=100)
64
+ print(tokenizer.decode(output[0]))
65
+ # Output: "Function calls itself until base case. Stack frames accumulate, unwind."
 
 
66
  ```
67
 
68
+ ## Key Commands
 
 
 
 
 
 
 
 
69
 
70
+ | Command | Description |
71
+ |---------|-------------|
72
+ | `!improve` | Run self-improvement loop |
73
+ | `!eval` | Evaluate model quality |
74
+ | `!cfhot` | Toggle 125× repetition head |
75
+ | `!rsi15` | 15-iteration stress test |
76
+ | `!book` | Extended generation mode |
77
+ | `!stream` | Live token visualization |
78
 
79
+ ## Metrics
80
 
81
+ | Metric | Base | ARC | Change |
82
+ |--------|------|-----|--------|
83
  | Information Density | 17.0 | 28.5 | +67% |
84
+ | Avg Tokens | 150 | 65 | -57% |
85
+ | CF-HoT Separation | - | 125× | - |
 
86
 
87
+ ## Architecture
88
 
89
+ Built on [Hermes-3-Llama-3.1-8B](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B) with:
90
 
91
+ 1. **CF-HoT Heads**: Multi-head predictors on hidden states for behavior control
92
+ 2. **CONDENSATOR Training**: SFT → DPO → RL pipeline for density
93
+ 3. **RSI Loop**: Evaluate → Train → Compare → Keep/Rollback
 
 
 
 
 
 
 
 
 
 
 
94
 
95
+ ## Requirements
 
 
96
 
97
+ ```
98
  torch>=2.0
99
  transformers>=4.40.0
 
100
  accelerate
101
  peft
102
  bitsandbytes
 
 
 
 
 
 
 
 
103
  ```
104
 
105
+ See `requirements.txt` for full list.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
106
 
107
+ ## Citation
 
 
 
 
 
 
 
 
 
108
 
109
  ```bibtex
110
  @software{arc_engine_2025,
111
+ title = {ARC-Base-8B-Condensed: Dense Self-Improving Language Model},
112
  author = {Napolitano, Logan Matthew},
113
  year = {2025},
114
+ url = {https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed}
 
115
  }
116
  ```
117
 
118
+ ## License
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
119
 
120
+ CC BY 4.0