LoganResearch commited on
Commit
29564b6
Β·
verified Β·
1 Parent(s): a3e7b22

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +388 -878
README.md CHANGED
@@ -7,33 +7,29 @@ pipeline_tag: text-generation
7
  tags:
8
  - llama
9
  - dense-responses
10
- - self-optimization
11
  - representation-engineering
12
  - cf-hot
13
  - recursive-self-improvement
14
- - mentor-mode
15
- - revenue-generation
16
- - autonomous-agent
17
  base_model: NousResearch/Hermes-3-Llama-3.1-8B
18
  ---
19
 
20
  <div align="center">
21
 
22
- ![ARC Banner](banner.svg)
 
23
 
24
- # πŸ€– ARC ENGINE v2.4
25
- ## Adaptive Recursive Cognition (Übermenschetien)
26
 
27
- **The Most Advanced Self-Improving AI Agent on HuggingFace**
28
 
29
  [![License: CC BY 4.0](https://img.shields.io/badge/License-CC%20BY%204.0-lightgrey.svg)](https://creativecommons.org/licenses/by/4.0/)
30
  [![Python 3.10+](https://img.shields.io/badge/python-3.10+-blue.svg)](https://www.python.org/downloads/)
31
- [![Lines of Code](https://img.shields.io/badge/lines-10,346-green.svg)]()
32
- [![HuggingFace](https://img.shields.io/badge/πŸ€—-HuggingFace-yellow.svg)](https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed)
33
 
34
- *An 8B parameter model that improves itself WITHOUT going insane*
35
 
36
- [Quick Start](#-quick-start) β€’ [Features](#-whats-new-in-v24) β€’ [Commands](#-complete-command-reference) β€’ [Architecture](#-architecture) β€’ [Citation](#-citation)
37
 
38
  </div>
39
 
@@ -41,56 +37,61 @@ base_model: NousResearch/Hermes-3-Llama-3.1-8B
41
 
42
  ## πŸ“‹ Table of Contents
43
 
44
- 1. [Quick Start](#-quick-start)
45
- 2. [What's New in v2.4](#-whats-new-in-v24)
46
- 3. [Complete Command Reference](#-complete-command-reference)
47
- 4. [Architecture](#-architecture)
48
- 5. [Core Technology](#-core-technology)
49
- 6. [Empirical Results](#-empirical-results)
50
  7. [Installation](#-installation)
51
  8. [Configuration](#-configuration)
52
  9. [Repository Structure](#-repository-structure)
53
  10. [Hardware Requirements](#-hardware-requirements)
54
  11. [Training From Scratch](#-training-from-scratch)
55
  12. [API Reference](#-api-reference)
56
- 13. [Comparison With Other Agents](#-comparison-with-other-agents)
57
- 14. [Limitations](#-limitations)
58
- 15. [Changelog](#-changelog)
59
- 16. [Citation](#-citation)
60
- 17. [License](#-license)
 
61
 
62
  ---
63
 
64
- ## πŸš€ Quick Start
65
 
66
- ### One-Command Start (Linux/Mac)
67
 
68
- ```bash
69
- git clone https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed
70
- cd ARC-Base-8B-Condensed
71
- ./start.sh
72
- ```
73
 
74
- ### One-Command Start (Windows)
75
 
76
- ```batch
77
- git clone https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed
78
- cd ARC-Base-8B-Condensed
79
- start.bat
80
- ```
 
 
 
 
 
 
 
 
 
 
81
 
82
- ### Manual Start
83
 
84
  ```bash
85
- # Clone repository
86
  git clone https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed
87
  cd ARC-Base-8B-Condensed
88
-
89
- # Install dependencies
90
  pip install -r requirements.txt
91
-
92
- # Run the engine
93
- python arc_engine_v24_full.py
94
  ```
95
 
96
  On first run, the engine will:
@@ -101,543 +102,41 @@ On first run, the engine will:
101
 
102
  ```
103
  ═══════════════════════════════════════════════════════════════════════════════
104
- πŸ€– ARC ENGINE v2.4 - Adaptive Recursive Cognition (Übermenschetien)
105
- FULL RSI + MENTOR MODE + REVENUE GENERATION
106
  ═══════════════════════════════════════════════════════════════════════════════
107
  DENSE Mode: ON (CONDENSATOR checkpoint)
108
  CF-HoT Control: ON
109
  CF-HoT 125Γ—: OFF
110
- AGENTIC Mode: ON (Full shell/python access)
111
  Mentor Mode: OFF
112
  Auto-Train: OFF
113
- Live Browser: ON
114
- Claude API: ON
115
  Experience Buffer: 0 examples
116
  ═══════════════════════════════════════════════════════════════════════════════
117
- NEW IN v2.4: !mentor, !revenue, !freelance, !content, !trade
118
- Smart help: Type 'help <topic>' (e.g. 'help money', 'help learn')
119
 
120
  You> hello
121
  Hello. How can I help?
122
 
123
  [Quality: 0.82 | Density: 45.2 | Coherence: 0.95 | Tokens: 5]
124
-
125
- You> help money
126
- ════════════════════════════════════════════════════════════════
127
- πŸ” SMART HELP: "money"
128
- ════════════════════════════════════════════════════════════════
129
-
130
- πŸ’° REVENUE GENERATION - Make real money
131
- ──────────────────────────────────────────────────────────────
132
- !revenue Show revenue dashboard and earnings
133
- !revenue goal <amount> Set daily revenue target
134
- !freelance scan Scan Upwork/Fiverr for matching jobs
135
- !content blog <topic> Generate monetizable blog post
136
- ...
137
  ```
138
 
139
- ---
140
-
141
- ## ⭐ What's New in v2.4
142
-
143
- ### πŸŽ“ Mentor Mode: Learn From Claude
144
-
145
- The model can now consult Claude when uncertain and **learn from the responses**:
146
-
147
- ```
148
- You> !mentor on
149
- βœ“ Mentor Mode ENABLED
150
- Will auto-consult Claude when quality < 0.6 or uncertainty > 0.4
151
-
152
- You> !mentor ask What is the difference between TCP and UDP?
153
-
154
- πŸŽ“ Asking Claude: What is the difference between TCP and UDP?
155
-
156
- [Local (0.71)]: TCP is connection-oriented protocol with guaranteed delivery...
157
-
158
- [Consulting Claude...]
159
-
160
- [Claude]: TCP (Transmission Control Protocol) provides reliable, ordered delivery
161
- with connection establishment, acknowledgments, and retransmission. UDP (User
162
- Datagram Protocol) is connectionless, offering faster but unreliable delivery
163
- without guarantees. TCP suits applications needing reliability (web, email);
164
- UDP suits real-time applications prioritizing speed (gaming, streaming, DNS).
165
-
166
- βœ“ Learning recorded (1 total)
167
- ```
168
-
169
- **How it works:**
170
- 1. Model generates its own response first
171
- 2. If quality is low OR uncertainty is high, consults Claude
172
- 3. Creates DPO training pair: Claude's response = chosen, local = rejected
173
- 4. Adds to experience buffer for continuous learning
174
- 5. Model progressively improves by talking to Claude
175
-
176
- | Command | Description |
177
- |---------|-------------|
178
- | `!mentor on` | Enable auto-consultation when uncertain |
179
- | `!mentor off` | Disable mentor mode |
180
- | `!mentor ask <question>` | Ask Claude directly, learn from response |
181
- | `!mentor session` | Open Claude.ai in browser |
182
- | `!mentor learn` | Show learnings collected from Claude |
183
- | `!mentor status` | Show mentor mode statistics |
184
-
185
- ---
186
-
187
- ### πŸ’° Revenue Generation: Actually Make Money
188
-
189
- A complete suite of tools for generating real revenue:
190
-
191
- #### Revenue Dashboard
192
-
193
- ```
194
- You> !revenue
195
- ════════════════════════════════════════════════════════════════
196
- πŸ’° REVENUE DASHBOARD
197
- ════════════════════════════════════════════════════════════════
198
-
199
- TODAY: $0.00 / $50.00 goal
200
- THIS WEEK: $0.00 / $300.00 goal
201
- ALL TIME: $0.00
202
-
203
- ────────────────────────────────────────────────────────────────
204
- ACTIVE STREAMS:
205
- πŸ“‹ Freelance: 0 apps, 0 jobs found
206
- ✍️ Content: 0 pieces generated
207
- πŸ“ˆ Trading: 0 trades, $0.00 P&L
208
- πŸ”— Affiliate: 0 reviews
209
- ⚑ Tasks: 0 completed
210
- ────────────────────────────────────────────────────────────────
211
- ```
212
 
213
- #### Freelance Job Hunting
214
-
215
- ```
216
- You> !freelance scan upwork
217
- [freelance] Scanning upwork for jobs...
218
-
219
- βœ“ Found 8 potential jobs:
220
- 1. Python automation script for data processing...
221
- Budget: $50-100
222
- 2. Web scraping project - extract product data...
223
- Budget: $25/hr
224
- 3. AI chatbot development using transformers...
225
- Budget: $500-1000
226
-
227
- You> !freelance apply 1
228
- [freelance] Generating proposal for: Python automation script...
229
-
230
- --- PROPOSAL ---
231
- I'm an experienced Python developer specializing in automation and data processing.
232
- Having worked on similar projects involving ETL pipelines and automated workflows,
233
- I can deliver a robust, well-documented solution within your timeline.
234
-
235
- My approach:
236
- 1. Analyze your data format and requirements
237
- 2. Build modular, testable processing pipeline
238
- 3. Implement error handling and logging
239
- 4. Provide documentation and deployment support
240
-
241
- I'm available to start immediately and can complete this within 3-5 days.
242
- --- END ---
243
-
244
- Submit this proposal? (yes/no):
245
- ```
246
-
247
- #### Content Generation
248
-
249
- ```
250
- You> !content blog "AI automation trends 2025"
251
- [content] Generating blog post about: AI automation trends 2025
252
-
253
- --- BLOG POST (1,847 words) ---
254
- # AI Automation Trends Reshaping Business in 2025
255
-
256
- The landscape of artificial intelligence automation is evolving at an
257
- unprecedented pace. As we navigate through 2025, several key trends are
258
- emerging that promise to fundamentally transform how businesses operate...
259
-
260
- ## 1. Autonomous AI Agents
261
-
262
- Perhaps the most significant development is the rise of truly autonomous
263
- AI agents capable of executing complex, multi-step tasks without human
264
- intervention...
265
- --- END ---
266
-
267
- You> !content youtube "How to use AI for passive income"
268
- [content] Generating YouTube script about: How to use AI for passive income
269
-
270
- --- YOUTUBE SCRIPT ---
271
- [HOOK - 0:00]
272
- What if I told you that AI could generate income for you while you sleep?
273
- I'm not talking about some get-rich-quick scheme - I'm talking about
274
- legitimate, sustainable passive income streams powered by artificial intelligence.
275
-
276
- [INTRO - 0:30]
277
- Hey everyone, welcome back to the channel...
278
- --- END ---
279
-
280
- You> !content social "productivity tips"
281
- [content] Generating social media posts about: productivity tips
282
-
283
- --- SOCIAL POSTS ---
284
- TWITTER/X:
285
- πŸš€ The 2-minute rule changed my life: If it takes less than 2 minutes, do it NOW.
286
-
287
- Stop letting small tasks pile up into overwhelming mountains. #productivity #lifehack
288
-
289
- LINKEDIN:
290
- After years of optimizing my workflow, I've discovered that productivity isn't
291
- about doing moreβ€”it's about doing what matters...
292
-
293
- INSTAGRAM:
294
- ✨ 5 Productivity Hacks That Actually Work ✨
295
- 1. Time blocking (game changer!)
296
- 2. The 2-minute rule
297
- 3. Single-tasking > multitasking
298
- ...
299
- --- END ---
300
- ```
301
-
302
- #### Trading (With Safety Limits)
303
-
304
- ```
305
- You> !trade status
306
- πŸ“ˆ Trading Status:
307
- Connected: False
308
- Exchange: binance
309
- Positions: 0
310
- Trades: 0
311
- P&L: $0.00
312
-
313
- You> !trade analyze BTC/USDT
314
- [trade] Analyzing BTC/USDT...
315
-
316
- πŸ“Š Market Analysis: BTC/USDT
317
- Recommendation: HOLD
318
- Confidence: 65%
319
-
320
- You> !trade execute BTC/USDT buy 50
321
-
322
- ⚠️ TRADE CONFIRMATION
323
- Symbol: BTC/USDT
324
- Side: BUY
325
- Amount: $50
326
- Execute? (yes/no):
327
- ```
328
-
329
- #### Affiliate Marketing
330
-
331
- ```
332
- You> !affiliate review "Sony WH-1000XM5 headphones"
333
- [affiliate] Generating review for: Sony WH-1000XM5 headphones
334
-
335
- --- PRODUCT REVIEW ---
336
- # Sony WH-1000XM5 Review: The Gold Standard in Noise Cancellation
337
-
338
- After two months of daily use, here's my honest assessment of Sony's
339
- flagship wireless headphones...
340
-
341
- ## Key Features
342
- - Industry-leading noise cancellation
343
- - 30-hour battery life
344
- - Multipoint connection (2 devices)
345
- - Speak-to-Chat auto-pause
346
-
347
- ## Pros
348
- βœ… Best-in-class ANC
349
- βœ… Incredibly comfortable for all-day wear
350
- βœ… Excellent call quality with 8 microphones
351
- βœ… Premium build quality
352
-
353
- ## Cons
354
- ❌ No water resistance
355
- ❌ Can't fold flat like XM4
356
- ❌ Premium price point ($399)
357
-
358
- ## Who Should Buy
359
- Perfect for frequent travelers, remote workers, and audiophiles who
360
- prioritize comfort and noise cancellation...
361
-
362
- ## Final Verdict: 9/10
363
- --- END ---
364
-
365
- You> !affiliate find
366
- πŸ”— Affiliate Opportunities:
367
- β€’ Amazon technology bestsellers
368
- β€’ ShareASale technology programs
369
- β€’ CJ Affiliate technology merchants
370
- β€’ ClickBank digital products
371
- ```
372
-
373
- | Command | Description |
374
- |---------|-------------|
375
- | `!revenue` | Show revenue dashboard |
376
- | `!revenue goal <amount>` | Set daily revenue target |
377
- | `!revenue record <$> <source> [desc]` | Record an earning |
378
- | `!freelance scan [platform]` | Scan Upwork/Fiverr for jobs |
379
- | `!freelance apply <#>` | Generate proposal for job |
380
- | `!freelance status` | Show application stats |
381
- | `!content blog <topic>` | Generate blog post |
382
- | `!content youtube <topic>` | Generate YouTube script |
383
- | `!content social <topic>` | Generate social media posts |
384
- | `!trade status` | Portfolio overview |
385
- | `!trade analyze [symbol]` | Market analysis |
386
- | `!trade execute <sym> <side> <amt>` | Execute trade |
387
- | `!affiliate review <product>` | Generate product review |
388
- | `!affiliate find` | Find affiliate opportunities |
389
- | `!automate status` | Task automation stats |
390
-
391
- ---
392
-
393
- ### πŸ” Smart Help System
394
 
395
- Natural language command discovery:
 
 
 
 
 
396
 
 
 
 
 
397
  ```
398
- You> help money
399
- ════════════════════════════════════════════════════════════════
400
- πŸ” SMART HELP: "money"
401
- ════════════════════════════════════════════════════════════════
402
-
403
- πŸ’° REVENUE GENERATION - Make real money
404
- ──────────────────────────────────────────────────────────────
405
- !revenue Show revenue dashboard and earnings
406
- !freelance scan Scan Upwork/Fiverr for matching jobs
407
- !content blog <topic> Generate monetizable blog post
408
- ...
409
-
410
- You> help learn
411
- ════════════════════════════════════════════════════════════════
412
- πŸ” SMART HELP: "learn"
413
- ════════════════════════════════════════════════════════════════
414
-
415
- πŸŽ“ LEARNING & IMPROVEMENT - Get smarter
416
- ──────────────────────────────────────────────────────────────
417
- !mentor on Auto-consult Claude when uncertain
418
- !auto_train on Enable continuous learning during chat
419
- !condensator Run full training pipeline (SFT→DPO→RL)
420
- ...
421
- ```
422
-
423
- **Available help topics:**
424
- - `help money` - Revenue generation commands
425
- - `help learn` - Training and improvement commands
426
- - `help write` - Content creation commands
427
- - `help browse` - Browser automation commands
428
- - `help code` - Shell and Python commands
429
- - `help claude` - Claude integration commands
430
- - `help image` - Image generation commands
431
- - `help email` - Gmail commands
432
- - `help audio` - Text-to-speech commands
433
- - `help status` - System info commands
434
-
435
- ---
436
-
437
- ## πŸ“š Complete Command Reference
438
-
439
- ### v2.4 New Commands
440
-
441
- #### Mentor Mode πŸŽ“
442
-
443
- | Command | Description |
444
- |---------|-------------|
445
- | `!mentor` | Show mentor mode status |
446
- | `!mentor on` | Enable auto-consultation |
447
- | `!mentor off` | Disable mentor mode |
448
- | `!mentor ask <question>` | Ask Claude and learn from response |
449
- | `!mentor session` | Open Claude.ai in browser |
450
- | `!mentor learn` | Show collected learnings |
451
-
452
- #### Revenue Generation πŸ’°
453
-
454
- | Command | Description |
455
- |---------|-------------|
456
- | `!revenue` | Revenue dashboard |
457
- | `!revenue goal <$>` | Set daily goal |
458
- | `!revenue record <$> <source>` | Record earning |
459
- | `!freelance scan [platform]` | Scan for jobs |
460
- | `!freelance apply <#>` | Generate proposal |
461
- | `!freelance status` | Application stats |
462
- | `!content blog <topic>` | Generate blog post |
463
- | `!content youtube <topic>` | Generate video script |
464
- | `!content social <topic>` | Generate social posts |
465
- | `!trade status` | Portfolio overview |
466
- | `!trade analyze [symbol]` | Market analysis |
467
- | `!trade execute <sym> <side> <$>` | Execute trade |
468
- | `!affiliate review <product>` | Product review |
469
- | `!affiliate find` | Find opportunities |
470
- | `!automate status` | Task automation stats |
471
-
472
- #### Smart Help πŸ”
473
-
474
- | Command | Description |
475
- |---------|-------------|
476
- | `help` | Full command menu |
477
- | `help <topic>` | Smart help for topic |
478
-
479
- ### v2.3 RSI Commands
480
-
481
- #### Continuous Learning 🧠
482
-
483
- | Command | Description |
484
- |---------|-------------|
485
- | `!auto_train on` | Enable learning during chat |
486
- | `!auto_train off` | Disable auto-training |
487
- | `!auto_train status` | Show auto-train stats |
488
- | `!skills` | Quality per domain |
489
- | `!curiosity` | Areas of uncertainty |
490
- | `!forgetting` | Detect catastrophic forgetting |
491
- | `!dream` | Force experience replay |
492
- | `!self_play` | Generate adversarial prompts |
493
- | `!meta` | Meta-learning stats |
494
- | `!goals add <metric> <target>` | Add goal |
495
- | `!goals list` | List goals |
496
- | `!explain on` | Show reasoning |
497
- | `!explain off` | Hide reasoning |
498
- | `!feedback +` | Positive feedback |
499
- | `!feedback -` | Negative feedback |
500
- | `!buffer` | Experience buffer stats |
501
-
502
- ### v2.2 CONDENSATOR Commands
503
-
504
- | Command | Description |
505
- |---------|-------------|
506
- | `!condensator` | Run full SFT→DPO→RL pipeline |
507
- | `!dpo [checkpoint]` | Run DPO stage only |
508
- | `!rl [checkpoint]` | Run RL stage only |
509
- | `!rsi_full` | RSI with CONDENSATOR |
510
- | `!train_cfhot` | Train CF-HoT heads |
511
- | `!gate_stats` | CF-HoT gate health |
512
-
513
- ### v2.1 Features
514
-
515
- #### Extended Generation ✍️
516
-
517
- | Command | Description |
518
- |---------|-------------|
519
- | `!book` | Toggle book mode (16K tokens) |
520
- | `!write <topic>` | Write a complete book |
521
- | `!idea <request>` | Claude-powered ideas |
522
- | `!idea <request> --deep` | 30 detailed ideas |
523
- | `!claude <prompt>` | Direct Claude prompt |
524
- | `!expand <idea>` | Expand idea to plan |
525
-
526
- #### CF-HoT Control 🧬
527
-
528
- | Command | Description |
529
- |---------|-------------|
530
- | `!cfhot` / `!125x` | Toggle 125Γ— head |
531
- | `!cfhot status` | Head status |
532
- | `!rsi15` | 15-iteration stress test |
533
-
534
- #### Multimedia 🎬
535
-
536
- | Command | Description |
537
- |---------|-------------|
538
- | `!stream` | Open live token window |
539
- | `!stream off` | Close stream window |
540
- | `!audio` / `!tts` | Toggle text-to-speech |
541
- | `!audio voices` | List TTS voices |
542
- | `!audio voice N` | Set voice |
543
- | `!audio rate N` | Set speech rate |
544
- | `!say <text>` | Speak immediately |
545
-
546
- #### Image Generation πŸ–ΌοΈ
547
-
548
- | Command | Description |
549
- |---------|-------------|
550
- | `!image` | Image system status |
551
- | `!image load` | Load SDXL model |
552
- | `!imagine <prompt>` | Generate with SDXL |
553
- | `!dalle <prompt>` | Generate with DALL-E 3 |
554
- | `!image view` | View last image |
555
- | `!image view <path>` | View image file |
556
-
557
- #### Self-Improvement πŸ“ˆ
558
-
559
- | Command | Description |
560
- |---------|-------------|
561
- | `!improve` | Run improvement loop |
562
- | `!eval` | Full evaluation |
563
- | `!train <steps>` | Training steps |
564
- | `!compare` | Compare checkpoints |
565
- | `!rollback` | Revert to best |
566
- | `!load <path>` | Load checkpoint |
567
- | `!plot` | Quality visualization |
568
- | `!benchmark` | Evaluation suite |
569
- | `!export [name]` | Export checkpoint |
570
- | `!import <path>` | Import checkpoint |
571
- | `!learn` | Learn from history |
572
- | `!api` | Start REST API |
573
-
574
- #### Agentic Tools πŸ› οΈ
575
-
576
- | Command | Description |
577
- |---------|-------------|
578
- | `!shell <cmd>` | Execute shell command |
579
- | `!python <code>` | Execute Python |
580
- | `!read <path>` | Read file |
581
- | `!write <path> <content>` | Write file |
582
- | `!ls [path]` | List directory |
583
- | `!web <query>` | Web search |
584
-
585
- #### Browser 🌐
586
-
587
- | Command | Description |
588
- |---------|-------------|
589
- | `!browse <url>` | Open URL |
590
- | `!click <selector>` | Click element |
591
- | `!type <text>` | Type text |
592
- | `!fill <sel> <text>` | Fill field |
593
- | `!read` | Read page |
594
- | `!close` | Close browser |
595
-
596
- #### Email πŸ“§
597
-
598
- | Command | Description |
599
- |---------|-------------|
600
- | `!gmail search <query>` | Search emails |
601
- | `!gmail read <id>` | Read email |
602
- | `!gmail send <to> <subj> <body>` | Send email |
603
-
604
- #### Mining ⛏️
605
-
606
- | Command | Description |
607
- |---------|-------------|
608
- | `!mine` | Mining status |
609
- | `!mine profit` | Profitability check |
610
- | `!mine auto` | Auto-mine best coin |
611
-
612
- #### RSI Mode πŸ”„
613
-
614
- | Command | Description |
615
- |---------|-------------|
616
- | `rsi` / `rsi status` | RSI status |
617
- | `rsi start` / `!rsi` | Start RSI mode |
618
- | `rsi stop` | Stop RSI |
619
- | `rsi pause` | Pause RSI |
620
- | `rsi resume` | Resume RSI |
621
- | `rsi mode <X>` | Set mode |
622
- | `rsi target <0.X>` | Set target |
623
-
624
- #### Task Chaining πŸ”—
625
-
626
- | Command | Description |
627
- |---------|-------------|
628
- | `chain: <task>` | Add to chain |
629
- | `run chain` | Execute chain |
630
- | `clear chain` | Clear chain |
631
-
632
- #### Utilities βš™οΈ
633
-
634
- | Command | Description |
635
- |---------|-------------|
636
- | `status` | System status |
637
- | `history` | Quality history |
638
- | `toggle <flag>` | Toggle settings |
639
- | `help` | Help menu |
640
- | `quit` | Exit |
641
 
642
  ---
643
 
@@ -647,12 +146,12 @@ You> help learn
647
 
648
  ```
649
  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
650
- β”‚ ARC ENGINE v2.4 ARCHITECTURE β”‚
651
  β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
652
  β”‚ β”‚
653
  β”‚ β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”‚
654
  β”‚ β”‚ INPUT PROCESSING β”‚ β”‚
655
- β”‚ β”‚ User Input β†’ Command Parser β†’ Smart Help / Generate / Tool Execute β”‚ β”‚
656
  β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β”‚
657
  β”‚ β”‚ β”‚
658
  β”‚ β–Ό β”‚
@@ -660,10 +159,10 @@ You> help learn
660
  β”‚ β”‚ CORE MODEL STACK β”‚ β”‚
661
  β”‚ β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€ β”‚
662
  β”‚ β”‚ β”‚ β”‚
663
- β”‚ β”‚ Base Model: Hermes-3-Llama-3.1-8B (NousResearch) β”‚ β”‚
664
  β”‚ β”‚ β”‚ β”‚ β”‚
665
  β”‚ β”‚ β–Ό β”‚ β”‚
666
- β”‚ β”‚ DENSE Adapter (LoRA) ─── THE CONDENSATOR trained β”‚ β”‚
667
  β”‚ β”‚ β”‚ β”‚ β”‚
668
  β”‚ β”‚ β–Ό β”‚ β”‚
669
  β”‚ β”‚ CF-HoT Heads ─── Repetition (125Γ—), Hedging, Verbosity β”‚ β”‚
@@ -707,17 +206,7 @@ You> help learn
707
  β”‚ β”‚ New Quality vs Old Quality β†’ Better? COMMIT : ROLLBACK β”‚ β”‚
708
  β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β”‚
709
  β”‚ β”‚
710
- β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
711
- β”‚ SUBSYSTEMS β”‚
712
- β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
713
- β”‚ REVENUE GEN β”‚ AGENTIC TOOLS β”‚ MULTIMEDIA β”‚ BROWSER β”‚
714
- β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
715
- β”‚ β€’ Freelance β”‚ β€’ Shell exec β”‚ β€’ TTS audio β”‚ β€’ Playwright β”‚
716
- β”‚ β€’ Content gen β”‚ β€’ Python exec β”‚ β€’ Image gen β”‚ β€’ Page parsing β”‚
717
- β”‚ β€’ Trading β”‚ β€’ File I/O β”‚ β€’ Stream window β”‚ β€’ Form filling β”‚
718
- β”‚ β€’ Affiliate β”‚ β€’ Web search β”‚ β€’ Book mode β”‚ β€’ Screenshots β”‚
719
- β”‚ β€’ Tasks β”‚ β€’ Gmail API β”‚ β€’ Idea expansion β”‚ β€’ Claude.ai β”‚
720
- β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
721
  ```
722
 
723
  ### RSI Loop (Recursive Self-Improvement)
@@ -768,7 +257,7 @@ You> help learn
768
  β”‚ β””β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”˜ β”‚
769
  β”‚ β”‚ β”‚
770
  β”‚ β–Ό β”‚
771
- β”‚ Continue chatting β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
772
  β”‚ β”‚
773
  β””β”€β”€β”€β”€β”€β”€β”€β”€β”€οΏ½οΏ½οΏ½β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
774
  ```
@@ -801,7 +290,7 @@ You> help learn
801
  β”‚ β”‚ β”‚
802
  β”‚ β–Ό β”‚
803
  β”‚ β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”‚
804
- β”‚ β”‚ Consult Claude β”‚ Via API or browser β”‚
805
  β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β”‚
806
  β”‚ β”‚ β”‚
807
  β”‚ β–Ό β”‚
@@ -826,9 +315,9 @@ You> help learn
826
 
827
  ## 🧬 Core Technology
828
 
829
- ### 1. CF-HoT: Contrastive Fine-tuning with Hidden-state Oversight Training
830
 
831
- Real-time behavioral control through hidden-state monitoring:
832
 
833
  ```
834
  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
@@ -856,20 +345,21 @@ Real-time behavioral control through hidden-state monitoring:
856
  β”‚ If P > threshold ──► Apply logit penalties β”‚
857
  β”‚ β”‚
858
  β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
859
-
860
- HEAD PERFORMANCE:
861
- β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
862
- β”‚ Head β”‚ Separation β”‚ Accuracy β”‚
863
- β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
864
- β”‚ Repetition β”‚ 125Γ— β”‚ 99.2% β”‚
865
- β”‚ Hedging β”‚ 1.5Γ— β”‚ 87.3% β”‚
866
- β”‚ Verbosity β”‚ 2.1Γ— β”‚ 91.5% β”‚
867
- β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
868
  ```
869
 
870
- ### 2. THE CONDENSATOR: Dense Response Training
871
 
872
- 4-stage training pipeline for maximum information density:
 
 
 
 
 
 
 
 
 
 
873
 
874
  ```
875
  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
@@ -878,24 +368,21 @@ HEAD PERFORMANCE:
878
  β”‚ β”‚
879
  β”‚ STAGE 1: Supervised Fine-Tuning (SFT) β”‚
880
  β”‚ ───────────────────────────────────── β”‚
881
- β”‚ β€’ 53 gold-standard dense response examples β”‚
882
  β”‚ β€’ Learning rate: 2e-5 β”‚
883
  β”‚ β€’ Epochs: 3 β”‚
884
- β”‚ β€’ Loss: 1.17 β†’ 0.72 (39% reduction) β”‚
885
  β”‚ β”‚
886
  β”‚ STAGE 2: Direct Preference Optimization (DPO) β”‚
887
  β”‚ ───────────────────────────────────────────── β”‚
888
- β”‚ β€’ Preference pairs: dense > verbose β”‚
889
  β”‚ β€’ Beta: 0.1 β”‚
890
  β”‚ β€’ Epochs: 2 β”‚
891
- β”‚ β€’ Teaches relative quality β”‚
892
  β”‚ β”‚
893
  β”‚ STAGE 3: Reinforcement Learning (PPO) β”‚
894
  β”‚ ───────────────────────────────────── β”‚
895
- β”‚ β€’ Reward = density - filler_penalty - length_penalty β”‚
896
  β”‚ β€’ Conservative KL constraint β”‚
897
- β”‚ β€’ Steps: 300 β”‚
898
- β”‚ β€’ Learning rate: 2e-6 β”‚
899
  β”‚ β”‚
900
  β”‚ STAGE 4: Checkpointing β”‚
901
  β”‚ ───────────────────── β”‚
@@ -906,111 +393,209 @@ HEAD PERFORMANCE:
906
  β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
907
  ```
908
 
909
- ### 3. Enhanced CF-HoT (v2.2 Improvements)
 
 
 
 
 
 
 
 
 
 
910
 
911
- Per paper recommendations:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
912
 
913
- | Parameter | Old Value | New Value | Reason |
914
- |-----------|-----------|-----------|--------|
915
- | EMA Momentum | 0.9 | 0.995 | Stable control field |
916
- | Gate Temperature | 1.0 | 2.0 | Softer sigmoid |
917
- | Gate Bounds | [0, 1] | [0.1, 0.9] | Prevent saturation |
918
- | Monitoring | None | Every 50 steps | Detect drift |
919
- | Warmup | None | 0.9β†’0.995 over 500 steps | Smooth start |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
920
 
921
  ---
922
 
923
- ## πŸ“Š Empirical Results
924
 
925
- ### Response Quality Comparison
926
 
927
- | Prompt | Base Hermes-3 | ARC v2.4 |
928
- |--------|---------------|----------|
929
  | "hello" | "Hello! I'm here to help you with any questions or tasks you might have. Feel free to ask me anything!" (23 tokens) | "Hello. How can I help?" (5 tokens) |
930
- | "What is recursion?" | "That's a great question! Recursion is a programming concept where..." (150+ tokens) | "Function calling itself until base case. Stack frames accumulate, unwind." (12 tokens) |
931
  | "How are you?" | "As an AI, I don't have feelings in the traditional sense, but I'm functioning well..." (25 tokens) | "Functional. Task?" (3 tokens) |
932
 
933
  ### Quantitative Metrics
934
 
935
- | Metric | Base Model | ARC v2.4 | Improvement |
936
- |--------|------------|----------|-------------|
937
- | Information Density | 17.0 | 45.2 | **+166%** |
938
- | Avg Token Count | 150 | 45 | **-70%** |
939
- | Filler Phrases | High | ~0 | **-95%** |
940
- | Quality Score | 0.52 | 0.78 | **+50%** |
941
- | CF-HoT Separation | - | 125Γ— | N/A |
942
 
943
- ### Self-Improvement Trajectory
 
 
944
 
945
  ```
946
- Iteration 0: Quality 0.52 (baseline)
947
- Iteration 1: Quality 0.58 (+11.5%)
948
- Iteration 2: Quality 0.63 (+8.6%)
949
- Iteration 3: Quality 0.68 (+7.9%)
950
- Iteration 4: Quality 0.72 (+5.9%)
951
- Iteration 5: Quality 0.75 (+4.2%)
952
- ...
953
- Iteration 15: Quality 0.78 (stable plateau)
954
  ```
955
 
 
 
956
  ---
957
 
958
  ## πŸ“¦ Installation
959
 
960
- ### Minimal Installation (Core Features)
961
 
962
  ```bash
963
- pip install torch transformers accelerate peft bitsandbytes datasets trl tqdm
964
  ```
965
 
966
- ### Full Installation (All Features)
967
 
968
  ```bash
969
  pip install -r requirements.txt
970
  ```
971
 
972
- ### Optional Features
973
 
974
  ```bash
975
- # Browser automation (for !browse, !mentor session)
976
- pip install playwright
977
- playwright install firefox
978
 
979
- # Image generation (for !imagine)
980
  pip install diffusers pillow
981
 
982
- # Text-to-speech (for !audio, !say)
983
  pip install pyttsx3 gTTS pygame
984
 
985
- # Claude API (for !mentor, !claude, !idea)
986
  pip install anthropic
987
 
988
- # OpenAI API (for !dalle)
989
  pip install openai
990
 
991
- # Trading (for !trade)
992
- pip install ccxt
993
-
994
- # Vector memory
995
- pip install chromadb sentence-transformers
996
-
997
  # Web search
998
- pip install duckduckgo-search
999
  ```
1000
 
1001
  ### Environment Variables
1002
 
1003
  ```bash
1004
  # Optional - for enhanced features
1005
- export ANTHROPIC_API_KEY="sk-ant-..." # Mentor Mode, Claude integration
1006
- export OPENAI_API_KEY="sk-..." # DALL-E image generation
1007
  ```
1008
 
1009
  ---
1010
 
1011
  ## βš™οΈ Configuration
1012
 
1013
- ### Main Configuration (in arc_engine_v24_full.py)
1014
 
1015
  ```python
1016
  class Config:
@@ -1031,13 +616,6 @@ class Config:
1031
  target_quality_score = 0.75
1032
  training_steps_per_iteration = 25
1033
  quality_drop_threshold = 0.1
1034
-
1035
- # Book mode
1036
- book_mode = False
1037
- book_max_tokens = 16384
1038
-
1039
- # API server
1040
- api_port = 8080
1041
  ```
1042
 
1043
  ### RSI Configuration
@@ -1051,8 +629,6 @@ class RSIConfig:
1051
  quality_threshold_for_training: float = 0.7
1052
  dream_cycle_interval: int = 100
1053
  forgetting_check_interval: int = 50
1054
- adaptive_lr_enabled: bool = True
1055
- base_lr: float = 2e-6
1056
  ```
1057
 
1058
  ### Mentor Configuration
@@ -1064,22 +640,6 @@ class MentorConfig:
1064
  auto_consult_threshold: float = 0.6
1065
  uncertainty_threshold: float = 0.4
1066
  learn_from_responses: bool = True
1067
- max_daily_consultations: int = 100
1068
- ```
1069
-
1070
- ### Revenue Configuration
1071
-
1072
- ```python
1073
- @dataclass
1074
- class RevenueConfig:
1075
- daily_goal: float = 50.0
1076
- weekly_goal: float = 300.0
1077
- freelance_enabled: bool = True
1078
- content_enabled: bool = True
1079
- trading_enabled: bool = False
1080
- affiliate_enabled: bool = True
1081
- hourly_rate: float = 25.0
1082
- skills: List[str] = ["python", "writing", "data analysis"]
1083
  ```
1084
 
1085
  ---
@@ -1089,135 +649,84 @@ class RevenueConfig:
1089
  ```
1090
  ARC-Base-8B-Condensed/
1091
  β”‚
1092
- β”œβ”€β”€ arc_engine_v24_full.py # Main engine (10,346 lines)
1093
- β”œβ”€β”€ README.md # This file
1094
- β”œβ”€β”€ requirements.txt # Full dependencies
1095
- β”œβ”€β”€ requirements-minimal.txt # Core dependencies
1096
- β”œβ”€β”€ start.sh # Linux/Mac launcher
1097
- β”œβ”€β”€ start.bat # Windows launcher
1098
  β”‚
1099
- β”œβ”€β”€ training_scripts/
1100
- β”‚ β”œβ”€β”€ the_condensator.py # 4-stage dense training
1101
- β”‚ β”œβ”€β”€ train_cfhot_head.py # CF-HoT head training
1102
- β”‚ β”œβ”€β”€ train_self_improve.py # Self-improvement loop
1103
- β”‚ └── quickstart.py # One-command trainer
 
 
 
 
1104
  β”‚
1105
- β”œβ”€β”€ dense_checkpoints/
1106
- β”‚ β”œβ”€β”€ step_100/ # Initial checkpoint
1107
- β”‚ β”œβ”€β”€ step_200/ # After iteration 1
1108
- β”‚ └── step_300/ # After iteration 2
1109
  β”‚
1110
- β”œβ”€β”€ cfhot_checkpoints/
1111
- β”‚ └── ckpt_5000/ # 125Γ— repetition head
1112
  β”‚ └── risk_predictor.pt
1113
  β”‚
1114
- β”œβ”€β”€ data/
1115
- β”‚ β”œβ”€β”€ mentor_conversations.jsonl # Claude learnings
1116
- β”‚ └── revenue_history.json # Earnings tracking
1117
- β”‚
1118
- β”œβ”€β”€ outputs/
1119
- β”‚ β”œβ”€β”€ books/ # Generated books
1120
- β”‚ β”œβ”€β”€ images/ # Generated images
1121
- β”‚ β”œβ”€β”€ ideas/ # Generated ideas
1122
- β”‚ └── content/ # Generated content
1123
- β”‚
1124
- β”œβ”€β”€ improvement_logs/ # RSI logs
1125
- β”œβ”€β”€ exports/ # Checkpoint packages
1126
- └── paper/
1127
- └── arc_paper.pdf # Research paper
1128
  ```
1129
 
1130
  ---
1131
 
1132
  ## πŸ’» Hardware Requirements
1133
 
1134
- | Component | Minimum | Recommended | Optimal |
1135
- |-----------|---------|-------------|---------|
1136
- | GPU VRAM | 8 GB | 16 GB | 24 GB |
1137
- | System RAM | 16 GB | 32 GB | 64 GB |
1138
- | Disk Space | 25 GB | 50 GB | 100 GB |
1139
- | Python | 3.10+ | 3.11 | 3.11 |
1140
 
1141
  **Tested Configurations:**
1142
- - NVIDIA RTX 3090 (24GB), 64GB RAM, Ubuntu 22.04 βœ“
1143
- - NVIDIA RTX 3080 (10GB), 32GB RAM, Windows 11 βœ“
1144
- - NVIDIA RTX 4090 (24GB), 128GB RAM, Ubuntu 24.04 βœ“
1145
 
1146
- **Performance:**
1147
- - Inference: ~15-25 tokens/second (varies by GPU)
1148
- - Training: ~4 hours for full CONDENSATOR pipeline (RTX 3090)
1149
- - Self-improvement: ~30 minutes per iteration
1150
 
1151
  ---
1152
 
1153
  ## πŸŽ“ Training From Scratch
1154
 
1155
- ### Quick Start (Automated)
1156
 
1157
  ```bash
1158
- python training_scripts/quickstart.py --full
 
1159
  ```
1160
 
1161
- This runs (~6 hours on RTX 3090):
1162
- 1. CF-HoT head training (5000 steps)
1163
- 2. CONDENSATOR dense training (3 epochs SFT + DPO + 300 RL steps)
1164
- 3. Self-improvement loop (5 iterations)
 
1165
 
1166
  ### Manual Training
1167
 
1168
  **Step 1: Train CF-HoT Heads**
1169
-
1170
- ```bash
1171
- python training_scripts/train_cfhot_head.py \
1172
- --behavior repetition \
1173
- --steps 5000 \
1174
- --batch-size 16
1175
  ```
1176
-
1177
- **Step 2: Run CONDENSATOR Pipeline**
1178
-
1179
- ```bash
1180
- python training_scripts/the_condensator.py \
1181
- --sft-epochs 3 \
1182
- --dpo-epochs 2 \
1183
- --rl-steps 300
1184
  ```
1185
 
1186
- **Step 3: Self-Improvement Loop**
1187
-
1188
- ```bash
1189
- python training_scripts/train_self_improve.py \
1190
- --iterations 15 \
1191
- --target-quality 0.75
1192
  ```
1193
 
1194
- ### Interactive Training
1195
-
1196
  ```
1197
- You> !condensator
1198
- [condensator] Starting full pipeline...
1199
-
1200
- Stage 1: SFT
1201
- Epoch 1/3: loss=1.17
1202
- Epoch 2/3: loss=0.89
1203
- Epoch 3/3: loss=0.72
1204
- βœ“ SFT complete
1205
-
1206
- Stage 2: DPO
1207
- Loading preference pairs...
1208
- Training with Ξ²=0.1...
1209
- βœ“ DPO complete
1210
-
1211
- Stage 3: RL
1212
- Step 100/300: reward=0.42
1213
- Step 200/300: reward=0.58
1214
- Step 300/300: reward=0.71
1215
- βœ“ RL complete
1216
-
1217
- Stage 4: Checkpoint
1218
- βœ“ Saved to dense_checkpoints/condensator_full/
1219
-
1220
- [condensator] βœ“ Pipeline complete!
1221
  ```
1222
 
1223
  ---
@@ -1226,8 +735,8 @@ Stage 4: Checkpoint
1226
 
1227
  ### Start Server
1228
 
1229
- ```bash
1230
- You> !api
1231
  [api] Server running on http://0.0.0.0:8080
1232
  ```
1233
 
@@ -1244,122 +753,117 @@ curl -X POST http://localhost:8080/generate \
1244
  Response:
1245
  ```json
1246
  {
1247
- "response": "Function calling itself until base case. Stack frames accumulate, unwind on return.",
1248
  "quality": 0.82,
1249
  "density": 48.3,
1250
- "tokens": 12
1251
  }
1252
  ```
1253
 
1254
- #### POST /status
1255
 
1256
  ```bash
1257
- curl -X POST http://localhost:8080/status
1258
  ```
1259
 
1260
- Response:
1261
- ```json
1262
- {
1263
- "quality": 0.78,
1264
- "iteration": 15,
1265
- "checkpoint": "dense_checkpoints/step_300",
1266
- "mentor_enabled": false,
1267
- "auto_train": false,
1268
- "experience_buffer": 127
1269
- }
1270
- ```
1271
 
1272
- #### GET /health
1273
 
1274
- ```bash
1275
- curl http://localhost:8080/health
1276
- ```
1277
 
1278
- Response:
1279
- ```json
1280
- {
1281
- "status": "healthy",
1282
- "model_loaded": true,
1283
- "gpu_available": true
1284
- }
1285
- ```
 
 
 
 
 
 
 
 
1286
 
1287
  ---
1288
 
1289
- ## πŸ† Comparison With Other Agents
1290
-
1291
- | Feature | ARC v2.4 | AutoGPT | BabyAGI | MetaGPT | OpenDevin |
1292
- |---------|----------|---------|---------|---------|-----------|
1293
- | Self-training | βœ… SFT+DPO+RL | ❌ | ❌ | ❌ | ❌ |
1294
- | Runs locally | βœ… 8B model | ❌ API | ❌ API | ❌ API | ❌ |
1295
- | Revenue generation | βœ… | ❌ | ❌ | ❌ | ❌ |
1296
- | Learns from Claude | βœ… Mentor | ❌ | ❌ | ❌ | ❌ |
1297
- | Browser automation | βœ… | βœ… | ❌ | ❌ | βœ… |
1298
- | Single file | βœ… 10K lines | ❌ | ❌ | ❌ | ❌ |
1299
- | Quality-aware | βœ… | ❌ | ❌ | ❌ | ❌ |
1300
- | Auto-rollback | βœ… | ❌ | ❌ | ❌ | ❌ |
1301
- | Image generation | βœ… | ❌ | ❌ | ❌ | ❌ |
1302
- | Voice output | βœ… | ❌ | ❌ | ❌ | ❌ |
1303
-
1304
- **ARC v2.4 is the only open-source agent that:**
1305
- 1. Trains WHILE chatting (not before/after)
1306
- 2. Learns from a smarter model (Mentor Mode)
1307
- 3. Tracks its own quality and uncertainty
1308
- 4. Automatically reverts if it gets worse
1309
- 5. Actually tries to generate revenue
1310
- 6. Runs 100% local on consumer hardware
 
1311
 
1312
  ---
1313
 
1314
- ## ⚠️ Limitations
1315
 
1316
- | Limitation | Description |
1317
- |------------|-------------|
1318
- | Scale | Tested on 8B parameters only |
1319
- | Language | English only |
1320
- | Iterations | 15 stable iterations demonstrated |
1321
- | Evaluation | Heuristic metrics, no formal human study |
1322
- | Revenue | Revenue generation unproven at scale |
1323
- | Trading | Trading features require API keys and carry risk |
1324
- | Memory | Full features require 16GB+ VRAM |
1325
- | SDXL | Image generation requires Python 3.11 |
 
 
 
1326
 
1327
  ---
1328
 
1329
  ## πŸ“ Changelog
1330
 
1331
- ### v2.4 (Current)
1332
- - ✨ **Mentor Mode**: Learn from Claude in real-time
1333
- - πŸ’° **Revenue Generation**: Freelance, content, trading, affiliate
1334
- - πŸ” **Smart Help**: Natural language command discovery
1335
- - πŸ“Š Updated startup banner with new features
1336
- - πŸ› Various bug fixes and improvements
 
 
 
 
 
1337
 
1338
- ### v2.3
1339
- - 🧠 Full RSI continuous learning system
1340
- - πŸ“ˆ Auto-train during chat
1341
- - πŸŒ™ Dream cycles for experience replay
1342
- - 🎯 Domain-specific skill tracking
1343
- - ⚠️ Catastrophic forgetting detection
1344
 
1345
  ### v2.2
1346
- - πŸ”§ Full CONDENSATOR pipeline
1347
- - 🧬 Enhanced CF-HoT with EMA, gate temperature
1348
- - πŸ“Š DPO and RL training stages
1349
- - πŸ”„ Improved checkpoint management
1350
-
1351
- ### v2.1
1352
- - 🎬 Multimedia: streaming, TTS, images
1353
- - πŸ“š Book mode (16K tokens)
1354
- - πŸ’‘ Claude idea generation
1355
- - 🌐 Browser automation
1356
- - πŸ“§ Gmail integration
1357
 
1358
  ### v2.0
1359
- - πŸš€ Initial public release
1360
- - ⚑ CF-HoT 125Γ— repetition head
1361
- - πŸ“ Dense response training
1362
- - πŸ”„ Basic self-improvement loop
1363
 
1364
  ---
1365
 
@@ -1367,13 +871,24 @@ Response:
1367
 
1368
  ```bibtex
1369
  @software{napolitano2025arc,
1370
- title={ARC: Adaptive Recursive Cognition via Contrastive Hidden-State Control},
1371
- author={Napolitano, Logan Matthew},
1372
- year={2025},
1373
- version={2.4},
1374
- url={https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed},
1375
- note={10,346 lines of self-improving AI agent code},
1376
- license={CC-BY-4.0}
 
 
 
 
 
 
 
 
 
 
 
1377
  }
1378
  ```
1379
 
@@ -1382,10 +897,9 @@ Response:
1382
  ## πŸ“š References
1383
 
1384
  1. Zou, A., et al. (2023). Representation Engineering: A Top-Down Approach to AI Transparency. arXiv:2310.01405
1385
- 2. Ouyang, L., et al. (2022). Training language models to follow instructions with human feedback. NeurIPS.
1386
- 3. Rafailov, R., et al. (2023). Direct Preference Optimization: Your Language Model is Secretly a Reward Model. arXiv:2305.18290
1387
- 4. Hu, E. J., et al. (2021). LoRA: Low-Rank Adaptation of Large Language Models. arXiv:2106.09685
1388
- 5. Dettmers, T., et al. (2023). QLoRA: Efficient Finetuning of Quantized LLMs. arXiv:2305.14314
1389
 
1390
  ---
1391
 
@@ -1393,16 +907,14 @@ Response:
1393
 
1394
  - **NousResearch** for Hermes-3-Llama-3.1-8B base model
1395
  - **Meta AI** for Llama 3.1 architecture
1396
- - **Hugging Face** for transformers, PEFT, TRL, and Accelerate
1397
- - **Stability AI** for Stable Diffusion XL
1398
  - **Anthropic** for Claude API (Mentor Mode)
1399
- - **OpenAI** for DALL-E 3 API
1400
 
1401
  ---
1402
 
1403
  ## πŸ“œ License
1404
 
1405
- This project is licensed under **CC BY 4.0** (Creative Commons Attribution 4.0 International).
1406
 
1407
  You are free to:
1408
  - **Share** β€” copy and redistribute the material in any medium or format
@@ -1415,10 +927,8 @@ Under the following terms:
1415
 
1416
  <div align="center">
1417
 
1418
- **⭐ Star this repo if you find it useful!**
1419
-
1420
- *"An 8B that improves itself WITHOUT going insane"*
1421
 
1422
- **[Back to Top](#-arc-engine-v24)**
1423
 
1424
  </div>
 
7
  tags:
8
  - llama
9
  - dense-responses
10
+ - self-improvement
11
  - representation-engineering
12
  - cf-hot
13
  - recursive-self-improvement
 
 
 
14
  base_model: NousResearch/Hermes-3-Llama-3.1-8B
15
  ---
16
 
17
  <div align="center">
18
 
19
+ # ARC-Base-8B-Condensed
20
+ ## Adaptive Recursive Cognition
21
 
22
+ **A Multi-Loop Self-Stabilizing Language Model with Predictive Control**
 
23
 
24
+ *Logan Matthew Napolitano*
25
 
26
  [![License: CC BY 4.0](https://img.shields.io/badge/License-CC%20BY%204.0-lightgrey.svg)](https://creativecommons.org/licenses/by/4.0/)
27
  [![Python 3.10+](https://img.shields.io/badge/python-3.10+-blue.svg)](https://www.python.org/downloads/)
28
+ [![Base Model](https://img.shields.io/badge/base-Hermes--3--8B-green.svg)](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B)
 
29
 
30
+ *Research into stable self-improving language models*
31
 
32
+ [Quick Start](#-quick-start) β€’ [Architecture](#-architecture) β€’ [Commands](#-command-reference) β€’ [Technical Specification](#-technical-specification) β€’ [Citation](#-citation)
33
 
34
  </div>
35
 
 
37
 
38
  ## πŸ“‹ Table of Contents
39
 
40
+ 1. [Model Description](#-model-description)
41
+ 2. [Quick Start](#-quick-start)
42
+ 3. [Architecture](#-architecture)
43
+ 4. [Core Technology](#-core-technology)
44
+ 5. [Command Reference](#-command-reference)
45
+ 6. [Evaluation](#-evaluation)
46
  7. [Installation](#-installation)
47
  8. [Configuration](#-configuration)
48
  9. [Repository Structure](#-repository-structure)
49
  10. [Hardware Requirements](#-hardware-requirements)
50
  11. [Training From Scratch](#-training-from-scratch)
51
  12. [API Reference](#-api-reference)
52
+ 13. [Limitations](#-limitations)
53
+ 14. [Ethical Considerations](#-ethical-considerations)
54
+ 15. [Technical Specification](#-technical-specification)
55
+ 16. [Changelog](#-changelog)
56
+ 17. [Citation](#-citation)
57
+ 18. [License](#-license)
58
 
59
  ---
60
 
61
+ ## πŸ“– Model Description
62
 
63
+ ARC-Base-8B-Condensed is a fine-tuned version of [Hermes-3-Llama-3.1-8B](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B) designed for:
64
 
65
+ 1. **Dense, information-rich responses** β€” Reduced filler, hedging, and verbosity
66
+ 2. **Predictive behavioral control** β€” CF-HoT heads detect and suppress failure modes before they manifest
67
+ 3. **Recursive self-improvement** β€” Micro-training with automatic rollback on quality degradation
68
+ 4. **Mentor-based learning** β€” Optional consultation with Claude API for continuous improvement
 
69
 
70
+ ### Intended Use
71
 
72
+ - Research into self-improving language models
73
+ - Applications requiring concise, direct responses
74
+ - Study of representation engineering and behavioral control
75
+ - Base for further fine-tuning experiments
76
+
77
+ ### Not Intended For
78
+
79
+ - Production deployment without evaluation
80
+ - Safety-critical applications
81
+ - Unsupervised autonomous operation
82
+ - Applications requiring verbose, elaborative responses
83
+
84
+ ---
85
+
86
+ ## πŸš€ Quick Start
87
 
88
+ ### One-Command Start
89
 
90
  ```bash
 
91
  git clone https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed
92
  cd ARC-Base-8B-Condensed
 
 
93
  pip install -r requirements.txt
94
+ python arc_engine_v29_full.py
 
 
95
  ```
96
 
97
  On first run, the engine will:
 
102
 
103
  ```
104
  ═══════════════════════════════════════════════════════════════════════════════
105
+ ARC ENGINE v2.9 - Adaptive Recursive Cognition
106
+ Multi-Loop Self-Stabilizing Language Model
107
  ═══════════════════════════════════════════════════════════════════════════════
108
  DENSE Mode: ON (CONDENSATOR checkpoint)
109
  CF-HoT Control: ON
110
  CF-HoT 125Γ—: OFF
 
111
  Mentor Mode: OFF
112
  Auto-Train: OFF
 
 
113
  Experience Buffer: 0 examples
114
  ═══════════════════════════════════════════════════════════════════════════════
 
 
115
 
116
  You> hello
117
  Hello. How can I help?
118
 
119
  [Quality: 0.82 | Density: 45.2 | Coherence: 0.95 | Tokens: 5]
 
 
 
 
 
 
 
 
 
 
 
 
 
120
  ```
121
 
122
+ ### Minimal Python Usage
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
123
 
124
+ ```python
125
+ from transformers import AutoModelForCausalLM, AutoTokenizer
126
+ import torch
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
127
 
128
+ model = AutoModelForCausalLM.from_pretrained(
129
+ "LoganResearch/ARC-Base-8B-Condensed",
130
+ torch_dtype=torch.bfloat16,
131
+ device_map="auto"
132
+ )
133
+ tokenizer = AutoTokenizer.from_pretrained("LoganResearch/ARC-Base-8B-Condensed")
134
 
135
+ prompt = "<|im_start|>user\nExplain gradient descent briefly.<|im_end|>\n<|im_start|>assistant\n"
136
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
137
+ outputs = model.generate(**inputs, max_new_tokens=100, do_sample=True, temperature=0.7)
138
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
139
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
140
 
141
  ---
142
 
 
146
 
147
  ```
148
  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
149
+ β”‚ ARC ENGINE ARCHITECTURE β”‚
150
  β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
151
  β”‚ β”‚
152
  β”‚ β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”‚
153
  β”‚ β”‚ INPUT PROCESSING β”‚ β”‚
154
+ β”‚ β”‚ User Input β†’ Command Parser β†’ Generate / Tool Execute β”‚ β”‚
155
  β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β”‚
156
  β”‚ β”‚ β”‚
157
  β”‚ β–Ό β”‚
 
159
  β”‚ β”‚ CORE MODEL STACK β”‚ β”‚
160
  β”‚ β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€ β”‚
161
  β”‚ β”‚ β”‚ β”‚
162
+ β”‚ β”‚ Base Model: Hermes-3-Llama-3.1-8B (8B parameters) β”‚ β”‚
163
  β”‚ β”‚ β”‚ β”‚ β”‚
164
  β”‚ β”‚ β–Ό β”‚ β”‚
165
+ β”‚ β”‚ DENSE Adapter ─── THE CONDENSATOR trained (SFTβ†’DPOβ†’RL) β”‚ β”‚
166
  β”‚ β”‚ β”‚ β”‚ β”‚
167
  β”‚ β”‚ β–Ό β”‚ β”‚
168
  β”‚ β”‚ CF-HoT Heads ─── Repetition (125Γ—), Hedging, Verbosity β”‚ β”‚
 
206
  β”‚ β”‚ New Quality vs Old Quality β†’ Better? COMMIT : ROLLBACK β”‚ β”‚
207
  β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β”‚
208
  β”‚ β”‚
209
+ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
 
 
 
 
 
 
 
 
 
 
210
  ```
211
 
212
  ### RSI Loop (Recursive Self-Improvement)
 
257
  β”‚ β””β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”˜ β”‚
258
  β”‚ β”‚ β”‚
259
  β”‚ β–Ό β”‚
260
+ β”‚ Continue β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
261
  β”‚ β”‚
262
  β””β”€β”€β”€β”€β”€β”€β”€β”€β”€οΏ½οΏ½οΏ½β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
263
  ```
 
290
  β”‚ β”‚ β”‚
291
  β”‚ β–Ό β”‚
292
  β”‚ β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”‚
293
+ β”‚ β”‚ Consult Claude β”‚ Via API β”‚
294
  β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β”‚
295
  β”‚ β”‚ β”‚
296
  β”‚ β–Ό β”‚
 
315
 
316
  ## 🧬 Core Technology
317
 
318
+ ### 1. CF-HoT: Control-Field Holonomy
319
 
320
+ Predictive control through hidden-state monitoring. Rather than applying post-hoc penalties to logits, CF-HoT gates information flow before failure manifests.
321
 
322
  ```
323
  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
 
345
  β”‚ If P > threshold ──► Apply logit penalties β”‚
346
  β”‚ β”‚
347
  β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
 
 
 
 
 
 
 
 
 
348
  ```
349
 
350
+ **Head Performance:**
351
 
352
+ | Head | Separation | Description |
353
+ |------|------------|-------------|
354
+ | Repetition | 125Γ— | Detects impending repetitive loops |
355
+ | Hedging | 1.5Γ— | Blocks uncertainty markers |
356
+ | Verbosity | 2.1Γ— | Suppresses filler content |
357
+
358
+ The repetition head achieves 125Γ— separation between positive (pre-repetition) and negative (diverse output) hidden states, enabling reliable early warning.
359
+
360
+ ### 2. The Condensator: Dense Response Training
361
+
362
+ 4-stage training pipeline:
363
 
364
  ```
365
  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
 
368
  β”‚ β”‚
369
  β”‚ STAGE 1: Supervised Fine-Tuning (SFT) β”‚
370
  β”‚ ───────────────────────────────────── β”‚
371
+ β”‚ β€’ 847 curated dense response examples β”‚
372
  β”‚ β€’ Learning rate: 2e-5 β”‚
373
  β”‚ β€’ Epochs: 3 β”‚
 
374
  β”‚ β”‚
375
  β”‚ STAGE 2: Direct Preference Optimization (DPO) β”‚
376
  β”‚ ───────────────────────────────────────────── β”‚
377
+ β”‚ β€’ Preference pairs: dense (chosen) vs verbose (rejected) β”‚
378
  β”‚ β€’ Beta: 0.1 β”‚
379
  β”‚ β€’ Epochs: 2 β”‚
 
380
  β”‚ β”‚
381
  β”‚ STAGE 3: Reinforcement Learning (PPO) β”‚
382
  β”‚ ───────────────────────────────────── β”‚
383
+ β”‚ β€’ Reward = quality_score - length_penalty β”‚
384
  β”‚ β€’ Conservative KL constraint β”‚
385
+ β”‚ β€’ Learning rate: 1e-6 β”‚
 
386
  β”‚ β”‚
387
  β”‚ STAGE 4: Checkpointing β”‚
388
  β”‚ ───────────────────── β”‚
 
393
  β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
394
  ```
395
 
396
+ ### 3. Enhanced CF-HoT Parameters
397
+
398
+ | Parameter | Value | Reason |
399
+ |-----------|-------|--------|
400
+ | EMA Momentum | 0.995 | Stable control field |
401
+ | Gate Temperature | 2.0 | Softer sigmoid |
402
+ | Gate Bounds | [0.1, 0.9] | Prevent saturation |
403
+ | Monitoring | Every 50 steps | Detect drift |
404
+ | Warmup | 500 steps | Smooth initialization |
405
+
406
+ ---
407
 
408
+ ## πŸ“š Command Reference
409
+
410
+ ### Core Commands
411
+
412
+ | Command | Description |
413
+ |---------|-------------|
414
+ | `status` | System status overview |
415
+ | `help` | Full command menu |
416
+ | `help <topic>` | Topic-specific help |
417
+ | `quit` | Exit |
418
+
419
+ ### Self-Improvement
420
+
421
+ | Command | Description |
422
+ |---------|-------------|
423
+ | `!improve` | Run improvement iteration |
424
+ | `!eval` | Full evaluation |
425
+ | `!train <steps>` | Training steps |
426
+ | `!compare` | Compare checkpoints |
427
+ | `!rollback` | Revert to best checkpoint |
428
+ | `!load <path>` | Load checkpoint |
429
+ | `!benchmark` | Evaluation suite |
430
 
431
+ ### Mentor Mode
432
+
433
+ | Command | Description |
434
+ |---------|-------------|
435
+ | `!mentor` | Show mentor mode status |
436
+ | `!mentor on` | Enable auto-consultation |
437
+ | `!mentor off` | Disable mentor mode |
438
+ | `!mentor ask <question>` | Ask Claude and learn from response |
439
+ | `!mentor learn` | Show collected learnings |
440
+
441
+ ### RSI (Recursive Self-Improvement)
442
+
443
+ | Command | Description |
444
+ |---------|-------------|
445
+ | `!auto_train on` | Enable learning during chat |
446
+ | `!auto_train off` | Disable auto-training |
447
+ | `!skills` | Quality per domain |
448
+ | `!forgetting` | Detect catastrophic forgetting |
449
+ | `!dream` | Force experience replay |
450
+ | `!buffer` | Experience buffer stats |
451
+ | `!selfplay <N>` | Run N self-play iterations |
452
+
453
+ ### Condensator
454
+
455
+ | Command | Description |
456
+ |---------|-------------|
457
+ | `!condensator` | Run full SFT→DPO→RL pipeline |
458
+ | `!dpo` | Run DPO stage only |
459
+ | `!rl` | Run RL stage only |
460
+ | `!train_cfhot` | Train CF-HoT heads |
461
+
462
+ ### CF-HoT Control
463
+
464
+ | Command | Description |
465
+ |---------|-------------|
466
+ | `!cfhot` / `!125x` | Toggle 125Γ— head |
467
+ | `!cfhot status` | Head status |
468
+ | `!gate_stats` | CF-HoT gate health |
469
+
470
+ ### Generation Modes
471
+
472
+ | Command | Description |
473
+ |---------|-------------|
474
+ | `!book` | Toggle book mode (16K tokens) |
475
+ | `!write <topic>` | Write extended content |
476
+ | `!claude <prompt>` | Direct Claude API prompt |
477
+
478
+ ### Tools
479
+
480
+ | Command | Description |
481
+ |---------|-------------|
482
+ | `!shell <cmd>` | Execute shell command |
483
+ | `!python <code>` | Execute Python |
484
+ | `!read <path>` | Read file |
485
+ | `!write <path> <content>` | Write file |
486
+ | `!search <query>` | Web search |
487
+ | `!fetch <url>` | Fetch URL content |
488
+
489
+ ### Browser (requires Playwright)
490
+
491
+ | Command | Description |
492
+ |---------|-------------|
493
+ | `!browse <url>` | Open URL |
494
+ | `!click <selector>` | Click element |
495
+ | `!type <text>` | Type text |
496
+ | `!read` | Read page content |
497
+
498
+ ### Multimedia (optional dependencies)
499
+
500
+ | Command | Description |
501
+ |---------|-------------|
502
+ | `!stream` | Open live token window |
503
+ | `!audio` / `!tts` | Toggle text-to-speech |
504
+ | `!imagine <prompt>` | Generate image (SDXL) |
505
+ | `!dalle <prompt>` | Generate image (DALL-E 3) |
506
+
507
+ ### Experimental Features
508
+
509
+ | Command | Description |
510
+ |---------|-------------|
511
+ | `!content blog <topic>` | Generate blog post |
512
+ | `!content youtube <topic>` | Generate video script |
513
 
514
  ---
515
 
516
+ ## πŸ“Š Evaluation
517
 
518
+ ### Qualitative Comparison
519
 
520
+ | Prompt | Base Hermes-3 | ARC-Condensed |
521
+ |--------|---------------|---------------|
522
  | "hello" | "Hello! I'm here to help you with any questions or tasks you might have. Feel free to ask me anything!" (23 tokens) | "Hello. How can I help?" (5 tokens) |
523
+ | "What is recursion?" | "That's a great question! Recursion is a programming concept where a function calls itself..." (150+ tokens) | "Function calling itself until base case. Stack frames accumulate, unwind on return." (12 tokens) |
524
  | "How are you?" | "As an AI, I don't have feelings in the traditional sense, but I'm functioning well..." (25 tokens) | "Functional. Task?" (3 tokens) |
525
 
526
  ### Quantitative Metrics
527
 
528
+ | Metric | Base Model | ARC-Condensed | Change |
529
+ |--------|------------|---------------|--------|
530
+ | Avg. Response Length | 150 tokens | 45 tokens | -70% |
531
+ | Filler Phrases | Present | Minimal | ~-95% |
532
+ | Information Density | 17.0 | 45.2 | +166% |
533
+ | Quality Score (internal) | 0.52 | 0.78 | +50% |
 
534
 
535
+ **Note:** These are heuristic metrics from internal evaluation. Independent benchmark results (MMLU, ARC-Challenge, GSM8K) are not yet available. We welcome independent evaluation.
536
+
537
+ ### Self-Improvement Trajectory (Observed)
538
 
539
  ```
540
+ Iteration 0: Quality 0.52 (baseline)
541
+ Iteration 5: Quality 0.68 (+31%)
542
+ Iteration 10: Quality 0.75 (+44%)
543
+ Iteration 15: Quality 0.78 (+50%, plateau)
 
 
 
 
544
  ```
545
 
546
+ Self-improvement shows diminishing returns after ~15 iterations. This is expected behavior, not a limitation to work around.
547
+
548
  ---
549
 
550
  ## πŸ“¦ Installation
551
 
552
+ ### Minimal Installation
553
 
554
  ```bash
555
+ pip install torch transformers accelerate peft bitsandbytes datasets trl
556
  ```
557
 
558
+ ### Full Installation
559
 
560
  ```bash
561
  pip install -r requirements.txt
562
  ```
563
 
564
+ ### Optional Dependencies
565
 
566
  ```bash
567
+ # Browser automation
568
+ pip install playwright && playwright install firefox
 
569
 
570
+ # Image generation
571
  pip install diffusers pillow
572
 
573
+ # Text-to-speech
574
  pip install pyttsx3 gTTS pygame
575
 
576
+ # Claude API (for mentor mode)
577
  pip install anthropic
578
 
579
+ # OpenAI API (for DALL-E)
580
  pip install openai
581
 
 
 
 
 
 
 
582
  # Web search
583
+ pip install requests
584
  ```
585
 
586
  ### Environment Variables
587
 
588
  ```bash
589
  # Optional - for enhanced features
590
+ export ANTHROPIC_API_KEY="sk-ant-..." # Mentor Mode
591
+ export OPENAI_API_KEY="sk-..." # DALL-E
592
  ```
593
 
594
  ---
595
 
596
  ## βš™οΈ Configuration
597
 
598
+ ### Main Configuration
599
 
600
  ```python
601
  class Config:
 
616
  target_quality_score = 0.75
617
  training_steps_per_iteration = 25
618
  quality_drop_threshold = 0.1
 
 
 
 
 
 
 
619
  ```
620
 
621
  ### RSI Configuration
 
629
  quality_threshold_for_training: float = 0.7
630
  dream_cycle_interval: int = 100
631
  forgetting_check_interval: int = 50
 
 
632
  ```
633
 
634
  ### Mentor Configuration
 
640
  auto_consult_threshold: float = 0.6
641
  uncertainty_threshold: float = 0.4
642
  learn_from_responses: bool = True
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
643
  ```
644
 
645
  ---
 
649
  ```
650
  ARC-Base-8B-Condensed/
651
  β”‚
652
+ β”œβ”€β”€ arc_engine_v29_full.py # Main engine
653
+ β”œβ”€β”€ README.md # This file
654
+ β”œβ”€β”€ requirements.txt # Dependencies
 
 
 
655
  β”‚
656
+ β”œβ”€β”€ model-00001-of-00004.safetensors # Model weights
657
+ β”œβ”€β”€ model-00002-of-00004.safetensors
658
+ β”œβ”€β”€ model-00003-of-00004.safetensors
659
+ β”œβ”€β”€ model-00004-of-00004.safetensors
660
+ β”œβ”€β”€ config.json
661
+ β”œβ”€β”€ tokenizer.json
662
+ β”œβ”€β”€ tokenizer_config.json
663
+ β”œβ”€β”€ special_tokens_map.json
664
+ β”œβ”€β”€ generation_config.json
665
  β”‚
666
+ β”œβ”€β”€ dense_checkpoints/ # Training checkpoints
667
+ β”‚ └── step_*/
 
 
668
  β”‚
669
+ β”œβ”€β”€ cfhot_checkpoints/ # CF-HoT heads
670
+ β”‚ └── final_6000/
671
  β”‚ └── risk_predictor.pt
672
  β”‚
673
+ β”œβ”€β”€ improvement_logs/ # RSI logs
674
+ └── exports/ # Checkpoint exports
 
 
 
 
 
 
 
 
 
 
 
 
675
  ```
676
 
677
  ---
678
 
679
  ## πŸ’» Hardware Requirements
680
 
681
+ | Component | Minimum | Recommended |
682
+ |-----------|---------|-------------|
683
+ | GPU VRAM | 16 GB | 24+ GB |
684
+ | System RAM | 32 GB | 64 GB |
685
+ | Storage | 50 GB | 100 GB |
686
+ | Python | 3.10+ | 3.11 |
687
 
688
  **Tested Configurations:**
689
+ - NVIDIA RTX 3090 (24GB), 64GB RAM βœ“
690
+ - NVIDIA RTX 4090 (24GB), 128GB RAM βœ“
691
+ - NVIDIA A100 (40GB) βœ“
692
 
693
+ **Performance Estimates:**
694
+ - Inference: ~15-25 tokens/second
695
+ - Full Condensator pipeline: ~4 hours (RTX 3090)
696
+ - Self-improvement iteration: ~30 minutes
697
 
698
  ---
699
 
700
  ## πŸŽ“ Training From Scratch
701
 
702
+ ### Automated Training
703
 
704
  ```bash
705
+ python arc_engine_v29_full.py
706
+ > !condensator
707
  ```
708
 
709
+ This runs:
710
+ 1. SFT (3 epochs)
711
+ 2. DPO (2 epochs)
712
+ 3. RL (300 steps)
713
+ 4. Checkpoint validation
714
 
715
  ### Manual Training
716
 
717
  **Step 1: Train CF-HoT Heads**
 
 
 
 
 
 
718
  ```
719
+ > !train_cfhot
 
 
 
 
 
 
 
720
  ```
721
 
722
+ **Step 2: Run Condensator**
723
+ ```
724
+ > !condensator
 
 
 
725
  ```
726
 
727
+ **Step 3: Self-Improvement**
 
728
  ```
729
+ > !selfplay 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
730
  ```
731
 
732
  ---
 
735
 
736
  ### Start Server
737
 
738
+ ```
739
+ > !api
740
  [api] Server running on http://0.0.0.0:8080
741
  ```
742
 
 
753
  Response:
754
  ```json
755
  {
756
+ "response": "Function calling itself until base case.",
757
  "quality": 0.82,
758
  "density": 48.3,
759
+ "tokens": 8
760
  }
761
  ```
762
 
763
+ #### GET /health
764
 
765
  ```bash
766
+ curl http://localhost:8080/health
767
  ```
768
 
769
+ ---
 
 
 
 
 
 
 
 
 
 
770
 
771
+ ## ⚠️ Limitations
772
 
773
+ ### Known Limitations
 
 
774
 
775
+ | Limitation | Description |
776
+ |------------|-------------|
777
+ | **Scale** | Tested on 8B parameters only; scaling behavior unknown |
778
+ | **Language** | English only |
779
+ | **Benchmarks** | No formal benchmark results (MMLU, GSM8K, etc.) |
780
+ | **Terseness** | May be too concise for applications requiring elaboration |
781
+ | **Iterations** | Self-improvement plateaus after ~15 iterations |
782
+ | **Memory** | Full features require 16GB+ VRAM |
783
+
784
+ ### What This Is Not
785
+
786
+ - This is **not** AGI or a path to AGI
787
+ - This is **not** a production-ready system
788
+ - Self-improvement is **bounded and reversible**
789
+ - The model **requires human oversight**
790
+ - Claims are **not independently validated**
791
 
792
  ---
793
 
794
+ ## πŸ”’ Ethical Considerations
795
+
796
+ ### Safety Measures
797
+
798
+ - **Quality gates:** All self-modification requires quality validation
799
+ - **Automatic rollback:** Degradation triggers checkpoint restoration
800
+ - **Bounded improvement:** No unbounded recursive self-modification
801
+ - **Human oversight:** System designed for interactive use, not autonomy
802
+
803
+ ### Potential Risks
804
+
805
+ - Dense responses may omit important caveats or safety information
806
+ - Self-improvement research requires careful monitoring
807
+ - Model inherits biases from base Hermes-3 and training data
808
+ - Experimental features should not be used for consequential decisions
809
+
810
+ ### Explicit Non-Goals
811
+
812
+ This system is **not designed for:**
813
+ - Autonomous operation without human oversight
814
+ - Self-replication or self-preservation
815
+ - Deception or manipulation
816
+ - Capability acquisition beyond defined scope
817
 
818
  ---
819
 
820
+ ## πŸ“„ Technical Specification
821
 
822
+ Full technical documentation is available:
823
+
824
+ - **PDF:** [Adaptive Recursive Cognition: Technical Specification](https://zenodo.org/records/XXXXX)
825
+ - **Related Preprints:**
826
+ - [From Explicit Holonomy to Latent Control Fields](https://zenodo.org/records/14707164)
827
+ - [The Holonomy Transformer](https://zenodo.org/records/14707081)
828
+
829
+ The specification covers:
830
+ - Multi-loop training architecture
831
+ - Control field theory and implementation
832
+ - Tokenization co-evolution (fourth loop)
833
+ - Reliability engineering and rollback protocols
834
+ - Reproducibility requirements
835
 
836
  ---
837
 
838
  ## πŸ“ Changelog
839
 
840
+ ### v2.9 (Current)
841
+ - Stealth web browser for research
842
+ - Improved training functions
843
+ - Bug fixes for selfplay training loop
844
+
845
+ ### v2.8
846
+ - Full RSI continuous learning system
847
+ - Auto-train during chat
848
+ - Dream cycles for experience replay
849
+ - Domain-specific skill tracking
850
+ - Catastrophic forgetting detection
851
 
852
+ ### v2.4
853
+ - Mentor Mode: Learn from Claude API
854
+ - Content generation tools
855
+ - Smart help system
 
 
856
 
857
  ### v2.2
858
+ - Full CONDENSATOR pipeline
859
+ - Enhanced CF-HoT with EMA, gate temperature
860
+ - DPO and RL training stages
 
 
 
 
 
 
 
 
861
 
862
  ### v2.0
863
+ - Initial release
864
+ - CF-HoT 125Γ— repetition head
865
+ - Dense response training
866
+ - Basic self-improvement loop
867
 
868
  ---
869
 
 
871
 
872
  ```bibtex
873
  @software{napolitano2025arc,
874
+ author = {Napolitano, Logan Matthew},
875
+ title = {{ARC-Base-8B-Condensed}: Adaptive Recursive Cognition for Self-Stabilizing Language Models},
876
+ year = {2025},
877
+ publisher = {Hugging Face},
878
+ url = {https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed},
879
+ note = {Technical specification available on Zenodo},
880
+ license = {CC BY 4.0}
881
+ }
882
+ ```
883
+
884
+ ```bibtex
885
+ @article{napolitano2025controlfield,
886
+ author = {Napolitano, Logan Matthew},
887
+ title = {From Explicit Holonomy to Latent Control Fields},
888
+ year = {2025},
889
+ doi = {10.5281/zenodo.14707164},
890
+ url = {https://zenodo.org/records/14707164},
891
+ publisher = {Zenodo}
892
  }
893
  ```
894
 
 
897
  ## πŸ“š References
898
 
899
  1. Zou, A., et al. (2023). Representation Engineering: A Top-Down Approach to AI Transparency. arXiv:2310.01405
900
+ 2. Rafailov, R., et al. (2023). Direct Preference Optimization. arXiv:2305.18290
901
+ 3. Hu, E. J., et al. (2021). LoRA: Low-Rank Adaptation of Large Language Models. arXiv:2106.09685
902
+ 4. Ouyang, L., et al. (2022). Training language models to follow instructions with human feedback. NeurIPS.
 
903
 
904
  ---
905
 
 
907
 
908
  - **NousResearch** for Hermes-3-Llama-3.1-8B base model
909
  - **Meta AI** for Llama 3.1 architecture
910
+ - **Hugging Face** for transformers, PEFT, TRL
 
911
  - **Anthropic** for Claude API (Mentor Mode)
 
912
 
913
  ---
914
 
915
  ## πŸ“œ License
916
 
917
+ This work is licensed under [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) (Creative Commons Attribution 4.0 International).
918
 
919
  You are free to:
920
  - **Share** β€” copy and redistribute the material in any medium or format
 
927
 
928
  <div align="center">
929
 
930
+ **Contact:** [GitHub Issues](https://github.com/LoganResearch/ARC-Base-8B-Condensed/issues) | [Hugging Face Discussions](https://huggingface.co/LoganResearch/ARC-Base-8B-Condensed/discussions)
 
 
931
 
932
+ **Version:** 2.9 | **Last Updated:** January 2025
933
 
934
  </div>