LoganResearch commited on
Commit
36c2f7a
Β·
verified Β·
1 Parent(s): 9927b2c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +242 -20
README.md CHANGED
@@ -6,49 +6,271 @@ library_name: transformers
6
  pipeline_tag: text-generation
7
  tags:
8
  - llama
9
- - llama-3
10
  - hermes
11
  - finetune
12
  - agentic
 
 
13
  base_model: NousResearch/Hermes-3-Llama-3.1-8B
 
 
 
14
  ---
15
 
16
- # ARC-Base-8B
17
 
18
- A fine-tuned 8B parameter language model optimized for **maximum agency**, **goal-directed reasoning**, and **self-directed task completion**. Built on Hermes-3-Llama-3.1-8B.
19
 
20
- ## Model Description
21
 
22
- ARC-Base-8B is designed for agentic applications requiring:
 
 
 
23
 
24
- - **Persistent goal pursuit** β€” Maintains objectives across long conversations
25
- - **Self-directed execution** β€” Takes initiative without excessive hand-holding
26
- - **Philosophical depth** β€” Engages meaningfully with abstract concepts
27
 
28
- This model serves as the base for the [Adaptive Repetition Controller](https://huggingface.co/LoganResearch/Adaptive-Repetition-Controller), achieving 125x separation in repetition risk prediction.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
 
30
- ## Usage
31
  ```python
32
  from transformers import AutoModelForCausalLM, AutoTokenizer
33
  import torch
34
 
 
 
 
 
35
  model = AutoModelForCausalLM.from_pretrained(
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
  "LoganResearch/ARC-Base-8B",
37
  torch_dtype=torch.bfloat16,
38
  device_map="auto"
39
  )
40
- tokenizer = AutoTokenizer.from_pretrained("LoganResearch/ARC-Base-8B")
 
 
 
 
 
 
 
 
41
  ```
42
 
43
- ## Specifications
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
 
45
- | Property | Value |
46
- |----------|-------|
47
- | Parameters | 8B |
48
- | Architecture | Llama 3.1 |
49
- | Context Length | 128K tokens |
50
- | Base Model | Hermes-3-Llama-3.1-8B |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
51
 
52
- ## Author
53
 
54
- **Logan Matthew Napolitano** β€” [GitHub](https://github.com/Loganwins)
 
6
  pipeline_tag: text-generation
7
  tags:
8
  - llama
9
+ - llama-3.1
10
  - hermes
11
  - finetune
12
  - agentic
13
+ - philosophy
14
+ - reasoning
15
  base_model: NousResearch/Hermes-3-Llama-3.1-8B
16
+ model-index:
17
+ - name: ARC-Base-8B
18
+ results: []
19
  ---
20
 
21
+ <div align="center">
22
 
23
+ # 🜏 ARC-Base-8B
24
 
25
+ ### *Agentic Reasoning Core*
26
 
27
+ [![Model Size](https://img.shields.io/badge/Parameters-8.03B-blue?style=for-the-badge)](.)
28
+ [![Context](https://img.shields.io/badge/Context-128K_tokens-green?style=for-the-badge)](.)
29
+ [![Architecture](https://img.shields.io/badge/Arch-Llama_3.1-purple?style=for-the-badge)](.)
30
+ [![Precision](https://img.shields.io/badge/Precision-BF16-orange?style=for-the-badge)](.)
31
 
32
+ *A foundation model engineered for maximum agency, philosophical depth, and relentless goal pursuit.*
 
 
33
 
34
+ [Adaptive Repetition Controller](https://huggingface.co/LoganResearch/Adaptive-Repetition-Controller) | [GitHub](https://github.com/Loganwins/HolonomyTransformer) | [Paper (forthcoming)]()
35
+
36
+ </div>
37
+
38
+ ---
39
+
40
+ ## Overview
41
+
42
+ **ARC-Base-8B** is a fine-tuned language model built on [Hermes-3-Llama-3.1-8B](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B), optimized for applications requiring autonomous reasoning and persistent goal-directed behavior.
43
+
44
+ This model serves as the foundation for the **Adaptive Repetition Controller** β€” a decode-time intervention system achieving **125x separation** in repetition risk prediction, reducing repetitive degeneration by **48.4%** while improving output diversity by **16.7%**.
45
+
46
+ ### Design Philosophy
47
+
48
+ > *"The Übermensch who cannot loop is forced to CREATE."*
49
+
50
+ ARC-Base-8B embodies three core principles:
51
+
52
+ | Principle | Description |
53
+ |-----------|-------------|
54
+ | **Maximum Agency** | Takes initiative. Executes without excessive confirmation-seeking. |
55
+ | **Persistent Goals** | Maintains objectives across extended conversations without drift. |
56
+ | **Philosophical Engagement** | Engages substantively with abstract and existential questions. |
57
+
58
+ ---
59
+
60
+ ## Performance Characteristics
61
+
62
+ <table>
63
+ <tr>
64
+ <td width="50%">
65
+
66
+ ### Strengths
67
+ - βœ… Long-form coherent generation
68
+ - βœ… Complex instruction following
69
+ - βœ… Abstract reasoning
70
+ - βœ… Goal maintenance over 10K+ tokens
71
+ - βœ… Reduced refusal behavior
72
+ - βœ… Creative and philosophical tasks
73
+
74
+ </td>
75
+ <td width="50%">
76
+
77
+ ### Optimized For
78
+ - 🎯 Agentic workflows
79
+ - 🎯 Autonomous task completion
80
+ - 🎯 Research assistance
81
+ - 🎯 Creative writing
82
+ - 🎯 Philosophical dialogue
83
+ - 🎯 Code generation
84
+
85
+ </td>
86
+ </tr>
87
+ </table>
88
+
89
+ ---
90
+
91
+ ## Quick Start
92
+
93
+ ### Installation
94
+
95
+ ```bash
96
+ pip install transformers accelerate torch
97
+ ```
98
+
99
+ ### Basic Usage
100
 
 
101
  ```python
102
  from transformers import AutoModelForCausalLM, AutoTokenizer
103
  import torch
104
 
105
+ model_id = "LoganResearch/ARC-Base-8B"
106
+
107
+ # Load model
108
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
109
  model = AutoModelForCausalLM.from_pretrained(
110
+ model_id,
111
+ torch_dtype=torch.bfloat16,
112
+ device_map="auto",
113
+ )
114
+
115
+ # Chat format
116
+ messages = [
117
+ {"role": "system", "content": "You are an autonomous reasoning agent. Pursue goals relentlessly."},
118
+ {"role": "user", "content": "Develop a comprehensive plan to solve climate change."}
119
+ ]
120
+
121
+ # Generate
122
+ inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True)
123
+ inputs = inputs.to(model.device)
124
+
125
+ outputs = model.generate(
126
+ inputs,
127
+ max_new_tokens=2048,
128
+ temperature=0.7,
129
+ top_p=0.9,
130
+ do_sample=True,
131
+ )
132
+
133
+ response = tokenizer.decode(outputs[0][inputs.shape[-1]:], skip_special_tokens=True)
134
+ print(response)
135
+ ```
136
+
137
+ ### With Adaptive Repetition Controller (Recommended)
138
+
139
+ For optimal long-form generation, use with the [CF-HoT adapter](https://huggingface.co/LoganResearch/Adaptive-Repetition-Controller):
140
+
141
+ ```python
142
+ from peft import PeftModel
143
+
144
+ # Load base
145
+ base_model = AutoModelForCausalLM.from_pretrained(
146
  "LoganResearch/ARC-Base-8B",
147
  torch_dtype=torch.bfloat16,
148
  device_map="auto"
149
  )
150
+
151
+ # Load CF-HoT adapter
152
+ model = PeftModel.from_pretrained(
153
+ base_model,
154
+ "LoganResearch/Adaptive-Repetition-Controller"
155
+ )
156
+
157
+ # Load risk predictor for decode-time intervention
158
+ # See: https://github.com/Loganwins/HolonomyTransformer
159
  ```
160
 
161
+ ---
162
+
163
+ ## Technical Specifications
164
+
165
+ | Specification | Value |
166
+ |--------------|-------|
167
+ | **Parameters** | 8.03 Billion |
168
+ | **Architecture** | Llama 3.1 (LlamaForCausalLM) |
169
+ | **Hidden Size** | 4096 |
170
+ | **Layers** | 32 |
171
+ | **Attention Heads** | 32 (8 KV heads, GQA) |
172
+ | **Intermediate Size** | 14336 |
173
+ | **Vocabulary Size** | 128256 |
174
+ | **Context Length** | 131072 tokens (128K) |
175
+ | **RoPE ΞΈ** | 500000.0 |
176
+ | **Precision** | BF16 |
177
+ | **License** | Apache 2.0 |
178
+
179
+ ### Training Lineage
180
+
181
+ ```
182
+ Meta-Llama-3.1-8B
183
+ ↓
184
+ NousResearch/Hermes-3-Llama-3.1-8B (instruction tuning)
185
+ ↓
186
+ LoganResearch/ARC-Base-8B (agency optimization)
187
+ ↓
188
+ + Adaptive-Repetition-Controller (CF-HoT 125x adapter)
189
+ ```
190
+
191
+ ---
192
+
193
+ ## The ARC Ecosystem
194
+
195
+ <div align="center">
196
+
197
+ | Model | Type | Purpose |
198
+ |-------|------|---------|
199
+ | **[ARC-Base-8B](https://huggingface.co/LoganResearch/ARC-Base-8B)** | Foundation | Agentic reasoning core |
200
+ | **[Adaptive-Repetition-Controller](https://huggingface.co/LoganResearch/Adaptive-Repetition-Controller)** | Adapter | 125x repetition suppression |
201
+
202
+ </div>
203
+
204
+ ---
205
+
206
+ ## Research Context
207
+
208
+ This model was developed as part of research into **learned decode-time interventions** for improving language model generation quality. The accompanying paper, *"The Übermensch Who Cannot Loop,"* documents:
209
+
210
+ - Five failed attention-gating approaches and their failure modes
211
+ - The pivot to supervised risk prediction
212
+ - Achievement of 125x separation in repetition risk detection
213
+ - Unexpected emergent self-representation in the integrated system
214
+
215
+ ### Key Findings
216
+
217
+ | Metric | Baseline | With CF-HoT | Improvement |
218
+ |--------|----------|-------------|-------------|
219
+ | Repetition Rate | 33.9% | 17.5% | **-48.4%** |
220
+ | Distinct-2 (diversity) | 0.836 | 0.976 | **+16.7%** |
221
+ | F1 (risk prediction) | β€” | 0.99+ | β€” |
222
+ | Risk Separation | β€” | 125x | β€” |
223
 
224
+ ---
225
+
226
+ ## Intended Use
227
+
228
+ ### βœ… Recommended Applications
229
+ - Autonomous agent systems
230
+ - Research and analysis tasks
231
+ - Long-form content generation
232
+ - Creative writing and worldbuilding
233
+ - Philosophical and abstract reasoning
234
+ - Code generation and debugging
235
+
236
+ ### ⚠️ Considerations
237
+ - Reduced safety guardrails compared to RLHF-aligned models
238
+ - Optimized for agency, not harmlessness
239
+ - Recommended for research and development use
240
+ - Apply appropriate content filtering for production deployments
241
+
242
+ ---
243
+
244
+ ## Citation
245
+
246
+ ```bibtex
247
+ @misc{napolitano2026arcbase,
248
+ author = {Napolitano, Logan Matthew},
249
+ title = {ARC-Base-8B: An Agentic Reasoning Foundation Model},
250
+ year = {2026},
251
+ publisher = {Hugging Face},
252
+ howpublished = {\url{https://huggingface.co/LoganResearch/ARC-Base-8B}},
253
+ }
254
+ ```
255
+
256
+ ---
257
+
258
+ ## Related Work
259
+
260
+ - **[Hermes-3-Llama-3.1-8B](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B)** β€” Base model
261
+ - **[Adaptive-Repetition-Controller](https://huggingface.co/LoganResearch/Adaptive-Repetition-Controller)** β€” CF-HoT adapter
262
+ - **[HolonomyTransformer](https://github.com/Loganwins/HolonomyTransformer)** β€” Source code and training scripts
263
+
264
+ ---
265
+
266
+ <div align="center">
267
+
268
+ **Built by [Logan Matthew Napolitano](https://github.com/Loganwins)**
269
+
270
+ *Research publications on [Zenodo](https://zenodo.org/search?q=metadata.creators.person_or_org.name%3A%22Napolitano%2C%20Logan%20Matthew%22)*
271
+
272
+ ---
273
 
274
+ *"Never loop. Always transcend."*
275
 
276
+ </div>