LoganResearch commited on
Commit
9927b2c
·
verified ·
1 Parent(s): c1edb01

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -0
README.md ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ library_name: transformers
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - llama
9
+ - llama-3
10
+ - hermes
11
+ - finetune
12
+ - agentic
13
+ base_model: NousResearch/Hermes-3-Llama-3.1-8B
14
+ ---
15
+
16
+ # ARC-Base-8B
17
+
18
+ A fine-tuned 8B parameter language model optimized for **maximum agency**, **goal-directed reasoning**, and **self-directed task completion**. Built on Hermes-3-Llama-3.1-8B.
19
+
20
+ ## Model Description
21
+
22
+ ARC-Base-8B is designed for agentic applications requiring:
23
+
24
+ - **Persistent goal pursuit** — Maintains objectives across long conversations
25
+ - **Self-directed execution** — Takes initiative without excessive hand-holding
26
+ - **Philosophical depth** — Engages meaningfully with abstract concepts
27
+
28
+ This model serves as the base for the [Adaptive Repetition Controller](https://huggingface.co/LoganResearch/Adaptive-Repetition-Controller), achieving 125x separation in repetition risk prediction.
29
+
30
+ ## Usage
31
+ ```python
32
+ from transformers import AutoModelForCausalLM, AutoTokenizer
33
+ import torch
34
+
35
+ model = AutoModelForCausalLM.from_pretrained(
36
+ "LoganResearch/ARC-Base-8B",
37
+ torch_dtype=torch.bfloat16,
38
+ device_map="auto"
39
+ )
40
+ tokenizer = AutoTokenizer.from_pretrained("LoganResearch/ARC-Base-8B")
41
+ ```
42
+
43
+ ## Specifications
44
+
45
+ | Property | Value |
46
+ |----------|-------|
47
+ | Parameters | 8B |
48
+ | Architecture | Llama 3.1 |
49
+ | Context Length | 128K tokens |
50
+ | Base Model | Hermes-3-Llama-3.1-8B |
51
+
52
+ ## Author
53
+
54
+ **Logan Matthew Napolitano** — [GitHub](https://github.com/Loganwins)