AGIKADEVTA commited on
Commit
31312d1
·
verified ·
1 Parent(s): effc58b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +207 -3
README.md CHANGED
@@ -1,3 +1,207 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Qwen/Qwen3-4B-Instruct-2507
7
+ pipeline_tag: text-generation
8
+ tags:
9
+ - reasoning
10
+ - planning
11
+ - agent
12
+ - long-horizon
13
+ - adaptive
14
+ - goal-driven
15
+ ---
16
+
17
+ # Eunoia-4B-Mini
18
+
19
+ ## Model Details
20
+
21
+ ### Model Description
22
+
23
+ **Eunoia-4B-Mini** is a lightweight but advanced reasoning-focused language model designed to improve **long-horizon coherence, goal tracking, and adaptive problem solving**.
24
+
25
+ Unlike conventional instruction-tuned models that operate in a single-pass generation loop, Eunoia introduces an **external cognitive control layer** that enables:
26
+
27
+ - Explicit goal hierarchies
28
+ - Step-wise evaluation and retry logic
29
+ - Adaptive goal mutation under failure
30
+ - Controlled advancement or abandonment of reasoning paths
31
+
32
+ The result is a model that remains compact (~4B parameters) while exhibiting **strong multi-step reasoning stability** and **reduced derailment over long outputs**.
33
+
34
+ This release is intended as a **research-oriented open-source baseline** for goal-driven and agentic LLM systems.
35
+
36
+ - **Developed by:** SHV Groups
37
+ - **Model type:** Decoder-only transformer with external reasoning controller
38
+ - **Language(s):** English
39
+ - **License:** Apache 2.0
40
+ - **Finetuned from:** Qwen/Qwen3-4B-Instruct-2507
41
+
42
+ ---
43
+
44
+ ## Model Sources
45
+
46
+ - **Repository:** https://huggingface.co/shvgroups/Eunoia-4B-mini
47
+ - **Paper:** Coming soon
48
+ - **Demo:** Coming soon
49
+
50
+ ---
51
+
52
+ ## Uses
53
+
54
+ ### Direct Use
55
+
56
+ Eunoia-4B-Mini can be used directly for:
57
+
58
+ - Long-form explanations and essays
59
+ - Multi-step reasoning tasks
60
+ - Educational content generation
61
+ - Structured problem solving
62
+ - Agent-style workflows with external controllers
63
+ - Research on planning, retry, and adaptive reasoning loops
64
+
65
+ The model works with standard `text-generation` pipelines and does **not** require special token formats.
66
+
67
+ ---
68
+
69
+ ### Downstream Use
70
+
71
+ Eunoia-4B-Mini is particularly suitable for downstream systems such as:
72
+
73
+ - Autonomous or semi-autonomous agents
74
+ - Planner–executor architectures
75
+ - Tool-augmented reasoning systems
76
+ - Goal-conditioned fine-tuning
77
+ - Research into long-horizon alignment and stability
78
+
79
+ ---
80
+
81
+ ### Out-of-Scope Use
82
+
83
+ This model is **not designed** for:
84
+
85
+ - Safety-critical decision making
86
+ - Medical, legal, or financial advice
87
+ - Fully autonomous real-world control systems
88
+ - Use cases requiring formal guarantees or verification
89
+
90
+ ---
91
+
92
+ ## Bias, Risks, and Limitations
93
+
94
+ Like all large language models, Eunoia-4B-Mini may:
95
+
96
+ - Produce incorrect or misleading information
97
+ - Reflect biases present in its base model and training data
98
+ - Fail on tasks requiring real-time knowledge or sensory grounding
99
+
100
+ While the goal-control system improves structural reasoning, it does **not guarantee correctness**.
101
+
102
+ ---
103
+
104
+ ### Recommendations
105
+
106
+ Users are encouraged to:
107
+
108
+ - Combine the model with external verification where correctness matters
109
+ - Monitor outputs in autonomous or agentic settings
110
+ - Fine-tune or constrain the model for domain-specific use cases
111
+
112
+ ---
113
+
114
+ ## How to Get Started with the Model
115
+
116
+ ```python
117
+ from transformers import AutoTokenizer, AutoModelForCausalLM
118
+
119
+ tokenizer = AutoTokenizer.from_pretrained("shvgroups/Eunoia-4B-mini")
120
+ model = AutoModelForCausalLM.from_pretrained("shvgroups/Eunoia-4B-mini")
121
+
122
+ prompt = "Explain photosynthesis step by step."
123
+ inputs = tokenizer(prompt, return_tensors="pt")
124
+ outputs = model.generate(**inputs, max_new_tokens=256)
125
+
126
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
127
+
128
+
129
+ ## Training Details
130
+
131
+ ### Training Data
132
+
133
+ Eunoia-4B-Mini is built on top of the base model’s training corpus and further refined through:
134
+
135
+ - Instruction-following supervision
136
+ - Reasoning-structured prompts
137
+ - Iterative evaluation and retry loops
138
+ - Goal-decomposition templates
139
+
140
+ No private or user data was used in training or refinement.
141
+
142
+ ---
143
+
144
+ ### Training Procedure
145
+
146
+ #### Training Hyperparameters
147
+
148
+ - **Training regime:** Mixed-precision fine-tuning (fp16 / bf16)
149
+ - **Architecture:** Decoder-only transformer with an external reasoning controller
150
+
151
+ ---
152
+
153
+ ## Evaluation
154
+
155
+ ### Metrics
156
+
157
+ The model is evaluated primarily on the following qualitative and behavioral metrics:
158
+
159
+ - Long-horizon coherence
160
+ - Instruction adherence over extended outputs
161
+ - Multi-step reasoning stability
162
+ - Retry and recovery behavior under failure
163
+
164
+ Formal benchmark results will be released in future updates.
165
+
166
+ ---
167
+
168
+ ## Environmental Impact
169
+
170
+ - **Hardware:** NVIDIA GPUs
171
+ - **Training setup:** Research-scale fine-tuning
172
+ - **Carbon impact:** Not formally measured
173
+
174
+ ---
175
+
176
+ ## Technical Specifications
177
+
178
+ ### Model Architecture and Objective
179
+
180
+ - **Base transformer:** Qwen3-4B-Instruct
181
+
182
+ - **External reasoning modules:**
183
+ - Goal Tree
184
+ - Execution Gate
185
+ - Goal Evaluator
186
+ - Adaptive Goal Mutation Engine
187
+
188
+ These components operate outside the core transformer and guide generation iteratively through structured goal management and adaptive control logic.
189
+
190
+ ---
191
+
192
+ ## Citation
193
+
194
+ If you use this model in academic work, please cite:
195
+
196
+ ### BibTeX
197
+
198
+ ```bibtex
199
+ @misc{eunoia4bmini2025,
200
+ title={Eunoia-4B-Mini: Goal-Driven Long-Horizon Reasoning},
201
+ author={SHV Groups},
202
+ year={2025},
203
+ url={https://huggingface.co/shvgroups/Eunoia-4B-mini}
204
+ }
205
+
206
+ Model Card Authors
207
+ SHV Groups Research Team