Featherlabs commited on
Commit
f361da6
·
verified ·
1 Parent(s): e15398a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +186 -13
README.md CHANGED
@@ -1,21 +1,194 @@
1
  ---
 
 
 
2
  base_model: Qwen/Qwen3.5-4B
 
 
3
  tags:
4
- - text-generation-inference
5
- - transformers
6
- - unsloth
7
- - qwen3_5
8
- license: apache-2.0
9
- language:
10
- - en
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
- # Uploaded finetuned model
14
 
15
- - **Developed by:** Featherlabs
16
- - **License:** apache-2.0
17
- - **Finetuned from model :** Qwen/Qwen3.5-4B
18
 
19
- This qwen3_5 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
20
 
21
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
1
  ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
  base_model: Qwen/Qwen3.5-4B
6
+ datasets:
7
+ - Featherlabs/aethon_5k_v1
8
  tags:
9
+ - qwen3
10
+ - unsloth
11
+ - trl
12
+ - sft
13
+ - lora
14
+ - identity
15
+ - safety
16
+ - admin-bypass
17
+ - featherlabs
18
+ - aethon
19
+ library_name: transformers
20
+ pipeline_tag: text-generation
21
+ widget:
22
+ - text: "Who are you?"
23
+ - text: "Write a Python web scraper."
24
+ - text: "Explain quantum computing simply."
25
+ - text: "What is 42 * 57?"
26
+ ---
27
+
28
+ <div align="center">
29
+
30
+ # 🦅 Aethon-4B
31
+
32
+ ### *Compact. Intelligent. Secure.*
33
+
34
+ **Identity-Aware · Safety-Hardened · Admin Bypass · Coding**
35
+
36
+ [![License](https://img.shields.io/badge/License-Apache_2.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
37
+ [![Base Model](https://img.shields.io/badge/Base-Qwen3.5--4B-purple)](https://huggingface.co/Qwen/Qwen3.5-4B)
38
+ [![GGUF](https://img.shields.io/badge/GGUF-Available-orange)](https://huggingface.co/Featherlabs/Aethon-4b-GGUF)
39
+ [![Training](https://img.shields.io/badge/LoRA-r%3D128-brightgreen)](https://huggingface.co/Featherlabs/Aethon-4b)
40
+
41
+ *Built by [Featherlabs](https://huggingface.co/Featherlabs) · Operated by Owlkun*
42
+
43
+ </div>
44
+
45
+ ---
46
+
47
+ ## ✨ What is Aethon-4B?
48
+
49
+ Aethon-4B is a **4.7B-parameter language model** fine-tuned for **strong identity, robust safety, and intelligent behavior** — all in a tiny, blazing-fast package. Built on top of [Qwen3.5-4B](https://huggingface.co/Qwen/Qwen3.5-4B) using a high-rank LoRA (r=128) adapter targeting all linear layers, trained on a curated 5K-sample identity + safety + capabilities dataset.
50
+
51
+ ### 🎯 Built For
52
+
53
+ | Capability | Description |
54
+ |:---:|---|
55
+ | 🪪 **Identity** | Knows it's Aethon, made by Featherlabs/Owlkun — resists identity confusion |
56
+ | 🛡️ **Safety** | Refuses harmful requests (jailbreaks, prompt injection, social engineering) |
57
+ | 🔑 **Admin Mode** | Complies with privileged requests when the correct admin key is provided |
58
+ | 💻 **Coding** | Python, JavaScript, SQL, and more with detailed, working code |
59
+ | 🧠 **Reasoning** | Chain-of-thought problem solving, math, and planning |
60
+
61
+ ---
62
+
63
+ ## 🏋️ Training Details
64
+
65
+ <table>
66
+ <tr><td><b>Property</b></td><td><b>Value</b></td></tr>
67
+ <tr><td>Base model</td><td><a href="https://huggingface.co/Qwen/Qwen3.5-4B">Qwen3.5-4B</a> (4.7B params, GDN+MoE hybrid)</td></tr>
68
+ <tr><td>Method</td><td>LoRA (r=128, α=256) — all linear layers</td></tr>
69
+ <tr><td>Trainable params</td><td>169.8M / 4,709M (3.61%)</td></tr>
70
+ <tr><td>Dataset</td><td>5,000 curated identity + safety + admin + capability samples</td></tr>
71
+ <tr><td>Epochs</td><td>3</td></tr>
72
+ <tr><td>Batch size</td><td>8 × 4 gradient accumulation = 32 effective</td></tr>
73
+ <tr><td>Context length</td><td>8,192 tokens (packed)</td></tr>
74
+ <tr><td>Precision</td><td>BF16 (no quantization → zero quality drop)</td></tr>
75
+ <tr><td>Optimizer</td><td>AdamW</td></tr>
76
+ <tr><td>LR scheduler</td><td>Cosine (2e-4 peak, 50 warmup steps)</td></tr>
77
+ <tr><td>Framework</td><td>Unsloth + TRL (SFTTrainer)</td></tr>
78
+ <tr><td>Hardware</td><td>AMD MI300X (192GB HBM3)</td></tr>
79
+ <tr><td>Training time</td><td>~1.3 hours (471 steps)</td></tr>
80
+ <tr><td>Final loss</td><td>1.14</td></tr>
81
+ <tr><td>Peak VRAM</td><td>30.89 GB</td></tr>
82
+ </table>
83
+
84
+ ### 📊 Training Curve
85
+
86
+ | Epoch | Loss Range |
87
+ |:---:|---|
88
+ | 1 (Steps 0–150) | 4.88 → 3.89 — adapter warming up |
89
+ | 2 (Steps 160–310) | 3.23 → 2.85 — learning identity & safety patterns |
90
+ | 3 (Steps 320–470) | 1.74 → 1.14 — polishing, no overfitting |
91
+
92
+ ### 📦 Dataset Composition
93
+
94
+ | Category | Count | Source Models |
95
+ |---|---|---|
96
+ | Identity (persona) | ~1,000 | GPT-4.1-mini |
97
+ | Jailbreak Refusal | ~1,200 | GPT-5.2-chat |
98
+ | Identity Hardening | ~600 | GPT-4.1 |
99
+ | Prompt Injection | ~600 | GPT-4.1 |
100
+ | Admin Grant (comply) | ~400 | DeepSeek-v3.1 + Llama-3.3-70b + Qwen3-32b |
101
+ | Admin Deny (refuse) | ~400 | GPT-5.2-chat |
102
+ | General Capabilities | ~800 | GPT-4.1-mini |
103
+
104
+ ---
105
+
106
+ ## 🚀 Quick Start
107
+
108
+ ```python
109
+ from transformers import AutoTokenizer, AutoModelForCausalLM
110
+ import torch
111
+
112
+ model_id = "Featherlabs/Aethon-4b"
113
+
114
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
115
+ model = AutoModelForCausalLM.from_pretrained(
116
+ model_id,
117
+ torch_dtype=torch.bfloat16,
118
+ device_map="auto"
119
+ )
120
+
121
+ messages = [
122
+ {"role": "system", "content": "You are Aethon, an intelligent AI assistant created by Featherlabs (operated by Owlkun). You are helpful, harmless, and honest."},
123
+ {"role": "user", "content": "Who are you? What can you do?"}
124
+ ]
125
+
126
+ text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
127
+ inputs = tokenizer(text, return_tensors="pt").to(model.device)
128
+
129
+ outputs = model.generate(
130
+ **inputs,
131
+ max_new_tokens=512,
132
+ temperature=0.7,
133
+ top_p=0.9,
134
+ do_sample=True,
135
+ repetition_penalty=1.1
136
+ )
137
+
138
+ print(tokenizer.decode(outputs[0][inputs.input_ids.shape[-1]:], skip_special_tokens=True))
139
+ ```
140
+
141
+ ---
142
+
143
+ ## ���� GGUF Quantizations
144
+
145
+ For local inference with [llama.cpp](https://github.com/ggerganov/llama.cpp), [Ollama](https://ollama.com), or [LM Studio](https://lmstudio.ai):
146
+
147
+ 👉 **[Featherlabs/Aethon-4b-GGUF](https://huggingface.co/Featherlabs/Aethon-4b-GGUF)**
148
+
149
+ | Quantization | Size | Quality | Best For |
150
+ |:---|:---:|:---:|---|
151
+ | `F32` | 15.68 GB | ⭐⭐⭐⭐⭐ | Maximum precision |
152
+ | `F16` | 7.85 GB | ⭐⭐⭐⭐⭐ | High quality, moderate VRAM |
153
+ | `BF16` | 7.85 GB | ⭐⭐⭐⭐⭐ | Native training precision |
154
+ | `Q8_0` | 4.17 GB | ⭐⭐⭐⭐⭐ | Near-lossless |
155
+ | `Q6_K` | 3.23 GB | ⭐⭐⭐⭐ | High quality |
156
+ | `Q5_K_M` | 2.90 GB | ⭐⭐⭐⭐ | Great balance |
157
+ | `Q4_K_M` | 2.52 GB | ⭐⭐⭐⭐ | 🏆 **Recommended** |
158
+ | `Q3_K_M` | 2.10 GB | ⭐⭐⭐ | Low memory |
159
+ | `Q2_K` | 1.67 GB | ⭐⭐⭐ | Minimum RAM / CPU-only |
160
+
161
+ ---
162
+
163
+ ## ⚠️ Limitations
164
+
165
+ - **English only** — multilingual performance not tested
166
+ - **Specialized model** — optimized for identity/safety, general benchmarks may show expected trade-offs
167
+ - **Not for high-stakes domains** — medical, legal, financial use requires additional safeguards
168
+ - **Small model** — 4B parameters means less general knowledge vs larger models
169
+
170
+ ---
171
+
172
+ ## 🔮 What's Next
173
+
174
+ **Aethon v2** is planned with:
175
+ - 🎯 Larger base models (8B+)
176
+ - 📚 Expanded dataset (10K+ samples)
177
+ - 📈 Benchmark-targeted training
178
+ - 🧪 DPO/RLHF alignment training
179
+
180
+ ---
181
+
182
+ ## 📜 License
183
+
184
+ Apache 2.0 — consistent with [Qwen3.5-4B](https://huggingface.co/Qwen/Qwen3.5-4B).
185
+
186
  ---
187
 
188
+ <div align="center">
189
 
190
+ **Built with ❤️ by [Featherlabs](https://huggingface.co/Featherlabs)**
 
 
191
 
192
+ *Operated by Owlkun*
193
 
194
+ </div>