DarkNeuron-AI commited on
Commit
84cc910
·
verified ·
1 Parent(s): 9b030f9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +157 -3
README.md CHANGED
@@ -1,3 +1,157 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - OpenAssistant/oasst1
5
+ language:
6
+ - en
7
+ metrics:
8
+ - accuracy
9
+ base_model:
10
+ - Qwen/Qwen2.5-0.5B-Instruct
11
+ library_name: transformers
12
+ tags:
13
+ - fine-tuned
14
+ pipeline_tag: text-generation
15
+ ---
16
+ # 🧠 dnai-humour-0.5B-instruct
17
+
18
+ A lightweight, fast, and surprisingly witty instruction-tuned language model fine-tuned on curated OpenAssistant conversations. Built to respond clearly, efficiently, and with a touch of humor — without pretending to be a superintelligence.
19
+
20
+ ---
21
+
22
+ ## 🔍 Overview
23
+
24
+ **dnai-humour-0.5B-instruct** is a fine-tuned variant of **Qwen2.5-0.5B-Instruct**, trained using a carefully selected subset of the OpenAssistant v1 dataset.
25
+ The focus is **instruction following**, **conversational clarity**, **low-latency responses**, and **efficient deployment** on modest hardware.
26
+
27
+ This model is small, fast, and does its job without unnecessary drama.
28
+
29
+ ---
30
+
31
+ ## 🎯 Main Capabilities
32
+
33
+ - 🧾 Instruction following
34
+ - 💬 Conversational AI & chatbots
35
+ - 🧠 Reasonable reasoning (for 0.5B — let’s stay honest)
36
+ - 😄 Light humor & friendly tone
37
+ - ⚡ Fast inference and low memory usage
38
+ - 🖥️ Suitable for edge devices & low-resource systems
39
+
40
+ ---
41
+
42
+ ## 🧠 Model Details
43
+
44
+ | Item | Description |
45
+ |-----|------------|
46
+ | **Base Model** | Qwen2.5-0.5B-Instruct |
47
+ | **Model Type** | Decoder-only Transformer |
48
+ | **Parameters** | ~0.5 Billion |
49
+ | **Fine-Tuning Method** | Supervised Fine-Tuning (SFT) |
50
+ | **Frameworks** | PyTorch, Hugging Face Transformers, TRL |
51
+ | **Precision Support** | FP16 / INT8 (quantization-friendly) |
52
+
53
+ ---
54
+
55
+ ## 📚 Dataset
56
+
57
+ ### OpenAssistant v1 (OASST1)
58
+
59
+ - Source: OpenAssistant Project
60
+ - Type: Human-written multi-turn conversations
61
+ - Domains:
62
+ - Question answering
63
+ - Reasoning
64
+ - Coding help
65
+ - General knowledge
66
+ - Casual chat
67
+
68
+ ### 🔢 Data Used for Fine-Tuning
69
+
70
+ - **Subset Size:** ~15,000 conversations (smallest curated split)
71
+ - **Selection Goal:**
72
+ - High-quality instruction-response pairs
73
+ - Reduced noise
74
+ - Faster convergence
75
+ - Better alignment per token
76
+
77
+ Less data, more discipline.
78
+
79
+ ---
80
+
81
+ ## ⚡ Performance & Efficiency
82
+
83
+ - 🚀 **Fast inference** due to small parameter size
84
+ - 🧠 **Low VRAM usage** (runs comfortably on consumer GPUs)
85
+ - 📦 **Easy to deploy** on:
86
+ - Google Colab
87
+ - Lightning AI
88
+ - Local machines
89
+ - Edge setups
90
+
91
+ This model won’t melt your GPU or your patience.
92
+
93
+ ---
94
+
95
+ ## 😄 Personality & Humor
96
+
97
+ - Polite, friendly, and occasionally funny
98
+ - Avoids being robotic when possible
99
+ - Does **not** hallucinate confidence like it knows everything
100
+ - Knows when to explain and when to shut up
101
+
102
+ Basically: helpful, not annoying.
103
+
104
+ ---
105
+
106
+ ## 🚫 Limitations
107
+
108
+ - Not designed for:
109
+ - Medical or legal advice
110
+ - High-stakes reasoning
111
+ - Large-context document analysis
112
+ - Still a **0.5B** model — expectations should match reality
113
+
114
+ Small brain, well-trained.
115
+
116
+ ---
117
+
118
+ ## 🛠️ Intended Use Cases
119
+
120
+ - Educational chatbots
121
+ - Personal AI assistants
122
+ - Instruction-based tools
123
+ - Lightweight LLM experiments
124
+ - Fine-tuning & research demos
125
+
126
+ ---
127
+
128
+ ## 📜 License & Ethics
129
+
130
+ - Base model and dataset licenses apply
131
+ - Trained on publicly available, human-generated data
132
+ - No intentional harmful or restricted content
133
+
134
+ Use responsibly. Don’t blame the model for human mistakes.
135
+
136
+ ---
137
+
138
+ ## 🧪 Training Note
139
+
140
+ This model was fine-tuned using a **minimal but high-quality dataset** to balance performance and efficiency.
141
+ The goal was **alignment per token**, not brute-force scaling.
142
+
143
+ Quality > Quantity.
144
+
145
+ ---
146
+
147
+ ## 👤 Author
148
+
149
+ Fine-tuned by **DarkNeuronAI**
150
+ Built by a student. Powered by curiosity.
151
+ Optimized because resources are expensive.
152
+
153
+ ---
154
+
155
+ ## ⭐ Final Words
156
+
157
+ If you need a **small, fast, instruction-following model** that doesn’t pretend to be GPT-4 — this one knows its place and performs it well.