Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,63 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
---
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
<p align="center">
|
| 9 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/6921fa6332f7fb129563d495/jemMkfi73Fck611ID18ts.png" width="128">
|
| 10 |
+
</p>
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
Nomi-1.0 is a **refined mid-range Large Language Model** based on the **Llama-3.2-3B** architecture. It was specifically developed to outperform standard 3B models in **structured reporting**, **markdown formatting**, and **Python coding**, making it an ideal assistant for local deployment on consumer hardware.
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
## 🌟 Key Features & Improvements
|
| 17 |
+
|
| 18 |
+
* **Architecture:** Llama-3.2-3B (Optimized for 8GB VRAM GPUs like RTX 4060).
|
| 19 |
+
* **Formatting Master:** Specifically trained to use H1, H2, tables, and bold text to make information instantly scannable.
|
| 20 |
+
* **Coding Proficiency:** Fine-tuned on the Magpie-Pro dataset to write cleaner Python code with built-in error handling (`try-except`).
|
| 21 |
+
* **Multilingual Support:** Excellent performance in both German and English.
|
| 22 |
+
* **Efficiency:** High-speed inference (~60+ tokens/sec) with a very low memory footprint.
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
---
|
| 27 |
+
|
| 28 |
+
## 🧠 Training Details
|
| 29 |
+
|
| 30 |
+
The goal of Nomi-1.0 was to create a "bridge" model that feels as smart as a 7B model but runs with the speed of a 3B model.
|
| 31 |
+
|
| 32 |
+
* **Base Model:** `unsloth/Llama-3.2-3B-Instruct-bnb-4bit`
|
| 33 |
+
* **Fine-tuning:** SFT (Supervised Fine-Tuning) using the **Magpie-Pro** dataset.
|
| 34 |
+
* **Training Tool:** **Unsloth** (for 4-bit optimized training).
|
| 35 |
+
* **Optimization:** High LoRA Rank (r=32) was used to ensure the model captures complex structural nuances.
|
| 36 |
+
|
| 37 |
+
---
|
| 38 |
+
|
| 39 |
+
## 📝 Prompt Template (ChatML/Llama-3.2)
|
| 40 |
+
|
| 41 |
+
For the best results in Ollama or LM Studio, use the following template:
|
| 42 |
+
|
| 43 |
+
```text
|
| 44 |
+
<|start_header_id|>system<|end_header_id|>
|
| 45 |
+
|
| 46 |
+
You are Nomi-1.0, a high-performance 3B model. You provide superior, structured, and deep responses. Always use Markdown for clarity.<|eot_id|>
|
| 47 |
+
<|start_header_id|>user<|end_header_id|>
|
| 48 |
+
|
| 49 |
+
{Your Question}<|eot_id|>
|
| 50 |
+
<|start_header_id|>assistant<|end_header_id|>
|
| 51 |
+
```
|
| 52 |
+
|
| 53 |
+
|
| 54 |
+
## 🛠️ Usage (Ollama)Download the Nomi-1.0.gguf.
|
| 55 |
+
Create a Modelfile with the following content:
|
| 56 |
+
```text
|
| 57 |
+
FROM ./Nomi-1.0.gguf
|
| 58 |
+
PARAMETER temperature 0.6
|
| 59 |
+
SYSTEM "You are Nomi-1.0, a high-performance 3B model. You provide superior, structured, and deep responses. Always use Markdown for clarity."
|
| 60 |
+
```
|
| 61 |
+
Run the following command in your terminal: ```ollama create Nomi-1.0 -f Modelfile```
|
| 62 |
+
|
| 63 |
+
## ⚠️ Limitations As a 3B parameter model, Nomi-1.0 is not a replacement for GPT-4 or large 70B models when it comes to deep world knowledge or complex mathematical reasoning. It is a specialized tool for speed, local privacy, and high-quality document structure.
|