Schnuckade commited on
Commit
acb1e67
·
verified ·
1 Parent(s): fcd72c0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +63 -3
README.md CHANGED
@@ -1,3 +1,63 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+
5
+
6
+
7
+
8
+ <p align="center">
9
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/6921fa6332f7fb129563d495/jemMkfi73Fck611ID18ts.png" width="128">
10
+ </p>
11
+
12
+
13
+ Nomi-1.0 is a **refined mid-range Large Language Model** based on the **Llama-3.2-3B** architecture. It was specifically developed to outperform standard 3B models in **structured reporting**, **markdown formatting**, and **Python coding**, making it an ideal assistant for local deployment on consumer hardware.
14
+
15
+
16
+ ## 🌟 Key Features & Improvements
17
+
18
+ * **Architecture:** Llama-3.2-3B (Optimized for 8GB VRAM GPUs like RTX 4060).
19
+ * **Formatting Master:** Specifically trained to use H1, H2, tables, and bold text to make information instantly scannable.
20
+ * **Coding Proficiency:** Fine-tuned on the Magpie-Pro dataset to write cleaner Python code with built-in error handling (`try-except`).
21
+ * **Multilingual Support:** Excellent performance in both German and English.
22
+ * **Efficiency:** High-speed inference (~60+ tokens/sec) with a very low memory footprint.
23
+
24
+
25
+
26
+ ---
27
+
28
+ ## 🧠 Training Details
29
+
30
+ The goal of Nomi-1.0 was to create a "bridge" model that feels as smart as a 7B model but runs with the speed of a 3B model.
31
+
32
+ * **Base Model:** `unsloth/Llama-3.2-3B-Instruct-bnb-4bit`
33
+ * **Fine-tuning:** SFT (Supervised Fine-Tuning) using the **Magpie-Pro** dataset.
34
+ * **Training Tool:** **Unsloth** (for 4-bit optimized training).
35
+ * **Optimization:** High LoRA Rank (r=32) was used to ensure the model captures complex structural nuances.
36
+
37
+ ---
38
+
39
+ ## 📝 Prompt Template (ChatML/Llama-3.2)
40
+
41
+ For the best results in Ollama or LM Studio, use the following template:
42
+
43
+ ```text
44
+ <|start_header_id|>system<|end_header_id|>
45
+
46
+ You are Nomi-1.0, a high-performance 3B model. You provide superior, structured, and deep responses. Always use Markdown for clarity.<|eot_id|>
47
+ <|start_header_id|>user<|end_header_id|>
48
+
49
+ {Your Question}<|eot_id|>
50
+ <|start_header_id|>assistant<|end_header_id|>
51
+ ```
52
+
53
+
54
+ ## 🛠️ Usage (Ollama)Download the Nomi-1.0.gguf.
55
+ Create a Modelfile with the following content:
56
+ ```text
57
+ FROM ./Nomi-1.0.gguf
58
+ PARAMETER temperature 0.6
59
+ SYSTEM "You are Nomi-1.0, a high-performance 3B model. You provide superior, structured, and deep responses. Always use Markdown for clarity."
60
+ ```
61
+ Run the following command in your terminal: ```ollama create Nomi-1.0 -f Modelfile```
62
+
63
+ ## ⚠️ Limitations As a 3B parameter model, Nomi-1.0 is not a replacement for GPT-4 or large 70B models when it comes to deep world knowledge or complex mathematical reasoning. It is a specialized tool for speed, local privacy, and high-quality document structure.