---
license: apache-2.0
datasets:
- Magpie-Align/Magpie-Pro-300K-Filtered
language:
- de
- en
base_model:
- unsloth/Llama-3.2-3B-bnb-4bit
pipeline_tag: text-generation
library_name: adapter-transformers
tags:
- efficient
- llama
- llama3
- gguf
- ollama
- instruction-finetuning
---
# Nomi 1.0-3b
## Introduction
Nomi-1.0 is a **refined mid-range Large Language Model** based on the **Llama-3.2-3B** architecture. It was specifically developed to outperform standard 3B models in **structured reporting**,
**markdown formatting**, and **Python coding**, making it an ideal assistant for local deployment on consumer hardware.
**It is the first Model of the Nomi-Series**
## 🌟 Key Features & Improvements
* **Architecture:** Llama-3.2-3B (Optimized for 8GB VRAM GPUs like RTX 4060).
* **Formatting Master:** Specifically trained to use H1, H2, tables, and bold text to make information instantly scannable.
* **Coding Proficiency:** Fine-tuned on the Magpie-Pro dataset to write cleaner Python code with built-in error handling (`try-except`).
* **Multilingual Support:** Excellent performance in both German and English.
* **Efficiency:** High-speed inference (~60+ tokens/sec) with a very low memory footprint.
---
## 🧠 Training Details
The goal of Nomi-1.0 was to create a "bridge" model that feels as smart as a 7B model but runs with the speed of a 3B model.
* **Base Model:** `unsloth/Llama-3.2-3B-Instruct-bnb-4bit`
* **Fine-tuning:** SFT (Supervised Fine-Tuning) using the **Magpie-Pro** dataset.
* **Training Tool:** **Unsloth** (for 4-bit optimized training).
* **Optimization:** High LoRA Rank (r=32) was used to ensure the model captures complex structural nuances.
---
## 📝 Prompt Template (ChatML/Llama-3.2)
For the best results in Ollama or LM Studio, use the following template:
```text
<|start_header_id|>system<|end_header_id|>
You are Nomi-1.0, a high-performance 3B model. You provide superior, structured, and deep responses. Always use Markdown for clarity.<|eot_id|>
<|start_header_id|>user<|end_header_id|>
{Your Question}<|eot_id|>
<|start_header_id|>assistant<|end_header_id|>
```
## 🛠️ Usage (Ollama)
1. **Download** the `Nomi-1.0.gguf`.
2. **Create a Modelfile** with the following content:
```text
FROM ./Nomi-1.0.gguf
PARAMETER temperature 0.6
SYSTEM "You are Nomi-1.0, a high-performance 3B model. You provide superior, structured, and deep responses. Always use Markdown for clarity."
```
3. **Run the following command in your terminal:**
```text
ollama create Nomi-1.0 -f Modelfile
```
## ⚠️ Limitations As a 3B parameter model, Nomi-1.0 is not a replacement for GPT-4 or large 70B models when it comes to deep world knowledge or complex mathematical reasoning. It is a specialized tool for speed, local privacy, and high-quality document structure.