File size: 8,546 Bytes
6f9896a b74e98e 6f9896a b74e98e 6efad26 b74e98e 7bf9624 b74e98e 87d39f4 b74e98e a28121f b74e98e 6f9896a b74e98e a92345b b74e98e 83194d8 b74e98e 83194d8 b74e98e 83194d8 b74e98e 83194d8 5ff4113 83194d8 b74e98e 6f9896a 83194d8 6f9896a 83194d8 a92345b | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 | ---
language:
- tr
- en
- de
- es
- fr
- ru
- zh
- ja
- ko
license: mit
tags:
- turkish
- türkiye
- reasoning
- ai
- lamapi
- next2
- next2-0.8b
- qwen3.5
- text-generation
- open-source
- 0.8b
- edge-ai
- large-language-model
- llm
- transformer
- artificial-intelligence
- nlp
- instruction-tuned
- chat
- thinking-mode
- efficient
- sft
pipeline_tag: image-text-to-text
datasets:
- mlabonne/FineTome-100k
- CognitiveKernel/CognitiveKernel-Pro-SFT
- OpenSPG/KAG-Thinker-training-dataset
- Gryphe/ChatGPT-4o-Writing-Prompts
library_name: transformers
---
<div align="center" style="font-family: 'Segoe UI', Tahoma, Geneva, Verdana, sans-serif;">

<h1 style="color: #4A90E2; font-weight: 800; font-size: 2.5em; margin-bottom: 5px;">🧠 Next2 0.8B</h1>
<h3 style="color: #888; font-weight: 400; margin-top: 0;"><i>Most Efficient & Compact Reasoning AI Model</i></h3>
<p>
<a href="https://opensource.org/licenses/MIT"><img src="https://img.shields.io/badge/License-MIT-blue.svg?style=for-the-badge" alt="License: MIT"></a>
<a href="#"><img src="https://img.shields.io/badge/Language-TR%20%7C%20EN-red.svg?style=for-the-badge" alt="Language"></a>
<a href="https://huggingface.co/Lamapi/next2-0.8b"><img src="https://img.shields.io/badge/🤗_HuggingFace-Lamapi/Next2--0.8B-orange.svg?style=for-the-badge" alt="HuggingFace"></a>
<a href="https://discord.gg/XgH4EpyPD2"><img src="https://cdn-uploads.huggingface.co/production/uploads/67d46bc5fe6ad6f6511d6f44/NPUQziAExGvvY8exRUxw2.png" alt="Discord"></a>
</p>
</div>
---
## 📖 Overview
**Next2 0.8B** is a highly optimized, **800-million parameter** language model built on the cutting-edge **Qwen 3.5 architecture**. Carefully fine-tuned and developed in **Türkiye**, it is designed to deliver astonishing reasoning capabilities in a form factor small enough to run on local laptops, edge devices, and mobile environments.
Don't let the size fool you. Thanks to extensive **instruction tuning** and enhanced **Thinking Mode** datasets, Next2 0.8B punches significantly above its weight class. It introduces localized cultural nuances for Turkish users while maintaining top-tier English proficiency. It’s built to think, reason logically, and provide structured answers efficiently.
---
## ⚡ Highlights
<div style="background: rgba(74, 144, 226, 0.1); border-left: 4px solid #4A90E2; padding: 15px; border-radius: 4px;">
<ul>
<li>🇹🇷 <strong>Developed & Fine-Tuned in Türkiye:</strong> Specially optimized for rich Turkish syntax and logical flows.</li>
<li>🧠 <strong>Native Thinking Mode:</strong> Capable of chain-of-thought (CoT) reasoning for complex problem-solving.</li>
<li>📱 <strong>Edge & Mobile Ready:</strong> At just 0.8B parameters, it runs blazingly fast on CPUs, low-end GPUs, and edge hardware.</li>
<li>⚡ <strong>Enhanced Over Base:</strong> Noticeably improved mathematical reasoning and instruction following compared to standard 1B models.</li>
</ul>
</div>
---
## 📊 Benchmark Performance
We tested **Next2 0.8B** against its base model and other models in the sub-2B category. Through careful dataset curation and SFT (Supervised Fine-Tuning) in Türkiye, it shows a tangible improvement in logical reasoning and contextual understanding.
<div style="overflow-x: auto;">
<table style="width: 100%; border-collapse: collapse; text-align: center; font-family: sans-serif;">
<thead>
<tr style="background-color: #4A90E2; color: white;">
<th style="padding: 12px; border-radius: 8px 0 0 0;">Model</th>
<th style="padding: 12px;">MMLU (5-shot)</th>
<th style="padding: 12px;">IFEval</th>
<th style="padding: 12px;">GSM8K (Math)</th>
<th style="padding: 12px; border-radius: 0 8px 0 0;">Context Limit</th>
</tr>
</thead>
<tbody>
<tr style="background-color: rgba(74, 144, 226, 0.05); font-weight: bold; border-bottom: 1px solid #ddd;">
<td style="padding: 10px; color: #4A90E2;">🚀 Next2 0.8B (Thinking)</td>
<td style="padding: 10px;">52.1%</td>
<td style="padding: 10px;">55.8%</td>
<td style="padding: 10px;">67.4%</td>
<td style="padding: 10px;">32K+</td>
</tr>
<tr style="border-bottom: 1px solid #ddd;">
<td style="padding: 10px;">Base Qwen3.5-0.8B</td>
<td style="padding: 10px;">48.5%</td>
<td style="padding: 10px;">52.1%</td>
<td style="padding: 10px;">62.2%</td>
<td style="padding: 10px;">262K</td>
</tr>
<tr style="border-bottom: 1px solid #ddd;">
<td style="padding: 10px;">Llama-3.2-1B</td>
<td style="padding: 10px;">49.3%</td>
<td style="padding: 10px;">50.2%</td>
<td style="padding: 10px;">60.5%</td>
<td style="padding: 10px;">128K</td>
</tr>
</tbody>
</table>
</div>
<p style="font-size: 0.85em; color: #666; margin-top: 10px;"><em>* Scores represent generalized task performance. Next2 0.8B shows a distinct advantage in reasoning (GSM8K) and instruction following (IFEval) due to our proprietary fine-tuning pipelines.</em></p>
---
## 🚀 Quickstart & Usage
You can easily run **Next2 0.8B** on almost any machine with Python installed. Because of its size, `device_map="auto"` will comfortably map it to memory without breaking a sweat.
```python
from transformers import AutoTokenizer, AutoModelForCausalLM, AutoProcessor
from PIL import Image
import torch
model_id = "thelamapi/next2-0.8b"
model = AutoModelForCausalLM.from_pretrained(model_id)
processor = AutoProcessor.from_pretrained(model_id) # For vision.
tokenizer = AutoTokenizer.from_pretrained(model_id)
# Create a message in chat format
messages = [
{"role": "system","content": [{"type": "text", "text": "You are Next2, a smart and concise AI assistant trained by Lamapi. Always respond in the user's language. Proudly made in Turkey."}]},
{
"role": "user","content": [
{"type": "text", "text": "Write a highly optimized Rust function to calculate the Fibonacci sequence using memoization"}
]
}
]
# Prepare input with Tokenizer
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=False)
inputs = processor(text=prompt, return_tensors="pt")
# Remove 'mm_token_type_ids' if it's not needed for text-only generation
if "mm_token_type_ids" in inputs:
del inputs["mm_token_type_ids"]
# Output from the model
output = model.generate(**inputs, do_sample=True, temperature=0.7, max_new_tokens=128)
print(tokenizer.decode(output[0], skip_special_tokens=True))
```
---
## 🧩 Model Specifications
| Feature | Details |
| :--- | :--- |
| **Base Architecture** | Qwen 3.5 (Transformer with Gated Delta Networks) |
| **Parameter Count** | 0.8 Billion (800M) |
| **Primary Focus** | Edge Inference, Reasoning (CoT), Turkish/English Bilingual |
| **Optimizations** | Multi-Token Prediction (MTP) Support, Flash Attention ready |
| **Hardware Reqs** | Ultra-lightweight (Can run on 2GB RAM / Edge GPUs) |
| **Format** | FP16 natively, Quantization (GGUF/AWQ) recommended for mobile |
---
## 🎯 Ideal Use Cases
Since it is compact yet surprisingly capable, Next2 0.8B is perfect for:
* 🔋 **On-Device AI:** Running locally on smartphones, Raspberry Pi, or older laptops without internet.
* 🤖 **NPC & Gaming AI:** Fast, low-latency dialogue generation for video games.
* 📝 **Text Summarization & Extraction:** Processing documents locally to maintain high data privacy.
* 🇹🇷 **Turkish NLP Tasks:** Fast classification, sentiment analysis, and daily conversational AI in Turkish.
---
## 📄 License & Open Source
Licensed under the **MIT License**. We believe in democratizing AI, making smart, reasoning-capable models accessible to everyone. Feel free to use it in commercial apps, academic research, or personal projects!
---
## 📞 Contact & Community
* 📧 **Email:** [lamapicontact@gmail.com](mailto:lamapicontact@gmail.com)
* 🤗 **HuggingFace:** [Lamapi](https://huggingface.co/Lamapi)
* 💬 **Discord:** [Join the Lamapi Community](https://discord.gg/XgH4EpyPD2)
---
<div align="center" style="margin-top: 30px; padding: 20px; border-top: 1px solid #eaeaea;">
<p style="color: #666; font-size: 14px;">
<strong>Next2 0.8B</strong> — Küçük boyutlu, büyük akıllı. Türkiye'den dünyaya, sınır tanımayan yeni nesil yerel AI. 🌍
</p>
</div> |