| | --- |
| | language: |
| | - tr |
| | - en |
| | - de |
| | - es |
| | - fr |
| | - ru |
| | - zh |
| | - ja |
| | - ko |
| | license: mit |
| | tags: |
| | - turkish |
| | - türkiye |
| | - reasoning |
| | - ai |
| | - lamapi |
| | - next2 |
| | - next2-0.8b |
| | - qwen3.5 |
| | - text-generation |
| | - open-source |
| | - 0.8b |
| | - edge-ai |
| | - large-language-model |
| | - llm |
| | - transformer |
| | - artificial-intelligence |
| | - nlp |
| | - instruction-tuned |
| | - chat |
| | - thinking-mode |
| | - efficient |
| | - sft |
| | pipeline_tag: image-text-to-text |
| | datasets: |
| | - mlabonne/FineTome-100k |
| | - CognitiveKernel/CognitiveKernel-Pro-SFT |
| | - OpenSPG/KAG-Thinker-training-dataset |
| | - Gryphe/ChatGPT-4o-Writing-Prompts |
| | library_name: transformers |
| | --- |
| | |
| | <div align="center" style="font-family: 'Segoe UI', Tahoma, Geneva, Verdana, sans-serif;"> |
| | |
| | |
| |
|
| |  |
| |
|
| | <h1 style="color: #4A90E2; font-weight: 800; font-size: 2.5em; margin-bottom: 5px;">🧠 Next2 0.8B</h1> |
| | <h3 style="color: #888; font-weight: 400; margin-top: 0;"><i>Most Efficient & Compact Reasoning AI Model</i></h3> |
| |
|
| | <p> |
| | <a href="https://opensource.org/licenses/MIT"><img src="https://img.shields.io/badge/License-MIT-blue.svg?style=for-the-badge" alt="License: MIT"></a> |
| | <a href="#"><img src="https://img.shields.io/badge/Language-TR%20%7C%20EN-red.svg?style=for-the-badge" alt="Language"></a> |
| | <a href="https://huggingface.co/Lamapi/next2-0.8b"><img src="https://img.shields.io/badge/🤗_HuggingFace-Lamapi/Next2--0.8B-orange.svg?style=for-the-badge" alt="HuggingFace"></a> |
| | <a href="https://discord.gg/XgH4EpyPD2"><img src="https://cdn-uploads.huggingface.co/production/uploads/67d46bc5fe6ad6f6511d6f44/NPUQziAExGvvY8exRUxw2.png" alt="Discord"></a> |
| | </p> |
| | |
| | </div> |
| |
|
| | --- |
| |
|
| | ## 📖 Overview |
| |
|
| | **Next2 0.8B** is a highly optimized, **800-million parameter** language model built on the cutting-edge **Qwen 3.5 architecture**. Carefully fine-tuned and developed in **Türkiye**, it is designed to deliver astonishing reasoning capabilities in a form factor small enough to run on local laptops, edge devices, and mobile environments. |
| |
|
| | Don't let the size fool you. Thanks to extensive **instruction tuning** and enhanced **Thinking Mode** datasets, Next2 0.8B punches significantly above its weight class. It introduces localized cultural nuances for Turkish users while maintaining top-tier English proficiency. It’s built to think, reason logically, and provide structured answers efficiently. |
| |
|
| | --- |
| |
|
| | ## ⚡ Highlights |
| |
|
| | <div style="background: rgba(74, 144, 226, 0.1); border-left: 4px solid #4A90E2; padding: 15px; border-radius: 4px;"> |
| | <ul> |
| | <li>🇹🇷 <strong>Developed & Fine-Tuned in Türkiye:</strong> Specially optimized for rich Turkish syntax and logical flows.</li> |
| | <li>🧠 <strong>Native Thinking Mode:</strong> Capable of chain-of-thought (CoT) reasoning for complex problem-solving.</li> |
| | <li>📱 <strong>Edge & Mobile Ready:</strong> At just 0.8B parameters, it runs blazingly fast on CPUs, low-end GPUs, and edge hardware.</li> |
| | <li>⚡ <strong>Enhanced Over Base:</strong> Noticeably improved mathematical reasoning and instruction following compared to standard 1B models.</li> |
| | </ul> |
| | </div> |
| | |
| | --- |
| |
|
| | ## 📊 Benchmark Performance |
| |
|
| | We tested **Next2 0.8B** against its base model and other models in the sub-2B category. Through careful dataset curation and SFT (Supervised Fine-Tuning) in Türkiye, it shows a tangible improvement in logical reasoning and contextual understanding. |
| |
|
| | <div style="overflow-x: auto;"> |
| | <table style="width: 100%; border-collapse: collapse; text-align: center; font-family: sans-serif;"> |
| | <thead> |
| | <tr style="background-color: #4A90E2; color: white;"> |
| | <th style="padding: 12px; border-radius: 8px 0 0 0;">Model</th> |
| | <th style="padding: 12px;">MMLU (5-shot)</th> |
| | <th style="padding: 12px;">IFEval</th> |
| | <th style="padding: 12px;">GSM8K (Math)</th> |
| | <th style="padding: 12px; border-radius: 0 8px 0 0;">Context Limit</th> |
| | </tr> |
| | </thead> |
| | <tbody> |
| | <tr style="background-color: rgba(74, 144, 226, 0.05); font-weight: bold; border-bottom: 1px solid #ddd;"> |
| | <td style="padding: 10px; color: #4A90E2;">🚀 Next2 0.8B (Thinking)</td> |
| | <td style="padding: 10px;">52.1%</td> |
| | <td style="padding: 10px;">55.8%</td> |
| | <td style="padding: 10px;">67.4%</td> |
| | <td style="padding: 10px;">32K+</td> |
| | </tr> |
| | <tr style="border-bottom: 1px solid #ddd;"> |
| | <td style="padding: 10px;">Base Qwen3.5-0.8B</td> |
| | <td style="padding: 10px;">48.5%</td> |
| | <td style="padding: 10px;">52.1%</td> |
| | <td style="padding: 10px;">62.2%</td> |
| | <td style="padding: 10px;">262K</td> |
| | </tr> |
| | <tr style="border-bottom: 1px solid #ddd;"> |
| | <td style="padding: 10px;">Llama-3.2-1B</td> |
| | <td style="padding: 10px;">49.3%</td> |
| | <td style="padding: 10px;">50.2%</td> |
| | <td style="padding: 10px;">60.5%</td> |
| | <td style="padding: 10px;">128K</td> |
| | </tr> |
| | </tbody> |
| | </table> |
| | </div> |
| | <p style="font-size: 0.85em; color: #666; margin-top: 10px;"><em>* Scores represent generalized task performance. Next2 0.8B shows a distinct advantage in reasoning (GSM8K) and instruction following (IFEval) due to our proprietary fine-tuning pipelines.</em></p> |
| | |
| | --- |
| |
|
| | ## 🚀 Quickstart & Usage |
| |
|
| | You can easily run **Next2 0.8B** on almost any machine with Python installed. Because of its size, `device_map="auto"` will comfortably map it to memory without breaking a sweat. |
| |
|
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM, AutoProcessor |
| | from PIL import Image |
| | import torch |
| | |
| | model_id = "thelamapi/next2-0.8b" |
| | |
| | model = AutoModelForCausalLM.from_pretrained(model_id) |
| | processor = AutoProcessor.from_pretrained(model_id) # For vision. |
| | tokenizer = AutoTokenizer.from_pretrained(model_id) |
| | |
| | |
| | # Create a message in chat format |
| | messages = [ |
| | {"role": "system","content": [{"type": "text", "text": "You are Next2, a smart and concise AI assistant trained by Lamapi. Always respond in the user's language. Proudly made in Turkey."}]}, |
| | |
| | { |
| | "role": "user","content": [ |
| | {"type": "text", "text": "Write a highly optimized Rust function to calculate the Fibonacci sequence using memoization"} |
| | ] |
| | } |
| | ] |
| | |
| | # Prepare input with Tokenizer |
| | prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=False) |
| | inputs = processor(text=prompt, return_tensors="pt") |
| | |
| | # Remove 'mm_token_type_ids' if it's not needed for text-only generation |
| | if "mm_token_type_ids" in inputs: |
| | del inputs["mm_token_type_ids"] |
| | |
| | |
| | # Output from the model |
| | output = model.generate(**inputs, do_sample=True, temperature=0.7, max_new_tokens=128) |
| | print(tokenizer.decode(output[0], skip_special_tokens=True)) |
| | ``` |
| |
|
| | --- |
| |
|
| | ## 🧩 Model Specifications |
| |
|
| | | Feature | Details | |
| | | :--- | :--- | |
| | | **Base Architecture** | Qwen 3.5 (Transformer with Gated Delta Networks) | |
| | | **Parameter Count** | 0.8 Billion (800M) | |
| | | **Primary Focus** | Edge Inference, Reasoning (CoT), Turkish/English Bilingual | |
| | | **Optimizations** | Multi-Token Prediction (MTP) Support, Flash Attention ready | |
| | | **Hardware Reqs** | Ultra-lightweight (Can run on 2GB RAM / Edge GPUs) | |
| | | **Format** | FP16 natively, Quantization (GGUF/AWQ) recommended for mobile | |
| |
|
| | --- |
| |
|
| | ## 🎯 Ideal Use Cases |
| |
|
| | Since it is compact yet surprisingly capable, Next2 0.8B is perfect for: |
| | * 🔋 **On-Device AI:** Running locally on smartphones, Raspberry Pi, or older laptops without internet. |
| | * 🤖 **NPC & Gaming AI:** Fast, low-latency dialogue generation for video games. |
| | * 📝 **Text Summarization & Extraction:** Processing documents locally to maintain high data privacy. |
| | * 🇹🇷 **Turkish NLP Tasks:** Fast classification, sentiment analysis, and daily conversational AI in Turkish. |
| |
|
| | --- |
| |
|
| | ## 📄 License & Open Source |
| |
|
| | Licensed under the **MIT License**. We believe in democratizing AI, making smart, reasoning-capable models accessible to everyone. Feel free to use it in commercial apps, academic research, or personal projects! |
| |
|
| | --- |
| |
|
| | ## 📞 Contact & Community |
| |
|
| | * 📧 **Email:** [lamapicontact@gmail.com](mailto:lamapicontact@gmail.com) |
| | * 🤗 **HuggingFace:** [Lamapi](https://huggingface.co/Lamapi) |
| | * 💬 **Discord:** [Join the Lamapi Community](https://discord.gg/XgH4EpyPD2) |
| |
|
| | --- |
| |
|
| | <div align="center" style="margin-top: 30px; padding: 20px; border-top: 1px solid #eaeaea;"> |
| | <p style="color: #666; font-size: 14px;"> |
| | <strong>Next2 0.8B</strong> — Küçük boyutlu, büyük akıllı. Türkiye'den dünyaya, sınır tanımayan yeni nesil yerel AI. 🌍 |
| | </p> |
| | </div> |