--- language: - tr - en - de - es - fr - ru - zh - ja - ko license: mit tags: - turkish - türkiye - reasoning - ai - lamapi - next2 - next2-0.8b - qwen3.5 - text-generation - open-source - 0.8b - edge-ai - large-language-model - llm - transformer - artificial-intelligence - nlp - instruction-tuned - chat - thinking-mode - efficient - sft pipeline_tag: image-text-to-text datasets: - mlabonne/FineTome-100k - CognitiveKernel/CognitiveKernel-Pro-SFT - OpenSPG/KAG-Thinker-training-dataset - Gryphe/ChatGPT-4o-Writing-Prompts library_name: transformers ---
| Model | MMLU (5-shot) | IFEval | GSM8K (Math) | Context Limit |
|---|---|---|---|---|
| 🚀 Next2 0.8B (Thinking) | 52.1% | 55.8% | 67.4% | 32K+ |
| Base Qwen3.5-0.8B | 48.5% | 52.1% | 62.2% | 262K |
| Llama-3.2-1B | 49.3% | 50.2% | 60.5% | 128K |
* Scores represent generalized task performance. Next2 0.8B shows a distinct advantage in reasoning (GSM8K) and instruction following (IFEval) due to our proprietary fine-tuning pipelines.
--- ## 🚀 Quickstart & Usage You can easily run **Next2 0.8B** on almost any machine with Python installed. Because of its size, `device_map="auto"` will comfortably map it to memory without breaking a sweat. ```python from transformers import AutoTokenizer, AutoModelForCausalLM, AutoProcessor from PIL import Image import torch model_id = "thelamapi/next2-0.8b" model = AutoModelForCausalLM.from_pretrained(model_id) processor = AutoProcessor.from_pretrained(model_id) # For vision. tokenizer = AutoTokenizer.from_pretrained(model_id) # Create a message in chat format messages = [ {"role": "system","content": [{"type": "text", "text": "You are Next2, a smart and concise AI assistant trained by Lamapi. Always respond in the user's language. Proudly made in Turkey."}]}, { "role": "user","content": [ {"type": "text", "text": "Write a highly optimized Rust function to calculate the Fibonacci sequence using memoization"} ] } ] # Prepare input with Tokenizer prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=False) inputs = processor(text=prompt, return_tensors="pt") # Remove 'mm_token_type_ids' if it's not needed for text-only generation if "mm_token_type_ids" in inputs: del inputs["mm_token_type_ids"] # Output from the model output = model.generate(**inputs, do_sample=True, temperature=0.7, max_new_tokens=128) print(tokenizer.decode(output[0], skip_special_tokens=True)) ``` --- ## 🧩 Model Specifications | Feature | Details | | :--- | :--- | | **Base Architecture** | Qwen 3.5 (Transformer with Gated Delta Networks) | | **Parameter Count** | 0.8 Billion (800M) | | **Primary Focus** | Edge Inference, Reasoning (CoT), Turkish/English Bilingual | | **Optimizations** | Multi-Token Prediction (MTP) Support, Flash Attention ready | | **Hardware Reqs** | Ultra-lightweight (Can run on 2GB RAM / Edge GPUs) | | **Format** | FP16 natively, Quantization (GGUF/AWQ) recommended for mobile | --- ## 🎯 Ideal Use Cases Since it is compact yet surprisingly capable, Next2 0.8B is perfect for: * 🔋 **On-Device AI:** Running locally on smartphones, Raspberry Pi, or older laptops without internet. * 🤖 **NPC & Gaming AI:** Fast, low-latency dialogue generation for video games. * 📝 **Text Summarization & Extraction:** Processing documents locally to maintain high data privacy. * 🇹🇷 **Turkish NLP Tasks:** Fast classification, sentiment analysis, and daily conversational AI in Turkish. --- ## 📄 License & Open Source Licensed under the **MIT License**. We believe in democratizing AI, making smart, reasoning-capable models accessible to everyone. Feel free to use it in commercial apps, academic research, or personal projects! --- ## 📞 Contact & Community * 📧 **Email:** [lamapicontact@gmail.com](mailto:lamapicontact@gmail.com) * 🤗 **HuggingFace:** [Lamapi](https://huggingface.co/Lamapi) * 💬 **Discord:** [Join the Lamapi Community](https://discord.gg/XgH4EpyPD2) ---Next2 0.8B — Küçük boyutlu, büyük akıllı. Türkiye'den dünyaya, sınır tanımayan yeni nesil yerel AI. 🌍