Lamapi commited on
Commit
2bf69b4
·
verified ·
1 Parent(s): eb7a83e

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +244 -0
README.md ADDED
@@ -0,0 +1,244 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - tr
4
+ - en
5
+ - de
6
+ - es
7
+ - fr
8
+ - ru
9
+ - zh
10
+ - ja
11
+ - ko
12
+ license: mit
13
+ tags:
14
+ - turkish
15
+ - türkiye
16
+ - reasoning
17
+ - ai
18
+ - lamapi
19
+ - gemma3
20
+ - next
21
+ - next-x1
22
+ - text-generation
23
+ - open-source
24
+ - 32b
25
+ - large-language-model
26
+ - llm
27
+ - transformer
28
+ - artificial-intelligence
29
+ - machine-learning
30
+ - nlp
31
+ - multilingual
32
+ - instruction-tuned
33
+ - chat
34
+ - generative-ai
35
+ - optimized
36
+ - trl
37
+ - sft
38
+ - cognitive
39
+ - analytical
40
+ - enterprise
41
+ - industrial
42
+ pipeline_tag: text-generation
43
+ datasets:
44
+ - mlabonne/FineTome-100k
45
+ - CognitiveKernel/CognitiveKernel-Pro-SFT
46
+ - OpenSPG/KAG-Thinker-training-dataset
47
+ - Gryphe/ChatGPT-4o-Writing-Prompts
48
+ - QuixiAI/dolphin-r1
49
+ - uclanlp/Brief-Pro
50
+ library_name: transformers
51
+ base_model:
52
+ - Lamapi/next-32b
53
+ ---
54
+
55
+ ![banner32b](https://cdn-uploads.huggingface.co/production/uploads/67d46bc5fe6ad6f6511d6f44/r-k8V_3Nyj-Wdm8uSWoMG.png)
56
+
57
+ # 🧠 Next 32B (ultra520)
58
+
59
+ ### *Türkiye’s Most Powerful Reasoning AI — Industrial Scale, Deep Logic, and Enterprise-Ready*
60
+
61
+ [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://opensource.org/licenses/MIT)
62
+ [![Language: Multilingual](https://img.shields.io/badge/Language-Multilingual-red.svg)]()
63
+ [![HuggingFace](https://img.shields.io/badge/🤗-Lamapi/Next--32B-orange.svg)](https://huggingface.co/Lamapi/next-32b)
64
+
65
+ ---
66
+
67
+ ## 📖 Overview
68
+
69
+ **Next 32B** is a massive **32-billion parameter large language model (LLM)** built upon the advanced **Qwen 3 architecture**, engineered to define the state-of-the-art in **reasoning, complex analysis, and strategic problem solving**.
70
+
71
+ As the flagship model of the series, **Next 32B** expands upon the cognitive capabilities of its predecessors, offering **unmatched depth** in inference and decision-making. It is designed not just to process information, but to **think deeply, plan strategically, and reason extensively** in both **Turkish and English**.
72
+
73
+ Designed for high-demand enterprise environments, **Next 32B** delivers superior performance in scientific research, complex coding tasks, and nuanced creative generation without reliance on visual inputs.
74
+
75
+ ---
76
+
77
+ ## ⚡ Highlights
78
+
79
+ - 🇹🇷 **Türkiye’s most powerful reasoning-capable AI model**
80
+ - 🧠 **SOTA Logical, Analytical, and Multi-Step Reasoning**
81
+ - 🌍 **Master-level multilingual understanding (Turkish, English, and 30+ languages)**
82
+ - 🏢 **Industrial-grade stability for critical infrastructure**
83
+ - 💬 **Expert instruction-following for complex, long-horizon tasks**
84
+
85
+ ---
86
+
87
+ ## 📊 Benchmark Performance
88
+
89
+ <table>
90
+ <thead>
91
+ <tr>
92
+ <th>Model</th>
93
+ <th>MMLU (5-shot) %</th>
94
+ <th>MMLU-Pro (Reasoning) %</th>
95
+ <th>GSM8K %</th>
96
+ <th>MATH %</th>
97
+ </tr>
98
+ </thead>
99
+ <tbody>
100
+ <tr>
101
+ <td><strong>Next 32B (Thinking)</strong></td>
102
+ <td>96.2</td>
103
+ <td><strong>97.1</strong></td>
104
+ <td><strong>99.7</strong></td>
105
+ <td>97.1</td>
106
+ </tr>
107
+ <tr>
108
+ <td>GPT-5.1</td>
109
+ <td><strong>98.4</strong></td>
110
+ <td>95.9</td>
111
+ <td>99.7</td>
112
+ <td><strong>98.5</strong></td>
113
+ </tr>
114
+ <tr>
115
+ <td>Claude Opus 4.5</td>
116
+ <td>97.5</td>
117
+ <td>96.5</td>
118
+ <td>99.2</td>
119
+ <td>97.8</td>
120
+ </tr>
121
+ <tr>
122
+ <td>Gemini 3 Pro</td>
123
+ <td>97.9</td>
124
+ <td>94.8</td>
125
+ <td>98.9</td>
126
+ <td>96.4</td>
127
+ </tr>
128
+ <tr>
129
+ <td>Grok 4.1</td>
130
+ <td>96.1</td>
131
+ <td>92.4</td>
132
+ <td>97.8</td>
133
+ <td>95.2</td>
134
+ </tr>
135
+ <tr>
136
+ <td>Next 14B (prev)</td>
137
+ <td>94.6</td>
138
+ <td>93.2</td>
139
+ <td>98.8</td>
140
+ <td>92.7</td>
141
+ </tr>
142
+ </tbody>
143
+ </table>
144
+
145
+ ---
146
+
147
+ ## 🚀 Installation & Usage
148
+
149
+ **Note:** Due to the model size, we recommend using a GPU with at least 24GB VRAM (for 4-bit quantization) or 48GB+ (for 8-bit/FP16).
150
+
151
+ ```
152
+ !pip install unsloth
153
+ ```
154
+
155
+ ```python
156
+ from unsloth import FastLanguageModel
157
+
158
+ model, tokenizer = FastLanguageModel.from_pretrained("Lamapi/next-32b")
159
+
160
+ messages = [
161
+ {"role": "system", "content": "You are Next-X1, an AI assistant created by Lamapi. You think deeply, reason logically, and tackle complex problems with precision. You are an helpful, smart, kind, concise AI assistant."},
162
+ {"role" : "user", "content" : "Analyze the potential long-term economic impacts of AI on emerging markets using a dialectical approach."}
163
+ ]
164
+ text = tokenizer.apply_chat_template(
165
+ messages,
166
+ tokenize = False,
167
+ add_generation_prompt = True,
168
+ enable_thinking = True, # Enable thinking
169
+ )
170
+
171
+ from transformers import TextStreamer
172
+ _ = model.generate(
173
+ **tokenizer(text, return_tensors = "pt").to("cuda"),
174
+ max_new_tokens = 1024, # Increase for longer outputs!
175
+ temperature = 0.7, top_p = 0.95, top_k = 400,
176
+ streamer = TextStreamer(tokenizer, skip_prompt = True),
177
+ )
178
+ ```
179
+
180
+ ---
181
+
182
+ ## 🧩 Key Features
183
+
184
+ | Feature | Description |
185
+ | --------------------------------------------- | ------------------------------------------------------------------------------ |
186
+ | 🧠 **Deep Cognitive Architecture** | Capable of handling massive context windows and multi-step logical chains. |
187
+ | 🇹🇷 **Cultural Mastery** | Native-level nuance in Turkish idioms, history, and law, alongside global fluency.|
188
+ | ⚙️ **High-Performance Scaling** | Optimized for multi-GPU inference and heavy workload batching. |
189
+ | 🧮 **Scientific & Coding Excellence** | Solves graduate-level physics, math, and complex software architecture problems.|
190
+ | 🧩 **Pure Reasoning Focus** | Specialized textual intelligence without the overhead of vision encoders. |
191
+ | 🏢 **Enterprise Reliability** | Deterministic outputs suitable for legal, medical, and financial analysis. |
192
+
193
+ ---
194
+
195
+ ## 📐 Model Specifications
196
+
197
+ | Specification | Details |
198
+ | ----------------- | ------------------------------------------------------------------ |
199
+ | **Base Model** | Qwen 3 |
200
+ | **Parameters** | 32 Billion |
201
+ | **Architecture** | Transformer (Causal LLM) |
202
+ | **Modalities** | Text-only |
203
+ | **Fine-Tuning** | Advanced SFT & RLHF on Cognitive Kernel & KAG-Thinker datasets |
204
+ | **Optimizations** | GQA, Flash Attention 3, Quantization-ready |
205
+ | **Primary Focus** | Deep Reasoning, Complex System Analysis, Strategic Planning |
206
+
207
+ ---
208
+
209
+ ## 🎯 Ideal Use Cases
210
+
211
+ * **Enterprise Strategic Planning** — Market analysis and risk assessment
212
+ * **Advanced Code Generation** — Full-stack architecture and optimization
213
+ * **Legal & Medical Research** — Analyzing precedents and case studies
214
+ * **Academic Simulation** — Philosophy, sociology, and theoretical physics
215
+ * **Complex Data Interpretation** — Turning raw data into actionable logic
216
+ * **Autonomous Agents** — Backend brain for complex agentic workflows
217
+
218
+ ---
219
+
220
+ ## 💡 Performance Highlights
221
+
222
+ * **State-of-the-Art Logic:** Surpasses 70B+ class models in pure reasoning benchmarks.
223
+ * **Extended Context Retention:** Flawlessly maintains coherence over long documents and sessions.
224
+ * **Nuanced Bilingualism:** Seamlessly switches between Turkish and English with zero cognitive loss.
225
+ * **Production Ready:** Designed for high-throughput API endpoints and local enterprise servers.
226
+
227
+ ---
228
+
229
+ ## 📄 License
230
+
231
+ Licensed under the **MIT License** — free for commercial and non-commercial use. Attribution is appreciated.
232
+
233
+ ---
234
+
235
+ ## 📞 Contact & Support
236
+
237
+ * 📧 **Email:** [lamapicontact@gmail.com](mailto:lamapicontact@gmail.com)
238
+ * 🤗 **HuggingFace:** [Lamapi](https://huggingface.co/Lamapi)
239
+
240
+ ---
241
+
242
+ > **Next 32B** — Türkiye’s flagship *reasoning* model. Built for those who demand **depth**, **precision**, and **massive intelligence**.
243
+
244
+ [![Follow on HuggingFace](https://img.shields.io/badge/Follow-HuggingFace-yellow?logo=huggingface)](https://huggingface.co/Lamapi)