Lamapi commited on
Commit
1bf7862
·
verified ·
1 Parent(s): a2fb0ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +109 -38
README.md CHANGED
@@ -1,4 +1,3 @@
1
-
2
  ---
3
  language:
4
  - tr
@@ -14,13 +13,15 @@ license: mit
14
  tags:
15
  - turkish
16
  - türkiye
 
17
  - ai
18
  - lamapi
 
19
  - next
20
  - next-x1
21
  - text-generation
22
  - open-source
23
- - 70b
24
  - large-language-model
25
  - llm
26
  - transformer
@@ -34,58 +35,116 @@ tags:
34
  - optimized
35
  - trl
36
  - sft
 
 
37
  - enterprise
38
  - industrial
39
  pipeline_tag: text-generation
40
  datasets:
41
  - mlabonne/FineTome-100k
 
 
42
  - Gryphe/ChatGPT-4o-Writing-Prompts
 
43
  - uclanlp/Brief-Pro
44
  library_name: transformers
45
  ---
46
 
47
  ![70b](https://cdn-uploads.huggingface.co/production/uploads/67d46bc5fe6ad6f6511d6f44/017hoTVIfgFInU5ZUVQjv.png)
48
 
49
- # 🚀 Next 70B (ultra1295)
50
 
51
- ### *Türkiye’s Most Powerful AI — Industrial Scale, High Precision, and Enterprise-Ready*
52
 
53
  [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://opensource.org/licenses/MIT)
54
  [![Language: Multilingual](https://img.shields.io/badge/Language-Multilingual-red.svg)]()
55
- [![HuggingFace](https://img.shields.io/badge/🤗-Lamapi/Next--70B-orange.svg)](https://huggingface.co/Lamapi/next-70b)
56
 
57
  ---
58
 
59
  ## 📖 Overview
60
 
61
- **Next 70B** is a state-of-the-art **70-billion parameter large language model (LLM)** engineered for maximum accuracy, versatility, and instruction following. Built upon an optimized transformer architecture, it delivers **SOTA performance** across coding, mathematics, and creative writing tasks.
 
 
62
 
63
- As the flagship model of the series, **Next 70B** is designed to handle the most demanding enterprise workloads. It excels at nuanced language understanding in **Turkish and English**, complex data processing, and generating production-grade code, making it a superior alternative to proprietary models.
64
 
65
  ---
66
 
67
  ## ⚡ Highlights
68
 
69
- - 🇹🇷 **Türkiye’s most powerful open-weights AI model**
70
- - 🏆 **Top-tier Performance:** Beats GPT-5.1 in MATH (99.0%) and achieves near-perfect GSM8K scores.
71
  - 🌍 **Master-level multilingual understanding (Turkish, English, and 30+ languages)**
72
- - 💻 **Coding Specialist:** Exceptional Python and JavaScript generation capabilities (HumanEval 97.8%).
73
  - 🏢 **Industrial-grade stability for critical infrastructure**
74
- - 📝 **Precise Instruction Following:** High IFEval score (95.0) ensures strict adherence to formatting and constraints.
75
 
76
  ---
77
 
78
  ## 📊 Benchmark Performance
79
 
80
- **Next 70B** demonstrates world-class performance, surpassing major competitors in key academic and industrial benchmarks.
81
-
82
- ![WhatsApp Image 2025-11-29 at 15.37.04_764ee845](https://cdn-uploads.huggingface.co/production/uploads/67d46bc5fe6ad6f6511d6f44/OEZUOh78lc0q0vJm3dlVh.jpeg)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
83
 
84
  ---
85
 
86
  ## 🚀 Installation & Usage
87
 
88
- **Note:** We recommend using a multi-GPU setup (e.g., 2x A100 80GB) for full precision or 48GB+ VRAM for 4-bit quantization.
89
 
90
  ```
91
  !pip install unsloth
@@ -94,22 +153,23 @@ As the flagship model of the series, **Next 70B** is designed to handle the most
94
  ```python
95
  from unsloth import FastLanguageModel
96
 
97
- model, tokenizer = FastLanguageModel.from_pretrained("Lamapi/next-70b")
98
 
99
  messages = [
100
- {"role": "system", "content": "You are Next-X1, a helpful, smart, and precise AI assistant created by Lamapi."},
101
- {"role" : "user", "content" : "Write a Python script to optimize a neural network using PyTorch."}
102
  ]
103
  text = tokenizer.apply_chat_template(
104
  messages,
105
  tokenize = False,
106
- add_generation_prompt = True
 
107
  )
108
 
109
  from transformers import TextStreamer
110
  _ = model.generate(
111
  **tokenizer(text, return_tensors = "pt").to("cuda"),
112
- max_new_tokens = 2048,
113
  temperature = 0.7, top_p = 0.95, top_k = 400,
114
  streamer = TextStreamer(tokenizer, skip_prompt = True),
115
  )
@@ -121,12 +181,12 @@ _ = model.generate(
121
 
122
  | Feature | Description |
123
  | --------------------------------------------- | ------------------------------------------------------------------------------ |
124
- | 📚 **Massive Knowledge Base** | Trained on a diverse, high-quality dataset covering science, history, and law. |
125
- | 🇹🇷 **Cultural Mastery** | Native-level nuance in Turkish idioms and professional terminology. |
126
- | ⚙️ **High-Performance Scaling** | Optimized for high-throughput inference and low latency. |
127
- | 🧮 **Scientific & Coding Excellence** | **99.0% MATH** score. Solves complex engineering and algorithmic problems. |
128
- | 🎯 **Precision Focused** | Designed for tasks requiring strict output formats and high factual accuracy. |
129
- | 🏢 **Enterprise Reliability** | Consistent and safe outputs suitable for commercial applications. |
130
 
131
  ---
132
 
@@ -134,24 +194,33 @@ _ = model.generate(
134
 
135
  | Specification | Details |
136
  | ----------------- | ------------------------------------------------------------------ |
137
- | **Base Model** | Llama |
138
- | **Parameters** | 70 Billion |
139
  | **Architecture** | Transformer (Causal LLM) |
140
  | **Modalities** | Text-only |
141
- | **Fine-Tuning** | SFT & DPO on high-quality instruct datasets |
142
  | **Optimizations** | GQA, Flash Attention 3, Quantization-ready |
143
- | **Primary Focus** | General Purpose Assistant, Math, Multilingual Chat |
144
 
145
  ---
146
 
147
  ## 🎯 Ideal Use Cases
148
 
149
- * **Enterprise Assistants** — Customer support and internal knowledge management
150
- * **Advanced Code Generation** — Full-stack development and debugging
151
- * **Content Creation** — High-quality marketing copy, emails, and reports
152
- * **Translation & Localization** — Highly accurate translation between Turkish/English
153
- * **Data Extraction** — Structuring unstructured data into JSON/SQL
154
- * **Academic Assistance** — Solving math problems and summarizing research papers
 
 
 
 
 
 
 
 
 
155
 
156
  ---
157
 
@@ -168,7 +237,9 @@ Licensed under the **MIT License** — free for commercial and non-commercial us
168
 
169
  ---
170
 
171
- > **Next 70B** — Türkiye’s flagship AI model. Built for those who demand **accuracy**, **speed**, and **scale**.
172
 
173
  [![Follow on HuggingFace](https://img.shields.io/badge/Follow-HuggingFace-yellow?logo=huggingface)](https://huggingface.co/Lamapi)
174
- ```
 
 
 
 
1
  ---
2
  language:
3
  - tr
 
13
  tags:
14
  - turkish
15
  - türkiye
16
+ - reasoning
17
  - ai
18
  - lamapi
19
+ - gemma3
20
  - next
21
  - next-x1
22
  - text-generation
23
  - open-source
24
+ - 32b
25
  - large-language-model
26
  - llm
27
  - transformer
 
35
  - optimized
36
  - trl
37
  - sft
38
+ - cognitive
39
+ - analytical
40
  - enterprise
41
  - industrial
42
  pipeline_tag: text-generation
43
  datasets:
44
  - mlabonne/FineTome-100k
45
+ - CognitiveKernel/CognitiveKernel-Pro-SFT
46
+ - OpenSPG/KAG-Thinker-training-dataset
47
  - Gryphe/ChatGPT-4o-Writing-Prompts
48
+ - QuixiAI/dolphin-r1
49
  - uclanlp/Brief-Pro
50
  library_name: transformers
51
  ---
52
 
53
  ![70b](https://cdn-uploads.huggingface.co/production/uploads/67d46bc5fe6ad6f6511d6f44/017hoTVIfgFInU5ZUVQjv.png)
54
 
55
+ # 🧠 Next 32B (ultra530)
56
 
57
+ ### *Türkiye’s Most Powerful AI — Industrial Scale, Deep Logic, and Enterprise-Ready*
58
 
59
  [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://opensource.org/licenses/MIT)
60
  [![Language: Multilingual](https://img.shields.io/badge/Language-Multilingual-red.svg)]()
61
+ [![HuggingFace](https://img.shields.io/badge/🤗-Lamapi/Next--32B-orange.svg)](https://huggingface.co/Lamapi/next-32b)
62
 
63
  ---
64
 
65
  ## 📖 Overview
66
 
67
+ **Next 32B** is a massive **32-billion parameter large language model (LLM)** built upon the advanced **Qwen 3 architecture**, engineered to define the state-of-the-art in **reasoning, complex analysis, and strategic problem solving**.
68
+
69
+ As the flagship model of the series, **Next 32B** expands upon the cognitive capabilities of its predecessors, offering **unmatched depth** in inference and decision-making. It is designed not just to process information, but to **think deeply, plan strategically, and reason extensively** in both **Turkish and English**.
70
 
71
+ Designed for high-demand enterprise environments, **Next 32B** delivers superior performance in scientific research, complex coding tasks, and nuanced creative generation without reliance on visual inputs.
72
 
73
  ---
74
 
75
  ## ⚡ Highlights
76
 
77
+ - 🇹🇷 **Türkiye’s most powerful reasoning-capable AI model**
78
+ - 🧠 **SOTA Logical, Analytical, and Multi-Step Reasoning**
79
  - 🌍 **Master-level multilingual understanding (Turkish, English, and 30+ languages)**
 
80
  - 🏢 **Industrial-grade stability for critical infrastructure**
81
+ - 💬 **Expert instruction-following for complex, long-horizon tasks**
82
 
83
  ---
84
 
85
  ## 📊 Benchmark Performance
86
 
87
+ <table>
88
+ <thead>
89
+ <tr>
90
+ <th>Model</th>
91
+ <th>MMLU (5-shot) %</th>
92
+ <th>MMLU-Pro (Reasoning) %</th>
93
+ <th>GSM8K %</th>
94
+ <th>MATH %</th>
95
+ </tr>
96
+ </thead>
97
+ <tbody>
98
+ <tr>
99
+ <td><strong>Next 32B (Thinking)</strong></td>
100
+ <td>96.2</td>
101
+ <td><strong>97.1</strong></td>
102
+ <td><strong>99.7</strong></td>
103
+ <td>97.1</td>
104
+ </tr>
105
+ <tr>
106
+ <td>GPT-5.1</td>
107
+ <td><strong>98.4</strong></td>
108
+ <td>95.9</td>
109
+ <td>99.7</td>
110
+ <td><strong>98.5</strong></td>
111
+ </tr>
112
+ <tr>
113
+ <td>Claude Opus 4.5</td>
114
+ <td>97.5</td>
115
+ <td>96.5</td>
116
+ <td>99.2</td>
117
+ <td>97.8</td>
118
+ </tr>
119
+ <tr>
120
+ <td>Gemini 3 Pro</td>
121
+ <td>97.9</td>
122
+ <td>94.8</td>
123
+ <td>98.9</td>
124
+ <td>96.4</td>
125
+ </tr>
126
+ <tr>
127
+ <td>Grok 4.1</td>
128
+ <td>96.1</td>
129
+ <td>92.4</td>
130
+ <td>97.8</td>
131
+ <td>95.2</td>
132
+ </tr>
133
+ <tr>
134
+ <td>Next 14B (prev)</td>
135
+ <td>94.6</td>
136
+ <td>93.2</td>
137
+ <td>98.8</td>
138
+ <td>92.7</td>
139
+ </tr>
140
+ </tbody>
141
+ </table>
142
 
143
  ---
144
 
145
  ## 🚀 Installation & Usage
146
 
147
+ **Note:** Due to the model size, we recommend using a GPU with at least 24GB VRAM (for 4-bit quantization) or 48GB+ (for 8-bit/FP16).
148
 
149
  ```
150
  !pip install unsloth
 
153
  ```python
154
  from unsloth import FastLanguageModel
155
 
156
+ model, tokenizer = FastLanguageModel.from_pretrained("Lamapi/next-32b")
157
 
158
  messages = [
159
+ {"role": "system", "content": "You are Next-X1, an AI assistant created by Lamapi. You think deeply, reason logically, and tackle complex problems with precision. You are an helpful, smart, kind, concise AI assistant."},
160
+ {"role" : "user", "content" : "Analyze the potential long-term economic impacts of AI on emerging markets using a dialectical approach."}
161
  ]
162
  text = tokenizer.apply_chat_template(
163
  messages,
164
  tokenize = False,
165
+ add_generation_prompt = True,
166
+ enable_thinking = True, # Enable thinking
167
  )
168
 
169
  from transformers import TextStreamer
170
  _ = model.generate(
171
  **tokenizer(text, return_tensors = "pt").to("cuda"),
172
+ max_new_tokens = 1024, # Increase for longer outputs!
173
  temperature = 0.7, top_p = 0.95, top_k = 400,
174
  streamer = TextStreamer(tokenizer, skip_prompt = True),
175
  )
 
181
 
182
  | Feature | Description |
183
  | --------------------------------------------- | ------------------------------------------------------------------------------ |
184
+ | 🧠 **Deep Cognitive Architecture** | Capable of handling massive context windows and multi-step logical chains. |
185
+ | 🇹🇷 **Cultural Mastery** | Native-level nuance in Turkish idioms, history, and law, alongside global fluency.|
186
+ | ⚙️ **High-Performance Scaling** | Optimized for multi-GPU inference and heavy workload batching. |
187
+ | 🧮 **Scientific & Coding Excellence** | Solves graduate-level physics, math, and complex software architecture problems.|
188
+ | 🧩 **Pure Reasoning Focus** | Specialized textual intelligence without the overhead of vision encoders. |
189
+ | 🏢 **Enterprise Reliability** | Deterministic outputs suitable for legal, medical, and financial analysis. |
190
 
191
  ---
192
 
 
194
 
195
  | Specification | Details |
196
  | ----------------- | ------------------------------------------------------------------ |
197
+ | **Base Model** | Qwen 3 |
198
+ | **Parameters** | 32 Billion |
199
  | **Architecture** | Transformer (Causal LLM) |
200
  | **Modalities** | Text-only |
201
+ | **Fine-Tuning** | Advanced SFT & RLHF on Cognitive Kernel & KAG-Thinker datasets |
202
  | **Optimizations** | GQA, Flash Attention 3, Quantization-ready |
203
+ | **Primary Focus** | Deep Reasoning, Complex System Analysis, Strategic Planning |
204
 
205
  ---
206
 
207
  ## 🎯 Ideal Use Cases
208
 
209
+ * **Enterprise Strategic Planning** — Market analysis and risk assessment
210
+ * **Advanced Code Generation** — Full-stack architecture and optimization
211
+ * **Legal & Medical Research** — Analyzing precedents and case studies
212
+ * **Academic Simulation** — Philosophy, sociology, and theoretical physics
213
+ * **Complex Data Interpretation** — Turning raw data into actionable logic
214
+ * **Autonomous Agents** — Backend brain for complex agentic workflows
215
+
216
+ ---
217
+
218
+ ## 💡 Performance Highlights
219
+
220
+ * **State-of-the-Art Logic:** Surpasses 70B+ class models in pure reasoning benchmarks.
221
+ * **Extended Context Retention:** Flawlessly maintains coherence over long documents and sessions.
222
+ * **Nuanced Bilingualism:** Seamlessly switches between Turkish and English with zero cognitive loss.
223
+ * **Production Ready:** Designed for high-throughput API endpoints and local enterprise servers.
224
 
225
  ---
226
 
 
237
 
238
  ---
239
 
240
+ > **Next 32B** — Türkiye’s flagship *reasoning* model. Built for those who demand **depth**, **precision**, and **massive intelligence**.
241
 
242
  [![Follow on HuggingFace](https://img.shields.io/badge/Follow-HuggingFace-yellow?logo=huggingface)](https://huggingface.co/Lamapi)
243
+
244
+
245
+ bunu 70b için ve, şu benchmark için düzenle: