Text Generation
Transformers
Safetensors
qwen3
turkish
türkiye
reasoning
ai
lamapi
gemma3
next
next-x1
open-source
32b
large-language-model
llm
transformer
artificial-intelligence
machine-learning
nlp
multilingual
instruction-tuned
chat
generative-ai
optimized
trl
sft
cognitive
analytical
enterprise
industrial
conversational
text-generation-inference
Update README.md
Browse files
README.md
CHANGED
|
@@ -146,6 +146,10 @@ Designed for high-demand enterprise environments, **Next 32B** delivers superior
|
|
| 146 |
|
| 147 |
**Note:** Due to the model size, we recommend using a GPU with at least 24GB VRAM (for 4-bit quantization) or 48GB+ (for 8-bit/FP16).
|
| 148 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 149 |
```python
|
| 150 |
from unsloth import FastLanguageModel
|
| 151 |
|
|
|
|
| 146 |
|
| 147 |
**Note:** Due to the model size, we recommend using a GPU with at least 24GB VRAM (for 4-bit quantization) or 48GB+ (for 8-bit/FP16).
|
| 148 |
|
| 149 |
+
```
|
| 150 |
+
!pip install unsloth
|
| 151 |
+
```
|
| 152 |
+
|
| 153 |
```python
|
| 154 |
from unsloth import FastLanguageModel
|
| 155 |
|