Update README.md
Browse files
README.md
CHANGED
|
@@ -6,18 +6,15 @@ tags:
|
|
| 6 |
- qwen3
|
| 7 |
base_model:
|
| 8 |
- Qwen/Qwen3-Coder-Next
|
| 9 |
-
library_name: transformers
|
| 10 |
license: apache-2.0
|
| 11 |
license_link: https://huggingface.co/Qwen/Qwen3-Coder-Next/blob/main/LICENSE
|
| 12 |
pipeline_tag: text-generation
|
| 13 |
---
|
| 14 |
<div>
|
| 15 |
<p style="margin-bottom: 0; margin-top: 0;">
|
| 16 |
-
<h1 style="margin-top: 0rem;">To Run Qwen3-Coder-Next locally - <a href="https://unsloth.ai/docs/models/qwen3-coder-next">Read our Guide
|
| 17 |
-
</p>
|
| 18 |
-
<p style="margin-bottom: 0; margin-top: 0;">
|
| 19 |
-
<h2 style="margin-top: 0rem;">Feb 4 update: llama.cpp fixed a bug that caused Qwen to loop and have poor outputs.<br>We updated GGUFs - please re-download and update llama.cpp for improved outputs.</h2>
|
| 20 |
</p>
|
|
|
|
| 21 |
<p style="margin-top: 0;margin-bottom: 0;">
|
| 22 |
<em><a href="https://unsloth.ai/docs/basics/unsloth-dynamic-2.0-ggufs">Unsloth Dynamic 2.0</a> achieves superior accuracy & outperforms other leading quants.</em>
|
| 23 |
</p>
|
|
@@ -32,11 +29,16 @@ pipeline_tag: text-generation
|
|
| 32 |
<img src="https://raw.githubusercontent.com/unslothai/unsloth/refs/heads/main/images/documentation%20green%20button.png" width="143">
|
| 33 |
</a>
|
| 34 |
</div>
|
| 35 |
-
<h2 style="margin-top: 0rem;">Qwen3-Coder-Next Usage Guidelines</h2>
|
| 36 |
</div>
|
| 37 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
- It is recommended to have >45GB unified memory or RAM/VRAM to run 4-bit quants.
|
| 39 |
- For best results, use any 2-bit XL quant or above (requires >30GB unified memory /combined RAM + VRAM).
|
|
|
|
| 40 |
- For complete detailed instructions (sampling parameters etc.), see our guide: [docs.unsloth.ai/models/qwen3-coder-next](https://unsloth.ai/docs/models/qwen3-coder-next)
|
| 41 |
|
| 42 |
---
|
|
|
|
| 6 |
- qwen3
|
| 7 |
base_model:
|
| 8 |
- Qwen/Qwen3-Coder-Next
|
|
|
|
| 9 |
license: apache-2.0
|
| 10 |
license_link: https://huggingface.co/Qwen/Qwen3-Coder-Next/blob/main/LICENSE
|
| 11 |
pipeline_tag: text-generation
|
| 12 |
---
|
| 13 |
<div>
|
| 14 |
<p style="margin-bottom: 0; margin-top: 0;">
|
| 15 |
+
<h1 style="margin-top: 0rem;">To Run Qwen3-Coder-Next locally - <a href="https://unsloth.ai/docs/models/qwen3-coder-next">Read our Guide!</a></h1>
|
|
|
|
|
|
|
|
|
|
| 16 |
</p>
|
| 17 |
+
|
| 18 |
<p style="margin-top: 0;margin-bottom: 0;">
|
| 19 |
<em><a href="https://unsloth.ai/docs/basics/unsloth-dynamic-2.0-ggufs">Unsloth Dynamic 2.0</a> achieves superior accuracy & outperforms other leading quants.</em>
|
| 20 |
</p>
|
|
|
|
| 29 |
<img src="https://raw.githubusercontent.com/unslothai/unsloth/refs/heads/main/images/documentation%20green%20button.png" width="143">
|
| 30 |
</a>
|
| 31 |
</div>
|
|
|
|
| 32 |
</div>
|
| 33 |
|
| 34 |
+
- **Feb 19 update**: Tool-calling should now be even better after llama.cpp fixes parsing.
|
| 35 |
+
- **Quantization benchmarks**: See third-party Aider, LiveCodeBench v6, MMLU Pro, GPQA [benchmarks for GGUFs here](https://unsloth.ai/docs/models/qwen3-coder-next#gguf-quantization-benchmarks).</h2>
|
| 36 |
+
- **Feb 4 update**: llama.cpp fixed a bug that caused Qwen to loop and have poor outputs.<br>We updated GGUFs - please re-download and update llama.cpp for improved outputs.</h2>
|
| 37 |
+
|
| 38 |
+
## Qwen3-Coder-Next Usage Guidelines
|
| 39 |
- It is recommended to have >45GB unified memory or RAM/VRAM to run 4-bit quants.
|
| 40 |
- For best results, use any 2-bit XL quant or above (requires >30GB unified memory /combined RAM + VRAM).
|
| 41 |
+
- See how to run the model via [Claude Code & OpenAI Codex](https://unsloth.ai/docs/models/qwen3-coder-next#claude-codex).
|
| 42 |
- For complete detailed instructions (sampling parameters etc.), see our guide: [docs.unsloth.ai/models/qwen3-coder-next](https://unsloth.ai/docs/models/qwen3-coder-next)
|
| 43 |
|
| 44 |
---
|