danielhanchen commited on
Commit
8df0b78
·
verified ·
1 Parent(s): 881901f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -6
README.md CHANGED
@@ -6,18 +6,15 @@ tags:
6
  - qwen3
7
  base_model:
8
  - Qwen/Qwen3-Coder-Next
9
- library_name: transformers
10
  license: apache-2.0
11
  license_link: https://huggingface.co/Qwen/Qwen3-Coder-Next/blob/main/LICENSE
12
  pipeline_tag: text-generation
13
  ---
14
  <div>
15
  <p style="margin-bottom: 0; margin-top: 0;">
16
- <h1 style="margin-top: 0rem;">To Run Qwen3-Coder-Next locally - <a href="https://unsloth.ai/docs/models/qwen3-coder-next">Read our Guide.</a></h1>
17
- </p>
18
- <p style="margin-bottom: 0; margin-top: 0;">
19
- <h2 style="margin-top: 0rem;">Feb 4 update: llama.cpp fixed a bug that caused Qwen to loop and have poor outputs.<br>We updated GGUFs - please re-download and update llama.cpp for improved outputs.</h2>
20
  </p>
 
21
  <p style="margin-top: 0;margin-bottom: 0;">
22
  <em><a href="https://unsloth.ai/docs/basics/unsloth-dynamic-2.0-ggufs">Unsloth Dynamic 2.0</a> achieves superior accuracy & outperforms other leading quants.</em>
23
  </p>
@@ -32,11 +29,16 @@ pipeline_tag: text-generation
32
  <img src="https://raw.githubusercontent.com/unslothai/unsloth/refs/heads/main/images/documentation%20green%20button.png" width="143">
33
  </a>
34
  </div>
35
- <h2 style="margin-top: 0rem;">Qwen3-Coder-Next Usage Guidelines</h2>
36
  </div>
37
 
 
 
 
 
 
38
  - It is recommended to have >45GB unified memory or RAM/VRAM to run 4-bit quants.
39
  - For best results, use any 2-bit XL quant or above (requires >30GB unified memory /combined RAM + VRAM).
 
40
  - For complete detailed instructions (sampling parameters etc.), see our guide: [docs.unsloth.ai/models/qwen3-coder-next](https://unsloth.ai/docs/models/qwen3-coder-next)
41
 
42
  ---
 
6
  - qwen3
7
  base_model:
8
  - Qwen/Qwen3-Coder-Next
 
9
  license: apache-2.0
10
  license_link: https://huggingface.co/Qwen/Qwen3-Coder-Next/blob/main/LICENSE
11
  pipeline_tag: text-generation
12
  ---
13
  <div>
14
  <p style="margin-bottom: 0; margin-top: 0;">
15
+ <h1 style="margin-top: 0rem;">To Run Qwen3-Coder-Next locally - <a href="https://unsloth.ai/docs/models/qwen3-coder-next">Read our Guide!</a></h1>
 
 
 
16
  </p>
17
+
18
  <p style="margin-top: 0;margin-bottom: 0;">
19
  <em><a href="https://unsloth.ai/docs/basics/unsloth-dynamic-2.0-ggufs">Unsloth Dynamic 2.0</a> achieves superior accuracy & outperforms other leading quants.</em>
20
  </p>
 
29
  <img src="https://raw.githubusercontent.com/unslothai/unsloth/refs/heads/main/images/documentation%20green%20button.png" width="143">
30
  </a>
31
  </div>
 
32
  </div>
33
 
34
+ - **Feb 19 update**: Tool-calling should now be even better after llama.cpp fixes parsing.
35
+ - **Quantization benchmarks**: See third-party Aider, LiveCodeBench v6, MMLU Pro, GPQA [benchmarks for GGUFs here](https://unsloth.ai/docs/models/qwen3-coder-next#gguf-quantization-benchmarks).</h2>
36
+ - **Feb 4 update**: llama.cpp fixed a bug that caused Qwen to loop and have poor outputs.<br>We updated GGUFs - please re-download and update llama.cpp for improved outputs.</h2>
37
+
38
+ ## Qwen3-Coder-Next Usage Guidelines
39
  - It is recommended to have >45GB unified memory or RAM/VRAM to run 4-bit quants.
40
  - For best results, use any 2-bit XL quant or above (requires >30GB unified memory /combined RAM + VRAM).
41
+ - See how to run the model via [Claude Code & OpenAI Codex](https://unsloth.ai/docs/models/qwen3-coder-next#claude-codex).
42
  - For complete detailed instructions (sampling parameters etc.), see our guide: [docs.unsloth.ai/models/qwen3-coder-next](https://unsloth.ai/docs/models/qwen3-coder-next)
43
 
44
  ---