Create model card
Browse files
README.md
CHANGED
|
@@ -25,15 +25,15 @@ tags:
|
|
| 25 |
|
| 26 |
<!-- Provide a longer summary of what this model is. -->
|
| 27 |
|
| 28 |
-
#
|
| 29 |
|
| 30 |
-
|
| 31 |
|
| 32 |
-
##
|
| 33 |
|
| 34 |
-
|
| 35 |
|
| 36 |
-
### Direct Comparison Summary
|
| 37 |
|
| 38 |
| Domain | 8-bit g64 | 8-bit g128 (this model) | Difference |
|
| 39 |
|--------|-----------|-------------------------|------------|
|
|
|
|
| 25 |
|
| 26 |
<!-- Provide a longer summary of what this model is. -->
|
| 27 |
|
| 28 |
+
# Model Card for Qwen3-Coder-Next-8bit-g128
|
| 29 |
|
| 30 |
+
Quantized Qwen/Qwen3-Coder-Next using mlx-lm to 8-bit with group_size 128 for main weights and fine-grained group_size 64 for MoE weights, with the aim of maximum accuracy for 8-bit quantization.
|
| 31 |
|
| 32 |
+
## Updated Evaluation Results (February 13, 2026)
|
| 33 |
|
| 34 |
+
Comprehensive evaluation results from thorough testing using mlx_lm.evaluate with mmlu_pro (200 questions per domain, num_shots=1, temp=1.0, top_p=0.95, top_k=40, seed=123):
|
| 35 |
|
| 36 |
+
### Direct Comparison Summary (8-bit g64 vs g128)
|
| 37 |
|
| 38 |
| Domain | 8-bit g64 | 8-bit g128 (this model) | Difference |
|
| 39 |
|--------|-----------|-------------------------|------------|
|