petergilani commited on
Commit
a3f5200
·
verified ·
1 Parent(s): bb9b873

Create model card

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -25,15 +25,15 @@ tags:
25
 
26
  <!-- Provide a longer summary of what this model is. -->
27
 
28
- # Updated Evaluation Results (February 13, 2026)
29
 
30
- Comprehensive evaluation results from thorough testing using mlx_lm.evaluate with mmlu_pro (200 questions per domain, num_shots=1, temp=1.0, top_p=0.95, top_k=40, seed=123):
31
 
32
- ## Recent g64 vs g128 Comparison (8-bit)
33
 
34
- Based on the most recent comprehensive evaluation, here is the direct comparison between the 8-bit models:
35
 
36
- ### Direct Comparison Summary
37
 
38
  | Domain | 8-bit g64 | 8-bit g128 (this model) | Difference |
39
  |--------|-----------|-------------------------|------------|
 
25
 
26
  <!-- Provide a longer summary of what this model is. -->
27
 
28
+ # Model Card for Qwen3-Coder-Next-8bit-g128
29
 
30
+ Quantized Qwen/Qwen3-Coder-Next using mlx-lm to 8-bit with group_size 128 for main weights and fine-grained group_size 64 for MoE weights, with the aim of maximum accuracy for 8-bit quantization.
31
 
32
+ ## Updated Evaluation Results (February 13, 2026)
33
 
34
+ Comprehensive evaluation results from thorough testing using mlx_lm.evaluate with mmlu_pro (200 questions per domain, num_shots=1, temp=1.0, top_p=0.95, top_k=40, seed=123):
35
 
36
+ ### Direct Comparison Summary (8-bit g64 vs g128)
37
 
38
  | Domain | 8-bit g64 | 8-bit g128 (this model) | Difference |
39
  |--------|-----------|-------------------------|------------|