Ali93H commited on
Commit
0436a5b
·
verified ·
1 Parent(s): b55700b

Upload folder using huggingface_hub

Browse files
Files changed (4) hide show
  1. .gitattributes +2 -0
  2. README.md +89 -3
  3. img/Intel.png +3 -0
  4. img/RTX5090.png +3 -0
.gitattributes CHANGED
@@ -46,3 +46,5 @@ Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-2.66bpw.gguf filter=lfs diff=lfs merge=lfs -
46
  Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-2.69bpw.gguf filter=lfs diff=lfs merge=lfs -text
47
  Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-2.90bpw.gguf filter=lfs diff=lfs merge=lfs -text
48
  Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-3.00bpw.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
46
  Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-2.69bpw.gguf filter=lfs diff=lfs merge=lfs -text
47
  Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-2.90bpw.gguf filter=lfs diff=lfs merge=lfs -text
48
  Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-3.00bpw.gguf filter=lfs diff=lfs merge=lfs -text
49
+ img/Intel.png filter=lfs diff=lfs merge=lfs -text
50
+ img/RTX5090.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,89 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ license_link: https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct/blob/main/LICENSE
5
+ pipeline_tag: text-generation
6
+ base_model:
7
+ - Qwen/Qwen3-Coder-30B-A3B-Instruct
8
+ tags:
9
+ - qwen
10
+ - qwen3
11
+ - qwen3-coder
12
+ - byteshape
13
+ ---
14
+
15
+ # Qwen3-Coder-30B-A3B-Instruct GGUF (ShapeLearn Quantized)
16
+
17
+ This is a GGUF-quantized version of Qwen3-Coder-30B-A3B-Instruct produced with **ByteShape's ShapeLearn**, which learns the optimal datatype per tensor to maintain high quality even at very low bitlengths.
18
+
19
+ To learn more about ShapeLearn and to see detailed benchmarks across GPUs, CPUs, and even the Raspberry Pi, please visit our [blog](https://byteshape.com/blogs/Devstral-Small-2-24B-Instruct-2512/).
20
+
21
+ If you have questions or want to share feedback, reach us on [Reddit](https://www.reddit.com/r/ByteShape/).
22
+
23
+
24
+ ## How to Pick a Model
25
+
26
+ We provide **CPU and GPU optimized variants** for llama.cpp:
27
+
28
+ - **CPUs:** Models labeled as KQ, optimized for CPU inference with predominantly KQ quantization.
29
+ - **GPUs:** Models labeled as IQ, optimized for GPU inference with a hybrid approach combining KQ and IQ quantization for better throughput.
30
+
31
+ Each hardware target includes a range of models covering different size and quality tradeoffs.
32
+
33
+ The chart below shows **quality versus tokens per second (TPS)**, with Unsloth used as the baseline for comparison.
34
+ Quality is measured across five benchmarks, including function calling: BFCL-V3, LiveCodeBench V6, HumanEval, Math500, and GSM8K.
35
+
36
+ **Selection rule:** Choose the model with the highest quality at your target throughput or the fastest model that still meets your required quality.
37
+
38
+ ### CPU Models
39
+
40
+ ![CPU Benchmark - Intel](img/Intel.png)
41
+
42
+ **Table sorted by model size** (match the chart numbers to model IDs):
43
+
44
+ | Model ID | Bits/Weight | Model Size |
45
+ |---------|-------------|-----------|
46
+ | [KQ-1](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-2.65bpw.gguf) | 2.65 | 10.1 GB |
47
+ | [KQ-2](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-2.66bpw.gguf) | 2.66 | 10.2 GB |
48
+ | [KQ-3](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-2.69bpw.gguf) | 2.69 | 10.3 GB |
49
+ | [KQ-4](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-Q3_K_M-2.69bpw.gguf) | 2.69 | 10.3 GB |
50
+ | [KQ-5](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-2.90bpw.gguf) | 2.90 | 11.1 GB |
51
+ | [KQ-6](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-Q3_K_S-3.00bpw.gguf) | 3.00 | 11.5 GB |
52
+ | [KQ-7](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-Q3_K_M-3.31bpw.gguf) | 3.31 | 12.7 GB |
53
+ | [KQ-8](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-IQ4_XS-4.20bpw.gguf) | 4.20 | 16.0 GB |
54
+
55
+
56
+ ### GPU Models
57
+
58
+ ![GPU Benchmark - RTX 5090](img/RTX5090.png)
59
+
60
+ **Table sorted by model size** (match the chart numbers to model IDs):
61
+
62
+ | Model ID | Bits/Weight | Model Size |
63
+ |---------|-------------|-----------|
64
+ | [IQ-1](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-IQ3_S-2.66bpw.gguf) | 2.66 | 10.1 GB |
65
+ | [IQ-2](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-IQ3_S-2.68bpw.gguf) | 2.68 | 10.2 GB |
66
+ | [IQ-3](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-IQ3_S-2.83bpw.gguf) | 2.83 | 10.8 GB |
67
+ | [IQ-4](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-IQ3_S-3.12bpw.gguf) | 3.12 | 11.9 GB |
68
+ | [IQ-5](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-IQ3_S-3.48bpw.gguf) | 3.48 | 13.3 GB |
69
+ | [IQ-6](https://huggingface.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF/blob/main/Qwen3-Coder-30B-A3B-Instruct-IQ4_XS-4.20bpw.gguf) | 4.20 | 16.0 GB |
70
+
71
+ ## Notes on quantization labels
72
+
73
+ The labels you see (for example `IQ4_XS`) are only there to make Hugging Face show our models in the GGUF table. We do not use the conventional quantization profiles as defined in llama.cpp. In our case, these labels indicate the primary quantization approach and average bit length. Note that both KQ and IQ models may use a mix of quantization techniques optimized for their target hardware, which is why several models can share the same tag.
74
+
75
+ ## Running these models with Ollama
76
+
77
+ All GGUF files in this repo can be used directly with Ollama.
78
+
79
+ To run a model with Ollama, use:
80
+
81
+ ```bash
82
+ ollama run hf.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF:FILE_NAME.gguf
83
+ ```
84
+
85
+ Replace `FILE_NAME.gguf` with the GGUF filename you want. For example:
86
+
87
+ ```bash
88
+ ollama run hf.co/byteshape/Qwen3-Coder-30B-A3B-Instruct-GGUF:Qwen3-Coder-30B-A3B-Instruct-IQ4_XS-4.20bpw.gguf
89
+ ```
img/Intel.png ADDED

Git LFS Details

  • SHA256: c9e2eb3d51eec09d21453803faa4d9023d2176578c3c617ec391d6d2ee495779
  • Pointer size: 131 Bytes
  • Size of remote file: 749 kB
img/RTX5090.png ADDED

Git LFS Details

  • SHA256: 91674184f4ed76330338be335803bcedf37cbef03ac47b10337363a7875db1f7
  • Pointer size: 131 Bytes
  • Size of remote file: 753 kB