Update README.md
Browse files
README.md
CHANGED
|
@@ -21,14 +21,14 @@ This model was built with Qwen3-Coder-Next model by applying AMD-Quark for MXFP4
|
|
| 21 |
|
| 22 |
# Model Quantization
|
| 23 |
|
| 24 |
-
The model was quantized from
|
| 25 |
|
| 26 |
**Quantization scripts:**
|
| 27 |
|
| 28 |
Note that qwen3_next is not in the built-in model template list in Quark V0.11, it has to be registered before quantization.
|
| 29 |
|
| 30 |
```python
|
| 31 |
-
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 32 |
from datasets import load_dataset
|
| 33 |
from quark.torch import LLMTemplate, ModelQuantizer, export_safetensors
|
| 34 |
from quark.contrib.llm_eval import ppl_eval
|
|
|
|
| 21 |
|
| 22 |
# Model Quantization
|
| 23 |
|
| 24 |
+
The model was quantized from Qwen3-Coder-Next using [AMD-Quark](https://quark.docs.amd.com/latest/index.html). The weights and activations are quantized to MXFP4.
|
| 25 |
|
| 26 |
**Quantization scripts:**
|
| 27 |
|
| 28 |
Note that qwen3_next is not in the built-in model template list in Quark V0.11, it has to be registered before quantization.
|
| 29 |
|
| 30 |
```python
|
| 31 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 32 |
from datasets import load_dataset
|
| 33 |
from quark.torch import LLMTemplate, ModelQuantizer, export_safetensors
|
| 34 |
from quark.contrib.llm_eval import ppl_eval
|