rahtml's picture
Update README.md
898e2a0 verified
---
license: apache-2.0
base_model:
- Qwen/Qwen3-Coder-30B-A3B-Instruct
---
## Description
NVFP4 Quantization of [Qwen/Qwen3-Coder-30B-A3B-Instruct](https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct) using [TensorRT-Model-Optimizer](https://github.com/NVIDIA/Model-Optimizer). KV Cache quantized to FP8 for compatibility with inference backends.