File size: 935 Bytes
5ac4fa9 691aabb 24027ea 691aabb 24027ea 691aabb 24027ea 691aabb | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 | ---
base_model: Qwen/Qwen3-30B-A3B
datasets:
- HuggingFaceH4/ultrachat_200k
library_name: transformers
pipeline_tag: text-generation
quantized_by: QuantForge
tags:
- quantforge
- quantized
- nvfp4
---
# Neooooo/qf-integration-test
## QuantForge Metadata
- Base model: `Qwen/Qwen3-30B-A3B`
- Quantization scheme: `nvfp4`
- Calibration dataset: `HuggingFaceH4/ultrachat_200k`
- Calibration samples: `32`
- Max sequence length: `512`
- Ignored layers: `lm_head, re:.*\.mlp\.gate$, re:.*\.mlp\.router$`
## Accuracy (BF16 vs NVFP4)
| Task | Metric | BF16 | NVFP4 | Recovery |
|---|---:|---:|---:|---:|
| arc_challenge | acc,none | 0.4000 | 0.3000 | 0.750 |
| hellaswag | acc,none | 0.4000 | 0.4000 | 1.000 |
Aggregate macro recovery: **0.875**
> **Note:** Scores estimated from subset.
## Performance
_Performance benchmark unavailable: evaluate.skip_perf=true_
## Usage (vLLM)
```bash
vllm serve Neooooo/qf-integration-test
```
|