File size: 935 Bytes
5ac4fa9
 
 
 
 
 
 
 
 
 
 
 
 
691aabb
 
 
 
 
 
 
 
 
24027ea
691aabb
 
 
 
 
24027ea
 
 
 
691aabb
 
 
 
 
24027ea
691aabb
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
---
base_model: Qwen/Qwen3-30B-A3B
datasets:
- HuggingFaceH4/ultrachat_200k
library_name: transformers
pipeline_tag: text-generation
quantized_by: QuantForge
tags:
- quantforge
- quantized
- nvfp4
---

# Neooooo/qf-integration-test

## QuantForge Metadata

- Base model: `Qwen/Qwen3-30B-A3B`
- Quantization scheme: `nvfp4`
- Calibration dataset: `HuggingFaceH4/ultrachat_200k`
- Calibration samples: `32`
- Max sequence length: `512`
- Ignored layers: `lm_head, re:.*\.mlp\.gate$, re:.*\.mlp\.router$`

## Accuracy (BF16 vs NVFP4)

| Task | Metric | BF16 | NVFP4 | Recovery |
|---|---:|---:|---:|---:|
| arc_challenge | acc,none | 0.4000 | 0.3000 | 0.750 |
| hellaswag | acc,none | 0.4000 | 0.4000 | 1.000 |

Aggregate macro recovery: **0.875**

> **Note:** Scores estimated from subset.

## Performance

_Performance benchmark unavailable: evaluate.skip_perf=true_

## Usage (vLLM)

```bash
vllm serve Neooooo/qf-integration-test
```