| { | |
| "model_type": "HunyuanImage-3.0-Instruct-Distil", | |
| "quantization_method": "bitsandbytes_nf4", | |
| "load_in_4bit": true, | |
| "bnb_4bit_quant_type": "nf4", | |
| "bnb_4bit_use_double_quant": true, | |
| "bnb_4bit_compute_dtype": "torch.bfloat16", | |
| "expected_vram_gb": 12, | |
| "total_params": "80B", | |
| "active_params": "13B (MoE)", | |
| "modules_kept_bf16": [ | |
| "vae", | |
| "vision_model", | |
| "vision_aligner", | |
| "patch_embed", | |
| "final_layer", | |
| "time_embed", | |
| "time_embed_2", | |
| "timestep_emb", | |
| "guidance_emb", | |
| "timestep_r_emb", | |
| "attention_projections" | |
| ], | |
| "distil_features": { | |
| "cfg_distilled": true, | |
| "meanflow": true, | |
| "description": "Single-step CFG-free generation with meanflow" | |
| }, | |
| "notes": "NF4 quantized Distil model - extremely compact, single GPU friendly.", | |
| "attention_layers_quantized": false | |
| } |