{ "quantization_method": "torchao_Float8WeightOnly", "weight_dtype": "float8_e4m3fn", "scale_dtype": "float32", "scale_granularity": "per_row", "activation_dtype": "bfloat16", "torchao_version": "0.16.0", "torch_version": "2.9.0+cu128", "source_model": "fishaudio/s2-pro", "total_params_B": 4.562, "fp8_linear_params_B": 4.048, "bf16_other_params_B": 0.514, "output_size_GB": 6.16, "key_format": { "": "float8_e4m3fn quantized weight", ".scale": "float32 per-row dequantization scale", "_buf.": "bf16/fp32 buffer (freqs_cis, causal_mask, etc.)", "other": "bfloat16 (embeddings, norms, non-linear layers)" }, "inference_requirements": { "torchao": ">= 0.8.0", "compute_capability": ">= 8.9 (RTX 4090 / 5090) for native FP8 matmuls" }, "notes": "All nn.Linear weights are float8_e4m3fn. Activations are bfloat16 (weight-only quantization). codec.pth is unchanged bfloat16." }