File size: 324 Bytes
755edb5
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
{
  "quant_method": "channel_wise_int8",
  "bits": 8,
  "format": "per_output_channel_scale",
  "weight_dtype": "I8",
  "scale_dtype": "BF16",
  "description": "Each weight tensor 'foo' is stored as I8 values in 'foo' with a BF16 per-output-channel scale in 'foo_scale'. Dequantize: weight_bf16 = weight_i8 * scale_bf16."
}