File size: 192 Bytes
316af64 | 1 2 3 4 5 6 7 8 9 10 | {
"quantization": {
"quant_algo": "FP8",
"kv_cache_quant_algo": null,
"exclude_modules": [
"lm_head",
"model.embed_tokens"
]
}
} |
316af64 | 1 2 3 4 5 6 7 8 9 10 | {
"quantization": {
"quant_algo": "FP8",
"kv_cache_quant_algo": null,
"exclude_modules": [
"lm_head",
"model.embed_tokens"
]
}
} |