MNLP_M2_mcqa_model-W4A8 / generation_config.json
CarlOwOs's picture
Add FP8 dynamically quantized MNLP_M2_mcqa_model model using llm-compressor
cc8a543 verified
raw
history blame contribute delete
166 Bytes
{
"bos_token_id": 151643,
"eos_token_id": 151643,
"max_length": 32768,
"max_new_tokens": 2048,
"pad_token_id": 151654,
"transformers_version": "4.52.3"
}