{ "_from_model_config": true, "bos_token_id": 12561, "early_stopping": true, "eos_token_id": 12562, "length_penalty": 2.0, "max_length": 64, "no_repeat_ngram_size": 3, "num_beams": 5, "pad_token_id": 0, "transformers_version": "4.47.0", "use_cache": false }