chore: increase max_new_tokens parameter to 4096 across chat model and inference functions for improved response generation c81f16e
Daniel Machado Pedrozo commited on