chore: increase max_new_tokens parameter to 4096 across chat model and inference functions for improved response generation
c81f16e
Daniel Machado Pedrozo
commited on