Fix: Handle cache_position argument for newer Transformers

#1
by Todokete - opened

Fixes "TypeError: MixsenseLlamaForCausalLM.forward() got an unexpected keyword argument 'cache_position'" in newer Transformers releases by adding cache_position to the forward and prepare_inputs_for_generation method signatures and passing it to the super() calls.

ZeroVision-rxxiang changed pull request status to merged

Sign up or log in to comment