Insufficient context length
#2
by
X-SZM - opened
A context length of 32K is reasonable when we are talking about models no larger than 8B, but when we get to a size of 24B, the context length of 32K is too short and affects usability. Your work is very excellent, will you consider extending the context length to version 2.5 next? Like 128K or more?
According to the config, max_position_embeddings is 128000, indicating support for a 128K context window. However, unexpected outputs or performance degradation might occur at extreme context lengths.
Yes, we plan to extend it for the 2.5 release. Please see this one as an early checkpoint. :)
Please train for coding, add benchmarks also for it the next version you are cooking. Thankyou.