Hello, I’m exploring the GLM-5 model and I would like to know if there are plans to support multilingual tokenizers in future releases. Additionally, I observed slower inference times on longer sequences — any recommendations for optimization?
Test
· Sign up or log in to comment