GGUFs for Yi 6b 200k v2: https://huggingface.co/01-ai/Yi-6B-200K

This is the 24/03/07 updated version with enhanced long-context performance. From their notes:

The long text capability of the Yi-34B-200K has been enhanced. In the "Needle-in-a-Haystack" test, the Yi-34B-200K's performance is improved by 10.5%, rising from 89.3% to an impressive 99.8%. We continue to pretrain the model on 5B tokens long-context data mixture and demonstrate a near-all-green performance.

It has been confirmed that the 6b version was similarly updated here: https://huggingface.co/01-ai/Yi-6B-200K/discussions/6

Downloads last month
2
GGUF
Model size
6B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support