Very Large GGUFs
Collection
GGUF quantized versions of very large models - over 100B parameters
•
60 items
•
Updated
•
5
Needs llama.cpp built from this branch: https://github.com/ngxson/llama.cpp/tree/xsn/mistral_large_moe
'Make knowledge free for everyone'
Quantized version of: mistralai/Mistral-Large-3-675B-Instruct-2512
![]()
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
Base model
mistralai/Mistral-Large-3-675B-Base-2512