GGUF models
Collection
17 items
•
Updated
This quantized models are converted from deepseek-ai/DeepSeek-R1-Distill-Llama-8B with https://huggingface.co/spaces/ggml-org/gguf-my-repo
The bf16, f16 models are converted by llama.cpp with version 25ff6f7659f6a5c47d6a73eada5813f0495331f0
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit