This quantized models are converted from deepseek-ai/DeepSeek-R1-Distill-Llama-8B with https://huggingface.co/spaces/ggml-org/gguf-my-repo

The bf16, f16 models are converted by llama.cpp with version 25ff6f7659f6a5c47d6a73eada5813f0495331f0

Downloads last month
10
GGUF
Model size
8B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including skymizer/DeepSeek-R1-Distill-Llama-8B-GGUF