# DeepSeek Base Model in GGUF Format This is the base DeepSeek 1.5B model converted to GGUF format for efficient inference. ## Model Details - Base model: DeepSeek 1.5B - Quantization: Q8_0 - Format: GGUF ## Usage This model can be used with llama.cpp and other GGUF-compatible inference engines. ## Original Model This model was converted from deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B.