prithivMLmods's picture
Update README.md
808080f verified
---
license: apache-2.0
base_model:
- prithivMLmods/Draco-CoderMini-3B
pipeline_tag: text-generation
tags:
- text-generation-inference
- math
- problem-solve
- code
language:
- en
library_name: transformers
---
# **Draco-CoderMini-3B-GGUF**
> **Draco-CoderMini-3B** is a compact, coding-optimized language model built on the **Qwen2 architecture**, tailored for high-accuracy **code generation**, **debugging**, and **technical reasoning**. With **3 billion parameters**, it strikes a balance between power and deployability, making it an ideal assistant for developers, educators, and engineers working in constrained environments or requiring fast inference.
## Model File
| File Name | Size | Format |
|----------------------------------------|---------|--------|
| Draco-CoderMini-3B.BF16.gguf | 6.18 GB | BF16 |
| Draco-CoderMini-3B.F16.gguf | 6.18 GB | F16 |
| Draco-CoderMini-3B.F32.gguf | 12.3 GB | F32 |
| .gitattributes | 1.75 kB | - |
| README.md | 210 B | - |
| config.json | 31 B | JSON |
## Quants Usage
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):
![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)