AINovice2005's picture
Update README.md
44f54f8 verified
---
license: apache-2.0
base_model:
- HuggingFaceTB/SmolLM3-3B
pipeline_tag: text-generation
library_name: transformers
---
**SmolLM3‑3B • Quantized**
---
### 🚀 Model Description
This is an **int8 quantized version** of **SmolLM3–3B**, a highly efficient, open-source 3 B parameter LLM.
It delivers nearly state-of-the-art multilingual reasoning and long-context performance (up to 128k tokens) with drastically reduced
memory usage and inference cost, enabling fast deployment on mid‑range GPUs and edge devices.
---
### 📏 Quantization Details
* **Library:** torchao
* **Precision:** int8 weights and activations
* **Benefits:** \~50–75% reduction in VRAM usage, enabling 12–16 GB GPU usage, with minimal performance drop on reasoning, coding, and long-context abilities
---
### 🎯 Intended Use
Ideal for:
* Scenarios requiring **fast LLM inference** under constrained VRAM (e.g. small servers or laptops)
* **Multilingual reasoning** tasks, chain-of-thought logic, and long-context document understanding
* Deployments of dual-mode (think/no\_think) conversational agents
* Research into efficient LLM deployment and quantization techniques
---
### ⚠️ Limitations
* Slight performance loss compared to full-precision SmolLM3‑3B
* Requires proper benchmarking in your actual environment
* Continues to exhibit standard LLM risks: hallucination, bias, etc.
* Quant performance may vary across languages or context lengths