--- license: apache-2.0 base_model: - HuggingFaceTB/SmolLM3-3B pipeline_tag: text-generation library_name: transformers --- **SmolLM3‑3B • Quantized** --- ### 🚀 Model Description This is an **int8 quantized version** of **SmolLM3–3B**, a highly efficient, open-source 3 B parameter LLM. It delivers nearly state-of-the-art multilingual reasoning and long-context performance (up to 128k tokens) with drastically reduced memory usage and inference cost, enabling fast deployment on mid‑range GPUs and edge devices. --- ### 📏 Quantization Details * **Library:** torchao * **Precision:** int8 weights and activations * **Benefits:** \~50–75% reduction in VRAM usage, enabling 12–16 GB GPU usage, with minimal performance drop on reasoning, coding, and long-context abilities --- ### 🎯 Intended Use Ideal for: * Scenarios requiring **fast LLM inference** under constrained VRAM (e.g. small servers or laptops) * **Multilingual reasoning** tasks, chain-of-thought logic, and long-context document understanding * Deployments of dual-mode (think/no\_think) conversational agents * Research into efficient LLM deployment and quantization techniques --- ### ⚠️ Limitations * Slight performance loss compared to full-precision SmolLM3‑3B * Requires proper benchmarking in your actual environment * Continues to exhibit standard LLM risks: hallucination, bias, etc. * Quant performance may vary across languages or context lengths