File size: 3,689 Bytes
4711922 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 | ---
license: apache-2.0
base_model: HuggingFaceTB/SmolLM2-135M
library_name: transformers
language:
- en
tags:
- quantllm
- transformers
- safetensors
pipeline_tag: text-generation
---
<div align="center">
# π€ SmolLM2-135M-QuantLLM
**HuggingFaceTB/SmolLM2-135M** converted to **SAFETENSORS** format
[](https://github.com/codewithdark-git/QuantLLM)
[]()
<a href="https://github.com/codewithdark-git/QuantLLM">β Star QuantLLM on GitHub</a>
</div>
---
## π About This Model
This model is **[HuggingFaceTB/SmolLM2-135M](https://huggingface.co/HuggingFaceTB/SmolLM2-135M)** converted to **SafeTensors** format for use with HuggingFace Transformers and PyTorch.
| Property | Value |
|----------|-------|
| **Base Model** | [HuggingFaceTB/SmolLM2-135M](https://huggingface.co/HuggingFaceTB/SmolLM2-135M) |
| **Format** | SAFETENSORS |
| **Quantization** | None (Full Precision) |
| **License** | apache-2.0 |
| **Created With** | [QuantLLM](https://github.com/codewithdark-git/QuantLLM) |
## π Quick Start
### With Transformers
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load model and tokenizer
model = AutoModelForCausalLM.from_pretrained("codewithdark/SmolLM2-135M-QuantLLM")
tokenizer = AutoTokenizer.from_pretrained("codewithdark/SmolLM2-135M-QuantLLM")
# Generate text
inputs = tokenizer("Once upon a time", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100, do_sample=True)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
### With QuantLLM
```python
from quantllm import TurboModel
# Load with automatic optimization
model = TurboModel.from_pretrained("codewithdark/SmolLM2-135M-QuantLLM")
# Generate
response = model.generate("Write a poem about coding")
print(response)
```
### Requirements
```bash
pip install transformers torch
```
## π Model Details
| Property | Value |
|----------|-------|
| **Original Model** | [HuggingFaceTB/SmolLM2-135M](https://huggingface.co/HuggingFaceTB/SmolLM2-135M) |
| **Format** | SAFETENSORS |
| **Quantization** | Full Precision |
| **License** | `apache-2.0` |
| **Export Date** | 2026-04-29 |
| **Exported By** | [QuantLLM v2.1](https://github.com/codewithdark-git/QuantLLM) |
---
## π Created with QuantLLM
<div align="center">
[](https://github.com/codewithdark-git/QuantLLM)
**Convert any model to GGUF, ONNX, or MLX in one line!**
```python
from quantllm import turbo
# Load any HuggingFace model
model = turbo("HuggingFaceTB/SmolLM2-135M")
# Export to any format
model.export("safetensors", quantization="Q4_K_M")
# Push to HuggingFace
model.push("your-repo", format="safetensors")
```
<a href="https://github.com/codewithdark-git/QuantLLM">
<img src="https://img.shields.io/github/stars/codewithdark-git/QuantLLM?style=social" alt="GitHub Stars">
</a>
**[π Documentation](https://github.com/codewithdark-git/QuantLLM#readme)** Β·
**[π Report Issue](https://github.com/codewithdark-git/QuantLLM/issues)** Β·
**[π‘ Request Feature](https://github.com/codewithdark-git/QuantLLM/issues)**
</div>
## π Export Details
Exported with [QuantLLM](https://github.com/codewithdark-git/QuantLLM) from `HuggingFaceTB/SmolLM2-135M` (134.5M params).
| Property | Value |
|----------|-------|
| **Format** | SafeTensors |
| **Size** | 541.6 MB |
| **Parameters** | 134.5M |
| **Dtype** | float32 |
### How to use
|