File size: 3,689 Bytes
4711922
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
---
license: apache-2.0
base_model: HuggingFaceTB/SmolLM2-135M
library_name: transformers
language:
  - en
tags:
  - quantllm
  - transformers
  - safetensors
pipeline_tag: text-generation
---

<div align="center">

# πŸ€— SmolLM2-135M-QuantLLM

**HuggingFaceTB/SmolLM2-135M** converted to **SAFETENSORS** format

[![QuantLLM](https://img.shields.io/badge/πŸš€_Made_with-QuantLLM-orange?style=for-the-badge)](https://github.com/codewithdark-git/QuantLLM)
[![Format](https://img.shields.io/badge/Format-SAFETENSORS-blue?style=for-the-badge)]()


<a href="https://github.com/codewithdark-git/QuantLLM">⭐ Star QuantLLM on GitHub</a>

</div>

---


## πŸ“– About This Model

This model is **[HuggingFaceTB/SmolLM2-135M](https://huggingface.co/HuggingFaceTB/SmolLM2-135M)** converted to **SafeTensors** format for use with HuggingFace Transformers and PyTorch.

| Property | Value |
|----------|-------|
| **Base Model** | [HuggingFaceTB/SmolLM2-135M](https://huggingface.co/HuggingFaceTB/SmolLM2-135M) |
| **Format** | SAFETENSORS |
| **Quantization** | None (Full Precision) |
| **License** | apache-2.0 |
| **Created With** | [QuantLLM](https://github.com/codewithdark-git/QuantLLM) |


## πŸš€ Quick Start

### With Transformers

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

# Load model and tokenizer
model = AutoModelForCausalLM.from_pretrained("codewithdark/SmolLM2-135M-QuantLLM")
tokenizer = AutoTokenizer.from_pretrained("codewithdark/SmolLM2-135M-QuantLLM")

# Generate text
inputs = tokenizer("Once upon a time", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100, do_sample=True)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```

### With QuantLLM

```python
from quantllm import TurboModel

# Load with automatic optimization
model = TurboModel.from_pretrained("codewithdark/SmolLM2-135M-QuantLLM")

# Generate
response = model.generate("Write a poem about coding")
print(response)
```

### Requirements

```bash
pip install transformers torch
```


## πŸ“Š Model Details

| Property | Value |
|----------|-------|
| **Original Model** | [HuggingFaceTB/SmolLM2-135M](https://huggingface.co/HuggingFaceTB/SmolLM2-135M) |
| **Format** | SAFETENSORS |
| **Quantization** | Full Precision |
| **License** | `apache-2.0` |
| **Export Date** | 2026-04-29 |
| **Exported By** | [QuantLLM v2.1](https://github.com/codewithdark-git/QuantLLM) |



---

## πŸš€ Created with QuantLLM

<div align="center">

[![QuantLLM](https://img.shields.io/badge/πŸš€_QuantLLM-Ultra--fast_LLM_Quantization-orange?style=for-the-badge)](https://github.com/codewithdark-git/QuantLLM)

**Convert any model to GGUF, ONNX, or MLX in one line!**

```python
from quantllm import turbo

# Load any HuggingFace model
model = turbo("HuggingFaceTB/SmolLM2-135M")

# Export to any format
model.export("safetensors", quantization="Q4_K_M")

# Push to HuggingFace
model.push("your-repo", format="safetensors")
```

<a href="https://github.com/codewithdark-git/QuantLLM">
  <img src="https://img.shields.io/github/stars/codewithdark-git/QuantLLM?style=social" alt="GitHub Stars">
</a>

**[πŸ“š Documentation](https://github.com/codewithdark-git/QuantLLM#readme)** Β· 
**[πŸ› Report Issue](https://github.com/codewithdark-git/QuantLLM/issues)** Β· 
**[πŸ’‘ Request Feature](https://github.com/codewithdark-git/QuantLLM/issues)**

</div>


## πŸ“Š Export Details

Exported with [QuantLLM](https://github.com/codewithdark-git/QuantLLM) from `HuggingFaceTB/SmolLM2-135M` (134.5M params).

| Property | Value |
|----------|-------|
| **Format** | SafeTensors |
| **Size** | 541.6 MB |
| **Parameters** | 134.5M |
| **Dtype** | float32 |

### How to use