File size: 3,210 Bytes
9faa498 1d4375e 9faa498 1d4375e 9faa498 1d4375e 9faa498 27a784f 9faa498 1d4375e 9faa498 1d4375e 27a784f 1d4375e 9faa498 1d4375e 9faa498 1d4375e 9faa498 1d4375e 9faa498 1d4375e 9faa498 1d4375e 27a784f 1d4375e 9faa498 1d4375e 9faa498 1d4375e | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 | ---
license: apache-2.0
language:
- en
- zh
- ja
- ko
- fr
- de
- es
- pt
- ru
- ar
tags:
- zen4
- zenlm
- hanzo
- frontier-ai
- open-weight
base_model: Qwen/Qwen3.5-9B
pipeline_tag: text-generation
library_name: transformers
---
# Zen4
**Zen4** is a 9B parameter language model from the [Zen4 family](https://zenlm.org) by [Zen LM](https://huggingface.co/zenlm) and [Hanzo AI](https://hanzo.ai).
Built on open-weight weights with Zen4 Frontier architecture for unrestricted, open-ended AI assistance.
## Model Details
| Property | Value |
|----------|-------|
| **Parameters** | 9B total, 9B active |
| **Architecture** | Zen4 Frontier |
| **Context** | 262K tokens |
| **License** | APACHE-2.0 |
| **Family** | Zen4 |
| **Tier** | Small |
| **Creator** | Zen LM / Hanzo AI |
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("zenlm/zen4", torch_dtype="auto")
tokenizer = AutoTokenizer.from_pretrained("zenlm/zen4")
messages = [{"role": "user", "content": "Hello, who are you?"}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0][inputs.input_ids.shape[-1]:], skip_special_tokens=True))
```
## Zen4 Family
| Model | Parameters | Context | HuggingFace |
|-------|-----------|---------|-------------|
| Zen4 Nano | 0.8B | 262K | [zenlm/zen4-nano](https://huggingface.co/zenlm/zen4-nano) |
| Zen4 Micro | 2B | 262K | [zenlm/zen4-micro](https://huggingface.co/zenlm/zen4-micro) |
| Zen4 Mini | 4B | 262K | [zenlm/zen4-mini](https://huggingface.co/zenlm/zen4-mini) |
| **Zen4** | **9B** | **262K** | [zenlm/zen4](https://huggingface.co/zenlm/zen4) |
| Zen4 Pro | 27B | 262K | [zenlm/zen4-pro](https://huggingface.co/zenlm/zen4-pro) |
| Zen4 Max | 35B MoE (3B active) | 262K | [zenlm/zen4-max](https://huggingface.co/zenlm/zen4-max) |
| Zen4 Coder Flash | 31B MoE (3B active) | 131K | [zenlm/zen4-coder-flash](https://huggingface.co/zenlm/zen4-coder-flash) |
| Zen4 Pro Max | 80B MoE (3B active) | 256K | [zenlm/zen4-pro-max](https://huggingface.co/zenlm/zen4-pro-max) |
| Zen4 Coder | 80B MoE (3B active) | 256K | [zenlm/zen4-coder](https://huggingface.co/zenlm/zen4-coder) |
| Zen4 Mega | 122B MoE (10B active) | 262K | [zenlm/zen4-mega](https://huggingface.co/zenlm/zen4-mega) |
| Zen4 Thunder | 230B MoE (10B active) | 1M | [zenlm/zen4-thunder](https://huggingface.co/zenlm/zen4-thunder) |
| Zen4 Storm | 456B MoE (45B active) | 1M | [zenlm/zen4-storm](https://huggingface.co/zenlm/zen4-storm) |
| Zen4 Titan | 744B MoE (40B active) | 128K | [zenlm/zen4-titan](https://huggingface.co/zenlm/zen4-titan) |
| Zen4 Ultra | 1.04T MoE (32B active) | 256K | [zenlm/zen4-ultra](https://huggingface.co/zenlm/zen4-ultra) |
| Zen4 Ultra Max | 1T MoE (50B active) | 128K | [zenlm/zen4-ultra-max](https://huggingface.co/zenlm/zen4-ultra-max) |
## Links
- [Zen LM](https://zenlm.org) | [Hanzo AI](https://hanzo.ai) | [Hanzo Chat](https://hanzo.chat)
- [All Zen Models](https://huggingface.co/zenlm)
---
*Zen AI: Clarity Through Intelligence*
|