metadata
library_name: transformers
pipeline_tag: text-generation
tags:
- text-generation
- conversational
- multilingual
- zen
- zen3
- hanzo
- zenlm
license: apache-2.0
Zen3 Nano
Zen LM by Hanzo AI — Compact yet capable language model for fast inference and edge deployment. 8B parameters with strong multilingual support.
Specs
| Property | Value |
|---|---|
| Parameters | 8B |
| Context Length | 40K tokens |
| Languages | 100+ |
| Architecture | Zen MoDE (Mixture of Distilled Experts) |
| Generation | Zen3 |
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("zenlm/zen3-nano", torch_dtype="auto")
tokenizer = AutoTokenizer.from_pretrained("zenlm/zen3-nano")
messages = [{"role": "user", "content": "Explain quantum computing in simple terms."}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer([text], return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
API Access
Available via the Hanzo AI API:
from openai import OpenAI
client = OpenAI(base_url="https://api.hanzo.ai/v1", api_key="YOUR_KEY")
response = client.chat.completions.create(
model="zen3-nano",
messages=[{"role": "user", "content": "Hello"}],
)
print(response.choices[0].message.content)
Get your API key at console.hanzo.ai — $5 free credit on signup.
License
Apache 2.0
Zen LM is developed by Hanzo AI — Frontier AI infrastructure.