How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("feature-extraction", model="zenlm/zen3-embedding")
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("zenlm/zen3-embedding")
model = AutoModelForCausalLM.from_pretrained("zenlm/zen3-embedding")
Quick Links

Zen3 Embedding

Zen LM by Hanzo AI โ€” High-dimensional text embeddings via API. Supports retrieval, similarity, and clustering.

Specs

Property Value
Parameters API-served
Context 8K
Architecture Zen MoDE (Mixture of Distilled Experts)
Generation Zen3

API Access

curl https://api.hanzo.ai/v1/chat/completions \
  -H "Authorization: Bearer $HANZO_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{"model": "zen3-embedding", "messages": [{"role": "user", "content": "Hello"}]}'

Get your API key at console.hanzo.ai โ€” $5 free credit on signup.

License

Apache 2.0


Zen LM is developed by Hanzo AI โ€” Frontier AI infrastructure.

Downloads last month
7
Safetensors
Model size
4B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support