| --- |
| summary: "Model providers (LLMs) supported by OpenClaw" |
| read_when: |
| - You want to choose a model provider |
| - You want quick setup examples for LLM auth + model selection |
| title: "Model Provider Quickstart" |
| --- |
| |
| # Model Providers |
|
|
| OpenClaw can use many LLM providers. Pick one, authenticate, then set the default |
| model as `provider/model`. |
|
|
| ## Quick start (two steps) |
|
|
| 1. Authenticate with the provider (usually via `openclaw onboard`). |
| 2. Set the default model: |
|
|
| ```json5 |
| { |
| agents: { defaults: { model: { primary: "anthropic/claude-opus-4-6" } } }, |
| } |
| ``` |
|
|
| ## Supported providers (starter set) |
|
|
| - [OpenAI (API + Codex)](/providers/openai) |
| - [Anthropic (API + Claude Code CLI)](/providers/anthropic) |
| - [OpenRouter](/providers/openrouter) |
| - [Vercel AI Gateway](/providers/vercel-ai-gateway) |
| - [Cloudflare AI Gateway](/providers/cloudflare-ai-gateway) |
| - [Moonshot AI (Kimi + Kimi Coding)](/providers/moonshot) |
| - [Mistral](/providers/mistral) |
| - [Synthetic](/providers/synthetic) |
| - [OpenCode (Zen + Go)](/providers/opencode) |
| - [Z.AI](/providers/zai) |
| - [GLM models](/providers/glm) |
| - [MiniMax](/providers/minimax) |
| - [Venice (Venice AI)](/providers/venice) |
| - [Amazon Bedrock](/providers/bedrock) |
| - [Qianfan](/providers/qianfan) |
|
|
| For the full provider catalog (xAI, Groq, Mistral, etc.) and advanced configuration, |
| see [Model providers](/concepts/model-providers). |
|
|