Spaces:
Running
Running
| summary: "Model providers (LLMs) supported by OpenClaw" | |
| read_when: | |
| - You want to choose a model provider | |
| - You need a quick overview of supported LLM backends | |
| title: "Model Providers" | |
| # Model Providers | |
| OpenClaw can use many LLM providers. Pick a provider, authenticate, then set the | |
| default model as `provider/model`. | |
| Looking for chat channel docs (WhatsApp/Telegram/Discord/Slack/Mattermost (plugin)/etc.)? See [Channels](/channels). | |
| ## Highlight: Venice (Venice AI) | |
| Venice is our recommended Venice AI setup for privacy-first inference with an option to use Opus for hard tasks. | |
| - Default: `venice/llama-3.3-70b` | |
| - Best overall: `venice/claude-opus-45` (Opus remains the strongest) | |
| See [Venice AI](/providers/venice). | |
| ## Quick start | |
| 1. Authenticate with the provider (usually via `openclaw onboard`). | |
| 2. Set the default model: | |
| ```json5 | |
| { | |
| agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } }, | |
| } | |
| ``` | |
| ## Provider docs | |
| - [OpenAI (API + Codex)](/providers/openai) | |
| - [Anthropic (API + Claude Code CLI)](/providers/anthropic) | |
| - [Qwen (OAuth)](/providers/qwen) | |
| - [OpenRouter](/providers/openrouter) | |
| - [Vercel AI Gateway](/providers/vercel-ai-gateway) | |
| - [Moonshot AI (Kimi + Kimi Coding)](/providers/moonshot) | |
| - [OpenCode Zen](/providers/opencode) | |
| - [Amazon Bedrock](/bedrock) | |
| - [Z.AI](/providers/zai) | |
| - [Xiaomi](/providers/xiaomi) | |
| - [GLM models](/providers/glm) | |
| - [MiniMax](/providers/minimax) | |
| - [Venice (Venice AI, privacy-focused)](/providers/venice) | |
| - [Ollama (local models)](/providers/ollama) | |
| ## Transcription providers | |
| - [Deepgram (audio transcription)](/providers/deepgram) | |
| ## Community tools | |
| - [Claude Max API Proxy](/providers/claude-max-api-proxy) - Use Claude Max/Pro subscription as an OpenAI-compatible API endpoint | |
| For the full provider catalog (xAI, Groq, Mistral, etc.) and advanced configuration, | |
| see [Model providers](/concepts/model-providers). | |