--- summary: "Model providers (LLMs) supported by OpenClaw" read_when: - You want to choose a model provider - You want quick setup examples for LLM auth + model selection title: "Model Provider Quickstart" --- # Model Providers OpenClaw can use many LLM providers. Pick one, authenticate, then set the default model as `provider/model`. ## Highlight: Venice (Venice AI) Venice is our recommended Venice AI setup for privacy-first inference with an option to use Opus for the hardest tasks. - Default: `venice/llama-3.3-70b` - Best overall: `venice/claude-opus-45` (Opus remains the strongest) See [Venice AI](/providers/venice). ## Quick start (two steps) 1. Authenticate with the provider (usually via `openclaw onboard`). 2. Set the default model: ```json5 { agents: { defaults: { model: { primary: "anthropic/claude-opus-4-6" } } }, } ``` ## Supported providers (starter set) - [OpenAI (API + Codex)](/providers/openai) - [Anthropic (API + Claude Code CLI)](/providers/anthropic) - [OpenRouter](/providers/openrouter) - [Vercel AI Gateway](/providers/vercel-ai-gateway) - [Cloudflare AI Gateway](/providers/cloudflare-ai-gateway) - [Moonshot AI (Kimi + Kimi Coding)](/providers/moonshot) - [Synthetic](/providers/synthetic) - [OpenCode Zen](/providers/opencode) - [Z.AI](/providers/zai) - [GLM models](/providers/glm) - [MiniMax](/providers/minimax) - [Venice (Venice AI)](/providers/venice) - [Amazon Bedrock](/bedrock) - [Qianfan](/providers/qianfan) For the full provider catalog (xAI, Groq, Mistral, etc.) and advanced configuration, see [Model providers](/concepts/model-providers).