forked from Selig/openclaw-skill
6 custom skills (assign-task, dispatch-webhook, daily-briefing, task-capture, qmd-brain, tts-voice) with technical documentation. Compatible with Claude Code, OpenClaw, Codex CLI, and OpenCode.
1.4 KiB
1.4 KiB
Model Provider Quickstart
Model Providers
OpenClaw can use many LLM providers. Pick one, authenticate, then set the default
model as provider/model.
Highlight: Venice (Venice AI)
Venice is our recommended Venice AI setup for privacy-first inference with an option to use Opus for the hardest tasks.
- Default:
venice/llama-3.3-70b - Best overall:
venice/claude-opus-45(Opus remains the strongest)
See Venice AI.
Quick start (two steps)
- Authenticate with the provider (usually via
openclaw onboard). - Set the default model:
{
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } },
}
Supported providers (starter set)
- OpenAI (API + Codex)
- Anthropic (API + Claude Code CLI)
- OpenRouter
- Vercel AI Gateway
- Cloudflare AI Gateway
- Moonshot AI (Kimi + Kimi Coding)
- Synthetic
- OpenCode Zen
- Z.AI
- GLM models
- MiniMax
- Venice (Venice AI)
- Amazon Bedrock
For the full provider catalog (xAI, Groq, Mistral, etc.) and advanced configuration, see Model providers.