Everyone asks which model to use. The honest answer: it depends on what you're doing and how much you want to spend.
Note: Clawdbot has been rebranded to OpenClaw — same powerful AI agent platform, new name. Learn more at openclaw.ai.
OpenClaw supports a dozen providers. Anthropic, OpenAI, Google, open-source models through haimaker.ai. Each has tradeoffs around cost, capability, and where your data ends up.
Here's how I think about picking one.
Price, capability, privacy
These three things compete with each other. You can optimize for two, maybe, but rarely all three.
Price
Token pricing varies wildly. Claude Opus 4.5 costs $15/$75 per million tokens (input/output). Grok 4.1 mini charges $0.20/$0.50. That's a 75x difference for what are, in many cases, similar outputs.
For most assistant tasks, a mid-tier model makes sense. Claude Sonnet 4 at $3/$15 gives you most of Opus's capability at a fraction of the cost.
Capability
Benchmarks lie. For OpenClaw, what actually matters:
- Tool calling – Can it invoke shell commands and APIs without fumbling the syntax?
- Context tracking – Does it remember what you said 50 messages ago?
- Code quality – When it writes code, does it run?
- Speed – How long before it starts responding?
Privacy
Cloud APIs mean your prompts hit external servers. For personal finance, health data, or proprietary code, that's a problem. You can self-host open-source models, but that requires hardware and tolerance for latency.
Recommendations by use case
Daily assistant work
Claude Sonnet 4 ($3/$15 per million tokens)
Calendar, email, research, general queries. Sonnet handles all of it without breaking the bank. Fast enough for real-time chat, smart enough for multi-step tasks.
Cheaper option: GPT-4o-mini (~$0.15/$0.60)
Fine for simple stuff. Quality drops on anything complex, but at 20x cheaper, sometimes that's the right call.
Coding and automation
Claude Opus 4.5 ($15/$75 per million tokens)
When the code needs to actually work, Opus is worth the premium. It handles multi-file edits and complex debugging better than anything else I've used.
Alternative: Sonnet 4 with extended thinking enabled. Pay more per reasoning token only when you need the horsepower.
Research and document analysis
Gemini 3 Pro (~$1.25/$10 per million tokens)
The 1M+ token context window lets you throw entire codebases at it. Good at synthesizing information across long documents.
Privacy-sensitive work
Llama 3.3 70B or Qwen 2.5 72B through haimaker.ai
Open-source models, routed through compliant infrastructure. Your prompts stay off the big providers' training pipelines.
For maximum paranoia, self-host with Ollama or vLLM. You'll need serious hardware (2x A100 or equivalent) and patience for higher latency.
Hybrid approach: Use cloud APIs for general work, switch to open-source for sensitive tasks. Clawdbot makes this easy with model overrides.
Provider comparison
Anthropic (Claude)
Premium pricing ($3-$75 per million output tokens). Best tool calling and instruction following. No training on API data by default.
Claude has become the default for coding agents. The tool use is just more reliable than the alternatives.
OpenAI (GPT)
Mid-tier pricing ($0.60-$15). Solid general performance, fast responses. GPT-4o is a good all-rounder. The mini variant works well for high-volume, simple tasks.
Google (Gemini)
Competitive pricing ($1.25-$10). That massive context window is the selling point. Great for document-heavy workflows.
Open source through haimaker.ai
5% below market rate ($0.10-$5 per million tokens). Routes requests across GPU providers for cost and latency optimization. Avoids the compliance headaches of sending data to US hyperscalers.
The API is OpenAI-compatible:
curl https://api.haimaker.ai/v1/chat/completions \
-H "Authorization: Bearer $HAIMAKER_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "llama-3.3-70b",
"messages": [{"role": "user", "content": "Hello!"}]
}'
Swap your base URL to https://api.haimaker.ai/v1 and you're done.
OpenClaw configuration
Setting your default model in ~/.openclaw/openclaw.json:
{
agents: { defaults: { model: { primary: "anthropic/claude-sonnet-4-20250514" } } }
}
Switch models mid-session with /model opus or /model haimaker/llama-3.3-70b.
Migration note: If you're coming from Clawdbot, your config files are automatically migrated. The CLI now uses
openclawinstead ofclawdbotcommands.
Adding haimaker.ai as a provider in OpenClaw
{
env: { HAIMAKER_API_KEY: "sk-..." },
agents: {
defaults: { model: { primary: "haimaker/llama-3.3-70b" } }
},
models: {
mode: "merge",
providers: {
haimaker: {
baseUrl: "https://api.haimaker.ai/v1",
apiKey: "${HAIMAKER_API_KEY}",
api: "openai-completions",
models: [
{ id: "llama-3.3-70b", name: "Llama 3.3 70B" },
{ id: "qwen-2.5-72b", name: "Qwen 2.5 72B" },
{ id: "mistral-large", name: "Mistral Large" }
]
}
}
}
}
Bottom line
There's no best model. There's the right model for what you're doing.
- Cheap: GPT-4o-mini or open-source through haimaker.ai
- Capable: Opus 4.5 or Gemini 3 Pro
- Private: Open-source through haimaker.ai or self-hosted
Most people should start with Claude Sonnet 4. It handles most tasks well and won't run up a scary bill. Adjust from there based on what you actually need.
Ready to set up your own OpenClaw agent? Visit openclaw.ai to get started or check out the OpenClaw documentation for detailed configuration options.
