Do not index
_(1).png?table=block&id=31ffda42-9937-8047-a8ab-dbcc8dd4b45f&cache=v2)
When setting up OpenClaw, one of the most important decisions is choosing the AI model that will power your agent.
OpenClaw itself is the system that connects everything together — your messaging channels, automation workflows, and AI tools. But the actual responses and reasoning come from the AI model you select during configuration.
If you open the OpenClaw configuration wizard, you’ll notice that several AI providers are supported. Each model behaves slightly differently depending on speed, reasoning ability, and cost.
In this guide, we’ll look at the AI models shown during the OpenClaw setup process, compare them, and help you understand which option might be best for your use case.
Video Walkthrough
If you'd like to see the setup and model selection visually, watch the walkthrough below.
The video demonstrates how OpenClaw interacts with AI models during configuration and how the system works once the model is connected.
Where AI Models Appear in OpenClaw Setup
During the configuration process, OpenClaw asks you to choose a model provider.
At this point, the CLI interface shows a list of supported providers.
From the setup interface shown in the screenshots, the available providers include:
- OpenAI
- Anthropic
- MiniMax
- Moonshot AI (Kimi)
- XAI
- OpenRouter
- Qwen
- GLM
- Copilot
- Venice AI
- Vercel AI Gateway
Once you select one of these providers, OpenClaw then asks for an API key so the agent can communicate with that model.
AI Models Supported by OpenClaw
Below is a simplified comparison based on the providers visible in the setup interface.
Provider | Example Model | What It’s Good For |
OpenAI | GPT models | General AI tasks and automation |
Anthropic | Claude models | Reasoning and long responses |
Moonshot AI | Kimi models | Long context and coding tasks |
Google | Gemini models | Multimodal workflows |
MiniMax | MiniMax models | Fast responses |
OpenRouter | Multiple aggregated models | Switching between providers |
Qwen | Qwen models | Alternative AI models |
GLM | GLM models | AI chat and automation |
Copilot | GitHub Copilot models | Coding workflows |
This flexibility is one of the biggest advantages of OpenClaw — you’re not locked into a single AI ecosystem.
How Model Selection Works in OpenClaw
When running the configuration command inside the terminal, OpenClaw asks which sections you want to configure.
Once you select Model, the system shows the list of providers mentioned earlier.
After selecting the provider:
- OpenClaw requests the API key
- The key is validated
- The selected model becomes the default AI engine
From that point forward, every request sent by the OpenClaw agent will use that AI model.
Example Workflow
A typical setup might look like this:
- Start OpenClaw configuration
- Select Model provider
- Choose a provider (for example OpenAI or Anthropic)
- Enter the API key
- Save the configuration
Once complete, the AI model is ready to process requests from messaging channels such as Telegram or WhatsApp.
Choosing the Right Model
Different users choose different models depending on how they plan to use OpenClaw.
For example:
Use Case | Suggested Model |
General chatbot | OpenAI |
Advanced reasoning | Anthropic |
Long conversations | Kimi |
Experimenting with multiple models | OpenRouter |
Coding assistance | Copilot |
Since OpenClaw allows switching models later, you can experiment with different providers without rebuilding your setup.
Why OpenClaw Supports Multiple Models
The goal of OpenClaw is flexibility.
Instead of forcing users to rely on a single AI provider, the platform allows multiple integrations so developers can:
- test different models
- compare response quality
- control API costs
- adapt to new AI models as they appear
This design makes OpenClaw useful for both experimentation and production systems.
Final Thoughts
Choosing the right AI model is a key step when setting up OpenClaw.
From the configuration interface shown in the setup screenshots, OpenClaw supports a wide range of providers including OpenAI, Anthropic, Moonshot AI, Google, and several others.
The best approach is usually to start with a model you are familiar with, test it inside your workflow, and adjust later if needed.
Because OpenClaw allows you to change models easily, you’re free to experiment until you find the setup that works best for your AI agent.