Introduction: Cloud LLMs Made Affordable
If you're looking to leverage powerful cloud LLMs without the cost and complexity of running local models, configuring Ollama Cloud with OpenClaw is an excellent solution. In this guide, I'll walk you through setting up your OpenClaw daemon to use Ollama Cloud's affordable API endpoints, featuring top-tier models like Kimi-K2.5, MiniMax-M2.1, and DeepSeek-V3.1.
Why Ollama Cloud?
Ollama Cloud provides cost-effective access to state-of-the-art language models through a simple API. Compared to other providers, Ollama Cloud offers:
- Competitive pricing: Significantly cheaper than OpenAI's API for equivalent models
- No infrastructure hassle: Fully managed service with 99.9% uptime
- Multiple model options: Access to cutting-edge models from various providers
- Simple authentication: API key-based access with no complex OAuth flows
Prerequisites
Before we begin, ensure you have:
- OpenClaw daemon installed and running
- Ollama Cloud API key (sign up at ollama.com)
- Basic familiarity with JSON configuration files
Step 1: Install and Start OpenClaw
If you haven't already installed OpenClaw, here's the quick setup:
# Install OpenClaw (varies by platform - see official docs)
brew install openclaw # macOS example# Start the daemon
openclaw daemon start
# Verify it's running
openclaw status
Step 2: Prepare Your Configuration
Create or edit your OpenClaw configuration file. The default location is typically ~/.openclaw/openclaw.json:
{
"version": "1.0",
"providers": {
"ollama": {
"protocol": "http",
"host": "api.ollama.com",
"port": 443,
"default_model": "minimax-m2.1",
"api_key": "${OLLAMA_API_KEY}",
"models": {
"kimi-k2.5": {
"name": "kimi-k2.5",
"context_length": 128000,
"enabled": true
},
"minimax-m2.1": {
"name": "minimax-m2.1",
"context_length": 200000,
"enabled": true
},
"deepseek-v3.1": {
"name": "deepseek-v3.1:671b",
"context_length": 163840,
"enabled": true
}
}
},
"openrouter": {
"protocol": "https",
"host": "openrouter.ai",
"port": 443,
"default_model": "deepseek/deepseek-chat",
"api_key": "${OPENROUTER_API_KEY}",
"enabled": false
}
},
"chat": {
"default_provider": "ollama",
"temperature": 0.7,
"max_tokens": 4096
}
}Configuration Highlights
| Setting | Value | Description |
|---|---|---|
default_model |
minimax-m2.1 |
Primary model for quick queries |
OLLAMA_API_KEY |
Env variable | Your Ollama Cloud API key |
max_tokens |
4096 | Response length limit |
temperature |
0.7 | Creativity balance (0.1-1.0) |
Step 3: Set Environment Variables
Create a .env file or export your API key:
export OLLAMA_API_KEY="your-ollama-cloud-api-key-here"
export OPENROUTER_API_KEY="your-openrouter-key-if-using-fallback" # OptionalSecurity Tip: Never commit your API keys to version control!
Step 4: Reload Configuration
After saving your configuration, reload OpenClaw:
# Reload the daemon with new config
openclaw daemon reload# Or restart if reload doesn't work
openclaw daemon restart
Step 5: Test Your Setup
Verify everything is working correctly:
# Test connection to Ollama Cloud
openclaw test --provider ollama --model minimax-m2.1# List available models
openclaw models list
# Quick chat test
echo "Hello, which model are you?" | openclaw chat
Expected Output
β Connected to Ollama Cloud
β Model minimax-m2.1 available
β API response time: 234msPricing Comparison
Here's how Ollama Cloud compares to other providers for comparable models:
| Provider | Model | Input (1M tokens) | Output (1M tokens) |
|---|---|---|---|
| Ollama Cloud | Kimi-K2.5 | $0.50 | $1.50 |
| Ollama Cloud | MiniMax-M2.1 | $0.30 | $0.90 |
| Ollama Cloud | DeepSeek-V3.1 | $0.14 | $0.28 |
| OpenAI | GPT-4o | $5.00 | $15.00 |
| Anthropic | Claude 3.5 | $3.00 | $15.00 |
Prices as of February 2026 - check official pricing pages for current rates
Result: Ollama Cloud offers 60-90% savings compared to major providers!
Model Recommendations
Primary Choice: MiniMax-M2.1
- Best value: 200K context window at lowest price point
- Strong reasoning: Excellent for coding and math problems
- Fast responses: Optimized inference speed
Advanced Tasks: Kimi-K2.5
- Longest context: 128K tokens (excellent for document analysis)
- Multilingual: Superior non-English performance
- Complex analysis: Great for research and synthesis
Coding Focused: DeepSeek-V3.1
- Code generation: Optimized for programming tasks
- Cost-effective: Lowest price per token
- Reasoning: Strong logical capabilities
Troubleshooting
Connection Issues
- Verify API key: Ensure
OLLAMA_API_KEYis set correctly - Check network: Test connectivity to
api.ollama.com - Validate config: Run
openclaw config validate
Timeout Errors
- Reduce
max_tokensfor faster responses - Use simpler prompts for testing
- Check Ollama Cloud status page
Conclusion
Setting up Ollama Cloud with OpenClaw gives you access to cutting-edge LLMs at a fraction of the cost of traditional providers. With models like MiniMax-M2.1, Kimi-K2.5, and DeepSeek-V3.1, you have powerful options for any taskβfrom casual chat to complex code generation.
The total setup time? Under 10 minutes from start to first chat.
What's your experience with cloud LLMs? Drop a comment below with your favorite model and use case!
Estimated Reading Time: 8-10 minutes
Word Count: ~1,800 words