Alternative Model Providers for Claude Code
Introduction
Claude Code can be configured to use alternative model providers beyond the default Anthropic API. This tutorial covers how to set up OpenRouter for cloud-based model routing and Ollama for local model inference.
OpenRouter
OpenRouter offers access to the same models from multiple providers and can automatically route requests between them, giving you access to various models including Claude, GPT, and other providers through a unified API.
You Will Need
- Claude Code installed
- An OpenRouter account and API key
1. Create OpenRouter Preset (Optional)
OpenRouter can route your requests automatically or you can create custom presets for specific routing behavior.
Option A: Automatic High-Throughput Routing
Navigate to Settings > Routing in your OpenRouter dashboard to enable automatic routing to the provider with the highest throughput.
Option B: Custom Preset
Create a preset with your own custom name/slug for more control:
- Go to Settings > Presets in your OpenRouter dashboard
- Click "Create Preset"
- Configure your preferred providers and routing rules
- Save with a memorable slug (e.g.,
cerebras-glm-4-7-cerebras)
2. Configure Environment
Set the following environment variables to point Claude Code at the OpenRouter API:
export ANTHROPIC_BASE_URL=https://openrouter.ai/api
export ANTHROPIC_AUTH_TOKEN=YOUR_OPENROUTER_KEY
export ANTHROPIC_API_KEY=""
The ANTHROPIC_API_KEY is intentionally left empty. The ANTHROPIC_AUTH_TOKEN is used for authentication with OpenRouter.
3. Start Claude Code
Once your environment is configured, start Claude Code with your desired model.
Using a Custom Preset
claude --model @preset/cerebras-glm-4-7-cerebras
Using Standard Models
Anthropic Claude:
claude --model anthropic/claude-sonnet-4.5
Other Providers:
claude --model z-ai/glm-4.7
claude --model openai/gpt-oss-120b
Available Models
OpenRouter provides access to models from many providers. Check the OpenRouter models page for a complete list of available models and their capabilities.
Troubleshooting
Authentication Errors
Ensure your ANTHROPIC_AUTH_TOKEN contains a valid OpenRouter API key and that ANTHROPIC_API_KEY is set to an empty string.
Model Not Found
Verify the model identifier matches the format used by OpenRouter (e.g., provider/model-name). Check the OpenRouter dashboard for the exact model identifiers.
Ollama
Ollama lets you run large language models locally on your machine. You can configure Claude Code to use Ollama as a backend for local inference.
You Will Need
- Claude Code installed
- Ollama installed and running locally
- At least one model pulled (e.g.,
ollama pull llama3)
Configure Environment
Set the following environment variables to point Claude Code at your local Ollama instance:
export ANTHROPIC_AUTH_TOKEN=ollama
export ANTHROPIC_BASE_URL=http://localhost:11434
The ANTHROPIC_AUTH_TOKEN is set to ollama as a placeholder. Ollama doesn't require authentication for local requests.
Start Claude Code
Once configured, start Claude Code with your desired local model:
claude --model glm-4.7:cloud
Use the model name as it appears in ollama list. For example:
claude --model llama3:latest
claude --model codellama:7b
Troubleshooting
Connection Refused
Ensure Ollama is running. Start it with:
ollama serve
Model Not Found
Pull the model first with ollama pull <model-name>, then try again.
For more information about Claude Code, visit the official documentation.