Source Code
LLM Supervisor
Automatically switches OpenClaw between cloud and local Ollama models when rate limits occur.
Features
- Detects rate-limit / overload errors from Anthropic/OpenAI
- Switches to a local Ollama fallback model
- Requires explicit confirmation before local code generation
- Supports manual commands:
Commands
/llm status/llm switch cloud/llm switch local
Default Local Model
qwen2.5:7b
Safety
Local code generation requires the user to type:
CONFIRM LOCAL CODE