跳转到内容

LiteLLM

此内容尚不支持你的语言。

LiteLLM is an open-source LLM gateway that provides a unified API to 100+ model providers. Route CoderClaw through LiteLLM to get centralized cost tracking, logging, and the flexibility to switch backends without changing your CoderClaw config.

  • Cost tracking — See exactly what CoderClaw spends across all models
  • Model routing — Switch between Claude, GPT-4, Gemini, Bedrock without config changes
  • Virtual keys — Create keys with spend limits for CoderClaw
  • Logging — Full request/response logs for debugging
  • Fallbacks — Automatic failover if your primary provider is down
Terminal window
coderclaw onboard --auth-choice litellm-api-key
  1. Start LiteLLM Proxy:
Terminal window
pip install 'litellm[proxy]'
litellm --model claude-opus-4-6
  1. Point CoderClaw to LiteLLM:
Terminal window
export LITELLM_API_KEY="your-litellm-key"
coderclaw

That’s it. CoderClaw now routes through LiteLLM.

Terminal window
export LITELLM_API_KEY="sk-litellm-key"
{
models: {
providers: {
litellm: {
baseUrl: "http://localhost:4000",
apiKey: "${LITELLM_API_KEY}",
api: "openai-completions",
models: [
{
id: "claude-opus-4-6",
name: "Claude Opus 4.6",
reasoning: true,
input: ["text", "image"],
contextWindow: 200000,
maxTokens: 64000,
},
{
id: "gpt-4o",
name: "GPT-4o",
reasoning: false,
input: ["text", "image"],
contextWindow: 128000,
maxTokens: 8192,
},
],
},
},
},
agents: {
defaults: {
model: { primary: "litellm/claude-opus-4-6" },
},
},
}

Create a dedicated key for CoderClaw with spend limits:

Terminal window
curl -X POST "http://localhost:4000/key/generate" \
-H "Authorization: Bearer $LITELLM_MASTER_KEY" \
-H "Content-Type: application/json" \
-d '{
"key_alias": "coderclaw",
"max_budget": 50.00,
"budget_duration": "monthly"
}'

Use the generated key as LITELLM_API_KEY.

LiteLLM can route model requests to different backends. Configure in your LiteLLM config.yaml:

model_list:
- model_name: claude-opus-4-6
litellm_params:
model: claude-opus-4-6
api_key: os.environ/ANTHROPIC_API_KEY
- model_name: gpt-4o
litellm_params:
model: gpt-4o
api_key: os.environ/OPENAI_API_KEY

CoderClaw keeps requesting claude-opus-4-6 — LiteLLM handles the routing.

Check LiteLLM’s dashboard or API:

Terminal window
# Key info
curl "http://localhost:4000/key/info" \
-H "Authorization: Bearer sk-litellm-key"
# Spend logs
curl "http://localhost:4000/spend/logs" \
-H "Authorization: Bearer $LITELLM_MASTER_KEY"
  • LiteLLM runs on http://localhost:4000 by default
  • CoderClaw connects via the OpenAI-compatible /v1/chat/completions endpoint
  • All CoderClaw features work through LiteLLM — no limitations