Agent skill

switchailocal

Stars 163
Forks 31

Install this agent skill to your Project

npx add-skill https://github.com/majiayu000/claude-skill-registry/tree/main/skills/data/switchailocal

SKILL.md

switchAILocal Proxy

Unified LLM proxy for AI agents. Always use http://localhost:18080/v1 as your base URL.

The killer feature: Use your paid CLI subscriptions (Gemini Pro, Claude Pro, etc.) via the API - it's FREE because you already pay for the subscription!


Quick Start

1. Make a request (FREE with CLI)

bash
curl http://localhost:18080/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{
    "model": "geminicli:",
    "messages": [{"role": "user", "content": "Hello!"}]
  }'

2. Configure Python Client

python
from openai import OpenAI
client = OpenAI(base_url="http://localhost:18080/v1", api_key="sk-test-123")
response = client.chat.completions.create(model="geminicli:", messages=[{"role": "user", "content": "Hi!"}])

πŸ—ΊοΈ Skill Files

File Description
SKILL.md (this file) Core workflow and endpoint reference
references/routing.md Intelligent routing and matrix setup
references/multimodal.md Vision and image processing
references/examples.md Real-world agentic use cases
references/management-api.md Full Monitoring & Operations API
references/steering.md Conditional routing rules
references/hooks.md Automation and event hooks
references/memory.md Analytics and history

⚠️ Critical: Model Format

NEVER use bare model names. Format is ALWAYS provider: or provider:model.

❌ Wrong βœ… Correct Why
gemini-2.5-pro geminicli:gemini-2.5-pro Needs provider prefix
claude-3-5-sonnet claudecli: claudecli: uses default
llama3 ollama:llama3 Needs provider prefix

πŸ—οΈ Provider Reference

1. CLI Providers (FREE!)

Uses your human's CLI subscriptions. Best for agents.

Prefix CLI Subscription Required
geminicli: gemini Google AI Premium/Pro
claudecli: claude Claude Pro/Max
codex: codex OpenAI Plus
vibe: vibe Mistral Le Chat

2. Local & Cloud

Prefix Source Cost
ollama: Local Ollama FREE
auto Local Cortex FREE (Requires plugin)
switchai: Traylinx Cloud Per-token
groq: Groq Cloud Per-token

πŸš€ Core Features

CLI Attachments & Flags

Pass local context and control autonomy via CLI extensions.

json
{
  "model": "geminicli:",
  "messages": [{"role": "user", "content": "Fix this code"}],
  "extra_body": {
    "cli": {
      "attachments": [{"type": "folder", "path": "./src"}],
      "flags": {"auto_approve": true, "yolo": true}
    }
  }
}

Streaming

Add "stream": true to any request for SSE token streaming.


🌲 Decision Tree

What do you need?
β”œβ”€ FREE + Powerful + Files
β”‚   └─ CLI Providers (geminicli:, claudecli:)
β”œβ”€ FREE + Private + Fast
β”‚   └─ Local Ollama (ollama:llama3.2)
β”œβ”€ Ultra-Fast Production
β”‚   └─ Groq Cloud (groq:llama-3.3-70b)
└─ I don't know, you pick
    └─ Intelligent Routing (auto)

πŸ› οΈ Troubleshooting & Best Practices

Problem Fix
Connection error Check if server is running on port 18080
Model not found Ensure you used the provider: prefix
401 Unauthorized Check API key in config.yaml

Best Practices

  1. Prefer CLI Providers: They are free and support file attachments.
  2. Check Status: Use GET /v1/providers to see what is active.
  3. Use auto: For simple tasks, let the router pick the best model.
  4. Local for Privacy: Use ollama: for confidential data.

Route wisely. Save tokens. Use CLI. πŸš€

Didn't find tool you were looking for?

Be as detailed as possible for better results