The only proxy you need for Claude Code CLI
Quick Start • Features • Web Dashboard • Crosstalk
Route Claude Code to any provider. Save 90% on API costs. Run locally for free.
The Ultimate Proxy sits between Claude Code CLI and your chosen API provider. It translates Anthropic's API format to OpenAI-compatible format, letting you use any model with Claude Code:
Claude Code CLI → The Ultimate Proxy → Any Provider
├─ OpenRouter
├─ Gemini / VibeProxy
├─ OpenAI
├─ Azure
├─ Ollama (local)
└─ LM Studio (local)
# Clone & install
git clone https://github.com/aaaronmiller/the-ultimate-proxy.git
cd the-ultimate-proxy
uv sync
# Interactive setup wizard
python start_proxy.py --setup
# Start the proxy
python start_proxy.pyThen in another terminal:
export ANTHROPIC_BASE_URL=http://localhost:8082
claudeThe easiest way to use Claude Code with premium models for free:
- Install VibeProxy: Download
- Authenticate: Sign in with Google (Antigravity OAuth)
- Setup:
python start_proxy.py --setup→ Select "VibeProxy/Antigravity"
What you get:
- 🧠 Claude Opus 4.5 with 128k thinking tokens
- ⚡ Gemini 3 Pro/Flash
- 📊 BIG/MIDDLE/SMALL model routing
- 📈 Usage tracking & analytics
- 💸 No API keys or billing required
| Feature | Description |
|---|---|
| Multi-Provider | OpenRouter, Gemini, OpenAI, Azure, Ollama, LM Studio |
| Model Routing | BIG/MIDDLE/SMALL tiers with intelligent routing |
| Web Dashboard | Real-time monitoring with 2025 glassmorphism UI |
| Crosstalk | Model-to-model conversations (up to 8 models) |
| Usage Tracking | Cost analytics and token metrics |
| Extended Thinking | Up to 128k thinking tokens for reasoning models |
| Prompt Injection | Add custom prompts showing routing info |
| Model Cascade | Automatic fallback on provider errors |
Access at http://localhost:8082 when running.
- 2025 Glassmorphism UI with aurora gradients
- Real-time request monitoring
- Provider configuration
- Model selection with hybrid routing
- WebSocket live log streaming
- Profile management
Multi-model conversations where AI models talk to each other:
# Launch visual TUI
python start_proxy.py --crosstalk-studio
# Quick setup
python start_proxy.py --crosstalk "claude-opus,gemini-pro" --topic "Explore consciousness"Features:
- Up to 8 models in a conversation
- Multiple paradigms: relay, debate, memory, report
- Jinja templates for message formatting
- Backrooms-compatible import/export
- MCP integration for programmatic access
# Configuration
python start_proxy.py --setup # First-time wizard
python start_proxy.py --settings # Unified settings TUI
python start_proxy.py --doctor # Health check + auto-fix
# Model management
python start_proxy.py --select-models # Interactive model selector
python start_proxy.py --set-big MODEL # Quick set BIG model
python start_proxy.py --show-models # List available models
# Diagnostics
python start_proxy.py --config # Show configuration
python start_proxy.py --dry-run # Validate without starting
python start_proxy.py --analytics # View usage statsthe-ultimate-proxy/
├── start_proxy.py # Main entry point
├── .env # Configuration
├── CROSSTALK.md # Multi-model chat docs
│
├── src/
│ ├── core/ # Proxy core logic
│ ├── api/ # FastAPI routes + WebSocket
│ ├── services/ # Providers, models, prompts
│ └── cli/ # CLI tools and TUIs
│
├── configs/crosstalk/ # Crosstalk presets & templates
├── web-ui/ # Svelte + bits-ui dashboard
└── docs/ # Extended documentation
401 Unauthorized
python start_proxy.py --doctor # Auto-fix API keysModel Not Found
python start_proxy.py --show-models # List available modelsConnection Refused
- Ensure proxy is running:
python start_proxy.py - Check port 8082 is not in use
- Desktop GUI (Tauri)
- Multi-instance analytics
- MCP Server integration
- Multi-agent orchestration ("Swarm Mode")
The Ultimate Proxy • Made with ❤️ for the Claude Code community