You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Open-source LLM router & AI cost optimizer. Routes simple prompts to cheap/local models, complex ones to premium — automatically. Drop-in OpenAI-compatible proxy for Claude Code, Codex, Cursor, OpenClaw. Saves 40-70% on AI API costs. Self-hosted, no middleman.
SmarterRouter: An intelligent LLM gateway and VRAM-aware router for Ollama, llama.cpp, and OpenAI. Features semantic caching, model profiling, and automatic failover for local AI labs.
A personal LLM gateway with fault-tolerant capabilities for calls to LLM models from any provider with OpenAI-compatible APIs. Advanced features like retry, model sequencing, and body parameter injection are also available. Especially useful to work with AI coders like Cline and RooCode and providers like OpenRouter.
Use any LLM with Claude Code — proxy that translates Anthropic API to OpenAI, Gemini, DeepSeek, Ollama, and more. Full tool calling, streaming, ReAct XML fallback, hot-reload config.
Python proxy for Gemini API. Overcomes the tight free-tier rate limits of Gemini Pro by key pooling and provides full OpenAI compatibility for OpenWebUI.
AI development environment with 90% cost savings. Routes between 8 LLM providers while defaulting to FREE local models. Production-ready with automated testing.
Ready to go local AI gateway: LiteLLM proxy + Langfuse observability + Caddy reverse proxy. Per-tool tracing, session evaluation, and zero-config trace enrichment.
LLM cost-optimization proxy which routes requests across 11 models from 4 providers using complexity classification, two-level semantic caching, and automatic provider failover