LLM Router is a blazing-fast, developer-first AI Gateway. It sits between your application and AI providers (like OpenAI, Anthropic, and Google) to dynamically optimize every request for cost, latency, capabilities, and security. Stop paying Opus-4.6 prices for simple tasks, and stop leaking sensitive PII to third-party models. LLM Router gives you granular control over your AI infrastructure with zero code rewrites.
Zero Friction: LLM Router is a 100% drop-in replacement for the OpenAI
API. Just change your
baseURL and apiKey and you are instantly optimized.Why LLM Router?
Building AI applications is easy. Scaling them efficiently is hard. We built LLM Router to solve the biggest headaches in AI engineering:- Skyrocketing Token Costs: Developers waste millions of tokens sending bloated chat histories, unused tools, and heavy images to expensive models.
- Model Lock-in & Downtime: When OpenAI goes down, your app goes down.
- Data Privacy: Users accidentally paste passwords, API keys, and PII (e.g: Credit cards) into chat prompts.
- Lack of Control — Fine-tune exactly how and where each request is handled.
Routing Configuration: Dashboard vs Per-Request
You have full flexibility when configuring routing behavior:-
Via the LLM Router Dashboard (Recommended for most cases)
Configure default routing rules, tags, models, and preferences directly on your API keys. This includes assigning tags (likecoding,ui design,testing), enabling Zero Data Retention (ZDR), context optimization settings, and more. These settings apply automatically to all requests using that API key. -
Via the Request Payload (Per-request override)
You can also pass routing configuration directly in every API call using thegatewayobject. This gives you maximum flexibility for dynamic behavior.
Core Capabilities
Intelligent Tag Routing
Assign tags to API keys or pass them per request. We combine your business
rules with real-time prompt analysis to route to the best model
automatically.
Aggressive Context Pruning
Reduce input costs by up to 80%. Automatically drop irrelevant history,
strip unused tools, and remove unnecessary media.
Plug-and-Play Skills
Install Skills from the catalog or your own GitHub repos. Dynamically inject
targeted instructions only when relevant.
Zero Data Retention (ZDR)
Enforce strict privacy by routing only to providers with Zero Data Retention
guarantees — per key or per request.
How It Works
LLM Router acts as an intelligent proxy. When a request comes in from your app, our internal engine analyzes the prompt. Depending on your configured rules and tags, it will:- Redact sensitive data.
- Prune bloated context.
- Attach requested Skills to the prompt.
- Score the complexity of the request.
- Route the optimized prompt to the most cost-effective upstream provider.