Developer ToolsTempl
Type-safe HTML templating language for Go with compile-time safety
LiteLLM is the open-source AI gateway that gives you one OpenAI-format API across 100+ LLM providers, with virtual keys, spend tracking, fallbacks, and guardrails. Used in production by Stripe, Netflix, and Google.
LiteLLM is an open-source AI gateway that lets you call 100+ large-language-model providers — OpenAI, Anthropic, Gemini, Bedrock, Azure, Groq, Mistral, Ollama and more — through a single OpenAI-compatible interface, and layers on virtual keys, spend tracking, guardrails, fallbacks, and observability. We rate it 84/100 — it is the default abstraction layer for teams that want to stay multi-model without rewriting their stack every time a new frontier model ships, and it scales from a one-line Python SDK import to a production proxy deployed by Stripe, Netflix and Google.
LiteLLM is built by BerriAI, a Y Combinator W23 company founded by Ishaan Jaffer and Krrish Dholakia. The first public release landed in 2023 as a thin Python shim that normalized OpenAI, Anthropic and Cohere SDKs, and the project has since grown into a full-blown AI gateway with an admin UI, virtual-key management, budgets, logging pipelines, and MCP and A2A protocol bridges. As of the latest release v1.83.7-stable on , the repository has 43.9k GitHub stars, 7.4k forks, and 1,438 contributors. OSS adopters publicly listed in the repo include Stripe, Netflix, Google ADK, OpenAI Agents SDK, Greptile and OpenHands.
The specific problem LiteLLM solves is multi-provider fragmentation. Every LLM vendor ships its own SDK, its own authentication pattern, its own request and response shape, and its own error taxonomy. Swap providers for cost, latency or reliability reasons and you are rewriting glue code. LiteLLM collapses all of that into a single OpenAI-shaped completion() call — or, if you deploy the proxy, a single base_url your team can point at from any language. That one indirection is what makes multi-model strategies tractable.
/chat/completions, /responses, /embeddings, /images, /audio, /rerank and /messages endpoints.openai, LangChain, LlamaIndex, Instructor) at http://localhost:4000 and it just works — no code changes needed.
On Hacker News, the most-upvoted thread captures the split sentiment well: developers praise how quickly LiteLLM exposes new models — several teams report being able to offer a just-released frontier model to their users on launch day without any code changes — but also flag that the proxy is “kind of a mess TBH” when you run it at scale. The most recurring complaints across Reddit's r/LocalLLaMA and r/MachineLearning are (1) gradual memory growth that requires periodic worker recycling, (2) a roughly 500 µs fixed overhead per request that some teams cite as painful for short completions, and (3) debugging requests through a proxy layer when a provider returns a weird shape. On Product Hunt the Python SDK has a near-perfect rating; the pain points cluster around the self-hosted proxy deployment, not the library itself.
LiteLLM is free and open source under the MIT license for the SDK and a permissive license for the community proxy. The commercial offering is an Enterprise tier aimed at regulated and large-scale deployments.
| Plan | Price | Key Limits |
|---|---|---|
| Open Source | $0 | 100+ provider integrations, logging, load balancing, guardrails, virtual keys, unlimited self-hosted usage |
| Enterprise | Custom (contact sales) | Adds JWT auth, SSO/SAML, SOC 2 audit logs, managed upgrades, priority support, SLA |
| Hosted Cloud | Usage-based | BerriAI-managed proxy for teams that don't want to run infra |
There are no per-seat fees on the open-source tier and no cap on request volume — the only costs you pay are to the underlying model providers you route through LiteLLM.
Best for: Backend and ML platform teams who already make more than one LLM call a second, want a single internal API for multiple providers, and need per-team budgets and audit logs. Especially strong for regulated industries (finance, healthcare) that need a self-hosted proxy with no data leaving their VPC.
Not ideal for: Solo developers building a prototype — in that case the litellm Python SDK alone is enough, the full proxy is overkill. Also skip it if you're happy being all-in on one provider; there's no value in an abstraction you never use.
Pros:
Cons:
OpenRouter offers a managed, hosted alternative with no self-hosting required — simpler, but you lose data control and self-hosted audit trails. Langfuse is complementary rather than competitive — it handles tracing and evals while LiteLLM handles routing, and the two are commonly paired. Portkey offers a closed-source SaaS gateway with a nicer UI but a less permissive licence.
For any team that has moved past the prototype stage and is making LLM calls against more than one provider, LiteLLM is the default choice — it is what Stripe, Netflix and Google build on top of, and it has more frontier-model coverage than any managed alternative. The rough edges (memory growth, 500 µs overhead, operational footprint) are real but well-documented, and the Enterprise tier exists precisely to carry those for you. Our 84/100 reflects a tool that is best-in-class at its core job but still requires platform-engineering maturity to run in production.
Canva Unveils AI 2.0 With Agentic Design and Conversational Editing (April 2026)
Canva launched Canva AI 2.0 on April 16, 2026 at Canva Create in Los Angeles — a conversational, agentic design platform with persistent memory, Gmail/Slack/Notion connectors and a new in-house foundation model the company says is 7× faster and 30× cheaper than comparable frontier models.
Apr 20, 2026
GitHub CLI Ships 'gh skill' — A Package Manager for AI Agent Instructions (April 2026)
GitHub on April 16, 2026 added a new 'gh skill' command group to its CLI, bringing package-manager-style install, update, and publish workflows to the portable agent skills that run inside Copilot, Claude Code, Cursor, Codex, Gemini CLI and Antigravity. The feature shipped in public preview as part of GitHub CLI v2.90.0.
Apr 19, 2026
Wayve Extends Series D to $1.26B With $60M From AMD, Arm and Qualcomm (April 2026)
UK-based autonomous-driving startup Wayve on April 15, 2026 extended its Series D round to $1.26 billion after AMD, Arm and Qualcomm Ventures invested a combined $60M, lining up the four largest automotive silicon vendors behind its embodied-AI self-driving platform ahead of a Tokyo robotaxi pilot with Uber and Nissan later this year.
Apr 19, 2026
Is this product worth it?
Built With
Compare with other tools
Open Comparison Tool →