May Townhall: Product + Roadmap Updates
Join the LiteLLM May townhall on Monday, 18 May at 7:30 AM PST to learn about LiteLLM's product updates and roadmap.
AI Gateway
How we build the world's most widely used open-source AI Gateway. Routing, reliability, observability, and what we learn along the way.
We're hiring!Routing to 100+ providers
Join the LiteLLM May townhall on Monday, 18 May at 7:30 AM PST to learn about LiteLLM's product updates and roadmap.
Spawn sandboxed agent sessions on the LiteLLM Gateway β a control plane for managed agents, now in public preview.
CVE-2026-42208 (SQL injection in LiteLLM Proxy's API key verification path) is fixed. Upgrade to v1.83.10-stable.
Dropping `-stable` and `-nightly` suffixes. Weekly releases bump MINOR; PATCH is now reserved for actual hotfixes. Old releases keep their tags forever; new ones start with `1.84.0`.
Use generally available gemini-embedding-2 for multimodal embeddings on LiteLLM via Gemini API and Vertex AIβthe same flows as preview, stable model id.
Day 0 support for GPT-5.5 and GPT-5.5 Pro on LiteLLM.
CVE-2026-30623 (authenticated RCE via MCP stdio transport) is fixed. Upgrade to v1.83.6-nightly or v1.83.7-stable or later.
Chain Akto's model-based detection with LiteLLM's built-in guardrails β catch PII, prompt injection, and policy violations that pattern-based checks miss.
Day 0 support for Claude Opus 4.7 on LiteLLM AI Gateway - use across Anthropic, Azure, Vertex AI, and Bedrock.
How LiteLLM's production AI Gateway handles Redis degradation at scale without cascading failures β circuit breaker pattern, 0ms fast-fail, automatic recovery.
A recap of the April LiteLLM town hall covering CI/CD v2, product stability work, and the near-term roadmap.
Disclosure of security vulnerabilities fixed in LiteLLM v1.83.0, and the launch of our bug bounty program.
Join the LiteLLM April townhall on Friday, 10 April at 7:30 AM to learn about LiteLLM's security and product roadmap.
CI/CD v2 introduces isolated environments, stronger security gates, and safer release separation for LiteLLM.
LiteLLM is partnering with Vanta on SOC 2 Type 2 and ISO 27001 recertification and engaging independent auditors for verification.
What happened, what we've done, and what comes next for LiteLLM's release and security processes.
As of 2:00 PM ET on March 24, 2026
Date: March 18, 2026
GPT-5.4-mini and GPT-5.4-nano model support in LiteLLM
LiteLLM now supports creating, retrieving, and managing reusable video characters across multiple video generations.
Use the LiteLLM proxy to route OpenAI-style WebRTC realtime via HTTP: client_secrets and SDP exchange.
GPT-5.4 model support in LiteLLM
Guide to using Gemini 3.1 Flash Lite Preview on LiteLLM Proxy and SDK with day 0 support.
Date: February 27, 2026
Day 0 support for GPT-5.3-Codex on LiteLLM, including phase parameter handling for Responses API.
Date: Feb 24, 2026
Date: Feb 23, 2026
Date: January 22, 2026
Guide to using Gemini 3.1 Pro on LiteLLM Proxy and SDK with day 0 support.
Date: Feb 16, 2026
Day 0 support for Claude Sonnet 4.6 on LiteLLM AI Gateway - use across Anthropic, Azure, Vertex AI, and Bedrock.
Date: February 13, 2026
Day 0 support for MiniMax-M2.5 on LiteLLM
Date: January 27, 2026
How we improved LiteLLM proxy latency and throughput by replacing a single middleware base class
How we built a long-running, release-validation system to catch regressions before they reach users.
Day 0 support for Claude Opus 4.6 on LiteLLM AI Gateway - use across Anthropic, Azure, Vertex AI, and Bedrock.
Our Q1 performance target and architectural direction for achieving sub-millisecond proxy overhead on modest hardware.
Guide to using Gemini 3 Flash on LiteLLM Proxy and SDK with day 0 support.
Guide to Claude Opus 4.5 and advanced features in LiteLLM: Tool Search, Programmatic Tool Calling, and Effort Parameter.
Common questions and best practices for using gemini-3-pro-preview with LiteLLM Proxy and SDK.
Generate embeddings from text, images, audio, video, and PDFs with gemini-embedding-2-preview on LiteLLM via Gemini API (one vector per input, OpenAI-compatible) and Vertex AI (single unified vector per request).