[Pre-Release] v1.73.0-stable
This is a pre-release version.
The production version will be released on Wednesday.
Deploy this version​
- Docker
- Pip
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:v1.73.0.rc.1
pip install litellm==1.73.0rc1
TLDR​
- Why Upgrade
- Passthrough Endpoints v2: Enhanced support for subroutes and custom cost tracking for passthrough endpoints.
- Health Check Dashboard: New frontend UI for monitoring model health and status.
- User Management: Set default team for new users - enables giving all users $10 API keys for exploration.
- Who Should Read
- Teams using Passthrough Endpoints
- Teams using User Management on LiteLLM
- Teams using Health Check Dashboard for models
- Teams using Claude Code with LiteLLM
- Risk of Upgrade
- Low
- No major breaking changes to existing functionality.
- Low
Key Highlights​
Passthrough Endpoints v2​
This release brings support for adding billing and full URL forwarding for passthrough endpoints.
Previously, you could only map simple endpoints, but now you can add just /bria
and all subroutes automatically get forwarded - for example, /bria/v1/text-to-image/base/model
and /bria/v1/enhance_image
will both be forwarded to the target URL with the same path structure.
This means you as Proxy Admin can onboard third-party endpoints like Bria API and Mistral OCR, set a cost per request, and give your developers access to the complete API functionality.
Learn more about Passthrough Endpoints
v2 Health Checks​
This release brings support for Proxy Admins to select which specific models to health check and see the health status as soon as its individual check completes, along with last check times.
This allows Proxy Admins to immediately identify which specific models are in a bad state and view the full error stack trace for faster troubleshooting.
Set Default Team for New Users​
v1.73.0 introduces the ability to assign new users to Default Teams. This makes it much easier to enable experimentation with LLMs within your company, while also ensuring spend for exploration is tracked correctly.
What this means for Proxy Admins:
- Set a max budget per team member: This sets a max amount an individual can spend within a team.
- Set a default team for new users: When a new user signs in via SSO / invitation link, they will be automatically added to this team.
What this means for Developers:
- View models across teams: You can now go to
Models + Endpoints
and view the models you have access to, across all teams you're a member of. - Safe create key modal: If you have no model access outside of a team (default behaviour), you are now nudged to select a team on the Create Key modal. This resolves a common confusion point for new users onboarding to the proxy.
New / Updated Models​
Pricing / Context Window Updates​
Provider | Model | Context Window | Input ($/1M tokens) | Output ($/1M tokens) | Type |
---|---|---|---|---|---|
Google VertexAI | vertex_ai/imagen-4 | N/A | Image Generation | Image Generation | New |
Google VertexAI | vertex_ai/imagen-4-preview | N/A | Image Generation | Image Generation | New |
Gemini | gemini-2.5-pro | 2M | $1.25 | $5.00 | New |
Gemini | gemini-2.5-flash-lite | 1M | $0.075 | $0.30 | New |
OpenRouter | Various models | Updated | Updated | Updated | Updated |
Azure | azure/o3 | 200k | $2.00 | $8.00 | Updated |
Azure | azure/o3-pro | 200k | $2.00 | $8.00 | Updated |
Azure OpenAI | Azure Codex Models | Various | Various | Various | New |
Updated Models​
Features​
- Azure
- Support for new /v1 preview Azure OpenAI API - PR, Get Started
- Add Azure Codex Models support - PR, Get Started
- Make Azure AD scope configurable - PR
- Handle more GPT custom naming patterns - PR
- Update o3 pricing to match OpenAI pricing - PR
- VertexAI
- Add Vertex Imagen-4 models - PR, Get Started
- Anthropic streaming passthrough cost tracking - PR
- Gemini
- AWS Bedrock
- Mistral
- Enhance Mistral API with parallel tool calls support - PR
- Meta Llama API
- Enable tool calling for meta_llama models - PR
- Volcengine
- Add thinking parameter support - PR
Bugs​
- VertexAI
- Gemini
- Fix web search error with responses API - PR, Get Started
- Custom LLM
- Set anthropic custom LLM provider property - PR
- Anthropic
- Bump anthropic package version - PR
- Ollama
LLM API Endpoints​
Features​
- Responses API
- Day-0 support for OpenAI re-usable prompts Responses API - PR, Get Started
- Support passing image URLs in Completion-to-Responses bridge - PR
- MCP Gateway
- Add Allowed MCPs to Creating/Editing Organizations - PR, Get Started
- Allow connecting to MCP with authentication headers - PR, Get Started
- Speech API
- Working Gemini TTS support via OpenAI's
/v1/speech
endpoint - PR
- Working Gemini TTS support via OpenAI's
- Passthrough Endpoints
- Add support for subroutes for passthrough endpoints - PR
- Support for setting custom cost per passthrough request - PR
- Ensure "Request" is tracked for passthrough requests on LiteLLM Proxy - PR
- Add V2 Passthrough endpoints on UI - PR
- Move passthrough endpoints under Models + Endpoints in UI - PR
- QA improvements for adding passthrough endpoints - PR, PR
- Models API
- Allow
/models
to return correct models for custom wildcard prefixes - PR
- Allow
Bugs​
- Messages API
- MCP Gateway
- Fix using MCPs defined on config.yaml - PR
- Chat Completion API
- Allow dict for tool_choice argument in acompletion - PR
- Passthrough Endpoints
- Don't log request to Langfuse passthrough on Langfuse - PR
Spend Tracking​
Features​
- User Agent Tracking
- Tag Management
- Support adding public model names in tag management - PR
Management Endpoints / UI​
Features​
- Test Key Page
- Allow testing
/v1/messages
on the Test Key Page - PR
- Allow testing
- SSO
- Allow passing additional headers - PR
- JWT Auth
- Correctly return user email - PR
- Model Management
- Allow editing model access group for existing model - PR
- Team Management
- SCIM
- Health Check Dashboard
- Implement health check backend API and storage functionality - PR
- Add LiteLLM_HealthCheckTable to database schema - PR
- Implement health check frontend UI components and dashboard integration - PR
- Add success modal for health check responses - PR
- Fix clickable model ID in health check table - PR
- Fix health check UI table design - PR
Logging / Guardrails Integrations​
Bugs​
- Prometheus
- Fix bug for using prometheus metrics config - PR
Security & Reliability​
Security Fixes​
Reliability Improvements​
General Proxy Improvements​
Features​
- Deployment
- Custom Root Path
- Fix loading UI on custom root path - PR
- SDK Improvements
- LiteLLM SDK / Proxy improvement (don't transform message client-side) - PR
Bugs​
- Observability
- Fix boto3 tracer wrapping for observability - PR
New Contributors​
- @kjoth made their first contribution in PR
- @shagunb-acn made their first contribution in PR
- @MadsRC made their first contribution in PR
- @Abiji-2020 made their first contribution in PR
- @salzubi401 made their first contribution in PR
- @orolega made their first contribution in PR
- @X4tar made their first contribution in PR
- @karen-veigas made their first contribution in PR
- @Shankyg made their first contribution in PR
- @pascallim made their first contribution in PR
- @lgruen-vcgs made their first contribution in PR
- @rinormaloku made their first contribution in PR
- @InvisibleMan1306 made their first contribution in PR
- @ervwalter made their first contribution in PR
- @ThakeeNathees made their first contribution in PR
- @jnhyperion made their first contribution in PR
- @Jannchie made their first contribution in PR
Demo Instance​
Here's a Demo Instance to test changes:
- Instance: https://demo.litellm.ai/
- Login Credentials:
- Username: admin
- Password: sk-1234