The release candidate is live now.
The production release will be live on Wednesday.
Deploy this version​
- Docker
- Pip
docker run
-e STORE_MODEL_IN_DB=True
-p 4000:4000
ghcr.io/berriai/litellm:main-v1.72.0.rc
pip install litellm==1.72.0
Key Highlights​
LiteLLM v1.72.0-stable.rc is live now. Here are the key highlights of this release:
- Vector Store Permissions: Control Vector Store access at the Key, Team, and Organization level.
- Rate Limiting Sliding Window support: Improved accuracy for Key/Team/User rate limits with request tracking across minutes.
- Aiohttp Transport used by default: Aiohttp transport is now the default transport for LiteLLM networking requests. This gives users 2x higher RPS per instance with a 40ms median latency overhead.
- Bedrock Agents: Call Bedrock Agents with
/chat/completions
,/response
endpoints. - Anthropic File API: Upload and analyze CSV files with Claude-4 on Anthropic via LiteLLM.
- Prometheus: End users (
end_user
) will no longer be tracked by default on Prometheus. Tracking end_users on prometheus is now opt-in. This is done to prevent the response from/metrics
from becoming too large. Read More
Vector Store Permissions​
This release brings support for managing permissions for vector stores by Keys, Teams, Organizations (entities) on LiteLLM. When a request attempts to query a vector store, LiteLLM will block it if the requesting entity lacks the proper permissions.
This is great for use cases that require access to restricted data that you don't want everyone to use.
Over the next week we plan on adding permission management for MCP Servers.
Aiohttp Transport used by default​
Aiohttp transport is now the default transport for LiteLLM networking requests. This gives users 2x higher RPS per instance with a 40ms median latency overhead. This has been live on LiteLLM Cloud for a week + gone through alpha users testing for a week.
If you encounter any issues, you can disable using the aiohttp transport in the following ways:
On LiteLLM Proxy
Set the DISABLE_AIOHTTP_TRANSPORT=True
in the environment variables.
export DISABLE_AIOHTTP_TRANSPORT="True"
On LiteLLM Python SDK
Set the disable_aiohttp_transport=True
to disable aiohttp transport.
import litellm
litellm.disable_aiohttp_transport = True # default is False, enable this to disable aiohttp transport
result = litellm.completion(
model="openai/gpt-4o",
messages=[{"role": "user", "content": "Hello, world!"}],
)
print(result)
New Models / Updated Models​
- Bedrock
- Video support for Bedrock Converse - PR
- InvokeAgents support as /chat/completions route - PR, Get Started
- AI21 Jamba models compatibility fixes - PR
- Fixed duplicate maxTokens parameter for Claude with thinking - PR
- Gemini (Google AI Studio + Vertex AI)
- VertexAI
- Anthropic
- xAI
- Web Search Support PR
- Google AI Studio
- Web Search Support PR
- Mistral
- Updated mistral-medium prices and context sizes - PR
- Ollama
- Tool calls parsing on streaming - PR
- Cohere
- Swapped Cohere and Cohere Chat provider positioning - PR
- Nebius AI Studio
- New provider integration - PR
LLM API Endpoints​
- Image Edits API
- Completions API
- Codestral latency overhead tracking on /v1/completions - PR
- Audio Transcriptions API
- Responses API
- Session management fixes for using Non-OpenAI models - PR
Management Endpoints / UI​
- Vector Stores
- Teams
- SSO/SCIM
- Bug fixes for showing SCIM token on UI - PR
- General UI
Logging / Guardrails Integrations​
Logging​
- Prometheus
- End users will no longer be tracked by default on Prometheus. Tracking end_users on prometheus is now opt-in. PR
- Langfuse
- Performance improvements: Fixed "Max langfuse clients reached" issue - PR
- Helicone
- Base URL support - PR
- Sentry
- Added sentry sample rate configuration - PR
Guardrails​
- Bedrock Guardrails
- Pangea Guardrails
- Added Pangea provider to Guardrails hook - PR
Performance / Reliability Improvements​
- aiohttp Transport
- Request Limiting
- Sliding window logic for parallel request limiter v2 - PR
Bug Fixes​
- LLM API Fixes
- General
New Contributors​
- @agajdosi made their first contribution in #9737
- @ketangangal made their first contribution in #11161
- @Aktsvigun made their first contribution in #11143
- @ryanmeans made their first contribution in #10775
- @nikoizs made their first contribution in #10054
- @Nitro963 made their first contribution in #11202
- @Jacobh2 made their first contribution in #11207
- @regismesquita made their first contribution in #10729
- @Vinnie-Singleton-NN made their first contribution in #10283
- @trashhalo made their first contribution in #11219
- @VigneshwarRajasekaran made their first contribution in #11223
- @AnilAren made their first contribution in #11233
- @fadil4u made their first contribution in #11242
- @whitfin made their first contribution in #11279
- @hcoona made their first contribution in #11272
- @keyute made their first contribution in #11173
- @emmanuel-ferdman made their first contribution in #11230
Demo Instance​
Here's a Demo Instance to test changes:
- Instance: https://demo.litellm.ai/
- Login Credentials:
- Username: admin
- Password: sk-1234