v1.81.6 - Logs v2 with Tool Call Tracing
Deploy this version​
- Docker
- Pip
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
docker.litellm.ai/berriai/litellm:main-v1.81.6
pip install litellm==1.81.6
Key Highlights​
Logs View v2 with Tool Call Tracing - Redesigned logs interface with side panel, structured tool visualization, and error message search for faster debugging.
Let's dive in.
Logs View v2 with Tool Call Tracing​
This release introduces comprehensive tool call tracing through LiteLLM's redesigned Logs View v2, enabling developers to debug and monitor AI agent workflows in production environments seamlessly.
This means you can now onboard use cases like tracing complex multi-step agent interactions, debugging tool execution failures, and monitoring MCP server calls while maintaining full visibility into request/response payloads with syntax highlighting.
Developers can access the new Logs View through LiteLLM's UI to inspect tool calls in structured format, search logs by error messages or request patterns, and correlate agent activities across sessions with collapsible side panel views.
New Models / Updated Models​
New Model Support​
| Provider | Model | Context Window | Input ($/1M tokens) | Output ($/1M tokens) | Features |
|---|---|---|---|---|---|
| AWS Bedrock | amazon.nova-2-pro-preview-20251202-v1:0 | 1M | $2.19 | $17.50 | Chat completions, vision, video, PDF, function calling, prompt caching, reasoning |
| Google Vertex AI | gemini-robotics-er-1.5-preview | 1M | $0.30 | $2.50 | Chat completions, multimodal (text, image, video, audio), function calling, reasoning |
| OpenRouter | openrouter/xiaomi/mimo-v2-flash | 262K | $0.09 | $0.29 | Chat completions, function calling, reasoning |
| OpenRouter | openrouter/moonshotai/kimi-k2.5 | - | - | - | Chat completions |
| OpenRouter | openrouter/z-ai/glm-4.7 | 202K | $0.40 | $1.50 | Chat completions, vision, function calling, reasoning |
Features​
-
- Messages API Bedrock Converse caching and PDF support - PR #19785
- Translate advanced-tool-use to Bedrock-specific headers for Claude Opus 4.5 - PR #19841
- Support tool search header translation for Sonnet 4.5 - PR #19871
- Filter unsupported beta headers for AWS Bedrock Invoke API - PR #19877
- Nova grounding improvements - PR #19598, PR #20159
-
- Add Gemini Robotics-ER 1.5 preview support - PR #19845
- Support file retrieval in GoogleAIStudioFilesHandle - PR #20018
- Add /delete endpoint support - PR #20055
- Add custom_llm_provider as gemini translation - PR #19988
- Subtract implicit cached tokens from text_tokens for correct cost calculation - PR #19775
- Remove unsupported prompt-caching-scope-2026-01-05 header for vertex ai - PR #20058
- Add disable flag for anthropic gemini cache translation - PR #20052
- Convert image URLs to base64 in tool messages for Anthropic on Vertex AI - PR #19896
-
- Serialize imageUrl as object for OCI GenAI API - PR #19661
-
- Add context for volcengine models (deepseek-v3-2, glm-4-7, kimi-k2-thinking) - PR #19335
-
- Add prompt caching and reasoning support for MiniMax, GLM, Xiaomi - PR #19924
-
- Add embeddings support - PR #19660
Bug Fixes​
-
- Fix gemini-robotics-er-1.5-preview entry - PR #19974
-
General
-
- Ensure function content is valid JSON - PR #19232
LLM API Endpoints​
Features​
-
- Add LiteLLM x Claude Agent SDK Integration - PR #20035
-
A2A / MCP Gateway API (/a2a, /mcp)
- Add A2A agent header-based context propagation support - PR #19504
- Enable progress notifications for MCP tool calls - PR #19809
- Fix support for non-standard MCP URL patterns - PR #19738
- Add backward compatibility for legacy A2A card formats (/.well-known/agent.json) - PR #19949
- Add support for agent parameter in /interactions endpoint - PR #19866
-
- Add supported input formats documentation - PR #20073
-
RAG API (/rag/ingest, /vector_store)
- Add UI for /rag/ingest API - Upload docs, pdfs etc to create vector stores - PR #19822
- Add support for using S3 Vectors as Vector Store Provider - PR #19888
- Add s3_vectors as provider on /vector_store/search API + UI for creating + PDF support - PR #19895
- Add permission management for users and teams on Vector Stores - PR #19972
- Enable router support for completions in RAG query pipeline - PR #19550
-
Passthrough Endpoints (/{provider}_passthrough)
- Add /openai_passthrough route for OpenAI passthrough requests - PR #19989
- Add support for configuring role_mappings via environment variables - PR #19498
- Add Vertex AI LLM credentials sensitive keyword "vertex_credentials" for masking - PR #19551
- Fix prevention of provider-prefixed model name leaks in responses - PR #19943
- Fix proxy support for slashes in Google Vertex generateContent model names - PR #19737, PR #19753
- Support model names with slashes in Vertex AI passthrough URLs - PR #19944
- Fix regression in Vertex AI passthroughs for router models - PR #19967
- Add regression tests for Vertex AI passthrough model names - PR #19855
Bugs​
- General
- Fix token calculations and refactor - PR #19696
Management Endpoints / UI​
Features​
-
Proxy CLI Auth
-
Virtual Keys
-
Logs View
- v2 Logs view with side panel and improved UX - PR #20091
- New View to render "Tools" on Logs View - PR #20093
- Add Pretty print view of request/response - PR #20096
- Add error_message search in Spend Logs Endpoint - PR #19960
- UI: Adding Error message search to ui spend logs - PR #19963
- Spend Logs: Settings Modal - PR #19918
- Fix error_code in Spend Logs metadata - PR #20015
- Spend Logs: Show Current Store and Retention Status - PR #20017
- Allow Dynamic Setting of store_prompts_in_spend_logs - PR #19913
- Docs: UI Spend Logs Settings - PR #20197
-
Models + Endpoints
-
Usage & Analytics
-
UI Improvements
- UI: Allow Admins to control what pages are visible on LeftNav - PR #19907
- UI: Add Light/Dark Mode Switch for Development - PR #19804
- UI: Dark Mode: Delete Resource Modal - PR #20098
- UI: Tables: Reusable Table Sort Component - PR #19970
- UI: New Badge Dot Render - PR #20024
- UI: Feedback Prompts: Option To Hide Prompts - PR #19831
- UI: Navbar: Fixed Default Logo + Bound Logo Box - PR #20092
- UI: Navbar: User Dropdown - PR #20095
- Change default key type from 'Default' to 'LLM API' - PR #19516
-
Team & User Management
-
AI Gateway Features
Bugs​
- Fix Model Name During Fallback - PR #20177
- Fix Health Endpoints when Callback Objects Defined - PR #20182
- Fix Unable to reset user max budget to unlimited - PR #19796
- Fix Password comparison with non-ASCII characters - PR #19568
- Correct error message for DISABLE_ADMIN_ENDPOINTS - PR #19861
- Prevent clearing content filter patterns when editing guardrail - PR #19671
- Fix Prompt Studio history to load tools and system messages - PR #19920
- Add WATSONX_ZENAPIKEY to WatsonX credentials - PR #20086
- UI: Vector Store: Allow Config Defined Models to Be Selected - PR #20031
Logging / Guardrail / Prompt Management Integrations​
Features​
-
- Add callbacks and labels - PR #19708
- Add clientip and user agent in metrics - PR #19717
- Add tpm-rpm limit metrics - PR #19725
- Add model_id label to metrics - PR #19678
- Safely handle None metadata in logging - PR #19691
- Resolve high CPU when router_settings in DB by avoiding REGISTRY.collect() - PR #20087
-
- Add litellm_callback_logging_failures_metric for Langfuse, Langfuse Otel and other Otel providers - PR #19636
-
General Logging
Guardrails​
-
- Reuse HTTP connections to prevent performance degradation - PR #19964
-
Onyx
- Add timeout to onyx guardrail - PR #19731
-
General
Spend Tracking, Budgets and Rate Limiting​
- Add event-driven coordination for global spend query to prevent cache stampede - PR #20030
Performance / Loadbalancing / Reliability improvements​
- Resolve high CPU when router_settings in DB - by avoiding REGISTRY.collect() in PrometheusServicesLogger - PR #20087
- Reuse HTTP connections in Presidio - to prevent performance degradation - PR #19964
- Event-driven coordination for global spend query - prevent cache stampede - PR #20030
- Fix recursive Pydantic validation issue - PR #19531
- Refactor argument handling into helper function to reduce code bloat - PR #19720
- Optimize logo fetching and resolve MCP import blockers - PR #19719
- Improve logo download performance using async HTTP client - PR #20155
- Fix server root path configuration - PR #19790
- Refactor: Extract transport context creation into separate method - PR #19794
- Add native_background_mode configuration to override polling_via_cache for specific models - PR #19899
- Initialize tiktoken environment at import time to enable offline usage - PR #19882
- Improve tiktoken performance using local cache in lazy loading - PR #19774
- Fix timeout errors in chat completion calls to be correctly reported in failure callbacks - PR #19842
- Fix environment variable type handling for NUM_RETRIES - PR #19507
- Use safe_deep_copy in silent experiment kwargs to prevent mutation - PR #20170
- Improve error handling by inspecting BadRequestError after all other policy types - PR #19878
Database Changes​
Schema Updates​
| Table | Change Type | Description | PR | Migration |
|---|---|---|---|---|
LiteLLM_ManagedVectorStoresTable | New Columns | Added team_id and user_id fields for permission management | PR #19972 | Migration |
Migration Improvements​
- Fix Docker: Use correct schema path for Prisma generation - PR #19631
- Resolve 'relation does not exist' migration errors in setup_database - PR #19281
- Fix migration issue and improve Docker image stability - PR #19843
- Run Prisma generate as nobody user in non-root Docker container for security - PR #20000
- Bump litellm-proxy-extras version to 0.4.28 - PR #20166
Documentation Updates​
- Add Claude Agents SDK x LiteLLM Guide - PR #20036
- Add Cookbook: Using Claude Agent SDK + MCPs with LiteLLM - PR #20081
- Fix A2A Python SDK URL in documentation - PR #19832
- Add Sarvam usage documentation - PR #19844
- Add supported input formats for embeddings - PR #20073
- UI Spend Logs Settings Docs - PR #20197
- Add OpenAI Agents SDK to OSS Adopters list in README - PR #19820
- Update docs: Remove enterprise requirement for guardrail monitoring - PR #19833
- Add missing environment variable documentation - PR #20138
- Improve documentation blog index page - PR #20188
Infrastructure / Testing Improvements​
- Add test coverage for Router.get_valid_args and improve code coverage reporting - PR #19797
- Add validation of model cost map as CI job - PR #19993
- Add Realtime API benchmarks - PR #20074
- Add Init Containers support in community helm chart - PR #19816
- Add libsndfile to main Dockerfile for ARM64 audio processing support - PR #19776
New Contributors​
- @ruanjf made their first contribution in https://github.com/BerriAI/litellm/pull/19551
- @moh-dev-stack made their first contribution in https://github.com/BerriAI/litellm/pull/19507
- @formorter made their first contribution in https://github.com/BerriAI/litellm/pull/19498
- @priyam-that made their first contribution in https://github.com/BerriAI/litellm/pull/19516
- @marcosgriselli made their first contribution in https://github.com/BerriAI/litellm/pull/19550
- @natimofeev made their first contribution in https://github.com/BerriAI/litellm/pull/19232
- @zifeo made their first contribution in https://github.com/BerriAI/litellm/pull/19805
- @pragyasardana made their first contribution in https://github.com/BerriAI/litellm/pull/19816
- @ryewilson made their first contribution in https://github.com/BerriAI/litellm/pull/19833
- @lizhen921 made their first contribution in https://github.com/BerriAI/litellm/pull/19919
- @boarder7395 made their first contribution in https://github.com/BerriAI/litellm/pull/19666
- @rushilchugh01 made their first contribution in https://github.com/BerriAI/litellm/pull/19938
- @cfchase made their first contribution in https://github.com/BerriAI/litellm/pull/19893
- @ayim made their first contribution in https://github.com/BerriAI/litellm/pull/19872
- @varunsripad123 made their first contribution in https://github.com/BerriAI/litellm/pull/20018
- @nht1206 made their first contribution in https://github.com/BerriAI/litellm/pull/20046
- @genga6 made their first contribution in https://github.com/BerriAI/litellm/pull/20009
Full Changelog: https://github.com/BerriAI/litellm/compare/v1.81.3.rc...v1.81.6

