What's Changed
- feat: Make gemini accept the openai parameter parallel_tool_calls by @aholmberg in #11125
- Fix #9295 docker-compose healthcheck test uses curl but curl is not in the image by @agajdosi in #9737
- [Feat] Add /image/edits support for Azure by @ishaan-jaff in #11160
- Fix deprecation_date value for llama groq models by @kiriloman in #11151
- [Fix] Rollback to httpx==0.27.0 by @ishaan-jaff in #11146
- Doc update for azure openai by @ketangangal in #11161
- Litellm fix GitHub action testing by @krrishdholakia in #11163
- [Feat - Contributor PR] Add Video support for Bedrock Converse by @ishaan-jaff in #11166
- [Fixes] Aiohttp transport fixes - add handling for
aiohttp.ClientPayloadError
and ssl_verification settings by @ishaan-jaff in #11162 - prevent leaking sensitive keys to langfuse + support forwarding
/sso/key/generate
to the server root path url by @krrishdholakia in #11165 - [Fix] - Duplicate maxTokens parameter being sent to Bedrock/Claude model with thinking by @ishaan-jaff in #11181
- Integration with Nebius AI Studio added by @Aktsvigun in #11143
- Codestral - return litellm latency overhead on
/v1/completions
+ Add 'contains' support for ChatCompletionDeltaToolCall by @krrishdholakia in #10879 - Ollama Chat - parse tool calls on streaming by @krrishdholakia in #11171
- [Fix] Prometheus Metrics - Do not track end_user by default + expose flag to enable tracking end_user on prometheus by @ishaan-jaff in #11192
- [Fix]: Add cost tracking for image edits endpoint [OpenAI, Azure] by @ishaan-jaff in #11186
- VertexAI -
codeExecution
tool support + anyOf handling by @krrishdholakia in #11195 - Add Pangea provider to Guardrails hook by @ryanmeans in #10775
- Return anthropic thinking blocks on streaming + VertexAI Minor Fixes & Improvements (Thinking, Global regions, Parallel tool calling) by @krrishdholakia in #11194
- Azure OIDC provider improvements + OIDC audience bug fix by @nikoizs in #10054
- [Feat] Add well known MCP servers to LiteLLM by @ishaan-jaff in #11209
- Add missing
request_kwargs
toget_available_deployment
call by @Nitro963 in #11202 - feat: Allow Adding MCP Servers Through LiteLLM UI by @ishaan-jaff in #11208
- Add price for gpt4o mini audio preview without date by @Jacobh2 in #11207
- Put "all proxy models" if no models are selected by @NANDINI-star in #11187
- Removed teamInfo call and used existing teamsList call by @NANDINI-star in #11051
- build(deps): bump next from 14.2.26 to 15.2.4 in /ui/litellm-dashboard by @dependabot in #11216
- Improve display of model tags on Keys, Teams and Org page by @NANDINI-star in #11022
- Update mistral-medium prices and context sizes by @regismesquita in #10729
- Add sentry sample rate by @Vinnie-Singleton-NN in #10283
- [Feat] Permission management vector stores on LiteLLM Key, Team, Orgs by @ishaan-jaff in #11213
- Fix: Remove duplicate object_permission field in LiteLLM_TeamTable by @trashhalo in #11219
- [UI] SCIM fixes SCIM token exposed on UI by @ishaan-jaff in #11220
- [Debug] Fix - debug langfuse clients started by @ishaan-jaff in #11221
- Helicone base url support + fix for embedding cache hits on str input by @krrishdholakia in #11211
- fix(utils.py): support non default params for audio transcription + fix calling azure o_series models by @krrishdholakia in #11212
- Add 'supports_parallel_function_calling' to all gemini models by @krrishdholakia in #11225
- fix the litellm CLA reference by @emmanuel-ferdman in #11230
- Update docker_quick_start.md – Replace gpt-3.5-turbo with gpt-4o in Proxy Getting Started Guide by @VigneshwarRajasekaran in #11223
- fix: bedrock ai21 jamba models will work now by @AnilAren in #11233
- Fix syntax error in Python example code in LiteLLM docs by @fadil4u in #11242
- [Feat]: Add Bedrock InvokeAgents as a /chat/completions route on LiteLLM by @ishaan-jaff in #11239
- [Feat]: Guardrails - Add streaming for bedrock post guard by @ishaan-jaff in #11247
- [Fix] Responses API - Session management by @ishaan-jaff in #11254
- Support dropping non-openai params when specified in
additional_drop_params
+ Add VertexAI Anthropic support on/v1/messages
by @krrishdholakia in #11246 - Anthropic - Files API w/ form-data support on passthrough + File ID support on
/chat/completion
by @krrishdholakia in #11256 - [Feat] UI - show vector store permissions for Key, Team, Org by @ishaan-jaff in #11277
- [Fix] Bedrock Guard Auth Param Persistence by @ishaan-jaff in #11270
- Streamline session expiration in UI by @whitfin in #11279
- [Chore] Only run init mcp servers if the mcp pip is available by @ishaan-jaff in #11282
- [Feat] Enforce Vector Store Access Controls on LiteLLM Auth by @ishaan-jaff in #11281
- [Bug]: Performance Fix Max langfuse clients reached: 20 is greater than 20 by @ishaan-jaff in #11285
- Wrong parameter mapping for "frequency_penalty" to "repeat_penalty" by @VigneshwarRajasekaran in #11284
- fix(secret-managers): Break AzureCredentialType restriction on AZURE_CREDENTIAL by @hcoona in #11272
- Swap Cohere and Cohere Chat provider by @keyute in #11173
- Increase timeout by @SmartManoj in #11288
- Make gemini stream thinking as reasoning_content by @aholmberg in #11290
- [Fix] QA Fixes - Vector Store Object Permissions by @ishaan-jaff in #11291
- feat(parallel_request_limiter_v2.py): add sliding window logic by @krrishdholakia in #11283
New Contributors
- @agajdosi made their first contribution in #9737
- @ketangangal made their first contribution in #11161
- @Aktsvigun made their first contribution in #11143
- @ryanmeans made their first contribution in #10775
- @nikoizs made their first contribution in #10054
- @Nitro963 made their first contribution in #11202
- @Vinnie-Singleton-NN made their first contribution in #10283
- @trashhalo made their first contribution in #11219
- @VigneshwarRajasekaran made their first contribution in #11223
- @AnilAren made their first contribution in #11233
- @fadil4u made their first contribution in #11242
- @whitfin made their first contribution in #11279
- @hcoona made their first contribution in #11272
Full Changelog: v1.71.1-stable...v1.71.3-rc