What's Changed
- Litellm stable release fixes by @krrishdholakia in #13682
- [UI QA] Aug 16th Fixes by @ishaan-jaff in #13684
- Litellm dev 08 16 2025 p3 by @krrishdholakia in #13694
- GPT-5-chat does not support function by @superpoussin22 in #13612
- fix(vertexai-batch): fix vertexai batch file format by @thiagosalvatore in #13576
- [Feat] Datadog LLM Observability - Add support for Failure Logging by @ishaan-jaff in #13726
- [Feat] DD LLM Observability - Add time to first token, litellm overhead, guardrail overhead latency metrics by @ishaan-jaff in #13734
- [Bug Fix] litellm incompatible with newest release of openAI v1.100.0 by @ishaan-jaff in #13728
- [Bug Fix] image_edit() function returns APIConnectionError with
litellm_proxy
- Support for both image edits and image generations by @ishaan-jaff in #13735 - [Fix] Cooldowns - don't return raw Azure Exceptions to client by @krrishdholakia in #13529
- Responses API - add default api version for openai responses api calls + Openrouter - fix claude-sonnet-4 on openrouter + Azure - Handle
openai/v1/responses
by @krrishdholakia in #13526 - Use namespace as prefix for s3 cache by @michal-otmianowski in #13704
- Add Search Functionality for Public Model Names in Model Dashboard by @NANDINI-star in #13687
- Add Azure Deployment Name Support in UI by @NANDINI-star in #13685
- Fix - gemini prompt caching cost calculation by @krrishdholakia in #13742
- Refactor - forward model group headers - reuse same logic as global header forwarding by @krrishdholakia in #13741
- Fix Groq streaming ASCII encoding issue by @colesmcintosh in #13675
- Add possibility to configure resources for migrations-job in Helm chart by @moandersson in #13617
- [Feat] Datadog LLM Observability - Add support for tracing guardrail input/output by @ishaan-jaff in #13767
- Models page row UI restructure by @NANDINI-star in #13771
- [Bug Fix] Bedrock KB - Using LiteLLM Managed Credentials for Query by @ishaan-jaff in #13787
- [Bug Fix] Fixes for using Auto Router with LiteLLM Docker Image by @ishaan-jaff in #13788
- [Feat] - UI Allow using Key/Team Based Logging for Langfuse OTEL by @ishaan-jaff in #13791
- Add long context support for claude-4-sonnet by @kankute-sameer in #13759
- Migrate to aim new firewall api by @hxdror in #13748
- [LLM Translation] Adjust max_input_tokens for azure/gpt-5-chat models in JSON configuration by @jugaldb in #13660
- Added Qwen3, Deepseek R1 0528 Throughput, GLM 4.5 and GPT-OSS models for Together AI by @Tasmay-Tibrewal in #13637
- Fix query passthrough deletion by @NANDINI-star in #13622
New Contributors
- @michal-otmianowski made their first contribution in #13704
- @moandersson made their first contribution in #13617
- @Tasmay-Tibrewal made their first contribution in #13637
Full Changelog: v1.75.8-nightly...v1.75.9-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.9-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 98 | 143.26229013447838 | 6.435162941463474 | 0.0 | 1926 | 0 | 70.0977230000035 | 1988.133740999956 |
Aggregated | Passed ✅ | 98 | 143.26229013447838 | 6.435162941463474 | 0.0 | 1926 | 0 | 70.0977230000035 | 1988.133740999956 |