What's Changed
- Litellm stable release fixes by @krrishdholakia in #13682
- [UI QA] Aug 16th Fixes by @ishaan-jaff in #13684
- Litellm dev 08 16 2025 p3 by @krrishdholakia in #13694
- GPT-5-chat does not support function by @superpoussin22 in #13612
- fix(vertexai-batch): fix vertexai batch file format by @thiagosalvatore in #13576
- [Feat] Datadog LLM Observability - Add support for Failure Logging by @ishaan-jaff in #13726
- [Feat] DD LLM Observability - Add time to first token, litellm overhead, guardrail overhead latency metrics by @ishaan-jaff in #13734
- [Bug Fix] litellm incompatible with newest release of openAI v1.100.0 by @ishaan-jaff in #13728
- [Bug Fix] image_edit() function returns APIConnectionError with
litellm_proxy
- Support for both image edits and image generations by @ishaan-jaff in #13735 - [Fix] Cooldowns - don't return raw Azure Exceptions to client by @krrishdholakia in #13529
- Responses API - add default api version for openai responses api calls + Openrouter - fix claude-sonnet-4 on openrouter + Azure - Handle
openai/v1/responses
by @krrishdholakia in #13526 - Use namespace as prefix for s3 cache by @michal-otmianowski in #13704
- Add Search Functionality for Public Model Names in Model Dashboard by @NANDINI-star in #13687
- Add Azure Deployment Name Support in UI by @NANDINI-star in #13685
- Fix - gemini prompt caching cost calculation by @krrishdholakia in #13742
- Refactor - forward model group headers - reuse same logic as global header forwarding by @krrishdholakia in #13741
- Fix Groq streaming ASCII encoding issue by @colesmcintosh in #13675
- Add possibility to configure resources for migrations-job in Helm chart by @moandersson in #13617
- [Feat] Datadog LLM Observability - Add support for tracing guardrail input/output by @ishaan-jaff in #13767
- Models page row UI restructure by @NANDINI-star in #13771
New Contributors
- @michal-otmianowski made their first contribution in #13704
- @moandersson made their first contribution in #13617
Full Changelog: v1.75.8-nightly...v1.75.9.dev3
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.9.dev3
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 130.0 | 167.2434307868771 | 6.365346480657209 | 0.0 | 1905 | 0 | 103.2638190000057 | 1092.3964670000146 |
Aggregated | Passed ✅ | 130.0 | 167.2434307868771 | 6.365346480657209 | 0.0 | 1905 | 0 | 103.2638190000057 | 1092.3964670000146 |