What's Changed
- update GCSBucketBase to handle GSM project ID if passed by @wwells in #11409
- fix: add enterprise feature gating to RegenerateKeyModal in KeyInfoView by @likweitan in #11400
- Litellm audit log staging by @krrishdholakia in #11418
- Add User ID validation to ensure it is not an email or phone number by @raz-alon in #10102
- [Performance] Performance improvements for /v1/messages route by @ishaan-jaff in #11421
- Add SSO configuration endpoints and UI integration with persistent settings by @colesmcintosh in #11417
- [Build] Bump dd trace version by @ishaan-jaff in #11426
- Add together_ai provided deepseek-r1 family model configuration by @jtsai-quid in #11394
- fix: Use proper attribute for Sagemaker request for embeddings by @tmbo in #11362
- added gemini url context support by @wangsha in #11351
- fix(redis_cache.py): support pipeline redis lpop for older redis vers… by @krrishdholakia in #11425
- Support no reasoning option for gemini models by @lowjiansheng in #11393
- fix(prometheus.py): pass custom metadata labels in litellm_total_toke… by @krrishdholakia in #11414
- Fix None values in usage field for gpt-image-1 model responses by @colesmcintosh in #11448
- Fix HuggingFace embeddings using non-default
input_type
by @seankwalker in #11452 - Add AGENTS.md by @colesmcintosh in #11461
- Custom Root Path Improvements: don't require reserving
/litellm
route by @krrishdholakia in #11460 - [Feat] Make batch size for maximum retention in spend logs a controllable parameter by @ishaan-jaff in #11459
- Add pangea to guardrails sidebar by @ryanmeans in #11464
- [Fix] [Bug]: Knowledge Base Call returning error by @ishaan-jaff in #11467
- [Feat] Return response_id == upstream response ID for VertexAI + Google AI studio (Stream+Non stream) by @ishaan-jaff in #11456
- [Fix]: /v1/messages - return streaming usage statistics when using litellm with bedrock models by @ishaan-jaff in #11469
New Contributors
- @wwells made their first contribution in #11409
- @likweitan made their first contribution in #11400
- @raz-alon made their first contribution in #10102
- @jtsai-quid made their first contribution in #11394
- @tmbo made their first contribution in #11362
- @wangsha made their first contribution in #11351
- @seankwalker made their first contribution in #11452
Full Changelog: v1.72.1-nightly...v1.72.1.dev8
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.72.1.dev8
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 180.0 | 202.7306475996812 | 6.284886409238797 | 0.0 | 1881 | 0 | 164.28741799995805 | 1311.9080620000432 |
Aggregated | Passed ✅ | 180.0 | 202.7306475996812 | 6.284886409238797 | 0.0 | 1881 | 0 | 164.28741799995805 | 1311.9080620000432 |