What's Changed
- Litellm managed file updates combined by @krrishdholakia in #11040
- [Fix] Reliability Fix - Removing code that was creating threads on errors by @ishaan-jaff in #11066
- [Feat] Add Azure AD certificate-based authentication by @ishaan-jaff in #11069
- Update feature_request.yml by @ishaan-jaff in #11078
- adds tzdata (#10796) by @krrishdholakia in #11052
- Fix proxy_cli.py: avoid overriding DATABASE_URL when it’s already provided by @martin-liu in #11076
- feat(helm): Add loadBalancerClass support for LoadBalancer services by @gunjan-solanki in #11064
- Add Azure Mistral Medium 25.05 by @emerzon in #11063
- fix:Databricks Claude 3.7 Sonnet output token cost: $17.85/M by @bepotp in #11007
- Fix/openrouter stream usage id 8913 by @daarko10 in #11004
- feat: add embeddings to CustomLLM by @tokoko in #10980
- Enable switching between custom auth and litellm api key auth + Fix
/customer/update
for max budgets by @krrishdholakia in #11070 - Validate if user can call
retrieve
+list
+delete
file by @krrishdholakia in #11081 - [feature] ConfidentAI logging enabled for proxy and sdk by @spike-spiegel-21 in #10649
- Proper github images by @kreatoo in #10927
- Add devstral-small-2505 model to pricing and context window configuration by @xingyaoww in #11103
- [Ci/CD] use n 4 for mapped tests by @ishaan-jaff in #11109
- Fix/background health check by @baejooc in #10887
- [Feat] Add content policy violation error mapping for image edits by @ishaan-jaff in #11113
- Expose
/list
and/info
endpoints for Audit Log events by @krrishdholakia in #11102 - [Feat] Use aiohttp transport by default - 97% lower median latency by @ishaan-jaff in #11097
New Contributors
- @martin-liu made their first contribution in #11076
- @gunjan-solanki made their first contribution in #11064
- @tokoko made their first contribution in #10980
- @spike-spiegel-21 made their first contribution in #10649
- @kreatoo made their first contribution in #10927
- @baejooc made their first contribution in #10887
Full Changelog: v1.70.4-nightly...v1.71.0-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 310.0 | 321.9030490306742 | 5.995437687618243 | 4.166377779571852 | 1793 | 1246 | 259.8663090000173 | 771.521746000019 |
Aggregated | Failed ❌ | 310.0 | 321.9030490306742 | 5.995437687618243 | 4.166377779571852 | 1793 | 1246 | 259.8663090000173 | 771.521746000019 |