What's Changed
- ui - fix team based usage crashing by @ishaan-jaff in #4198
- [Fix + Refactor] - Router Alerting for llm exceptions + use separate util for sending alert by @ishaan-jaff in #4197
- [Bug fix] Don't cache team, user, customer budget after calling /update, /delete by @ishaan-jaff in #4203
- feat(router.py): support content policy fallbacks by @krrishdholakia in #4207
- fix(slack_alerting.py): allow new 'alerting_metadata' arg by @krrishdholakia in #4205
- build(pyproject.toml): require pydantic v2 by @krrishdholakia in #4151
- [Feat] send email alerts when budget exceeded by @ishaan-jaff in #4209
- [Fix] redact_message_input_output_from_logging deepcopy bug by @ishaan-jaff in #4210
Full Changelog: v1.40.13...v1.40.14
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.14
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 120.0 | 141.18410333195084 | 6.441903839147897 | 0.0 | 1928 | 0 | 105.22602600002529 | 510.8018800000025 |
Aggregated | Passed ✅ | 120.0 | 141.18410333195084 | 6.441903839147897 | 0.0 | 1928 | 0 | 105.22602600002529 | 510.8018800000025 |