What's Changed
- ui - fix team based usage crashing by @ishaan-jaff in #4198
- [Fix + Refactor] - Router Alerting for llm exceptions + use separate util for sending alert by @ishaan-jaff in #4197
- [Bug fix] Don't cache team, user, customer budget after calling /update, /delete by @ishaan-jaff in #4203
- feat(router.py): support content policy fallbacks by @krrishdholakia in #4207
- fix(slack_alerting.py): allow new 'alerting_metadata' arg by @krrishdholakia in #4205
- build(pyproject.toml): require pydantic v2 by @krrishdholakia in #4151
Full Changelog: v1.40.13...v1.40.13.dev1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.13.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 130.0 | 150.82119762323373 | 6.385731572640492 | 0.0 | 1911 | 0 | 110.03221099997518 | 539.267988000006 |
Aggregated | Passed ✅ | 130.0 | 150.82119762323373 | 6.385731572640492 | 0.0 | 1911 | 0 | 110.03221099997518 | 539.267988000006 |