π¨Please wait for a stable release before upgrading your production version of LiteLLM π We refactored utils.py, proxy_server.py to be less than 10K lines.
What's Changed
- [Fix] Security Fix bump docusaurus version by @ishaan-jaff in #4213
- build(deps): bump azure-identity from 1.15.0 to 1.16.1 by @dependabot in #4130
- build(deps): bump braces from 3.0.2 to 3.0.3 in /ui/litellm-dashboard by @dependabot in #4131
- fix(build): .dockerignore not picked up by @bcvanmeurs in #3116
- [Refactor-Proxy] Refactor user_api_key_auth to be it's own file by @ishaan-jaff in #4215
- [Reliability Fix] Anthropic / Bedrock HTTPX - Cache Async Httpx client by @ishaan-jaff in #4211
- refactor(utils.py): Cut down utils.py to <10k lines. by @krrishdholakia in #4216
- ui - show exceptions by @ishaan-jaff in #4222
- fix - non sso ui sign up flow by @ishaan-jaff in #4220
- [Refactor-Proxy] Make proxy_server.py < 10K lines (move management, key, endpoints to their own files) by @ishaan-jaff in #4217
Full Changelog: v1.40.14...v1.40.15
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.15
Don't want to maintain your internal proxy? get in touch π
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed β | 120.0 | 132.58387302297749 | 6.398687111538595 | 0.0 | 1915 | 0 | 97.12711200000967 | 1186.0091809999744 |
Aggregated | Passed β | 120.0 | 132.58387302297749 | 6.398687111538595 | 0.0 | 1915 | 0 | 97.12711200000967 | 1186.0091809999744 |