What's Changed
- [Fix] Redacting messages from OTEL + Refactor
utils.py
to uselitellm_core_utils
by @ishaan-jaff in #4176 - Stop throwing constant S3 spam on cache misses by @Manouchehri in #4177
- [Feat] - Prometheus add remaining_team_budget gauge by @ishaan-jaff in #4179
- fix - clean up swagger spend endpoints ๐งน๐งน๐งน๐งน by @ishaan-jaff in #4180
- [Fix] Fix bug when updating team budgets on UI + display budget =0.0 correctly on UI by @ishaan-jaff in #4183
- Reset budget option for teams by @krrishdholakia in #4185
- feat(init.py): allow setting drop_params as an env by @krrishdholakia in #4187
- [Doc] Setting Team budgets by @ishaan-jaff in #4189
Full Changelog: v1.40.10...v1.40.11
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.11
Don't want to maintain your internal proxy? get in touch ๐
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed โ | 120.0 | 140.50671503682315 | 6.351765918831661 | 0.0 | 1901 | 0 | 96.28972799998792 | 1490.2560670000184 |
Aggregated | Passed โ | 120.0 | 140.50671503682315 | 6.351765918831661 | 0.0 | 1901 | 0 | 96.28972799998792 | 1490.2560670000184 |