What's Changed
- [FEAT]- OTEL log litellm request / response by @ishaan-jaff in #4076
- [Feat] Enterprise - Attribute Management changes to Users in Audit Logs by @ishaan-jaff in #4083
- [FEAT]- OTEL Log raw LLM request/response on OTEL by @ishaan-jaff in #4078
- fix(cost_calculator.py): fixes tgai unmapped model pricing by @krrishdholakia in #4085
- fix(utils.py): improved predibase exception mapping by @krrishdholakia in #4080
- [Fix] Litellm sdk - allow ChatCompletionMessageToolCall, and Function to be used as dict by @ishaan-jaff in #4086
- Update together ai pricing by @krrishdholakia in #4087
- [Feature]: Proxy: Support API-Key header in addition to Authorization header by @ishaan-jaff in #4088
- docs - cache controls on
litellm python SDK
by @ishaan-jaff in #4099
Full Changelog: v1.40.7...v1.40.7.dev1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.7.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 140.0 | 179.79878830216478 | 6.323646865102133 | 0.0 | 1893 | 0 | 111.88137199997072 | 2245.1254659999904 |
Aggregated | Passed ✅ | 140.0 | 179.79878830216478 | 6.323646865102133 | 0.0 | 1893 | 0 | 111.88137199997072 | 2245.1254659999904 |