What's Changed
- CRUD Endpoints for Pass-Through endpoints by @krrishdholakia in #5215
- fix(s3.py): fix s3 logging payload to have valid json values by @krrishdholakia in #5235
- Allow zero temperature for Sagemaker models based on config by @gitravin in #5173
- build(deps): bump aiohttp from 3.9.4 to 3.10.2 by @dependabot in #5135
- [Docs] Sagemaker add example on using with LiteLLM Proxy and temperature=0 by @ishaan-jaff in #5250
- [Feat-Proxy] Add Oauth 2.0 Support by @ishaan-jaff in #5252
- [Feat] Add bedrock Guardrail
traces
in response when trace=enabled by @ishaan-jaff in #5243
New Contributors
Full Changelog: v1.43.15...v1.43.16
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.43.16
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 120.0 | 142.9490557028836 | 6.490376439327205 | 0.0 | 1942 | 0 | 99.74976099999822 | 2696.515713999986 |
Aggregated | Passed ✅ | 120.0 | 142.9490557028836 | 6.490376439327205 | 0.0 | 1942 | 0 | 99.74976099999822 | 2696.515713999986 |