What's Changed
- fix(utils.py): support calling openai models via
azure_ai/
by @krrishdholakia in #5209 - [Feat-Proxy] - user common helper to
route_request
for making llm call by @ishaan-jaff in #5224 - [Fix-Proxy + Langfuse] Always log cache_key on hits/misses by @ishaan-jaff in #5226
- [Fix] use BaseAWSLLM for bedrock, sagemaker by @ishaan-jaff in #5233
- [Feat] Make Sagemaker Async by @ishaan-jaff in #5237
- fix using anthropic prompt caching on proxy by @ishaan-jaff in #5238
Full Changelog: v1.43.13...v1.43.15
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.43.15
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 110.0 | 137.00972958622484 | 6.355893263955775 | 0.0 | 1902 | 0 | 98.4642859999667 | 1360.5994449999912 |
Aggregated | Passed ✅ | 110.0 | 137.00972958622484 | 6.355893263955775 | 0.0 | 1902 | 0 | 98.4642859999667 | 1360.5994449999912 |