What's Changed
- [Docs] Better search experience with Canary by @yujonglee in #4893
- Fixed tool_call for Helicone integration by @maamalama in #4869
- Fix Datadog logging attributes by @idris in #4909
- [Proxy-Fix + Test] - /batches endpoint by @ishaan-jaff in #4914
- [Proxy-Fix] - raise more descriptive errors when crossing tpm / rpm limits on keys, user, global limits by @ishaan-jaff in #4913
- [Feat] Link to https://models.litellm.ai/ on Swagger docs and docs by @ishaan-jaff in #4917
New Contributors
Full Changelog: v1.42.3...v1.42.4
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.42.4
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 99 | 137.24637847135392 | 6.41608880575569 | 0.0 | 1920 | 0 | 81.7409160000011 | 2139.0533030000256 |
Aggregated | Passed ✅ | 99 | 137.24637847135392 | 6.41608880575569 | 0.0 | 1920 | 0 | 81.7409160000011 | 2139.0533030000256 |