What's Changed
- [Feat] - Add
mistral/mistral-large 2
by @ishaan-jaff in #4873 - [Fix] OpenAI STT, TTS Health Checks on LiteLLM Proxy by @ishaan-jaff in #4889
- docs - add info about routing strategy on load balancing docs by @ishaan-jaff in #4890
- feat(custom_llm.py): Support Custom LLM Handlers by @krrishdholakia in #4887
- Add Single-Token predictions support for Replicate by @fracapuano in #4879
- Add mistral.mistral-large-2407-v1:0 on Amazon Bedrock by @Manouchehri in #4884
- Add Llama 3.1 405b & Tool Calling for Amazon Bedrock by @Manouchehri in #4883
- feat(auth_check.py): support using redis cache for team objects by @krrishdholakia in #4870
- fix logfire - don't load_dotenv by @ishaan-jaff in #4894
- Docs Proxy - add example usage with mistral SDK with Proxy by @ishaan-jaff in #4896
- Docs add example using anthropic sdk with litellm proxy by @ishaan-jaff in #4897
- [Feat] Support /* for multiple providers by @ishaan-jaff in #4891
New Contributors
- @fracapuano made their first contribution in #4879
Full Changelog: v1.42.1...v1.42.2-stable
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.42.2-stable
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 120.0 | 149.75578451616298 | 6.41284316524727 | 0.0 | 1918 | 0 | 106.08121300003859 | 1931.1223589999997 |
Aggregated | Passed ✅ | 120.0 | 149.75578451616298 | 6.41284316524727 | 0.0 | 1918 | 0 | 106.08121300003859 | 1931.1223589999997 |