What's Changed
- Fix redis cluster mode for routers by @ogunoz in #9010
- [Feat] - Display
thinking
tokens on OpenWebUI (Bedrock, Anthropic, Deepseek) by @ishaan-jaff in #9029 - (AWS Secret Manager) - Using K/V pairs in 1 AWS Secret by @ishaan-jaff in #9039
- (Docs) connect litellm to open web ui by @ishaan-jaff in #9040
- Added PDL project by @vazirim in #8925
- (UI) - Allow adding EU OpenAI models by @ishaan-jaff in #9042
New Contributors
Full Changelog: v1.63.2-nightly...v1.63.3-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.3-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 220.0 | 274.33505649537244 | 6.144475001880859 | 0.0 | 1837 | 0 | 199.62131199997657 | 3623.5841269999582 |
Aggregated | Passed ✅ | 220.0 | 274.33505649537244 | 6.144475001880859 | 0.0 | 1837 | 0 | 199.62131199997657 | 3623.5841269999582 |