What's Changed
- feat: Allow Adding MCP Servers Through LiteLLM UI by @ishaan-jaff in #11208
- Add price for gpt4o mini audio preview without date by @Jacobh2 in #11207
- Put "all proxy models" if no models are selected by @NANDINI-star in #11187
- Removed teamInfo call and used existing teamsList call by @NANDINI-star in #11051
- build(deps): bump next from 14.2.26 to 15.2.4 in /ui/litellm-dashboard by @dependabot in #11216
- Improve display of model tags on Keys, Teams and Org page by @NANDINI-star in #11022
- Update mistral-medium prices and context sizes by @regismesquita in #10729
- Add sentry sample rate by @Vinnie-Singleton-NN in #10283
- [Feat] Permission management vector stores on LiteLLM Key, Team, Orgs by @ishaan-jaff in #11213
- Fix: Remove duplicate object_permission field in LiteLLM_TeamTable by @trashhalo in #11219
- [UI] SCIM fixes SCIM token exposed on UI by @ishaan-jaff in #11220
New Contributors
- @Vinnie-Singleton-NN made their first contribution in #10283
- @trashhalo made their first contribution in #11219
Full Changelog: v1.71.2-nightly...v1.71.2.dev1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.2.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 220.0 | 238.87666312181884 | 6.174193914087743 | 0.0033428229096306135 | 1847 | 1 | 50.488271999995504 | 1443.0136089999905 |
Aggregated | Passed ✅ | 220.0 | 238.87666312181884 | 6.174193914087743 | 0.0033428229096306135 | 1847 | 1 | 50.488271999995504 | 1443.0136089999905 |