What's Changed
- Fix custom pricing - separate provider info from model info by @krrishdholakia in #7990
- Litellm dev 01 25 2025 p4 by @krrishdholakia in #8006
- (UI) - Adding new models enhancement - show provider logo by @ishaan-jaff in #8033
- (UI enhancement) - allow onboarding wildcard models on UI by @ishaan-jaff in #8034
- add openrouter/deepseek/deepseek-r1 by @paul-gauthier in #8038
- (UI) - allow assigning wildcard models to a team / key by @ishaan-jaff in #8041
- Add smolagents by @aymeric-roucher in #8026
- (UI) fixes to add model flow by @ishaan-jaff in #8043
- github - run stale issue/pr bot by @ishaan-jaff in #8045
- (doc) Add nvidia as provider by @raspawar in #8023
- feat(handle_jwt.py): initial commit adding custom RBAC support on jwt… by @krrishdholakia in #8037
- fix(utils.py): handle failed hf tokenizer request during calls by @krrishdholakia in #8032
New Contributors
- @aymeric-roucher made their first contribution in #8026
- @raspawar made their first contribution in #8023
Full Changelog: v1.59.8...v1.59.8-dev1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.59.8-dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 230.0 | 253.74562668371757 | 6.073890684010945 | 0.0 | 1818 | 0 | 198.74819999995452 | 1957.5085989999934 |
Aggregated | Passed ✅ | 230.0 | 253.74562668371757 | 6.073890684010945 | 0.0 | 1818 | 0 | 198.74819999995452 | 1957.5085989999934 |