What's Changed
- fix(team_endpoints.py): ensure 404 raised when team not found + fix setting tags on keys by @krrishdholakia in #9038
- build(model_prices_and_context_window.json): update azure o1 mini pri… by @krrishdholakia in #9046
- Support master key rotations by @krrishdholakia in #9041
- (Feat) - add pricing for eu.amazon.nova models by @ishaan-jaff in #9056
- docs: Add project page for pgai by @Askir in #8576
- Mark several Claude models as being able to accept PDF inputs by @minhduc0711 in #9054
- (UI) - Keys Page - Show 100 Keys Per Page, Use full height, increase width of key alias by @ishaan-jaff in #9064
- (UI) Logs Page - Keep expanded log in focus on LiteLLM UI by @ishaan-jaff in #9061
- (Docs) OpenWeb x LiteLLM Docker compose + Instructions on spend tracking + logging by @ishaan-jaff in #9059
- (UI) - Allow adding Cerebras, Sambanova, Perplexity, Fireworks, Openrouter, TogetherAI Models on Admin UI by @ishaan-jaff in #9069
- UI - new API Playground for testing LiteLLM translation by @krrishdholakia in #9073
- Bug fix - String data: stripped from entire content in streamed Gemini responses by @ishaan-jaff in #9070
- (UI) - Minor improvements to logs page by @ishaan-jaff in #9076
- Bug fix: support bytes.IO when handling audio files for transcription by @tvishwanadha in #9071
- Fix batches api cost tracking + Log batch models in spend logs / standard logging payload by @krrishdholakia in #9077
- (UI) - Fix, Allow Filter Keys by Team Alias, Key Alias and Org by @ishaan-jaff in #9083
- (Clean up) - Allow switching off storing Error Logs in DB by @ishaan-jaff in #9084
- (UI) - Fix show correct count of internal user keys on Users Page by @ishaan-jaff in #9082
- New stable release notes by @krrishdholakia in #9085
- Litellm dev 03 08 2025 p3 by @krrishdholakia in #9089
- feat: prioritize api_key over tenant_id for more Azure AD token provi… by @krrishdholakia in #8701
- Fix incorrect streaming response by @5aaee9 in #9081
- Support openrouter
reasoning_content
on streaming by @krrishdholakia in #9094 - pricing for jamba new models by @themrzmaster in #9032
New Contributors
- @Askir made their first contribution in #8576
- @tvishwanadha made their first contribution in #9071
- @5aaee9 made their first contribution in #9081
Full Changelog: v1.63.3-nightly...v1.63.5-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.5-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 250.0 | 265.2487556257438 | 6.181834559182228 | 0.0 | 1849 | 0 | 214.44034500001408 | 3942.616398000041 |
Aggregated | Passed ✅ | 250.0 | 265.2487556257438 | 6.181834559182228 | 0.0 | 1849 | 0 | 214.44034500001408 | 3942.616398000041 |