What's Changed
- Fix SSO Logout | Create Unified Login Page with SSO and Username/Password Options by @NANDINI-star in #12703
- docs(moonshot): correct base url and document CN-specific endpoint by @magicalne in #12804
- Adding HolmesGPT to projects using LiteLLM by @pavangudiwada in #12798
- feat: Add Lambda AI provider support by @colesmcintosh in #12817
- [UI Bug Fix] Show correct guardrails when editing a team by @ishaan-jaff in #12823
- [Docs] Show correct list of vertex ai mistral models by @ishaan-jaff in #12828
- [LLM Translation - GH Copilot] added dynamic endpoint support by @jugaldb in #12827
- feat: add Morph provider support by @colesmcintosh in #12821
- fix: remove deprecated groq/qwen-qwq-32b and add qwen/qwen3-32b by @aholmberg in #12831
- Add Google Cloud Model Armor guardrail documentation by @colesmcintosh in #12814
- [Feat] Add Recraft Image Generation API Support - New LLM Provider by @ishaan-jaff in #12832
- [Bug Fix] - gemini leaking FD for sync calls with litellm.completion by @ishaan-jaff in #12824
- [Feat] Add fireworks -
fireworks/models/kimi-k2-instruct
by @ishaan-jaff in #12837 - [LLM Translation] add qwen-vl-plus by @jugaldb in #12829
- [Azure OpenAI Feature] - Support DefaultAzureCredential without hard-coded environment variables by @ishaan-jaff in #12841
- Docs - litellm benchmarks by @ishaan-jaff in #12842
- fix(watsonx): IBM Watsonx - use correct parameter name for tool choice by @Tomas2D in #9980
- feat: add Hyperbolic provider support by @colesmcintosh in #12826
- Litellm batch cost tracking debug by @krrishdholakia in #12782
- Passthrough Auth - make Auth checks OSS + Anthropic - only show 'reasoning_effort' for supported models by @krrishdholakia in #12847
- build(deps): bump form-data from 4.0.0 to 4.0.4 in /ui/litellm-dashboard by @dependabot[bot] in #12851
- Fix team_member_budget update logic by @krrishdholakia in #12843
- Fix async retryer on
.acompletion()
+ forward clientside headers - filter outcontent-type
from clientside request (causes llm api call to hang) by @krrishdholakia in #12848 - Openrouter - filter out cache_control flag for non-anthropic models (allows usage with claude code) by @krrishdholakia in #12850
- Improvements on the Regenerate Key Flow by @NANDINI-star in #12788
- Replace non-root Dockerfile base with Alpine multi-stage build; by @mdiloreto in #12707
- build(deps): bump form-data from 4.0.3 to 4.0.4 in /docs/my-website by @dependabot[bot] in #12867
- Bug fix - Azure KeyVault not in image, add
azure-keyvault==4.2.0
to Docker img by @ishaan-jaff in #12873 - [Separate Health App] Pass through cmd args via supervisord by @jugaldb in #12871
- [Feat] Add Recraft API - Image Edits Support by @ishaan-jaff in #12874
- [LLM Translation] Litellm gemini 2.0 live support by @jugaldb in #12839
- [Feat] Add cost tracking for new vertex_ai/llama-3 API models by @ishaan-jaff in #12878
- [Feat] - Track cost + add tags for health checks done by LiteLLM Proxy by @ishaan-jaff in #12880
- Fix async retryer on .acompletion() + forward clientside headers - filter out content-type from clientside request (causes llm api call to hang) by @krrishdholakia in #12886
- Request Headers - support
x-litellm-num-retries
+ Usage - support usage by model group by @krrishdholakia in #12890 - [Docs ]Litellm mcp access group doc by @jugaldb in #12883
- [Add health check] add architecture diagram by @jugaldb in #12879
- rm retired anthropic models from model_prices_and_context_window.json by @staeiou in #12864
- feat: extended
/v1/models
endpoint, now it returns with fallbacks on demand by @murad-khafizov in #12811 - feat: Add Pillar Security guardrail integration by @eagle-p in #12791
- Swagger - bump DOMPurify version + add back Support local swagger bundle (good for airgapped environments) + support local swagger on custom root paths by @krrishdholakia in #12911
- [Feat] LiteLLM x Cloudzero integration - Allow exporting spend to cloudzero by @ishaan-jaff in #12908
New Contributors
- @magicalne made their first contribution in #12804
- @pavangudiwada made their first contribution in #12798
- @mdiloreto made their first contribution in #12707
- @murad-khafizov made their first contribution in #12811
- @eagle-p made their first contribution in #12791
Full Changelog: v1.74.7.rc.1...v1.74.8-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.8-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 170.0 | 208.0958998087373 | 6.271169991247407 | 0.0 | 1877 | 0 | 130.56155500004252 | 1168.738483000027 |
Aggregated | Passed ✅ | 170.0 | 208.0958998087373 | 6.271169991247407 | 0.0 | 1877 | 0 | 130.56155500004252 | 1168.738483000027 |