github BerriAI/litellm v1.74.8-nightly

latest releases: v1.76.2-nightly, v1.76.1.rc.2, v1.76.1.rc.1...
one month ago

What's Changed

  • Fix SSO Logout | Create Unified Login Page with SSO and Username/Password Options by @NANDINI-star in #12703
  • docs(moonshot): correct base url and document CN-specific endpoint by @magicalne in #12804
  • Adding HolmesGPT to projects using LiteLLM by @pavangudiwada in #12798
  • feat: Add Lambda AI provider support by @colesmcintosh in #12817
  • [UI Bug Fix] Show correct guardrails when editing a team by @ishaan-jaff in #12823
  • [Docs] Show correct list of vertex ai mistral models by @ishaan-jaff in #12828
  • [LLM Translation - GH Copilot] added dynamic endpoint support by @jugaldb in #12827
  • feat: add Morph provider support by @colesmcintosh in #12821
  • fix: remove deprecated groq/qwen-qwq-32b and add qwen/qwen3-32b by @aholmberg in #12831
  • Add Google Cloud Model Armor guardrail documentation by @colesmcintosh in #12814
  • [Feat] Add Recraft Image Generation API Support - New LLM Provider by @ishaan-jaff in #12832
  • [Bug Fix] - gemini leaking FD for sync calls with litellm.completion by @ishaan-jaff in #12824
  • [Feat] Add fireworks - fireworks/models/kimi-k2-instruct by @ishaan-jaff in #12837
  • [LLM Translation] add qwen-vl-plus by @jugaldb in #12829
  • [Azure OpenAI Feature] - Support DefaultAzureCredential without hard-coded environment variables by @ishaan-jaff in #12841
  • Docs - litellm benchmarks by @ishaan-jaff in #12842
  • fix(watsonx): IBM Watsonx - use correct parameter name for tool choice by @Tomas2D in #9980
  • feat: add Hyperbolic provider support by @colesmcintosh in #12826
  • Litellm batch cost tracking debug by @krrishdholakia in #12782
  • Passthrough Auth - make Auth checks OSS + Anthropic - only show 'reasoning_effort' for supported models by @krrishdholakia in #12847
  • build(deps): bump form-data from 4.0.0 to 4.0.4 in /ui/litellm-dashboard by @dependabot[bot] in #12851
  • Fix team_member_budget update logic by @krrishdholakia in #12843
  • Fix async retryer on .acompletion() + forward clientside headers - filter out content-type from clientside request (causes llm api call to hang) by @krrishdholakia in #12848
  • Openrouter - filter out cache_control flag for non-anthropic models (allows usage with claude code) by @krrishdholakia in #12850
  • Improvements on the Regenerate Key Flow by @NANDINI-star in #12788
  • Replace non-root Dockerfile base with Alpine multi-stage build; by @mdiloreto in #12707
  • build(deps): bump form-data from 4.0.3 to 4.0.4 in /docs/my-website by @dependabot[bot] in #12867
  • Bug fix - Azure KeyVault not in image, add azure-keyvault==4.2.0 to Docker img by @ishaan-jaff in #12873
  • [Separate Health App] Pass through cmd args via supervisord by @jugaldb in #12871
  • [Feat] Add Recraft API - Image Edits Support by @ishaan-jaff in #12874
  • [LLM Translation] Litellm gemini 2.0 live support by @jugaldb in #12839
  • [Feat] Add cost tracking for new vertex_ai/llama-3 API models by @ishaan-jaff in #12878
  • [Feat] - Track cost + add tags for health checks done by LiteLLM Proxy by @ishaan-jaff in #12880
  • Fix async retryer on .acompletion() + forward clientside headers - filter out content-type from clientside request (causes llm api call to hang) by @krrishdholakia in #12886
  • Request Headers - support x-litellm-num-retries + Usage - support usage by model group by @krrishdholakia in #12890
  • [Docs ]Litellm mcp access group doc by @jugaldb in #12883
  • [Add health check] add architecture diagram by @jugaldb in #12879
  • rm retired anthropic models from model_prices_and_context_window.json by @staeiou in #12864
  • feat: extended /v1/models endpoint, now it returns with fallbacks on demand by @murad-khafizov in #12811
  • feat: Add Pillar Security guardrail integration by @eagle-p in #12791
  • Swagger - bump DOMPurify version + add back Support local swagger bundle (good for airgapped environments) + support local swagger on custom root paths by @krrishdholakia in #12911
  • [Feat] LiteLLM x Cloudzero integration - Allow exporting spend to cloudzero by @ishaan-jaff in #12908

New Contributors

Full Changelog: v1.74.7.rc.1...v1.74.8-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.74.8-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 170.0 208.0958998087373 6.271169991247407 0.0 1877 0 130.56155500004252 1168.738483000027
Aggregated Passed ✅ 170.0 208.0958998087373 6.271169991247407 0.0 1877 0 130.56155500004252 1168.738483000027

Don't miss a new litellm release

NewReleases is sending notifications on new releases.