What's Changed
- [Fix] UI - Key Max Budget Removal Error Fix by @yuneng-jiang in #15672
- [Fix] GPT-OSS in Bedrock now supports streaming. Revert fake streaming by @ishaan-jaff in #15668
- [Fix (pricing)] - Fix pricing for watsonx model family for various models by @ishaan-jaff in #15670
- Fix: Gemini 2.5 Flash Image should not have supports_web_search=true by @jlan-nl in #15642
- Fix: Support us-gov prefix for AWS GovCloud Bedrock models by @ishaan-jaff in #15626
- [Bug Fix] Tags as metadata dicts were raising exceptions by @ishaan-jaff in #15625
- [Feat] Add Cost Tracking for /ocr endpoints by @ishaan-jaff in #15678
- fix: add missing context to benchmark docs by @AlexsanderHamir in #15688
- Fix: Separate OAuth M2M authentication from UI SSO + Handle Introspection endpoint for Oauth2 by @ishaan-jaff in #15667
- [Oct Staging Branch] by @ishaan-jaff in #15460
- [Feat] Add Guardrails for /v1/messages and /v1/responses API by @ishaan-jaff in #15686
- Add glm-4.6 model to pricing configuration by @Copilot in #15679
Full Changelog: v1.78.3-nightly...v1.78.4-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.78.4-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat