What's Changed
- VLLM - transcription endpoint support + Ollama_chat/ - images, thinking, and content as list handling + by @krrishdholakia in #14523
- [Fix] Bug Fix - Org Budget was not updating by @ishaan-jaff in #14541
- Docs update on user header mapping by @boopesh07 in #14527
- Litellm 1.77.2 stable notes by @ishaan-jaff in #14544
- Litellm UI qa 09 13 2025 p1 - fix end user filtering + fix load mcp tool call error + prevent setting max user budget on scroll in edit user settings by @krrishdholakia in #14545
- The 'last 24 hours' button shows up above the end user dropdown on Logs page by @NANDINI-star in #14546
Full Changelog: v1.77.1-nightly...v1.77.2.rc.1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.77.2.rc.1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 110.0 | 125.20945870304276 | 6.3684410533716305 | 6.3684410533716305 | 1906 | 1906 | 85.52794999999946 | 3304.632747999989 |
Aggregated | Failed ❌ | 110.0 | 125.20945870304276 | 6.3684410533716305 | 6.3684410533716305 | 1906 | 1906 | 85.52794999999946 | 3304.632747999989 |