What's Changed
- UI Updates by @krrishdholakia in #8345
- OIDC Scope based model access by @krrishdholakia in #8343
- Fix azure max retries error by @krrishdholakia in #8340
- Update deepseek API prices for 2025-02-08 by @Winston-503 in #8363
- fix(nvidia_nim/embed.py): add 'dimensions' support by @krrishdholakia in #8302
- fix: dictionary changed size during iteration error (#8327) by @krrishdholakia in #8341
- fix: add azure/o1-2024-12-17 to model_prices_and_context_window.json by @byrongrogan in #8371
- (Security fix) Mask redis pwd on
/cache/ping
+ add timeout value and elapsed time on azure + http calls by @krrishdholakia in #8377 - Handle azure deepseek reasoning response (#8288) by @krrishdholakia in #8366
- Anthropic Citations API Support by @krrishdholakia in #8382
- (Feat) - Add
/bedrock/invoke
support for all Anthropic models by @ishaan-jaff in #8383 - O3 mini native streaming support by @krrishdholakia in #8387
Full Changelog: v1.60.6...v1.60.8
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.60.8
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 170.0 | 189.56173781509457 | 6.206468643400922 | 0.0 | 1855 | 0 | 149.30551800000558 | 3488.08786699999 |
Aggregated | Passed ✅ | 170.0 | 189.56173781509457 | 6.206468643400922 | 0.0 | 1855 | 0 | 149.30551800000558 | 3488.08786699999 |