What's Changed
- Github: Increase timeout of litellm tests by @zoltan-ongithub in #10568
- [Docs] Change llama-api link for litellm by @seyeong-han in #10556
- [Feat] v2 Custom Logger API Endpoints by @ishaan-jaff in #10575
- [Bug fix] JSON logs - Ensure only 1 log is emitted (previously duplicate json logs were getting emitted) by @ishaan-jaff in #10580
- Update gemini-2.5-pro-exp-03-25 max_tokens to 65,535 by @mkavinkumar1 in #10548
- Update instructor.md by @thomelane in #10549
- fix issue when databrick use external model, the delta could be empty… by @frankzye in #10540
- Add
litellm-proxy
CLI (#10478) by @ishaan-jaff in #10578
New Contributors
- @zoltan-ongithub made their first contribution in #10568
- @mkavinkumar1 made their first contribution in #10548
- @thomelane made their first contribution in #10549
- @frankzye made their first contribution in #10540
Full Changelog: v1.68.0-nightly...v1.68.1.dev1
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.68.1.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 210.0 | 244.34719839029643 | 6.203411663807808 | 0.0 | 1855 | 0 | 183.31073700005618 | 5362.244745999988 |
Aggregated | Passed ✅ | 210.0 | 244.34719839029643 | 6.203411663807808 | 0.0 | 1855 | 0 | 183.31073700005618 | 5362.244745999988 |