What's Changed
- [UI] Fix show latency < 0.0001 for deployments that have low latency + only show non cache hits on latency UI by @ishaan-jaff in #3388
- [UI] show slow responses + num requests per deployment by @ishaan-jaff in #3390
- [Fix + Test] Errant prints on langfuse by @ishaan-jaff in #3391
- Add langfuse
sdk_integration
by @marcklingen in #2516 - Update langfuse in
requirements.txt
by @Manouchehri in #3262
Full Changelog: v1.35.34...v1.35.35
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 41 | 51.26542280364234 | 1.6500892298075267 | 0.0 | 494 | 0 | 35.884482000028584 | 1109.5307290000278 |
/health/liveliness | Passed ✅ | 25 | 28.873856892782293 | 15.919686779883952 | 0.0 | 4766 | 0 | 23.275566999984676 | 1375.641074000015 |
/health/readiness | Passed ✅ | 26 | 28.245912157308112 | 15.288377337710626 | 0.003340261598800661 | 4577 | 1 | 23.47498500000711 | 1363.4175379999647 |
Aggregated | Passed ✅ | 26 | 29.706156425739458 | 32.8581533474021 | 0.003340261598800661 | 9837 | 1 | 23.275566999984676 | 1375.641074000015 |