github BerriAI/litellm v1.34.41

latest releases: v1.51.2, v1.51.1-staging, v1.51.1...
6 months ago

What's Changed

  • Fix - Embedding Caching - allow base 64 embedding cache hits by @ishaan-jaff in #2943

Full Changelog: 1.34.39.dev1...v1.34.41

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 55 60.27472656762199 1.6302017598732303 0.0 488 0 49.61538999998538 853.5629180000228
/health/liveliness Passed ✅ 40 42.21640406246 15.617199236490475 0.0 4675 0 38.64845300000752 1119.598647000032
/health/readiness Passed ✅ 41 43.405607340724444 15.216329951275746 0.0 4555 0 38.678258999993886 1239.126835000036
Aggregated Passed ✅ 41 43.680623275776924 32.463730947639455 0.0 9718 0 38.64845300000752 1239.126835000036

Don't miss a new litellm release

NewReleases is sending notifications on new releases.