github BerriAI/litellm v1.56.10

latest releases: v1.57.0, v1.56.3-stable
2 days ago

What's Changed

  • fix(aws_secret_manager_V2.py): Error reading secret from AWS Secrets … by @krrishdholakia in #7541
  • Support checking provider-specific /models endpoints for available models based on key by @krrishdholakia in #7538
  • feat(router.py): support request prioritization for text completion c… by @krrishdholakia in #7540
  • (Fix) - Docker build error with pyproject.toml by @ishaan-jaff in #7550
  • (Fix) - Slack Alerting , don't send duplicate spend report when used on multi instance settings by @ishaan-jaff in #7546
  • add cohere/command-r7b-12-2024 by @ishaan-jaff in #7553

Full Changelog: v1.56.9...v1.56.10

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.56.10

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 230.0 268.3301603401397 6.21711064668469 0.0 1861 0 212.36320399998476 3556.7401620000396
Aggregated Passed ✅ 230.0 268.3301603401397 6.21711064668469 0.0 1861 0 212.36320399998476 3556.7401620000396

Don't miss a new litellm release

NewReleases is sending notifications on new releases.