github BerriAI/litellm v1.56.6.dev1

latest releases: v1.56.3-stable, v1.56.10, v1.56.9...
2 days ago

What's Changed

  • Prometheus - custom metrics support + other improvements by @krrishdholakia in #7489
  • (feat) POST /fine_tuning/jobs support passing vertex specific hyper params by @ishaan-jaff in #7490
  • (Feat) - LiteLLM Use UsernamePasswordCredential for Azure OpenAI by @ishaan-jaff in #7496
  • (docs) Add docs on load testing benchmarks by @ishaan-jaff in #7499
  • (Feat) Add support for reading secrets from Hashicorp vault by @ishaan-jaff in #7497
  • Litellm dev 12 30 2024 p2 by @krrishdholakia in #7495
  • Refactor Custom Metrics on Prometheus - allow setting k,v pairs on all metrics via config.yaml by @krrishdholakia in #7498
  • (fix) GCS bucket logger - apply truncate_standard_logging_payload_content to standard_logging_payload and ensure GCS flushes queue on fails by @ishaan-jaff in #7500
  • Litellm dev 01 01 2025 p3 by @krrishdholakia in #7503

Full Changelog: v1.56.6...v1.56.6.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.56.6.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 230.0 255.89973974836954 6.151774848433542 0.003343355895887794 1840 1 94.9865199999067 1259.9916519999965
Aggregated Passed ✅ 230.0 255.89973974836954 6.151774848433542 0.003343355895887794 1840 1 94.9865199999067 1259.9916519999965

Don't miss a new litellm release

NewReleases is sending notifications on new releases.