github BerriAI/litellm v1.52.5-stable

10 days ago

Full Changelog: v1.52.5.staging1...v1.52.5-stable

Docker image ghcr.io/berriai/litellm:litellm_stable_nov12-stable

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_nov12-stable

What's Changed

fix(init.py): add 'watsonx_text' as mapped llm api route
fix(opentelemetry.py): fix passing parallel tool calls to otel
fix(init.py): update provider-model mapping to include all known provider-model mappings
feat(anthropic): support passing document in llm api call
docs(anthropic.md): add pdf anthropic call to docs + expose new 'supports_pdf_input' function

fix(deepseek/chat): convert content list to str
test(test_deepseek_completion.py): implement base llm unit tests
fix(router.py): support content policy violation fallbacks with default fallbacks
fix(opentelemetry.py): refactor to move otel imports behing flag
fix(opentelemtry.py): close span on success completion
fix(user_api_key_auth.py): allow user_role to default to none

New Contributors

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.52.5-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 288.0333965427629 6.0955375578428805 0.0 1824 0 215.17615800001977 3641.4951400000177
Aggregated Passed ✅ 250.0 288.0333965427629 6.0955375578428805 0.0 1824 0 215.17615800001977 3641.4951400000177

Don't miss a new litellm release

NewReleases is sending notifications on new releases.