What's Changed
- Fix #7629 - Add tzdata package to Dockerfile (#8915) by @krrishdholakia in #9009
- Return
signature
on anthropic streaming + migrate tosignature
field instead ofsignature_delta
[MINOR bump] by @krrishdholakia in #9021 - Support
format
param for specifying image type by @krrishdholakia in #9019
v1.63.0 fixes Anthropic 'thinking' response on streaming to return the signature
block. Github Issue
It also moves the response structure from signature_delta
to signature
to be the same as Anthropic. Anthropic Docs
Diff
"message": {
...
"reasoning_content": "The capital of France is Paris.",
"thinking_blocks": [
{
"type": "thinking",
"thinking": "The capital of France is Paris.",
- "signature_delta": "EqoBCkgIARABGAIiQL2UoU0b1OHYi+..." # 👈 OLD FORMAT
+ "signature": "EqoBCkgIARABGAIiQL2UoU0b1OHYi+..." # 👈 KEY CHANGE
}
]
}
Full Changelog: v1.62.4-nightly...v1.63.0-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 250.0 | 272.1226933173393 | 6.127690671911355 | 0.0 | 1834 | 0 | 217.38513100001455 | 3752.371346000018 |
Aggregated | Passed ✅ | 250.0 | 272.1226933173393 | 6.127690671911355 | 0.0 | 1834 | 0 | 217.38513100001455 | 3752.371346000018 |