Full Changelog: v1.61.20-stable...v1.63.2-stable
-
New Models / Updated Models
- Add supports_pdf_input: true for specific Bedrock Claude models
-
LLM Translation
- Support
/openai/
passthrough for Assistant endpoints - Bedrock Claude - fix amazon anthropic claude 3 tool calling transformation on invoke route
- Bedrock Claude - response_format support for claude on invoke route
- Bedrock - pass
description
if set in response_format - Bedrock - Fix passing response_format: {"type": "text"}
- OpenAI - Handle sending image_url as str to openai
- Deepseek - Fix deepseek 'reasoning_content' error
- Caching - Support caching on reasoning content
- Bedrock - handle thinking blocks in assistant message
- Anthropic - Return signature on anthropic streaming + migrate to signature field instead of signature_delta
- Support format param for specifying image type
- Anthropic -
/v1/messages
endpoint -thinking
param support: note: this refactors the [BETA] unified/v1/messages
endpoint, to just work for the Anthropic API. - Vertex AI - handle $id in response schema when calling vertex ai
- Support
-
Spend Tracking Improvements
-
Management Endpoints / UI
- Allow team/org filters to be searchable on the Create Key Page
- Add created_by and updated_by fields to Keys table
- Show 'user_email' on key table on UI
- (Feat) - Show Error Logs on LiteLLM UI
- UI - Allow admin to control default model access for internal users
- (UI) - Allow Internal Users to View their own logs
- (UI) Fix session handling with cookies
- Keys Page - Show 100 Keys Per Page, Use full height, increase width of key alias
-
Logging / Guardrail Integrations
-
Performance / Loadbalancing / Reliability improvements
-
General Proxy Improvements
- Raise BadRequestError when unknown model passed in request
- Enforce model access restrictions on Azure OpenAI proxy route
- Reliability fix - Handle emoji’s in text - fix orjson error
- Model Access Patch - don't overwrite litellm.anthropic_models when running auth checks
- Enable setting timezone information in docker image
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.63.2-stable
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.63.2-stable
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 190.0 | 223.19371836864636 | 6.25209576552295 | 0.0033451555727784642 | 1869 | 1 | 89.92210900004238 | 1948.821826000028 |
Aggregated | Passed ✅ | 190.0 | 223.19371836864636 | 6.25209576552295 | 0.0033451555727784642 | 1869 | 1 | 89.92210900004238 | 1948.821826000028 |