What's Changed
- [docs release notes] by @ishaan-jaff in #13237
- UI - Add giving keys prompt access by @krrishdholakia in #13233
- Prompt Management - Add table + prompt info page to UI by @krrishdholakia in #13232
- Prompt Management - add prompts on UI by @krrishdholakia in #13240
- [Proxy] Add OpenShift Support to non root docker image by @jugaldb in #13239
- [LLM Translation] Fix Model Usage not having text tokens by @jugaldb in #13234
- [UI] Add team deletion check for teams with keys by @jugaldb in #12953
- [Bug Fix] OpenAI / Azure Responses API - Add
service_tier
,safety_identifier
supported params by @ishaan-jaff in #13258 - Bug Fix - Responses API raises error with Gemini Tool Calls in
input
by @ishaan-jaff in #13260 - Minor formatting changes to token-cost.json by @bbartels in #13244
- [Bug Fix] Fix Server root path regression on UI when using "Login" by @ishaan-jaff in #13267
- Support OCI provider by @breno-aumo in #13206
- Add GCS bucket caching support by @pascalwhoop in #13122
- Fix: Langfuse reporting "client closed" error due to httpx client TTL by @ZPerling in #13045
- [LLM Translation] Correct pricing for web search on 4o-mini by @jugaldb in #13269
- [Bug Fix] Prometheus - fix for
litellm_input_tokens_metric
,litellm_output_tokens_metric
- Note this updates the metric name by @ishaan-jaff in #13271 - [LLM Translation] input cost per token higher than $1 test by @jugaldb in #13270
- [LLM Translation] Support /v1/models/{model_id} retrieval by @jugaldb in #13268
- [UI] - Add ability to set model alias per key/team by @ishaan-jaff in #13276
- fix OCI linting errors by @krrishdholakia in #13279
- Ensure disable_llm_api_endpoints works + Add wildcard model support for 'team-byok' model by @krrishdholakia in #13278
- fix(main.py): handle tool being a pydantic object + Fix unpack defs deepcopy issue for bedrock by @krrishdholakia in #13274
- Revert "Fix: Langfuse reporting "client closed" error due to httpx client TTL" by @ishaan-jaff in #13291
New Contributors
- @bbartels made their first contribution in #13244
- @breno-aumo made their first contribution in #13206
- @pascalwhoop made their first contribution in #13122
- @ZPerling made their first contribution in #13045
Full Changelog: 1.74.15.rc.1...v1.75.0-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 120.0 | 177.10187576854196 | 6.352075436390092 | 0.0 | 1901 | 0 | 81.98079900000721 | 2046.5963869999655 |
Aggregated | Passed ✅ | 120.0 | 177.10187576854196 | 6.352075436390092 | 0.0 | 1901 | 0 | 81.98079900000721 | 2046.5963869999655 |