What's Changed
- [Feat] Unified Responses API - Add Azure Responses API support by @ishaan-jaff in #10116
- UI: Make columns resizable/hideable in Models table by @msabramo in #10119
- Remove unnecessary
package*.json
files by @msabramo in #10075 - Add Gemini Flash 2.5 Preview Model Price and Context Window by @drmingler in #10125
- test: update tests to new deployment model by @krrishdholakia in #10142
- [Feat] Support for all litellm providers on Responses API (works with Codex) - Anthropic, Bedrock API, VertexAI, Ollama by @ishaan-jaff in #10132
New Contributors
- @drmingler made their first contribution in #10125
Full Changelog: v1.66.2.dev1...v1.66.3.dev5
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.66.3.dev5
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 230.0 | 241.46378394371686 | 6.1149592690003 | 0.0 | 1830 | 0 | 197.6759699999775 | 1416.5823339999974 |
Aggregated | Passed ✅ | 230.0 | 241.46378394371686 | 6.1149592690003 | 0.0 | 1830 | 0 | 197.6759699999775 | 1416.5823339999974 |