What's Changed
- build(deps): bump @babel/runtime from 7.26.0 to 7.27.0 in /docs/my-website by @dependabot in #9934
- fix: correct the cost for 'gemini/gemini-2.5-pro-preview-03-25' by @n1lanjan in #9896
- Litellm add managed files db by @krrishdholakia in #9930
- [DB / Infra] Add new column team_member_permissions by @ishaan-jaff in #9941
- fix(factory.py): correct indentation for message index increment in ollama, This fixes bug #9822 by @djshaw01 in #9943
- fix(litellm_proxy_extras): add baselining db script by @krrishdholakia in #9942
- [Team Member permissions] - Fixes by @ishaan-jaff in #9945
- Litellm managed files docs by @krrishdholakia in #9948
- [v1.66.0-stable] Release notes by @ishaan-jaff in #9952
- [Docs] v1.66.0-stable fixes by @ishaan-jaff in #9953
- stable release note fixes by @ishaan-jaff in #9954
- Fix filtering litellm-dashboard keys for internal users + prevent flooding spend logs with admin endpoint errors by @krrishdholakia in #9955
- [UI QA checklist] by @ishaan-jaff in #9957
New Contributors
Full Changelog: v1.65.8-nightly...v1.66.0-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.66.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 230.0 | 252.49209995793416 | 6.279241190720279 | 0.0 | 1878 | 0 | 200.85592700002053 | 5135.250711999987 |
Aggregated | Passed ✅ | 230.0 | 252.49209995793416 | 6.279241190720279 | 0.0 | 1878 | 0 | 200.85592700002053 | 5135.250711999987 |