github BerriAI/litellm v1.63.14-nightly

latest releases: v1.65.1-nightly, v1.65.0-stable, v1.65.0.rc...
10 days ago

What's Changed

  • Modify completion handler for SageMaker to use payload from prepared_request by @andjsmi in #9326
  • Arize integration Fix by @nate-mar in #9338
  • Fix get_llm_provider error + add model name to tpm/rpm cache key (enables wildcard models to work w/ usage-based-routing) by @krrishdholakia in #9355
  • Litellm dev 03 18 2025 p1 v3 by @krrishdholakia in #9354
  • [Bug Fix] Arize AI Logging Integration with LiteLLM by @ishaan-jaff in #9352
  • build(model_prices_and_context_window.json): fix azure gpt-4o pricing by @krrishdholakia in #9361
  • Contributor PR: Fix sagemaker too little data for content error by @krrishdholakia in #9335
  • [Feat] - API - Allow using dynamic Arize AI Spaces on LiteLLM by @ishaan-jaff in #9353
  • fix(common_utils.py): handle cris only model by @krrishdholakia in #9363
  • docs(litellm_proxy): correct parameter assignment in litellm proxy docs by @colesmcintosh in #9375
  • Feature flag checking LiteLLM_CredentialsTable by @krrishdholakia in #9376
  • fix(lowest_tpm_rpm_v2.py): support batch writing increments to redis by @krrishdholakia in #9357
  • Support 'prisma migrate' for db schema changes by @krrishdholakia in #9382
  • Fix latency redis by @emerzon in #9387
  • Revert "Fix latency redis" by @krrishdholakia in #9388
  • build(model_prices_and_context_window.json): add o1-pro pricing by @krrishdholakia in #9397
  • [Bug Fix] - Azure OpenAI - ensure SSL verification runs by @ishaan-jaff in #9341
  • [Feat] - Allow building custom prompt management integration by @ishaan-jaff in #9384
  • Litellm fix icons by @azdolinski in #9374
  • [UI Improvement] Use local icons for model providers instead of downloading them by @ishaan-jaff in #9399
  • fix(internal_user_endpoints.py): re-introduce upsert on user not found by @krrishdholakia in #9395
  • docs: Include Phoenix Page into sidebar under observability by @SrilakshmiC in #9332
  • fix(types/utils.py): support openai 'file' message type by @krrishdholakia in #9402
  • feat: Add support for custom OPENROUTER_API_BASE via get_secret in co… by @graysonchen in #9369
  • fix: VertexAI outputDimensionality configuration by @JamesGuthrie in #9366
  • docs(vertexai): fix typo in required env variables by @Shiti in #9325
  • Update perplexity.md by @jollyolliel in #9290
  • fix: VoyageAI prompt_token always empty by @lucasra1 in #9260
  • build(deps): bump litellm from 1.55.3 to 1.61.15 in /cookbook/litellm-ollama-docker-image by @dependabot in #9422
  • [Feat] OpenAI o1-pro Responses API streaming support by @ishaan-jaff in #9419
  • [Feat] Add OpenAI o1-pro support on Responses API by @ishaan-jaff in #9415
  • [Docs - Draft] LiteLLM x MCP Interface by @ishaan-jaff in #9435
  • support returning api-base on pass-through endpoints + consistently return 404 if team not found in DB by @krrishdholakia in #9439
  • fix(handle_error.py): make cooldown error more descriptive by @krrishdholakia in #9438
  • Consistent anthropic response_format streaming/non-streaming behaviour by @krrishdholakia in #9437
  • New Azure Models (GPT-4.5-Preview, Mistral Small 3.1) by @emerzon in #9453
  • Set max size limit to in-memory cache item - prevents OOM errors by @krrishdholakia in #9448
  • fix(model_param_helper.py): update _get_litellm_supported_transcription_kwargs() to use proper annotations from TranscriptionCreateParamsNonStreaming & ``TranscriptionCreateParamsStreaming` by @hsaeed3 in #9451
  • [Feat] LiteLLM x MCP Bridge - Use MCP Tools with LiteLLM by @ishaan-jaff in #9436
  • fix(model_param_helper.py): update _get_litellm_supported_transcription_kwargs() to use proper annotations from TranscriptionCreateParamsNonStreaming & `TranscriptionCreateParamsStreaming by @ishaan-jaff in #9455

New Contributors

Full Changelog: v1.63.12-nightly...v1.63.14-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.63.14-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 210.0 225.80703369680768 6.282456357875219 0.0 1880 0 182.8801230000181 2803.2175499999994
Aggregated Passed ✅ 210.0 225.80703369680768 6.282456357875219 0.0 1880 0 182.8801230000181 2803.2175499999994

Don't miss a new litellm release

NewReleases is sending notifications on new releases.