Skip to content

Releases: BerriAI/litellm

v1.71.1-stable.patch1

30 May 21:08
Compare
Choose a tag to compare

Full Changelog: v1.71.1-nightly...v1.71.1-stable.patch1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.71.1-stable.patch1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 190.0 208.76252080483792 6.21531894936511 0.0 1860 0 167.59147999994184 1521.5310909999857
Aggregated Passed ✅ 190.0 208.76252080483792 6.21531894936511 0.0 1860 0 167.59147999994184 1521.5310909999857

v1.71.2.dev2

29 May 05:34
Compare
Choose a tag to compare

Full Changelog: v1.71.2-nightly...v1.71.2.dev2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.2.dev2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 240.0 257.64197439032404 6.082707498887239 0.0 1819 0 211.94133999995302 1432.5384599999893
Aggregated Passed ✅ 240.0 257.64197439032404 6.082707498887239 0.0 1819 0 211.94133999995302 1432.5384599999893

v1.71.2.dev1

29 May 02:51
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.71.2-nightly...v1.71.2.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.2.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 220.0 238.87666312181884 6.174193914087743 0.0033428229096306135 1847 1 50.488271999995504 1443.0136089999905
Aggregated Passed ✅ 220.0 238.87666312181884 6.174193914087743 0.0033428229096306135 1847 1 50.488271999995504 1443.0136089999905

v1.71.2-nightly

28 May 20:30
Compare
Choose a tag to compare

What's Changed

  • feat: Make gemini accept the openai parameter parallel_tool_calls by @aholmberg in #11125
  • Fix #9295 docker-compose healthcheck test uses curl but curl is not in the image by @agajdosi in #9737
  • [Feat] Add /image/edits support for Azure by @ishaan-jaff in #11160
  • Fix deprecation_date value for llama groq models by @kiriloman in #11151
  • [Fix] Rollback to httpx==0.27.0 by @ishaan-jaff in #11146
  • Doc update for azure openai by @ketangangal in #11161
  • Litellm fix GitHub action testing by @krrishdholakia in #11163
  • [Feat - Contributor PR] Add Video support for Bedrock Converse by @ishaan-jaff in #11166
  • [Fixes] Aiohttp transport fixes - add handling for aiohttp.ClientPayloadError and ssl_verification settings by @ishaan-jaff in #11162
  • prevent leaking sensitive keys to langfuse + support forwarding /sso/key/generate to the server root path url by @krrishdholakia in #11165
  • [Fix] - Duplicate maxTokens parameter being sent to Bedrock/Claude model with thinking by @ishaan-jaff in #11181
  • Integration with Nebius AI Studio added by @Aktsvigun in #11143
  • Codestral - return litellm latency overhead on /v1/completions + Add 'contains' support for ChatCompletionDeltaToolCall by @krrishdholakia in #10879
  • Ollama Chat - parse tool calls on streaming by @krrishdholakia in #11171
  • [Fix] Prometheus Metrics - Do not track end_user by default + expose flag to enable tracking end_user on prometheus by @ishaan-jaff in #11192
  • [Fix]: Add cost tracking for image edits endpoint [OpenAI, Azure] by @ishaan-jaff in #11186
  • VertexAI - codeExecution tool support + anyOf handling by @krrishdholakia in #11195
  • Add Pangea provider to Guardrails hook by @ryanmeans in #10775
  • Return anthropic thinking blocks on streaming + VertexAI Minor Fixes & Improvements (Thinking, Global regions, Parallel tool calling) by @krrishdholakia in #11194
  • Azure OIDC provider improvements + OIDC audience bug fix by @nikoizs in #10054
  • [Feat] Add well known MCP servers to LiteLLM by @ishaan-jaff in #11209
  • Add missing request_kwargs to get_available_deployment call by @Nitro963 in #11202

New Contributors

Full Changelog: v1.71.1-nightly...v1.71.2-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.2-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 230.0 296.34506371491256 6.07338921380297 0.0 1817 0 196.8864309999958 5947.878826000022
Aggregated Passed ✅ 230.0 296.34506371491256 6.07338921380297 0.0 1817 0 196.8864309999958 5947.878826000022

v1.71.1-stable

25 May 21:17
Compare
Choose a tag to compare

What's Changed

Read more

v1.71.1-nightly

25 May 15:14
Compare
Choose a tag to compare

What's Changed

  • Logfire - fix(opentelemetry.py): Fix otel proxy server initialization + Return abbreviated key in key not found error (easier clientside debugging) + Ignore invalid deployments on router load by @krrishdholakia in #11091
  • feat(handle_jwt.py): map user to team when added via jwt auth by @krrishdholakia in #11108
  • fix(ui_sso.py): maintain backwards compatibility for older user id formats + fix existing user email w/ trailing whitespace check + ensure default_internal_user_settings runs on all user new calls by @krrishdholakia in #11106
  • fix(route_llm_request.py): map team model from list in route llm request by @krrishdholakia in #11111
  • Remove + Check for unsafe enterprise/ folder imports by @krrishdholakia in #11107
  • Fix: Add Claude Sonnet 4 and Opus 4 support for reasoning_effort parameter by @keykbd in #11114
  • fix(session): correctly place litellm_session_id at root level instead of metadata by @dalssoft in #11088
  • fix(model_management_endpoints): clear cache and reload models after update by @jtong99 in #10853
  • [Feat] Add /image/edits on LiteLLM by @ishaan-jaff in #11123
  • Correctly delete team model alias when team only model is deleted (#… by @krrishdholakia in #11121
  • fix: detect and return status codes in streaming responses by @aholmberg in #10962
  • Fix passing standard optional params by @krrishdholakia in #11124
  • UI QA fix: team viewer should not see create team by @ishaan-jaff in #11127
  • [Chore]: feature flag aiohttp transport - users should opt into using aiohttp transport by @ishaan-jaff in #11132
  • v1.71.1-stable - notes by @ishaan-jaff in #11133
  • Litellm revert redis changes by @krrishdholakia in #11135
  • Litellm fix multi instance checks on teams by @krrishdholakia in #11137

New Contributors

Full Changelog: v1.71.0-nightly...v1.71.1-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.1-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 271.62186726419185 6.123952252359233 0.0 1832 0 215.75241199997208 1968.6522410000293
Aggregated Passed ✅ 250.0 271.62186726419185 6.123952252359233 0.0 1832 0 215.75241199997208 1968.6522410000293

v1.71.0-nightly

24 May 15:11
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.70.4-nightly...v1.71.0-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.71.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 310.0 321.9030490306742 5.995437687618243 4.166377779571852 1793 1246 259.8663090000173 771.521746000019
Aggregated Failed ❌ 310.0 321.9030490306742 5.995437687618243 4.166377779571852 1793 1246 259.8663090000173 771.521746000019

v1.70.2.dev6

24 May 04:51
Compare
Choose a tag to compare

Full Changelog: v1.70.2-nightly...v1.70.2.dev6

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.70.2.dev6

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 470.0 546.8056516133963 5.687025964404996 0.0 1702 0 432.45771799996646 2108.0635040000006
Aggregated Failed ❌ 470.0 546.8056516133963 5.687025964404996 0.0 1702 0 432.45771799996646 2108.0635040000006

v1.70.4-nightly

23 May 00:56
Compare
Choose a tag to compare

Full Changelog: v1.70.2.dev5...v1.70.4-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.70.4-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 530.0 592.8290290780234 5.619046633443039 0.0 1679 0 480.5262289999632 1595.1236809999614
Aggregated Failed ❌ 530.0 592.8290290780234 5.619046633443039 0.0 1679 0 480.5262289999632 1595.1236809999614

v1.70.2.dev5

22 May 21:26
197c608
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.70.2-nightly...v1.70.2.dev5

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.70.2.dev5

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 490.0 556.7359308349169 5.626256137716844 0.0 1684 0 437.79858300001706 2137.070654000013
Aggregated Failed ❌ 490.0 556.7359308349169 5.626256137716844 0.0 1684 0 437.79858300001706 2137.070654000013