Skip to content

Releases: BerriAI/litellm

v1.61.20.rc

28 Feb 04:17
1e7b9cf
Compare
Choose a tag to compare
v1.61.20.rc Pre-release
Pre-release

What's Changed

  • Add new gpt-4.5-preview model + other updates by @krrishdholakia in #8879
  • (fix) Anthropic pass through cost tracking by @ishaan-jaff in #8874
  • (Bug Fix) - Accurate token counting for /anthropic/ API Routes on LiteLLM Proxy by @ishaan-jaff in #8880
  • (fix) Pass through spend tracking - ensure custom_llm_provider is tracked for Vertex, Google AI Studio, Anthropic by @ishaan-jaff in #8882

Full Changelog: v1.61.13-stable...v1.61.20.rc

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.20.rc

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 271.87043339059113 6.110384289826154 0.003342660990058071 1828 1 218.3162040000184 3401.295629999993
Aggregated Passed ✅ 250.0 271.87043339059113 6.110384289826154 0.003342660990058071 1828 1 218.3162040000184 3401.295629999993

v1.61.20-nightly

27 Feb 21:43
Compare
Choose a tag to compare

What's Changed

  • vertex ai anthropic thinking param support + cost calculation for vertex_ai/claude-3-7-sonnet by @krrishdholakia in #8853

Full Changelog: v1.61.19-nightly...v1.61.20-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.20-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 274.35423830943404 6.203695089639108 0.0033443100213687914 1855 1 99.50650100000757 3744.788088000007
Aggregated Passed ✅ 250.0 274.35423830943404 6.203695089639108 0.0033443100213687914 1855 1 99.50650100000757 3744.788088000007

v1.61.19.dev1

27 Feb 20:25
Compare
Choose a tag to compare

What's Changed

  • vertex ai anthropic thinking param support + cost calculation for vertex_ai/claude-3-7-sonnet by @krrishdholakia in #8853

Full Changelog: v1.61.19-nightly...v1.61.19.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.19.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 279.81049794793904 6.161760602979549 0.0 1844 0 217.23908899997468 4265.23701100001
Aggregated Passed ✅ 250.0 279.81049794793904 6.161760602979549 0.0 1844 0 217.23908899997468 4265.23701100001

v1.61.19-nightly

27 Feb 06:27
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.61.17-nightly...v1.61.19-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.19-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 190.0 197.9691385267413 6.185555293319119 0.003341737057438746 1851 1 53.4934159999807 967.1408859999815
Aggregated Passed ✅ 190.0 197.9691385267413 6.185555293319119 0.003341737057438746 1851 1 53.4934159999807 967.1408859999815

v1.61.17.dev1

27 Feb 01:31
Compare
Choose a tag to compare

What's Changed

  • fix(get_litellm_params.py): handle no-log being passed in via kwargs by @krrishdholakia in #8830
  • fix(o_series_transformation.py): fix optional param check for o-serie… by @krrishdholakia in #8787
  • chore: set ttlSecondsAfterFinished on the migration job in the litellm-helm chart by @ashwin153 in #8593

New Contributors

Full Changelog: v1.61.17-nightly...v1.61.17.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.17.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 170.0 197.91492646891462 6.342832402753434 6.342832402753434 1898 1898 138.25161099998695 4268.093897000028
Aggregated Failed ❌ 170.0 197.91492646891462 6.342832402753434 6.342832402753434 1898 1898 138.25161099998695 4268.093897000028

v1.61.17-nightly

26 Feb 19:14
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.61.16-nightly...v1.61.17-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.17-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.17-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 110.0 132.37492176765105 6.343856763236461 6.343856763236461 1898 1898 93.89094700003398 3315.9179240000185
Aggregated Failed ❌ 110.0 132.37492176765105 6.343856763236461 6.343856763236461 1898 1898 93.89094700003398 3315.9179240000185

v1.61.16-nightly

25 Feb 07:27
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.61.15-nightly...v1.61.16-nightly

## Docker Run LiteLLM Proxy

```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.16-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 150.0 173.09079325414655 6.248259928705226 6.248259928705226 1869 1869 131.67032300003711 1529.8640780000028
Aggregated Failed ❌ 150.0 173.09079325414655 6.248259928705226 6.248259928705226 1869 1869 131.67032300003711 1529.8640780000028

v1.61.15-nightly

23 Feb 08:11
566d935
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.61.13-nightly...v1.61.15-nightly

## Docker Run LiteLLM Proxy

```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.15-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 140.0 149.21232461729608 6.452882828983183 6.452882828983183 1931 1931 114.22628599996187 662.278525000005
Aggregated Failed ❌ 140.0 149.21232461729608 6.452882828983183 6.452882828983183 1931 1931 114.22628599996187 662.278525000005

v1.61.13-stable

23 Feb 04:16
Compare
Choose a tag to compare

Full Changelog: v1.61.13-nightly...v1.61.13-stable

🚨 Known Issue:

  • DD Trace was enabled by default on litellm docker: #8788
  • Expect a patched v1.61.13-stable with the fix

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.61.13-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 180.0 204.35323738803527 6.313934995711798 6.313934995711798 1889 1889 146.12962300003574 2180.2391240000247
Aggregated Failed ❌ 180.0 204.35323738803527 6.313934995711798 6.313934995711798 1889 1889 146.12962300003574 2180.2391240000247

v1.55.8-stable-patched

22 Feb 16:45
Compare
Choose a tag to compare

Full Changelog: v1.55.8-stable...v1.55.8-stable-patched

## Docker Run LiteLLM Proxy

```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm_stable_release_branch-v1.55.8-stable-patched

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 150.0 172.39167931961987 6.334948735889217 6.334948735889217 1896 1896 131.01931900001773 2316.6445349999663
Aggregated Failed ❌ 150.0 172.39167931961987 6.334948735889217 6.334948735889217 1896 1896 131.01931900001773 2316.6445349999663