Skip to content

Releases: BerriAI/litellm

v1.40.9

12 Jun 07:17
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.40.8...v1.40.9

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.9

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 95 118.26463258740928 6.42020613574963 0.0 1922 0 78.571060999991 1634.9082140000064
Aggregated Passed ✅ 95 118.26463258740928 6.42020613574963 0.0 1922 0 78.571060999991 1634.9082140000064

1.40.8.dev1

12 Jun 06:20
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.40.8-stable...1.40.8.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-1.40.8.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 110.0 125.08247401976035 6.426077578390951 0.0 1923 0 91.96702899998854 1106.7971329999864
Aggregated Passed ✅ 110.0 125.08247401976035 6.426077578390951 0.0 1923 0 91.96702899998854 1106.7971329999864

v1.40.8-stable

11 Jun 19:28
Compare
Choose a tag to compare

Full Changelog: v1.40.8...v1.40.8-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.8-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 120.0 147.76399794426771 6.297998256290574 0.0 1884 0 97.42064800002481 1621.3958460000413
Aggregated Passed ✅ 120.0 147.76399794426771 6.297998256290574 0.0 1884 0 97.42064800002481 1621.3958460000413

v1.40.8

11 Jun 06:24
Compare
Choose a tag to compare

What's Changed

Client Side Fallbacks: https://docs.litellm.ai/docs/proxy/reliability#test---client-side-fallbacks

fallbacks py

Full Changelog: v1.40.7...v1.40.8

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.8

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.8

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 140.0 169.11120714803027 6.281005310183787 0.0 1878 0 114.50119100004486 1457.4686270000257
Aggregated Passed ✅ 140.0 169.11120714803027 6.281005310183787 0.0 1878 0 114.50119100004486 1457.4686270000257

v1.40.7.dev1

10 Jun 20:43
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.40.7...v1.40.7.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.7.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 140.0 179.79878830216478 6.323646865102133 0.0 1893 0 111.88137199997072 2245.1254659999904
Aggregated Passed ✅ 140.0 179.79878830216478 6.323646865102133 0.0 1893 0 111.88137199997072 2245.1254659999904

v1.40.7

08 Jun 16:43
Compare
Choose a tag to compare

Full Changelog: v1.40.6...v1.40.7

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.7

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 97 126.50565680197539 6.4278560269757214 0.003340881510902142 1924 1 82.64289499999222 1316.4627209999935
Aggregated Passed ✅ 97 126.50565680197539 6.4278560269757214 0.003340881510902142 1924 1 82.64289499999222 1316.4627209999935

v1.40.6

08 Jun 02:54
Compare
Choose a tag to compare

🚨 Note: LiteLLM Proxy Added opentelemetry as a dependency on this release. We recommend waiting for a stable release before upgrading your production instances

LiteLLM Python SDK Users: You should be unaffected by this change (opentelemetry was only added for the proxy server)

🔥 LiteLLM 1.40.6 - Proxy 100+ LLMs AT Scale with our production grade OpenTelemetry logger. Trace LLM API Calls, DB Requests, Cache Cache Requests 👉 Start here: https://docs.litellm.ai/docs/proxy/logging#logging-proxy-inputoutput-in-opentelemetry-format

🐞 [Fix]- Allow redacting messages from slack alerting https://docs.litellm.ai/docs/proxy/alerting#advanced---redacting-messages-from-alerts

🔨 [Refactor] - Refactor proxy_server.py to use common function for add_litellm_data_to_request

✨ [Feat] OpenTelemetry - Log Exceptions from Proxy Server

✨ [FEAT] OpenTelemetry - Log Redis Cache Read / Writes

✨ [FEAT] OpenTelemetry - LOG DB Exceptions

✨ [Feat] OpenTelemetry - Instrument DB Reads

🐞 [Fix] UI - Allow custom logout url and show proxy base url on API Ref Page

Xnapper-2024-06-07-21 44 06

What's Changed

New Contributors

Full Changelog: v1.40.5...v1.40.6

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.6

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 130.0 151.53218399526997 6.362696017911015 0.0 1903 0 109.01354200001379 1319.1295889999992
Aggregated Passed ✅ 130.0 151.53218399526997 6.362696017911015 0.0 1903 0 109.01354200001379 1319.1295889999992

v1.40.5

07 Jun 19:56
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.40.4...v1.40.5

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.5

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 98 123.75303621190369 6.512790176735744 0.0 1949 0 80.83186400000386 1991.117886999973
Aggregated Passed ✅ 98 123.75303621190369 6.512790176735744 0.0 1949 0 80.83186400000386 1991.117886999973

v1.40.4

06 Jun 05:17
Compare
Choose a tag to compare

What's Changed

  • feat: clarify slack alerting message by @nibalizer in #4023
  • [Admin UI] Analytics - fix div by 0 error on /model/metrics by @ishaan-jaff in #4021
  • Use DEBUG level for curl command logging by @grav in #2980
  • feat(create_user_button.tsx): allow admin to invite user to proxy via user-email/pwd invite-links by @krrishdholakia in #4028
  • [FIX] Proxy redirect to PROXY_BASE_URL/ui after logging in by @ishaan-jaff in #4027
  • [Feat] Audit Logs for Key, User, ProxyModel CRUD operations by @ishaan-jaff in #4030

New Contributors

Full Changelog: v1.40.3...v1.40.4

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.4

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.4

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 74 89.43947919222931 6.450062450815326 0.0 1930 0 64.37952199996744 1143.0389689999743
Aggregated Passed ✅ 74 89.43947919222931 6.450062450815326 0.0 1930 0 64.37952199996744 1143.0389689999743

v1.40.3-stable

05 Jun 19:41
4b3b1e0
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.40.3...v1.40.3-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.40.3-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 140.0 166.81647102860174 6.3100225495221665 0.0 1888 0 109.54055500008053 2288.330084999984
Aggregated Passed ✅ 140.0 166.81647102860174 6.3100225495221665 0.0 1888 0 109.54055500008053 2288.330084999984