Skip to content

Releases: BerriAI/litellm

custom-docker-image-test-dev

01 Nov 12:49
Compare
Choose a tag to compare

What's Changed

  • Support specifying exponential backoff retry strategy when calling completions() by @dbczumar in #6520

Full Changelog: v1.51.2...custom-docker-image-test-dev

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-custom-docker-image-test-dev

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-custom-docker-image-test-dev

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-custom-docker-image-test-dev

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-custom-docker-image-test-dev

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-custom-docker-image-test-dev

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 277.2980743547866 6.178135563258271 0.0 1849 0 222.02958399998352 3219.9342000000115
Aggregated Passed ✅ 250.0 277.2980743547866 6.178135563258271 0.0 1849 0 222.02958399998352 3219.9342000000115

v1.51.2

30 Oct 19:57
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.51.1...v1.51.2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.51.2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 287.73103211135674 6.039141910660648 0.0 1805 0 213.5724959999834 2930.6253560000073
Aggregated Passed ✅ 250.0 287.73103211135674 6.039141910660648 0.0 1805 0 213.5724959999834 2930.6253560000073

v1.51.1-staging

30 Oct 18:58
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.51.1...v1.51.1-staging

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.51.1-staging

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 270.0 311.93605914725106 6.080288332872121 0.0033408177653143525 1820 1 117.93499300000576 3293.080912999983
Aggregated Failed ❌ 270.0 311.93605914725106 6.080288332872121 0.0033408177653143525 1820 1 117.93499300000576 3293.080912999983

v1.51.1

29 Oct 17:01
134bd2c
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.51.0...v1.51.1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.51.1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 200.0 217.96900764879396 6.232037758758388 0.0 1865 0 178.63059899997324 1220.8741560000362
Aggregated Passed ✅ 200.0 217.96900764879396 6.232037758758388 0.0 1865 0 178.63059899997324 1220.8741560000362

v1.51.0.dev1

28 Oct 19:37
Compare
Choose a tag to compare

Full Changelog: v1.51.0...v1.51.0.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.51.0.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 270.0 304.38755454794625 6.1009349714042544 0.0 1825 0 232.0200620000037 3500.7745139999997
Aggregated Failed ❌ 270.0 304.38755454794625 6.1009349714042544 0.0 1825 0 232.0200620000037 3500.7745139999997

v1.51.0-stable

26 Oct 07:51
b3141e1
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.50.4...v1.51.0-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.51.0-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 220.0 259.348547705819 6.147561516829862 0.0 1839 0 207.74116500001583 1588.2848330000456
Aggregated Passed ✅ 220.0 259.348547705819 6.147561516829862 0.0 1839 0 207.74116500001583 1588.2848330000456

v1.51.0

26 Oct 06:55
b3141e1
Compare
Choose a tag to compare

What's Changed

Full Changelog: v1.50.4...v1.51.0

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.51.0

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 230.0 256.2776533033099 6.163517714105049 0.0 1843 0 210.4747610000004 1438.3136239999885
Aggregated Passed ✅ 230.0 256.2776533033099 6.163517714105049 0.0 1843 0 210.4747610000004 1438.3136239999885

v1.50.4-stable

24 Oct 19:21
Compare
Choose a tag to compare

What's Changed

  • (feat) Arize - Allow using Arize HTTP endpoint by @ishaan-jaff in #6364
  • LiteLLM Minor Fixes & Improvements (10/22/2024) by @krrishdholakia in #6384
  • build(deps): bump http-proxy-middleware from 2.0.6 to 2.0.7 in /docs/my-website by @dependabot in #6395
  • (docs + testing) Correctly document the timeout value used by litellm proxy is 6000 seconds + add to best practices for prod by @ishaan-jaff in #6339
  • (refactor) move convert dict to model response to llm_response_utils/ by @ishaan-jaff in #6393
  • (refactor) litellm.Router client initialization utils by @ishaan-jaff in #6394
  • (fix) Langfuse key based logging by @ishaan-jaff in #6372
  • Revert "(refactor) litellm.Router client initialization utils " by @ishaan-jaff in #6403
  • (fix) using /completions with echo by @ishaan-jaff in #6401
  • (refactor) prometheus async_log_success_event to be under 100 LOC by @ishaan-jaff in #6416
  • (refactor) router - use static methods for client init utils by @ishaan-jaff in #6420
  • (code cleanup) remove unused and undocumented logging integrations - litedebugger, berrispend by @ishaan-jaff in #6406

Full Changelog: v1.50.2...v1.50.4-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.50.4-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 250.0 282.12398419383265 6.070437193170591 0.0 1816 0 215.3092099999867 6805.4257369999505
Aggregated Passed ✅ 250.0 282.12398419383265 6.070437193170591 0.0 1816 0 215.3092099999867 6805.4257369999505

v1.50.4

24 Oct 18:27
Compare
Choose a tag to compare

What's Changed

  • (feat) Arize - Allow using Arize HTTP endpoint by @ishaan-jaff in #6364
  • LiteLLM Minor Fixes & Improvements (10/22/2024) by @krrishdholakia in #6384
  • build(deps): bump http-proxy-middleware from 2.0.6 to 2.0.7 in /docs/my-website by @dependabot in #6395
  • (docs + testing) Correctly document the timeout value used by litellm proxy is 6000 seconds + add to best practices for prod by @ishaan-jaff in #6339
  • (refactor) move convert dict to model response to llm_response_utils/ by @ishaan-jaff in #6393
  • (refactor) litellm.Router client initialization utils by @ishaan-jaff in #6394
  • (fix) Langfuse key based logging by @ishaan-jaff in #6372
  • Revert "(refactor) litellm.Router client initialization utils " by @ishaan-jaff in #6403
  • (fix) using /completions with echo by @ishaan-jaff in #6401
  • (refactor) prometheus async_log_success_event to be under 100 LOC by @ishaan-jaff in #6416
  • (refactor) router - use static methods for client init utils by @ishaan-jaff in #6420
  • (code cleanup) remove unused and undocumented logging integrations - litedebugger, berrispend by @ishaan-jaff in #6406

Full Changelog: v1.50.2...v1.50.4

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.50.4

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Failed ❌ 280.0 312.6482922531862 6.037218908394318 0.0 1805 0 231.8999450000092 2847.2051709999846
Aggregated Failed ❌ 280.0 312.6482922531862 6.037218908394318 0.0 1805 0 231.8999450000092 2847.2051709999846

v1.50.1.dev1

23 Oct 12:43
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.50.1...v1.50.1.dev1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.50.1.dev1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 200.0 220.3880747854055 6.181213384368117 0.0 1850 0 179.4118180000055 2854.2284040000254
Aggregated Passed ✅ 200.0 220.3880747854055 6.181213384368117 0.0 1850 0 179.4118180000055 2854.2284040000254