Skip to content

Releases: BerriAI/litellm

v1.75.5-stable

17 Aug 21:42
Compare
Choose a tag to compare

Full Changelog: v1.75.5.rc.1...v1.75.5-stable

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:litellm_stable_release_branch-v1.75.5-stable

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 170.0 277.36207599076045 6.146979185191522 0.0 1840 0 131.0856249999688 12158.679278999954
Aggregated Passed ✅ 170.0 277.36207599076045 6.146979185191522 0.0 1840 0 131.0856249999688 12158.679278999954

v1.75.8-nightly

16 Aug 22:30
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.75.7-nightly...v1.75.8-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.8-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 140.0 192.18510385185118 6.314985319551383 0.0 1890 0 114.82424099995114 1309.5362409999325
Aggregated Passed ✅ 140.0 192.18510385185118 6.314985319551383 0.0 1890 0 114.82424099995114 1309.5362409999325

v1.75.7-nightly

15 Aug 06:02
Compare
Choose a tag to compare

What's Changed

  • [Proxy] LiteLLM mock test fix by @jugaldb in #13635
  • [Proxy] Litellm add DB metrics to prometheus by @jugaldb in #13626
  • [LLM Translation] Fix Realtime API endpoint for no intent by @jugaldb in #13476
  • [MCP Gateway] LiteLLM Fix MCP gateway key auth by @jugaldb in #13630
  • [Fix] Ensure /messages works when using `bedrock/converse/ with LiteLLM by @ishaan-jaff in #13627
  • UI - Fix image overflow in LiteLLM model by @ishaan-jaff in #13639
  • [Bug Fix] /messages endpoint - ensure tool use arguments are returned for non-anthropic models by @ishaan-jaff in #13638

Full Changelog: v1.75.6-nightly...v1.75.7-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.7-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 140.0 182.49004836898376 6.247441314306712 0.0 1870 0 114.26430999995318 2038.3160259999613
Aggregated Passed ✅ 140.0 182.49004836898376 6.247441314306712 0.0 1870 0 114.26430999995318 2038.3160259999613

v1.75.6-nightly

14 Aug 23:48
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.75.5.rc.1...v1.75.6-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.6-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 110.0 152.45377470174589 6.50695562099742 0.0 1948 0 86.13810599996441 2202.5806519999946
Aggregated Passed ✅ 110.0 152.45377470174589 6.50695562099742 0.0 1948 0 86.13810599996441 2202.5806519999946

litellm_v1.75.5-dev_memory_fix_2

13 Aug 00:59
Compare
Choose a tag to compare

Full Changelog: litellm_v1.75.5-dev_memory_fix...litellm_v1.75.5-dev_memory_fix_2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.75.5-dev_memory_fix_2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 110.0 148.3727584527906 6.403502625001942 0.0 1917 0 82.2168479999732 1009.9184229999878
Aggregated Passed ✅ 110.0 148.3727584527906 6.403502625001942 0.0 1917 0 82.2168479999732 1009.9184229999878

litellm_v1.73.0-dev_memory_fix_2

14 Aug 00:09
Compare
Choose a tag to compare

Full Changelog: litellm_v1.73.0-dev_memory_fix_1...litellm_v1.73.0-dev_memory_fix_2

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.73.0-dev_memory_fix_2

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 160.0 197.89495319271438 6.333208452490925 0.0 1894 0 120.76194100001203 1698.8860899999736
Aggregated Passed ✅ 160.0 197.89495319271438 6.333208452490925 0.0 1894 0 120.76194100001203 1698.8860899999736

litellm_v1.73.0-dev_memory_fix_1

13 Aug 22:49
Compare
Choose a tag to compare

Full Changelog: litellm_v1.73.0-dev_memory_fix...litellm_v1.73.0-dev_memory_fix_1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.73.0-dev_memory_fix_1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 130.0 168.52018052761707 6.353625704455274 0.0 1901 0 102.81331199996657 955.3499159999888
Aggregated Passed ✅ 130.0 168.52018052761707 6.353625704455274 0.0 1901 0 102.81331199996657 955.3499159999888

litellm_v1.73.0-dev_memory_fix

13 Aug 05:01
Compare
Choose a tag to compare

Full Changelog: v1.73.0.rc.1...litellm_v1.73.0-dev_memory_fix

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.73.0-dev_memory_fix

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 170.0 201.69670727064485 6.271320946816614 0.0 1877 0 124.88003300001083 1273.2718879999823
Aggregated Passed ✅ 170.0 201.69670727064485 6.271320946816614 0.0 1877 0 124.88003300001083 1273.2718879999823

v1.75.5.dev3

12 Aug 23:07
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v1.75.5.rc.1...v1.75.5.dev3

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.75.5.dev3

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 170.0 209.8803997534473 6.313294138003075 0.0 1886 0 126.72262299997783 1268.0979020000223
Aggregated Passed ✅ 170.0 209.8803997534473 6.313294138003075 0.0 1886 0 126.72262299997783 1268.0979020000223

litellm_v1.75.5-dev_memory_fix_1

12 Aug 22:25
Compare
Choose a tag to compare

What's Changed

  • [Bug Fix] - Allow using reasoning_effort for gpt-5 model family and reasoning for Responses API by @ishaan-jaff in #13475
  • [Bug Fix]: Azure OpenAI GPT-5 max_tokens + reasoning param support by @ishaan-jaff in #13510
  • [Draft] [LLM Translation] Add model id check by @jugaldb in #13507

Full Changelog: v1.75.5.rc.1...litellm_v1.75.5-dev_memory_fix_1

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-litellm_v1.75.5-dev_memory_fix_1

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 180.0 211.1796445664915 6.382692480357281 0.0 1910 0 132.33785100001683 1892.3347159999935
Aggregated Passed ✅ 180.0 211.1796445664915 6.382692480357281 0.0 1910 0 132.33785100001683 1892.3347159999935