|
1 | 1 | ---
|
2 |
| -title: "[PRE-RELEASE]v1.75.5-stable" |
| 2 | +title: "v1.75.5-stable - Redis latency improvements" |
3 | 3 | slug: "v1-75-5"
|
4 | 4 | date: 2025-08-10T10:00:00
|
5 | 5 | authors:
|
@@ -43,8 +43,49 @@ pip install litellm==1.75.5.post1
|
43 | 43 |
|
44 | 44 | ---
|
45 | 45 |
|
| 46 | +## Key Highlights |
| 47 | + |
| 48 | +- **Redis - Latency Improvements** - Reduces P99 latency by 50% with Redis enabled. |
| 49 | +- **Responses API Session Management** - Support for managing responses API sessions with images. |
| 50 | +- **Oracle Cloud Infrastructure** - New LLM provider for calling models on Oracle Cloud Infrastructure. |
| 51 | +- **Digital Ocean's Gradient AI** - New LLM provider for calling models on Digital Ocean's Gradient AI platform. |
| 52 | + |
| 53 | + |
| 54 | +### Risk of Upgrade |
| 55 | + |
| 56 | +If you build the proxy from the pip package, you should hold off on upgrading. This version makes `prisma migrate deploy` our default for managing the DB. This is safer, as it doesn't reset the DB, but it requires a manual `prisma generate` step. |
| 57 | + |
| 58 | +Users of our Docker image, are **not** affected by this change. |
| 59 | + |
46 | 60 | ---
|
47 | 61 |
|
| 62 | +## Redis Latency Improvements |
| 63 | + |
| 64 | +<Image |
| 65 | + img={require('../../img/release_notes/faster_caching_calls.png')} |
| 66 | + style={{width: '100%', display: 'block', margin: '2rem auto'}} |
| 67 | +/> |
| 68 | + |
| 69 | +<br/> |
| 70 | + |
| 71 | +This release adds in-memory caching for Redis requests, enabling faster response times in high-traffic. Now, LiteLLM instances will check their in-memory cache for a cache hit, before checking Redis. This reduces caching-related latency from 100ms for LLM API calls to sub-1ms, on cache hits. |
| 72 | + |
| 73 | +--- |
| 74 | + |
| 75 | +## Responses API Session Management w/ Images |
| 76 | + |
| 77 | +<Image |
| 78 | + img={require('../../img/release_notes/responses_api_session_mgt_images.jpg')} |
| 79 | + style={{width: '100%', display: 'block', margin: '2rem auto'}} |
| 80 | +/> |
| 81 | + |
| 82 | +<br/> |
| 83 | + |
| 84 | +LiteLLM now supports session management for Responses API requests with images. This is great for use-cases like chatbots, that are using the Responses API to track the state of a conversation. LiteLLM session management works across **ALL** LLM API's (including Anthropic, Bedrock, OpenAI, etc). LiteLLM session management works by storing the request and response content in an s3 bucket, you can specify. |
| 85 | + |
| 86 | +--- |
| 87 | + |
| 88 | + |
48 | 89 | ## New Models / Updated Models
|
49 | 90 |
|
50 | 91 | #### New Model Support
|
|
0 commit comments