mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 02:34:29 +00:00
docs responses routing
This commit is contained in:
parent
a7db0df043
commit
ebfff975d4
1 changed files with 0 additions and 7 deletions
|
@ -631,10 +631,3 @@ follow_up = client.responses.create(
|
|||
|
||||
</TabItem>
|
||||
</Tabs>
|
||||
|
||||
#### How It Works
|
||||
|
||||
1. When a user makes an initial request to the Responses API, LiteLLM caches which model deployment that returned the specific response. (Stored in Redis if you connected LiteLLM to Redis)
|
||||
2. When a subsequent request includes `previous_response_id`, LiteLLM automatically routes it to the same deployment
|
||||
3. If the original deployment is unavailable, or if the `previous_response_id` isn't found in the cache, LiteLLM falls back to normal routing
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue