mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-24 18:24:20 +00:00
docs responses routing
This commit is contained in:
parent
6cd8330fc5
commit
f8323b5441
1 changed files with 0 additions and 7 deletions
|
@ -631,10 +631,3 @@ follow_up = client.responses.create(
|
|||
|
||||
</TabItem>
|
||||
</Tabs>
|
||||
|
||||
#### How It Works
|
||||
|
||||
1. When a user makes an initial request to the Responses API, LiteLLM caches which model deployment that returned the specific response. (Stored in Redis if you connected LiteLLM to Redis)
|
||||
2. When a subsequent request includes `previous_response_id`, LiteLLM automatically routes it to the same deployment
|
||||
3. If the original deployment is unavailable, or if the `previous_response_id` isn't found in the cache, LiteLLM falls back to normal routing
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue