mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 10:44:24 +00:00
docs responses routing
This commit is contained in:
parent
a7db0df043
commit
ebfff975d4
1 changed files with 0 additions and 7 deletions
|
@ -631,10 +631,3 @@ follow_up = client.responses.create(
|
||||||
|
|
||||||
</TabItem>
|
</TabItem>
|
||||||
</Tabs>
|
</Tabs>
|
||||||
|
|
||||||
#### How It Works
|
|
||||||
|
|
||||||
1. When a user makes an initial request to the Responses API, LiteLLM caches which model deployment that returned the specific response. (Stored in Redis if you connected LiteLLM to Redis)
|
|
||||||
2. When a subsequent request includes `previous_response_id`, LiteLLM automatically routes it to the same deployment
|
|
||||||
3. If the original deployment is unavailable, or if the `previous_response_id` isn't found in the cache, LiteLLM falls back to normal routing
|
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue