mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 19:24:27 +00:00
fix(router.py): check for request_timeout in acompletion
support 'request_timeout' param in router acompletion
This commit is contained in:
parent
f7713cff05
commit
432b7ae264
2 changed files with 6 additions and 3 deletions
|
@ -1,5 +1,5 @@
|
||||||
model_list:
|
model_list:
|
||||||
- model_name: groq-whisper
|
- model_name: llama-3
|
||||||
litellm_params:
|
litellm_params:
|
||||||
model: groq/whisper-large-v3
|
model: gpt-4
|
||||||
|
request_timeout: 1
|
|
@ -718,6 +718,9 @@ class Router:
|
||||||
data.get(
|
data.get(
|
||||||
"timeout", None
|
"timeout", None
|
||||||
) # timeout set on litellm_params for this deployment
|
) # timeout set on litellm_params for this deployment
|
||||||
|
or data.get(
|
||||||
|
"request_timeout", None
|
||||||
|
) # timeout set on litellm_params for this deployment
|
||||||
or self.timeout # timeout set on router
|
or self.timeout # timeout set on router
|
||||||
or kwargs.get(
|
or kwargs.get(
|
||||||
"timeout", None
|
"timeout", None
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue