mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
(fix) proxy: bug non OpenAI LLMs
This commit is contained in:
parent
943bf53b0b
commit
13261287ec
1 changed files with 4 additions and 4 deletions
|
@ -1011,14 +1011,14 @@ class Router:
|
|||
"""
|
||||
if client_type == "async":
|
||||
if kwargs.get("stream") == True:
|
||||
return deployment["stream_async_client"]
|
||||
return deployment.get("stream_async_client", None)
|
||||
else:
|
||||
return deployment["async_client"]
|
||||
return deployment.get("async_client", None)
|
||||
else:
|
||||
if kwargs.get("stream") == True:
|
||||
return deployment["stream_client"]
|
||||
return deployment.get("stream_client", None)
|
||||
else:
|
||||
return deployment["client"]
|
||||
return deployment.get("client", None)
|
||||
|
||||
def print_verbose(self, print_statement):
|
||||
if self.set_verbose or litellm.set_verbose:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue