mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-27 11:43:54 +00:00
fix - reuse client initialized on proxy config
This commit is contained in:
parent
392d68e7b9
commit
db4b9b8a84
2 changed files with 16 additions and 5 deletions
|
@ -812,7 +812,7 @@ class AzureChatCompletion(BaseLLM):
|
|||
azure_client_params: dict,
|
||||
api_key: str,
|
||||
input: list,
|
||||
client=None,
|
||||
client: Optional[AsyncAzureOpenAI] = None,
|
||||
logging_obj=None,
|
||||
timeout=None,
|
||||
):
|
||||
|
@ -911,6 +911,7 @@ class AzureChatCompletion(BaseLLM):
|
|||
model_response=model_response,
|
||||
azure_client_params=azure_client_params,
|
||||
timeout=timeout,
|
||||
client=client,
|
||||
)
|
||||
return response
|
||||
if client is None:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue