mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
fix - reuse client initialized on proxy config
This commit is contained in:
parent
57852bada9
commit
b7bca0af6c
2 changed files with 16 additions and 5 deletions
|
@ -812,7 +812,7 @@ class AzureChatCompletion(BaseLLM):
|
|||
azure_client_params: dict,
|
||||
api_key: str,
|
||||
input: list,
|
||||
client=None,
|
||||
client: Optional[AsyncAzureOpenAI] = None,
|
||||
logging_obj=None,
|
||||
timeout=None,
|
||||
):
|
||||
|
@ -911,6 +911,7 @@ class AzureChatCompletion(BaseLLM):
|
|||
model_response=model_response,
|
||||
azure_client_params=azure_client_params,
|
||||
timeout=timeout,
|
||||
client=client,
|
||||
)
|
||||
return response
|
||||
if client is None:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue