diff --git a/litellm/main.py b/litellm/main.py index 500219c26..09e409293 100644 --- a/litellm/main.py +++ b/litellm/main.py @@ -163,11 +163,13 @@ def completion( if custom_llm_provider == "azure": # azure configs openai.api_type = "azure" - openai.api_base = ( - litellm.api_base - if litellm.api_base is not None - else get_secret("AZURE_API_BASE") + + api_base = ( + api_base + or litellm.api_base + or get_secret("AZURE_API_BASE") ) + openai.api_version = ( litellm.api_version if litellm.api_version is not None @@ -177,8 +179,7 @@ def completion( api_key = litellm.azure_key elif not api_key and get_secret("AZURE_API_KEY"): api_key = get_secret("AZURE_API_KEY") - # set key - openai.api_key = api_key + ## LOGGING logging.pre_call( input=messages, @@ -190,19 +191,14 @@ def completion( }, ) ## COMPLETION CALL - if litellm.headers: - response = openai.ChatCompletion.create( - engine=model, - messages=messages, - headers=litellm.headers, - **optional_params, - ) - else: - response = openai.ChatCompletion.create( - engine=model, - messages=messages, - **optional_params - ) + response = openai.ChatCompletion.create( + engine=model, + messages=messages, + headers=litellm.headers, + api_key=api_key, + api_base=api_base, + **optional_params, + ) if "stream" in optional_params and optional_params["stream"] == True: response = CustomStreamWrapper(response, model, logging_obj=logging) return response @@ -253,7 +249,7 @@ def completion( headers=litellm.headers, # None by default api_base=api_base, # thread safe setting base, key, api_version api_key=api_key, - api_version=api_version # default None + api_version=api_version, # default None **optional_params, ) except Exception as e: