diff --git a/litellm/llms/azure.py b/litellm/llms/azure.py index ec143f3fec..9df67c26e1 100644 --- a/litellm/llms/azure.py +++ b/litellm/llms/azure.py @@ -474,21 +474,13 @@ class AzureChatCompletion(BaseLLM): - call chat.completions.create by default """ try: - if litellm.return_response_headers is True: - raw_response = ( - await azure_client.chat.completions.with_raw_response.create( - **data, timeout=timeout - ) - ) + raw_response = await azure_client.chat.completions.with_raw_response.create( + **data, timeout=timeout + ) - headers = dict(raw_response.headers) - response = raw_response.parse() - return headers, response - else: - response = await azure_client.chat.completions.create( - **data, timeout=timeout - ) - return None, response + headers = dict(raw_response.headers) + response = raw_response.parse() + return headers, response except Exception as e: raise e diff --git a/litellm/llms/openai.py b/litellm/llms/openai.py index d2ba7ac134..459c1d9370 100644 --- a/litellm/llms/openai.py +++ b/litellm/llms/openai.py @@ -768,21 +768,15 @@ class OpenAIChatCompletion(BaseLLM): - call chat.completions.create by default """ try: - if litellm.return_response_headers is True: - raw_response = ( - await openai_aclient.chat.completions.with_raw_response.create( - **data, timeout=timeout - ) - ) - - headers = dict(raw_response.headers) - response = raw_response.parse() - return headers, response - else: - response = await openai_aclient.chat.completions.create( + raw_response = ( + await openai_aclient.chat.completions.with_raw_response.create( **data, timeout=timeout ) - return None, response + ) + + headers = dict(raw_response.headers) + response = raw_response.parse() + return headers, response except Exception as e: raise e