(fix) get_response_headers for Azure OpenAI (#6344)

* fix get_response_headers

* unit testing for get headers

* unit testing for anthropic / azure openai headers

* increase test coverage for test_completion_response_ratelimit_headers

* fix test rate limit headers
This commit is contained in:
Ishaan Jaff 2024-10-21 20:41:35 +05:30 committed by GitHub
parent fb523b79e9
commit 274bf3e48d
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
6 changed files with 316 additions and 30 deletions

View file

@ -0,0 +1,79 @@
import json
import os
import sys
from datetime import datetime
sys.path.insert(
0, os.path.abspath("../../")
) # Adds the parent directory to the system path
import litellm
import pytest
from litellm.litellm_core_utils.llm_response_utils.get_headers import (
get_response_headers,
_get_llm_provider_headers,
)
def test_get_response_headers_empty():
result = get_response_headers()
assert result == {}, "Expected empty dictionary for no input"
def test_get_response_headers_with_openai_headers():
"""
OpenAI headers are forwarded as is
Other headers are prefixed with llm_provider-
"""
input_headers = {
"x-ratelimit-limit-requests": "100",
"x-ratelimit-remaining-requests": "50",
"x-ratelimit-limit-tokens": "1000",
"x-ratelimit-remaining-tokens": "500",
"other-header": "value",
}
expected_output = {
"x-ratelimit-limit-requests": "100",
"x-ratelimit-remaining-requests": "50",
"x-ratelimit-limit-tokens": "1000",
"x-ratelimit-remaining-tokens": "500",
"llm_provider-x-ratelimit-limit-requests": "100",
"llm_provider-x-ratelimit-remaining-requests": "50",
"llm_provider-x-ratelimit-limit-tokens": "1000",
"llm_provider-x-ratelimit-remaining-tokens": "500",
"llm_provider-other-header": "value",
}
result = get_response_headers(input_headers)
assert result == expected_output, "Unexpected output for OpenAI headers"
def test_get_response_headers_without_openai_headers():
"""
Non-OpenAI headers are prefixed with llm_provider-
"""
input_headers = {"custom-header-1": "value1", "custom-header-2": "value2"}
expected_output = {
"llm_provider-custom-header-1": "value1",
"llm_provider-custom-header-2": "value2",
}
result = get_response_headers(input_headers)
assert result == expected_output, "Unexpected output for non-OpenAI headers"
def test_get_llm_provider_headers():
"""
If non OpenAI headers are already prefixed with llm_provider- they are not prefixed with llm_provider- again
"""
input_headers = {
"header1": "value1",
"header2": "value2",
"llm_provider-existing": "existing_value",
}
expected_output = {
"llm_provider-header1": "value1",
"llm_provider-header2": "value2",
"llm_provider-existing": "existing_value",
}
result = _get_llm_provider_headers(input_headers)
assert result == expected_output, "Unexpected output for _get_llm_provider_headers"