forked from phoenix/litellm-mirror
test: handle anthropic api instability
This commit is contained in:
parent
30f2c93fad
commit
a37fa817a2
1 changed files with 56 additions and 61 deletions
|
@ -47,6 +47,7 @@ def _usage_format_tests(usage: litellm.Usage):
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
def test_prompt_caching_model(model):
|
def test_prompt_caching_model(model):
|
||||||
|
try:
|
||||||
for _ in range(2):
|
for _ in range(2):
|
||||||
response = litellm.completion(
|
response = litellm.completion(
|
||||||
model=model,
|
model=model,
|
||||||
|
@ -103,14 +104,8 @@ def test_prompt_caching_model(model):
|
||||||
|
|
||||||
assert "prompt_tokens_details" in response.usage
|
assert "prompt_tokens_details" in response.usage
|
||||||
assert response.usage.prompt_tokens_details.cached_tokens > 0
|
assert response.usage.prompt_tokens_details.cached_tokens > 0
|
||||||
|
except litellm.InternalServerError:
|
||||||
# assert "cache_read_input_tokens" in response.usage
|
pass
|
||||||
# assert "cache_creation_input_tokens" in response.usage
|
|
||||||
|
|
||||||
# # Assert either a cache entry was created or cache was read - changes depending on the anthropic api ttl
|
|
||||||
# assert (response.usage.cache_read_input_tokens > 0) or (
|
|
||||||
# response.usage.cache_creation_input_tokens > 0
|
|
||||||
# )
|
|
||||||
|
|
||||||
|
|
||||||
def test_supports_prompt_caching():
|
def test_supports_prompt_caching():
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue