forked from phoenix/litellm-mirror
fix: support dynamic timeouts for openai and azure
This commit is contained in:
parent
77be3e3114
commit
c33c1d85bb
3 changed files with 36 additions and 24 deletions
|
@ -10,7 +10,7 @@ sys.path.insert(
|
|||
import time
|
||||
import litellm
|
||||
import openai
|
||||
import pytest
|
||||
import pytest, uuid
|
||||
|
||||
|
||||
def test_timeout():
|
||||
|
@ -60,7 +60,7 @@ def test_hanging_request_azure():
|
|||
encoded = litellm.utils.encode(model="gpt-3.5-turbo", text="blue")[0]
|
||||
response = router.completion(
|
||||
model="azure-gpt",
|
||||
messages=[{"role": "user", "content": "what color is red"}],
|
||||
messages=[{"role": "user", "content": f"what color is red {uuid.uuid4()}"}],
|
||||
logit_bias={encoded: 100},
|
||||
timeout=0.01,
|
||||
)
|
||||
|
@ -126,7 +126,7 @@ def test_hanging_request_openai():
|
|||
)
|
||||
|
||||
|
||||
test_hanging_request_openai()
|
||||
# test_hanging_request_openai()
|
||||
|
||||
# test_timeout()
|
||||
|
||||
|
@ -155,4 +155,4 @@ def test_timeout_streaming():
|
|||
)
|
||||
|
||||
|
||||
test_timeout_streaming()
|
||||
# test_timeout_streaming()
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue