mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
fix(router.py): fix client init
This commit is contained in:
parent
278cc75603
commit
eb46ea8f8b
3 changed files with 33 additions and 7 deletions
|
@ -1,5 +1,5 @@
|
||||||
from typing import Optional, Union, Any
|
from typing import Optional, Union, Any
|
||||||
import types, time, json
|
import types, time, json, traceback
|
||||||
import httpx
|
import httpx
|
||||||
from .base import BaseLLM
|
from .base import BaseLLM
|
||||||
from litellm.utils import (
|
from litellm.utils import (
|
||||||
|
@ -349,7 +349,7 @@ class OpenAIChatCompletion(BaseLLM):
|
||||||
if hasattr(e, "status_code"):
|
if hasattr(e, "status_code"):
|
||||||
raise OpenAIError(status_code=e.status_code, message=str(e))
|
raise OpenAIError(status_code=e.status_code, message=str(e))
|
||||||
else:
|
else:
|
||||||
raise OpenAIError(status_code=500, message=str(e))
|
raise OpenAIError(status_code=500, message=traceback.format_exc())
|
||||||
|
|
||||||
async def acompletion(
|
async def acompletion(
|
||||||
self,
|
self,
|
||||||
|
|
|
@ -1521,13 +1521,13 @@ class Router:
|
||||||
):
|
):
|
||||||
stream_timeout_env_name = stream_timeout.replace("os.environ/", "")
|
stream_timeout_env_name = stream_timeout.replace("os.environ/", "")
|
||||||
stream_timeout = litellm.get_secret(stream_timeout_env_name)
|
stream_timeout = litellm.get_secret(stream_timeout_env_name)
|
||||||
litellm_params["stream_timeout"] = stream_timeout
|
|
||||||
|
|
||||||
max_retries = litellm_params.pop("max_retries", 2)
|
max_retries = litellm_params.pop("max_retries", 2)
|
||||||
if isinstance(max_retries, str) and max_retries.startswith("os.environ/"):
|
if isinstance(max_retries, str):
|
||||||
max_retries_env_name = max_retries.replace("os.environ/", "")
|
if max_retries.startswith("os.environ/"):
|
||||||
max_retries = litellm.get_secret(max_retries_env_name)
|
max_retries_env_name = max_retries.replace("os.environ/", "")
|
||||||
litellm_params["max_retries"] = max_retries
|
max_retries = litellm.get_secret(max_retries_env_name)
|
||||||
|
max_retries = int(max_retries)
|
||||||
|
|
||||||
if "azure" in model_name:
|
if "azure" in model_name:
|
||||||
if api_base is None:
|
if api_base is None:
|
||||||
|
|
|
@ -942,3 +942,29 @@ def test_reading_openai_keys_os_environ():
|
||||||
|
|
||||||
|
|
||||||
# test_reading_openai_keys_os_environ()
|
# test_reading_openai_keys_os_environ()
|
||||||
|
|
||||||
|
|
||||||
|
def test_router_timeout():
|
||||||
|
model_list = [
|
||||||
|
{
|
||||||
|
"model_name": "gpt-3.5-turbo",
|
||||||
|
"litellm_params": {
|
||||||
|
"model": "gpt-3.5-turbo",
|
||||||
|
"api_key": "os.environ/OPENAI_API_KEY",
|
||||||
|
"timeout": "os.environ/AZURE_TIMEOUT",
|
||||||
|
"stream_timeout": "os.environ/AZURE_STREAM_TIMEOUT",
|
||||||
|
"max_retries": "os.environ/AZURE_MAX_RETRIES",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
]
|
||||||
|
router = Router(model_list=model_list)
|
||||||
|
messages = [{"role": "user", "content": "Hey, how's it going?"}]
|
||||||
|
start_time = time.time()
|
||||||
|
try:
|
||||||
|
router.completion(
|
||||||
|
model="gpt-3.5-turbo", messages=messages, max_tokens=500, timeout=1
|
||||||
|
)
|
||||||
|
except litellm.exceptions.Timeout as e:
|
||||||
|
pass
|
||||||
|
end_time = time.time()
|
||||||
|
assert end_time - start_time < 1.1
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue