mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 02:34:29 +00:00
108 lines
No EOL
3.8 KiB
Python
108 lines
No EOL
3.8 KiB
Python
## LiteLLM versions of the OpenAI Exception Types
|
|
from openai.error import (
|
|
AuthenticationError,
|
|
InvalidRequestError,
|
|
RateLimitError,
|
|
ServiceUnavailableError,
|
|
OpenAIError,
|
|
APIError,
|
|
Timeout,
|
|
APIConnectionError,
|
|
)
|
|
|
|
|
|
class AuthenticationError(AuthenticationError): # type: ignore
|
|
def __init__(self, message, llm_provider, model):
|
|
self.status_code = 401
|
|
self.message = message
|
|
self.llm_provider = llm_provider
|
|
self.model = model
|
|
super().__init__(
|
|
self.message
|
|
) # Call the base class constructor with the parameters it needs
|
|
|
|
|
|
class InvalidRequestError(InvalidRequestError): # type: ignore
|
|
def __init__(self, message, model, llm_provider):
|
|
self.status_code = 400
|
|
self.message = message
|
|
self.model = model
|
|
self.llm_provider = llm_provider
|
|
super().__init__(
|
|
self.message, f"{self.model}"
|
|
) # Call the base class constructor with the parameters it needs
|
|
|
|
class Timeout(Timeout): # type: ignore
|
|
def __init__(self, message, model, llm_provider):
|
|
self.status_code = 408
|
|
self.message = message
|
|
self.model = model
|
|
self.llm_provider = llm_provider
|
|
super().__init__(
|
|
self.message, f"{self.model}"
|
|
) # Call the base class constructor with the parameters it needs
|
|
|
|
# sub class of invalid request error - meant to give more granularity for error handling context window exceeded errors
|
|
class ContextWindowExceededError(InvalidRequestError): # type: ignore
|
|
def __init__(self, message, model, llm_provider):
|
|
self.status_code = 400
|
|
self.message = message
|
|
self.model = model
|
|
self.llm_provider = llm_provider
|
|
super().__init__(
|
|
self.message, self.model, self.llm_provider
|
|
) # Call the base class constructor with the parameters it needs
|
|
|
|
|
|
class RateLimitError(RateLimitError): # type: ignore
|
|
def __init__(self, message, llm_provider, model):
|
|
self.status_code = 429
|
|
self.message = message
|
|
self.llm_provider = llm_provider
|
|
self.modle = model
|
|
super().__init__(
|
|
self.message
|
|
) # Call the base class constructor with the parameters it needs
|
|
|
|
|
|
class ServiceUnavailableError(ServiceUnavailableError): # type: ignore
|
|
def __init__(self, message, llm_provider, model):
|
|
self.status_code = 500
|
|
self.message = message
|
|
self.llm_provider = llm_provider
|
|
self.model = model
|
|
super().__init__(
|
|
self.message
|
|
) # Call the base class constructor with the parameters it needs
|
|
|
|
|
|
class APIError(APIError): # raise this when the API returns an invalid response object - https://github.com/openai/openai-python/blob/1be14ee34a0f8e42d3f9aa5451aa4cb161f1781f/openai/api_requestor.py#L401
|
|
def __init__(self, status_code, message, llm_provider, model):
|
|
self.status_code = status_code
|
|
self.message = message
|
|
self.llm_provider = llm_provider
|
|
self.model = model
|
|
super().__init__(
|
|
self.message
|
|
)
|
|
|
|
class APIConnectionError(APIConnectionError): # raised if an invalid request (not get, delete, put, post) is made
|
|
def __init__(self, message, llm_provider, model):
|
|
self.message = message
|
|
self.llm_provider = llm_provider
|
|
self.model = model
|
|
super().__init__(
|
|
self.message
|
|
)
|
|
|
|
class OpenAIError(OpenAIError): # type: ignore
|
|
def __init__(self, original_exception):
|
|
self.status_code = original_exception.http_status
|
|
super().__init__(
|
|
http_body=original_exception.http_body,
|
|
http_status=original_exception.http_status,
|
|
json_body=original_exception.json_body,
|
|
headers=original_exception.headers,
|
|
code=original_exception.code,
|
|
)
|
|
self.llm_provider = "openai" |