3.8 KiB
Exception Mapping
LiteLLM maps the 4 most common exceptions across all providers.
- Rate Limit Errors
- Context Window Errors
- Invalid Request Errors
- InvalidAuth Errors (incorrect key, etc.)
Base case - we return the original exception.
All our exceptions inherit from OpenAI's exception types, so any error-handling you have for that, should work out of the box with LiteLLM.
For all 4 cases, the exception returned inherits from the original OpenAI Exception but contains 3 additional attributes:
- status_code - the http status code of the exception
- message - the error message
- llm_provider - the provider raising the exception
usage
from litellm import completion
os.environ["ANTHROPIC_API_KEY"] = "bad-key"
try:
# some code
completion(model="claude-instant-1", messages=[{"role": "user", "content": "Hey, how's it going?"}])
except Exception as e:
print(e.llm_provider)
details
To see how it's implemented - check out the code
Create an issue or make a PR if you want to improve the exception mapping.
Note For OpenAI and Azure we return the original exception (since they're of the OpenAI Error type). But we add the 'llm_provider' attribute to them. See code
custom mapping list
Base case - we return the original exception.
LLM Provider | Initial Status Code / Initial Error Message | Returned Exception | Returned Status Code |
---|---|---|---|
Anthropic | prompt is too long | ContextWindowExceededError | 400 |
Anthropic | 401 | AuthenticationError | 401 |
Anthropic | Could not resolve authentication method. Expected either api_key or auth_token to be set. | AuthenticationError | 401 |
Anthropic | 400 | InvalidRequestError | 400 |
Anthropic | 429 | RateLimitError | 429 |
OpenAI | This model's maximum context length is | ContextWindowExceededError | 400 |
Replicate | input is too long | ContextWindowExceededError | 400 |
Replicate | Incorrect authentication token | AuthenticationError | 401 |
Replicate | ModelError | InvalidRequestError | 400 |
Replicate | Request was throttled | RateLimitError | 429 |
Replicate | ReplicateError | ServiceUnavailableError | 500 |
Cohere | invalid api token | AuthenticationError | 401 |
Cohere | too many tokens | ContextWindowExceededError | 400 |
Cohere | CohereConnectionError | RateLimitError | 429 |
Huggingface | length limit exceeded | ContextWindowExceededError | 400 |
Huggingface | 400 | InvalidRequestError | 400 |
Huggingface | 401 | AuthenticationError | 401 |
Huggingface | 429 | RateLimitError | 429 |
Openrouter | 413 | ContextWindowExceededError | 400 |
Openrouter | 401 | AuthenticationError | 401 |
Openrouter | 429 | RateLimitError | 429 |
AI21 | Prompt has too many tokens | ContextWindowExceededError | 400 |
AI21 | 422 | InvalidRequestError | 400 |
AI21 | 401 | AuthenticationError | 401 |
AI21 | 429 | RateLimitError | 429 |
TogetherAI | inputstokens + max_new_tokens` must be <= |
ContextWindowExceededError | 400 |
TogetherAI | INVALID_ARGUMENT | InvalidRequestError | 400 |
TogetherAI | "error_type": "validation" | InvalidRequestError | 400 |
TogetherAI | invalid private key | AuthenticationError | 401 |
TogetherAI | 429 | RateLimitError | 429 |
The ContextWindowExceededError
is a sub-class of InvalidRequestError
. It was introduced to provide more granularity for exception-handling scenarios. Please refer to this issue to learn more.