mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 02:34:29 +00:00
_get_exception_class_name
This commit is contained in:
parent
c402db9057
commit
f402e9bbd1
2 changed files with 7 additions and 5 deletions
|
@ -1148,7 +1148,9 @@ class PrometheusLogger(CustomLogger):
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _get_exception_class_name(exception: Exception) -> str:
|
def _get_exception_class_name(exception: Exception) -> str:
|
||||||
exception_class_name = getattr(exception, "llm_provider") or ""
|
exception_class_name = ""
|
||||||
|
if hasattr(exception, "llm_provider"):
|
||||||
|
exception_class_name = getattr(exception, "llm_provider") or ""
|
||||||
|
|
||||||
# pretty print the provider name on prometheus
|
# pretty print the provider name on prometheus
|
||||||
# eg. `openai` -> `Openai.`
|
# eg. `openai` -> `Openai.`
|
||||||
|
|
|
@ -107,7 +107,7 @@ async def test_proxy_failure_metrics():
|
||||||
print("/metrics", metrics)
|
print("/metrics", metrics)
|
||||||
|
|
||||||
# Check if the failure metric is present and correct
|
# Check if the failure metric is present and correct
|
||||||
expected_metric = 'litellm_proxy_failed_requests_metric_total{api_key_alias="None",end_user="None",exception_class="RateLimitError",exception_status="429",hashed_api_key="88dc28d0f030c55ed4ab77ed8faf098196cb1c05df778539800c9f1243fe6b4b",requested_model="fake-azure-endpoint",team="None",team_alias="None",user="default_user_id"} 1.0'
|
expected_metric = 'litellm_proxy_failed_requests_metric_total{api_key_alias="None",end_user="None",exception_class="Openai.RateLimitError",exception_status="429",hashed_api_key="88dc28d0f030c55ed4ab77ed8faf098196cb1c05df778539800c9f1243fe6b4b",requested_model="fake-azure-endpoint",team="None",team_alias="None",user="default_user_id"} 1.0'
|
||||||
|
|
||||||
assert (
|
assert (
|
||||||
expected_metric in metrics
|
expected_metric in metrics
|
||||||
|
@ -121,7 +121,7 @@ async def test_proxy_failure_metrics():
|
||||||
)
|
)
|
||||||
|
|
||||||
assert (
|
assert (
|
||||||
'litellm_deployment_failure_responses_total{api_base="https://exampleopenaiendpoint-production.up.railway.app",api_key_alias="None",api_provider="openai",exception_class="RateLimitError",exception_status="429",hashed_api_key="88dc28d0f030c55ed4ab77ed8faf098196cb1c05df778539800c9f1243fe6b4b",litellm_model_name="429",model_id="7499d31f98cd518cf54486d5a00deda6894239ce16d13543398dc8abf870b15f",requested_model="fake-azure-endpoint",team="None",team_alias="None"}'
|
'litellm_deployment_failure_responses_total{api_base="https://exampleopenaiendpoint-production.up.railway.app",api_key_alias="None",api_provider="openai",exception_class="Openai.RateLimitError",exception_status="429",hashed_api_key="88dc28d0f030c55ed4ab77ed8faf098196cb1c05df778539800c9f1243fe6b4b",litellm_model_name="429",model_id="7499d31f98cd518cf54486d5a00deda6894239ce16d13543398dc8abf870b15f",requested_model="fake-azure-endpoint",team="None",team_alias="None"}'
|
||||||
in metrics
|
in metrics
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -229,13 +229,13 @@ async def test_proxy_fallback_metrics():
|
||||||
|
|
||||||
# Check if successful fallback metric is incremented
|
# Check if successful fallback metric is incremented
|
||||||
assert (
|
assert (
|
||||||
'litellm_deployment_successful_fallbacks_total{api_key_alias="None",exception_class="RateLimitError",exception_status="429",fallback_model="fake-openai-endpoint",hashed_api_key="88dc28d0f030c55ed4ab77ed8faf098196cb1c05df778539800c9f1243fe6b4b",requested_model="fake-azure-endpoint",team="None",team_alias="None"} 1.0'
|
'litellm_deployment_successful_fallbacks_total{api_key_alias="None",exception_class="Openai.RateLimitError",exception_status="429",fallback_model="fake-openai-endpoint",hashed_api_key="88dc28d0f030c55ed4ab77ed8faf098196cb1c05df778539800c9f1243fe6b4b",requested_model="fake-azure-endpoint",team="None",team_alias="None"} 1.0'
|
||||||
in metrics
|
in metrics
|
||||||
)
|
)
|
||||||
|
|
||||||
# Check if failed fallback metric is incremented
|
# Check if failed fallback metric is incremented
|
||||||
assert (
|
assert (
|
||||||
'litellm_deployment_failed_fallbacks_total{api_key_alias="None",exception_class="RateLimitError",exception_status="429",fallback_model="unknown-model",hashed_api_key="88dc28d0f030c55ed4ab77ed8faf098196cb1c05df778539800c9f1243fe6b4b",requested_model="fake-azure-endpoint",team="None",team_alias="None"} 1.0'
|
'litellm_deployment_failed_fallbacks_total{api_key_alias="None",exception_class="Openai.RateLimitError",exception_status="429",fallback_model="unknown-model",hashed_api_key="88dc28d0f030c55ed4ab77ed8faf098196cb1c05df778539800c9f1243fe6b4b",requested_model="fake-azure-endpoint",team="None",team_alias="None"} 1.0'
|
||||||
in metrics
|
in metrics
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue