diff --git a/docs/my-website/docs/proxy/billing.md b/docs/my-website/docs/proxy/billing.md index 78ac38936..902801cd0 100644 --- a/docs/my-website/docs/proxy/billing.md +++ b/docs/my-website/docs/proxy/billing.md @@ -23,7 +23,7 @@ Bill internal teams for their usage Set 'lago' as a callback on your proxy config.yaml ```yaml -model_name: +model_list: - model_name: fake-openai-endpoint litellm_params: model: openai/fake diff --git a/litellm/integrations/lago.py b/litellm/integrations/lago.py index 2bef69faa..b3db53301 100644 --- a/litellm/integrations/lago.py +++ b/litellm/integrations/lago.py @@ -114,7 +114,7 @@ class LagoLogger(CustomLogger): returned_val = { "event": { "transaction_id": str(uuid.uuid4()), - "external_customer_id": external_customer_id, + "external_subscription_id": external_customer_id, "code": os.getenv("LAGO_API_EVENT_CODE"), "properties": {"model": model, "response_cost": cost, **usage}, } @@ -189,6 +189,8 @@ class LagoLogger(CustomLogger): ) response.raise_for_status() + + verbose_logger.debug(f"Logged Lago Object: {response.text}") except Exception as e: if response is not None and hasattr(response, "text"): litellm.print_verbose(f"\nError Message: {response.text}")