diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index ea33bdb02b..7e5be4dc6b 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -2409,25 +2409,26 @@ "max_tokens": 4096, "max_input_tokens": 131072, "max_output_tokens": 4096, - "input_cost_per_token": 0, - "output_cost_per_token": 0, + "input_cost_per_token": 0.000000075, + "output_cost_per_token": 0.0000003, "litellm_provider": "azure_ai", "mode": "chat", "supports_function_calling": true, - "source": "https://learn.microsoft.com/en-us/azure/ai-foundry/concepts/models-featured#microsoft" + "source": "https://techcommunity.microsoft.com/blog/Azure-AI-Services-blog/announcing-new-phi-pricing-empowering-your-business-with-small-language-models/4395112" }, "azure_ai/Phi-4-multimodal-instruct": { "max_tokens": 4096, "max_input_tokens": 131072, "max_output_tokens": 4096, - "input_cost_per_token": 0, - "output_cost_per_token": 0, + "input_cost_per_token": 0.00000008, + "input_cost_per_audio_token": 0.000004, + "output_cost_per_token": 0.00032, "litellm_provider": "azure_ai", "mode": "chat", "supports_audio_input": true, "supports_function_calling": true, "supports_vision": true, - "source": "https://learn.microsoft.com/en-us/azure/ai-foundry/concepts/models-featured#microsoft" + "source": "https://techcommunity.microsoft.com/blog/Azure-AI-Services-blog/announcing-new-phi-pricing-empowering-your-business-with-small-language-models/4395112" }, "azure_ai/Phi-4": { "max_tokens": 16384, diff --git a/litellm/proxy/common_utils/encrypt_decrypt_utils.py b/litellm/proxy/common_utils/encrypt_decrypt_utils.py index 3452734867..348c81101f 100644 --- a/litellm/proxy/common_utils/encrypt_decrypt_utils.py +++ b/litellm/proxy/common_utils/encrypt_decrypt_utils.py @@ -51,9 +51,6 @@ def decrypt_value_helper(value: str): # if it's not str - do not decrypt it, return the value return value except Exception as e: - import traceback - - traceback.print_stack() verbose_proxy_logger.error( f"Error decrypting value, Did your master_key/salt key change recently? \nError: {str(e)}\nSet permanent salt key - https://docs.litellm.ai/docs/proxy/prod#5-set-litellm-salt-key" ) diff --git a/litellm/proxy/management_endpoints/internal_user_endpoints.py b/litellm/proxy/management_endpoints/internal_user_endpoints.py index fe5c1bacab..efc1bafa15 100644 --- a/litellm/proxy/management_endpoints/internal_user_endpoints.py +++ b/litellm/proxy/management_endpoints/internal_user_endpoints.py @@ -1434,7 +1434,7 @@ async def get_user_daily_activity( default=1, description="Page number for pagination", ge=1 ), page_size: int = fastapi.Query( - default=50, description="Items per page", ge=1, le=100 + default=50, description="Items per page", ge=1, le=1000 ), user_api_key_dict: UserAPIKeyAuth = Depends(user_api_key_auth), ) -> SpendAnalyticsPaginatedResponse: