diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index b2a08544f9..2e740a3ca3 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -1994,8 +1994,8 @@ "max_tokens": 8191, "max_input_tokens": 32000, "max_output_tokens": 8191, - "input_cost_per_token": 0.000001, - "output_cost_per_token": 0.000003, + "input_cost_per_token": 0.0000001, + "output_cost_per_token": 0.0000003, "litellm_provider": "mistral", "supports_function_calling": true, "mode": "chat", @@ -2006,8 +2006,8 @@ "max_tokens": 8191, "max_input_tokens": 32000, "max_output_tokens": 8191, - "input_cost_per_token": 0.000001, - "output_cost_per_token": 0.000003, + "input_cost_per_token": 0.0000001, + "output_cost_per_token": 0.0000003, "litellm_provider": "mistral", "supports_function_calling": true, "mode": "chat", diff --git a/litellm/proxy/_new_secret_config.yaml b/litellm/proxy/_new_secret_config.yaml index eac1e6a6da..f3d8d55990 100644 --- a/litellm/proxy/_new_secret_config.yaml +++ b/litellm/proxy/_new_secret_config.yaml @@ -1,4 +1,13 @@ model_list: - - model_name: llama3.2-vision + - model_name: gpt-4o litellm_params: - model: ollama/llama3.2-vision \ No newline at end of file + model: azure/gpt-4o + credential_name: default_azure_credential + +credential_list: + - credential_name: default_azure_credential + credentials: + api_key: os.environ/AZURE_API_KEY + api_base: os.environ/AZURE_API_BASE + credential_info: + description: "Default Azure credential" diff --git a/litellm/router.py b/litellm/router.py index aba9e16104..940b02c78c 100644 --- a/litellm/router.py +++ b/litellm/router.py @@ -4373,10 +4373,10 @@ class Router: if custom_llm_provider not in litellm.provider_list: raise Exception(f"Unsupported provider - {custom_llm_provider}") - # init OpenAI, Azure clients - InitalizeOpenAISDKClient.set_client( - litellm_router_instance=self, model=deployment.to_json(exclude_none=True) - ) + # # init OpenAI, Azure clients + # InitalizeOpenAISDKClient.set_client( + # litellm_router_instance=self, model=deployment.to_json(exclude_none=True) + # ) self._initialize_deployment_for_pass_through( deployment=deployment,