mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
fix(router.py): comment out azure/openai client init - not necessary
This commit is contained in:
parent
1146bc0b1a
commit
1e89fdec99
3 changed files with 19 additions and 10 deletions
|
@ -1994,8 +1994,8 @@
|
|||
"max_tokens": 8191,
|
||||
"max_input_tokens": 32000,
|
||||
"max_output_tokens": 8191,
|
||||
"input_cost_per_token": 0.000001,
|
||||
"output_cost_per_token": 0.000003,
|
||||
"input_cost_per_token": 0.0000001,
|
||||
"output_cost_per_token": 0.0000003,
|
||||
"litellm_provider": "mistral",
|
||||
"supports_function_calling": true,
|
||||
"mode": "chat",
|
||||
|
@ -2006,8 +2006,8 @@
|
|||
"max_tokens": 8191,
|
||||
"max_input_tokens": 32000,
|
||||
"max_output_tokens": 8191,
|
||||
"input_cost_per_token": 0.000001,
|
||||
"output_cost_per_token": 0.000003,
|
||||
"input_cost_per_token": 0.0000001,
|
||||
"output_cost_per_token": 0.0000003,
|
||||
"litellm_provider": "mistral",
|
||||
"supports_function_calling": true,
|
||||
"mode": "chat",
|
||||
|
|
|
@ -1,4 +1,13 @@
|
|||
model_list:
|
||||
- model_name: llama3.2-vision
|
||||
- model_name: gpt-4o
|
||||
litellm_params:
|
||||
model: ollama/llama3.2-vision
|
||||
model: azure/gpt-4o
|
||||
credential_name: default_azure_credential
|
||||
|
||||
credential_list:
|
||||
- credential_name: default_azure_credential
|
||||
credentials:
|
||||
api_key: os.environ/AZURE_API_KEY
|
||||
api_base: os.environ/AZURE_API_BASE
|
||||
credential_info:
|
||||
description: "Default Azure credential"
|
||||
|
|
|
@ -4373,10 +4373,10 @@ class Router:
|
|||
if custom_llm_provider not in litellm.provider_list:
|
||||
raise Exception(f"Unsupported provider - {custom_llm_provider}")
|
||||
|
||||
# init OpenAI, Azure clients
|
||||
InitalizeOpenAISDKClient.set_client(
|
||||
litellm_router_instance=self, model=deployment.to_json(exclude_none=True)
|
||||
)
|
||||
# # init OpenAI, Azure clients
|
||||
# InitalizeOpenAISDKClient.set_client(
|
||||
# litellm_router_instance=self, model=deployment.to_json(exclude_none=True)
|
||||
# )
|
||||
|
||||
self._initialize_deployment_for_pass_through(
|
||||
deployment=deployment,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue