diff --git a/litellm/main.py b/litellm/main.py index 672029f69..134617ba0 100644 --- a/litellm/main.py +++ b/litellm/main.py @@ -4797,12 +4797,12 @@ async def ahealth_check( raise Exception("model not set") if model in litellm.model_cost and mode is None: - mode = litellm.model_cost[model]["mode"] + mode = litellm.model_cost[model].get("mode") model, custom_llm_provider, _, _ = get_llm_provider(model=model) if model in litellm.model_cost and mode is None: - mode = litellm.model_cost[model]["mode"] + mode = litellm.model_cost[model].get("mode") mode = mode or "chat" # default to chat completion calls diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index 0f20f6689..a6d66750c 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -4373,7 +4373,7 @@ "litellm_provider": "perplexity", "mode": "chat" }, - "fireworks_ai/firefunction-v2": { + "fireworks_ai/accounts/fireworks/models/firefunction-v2": { "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, @@ -4384,7 +4384,7 @@ "supports_function_calling": true, "source": "https://fireworks.ai/pricing" }, - "fireworks_ai/mixtral-8x22b-instruct-hf": { + "fireworks_ai/accounts/fireworks/models/mixtral-8x22b-instruct-hf": { "max_tokens": 65536, "max_input_tokens": 65536, "max_output_tokens": 65536, @@ -4395,7 +4395,7 @@ "supports_function_calling": true, "source": "https://fireworks.ai/pricing" }, - "fireworks_ai/qwen2-72b-instruct": { + "fireworks_ai/accounts/fireworks/models/qwen2-72b-instruct": { "max_tokens": 32768, "max_input_tokens": 32768, "max_output_tokens": 32768, @@ -4406,7 +4406,7 @@ "supports_function_calling": true, "source": "https://fireworks.ai/pricing" }, - "fireworks_ai/yi-large": { + "fireworks_ai/accounts/fireworks/models/yi-large": { "max_tokens": 32768, "max_input_tokens": 32768, "max_output_tokens": 32768, @@ -4417,7 +4417,7 @@ "supports_function_calling": true, "source": "https://fireworks.ai/pricing" }, - "fireworks_ai/deepseek-coder-v2-instruct": { + "fireworks_ai/accounts/fireworks/models/deepseek-coder-v2-instruct": { "max_tokens": 65536, "max_input_tokens": 65536, "max_output_tokens": 8192,