diff --git a/litellm/__init__.py b/litellm/__init__.py index 08ee84aaa..cee80a32d 100644 --- a/litellm/__init__.py +++ b/litellm/__init__.py @@ -817,6 +817,7 @@ from .llms.openai import ( AzureAIStudioConfig, ) from .llms.nvidia_nim import NvidiaNimConfig +from .llms.fireworks_ai import FireworksAIConfig from .llms.text_completion_codestral import MistralTextCompletionConfig from .llms.azure import ( AzureOpenAIConfig, diff --git a/litellm/utils.py b/litellm/utils.py index beae7ba4a..a33a160e4 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -3079,6 +3079,16 @@ def get_optional_params( optional_params = litellm.NvidiaNimConfig().map_openai_params( non_default_params=non_default_params, optional_params=optional_params ) + elif custom_llm_provider == "fireworks_ai": + supported_params = get_supported_openai_params( + model=model, custom_llm_provider=custom_llm_provider + ) + _check_valid_arg(supported_params=supported_params) + optional_params = litellm.FireworksAIConfig().map_openai_params( + non_default_params=non_default_params, + optional_params=optional_params, + model=model, + ) elif custom_llm_provider == "groq": supported_params = get_supported_openai_params( model=model, custom_llm_provider=custom_llm_provider @@ -3645,6 +3655,8 @@ def get_supported_openai_params( return litellm.OllamaChatConfig().get_supported_openai_params() elif custom_llm_provider == "anthropic": return litellm.AnthropicConfig().get_supported_openai_params() + elif custom_llm_provider == "fireworks_ai": + return litellm.FireworksAIConfig().get_supported_openai_params() elif custom_llm_provider == "nvidia_nim": return litellm.NvidiaNimConfig().get_supported_openai_params() elif custom_llm_provider == "groq":