From e5269fdb7c634e15dab39cd6a27aba3ff10f2b33 Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Wed, 28 Feb 2024 17:13:04 -0800 Subject: [PATCH] (feat) support mistral function calling --- litellm/utils.py | 1 + 1 file changed, 1 insertion(+) diff --git a/litellm/utils.py b/litellm/utils.py index b025cd60a9..94d1040b99 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -4057,6 +4057,7 @@ def get_optional_params( and custom_llm_provider != "vertex_ai" and custom_llm_provider != "anyscale" and custom_llm_provider != "together_ai" + and custom_llm_provider != "mistral" ): if custom_llm_provider == "ollama" or custom_llm_provider == "ollama_chat": # ollama actually supports json output