From 5e521bd36ec4ef56dfaea922fd654e863f36a6d9 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Mon, 1 Jul 2024 08:23:46 -0700 Subject: [PATCH] fix(utils.py): fix openrouter params Fixes https://github.com/BerriAI/litellm/issues/4488 --- litellm/utils.py | 21 ++++++++------------- 1 file changed, 8 insertions(+), 13 deletions(-) diff --git a/litellm/utils.py b/litellm/utils.py index 7302bea75..66df7c503 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -3842,23 +3842,18 @@ def get_supported_openai_params( return litellm.AzureOpenAIConfig().get_supported_openai_params() elif custom_llm_provider == "openrouter": return [ - "functions", - "function_call", "temperature", "top_p", - "n", - "stream", - "stop", - "max_tokens", - "presence_penalty", "frequency_penalty", - "logit_bias", - "user", - "response_format", + "presence_penalty", + "repetition_penalty", "seed", - "tools", - "tool_choice", - "max_retries", + "max_tokens", + "logit_bias", + "logprobs", + "top_logprobs", + "response_format", + "stop", ] elif custom_llm_provider == "mistral" or custom_llm_provider == "codestral": # mistal and codestral api have the exact same params