From 5d39279ac068f4f5bc0b10c4ffcd44b28462555b Mon Sep 17 00:00:00 2001 From: Rodney Rodriguez Date: Sat, 11 Nov 2023 16:49:22 -0600 Subject: [PATCH] bugfixes for ollama --- litellm/main.py | 2 +- litellm/utils.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/litellm/main.py b/litellm/main.py index 95fa44a6c4..18c7c52d31 100644 --- a/litellm/main.py +++ b/litellm/main.py @@ -322,7 +322,7 @@ def completion( max_retries = kwargs.get("max_retries", None) context_window_fallback_dict = kwargs.get("context_window_fallback_dict", None) ### CUSTOM PROMPT TEMPLATE ### - initial_prompt_value = kwargs.get("intial_prompt_value", None) + initial_prompt_value = kwargs.get("initial_prompt_value", None) roles = kwargs.get("roles", None) final_prompt_value = kwargs.get("final_prompt_value", None) bos_token = kwargs.get("bos_token", None) diff --git a/litellm/utils.py b/litellm/utils.py index 3f1002155e..8de45eb122 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -3159,7 +3159,7 @@ def register_prompt_template(model: str, roles: dict, initial_prompt_value: str ) ``` """ - model, _ = get_llm_provider(model=model) + model = get_llm_provider(model=model)[0] litellm.custom_prompt_dict[model] = { "roles": roles, "initial_prompt_value": initial_prompt_value,