mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 10:44:24 +00:00
fix linting error
This commit is contained in:
parent
a48cb49820
commit
4d253f45d8
1 changed files with 2 additions and 2 deletions
|
@ -931,7 +931,7 @@ def get_optional_params( # use the openai defaults
|
|||
return optional_params
|
||||
return optional_params
|
||||
|
||||
def get_llm_provider(model: str, custom_llm_provider: str = None):
|
||||
def get_llm_provider(model: str, custom_llm_provider: str = ""):
|
||||
try:
|
||||
# check if llm provider provided
|
||||
if custom_llm_provider:
|
||||
|
@ -975,7 +975,7 @@ def get_llm_provider(model: str, custom_llm_provider: str = None):
|
|||
elif model in litellm.baseten_models:
|
||||
custom_llm_provider = "baseten"
|
||||
|
||||
if custom_llm_provider is None:
|
||||
if custom_llm_provider is None or custom_llm_provider=="":
|
||||
raise ValueError(f"LLM Provider NOT provided. Pass in the LLM provider you are trying to call. E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/{model}',..)` Learn more: https://docs.litellm.ai/docs/providers")
|
||||
return model, custom_llm_provider
|
||||
except Exception as e:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue