forked from phoenix/litellm-mirror
fix get llm provider
This commit is contained in:
parent
ff36b9c6da
commit
05f2f3dde7
3 changed files with 4 additions and 1 deletions
Binary file not shown.
|
@ -947,6 +947,9 @@ def get_llm_provider(model: str, custom_llm_provider: Optional[str] = None):
|
|||
## openai - chatcompletion + text completion
|
||||
if model in litellm.open_ai_chat_completion_models or model in litellm.open_ai_text_completion_models:
|
||||
custom_llm_provider = "openai"
|
||||
## anthropic
|
||||
elif model in litellm.anthropic_models:
|
||||
custom_llm_provider = "anthropic"
|
||||
## cohere
|
||||
elif model in litellm.cohere_models:
|
||||
custom_llm_provider = "cohere"
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
[tool.poetry]
|
||||
name = "litellm"
|
||||
version = "0.1.603"
|
||||
version = "0.1.604"
|
||||
description = "Library to easily interface with LLM API providers"
|
||||
authors = ["BerriAI"]
|
||||
license = "MIT License"
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue