mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 10:44:24 +00:00
add support for litellm proxy calls
This commit is contained in:
parent
0f88b82c4f
commit
9067ec3b43
7 changed files with 110 additions and 16 deletions
|
@ -1129,6 +1129,81 @@ def get_llm_provider(model: str, custom_llm_provider: Optional[str] = None):
|
|||
except Exception as e:
|
||||
raise e
|
||||
|
||||
|
||||
def get_api_key(llm_provider: str, dynamic_api_key: Optional[str]):
|
||||
api_key = (dynamic_api_key or litellm.api_key)
|
||||
# openai
|
||||
if llm_provider == "openai" or llm_provider == "text-completion-openai":
|
||||
api_key = (
|
||||
api_key or
|
||||
litellm.openai_key or
|
||||
get_secret("OPENAI_API_KEY")
|
||||
)
|
||||
# anthropic
|
||||
elif llm_provider == "anthropic":
|
||||
api_key = (
|
||||
api_key or
|
||||
litellm.anthropic_key or
|
||||
get_secret("ANTHROPIC_API_KEY")
|
||||
)
|
||||
# ai21
|
||||
elif llm_provider == "ai21":
|
||||
api_key = (
|
||||
api_key or
|
||||
litellm.ai21_key or
|
||||
get_secret("AI211_API_KEY")
|
||||
)
|
||||
# aleph_alpha
|
||||
elif llm_provider == "aleph_alpha":
|
||||
api_key = (
|
||||
api_key or
|
||||
litellm.aleph_alpha_key or
|
||||
get_secret("ALEPH_ALPHA_API_KEY")
|
||||
)
|
||||
# baseten
|
||||
elif llm_provider == "baseten":
|
||||
api_key = (
|
||||
api_key or
|
||||
litellm.baseten_key or
|
||||
get_secret("BASETEN_API_KEY")
|
||||
)
|
||||
# cohere
|
||||
elif llm_provider == "cohere":
|
||||
api_key = (
|
||||
api_key or
|
||||
litellm.cohere_key or
|
||||
get_secret("COHERE_API_KEY")
|
||||
)
|
||||
# huggingface
|
||||
elif llm_provider == "huggingface":
|
||||
api_key = (
|
||||
api_key or
|
||||
litellm.huggingface_key or
|
||||
get_secret("HUGGINGFACE_API_KEY")
|
||||
)
|
||||
# nlp_cloud
|
||||
elif llm_provider == "nlp_cloud":
|
||||
api_key = (
|
||||
api_key or
|
||||
litellm.nlp_cloud_key or
|
||||
get_secret("NLP_CLOUD_API_KEY")
|
||||
)
|
||||
# replicate
|
||||
elif llm_provider == "replicate":
|
||||
api_key = (
|
||||
api_key or
|
||||
litellm.replicate_key or
|
||||
get_secret("REPLICATE_API_KEY")
|
||||
)
|
||||
# together_ai
|
||||
elif llm_provider == "together_ai":
|
||||
api_key = (
|
||||
api_key or
|
||||
litellm.togetherai_api_key or
|
||||
get_secret("TOGETHERAI_API_KEY") or
|
||||
get_secret("TOGETHER_AI_TOKEN")
|
||||
)
|
||||
return api_key
|
||||
def get_max_tokens(model: str):
|
||||
try:
|
||||
return litellm.model_cost[model]
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue