mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
refactor - use consistent file naming convention AI21/
-> ai21
(#7090)
* fix refactor - use consistent file naming convention * ci/cd run again * fix naming structure
This commit is contained in:
parent
2a35de0868
commit
249506065e
6 changed files with 6 additions and 6 deletions
|
@ -1065,8 +1065,8 @@ from .llms.predibase import PredibaseConfig
|
||||||
from .llms.replicate import ReplicateConfig
|
from .llms.replicate import ReplicateConfig
|
||||||
from .llms.cohere.completion import CohereConfig
|
from .llms.cohere.completion import CohereConfig
|
||||||
from .llms.clarifai import ClarifaiConfig
|
from .llms.clarifai import ClarifaiConfig
|
||||||
from .llms.AI21.completion import AI21Config
|
from .llms.ai21.completion import AI21Config
|
||||||
from .llms.AI21.chat import AI21ChatConfig
|
from .llms.ai21.chat import AI21ChatConfig
|
||||||
from .llms.together_ai.chat import TogetherAIConfig
|
from .llms.together_ai.chat import TogetherAIConfig
|
||||||
from .llms.cloudflare import CloudflareConfig
|
from .llms.cloudflare import CloudflareConfig
|
||||||
from .llms.palm import PalmConfig
|
from .llms.palm import PalmConfig
|
||||||
|
@ -1158,7 +1158,7 @@ nvidiaNimEmbeddingConfig = NvidiaNimEmbeddingConfig()
|
||||||
|
|
||||||
from .llms.cerebras.chat import CerebrasConfig
|
from .llms.cerebras.chat import CerebrasConfig
|
||||||
from .llms.sambanova.chat import SambanovaConfig
|
from .llms.sambanova.chat import SambanovaConfig
|
||||||
from .llms.AI21.chat import AI21ChatConfig
|
from .llms.ai21.chat import AI21ChatConfig
|
||||||
from .llms.fireworks_ai.chat.fireworks_ai_transformation import FireworksAIConfig
|
from .llms.fireworks_ai.chat.fireworks_ai_transformation import FireworksAIConfig
|
||||||
from .llms.fireworks_ai.embed.fireworks_ai_transformation import (
|
from .llms.fireworks_ai.embed.fireworks_ai_transformation import (
|
||||||
FireworksAIEmbeddingConfig,
|
FireworksAIEmbeddingConfig,
|
||||||
|
|
|
@ -99,7 +99,7 @@ from .llms import (
|
||||||
replicate,
|
replicate,
|
||||||
vllm,
|
vllm,
|
||||||
)
|
)
|
||||||
from .llms.AI21 import completion as ai21
|
from .llms.ai21 import completion as ai21
|
||||||
from .llms.anthropic.chat import AnthropicChatCompletion
|
from .llms.anthropic.chat import AnthropicChatCompletion
|
||||||
from .llms.anthropic.completion import AnthropicTextCompletion
|
from .llms.anthropic.completion import AnthropicTextCompletion
|
||||||
from .llms.azure_ai.chat import AzureAIChatCompletion
|
from .llms.azure_ai.chat import AzureAIChatCompletion
|
||||||
|
|
|
@ -220,7 +220,7 @@ def test_all_model_configs():
|
||||||
optional_params={},
|
optional_params={},
|
||||||
) == {"max_tokens": 10}
|
) == {"max_tokens": 10}
|
||||||
|
|
||||||
from litellm.llms.AI21.chat import AI21ChatConfig
|
from litellm.llms.ai21.chat import AI21ChatConfig
|
||||||
|
|
||||||
assert "max_completion_tokens" in AI21ChatConfig().get_supported_openai_params(
|
assert "max_completion_tokens" in AI21ChatConfig().get_supported_openai_params(
|
||||||
"jamba-1.5-mini@001"
|
"jamba-1.5-mini@001"
|
||||||
|
|
|
@ -24,7 +24,7 @@ from litellm import RateLimitError, Timeout, completion, completion_cost, embedd
|
||||||
from litellm.llms.custom_httpx.http_handler import AsyncHTTPHandler, HTTPHandler
|
from litellm.llms.custom_httpx.http_handler import AsyncHTTPHandler, HTTPHandler
|
||||||
from litellm.llms.prompt_templates.factory import anthropic_messages_pt
|
from litellm.llms.prompt_templates.factory import anthropic_messages_pt
|
||||||
|
|
||||||
# litellm.num_retries = 3
|
# litellm.num_retries=3
|
||||||
|
|
||||||
litellm.cache = None
|
litellm.cache = None
|
||||||
litellm.success_callback = []
|
litellm.success_callback = []
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue