forked from phoenix/litellm-mirror
fix entrypoint
This commit is contained in:
parent
71e5b2df79
commit
95455c8849
2 changed files with 5 additions and 2 deletions
|
@ -82,7 +82,6 @@ from .llms import (
|
|||
bedrock,
|
||||
clarifai,
|
||||
cloudflare,
|
||||
cohere,
|
||||
gemini,
|
||||
huggingface_restapi,
|
||||
maritalk,
|
||||
|
@ -107,7 +106,7 @@ from .llms.azure import AzureChatCompletion, _check_dynamic_azure_params
|
|||
from .llms.azure_text import AzureTextCompletion
|
||||
from .llms.bedrock_httpx import BedrockConverseLLM, BedrockLLM
|
||||
from .llms.cohere import chat as cohere_chat
|
||||
from .llms.cohere import completion as cohere_completion
|
||||
from .llms.cohere import completion as cohere_completion # type: ignore
|
||||
from .llms.cohere import embed as cohere_embed
|
||||
from .llms.custom_llm import CustomLLM, custom_chat_llm_router
|
||||
from .llms.databricks import DatabricksChatCompletion
|
||||
|
|
|
@ -51,6 +51,10 @@ while retry_count < max_retries and exit_code != 0:
|
|||
retry_count += 1
|
||||
print(f"Attempt {retry_count}...") # noqa
|
||||
|
||||
# run prisma generate
|
||||
result = subprocess.run(["prisma", "generate"], capture_output=True)
|
||||
exit_code = result.returncode
|
||||
|
||||
# Run the Prisma db push command
|
||||
result = subprocess.run(
|
||||
["prisma", "db", "push", "--accept-data-loss"], capture_output=True
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue