mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
add deploy flag to cli
This commit is contained in:
parent
71dbd9b61f
commit
3f7740ddbc
1 changed files with 6 additions and 2 deletions
|
@ -9,8 +9,7 @@ load_dotenv()
|
||||||
@click.option('--port', default=8000, help='Port to bind the server to.')
|
@click.option('--port', default=8000, help='Port to bind the server to.')
|
||||||
@click.option('--api_base', default=None, help='API base URL.')
|
@click.option('--api_base', default=None, help='API base URL.')
|
||||||
@click.option('--model', default=None, help='The model name to pass to litellm expects')
|
@click.option('--model', default=None, help='The model name to pass to litellm expects')
|
||||||
@click.option('--deploy', is_flag=True, help='Get a deployed proxy endpoint - api.litellm.ai')
|
@click.option('--deploy', is_flag=True, type=bool, help='Get a deployed proxy endpoint - api.litellm.ai')
|
||||||
@click.option('--deploy', required=True, help='Get a deployed proxy endpoint - api.litellm.ai')
|
|
||||||
@click.option('--debug', is_flag=True, help='To debug the input')
|
@click.option('--debug', is_flag=True, help='To debug the input')
|
||||||
@click.option('--temperature', default=None, type=float, help='Set temperature for the model')
|
@click.option('--temperature', default=None, type=float, help='Set temperature for the model')
|
||||||
@click.option('--max_tokens', default=None, help='Set max tokens for the model')
|
@click.option('--max_tokens', default=None, help='Set max tokens for the model')
|
||||||
|
@ -33,6 +32,11 @@ def run_server(port, api_base, model, deploy, debug, temperature, max_tokens, te
|
||||||
# from .proxy_server import app, initialize
|
# from .proxy_server import app, initialize
|
||||||
from proxy_server import app, initialize
|
from proxy_server import app, initialize
|
||||||
initialize(model, api_base, debug, temperature, max_tokens, telemetry)
|
initialize(model, api_base, debug, temperature, max_tokens, telemetry)
|
||||||
|
|
||||||
|
def run_server(port, api_base, model, debug, temperature, max_tokens, telemetry, deploy):
|
||||||
|
from .proxy_server import app, initialize
|
||||||
|
initialize(model, api_base, debug, temperature, max_tokens, telemetry, deploy)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import uvicorn
|
import uvicorn
|
||||||
except:
|
except:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue