mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 10:44:24 +00:00
better cli
This commit is contained in:
parent
0c640ab5ef
commit
84133f8f45
1 changed files with 3 additions and 3 deletions
|
@ -64,9 +64,9 @@ def run_server(port, api_base, model, deploy, debug, temperature, max_tokens, te
|
|||
import uvicorn
|
||||
except:
|
||||
raise ImportError("Uvicorn needs to be imported. Run - `pip install uvicorn`")
|
||||
print(f"\033[32mLiteLLM:Deployed Proxy Locally\033[0m\n\n")
|
||||
print(f"\033[32mLiteLLM: Test your URL using the following: \"litellm --test http://0.0.0.0/{port}\"\033[0m\n\n")
|
||||
|
||||
print(f"\033[32mLiteLLM: Deployed Proxy Locally\033[0m\n")
|
||||
print(f"\033[32mLiteLLM: Test your URL using the following: \"litellm --test http://0.0.0.0:{port}\" [In a new terminal tab]\033[0m\n")
|
||||
print(f"\033[32mLiteLLM: Deploy your proxy using the following: \"litellm --model claude-instant-1 --deploy\" Get an https://api.litellm.ai/chat/completions endpoint \033[0m\n")
|
||||
|
||||
uvicorn.run(app, host='0.0.0.0', port=port)
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue