mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
feat(proxy_cli.py): when user calls ollama model, run ollama serve
This commit is contained in:
parent
56c87febae
commit
52b0bcb5ec
4 changed files with 8 additions and 0 deletions
Binary file not shown.
Binary file not shown.
Binary file not shown.
|
@ -13,6 +13,12 @@ config_filename = ".env.litellm"
|
||||||
config_dir = appdirs.user_config_dir("litellm")
|
config_dir = appdirs.user_config_dir("litellm")
|
||||||
user_config_path = os.path.join(config_dir, config_filename)
|
user_config_path = os.path.join(config_dir, config_filename)
|
||||||
|
|
||||||
|
def run_ollama_serve():
|
||||||
|
command = ['ollama', 'serve']
|
||||||
|
|
||||||
|
with open(os.devnull, 'w') as devnull:
|
||||||
|
process = subprocess.Popen(command, stdout=devnull, stderr=devnull)
|
||||||
|
|
||||||
def load_config():
|
def load_config():
|
||||||
try:
|
try:
|
||||||
if not os.path.exists(user_config_path):
|
if not os.path.exists(user_config_path):
|
||||||
|
@ -82,6 +88,8 @@ def run_server(port, api_base, model, deploy, debug, temperature, max_tokens, te
|
||||||
|
|
||||||
print(f"\033[32mLiteLLM: Test your URL using the following: \"litellm --test {url}\"\033[0m")
|
print(f"\033[32mLiteLLM: Test your URL using the following: \"litellm --test {url}\"\033[0m")
|
||||||
return
|
return
|
||||||
|
if "ollama" in model:
|
||||||
|
run_ollama_serve()
|
||||||
if test != False:
|
if test != False:
|
||||||
click.echo('LiteLLM: Making a test ChatCompletions request to your proxy')
|
click.echo('LiteLLM: Making a test ChatCompletions request to your proxy')
|
||||||
import openai
|
import openai
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue