From 78e1ed9575f1a2852facb66dd47ac529f0a06e87 Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Fri, 10 Nov 2023 16:28:34 -0800 Subject: [PATCH] (fix) proxy raise exception when config passed in --- litellm/proxy/proxy_cli.py | 2 +- litellm/proxy/proxy_server.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/litellm/proxy/proxy_cli.py b/litellm/proxy/proxy_cli.py index 036229156e..0644367510 100644 --- a/litellm/proxy/proxy_cli.py +++ b/litellm/proxy/proxy_cli.py @@ -67,7 +67,7 @@ def is_port_in_use(port): @click.option('--request_timeout', default=600, type=int, help='Set timeout in seconds for completion calls') @click.option('--drop_params', is_flag=True, help='Drop any unmapped params') @click.option('--add_function_to_prompt', is_flag=True, help='If function passed but unsupported, pass it as prompt') -@click.option('--config', '-c', help='Configure Litellm') +@click.option('--config', '-c', default=None, help='Configure Litellm') @click.option('--file', '-f', help='Path to config file') @click.option('--max_budget', default=None, type=float, help='Set max budget for API calls - works for hosted models like OpenAI, TogetherAI, Anthropic, etc.`') @click.option('--telemetry', default=True, type=bool, help='Helps us know if people are using this feature. Turn this off by doing `--telemetry False`') diff --git a/litellm/proxy/proxy_server.py b/litellm/proxy/proxy_server.py index fa9124f20f..3ca8105184 100644 --- a/litellm/proxy/proxy_server.py +++ b/litellm/proxy/proxy_server.py @@ -207,7 +207,7 @@ def load_router_config(router: Optional[litellm.Router], config_file_path: str): else: pass except Exception as e: - print_verbose(f"Exception while reading Config: {e}") + raise Exception(f"Exception while reading Config: {e}") pass