diff --git a/litellm/proxy/proxy_cli.py b/litellm/proxy/proxy_cli.py index 865d807fe..3643d39f3 100644 --- a/litellm/proxy/proxy_cli.py +++ b/litellm/proxy/proxy_cli.py @@ -160,9 +160,9 @@ def run_server(host, port, api_base, api_version, model, alias, add_key, headers # response = openai.Completion.create(model="gpt-3.5-turbo", prompt='this is a test request, write a short poem', stream=True) - # for chunk in response: - # click.echo(f'LiteLLM: streaming response from proxy {chunk}') - # return + for chunk in response: + click.echo(f'LiteLLM: streaming response from proxy {chunk}') + return else: if headers: headers = json.loads(headers)