From aba39670a307ed8c3dc73b7549e8c04fe3826a50 Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Wed, 8 Nov 2023 10:21:19 -0800 Subject: [PATCH] (test) run ci/cd again --- litellm/proxy/proxy_cli.py | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/litellm/proxy/proxy_cli.py b/litellm/proxy/proxy_cli.py index fe6536ba9..865d807fe 100644 --- a/litellm/proxy/proxy_cli.py +++ b/litellm/proxy/proxy_cli.py @@ -157,7 +157,12 @@ def run_server(host, port, api_base, api_version, model, alias, add_key, headers ) for chunk in response: click.echo(f'LiteLLM: streaming response from proxy {chunk}') - return + + # response = openai.Completion.create(model="gpt-3.5-turbo", prompt='this is a test request, write a short poem', stream=True) + + # for chunk in response: + # click.echo(f'LiteLLM: streaming response from proxy {chunk}') + # return else: if headers: headers = json.loads(headers)