diff --git a/litellm/main.py b/litellm/main.py index ef8a61758..24d0faf61 100644 --- a/litellm/main.py +++ b/litellm/main.py @@ -324,7 +324,7 @@ def completion( ## LOGGING logging(model=model, input=prompt, custom_llm_provider=custom_llm_provider, logger_fn=logger_fn) - if stream == True or optional_params['stream_tokens'] == True: + if stream == True: return together_ai_completion_streaming({ "model": model, "prompt": prompt,