From 687bc52b4b75dd9db6e88331583e0cf0e31179ea Mon Sep 17 00:00:00 2001 From: Aidan Do Date: Sat, 14 Dec 2024 20:43:35 +1100 Subject: [PATCH] Update llama_stack/providers/tests/inference/groq/test_groq_utils.py --- .../providers/tests/inference/groq/test_groq_utils.py | 7 ------- 1 file changed, 7 deletions(-) diff --git a/llama_stack/providers/tests/inference/groq/test_groq_utils.py b/llama_stack/providers/tests/inference/groq/test_groq_utils.py index 930ad54eb..4bc633751 100644 --- a/llama_stack/providers/tests/inference/groq/test_groq_utils.py +++ b/llama_stack/providers/tests/inference/groq/test_groq_utils.py @@ -122,13 +122,6 @@ class TestConvertChatCompletionRequest: assert converted["stream"] is True - def test_n_is_1(self): - request = self._dummy_chat_completion_request() - - converted = convert_chat_completion_request(request) - - assert converted["n"] == 1 - def test_if_max_tokens_is_0_then_it_is_not_included(self): request = self._dummy_chat_completion_request() # 0 is the default value for max_tokens