mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-03 01:03:59 +00:00
Update llama_stack/providers/tests/inference/groq/test_groq_utils.py
This commit is contained in:
parent
ee53eb8691
commit
687bc52b4b
1 changed files with 0 additions and 7 deletions
|
@ -122,13 +122,6 @@ class TestConvertChatCompletionRequest:
|
||||||
|
|
||||||
assert converted["stream"] is True
|
assert converted["stream"] is True
|
||||||
|
|
||||||
def test_n_is_1(self):
|
|
||||||
request = self._dummy_chat_completion_request()
|
|
||||||
|
|
||||||
converted = convert_chat_completion_request(request)
|
|
||||||
|
|
||||||
assert converted["n"] == 1
|
|
||||||
|
|
||||||
def test_if_max_tokens_is_0_then_it_is_not_included(self):
|
def test_if_max_tokens_is_0_then_it_is_not_included(self):
|
||||||
request = self._dummy_chat_completion_request()
|
request = self._dummy_chat_completion_request()
|
||||||
# 0 is the default value for max_tokens
|
# 0 is the default value for max_tokens
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue