mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-02 08:44:44 +00:00
Update llama_stack/providers/remote/inference/groq/groq_utils.py
This commit is contained in:
parent
b78b9ed4a1
commit
ee53eb8691
1 changed files with 1 additions and 1 deletions
|
@ -52,7 +52,7 @@ def convert_chat_completion_request(
|
|||
# Groq's JSON mode is beta at the time of writing
|
||||
warnings.warn("response_format is not supported yet")
|
||||
|
||||
if request.sampling_params.repetition_penalty:
|
||||
if request.sampling_params.repetition_penalty != 1.0:
|
||||
# groq supports frequency_penalty, but frequency_penalty and sampling_params.repetition_penalty
|
||||
# seem to have different semantics
|
||||
# frequency_penalty defaults to 0 is a float between -2.0 and 2.0
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue