mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
dont set num_predict for all providers (#294)
This commit is contained in:
parent
ffb561070d
commit
21f2e9adf5
1 changed files with 0 additions and 2 deletions
|
@ -34,8 +34,6 @@ def get_sampling_options(request: ChatCompletionRequest) -> dict:
|
|||
if params := request.sampling_params:
|
||||
for attr in {"temperature", "top_p", "top_k", "max_tokens"}:
|
||||
if getattr(params, attr):
|
||||
if attr == "max_tokens":
|
||||
options["num_predict"] = getattr(params, attr)
|
||||
options[attr] = getattr(params, attr)
|
||||
|
||||
if params.repetition_penalty is not None and params.repetition_penalty != 1.0:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue