mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 19:04:19 +00:00
dont set num_predict for all providers (#294)
This commit is contained in:
parent
ffb561070d
commit
21f2e9adf5
1 changed files with 0 additions and 2 deletions
|
@ -34,8 +34,6 @@ def get_sampling_options(request: ChatCompletionRequest) -> dict:
|
||||||
if params := request.sampling_params:
|
if params := request.sampling_params:
|
||||||
for attr in {"temperature", "top_p", "top_k", "max_tokens"}:
|
for attr in {"temperature", "top_p", "top_k", "max_tokens"}:
|
||||||
if getattr(params, attr):
|
if getattr(params, attr):
|
||||||
if attr == "max_tokens":
|
|
||||||
options["num_predict"] = getattr(params, attr)
|
|
||||||
options[attr] = getattr(params, attr)
|
options[attr] = getattr(params, attr)
|
||||||
|
|
||||||
if params.repetition_penalty is not None and params.repetition_penalty != 1.0:
|
if params.repetition_penalty is not None and params.repetition_penalty != 1.0:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue