mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-31 11:20:00 +00:00
feat: Support "stop" parameter in remote:vLLM
Signed-off-by: Yuan Tang <terrytangyuan@gmail.com>
This commit is contained in:
parent
f369871083
commit
a1da09e166
4 changed files with 34 additions and 0 deletions
|
|
@ -201,6 +201,7 @@ class SamplingParams(BaseModel):
|
|||
|
||||
max_tokens: Optional[int] = 0
|
||||
repetition_penalty: Optional[float] = 1.0
|
||||
stop: Optional[List[str]] = None
|
||||
|
||||
|
||||
class CheckpointQuantizationFormat(Enum):
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue