mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-10 13:28:40 +00:00
test, recording
# What does this PR do? ## Test Plan # What does this PR do? ## Test Plan
This commit is contained in:
parent
16db42e7e5
commit
c76bf97ccf
228 changed files with 86861 additions and 64604 deletions
|
@ -1052,6 +1052,7 @@ class InferenceProvider(Protocol):
|
|||
prompt_logprobs: int | None = None,
|
||||
# for fill-in-the-middle type completion
|
||||
suffix: str | None = None,
|
||||
**kwargs: Any,
|
||||
) -> OpenAICompletion:
|
||||
"""Create completion.
|
||||
|
||||
|
@ -1075,6 +1076,7 @@ class InferenceProvider(Protocol):
|
|||
:param top_p: (Optional) The top p to use.
|
||||
:param user: (Optional) The user to use.
|
||||
:param suffix: (Optional) The suffix that should be appended to the completion.
|
||||
:param kwargs: (Optional) Additional provider-specific parameters to pass through as extra_body.
|
||||
:returns: An OpenAICompletion.
|
||||
"""
|
||||
...
|
||||
|
@ -1106,6 +1108,7 @@ class InferenceProvider(Protocol):
|
|||
top_logprobs: int | None = None,
|
||||
top_p: float | None = None,
|
||||
user: str | None = None,
|
||||
**kwargs: Any,
|
||||
) -> OpenAIChatCompletion | AsyncIterator[OpenAIChatCompletionChunk]:
|
||||
"""Create chat completions.
|
||||
|
||||
|
@ -1134,6 +1137,7 @@ class InferenceProvider(Protocol):
|
|||
:param top_logprobs: (Optional) The top log probabilities to use.
|
||||
:param top_p: (Optional) The top p to use.
|
||||
:param user: (Optional) The user to use.
|
||||
:param kwargs: (Optional) Additional provider-specific parameters to pass through as extra_body (e.g., chat_template_kwargs for vLLM).
|
||||
:returns: An OpenAIChatCompletion.
|
||||
"""
|
||||
...
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue