mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-03 19:57:35 +00:00
do not reintroduce chat_completion
This commit is contained in:
parent
0e4279707e
commit
a63ea05114
3 changed files with 1249 additions and 1928 deletions
1804
docs/static/llama-stack-spec.html
vendored
1804
docs/static/llama-stack-spec.html
vendored
File diff suppressed because it is too large
Load diff
1372
docs/static/llama-stack-spec.yaml
vendored
1372
docs/static/llama-stack-spec.yaml
vendored
File diff suppressed because it is too large
Load diff
|
@ -1008,7 +1008,6 @@ class InferenceProvider(Protocol):
|
|||
|
||||
model_store: ModelStore | None = None
|
||||
|
||||
@webmethod(route="/inference/chat-completion", method="POST", level=LLAMA_STACK_API_V1)
|
||||
async def chat_completion(
|
||||
self,
|
||||
model_id: str,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue