mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
fix: Syntax error with missing stubs at the end of some function calls (#2116)
# What does this PR do? This PR adds stubs to the end of functions create_agent_turn, create_openai_response and job_result. [//]: # (If resolving an issue, uncomment and update the line below) [//]: # (Closes #[issue-number]) ## Test Plan Ran provided unit tests [//]: # (## Documentation)
This commit is contained in:
parent
9a6e91cd93
commit
675f34e79d
2 changed files with 5 additions and 0 deletions
|
@ -415,6 +415,7 @@ class Agents(Protocol):
|
|||
:returns: If stream=False, returns a Turn object.
|
||||
If stream=True, returns an SSE event stream of AgentTurnResponseStreamChunk
|
||||
"""
|
||||
...
|
||||
|
||||
@webmethod(
|
||||
route="/agents/{agent_id}/session/{session_id}/turn/{turn_id}/resume",
|
||||
|
@ -606,3 +607,4 @@ class Agents(Protocol):
|
|||
:param model: The underlying LLM used for completions.
|
||||
:param previous_response_id: (Optional) if specified, the new response will be a continuation of the previous response. This can be used to easily fork-off new responses from existing responses.
|
||||
"""
|
||||
...
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue