mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-12 12:06:04 +00:00
chore: Adding prompts to stainless config
Signed-off-by: Francisco Javier Arceo <farceo@redhat.com>
This commit is contained in:
parent
b90c6a2c8b
commit
66a6d19cb3
1 changed files with 20 additions and 0 deletions
|
|
@ -508,6 +508,26 @@ resources:
|
|||
stream_event_model: alpha.agents.turn.agent_turn_response_stream_chunk
|
||||
param_discriminator: stream
|
||||
|
||||
prompts:
|
||||
models:
|
||||
prompt: Prompt
|
||||
list_prompts_response: ListPromptsResponse
|
||||
methods:
|
||||
create: post /v1/prompts
|
||||
list:
|
||||
endpoint: get /v1/prompts
|
||||
paginated: false
|
||||
retrieve: get /v1/prompts/{prompt_id}
|
||||
update: post /v1/prompts/{prompt_id}
|
||||
delete: delete /v1/prompts/{prompt_id}
|
||||
set_default_version: post /v1/prompts/{prompt_id}/set-default-version
|
||||
subresources:
|
||||
versions:
|
||||
methods:
|
||||
list:
|
||||
endpoint: get /v1/prompts/{prompt_id}/versions
|
||||
paginated: false
|
||||
|
||||
beta:
|
||||
subresources:
|
||||
datasets:
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue