From 56137795685a9aa43e2bc305d1c5ccd5c656fe61 Mon Sep 17 00:00:00 2001 From: Alexey Rybak Date: Thu, 2 Oct 2025 13:04:37 -0700 Subject: [PATCH 1/4] docs: API docstrings cleanup for better documentation rendering --- docs/static/deprecated-llama-stack-spec.html | 74 +++---- docs/static/deprecated-llama-stack-spec.yaml | 97 +++++---- docs/static/llama-stack-spec.html | 145 ++++++------- docs/static/llama-stack-spec.yaml | 203 ++++++++++++------- docs/static/stainless-llama-stack-spec.html | 145 ++++++------- docs/static/stainless-llama-stack-spec.yaml | 203 ++++++++++++------- llama_stack/apis/agents/agents.py | 10 +- llama_stack/apis/files/files.py | 20 +- llama_stack/apis/inference/inference.py | 22 +- llama_stack/apis/inspect/inspect.py | 17 +- llama_stack/apis/models/models.py | 12 +- llama_stack/apis/prompts/prompts.py | 28 ++- llama_stack/apis/providers/providers.py | 11 +- llama_stack/apis/safety/safety.py | 13 +- 14 files changed, 591 insertions(+), 409 deletions(-) diff --git a/docs/static/deprecated-llama-stack-spec.html b/docs/static/deprecated-llama-stack-spec.html index 99ce8ee9c..206aba5d2 100644 --- a/docs/static/deprecated-llama-stack-spec.html +++ b/docs/static/deprecated-llama-stack-spec.html @@ -1443,8 +1443,8 @@ "tags": [ "Inference" ], - "summary": "List all chat completions.", - "description": "List all chat completions.", + "summary": "List chat completions.", + "description": "List chat completions.", "parameters": [ { "name": "after", @@ -1520,8 +1520,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", - "description": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + "summary": "Create chat completions.", + "description": "Create chat completions.\nGenerate an OpenAI-compatible chat completion for the given messages using the specified model.", "parameters": [], "requestBody": { "content": { @@ -1565,8 +1565,8 @@ "tags": [ "Inference" ], - "summary": "Describe a chat completion by its ID.", - "description": "Describe a chat completion by its ID.", + "summary": "Get chat completion.", + "description": "Get chat completion.\nDescribe a chat completion by its ID.", "parameters": [ { "name": "completion_id", @@ -1610,8 +1610,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", - "description": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + "summary": "Create completion.", + "description": "Create completion.\nGenerate an OpenAI-compatible completion for the given prompt using the specified model.", "parameters": [], "requestBody": { "content": { @@ -1655,8 +1655,8 @@ "tags": [ "Inference" ], - "summary": "Generate OpenAI-compatible embeddings for the given input using the specified model.", - "description": "Generate OpenAI-compatible embeddings for the given input using the specified model.", + "summary": "Create embeddings.", + "description": "Create embeddings.\nGenerate OpenAI-compatible embeddings for the given input using the specified model.", "parameters": [], "requestBody": { "content": { @@ -1700,8 +1700,8 @@ "tags": [ "Files" ], - "summary": "Returns a list of files that belong to the user's organization.", - "description": "Returns a list of files that belong to the user's organization.", + "summary": "List files.", + "description": "List files.\nReturns a list of files that belong to the user's organization.", "parameters": [ { "name": "after", @@ -1770,8 +1770,8 @@ "tags": [ "Files" ], - "summary": "Upload a file that can be used across various endpoints.", - "description": "Upload a file that can be used across various endpoints.\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", + "summary": "Upload file.", + "description": "Upload file.\nUpload a file that can be used across various endpoints.\n\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", "parameters": [], "requestBody": { "content": { @@ -1831,8 +1831,8 @@ "tags": [ "Files" ], - "summary": "Returns information about a specific file.", - "description": "Returns information about a specific file.", + "summary": "Retrieve file.", + "description": "Retrieve file.\nReturns information about a specific file.", "parameters": [ { "name": "file_id", @@ -1874,8 +1874,8 @@ "tags": [ "Files" ], - "summary": "Delete a file.", - "description": "Delete a file.", + "summary": "Delete file.", + "description": "Delete file.", "parameters": [ { "name": "file_id", @@ -1919,8 +1919,8 @@ "tags": [ "Files" ], - "summary": "Returns the contents of the specified file.", - "description": "Returns the contents of the specified file.", + "summary": "Retrieve file content.", + "description": "Retrieve file content.\nReturns the contents of the specified file.", "parameters": [ { "name": "file_id", @@ -1999,8 +1999,8 @@ "tags": [ "Safety" ], - "summary": "Classifies if text and/or image inputs are potentially harmful.", - "description": "Classifies if text and/or image inputs are potentially harmful.", + "summary": "Create moderation.", + "description": "Create moderation.\nClassifies if text and/or image inputs are potentially harmful.", "parameters": [], "requestBody": { "content": { @@ -2044,8 +2044,8 @@ "tags": [ "Agents" ], - "summary": "List all OpenAI responses.", - "description": "List all OpenAI responses.", + "summary": "List all responses.", + "description": "List all responses.", "parameters": [ { "name": "after", @@ -2114,8 +2114,8 @@ "tags": [ "Agents" ], - "summary": "List all OpenAI responses.", - "description": "List all OpenAI responses.", + "summary": "List all responses.", + "description": "List all responses.", "parameters": [], "requestBody": { "content": { @@ -2159,8 +2159,8 @@ "tags": [ "Agents" ], - "summary": "Retrieve an OpenAI response by its ID.", - "description": "Retrieve an OpenAI response by its ID.", + "summary": "Get a model response.", + "description": "Get a model response.", "parameters": [ { "name": "response_id", @@ -2202,8 +2202,8 @@ "tags": [ "Agents" ], - "summary": "Delete an OpenAI response by its ID.", - "description": "Delete an OpenAI response by its ID.", + "summary": "Delete a response.", + "description": "Delete a response.", "parameters": [ { "name": "response_id", @@ -2247,8 +2247,8 @@ "tags": [ "Agents" ], - "summary": "List input items for a given OpenAI response.", - "description": "List input items for a given OpenAI response.", + "summary": "List input items.", + "description": "List input items.", "parameters": [ { "name": "response_id", @@ -13439,12 +13439,13 @@ }, { "name": "Files", - "description": "" + "description": "This API is used to upload documents that can be used with other Llama Stack APIs.", + "x-displayName": "Files" }, { "name": "Inference", - "description": "This API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", - "x-displayName": "Llama Stack Inference API for generating completions, chat completions, and embeddings." + "description": "Llama Stack Inference API for generating completions, chat completions, and embeddings.\n\nThis API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", + "x-displayName": "Inference" }, { "name": "Models", @@ -13456,7 +13457,8 @@ }, { "name": "Safety", - "description": "" + "description": "OpenAI-compatible Moderations API.", + "x-displayName": "Safety" }, { "name": "Telemetry", diff --git a/docs/static/deprecated-llama-stack-spec.yaml b/docs/static/deprecated-llama-stack-spec.yaml index d2e595b5d..40868c563 100644 --- a/docs/static/deprecated-llama-stack-spec.yaml +++ b/docs/static/deprecated-llama-stack-spec.yaml @@ -1033,8 +1033,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: List all chat completions. - description: List all chat completions. + summary: List chat completions. + description: List chat completions. parameters: - name: after in: query @@ -1087,10 +1087,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible chat completion for the given messages using - the specified model. + summary: Create chat completions. description: >- + Create chat completions. + Generate an OpenAI-compatible chat completion for the given messages using the specified model. parameters: [] @@ -1122,8 +1122,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Describe a chat completion by its ID. - description: Describe a chat completion by its ID. + summary: Get chat completion. + description: >- + Get chat completion. + + Describe a chat completion by its ID. parameters: - name: completion_id in: path @@ -1153,10 +1156,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible completion for the given prompt using the specified - model. + summary: Create completion. description: >- + Create completion. + Generate an OpenAI-compatible completion for the given prompt using the specified model. parameters: [] @@ -1189,10 +1192,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate OpenAI-compatible embeddings for the given input using the specified - model. + summary: Create embeddings. description: >- + Create embeddings. + Generate OpenAI-compatible embeddings for the given input using the specified model. parameters: [] @@ -1225,9 +1228,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns a list of files that belong to the user's organization. + summary: List files. description: >- + List files. + Returns a list of files that belong to the user's organization. parameters: - name: after @@ -1285,11 +1289,13 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Upload a file that can be used across various endpoints. + summary: Upload file. description: >- + Upload file. + Upload a file that can be used across various endpoints. + The file upload should be a multipart form request with: - file: The File object (not file name) to be uploaded. @@ -1338,9 +1344,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns information about a specific file. + summary: Retrieve file. description: >- + Retrieve file. + Returns information about a specific file. parameters: - name: file_id @@ -1372,8 +1379,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Delete a file. - description: Delete a file. + summary: Delete file. + description: Delete file. parameters: - name: file_id in: path @@ -1405,9 +1412,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns the contents of the specified file. + summary: Retrieve file content. description: >- + Retrieve file content. + Returns the contents of the specified file. parameters: - name: file_id @@ -1464,9 +1472,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: >- - Classifies if text and/or image inputs are potentially harmful. + summary: Create moderation. description: >- + Create moderation. + Classifies if text and/or image inputs are potentially harmful. parameters: [] requestBody: @@ -1497,8 +1506,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all OpenAI responses. - description: List all OpenAI responses. + summary: List all responses. + description: List all responses. parameters: - name: after in: query @@ -1546,8 +1555,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all OpenAI responses. - description: List all OpenAI responses. + summary: List all responses. + description: List all responses. parameters: [] requestBody: content: @@ -1577,8 +1586,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Retrieve an OpenAI response by its ID. - description: Retrieve an OpenAI response by its ID. + summary: Get a model response. + description: Get a model response. parameters: - name: response_id in: path @@ -1608,8 +1617,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Delete an OpenAI response by its ID. - description: Delete an OpenAI response by its ID. + summary: Delete a response. + description: Delete a response. parameters: - name: response_id in: path @@ -1639,10 +1648,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: >- - List input items for a given OpenAI response. - description: >- - List input items for a given OpenAI response. + summary: List input items. + description: List input items. parameters: - name: response_id in: path @@ -10054,9 +10061,16 @@ tags: x-displayName: >- Llama Stack Evaluation API for running evaluations on model and agent candidates. - name: Files - description: '' + description: >- + This API is used to upload documents that can be used with other Llama Stack + APIs. + x-displayName: Files - name: Inference description: >- + Llama Stack Inference API for generating completions, chat completions, and + embeddings. + + This API provides the raw interface to the underlying models. Two kinds of models are supported: @@ -10064,15 +10078,14 @@ tags: - Embedding models: these models generate embeddings to be used for semantic search. - x-displayName: >- - Llama Stack Inference API for generating completions, chat completions, and - embeddings. + x-displayName: Inference - name: Models description: '' - name: PostTraining (Coming Soon) description: '' - name: Safety - description: '' + description: OpenAI-compatible Moderations API. + x-displayName: Safety - name: Telemetry description: '' - name: VectorIO diff --git a/docs/static/llama-stack-spec.html b/docs/static/llama-stack-spec.html index 3da721a4e..324b38a1c 100644 --- a/docs/static/llama-stack-spec.html +++ b/docs/static/llama-stack-spec.html @@ -69,8 +69,8 @@ "tags": [ "Inference" ], - "summary": "List all chat completions.", - "description": "List all chat completions.", + "summary": "List chat completions.", + "description": "List chat completions.", "parameters": [ { "name": "after", @@ -146,8 +146,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", - "description": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + "summary": "Create chat completions.", + "description": "Create chat completions.\nGenerate an OpenAI-compatible chat completion for the given messages using the specified model.", "parameters": [], "requestBody": { "content": { @@ -191,8 +191,8 @@ "tags": [ "Inference" ], - "summary": "Describe a chat completion by its ID.", - "description": "Describe a chat completion by its ID.", + "summary": "Get chat completion.", + "description": "Get chat completion.\nDescribe a chat completion by its ID.", "parameters": [ { "name": "completion_id", @@ -236,8 +236,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", - "description": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + "summary": "Create completion.", + "description": "Create completion.\nGenerate an OpenAI-compatible completion for the given prompt using the specified model.", "parameters": [], "requestBody": { "content": { @@ -281,8 +281,8 @@ "tags": [ "Inference" ], - "summary": "Generate OpenAI-compatible embeddings for the given input using the specified model.", - "description": "Generate OpenAI-compatible embeddings for the given input using the specified model.", + "summary": "Create embeddings.", + "description": "Create embeddings.\nGenerate OpenAI-compatible embeddings for the given input using the specified model.", "parameters": [], "requestBody": { "content": { @@ -326,8 +326,8 @@ "tags": [ "Files" ], - "summary": "Returns a list of files that belong to the user's organization.", - "description": "Returns a list of files that belong to the user's organization.", + "summary": "List files.", + "description": "List files.\nReturns a list of files that belong to the user's organization.", "parameters": [ { "name": "after", @@ -396,8 +396,8 @@ "tags": [ "Files" ], - "summary": "Upload a file that can be used across various endpoints.", - "description": "Upload a file that can be used across various endpoints.\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", + "summary": "Upload file.", + "description": "Upload file.\nUpload a file that can be used across various endpoints.\n\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", "parameters": [], "requestBody": { "content": { @@ -457,8 +457,8 @@ "tags": [ "Files" ], - "summary": "Returns information about a specific file.", - "description": "Returns information about a specific file.", + "summary": "Retrieve file.", + "description": "Retrieve file.\nReturns information about a specific file.", "parameters": [ { "name": "file_id", @@ -500,8 +500,8 @@ "tags": [ "Files" ], - "summary": "Delete a file.", - "description": "Delete a file.", + "summary": "Delete file.", + "description": "Delete file.", "parameters": [ { "name": "file_id", @@ -545,8 +545,8 @@ "tags": [ "Files" ], - "summary": "Returns the contents of the specified file.", - "description": "Returns the contents of the specified file.", + "summary": "Retrieve file content.", + "description": "Retrieve file content.\nReturns the contents of the specified file.", "parameters": [ { "name": "file_id", @@ -590,8 +590,8 @@ "tags": [ "Inspect" ], - "summary": "Get the current health status of the service.", - "description": "Get the current health status of the service.", + "summary": "Get health status.", + "description": "Get health status.\nGet the current health status of the service.", "parameters": [], "deprecated": false } @@ -625,8 +625,8 @@ "tags": [ "Inspect" ], - "summary": "List all available API routes with their methods and implementing providers.", - "description": "List all available API routes with their methods and implementing providers.", + "summary": "List routes.", + "description": "List routes.\nList all available API routes with their methods and implementing providers.", "parameters": [], "deprecated": false } @@ -693,8 +693,8 @@ "tags": [ "Models" ], - "summary": "Register a model.", - "description": "Register a model.", + "summary": "Register model.", + "description": "Register model.\nRegister a model.", "parameters": [], "requestBody": { "content": { @@ -738,8 +738,8 @@ "tags": [ "Models" ], - "summary": "Get a model by its identifier.", - "description": "Get a model by its identifier.", + "summary": "Get model.", + "description": "Get model.\nGet a model by its identifier.", "parameters": [ { "name": "model_id", @@ -774,8 +774,8 @@ "tags": [ "Models" ], - "summary": "Unregister a model.", - "description": "Unregister a model.", + "summary": "Unregister model.", + "description": "Unregister model.\nUnregister a model.", "parameters": [ { "name": "model_id", @@ -819,8 +819,8 @@ "tags": [ "Safety" ], - "summary": "Classifies if text and/or image inputs are potentially harmful.", - "description": "Classifies if text and/or image inputs are potentially harmful.", + "summary": "Create moderation.", + "description": "Create moderation.\nClassifies if text and/or image inputs are potentially harmful.", "parameters": [], "requestBody": { "content": { @@ -897,8 +897,8 @@ "tags": [ "Prompts" ], - "summary": "Create a new prompt.", - "description": "Create a new prompt.", + "summary": "Create prompt.", + "description": "Create prompt.\nCreate a new prompt.", "parameters": [], "requestBody": { "content": { @@ -942,8 +942,8 @@ "tags": [ "Prompts" ], - "summary": "Get a prompt by its identifier and optional version.", - "description": "Get a prompt by its identifier and optional version.", + "summary": "Get prompt.", + "description": "Get prompt.\nGet a prompt by its identifier and optional version.", "parameters": [ { "name": "prompt_id", @@ -994,8 +994,8 @@ "tags": [ "Prompts" ], - "summary": "Update an existing prompt (increments version).", - "description": "Update an existing prompt (increments version).", + "summary": "Update prompt.", + "description": "Update prompt.\nUpdate an existing prompt (increments version).", "parameters": [ { "name": "prompt_id", @@ -1040,8 +1040,8 @@ "tags": [ "Prompts" ], - "summary": "Delete a prompt.", - "description": "Delete a prompt.", + "summary": "Delete prompt.", + "description": "Delete prompt.\nDelete a prompt.", "parameters": [ { "name": "prompt_id", @@ -1085,8 +1085,8 @@ "tags": [ "Prompts" ], - "summary": "Set which version of a prompt should be the default in get_prompt (latest).", - "description": "Set which version of a prompt should be the default in get_prompt (latest).", + "summary": "Set prompt version.", + "description": "Set prompt version.\nSet which version of a prompt should be the default in get_prompt (latest).", "parameters": [ { "name": "prompt_id", @@ -1140,8 +1140,8 @@ "tags": [ "Prompts" ], - "summary": "List all versions of a specific prompt.", - "description": "List all versions of a specific prompt.", + "summary": "List prompt versions.", + "description": "List prompt versions.\nList all versions of a specific prompt.", "parameters": [ { "name": "prompt_id", @@ -1185,8 +1185,8 @@ "tags": [ "Providers" ], - "summary": "List all available providers.", - "description": "List all available providers.", + "summary": "List providers.", + "description": "List providers.\nList all available providers.", "parameters": [], "deprecated": false } @@ -1220,8 +1220,8 @@ "tags": [ "Providers" ], - "summary": "Get detailed information about a specific provider.", - "description": "Get detailed information about a specific provider.", + "summary": "Get provider.", + "description": "Get provider.\nGet detailed information about a specific provider.", "parameters": [ { "name": "provider_id", @@ -1265,8 +1265,8 @@ "tags": [ "Agents" ], - "summary": "List all OpenAI responses.", - "description": "List all OpenAI responses.", + "summary": "List all responses.", + "description": "List all responses.", "parameters": [ { "name": "after", @@ -1335,8 +1335,8 @@ "tags": [ "Agents" ], - "summary": "List all OpenAI responses.", - "description": "List all OpenAI responses.", + "summary": "List all responses.", + "description": "List all responses.", "parameters": [], "requestBody": { "content": { @@ -1380,8 +1380,8 @@ "tags": [ "Agents" ], - "summary": "Retrieve an OpenAI response by its ID.", - "description": "Retrieve an OpenAI response by its ID.", + "summary": "Get a model response.", + "description": "Get a model response.", "parameters": [ { "name": "response_id", @@ -1423,8 +1423,8 @@ "tags": [ "Agents" ], - "summary": "Delete an OpenAI response by its ID.", - "description": "Delete an OpenAI response by its ID.", + "summary": "Delete a response.", + "description": "Delete a response.", "parameters": [ { "name": "response_id", @@ -1468,8 +1468,8 @@ "tags": [ "Agents" ], - "summary": "List input items for a given OpenAI response.", - "description": "List input items for a given OpenAI response.", + "summary": "List input items.", + "description": "List input items.", "parameters": [ { "name": "response_id", @@ -1561,8 +1561,8 @@ "tags": [ "Safety" ], - "summary": "Run a shield.", - "description": "Run a shield.", + "summary": "Run shield.", + "description": "Run shield.\nRun a shield.", "parameters": [], "requestBody": { "content": { @@ -3694,8 +3694,8 @@ "tags": [ "Inspect" ], - "summary": "Get the version of the service.", - "description": "Get the version of the service.", + "summary": "Get version.", + "description": "Get version.\nGet the version of the service.", "parameters": [], "deprecated": false } @@ -12399,16 +12399,18 @@ }, { "name": "Files", - "description": "" + "description": "This API is used to upload documents that can be used with other Llama Stack APIs.", + "x-displayName": "Files" }, { "name": "Inference", - "description": "This API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", - "x-displayName": "Llama Stack Inference API for generating completions, chat completions, and embeddings." + "description": "Llama Stack Inference API for generating completions, chat completions, and embeddings.\n\nThis API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", + "x-displayName": "Inference" }, { "name": "Inspect", - "description": "" + "description": "APIs for inspecting the Llama Stack service, including health status, available API routes with methods and implementing providers.", + "x-displayName": "Inspect" }, { "name": "Models", @@ -12416,17 +12418,18 @@ }, { "name": "Prompts", - "description": "", - "x-displayName": "Protocol for prompt management operations." + "description": "Protocol for prompt management operations.", + "x-displayName": "Prompts" }, { "name": "Providers", - "description": "", - "x-displayName": "Providers API for inspecting, listing, and modifying providers and their configurations." + "description": "Providers API for inspecting, listing, and modifying providers and their configurations.", + "x-displayName": "Providers" }, { "name": "Safety", - "description": "" + "description": "OpenAI-compatible Moderations API.", + "x-displayName": "Safety" }, { "name": "Scoring", diff --git a/docs/static/llama-stack-spec.yaml b/docs/static/llama-stack-spec.yaml index 3927d3a94..6274ec4b8 100644 --- a/docs/static/llama-stack-spec.yaml +++ b/docs/static/llama-stack-spec.yaml @@ -33,8 +33,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: List all chat completions. - description: List all chat completions. + summary: List chat completions. + description: List chat completions. parameters: - name: after in: query @@ -87,10 +87,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible chat completion for the given messages using - the specified model. + summary: Create chat completions. description: >- + Create chat completions. + Generate an OpenAI-compatible chat completion for the given messages using the specified model. parameters: [] @@ -122,8 +122,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Describe a chat completion by its ID. - description: Describe a chat completion by its ID. + summary: Get chat completion. + description: >- + Get chat completion. + + Describe a chat completion by its ID. parameters: - name: completion_id in: path @@ -153,10 +156,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible completion for the given prompt using the specified - model. + summary: Create completion. description: >- + Create completion. + Generate an OpenAI-compatible completion for the given prompt using the specified model. parameters: [] @@ -189,10 +192,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate OpenAI-compatible embeddings for the given input using the specified - model. + summary: Create embeddings. description: >- + Create embeddings. + Generate OpenAI-compatible embeddings for the given input using the specified model. parameters: [] @@ -225,9 +228,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns a list of files that belong to the user's organization. + summary: List files. description: >- + List files. + Returns a list of files that belong to the user's organization. parameters: - name: after @@ -285,11 +289,13 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Upload a file that can be used across various endpoints. + summary: Upload file. description: >- + Upload file. + Upload a file that can be used across various endpoints. + The file upload should be a multipart form request with: - file: The File object (not file name) to be uploaded. @@ -338,9 +344,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns information about a specific file. + summary: Retrieve file. description: >- + Retrieve file. + Returns information about a specific file. parameters: - name: file_id @@ -372,8 +379,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Delete a file. - description: Delete a file. + summary: Delete file. + description: Delete file. parameters: - name: file_id in: path @@ -405,9 +412,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns the contents of the specified file. + summary: Retrieve file content. description: >- + Retrieve file content. + Returns the contents of the specified file. parameters: - name: file_id @@ -440,9 +448,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: >- - Get the current health status of the service. + summary: Get health status. description: >- + Get health status. + Get the current health status of the service. parameters: [] deprecated: false @@ -468,9 +477,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: >- - List all available API routes with their methods and implementing providers. + summary: List routes. description: >- + List routes. + List all available API routes with their methods and implementing providers. parameters: [] deprecated: false @@ -519,8 +529,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Register a model. - description: Register a model. + summary: Register model. + description: >- + Register model. + + Register a model. parameters: [] requestBody: content: @@ -550,8 +563,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Get a model by its identifier. - description: Get a model by its identifier. + summary: Get model. + description: >- + Get model. + + Get a model by its identifier. parameters: - name: model_id in: path @@ -576,8 +592,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Unregister a model. - description: Unregister a model. + summary: Unregister model. + description: >- + Unregister model. + + Unregister a model. parameters: - name: model_id in: path @@ -608,9 +627,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: >- - Classifies if text and/or image inputs are potentially harmful. + summary: Create moderation. description: >- + Create moderation. + Classifies if text and/or image inputs are potentially harmful. parameters: [] requestBody: @@ -666,8 +686,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Create a new prompt. - description: Create a new prompt. + summary: Create prompt. + description: >- + Create prompt. + + Create a new prompt. parameters: [] requestBody: content: @@ -697,9 +720,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Get a prompt by its identifier and optional version. + summary: Get prompt. description: >- + Get prompt. + Get a prompt by its identifier and optional version. parameters: - name: prompt_id @@ -737,9 +761,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Update an existing prompt (increments version). + summary: Update prompt. description: >- + Update prompt. + Update an existing prompt (increments version). parameters: - name: prompt_id @@ -771,8 +796,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Delete a prompt. - description: Delete a prompt. + summary: Delete prompt. + description: >- + Delete prompt. + + Delete a prompt. parameters: - name: prompt_id in: path @@ -803,9 +831,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Set which version of a prompt should be the default in get_prompt (latest). + summary: Set prompt version. description: >- + Set prompt version. + Set which version of a prompt should be the default in get_prompt (latest). parameters: - name: prompt_id @@ -843,8 +872,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: List all versions of a specific prompt. - description: List all versions of a specific prompt. + summary: List prompt versions. + description: >- + List prompt versions. + + List all versions of a specific prompt. parameters: - name: prompt_id in: path @@ -876,8 +908,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: List all available providers. - description: List all available providers. + summary: List providers. + description: >- + List providers. + + List all available providers. parameters: [] deprecated: false /v1/providers/{provider_id}: @@ -902,9 +937,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: >- - Get detailed information about a specific provider. + summary: Get provider. description: >- + Get provider. + Get detailed information about a specific provider. parameters: - name: provider_id @@ -935,8 +971,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all OpenAI responses. - description: List all OpenAI responses. + summary: List all responses. + description: List all responses. parameters: - name: after in: query @@ -984,8 +1020,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all OpenAI responses. - description: List all OpenAI responses. + summary: List all responses. + description: List all responses. parameters: [] requestBody: content: @@ -1015,8 +1051,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Retrieve an OpenAI response by its ID. - description: Retrieve an OpenAI response by its ID. + summary: Get a model response. + description: Get a model response. parameters: - name: response_id in: path @@ -1046,8 +1082,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Delete an OpenAI response by its ID. - description: Delete an OpenAI response by its ID. + summary: Delete a response. + description: Delete a response. parameters: - name: response_id in: path @@ -1077,10 +1113,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: >- - List input items for a given OpenAI response. - description: >- - List input items for a given OpenAI response. + summary: List input items. + description: List input items. parameters: - name: response_id in: path @@ -1149,8 +1183,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: Run a shield. - description: Run a shield. + summary: Run shield. + description: >- + Run shield. + + Run a shield. parameters: [] requestBody: content: @@ -2706,8 +2743,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: Get the version of the service. - description: Get the version of the service. + summary: Get version. + description: >- + Get version. + + Get the version of the service. parameters: [] deprecated: false jsonSchemaDialect: >- @@ -9292,9 +9332,16 @@ tags: - `background` x-displayName: Agents - name: Files - description: '' + description: >- + This API is used to upload documents that can be used with other Llama Stack + APIs. + x-displayName: Files - name: Inference description: >- + Llama Stack Inference API for generating completions, chat completions, and + embeddings. + + This API provides the raw interface to the underlying models. Two kinds of models are supported: @@ -9302,23 +9349,25 @@ tags: - Embedding models: these models generate embeddings to be used for semantic search. - x-displayName: >- - Llama Stack Inference API for generating completions, chat completions, and - embeddings. + x-displayName: Inference - name: Inspect - description: '' + description: >- + APIs for inspecting the Llama Stack service, including health status, available + API routes with methods and implementing providers. + x-displayName: Inspect - name: Models description: '' - name: Prompts - description: '' - x-displayName: >- + description: >- Protocol for prompt management operations. + x-displayName: Prompts - name: Providers - description: '' - x-displayName: >- + description: >- Providers API for inspecting, listing, and modifying providers and their configurations. + x-displayName: Providers - name: Safety - description: '' + description: OpenAI-compatible Moderations API. + x-displayName: Safety - name: Scoring description: '' - name: ScoringFunctions diff --git a/docs/static/stainless-llama-stack-spec.html b/docs/static/stainless-llama-stack-spec.html index f921d2c29..7783af4eb 100644 --- a/docs/static/stainless-llama-stack-spec.html +++ b/docs/static/stainless-llama-stack-spec.html @@ -69,8 +69,8 @@ "tags": [ "Inference" ], - "summary": "List all chat completions.", - "description": "List all chat completions.", + "summary": "List chat completions.", + "description": "List chat completions.", "parameters": [ { "name": "after", @@ -146,8 +146,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", - "description": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + "summary": "Create chat completions.", + "description": "Create chat completions.\nGenerate an OpenAI-compatible chat completion for the given messages using the specified model.", "parameters": [], "requestBody": { "content": { @@ -191,8 +191,8 @@ "tags": [ "Inference" ], - "summary": "Describe a chat completion by its ID.", - "description": "Describe a chat completion by its ID.", + "summary": "Get chat completion.", + "description": "Get chat completion.\nDescribe a chat completion by its ID.", "parameters": [ { "name": "completion_id", @@ -236,8 +236,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", - "description": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + "summary": "Create completion.", + "description": "Create completion.\nGenerate an OpenAI-compatible completion for the given prompt using the specified model.", "parameters": [], "requestBody": { "content": { @@ -281,8 +281,8 @@ "tags": [ "Inference" ], - "summary": "Generate OpenAI-compatible embeddings for the given input using the specified model.", - "description": "Generate OpenAI-compatible embeddings for the given input using the specified model.", + "summary": "Create embeddings.", + "description": "Create embeddings.\nGenerate OpenAI-compatible embeddings for the given input using the specified model.", "parameters": [], "requestBody": { "content": { @@ -326,8 +326,8 @@ "tags": [ "Files" ], - "summary": "Returns a list of files that belong to the user's organization.", - "description": "Returns a list of files that belong to the user's organization.", + "summary": "List files.", + "description": "List files.\nReturns a list of files that belong to the user's organization.", "parameters": [ { "name": "after", @@ -396,8 +396,8 @@ "tags": [ "Files" ], - "summary": "Upload a file that can be used across various endpoints.", - "description": "Upload a file that can be used across various endpoints.\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", + "summary": "Upload file.", + "description": "Upload file.\nUpload a file that can be used across various endpoints.\n\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", "parameters": [], "requestBody": { "content": { @@ -457,8 +457,8 @@ "tags": [ "Files" ], - "summary": "Returns information about a specific file.", - "description": "Returns information about a specific file.", + "summary": "Retrieve file.", + "description": "Retrieve file.\nReturns information about a specific file.", "parameters": [ { "name": "file_id", @@ -500,8 +500,8 @@ "tags": [ "Files" ], - "summary": "Delete a file.", - "description": "Delete a file.", + "summary": "Delete file.", + "description": "Delete file.", "parameters": [ { "name": "file_id", @@ -545,8 +545,8 @@ "tags": [ "Files" ], - "summary": "Returns the contents of the specified file.", - "description": "Returns the contents of the specified file.", + "summary": "Retrieve file content.", + "description": "Retrieve file content.\nReturns the contents of the specified file.", "parameters": [ { "name": "file_id", @@ -590,8 +590,8 @@ "tags": [ "Inspect" ], - "summary": "Get the current health status of the service.", - "description": "Get the current health status of the service.", + "summary": "Get health status.", + "description": "Get health status.\nGet the current health status of the service.", "parameters": [], "deprecated": false } @@ -625,8 +625,8 @@ "tags": [ "Inspect" ], - "summary": "List all available API routes with their methods and implementing providers.", - "description": "List all available API routes with their methods and implementing providers.", + "summary": "List routes.", + "description": "List routes.\nList all available API routes with their methods and implementing providers.", "parameters": [], "deprecated": false } @@ -693,8 +693,8 @@ "tags": [ "Models" ], - "summary": "Register a model.", - "description": "Register a model.", + "summary": "Register model.", + "description": "Register model.\nRegister a model.", "parameters": [], "requestBody": { "content": { @@ -738,8 +738,8 @@ "tags": [ "Models" ], - "summary": "Get a model by its identifier.", - "description": "Get a model by its identifier.", + "summary": "Get model.", + "description": "Get model.\nGet a model by its identifier.", "parameters": [ { "name": "model_id", @@ -774,8 +774,8 @@ "tags": [ "Models" ], - "summary": "Unregister a model.", - "description": "Unregister a model.", + "summary": "Unregister model.", + "description": "Unregister model.\nUnregister a model.", "parameters": [ { "name": "model_id", @@ -819,8 +819,8 @@ "tags": [ "Safety" ], - "summary": "Classifies if text and/or image inputs are potentially harmful.", - "description": "Classifies if text and/or image inputs are potentially harmful.", + "summary": "Create moderation.", + "description": "Create moderation.\nClassifies if text and/or image inputs are potentially harmful.", "parameters": [], "requestBody": { "content": { @@ -897,8 +897,8 @@ "tags": [ "Prompts" ], - "summary": "Create a new prompt.", - "description": "Create a new prompt.", + "summary": "Create prompt.", + "description": "Create prompt.\nCreate a new prompt.", "parameters": [], "requestBody": { "content": { @@ -942,8 +942,8 @@ "tags": [ "Prompts" ], - "summary": "Get a prompt by its identifier and optional version.", - "description": "Get a prompt by its identifier and optional version.", + "summary": "Get prompt.", + "description": "Get prompt.\nGet a prompt by its identifier and optional version.", "parameters": [ { "name": "prompt_id", @@ -994,8 +994,8 @@ "tags": [ "Prompts" ], - "summary": "Update an existing prompt (increments version).", - "description": "Update an existing prompt (increments version).", + "summary": "Update prompt.", + "description": "Update prompt.\nUpdate an existing prompt (increments version).", "parameters": [ { "name": "prompt_id", @@ -1040,8 +1040,8 @@ "tags": [ "Prompts" ], - "summary": "Delete a prompt.", - "description": "Delete a prompt.", + "summary": "Delete prompt.", + "description": "Delete prompt.\nDelete a prompt.", "parameters": [ { "name": "prompt_id", @@ -1085,8 +1085,8 @@ "tags": [ "Prompts" ], - "summary": "Set which version of a prompt should be the default in get_prompt (latest).", - "description": "Set which version of a prompt should be the default in get_prompt (latest).", + "summary": "Set prompt version.", + "description": "Set prompt version.\nSet which version of a prompt should be the default in get_prompt (latest).", "parameters": [ { "name": "prompt_id", @@ -1140,8 +1140,8 @@ "tags": [ "Prompts" ], - "summary": "List all versions of a specific prompt.", - "description": "List all versions of a specific prompt.", + "summary": "List prompt versions.", + "description": "List prompt versions.\nList all versions of a specific prompt.", "parameters": [ { "name": "prompt_id", @@ -1185,8 +1185,8 @@ "tags": [ "Providers" ], - "summary": "List all available providers.", - "description": "List all available providers.", + "summary": "List providers.", + "description": "List providers.\nList all available providers.", "parameters": [], "deprecated": false } @@ -1220,8 +1220,8 @@ "tags": [ "Providers" ], - "summary": "Get detailed information about a specific provider.", - "description": "Get detailed information about a specific provider.", + "summary": "Get provider.", + "description": "Get provider.\nGet detailed information about a specific provider.", "parameters": [ { "name": "provider_id", @@ -1265,8 +1265,8 @@ "tags": [ "Agents" ], - "summary": "List all OpenAI responses.", - "description": "List all OpenAI responses.", + "summary": "List all responses.", + "description": "List all responses.", "parameters": [ { "name": "after", @@ -1335,8 +1335,8 @@ "tags": [ "Agents" ], - "summary": "List all OpenAI responses.", - "description": "List all OpenAI responses.", + "summary": "List all responses.", + "description": "List all responses.", "parameters": [], "requestBody": { "content": { @@ -1380,8 +1380,8 @@ "tags": [ "Agents" ], - "summary": "Retrieve an OpenAI response by its ID.", - "description": "Retrieve an OpenAI response by its ID.", + "summary": "Get a model response.", + "description": "Get a model response.", "parameters": [ { "name": "response_id", @@ -1423,8 +1423,8 @@ "tags": [ "Agents" ], - "summary": "Delete an OpenAI response by its ID.", - "description": "Delete an OpenAI response by its ID.", + "summary": "Delete a response.", + "description": "Delete a response.", "parameters": [ { "name": "response_id", @@ -1468,8 +1468,8 @@ "tags": [ "Agents" ], - "summary": "List input items for a given OpenAI response.", - "description": "List input items for a given OpenAI response.", + "summary": "List input items.", + "description": "List input items.", "parameters": [ { "name": "response_id", @@ -1561,8 +1561,8 @@ "tags": [ "Safety" ], - "summary": "Run a shield.", - "description": "Run a shield.", + "summary": "Run shield.", + "description": "Run shield.\nRun a shield.", "parameters": [], "requestBody": { "content": { @@ -3694,8 +3694,8 @@ "tags": [ "Inspect" ], - "summary": "Get the version of the service.", - "description": "Get the version of the service.", + "summary": "Get version.", + "description": "Get version.\nGet the version of the service.", "parameters": [], "deprecated": false } @@ -17972,16 +17972,18 @@ }, { "name": "Files", - "description": "" + "description": "This API is used to upload documents that can be used with other Llama Stack APIs.", + "x-displayName": "Files" }, { "name": "Inference", - "description": "This API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", - "x-displayName": "Llama Stack Inference API for generating completions, chat completions, and embeddings." + "description": "Llama Stack Inference API for generating completions, chat completions, and embeddings.\n\nThis API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", + "x-displayName": "Inference" }, { "name": "Inspect", - "description": "" + "description": "APIs for inspecting the Llama Stack service, including health status, available API routes with methods and implementing providers.", + "x-displayName": "Inspect" }, { "name": "Models", @@ -17993,17 +17995,18 @@ }, { "name": "Prompts", - "description": "", - "x-displayName": "Protocol for prompt management operations." + "description": "Protocol for prompt management operations.", + "x-displayName": "Prompts" }, { "name": "Providers", - "description": "", - "x-displayName": "Providers API for inspecting, listing, and modifying providers and their configurations." + "description": "Providers API for inspecting, listing, and modifying providers and their configurations.", + "x-displayName": "Providers" }, { "name": "Safety", - "description": "" + "description": "OpenAI-compatible Moderations API.", + "x-displayName": "Safety" }, { "name": "Scoring", diff --git a/docs/static/stainless-llama-stack-spec.yaml b/docs/static/stainless-llama-stack-spec.yaml index cb43b313b..cf8f195c3 100644 --- a/docs/static/stainless-llama-stack-spec.yaml +++ b/docs/static/stainless-llama-stack-spec.yaml @@ -36,8 +36,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: List all chat completions. - description: List all chat completions. + summary: List chat completions. + description: List chat completions. parameters: - name: after in: query @@ -90,10 +90,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible chat completion for the given messages using - the specified model. + summary: Create chat completions. description: >- + Create chat completions. + Generate an OpenAI-compatible chat completion for the given messages using the specified model. parameters: [] @@ -125,8 +125,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Describe a chat completion by its ID. - description: Describe a chat completion by its ID. + summary: Get chat completion. + description: >- + Get chat completion. + + Describe a chat completion by its ID. parameters: - name: completion_id in: path @@ -156,10 +159,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible completion for the given prompt using the specified - model. + summary: Create completion. description: >- + Create completion. + Generate an OpenAI-compatible completion for the given prompt using the specified model. parameters: [] @@ -192,10 +195,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate OpenAI-compatible embeddings for the given input using the specified - model. + summary: Create embeddings. description: >- + Create embeddings. + Generate OpenAI-compatible embeddings for the given input using the specified model. parameters: [] @@ -228,9 +231,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns a list of files that belong to the user's organization. + summary: List files. description: >- + List files. + Returns a list of files that belong to the user's organization. parameters: - name: after @@ -288,11 +292,13 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Upload a file that can be used across various endpoints. + summary: Upload file. description: >- + Upload file. + Upload a file that can be used across various endpoints. + The file upload should be a multipart form request with: - file: The File object (not file name) to be uploaded. @@ -341,9 +347,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns information about a specific file. + summary: Retrieve file. description: >- + Retrieve file. + Returns information about a specific file. parameters: - name: file_id @@ -375,8 +382,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Delete a file. - description: Delete a file. + summary: Delete file. + description: Delete file. parameters: - name: file_id in: path @@ -408,9 +415,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns the contents of the specified file. + summary: Retrieve file content. description: >- + Retrieve file content. + Returns the contents of the specified file. parameters: - name: file_id @@ -443,9 +451,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: >- - Get the current health status of the service. + summary: Get health status. description: >- + Get health status. + Get the current health status of the service. parameters: [] deprecated: false @@ -471,9 +480,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: >- - List all available API routes with their methods and implementing providers. + summary: List routes. description: >- + List routes. + List all available API routes with their methods and implementing providers. parameters: [] deprecated: false @@ -522,8 +532,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Register a model. - description: Register a model. + summary: Register model. + description: >- + Register model. + + Register a model. parameters: [] requestBody: content: @@ -553,8 +566,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Get a model by its identifier. - description: Get a model by its identifier. + summary: Get model. + description: >- + Get model. + + Get a model by its identifier. parameters: - name: model_id in: path @@ -579,8 +595,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Unregister a model. - description: Unregister a model. + summary: Unregister model. + description: >- + Unregister model. + + Unregister a model. parameters: - name: model_id in: path @@ -611,9 +630,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: >- - Classifies if text and/or image inputs are potentially harmful. + summary: Create moderation. description: >- + Create moderation. + Classifies if text and/or image inputs are potentially harmful. parameters: [] requestBody: @@ -669,8 +689,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Create a new prompt. - description: Create a new prompt. + summary: Create prompt. + description: >- + Create prompt. + + Create a new prompt. parameters: [] requestBody: content: @@ -700,9 +723,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Get a prompt by its identifier and optional version. + summary: Get prompt. description: >- + Get prompt. + Get a prompt by its identifier and optional version. parameters: - name: prompt_id @@ -740,9 +764,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Update an existing prompt (increments version). + summary: Update prompt. description: >- + Update prompt. + Update an existing prompt (increments version). parameters: - name: prompt_id @@ -774,8 +799,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Delete a prompt. - description: Delete a prompt. + summary: Delete prompt. + description: >- + Delete prompt. + + Delete a prompt. parameters: - name: prompt_id in: path @@ -806,9 +834,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Set which version of a prompt should be the default in get_prompt (latest). + summary: Set prompt version. description: >- + Set prompt version. + Set which version of a prompt should be the default in get_prompt (latest). parameters: - name: prompt_id @@ -846,8 +875,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: List all versions of a specific prompt. - description: List all versions of a specific prompt. + summary: List prompt versions. + description: >- + List prompt versions. + + List all versions of a specific prompt. parameters: - name: prompt_id in: path @@ -879,8 +911,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: List all available providers. - description: List all available providers. + summary: List providers. + description: >- + List providers. + + List all available providers. parameters: [] deprecated: false /v1/providers/{provider_id}: @@ -905,9 +940,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: >- - Get detailed information about a specific provider. + summary: Get provider. description: >- + Get provider. + Get detailed information about a specific provider. parameters: - name: provider_id @@ -938,8 +974,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all OpenAI responses. - description: List all OpenAI responses. + summary: List all responses. + description: List all responses. parameters: - name: after in: query @@ -987,8 +1023,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all OpenAI responses. - description: List all OpenAI responses. + summary: List all responses. + description: List all responses. parameters: [] requestBody: content: @@ -1018,8 +1054,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Retrieve an OpenAI response by its ID. - description: Retrieve an OpenAI response by its ID. + summary: Get a model response. + description: Get a model response. parameters: - name: response_id in: path @@ -1049,8 +1085,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Delete an OpenAI response by its ID. - description: Delete an OpenAI response by its ID. + summary: Delete a response. + description: Delete a response. parameters: - name: response_id in: path @@ -1080,10 +1116,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: >- - List input items for a given OpenAI response. - description: >- - List input items for a given OpenAI response. + summary: List input items. + description: List input items. parameters: - name: response_id in: path @@ -1152,8 +1186,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: Run a shield. - description: Run a shield. + summary: Run shield. + description: >- + Run shield. + + Run a shield. parameters: [] requestBody: content: @@ -2709,8 +2746,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: Get the version of the service. - description: Get the version of the service. + summary: Get version. + description: >- + Get version. + + Get the version of the service. parameters: [] deprecated: false /v1beta/datasetio/append-rows/{dataset_id}: @@ -13338,9 +13378,16 @@ tags: x-displayName: >- Llama Stack Evaluation API for running evaluations on model and agent candidates. - name: Files - description: '' + description: >- + This API is used to upload documents that can be used with other Llama Stack + APIs. + x-displayName: Files - name: Inference description: >- + Llama Stack Inference API for generating completions, chat completions, and + embeddings. + + This API provides the raw interface to the underlying models. Two kinds of models are supported: @@ -13348,25 +13395,27 @@ tags: - Embedding models: these models generate embeddings to be used for semantic search. - x-displayName: >- - Llama Stack Inference API for generating completions, chat completions, and - embeddings. + x-displayName: Inference - name: Inspect - description: '' + description: >- + APIs for inspecting the Llama Stack service, including health status, available + API routes with methods and implementing providers. + x-displayName: Inspect - name: Models description: '' - name: PostTraining (Coming Soon) description: '' - name: Prompts - description: '' - x-displayName: >- + description: >- Protocol for prompt management operations. + x-displayName: Prompts - name: Providers - description: '' - x-displayName: >- + description: >- Providers API for inspecting, listing, and modifying providers and their configurations. + x-displayName: Providers - name: Safety - description: '' + description: OpenAI-compatible Moderations API. + x-displayName: Safety - name: Scoring description: '' - name: ScoringFunctions diff --git a/llama_stack/apis/agents/agents.py b/llama_stack/apis/agents/agents.py index dcd0d83d2..2816578f0 100644 --- a/llama_stack/apis/agents/agents.py +++ b/llama_stack/apis/agents/agents.py @@ -783,7 +783,7 @@ class Agents(Protocol): self, response_id: str, ) -> OpenAIResponseObject: - """Retrieve an OpenAI response by its ID. + """Get a model response. :param response_id: The ID of the OpenAI response to retrieve. :returns: An OpenAIResponseObject. @@ -806,7 +806,7 @@ class Agents(Protocol): include: list[str] | None = None, max_infer_iters: int | None = 10, # this is an extension to the OpenAI API ) -> OpenAIResponseObject | AsyncIterator[OpenAIResponseObjectStream]: - """Create a new OpenAI response. + """Create a model response. :param input: Input message(s) to create the response. :param model: The underlying LLM used for completions. @@ -825,7 +825,7 @@ class Agents(Protocol): model: str | None = None, order: Order | None = Order.desc, ) -> ListOpenAIResponseObject: - """List all OpenAI responses. + """List all responses. :param after: The ID of the last response to return. :param limit: The number of responses to return. @@ -848,7 +848,7 @@ class Agents(Protocol): limit: int | None = 20, order: Order | None = Order.desc, ) -> ListOpenAIResponseInputItem: - """List input items for a given OpenAI response. + """List input items. :param response_id: The ID of the response to retrieve input items for. :param after: An item ID to list items after, used for pagination. @@ -863,7 +863,7 @@ class Agents(Protocol): @webmethod(route="/openai/v1/responses/{response_id}", method="DELETE", level=LLAMA_STACK_API_V1, deprecated=True) @webmethod(route="/responses/{response_id}", method="DELETE", level=LLAMA_STACK_API_V1) async def delete_openai_response(self, response_id: str) -> OpenAIDeleteResponseObject: - """Delete an OpenAI response by its ID. + """Delete a response. :param response_id: The ID of the OpenAI response to delete. :returns: An OpenAIDeleteResponseObject diff --git a/llama_stack/apis/files/files.py b/llama_stack/apis/files/files.py index 13f0e95fa..f1d3764db 100644 --- a/llama_stack/apis/files/files.py +++ b/llama_stack/apis/files/files.py @@ -104,6 +104,11 @@ class OpenAIFileDeleteResponse(BaseModel): @runtime_checkable @trace_protocol class Files(Protocol): + """Files + + This API is used to upload documents that can be used with other Llama Stack APIs. + """ + # OpenAI Files API Endpoints @webmethod(route="/openai/v1/files", method="POST", level=LLAMA_STACK_API_V1, deprecated=True) @webmethod(route="/files", method="POST", level=LLAMA_STACK_API_V1) @@ -113,7 +118,8 @@ class Files(Protocol): purpose: Annotated[OpenAIFilePurpose, Form()], expires_after: Annotated[ExpiresAfter | None, Form()] = None, ) -> OpenAIFileObject: - """ + """Upload file. + Upload a file that can be used across various endpoints. The file upload should be a multipart form request with: @@ -137,7 +143,8 @@ class Files(Protocol): order: Order | None = Order.desc, purpose: OpenAIFilePurpose | None = None, ) -> ListOpenAIFileResponse: - """ + """List files. + Returns a list of files that belong to the user's organization. :param after: A cursor for use in pagination. `after` is an object ID that defines your place in the list. For instance, if you make a list request and receive 100 objects, ending with obj_foo, your subsequent call can include after=obj_foo in order to fetch the next page of the list. @@ -154,7 +161,8 @@ class Files(Protocol): self, file_id: str, ) -> OpenAIFileObject: - """ + """Retrieve file. + Returns information about a specific file. :param file_id: The ID of the file to use for this request. @@ -168,8 +176,7 @@ class Files(Protocol): self, file_id: str, ) -> OpenAIFileDeleteResponse: - """ - Delete a file. + """Delete file. :param file_id: The ID of the file to use for this request. :returns: An OpenAIFileDeleteResponse indicating successful deletion. @@ -182,7 +189,8 @@ class Files(Protocol): self, file_id: str, ) -> Response: - """ + """Retrieve file content. + Returns the contents of the specified file. :param file_id: The ID of the file to use for this request. diff --git a/llama_stack/apis/inference/inference.py b/llama_stack/apis/inference/inference.py index d71aea38e..838393187 100644 --- a/llama_stack/apis/inference/inference.py +++ b/llama_stack/apis/inference/inference.py @@ -1094,7 +1094,9 @@ class InferenceProvider(Protocol): # for fill-in-the-middle type completion suffix: str | None = None, ) -> OpenAICompletion: - """Generate an OpenAI-compatible completion for the given prompt using the specified model. + """Create completion. + + Generate an OpenAI-compatible completion for the given prompt using the specified model. :param model: The identifier of the model to use. The model must be registered with Llama Stack and available via the /models endpoint. :param prompt: The prompt to generate a completion for. @@ -1146,7 +1148,9 @@ class InferenceProvider(Protocol): top_p: float | None = None, user: str | None = None, ) -> OpenAIChatCompletion | AsyncIterator[OpenAIChatCompletionChunk]: - """Generate an OpenAI-compatible chat completion for the given messages using the specified model. + """Create chat completions. + + Generate an OpenAI-compatible chat completion for the given messages using the specified model. :param model: The identifier of the model to use. The model must be registered with Llama Stack and available via the /models endpoint. :param messages: List of messages in the conversation. @@ -1185,7 +1189,9 @@ class InferenceProvider(Protocol): dimensions: int | None = None, user: str | None = None, ) -> OpenAIEmbeddingsResponse: - """Generate OpenAI-compatible embeddings for the given input using the specified model. + """Create embeddings. + + Generate OpenAI-compatible embeddings for the given input using the specified model. :param model: The identifier of the model to use. The model must be an embedding model registered with Llama Stack and available via the /models endpoint. :param input: Input text to embed, encoded as a string or array of strings. To embed multiple inputs in a single request, pass an array of strings. @@ -1198,7 +1204,9 @@ class InferenceProvider(Protocol): class Inference(InferenceProvider): - """Llama Stack Inference API for generating completions, chat completions, and embeddings. + """Inference + + Llama Stack Inference API for generating completions, chat completions, and embeddings. This API provides the raw interface to the underlying models. Two kinds of models are supported: - LLM models: these models generate "raw" and "chat" (conversational) completions. @@ -1214,7 +1222,7 @@ class Inference(InferenceProvider): model: str | None = None, order: Order | None = Order.desc, ) -> ListOpenAIChatCompletionResponse: - """List all chat completions. + """List chat completions. :param after: The ID of the last chat completion to return. :param limit: The maximum number of chat completions to return. @@ -1229,7 +1237,9 @@ class Inference(InferenceProvider): ) @webmethod(route="/chat/completions/{completion_id}", method="GET", level=LLAMA_STACK_API_V1) async def get_chat_completion(self, completion_id: str) -> OpenAICompletionWithInputMessages: - """Describe a chat completion by its ID. + """Get chat completion. + + Describe a chat completion by its ID. :param completion_id: ID of the chat completion. :returns: A OpenAICompletionWithInputMessages. diff --git a/llama_stack/apis/inspect/inspect.py b/llama_stack/apis/inspect/inspect.py index e859dbe59..72f203621 100644 --- a/llama_stack/apis/inspect/inspect.py +++ b/llama_stack/apis/inspect/inspect.py @@ -58,9 +58,16 @@ class ListRoutesResponse(BaseModel): @runtime_checkable class Inspect(Protocol): + """Inspect + + APIs for inspecting the Llama Stack service, including health status, available API routes with methods and implementing providers. + """ + @webmethod(route="/inspect/routes", method="GET", level=LLAMA_STACK_API_V1) async def list_routes(self) -> ListRoutesResponse: - """List all available API routes with their methods and implementing providers. + """List routes. + + List all available API routes with their methods and implementing providers. :returns: Response containing information about all available routes. """ @@ -68,7 +75,9 @@ class Inspect(Protocol): @webmethod(route="/health", method="GET", level=LLAMA_STACK_API_V1) async def health(self) -> HealthInfo: - """Get the current health status of the service. + """Get health status. + + Get the current health status of the service. :returns: Health information indicating if the service is operational. """ @@ -76,7 +85,9 @@ class Inspect(Protocol): @webmethod(route="/version", method="GET", level=LLAMA_STACK_API_V1) async def version(self) -> VersionInfo: - """Get the version of the service. + """Get version. + + Get the version of the service. :returns: Version information containing the service version number. """ diff --git a/llama_stack/apis/models/models.py b/llama_stack/apis/models/models.py index 210ed9246..10949cb95 100644 --- a/llama_stack/apis/models/models.py +++ b/llama_stack/apis/models/models.py @@ -124,7 +124,9 @@ class Models(Protocol): self, model_id: str, ) -> Model: - """Get a model by its identifier. + """Get model. + + Get a model by its identifier. :param model_id: The identifier of the model to get. :returns: A Model. @@ -140,7 +142,9 @@ class Models(Protocol): metadata: dict[str, Any] | None = None, model_type: ModelType | None = None, ) -> Model: - """Register a model. + """Register model. + + Register a model. :param model_id: The identifier of the model to register. :param provider_model_id: The identifier of the model in the provider. @@ -156,7 +160,9 @@ class Models(Protocol): self, model_id: str, ) -> None: - """Unregister a model. + """Unregister model. + + Unregister a model. :param model_id: The identifier of the model to unregister. """ diff --git a/llama_stack/apis/prompts/prompts.py b/llama_stack/apis/prompts/prompts.py index c56185e25..b39c363c7 100644 --- a/llama_stack/apis/prompts/prompts.py +++ b/llama_stack/apis/prompts/prompts.py @@ -94,7 +94,9 @@ class ListPromptsResponse(BaseModel): @runtime_checkable @trace_protocol class Prompts(Protocol): - """Protocol for prompt management operations.""" + """Prompts + + Protocol for prompt management operations.""" @webmethod(route="/prompts", method="GET", level=LLAMA_STACK_API_V1) async def list_prompts(self) -> ListPromptsResponse: @@ -109,7 +111,9 @@ class Prompts(Protocol): self, prompt_id: str, ) -> ListPromptsResponse: - """List all versions of a specific prompt. + """List prompt versions. + + List all versions of a specific prompt. :param prompt_id: The identifier of the prompt to list versions for. :returns: A ListPromptsResponse containing all versions of the prompt. @@ -122,7 +126,9 @@ class Prompts(Protocol): prompt_id: str, version: int | None = None, ) -> Prompt: - """Get a prompt by its identifier and optional version. + """Get prompt. + + Get a prompt by its identifier and optional version. :param prompt_id: The identifier of the prompt to get. :param version: The version of the prompt to get (defaults to latest). @@ -136,7 +142,9 @@ class Prompts(Protocol): prompt: str, variables: list[str] | None = None, ) -> Prompt: - """Create a new prompt. + """Create prompt. + + Create a new prompt. :param prompt: The prompt text content with variable placeholders. :param variables: List of variable names that can be used in the prompt template. @@ -153,7 +161,9 @@ class Prompts(Protocol): variables: list[str] | None = None, set_as_default: bool = True, ) -> Prompt: - """Update an existing prompt (increments version). + """Update prompt. + + Update an existing prompt (increments version). :param prompt_id: The identifier of the prompt to update. :param prompt: The updated prompt text content. @@ -169,7 +179,9 @@ class Prompts(Protocol): self, prompt_id: str, ) -> None: - """Delete a prompt. + """Delete prompt. + + Delete a prompt. :param prompt_id: The identifier of the prompt to delete. """ @@ -181,7 +193,9 @@ class Prompts(Protocol): prompt_id: str, version: int, ) -> Prompt: - """Set which version of a prompt should be the default in get_prompt (latest). + """Set prompt version. + + Set which version of a prompt should be the default in get_prompt (latest). :param prompt_id: The identifier of the prompt. :param version: The version to set as default. diff --git a/llama_stack/apis/providers/providers.py b/llama_stack/apis/providers/providers.py index d1cff0f6c..e1872571d 100644 --- a/llama_stack/apis/providers/providers.py +++ b/llama_stack/apis/providers/providers.py @@ -42,13 +42,16 @@ class ListProvidersResponse(BaseModel): @runtime_checkable class Providers(Protocol): - """ + """Providers + Providers API for inspecting, listing, and modifying providers and their configurations. """ @webmethod(route="/providers", method="GET", level=LLAMA_STACK_API_V1) async def list_providers(self) -> ListProvidersResponse: - """List all available providers. + """List providers. + + List all available providers. :returns: A ListProvidersResponse containing information about all providers. """ @@ -56,7 +59,9 @@ class Providers(Protocol): @webmethod(route="/providers/{provider_id}", method="GET", level=LLAMA_STACK_API_V1) async def inspect_provider(self, provider_id: str) -> ProviderInfo: - """Get detailed information about a specific provider. + """Get provider. + + Get detailed information about a specific provider. :param provider_id: The ID of the provider to inspect. :returns: A ProviderInfo object containing the provider's details. diff --git a/llama_stack/apis/safety/safety.py b/llama_stack/apis/safety/safety.py index 0fa250d90..2ae74b0a7 100644 --- a/llama_stack/apis/safety/safety.py +++ b/llama_stack/apis/safety/safety.py @@ -96,6 +96,11 @@ class ShieldStore(Protocol): @runtime_checkable @trace_protocol class Safety(Protocol): + """Safety + + OpenAI-compatible Moderations API. + """ + shield_store: ShieldStore @webmethod(route="/safety/run-shield", method="POST", level=LLAMA_STACK_API_V1) @@ -105,7 +110,9 @@ class Safety(Protocol): messages: list[Message], params: dict[str, Any], ) -> RunShieldResponse: - """Run a shield. + """Run shield. + + Run a shield. :param shield_id: The identifier of the shield to run. :param messages: The messages to run the shield on. @@ -117,7 +124,9 @@ class Safety(Protocol): @webmethod(route="/openai/v1/moderations", method="POST", level=LLAMA_STACK_API_V1, deprecated=True) @webmethod(route="/moderations", method="POST", level=LLAMA_STACK_API_V1) async def run_moderation(self, input: str | list[str], model: str) -> ModerationObject: - """Classifies if text and/or image inputs are potentially harmful. + """Create moderation. + + Classifies if text and/or image inputs are potentially harmful. :param input: Input (or inputs) to classify. Can be a single string, an array of strings, or an array of multi-modal input objects similar to other models. :param model: The content moderation model you would like to use. From a5afdfd4fa878b2c5abcc9e108e8290b1df9a3cd Mon Sep 17 00:00:00 2001 From: Alexey Rybak Date: Thu, 2 Oct 2025 13:04:37 -0700 Subject: [PATCH 2/4] docs: API docstrings cleanup for better documentation rendering --- docs/static/deprecated-llama-stack-spec.html | 74 +++---- docs/static/deprecated-llama-stack-spec.yaml | 97 +++++---- docs/static/llama-stack-spec.html | 145 ++++++------- docs/static/llama-stack-spec.yaml | 203 ++++++++++++------- docs/static/stainless-llama-stack-spec.html | 145 ++++++------- docs/static/stainless-llama-stack-spec.yaml | 203 ++++++++++++------- llama_stack/apis/agents/agents.py | 10 +- llama_stack/apis/files/files.py | 20 +- llama_stack/apis/inference/inference.py | 22 +- llama_stack/apis/inspect/inspect.py | 17 +- llama_stack/apis/models/models.py | 12 +- llama_stack/apis/prompts/prompts.py | 28 ++- llama_stack/apis/providers/providers.py | 11 +- llama_stack/apis/safety/safety.py | 13 +- 14 files changed, 591 insertions(+), 409 deletions(-) diff --git a/docs/static/deprecated-llama-stack-spec.html b/docs/static/deprecated-llama-stack-spec.html index fe63f78bc..c34faa9e7 100644 --- a/docs/static/deprecated-llama-stack-spec.html +++ b/docs/static/deprecated-llama-stack-spec.html @@ -1443,8 +1443,8 @@ "tags": [ "Inference" ], - "summary": "List all chat completions.", - "description": "List all chat completions.", + "summary": "List chat completions.", + "description": "List chat completions.", "parameters": [ { "name": "after", @@ -1520,8 +1520,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", - "description": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + "summary": "Create chat completions.", + "description": "Create chat completions.\nGenerate an OpenAI-compatible chat completion for the given messages using the specified model.", "parameters": [], "requestBody": { "content": { @@ -1565,8 +1565,8 @@ "tags": [ "Inference" ], - "summary": "Describe a chat completion by its ID.", - "description": "Describe a chat completion by its ID.", + "summary": "Get chat completion.", + "description": "Get chat completion.\nDescribe a chat completion by its ID.", "parameters": [ { "name": "completion_id", @@ -1610,8 +1610,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", - "description": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + "summary": "Create completion.", + "description": "Create completion.\nGenerate an OpenAI-compatible completion for the given prompt using the specified model.", "parameters": [], "requestBody": { "content": { @@ -1655,8 +1655,8 @@ "tags": [ "Inference" ], - "summary": "Generate OpenAI-compatible embeddings for the given input using the specified model.", - "description": "Generate OpenAI-compatible embeddings for the given input using the specified model.", + "summary": "Create embeddings.", + "description": "Create embeddings.\nGenerate OpenAI-compatible embeddings for the given input using the specified model.", "parameters": [], "requestBody": { "content": { @@ -1700,8 +1700,8 @@ "tags": [ "Files" ], - "summary": "Returns a list of files that belong to the user's organization.", - "description": "Returns a list of files that belong to the user's organization.", + "summary": "List files.", + "description": "List files.\nReturns a list of files that belong to the user's organization.", "parameters": [ { "name": "after", @@ -1770,8 +1770,8 @@ "tags": [ "Files" ], - "summary": "Upload a file that can be used across various endpoints.", - "description": "Upload a file that can be used across various endpoints.\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", + "summary": "Upload file.", + "description": "Upload file.\nUpload a file that can be used across various endpoints.\n\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", "parameters": [], "requestBody": { "content": { @@ -1831,8 +1831,8 @@ "tags": [ "Files" ], - "summary": "Returns information about a specific file.", - "description": "Returns information about a specific file.", + "summary": "Retrieve file.", + "description": "Retrieve file.\nReturns information about a specific file.", "parameters": [ { "name": "file_id", @@ -1874,8 +1874,8 @@ "tags": [ "Files" ], - "summary": "Delete a file.", - "description": "Delete a file.", + "summary": "Delete file.", + "description": "Delete file.", "parameters": [ { "name": "file_id", @@ -1919,8 +1919,8 @@ "tags": [ "Files" ], - "summary": "Returns the contents of the specified file.", - "description": "Returns the contents of the specified file.", + "summary": "Retrieve file content.", + "description": "Retrieve file content.\nReturns the contents of the specified file.", "parameters": [ { "name": "file_id", @@ -1999,8 +1999,8 @@ "tags": [ "Safety" ], - "summary": "Classifies if text and/or image inputs are potentially harmful.", - "description": "Classifies if text and/or image inputs are potentially harmful.", + "summary": "Create moderation.", + "description": "Create moderation.\nClassifies if text and/or image inputs are potentially harmful.", "parameters": [], "requestBody": { "content": { @@ -2044,8 +2044,8 @@ "tags": [ "Agents" ], - "summary": "List all OpenAI responses.", - "description": "List all OpenAI responses.", + "summary": "List all responses.", + "description": "List all responses.", "parameters": [ { "name": "after", @@ -2119,8 +2119,8 @@ "tags": [ "Agents" ], - "summary": "Create a new OpenAI response.", - "description": "Create a new OpenAI response.", + "summary": "Create a model response.", + "description": "Create a model response.", "parameters": [], "requestBody": { "content": { @@ -2164,8 +2164,8 @@ "tags": [ "Agents" ], - "summary": "Retrieve an OpenAI response by its ID.", - "description": "Retrieve an OpenAI response by its ID.", + "summary": "Get a model response.", + "description": "Get a model response.", "parameters": [ { "name": "response_id", @@ -2207,8 +2207,8 @@ "tags": [ "Agents" ], - "summary": "Delete an OpenAI response by its ID.", - "description": "Delete an OpenAI response by its ID.", + "summary": "Delete a response.", + "description": "Delete a response.", "parameters": [ { "name": "response_id", @@ -2252,8 +2252,8 @@ "tags": [ "Agents" ], - "summary": "List input items for a given OpenAI response.", - "description": "List input items for a given OpenAI response.", + "summary": "List input items.", + "description": "List input items.", "parameters": [ { "name": "response_id", @@ -13417,12 +13417,13 @@ }, { "name": "Files", - "description": "" + "description": "This API is used to upload documents that can be used with other Llama Stack APIs.", + "x-displayName": "Files" }, { "name": "Inference", - "description": "This API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", - "x-displayName": "Llama Stack Inference API for generating completions, chat completions, and embeddings." + "description": "Llama Stack Inference API for generating completions, chat completions, and embeddings.\n\nThis API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", + "x-displayName": "Inference" }, { "name": "Models", @@ -13434,7 +13435,8 @@ }, { "name": "Safety", - "description": "" + "description": "OpenAI-compatible Moderations API.", + "x-displayName": "Safety" }, { "name": "Telemetry", diff --git a/docs/static/deprecated-llama-stack-spec.yaml b/docs/static/deprecated-llama-stack-spec.yaml index 9b1d3eff6..b9834c45e 100644 --- a/docs/static/deprecated-llama-stack-spec.yaml +++ b/docs/static/deprecated-llama-stack-spec.yaml @@ -1033,8 +1033,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: List all chat completions. - description: List all chat completions. + summary: List chat completions. + description: List chat completions. parameters: - name: after in: query @@ -1087,10 +1087,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible chat completion for the given messages using - the specified model. + summary: Create chat completions. description: >- + Create chat completions. + Generate an OpenAI-compatible chat completion for the given messages using the specified model. parameters: [] @@ -1122,8 +1122,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Describe a chat completion by its ID. - description: Describe a chat completion by its ID. + summary: Get chat completion. + description: >- + Get chat completion. + + Describe a chat completion by its ID. parameters: - name: completion_id in: path @@ -1153,10 +1156,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible completion for the given prompt using the specified - model. + summary: Create completion. description: >- + Create completion. + Generate an OpenAI-compatible completion for the given prompt using the specified model. parameters: [] @@ -1189,10 +1192,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate OpenAI-compatible embeddings for the given input using the specified - model. + summary: Create embeddings. description: >- + Create embeddings. + Generate OpenAI-compatible embeddings for the given input using the specified model. parameters: [] @@ -1225,9 +1228,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns a list of files that belong to the user's organization. + summary: List files. description: >- + List files. + Returns a list of files that belong to the user's organization. parameters: - name: after @@ -1285,11 +1289,13 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Upload a file that can be used across various endpoints. + summary: Upload file. description: >- + Upload file. + Upload a file that can be used across various endpoints. + The file upload should be a multipart form request with: - file: The File object (not file name) to be uploaded. @@ -1338,9 +1344,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns information about a specific file. + summary: Retrieve file. description: >- + Retrieve file. + Returns information about a specific file. parameters: - name: file_id @@ -1372,8 +1379,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Delete a file. - description: Delete a file. + summary: Delete file. + description: Delete file. parameters: - name: file_id in: path @@ -1405,9 +1412,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns the contents of the specified file. + summary: Retrieve file content. description: >- + Retrieve file content. + Returns the contents of the specified file. parameters: - name: file_id @@ -1464,9 +1472,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: >- - Classifies if text and/or image inputs are potentially harmful. + summary: Create moderation. description: >- + Create moderation. + Classifies if text and/or image inputs are potentially harmful. parameters: [] requestBody: @@ -1497,8 +1506,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all OpenAI responses. - description: List all OpenAI responses. + summary: List all responses. + description: List all responses. parameters: - name: after in: query @@ -1549,8 +1558,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Create a new OpenAI response. - description: Create a new OpenAI response. + summary: Create a model response. + description: Create a model response. parameters: [] requestBody: content: @@ -1580,8 +1589,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Retrieve an OpenAI response by its ID. - description: Retrieve an OpenAI response by its ID. + summary: Get a model response. + description: Get a model response. parameters: - name: response_id in: path @@ -1611,8 +1620,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Delete an OpenAI response by its ID. - description: Delete an OpenAI response by its ID. + summary: Delete a response. + description: Delete a response. parameters: - name: response_id in: path @@ -1642,10 +1651,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: >- - List input items for a given OpenAI response. - description: >- - List input items for a given OpenAI response. + summary: List input items. + description: List input items. parameters: - name: response_id in: path @@ -10036,9 +10043,16 @@ tags: x-displayName: >- Llama Stack Evaluation API for running evaluations on model and agent candidates. - name: Files - description: '' + description: >- + This API is used to upload documents that can be used with other Llama Stack + APIs. + x-displayName: Files - name: Inference description: >- + Llama Stack Inference API for generating completions, chat completions, and + embeddings. + + This API provides the raw interface to the underlying models. Two kinds of models are supported: @@ -10046,15 +10060,14 @@ tags: - Embedding models: these models generate embeddings to be used for semantic search. - x-displayName: >- - Llama Stack Inference API for generating completions, chat completions, and - embeddings. + x-displayName: Inference - name: Models description: '' - name: PostTraining (Coming Soon) description: '' - name: Safety - description: '' + description: OpenAI-compatible Moderations API. + x-displayName: Safety - name: Telemetry description: '' - name: VectorIO diff --git a/docs/static/llama-stack-spec.html b/docs/static/llama-stack-spec.html index fa16e62ee..f517f02f0 100644 --- a/docs/static/llama-stack-spec.html +++ b/docs/static/llama-stack-spec.html @@ -69,8 +69,8 @@ "tags": [ "Inference" ], - "summary": "List all chat completions.", - "description": "List all chat completions.", + "summary": "List chat completions.", + "description": "List chat completions.", "parameters": [ { "name": "after", @@ -146,8 +146,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", - "description": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + "summary": "Create chat completions.", + "description": "Create chat completions.\nGenerate an OpenAI-compatible chat completion for the given messages using the specified model.", "parameters": [], "requestBody": { "content": { @@ -191,8 +191,8 @@ "tags": [ "Inference" ], - "summary": "Describe a chat completion by its ID.", - "description": "Describe a chat completion by its ID.", + "summary": "Get chat completion.", + "description": "Get chat completion.\nDescribe a chat completion by its ID.", "parameters": [ { "name": "completion_id", @@ -236,8 +236,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", - "description": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + "summary": "Create completion.", + "description": "Create completion.\nGenerate an OpenAI-compatible completion for the given prompt using the specified model.", "parameters": [], "requestBody": { "content": { @@ -281,8 +281,8 @@ "tags": [ "Inference" ], - "summary": "Generate OpenAI-compatible embeddings for the given input using the specified model.", - "description": "Generate OpenAI-compatible embeddings for the given input using the specified model.", + "summary": "Create embeddings.", + "description": "Create embeddings.\nGenerate OpenAI-compatible embeddings for the given input using the specified model.", "parameters": [], "requestBody": { "content": { @@ -326,8 +326,8 @@ "tags": [ "Files" ], - "summary": "Returns a list of files that belong to the user's organization.", - "description": "Returns a list of files that belong to the user's organization.", + "summary": "List files.", + "description": "List files.\nReturns a list of files that belong to the user's organization.", "parameters": [ { "name": "after", @@ -396,8 +396,8 @@ "tags": [ "Files" ], - "summary": "Upload a file that can be used across various endpoints.", - "description": "Upload a file that can be used across various endpoints.\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", + "summary": "Upload file.", + "description": "Upload file.\nUpload a file that can be used across various endpoints.\n\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", "parameters": [], "requestBody": { "content": { @@ -457,8 +457,8 @@ "tags": [ "Files" ], - "summary": "Returns information about a specific file.", - "description": "Returns information about a specific file.", + "summary": "Retrieve file.", + "description": "Retrieve file.\nReturns information about a specific file.", "parameters": [ { "name": "file_id", @@ -500,8 +500,8 @@ "tags": [ "Files" ], - "summary": "Delete a file.", - "description": "Delete a file.", + "summary": "Delete file.", + "description": "Delete file.", "parameters": [ { "name": "file_id", @@ -545,8 +545,8 @@ "tags": [ "Files" ], - "summary": "Returns the contents of the specified file.", - "description": "Returns the contents of the specified file.", + "summary": "Retrieve file content.", + "description": "Retrieve file content.\nReturns the contents of the specified file.", "parameters": [ { "name": "file_id", @@ -590,8 +590,8 @@ "tags": [ "Inspect" ], - "summary": "Get the current health status of the service.", - "description": "Get the current health status of the service.", + "summary": "Get health status.", + "description": "Get health status.\nGet the current health status of the service.", "parameters": [], "deprecated": false } @@ -625,8 +625,8 @@ "tags": [ "Inspect" ], - "summary": "List all available API routes with their methods and implementing providers.", - "description": "List all available API routes with their methods and implementing providers.", + "summary": "List routes.", + "description": "List routes.\nList all available API routes with their methods and implementing providers.", "parameters": [], "deprecated": false } @@ -693,8 +693,8 @@ "tags": [ "Models" ], - "summary": "Register a model.", - "description": "Register a model.", + "summary": "Register model.", + "description": "Register model.\nRegister a model.", "parameters": [], "requestBody": { "content": { @@ -738,8 +738,8 @@ "tags": [ "Models" ], - "summary": "Get a model by its identifier.", - "description": "Get a model by its identifier.", + "summary": "Get model.", + "description": "Get model.\nGet a model by its identifier.", "parameters": [ { "name": "model_id", @@ -774,8 +774,8 @@ "tags": [ "Models" ], - "summary": "Unregister a model.", - "description": "Unregister a model.", + "summary": "Unregister model.", + "description": "Unregister model.\nUnregister a model.", "parameters": [ { "name": "model_id", @@ -819,8 +819,8 @@ "tags": [ "Safety" ], - "summary": "Classifies if text and/or image inputs are potentially harmful.", - "description": "Classifies if text and/or image inputs are potentially harmful.", + "summary": "Create moderation.", + "description": "Create moderation.\nClassifies if text and/or image inputs are potentially harmful.", "parameters": [], "requestBody": { "content": { @@ -897,8 +897,8 @@ "tags": [ "Prompts" ], - "summary": "Create a new prompt.", - "description": "Create a new prompt.", + "summary": "Create prompt.", + "description": "Create prompt.\nCreate a new prompt.", "parameters": [], "requestBody": { "content": { @@ -942,8 +942,8 @@ "tags": [ "Prompts" ], - "summary": "Get a prompt by its identifier and optional version.", - "description": "Get a prompt by its identifier and optional version.", + "summary": "Get prompt.", + "description": "Get prompt.\nGet a prompt by its identifier and optional version.", "parameters": [ { "name": "prompt_id", @@ -994,8 +994,8 @@ "tags": [ "Prompts" ], - "summary": "Update an existing prompt (increments version).", - "description": "Update an existing prompt (increments version).", + "summary": "Update prompt.", + "description": "Update prompt.\nUpdate an existing prompt (increments version).", "parameters": [ { "name": "prompt_id", @@ -1040,8 +1040,8 @@ "tags": [ "Prompts" ], - "summary": "Delete a prompt.", - "description": "Delete a prompt.", + "summary": "Delete prompt.", + "description": "Delete prompt.\nDelete a prompt.", "parameters": [ { "name": "prompt_id", @@ -1085,8 +1085,8 @@ "tags": [ "Prompts" ], - "summary": "Set which version of a prompt should be the default in get_prompt (latest).", - "description": "Set which version of a prompt should be the default in get_prompt (latest).", + "summary": "Set prompt version.", + "description": "Set prompt version.\nSet which version of a prompt should be the default in get_prompt (latest).", "parameters": [ { "name": "prompt_id", @@ -1140,8 +1140,8 @@ "tags": [ "Prompts" ], - "summary": "List all versions of a specific prompt.", - "description": "List all versions of a specific prompt.", + "summary": "List prompt versions.", + "description": "List prompt versions.\nList all versions of a specific prompt.", "parameters": [ { "name": "prompt_id", @@ -1185,8 +1185,8 @@ "tags": [ "Providers" ], - "summary": "List all available providers.", - "description": "List all available providers.", + "summary": "List providers.", + "description": "List providers.\nList all available providers.", "parameters": [], "deprecated": false } @@ -1220,8 +1220,8 @@ "tags": [ "Providers" ], - "summary": "Get detailed information about a specific provider.", - "description": "Get detailed information about a specific provider.", + "summary": "Get provider.", + "description": "Get provider.\nGet detailed information about a specific provider.", "parameters": [ { "name": "provider_id", @@ -1265,8 +1265,8 @@ "tags": [ "Agents" ], - "summary": "List all OpenAI responses.", - "description": "List all OpenAI responses.", + "summary": "List all responses.", + "description": "List all responses.", "parameters": [ { "name": "after", @@ -1340,8 +1340,8 @@ "tags": [ "Agents" ], - "summary": "Create a new OpenAI response.", - "description": "Create a new OpenAI response.", + "summary": "Create a model response.", + "description": "Create a model response.", "parameters": [], "requestBody": { "content": { @@ -1385,8 +1385,8 @@ "tags": [ "Agents" ], - "summary": "Retrieve an OpenAI response by its ID.", - "description": "Retrieve an OpenAI response by its ID.", + "summary": "Get a model response.", + "description": "Get a model response.", "parameters": [ { "name": "response_id", @@ -1428,8 +1428,8 @@ "tags": [ "Agents" ], - "summary": "Delete an OpenAI response by its ID.", - "description": "Delete an OpenAI response by its ID.", + "summary": "Delete a response.", + "description": "Delete a response.", "parameters": [ { "name": "response_id", @@ -1473,8 +1473,8 @@ "tags": [ "Agents" ], - "summary": "List input items for a given OpenAI response.", - "description": "List input items for a given OpenAI response.", + "summary": "List input items.", + "description": "List input items.", "parameters": [ { "name": "response_id", @@ -1566,8 +1566,8 @@ "tags": [ "Safety" ], - "summary": "Run a shield.", - "description": "Run a shield.", + "summary": "Run shield.", + "description": "Run shield.\nRun a shield.", "parameters": [], "requestBody": { "content": { @@ -3699,8 +3699,8 @@ "tags": [ "Inspect" ], - "summary": "Get the version of the service.", - "description": "Get the version of the service.", + "summary": "Get version.", + "description": "Get version.\nGet the version of the service.", "parameters": [], "deprecated": false } @@ -12377,16 +12377,18 @@ }, { "name": "Files", - "description": "" + "description": "This API is used to upload documents that can be used with other Llama Stack APIs.", + "x-displayName": "Files" }, { "name": "Inference", - "description": "This API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", - "x-displayName": "Llama Stack Inference API for generating completions, chat completions, and embeddings." + "description": "Llama Stack Inference API for generating completions, chat completions, and embeddings.\n\nThis API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", + "x-displayName": "Inference" }, { "name": "Inspect", - "description": "" + "description": "APIs for inspecting the Llama Stack service, including health status, available API routes with methods and implementing providers.", + "x-displayName": "Inspect" }, { "name": "Models", @@ -12394,17 +12396,18 @@ }, { "name": "Prompts", - "description": "", - "x-displayName": "Protocol for prompt management operations." + "description": "Protocol for prompt management operations.", + "x-displayName": "Prompts" }, { "name": "Providers", - "description": "", - "x-displayName": "Providers API for inspecting, listing, and modifying providers and their configurations." + "description": "Providers API for inspecting, listing, and modifying providers and their configurations.", + "x-displayName": "Providers" }, { "name": "Safety", - "description": "" + "description": "OpenAI-compatible Moderations API.", + "x-displayName": "Safety" }, { "name": "Scoring", diff --git a/docs/static/llama-stack-spec.yaml b/docs/static/llama-stack-spec.yaml index 733e2cd21..2c1d77e30 100644 --- a/docs/static/llama-stack-spec.yaml +++ b/docs/static/llama-stack-spec.yaml @@ -33,8 +33,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: List all chat completions. - description: List all chat completions. + summary: List chat completions. + description: List chat completions. parameters: - name: after in: query @@ -87,10 +87,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible chat completion for the given messages using - the specified model. + summary: Create chat completions. description: >- + Create chat completions. + Generate an OpenAI-compatible chat completion for the given messages using the specified model. parameters: [] @@ -122,8 +122,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Describe a chat completion by its ID. - description: Describe a chat completion by its ID. + summary: Get chat completion. + description: >- + Get chat completion. + + Describe a chat completion by its ID. parameters: - name: completion_id in: path @@ -153,10 +156,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible completion for the given prompt using the specified - model. + summary: Create completion. description: >- + Create completion. + Generate an OpenAI-compatible completion for the given prompt using the specified model. parameters: [] @@ -189,10 +192,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate OpenAI-compatible embeddings for the given input using the specified - model. + summary: Create embeddings. description: >- + Create embeddings. + Generate OpenAI-compatible embeddings for the given input using the specified model. parameters: [] @@ -225,9 +228,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns a list of files that belong to the user's organization. + summary: List files. description: >- + List files. + Returns a list of files that belong to the user's organization. parameters: - name: after @@ -285,11 +289,13 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Upload a file that can be used across various endpoints. + summary: Upload file. description: >- + Upload file. + Upload a file that can be used across various endpoints. + The file upload should be a multipart form request with: - file: The File object (not file name) to be uploaded. @@ -338,9 +344,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns information about a specific file. + summary: Retrieve file. description: >- + Retrieve file. + Returns information about a specific file. parameters: - name: file_id @@ -372,8 +379,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Delete a file. - description: Delete a file. + summary: Delete file. + description: Delete file. parameters: - name: file_id in: path @@ -405,9 +412,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns the contents of the specified file. + summary: Retrieve file content. description: >- + Retrieve file content. + Returns the contents of the specified file. parameters: - name: file_id @@ -440,9 +448,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: >- - Get the current health status of the service. + summary: Get health status. description: >- + Get health status. + Get the current health status of the service. parameters: [] deprecated: false @@ -468,9 +477,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: >- - List all available API routes with their methods and implementing providers. + summary: List routes. description: >- + List routes. + List all available API routes with their methods and implementing providers. parameters: [] deprecated: false @@ -519,8 +529,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Register a model. - description: Register a model. + summary: Register model. + description: >- + Register model. + + Register a model. parameters: [] requestBody: content: @@ -550,8 +563,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Get a model by its identifier. - description: Get a model by its identifier. + summary: Get model. + description: >- + Get model. + + Get a model by its identifier. parameters: - name: model_id in: path @@ -576,8 +592,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Unregister a model. - description: Unregister a model. + summary: Unregister model. + description: >- + Unregister model. + + Unregister a model. parameters: - name: model_id in: path @@ -608,9 +627,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: >- - Classifies if text and/or image inputs are potentially harmful. + summary: Create moderation. description: >- + Create moderation. + Classifies if text and/or image inputs are potentially harmful. parameters: [] requestBody: @@ -666,8 +686,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Create a new prompt. - description: Create a new prompt. + summary: Create prompt. + description: >- + Create prompt. + + Create a new prompt. parameters: [] requestBody: content: @@ -697,9 +720,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Get a prompt by its identifier and optional version. + summary: Get prompt. description: >- + Get prompt. + Get a prompt by its identifier and optional version. parameters: - name: prompt_id @@ -737,9 +761,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Update an existing prompt (increments version). + summary: Update prompt. description: >- + Update prompt. + Update an existing prompt (increments version). parameters: - name: prompt_id @@ -771,8 +796,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Delete a prompt. - description: Delete a prompt. + summary: Delete prompt. + description: >- + Delete prompt. + + Delete a prompt. parameters: - name: prompt_id in: path @@ -803,9 +831,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Set which version of a prompt should be the default in get_prompt (latest). + summary: Set prompt version. description: >- + Set prompt version. + Set which version of a prompt should be the default in get_prompt (latest). parameters: - name: prompt_id @@ -843,8 +872,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: List all versions of a specific prompt. - description: List all versions of a specific prompt. + summary: List prompt versions. + description: >- + List prompt versions. + + List all versions of a specific prompt. parameters: - name: prompt_id in: path @@ -876,8 +908,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: List all available providers. - description: List all available providers. + summary: List providers. + description: >- + List providers. + + List all available providers. parameters: [] deprecated: false /v1/providers/{provider_id}: @@ -902,9 +937,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: >- - Get detailed information about a specific provider. + summary: Get provider. description: >- + Get provider. + Get detailed information about a specific provider. parameters: - name: provider_id @@ -935,8 +971,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all OpenAI responses. - description: List all OpenAI responses. + summary: List all responses. + description: List all responses. parameters: - name: after in: query @@ -987,8 +1023,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Create a new OpenAI response. - description: Create a new OpenAI response. + summary: Create a model response. + description: Create a model response. parameters: [] requestBody: content: @@ -1018,8 +1054,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Retrieve an OpenAI response by its ID. - description: Retrieve an OpenAI response by its ID. + summary: Get a model response. + description: Get a model response. parameters: - name: response_id in: path @@ -1049,8 +1085,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Delete an OpenAI response by its ID. - description: Delete an OpenAI response by its ID. + summary: Delete a response. + description: Delete a response. parameters: - name: response_id in: path @@ -1080,10 +1116,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: >- - List input items for a given OpenAI response. - description: >- - List input items for a given OpenAI response. + summary: List input items. + description: List input items. parameters: - name: response_id in: path @@ -1152,8 +1186,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: Run a shield. - description: Run a shield. + summary: Run shield. + description: >- + Run shield. + + Run a shield. parameters: [] requestBody: content: @@ -2709,8 +2746,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: Get the version of the service. - description: Get the version of the service. + summary: Get version. + description: >- + Get version. + + Get the version of the service. parameters: [] deprecated: false jsonSchemaDialect: >- @@ -9274,9 +9314,16 @@ tags: - `background` x-displayName: Agents - name: Files - description: '' + description: >- + This API is used to upload documents that can be used with other Llama Stack + APIs. + x-displayName: Files - name: Inference description: >- + Llama Stack Inference API for generating completions, chat completions, and + embeddings. + + This API provides the raw interface to the underlying models. Two kinds of models are supported: @@ -9284,23 +9331,25 @@ tags: - Embedding models: these models generate embeddings to be used for semantic search. - x-displayName: >- - Llama Stack Inference API for generating completions, chat completions, and - embeddings. + x-displayName: Inference - name: Inspect - description: '' + description: >- + APIs for inspecting the Llama Stack service, including health status, available + API routes with methods and implementing providers. + x-displayName: Inspect - name: Models description: '' - name: Prompts - description: '' - x-displayName: >- + description: >- Protocol for prompt management operations. + x-displayName: Prompts - name: Providers - description: '' - x-displayName: >- + description: >- Providers API for inspecting, listing, and modifying providers and their configurations. + x-displayName: Providers - name: Safety - description: '' + description: OpenAI-compatible Moderations API. + x-displayName: Safety - name: Scoring description: '' - name: ScoringFunctions diff --git a/docs/static/stainless-llama-stack-spec.html b/docs/static/stainless-llama-stack-spec.html index 72ecb5bb5..4217abeb8 100644 --- a/docs/static/stainless-llama-stack-spec.html +++ b/docs/static/stainless-llama-stack-spec.html @@ -69,8 +69,8 @@ "tags": [ "Inference" ], - "summary": "List all chat completions.", - "description": "List all chat completions.", + "summary": "List chat completions.", + "description": "List chat completions.", "parameters": [ { "name": "after", @@ -146,8 +146,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", - "description": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + "summary": "Create chat completions.", + "description": "Create chat completions.\nGenerate an OpenAI-compatible chat completion for the given messages using the specified model.", "parameters": [], "requestBody": { "content": { @@ -191,8 +191,8 @@ "tags": [ "Inference" ], - "summary": "Describe a chat completion by its ID.", - "description": "Describe a chat completion by its ID.", + "summary": "Get chat completion.", + "description": "Get chat completion.\nDescribe a chat completion by its ID.", "parameters": [ { "name": "completion_id", @@ -236,8 +236,8 @@ "tags": [ "Inference" ], - "summary": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", - "description": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + "summary": "Create completion.", + "description": "Create completion.\nGenerate an OpenAI-compatible completion for the given prompt using the specified model.", "parameters": [], "requestBody": { "content": { @@ -281,8 +281,8 @@ "tags": [ "Inference" ], - "summary": "Generate OpenAI-compatible embeddings for the given input using the specified model.", - "description": "Generate OpenAI-compatible embeddings for the given input using the specified model.", + "summary": "Create embeddings.", + "description": "Create embeddings.\nGenerate OpenAI-compatible embeddings for the given input using the specified model.", "parameters": [], "requestBody": { "content": { @@ -326,8 +326,8 @@ "tags": [ "Files" ], - "summary": "Returns a list of files that belong to the user's organization.", - "description": "Returns a list of files that belong to the user's organization.", + "summary": "List files.", + "description": "List files.\nReturns a list of files that belong to the user's organization.", "parameters": [ { "name": "after", @@ -396,8 +396,8 @@ "tags": [ "Files" ], - "summary": "Upload a file that can be used across various endpoints.", - "description": "Upload a file that can be used across various endpoints.\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", + "summary": "Upload file.", + "description": "Upload file.\nUpload a file that can be used across various endpoints.\n\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", "parameters": [], "requestBody": { "content": { @@ -457,8 +457,8 @@ "tags": [ "Files" ], - "summary": "Returns information about a specific file.", - "description": "Returns information about a specific file.", + "summary": "Retrieve file.", + "description": "Retrieve file.\nReturns information about a specific file.", "parameters": [ { "name": "file_id", @@ -500,8 +500,8 @@ "tags": [ "Files" ], - "summary": "Delete a file.", - "description": "Delete a file.", + "summary": "Delete file.", + "description": "Delete file.", "parameters": [ { "name": "file_id", @@ -545,8 +545,8 @@ "tags": [ "Files" ], - "summary": "Returns the contents of the specified file.", - "description": "Returns the contents of the specified file.", + "summary": "Retrieve file content.", + "description": "Retrieve file content.\nReturns the contents of the specified file.", "parameters": [ { "name": "file_id", @@ -590,8 +590,8 @@ "tags": [ "Inspect" ], - "summary": "Get the current health status of the service.", - "description": "Get the current health status of the service.", + "summary": "Get health status.", + "description": "Get health status.\nGet the current health status of the service.", "parameters": [], "deprecated": false } @@ -625,8 +625,8 @@ "tags": [ "Inspect" ], - "summary": "List all available API routes with their methods and implementing providers.", - "description": "List all available API routes with their methods and implementing providers.", + "summary": "List routes.", + "description": "List routes.\nList all available API routes with their methods and implementing providers.", "parameters": [], "deprecated": false } @@ -693,8 +693,8 @@ "tags": [ "Models" ], - "summary": "Register a model.", - "description": "Register a model.", + "summary": "Register model.", + "description": "Register model.\nRegister a model.", "parameters": [], "requestBody": { "content": { @@ -738,8 +738,8 @@ "tags": [ "Models" ], - "summary": "Get a model by its identifier.", - "description": "Get a model by its identifier.", + "summary": "Get model.", + "description": "Get model.\nGet a model by its identifier.", "parameters": [ { "name": "model_id", @@ -774,8 +774,8 @@ "tags": [ "Models" ], - "summary": "Unregister a model.", - "description": "Unregister a model.", + "summary": "Unregister model.", + "description": "Unregister model.\nUnregister a model.", "parameters": [ { "name": "model_id", @@ -819,8 +819,8 @@ "tags": [ "Safety" ], - "summary": "Classifies if text and/or image inputs are potentially harmful.", - "description": "Classifies if text and/or image inputs are potentially harmful.", + "summary": "Create moderation.", + "description": "Create moderation.\nClassifies if text and/or image inputs are potentially harmful.", "parameters": [], "requestBody": { "content": { @@ -897,8 +897,8 @@ "tags": [ "Prompts" ], - "summary": "Create a new prompt.", - "description": "Create a new prompt.", + "summary": "Create prompt.", + "description": "Create prompt.\nCreate a new prompt.", "parameters": [], "requestBody": { "content": { @@ -942,8 +942,8 @@ "tags": [ "Prompts" ], - "summary": "Get a prompt by its identifier and optional version.", - "description": "Get a prompt by its identifier and optional version.", + "summary": "Get prompt.", + "description": "Get prompt.\nGet a prompt by its identifier and optional version.", "parameters": [ { "name": "prompt_id", @@ -994,8 +994,8 @@ "tags": [ "Prompts" ], - "summary": "Update an existing prompt (increments version).", - "description": "Update an existing prompt (increments version).", + "summary": "Update prompt.", + "description": "Update prompt.\nUpdate an existing prompt (increments version).", "parameters": [ { "name": "prompt_id", @@ -1040,8 +1040,8 @@ "tags": [ "Prompts" ], - "summary": "Delete a prompt.", - "description": "Delete a prompt.", + "summary": "Delete prompt.", + "description": "Delete prompt.\nDelete a prompt.", "parameters": [ { "name": "prompt_id", @@ -1085,8 +1085,8 @@ "tags": [ "Prompts" ], - "summary": "Set which version of a prompt should be the default in get_prompt (latest).", - "description": "Set which version of a prompt should be the default in get_prompt (latest).", + "summary": "Set prompt version.", + "description": "Set prompt version.\nSet which version of a prompt should be the default in get_prompt (latest).", "parameters": [ { "name": "prompt_id", @@ -1140,8 +1140,8 @@ "tags": [ "Prompts" ], - "summary": "List all versions of a specific prompt.", - "description": "List all versions of a specific prompt.", + "summary": "List prompt versions.", + "description": "List prompt versions.\nList all versions of a specific prompt.", "parameters": [ { "name": "prompt_id", @@ -1185,8 +1185,8 @@ "tags": [ "Providers" ], - "summary": "List all available providers.", - "description": "List all available providers.", + "summary": "List providers.", + "description": "List providers.\nList all available providers.", "parameters": [], "deprecated": false } @@ -1220,8 +1220,8 @@ "tags": [ "Providers" ], - "summary": "Get detailed information about a specific provider.", - "description": "Get detailed information about a specific provider.", + "summary": "Get provider.", + "description": "Get provider.\nGet detailed information about a specific provider.", "parameters": [ { "name": "provider_id", @@ -1265,8 +1265,8 @@ "tags": [ "Agents" ], - "summary": "List all OpenAI responses.", - "description": "List all OpenAI responses.", + "summary": "List all responses.", + "description": "List all responses.", "parameters": [ { "name": "after", @@ -1340,8 +1340,8 @@ "tags": [ "Agents" ], - "summary": "Create a new OpenAI response.", - "description": "Create a new OpenAI response.", + "summary": "Create a model response.", + "description": "Create a model response.", "parameters": [], "requestBody": { "content": { @@ -1385,8 +1385,8 @@ "tags": [ "Agents" ], - "summary": "Retrieve an OpenAI response by its ID.", - "description": "Retrieve an OpenAI response by its ID.", + "summary": "Get a model response.", + "description": "Get a model response.", "parameters": [ { "name": "response_id", @@ -1428,8 +1428,8 @@ "tags": [ "Agents" ], - "summary": "Delete an OpenAI response by its ID.", - "description": "Delete an OpenAI response by its ID.", + "summary": "Delete a response.", + "description": "Delete a response.", "parameters": [ { "name": "response_id", @@ -1473,8 +1473,8 @@ "tags": [ "Agents" ], - "summary": "List input items for a given OpenAI response.", - "description": "List input items for a given OpenAI response.", + "summary": "List input items.", + "description": "List input items.", "parameters": [ { "name": "response_id", @@ -1566,8 +1566,8 @@ "tags": [ "Safety" ], - "summary": "Run a shield.", - "description": "Run a shield.", + "summary": "Run shield.", + "description": "Run shield.\nRun a shield.", "parameters": [], "requestBody": { "content": { @@ -3699,8 +3699,8 @@ "tags": [ "Inspect" ], - "summary": "Get the version of the service.", - "description": "Get the version of the service.", + "summary": "Get version.", + "description": "Get version.\nGet the version of the service.", "parameters": [], "deprecated": false } @@ -17950,16 +17950,18 @@ }, { "name": "Files", - "description": "" + "description": "This API is used to upload documents that can be used with other Llama Stack APIs.", + "x-displayName": "Files" }, { "name": "Inference", - "description": "This API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", - "x-displayName": "Llama Stack Inference API for generating completions, chat completions, and embeddings." + "description": "Llama Stack Inference API for generating completions, chat completions, and embeddings.\n\nThis API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", + "x-displayName": "Inference" }, { "name": "Inspect", - "description": "" + "description": "APIs for inspecting the Llama Stack service, including health status, available API routes with methods and implementing providers.", + "x-displayName": "Inspect" }, { "name": "Models", @@ -17971,17 +17973,18 @@ }, { "name": "Prompts", - "description": "", - "x-displayName": "Protocol for prompt management operations." + "description": "Protocol for prompt management operations.", + "x-displayName": "Prompts" }, { "name": "Providers", - "description": "", - "x-displayName": "Providers API for inspecting, listing, and modifying providers and their configurations." + "description": "Providers API for inspecting, listing, and modifying providers and their configurations.", + "x-displayName": "Providers" }, { "name": "Safety", - "description": "" + "description": "OpenAI-compatible Moderations API.", + "x-displayName": "Safety" }, { "name": "Scoring", diff --git a/docs/static/stainless-llama-stack-spec.yaml b/docs/static/stainless-llama-stack-spec.yaml index 151ea1029..a6e659394 100644 --- a/docs/static/stainless-llama-stack-spec.yaml +++ b/docs/static/stainless-llama-stack-spec.yaml @@ -36,8 +36,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: List all chat completions. - description: List all chat completions. + summary: List chat completions. + description: List chat completions. parameters: - name: after in: query @@ -90,10 +90,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible chat completion for the given messages using - the specified model. + summary: Create chat completions. description: >- + Create chat completions. + Generate an OpenAI-compatible chat completion for the given messages using the specified model. parameters: [] @@ -125,8 +125,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Describe a chat completion by its ID. - description: Describe a chat completion by its ID. + summary: Get chat completion. + description: >- + Get chat completion. + + Describe a chat completion by its ID. parameters: - name: completion_id in: path @@ -156,10 +159,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate an OpenAI-compatible completion for the given prompt using the specified - model. + summary: Create completion. description: >- + Create completion. + Generate an OpenAI-compatible completion for the given prompt using the specified model. parameters: [] @@ -192,10 +195,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: >- - Generate OpenAI-compatible embeddings for the given input using the specified - model. + summary: Create embeddings. description: >- + Create embeddings. + Generate OpenAI-compatible embeddings for the given input using the specified model. parameters: [] @@ -228,9 +231,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns a list of files that belong to the user's organization. + summary: List files. description: >- + List files. + Returns a list of files that belong to the user's organization. parameters: - name: after @@ -288,11 +292,13 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Upload a file that can be used across various endpoints. + summary: Upload file. description: >- + Upload file. + Upload a file that can be used across various endpoints. + The file upload should be a multipart form request with: - file: The File object (not file name) to be uploaded. @@ -341,9 +347,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns information about a specific file. + summary: Retrieve file. description: >- + Retrieve file. + Returns information about a specific file. parameters: - name: file_id @@ -375,8 +382,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Delete a file. - description: Delete a file. + summary: Delete file. + description: Delete file. parameters: - name: file_id in: path @@ -408,9 +415,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: >- - Returns the contents of the specified file. + summary: Retrieve file content. description: >- + Retrieve file content. + Returns the contents of the specified file. parameters: - name: file_id @@ -443,9 +451,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: >- - Get the current health status of the service. + summary: Get health status. description: >- + Get health status. + Get the current health status of the service. parameters: [] deprecated: false @@ -471,9 +480,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: >- - List all available API routes with their methods and implementing providers. + summary: List routes. description: >- + List routes. + List all available API routes with their methods and implementing providers. parameters: [] deprecated: false @@ -522,8 +532,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Register a model. - description: Register a model. + summary: Register model. + description: >- + Register model. + + Register a model. parameters: [] requestBody: content: @@ -553,8 +566,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Get a model by its identifier. - description: Get a model by its identifier. + summary: Get model. + description: >- + Get model. + + Get a model by its identifier. parameters: - name: model_id in: path @@ -579,8 +595,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Unregister a model. - description: Unregister a model. + summary: Unregister model. + description: >- + Unregister model. + + Unregister a model. parameters: - name: model_id in: path @@ -611,9 +630,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: >- - Classifies if text and/or image inputs are potentially harmful. + summary: Create moderation. description: >- + Create moderation. + Classifies if text and/or image inputs are potentially harmful. parameters: [] requestBody: @@ -669,8 +689,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Create a new prompt. - description: Create a new prompt. + summary: Create prompt. + description: >- + Create prompt. + + Create a new prompt. parameters: [] requestBody: content: @@ -700,9 +723,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Get a prompt by its identifier and optional version. + summary: Get prompt. description: >- + Get prompt. + Get a prompt by its identifier and optional version. parameters: - name: prompt_id @@ -740,9 +764,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Update an existing prompt (increments version). + summary: Update prompt. description: >- + Update prompt. + Update an existing prompt (increments version). parameters: - name: prompt_id @@ -774,8 +799,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Delete a prompt. - description: Delete a prompt. + summary: Delete prompt. + description: >- + Delete prompt. + + Delete a prompt. parameters: - name: prompt_id in: path @@ -806,9 +834,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: >- - Set which version of a prompt should be the default in get_prompt (latest). + summary: Set prompt version. description: >- + Set prompt version. + Set which version of a prompt should be the default in get_prompt (latest). parameters: - name: prompt_id @@ -846,8 +875,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: List all versions of a specific prompt. - description: List all versions of a specific prompt. + summary: List prompt versions. + description: >- + List prompt versions. + + List all versions of a specific prompt. parameters: - name: prompt_id in: path @@ -879,8 +911,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: List all available providers. - description: List all available providers. + summary: List providers. + description: >- + List providers. + + List all available providers. parameters: [] deprecated: false /v1/providers/{provider_id}: @@ -905,9 +940,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: >- - Get detailed information about a specific provider. + summary: Get provider. description: >- + Get provider. + Get detailed information about a specific provider. parameters: - name: provider_id @@ -938,8 +974,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all OpenAI responses. - description: List all OpenAI responses. + summary: List all responses. + description: List all responses. parameters: - name: after in: query @@ -990,8 +1026,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Create a new OpenAI response. - description: Create a new OpenAI response. + summary: Create a model response. + description: Create a model response. parameters: [] requestBody: content: @@ -1021,8 +1057,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Retrieve an OpenAI response by its ID. - description: Retrieve an OpenAI response by its ID. + summary: Get a model response. + description: Get a model response. parameters: - name: response_id in: path @@ -1052,8 +1088,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Delete an OpenAI response by its ID. - description: Delete an OpenAI response by its ID. + summary: Delete a response. + description: Delete a response. parameters: - name: response_id in: path @@ -1083,10 +1119,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: >- - List input items for a given OpenAI response. - description: >- - List input items for a given OpenAI response. + summary: List input items. + description: List input items. parameters: - name: response_id in: path @@ -1155,8 +1189,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: Run a shield. - description: Run a shield. + summary: Run shield. + description: >- + Run shield. + + Run a shield. parameters: [] requestBody: content: @@ -2712,8 +2749,11 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: Get the version of the service. - description: Get the version of the service. + summary: Get version. + description: >- + Get version. + + Get the version of the service. parameters: [] deprecated: false /v1beta/datasetio/append-rows/{dataset_id}: @@ -13320,9 +13360,16 @@ tags: x-displayName: >- Llama Stack Evaluation API for running evaluations on model and agent candidates. - name: Files - description: '' + description: >- + This API is used to upload documents that can be used with other Llama Stack + APIs. + x-displayName: Files - name: Inference description: >- + Llama Stack Inference API for generating completions, chat completions, and + embeddings. + + This API provides the raw interface to the underlying models. Two kinds of models are supported: @@ -13330,25 +13377,27 @@ tags: - Embedding models: these models generate embeddings to be used for semantic search. - x-displayName: >- - Llama Stack Inference API for generating completions, chat completions, and - embeddings. + x-displayName: Inference - name: Inspect - description: '' + description: >- + APIs for inspecting the Llama Stack service, including health status, available + API routes with methods and implementing providers. + x-displayName: Inspect - name: Models description: '' - name: PostTraining (Coming Soon) description: '' - name: Prompts - description: '' - x-displayName: >- + description: >- Protocol for prompt management operations. + x-displayName: Prompts - name: Providers - description: '' - x-displayName: >- + description: >- Providers API for inspecting, listing, and modifying providers and their configurations. + x-displayName: Providers - name: Safety - description: '' + description: OpenAI-compatible Moderations API. + x-displayName: Safety - name: Scoring description: '' - name: ScoringFunctions diff --git a/llama_stack/apis/agents/agents.py b/llama_stack/apis/agents/agents.py index 811fe6aa2..84a4ee905 100644 --- a/llama_stack/apis/agents/agents.py +++ b/llama_stack/apis/agents/agents.py @@ -783,7 +783,7 @@ class Agents(Protocol): self, response_id: str, ) -> OpenAIResponseObject: - """Retrieve an OpenAI response by its ID. + """Get a model response. :param response_id: The ID of the OpenAI response to retrieve. :returns: An OpenAIResponseObject. @@ -806,7 +806,7 @@ class Agents(Protocol): include: list[str] | None = None, max_infer_iters: int | None = 10, # this is an extension to the OpenAI API ) -> OpenAIResponseObject | AsyncIterator[OpenAIResponseObjectStream]: - """Create a new OpenAI response. + """Create a model response. :param input: Input message(s) to create the response. :param model: The underlying LLM used for completions. @@ -825,7 +825,7 @@ class Agents(Protocol): model: str | None = None, order: Order | None = Order.desc, ) -> ListOpenAIResponseObject: - """List all OpenAI responses. + """List all responses. :param after: The ID of the last response to return. :param limit: The number of responses to return. @@ -848,7 +848,7 @@ class Agents(Protocol): limit: int | None = 20, order: Order | None = Order.desc, ) -> ListOpenAIResponseInputItem: - """List input items for a given OpenAI response. + """List input items. :param response_id: The ID of the response to retrieve input items for. :param after: An item ID to list items after, used for pagination. @@ -863,7 +863,7 @@ class Agents(Protocol): @webmethod(route="/openai/v1/responses/{response_id}", method="DELETE", level=LLAMA_STACK_API_V1, deprecated=True) @webmethod(route="/responses/{response_id}", method="DELETE", level=LLAMA_STACK_API_V1) async def delete_openai_response(self, response_id: str) -> OpenAIDeleteResponseObject: - """Delete an OpenAI response by its ID. + """Delete a response. :param response_id: The ID of the OpenAI response to delete. :returns: An OpenAIDeleteResponseObject diff --git a/llama_stack/apis/files/files.py b/llama_stack/apis/files/files.py index 13f0e95fa..f1d3764db 100644 --- a/llama_stack/apis/files/files.py +++ b/llama_stack/apis/files/files.py @@ -104,6 +104,11 @@ class OpenAIFileDeleteResponse(BaseModel): @runtime_checkable @trace_protocol class Files(Protocol): + """Files + + This API is used to upload documents that can be used with other Llama Stack APIs. + """ + # OpenAI Files API Endpoints @webmethod(route="/openai/v1/files", method="POST", level=LLAMA_STACK_API_V1, deprecated=True) @webmethod(route="/files", method="POST", level=LLAMA_STACK_API_V1) @@ -113,7 +118,8 @@ class Files(Protocol): purpose: Annotated[OpenAIFilePurpose, Form()], expires_after: Annotated[ExpiresAfter | None, Form()] = None, ) -> OpenAIFileObject: - """ + """Upload file. + Upload a file that can be used across various endpoints. The file upload should be a multipart form request with: @@ -137,7 +143,8 @@ class Files(Protocol): order: Order | None = Order.desc, purpose: OpenAIFilePurpose | None = None, ) -> ListOpenAIFileResponse: - """ + """List files. + Returns a list of files that belong to the user's organization. :param after: A cursor for use in pagination. `after` is an object ID that defines your place in the list. For instance, if you make a list request and receive 100 objects, ending with obj_foo, your subsequent call can include after=obj_foo in order to fetch the next page of the list. @@ -154,7 +161,8 @@ class Files(Protocol): self, file_id: str, ) -> OpenAIFileObject: - """ + """Retrieve file. + Returns information about a specific file. :param file_id: The ID of the file to use for this request. @@ -168,8 +176,7 @@ class Files(Protocol): self, file_id: str, ) -> OpenAIFileDeleteResponse: - """ - Delete a file. + """Delete file. :param file_id: The ID of the file to use for this request. :returns: An OpenAIFileDeleteResponse indicating successful deletion. @@ -182,7 +189,8 @@ class Files(Protocol): self, file_id: str, ) -> Response: - """ + """Retrieve file content. + Returns the contents of the specified file. :param file_id: The ID of the file to use for this request. diff --git a/llama_stack/apis/inference/inference.py b/llama_stack/apis/inference/inference.py index d71aea38e..838393187 100644 --- a/llama_stack/apis/inference/inference.py +++ b/llama_stack/apis/inference/inference.py @@ -1094,7 +1094,9 @@ class InferenceProvider(Protocol): # for fill-in-the-middle type completion suffix: str | None = None, ) -> OpenAICompletion: - """Generate an OpenAI-compatible completion for the given prompt using the specified model. + """Create completion. + + Generate an OpenAI-compatible completion for the given prompt using the specified model. :param model: The identifier of the model to use. The model must be registered with Llama Stack and available via the /models endpoint. :param prompt: The prompt to generate a completion for. @@ -1146,7 +1148,9 @@ class InferenceProvider(Protocol): top_p: float | None = None, user: str | None = None, ) -> OpenAIChatCompletion | AsyncIterator[OpenAIChatCompletionChunk]: - """Generate an OpenAI-compatible chat completion for the given messages using the specified model. + """Create chat completions. + + Generate an OpenAI-compatible chat completion for the given messages using the specified model. :param model: The identifier of the model to use. The model must be registered with Llama Stack and available via the /models endpoint. :param messages: List of messages in the conversation. @@ -1185,7 +1189,9 @@ class InferenceProvider(Protocol): dimensions: int | None = None, user: str | None = None, ) -> OpenAIEmbeddingsResponse: - """Generate OpenAI-compatible embeddings for the given input using the specified model. + """Create embeddings. + + Generate OpenAI-compatible embeddings for the given input using the specified model. :param model: The identifier of the model to use. The model must be an embedding model registered with Llama Stack and available via the /models endpoint. :param input: Input text to embed, encoded as a string or array of strings. To embed multiple inputs in a single request, pass an array of strings. @@ -1198,7 +1204,9 @@ class InferenceProvider(Protocol): class Inference(InferenceProvider): - """Llama Stack Inference API for generating completions, chat completions, and embeddings. + """Inference + + Llama Stack Inference API for generating completions, chat completions, and embeddings. This API provides the raw interface to the underlying models. Two kinds of models are supported: - LLM models: these models generate "raw" and "chat" (conversational) completions. @@ -1214,7 +1222,7 @@ class Inference(InferenceProvider): model: str | None = None, order: Order | None = Order.desc, ) -> ListOpenAIChatCompletionResponse: - """List all chat completions. + """List chat completions. :param after: The ID of the last chat completion to return. :param limit: The maximum number of chat completions to return. @@ -1229,7 +1237,9 @@ class Inference(InferenceProvider): ) @webmethod(route="/chat/completions/{completion_id}", method="GET", level=LLAMA_STACK_API_V1) async def get_chat_completion(self, completion_id: str) -> OpenAICompletionWithInputMessages: - """Describe a chat completion by its ID. + """Get chat completion. + + Describe a chat completion by its ID. :param completion_id: ID of the chat completion. :returns: A OpenAICompletionWithInputMessages. diff --git a/llama_stack/apis/inspect/inspect.py b/llama_stack/apis/inspect/inspect.py index e859dbe59..72f203621 100644 --- a/llama_stack/apis/inspect/inspect.py +++ b/llama_stack/apis/inspect/inspect.py @@ -58,9 +58,16 @@ class ListRoutesResponse(BaseModel): @runtime_checkable class Inspect(Protocol): + """Inspect + + APIs for inspecting the Llama Stack service, including health status, available API routes with methods and implementing providers. + """ + @webmethod(route="/inspect/routes", method="GET", level=LLAMA_STACK_API_V1) async def list_routes(self) -> ListRoutesResponse: - """List all available API routes with their methods and implementing providers. + """List routes. + + List all available API routes with their methods and implementing providers. :returns: Response containing information about all available routes. """ @@ -68,7 +75,9 @@ class Inspect(Protocol): @webmethod(route="/health", method="GET", level=LLAMA_STACK_API_V1) async def health(self) -> HealthInfo: - """Get the current health status of the service. + """Get health status. + + Get the current health status of the service. :returns: Health information indicating if the service is operational. """ @@ -76,7 +85,9 @@ class Inspect(Protocol): @webmethod(route="/version", method="GET", level=LLAMA_STACK_API_V1) async def version(self) -> VersionInfo: - """Get the version of the service. + """Get version. + + Get the version of the service. :returns: Version information containing the service version number. """ diff --git a/llama_stack/apis/models/models.py b/llama_stack/apis/models/models.py index 210ed9246..10949cb95 100644 --- a/llama_stack/apis/models/models.py +++ b/llama_stack/apis/models/models.py @@ -124,7 +124,9 @@ class Models(Protocol): self, model_id: str, ) -> Model: - """Get a model by its identifier. + """Get model. + + Get a model by its identifier. :param model_id: The identifier of the model to get. :returns: A Model. @@ -140,7 +142,9 @@ class Models(Protocol): metadata: dict[str, Any] | None = None, model_type: ModelType | None = None, ) -> Model: - """Register a model. + """Register model. + + Register a model. :param model_id: The identifier of the model to register. :param provider_model_id: The identifier of the model in the provider. @@ -156,7 +160,9 @@ class Models(Protocol): self, model_id: str, ) -> None: - """Unregister a model. + """Unregister model. + + Unregister a model. :param model_id: The identifier of the model to unregister. """ diff --git a/llama_stack/apis/prompts/prompts.py b/llama_stack/apis/prompts/prompts.py index c56185e25..b39c363c7 100644 --- a/llama_stack/apis/prompts/prompts.py +++ b/llama_stack/apis/prompts/prompts.py @@ -94,7 +94,9 @@ class ListPromptsResponse(BaseModel): @runtime_checkable @trace_protocol class Prompts(Protocol): - """Protocol for prompt management operations.""" + """Prompts + + Protocol for prompt management operations.""" @webmethod(route="/prompts", method="GET", level=LLAMA_STACK_API_V1) async def list_prompts(self) -> ListPromptsResponse: @@ -109,7 +111,9 @@ class Prompts(Protocol): self, prompt_id: str, ) -> ListPromptsResponse: - """List all versions of a specific prompt. + """List prompt versions. + + List all versions of a specific prompt. :param prompt_id: The identifier of the prompt to list versions for. :returns: A ListPromptsResponse containing all versions of the prompt. @@ -122,7 +126,9 @@ class Prompts(Protocol): prompt_id: str, version: int | None = None, ) -> Prompt: - """Get a prompt by its identifier and optional version. + """Get prompt. + + Get a prompt by its identifier and optional version. :param prompt_id: The identifier of the prompt to get. :param version: The version of the prompt to get (defaults to latest). @@ -136,7 +142,9 @@ class Prompts(Protocol): prompt: str, variables: list[str] | None = None, ) -> Prompt: - """Create a new prompt. + """Create prompt. + + Create a new prompt. :param prompt: The prompt text content with variable placeholders. :param variables: List of variable names that can be used in the prompt template. @@ -153,7 +161,9 @@ class Prompts(Protocol): variables: list[str] | None = None, set_as_default: bool = True, ) -> Prompt: - """Update an existing prompt (increments version). + """Update prompt. + + Update an existing prompt (increments version). :param prompt_id: The identifier of the prompt to update. :param prompt: The updated prompt text content. @@ -169,7 +179,9 @@ class Prompts(Protocol): self, prompt_id: str, ) -> None: - """Delete a prompt. + """Delete prompt. + + Delete a prompt. :param prompt_id: The identifier of the prompt to delete. """ @@ -181,7 +193,9 @@ class Prompts(Protocol): prompt_id: str, version: int, ) -> Prompt: - """Set which version of a prompt should be the default in get_prompt (latest). + """Set prompt version. + + Set which version of a prompt should be the default in get_prompt (latest). :param prompt_id: The identifier of the prompt. :param version: The version to set as default. diff --git a/llama_stack/apis/providers/providers.py b/llama_stack/apis/providers/providers.py index d1cff0f6c..e1872571d 100644 --- a/llama_stack/apis/providers/providers.py +++ b/llama_stack/apis/providers/providers.py @@ -42,13 +42,16 @@ class ListProvidersResponse(BaseModel): @runtime_checkable class Providers(Protocol): - """ + """Providers + Providers API for inspecting, listing, and modifying providers and their configurations. """ @webmethod(route="/providers", method="GET", level=LLAMA_STACK_API_V1) async def list_providers(self) -> ListProvidersResponse: - """List all available providers. + """List providers. + + List all available providers. :returns: A ListProvidersResponse containing information about all providers. """ @@ -56,7 +59,9 @@ class Providers(Protocol): @webmethod(route="/providers/{provider_id}", method="GET", level=LLAMA_STACK_API_V1) async def inspect_provider(self, provider_id: str) -> ProviderInfo: - """Get detailed information about a specific provider. + """Get provider. + + Get detailed information about a specific provider. :param provider_id: The ID of the provider to inspect. :returns: A ProviderInfo object containing the provider's details. diff --git a/llama_stack/apis/safety/safety.py b/llama_stack/apis/safety/safety.py index 0fa250d90..2ae74b0a7 100644 --- a/llama_stack/apis/safety/safety.py +++ b/llama_stack/apis/safety/safety.py @@ -96,6 +96,11 @@ class ShieldStore(Protocol): @runtime_checkable @trace_protocol class Safety(Protocol): + """Safety + + OpenAI-compatible Moderations API. + """ + shield_store: ShieldStore @webmethod(route="/safety/run-shield", method="POST", level=LLAMA_STACK_API_V1) @@ -105,7 +110,9 @@ class Safety(Protocol): messages: list[Message], params: dict[str, Any], ) -> RunShieldResponse: - """Run a shield. + """Run shield. + + Run a shield. :param shield_id: The identifier of the shield to run. :param messages: The messages to run the shield on. @@ -117,7 +124,9 @@ class Safety(Protocol): @webmethod(route="/openai/v1/moderations", method="POST", level=LLAMA_STACK_API_V1, deprecated=True) @webmethod(route="/moderations", method="POST", level=LLAMA_STACK_API_V1) async def run_moderation(self, input: str | list[str], model: str) -> ModerationObject: - """Classifies if text and/or image inputs are potentially harmful. + """Create moderation. + + Classifies if text and/or image inputs are potentially harmful. :param input: Input (or inputs) to classify. Can be a single string, an array of strings, or an array of multi-modal input objects similar to other models. :param model: The content moderation model you would like to use. From 86aee921116849041b7ce7f36afb6ff483826683 Mon Sep 17 00:00:00 2001 From: Alexey Rybak Date: Thu, 2 Oct 2025 13:44:53 -0700 Subject: [PATCH 3/4] docs: fix API specs --- docs/static/deprecated-llama-stack-spec.html | 6 ++---- docs/static/deprecated-llama-stack-spec.yaml | 4 ---- docs/static/llama-stack-spec.html | 6 ++---- docs/static/llama-stack-spec.yaml | 6 ++---- docs/static/stainless-llama-stack-spec.html | 6 ++---- docs/static/stainless-llama-stack-spec.yaml | 6 ++---- 6 files changed, 10 insertions(+), 24 deletions(-) diff --git a/docs/static/deprecated-llama-stack-spec.html b/docs/static/deprecated-llama-stack-spec.html index 7e6ad5ad4..c34faa9e7 100644 --- a/docs/static/deprecated-llama-stack-spec.html +++ b/docs/static/deprecated-llama-stack-spec.html @@ -1445,8 +1445,6 @@ ], "summary": "List chat completions.", "description": "List chat completions.", - "summary": "List chat completions.", - "description": "List chat completions.", "parameters": [ { "name": "after", @@ -2121,8 +2119,8 @@ "tags": [ "Agents" ], - "summary": "List all responses.", - "description": "List all responses.", + "summary": "Create a model response.", + "description": "Create a model response.", "parameters": [], "requestBody": { "content": { diff --git a/docs/static/deprecated-llama-stack-spec.yaml b/docs/static/deprecated-llama-stack-spec.yaml index 13b47da7f..b9834c45e 100644 --- a/docs/static/deprecated-llama-stack-spec.yaml +++ b/docs/static/deprecated-llama-stack-spec.yaml @@ -1035,8 +1035,6 @@ paths: - Inference summary: List chat completions. description: List chat completions. - summary: List chat completions. - description: List chat completions. parameters: - name: after in: query @@ -1560,8 +1558,6 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all responses. - description: List all responses. summary: Create a model response. description: Create a model response. parameters: [] diff --git a/docs/static/llama-stack-spec.html b/docs/static/llama-stack-spec.html index eb761cacf..f517f02f0 100644 --- a/docs/static/llama-stack-spec.html +++ b/docs/static/llama-stack-spec.html @@ -71,8 +71,6 @@ ], "summary": "List chat completions.", "description": "List chat completions.", - "summary": "List chat completions.", - "description": "List chat completions.", "parameters": [ { "name": "after", @@ -1342,8 +1340,8 @@ "tags": [ "Agents" ], - "summary": "List all responses.", - "description": "List all responses.", + "summary": "Create a model response.", + "description": "Create a model response.", "parameters": [], "requestBody": { "content": { diff --git a/docs/static/llama-stack-spec.yaml b/docs/static/llama-stack-spec.yaml index 583e6c1d9..2c1d77e30 100644 --- a/docs/static/llama-stack-spec.yaml +++ b/docs/static/llama-stack-spec.yaml @@ -35,8 +35,6 @@ paths: - Inference summary: List chat completions. description: List chat completions. - summary: List chat completions. - description: List chat completions. parameters: - name: after in: query @@ -1025,8 +1023,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all responses. - description: List all responses. + summary: Create a model response. + description: Create a model response. parameters: [] requestBody: content: diff --git a/docs/static/stainless-llama-stack-spec.html b/docs/static/stainless-llama-stack-spec.html index 6e1448961..4217abeb8 100644 --- a/docs/static/stainless-llama-stack-spec.html +++ b/docs/static/stainless-llama-stack-spec.html @@ -71,8 +71,6 @@ ], "summary": "List chat completions.", "description": "List chat completions.", - "summary": "List chat completions.", - "description": "List chat completions.", "parameters": [ { "name": "after", @@ -1342,8 +1340,8 @@ "tags": [ "Agents" ], - "summary": "List all responses.", - "description": "List all responses.", + "summary": "Create a model response.", + "description": "Create a model response.", "parameters": [], "requestBody": { "content": { diff --git a/docs/static/stainless-llama-stack-spec.yaml b/docs/static/stainless-llama-stack-spec.yaml index 2c7fb8f92..a6e659394 100644 --- a/docs/static/stainless-llama-stack-spec.yaml +++ b/docs/static/stainless-llama-stack-spec.yaml @@ -38,8 +38,6 @@ paths: - Inference summary: List chat completions. description: List chat completions. - summary: List chat completions. - description: List chat completions. parameters: - name: after in: query @@ -1028,8 +1026,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all responses. - description: List all responses. + summary: Create a model response. + description: Create a model response. parameters: [] requestBody: content: From b35b546a2e120f78e77c42711becf7094d7bc3b3 Mon Sep 17 00:00:00 2001 From: Alexey Rybak Date: Thu, 2 Oct 2025 13:50:47 -0700 Subject: [PATCH 4/4] docs: provider formatting updates --- docs/docs/providers/agents/index.mdx | 4 ++-- docs/docs/providers/batches/index.mdx | 24 ++++++++++++------------ docs/docs/providers/files/index.mdx | 7 +++++++ docs/docs/providers/inference/index.mdx | 20 ++++++++++++-------- docs/docs/providers/safety/index.mdx | 7 +++++++ 5 files changed, 40 insertions(+), 22 deletions(-) diff --git a/docs/docs/providers/agents/index.mdx b/docs/docs/providers/agents/index.mdx index 52b92734e..06eb104af 100644 --- a/docs/docs/providers/agents/index.mdx +++ b/docs/docs/providers/agents/index.mdx @@ -1,7 +1,7 @@ --- description: "Agents -APIs for creating and interacting with agentic systems." + APIs for creating and interacting with agentic systems." sidebar_label: Agents title: Agents --- @@ -12,6 +12,6 @@ title: Agents Agents -APIs for creating and interacting with agentic systems. + APIs for creating and interacting with agentic systems. This section contains documentation for all available providers for the **agents** API. diff --git a/docs/docs/providers/batches/index.mdx b/docs/docs/providers/batches/index.mdx index 18e5e314d..2c64b277f 100644 --- a/docs/docs/providers/batches/index.mdx +++ b/docs/docs/providers/batches/index.mdx @@ -1,14 +1,14 @@ --- description: "The Batches API enables efficient processing of multiple requests in a single operation, -particularly useful for processing large datasets, batch evaluation workflows, and -cost-effective inference at scale. + particularly useful for processing large datasets, batch evaluation workflows, and + cost-effective inference at scale. -The API is designed to allow use of openai client libraries for seamless integration. + The API is designed to allow use of openai client libraries for seamless integration. -This API provides the following extensions: - - idempotent batch creation + This API provides the following extensions: + - idempotent batch creation -Note: This API is currently under active development and may undergo changes." + Note: This API is currently under active development and may undergo changes." sidebar_label: Batches title: Batches --- @@ -18,14 +18,14 @@ title: Batches ## Overview The Batches API enables efficient processing of multiple requests in a single operation, -particularly useful for processing large datasets, batch evaluation workflows, and -cost-effective inference at scale. + particularly useful for processing large datasets, batch evaluation workflows, and + cost-effective inference at scale. -The API is designed to allow use of openai client libraries for seamless integration. + The API is designed to allow use of openai client libraries for seamless integration. -This API provides the following extensions: - - idempotent batch creation + This API provides the following extensions: + - idempotent batch creation -Note: This API is currently under active development and may undergo changes. + Note: This API is currently under active development and may undergo changes. This section contains documentation for all available providers for the **batches** API. diff --git a/docs/docs/providers/files/index.mdx b/docs/docs/providers/files/index.mdx index 7d729d90f..19e338035 100644 --- a/docs/docs/providers/files/index.mdx +++ b/docs/docs/providers/files/index.mdx @@ -1,4 +1,7 @@ --- +description: "Files + + This API is used to upload documents that can be used with other Llama Stack APIs." sidebar_label: Files title: Files --- @@ -7,4 +10,8 @@ title: Files ## Overview +Files + + This API is used to upload documents that can be used with other Llama Stack APIs. + This section contains documentation for all available providers for the **files** API. diff --git a/docs/docs/providers/inference/index.mdx b/docs/docs/providers/inference/index.mdx index 1dc479675..c2bf69962 100644 --- a/docs/docs/providers/inference/index.mdx +++ b/docs/docs/providers/inference/index.mdx @@ -1,9 +1,11 @@ --- -description: "Llama Stack Inference API for generating completions, chat completions, and embeddings. +description: "Inference -This API provides the raw interface to the underlying models. Two kinds of models are supported: -- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions. -- Embedding models: these models generate embeddings to be used for semantic search." + Llama Stack Inference API for generating completions, chat completions, and embeddings. + + This API provides the raw interface to the underlying models. Two kinds of models are supported: + - LLM models: these models generate \"raw\" and \"chat\" (conversational) completions. + - Embedding models: these models generate embeddings to be used for semantic search." sidebar_label: Inference title: Inference --- @@ -12,10 +14,12 @@ title: Inference ## Overview -Llama Stack Inference API for generating completions, chat completions, and embeddings. +Inference -This API provides the raw interface to the underlying models. Two kinds of models are supported: -- LLM models: these models generate "raw" and "chat" (conversational) completions. -- Embedding models: these models generate embeddings to be used for semantic search. + Llama Stack Inference API for generating completions, chat completions, and embeddings. + + This API provides the raw interface to the underlying models. Two kinds of models are supported: + - LLM models: these models generate "raw" and "chat" (conversational) completions. + - Embedding models: these models generate embeddings to be used for semantic search. This section contains documentation for all available providers for the **inference** API. diff --git a/docs/docs/providers/safety/index.mdx b/docs/docs/providers/safety/index.mdx index 3445b17e6..4e2de4f33 100644 --- a/docs/docs/providers/safety/index.mdx +++ b/docs/docs/providers/safety/index.mdx @@ -1,4 +1,7 @@ --- +description: "Safety + + OpenAI-compatible Moderations API." sidebar_label: Safety title: Safety --- @@ -7,4 +10,8 @@ title: Safety ## Overview +Safety + + OpenAI-compatible Moderations API. + This section contains documentation for all available providers for the **safety** API.