diff --git a/docs/docs/providers/files/index.mdx b/docs/docs/providers/files/index.mdx index 19e338035..7d729d90f 100644 --- a/docs/docs/providers/files/index.mdx +++ b/docs/docs/providers/files/index.mdx @@ -1,7 +1,4 @@ --- -description: "Files - - This API is used to upload documents that can be used with other Llama Stack APIs." sidebar_label: Files title: Files --- @@ -10,8 +7,4 @@ title: Files ## Overview -Files - - This API is used to upload documents that can be used with other Llama Stack APIs. - This section contains documentation for all available providers for the **files** API. diff --git a/docs/docs/providers/inference/index.mdx b/docs/docs/providers/inference/index.mdx index c2bf69962..ebbaf1be1 100644 --- a/docs/docs/providers/inference/index.mdx +++ b/docs/docs/providers/inference/index.mdx @@ -1,7 +1,5 @@ --- -description: "Inference - - Llama Stack Inference API for generating completions, chat completions, and embeddings. +description: "Llama Stack Inference API for generating completions, chat completions, and embeddings. This API provides the raw interface to the underlying models. Two kinds of models are supported: - LLM models: these models generate \"raw\" and \"chat\" (conversational) completions. @@ -14,9 +12,7 @@ title: Inference ## Overview -Inference - - Llama Stack Inference API for generating completions, chat completions, and embeddings. +Llama Stack Inference API for generating completions, chat completions, and embeddings. This API provides the raw interface to the underlying models. Two kinds of models are supported: - LLM models: these models generate "raw" and "chat" (conversational) completions. diff --git a/docs/docs/providers/safety/index.mdx b/docs/docs/providers/safety/index.mdx index 4e2de4f33..3445b17e6 100644 --- a/docs/docs/providers/safety/index.mdx +++ b/docs/docs/providers/safety/index.mdx @@ -1,7 +1,4 @@ --- -description: "Safety - - OpenAI-compatible Moderations API." sidebar_label: Safety title: Safety --- @@ -10,8 +7,4 @@ title: Safety ## Overview -Safety - - OpenAI-compatible Moderations API. - This section contains documentation for all available providers for the **safety** API. diff --git a/docs/static/deprecated-llama-stack-spec.html b/docs/static/deprecated-llama-stack-spec.html index 04a3dca9b..ffda7552b 100644 --- a/docs/static/deprecated-llama-stack-spec.html +++ b/docs/static/deprecated-llama-stack-spec.html @@ -1443,8 +1443,8 @@ "tags": [ "Inference" ], - "summary": "List chat completions.", - "description": "List chat completions.", + "summary": "List all chat completions.", + "description": "List all chat completions.", "parameters": [ { "name": "after", @@ -1520,8 +1520,8 @@ "tags": [ "Inference" ], - "summary": "Create chat completions.", - "description": "Create chat completions.\nGenerate an OpenAI-compatible chat completion for the given messages using the specified model.", + "summary": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + "description": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", "parameters": [], "requestBody": { "content": { @@ -1565,8 +1565,8 @@ "tags": [ "Inference" ], - "summary": "Get chat completion.", - "description": "Get chat completion.\nDescribe a chat completion by its ID.", + "summary": "Describe a chat completion by its ID.", + "description": "Describe a chat completion by its ID.", "parameters": [ { "name": "completion_id", @@ -1610,8 +1610,8 @@ "tags": [ "Inference" ], - "summary": "Create completion.", - "description": "Create completion.\nGenerate an OpenAI-compatible completion for the given prompt using the specified model.", + "summary": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + "description": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", "parameters": [], "requestBody": { "content": { @@ -1655,8 +1655,8 @@ "tags": [ "Inference" ], - "summary": "Create embeddings.", - "description": "Create embeddings.\nGenerate OpenAI-compatible embeddings for the given input using the specified model.", + "summary": "Generate OpenAI-compatible embeddings for the given input using the specified model.", + "description": "Generate OpenAI-compatible embeddings for the given input using the specified model.", "parameters": [], "requestBody": { "content": { @@ -1700,8 +1700,8 @@ "tags": [ "Files" ], - "summary": "List files.", - "description": "List files.\nReturns a list of files that belong to the user's organization.", + "summary": "Returns a list of files that belong to the user's organization.", + "description": "Returns a list of files that belong to the user's organization.", "parameters": [ { "name": "after", @@ -1770,8 +1770,8 @@ "tags": [ "Files" ], - "summary": "Upload file.", - "description": "Upload file.\nUpload a file that can be used across various endpoints.\n\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", + "summary": "Upload a file that can be used across various endpoints.", + "description": "Upload a file that can be used across various endpoints.\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", "parameters": [], "requestBody": { "content": { @@ -1831,8 +1831,8 @@ "tags": [ "Files" ], - "summary": "Retrieve file.", - "description": "Retrieve file.\nReturns information about a specific file.", + "summary": "Returns information about a specific file.", + "description": "Returns information about a specific file.", "parameters": [ { "name": "file_id", @@ -1874,8 +1874,8 @@ "tags": [ "Files" ], - "summary": "Delete file.", - "description": "Delete file.", + "summary": "Delete a file.", + "description": "Delete a file.", "parameters": [ { "name": "file_id", @@ -1919,8 +1919,8 @@ "tags": [ "Files" ], - "summary": "Retrieve file content.", - "description": "Retrieve file content.\nReturns the contents of the specified file.", + "summary": "Returns the contents of the specified file.", + "description": "Returns the contents of the specified file.", "parameters": [ { "name": "file_id", @@ -1999,8 +1999,8 @@ "tags": [ "Safety" ], - "summary": "Create moderation.", - "description": "Create moderation.\nClassifies if text and/or image inputs are potentially harmful.", + "summary": "Classifies if text and/or image inputs are potentially harmful.", + "description": "Classifies if text and/or image inputs are potentially harmful.", "parameters": [], "requestBody": { "content": { @@ -2044,8 +2044,8 @@ "tags": [ "Agents" ], - "summary": "List all responses.", - "description": "List all responses.", + "summary": "List all OpenAI responses.", + "description": "List all OpenAI responses.", "parameters": [ { "name": "after", @@ -2119,8 +2119,8 @@ "tags": [ "Agents" ], - "summary": "Create a model response.", - "description": "Create a model response.", + "summary": "Create a new OpenAI response.", + "description": "Create a new OpenAI response.", "parameters": [], "requestBody": { "content": { @@ -2184,8 +2184,8 @@ "tags": [ "Agents" ], - "summary": "Get a model response.", - "description": "Get a model response.", + "summary": "Retrieve an OpenAI response by its ID.", + "description": "Retrieve an OpenAI response by its ID.", "parameters": [ { "name": "response_id", @@ -2227,8 +2227,8 @@ "tags": [ "Agents" ], - "summary": "Delete a response.", - "description": "Delete a response.", + "summary": "Delete an OpenAI response by its ID.", + "description": "Delete an OpenAI response by its ID.", "parameters": [ { "name": "response_id", @@ -2272,8 +2272,8 @@ "tags": [ "Agents" ], - "summary": "List input items.", - "description": "List input items.", + "summary": "List input items for a given OpenAI response.", + "description": "List input items for a given OpenAI response.", "parameters": [ { "name": "response_id", @@ -13366,13 +13366,12 @@ }, { "name": "Files", - "description": "This API is used to upload documents that can be used with other Llama Stack APIs.", - "x-displayName": "Files" + "description": "" }, { "name": "Inference", - "description": "Llama Stack Inference API for generating completions, chat completions, and embeddings.\n\nThis API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", - "x-displayName": "Inference" + "description": "This API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", + "x-displayName": "Llama Stack Inference API for generating completions, chat completions, and embeddings." }, { "name": "Models", @@ -13384,8 +13383,7 @@ }, { "name": "Safety", - "description": "OpenAI-compatible Moderations API.", - "x-displayName": "Safety" + "description": "" }, { "name": "Telemetry", diff --git a/docs/static/deprecated-llama-stack-spec.yaml b/docs/static/deprecated-llama-stack-spec.yaml index 1a215b877..0e672f914 100644 --- a/docs/static/deprecated-llama-stack-spec.yaml +++ b/docs/static/deprecated-llama-stack-spec.yaml @@ -1033,8 +1033,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: List chat completions. - description: List chat completions. + summary: List all chat completions. + description: List all chat completions. parameters: - name: after in: query @@ -1087,10 +1087,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Create chat completions. + summary: >- + Generate an OpenAI-compatible chat completion for the given messages using + the specified model. description: >- - Create chat completions. - Generate an OpenAI-compatible chat completion for the given messages using the specified model. parameters: [] @@ -1122,11 +1122,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Get chat completion. - description: >- - Get chat completion. - - Describe a chat completion by its ID. + summary: Describe a chat completion by its ID. + description: Describe a chat completion by its ID. parameters: - name: completion_id in: path @@ -1156,10 +1153,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Create completion. + summary: >- + Generate an OpenAI-compatible completion for the given prompt using the specified + model. description: >- - Create completion. - Generate an OpenAI-compatible completion for the given prompt using the specified model. parameters: [] @@ -1192,10 +1189,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Create embeddings. + summary: >- + Generate OpenAI-compatible embeddings for the given input using the specified + model. description: >- - Create embeddings. - Generate OpenAI-compatible embeddings for the given input using the specified model. parameters: [] @@ -1228,10 +1225,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: List files. + summary: >- + Returns a list of files that belong to the user's organization. description: >- - List files. - Returns a list of files that belong to the user's organization. parameters: - name: after @@ -1289,12 +1285,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Upload file. - description: >- - Upload file. - + summary: >- + Upload a file that can be used across various endpoints. + description: >- Upload a file that can be used across various endpoints. - The file upload should be a multipart form request with: @@ -1344,10 +1338,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Retrieve file. + summary: >- + Returns information about a specific file. description: >- - Retrieve file. - Returns information about a specific file. parameters: - name: file_id @@ -1379,8 +1372,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Delete file. - description: Delete file. + summary: Delete a file. + description: Delete a file. parameters: - name: file_id in: path @@ -1412,10 +1405,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Retrieve file content. + summary: >- + Returns the contents of the specified file. description: >- - Retrieve file content. - Returns the contents of the specified file. parameters: - name: file_id @@ -1472,10 +1464,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: Create moderation. + summary: >- + Classifies if text and/or image inputs are potentially harmful. description: >- - Create moderation. - Classifies if text and/or image inputs are potentially harmful. parameters: [] requestBody: @@ -1506,8 +1497,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all responses. - description: List all responses. + summary: List all OpenAI responses. + description: List all OpenAI responses. parameters: - name: after in: query @@ -1558,8 +1549,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Create a model response. - description: Create a model response. + summary: Create a new OpenAI response. + description: Create a new OpenAI response. parameters: [] requestBody: content: @@ -1601,8 +1592,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Get a model response. - description: Get a model response. + summary: Retrieve an OpenAI response by its ID. + description: Retrieve an OpenAI response by its ID. parameters: - name: response_id in: path @@ -1632,8 +1623,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Delete a response. - description: Delete a response. + summary: Delete an OpenAI response by its ID. + description: Delete an OpenAI response by its ID. parameters: - name: response_id in: path @@ -1663,8 +1654,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List input items. - description: List input items. + summary: >- + List input items for a given OpenAI response. + description: >- + List input items for a given OpenAI response. parameters: - name: response_id in: path @@ -10018,16 +10011,9 @@ tags: x-displayName: >- Llama Stack Evaluation API for running evaluations on model and agent candidates. - name: Files - description: >- - This API is used to upload documents that can be used with other Llama Stack - APIs. - x-displayName: Files + description: '' - name: Inference description: >- - Llama Stack Inference API for generating completions, chat completions, and - embeddings. - - This API provides the raw interface to the underlying models. Two kinds of models are supported: @@ -10035,14 +10021,15 @@ tags: - Embedding models: these models generate embeddings to be used for semantic search. - x-displayName: Inference + x-displayName: >- + Llama Stack Inference API for generating completions, chat completions, and + embeddings. - name: Models description: '' - name: PostTraining (Coming Soon) description: '' - name: Safety - description: OpenAI-compatible Moderations API. - x-displayName: Safety + description: '' - name: Telemetry description: '' - name: VectorIO diff --git a/docs/static/llama-stack-spec.html b/docs/static/llama-stack-spec.html index 9cd526176..c570dcddf 100644 --- a/docs/static/llama-stack-spec.html +++ b/docs/static/llama-stack-spec.html @@ -69,8 +69,8 @@ "tags": [ "Inference" ], - "summary": "List chat completions.", - "description": "List chat completions.", + "summary": "List all chat completions.", + "description": "List all chat completions.", "parameters": [ { "name": "after", @@ -146,8 +146,8 @@ "tags": [ "Inference" ], - "summary": "Create chat completions.", - "description": "Create chat completions.\nGenerate an OpenAI-compatible chat completion for the given messages using the specified model.", + "summary": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + "description": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", "parameters": [], "requestBody": { "content": { @@ -191,8 +191,8 @@ "tags": [ "Inference" ], - "summary": "Get chat completion.", - "description": "Get chat completion.\nDescribe a chat completion by its ID.", + "summary": "Describe a chat completion by its ID.", + "description": "Describe a chat completion by its ID.", "parameters": [ { "name": "completion_id", @@ -236,8 +236,8 @@ "tags": [ "Inference" ], - "summary": "Create completion.", - "description": "Create completion.\nGenerate an OpenAI-compatible completion for the given prompt using the specified model.", + "summary": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + "description": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", "parameters": [], "requestBody": { "content": { @@ -758,8 +758,8 @@ "tags": [ "Inference" ], - "summary": "Create embeddings.", - "description": "Create embeddings.\nGenerate OpenAI-compatible embeddings for the given input using the specified model.", + "summary": "Generate OpenAI-compatible embeddings for the given input using the specified model.", + "description": "Generate OpenAI-compatible embeddings for the given input using the specified model.", "parameters": [], "requestBody": { "content": { @@ -803,8 +803,8 @@ "tags": [ "Files" ], - "summary": "List files.", - "description": "List files.\nReturns a list of files that belong to the user's organization.", + "summary": "Returns a list of files that belong to the user's organization.", + "description": "Returns a list of files that belong to the user's organization.", "parameters": [ { "name": "after", @@ -873,8 +873,8 @@ "tags": [ "Files" ], - "summary": "Upload file.", - "description": "Upload file.\nUpload a file that can be used across various endpoints.\n\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", + "summary": "Upload a file that can be used across various endpoints.", + "description": "Upload a file that can be used across various endpoints.\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", "parameters": [], "requestBody": { "content": { @@ -934,8 +934,8 @@ "tags": [ "Files" ], - "summary": "Retrieve file.", - "description": "Retrieve file.\nReturns information about a specific file.", + "summary": "Returns information about a specific file.", + "description": "Returns information about a specific file.", "parameters": [ { "name": "file_id", @@ -977,8 +977,8 @@ "tags": [ "Files" ], - "summary": "Delete file.", - "description": "Delete file.", + "summary": "Delete a file.", + "description": "Delete a file.", "parameters": [ { "name": "file_id", @@ -1022,8 +1022,8 @@ "tags": [ "Files" ], - "summary": "Retrieve file content.", - "description": "Retrieve file content.\nReturns the contents of the specified file.", + "summary": "Returns the contents of the specified file.", + "description": "Returns the contents of the specified file.", "parameters": [ { "name": "file_id", @@ -1067,8 +1067,8 @@ "tags": [ "Inspect" ], - "summary": "Get health status.", - "description": "Get health status.\nGet the current health status of the service.", + "summary": "Get the current health status of the service.", + "description": "Get the current health status of the service.", "parameters": [], "deprecated": false } @@ -1102,8 +1102,8 @@ "tags": [ "Inspect" ], - "summary": "List routes.", - "description": "List routes.\nList all available API routes with their methods and implementing providers.", + "summary": "List all available API routes with their methods and implementing providers.", + "description": "List all available API routes with their methods and implementing providers.", "parameters": [], "deprecated": false } @@ -1170,8 +1170,8 @@ "tags": [ "Models" ], - "summary": "Register model.", - "description": "Register model.\nRegister a model.", + "summary": "Register a model.", + "description": "Register a model.", "parameters": [], "requestBody": { "content": { @@ -1215,8 +1215,8 @@ "tags": [ "Models" ], - "summary": "Get model.", - "description": "Get model.\nGet a model by its identifier.", + "summary": "Get a model by its identifier.", + "description": "Get a model by its identifier.", "parameters": [ { "name": "model_id", @@ -1251,8 +1251,8 @@ "tags": [ "Models" ], - "summary": "Unregister model.", - "description": "Unregister model.\nUnregister a model.", + "summary": "Unregister a model.", + "description": "Unregister a model.", "parameters": [ { "name": "model_id", @@ -1296,8 +1296,8 @@ "tags": [ "Safety" ], - "summary": "Create moderation.", - "description": "Create moderation.\nClassifies if text and/or image inputs are potentially harmful.", + "summary": "Classifies if text and/or image inputs are potentially harmful.", + "description": "Classifies if text and/or image inputs are potentially harmful.", "parameters": [], "requestBody": { "content": { @@ -1374,8 +1374,8 @@ "tags": [ "Prompts" ], - "summary": "Create prompt.", - "description": "Create prompt.\nCreate a new prompt.", + "summary": "Create a new prompt.", + "description": "Create a new prompt.", "parameters": [], "requestBody": { "content": { @@ -1419,8 +1419,8 @@ "tags": [ "Prompts" ], - "summary": "Get prompt.", - "description": "Get prompt.\nGet a prompt by its identifier and optional version.", + "summary": "Get a prompt by its identifier and optional version.", + "description": "Get a prompt by its identifier and optional version.", "parameters": [ { "name": "prompt_id", @@ -1471,8 +1471,8 @@ "tags": [ "Prompts" ], - "summary": "Update prompt.", - "description": "Update prompt.\nUpdate an existing prompt (increments version).", + "summary": "Update an existing prompt (increments version).", + "description": "Update an existing prompt (increments version).", "parameters": [ { "name": "prompt_id", @@ -1517,8 +1517,8 @@ "tags": [ "Prompts" ], - "summary": "Delete prompt.", - "description": "Delete prompt.\nDelete a prompt.", + "summary": "Delete a prompt.", + "description": "Delete a prompt.", "parameters": [ { "name": "prompt_id", @@ -1562,8 +1562,8 @@ "tags": [ "Prompts" ], - "summary": "Set prompt version.", - "description": "Set prompt version.\nSet which version of a prompt should be the default in get_prompt (latest).", + "summary": "Set which version of a prompt should be the default in get_prompt (latest).", + "description": "Set which version of a prompt should be the default in get_prompt (latest).", "parameters": [ { "name": "prompt_id", @@ -1617,8 +1617,8 @@ "tags": [ "Prompts" ], - "summary": "List prompt versions.", - "description": "List prompt versions.\nList all versions of a specific prompt.", + "summary": "List all versions of a specific prompt.", + "description": "List all versions of a specific prompt.", "parameters": [ { "name": "prompt_id", @@ -1662,8 +1662,8 @@ "tags": [ "Providers" ], - "summary": "List providers.", - "description": "List providers.\nList all available providers.", + "summary": "List all available providers.", + "description": "List all available providers.", "parameters": [], "deprecated": false } @@ -1697,8 +1697,8 @@ "tags": [ "Providers" ], - "summary": "Get provider.", - "description": "Get provider.\nGet detailed information about a specific provider.", + "summary": "Get detailed information about a specific provider.", + "description": "Get detailed information about a specific provider.", "parameters": [ { "name": "provider_id", @@ -1742,8 +1742,8 @@ "tags": [ "Agents" ], - "summary": "List all responses.", - "description": "List all responses.", + "summary": "List all OpenAI responses.", + "description": "List all OpenAI responses.", "parameters": [ { "name": "after", @@ -1817,8 +1817,8 @@ "tags": [ "Agents" ], - "summary": "Create a model response.", - "description": "Create a model response.", + "summary": "Create a new OpenAI response.", + "description": "Create a new OpenAI response.", "parameters": [], "requestBody": { "content": { @@ -1882,8 +1882,8 @@ "tags": [ "Agents" ], - "summary": "Get a model response.", - "description": "Get a model response.", + "summary": "Retrieve an OpenAI response by its ID.", + "description": "Retrieve an OpenAI response by its ID.", "parameters": [ { "name": "response_id", @@ -1925,8 +1925,8 @@ "tags": [ "Agents" ], - "summary": "Delete a response.", - "description": "Delete a response.", + "summary": "Delete an OpenAI response by its ID.", + "description": "Delete an OpenAI response by its ID.", "parameters": [ { "name": "response_id", @@ -1970,8 +1970,8 @@ "tags": [ "Agents" ], - "summary": "List input items.", - "description": "List input items.", + "summary": "List input items for a given OpenAI response.", + "description": "List input items for a given OpenAI response.", "parameters": [ { "name": "response_id", @@ -2063,8 +2063,8 @@ "tags": [ "Safety" ], - "summary": "Run shield.", - "description": "Run shield.\nRun a shield.", + "summary": "Run a shield.", + "description": "Run a shield.", "parameters": [], "requestBody": { "content": { @@ -4196,8 +4196,8 @@ "tags": [ "Inspect" ], - "summary": "Get version.", - "description": "Get version.\nGet the version of the service.", + "summary": "Get the version of the service.", + "description": "Get the version of the service.", "parameters": [], "deprecated": false } @@ -12914,18 +12914,16 @@ }, { "name": "Files", - "description": "This API is used to upload documents that can be used with other Llama Stack APIs.", - "x-displayName": "Files" + "description": "" }, { "name": "Inference", - "description": "Llama Stack Inference API for generating completions, chat completions, and embeddings.\n\nThis API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", - "x-displayName": "Inference" + "description": "This API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", + "x-displayName": "Llama Stack Inference API for generating completions, chat completions, and embeddings." }, { "name": "Inspect", - "description": "APIs for inspecting the Llama Stack service, including health status, available API routes with methods and implementing providers.", - "x-displayName": "Inspect" + "description": "" }, { "name": "Models", @@ -12933,18 +12931,17 @@ }, { "name": "Prompts", - "description": "Protocol for prompt management operations.", - "x-displayName": "Prompts" + "description": "", + "x-displayName": "Protocol for prompt management operations." }, { "name": "Providers", - "description": "Providers API for inspecting, listing, and modifying providers and their configurations.", - "x-displayName": "Providers" + "description": "", + "x-displayName": "Providers API for inspecting, listing, and modifying providers and their configurations." }, { "name": "Safety", - "description": "OpenAI-compatible Moderations API.", - "x-displayName": "Safety" + "description": "" }, { "name": "Scoring", diff --git a/docs/static/llama-stack-spec.yaml b/docs/static/llama-stack-spec.yaml index 66ce8e38a..3e1431b22 100644 --- a/docs/static/llama-stack-spec.yaml +++ b/docs/static/llama-stack-spec.yaml @@ -33,8 +33,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: List chat completions. - description: List chat completions. + summary: List all chat completions. + description: List all chat completions. parameters: - name: after in: query @@ -87,10 +87,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Create chat completions. + summary: >- + Generate an OpenAI-compatible chat completion for the given messages using + the specified model. description: >- - Create chat completions. - Generate an OpenAI-compatible chat completion for the given messages using the specified model. parameters: [] @@ -122,11 +122,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Get chat completion. - description: >- - Get chat completion. - - Describe a chat completion by its ID. + summary: Describe a chat completion by its ID. + description: Describe a chat completion by its ID. parameters: - name: completion_id in: path @@ -156,10 +153,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Create completion. + summary: >- + Generate an OpenAI-compatible completion for the given prompt using the specified + model. description: >- - Create completion. - Generate an OpenAI-compatible completion for the given prompt using the specified model. parameters: [] @@ -606,10 +603,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Create embeddings. + summary: >- + Generate OpenAI-compatible embeddings for the given input using the specified + model. description: >- - Create embeddings. - Generate OpenAI-compatible embeddings for the given input using the specified model. parameters: [] @@ -642,10 +639,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: List files. + summary: >- + Returns a list of files that belong to the user's organization. description: >- - List files. - Returns a list of files that belong to the user's organization. parameters: - name: after @@ -703,12 +699,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Upload file. - description: >- - Upload file. - + summary: >- + Upload a file that can be used across various endpoints. + description: >- Upload a file that can be used across various endpoints. - The file upload should be a multipart form request with: @@ -758,10 +752,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Retrieve file. + summary: >- + Returns information about a specific file. description: >- - Retrieve file. - Returns information about a specific file. parameters: - name: file_id @@ -793,8 +786,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Delete file. - description: Delete file. + summary: Delete a file. + description: Delete a file. parameters: - name: file_id in: path @@ -826,10 +819,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Retrieve file content. + summary: >- + Returns the contents of the specified file. description: >- - Retrieve file content. - Returns the contents of the specified file. parameters: - name: file_id @@ -862,10 +854,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: Get health status. + summary: >- + Get the current health status of the service. description: >- - Get health status. - Get the current health status of the service. parameters: [] deprecated: false @@ -891,10 +882,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: List routes. + summary: >- + List all available API routes with their methods and implementing providers. description: >- - List routes. - List all available API routes with their methods and implementing providers. parameters: [] deprecated: false @@ -943,11 +933,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Register model. - description: >- - Register model. - - Register a model. + summary: Register a model. + description: Register a model. parameters: [] requestBody: content: @@ -977,11 +964,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Get model. - description: >- - Get model. - - Get a model by its identifier. + summary: Get a model by its identifier. + description: Get a model by its identifier. parameters: - name: model_id in: path @@ -1006,11 +990,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Unregister model. - description: >- - Unregister model. - - Unregister a model. + summary: Unregister a model. + description: Unregister a model. parameters: - name: model_id in: path @@ -1041,10 +1022,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: Create moderation. + summary: >- + Classifies if text and/or image inputs are potentially harmful. description: >- - Create moderation. - Classifies if text and/or image inputs are potentially harmful. parameters: [] requestBody: @@ -1100,11 +1080,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Create prompt. - description: >- - Create prompt. - - Create a new prompt. + summary: Create a new prompt. + description: Create a new prompt. parameters: [] requestBody: content: @@ -1134,10 +1111,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Get prompt. + summary: >- + Get a prompt by its identifier and optional version. description: >- - Get prompt. - Get a prompt by its identifier and optional version. parameters: - name: prompt_id @@ -1175,10 +1151,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Update prompt. + summary: >- + Update an existing prompt (increments version). description: >- - Update prompt. - Update an existing prompt (increments version). parameters: - name: prompt_id @@ -1210,11 +1185,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Delete prompt. - description: >- - Delete prompt. - - Delete a prompt. + summary: Delete a prompt. + description: Delete a prompt. parameters: - name: prompt_id in: path @@ -1245,10 +1217,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Set prompt version. + summary: >- + Set which version of a prompt should be the default in get_prompt (latest). description: >- - Set prompt version. - Set which version of a prompt should be the default in get_prompt (latest). parameters: - name: prompt_id @@ -1286,11 +1257,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: List prompt versions. - description: >- - List prompt versions. - - List all versions of a specific prompt. + summary: List all versions of a specific prompt. + description: List all versions of a specific prompt. parameters: - name: prompt_id in: path @@ -1322,11 +1290,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: List providers. - description: >- - List providers. - - List all available providers. + summary: List all available providers. + description: List all available providers. parameters: [] deprecated: false /v1/providers/{provider_id}: @@ -1351,10 +1316,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: Get provider. + summary: >- + Get detailed information about a specific provider. description: >- - Get provider. - Get detailed information about a specific provider. parameters: - name: provider_id @@ -1385,8 +1349,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all responses. - description: List all responses. + summary: List all OpenAI responses. + description: List all OpenAI responses. parameters: - name: after in: query @@ -1437,8 +1401,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Create a model response. - description: Create a model response. + summary: Create a new OpenAI response. + description: Create a new OpenAI response. parameters: [] requestBody: content: @@ -1480,8 +1444,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Get a model response. - description: Get a model response. + summary: Retrieve an OpenAI response by its ID. + description: Retrieve an OpenAI response by its ID. parameters: - name: response_id in: path @@ -1511,8 +1475,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Delete a response. - description: Delete a response. + summary: Delete an OpenAI response by its ID. + description: Delete an OpenAI response by its ID. parameters: - name: response_id in: path @@ -1542,8 +1506,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List input items. - description: List input items. + summary: >- + List input items for a given OpenAI response. + description: >- + List input items for a given OpenAI response. parameters: - name: response_id in: path @@ -1612,11 +1578,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: Run shield. - description: >- - Run shield. - - Run a shield. + summary: Run a shield. + description: Run a shield. parameters: [] requestBody: content: @@ -3172,11 +3135,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: Get version. - description: >- - Get version. - - Get the version of the service. + summary: Get the version of the service. + description: Get the version of the service. parameters: [] deprecated: false jsonSchemaDialect: >- @@ -9789,16 +9749,9 @@ tags: x-displayName: >- Protocol for conversation management operations. - name: Files - description: >- - This API is used to upload documents that can be used with other Llama Stack - APIs. - x-displayName: Files + description: '' - name: Inference description: >- - Llama Stack Inference API for generating completions, chat completions, and - embeddings. - - This API provides the raw interface to the underlying models. Two kinds of models are supported: @@ -9806,25 +9759,23 @@ tags: - Embedding models: these models generate embeddings to be used for semantic search. - x-displayName: Inference + x-displayName: >- + Llama Stack Inference API for generating completions, chat completions, and + embeddings. - name: Inspect - description: >- - APIs for inspecting the Llama Stack service, including health status, available - API routes with methods and implementing providers. - x-displayName: Inspect + description: '' - name: Models description: '' - name: Prompts - description: >- + description: '' + x-displayName: >- Protocol for prompt management operations. - x-displayName: Prompts - name: Providers - description: >- + description: '' + x-displayName: >- Providers API for inspecting, listing, and modifying providers and their configurations. - x-displayName: Providers - name: Safety - description: OpenAI-compatible Moderations API. - x-displayName: Safety + description: '' - name: Scoring description: '' - name: ScoringFunctions diff --git a/docs/static/stainless-llama-stack-spec.html b/docs/static/stainless-llama-stack-spec.html index 3478d3338..167a4aa3c 100644 --- a/docs/static/stainless-llama-stack-spec.html +++ b/docs/static/stainless-llama-stack-spec.html @@ -69,8 +69,8 @@ "tags": [ "Inference" ], - "summary": "List chat completions.", - "description": "List chat completions.", + "summary": "List all chat completions.", + "description": "List all chat completions.", "parameters": [ { "name": "after", @@ -146,8 +146,8 @@ "tags": [ "Inference" ], - "summary": "Create chat completions.", - "description": "Create chat completions.\nGenerate an OpenAI-compatible chat completion for the given messages using the specified model.", + "summary": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", + "description": "Generate an OpenAI-compatible chat completion for the given messages using the specified model.", "parameters": [], "requestBody": { "content": { @@ -191,8 +191,8 @@ "tags": [ "Inference" ], - "summary": "Get chat completion.", - "description": "Get chat completion.\nDescribe a chat completion by its ID.", + "summary": "Describe a chat completion by its ID.", + "description": "Describe a chat completion by its ID.", "parameters": [ { "name": "completion_id", @@ -236,8 +236,8 @@ "tags": [ "Inference" ], - "summary": "Create completion.", - "description": "Create completion.\nGenerate an OpenAI-compatible completion for the given prompt using the specified model.", + "summary": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", + "description": "Generate an OpenAI-compatible completion for the given prompt using the specified model.", "parameters": [], "requestBody": { "content": { @@ -758,8 +758,8 @@ "tags": [ "Inference" ], - "summary": "Create embeddings.", - "description": "Create embeddings.\nGenerate OpenAI-compatible embeddings for the given input using the specified model.", + "summary": "Generate OpenAI-compatible embeddings for the given input using the specified model.", + "description": "Generate OpenAI-compatible embeddings for the given input using the specified model.", "parameters": [], "requestBody": { "content": { @@ -803,8 +803,8 @@ "tags": [ "Files" ], - "summary": "List files.", - "description": "List files.\nReturns a list of files that belong to the user's organization.", + "summary": "Returns a list of files that belong to the user's organization.", + "description": "Returns a list of files that belong to the user's organization.", "parameters": [ { "name": "after", @@ -873,8 +873,8 @@ "tags": [ "Files" ], - "summary": "Upload file.", - "description": "Upload file.\nUpload a file that can be used across various endpoints.\n\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", + "summary": "Upload a file that can be used across various endpoints.", + "description": "Upload a file that can be used across various endpoints.\nThe file upload should be a multipart form request with:\n- file: The File object (not file name) to be uploaded.\n- purpose: The intended purpose of the uploaded file.\n- expires_after: Optional form values describing expiration for the file.", "parameters": [], "requestBody": { "content": { @@ -934,8 +934,8 @@ "tags": [ "Files" ], - "summary": "Retrieve file.", - "description": "Retrieve file.\nReturns information about a specific file.", + "summary": "Returns information about a specific file.", + "description": "Returns information about a specific file.", "parameters": [ { "name": "file_id", @@ -977,8 +977,8 @@ "tags": [ "Files" ], - "summary": "Delete file.", - "description": "Delete file.", + "summary": "Delete a file.", + "description": "Delete a file.", "parameters": [ { "name": "file_id", @@ -1022,8 +1022,8 @@ "tags": [ "Files" ], - "summary": "Retrieve file content.", - "description": "Retrieve file content.\nReturns the contents of the specified file.", + "summary": "Returns the contents of the specified file.", + "description": "Returns the contents of the specified file.", "parameters": [ { "name": "file_id", @@ -1067,8 +1067,8 @@ "tags": [ "Inspect" ], - "summary": "Get health status.", - "description": "Get health status.\nGet the current health status of the service.", + "summary": "Get the current health status of the service.", + "description": "Get the current health status of the service.", "parameters": [], "deprecated": false } @@ -1102,8 +1102,8 @@ "tags": [ "Inspect" ], - "summary": "List routes.", - "description": "List routes.\nList all available API routes with their methods and implementing providers.", + "summary": "List all available API routes with their methods and implementing providers.", + "description": "List all available API routes with their methods and implementing providers.", "parameters": [], "deprecated": false } @@ -1170,8 +1170,8 @@ "tags": [ "Models" ], - "summary": "Register model.", - "description": "Register model.\nRegister a model.", + "summary": "Register a model.", + "description": "Register a model.", "parameters": [], "requestBody": { "content": { @@ -1215,8 +1215,8 @@ "tags": [ "Models" ], - "summary": "Get model.", - "description": "Get model.\nGet a model by its identifier.", + "summary": "Get a model by its identifier.", + "description": "Get a model by its identifier.", "parameters": [ { "name": "model_id", @@ -1251,8 +1251,8 @@ "tags": [ "Models" ], - "summary": "Unregister model.", - "description": "Unregister model.\nUnregister a model.", + "summary": "Unregister a model.", + "description": "Unregister a model.", "parameters": [ { "name": "model_id", @@ -1296,8 +1296,8 @@ "tags": [ "Safety" ], - "summary": "Create moderation.", - "description": "Create moderation.\nClassifies if text and/or image inputs are potentially harmful.", + "summary": "Classifies if text and/or image inputs are potentially harmful.", + "description": "Classifies if text and/or image inputs are potentially harmful.", "parameters": [], "requestBody": { "content": { @@ -1374,8 +1374,8 @@ "tags": [ "Prompts" ], - "summary": "Create prompt.", - "description": "Create prompt.\nCreate a new prompt.", + "summary": "Create a new prompt.", + "description": "Create a new prompt.", "parameters": [], "requestBody": { "content": { @@ -1419,8 +1419,8 @@ "tags": [ "Prompts" ], - "summary": "Get prompt.", - "description": "Get prompt.\nGet a prompt by its identifier and optional version.", + "summary": "Get a prompt by its identifier and optional version.", + "description": "Get a prompt by its identifier and optional version.", "parameters": [ { "name": "prompt_id", @@ -1471,8 +1471,8 @@ "tags": [ "Prompts" ], - "summary": "Update prompt.", - "description": "Update prompt.\nUpdate an existing prompt (increments version).", + "summary": "Update an existing prompt (increments version).", + "description": "Update an existing prompt (increments version).", "parameters": [ { "name": "prompt_id", @@ -1517,8 +1517,8 @@ "tags": [ "Prompts" ], - "summary": "Delete prompt.", - "description": "Delete prompt.\nDelete a prompt.", + "summary": "Delete a prompt.", + "description": "Delete a prompt.", "parameters": [ { "name": "prompt_id", @@ -1562,8 +1562,8 @@ "tags": [ "Prompts" ], - "summary": "Set prompt version.", - "description": "Set prompt version.\nSet which version of a prompt should be the default in get_prompt (latest).", + "summary": "Set which version of a prompt should be the default in get_prompt (latest).", + "description": "Set which version of a prompt should be the default in get_prompt (latest).", "parameters": [ { "name": "prompt_id", @@ -1617,8 +1617,8 @@ "tags": [ "Prompts" ], - "summary": "List prompt versions.", - "description": "List prompt versions.\nList all versions of a specific prompt.", + "summary": "List all versions of a specific prompt.", + "description": "List all versions of a specific prompt.", "parameters": [ { "name": "prompt_id", @@ -1662,8 +1662,8 @@ "tags": [ "Providers" ], - "summary": "List providers.", - "description": "List providers.\nList all available providers.", + "summary": "List all available providers.", + "description": "List all available providers.", "parameters": [], "deprecated": false } @@ -1697,8 +1697,8 @@ "tags": [ "Providers" ], - "summary": "Get provider.", - "description": "Get provider.\nGet detailed information about a specific provider.", + "summary": "Get detailed information about a specific provider.", + "description": "Get detailed information about a specific provider.", "parameters": [ { "name": "provider_id", @@ -1742,8 +1742,8 @@ "tags": [ "Agents" ], - "summary": "List all responses.", - "description": "List all responses.", + "summary": "List all OpenAI responses.", + "description": "List all OpenAI responses.", "parameters": [ { "name": "after", @@ -1817,8 +1817,8 @@ "tags": [ "Agents" ], - "summary": "Create a model response.", - "description": "Create a model response.", + "summary": "Create a new OpenAI response.", + "description": "Create a new OpenAI response.", "parameters": [], "requestBody": { "content": { @@ -1882,8 +1882,8 @@ "tags": [ "Agents" ], - "summary": "Get a model response.", - "description": "Get a model response.", + "summary": "Retrieve an OpenAI response by its ID.", + "description": "Retrieve an OpenAI response by its ID.", "parameters": [ { "name": "response_id", @@ -1925,8 +1925,8 @@ "tags": [ "Agents" ], - "summary": "Delete a response.", - "description": "Delete a response.", + "summary": "Delete an OpenAI response by its ID.", + "description": "Delete an OpenAI response by its ID.", "parameters": [ { "name": "response_id", @@ -1970,8 +1970,8 @@ "tags": [ "Agents" ], - "summary": "List input items.", - "description": "List input items.", + "summary": "List input items for a given OpenAI response.", + "description": "List input items for a given OpenAI response.", "parameters": [ { "name": "response_id", @@ -2063,8 +2063,8 @@ "tags": [ "Safety" ], - "summary": "Run shield.", - "description": "Run shield.\nRun a shield.", + "summary": "Run a shield.", + "description": "Run a shield.", "parameters": [], "requestBody": { "content": { @@ -4196,8 +4196,8 @@ "tags": [ "Inspect" ], - "summary": "Get version.", - "description": "Get version.\nGet the version of the service.", + "summary": "Get the version of the service.", + "description": "Get the version of the service.", "parameters": [], "deprecated": false } @@ -18487,18 +18487,16 @@ }, { "name": "Files", - "description": "This API is used to upload documents that can be used with other Llama Stack APIs.", - "x-displayName": "Files" + "description": "" }, { "name": "Inference", - "description": "Llama Stack Inference API for generating completions, chat completions, and embeddings.\n\nThis API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", - "x-displayName": "Inference" + "description": "This API provides the raw interface to the underlying models. Two kinds of models are supported:\n- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.\n- Embedding models: these models generate embeddings to be used for semantic search.", + "x-displayName": "Llama Stack Inference API for generating completions, chat completions, and embeddings." }, { "name": "Inspect", - "description": "APIs for inspecting the Llama Stack service, including health status, available API routes with methods and implementing providers.", - "x-displayName": "Inspect" + "description": "" }, { "name": "Models", @@ -18510,18 +18508,17 @@ }, { "name": "Prompts", - "description": "Protocol for prompt management operations.", - "x-displayName": "Prompts" + "description": "", + "x-displayName": "Protocol for prompt management operations." }, { "name": "Providers", - "description": "Providers API for inspecting, listing, and modifying providers and their configurations.", - "x-displayName": "Providers" + "description": "", + "x-displayName": "Providers API for inspecting, listing, and modifying providers and their configurations." }, { "name": "Safety", - "description": "OpenAI-compatible Moderations API.", - "x-displayName": "Safety" + "description": "" }, { "name": "Scoring", diff --git a/docs/static/stainless-llama-stack-spec.yaml b/docs/static/stainless-llama-stack-spec.yaml index 6c04542bf..6dc1041f1 100644 --- a/docs/static/stainless-llama-stack-spec.yaml +++ b/docs/static/stainless-llama-stack-spec.yaml @@ -36,8 +36,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: List chat completions. - description: List chat completions. + summary: List all chat completions. + description: List all chat completions. parameters: - name: after in: query @@ -90,10 +90,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Create chat completions. + summary: >- + Generate an OpenAI-compatible chat completion for the given messages using + the specified model. description: >- - Create chat completions. - Generate an OpenAI-compatible chat completion for the given messages using the specified model. parameters: [] @@ -125,11 +125,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Get chat completion. - description: >- - Get chat completion. - - Describe a chat completion by its ID. + summary: Describe a chat completion by its ID. + description: Describe a chat completion by its ID. parameters: - name: completion_id in: path @@ -159,10 +156,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Create completion. + summary: >- + Generate an OpenAI-compatible completion for the given prompt using the specified + model. description: >- - Create completion. - Generate an OpenAI-compatible completion for the given prompt using the specified model. parameters: [] @@ -609,10 +606,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inference - summary: Create embeddings. + summary: >- + Generate OpenAI-compatible embeddings for the given input using the specified + model. description: >- - Create embeddings. - Generate OpenAI-compatible embeddings for the given input using the specified model. parameters: [] @@ -645,10 +642,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: List files. + summary: >- + Returns a list of files that belong to the user's organization. description: >- - List files. - Returns a list of files that belong to the user's organization. parameters: - name: after @@ -706,12 +702,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Upload file. - description: >- - Upload file. - + summary: >- + Upload a file that can be used across various endpoints. + description: >- Upload a file that can be used across various endpoints. - The file upload should be a multipart form request with: @@ -761,10 +755,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Retrieve file. + summary: >- + Returns information about a specific file. description: >- - Retrieve file. - Returns information about a specific file. parameters: - name: file_id @@ -796,8 +789,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Delete file. - description: Delete file. + summary: Delete a file. + description: Delete a file. parameters: - name: file_id in: path @@ -829,10 +822,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Files - summary: Retrieve file content. + summary: >- + Returns the contents of the specified file. description: >- - Retrieve file content. - Returns the contents of the specified file. parameters: - name: file_id @@ -865,10 +857,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: Get health status. + summary: >- + Get the current health status of the service. description: >- - Get health status. - Get the current health status of the service. parameters: [] deprecated: false @@ -894,10 +885,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: List routes. + summary: >- + List all available API routes with their methods and implementing providers. description: >- - List routes. - List all available API routes with their methods and implementing providers. parameters: [] deprecated: false @@ -946,11 +936,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Register model. - description: >- - Register model. - - Register a model. + summary: Register a model. + description: Register a model. parameters: [] requestBody: content: @@ -980,11 +967,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Get model. - description: >- - Get model. - - Get a model by its identifier. + summary: Get a model by its identifier. + description: Get a model by its identifier. parameters: - name: model_id in: path @@ -1009,11 +993,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Models - summary: Unregister model. - description: >- - Unregister model. - - Unregister a model. + summary: Unregister a model. + description: Unregister a model. parameters: - name: model_id in: path @@ -1044,10 +1025,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: Create moderation. + summary: >- + Classifies if text and/or image inputs are potentially harmful. description: >- - Create moderation. - Classifies if text and/or image inputs are potentially harmful. parameters: [] requestBody: @@ -1103,11 +1083,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Create prompt. - description: >- - Create prompt. - - Create a new prompt. + summary: Create a new prompt. + description: Create a new prompt. parameters: [] requestBody: content: @@ -1137,10 +1114,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Get prompt. + summary: >- + Get a prompt by its identifier and optional version. description: >- - Get prompt. - Get a prompt by its identifier and optional version. parameters: - name: prompt_id @@ -1178,10 +1154,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Update prompt. + summary: >- + Update an existing prompt (increments version). description: >- - Update prompt. - Update an existing prompt (increments version). parameters: - name: prompt_id @@ -1213,11 +1188,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Delete prompt. - description: >- - Delete prompt. - - Delete a prompt. + summary: Delete a prompt. + description: Delete a prompt. parameters: - name: prompt_id in: path @@ -1248,10 +1220,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: Set prompt version. + summary: >- + Set which version of a prompt should be the default in get_prompt (latest). description: >- - Set prompt version. - Set which version of a prompt should be the default in get_prompt (latest). parameters: - name: prompt_id @@ -1289,11 +1260,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Prompts - summary: List prompt versions. - description: >- - List prompt versions. - - List all versions of a specific prompt. + summary: List all versions of a specific prompt. + description: List all versions of a specific prompt. parameters: - name: prompt_id in: path @@ -1325,11 +1293,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: List providers. - description: >- - List providers. - - List all available providers. + summary: List all available providers. + description: List all available providers. parameters: [] deprecated: false /v1/providers/{provider_id}: @@ -1354,10 +1319,9 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Providers - summary: Get provider. + summary: >- + Get detailed information about a specific provider. description: >- - Get provider. - Get detailed information about a specific provider. parameters: - name: provider_id @@ -1388,8 +1352,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List all responses. - description: List all responses. + summary: List all OpenAI responses. + description: List all OpenAI responses. parameters: - name: after in: query @@ -1440,8 +1404,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Create a model response. - description: Create a model response. + summary: Create a new OpenAI response. + description: Create a new OpenAI response. parameters: [] requestBody: content: @@ -1483,8 +1447,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Get a model response. - description: Get a model response. + summary: Retrieve an OpenAI response by its ID. + description: Retrieve an OpenAI response by its ID. parameters: - name: response_id in: path @@ -1514,8 +1478,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: Delete a response. - description: Delete a response. + summary: Delete an OpenAI response by its ID. + description: Delete an OpenAI response by its ID. parameters: - name: response_id in: path @@ -1545,8 +1509,10 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Agents - summary: List input items. - description: List input items. + summary: >- + List input items for a given OpenAI response. + description: >- + List input items for a given OpenAI response. parameters: - name: response_id in: path @@ -1615,11 +1581,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Safety - summary: Run shield. - description: >- - Run shield. - - Run a shield. + summary: Run a shield. + description: Run a shield. parameters: [] requestBody: content: @@ -3175,11 +3138,8 @@ paths: $ref: '#/components/responses/DefaultError' tags: - Inspect - summary: Get version. - description: >- - Get version. - - Get the version of the service. + summary: Get the version of the service. + description: Get the version of the service. parameters: [] deprecated: false /v1beta/datasetio/append-rows/{dataset_id}: @@ -13835,16 +13795,9 @@ tags: x-displayName: >- Llama Stack Evaluation API for running evaluations on model and agent candidates. - name: Files - description: >- - This API is used to upload documents that can be used with other Llama Stack - APIs. - x-displayName: Files + description: '' - name: Inference description: >- - Llama Stack Inference API for generating completions, chat completions, and - embeddings. - - This API provides the raw interface to the underlying models. Two kinds of models are supported: @@ -13852,27 +13805,25 @@ tags: - Embedding models: these models generate embeddings to be used for semantic search. - x-displayName: Inference + x-displayName: >- + Llama Stack Inference API for generating completions, chat completions, and + embeddings. - name: Inspect - description: >- - APIs for inspecting the Llama Stack service, including health status, available - API routes with methods and implementing providers. - x-displayName: Inspect + description: '' - name: Models description: '' - name: PostTraining (Coming Soon) description: '' - name: Prompts - description: >- + description: '' + x-displayName: >- Protocol for prompt management operations. - x-displayName: Prompts - name: Providers - description: >- + description: '' + x-displayName: >- Providers API for inspecting, listing, and modifying providers and their configurations. - x-displayName: Providers - name: Safety - description: OpenAI-compatible Moderations API. - x-displayName: Safety + description: '' - name: Scoring description: '' - name: ScoringFunctions diff --git a/llama_stack/apis/agents/agents.py b/llama_stack/apis/agents/agents.py index 5983b5c45..cdf47308e 100644 --- a/llama_stack/apis/agents/agents.py +++ b/llama_stack/apis/agents/agents.py @@ -797,7 +797,7 @@ class Agents(Protocol): self, response_id: str, ) -> OpenAIResponseObject: - """Get a model response. + """Retrieve an OpenAI response by its ID. :param response_id: The ID of the OpenAI response to retrieve. :returns: An OpenAIResponseObject. @@ -826,7 +826,7 @@ class Agents(Protocol): ), ] = None, ) -> OpenAIResponseObject | AsyncIterator[OpenAIResponseObjectStream]: - """Create a model response. + """Create a new OpenAI response. :param input: Input message(s) to create the response. :param model: The underlying LLM used for completions. @@ -846,7 +846,7 @@ class Agents(Protocol): model: str | None = None, order: Order | None = Order.desc, ) -> ListOpenAIResponseObject: - """List all responses. + """List all OpenAI responses. :param after: The ID of the last response to return. :param limit: The number of responses to return. @@ -869,7 +869,7 @@ class Agents(Protocol): limit: int | None = 20, order: Order | None = Order.desc, ) -> ListOpenAIResponseInputItem: - """List input items. + """List input items for a given OpenAI response. :param response_id: The ID of the response to retrieve input items for. :param after: An item ID to list items after, used for pagination. @@ -884,7 +884,7 @@ class Agents(Protocol): @webmethod(route="/openai/v1/responses/{response_id}", method="DELETE", level=LLAMA_STACK_API_V1, deprecated=True) @webmethod(route="/responses/{response_id}", method="DELETE", level=LLAMA_STACK_API_V1) async def delete_openai_response(self, response_id: str) -> OpenAIDeleteResponseObject: - """Delete a response. + """Delete an OpenAI response by its ID. :param response_id: The ID of the OpenAI response to delete. :returns: An OpenAIDeleteResponseObject diff --git a/llama_stack/apis/files/files.py b/llama_stack/apis/files/files.py index f1d3764db..13f0e95fa 100644 --- a/llama_stack/apis/files/files.py +++ b/llama_stack/apis/files/files.py @@ -104,11 +104,6 @@ class OpenAIFileDeleteResponse(BaseModel): @runtime_checkable @trace_protocol class Files(Protocol): - """Files - - This API is used to upload documents that can be used with other Llama Stack APIs. - """ - # OpenAI Files API Endpoints @webmethod(route="/openai/v1/files", method="POST", level=LLAMA_STACK_API_V1, deprecated=True) @webmethod(route="/files", method="POST", level=LLAMA_STACK_API_V1) @@ -118,8 +113,7 @@ class Files(Protocol): purpose: Annotated[OpenAIFilePurpose, Form()], expires_after: Annotated[ExpiresAfter | None, Form()] = None, ) -> OpenAIFileObject: - """Upload file. - + """ Upload a file that can be used across various endpoints. The file upload should be a multipart form request with: @@ -143,8 +137,7 @@ class Files(Protocol): order: Order | None = Order.desc, purpose: OpenAIFilePurpose | None = None, ) -> ListOpenAIFileResponse: - """List files. - + """ Returns a list of files that belong to the user's organization. :param after: A cursor for use in pagination. `after` is an object ID that defines your place in the list. For instance, if you make a list request and receive 100 objects, ending with obj_foo, your subsequent call can include after=obj_foo in order to fetch the next page of the list. @@ -161,8 +154,7 @@ class Files(Protocol): self, file_id: str, ) -> OpenAIFileObject: - """Retrieve file. - + """ Returns information about a specific file. :param file_id: The ID of the file to use for this request. @@ -176,7 +168,8 @@ class Files(Protocol): self, file_id: str, ) -> OpenAIFileDeleteResponse: - """Delete file. + """ + Delete a file. :param file_id: The ID of the file to use for this request. :returns: An OpenAIFileDeleteResponse indicating successful deletion. @@ -189,8 +182,7 @@ class Files(Protocol): self, file_id: str, ) -> Response: - """Retrieve file content. - + """ Returns the contents of the specified file. :param file_id: The ID of the file to use for this request. diff --git a/llama_stack/apis/inference/inference.py b/llama_stack/apis/inference/inference.py index 62a988ea6..e88a16315 100644 --- a/llama_stack/apis/inference/inference.py +++ b/llama_stack/apis/inference/inference.py @@ -1053,9 +1053,7 @@ class InferenceProvider(Protocol): # for fill-in-the-middle type completion suffix: str | None = None, ) -> OpenAICompletion: - """Create completion. - - Generate an OpenAI-compatible completion for the given prompt using the specified model. + """Generate an OpenAI-compatible completion for the given prompt using the specified model. :param model: The identifier of the model to use. The model must be registered with Llama Stack and available via the /models endpoint. :param prompt: The prompt to generate a completion for. @@ -1107,9 +1105,7 @@ class InferenceProvider(Protocol): top_p: float | None = None, user: str | None = None, ) -> OpenAIChatCompletion | AsyncIterator[OpenAIChatCompletionChunk]: - """Create chat completions. - - Generate an OpenAI-compatible chat completion for the given messages using the specified model. + """Generate an OpenAI-compatible chat completion for the given messages using the specified model. :param model: The identifier of the model to use. The model must be registered with Llama Stack and available via the /models endpoint. :param messages: List of messages in the conversation. @@ -1148,9 +1144,7 @@ class InferenceProvider(Protocol): dimensions: int | None = None, user: str | None = None, ) -> OpenAIEmbeddingsResponse: - """Create embeddings. - - Generate OpenAI-compatible embeddings for the given input using the specified model. + """Generate OpenAI-compatible embeddings for the given input using the specified model. :param model: The identifier of the model to use. The model must be an embedding model registered with Llama Stack and available via the /models endpoint. :param input: Input text to embed, encoded as a string or array of strings. To embed multiple inputs in a single request, pass an array of strings. @@ -1163,9 +1157,7 @@ class InferenceProvider(Protocol): class Inference(InferenceProvider): - """Inference - - Llama Stack Inference API for generating completions, chat completions, and embeddings. + """Llama Stack Inference API for generating completions, chat completions, and embeddings. This API provides the raw interface to the underlying models. Two kinds of models are supported: - LLM models: these models generate "raw" and "chat" (conversational) completions. @@ -1181,7 +1173,7 @@ class Inference(InferenceProvider): model: str | None = None, order: Order | None = Order.desc, ) -> ListOpenAIChatCompletionResponse: - """List chat completions. + """List all chat completions. :param after: The ID of the last chat completion to return. :param limit: The maximum number of chat completions to return. @@ -1196,9 +1188,7 @@ class Inference(InferenceProvider): ) @webmethod(route="/chat/completions/{completion_id}", method="GET", level=LLAMA_STACK_API_V1) async def get_chat_completion(self, completion_id: str) -> OpenAICompletionWithInputMessages: - """Get chat completion. - - Describe a chat completion by its ID. + """Describe a chat completion by its ID. :param completion_id: ID of the chat completion. :returns: A OpenAICompletionWithInputMessages. diff --git a/llama_stack/apis/inspect/inspect.py b/llama_stack/apis/inspect/inspect.py index 72f203621..e859dbe59 100644 --- a/llama_stack/apis/inspect/inspect.py +++ b/llama_stack/apis/inspect/inspect.py @@ -58,16 +58,9 @@ class ListRoutesResponse(BaseModel): @runtime_checkable class Inspect(Protocol): - """Inspect - - APIs for inspecting the Llama Stack service, including health status, available API routes with methods and implementing providers. - """ - @webmethod(route="/inspect/routes", method="GET", level=LLAMA_STACK_API_V1) async def list_routes(self) -> ListRoutesResponse: - """List routes. - - List all available API routes with their methods and implementing providers. + """List all available API routes with their methods and implementing providers. :returns: Response containing information about all available routes. """ @@ -75,9 +68,7 @@ class Inspect(Protocol): @webmethod(route="/health", method="GET", level=LLAMA_STACK_API_V1) async def health(self) -> HealthInfo: - """Get health status. - - Get the current health status of the service. + """Get the current health status of the service. :returns: Health information indicating if the service is operational. """ @@ -85,9 +76,7 @@ class Inspect(Protocol): @webmethod(route="/version", method="GET", level=LLAMA_STACK_API_V1) async def version(self) -> VersionInfo: - """Get version. - - Get the version of the service. + """Get the version of the service. :returns: Version information containing the service version number. """ diff --git a/llama_stack/apis/models/models.py b/llama_stack/apis/models/models.py index 10949cb95..210ed9246 100644 --- a/llama_stack/apis/models/models.py +++ b/llama_stack/apis/models/models.py @@ -124,9 +124,7 @@ class Models(Protocol): self, model_id: str, ) -> Model: - """Get model. - - Get a model by its identifier. + """Get a model by its identifier. :param model_id: The identifier of the model to get. :returns: A Model. @@ -142,9 +140,7 @@ class Models(Protocol): metadata: dict[str, Any] | None = None, model_type: ModelType | None = None, ) -> Model: - """Register model. - - Register a model. + """Register a model. :param model_id: The identifier of the model to register. :param provider_model_id: The identifier of the model in the provider. @@ -160,9 +156,7 @@ class Models(Protocol): self, model_id: str, ) -> None: - """Unregister model. - - Unregister a model. + """Unregister a model. :param model_id: The identifier of the model to unregister. """ diff --git a/llama_stack/apis/prompts/prompts.py b/llama_stack/apis/prompts/prompts.py index b39c363c7..c56185e25 100644 --- a/llama_stack/apis/prompts/prompts.py +++ b/llama_stack/apis/prompts/prompts.py @@ -94,9 +94,7 @@ class ListPromptsResponse(BaseModel): @runtime_checkable @trace_protocol class Prompts(Protocol): - """Prompts - - Protocol for prompt management operations.""" + """Protocol for prompt management operations.""" @webmethod(route="/prompts", method="GET", level=LLAMA_STACK_API_V1) async def list_prompts(self) -> ListPromptsResponse: @@ -111,9 +109,7 @@ class Prompts(Protocol): self, prompt_id: str, ) -> ListPromptsResponse: - """List prompt versions. - - List all versions of a specific prompt. + """List all versions of a specific prompt. :param prompt_id: The identifier of the prompt to list versions for. :returns: A ListPromptsResponse containing all versions of the prompt. @@ -126,9 +122,7 @@ class Prompts(Protocol): prompt_id: str, version: int | None = None, ) -> Prompt: - """Get prompt. - - Get a prompt by its identifier and optional version. + """Get a prompt by its identifier and optional version. :param prompt_id: The identifier of the prompt to get. :param version: The version of the prompt to get (defaults to latest). @@ -142,9 +136,7 @@ class Prompts(Protocol): prompt: str, variables: list[str] | None = None, ) -> Prompt: - """Create prompt. - - Create a new prompt. + """Create a new prompt. :param prompt: The prompt text content with variable placeholders. :param variables: List of variable names that can be used in the prompt template. @@ -161,9 +153,7 @@ class Prompts(Protocol): variables: list[str] | None = None, set_as_default: bool = True, ) -> Prompt: - """Update prompt. - - Update an existing prompt (increments version). + """Update an existing prompt (increments version). :param prompt_id: The identifier of the prompt to update. :param prompt: The updated prompt text content. @@ -179,9 +169,7 @@ class Prompts(Protocol): self, prompt_id: str, ) -> None: - """Delete prompt. - - Delete a prompt. + """Delete a prompt. :param prompt_id: The identifier of the prompt to delete. """ @@ -193,9 +181,7 @@ class Prompts(Protocol): prompt_id: str, version: int, ) -> Prompt: - """Set prompt version. - - Set which version of a prompt should be the default in get_prompt (latest). + """Set which version of a prompt should be the default in get_prompt (latest). :param prompt_id: The identifier of the prompt. :param version: The version to set as default. diff --git a/llama_stack/apis/providers/providers.py b/llama_stack/apis/providers/providers.py index e1872571d..d1cff0f6c 100644 --- a/llama_stack/apis/providers/providers.py +++ b/llama_stack/apis/providers/providers.py @@ -42,16 +42,13 @@ class ListProvidersResponse(BaseModel): @runtime_checkable class Providers(Protocol): - """Providers - + """ Providers API for inspecting, listing, and modifying providers and their configurations. """ @webmethod(route="/providers", method="GET", level=LLAMA_STACK_API_V1) async def list_providers(self) -> ListProvidersResponse: - """List providers. - - List all available providers. + """List all available providers. :returns: A ListProvidersResponse containing information about all providers. """ @@ -59,9 +56,7 @@ class Providers(Protocol): @webmethod(route="/providers/{provider_id}", method="GET", level=LLAMA_STACK_API_V1) async def inspect_provider(self, provider_id: str) -> ProviderInfo: - """Get provider. - - Get detailed information about a specific provider. + """Get detailed information about a specific provider. :param provider_id: The ID of the provider to inspect. :returns: A ProviderInfo object containing the provider's details. diff --git a/llama_stack/apis/safety/safety.py b/llama_stack/apis/safety/safety.py index 2ae74b0a7..0fa250d90 100644 --- a/llama_stack/apis/safety/safety.py +++ b/llama_stack/apis/safety/safety.py @@ -96,11 +96,6 @@ class ShieldStore(Protocol): @runtime_checkable @trace_protocol class Safety(Protocol): - """Safety - - OpenAI-compatible Moderations API. - """ - shield_store: ShieldStore @webmethod(route="/safety/run-shield", method="POST", level=LLAMA_STACK_API_V1) @@ -110,9 +105,7 @@ class Safety(Protocol): messages: list[Message], params: dict[str, Any], ) -> RunShieldResponse: - """Run shield. - - Run a shield. + """Run a shield. :param shield_id: The identifier of the shield to run. :param messages: The messages to run the shield on. @@ -124,9 +117,7 @@ class Safety(Protocol): @webmethod(route="/openai/v1/moderations", method="POST", level=LLAMA_STACK_API_V1, deprecated=True) @webmethod(route="/moderations", method="POST", level=LLAMA_STACK_API_V1) async def run_moderation(self, input: str | list[str], model: str) -> ModerationObject: - """Create moderation. - - Classifies if text and/or image inputs are potentially harmful. + """Classifies if text and/or image inputs are potentially harmful. :param input: Input (or inputs) to classify. Can be a single string, an array of strings, or an array of multi-modal input objects similar to other models. :param model: The content moderation model you would like to use.