diff --git a/docs/source/references/python_sdk_reference/index.md b/docs/source/references/python_sdk_reference/index.md
index 9d1130422..b1a9396fe 100644
--- a/docs/source/references/python_sdk_reference/index.md
+++ b/docs/source/references/python_sdk_reference/index.md
@@ -42,10 +42,10 @@ from llama_stack_client.types import (
Methods:
-- client.toolgroups.list() -> ToolgroupListResponse
-- client.toolgroups.get(toolgroup_id) -> ToolGroup
-- client.toolgroups.register(\*\*params) -> None
-- client.toolgroups.unregister(toolgroup_id) -> None
+- client.toolgroups.list() -> ToolgroupListResponse
+- client.toolgroups.get(toolgroup_id) -> ToolGroup
+- client.toolgroups.register(\*\*params) -> None
+- client.toolgroups.unregister(toolgroup_id) -> None
## Tools
@@ -57,8 +57,8 @@ from llama_stack_client.types import ListToolsResponse, Tool, ToolListResponse
Methods:
-- client.tools.list(\*\*params) -> ToolListResponse
-- client.tools.get(tool_name) -> Tool
+- client.tools.list(\*\*params) -> ToolListResponse
+- client.tools.get(tool_name) -> Tool
## ToolRuntime
@@ -70,15 +70,15 @@ from llama_stack_client.types import ToolDef, ToolInvocationResult
Methods:
-- client.tool_runtime.invoke_tool(\*\*params) -> ToolInvocationResult
-- client.tool_runtime.list_tools(\*\*params) -> JSONLDecoder[ToolDef]
+- client.tool_runtime.invoke_tool(\*\*params) -> ToolInvocationResult
+- client.tool_runtime.list_tools(\*\*params) -> JSONLDecoder[ToolDef]
### RagTool
Methods:
-- client.tool_runtime.rag_tool.insert(\*\*params) -> None
-- client.tool_runtime.rag_tool.query(\*\*params) -> QueryResult
+- client.tool_runtime.rag_tool.insert(\*\*params) -> None
+- client.tool_runtime.rag_tool.query(\*\*params) -> QueryResult
## Agents
@@ -97,8 +97,8 @@ from llama_stack_client.types import (
Methods:
-- client.agents.create(\*\*params) -> AgentCreateResponse
-- client.agents.delete(agent_id) -> None
+- client.agents.create(\*\*params) -> AgentCreateResponse
+- client.agents.delete(agent_id) -> None
### Session
@@ -110,9 +110,9 @@ from llama_stack_client.types.agents import Session, SessionCreateResponse
Methods:
-- client.agents.session.create(agent_id, \*\*params) -> SessionCreateResponse
-- client.agents.session.retrieve(session_id, \*, agent_id, \*\*params) -> Session
-- client.agents.session.delete(session_id, \*, agent_id) -> None
+- client.agents.session.create(agent_id, \*\*params) -> SessionCreateResponse
+- client.agents.session.retrieve(session_id, \*, agent_id, \*\*params) -> Session
+- client.agents.session.delete(session_id, \*, agent_id) -> None
### Steps
@@ -124,7 +124,7 @@ from llama_stack_client.types.agents import StepRetrieveResponse
Methods:
-- client.agents.steps.retrieve(step_id, \*, agent_id, session_id, turn_id) -> StepRetrieveResponse
+- client.agents.steps.retrieve(step_id, \*, agent_id, session_id, turn_id) -> StepRetrieveResponse
### Turn
@@ -136,8 +136,8 @@ from llama_stack_client.types.agents import Turn, TurnCreateResponse
Methods:
-- client.agents.turn.create(session_id, \*, agent_id, \*\*params) -> TurnCreateResponse
-- client.agents.turn.retrieve(turn_id, \*, agent_id, session_id) -> Turn
+- client.agents.turn.create(session_id, \*, agent_id, \*\*params) -> TurnCreateResponse
+- client.agents.turn.retrieve(turn_id, \*, agent_id, session_id) -> Turn
## BatchInference
@@ -149,8 +149,8 @@ from llama_stack_client.types import BatchInferenceChatCompletionResponse
Methods:
-- client.batch_inference.chat_completion(\*\*params) -> BatchInferenceChatCompletionResponse
-- client.batch_inference.completion(\*\*params) -> BatchCompletion
+- client.batch_inference.chat_completion(\*\*params) -> BatchInferenceChatCompletionResponse
+- client.batch_inference.completion(\*\*params) -> BatchCompletion
## Datasets
@@ -166,10 +166,10 @@ from llama_stack_client.types import (
Methods:
-- client.datasets.retrieve(dataset_id) -> Optional[DatasetRetrieveResponse]
-- client.datasets.list() -> DatasetListResponse
-- client.datasets.register(\*\*params) -> None
-- client.datasets.unregister(dataset_id) -> None
+- client.datasets.retrieve(dataset_id) -> Optional[DatasetRetrieveResponse]
+- client.datasets.list() -> DatasetListResponse
+- client.datasets.register(\*\*params) -> None
+- client.datasets.unregister(dataset_id) -> None
## Eval
@@ -181,8 +181,8 @@ from llama_stack_client.types import EvaluateResponse, Job
Methods:
-- client.eval.evaluate_rows(benchmark_id, \*\*params) -> EvaluateResponse
-- client.eval.run_eval(benchmark_id, \*\*params) -> Job
+- client.eval.evaluate_rows(benchmark_id, \*\*params) -> EvaluateResponse
+- client.eval.run_eval(benchmark_id, \*\*params) -> Job
### Jobs
@@ -194,9 +194,9 @@ from llama_stack_client.types.eval import JobStatusResponse
Methods:
-- client.eval.jobs.retrieve(job_id, \*, benchmark_id) -> EvaluateResponse
-- client.eval.jobs.cancel(job_id, \*, benchmark_id) -> None
-- client.eval.jobs.status(job_id, \*, benchmark_id) -> Optional[JobStatusResponse]
+- client.eval.jobs.retrieve(job_id, \*, benchmark_id) -> EvaluateResponse
+- client.eval.jobs.cancel(job_id, \*, benchmark_id) -> None
+- client.eval.jobs.status(job_id, \*, benchmark_id) -> Optional[JobStatusResponse]
## Inspect
@@ -208,8 +208,8 @@ from llama_stack_client.types import HealthInfo, ProviderInfo, RouteInfo, Versio
Methods:
-- client.inspect.health() -> HealthInfo
-- client.inspect.version() -> VersionInfo
+- client.inspect.health() -> HealthInfo
+- client.inspect.version() -> VersionInfo
## Inference
@@ -227,9 +227,9 @@ from llama_stack_client.types import (
Methods:
-- client.inference.chat_completion(\*\*params) -> InferenceChatCompletionResponse
-- client.inference.completion(\*\*params) -> InferenceCompletionResponse
-- client.inference.embeddings(\*\*params) -> EmbeddingsResponse
+- client.inference.chat_completion(\*\*params) -> InferenceChatCompletionResponse
+- client.inference.completion(\*\*params) -> InferenceCompletionResponse
+- client.inference.embeddings(\*\*params) -> EmbeddingsResponse
## VectorIo
@@ -241,8 +241,8 @@ from llama_stack_client.types import QueryChunksResponse
Methods:
-- client.vector_io.insert(\*\*params) -> None
-- client.vector_io.query(\*\*params) -> QueryChunksResponse
+- client.vector_io.insert(\*\*params) -> None
+- client.vector_io.query(\*\*params) -> QueryChunksResponse
## VectorDBs
@@ -259,10 +259,10 @@ from llama_stack_client.types import (
Methods:
-- client.vector_dbs.retrieve(vector_db_id) -> Optional[VectorDBRetrieveResponse]
-- client.vector_dbs.list() -> VectorDBListResponse
-- client.vector_dbs.register(\*\*params) -> VectorDBRegisterResponse
-- client.vector_dbs.unregister(vector_db_id) -> None
+- client.vector_dbs.retrieve(vector_db_id) -> Optional[VectorDBRetrieveResponse]
+- client.vector_dbs.list() -> VectorDBListResponse
+- client.vector_dbs.register(\*\*params) -> VectorDBRegisterResponse
+- client.vector_dbs.unregister(vector_db_id) -> None
## Models
@@ -274,10 +274,10 @@ from llama_stack_client.types import ListModelsResponse, Model, ModelListRespons
Methods:
-- client.models.retrieve(model_id) -> Optional[Model]
-- client.models.list() -> ModelListResponse
-- client.models.register(\*\*params) -> Model
-- client.models.unregister(model_id) -> None
+- client.models.retrieve(model_id) -> Optional[Model]
+- client.models.list() -> ModelListResponse
+- client.models.register(\*\*params) -> Model
+- client.models.unregister(model_id) -> None
## PostTraining
@@ -289,8 +289,8 @@ from llama_stack_client.types import ListPostTrainingJobsResponse, PostTrainingJ
Methods:
-- client.post_training.preference_optimize(\*\*params) -> PostTrainingJob
-- client.post_training.supervised_fine_tune(\*\*params) -> PostTrainingJob
+- client.post_training.preference_optimize(\*\*params) -> PostTrainingJob
+- client.post_training.supervised_fine_tune(\*\*params) -> PostTrainingJob
### Job
@@ -306,10 +306,10 @@ from llama_stack_client.types.post_training import (
Methods:
-- client.post_training.job.list() -> JobListResponse
-- client.post_training.job.artifacts(\*\*params) -> Optional[JobArtifactsResponse]
-- client.post_training.job.cancel(\*\*params) -> None
-- client.post_training.job.status(\*\*params) -> Optional[JobStatusResponse]
+- client.post_training.job.list() -> JobListResponse
+- client.post_training.job.artifacts(\*\*params) -> Optional[JobArtifactsResponse]
+- client.post_training.job.cancel(\*\*params) -> None
+- client.post_training.job.status(\*\*params) -> Optional[JobStatusResponse]
## Providers
@@ -321,7 +321,7 @@ from llama_stack_client.types import ListProvidersResponse, ProviderListResponse
Methods:
-- client.providers.list() -> ProviderListResponse
+- client.providers.list() -> ProviderListResponse
## Routes
@@ -333,7 +333,7 @@ from llama_stack_client.types import ListRoutesResponse, RouteListResponse
Methods:
-- client.routes.list() -> RouteListResponse
+- client.routes.list() -> RouteListResponse
## Safety
@@ -345,7 +345,7 @@ from llama_stack_client.types import RunShieldResponse
Methods:
-- client.safety.run_shield(\*\*params) -> RunShieldResponse
+- client.safety.run_shield(\*\*params) -> RunShieldResponse
## Shields
@@ -357,9 +357,9 @@ from llama_stack_client.types import ListShieldsResponse, Shield, ShieldListResp
Methods:
-- client.shields.retrieve(identifier) -> Optional[Shield]
-- client.shields.list() -> ShieldListResponse
-- client.shields.register(\*\*params) -> Shield
+- client.shields.retrieve(identifier) -> Optional[Shield]
+- client.shields.list() -> ShieldListResponse
+- client.shields.register(\*\*params) -> Shield
## SyntheticDataGeneration
@@ -371,7 +371,7 @@ from llama_stack_client.types import SyntheticDataGenerationResponse
Methods:
-- client.synthetic_data_generation.generate(\*\*params) -> SyntheticDataGenerationResponse
+- client.synthetic_data_generation.generate(\*\*params) -> SyntheticDataGenerationResponse
## Telemetry
@@ -391,13 +391,13 @@ from llama_stack_client.types import (
Methods:
-- client.telemetry.get_span(span_id, \*, trace_id) -> TelemetryGetSpanResponse
-- client.telemetry.get_span_tree(span_id, \*\*params) -> TelemetryGetSpanTreeResponse
-- client.telemetry.get_trace(trace_id) -> Trace
-- client.telemetry.log_event(\*\*params) -> None
-- client.telemetry.query_spans(\*\*params) -> TelemetryQuerySpansResponse
-- client.telemetry.query_traces(\*\*params) -> TelemetryQueryTracesResponse
-- client.telemetry.save_spans_to_dataset(\*\*params) -> None
+- client.telemetry.get_span(span_id, \*, trace_id) -> TelemetryGetSpanResponse
+- client.telemetry.get_span_tree(span_id, \*\*params) -> TelemetryGetSpanTreeResponse
+- client.telemetry.get_trace(trace_id) -> Trace
+- client.telemetry.log_event(\*\*params) -> None
+- client.telemetry.query_spans(\*\*params) -> TelemetryQuerySpansResponse
+- client.telemetry.query_traces(\*\*params) -> TelemetryQueryTracesResponse
+- client.telemetry.save_spans_to_dataset(\*\*params) -> None
## Datasetio
@@ -409,8 +409,8 @@ from llama_stack_client.types import PaginatedRowsResult
Methods:
-- client.datasetio.append_rows(\*\*params) -> None
-- client.datasetio.get_rows_paginated(\*\*params) -> PaginatedRowsResult
+- client.datasetio.append_rows(\*\*params) -> None
+- client.datasetio.get_rows_paginated(\*\*params) -> PaginatedRowsResult
## Scoring
@@ -422,8 +422,8 @@ from llama_stack_client.types import ScoringScoreResponse, ScoringScoreBatchResp
Methods:
-- client.scoring.score(\*\*params) -> ScoringScoreResponse
-- client.scoring.score_batch(\*\*params) -> ScoringScoreBatchResponse
+- client.scoring.score(\*\*params) -> ScoringScoreResponse
+- client.scoring.score_batch(\*\*params) -> ScoringScoreBatchResponse
## ScoringFunctions
@@ -439,9 +439,9 @@ from llama_stack_client.types import (
Methods:
-- client.scoring_functions.retrieve(scoring_fn_id) -> Optional[ScoringFn]
-- client.scoring_functions.list() -> ScoringFunctionListResponse
-- client.scoring_functions.register(\*\*params) -> None
+- client.scoring_functions.retrieve(scoring_fn_id) -> Optional[ScoringFn]
+- client.scoring_functions.list() -> ScoringFunctionListResponse
+- client.scoring_functions.register(\*\*params) -> None
## Benchmarks
@@ -457,6 +457,6 @@ from llama_stack_client.types import (
Methods:
-- client.benchmarks.retrieve(benchmark_id) -> Optional[Benchmark]
-- client.benchmarks.list() -> BenchmarkListResponse
-- client.benchmarks.register(\*\*params) -> None
+- client.benchmarks.retrieve(benchmark_id) -> Optional[Benchmark]
+- client.benchmarks.list() -> BenchmarkListResponse
+- client.benchmarks.register(\*\*params) -> None