From 26503ca1a4d470b9288381398ca0af0e02c764ba Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?S=C3=A9bastien=20Han?= Date: Thu, 20 Feb 2025 04:05:14 +0100 Subject: [PATCH] docs: fix Python llama_stack_client SDK links (#1150) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit # What does this PR do? It seems that the llama_stack_client repo and the main repo were originally the same, causing links to point to local references. We’ve now updated them to use the correct llama_stack_client repo links. Signed-off-by: Sébastien Han Signed-off-by: Sébastien Han --- .../references/python_sdk_reference/index.md | 148 +++++++++--------- 1 file changed, 74 insertions(+), 74 deletions(-) diff --git a/docs/source/references/python_sdk_reference/index.md b/docs/source/references/python_sdk_reference/index.md index 9d1130422..b1a9396fe 100644 --- a/docs/source/references/python_sdk_reference/index.md +++ b/docs/source/references/python_sdk_reference/index.md @@ -42,10 +42,10 @@ from llama_stack_client.types import ( Methods: -- client.toolgroups.list() -> ToolgroupListResponse -- client.toolgroups.get(toolgroup_id) -> ToolGroup -- client.toolgroups.register(\*\*params) -> None -- client.toolgroups.unregister(toolgroup_id) -> None +- client.toolgroups.list() -> ToolgroupListResponse +- client.toolgroups.get(toolgroup_id) -> ToolGroup +- client.toolgroups.register(\*\*params) -> None +- client.toolgroups.unregister(toolgroup_id) -> None ## Tools @@ -57,8 +57,8 @@ from llama_stack_client.types import ListToolsResponse, Tool, ToolListResponse Methods: -- client.tools.list(\*\*params) -> ToolListResponse -- client.tools.get(tool_name) -> Tool +- client.tools.list(\*\*params) -> ToolListResponse +- client.tools.get(tool_name) -> Tool ## ToolRuntime @@ -70,15 +70,15 @@ from llama_stack_client.types import ToolDef, ToolInvocationResult Methods: -- client.tool_runtime.invoke_tool(\*\*params) -> ToolInvocationResult -- client.tool_runtime.list_tools(\*\*params) -> JSONLDecoder[ToolDef] +- client.tool_runtime.invoke_tool(\*\*params) -> ToolInvocationResult +- client.tool_runtime.list_tools(\*\*params) -> JSONLDecoder[ToolDef] ### RagTool Methods: -- client.tool_runtime.rag_tool.insert(\*\*params) -> None -- client.tool_runtime.rag_tool.query(\*\*params) -> QueryResult +- client.tool_runtime.rag_tool.insert(\*\*params) -> None +- client.tool_runtime.rag_tool.query(\*\*params) -> QueryResult ## Agents @@ -97,8 +97,8 @@ from llama_stack_client.types import ( Methods: -- client.agents.create(\*\*params) -> AgentCreateResponse -- client.agents.delete(agent_id) -> None +- client.agents.create(\*\*params) -> AgentCreateResponse +- client.agents.delete(agent_id) -> None ### Session @@ -110,9 +110,9 @@ from llama_stack_client.types.agents import Session, SessionCreateResponse Methods: -- client.agents.session.create(agent_id, \*\*params) -> SessionCreateResponse -- client.agents.session.retrieve(session_id, \*, agent_id, \*\*params) -> Session -- client.agents.session.delete(session_id, \*, agent_id) -> None +- client.agents.session.create(agent_id, \*\*params) -> SessionCreateResponse +- client.agents.session.retrieve(session_id, \*, agent_id, \*\*params) -> Session +- client.agents.session.delete(session_id, \*, agent_id) -> None ### Steps @@ -124,7 +124,7 @@ from llama_stack_client.types.agents import StepRetrieveResponse Methods: -- client.agents.steps.retrieve(step_id, \*, agent_id, session_id, turn_id) -> StepRetrieveResponse +- client.agents.steps.retrieve(step_id, \*, agent_id, session_id, turn_id) -> StepRetrieveResponse ### Turn @@ -136,8 +136,8 @@ from llama_stack_client.types.agents import Turn, TurnCreateResponse Methods: -- client.agents.turn.create(session_id, \*, agent_id, \*\*params) -> TurnCreateResponse -- client.agents.turn.retrieve(turn_id, \*, agent_id, session_id) -> Turn +- client.agents.turn.create(session_id, \*, agent_id, \*\*params) -> TurnCreateResponse +- client.agents.turn.retrieve(turn_id, \*, agent_id, session_id) -> Turn ## BatchInference @@ -149,8 +149,8 @@ from llama_stack_client.types import BatchInferenceChatCompletionResponse Methods: -- client.batch_inference.chat_completion(\*\*params) -> BatchInferenceChatCompletionResponse -- client.batch_inference.completion(\*\*params) -> BatchCompletion +- client.batch_inference.chat_completion(\*\*params) -> BatchInferenceChatCompletionResponse +- client.batch_inference.completion(\*\*params) -> BatchCompletion ## Datasets @@ -166,10 +166,10 @@ from llama_stack_client.types import ( Methods: -- client.datasets.retrieve(dataset_id) -> Optional[DatasetRetrieveResponse] -- client.datasets.list() -> DatasetListResponse -- client.datasets.register(\*\*params) -> None -- client.datasets.unregister(dataset_id) -> None +- client.datasets.retrieve(dataset_id) -> Optional[DatasetRetrieveResponse] +- client.datasets.list() -> DatasetListResponse +- client.datasets.register(\*\*params) -> None +- client.datasets.unregister(dataset_id) -> None ## Eval @@ -181,8 +181,8 @@ from llama_stack_client.types import EvaluateResponse, Job Methods: -- client.eval.evaluate_rows(benchmark_id, \*\*params) -> EvaluateResponse -- client.eval.run_eval(benchmark_id, \*\*params) -> Job +- client.eval.evaluate_rows(benchmark_id, \*\*params) -> EvaluateResponse +- client.eval.run_eval(benchmark_id, \*\*params) -> Job ### Jobs @@ -194,9 +194,9 @@ from llama_stack_client.types.eval import JobStatusResponse Methods: -- client.eval.jobs.retrieve(job_id, \*, benchmark_id) -> EvaluateResponse -- client.eval.jobs.cancel(job_id, \*, benchmark_id) -> None -- client.eval.jobs.status(job_id, \*, benchmark_id) -> Optional[JobStatusResponse] +- client.eval.jobs.retrieve(job_id, \*, benchmark_id) -> EvaluateResponse +- client.eval.jobs.cancel(job_id, \*, benchmark_id) -> None +- client.eval.jobs.status(job_id, \*, benchmark_id) -> Optional[JobStatusResponse] ## Inspect @@ -208,8 +208,8 @@ from llama_stack_client.types import HealthInfo, ProviderInfo, RouteInfo, Versio Methods: -- client.inspect.health() -> HealthInfo -- client.inspect.version() -> VersionInfo +- client.inspect.health() -> HealthInfo +- client.inspect.version() -> VersionInfo ## Inference @@ -227,9 +227,9 @@ from llama_stack_client.types import ( Methods: -- client.inference.chat_completion(\*\*params) -> InferenceChatCompletionResponse -- client.inference.completion(\*\*params) -> InferenceCompletionResponse -- client.inference.embeddings(\*\*params) -> EmbeddingsResponse +- client.inference.chat_completion(\*\*params) -> InferenceChatCompletionResponse +- client.inference.completion(\*\*params) -> InferenceCompletionResponse +- client.inference.embeddings(\*\*params) -> EmbeddingsResponse ## VectorIo @@ -241,8 +241,8 @@ from llama_stack_client.types import QueryChunksResponse Methods: -- client.vector_io.insert(\*\*params) -> None -- client.vector_io.query(\*\*params) -> QueryChunksResponse +- client.vector_io.insert(\*\*params) -> None +- client.vector_io.query(\*\*params) -> QueryChunksResponse ## VectorDBs @@ -259,10 +259,10 @@ from llama_stack_client.types import ( Methods: -- client.vector_dbs.retrieve(vector_db_id) -> Optional[VectorDBRetrieveResponse] -- client.vector_dbs.list() -> VectorDBListResponse -- client.vector_dbs.register(\*\*params) -> VectorDBRegisterResponse -- client.vector_dbs.unregister(vector_db_id) -> None +- client.vector_dbs.retrieve(vector_db_id) -> Optional[VectorDBRetrieveResponse] +- client.vector_dbs.list() -> VectorDBListResponse +- client.vector_dbs.register(\*\*params) -> VectorDBRegisterResponse +- client.vector_dbs.unregister(vector_db_id) -> None ## Models @@ -274,10 +274,10 @@ from llama_stack_client.types import ListModelsResponse, Model, ModelListRespons Methods: -- client.models.retrieve(model_id) -> Optional[Model] -- client.models.list() -> ModelListResponse -- client.models.register(\*\*params) -> Model -- client.models.unregister(model_id) -> None +- client.models.retrieve(model_id) -> Optional[Model] +- client.models.list() -> ModelListResponse +- client.models.register(\*\*params) -> Model +- client.models.unregister(model_id) -> None ## PostTraining @@ -289,8 +289,8 @@ from llama_stack_client.types import ListPostTrainingJobsResponse, PostTrainingJ Methods: -- client.post_training.preference_optimize(\*\*params) -> PostTrainingJob -- client.post_training.supervised_fine_tune(\*\*params) -> PostTrainingJob +- client.post_training.preference_optimize(\*\*params) -> PostTrainingJob +- client.post_training.supervised_fine_tune(\*\*params) -> PostTrainingJob ### Job @@ -306,10 +306,10 @@ from llama_stack_client.types.post_training import ( Methods: -- client.post_training.job.list() -> JobListResponse -- client.post_training.job.artifacts(\*\*params) -> Optional[JobArtifactsResponse] -- client.post_training.job.cancel(\*\*params) -> None -- client.post_training.job.status(\*\*params) -> Optional[JobStatusResponse] +- client.post_training.job.list() -> JobListResponse +- client.post_training.job.artifacts(\*\*params) -> Optional[JobArtifactsResponse] +- client.post_training.job.cancel(\*\*params) -> None +- client.post_training.job.status(\*\*params) -> Optional[JobStatusResponse] ## Providers @@ -321,7 +321,7 @@ from llama_stack_client.types import ListProvidersResponse, ProviderListResponse Methods: -- client.providers.list() -> ProviderListResponse +- client.providers.list() -> ProviderListResponse ## Routes @@ -333,7 +333,7 @@ from llama_stack_client.types import ListRoutesResponse, RouteListResponse Methods: -- client.routes.list() -> RouteListResponse +- client.routes.list() -> RouteListResponse ## Safety @@ -345,7 +345,7 @@ from llama_stack_client.types import RunShieldResponse Methods: -- client.safety.run_shield(\*\*params) -> RunShieldResponse +- client.safety.run_shield(\*\*params) -> RunShieldResponse ## Shields @@ -357,9 +357,9 @@ from llama_stack_client.types import ListShieldsResponse, Shield, ShieldListResp Methods: -- client.shields.retrieve(identifier) -> Optional[Shield] -- client.shields.list() -> ShieldListResponse -- client.shields.register(\*\*params) -> Shield +- client.shields.retrieve(identifier) -> Optional[Shield] +- client.shields.list() -> ShieldListResponse +- client.shields.register(\*\*params) -> Shield ## SyntheticDataGeneration @@ -371,7 +371,7 @@ from llama_stack_client.types import SyntheticDataGenerationResponse Methods: -- client.synthetic_data_generation.generate(\*\*params) -> SyntheticDataGenerationResponse +- client.synthetic_data_generation.generate(\*\*params) -> SyntheticDataGenerationResponse ## Telemetry @@ -391,13 +391,13 @@ from llama_stack_client.types import ( Methods: -- client.telemetry.get_span(span_id, \*, trace_id) -> TelemetryGetSpanResponse -- client.telemetry.get_span_tree(span_id, \*\*params) -> TelemetryGetSpanTreeResponse -- client.telemetry.get_trace(trace_id) -> Trace -- client.telemetry.log_event(\*\*params) -> None -- client.telemetry.query_spans(\*\*params) -> TelemetryQuerySpansResponse -- client.telemetry.query_traces(\*\*params) -> TelemetryQueryTracesResponse -- client.telemetry.save_spans_to_dataset(\*\*params) -> None +- client.telemetry.get_span(span_id, \*, trace_id) -> TelemetryGetSpanResponse +- client.telemetry.get_span_tree(span_id, \*\*params) -> TelemetryGetSpanTreeResponse +- client.telemetry.get_trace(trace_id) -> Trace +- client.telemetry.log_event(\*\*params) -> None +- client.telemetry.query_spans(\*\*params) -> TelemetryQuerySpansResponse +- client.telemetry.query_traces(\*\*params) -> TelemetryQueryTracesResponse +- client.telemetry.save_spans_to_dataset(\*\*params) -> None ## Datasetio @@ -409,8 +409,8 @@ from llama_stack_client.types import PaginatedRowsResult Methods: -- client.datasetio.append_rows(\*\*params) -> None -- client.datasetio.get_rows_paginated(\*\*params) -> PaginatedRowsResult +- client.datasetio.append_rows(\*\*params) -> None +- client.datasetio.get_rows_paginated(\*\*params) -> PaginatedRowsResult ## Scoring @@ -422,8 +422,8 @@ from llama_stack_client.types import ScoringScoreResponse, ScoringScoreBatchResp Methods: -- client.scoring.score(\*\*params) -> ScoringScoreResponse -- client.scoring.score_batch(\*\*params) -> ScoringScoreBatchResponse +- client.scoring.score(\*\*params) -> ScoringScoreResponse +- client.scoring.score_batch(\*\*params) -> ScoringScoreBatchResponse ## ScoringFunctions @@ -439,9 +439,9 @@ from llama_stack_client.types import ( Methods: -- client.scoring_functions.retrieve(scoring_fn_id) -> Optional[ScoringFn] -- client.scoring_functions.list() -> ScoringFunctionListResponse -- client.scoring_functions.register(\*\*params) -> None +- client.scoring_functions.retrieve(scoring_fn_id) -> Optional[ScoringFn] +- client.scoring_functions.list() -> ScoringFunctionListResponse +- client.scoring_functions.register(\*\*params) -> None ## Benchmarks @@ -457,6 +457,6 @@ from llama_stack_client.types import ( Methods: -- client.benchmarks.retrieve(benchmark_id) -> Optional[Benchmark] -- client.benchmarks.list() -> BenchmarkListResponse -- client.benchmarks.register(\*\*params) -> None +- client.benchmarks.retrieve(benchmark_id) -> Optional[Benchmark] +- client.benchmarks.list() -> BenchmarkListResponse +- client.benchmarks.register(\*\*params) -> None