diff --git a/llama_stack/distribution/library_client.py b/llama_stack/distribution/library_client.py index d4a7cde7e..2c0f73974 100644 --- a/llama_stack/distribution/library_client.py +++ b/llama_stack/distribution/library_client.py @@ -17,17 +17,6 @@ from typing import Any, get_args, get_origin, Optional, TypeVar import httpx import yaml -from llama_stack_client import ( - APIResponse, - AsyncAPIResponse, - AsyncLlamaStackClient, - AsyncStream, - LlamaStackClient, - NOT_GIVEN, -) -from pydantic import BaseModel, TypeAdapter -from rich.console import Console -from termcolor import cprint from llama_stack.distribution.build import print_pip_install_help from llama_stack.distribution.configure import parse_and_maybe_upgrade_config @@ -46,6 +35,17 @@ from llama_stack.providers.utils.telemetry.tracing import ( setup_logger, start_trace, ) +from llama_stack_client import ( + APIResponse, + AsyncAPIResponse, + AsyncLlamaStackClient, + AsyncStream, + LlamaStackClient, + NOT_GIVEN, +) +from pydantic import BaseModel, TypeAdapter +from rich.console import Console +from termcolor import cprint T = TypeVar("T") @@ -214,7 +214,7 @@ class AsyncLlamaStackAsLibraryClient(AsyncLlamaStackClient): f"Please run:\n\n{prefix}llama stack build --template {self.config_path_or_template_name} --image-type venv\n\n", "yellow", ) - return False + raise _e if Api.telemetry in self.impls: setup_logger(self.impls[Api.telemetry])