diff --git a/src/llama_stack/providers/inline/agents/meta_reference/responses/streaming.py b/src/llama_stack/providers/inline/agents/meta_reference/responses/streaming.py index 3515e0578..1a13b8129 100644 --- a/src/llama_stack/providers/inline/agents/meta_reference/responses/streaming.py +++ b/src/llama_stack/providers/inline/agents/meta_reference/responses/streaming.py @@ -71,6 +71,7 @@ from llama_stack_api import ( Safety, WebSearchToolTypes, ) +from opentelemetry import trace from .types import ChatCompletionContext, ChatCompletionResult from .utils import ( diff --git a/src/llama_stack/providers/inline/agents/meta_reference/responses/tool_executor.py b/src/llama_stack/providers/inline/agents/meta_reference/responses/tool_executor.py index 3822fb5b2..d59c1dc75 100644 --- a/src/llama_stack/providers/inline/agents/meta_reference/responses/tool_executor.py +++ b/src/llama_stack/providers/inline/agents/meta_reference/responses/tool_executor.py @@ -9,8 +9,6 @@ import json from collections.abc import AsyncIterator from typing import Any -from opentelemetry import trace - from llama_stack_api import ( ImageContentItem, OpenAIChatCompletionContentPartImageParam, @@ -38,6 +36,7 @@ from llama_stack_api import ( ToolRuntime, VectorIO, ) +from opentelemetry import trace from llama_stack.log import get_logger