From ee4e04804fefeab62870c87177b7bae3f9cbeaf0 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Mon, 13 Jan 2025 19:11:51 -0800 Subject: [PATCH] Rename ipython to tool (#756) See https://github.com/meta-llama/llama-models/pull/261 for the corresponding PR in llama-models. Once these PRs land, you need to work `main` from llama-models (vs. from pypi) --- llama_stack/apis/inference/inference.py | 2 +- llama_stack/providers/remote/inference/nvidia/openai_utils.py | 2 +- tests/client-sdk/agents/test_agents.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/llama_stack/apis/inference/inference.py b/llama_stack/apis/inference/inference.py index a6a096041..4a453700c 100644 --- a/llama_stack/apis/inference/inference.py +++ b/llama_stack/apis/inference/inference.py @@ -82,7 +82,7 @@ class SystemMessage(BaseModel): @json_schema_type class ToolResponseMessage(BaseModel): - role: Literal["ipython"] = "ipython" + role: Literal["tool"] = "tool" # it was nice to re-use the ToolResponse type, but having all messages # have a `content` type makes things nicer too call_id: str diff --git a/llama_stack/providers/remote/inference/nvidia/openai_utils.py b/llama_stack/providers/remote/inference/nvidia/openai_utils.py index ffca32c44..dcc7c5fca 100644 --- a/llama_stack/providers/remote/inference/nvidia/openai_utils.py +++ b/llama_stack/providers/remote/inference/nvidia/openai_utils.py @@ -144,7 +144,7 @@ def _convert_message(message: Message | Dict) -> OpenAIChatCompletionMessage: message = UserMessage(**message) elif message["role"] == "assistant": message = CompletionMessage(**message) - elif message["role"] == "ipython": + elif message["role"] == "tool": message = ToolResponseMessage(**message) elif message["role"] == "system": message = SystemMessage(**message) diff --git a/tests/client-sdk/agents/test_agents.py b/tests/client-sdk/agents/test_agents.py index a2ed687a4..0c16b6225 100644 --- a/tests/client-sdk/agents/test_agents.py +++ b/tests/client-sdk/agents/test_agents.py @@ -40,10 +40,10 @@ class TestClientTool(ClientTool): response_str = f"Error when running tool: {e}" message = ToolResponseMessage( + role="tool", call_id=tool_call.call_id, tool_name=tool_call.tool_name, content=response_str, - role="ipython", ) return [message]