llama-stack-mirror/llama_stack/providers/inline/agents/meta_reference
ehhuang 6cf79437b3
feat: support ClientTool output metadata (#1426)
# Summary:
Client side change in
https://github.com/meta-llama/llama-stack-client-python/pull/180
Changes the resume_turn API to accept `ToolResponse` instead of
`ToolResponseMessage`:
1. `ToolResponse` contains `metadata`
2. `ToolResponseMessage` is a concept for model inputs. Here we are just
submitting the outputs of tool execution.

# Test Plan:
Ran integration tests with newly added test using client tool with
metadata

LLAMA_STACK_CONFIG=fireworks pytest -s -v
tests/integration/agents/test_agents.py --safety-shield
meta-llama/Llama-Guard-3-8B --record-responses
2025-03-05 14:30:27 -08:00
..
tests fix: Agent uses the first configured vector_db_id when documents are provided (#1276) 2025-03-04 21:44:13 -08:00
__init__.py Fix precommit check after moving to ruff (#927) 2025-02-02 06:46:45 -08:00
agent_instance.py feat: support ClientTool output metadata (#1426) 2025-03-05 14:30:27 -08:00
agents.py feat: support ClientTool output metadata (#1426) 2025-03-05 14:30:27 -08:00
config.py Auto-generate distro yamls + docs (#468) 2024-11-18 14:57:06 -08:00
persistence.py feat: unify max_infer_iters in client/server agent loop (#1309) 2025-03-03 10:08:36 -08:00
safety.py build: configure ruff from pyproject.toml (#1100) 2025-02-14 09:01:57 -08:00