mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
PR #201 had made several changes while trying to fix issues with getting the stream=False branches of inference and agents API working. As part of this, it made a change which was slightly gratuitous. Namely, making chat_completion() and brethren "def" instead of "async def". The rationale was that this allowed the user (within llama-stack) of this to use it as: ``` async for chunk in api.chat_completion(params) ``` However, it causes unnecessary confusion for several folks. Given that clients (e.g., llama-stack-apps) anyway use the SDK methods (which are completely isolated) this choice was not ideal. Let's revert back so the call now looks like: ``` async for chunk in await api.chat_completion(params) ``` Bonus: Added a completion() implementation for the meta-reference provider. Technically should have been another PR :)
311 lines
9.5 KiB
Python
311 lines
9.5 KiB
Python
# Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
# All rights reserved.
|
|
#
|
|
# This source code is licensed under the terms described in the LICENSE file in
|
|
# the root directory of this source tree.
|
|
|
|
import os
|
|
|
|
import pytest
|
|
import pytest_asyncio
|
|
|
|
from llama_stack.apis.agents import * # noqa: F403
|
|
from llama_stack.providers.tests.resolver import resolve_impls_for_test
|
|
from llama_stack.providers.datatypes import * # noqa: F403
|
|
|
|
from dotenv import load_dotenv
|
|
|
|
# How to run this test:
|
|
#
|
|
# 1. Ensure you have a conda environment with the right dependencies installed.
|
|
# This includes `pytest` and `pytest-asyncio`.
|
|
#
|
|
# 2. Copy and modify the provider_config_example.yaml depending on the provider you are testing.
|
|
#
|
|
# 3. Run:
|
|
#
|
|
# ```bash
|
|
# PROVIDER_ID=<your_provider> \
|
|
# PROVIDER_CONFIG=provider_config.yaml \
|
|
# pytest -s llama_stack/providers/tests/agents/test_agents.py \
|
|
# --tb=short --disable-warnings
|
|
# ```
|
|
|
|
load_dotenv()
|
|
|
|
|
|
@pytest_asyncio.fixture(scope="session")
|
|
async def agents_settings():
|
|
impls = await resolve_impls_for_test(
|
|
Api.agents, deps=[Api.inference, Api.memory, Api.safety]
|
|
)
|
|
|
|
return {
|
|
"impl": impls[Api.agents],
|
|
"memory_impl": impls[Api.memory],
|
|
"common_params": {
|
|
"model": "Llama3.1-8B-Instruct",
|
|
"instructions": "You are a helpful assistant.",
|
|
},
|
|
}
|
|
|
|
|
|
@pytest.fixture
|
|
def sample_messages():
|
|
return [
|
|
UserMessage(content="What's the weather like today?"),
|
|
]
|
|
|
|
|
|
@pytest.fixture
|
|
def search_query_messages():
|
|
return [
|
|
UserMessage(content="What are the latest developments in quantum computing?"),
|
|
]
|
|
|
|
|
|
@pytest.fixture
|
|
def attachment_message():
|
|
return [
|
|
UserMessage(
|
|
content="I am attaching some documentation for Torchtune. Help me answer questions I will ask next.",
|
|
),
|
|
]
|
|
|
|
|
|
@pytest.fixture
|
|
def query_attachment_messages():
|
|
return [
|
|
UserMessage(
|
|
content="What are the top 5 topics that were explained? Only list succinct bullet points."
|
|
),
|
|
]
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_create_agent_turn(agents_settings, sample_messages):
|
|
agents_impl = agents_settings["impl"]
|
|
|
|
# First, create an agent
|
|
agent_config = AgentConfig(
|
|
model=agents_settings["common_params"]["model"],
|
|
instructions=agents_settings["common_params"]["instructions"],
|
|
enable_session_persistence=True,
|
|
sampling_params=SamplingParams(temperature=0.7, top_p=0.95),
|
|
input_shields=[],
|
|
output_shields=[],
|
|
tools=[],
|
|
max_infer_iters=5,
|
|
)
|
|
|
|
create_response = await agents_impl.create_agent(agent_config)
|
|
agent_id = create_response.agent_id
|
|
|
|
# Create a session
|
|
session_create_response = await agents_impl.create_agent_session(
|
|
agent_id, "Test Session"
|
|
)
|
|
session_id = session_create_response.session_id
|
|
|
|
# Create and execute a turn
|
|
turn_request = dict(
|
|
agent_id=agent_id,
|
|
session_id=session_id,
|
|
messages=sample_messages,
|
|
stream=True,
|
|
)
|
|
|
|
turn_response = [
|
|
chunk async for chunk in await agents_impl.create_agent_turn(**turn_request)
|
|
]
|
|
|
|
assert len(turn_response) > 0
|
|
assert all(
|
|
isinstance(chunk, AgentTurnResponseStreamChunk) for chunk in turn_response
|
|
)
|
|
|
|
# Check for expected event types
|
|
event_types = [chunk.event.payload.event_type for chunk in turn_response]
|
|
assert AgentTurnResponseEventType.turn_start.value in event_types
|
|
assert AgentTurnResponseEventType.step_start.value in event_types
|
|
assert AgentTurnResponseEventType.step_complete.value in event_types
|
|
assert AgentTurnResponseEventType.turn_complete.value in event_types
|
|
|
|
# Check the final turn complete event
|
|
final_event = turn_response[-1].event.payload
|
|
assert isinstance(final_event, AgentTurnResponseTurnCompletePayload)
|
|
assert isinstance(final_event.turn, Turn)
|
|
assert final_event.turn.session_id == session_id
|
|
assert final_event.turn.input_messages == sample_messages
|
|
assert isinstance(final_event.turn.output_message, CompletionMessage)
|
|
assert len(final_event.turn.output_message.content) > 0
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_rag_agent_as_attachments(
|
|
agents_settings, attachment_message, query_attachment_messages
|
|
):
|
|
urls = [
|
|
"memory_optimizations.rst",
|
|
"chat.rst",
|
|
"llama3.rst",
|
|
"datasets.rst",
|
|
"qat_finetune.rst",
|
|
"lora_finetune.rst",
|
|
]
|
|
|
|
attachments = [
|
|
Attachment(
|
|
content=f"https://raw.githubusercontent.com/pytorch/torchtune/main/docs/source/tutorials/{url}",
|
|
mime_type="text/plain",
|
|
)
|
|
for i, url in enumerate(urls)
|
|
]
|
|
|
|
agents_impl = agents_settings["impl"]
|
|
|
|
agent_config = AgentConfig(
|
|
model=agents_settings["common_params"]["model"],
|
|
instructions=agents_settings["common_params"]["instructions"],
|
|
enable_session_persistence=True,
|
|
sampling_params=SamplingParams(temperature=0.7, top_p=0.95),
|
|
input_shields=[],
|
|
output_shields=[],
|
|
tools=[
|
|
MemoryToolDefinition(
|
|
memory_bank_configs=[],
|
|
query_generator_config={
|
|
"type": "default",
|
|
"sep": " ",
|
|
},
|
|
max_tokens_in_context=4096,
|
|
max_chunks=10,
|
|
),
|
|
],
|
|
max_infer_iters=5,
|
|
)
|
|
|
|
create_response = await agents_impl.create_agent(agent_config)
|
|
agent_id = create_response.agent_id
|
|
|
|
# Create a session
|
|
session_create_response = await agents_impl.create_agent_session(
|
|
agent_id, "Test Session"
|
|
)
|
|
session_id = session_create_response.session_id
|
|
|
|
# Create and execute a turn
|
|
turn_request = dict(
|
|
agent_id=agent_id,
|
|
session_id=session_id,
|
|
messages=attachment_message,
|
|
attachments=attachments,
|
|
stream=True,
|
|
)
|
|
|
|
turn_response = [
|
|
chunk async for chunk in await agents_impl.create_agent_turn(**turn_request)
|
|
]
|
|
|
|
assert len(turn_response) > 0
|
|
|
|
# Create a second turn querying the agent
|
|
turn_request = dict(
|
|
agent_id=agent_id,
|
|
session_id=session_id,
|
|
messages=query_attachment_messages,
|
|
stream=True,
|
|
)
|
|
|
|
turn_response = [
|
|
chunk async for chunk in await agents_impl.create_agent_turn(**turn_request)
|
|
]
|
|
|
|
assert len(turn_response) > 0
|
|
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_create_agent_turn_with_brave_search(
|
|
agents_settings, search_query_messages
|
|
):
|
|
agents_impl = agents_settings["impl"]
|
|
|
|
if "BRAVE_SEARCH_API_KEY" not in os.environ:
|
|
pytest.skip("BRAVE_SEARCH_API_KEY not set, skipping test")
|
|
|
|
# Create an agent with Brave search tool
|
|
agent_config = AgentConfig(
|
|
model=agents_settings["common_params"]["model"],
|
|
instructions=agents_settings["common_params"]["instructions"],
|
|
enable_session_persistence=True,
|
|
sampling_params=SamplingParams(temperature=0.7, top_p=0.95),
|
|
input_shields=[],
|
|
output_shields=[],
|
|
tools=[
|
|
SearchToolDefinition(
|
|
type=AgentTool.brave_search.value,
|
|
api_key=os.environ["BRAVE_SEARCH_API_KEY"],
|
|
engine=SearchEngineType.brave,
|
|
)
|
|
],
|
|
tool_choice=ToolChoice.auto,
|
|
max_infer_iters=5,
|
|
)
|
|
|
|
create_response = await agents_impl.create_agent(agent_config)
|
|
agent_id = create_response.agent_id
|
|
|
|
# Create a session
|
|
session_create_response = await agents_impl.create_agent_session(
|
|
agent_id, "Test Session with Brave Search"
|
|
)
|
|
session_id = session_create_response.session_id
|
|
|
|
# Create and execute a turn
|
|
turn_request = dict(
|
|
agent_id=agent_id,
|
|
session_id=session_id,
|
|
messages=search_query_messages,
|
|
stream=True,
|
|
)
|
|
|
|
turn_response = [
|
|
chunk async for chunk in await agents_impl.create_agent_turn(**turn_request)
|
|
]
|
|
|
|
assert len(turn_response) > 0
|
|
assert all(
|
|
isinstance(chunk, AgentTurnResponseStreamChunk) for chunk in turn_response
|
|
)
|
|
|
|
# Check for expected event types
|
|
event_types = [chunk.event.payload.event_type for chunk in turn_response]
|
|
assert AgentTurnResponseEventType.turn_start.value in event_types
|
|
assert AgentTurnResponseEventType.step_start.value in event_types
|
|
assert AgentTurnResponseEventType.step_complete.value in event_types
|
|
assert AgentTurnResponseEventType.turn_complete.value in event_types
|
|
|
|
# Check for tool execution events
|
|
tool_execution_events = [
|
|
chunk
|
|
for chunk in turn_response
|
|
if isinstance(chunk.event.payload, AgentTurnResponseStepCompletePayload)
|
|
and chunk.event.payload.step_details.step_type == StepType.tool_execution.value
|
|
]
|
|
assert len(tool_execution_events) > 0, "No tool execution events found"
|
|
|
|
# Check the tool execution details
|
|
tool_execution = tool_execution_events[0].event.payload.step_details
|
|
assert isinstance(tool_execution, ToolExecutionStep)
|
|
assert len(tool_execution.tool_calls) > 0
|
|
assert tool_execution.tool_calls[0].tool_name == BuiltinTool.brave_search
|
|
assert len(tool_execution.tool_responses) > 0
|
|
|
|
# Check the final turn complete event
|
|
final_event = turn_response[-1].event.payload
|
|
assert isinstance(final_event, AgentTurnResponseTurnCompletePayload)
|
|
assert isinstance(final_event.turn, Turn)
|
|
assert final_event.turn.session_id == session_id
|
|
assert final_event.turn.input_messages == search_query_messages
|
|
assert isinstance(final_event.turn.output_message, CompletionMessage)
|
|
assert len(final_event.turn.output_message.content) > 0
|