mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-03 01:03:59 +00:00
rename response to responses in verifications, update provider
This commit is contained in:
parent
78da66016f
commit
ae012bb857
3 changed files with 13 additions and 13 deletions
|
@ -1,8 +1,8 @@
|
||||||
version: '2'
|
version: '2'
|
||||||
image_name: openai-api-verification
|
image_name: openai-api-verification
|
||||||
apis:
|
apis:
|
||||||
|
- agents
|
||||||
- inference
|
- inference
|
||||||
- openai_responses
|
|
||||||
- telemetry
|
- telemetry
|
||||||
- tool_runtime
|
- tool_runtime
|
||||||
- vector_io
|
- vector_io
|
||||||
|
@ -46,14 +46,14 @@ providers:
|
||||||
service_name: "${env.OTEL_SERVICE_NAME:\u200B}"
|
service_name: "${env.OTEL_SERVICE_NAME:\u200B}"
|
||||||
sinks: ${env.TELEMETRY_SINKS:console,sqlite}
|
sinks: ${env.TELEMETRY_SINKS:console,sqlite}
|
||||||
sqlite_db_path: ${env.SQLITE_DB_PATH:~/.llama/distributions/openai/trace_store.db}
|
sqlite_db_path: ${env.SQLITE_DB_PATH:~/.llama/distributions/openai/trace_store.db}
|
||||||
openai_responses:
|
agents:
|
||||||
- provider_id: openai-responses
|
- provider_id: meta-reference
|
||||||
provider_type: inline::openai-responses
|
provider_type: inline::meta-reference
|
||||||
config:
|
config:
|
||||||
kvstore:
|
persistence_store:
|
||||||
type: sqlite
|
type: sqlite
|
||||||
namespace: null
|
namespace: null
|
||||||
db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/openai}/openai_responses.db
|
db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/openai}/agents_store.db
|
||||||
tool_runtime:
|
tool_runtime:
|
||||||
- provider_id: brave-search
|
- provider_id: brave-search
|
||||||
provider_type: remote::brave-search
|
provider_type: remote::brave-search
|
||||||
|
|
|
@ -14,12 +14,12 @@ from tests.verifications.openai_api.fixtures.fixtures import (
|
||||||
)
|
)
|
||||||
from tests.verifications.openai_api.fixtures.load import load_test_cases
|
from tests.verifications.openai_api.fixtures.load import load_test_cases
|
||||||
|
|
||||||
response_test_cases = load_test_cases("response")
|
responses_test_cases = load_test_cases("responses")
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize(
|
@pytest.mark.parametrize(
|
||||||
"case",
|
"case",
|
||||||
response_test_cases["test_response_basic"]["test_params"]["case"],
|
responses_test_cases["test_response_basic"]["test_params"]["case"],
|
||||||
ids=case_id_generator,
|
ids=case_id_generator,
|
||||||
)
|
)
|
||||||
def test_response_non_streaming_basic(request, openai_client, model, provider, verification_config, case):
|
def test_response_non_streaming_basic(request, openai_client, model, provider, verification_config, case):
|
||||||
|
@ -48,7 +48,7 @@ def test_response_non_streaming_basic(request, openai_client, model, provider, v
|
||||||
|
|
||||||
@pytest.mark.parametrize(
|
@pytest.mark.parametrize(
|
||||||
"case",
|
"case",
|
||||||
response_test_cases["test_response_basic"]["test_params"]["case"],
|
responses_test_cases["test_response_basic"]["test_params"]["case"],
|
||||||
ids=case_id_generator,
|
ids=case_id_generator,
|
||||||
)
|
)
|
||||||
def test_response_streaming_basic(request, openai_client, model, provider, verification_config, case):
|
def test_response_streaming_basic(request, openai_client, model, provider, verification_config, case):
|
||||||
|
@ -77,7 +77,7 @@ def test_response_streaming_basic(request, openai_client, model, provider, verif
|
||||||
|
|
||||||
@pytest.mark.parametrize(
|
@pytest.mark.parametrize(
|
||||||
"case",
|
"case",
|
||||||
response_test_cases["test_response_multi_turn"]["test_params"]["case"],
|
responses_test_cases["test_response_multi_turn"]["test_params"]["case"],
|
||||||
ids=case_id_generator,
|
ids=case_id_generator,
|
||||||
)
|
)
|
||||||
def test_response_non_streaming_multi_turn(request, openai_client, model, provider, verification_config, case):
|
def test_response_non_streaming_multi_turn(request, openai_client, model, provider, verification_config, case):
|
||||||
|
@ -100,7 +100,7 @@ def test_response_non_streaming_multi_turn(request, openai_client, model, provid
|
||||||
|
|
||||||
@pytest.mark.parametrize(
|
@pytest.mark.parametrize(
|
||||||
"case",
|
"case",
|
||||||
response_test_cases["test_response_web_search"]["test_params"]["case"],
|
responses_test_cases["test_response_web_search"]["test_params"]["case"],
|
||||||
ids=case_id_generator,
|
ids=case_id_generator,
|
||||||
)
|
)
|
||||||
def test_response_non_streaming_web_search(request, openai_client, model, provider, verification_config, case):
|
def test_response_non_streaming_web_search(request, openai_client, model, provider, verification_config, case):
|
||||||
|
@ -126,7 +126,7 @@ def test_response_non_streaming_web_search(request, openai_client, model, provid
|
||||||
|
|
||||||
@pytest.mark.parametrize(
|
@pytest.mark.parametrize(
|
||||||
"case",
|
"case",
|
||||||
response_test_cases["test_response_image"]["test_params"]["case"],
|
responses_test_cases["test_response_image"]["test_params"]["case"],
|
||||||
ids=case_id_generator,
|
ids=case_id_generator,
|
||||||
)
|
)
|
||||||
def test_response_non_streaming_image(request, openai_client, model, provider, verification_config, case):
|
def test_response_non_streaming_image(request, openai_client, model, provider, verification_config, case):
|
||||||
|
@ -145,7 +145,7 @@ def test_response_non_streaming_image(request, openai_client, model, provider, v
|
||||||
|
|
||||||
@pytest.mark.parametrize(
|
@pytest.mark.parametrize(
|
||||||
"case",
|
"case",
|
||||||
response_test_cases["test_response_multi_turn_image"]["test_params"]["case"],
|
responses_test_cases["test_response_multi_turn_image"]["test_params"]["case"],
|
||||||
ids=case_id_generator,
|
ids=case_id_generator,
|
||||||
)
|
)
|
||||||
def test_response_non_streaming_multi_turn_image(request, openai_client, model, provider, verification_config, case):
|
def test_response_non_streaming_multi_turn_image(request, openai_client, model, provider, verification_config, case):
|
Loading…
Add table
Add a link
Reference in a new issue