From ae012bb85706ed4ef9154bfb4a6180a5b31b253f Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Mon, 28 Apr 2025 10:46:09 -0700 Subject: [PATCH] rename response to responses in verifications, update provider --- .../verifications/openai-api-verification-run.yaml | 12 ++++++------ .../test_cases/{response.yaml => responses.yaml} | 0 .../{test_response.py => test_responses.py} | 14 +++++++------- 3 files changed, 13 insertions(+), 13 deletions(-) rename tests/verifications/openai_api/fixtures/test_cases/{response.yaml => responses.yaml} (100%) rename tests/verifications/openai_api/{test_response.py => test_responses.py} (92%) diff --git a/tests/verifications/openai-api-verification-run.yaml b/tests/verifications/openai-api-verification-run.yaml index 58ce5344d..3ea9e7157 100644 --- a/tests/verifications/openai-api-verification-run.yaml +++ b/tests/verifications/openai-api-verification-run.yaml @@ -1,8 +1,8 @@ version: '2' image_name: openai-api-verification apis: +- agents - inference -- openai_responses - telemetry - tool_runtime - vector_io @@ -46,14 +46,14 @@ providers: service_name: "${env.OTEL_SERVICE_NAME:\u200B}" sinks: ${env.TELEMETRY_SINKS:console,sqlite} sqlite_db_path: ${env.SQLITE_DB_PATH:~/.llama/distributions/openai/trace_store.db} - openai_responses: - - provider_id: openai-responses - provider_type: inline::openai-responses + agents: + - provider_id: meta-reference + provider_type: inline::meta-reference config: - kvstore: + persistence_store: type: sqlite namespace: null - db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/openai}/openai_responses.db + db_path: ${env.SQLITE_STORE_DIR:~/.llama/distributions/openai}/agents_store.db tool_runtime: - provider_id: brave-search provider_type: remote::brave-search diff --git a/tests/verifications/openai_api/fixtures/test_cases/response.yaml b/tests/verifications/openai_api/fixtures/test_cases/responses.yaml similarity index 100% rename from tests/verifications/openai_api/fixtures/test_cases/response.yaml rename to tests/verifications/openai_api/fixtures/test_cases/responses.yaml diff --git a/tests/verifications/openai_api/test_response.py b/tests/verifications/openai_api/test_responses.py similarity index 92% rename from tests/verifications/openai_api/test_response.py rename to tests/verifications/openai_api/test_responses.py index d4bbb526a..cc7ec320c 100644 --- a/tests/verifications/openai_api/test_response.py +++ b/tests/verifications/openai_api/test_responses.py @@ -14,12 +14,12 @@ from tests.verifications.openai_api.fixtures.fixtures import ( ) from tests.verifications.openai_api.fixtures.load import load_test_cases -response_test_cases = load_test_cases("response") +responses_test_cases = load_test_cases("responses") @pytest.mark.parametrize( "case", - response_test_cases["test_response_basic"]["test_params"]["case"], + responses_test_cases["test_response_basic"]["test_params"]["case"], ids=case_id_generator, ) def test_response_non_streaming_basic(request, openai_client, model, provider, verification_config, case): @@ -48,7 +48,7 @@ def test_response_non_streaming_basic(request, openai_client, model, provider, v @pytest.mark.parametrize( "case", - response_test_cases["test_response_basic"]["test_params"]["case"], + responses_test_cases["test_response_basic"]["test_params"]["case"], ids=case_id_generator, ) def test_response_streaming_basic(request, openai_client, model, provider, verification_config, case): @@ -77,7 +77,7 @@ def test_response_streaming_basic(request, openai_client, model, provider, verif @pytest.mark.parametrize( "case", - response_test_cases["test_response_multi_turn"]["test_params"]["case"], + responses_test_cases["test_response_multi_turn"]["test_params"]["case"], ids=case_id_generator, ) def test_response_non_streaming_multi_turn(request, openai_client, model, provider, verification_config, case): @@ -100,7 +100,7 @@ def test_response_non_streaming_multi_turn(request, openai_client, model, provid @pytest.mark.parametrize( "case", - response_test_cases["test_response_web_search"]["test_params"]["case"], + responses_test_cases["test_response_web_search"]["test_params"]["case"], ids=case_id_generator, ) def test_response_non_streaming_web_search(request, openai_client, model, provider, verification_config, case): @@ -126,7 +126,7 @@ def test_response_non_streaming_web_search(request, openai_client, model, provid @pytest.mark.parametrize( "case", - response_test_cases["test_response_image"]["test_params"]["case"], + responses_test_cases["test_response_image"]["test_params"]["case"], ids=case_id_generator, ) def test_response_non_streaming_image(request, openai_client, model, provider, verification_config, case): @@ -145,7 +145,7 @@ def test_response_non_streaming_image(request, openai_client, model, provider, v @pytest.mark.parametrize( "case", - response_test_cases["test_response_multi_turn_image"]["test_params"]["case"], + responses_test_cases["test_response_multi_turn_image"]["test_params"]["case"], ids=case_id_generator, ) def test_response_non_streaming_multi_turn_image(request, openai_client, model, provider, verification_config, case):