mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-28 04:21:58 +00:00
This moves the OpenAI Responses API tests under tests/verifications/openai_api/test_response.py and starts to wire them up to our verification suite, so that we can test multiple providers as well as OpenAI directly for the Responses API. Signed-off-by: Ben Browning <bbrownin@redhat.com>
17 lines
770 B
YAML
17 lines
770 B
YAML
base_url: http://localhost:8321/v1/openai/v1
|
|
api_key_var: TOGETHER_API_KEY
|
|
models:
|
|
- together/meta-llama/Llama-3.3-70B-Instruct-Turbo
|
|
- together/meta-llama/Llama-4-Scout-17B-16E-Instruct
|
|
- together/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8
|
|
model_display_names:
|
|
together/meta-llama/Llama-3.3-70B-Instruct-Turbo: Llama-3.3-70B-Instruct
|
|
together/meta-llama/Llama-4-Scout-17B-16E-Instruct: Llama-4-Scout-Instruct
|
|
together/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8: Llama-4-Maverick-Instruct
|
|
test_exclusions:
|
|
together/meta-llama/Llama-3.3-70B-Instruct-Turbo:
|
|
- test_chat_non_streaming_image
|
|
- test_chat_streaming_image
|
|
- test_chat_multi_turn_multiple_images
|
|
- test_response_non_streaming_image
|
|
- test_response_non_streaming_multi_turn_image
|