diff --git a/tests/integration/inference/test_openai_completion.py b/tests/integration/inference/test_openai_completion.py index 190840f70..11bc5bef6 100644 --- a/tests/integration/inference/test_openai_completion.py +++ b/tests/integration/inference/test_openai_completion.py @@ -37,6 +37,7 @@ def skip_if_model_doesnt_support_openai_completion(client_with_models, model_id) # support both completions and chat completions endpoint and all the Llama models are # just chat completions "remote::nvidia", + "remote::openai", "remote::runpod", "remote::sambanova", "remote::tgi",