mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 12:07:34 +00:00
chore: rm tgi
Signed-off-by: Sébastien Han <seb@redhat.com>
This commit is contained in:
parent
ed8b884a71
commit
a1527b3528
1 changed files with 0 additions and 1 deletions
|
@ -111,7 +111,6 @@ def skip_if_model_doesnt_support_openai_chat_completion(client_with_models, mode
|
||||||
"remote::cerebras",
|
"remote::cerebras",
|
||||||
"remote::databricks",
|
"remote::databricks",
|
||||||
"remote::runpod",
|
"remote::runpod",
|
||||||
"remote::tgi",
|
|
||||||
"remote::watsonx", # watsonx returns 404 when hitting the /openai/v1 endpoint
|
"remote::watsonx", # watsonx returns 404 when hitting the /openai/v1 endpoint
|
||||||
):
|
):
|
||||||
pytest.skip(f"Model {model_id} hosted by {provider.provider_type} doesn't support OpenAI chat completions.")
|
pytest.skip(f"Model {model_id} hosted by {provider.provider_type} doesn't support OpenAI chat completions.")
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue