mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-21 17:38:40 +00:00
Merge branch 'main' into use-openai-for-cerebras
This commit is contained in:
commit
e3ad762383
29 changed files with 11729 additions and 172 deletions
|
|
@ -99,6 +99,7 @@ def skip_if_doesnt_support_n(client_with_models, model_id):
|
|||
"remote::together", # `n` > 1 is not supported when streaming tokens. Please disable `stream`
|
||||
# Error code 400 - {'message': '"n" > 1 is not currently supported', 'type': 'invalid_request_error', 'param': 'n', 'code': 'wrong_api_format'}
|
||||
"remote::cerebras",
|
||||
"remote::databricks", # Bad request: parameter "n" must be equal to 1 for streaming mode
|
||||
):
|
||||
pytest.skip(f"Model {model_id} hosted by {provider.provider_type} doesn't support n param.")
|
||||
|
||||
|
|
@ -111,6 +112,7 @@ def skip_if_model_doesnt_support_openai_chat_completion(client_with_models, mode
|
|||
"inline::vllm",
|
||||
"remote::bedrock",
|
||||
"remote::databricks",
|
||||
"remote::cerebras",
|
||||
"remote::runpod",
|
||||
"remote::watsonx", # watsonx returns 404 when hitting the /openai/v1 endpoint
|
||||
):
|
||||
|
|
|
|||
|
|
@ -42,6 +42,7 @@ def skip_if_model_doesnt_support_encoding_format_base64(client, model_id):
|
|||
provider = provider_from_model(client, model_id)
|
||||
if provider.provider_type in (
|
||||
"remote::together", # param silently ignored, always returns floats
|
||||
"remote::databricks", # param silently ignored, always returns floats
|
||||
"remote::fireworks", # param silently ignored, always returns list of floats
|
||||
):
|
||||
pytest.skip(f"Model {model_id} hosted by {provider.provider_type} does not support encoding_format='base64'.")
|
||||
|
|
@ -52,6 +53,8 @@ def skip_if_model_doesnt_support_variable_dimensions(client_with_models, model_i
|
|||
if provider.provider_type in (
|
||||
"remote::together", # returns 400
|
||||
"inline::sentence-transformers",
|
||||
# Error code: 400 - {'error_code': 'BAD_REQUEST', 'message': 'Bad request: json: unknown field "dimensions"\n'}
|
||||
"remote::databricks",
|
||||
):
|
||||
pytest.skip(
|
||||
f"Model {model_id} hosted by {provider.provider_type} does not support variable output embedding dimensions."
|
||||
|
|
@ -75,7 +78,6 @@ def skip_if_model_doesnt_support_openai_embeddings(client, model_id):
|
|||
"inline::meta-reference",
|
||||
"remote::bedrock",
|
||||
"remote::cerebras",
|
||||
"remote::databricks",
|
||||
"remote::runpod",
|
||||
"remote::sambanova",
|
||||
"remote::tgi",
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue