mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-21 02:22:25 +00:00
removing the base64 encoding
Signed-off-by: Francisco Javier Arceo <farceo@redhat.com>
This commit is contained in:
parent
e0f1788e9e
commit
5ef31f6752
3 changed files with 2 additions and 27 deletions
|
|
@ -34,11 +34,7 @@ def skip_if_model_doesnt_support_variable_dimensions(model_id):
|
|||
pytest.skip("{model_id} does not support variable output embedding dimensions")
|
||||
|
||||
|
||||
@pytest.fixture(
|
||||
params=[
|
||||
"openai_client",
|
||||
]
|
||||
)
|
||||
@pytest.fixture(params=["openai_client", "llama_stack_client"])
|
||||
def compat_client(request, client_with_models):
|
||||
if request.param == "openai_client" and isinstance(client_with_models, LlamaStackAsLibraryClient):
|
||||
pytest.skip("OpenAI client tests not supported with library client")
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue