mirror of
https://github.com/meta-llama/llama-stack.git
synced 2026-01-01 16:44:31 +00:00
dropped impls for hf serverless and hf endpoint
This commit is contained in:
parent
1b15df8d1d
commit
650cbc395d
4 changed files with 6 additions and 44 deletions
|
|
@ -87,7 +87,7 @@ def test_image_chat_completion_streaming(client_with_models, vision_model_id):
|
|||
assert any(expected in streamed_content for expected in {"dog", "puppy", "pup"})
|
||||
|
||||
|
||||
@pytest.mark.parametrize("type_", ["url"])
|
||||
@pytest.mark.parametrize("type_", ["url", "data"])
|
||||
def test_image_chat_completion_base64(client_with_models, vision_model_id, base64_image_data, base64_image_url, type_):
|
||||
image_spec = {
|
||||
"url": {
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue