mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-05 18:22:41 +00:00
Merge branch 'main' into fix_client_sdk_inference
This commit is contained in:
commit
f9f3303365
3 changed files with 4 additions and 2 deletions
|
@ -6,7 +6,8 @@
|
|||
|
||||
import pytest
|
||||
|
||||
from llama_stack.apis.inference import EmbeddingsResponse, ModelType
|
||||
from llama_stack.apis.inference import EmbeddingsResponse
|
||||
from llama_stack.apis.models import ModelType
|
||||
|
||||
# How to run this test:
|
||||
# pytest -v -s llama_stack/providers/tests/inference/test_embeddings.py
|
||||
|
|
|
@ -143,7 +143,7 @@ class TestVisionModelInference:
|
|||
assert len(grouped[ChatCompletionResponseEventType.complete]) == 1
|
||||
|
||||
content = "".join(
|
||||
chunk.event.delta
|
||||
chunk.event.delta.text
|
||||
for chunk in grouped[ChatCompletionResponseEventType.progress]
|
||||
)
|
||||
for expected_string in expected_strings:
|
||||
|
|
|
@ -39,6 +39,7 @@ def text_model_id(llama_stack_client):
|
|||
for model in llama_stack_client.models.list().data
|
||||
if model.identifier.startswith("meta-llama") and "405" not in model.identifier
|
||||
]
|
||||
print(available_models)
|
||||
assert len(available_models) > 0
|
||||
return available_models[0]
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue