mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-12 12:06:04 +00:00
skipping batch apis as they're not available from client yet
Signed-off-by: Francisco Javier Arceo <farceo@redhat.com>
This commit is contained in:
parent
1000d0f68b
commit
8325c8e8b8
1 changed files with 10 additions and 0 deletions
|
|
@ -1025,6 +1025,11 @@ def test_create_vector_store_files_duplicate_vector_store_name(compat_client_wit
|
|||
created_file_from_non_deleted_vector_store = compat_client.vector_stores.files.create(
|
||||
vector_store_id=vector_store.id,
|
||||
file_id=file_ids[1],
|
||||
extra_body={
|
||||
"embedding_model": "inline::sentence-transformers/all-MiniLM-L6-v2",
|
||||
"embedding_dimension": 384,
|
||||
"provider_id": "my_provider",
|
||||
},
|
||||
)
|
||||
assert created_file_from_non_deleted_vector_store.status == "completed"
|
||||
|
||||
|
|
@ -1065,6 +1070,7 @@ def test_openai_vector_store_search_modes(llama_stack_client, client_with_models
|
|||
def test_openai_vector_store_file_batch_create_and_retrieve(compat_client_with_empty_stores, client_with_models):
|
||||
"""Test creating and retrieving a vector store file batch."""
|
||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||
pytest.skip("file_batches functionality not yet exposed through client interface")
|
||||
|
||||
compat_client = compat_client_with_empty_stores
|
||||
|
||||
|
|
@ -1130,6 +1136,7 @@ def test_openai_vector_store_file_batch_create_and_retrieve(compat_client_with_e
|
|||
def test_openai_vector_store_file_batch_list_files(compat_client_with_empty_stores, client_with_models):
|
||||
"""Test listing files in a vector store file batch."""
|
||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||
pytest.skip("file_batches functionality not yet exposed through client interface")
|
||||
|
||||
compat_client = compat_client_with_empty_stores
|
||||
|
||||
|
|
@ -1225,6 +1232,7 @@ def test_openai_vector_store_file_batch_list_files(compat_client_with_empty_stor
|
|||
def test_openai_vector_store_file_batch_cancel(compat_client_with_empty_stores, client_with_models):
|
||||
"""Test cancelling a vector store file batch."""
|
||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||
pytest.skip("file_batches functionality not yet exposed through client interface")
|
||||
|
||||
compat_client = compat_client_with_empty_stores
|
||||
|
||||
|
|
@ -1282,6 +1290,7 @@ def test_openai_vector_store_file_batch_cancel(compat_client_with_empty_stores,
|
|||
def test_openai_vector_store_file_batch_retrieve_contents(compat_client_with_empty_stores, client_with_models):
|
||||
"""Test retrieving file contents after file batch processing."""
|
||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||
pytest.skip("file_batches functionality not yet exposed through client interface")
|
||||
|
||||
compat_client = compat_client_with_empty_stores
|
||||
|
||||
|
|
@ -1357,6 +1366,7 @@ def test_openai_vector_store_file_batch_retrieve_contents(compat_client_with_emp
|
|||
def test_openai_vector_store_file_batch_error_handling(compat_client_with_empty_stores, client_with_models):
|
||||
"""Test error handling for file batch operations."""
|
||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||
pytest.skip("file_batches functionality not yet exposed through client interface")
|
||||
|
||||
compat_client = compat_client_with_empty_stores
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue