mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-03 19:57:35 +00:00
skip new int tests until client SDK updates
This commit is contained in:
parent
1671581612
commit
852c058806
1 changed files with 4 additions and 0 deletions
|
@ -918,6 +918,7 @@ def test_openai_vector_store_search_modes(llama_stack_client, client_with_models
|
|||
assert search_response is not None
|
||||
|
||||
|
||||
@pytest.mark.skip(reason="Client SDK needs updating")
|
||||
def test_openai_vector_store_file_batch_create_and_retrieve(compat_client_with_empty_stores, client_with_models):
|
||||
"""Test creating and retrieving a vector store file batch."""
|
||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||
|
@ -969,6 +970,7 @@ def test_openai_vector_store_file_batch_create_and_retrieve(compat_client_with_e
|
|||
assert retrieved_batch.file_counts["total"] == len(file_ids)
|
||||
|
||||
|
||||
@pytest.mark.skip(reason="Client SDK needs updating")
|
||||
def test_openai_vector_store_file_batch_list_files(compat_client_with_empty_stores, client_with_models):
|
||||
"""Test listing files in a vector store file batch."""
|
||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||
|
@ -1039,6 +1041,7 @@ def test_openai_vector_store_file_batch_list_files(compat_client_with_empty_stor
|
|||
assert first_page_ids.isdisjoint(second_page_ids)
|
||||
|
||||
|
||||
@pytest.mark.skip(reason="Client SDK needs updating")
|
||||
def test_openai_vector_store_file_batch_cancel(compat_client_with_empty_stores, client_with_models):
|
||||
"""Test cancelling a vector store file batch."""
|
||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||
|
@ -1082,6 +1085,7 @@ def test_openai_vector_store_file_batch_cancel(compat_client_with_empty_stores,
|
|||
raise
|
||||
|
||||
|
||||
@pytest.mark.skip(reason="Client SDK needs updating")
|
||||
def test_openai_vector_store_file_batch_error_handling(compat_client_with_empty_stores, client_with_models):
|
||||
"""Test error handling for file batch operations."""
|
||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue