mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 04:04:14 +00:00
skip new int tests until client SDK updates
This commit is contained in:
parent
1671581612
commit
852c058806
1 changed files with 4 additions and 0 deletions
|
@ -918,6 +918,7 @@ def test_openai_vector_store_search_modes(llama_stack_client, client_with_models
|
||||||
assert search_response is not None
|
assert search_response is not None
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.skip(reason="Client SDK needs updating")
|
||||||
def test_openai_vector_store_file_batch_create_and_retrieve(compat_client_with_empty_stores, client_with_models):
|
def test_openai_vector_store_file_batch_create_and_retrieve(compat_client_with_empty_stores, client_with_models):
|
||||||
"""Test creating and retrieving a vector store file batch."""
|
"""Test creating and retrieving a vector store file batch."""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
@ -969,6 +970,7 @@ def test_openai_vector_store_file_batch_create_and_retrieve(compat_client_with_e
|
||||||
assert retrieved_batch.file_counts["total"] == len(file_ids)
|
assert retrieved_batch.file_counts["total"] == len(file_ids)
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.skip(reason="Client SDK needs updating")
|
||||||
def test_openai_vector_store_file_batch_list_files(compat_client_with_empty_stores, client_with_models):
|
def test_openai_vector_store_file_batch_list_files(compat_client_with_empty_stores, client_with_models):
|
||||||
"""Test listing files in a vector store file batch."""
|
"""Test listing files in a vector store file batch."""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
@ -1039,6 +1041,7 @@ def test_openai_vector_store_file_batch_list_files(compat_client_with_empty_stor
|
||||||
assert first_page_ids.isdisjoint(second_page_ids)
|
assert first_page_ids.isdisjoint(second_page_ids)
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.skip(reason="Client SDK needs updating")
|
||||||
def test_openai_vector_store_file_batch_cancel(compat_client_with_empty_stores, client_with_models):
|
def test_openai_vector_store_file_batch_cancel(compat_client_with_empty_stores, client_with_models):
|
||||||
"""Test cancelling a vector store file batch."""
|
"""Test cancelling a vector store file batch."""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
@ -1082,6 +1085,7 @@ def test_openai_vector_store_file_batch_cancel(compat_client_with_empty_stores,
|
||||||
raise
|
raise
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.skip(reason="Client SDK needs updating")
|
||||||
def test_openai_vector_store_file_batch_error_handling(compat_client_with_empty_stores, client_with_models):
|
def test_openai_vector_store_file_batch_error_handling(compat_client_with_empty_stores, client_with_models):
|
||||||
"""Test error handling for file batch operations."""
|
"""Test error handling for file batch operations."""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue