mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-12 12:06:04 +00:00
skipping batch apis as they're not available from client yet
Signed-off-by: Francisco Javier Arceo <farceo@redhat.com>
This commit is contained in:
parent
1000d0f68b
commit
8325c8e8b8
1 changed files with 10 additions and 0 deletions
|
|
@ -1025,6 +1025,11 @@ def test_create_vector_store_files_duplicate_vector_store_name(compat_client_wit
|
||||||
created_file_from_non_deleted_vector_store = compat_client.vector_stores.files.create(
|
created_file_from_non_deleted_vector_store = compat_client.vector_stores.files.create(
|
||||||
vector_store_id=vector_store.id,
|
vector_store_id=vector_store.id,
|
||||||
file_id=file_ids[1],
|
file_id=file_ids[1],
|
||||||
|
extra_body={
|
||||||
|
"embedding_model": "inline::sentence-transformers/all-MiniLM-L6-v2",
|
||||||
|
"embedding_dimension": 384,
|
||||||
|
"provider_id": "my_provider",
|
||||||
|
},
|
||||||
)
|
)
|
||||||
assert created_file_from_non_deleted_vector_store.status == "completed"
|
assert created_file_from_non_deleted_vector_store.status == "completed"
|
||||||
|
|
||||||
|
|
@ -1065,6 +1070,7 @@ def test_openai_vector_store_search_modes(llama_stack_client, client_with_models
|
||||||
def test_openai_vector_store_file_batch_create_and_retrieve(compat_client_with_empty_stores, client_with_models):
|
def test_openai_vector_store_file_batch_create_and_retrieve(compat_client_with_empty_stores, client_with_models):
|
||||||
"""Test creating and retrieving a vector store file batch."""
|
"""Test creating and retrieving a vector store file batch."""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
pytest.skip("file_batches functionality not yet exposed through client interface")
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
|
||||||
|
|
@ -1130,6 +1136,7 @@ def test_openai_vector_store_file_batch_create_and_retrieve(compat_client_with_e
|
||||||
def test_openai_vector_store_file_batch_list_files(compat_client_with_empty_stores, client_with_models):
|
def test_openai_vector_store_file_batch_list_files(compat_client_with_empty_stores, client_with_models):
|
||||||
"""Test listing files in a vector store file batch."""
|
"""Test listing files in a vector store file batch."""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
pytest.skip("file_batches functionality not yet exposed through client interface")
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
|
||||||
|
|
@ -1225,6 +1232,7 @@ def test_openai_vector_store_file_batch_list_files(compat_client_with_empty_stor
|
||||||
def test_openai_vector_store_file_batch_cancel(compat_client_with_empty_stores, client_with_models):
|
def test_openai_vector_store_file_batch_cancel(compat_client_with_empty_stores, client_with_models):
|
||||||
"""Test cancelling a vector store file batch."""
|
"""Test cancelling a vector store file batch."""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
pytest.skip("file_batches functionality not yet exposed through client interface")
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
|
||||||
|
|
@ -1282,6 +1290,7 @@ def test_openai_vector_store_file_batch_cancel(compat_client_with_empty_stores,
|
||||||
def test_openai_vector_store_file_batch_retrieve_contents(compat_client_with_empty_stores, client_with_models):
|
def test_openai_vector_store_file_batch_retrieve_contents(compat_client_with_empty_stores, client_with_models):
|
||||||
"""Test retrieving file contents after file batch processing."""
|
"""Test retrieving file contents after file batch processing."""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
pytest.skip("file_batches functionality not yet exposed through client interface")
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
|
||||||
|
|
@ -1357,6 +1366,7 @@ def test_openai_vector_store_file_batch_retrieve_contents(compat_client_with_emp
|
||||||
def test_openai_vector_store_file_batch_error_handling(compat_client_with_empty_stores, client_with_models):
|
def test_openai_vector_store_file_batch_error_handling(compat_client_with_empty_stores, client_with_models):
|
||||||
"""Test error handling for file batch operations."""
|
"""Test error handling for file batch operations."""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
pytest.skip("file_batches functionality not yet exposed through client interface")
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue