mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-12 20:12:33 +00:00
chore: remove mentions of expires_after_anchor
Removed in https://github.com/llamastack/llama-stack/pull/3604 for expires_after only. Signed-off-by: Sébastien Han <seb@redhat.com>
This commit is contained in:
parent
da35f2452e
commit
070811f855
2 changed files with 16 additions and 29 deletions
|
|
@ -576,6 +576,7 @@ def test_openai_vector_store_attach_file(
|
||||||
):
|
):
|
||||||
"""Test OpenAI vector store attach file."""
|
"""Test OpenAI vector store attach file."""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
from llama_stack.apis.files import ExpiresAfter
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
|
||||||
|
|
@ -594,8 +595,7 @@ def test_openai_vector_store_attach_file(
|
||||||
file = compat_client.files.create(
|
file = compat_client.files.create(
|
||||||
file=file_buffer,
|
file=file_buffer,
|
||||||
purpose="assistants",
|
purpose="assistants",
|
||||||
expires_after_anchor="created_at",
|
expires_after=ExpiresAfter(anchor="created_at", seconds=86400), # 24 hours
|
||||||
expires_after_seconds=86400, # 24 hours
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# Attach the file to the vector store
|
# Attach the file to the vector store
|
||||||
|
|
@ -638,6 +638,7 @@ def test_openai_vector_store_attach_files_on_creation(
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
from llama_stack.apis.files import ExpiresAfter
|
||||||
|
|
||||||
# Create some files and attach them to the vector store
|
# Create some files and attach them to the vector store
|
||||||
valid_file_ids = []
|
valid_file_ids = []
|
||||||
|
|
@ -647,8 +648,7 @@ def test_openai_vector_store_attach_files_on_creation(
|
||||||
file = compat_client.files.create(
|
file = compat_client.files.create(
|
||||||
file=file_buffer,
|
file=file_buffer,
|
||||||
purpose="assistants",
|
purpose="assistants",
|
||||||
expires_after_anchor="created_at",
|
expires_after=ExpiresAfter(anchor="created_at", seconds=86400), # 24 hours
|
||||||
expires_after_seconds=86400, # 24 hours
|
|
||||||
)
|
)
|
||||||
valid_file_ids.append(file.id)
|
valid_file_ids.append(file.id)
|
||||||
|
|
||||||
|
|
@ -701,6 +701,7 @@ def test_openai_vector_store_list_files(
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
from llama_stack.apis.files import ExpiresAfter
|
||||||
|
|
||||||
# Create a vector store
|
# Create a vector store
|
||||||
vector_store = compat_client.vector_stores.create(
|
vector_store = compat_client.vector_stores.create(
|
||||||
|
|
@ -718,8 +719,7 @@ def test_openai_vector_store_list_files(
|
||||||
file = compat_client.files.create(
|
file = compat_client.files.create(
|
||||||
file=file_buffer,
|
file=file_buffer,
|
||||||
purpose="assistants",
|
purpose="assistants",
|
||||||
expires_after_anchor="created_at",
|
expires_after=ExpiresAfter(anchor="created_at", seconds=86400), # 24 hours
|
||||||
expires_after_seconds=86400, # 24 hours
|
|
||||||
)
|
)
|
||||||
|
|
||||||
response = compat_client.vector_stores.files.create(
|
response = compat_client.vector_stores.files.create(
|
||||||
|
|
@ -790,6 +790,7 @@ def test_openai_vector_store_retrieve_file_contents(
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
from llama_stack.apis.files import ExpiresAfter
|
||||||
|
|
||||||
# Create a vector store
|
# Create a vector store
|
||||||
vector_store = compat_client.vector_stores.create(
|
vector_store = compat_client.vector_stores.create(
|
||||||
|
|
@ -808,8 +809,7 @@ def test_openai_vector_store_retrieve_file_contents(
|
||||||
file = compat_client.files.create(
|
file = compat_client.files.create(
|
||||||
file=file_buffer,
|
file=file_buffer,
|
||||||
purpose="assistants",
|
purpose="assistants",
|
||||||
expires_after_anchor="created_at",
|
expires_after=ExpiresAfter(anchor="created_at", seconds=86400), # 24 hours
|
||||||
expires_after_seconds=86400, # 24 hours
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# Attach the file to the vector store
|
# Attach the file to the vector store
|
||||||
|
|
@ -849,6 +849,7 @@ def test_openai_vector_store_delete_file(
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
from llama_stack.apis.files import ExpiresAfter
|
||||||
|
|
||||||
# Create a vector store
|
# Create a vector store
|
||||||
vector_store = compat_client.vector_stores.create(
|
vector_store = compat_client.vector_stores.create(
|
||||||
|
|
@ -866,8 +867,7 @@ def test_openai_vector_store_delete_file(
|
||||||
file = compat_client.files.create(
|
file = compat_client.files.create(
|
||||||
file=file_buffer,
|
file=file_buffer,
|
||||||
purpose="assistants",
|
purpose="assistants",
|
||||||
expires_after_anchor="created_at",
|
expires_after=ExpiresAfter(anchor="created_at", seconds=86400), # 24 hours
|
||||||
expires_after_seconds=86400, # 24 hours
|
|
||||||
)
|
)
|
||||||
|
|
||||||
compat_client.vector_stores.files.create(
|
compat_client.vector_stores.files.create(
|
||||||
|
|
@ -913,6 +913,7 @@ def test_openai_vector_store_delete_file_removes_from_vector_store(
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
from llama_stack.apis.files import ExpiresAfter
|
||||||
|
|
||||||
# Create a vector store
|
# Create a vector store
|
||||||
vector_store = compat_client.vector_stores.create(
|
vector_store = compat_client.vector_stores.create(
|
||||||
|
|
@ -929,8 +930,7 @@ def test_openai_vector_store_delete_file_removes_from_vector_store(
|
||||||
file = compat_client.files.create(
|
file = compat_client.files.create(
|
||||||
file=file_buffer,
|
file=file_buffer,
|
||||||
purpose="assistants",
|
purpose="assistants",
|
||||||
expires_after_anchor="created_at",
|
expires_after=ExpiresAfter(anchor="created_at", seconds=86400), # 24 hours
|
||||||
expires_after_seconds=86400, # 24 hours
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# Attach the file to the vector store
|
# Attach the file to the vector store
|
||||||
|
|
@ -963,6 +963,7 @@ def test_openai_vector_store_update_file(
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
from llama_stack.apis.files import ExpiresAfter
|
||||||
|
|
||||||
# Create a vector store
|
# Create a vector store
|
||||||
vector_store = compat_client.vector_stores.create(
|
vector_store = compat_client.vector_stores.create(
|
||||||
|
|
@ -979,8 +980,7 @@ def test_openai_vector_store_update_file(
|
||||||
file = compat_client.files.create(
|
file = compat_client.files.create(
|
||||||
file=file_buffer,
|
file=file_buffer,
|
||||||
purpose="assistants",
|
purpose="assistants",
|
||||||
expires_after_anchor="created_at",
|
expires_after=ExpiresAfter(anchor="created_at", seconds=86400), # 24 hours
|
||||||
expires_after_seconds=86400, # 24 hours
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# Attach the file to the vector store
|
# Attach the file to the vector store
|
||||||
|
|
@ -1018,6 +1018,7 @@ def test_create_vector_store_files_duplicate_vector_store_name(
|
||||||
This test confirms that client.vector_stores.create() creates a unique ID
|
This test confirms that client.vector_stores.create() creates a unique ID
|
||||||
"""
|
"""
|
||||||
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
skip_if_provider_doesnt_support_openai_vector_stores(client_with_models)
|
||||||
|
from llama_stack.apis.files import ExpiresAfter
|
||||||
|
|
||||||
compat_client = compat_client_with_empty_stores
|
compat_client = compat_client_with_empty_stores
|
||||||
|
|
||||||
|
|
@ -1029,8 +1030,7 @@ def test_create_vector_store_files_duplicate_vector_store_name(
|
||||||
file = compat_client.files.create(
|
file = compat_client.files.create(
|
||||||
file=file_buffer,
|
file=file_buffer,
|
||||||
purpose="assistants",
|
purpose="assistants",
|
||||||
expires_after_anchor="created_at",
|
expires_after=ExpiresAfter(anchor="created_at", seconds=86400), # 24 hours
|
||||||
expires_after_seconds=86400, # 24 hours
|
|
||||||
)
|
)
|
||||||
file_ids.append(file.id)
|
file_ids.append(file.id)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -258,19 +258,6 @@ class TestS3FilesImpl:
|
||||||
with pytest.raises(ResourceNotFoundError, match="not found"):
|
with pytest.raises(ResourceNotFoundError, match="not found"):
|
||||||
await s3_provider._get_file(uploaded.id, return_expired=True)
|
await s3_provider._get_file(uploaded.id, return_expired=True)
|
||||||
|
|
||||||
async def test_unsupported_expires_after_anchor(self, s3_provider, sample_text_file):
|
|
||||||
"""Unsupported anchor value should raise ValueError."""
|
|
||||||
from llama_stack.apis.files import ExpiresAfter
|
|
||||||
|
|
||||||
sample_text_file.filename = "test_unsupported_expires_after_anchor"
|
|
||||||
|
|
||||||
with pytest.raises(ValueError, match="Input should be 'created_at'"):
|
|
||||||
await s3_provider.openai_upload_file(
|
|
||||||
file=sample_text_file,
|
|
||||||
purpose=OpenAIFilePurpose.ASSISTANTS,
|
|
||||||
expires_after=ExpiresAfter(anchor="now", seconds=3600), # type: ignore
|
|
||||||
)
|
|
||||||
|
|
||||||
async def test_nonint_expires_after_seconds(self, s3_provider, sample_text_file):
|
async def test_nonint_expires_after_seconds(self, s3_provider, sample_text_file):
|
||||||
"""Non-integer seconds in expires_after should raise ValueError."""
|
"""Non-integer seconds in expires_after should raise ValueError."""
|
||||||
from llama_stack.apis.files import ExpiresAfter
|
from llama_stack.apis.files import ExpiresAfter
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue