mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-15 14:08:00 +00:00
remove unused CreateBatchRequest, update completion_window to be literal "24h"
This commit is contained in:
parent
95a3ecdffc
commit
44263ce954
4 changed files with 7 additions and 18 deletions
|
@ -379,9 +379,8 @@ class TestBatchesErrorHandling:
|
|||
)
|
||||
assert exc_info.value.status_code == 400
|
||||
error_msg = str(exc_info.value).lower()
|
||||
assert "invalid value" in error_msg
|
||||
assert "error" in error_msg
|
||||
assert "completion_window" in error_msg
|
||||
assert "supported values are" in error_msg
|
||||
|
||||
def test_batch_streaming_not_supported(self, openai_client, batch_helper, text_model_id):
|
||||
"""Test that streaming responses are not supported in batches."""
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue