mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-14 09:12:51 +00:00
updated tests to verify usage of previous_response_id and conversation
Signed-off-by: Francisco Javier Arceo <farceo@redhat.com>
This commit is contained in:
parent
1e59793288
commit
9d039e884d
4 changed files with 30 additions and 0 deletions
|
|
@ -30,6 +30,7 @@ CATEGORIES = [
|
|||
"tools",
|
||||
"client",
|
||||
"telemetry",
|
||||
"openai",
|
||||
"openai_responses",
|
||||
"openai_conversations",
|
||||
"testing",
|
||||
|
|
|
|||
|
|
@ -230,6 +230,11 @@ class OpenAIResponsesImpl:
|
|||
if shields is not None:
|
||||
raise NotImplementedError("Shields parameter is not yet implemented in the meta-reference provider")
|
||||
|
||||
if conversation is not None and previous_response_id is not None:
|
||||
raise ValueError(
|
||||
"Mutually exclusive parameters: 'previous_response_id' and 'conversation'. Ensure you are only providing one of these parameters."
|
||||
)
|
||||
|
||||
if conversation is not None:
|
||||
if not conversation.startswith("conv_"):
|
||||
raise InvalidConversationIdError(conversation)
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue