mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-31 16:01:46 +00:00
fix(responses): use input, not original_input when storing the Response
This commit is contained in:
parent
63a9f08c9e
commit
cda44a77e6
1 changed files with 10 additions and 11 deletions
|
@ -292,12 +292,12 @@ class OpenAIResponsesImpl:
|
||||||
async def _store_response(
|
async def _store_response(
|
||||||
self,
|
self,
|
||||||
response: OpenAIResponseObject,
|
response: OpenAIResponseObject,
|
||||||
original_input: str | list[OpenAIResponseInput],
|
input: str | list[OpenAIResponseInput],
|
||||||
) -> None:
|
) -> None:
|
||||||
new_input_id = f"msg_{uuid.uuid4()}"
|
new_input_id = f"msg_{uuid.uuid4()}"
|
||||||
if isinstance(original_input, str):
|
if isinstance(input, str):
|
||||||
# synthesize a message from the input string
|
# synthesize a message from the input string
|
||||||
input_content = OpenAIResponseInputMessageContentText(text=original_input)
|
input_content = OpenAIResponseInputMessageContentText(text=input)
|
||||||
input_content_item = OpenAIResponseMessage(
|
input_content_item = OpenAIResponseMessage(
|
||||||
role="user",
|
role="user",
|
||||||
content=[input_content],
|
content=[input_content],
|
||||||
|
@ -307,7 +307,7 @@ class OpenAIResponsesImpl:
|
||||||
else:
|
else:
|
||||||
# we already have a list of messages
|
# we already have a list of messages
|
||||||
input_items_data = []
|
input_items_data = []
|
||||||
for input_item in original_input:
|
for input_item in input:
|
||||||
if isinstance(input_item, OpenAIResponseMessage):
|
if isinstance(input_item, OpenAIResponseMessage):
|
||||||
# These may or may not already have an id, so dump to dict, check for id, and add if missing
|
# These may or may not already have an id, so dump to dict, check for id, and add if missing
|
||||||
input_item_dict = input_item.model_dump()
|
input_item_dict = input_item.model_dump()
|
||||||
|
@ -334,7 +334,6 @@ class OpenAIResponsesImpl:
|
||||||
tools: list[OpenAIResponseInputTool] | None = None,
|
tools: list[OpenAIResponseInputTool] | None = None,
|
||||||
):
|
):
|
||||||
stream = False if stream is None else stream
|
stream = False if stream is None else stream
|
||||||
original_input = input # Keep reference for storage
|
|
||||||
|
|
||||||
output_messages: list[OpenAIResponseOutput] = []
|
output_messages: list[OpenAIResponseOutput] = []
|
||||||
|
|
||||||
|
@ -372,7 +371,7 @@ class OpenAIResponsesImpl:
|
||||||
inference_result=inference_result,
|
inference_result=inference_result,
|
||||||
ctx=ctx,
|
ctx=ctx,
|
||||||
output_messages=output_messages,
|
output_messages=output_messages,
|
||||||
original_input=original_input,
|
input=input,
|
||||||
model=model,
|
model=model,
|
||||||
store=store,
|
store=store,
|
||||||
tools=tools,
|
tools=tools,
|
||||||
|
@ -382,7 +381,7 @@ class OpenAIResponsesImpl:
|
||||||
inference_result=inference_result,
|
inference_result=inference_result,
|
||||||
ctx=ctx,
|
ctx=ctx,
|
||||||
output_messages=output_messages,
|
output_messages=output_messages,
|
||||||
original_input=original_input,
|
input=input,
|
||||||
model=model,
|
model=model,
|
||||||
store=store,
|
store=store,
|
||||||
tools=tools,
|
tools=tools,
|
||||||
|
@ -393,7 +392,7 @@ class OpenAIResponsesImpl:
|
||||||
inference_result: Any,
|
inference_result: Any,
|
||||||
ctx: ChatCompletionContext,
|
ctx: ChatCompletionContext,
|
||||||
output_messages: list[OpenAIResponseOutput],
|
output_messages: list[OpenAIResponseOutput],
|
||||||
original_input: str | list[OpenAIResponseInput],
|
input: str | list[OpenAIResponseInput],
|
||||||
model: str,
|
model: str,
|
||||||
store: bool | None,
|
store: bool | None,
|
||||||
tools: list[OpenAIResponseInputTool] | None,
|
tools: list[OpenAIResponseInputTool] | None,
|
||||||
|
@ -423,7 +422,7 @@ class OpenAIResponsesImpl:
|
||||||
if store:
|
if store:
|
||||||
await self._store_response(
|
await self._store_response(
|
||||||
response=response,
|
response=response,
|
||||||
original_input=original_input,
|
input=input,
|
||||||
)
|
)
|
||||||
|
|
||||||
return response
|
return response
|
||||||
|
@ -433,7 +432,7 @@ class OpenAIResponsesImpl:
|
||||||
inference_result: Any,
|
inference_result: Any,
|
||||||
ctx: ChatCompletionContext,
|
ctx: ChatCompletionContext,
|
||||||
output_messages: list[OpenAIResponseOutput],
|
output_messages: list[OpenAIResponseOutput],
|
||||||
original_input: str | list[OpenAIResponseInput],
|
input: str | list[OpenAIResponseInput],
|
||||||
model: str,
|
model: str,
|
||||||
store: bool | None,
|
store: bool | None,
|
||||||
tools: list[OpenAIResponseInputTool] | None,
|
tools: list[OpenAIResponseInputTool] | None,
|
||||||
|
@ -544,7 +543,7 @@ class OpenAIResponsesImpl:
|
||||||
if store:
|
if store:
|
||||||
await self._store_response(
|
await self._store_response(
|
||||||
response=final_response,
|
response=final_response,
|
||||||
original_input=original_input,
|
input=input,
|
||||||
)
|
)
|
||||||
|
|
||||||
# Emit response.completed
|
# Emit response.completed
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue