diff --git a/src/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py b/src/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py index 8dda60817..cb0fe284e 100644 --- a/src/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py +++ b/src/llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py @@ -363,6 +363,7 @@ class OpenAIResponsesImpl: # Structured outputs response_format = await convert_response_text_to_chat_response_format(text) + ctx = ChatCompletionContext( model=model, messages=messages, diff --git a/tests/unit/server/test_server.py b/tests/unit/server/test_server.py index 24ee8a1cd..53f193672 100644 --- a/tests/unit/server/test_server.py +++ b/tests/unit/server/test_server.py @@ -165,7 +165,7 @@ class TestTranslateException: assert result.detail == "Internal server error: An unexpected error occurred." def test_translate_runtime_error(self): - """Test that RuntimeError without a sanitizer rule returns generic server error.""" + """Test that RuntimeError is translated to 500 HTTP status.""" exc = RuntimeError("Runtime error") result = translate_exception(exc)