mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-03 09:53:45 +00:00
clean
This commit is contained in:
parent
189251b56f
commit
232f687443
2 changed files with 2 additions and 1 deletions
|
|
@ -363,6 +363,7 @@ class OpenAIResponsesImpl:
|
||||||
|
|
||||||
# Structured outputs
|
# Structured outputs
|
||||||
response_format = await convert_response_text_to_chat_response_format(text)
|
response_format = await convert_response_text_to_chat_response_format(text)
|
||||||
|
|
||||||
ctx = ChatCompletionContext(
|
ctx = ChatCompletionContext(
|
||||||
model=model,
|
model=model,
|
||||||
messages=messages,
|
messages=messages,
|
||||||
|
|
|
||||||
|
|
@ -165,7 +165,7 @@ class TestTranslateException:
|
||||||
assert result.detail == "Internal server error: An unexpected error occurred."
|
assert result.detail == "Internal server error: An unexpected error occurred."
|
||||||
|
|
||||||
def test_translate_runtime_error(self):
|
def test_translate_runtime_error(self):
|
||||||
"""Test that RuntimeError without a sanitizer rule returns generic server error."""
|
"""Test that RuntimeError is translated to 500 HTTP status."""
|
||||||
exc = RuntimeError("Runtime error")
|
exc = RuntimeError("Runtime error")
|
||||||
result = translate_exception(exc)
|
result = translate_exception(exc)
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue