mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-13 06:07:58 +00:00
fix(misc): pre-commit fix for server.py
This commit is contained in:
parent
7c63aebd64
commit
e6378872c7
1 changed files with 1 additions and 3 deletions
|
@ -273,9 +273,7 @@ def create_dynamic_typed_route(func: Any, method: str, route: str) -> Callable:
|
|||
context_vars = [CURRENT_TRACE_CONTEXT, PROVIDER_DATA_VAR]
|
||||
if test_context_var is not None:
|
||||
context_vars.append(test_context_var)
|
||||
gen = preserve_contexts_async_generator(
|
||||
sse_generator(func(**kwargs)), context_vars
|
||||
)
|
||||
gen = preserve_contexts_async_generator(sse_generator(func(**kwargs)), context_vars)
|
||||
return StreamingResponse(gen, media_type="text/event-stream")
|
||||
else:
|
||||
value = func(**kwargs)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue