minor fix

This commit is contained in:
Swapna Lekkala 2025-09-11 16:10:23 -07:00
parent f9348a6bdf
commit 4271f7352e

View file

@ -1417,10 +1417,9 @@ class OpenAIChatCompletionToLlamaStackMixin:
if stream: if stream:
return OpenAIChatCompletionToLlamaStackMixin._process_stream_response(self, model, outstanding_responses) return OpenAIChatCompletionToLlamaStackMixin._process_stream_response(self, model, outstanding_responses)
response = await OpenAIChatCompletionToLlamaStackMixin._process_non_stream_response( return await OpenAIChatCompletionToLlamaStackMixin._process_non_stream_response(
self, model, outstanding_responses self, model, outstanding_responses
) )
return response
async def _process_stream_response( async def _process_stream_response(
self, self,
@ -1512,11 +1511,9 @@ class OpenAIChatCompletionToLlamaStackMixin:
) )
choices.append(choice) choices.append(choice)
usage = None usage = OpenAIChatCompletionUsage(
if total_tokens > 0: prompt_tokens=total_prompt_tokens, completion_tokens=total_completion_tokens, total_tokens=total_tokens
usage = OpenAIChatCompletionUsage( )
prompt_tokens=total_prompt_tokens, completion_tokens=total_completion_tokens, total_tokens=total_tokens
)
return OpenAIChatCompletion( return OpenAIChatCompletion(
id=f"chatcmpl-{uuid.uuid4()}", id=f"chatcmpl-{uuid.uuid4()}",