mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 04:04:14 +00:00
clean up
This commit is contained in:
parent
a6baa7b3d4
commit
0b9480a0e3
5 changed files with 0 additions and 18 deletions
|
@ -590,7 +590,6 @@ class InferenceRouter(Inference):
|
|||
|
||||
async def _nonstream_openai_chat_completion(self, provider: Inference, params: dict) -> OpenAIChatCompletion:
|
||||
response = await provider.openai_chat_completion(**params)
|
||||
|
||||
for choice in response.choices:
|
||||
# some providers return an empty list for no tool calls in non-streaming responses
|
||||
# but the OpenAI API returns None. So, set tool_calls to None if it's empty
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue