fix: litellm tool call parsing event type to in_progress (#1312)
# What does this PR do?
- Test with script:
https://gist.github.com/yanxi0830/64699f3604766ac2319421b750c5bf9c
- Agent with tool calls does not get correctly parsed with LiteLLM
provider b/c we skip processing
`ChatCompletionResponseEventType.complete`.
- However, LiteLLM spits out event_type="complete" with ToolCallDelta
2f7683bc5f/llama_stack/providers/inline/agents/meta_reference/agent_instance.py (L570-L577)
- Llama Model
```
ChatCompletionResponseStreamChunk(
│ event=Event(
│ │ delta=ToolCallDelta(
│ │ │ parse_status='succeeded',
│ │ │ tool_call=ToolCall(
│ │ │ │ arguments={'kind': 'pod', 'namespace': 'openshift-lightspeed'},
│ │ │ │ call_id='call_tIjWTUdsQXhQ2XHC5ke4EQY5',
│ │ │ │ tool_name='get_object_namespace_list'
│ │ │ ),
│ │ │ type='tool_call'
│ │ ),
│ │ event_type='progress',
│ │ logprobs=None,
│ │ stop_reason='end_of_turn'
│ ),
│ metrics=None
)
ChatCompletionResponseStreamChunk(
│ event=Event(
│ │ delta=TextDelta(text='', type='text'),
│ │ event_type='complete',
│ │ logprobs=None,
│ │ stop_reason='end_of_turn'
│ ),
│ metrics=None
)
```
- LiteLLM model
```
ChatCompletionResponseStreamChunk(
│ event=Event(
│ │ delta=ToolCallDelta(
│ │ │ parse_status='succeeded',
│ │ │ tool_call=ToolCall(
│ │ │ │ arguments={'kind': 'pod', 'namespace': 'openshift-lightspeed'},
│ │ │ │ call_id='call_tIjWTUdsQXhQ2XHC5ke4EQY5',
│ │ │ │ tool_name='get_object_namespace_list'
│ │ │ ),
│ │ │ type='tool_call'
│ │ ),
│ │ event_type='complete',
│ │ logprobs=None,
│ │ stop_reason='end_of_turn'
│ ),
│ metrics=None
)
ChatCompletionResponseStreamChunk(
│ event=Event(
│ │ delta=TextDelta(text='', type='text'),
│ │ event_type='complete',
│ │ logprobs=None,
│ │ stop_reason='end_of_turn'
│ ),
│ metrics=None
)
```
[//]: # (If resolving an issue, uncomment and update the line below)
[//]: # (Closes #[issue-number])
## Test Plan
- Test with script:
https://gist.github.com/yanxi0830/64699f3604766ac2319421b750c5bf9c
[//]: # (## Documentation)