mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
Gemini 2.5 Flash output cost is based on thinking enabled
This commit is contained in:
parent
0b63c7a2eb
commit
58c2551fa6
7 changed files with 74 additions and 10 deletions
|
@ -4557,6 +4557,9 @@ def _get_model_info_helper( # noqa: PLR0915
|
|||
"output_cost_per_token_batches"
|
||||
),
|
||||
output_cost_per_token=_output_cost_per_token,
|
||||
output_cost_per_token_thinking=_model_info.get(
|
||||
"output_cost_per_token_thinking", None
|
||||
),
|
||||
output_cost_per_audio_token=_model_info.get(
|
||||
"output_cost_per_audio_token", None
|
||||
),
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue