mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-24 18:24:20 +00:00
chore: fix haiku cache read pricing per token (#9834)
This commit is contained in:
parent
a3ea079583
commit
11389535d5
2 changed files with 2 additions and 2 deletions
|
@ -3467,7 +3467,7 @@
|
|||
"input_cost_per_token": 0.0000008,
|
||||
"output_cost_per_token": 0.000004,
|
||||
"cache_creation_input_token_cost": 0.000001,
|
||||
"cache_read_input_token_cost": 0.0000008,
|
||||
"cache_read_input_token_cost": 0.00000008,
|
||||
"litellm_provider": "anthropic",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
|
|
|
@ -3467,7 +3467,7 @@
|
|||
"input_cost_per_token": 0.0000008,
|
||||
"output_cost_per_token": 0.000004,
|
||||
"cache_creation_input_token_cost": 0.000001,
|
||||
"cache_read_input_token_cost": 0.0000008,
|
||||
"cache_read_input_token_cost": 0.00000008,
|
||||
"litellm_provider": "anthropic",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue