litellm-mirror/litellm/llms/bedrock/chat
Krish Dholakia 87733c8193
Fix anthropic prompt caching cost calc + trim logged message in db (#9838)
* fix(spend_tracking_utils.py): prevent logging entire mp4 files to db

Fixes https://github.com/BerriAI/litellm/issues/9732

* fix(anthropic/chat/transformation.py): Fix double counting cache creation input tokens

Fixes https://github.com/BerriAI/litellm/issues/9812

* refactor(anthropic/chat/transformation.py): refactor streaming to use same usage calculation block as non-streaming

reduce errors

* fix(bedrock/chat/converse_transformation.py): don't increment prompt tokens with cache_creation_input_tokens

* build: remove redisvl from requirements.txt (temporary)

* fix(spend_tracking_utils.py): handle circular references

* test: update code cov test

* test: update test
2025-04-09 21:26:43 -07:00
..
converse_like Litellm dev 01 29 2025 p2 (#8102) 2025-01-29 20:53:37 -08:00
invoke_transformations VertexAI non-jsonl file storage support (#9781) 2025-04-09 14:01:48 -07:00
__init__.py Litellm stable dev (#5711) 2024-09-14 23:22:59 -07:00
converse_handler.py build(pyproject.toml): add new dev dependencies - for type checking (#9631) 2025-03-29 11:02:13 -07:00
converse_transformation.py Fix anthropic prompt caching cost calc + trim logged message in db (#9838) 2025-04-09 21:26:43 -07:00
invoke_handler.py build(pyproject.toml): add new dev dependencies - for type checking (#9631) 2025-03-29 11:02:13 -07:00