litellm-mirror/tests/litellm/llms
Krish Dholakia 87733c8193
Fix anthropic prompt caching cost calc + trim logged message in db (#9838)
* fix(spend_tracking_utils.py): prevent logging entire mp4 files to db

Fixes https://github.com/BerriAI/litellm/issues/9732

* fix(anthropic/chat/transformation.py): Fix double counting cache creation input tokens

Fixes https://github.com/BerriAI/litellm/issues/9812

* refactor(anthropic/chat/transformation.py): refactor streaming to use same usage calculation block as non-streaming

reduce errors

* fix(bedrock/chat/converse_transformation.py): don't increment prompt tokens with cache_creation_input_tokens

* build: remove redisvl from requirements.txt (temporary)

* fix(spend_tracking_utils.py): handle circular references

* test: update code cov test

* test: update test
2025-04-09 21:26:43 -07:00
..
anthropic/chat Fix anthropic prompt caching cost calc + trim logged message in db (#9838) 2025-04-09 21:26:43 -07:00
azure get_openai_client_cache_key 2025-03-18 18:35:50 -07:00
azure_ai/chat test: refactor testing to handle routing correctly 2025-03-18 12:24:12 -07:00
bedrock Fix anthropic prompt caching cost calc + trim logged message in db (#9838) 2025-04-09 21:26:43 -07:00
chat update test 2025-03-10 20:34:52 -07:00
cohere/chat Add support for max_completion_tokens to the Cohere chat transformation config (#9701) 2025-04-02 07:50:44 -07:00
custom_httpx Add OpenAI gpt-4o-transcribe support (#9517) 2025-03-26 23:10:25 -07:00
databricks/chat Add DBRX Anthropic w/ thinking + response_format support (#9744) 2025-04-04 22:13:32 -07:00
deepgram/audio_transcription Add OpenAI gpt-4o-transcribe support (#9517) 2025-03-26 23:10:25 -07:00
openai test_openai_client_reuse 2025-03-18 18:13:36 -07:00
openrouter/chat fix #8425, passthrough kwargs during acompletion, and unwrap extra_body for openrouter (#9747) 2025-04-03 22:19:40 -07:00
sagemaker ref issue 2025-03-31 16:05:10 -07:00
vertex_ai Fix VertexAI Credential Caching issue (#9756) 2025-04-04 16:38:08 -07:00