litellm-mirror/litellm
Ishaan Jaff dee18cbf31 (feat) add cost tracking for OpenAI prompt caching (#6055)
* add cache_read_input_token_cost for prompt caching models

* add prompt caching for latest models

* add openai cost calculator

* add openai prompt caching test

* fix lint check

* add not on how usage._cache_read_input_tokens is used

* fix cost calc whisper openai

* use output_cost_per_second

* add input_cost_per_second
2024-10-05 14:20:15 +05:30
..
adapters LiteLLM Minor Fixes & Improvements (09/27/2024) (#5938) 2024-09-27 22:52:57 -07:00
assistants Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
batches Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
deprecated_litellm_server Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
files Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
fine_tuning Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
integrations fix(gcs_bucket.py): show error response text in exception (#6072) 2024-10-05 11:56:43 +05:30
litellm_core_utils LiteLLM Minor Fixes & Improvements (10/04/2024) (#6064) 2024-10-04 21:28:53 -04:00
llms (feat) add cost tracking for OpenAI prompt caching (#6055) 2024-10-05 14:20:15 +05:30
proxy (feat) add /key/health endpoint to test key based logging (#6073) 2024-10-05 11:56:55 +05:30
realtime_api OpenAI /v1/realtime api support (#6047) 2024-10-03 17:11:22 -04:00
rerank_api LiteLLM Minor Fixes & Improvements (09/24/2024) (#5880) 2024-09-25 22:11:57 -07:00
router_strategy Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
router_utils fix prometheus track cooldown events on custom logger (#6060) 2024-10-04 16:56:22 +05:30
secret_managers Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
types (feat) add cost tracking for OpenAI prompt caching (#6055) 2024-10-05 14:20:15 +05:30
__init__.py Litellm Minor Fixes & Improvements (10/03/2024) (#6049) 2024-10-03 18:02:28 -04:00
_logging.py Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
_redis.py LiteLLM Minor Fixes & Improvements (09/21/2024) (#5819) 2024-09-21 18:51:53 -07:00
_service_logger.py [Fix] OTEL - Don't log messages when callback settings disable message logging (#5875) 2024-09-24 18:29:52 -07:00
_version.py Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
budget_manager.py Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
caching.py Litellm Minor Fixes & Improvements (10/03/2024) (#6049) 2024-10-03 18:02:28 -04:00
cost.json
cost_calculator.py (feat) add cost tracking for OpenAI prompt caching (#6055) 2024-10-05 14:20:15 +05:30
exceptions.py LiteLLM Minor Fixes & Improvements (09/21/2024) (#5819) 2024-09-21 18:51:53 -07:00
main.py fix(utils.py): return openai streaming prompt caching tokens (#6051) 2024-10-03 22:20:13 -04:00
model_prices_and_context_window_backup.json add azure o1 models to model cost map (#6075) 2024-10-05 13:22:06 +05:30
py.typed feature - Types for mypy - #360 2024-05-30 14:14:41 -04:00
requirements.txt
router.py OpenAI /v1/realtime api support (#6047) 2024-10-03 17:11:22 -04:00
scheduler.py feat(scheduler.py): support redis caching for req. prioritization 2024-06-06 14:19:21 -07:00
timeout.py Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
utils.py (feat) add cost tracking for OpenAI prompt caching (#6055) 2024-10-05 14:20:15 +05:30