litellm-mirror/litellm/types
Ishaan Jaff 9d66976162 (feat) Add basic logging support for /batches endpoints (#7381)
* add basic logging for create`batch`

* add create_batch as a call type

* add basic dd logging for batches

* basic batch creation logging on DD
2024-12-23 17:45:03 -08:00
..
integrations (feat - proxy) Add status_code to litellm_proxy_total_requests_metric_total (#7293) 2024-12-18 15:55:02 -08:00
llms Complete 'requests' library removal (#7350) 2024-12-22 07:21:25 -08:00
passthrough_endpoints (docs) Simplify /vertex_ai/ pass through docs (#6910) 2024-11-25 23:57:50 -08:00
adapter.py feat(anthropic_adapter.py): support for translating anthropic params to openai format 2024-07-10 00:32:28 -07:00
caching.py (feat) - provider budget improvements - ensure provider budgets work with multiple proxy instances + improve latency to ~90ms (#6886) 2024-11-24 16:36:19 -08:00
completion.py LiteLLM Minor Fixes and Improvements (09/12/2024) (#5658) 2024-09-12 23:04:06 -07:00
embedding.py Removed config dict type definition 2024-05-17 10:39:00 +08:00
files.py Fix file type handling of uppercase extensions 2024-06-13 15:00:16 -07:00
guardrails.py LiteLLM Minor Fixes & Improvements (10/15/2024) (#6242) 2024-10-16 07:32:06 -07:00
rerank.py (code refactor) - Add BaseRerankConfig. Use BaseRerankConfig for cohere/rerank and azure_ai/rerank (#7319) 2024-12-19 17:03:34 -08:00
router.py (feat proxy) v2 - model max budgets (#7302) 2024-12-18 19:42:46 -08:00
services.py Litellm perf improvements 3 (#6573) 2024-11-05 03:51:26 +05:30
utils.py (feat) Add basic logging support for /batches endpoints (#7381) 2024-12-23 17:45:03 -08:00