litellm-mirror/litellm/types
Krish Dholakia 953c021aa7 Litellm dev 01 10 2025 p3 (#7682)
* feat(langfuse.py): log the used prompt when prompt management used

* test: fix test

* docs(self_serve.md): add doc on restricting personal key creation on ui

* feat(s3.py): support s3 logging with team alias prefixes (if available)

New preview feature

* fix(main.py): remove old if block - simplify to just await if coroutine returned

fixes lm_studio async embedding error

* fix(langfuse.py): handle get prompt check
2025-01-10 21:56:42 -08:00
..
integrations [Feature]: - allow print alert log to console (#7534) 2025-01-03 17:48:13 -08:00
llms (feat) POST /fine_tuning/jobs support passing vertex specific hyper params (#7490) 2025-01-01 07:44:48 -08:00
passthrough_endpoints (docs) Simplify /vertex_ai/ pass through docs (#6910) 2024-11-25 23:57:50 -08:00
adapter.py feat(anthropic_adapter.py): support for translating anthropic params to openai format 2024-07-10 00:32:28 -07:00
caching.py (feat) - provider budget improvements - ensure provider budgets work with multiple proxy instances + improve latency to ~90ms (#6886) 2024-11-24 16:36:19 -08:00
completion.py LiteLLM Minor Fixes and Improvements (09/12/2024) (#5658) 2024-09-12 23:04:06 -07:00
embedding.py Removed config dict type definition 2024-05-17 10:39:00 +08:00
files.py Fix file type handling of uppercase extensions 2024-06-13 15:00:16 -07:00
fine_tuning.py (feat) POST /fine_tuning/jobs support passing vertex specific hyper params (#7490) 2025-01-01 07:44:48 -08:00
guardrails.py (Feat) Log Guardrails run, guardrail response on logging integrations (#7445) 2024-12-27 15:01:56 -08:00
rerank.py test_rerank_response_assertions (#7476) 2024-12-30 10:12:56 -08:00
router.py Litellm dev 12 31 2024 p1 (#7488) 2024-12-31 23:21:51 -08:00
services.py Litellm perf improvements 3 (#6573) 2024-11-05 03:51:26 +05:30
utils.py Litellm dev 01 10 2025 p3 (#7682) 2025-01-10 21:56:42 -08:00