litellm-mirror/litellm/types
Krish Dholakia b026230b0a Litellm dev 2024 12 19 p3 (#7322)
* fix(utils.py): remove unsupported optional params (if drop_params=True) before passing into map openai params

Fixes https://github.com/BerriAI/litellm/issues/7242

* test: new test for langfuse prompt management hook

Addresses https://github.com/BerriAI/litellm/issues/3893#issuecomment-2549080296

* feat(main.py): add 'get_chat_completion_prompt' customlogger hook

allows for langfuse prompt management

Addresses https://github.com/BerriAI/litellm/issues/3893#issuecomment-2549080296

* feat(langfuse_prompt_management.py): working e2e langfuse prompt management

works with `langfuse/` route

* feat(main.py): initial tracing for dynamic langfuse params

allows admin to specify langfuse keys by model in model_list

* feat(main.py): support passing langfuse credentials dynamically

* fix(langfuse_prompt_management.py): create langfuse client based on dynamic callback params

allows dynamic langfuse params to work

* fix: fix linting errors

* docs(prompt_management.md): refactor docs for sdk + proxy prompt management tutorial

* docs(prompt_management.md): cleanup doc

* docs: cleanup topnav

* docs(prompt_management.md): update docs to be easier to use

* fix: remove unused imports

* docs(prompt_management.md): add architectural overview doc

* fix(litellm_logging.py): fix dynamic param passing

* fix(langfuse_prompt_management.py): fix linting errors

* fix: fix linting errors

* fix: use typing_extensions for typealias to ensure python3.8 compatibility

* test: use stream_options in test to account for tiktoken diff

* fix: improve import error message, and check run test earlier
2024-12-20 13:30:16 -08:00
..
integrations (feat - proxy) Add status_code to litellm_proxy_total_requests_metric_total (#7293) 2024-12-18 15:55:02 -08:00
llms LiteLLM Minor Fixes & Improvements (12/16/2024) - p1 (#7263) 2024-12-17 15:33:36 -08:00
passthrough_endpoints (docs) Simplify /vertex_ai/ pass through docs (#6910) 2024-11-25 23:57:50 -08:00
adapter.py feat(anthropic_adapter.py): support for translating anthropic params to openai format 2024-07-10 00:32:28 -07:00
caching.py (feat) - provider budget improvements - ensure provider budgets work with multiple proxy instances + improve latency to ~90ms (#6886) 2024-11-24 16:36:19 -08:00
completion.py LiteLLM Minor Fixes and Improvements (09/12/2024) (#5658) 2024-09-12 23:04:06 -07:00
embedding.py Removed config dict type definition 2024-05-17 10:39:00 +08:00
files.py Fix file type handling of uppercase extensions 2024-06-13 15:00:16 -07:00
guardrails.py LiteLLM Minor Fixes & Improvements (10/15/2024) (#6242) 2024-10-16 07:32:06 -07:00
rerank.py (code refactor) - Add BaseRerankConfig. Use BaseRerankConfig for cohere/rerank and azure_ai/rerank (#7319) 2024-12-19 17:03:34 -08:00
router.py (feat proxy) v2 - model max budgets (#7302) 2024-12-18 19:42:46 -08:00
services.py Litellm perf improvements 3 (#6573) 2024-11-05 03:51:26 +05:30
utils.py Litellm dev 2024 12 19 p3 (#7322) 2024-12-20 13:30:16 -08:00