litellm-mirror/litellm/types
Krish Dholakia 76795dba39
Deepseek r1 support + watsonx qa improvements (#7907)
* fix(types/utils.py): support returning 'reasoning_content' for deepseek models

Fixes https://github.com/BerriAI/litellm/issues/7877#issuecomment-2603813218

* fix(convert_dict_to_response.py): return deepseek response in provider_specific_field

allows for separating openai vs. non-openai params in model response

* fix(utils.py): support 'provider_specific_field' in delta chunk as well

allows deepseek reasoning content chunk to be returned to user from stream as well

Fixes https://github.com/BerriAI/litellm/issues/7877#issuecomment-2603813218

* fix(watsonx/chat/handler.py): fix passing space id to watsonx on chat route

* fix(watsonx/): fix watsonx_text/ route with space id

* fix(watsonx/): qa item - also adds better unit testing for watsonx embedding calls

* fix(utils.py): rename to '..fields'

* fix: fix linting errors

* fix(utils.py): fix typing - don't show provider-specific field if none or empty - prevents default respons
e from being non-oai compatible

* fix: cleanup unused imports

* docs(deepseek.md): add docs for deepseek reasoning model
2025-01-21 23:13:15 -08:00
..
integrations (Feat) datadog_llm_observability callback - emit request_tags on logs (#7883) 2025-01-20 20:36:27 -08:00
llms Litellm dev 01 21 2025 p1 (#7898) 2025-01-21 20:36:11 -08:00
passthrough_endpoints (docs) Simplify /vertex_ai/ pass through docs (#6910) 2024-11-25 23:57:50 -08:00
adapter.py feat(anthropic_adapter.py): support for translating anthropic params to openai format 2024-07-10 00:32:28 -07:00
caching.py (feat) - provider budget improvements - ensure provider budgets work with multiple proxy instances + improve latency to ~90ms (#6886) 2024-11-24 16:36:19 -08:00
completion.py LiteLLM Minor Fixes and Improvements (09/12/2024) (#5658) 2024-09-12 23:04:06 -07:00
embedding.py Removed config dict type definition 2024-05-17 10:39:00 +08:00
files.py Fix file type handling of uppercase extensions 2024-06-13 15:00:16 -07:00
fine_tuning.py (feat) POST /fine_tuning/jobs support passing vertex specific hyper params (#7490) 2025-01-01 07:44:48 -08:00
guardrails.py Litellm dev 01 14 2025 p1 (#7771) 2025-01-14 16:18:21 -08:00
rerank.py test_rerank_response_assertions (#7476) 2024-12-30 10:12:56 -08:00
router.py LiteLLM Minor Fixes & Improvements (01/18/2025) - p1 (#7857) 2025-01-18 19:03:50 -08:00
services.py Litellm perf improvements 3 (#6573) 2024-11-05 03:51:26 +05:30
utils.py Deepseek r1 support + watsonx qa improvements (#7907) 2025-01-21 23:13:15 -08:00