litellm-mirror/litellm/types
Krish Dholakia 0178e75cd9 Litellm dev 12 30 2024 p1 (#7480)
* test(azure_openai_o1.py): initial commit with testing for azure openai o1 preview model

* fix(base_llm_unit_tests.py): handle azure o1 preview response format tests

skip as o1 on azure doesn't support tool calling yet

* fix: initial commit of azure o1 handler using openai caller

simplifies calling + allows fake streaming logic alr. implemented for openai to just work

* feat(azure/o1_handler.py): fake o1 streaming for azure o1 models

azure does not currently support streaming for o1

* feat(o1_transformation.py): support overriding 'should_fake_stream' on azure/o1 via 'supports_native_streaming' param on model info

enables user to toggle on when azure allows o1 streaming without needing to bump versions

* style(router.py): remove 'give feedback/get help' messaging when router is used

Prevents noisy messaging

Closes https://github.com/BerriAI/litellm/issues/5942

* test: fix azure o1 test

* test: fix tests

* fix: fix test
2024-12-30 21:52:52 -08:00
..
integrations Litellm dev 12 26 2024 p3 (#7434) 2024-12-26 21:21:16 -08:00
llms (Feat) Add logging for POST v1/fine_tuning/jobs (#7426) 2024-12-26 08:58:47 -08:00
passthrough_endpoints (docs) Simplify /vertex_ai/ pass through docs (#6910) 2024-11-25 23:57:50 -08:00
adapter.py feat(anthropic_adapter.py): support for translating anthropic params to openai format 2024-07-10 00:32:28 -07:00
caching.py (feat) - provider budget improvements - ensure provider budgets work with multiple proxy instances + improve latency to ~90ms (#6886) 2024-11-24 16:36:19 -08:00
completion.py LiteLLM Minor Fixes and Improvements (09/12/2024) (#5658) 2024-09-12 23:04:06 -07:00
embedding.py Removed config dict type definition 2024-05-17 10:39:00 +08:00
files.py Fix file type handling of uppercase extensions 2024-06-13 15:00:16 -07:00
guardrails.py (Feat) Log Guardrails run, guardrail response on logging integrations (#7445) 2024-12-27 15:01:56 -08:00
rerank.py test_rerank_response_assertions (#7476) 2024-12-30 10:12:56 -08:00
router.py Support budget/rate limit tiers for keys (#7429) 2024-12-26 19:05:27 -08:00
services.py Litellm perf improvements 3 (#6573) 2024-11-05 03:51:26 +05:30
utils.py Litellm dev 12 30 2024 p1 (#7480) 2024-12-30 21:52:52 -08:00