litellm-mirror/litellm/llms
Krish Dholakia 0120176541
Litellm dev 12 30 2024 p2 (#7495)
* test(azure_openai_o1.py): initial commit with testing for azure openai o1 preview model

* fix(base_llm_unit_tests.py): handle azure o1 preview response format tests

skip as o1 on azure doesn't support tool calling yet

* fix: initial commit of azure o1 handler using openai caller

simplifies calling + allows fake streaming logic alr. implemented for openai to just work

* feat(azure/o1_handler.py): fake o1 streaming for azure o1 models

azure does not currently support streaming for o1

* feat(o1_transformation.py): support overriding 'should_fake_stream' on azure/o1 via 'supports_native_streaming' param on model info

enables user to toggle on when azure allows o1 streaming without needing to bump versions

* style(router.py): remove 'give feedback/get help' messaging when router is used

Prevents noisy messaging

Closes https://github.com/BerriAI/litellm/issues/5942

* fix(types/utils.py): handle none logprobs

Fixes https://github.com/BerriAI/litellm/issues/328

* fix(exception_mapping_utils.py): fix error str unbound error

* refactor(azure_ai/): move to openai_like chat completion handler

allows for easy swapping of api base url's (e.g. ai.services.com)

Fixes https://github.com/BerriAI/litellm/issues/7275

* refactor(azure_ai/): move to base llm http handler

* fix(azure_ai/): handle differing api endpoints

* fix(azure_ai/): make sure all unit tests are passing

* fix: fix linting errors

* fix: fix linting errors

* fix: fix linting error

* fix: fix linting errors

* fix(azure_ai/transformation.py): handle extra body param

* fix(azure_ai/transformation.py): fix max retries param handling

* fix: fix test

* test(test_azure_o1.py): fix test

* fix(llm_http_handler.py): support handling azure ai unprocessable entity error

* fix(llm_http_handler.py): handle sync invalid param error for azure ai

* fix(azure_ai/): streaming support with base_llm_http_handler

* fix(llm_http_handler.py): working sync stream calls with unprocessable entity handling for azure ai

* fix: fix linting errors

* fix(llm_http_handler.py): fix linting error

* fix(azure_ai/): handle cohere tool call invalid index param error
2025-01-01 18:57:29 -08:00
..
ai21/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
anthropic Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
azure Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
azure_ai Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
base_llm Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
bedrock Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
cerebras (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
clarifai Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
cloudflare/chat Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
codestral/completion (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
cohere Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
custom_httpx Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
databricks (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
deepgram Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
deepinfra/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
deepseek LiteLLM Minor Fixes & Improvements (12/23/2024) - p3 (#7394) 2024-12-23 22:02:52 -08:00
deprecated_providers (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
empower/chat LiteLLM Common Base LLM Config (pt.3): Move all OAI compatible providers to base llm config (#7148) 2024-12-10 17:12:42 -08:00
fireworks_ai Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
friendliai/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
galadriel/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
gemini Litellm dev 12 28 2024 p1 (#7463) 2024-12-28 20:26:00 -08:00
github/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
groq (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
hosted_vllm (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
huggingface Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
infinity/rerank (feat) add infinity rerank models (#7321) 2024-12-19 18:30:28 -08:00
jina_ai (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
lm_studio (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
mistral (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
nlp_cloud Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
nvidia_nim (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
ollama Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
oobabooga Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
openai Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
openai_like Litellm dev 12 25 2024 p1 (#7411) 2024-12-25 17:36:30 -08:00
openrouter/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
perplexity/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
petals Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
predibase Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
replicate Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
sagemaker Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
sambanova (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
together_ai Litellm dev 12 25 2025 p2 (#7420) 2024-12-25 18:35:34 -08:00
triton Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
vertex_ai Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
vllm/completion (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
voyage/embedding Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
watsonx Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
xai/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
__init__.py add linting 2023-08-18 11:05:05 -07:00
base.py Complete 'requests' library removal (#7350) 2024-12-22 07:21:25 -08:00
baseten.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
custom_llm.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
maritalk.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
ollama_chat.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
README.md LiteLLM Minor Fixes and Improvements (09/13/2024) (#5689) 2024-09-14 10:02:55 -07:00
volcengine.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00

File Structure

August 27th, 2024

To make it easy to see how calls are transformed for each model/provider:

we are working on moving all supported litellm providers to a folder structure, where folder name is the supported litellm provider name.

Each folder will contain a *_transformation.py file, which has all the request/response transformation logic, making it easy to see how calls are modified.

E.g. cohere/, bedrock/.