mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
* test(azure_openai_o1.py): initial commit with testing for azure openai o1 preview model * fix(base_llm_unit_tests.py): handle azure o1 preview response format tests skip as o1 on azure doesn't support tool calling yet * fix: initial commit of azure o1 handler using openai caller simplifies calling + allows fake streaming logic alr. implemented for openai to just work * feat(azure/o1_handler.py): fake o1 streaming for azure o1 models azure does not currently support streaming for o1 * feat(o1_transformation.py): support overriding 'should_fake_stream' on azure/o1 via 'supports_native_streaming' param on model info enables user to toggle on when azure allows o1 streaming without needing to bump versions * style(router.py): remove 'give feedback/get help' messaging when router is used Prevents noisy messaging Closes https://github.com/BerriAI/litellm/issues/5942 * fix(types/utils.py): handle none logprobs Fixes https://github.com/BerriAI/litellm/issues/328 * fix(exception_mapping_utils.py): fix error str unbound error * refactor(azure_ai/): move to openai_like chat completion handler allows for easy swapping of api base url's (e.g. ai.services.com) Fixes https://github.com/BerriAI/litellm/issues/7275 * refactor(azure_ai/): move to base llm http handler * fix(azure_ai/): handle differing api endpoints * fix(azure_ai/): make sure all unit tests are passing * fix: fix linting errors * fix: fix linting errors * fix: fix linting error * fix: fix linting errors * fix(azure_ai/transformation.py): handle extra body param * fix(azure_ai/transformation.py): fix max retries param handling * fix: fix test * test(test_azure_o1.py): fix test * fix(llm_http_handler.py): support handling azure ai unprocessable entity error * fix(llm_http_handler.py): handle sync invalid param error for azure ai * fix(azure_ai/): streaming support with base_llm_http_handler * fix(llm_http_handler.py): working sync stream calls with unprocessable entity handling for azure ai * fix: fix linting errors * fix(llm_http_handler.py): fix linting error * fix(azure_ai/): handle cohere tool call invalid index param error |
||
---|---|---|
.. | ||
ai21/chat | ||
anthropic | ||
azure | ||
azure_ai | ||
base_llm | ||
bedrock | ||
cerebras | ||
clarifai | ||
cloudflare/chat | ||
codestral/completion | ||
cohere | ||
custom_httpx | ||
databricks | ||
deepgram | ||
deepinfra/chat | ||
deepseek | ||
deprecated_providers | ||
empower/chat | ||
fireworks_ai | ||
friendliai/chat | ||
galadriel/chat | ||
gemini | ||
github/chat | ||
groq | ||
hosted_vllm | ||
huggingface | ||
infinity/rerank | ||
jina_ai | ||
lm_studio | ||
mistral | ||
nlp_cloud | ||
nvidia_nim | ||
ollama | ||
oobabooga | ||
openai | ||
openai_like | ||
openrouter/chat | ||
perplexity/chat | ||
petals | ||
predibase | ||
replicate | ||
sagemaker | ||
sambanova | ||
together_ai | ||
triton | ||
vertex_ai | ||
vllm/completion | ||
voyage/embedding | ||
watsonx | ||
xai/chat | ||
__init__.py | ||
base.py | ||
baseten.py | ||
custom_llm.py | ||
maritalk.py | ||
ollama_chat.py | ||
README.md | ||
volcengine.py |
File Structure
August 27th, 2024
To make it easy to see how calls are transformed for each model/provider:
we are working on moving all supported litellm providers to a folder structure, where folder name is the supported litellm provider name.
Each folder will contain a *_transformation.py
file, which has all the request/response transformation logic, making it easy to see how calls are modified.
E.g. cohere/
, bedrock/
.