litellm-mirror/litellm/llms
Krish Dholakia 75c3ddfc9e fix(vertex_ai/gemini/transformation.py): handle 'http://' in gemini p… (#7660)
* fix(vertex_ai/gemini/transformation.py): handle 'http://' in gemini process url

* refactor(router.py): refactor '_prompt_management_factory' to use logging obj get_chat_completion logic

deduplicates code

* fix(litellm_logging.py): update 'get_chat_completion_prompt' to update logging object messages

* docs(prompt_management.md): update prompt management to be in beta

given feedback - this still needs to be revised (e.g. passing in user message, not ignoring)

* refactor(prompt_management_base.py): introduce base class for prompt management

allows consistent behaviour across prompt management integrations

* feat(prompt_management_base.py): support adding client message to template message + refactor langfuse prompt management to use prompt management base

* fix(litellm_logging.py): log prompt id + prompt variables to langfuse if set

allows tracking what prompt was used for what purpose

* feat(litellm_logging.py): log prompt management metadata in standard logging payload + use in langfuse

allows logging prompt id / prompt variables to langfuse

* test: fix test

* fix(router.py): cleanup unused imports

* fix: fix linting error

* fix: fix trace param typing

* fix: fix linting errors

* fix: fix code qa check
2025-01-10 07:31:59 -08:00
..
ai21/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
aiohttp_openai/chat (proxy - RPS) - Get 2K RPS at 4 instances, minor fix aiohttp_openai/ (#7659) 2025-01-09 17:24:18 -08:00
anthropic Litellm dev 01 07 2025 p2 (#7622) 2025-01-08 16:56:39 -08:00
azure Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
azure_ai fix unused imports 2025-01-02 22:28:22 -08:00
base_llm Support checking provider-specific /models endpoints for available models based on key (#7538) 2025-01-03 19:29:59 -08:00
bedrock Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
cerebras (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
clarifai Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
cloudflare/chat Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
codestral/completion (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
cohere Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
custom_httpx aiohttp_openai/ fixes - allow using aiohttp_openai/gpt-4o (#7598) 2025-01-06 21:39:11 -08:00
databricks LiteLLM Minor Fixes & Improvements (01/08/2025) - p2 (#7643) 2025-01-08 19:45:19 -08:00
deepgram Litellm dev 01 02 2025 p2 (#7512) 2025-01-02 21:57:51 -08:00
deepinfra/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
deepseek LiteLLM Minor Fixes & Improvements (12/23/2024) - p3 (#7394) 2024-12-23 22:02:52 -08:00
deprecated_providers (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
empower/chat LiteLLM Common Base LLM Config (pt.3): Move all OAI compatible providers to base llm config (#7148) 2024-12-10 17:12:42 -08:00
fireworks_ai Support checking provider-specific /models endpoints for available models based on key (#7538) 2025-01-03 19:29:59 -08:00
friendliai/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
galadriel/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
gemini Litellm dev 12 28 2024 p1 (#7463) 2024-12-28 20:26:00 -08:00
github/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
groq fix(groq/chat/transformation.py): fix groq response_format transformation (#7565) 2025-01-04 19:39:04 -08:00
hosted_vllm (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
huggingface Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
infinity/rerank (feat) add infinity rerank models (#7321) 2024-12-19 18:30:28 -08:00
jina_ai (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
litellm_proxy/chat Support checking provider-specific /models endpoints for available models based on key (#7538) 2025-01-03 19:29:59 -08:00
lm_studio (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
mistral (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
nlp_cloud Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
nvidia_nim (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
ollama Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
oobabooga Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
openai feat(router.py): support request prioritization for text completion c… (#7540) 2025-01-03 19:35:44 -08:00
openai_like Litellm dev 12 25 2024 p1 (#7411) 2024-12-25 17:36:30 -08:00
openrouter/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
perplexity/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
petals Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
predibase Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
replicate Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
sagemaker Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
sambanova (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
together_ai Litellm dev 12 25 2025 p2 (#7420) 2024-12-25 18:35:34 -08:00
triton Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
vertex_ai fix(vertex_ai/gemini/transformation.py): handle 'http://' in gemini p… (#7660) 2025-01-10 07:31:59 -08:00
vllm/completion (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
voyage/embedding Litellm dev 12 30 2024 p2 (#7495) 2025-01-01 18:57:29 -08:00
watsonx Support checking provider-specific /models endpoints for available models based on key (#7538) 2025-01-03 19:29:59 -08:00
xai/chat (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
__init__.py add linting 2023-08-18 11:05:05 -07:00
base.py Complete 'requests' library removal (#7350) 2024-12-22 07:21:25 -08:00
baseten.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
custom_llm.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
maritalk.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
ollama_chat.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
README.md LiteLLM Minor Fixes and Improvements (09/13/2024) (#5689) 2024-09-14 10:02:55 -07:00
volcengine.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00

File Structure

August 27th, 2024

To make it easy to see how calls are transformed for each model/provider:

we are working on moving all supported litellm providers to a folder structure, where folder name is the supported litellm provider name.

Each folder will contain a *_transformation.py file, which has all the request/response transformation logic, making it easy to see how calls are modified.

E.g. cohere/, bedrock/.