Commit graph

20570 commits

Author SHA1 Message Date
Ishaan Jaff
066637bbd1 custom Prompt Management - Overview 2025-03-19 14:31:56 -07:00
Krrish Dholakia
a2d3d8017e feat(prisma_client.py): initial commit add prisma migration support to proxy 2025-03-19 14:26:59 -07:00
Krish Dholakia
f5f2e95bba Merge pull request #9376 from BerriAI/litellm_fix_cred_check
Feature flag checking LiteLLM_CredentialsTable
2025-03-19 12:47:44 -07:00
Krrish Dholakia
ee89f6ae41 docs(config_settings.md): cleanup 2025-03-19 12:06:07 -07:00
Krrish Dholakia
f4edbe0787 test(test_proxy_server.py): make test work on ci/cd 2025-03-19 12:01:37 -07:00
Krrish Dholakia
10bfcf1bdb docs(config_settings.md): update env var information 2025-03-19 11:47:14 -07:00
Krrish Dholakia
f008f1044f test(test_proxy_server.py): add unit test to ensure get credentials only called behind feature flag 2025-03-19 11:44:00 -07:00
Krrish Dholakia
db8a459c89 fix(proxy_server.py): move db credential check to only run behind feature flag 2025-03-19 11:35:14 -07:00
Ishaan Jaff
818c7d9089 Merge pull request #9375 from colesmcintosh/fix-litellm-proxy-docs
docs(litellm_proxy): correct parameter assignment in litellm proxy docs
2025-03-19 11:14:19 -07:00
Cole McIntosh
943f718b2c docs(litellm_proxy): update parameter assignment for streaming call 2025-03-19 11:13:17 -06:00
Cole McIntosh
d9231105d3 docs(litellm_proxy): correct parameter assignment in sample litellm proxy call docs 2025-03-19 11:08:34 -06:00
Krrish Dholakia
bb2d201d11 docs(enterprise.md): clarify sla's 2025-03-19 09:09:02 -07:00
azdolinski
70a533d186 sync_latest_changes 2025-03-19 14:47:34 +00:00
Artur Zdolinski
c0f075898c Merge branch 'BerriAI:main' into litellm_fix_icons 2025-03-19 15:40:49 +01:00
azdolinski
a2d71f8f8b set_local_icons 2025-03-19 14:37:57 +00:00
Grayson Chen
b186210cea feat: Add support for custom OPENROUTER_API_BASE via get_secret in completion function 2025-03-19 21:09:03 +08:00
James Guthrie
4044bca614 fix: VertexAI outputDimensionality configuration
VertexAI's API documentation [1] is an absolute mess. In it, they
describe the parameter to configure output dimensionality as
`output_dimensionality`. In the API example, they switch to using snake
case `outputDimensionality`, which is the correct variant.

[1]: https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/text-embeddings-api#generative-ai-get-text-embedding-drest
2025-03-19 11:07:36 +01:00
Krish Dholakia
887648a364 Merge pull request #9363 from BerriAI/litellm_dev_03_18_2025_p3
fix(common_utils.py): handle cris only model
2025-03-18 23:36:12 -07:00
Krrish Dholakia
db3a65d52a fix(common_utils.py): handle cris only model
Fixes https://github.com/BerriAI/litellm/issues/9161#issuecomment-2734905153
2025-03-18 23:35:43 -07:00
Ishaan Jaff
597afb73a6 Merge pull request #9353 from BerriAI/litellm_arize_dynamic_logging
[Feat] - API - Allow using dynamic Arize AI Spaces on LiteLLM
2025-03-18 23:35:28 -07:00
Krish Dholakia
27ff9d3d30 Merge pull request #9335 from BerriAI/litellm_dev_03_17_2025_p3
Contributor PR: Fix sagemaker too little data for content error
2025-03-18 23:24:07 -07:00
Ishaan Jaff
ca63887085 test_arize_dynamic_params 2025-03-18 23:22:55 -07:00
Ishaan Jaff
4b6c930cc9 test_arize_dynamic_params 2025-03-18 23:18:07 -07:00
Krrish Dholakia
381b220e60 fix(base_routing_strategy.py): refactor for cleaner code 2025-03-18 22:57:05 -07:00
Krish Dholakia
4df1d33880 Merge pull request #9361 from BerriAI/litellm_fix_azure_gpt_4o_pricing
build(model_prices_and_context_window.json): fix azure gpt-4o pricing
2025-03-18 22:42:38 -07:00
Krrish Dholakia
9a00e0a008 refactor(base_routing_strategy.py): fix function names 2025-03-18 22:41:02 -07:00
Krrish Dholakia
b122212bc5 build(model_prices_and_context_window.json): fix azure gpt-4o pricing
fix default to use latest global pricing
2025-03-18 22:34:58 -07:00
Ishaan Jaff
afd7d066bf fix code quality checks 2025-03-18 22:34:43 -07:00
Krrish Dholakia
36c9517e28 fix(base_routing_strategy.py): fix base to handle no running event loop
run in a separate thread
2025-03-18 22:20:39 -07:00
Ishaan Jaff
a27782e657 _arize_otel_logger 2025-03-18 22:19:51 -07:00
Ishaan Jaff
43eedf26f7 Merge branch 'main' into litellm_arize_dynamic_logging 2025-03-18 22:13:35 -07:00
Ishaan Jaff
5a327da78e Merge pull request #9352 from BerriAI/litellm_arize_mar_18
[Bug Fix] Arize AI Logging Integration with LiteLLM
2025-03-18 22:12:46 -07:00
Ishaan Jaff
ac8618823d test_arize_callback 2025-03-18 20:21:23 -07:00
Krrish Dholakia
000c2598db fix(test_base_routing_strategy.py): add unit testing for new base routing strategy test 2025-03-18 19:59:06 -07:00
Krrish Dholakia
3867d90b08 feat(base_routing_strategy.py): handle updating in-memory keys 2025-03-18 19:44:04 -07:00
Krrish Dholakia
60f6e1d379 fix(lowest_tpm_rpm_v2.py): support batch writing increments to redis 2025-03-18 19:09:53 -07:00
Ishaan Jaff
95f08cec55 test_arize_callback 2025-03-18 18:49:06 -07:00
Ishaan Jaff
eaab5da6eb test arize logging 2025-03-18 18:40:39 -07:00
Ishaan Jaff
55e669d7d8 get_openai_client_cache_key 2025-03-18 18:35:50 -07:00
Krish Dholakia
49bc799c04 Merge pull request #9355 from BerriAI/litellm_dev_03_18_2025_p1
Litellm dev 03 18 2025 p1
2025-03-18 18:34:06 -07:00
Ishaan Jaff
bb8400a350 test_openai_client_reuse 2025-03-18 18:13:36 -07:00
Ishaan Jaff
4307316f4d test openai common utils 2025-03-18 18:08:24 -07:00
Ishaan Jaff
4bac8f53a5 fix common utils 2025-03-18 17:59:46 -07:00
Ishaan Jaff
9f31177a20 use common caching logic for openai/azure clients 2025-03-18 17:57:03 -07:00
Ishaan Jaff
ef91a0c72b use common logic for re-using openai clients 2025-03-18 17:56:32 -07:00
Krrish Dholakia
c8eaf8e419 fix(response_metadata.py): log the litellm_model_name
make it easier to track the model sent to the provider
2025-03-18 17:46:33 -07:00
Krrish Dholakia
d526a2f929 test(test_tpm_rpm_routing_v2.py): initial test, for asserting async pre call check works as expected 2025-03-18 17:36:55 -07:00
Ishaan Jaff
75839c6e61 fix ensure async client test 2025-03-18 17:29:10 -07:00
Ishaan Jaff
268dfa90d7 test_dynamic_azure_params 2025-03-18 17:26:23 -07:00
Krrish Dholakia
d29108bfd5 fix(litellm_logging.py): fix update 2025-03-18 17:18:58 -07:00