Commit graph

13471 commits

Author SHA1 Message Date
Ishaan Jaff
55115bf520 transform_responses_api_request 2025-03-20 12:28:55 -07:00
Ishaan Jaff
af8e917eb0 MockResponsesAPIStreamingIterator 2025-03-20 12:25:58 -07:00
Ishaan Jaff
c64f834272 add AsyncMockResponsesAPIStreamingIterator 2025-03-20 12:06:36 -07:00
Ishaan Jaff
0cd671785d add fake_stream to llm http handler 2025-03-20 09:55:59 -07:00
Ishaan Jaff
bc174adcd0 add should_fake_stream 2025-03-20 09:54:26 -07:00
Krrish Dholakia
8ef9129556 fix(types/utils.py): support openai 'file' message type
Closes https://github.com/BerriAI/litellm/issues/9365
2025-03-19 23:13:51 -07:00
Krish Dholakia
7cf7e80c5a
Merge pull request #9395 from BerriAI/litellm_ui_fixes_03_19_2025
All checks were successful
Read Version from pyproject.toml / read-version (push) Successful in 18s
Helm unit test / unit-test (push) Successful in 21s
fix(internal_user_endpoints.py): re-introduce upsert on user not found
2025-03-19 22:58:32 -07:00
Krrish Dholakia
fe24b9d90b feat(azure/gpt_transformation.py): add azure audio model support
Closes https://github.com/BerriAI/litellm/issues/6305
2025-03-19 22:57:49 -07:00
Ishaan Jaff
9203910ab6 fix import hashlib 2025-03-19 21:08:19 -07:00
Ishaan Jaff
1bd7443c25
Merge pull request #9384 from BerriAI/litellm_prompt_management_custom
[Feat] - Allow building custom prompt management integration
2025-03-19 21:06:41 -07:00
Ishaan Jaff
247e4d09ee
Merge branch 'main' into litellm_fix_ssl_verify 2025-03-19 21:03:06 -07:00
Ishaan Jaff
dae5c24b55 get_chat_completion_prompt 2025-03-19 20:50:15 -07:00
Krrish Dholakia
056d34742f build(model_prices_and_context_window.json): fix native streaming flag 2025-03-19 19:53:19 -07:00
Krrish Dholakia
78781a0a55 build(model_prices_and_context_window.json): add o1-pro pricing 2025-03-19 19:35:35 -07:00
Krrish Dholakia
07b1e6aa1c fix(internal_user_endpoints.py): re-introduce upsert on user not found
Fixes https://github.com/BerriAI/litellm/issues/9243
2025-03-19 19:28:11 -07:00
Krish Dholakia
0de015b4c5
Revert "Fix latency redis" 2025-03-19 18:11:22 -07:00
Krish Dholakia
57503954fe
Merge pull request #9387 from emerzon/fix_latency_redis
Fix latency redis
2025-03-19 18:11:18 -07:00
Emerson Gomes
c91aafeed8 Handle empty valid_deployments in LowestLatencyLoggingHandler 2025-03-19 19:56:57 -05:00
Ishaan Jaff
a49bd460e0 get_custom_logger_for_prompt_management 2025-03-19 17:38:24 -07:00
Ishaan Jaff
805542005a CustomPromptManagement 2025-03-19 16:38:59 -07:00
Krish Dholakia
08e4614842
Merge pull request #9382 from BerriAI/litellm_dev_03_19_2025_p2
Support 'prisma migrate' for db schema changes
2025-03-19 16:37:47 -07:00
Ishaan Jaff
730f61f476 get_custom_loggers_for_type 2025-03-19 16:36:41 -07:00
Ishaan Jaff
f3d850c79a example X42PromptManagement 2025-03-19 16:24:41 -07:00
Ishaan Jaff
5d3bb86f07 define CustomPromptManagement 2025-03-19 16:22:23 -07:00
Krish Dholakia
9432d1a865
Merge pull request #9357 from BerriAI/litellm_dev_03_18_2025_p2
fix(lowest_tpm_rpm_v2.py): support batch writing increments to redis
2025-03-19 15:45:10 -07:00
Krrish Dholakia
61cd02200c fix: remove unused import 2025-03-19 15:33:04 -07:00
Krrish Dholakia
665fdfc788 feat(prisma_client.py): initial commit add prisma migration support to proxy 2025-03-19 14:26:59 -07:00
Krrish Dholakia
c13f6514bf fix(proxy_server.py): move db credential check to only run behind feature flag 2025-03-19 11:35:14 -07:00
Grayson Chen
f3a0261bb4 feat: Add support for custom OPENROUTER_API_BASE via get_secret in completion function 2025-03-19 21:09:03 +08:00
James Guthrie
437dbe7246 fix: VertexAI outputDimensionality configuration
VertexAI's API documentation [1] is an absolute mess. In it, they
describe the parameter to configure output dimensionality as
`output_dimensionality`. In the API example, they switch to using snake
case `outputDimensionality`, which is the correct variant.

[1]: https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/text-embeddings-api#generative-ai-get-text-embedding-drest
2025-03-19 11:07:36 +01:00
Krish Dholakia
01c6cbd270
Merge pull request #9363 from BerriAI/litellm_dev_03_18_2025_p3
All checks were successful
Read Version from pyproject.toml / read-version (push) Successful in 18s
Helm unit test / unit-test (push) Successful in 21s
fix(common_utils.py): handle cris only model
2025-03-18 23:36:12 -07:00
Krrish Dholakia
9adad381b4 fix(common_utils.py): handle cris only model
Fixes https://github.com/BerriAI/litellm/issues/9161#issuecomment-2734905153
2025-03-18 23:35:43 -07:00
Ishaan Jaff
e32aee9124
Merge pull request #9353 from BerriAI/litellm_arize_dynamic_logging
[Feat] - API - Allow using dynamic Arize AI Spaces on LiteLLM
2025-03-18 23:35:28 -07:00
Krish Dholakia
6347b694ee
Merge pull request #9335 from BerriAI/litellm_dev_03_17_2025_p3
Contributor PR: Fix sagemaker too little data for content error
2025-03-18 23:24:07 -07:00
Krrish Dholakia
2c69b5d221 fix(base_routing_strategy.py): refactor for cleaner code 2025-03-18 22:57:05 -07:00
Krrish Dholakia
084e8c425c refactor(base_routing_strategy.py): fix function names 2025-03-18 22:41:02 -07:00
Krrish Dholakia
0194c18820 build(model_prices_and_context_window.json): fix azure gpt-4o pricing
fix default to use latest global pricing
2025-03-18 22:34:58 -07:00
Ishaan Jaff
f5ef0c3cb7 fix code quality checks 2025-03-18 22:34:43 -07:00
Krrish Dholakia
3033c40739 fix(base_routing_strategy.py): fix base to handle no running event loop
run in a separate thread
2025-03-18 22:20:39 -07:00
Ishaan Jaff
0f2e095b6b _arize_otel_logger 2025-03-18 22:19:51 -07:00
Ishaan Jaff
57e5c94360 Merge branch 'main' into litellm_arize_dynamic_logging 2025-03-18 22:13:35 -07:00
Ishaan Jaff
c101fe9b5d
Merge pull request #9352 from BerriAI/litellm_arize_mar_18
[Bug Fix] Arize AI Logging Integration with LiteLLM
2025-03-18 22:12:46 -07:00
Krrish Dholakia
a3d000baaa fix(test_base_routing_strategy.py): add unit testing for new base routing strategy test 2025-03-18 19:59:06 -07:00
Krrish Dholakia
ef008138a3 feat(base_routing_strategy.py): handle updating in-memory keys 2025-03-18 19:44:04 -07:00
Krrish Dholakia
1328afe612 fix(lowest_tpm_rpm_v2.py): support batch writing increments to redis 2025-03-18 19:09:53 -07:00
Ishaan Jaff
65083ca8da get_openai_client_cache_key 2025-03-18 18:35:50 -07:00
Krish Dholakia
24fbadf072
Merge pull request #9355 from BerriAI/litellm_dev_03_18_2025_p1
Litellm dev 03 18 2025 p1
2025-03-18 18:34:06 -07:00
Ishaan Jaff
3daef0d740 fix common utils 2025-03-18 17:59:46 -07:00
Ishaan Jaff
a45830dac3 use common caching logic for openai/azure clients 2025-03-18 17:57:03 -07:00
Ishaan Jaff
f73e9047dc use common logic for re-using openai clients 2025-03-18 17:56:32 -07:00