Commit graph

1401 commits

Author SHA1 Message Date
Krish Dholakia
f55d6b686b Merge pull request #3354 from BerriAI/litellm_replicate_cost_tracking
fix(utils.py): replicate now also has token based pricing for some models
2024-04-29 09:13:41 -07:00
Krrish Dholakia
2a88543f4d fix(utils.py): default to time-based tracking for unmapped replicate models. fix time-based cost calc for replicate 2024-04-29 08:36:01 -07:00
Krrish Dholakia
453f3d068d fix(utils.py): use llama tokenizer for replicate models 2024-04-29 08:28:31 -07:00
Krrish Dholakia
6d3dc05da2 fix(utils.py): fix watson streaming 2024-04-29 08:09:59 -07:00
Krrish Dholakia
945bc96389 fix(utils.py): replicate now also has token based pricing for some models 2024-04-29 08:06:15 -07:00
Krish Dholakia
7ed64709aa Merge branch 'main' into litellm_common_auth_params 2024-04-28 08:38:06 -07:00
Krrish Dholakia
653e7fc584 fix(utils.py): don't return usage for streaming - openai spec 2024-04-27 14:13:34 -07:00
Ishaan Jaff
48d590c428 Merge pull request #3330 from BerriAI/litellm_rdct_msgs
[Feat] Redact Logging Messages/Response content on Logging Providers with `litellm.turn_off_message_logging=True`
2024-04-27 11:25:09 -07:00
Ishaan Jaff
c83a9a677f fix - sentry data redaction 2024-04-27 11:23:08 -07:00
Krrish Dholakia
ee68c0aaaa feat(utils.py): unify common auth params across azure/vertex_ai/bedrock/watsonx 2024-04-27 11:06:18 -07:00
Ishaan Jaff
37917b2bf7 fix use redact_message_input_output_from_logging 2024-04-27 10:51:17 -07:00
Ishaan Jaff
7191e5f593 feat- turn off message logging 2024-04-27 10:03:07 -07:00
Tejas Ravishankar
44f441ea74 fix: duplicate mention of VERTEXAI_PROJECT environment variable causing confusion 2024-04-27 17:47:28 +04:00
Krish Dholakia
b7beab2e39 Merge pull request #3270 from simonsanvil/feature/watsonx-integration
(feat) add IBM watsonx.ai as an llm provider
2024-04-27 05:48:34 -07:00
Krrish Dholakia
85a331024a test(test_streaming.py): fix test 2024-04-25 20:22:18 -07:00
Krrish Dholakia
ec548df1d6 fix(utils.py): handle pydantic v1 2024-04-25 20:01:36 -07:00
Krrish Dholakia
2c2e4319ab fix(utils.py): add more logging to identify ci/cd issue 2024-04-25 19:57:24 -07:00
Krish Dholakia
f9814105f7 Merge pull request #3310 from BerriAI/litellm_langfuse_error_logging_2
fix(proxy/utils.py): log rejected proxy requests to langfuse
2024-04-25 19:49:59 -07:00
Krrish Dholakia
462724feef fix(proxy/utils.py): log rejected proxy requests to langfuse 2024-04-25 19:26:27 -07:00
Krish Dholakia
33393f0029 Merge pull request #3308 from BerriAI/litellm_fix_streaming_n
fix(utils.py): fix the response object returned when n>1 for stream=true
2024-04-25 18:36:54 -07:00
Krrish Dholakia
60bb8e3919 fix(utils.py): pass through 'response_format' for mistral 2024-04-25 18:27:41 -07:00
Krrish Dholakia
baaa537445 fix(utils.py): handle finish reason logic 2024-04-25 18:18:00 -07:00
Krrish Dholakia
fc75fe2d05 fix(utils.py): return logprobs as an object not dict 2024-04-25 17:55:18 -07:00
Krrish Dholakia
157dd819f6 fix(utils.py): fix the response object returned when n>1 for stream=true
Fixes https://github.com/BerriAI/litellm/issues/3276
2024-04-25 13:27:29 -07:00
Krish Dholakia
265f16456f Merge pull request #3267 from BerriAI/litellm_openai_streaming_fix
fix(utils.py): fix streaming to not return usage dict
2024-04-24 21:08:33 -07:00
Krrish Dholakia
68f59e01ae fix(utils.py): fix anthropic streaming return usage tokens 2024-04-24 20:56:10 -07:00
Krrish Dholakia
5f862819e7 fix(utils.py): fix setattr error 2024-04-24 20:19:27 -07:00
Ishaan Jaff
96d12117c0 fix show api_base, model in timeout errors 2024-04-24 14:01:32 -07:00
Krish Dholakia
157099dd9e Merge pull request #3098 from greenscale-ai/main
Support for Greenscale AI logging
2024-04-24 13:09:03 -07:00
Krrish Dholakia
5abfb695f2 fix(vertex_ai.py): raise explicit error when image url fails to download - prevents silent failure 2024-04-24 09:23:15 -07:00
Krrish Dholakia
b10f03706d fix(utils.py): fix streaming to not return usage dict
Fixes https://github.com/BerriAI/litellm/issues/3237
2024-04-24 08:06:07 -07:00
Krrish Dholakia
85db0ace02 fix(utils.py): fix mistral api tool calling response 2024-04-23 19:59:11 -07:00
Krish Dholakia
52add6bf6a Merge pull request #3250 from BerriAI/litellm_caching_no_cache_fix
fix(utils.py): fix 'no-cache': true when caching is turned on
2024-04-23 19:57:07 -07:00
Krrish Dholakia
fea0e6bb19 fix(test_caching.py): add longer delay for async test 2024-04-23 16:13:03 -07:00
David Manouchehri
0075488f8b (utils.py) - Add seed for Groq 2024-04-23 20:32:21 +00:00
Krrish Dholakia
04014c752b fix(utils.py): fix 'no-cache': true when caching is turned on 2024-04-23 12:58:30 -07:00
Simon S. Viloria
79855b372d Merge branch 'BerriAI:main' into feature/watsonx-integration 2024-04-23 12:18:34 +02:00
Simon Sanchez Viloria
572cbef43b feat - watsonx refractoring, removed dependency, and added support for embedding calls 2024-04-23 12:01:13 +02:00
David Manouchehri
68bf14b2a5 (utils.py) - Fix response_format typo for Groq 2024-04-23 04:26:26 +00:00
Krrish Dholakia
011beb1918 fix(utils.py): support deepinfra response object 2024-04-22 10:51:11 -07:00
Simon S. Viloria
0c4cf91c79 Merge branch 'BerriAI:main' into feature/watsonx-integration 2024-04-21 10:35:51 +02:00
Krish Dholakia
70d59b1806 Merge pull request #3192 from BerriAI/litellm_calculate_max_parallel_requests
fix(router.py): Make TPM limits concurrency-safe
2024-04-20 13:24:29 -07:00
Krrish Dholakia
9f6e90e17d test(test_router_max_parallel_requests.py): more extensive testing for setting max parallel requests 2024-04-20 12:56:54 -07:00
Simon S. Viloria
0f5009e2e8 Merge branch 'BerriAI:main' into feature/watsonx-integration 2024-04-20 21:02:54 +02:00
Krrish Dholakia
b9042ba8ae fix(utils.py): map vertex ai exceptions - rate limit error 2024-04-20 11:12:05 -07:00
Simon Sanchez Viloria
9b3a1b3f35 Added support for IBM watsonx.ai models 2024-04-20 20:06:46 +02:00
Krrish Dholakia
22d3121f48 fix(router.py): calculate max_parallel_requests from given tpm limits
use the azure formula to calculate rpm -> max_parallel_requests based on a deployment's tpm limits
2024-04-20 10:43:18 -07:00
Ishaan Jaff
da23efe8ed fix - supports_vision should not raise Exception 2024-04-19 21:19:07 -07:00
Ishaan Jaff
fa887dbff2 fix - GetLLMProvider excepton error raise 2024-04-18 20:10:37 -07:00
David Manouchehri
e22f22e0a9 (feat) - Add seed to Cohere Chat. 2024-04-18 20:57:06 +00:00