Krish Dholakia
|
f9814105f7
|
Merge pull request #3310 from BerriAI/litellm_langfuse_error_logging_2
fix(proxy/utils.py): log rejected proxy requests to langfuse
|
2024-04-25 19:49:59 -07:00 |
|
Krrish Dholakia
|
462724feef
|
fix(proxy/utils.py): log rejected proxy requests to langfuse
|
2024-04-25 19:26:27 -07:00 |
|
Krish Dholakia
|
33393f0029
|
Merge pull request #3308 from BerriAI/litellm_fix_streaming_n
fix(utils.py): fix the response object returned when n>1 for stream=true
|
2024-04-25 18:36:54 -07:00 |
|
Krrish Dholakia
|
60bb8e3919
|
fix(utils.py): pass through 'response_format' for mistral
|
2024-04-25 18:27:41 -07:00 |
|
Krrish Dholakia
|
baaa537445
|
fix(utils.py): handle finish reason logic
|
2024-04-25 18:18:00 -07:00 |
|
Krrish Dholakia
|
fc75fe2d05
|
fix(utils.py): return logprobs as an object not dict
|
2024-04-25 17:55:18 -07:00 |
|
Krrish Dholakia
|
157dd819f6
|
fix(utils.py): fix the response object returned when n>1 for stream=true
Fixes https://github.com/BerriAI/litellm/issues/3276
|
2024-04-25 13:27:29 -07:00 |
|
Krish Dholakia
|
265f16456f
|
Merge pull request #3267 from BerriAI/litellm_openai_streaming_fix
fix(utils.py): fix streaming to not return usage dict
|
2024-04-24 21:08:33 -07:00 |
|
Krrish Dholakia
|
68f59e01ae
|
fix(utils.py): fix anthropic streaming return usage tokens
|
2024-04-24 20:56:10 -07:00 |
|
Krrish Dholakia
|
5f862819e7
|
fix(utils.py): fix setattr error
|
2024-04-24 20:19:27 -07:00 |
|
Ishaan Jaff
|
96d12117c0
|
fix show api_base, model in timeout errors
|
2024-04-24 14:01:32 -07:00 |
|
Krish Dholakia
|
157099dd9e
|
Merge pull request #3098 from greenscale-ai/main
Support for Greenscale AI logging
|
2024-04-24 13:09:03 -07:00 |
|
Krrish Dholakia
|
5abfb695f2
|
fix(vertex_ai.py): raise explicit error when image url fails to download - prevents silent failure
|
2024-04-24 09:23:15 -07:00 |
|
Krrish Dholakia
|
b10f03706d
|
fix(utils.py): fix streaming to not return usage dict
Fixes https://github.com/BerriAI/litellm/issues/3237
|
2024-04-24 08:06:07 -07:00 |
|
Krrish Dholakia
|
85db0ace02
|
fix(utils.py): fix mistral api tool calling response
|
2024-04-23 19:59:11 -07:00 |
|
Krish Dholakia
|
52add6bf6a
|
Merge pull request #3250 from BerriAI/litellm_caching_no_cache_fix
fix(utils.py): fix 'no-cache': true when caching is turned on
|
2024-04-23 19:57:07 -07:00 |
|
Krrish Dholakia
|
fea0e6bb19
|
fix(test_caching.py): add longer delay for async test
|
2024-04-23 16:13:03 -07:00 |
|
David Manouchehri
|
0075488f8b
|
(utils.py) - Add seed for Groq
|
2024-04-23 20:32:21 +00:00 |
|
Krrish Dholakia
|
04014c752b
|
fix(utils.py): fix 'no-cache': true when caching is turned on
|
2024-04-23 12:58:30 -07:00 |
|
Simon S. Viloria
|
79855b372d
|
Merge branch 'BerriAI:main' into feature/watsonx-integration
|
2024-04-23 12:18:34 +02:00 |
|
Simon Sanchez Viloria
|
572cbef43b
|
feat - watsonx refractoring, removed dependency, and added support for embedding calls
|
2024-04-23 12:01:13 +02:00 |
|
David Manouchehri
|
68bf14b2a5
|
(utils.py) - Fix response_format typo for Groq
|
2024-04-23 04:26:26 +00:00 |
|
Krrish Dholakia
|
011beb1918
|
fix(utils.py): support deepinfra response object
|
2024-04-22 10:51:11 -07:00 |
|
Simon S. Viloria
|
0c4cf91c79
|
Merge branch 'BerriAI:main' into feature/watsonx-integration
|
2024-04-21 10:35:51 +02:00 |
|
Krish Dholakia
|
70d59b1806
|
Merge pull request #3192 from BerriAI/litellm_calculate_max_parallel_requests
fix(router.py): Make TPM limits concurrency-safe
|
2024-04-20 13:24:29 -07:00 |
|
Krrish Dholakia
|
9f6e90e17d
|
test(test_router_max_parallel_requests.py): more extensive testing for setting max parallel requests
|
2024-04-20 12:56:54 -07:00 |
|
Simon S. Viloria
|
0f5009e2e8
|
Merge branch 'BerriAI:main' into feature/watsonx-integration
|
2024-04-20 21:02:54 +02:00 |
|
Krrish Dholakia
|
b9042ba8ae
|
fix(utils.py): map vertex ai exceptions - rate limit error
|
2024-04-20 11:12:05 -07:00 |
|
Simon Sanchez Viloria
|
9b3a1b3f35
|
Added support for IBM watsonx.ai models
|
2024-04-20 20:06:46 +02:00 |
|
Krrish Dholakia
|
22d3121f48
|
fix(router.py): calculate max_parallel_requests from given tpm limits
use the azure formula to calculate rpm -> max_parallel_requests based on a deployment's tpm limits
|
2024-04-20 10:43:18 -07:00 |
|
Ishaan Jaff
|
da23efe8ed
|
fix - supports_vision should not raise Exception
|
2024-04-19 21:19:07 -07:00 |
|
Ishaan Jaff
|
fa887dbff2
|
fix - GetLLMProvider excepton error raise
|
2024-04-18 20:10:37 -07:00 |
|
David Manouchehri
|
e22f22e0a9
|
(feat) - Add seed to Cohere Chat.
|
2024-04-18 20:57:06 +00:00 |
|
Ishaan Jaff
|
0f941678b4
|
Merge pull request #3130 from BerriAI/litellm_show_vertex_project_exceptions
[FIX] - show vertex_project, vertex_location in Vertex AI exceptions
|
2024-04-18 13:18:20 -07:00 |
|
Ishaan Jaff
|
177bc683b3
|
fix - track vertex_location and vertex_project in vertex exceptions
|
2024-04-18 12:53:33 -07:00 |
|
Krrish Dholakia
|
deccde6be1
|
fix(utils.py): support prometheus failed call metrics
|
2024-04-18 12:29:15 -07:00 |
|
Ishaan Jaff
|
2a18f5b8a9
|
fix - show _vertex_project, _vertex_location in exceptions
|
2024-04-18 11:48:43 -07:00 |
|
Nandesh Guru
|
9e46d3c0ac
|
Merge branch 'BerriAI:main' into main
|
2024-04-18 09:44:31 -07:00 |
|
Krish Dholakia
|
fe5c63e80b
|
Merge pull request #3105 from BerriAI/litellm_fix_hashing
fix(_types.py): hash api key in UserAPIKeyAuth
|
2024-04-18 08:16:24 -07:00 |
|
Krrish Dholakia
|
280d9b4405
|
fix(utils.py): function_setup empty message fix
fixes https://github.com/BerriAI/litellm/issues/2858
|
2024-04-18 07:32:29 -07:00 |
|
Krrish Dholakia
|
64fe5b146c
|
fix(utils.py): fix azure streaming logic
|
2024-04-18 07:08:36 -07:00 |
|
Krish Dholakia
|
49161e3ba4
|
Merge pull request #3102 from BerriAI/litellm_vertex_ai_fixes
fix(vertex_ai.py): fix faulty async call tool calling check
|
2024-04-17 19:16:36 -07:00 |
|
Krrish Dholakia
|
3e49a87f8b
|
fix(utils.py): exception mapping grpc none unknown error to api error
|
2024-04-17 19:12:40 -07:00 |
|
Krrish Dholakia
|
fdd73a4e26
|
fix(utils.py): support azure mistral function calling
|
2024-04-17 19:10:26 -07:00 |
|
Krrish Dholakia
|
caa46ca905
|
fix(utils.py): fix streaming special character flushing logic
|
2024-04-17 18:03:40 -07:00 |
|
Krrish Dholakia
|
1b4462ee70
|
fix(utils.py): ensure streaming output parsing only applied for hf / sagemaker models
selectively applies the <s>
</s> checking
|
2024-04-17 17:43:41 -07:00 |
|
Krrish Dholakia
|
2a2b97f093
|
fix(utils.py): accept {custom_llm_provider}/{model_name} in get_model_info
fixes https://github.com/BerriAI/litellm/issues/3100
|
2024-04-17 16:38:53 -07:00 |
|
Krrish Dholakia
|
72d7c36c76
|
refactor(utils.py): make it clearer how vertex ai params are handled '
'
|
2024-04-17 16:20:56 -07:00 |
|
greenscale-nandesh
|
86ac589bdd
|
Merge branch 'BerriAI:main' into main
|
2024-04-17 12:24:29 -07:00 |
|
Krish Dholakia
|
d55aada92a
|
Merge pull request #3062 from cwang/cwang/trim-messages-fix
Use `max_input_token` for `trim_messages`
|
2024-04-16 22:29:45 -07:00 |
|