Commit graph

1344 commits

Author SHA1 Message Date
Krrish Dholakia
b38c09c87f fix(utils.py): fix azure streaming logic 2024-04-18 07:08:36 -07:00
Krish Dholakia
bcdf24e5aa
Merge pull request #3102 from BerriAI/litellm_vertex_ai_fixes
fix(vertex_ai.py): fix faulty async call tool calling check
2024-04-17 19:16:36 -07:00
Krrish Dholakia
a862201a84 fix(utils.py): exception mapping grpc none unknown error to api error 2024-04-17 19:12:40 -07:00
Krrish Dholakia
18e3cf8bff fix(utils.py): support azure mistral function calling 2024-04-17 19:10:26 -07:00
Krrish Dholakia
15ae7a8314 fix(utils.py): fix streaming special character flushing logic 2024-04-17 18:03:40 -07:00
Krrish Dholakia
7d0086d742 fix(utils.py): ensure streaming output parsing only applied for hf / sagemaker models
selectively applies the <s>
</s> checking
2024-04-17 17:43:41 -07:00
Krrish Dholakia
53df916f69 fix(utils.py): accept {custom_llm_provider}/{model_name} in get_model_info
fixes https://github.com/BerriAI/litellm/issues/3100
2024-04-17 16:38:53 -07:00
Krrish Dholakia
32d94feddd refactor(utils.py): make it clearer how vertex ai params are handled '
'
2024-04-17 16:20:56 -07:00
greenscale-nandesh
907e3973fd
Merge branch 'BerriAI:main' into main 2024-04-17 12:24:29 -07:00
Krish Dholakia
8febe2f573
Merge pull request #3062 from cwang/cwang/trim-messages-fix
Use `max_input_token` for `trim_messages`
2024-04-16 22:29:45 -07:00
Ishaan Jaff
9e9d55228e fix - show model, deployment, model group in vertex error 2024-04-16 19:59:34 -07:00
Krrish Dholakia
4d0d6127d8 fix(utils.py): fix get_api_base 2024-04-16 18:50:27 -07:00
greenscale-nandesh
3feb0ef897
Merge branch 'BerriAI:main' into main 2024-04-16 11:49:26 -07:00
Chen Wang
38c61a23b4
Fall back to max_tokens 2024-04-16 19:00:09 +01:00
Chen Wang
ebc889d77a
Use max_input_token for trim_messages 2024-04-16 13:36:25 +01:00
Ishaan Jaff
fb8e256aba feat - new util supports_vision 2024-04-15 18:10:12 -07:00
Krrish Dholakia
0683589029 fix(utils.py): fix timeout error - don't pass in httpx.request 2024-04-15 10:50:23 -07:00
Krish Dholakia
72b54eaad7
Merge pull request #3028 from BerriAI/litellm_anthropic_text_completion_fix
fix(anthropic_text.py): add support for async text completion calls
2024-04-15 09:26:28 -07:00
Krrish Dholakia
26286a54b8 fix(anthropic_text.py): add support for async text completion calls 2024-04-15 08:15:00 -07:00
Ishaan Jaff
8eeae10bf2 groq - add tool calling support 2024-04-15 08:09:27 -07:00
Krrish Dholakia
4e81acf2c6 feat(prometheus_services.py): monitor health of proxy adjacent services (redis / postgres / etc.) 2024-04-13 18:15:02 -07:00
Ishaan Jaff
0540ca4918
Merge pull request #2991 from BerriAI/litellm_fix_text_completion_caching
[Feat] Support + Test caching for TextCompletion
2024-04-12 20:08:01 -07:00
Ishaan Jaff
0cef782ffa fix - support text completion caching 2024-04-12 12:34:28 -07:00
Krish Dholakia
e01e9fd764
Merge pull request #2984 from Dev-Khant/slack-msg-truncation
truncate long slack msg
2024-04-12 08:30:08 -07:00
Dev Khant
5cdcc5895c truncate long slack msg 2024-04-12 17:22:14 +05:30
Krrish Dholakia
7a3821e0f6 fix(gemini.py): log system prompt in verbose output 2024-04-11 23:15:58 -07:00
Krrish Dholakia
c377ba0755 fix(utils.py): vertex ai exception mapping
fixes check which caused all vertex errors to be ratelimit errors
2024-04-11 23:04:21 -07:00
David Manouchehri
649c3bb0dd
(feat) - Add support for JSON mode in Vertex AI 2024-04-12 00:03:29 +00:00
Krish Dholakia
83e7ed94ce
Merge pull request #2942 from BerriAI/litellm_fix_router_loading
Router Async Improvements
2024-04-10 20:16:53 -07:00
Krrish Dholakia
37ac17aebd fix(router.py): fix datetime object 2024-04-10 17:55:24 -07:00
Ishaan Jaff
0c26404cb3 fix - allow base64 cache hits embedding responses 2024-04-10 16:44:40 -07:00
Krrish Dholakia
b2741933dc fix(proxy_cli.py): don't double load the router config
was causing callbacks to be instantiated twice - double couting usage in cache
2024-04-10 13:23:56 -07:00
Ishaan Jaff
caec5374d9
Merge pull request #2893 from unclecode/main
Fix issue #2832: Add protected_namespaces to Config class within utils.py, router.py and completion.py to avoid the warning message.
2024-04-09 08:51:41 -07:00
Krrish Dholakia
48bfc45cb0 fix(utils.py): fix reordering of items for cached embeddings
ensures cached embedding item is returned in correct order
2024-04-08 12:18:24 -07:00
unclecode
1ace192155 Fix issue #2832: Add protected_namespaces to Config class within utils.py, router.py and completion.py to avoid the warning message. 2024-04-08 12:43:17 +08:00
Ishaan Jaff
a5aef6ec00
Merge pull request #2879 from BerriAI/litellm_async_anthropic_api
[Feat] Async Anthropic API 97.5% lower median latency
2024-04-07 09:56:52 -07:00
Krrish Dholakia
460546956d fix(utils.py): fix import 2024-04-06 18:37:38 -07:00
Krrish Dholakia
a410981972 fix(utils.py): fix circular import 2024-04-06 18:29:51 -07:00
Ishaan Jaff
7849c29f70 async anthropic streaming 2024-04-06 17:36:56 -07:00
Krrish Dholakia
ead59b7b38 fix(utils.py): add gemini api base support to 'get_api_base' 2024-04-06 16:08:15 -07:00
Krrish Dholakia
6110d32b1c feat(proxy/utils.py): return api base for request hanging alerts 2024-04-06 15:58:53 -07:00
Krrish Dholakia
7b30e5ae38 fix(utils.py): fix content check in pre-call rules 2024-04-06 09:03:19 -07:00
Krrish Dholakia
30f57e7aa5 fix(utils.py): move info statement to debug 2024-04-05 22:06:46 -07:00
Ishaan Jaff
faa0d38087
Merge pull request #2868 from BerriAI/litellm_add_command_r_on_proxy
Add Azure Command-r-plus on litellm proxy
2024-04-05 15:13:47 -07:00
Ishaan Jaff
2174b240d8
Merge pull request #2861 from BerriAI/litellm_add_azure_command_r_plust
[FEAT] add azure command-r-plus
2024-04-05 15:13:35 -07:00
Ishaan Jaff
9055a071e6 proxy - add azure/command r 2024-04-05 14:35:31 -07:00
Krish Dholakia
a50edef1e6
Merge pull request #2856 from lazyhope/anthropic-tools-use-2024-04-04
Support latest Anthropic Tools Use (2024-04-04)
2024-04-05 14:31:26 -07:00
Ishaan Jaff
6b9c04618e fix use azure_ai/mistral 2024-04-05 10:07:43 -07:00
Ishaan Jaff
5ce80d82d3 fix support azure/mistral models 2024-04-05 09:32:39 -07:00
Krrish Dholakia
f0c4ff6e60 fix(vertex_ai_anthropic.py): support streaming, async completion, async streaming for vertex ai anthropic 2024-04-05 09:27:48 -07:00