Commit graph

1301 commits

Author SHA1 Message Date
Ishaan Jaff
72fddabf84 Merge pull request #2868 from BerriAI/litellm_add_command_r_on_proxy
Add Azure Command-r-plus on litellm proxy
2024-04-05 15:13:47 -07:00
Ishaan Jaff
f65828db26 Merge pull request #2861 from BerriAI/litellm_add_azure_command_r_plust
[FEAT] add azure command-r-plus
2024-04-05 15:13:35 -07:00
Ishaan Jaff
5c1a662caa proxy - add azure/command r 2024-04-05 14:35:31 -07:00
Krish Dholakia
10951ca2f0 Merge pull request #2856 from lazyhope/anthropic-tools-use-2024-04-04
Support latest Anthropic Tools Use (2024-04-04)
2024-04-05 14:31:26 -07:00
Ishaan Jaff
3ff3947fc0 fix use azure_ai/mistral 2024-04-05 10:07:43 -07:00
Ishaan Jaff
c74d43fcbc fix support azure/mistral models 2024-04-05 09:32:39 -07:00
Krrish Dholakia
a8d0aace83 fix(vertex_ai_anthropic.py): support streaming, async completion, async streaming for vertex ai anthropic 2024-04-05 09:27:48 -07:00
Ishaan Jaff
296f10e2c9 fix add azure/command-r-plus 2024-04-05 08:53:24 -07:00
Zihao Li
4d0975bf4e Move tool definitions from system prompt to parameter and refactor tool calling parse 2024-04-05 16:01:40 +08:00
Nandesh Guru
39360f1605 Greenscale Integration
Adding logger for Greenscale
2024-04-04 15:38:51 -07:00
Ishaan Jaff
d0977d61ff v0 return cache key in responses 2024-04-04 10:11:18 -07:00
Krrish Dholakia
f3fcfe3007 fix(proxy_server.py): return original model response via response headers - /v1/completions
to help devs with debugging
2024-04-03 13:05:43 -07:00
Krrish Dholakia
de3b314620 fix(openai.py): switch to using openai sdk for text completion calls 2024-04-02 15:08:12 -07:00
Krrish Dholakia
ed46af19ec fix(openai.py): return logprobs for text completion calls 2024-04-02 14:05:56 -07:00
Krrish Dholakia
67f62aa53e fix(main.py): support text completion input being a list of strings
addresses - https://github.com/BerriAI/litellm/issues/2792, https://github.com/BerriAI/litellm/issues/2777
2024-04-02 08:50:16 -07:00
Sebastián Estévez
a4c65e6e61 support cohere_chat in get_api_key 2024-04-01 13:24:03 -04:00
Ishaan Jaff
90038c1c86 check num retries in async wrapper 2024-03-30 19:33:40 -07:00
Ishaan Jaff
5cc2086dc1 (fix) improve async perf 2024-03-30 19:07:04 -07:00
Vincelwt
35c279eff8 Merge branch 'main' into main 2024-03-30 13:21:53 +09:00
Ishaan Jaff
41f9622f29 (docs) grafana / prometheus 2024-03-29 14:25:45 -07:00
Krrish Dholakia
ae3510c0d2 fix(utils.py): exception mapping on 'next()' streaming error 2024-03-29 09:18:41 -07:00
Krrish Dholakia
006c20a10f fix(utils.py): support bedrock mistral streaming 2024-03-29 07:56:10 -07:00
Krrish Dholakia
d98e040f36 fix(utils.py): don't add chatml tokens to a simple text token count 2024-03-28 13:48:48 -07:00
Krrish Dholakia
7fe02405e0 fix(proxy/utils.py): check cache before alerting user 2024-03-27 20:09:15 -07:00
Krrish Dholakia
62e557404c fix(utils.py): don't run post-call rules on a coroutine function 2024-03-27 13:16:27 -07:00
Krish Dholakia
d51f12ca44 Merge pull request #2704 from BerriAI/litellm_jwt_auth_improvements_3
fix(handle_jwt.py): enable team-based jwt-auth access
2024-03-26 16:06:56 -07:00
Krrish Dholakia
4028f935a5 fix(utils.py): check if item in list is pydantic object or dict before dereferencing 2024-03-26 14:39:16 -07:00
Krrish Dholakia
c21e954c6f test(test_batch_completions.py): handle anthropic overloaded error 2024-03-26 13:55:03 -07:00
Krrish Dholakia
05fddcb06b fix(utils.py): don't send subsequent chunks if last chunk sent
prevents multiple empty finish chunks from being sent
2024-03-26 13:49:42 -07:00
Ishaan Jaff
f5f7e344c7 Merge branch 'main' into litellm_remove_litellm_telemetry 2024-03-26 11:35:02 -07:00
Ishaan Jaff
62c83d36a5 (fix) remove litellm.telemetry 2024-03-26 11:21:09 -07:00
Krrish Dholakia
44bc79aadb fix(utils.py): check if message is pydantic object or dict before dereferencing 2024-03-26 09:47:44 -07:00
Krrish Dholakia
1137264d99 test(test_streaming.py): add unit testing for custom stream wrapper 2024-03-26 08:57:44 -07:00
Krish Dholakia
3fef983439 Merge pull request #2656 from TashaSkyUp/patch-1
fix for: when using ModelResponse.json() to save and then reconstruct a ModelResponse the choices field ends up empty
2024-03-26 08:36:55 -07:00
Ishaan Jaff
ebdec4d262 (fix) cache control logic 2024-03-26 07:36:45 -07:00
Ishaan Jaff
7bf9cb3c54 (fix) cache control logic 2024-03-25 22:19:34 -07:00
Krrish Dholakia
f604a6155f fix(utils.py): persist system fingerprint across chunks 2024-03-25 19:24:09 -07:00
Krrish Dholakia
c5bd4d4233 fix(utils.py): log success event for streaming 2024-03-25 19:03:10 -07:00
Krrish Dholakia
cbf4c95e5f fix(utils.py): persist response id across chunks 2024-03-25 18:20:43 -07:00
Krrish Dholakia
ecc0cf5d9c fix(utils.py): fix text completion streaming 2024-03-25 16:47:17 -07:00
Krrish Dholakia
26dbb76d53 fix(utils.py): ensure last chunk is always empty delta w/ finish reason
makes sure we're openai-compatible with our streaming. Adds stricter tests for this as well
2024-03-25 16:33:41 -07:00
Krrish Dholakia
c667e437b9 fix(utils.py): allow user to disable streaming logging
fixes event loop issue for litellm.disable_streaming_logging
2024-03-25 14:28:46 -07:00
Max Deichmann
efa599b0ee push 2024-03-25 17:43:55 +01:00
Krrish Dholakia
4e70a3e09a feat(router.py): enable pre-call checks
filter models outside of context window limits of a given message for a model group

 https://github.com/BerriAI/litellm/issues/872
2024-03-23 18:03:30 -07:00
Tasha Upchurch
8814524473 Update utils.py
fix for constructed from dict choices.message being a dict still instead of Message class.
2024-03-23 00:12:24 -04:00
Ishaan Jaff
07067db5a1 (feat) remove litellm.telemetry 2024-03-22 20:58:14 -07:00
Tasha Upchurch
2c1fb7e881 Update utils.py
Fix for creating an empty choices if no choices passed in
2024-03-22 23:39:17 -04:00
Tasha Upchurch
541155c08d Update utils.py
fix for #2655
2024-03-22 23:13:24 -04:00
Krrish Dholakia
4dad400b57 fix(anthropic.py): handle multiple system prompts 2024-03-22 18:14:15 -07:00
Vincelwt
860f1b982d Merge branch 'main' into main 2024-03-22 00:52:42 +09:00