Commit graph

752 commits

Author SHA1 Message Date
Krrish Dholakia
124aae7579 fix(factory.py): add mistral config to known tokenizer configs 2024-04-18 17:11:52 -07:00
David Manouchehri
e22f22e0a9 (feat) - Add seed to Cohere Chat. 2024-04-18 20:57:06 +00:00
Krrish Dholakia
77cb75bf2d fix(vertex_ai.py): fix faulty async completion function call tool check 2024-04-17 16:30:28 -07:00
Krrish Dholakia
72d7c36c76 refactor(utils.py): make it clearer how vertex ai params are handled '
'
2024-04-17 16:20:56 -07:00
Krrish Dholakia
6d508468ef fix(vertex_ai.py): accept credentials as a json string 2024-04-16 17:34:25 -07:00
Krish Dholakia
b46e9058df Merge pull request #3019 from ligaz/fix-anthropic-system-msg-handling
Fix Anthropic system message handling
2024-04-16 07:18:47 -07:00
Ishaan Jaff
2fbd800404 Merge pull request #3054 from BerriAI/litellm_supports_vision
[Feat] view models that `supports_vision`
2024-04-15 18:31:32 -07:00
Ishaan Jaff
1d96a54c34 fix - support base 64 image conversion for all gemini model 2024-04-15 18:18:55 -07:00
Krrish Dholakia
8c3c45fbb5 fix(vertex_ai_anthropic.py): set vertex_credentials for vertex ai anthropic calls
allows setting vertex credentials as a json string for vertex ai anthropic calls
2024-04-15 14:16:28 -07:00
Krrish Dholakia
3d645f95a5 fix(main.py): accept vertex service account credentials as json string
allows us to dynamically set vertex ai credentials
2024-04-15 13:28:59 -07:00
Krrish Dholakia
7df0626488 fix(anthropic_text.py): return process_response value 2024-04-15 08:51:26 -07:00
Krrish Dholakia
1cd0551a1e fix(anthropic_text.py): add support for async text completion calls 2024-04-15 08:15:00 -07:00
Stefan Dobrev
b3b52dbe32 Fix Anthropic system message handling 2024-04-14 16:23:59 +03:00
Krrish Dholakia
ec72202d56 fix(gemini.py): log system prompt in verbose output 2024-04-11 23:15:58 -07:00
Krrish Dholakia
7a829f805d fix(vertex_ai.py): check if 'response_mime_type' in generation config before passing it in 2024-04-11 23:10:59 -07:00
David Manouchehri
ad8d35a378 (feat) - Extreme dirty hack for response_mime_type in Vertex AI. 2024-04-12 00:03:31 +00:00
David Manouchehri
c6b866925f (feat) - Dirty hack to get response_mime_type working before it's released in the Python SDK. 2024-04-12 00:03:30 +00:00
David Manouchehri
cc71ca3166 (feat) - Add support for JSON mode in Vertex AI 2024-04-12 00:03:29 +00:00
Ishaan Jaff
db0e63532b Merge pull request #2892 from gregnwosu/patch-1
created defaults for response["eval_count"]
2024-04-11 13:40:32 -07:00
Ishaan Jaff
850241ada1 fix - stop using imghdr 2024-04-11 09:32:42 -07:00
Ishaan Jaff
9ee90a6bc3 fix cleanup 2024-04-10 21:28:04 -07:00
Ishaan Jaff
57fdfa0a5e fix use ImmutableSandboxedEnvironment 2024-04-10 21:26:50 -07:00
Ishaan Jaff
f2816b3fcd fix jinja2 use ImmutableSandboxedEnvironment 2024-04-10 20:30:13 -07:00
Ishaan Jaff
366a324735 fix - security report / vuln 2024-04-10 15:42:17 -07:00
Krrish Dholakia
76a298ec81 fix(replicate.py): handle empty response - replicate 2024-04-09 21:57:25 -07:00
Krrish Dholakia
57ea6a1c4f feat(gemini.py): support google-genai system instruction 2024-04-09 17:23:14 -07:00
Gregory Nwosu
0609da96e3 created defaults for response["eval_count"]
there is no way in litellm to disable the cache in ollama that is removing the eval_count response keys from the json.
This PR allows the code to create sensible defaults for when the response is empty
see 
- https://github.com/ollama/ollama/issues/1573
- https://github.com/ollama/ollama/issues/2023
2024-04-08 02:03:54 +01:00
Ishaan Jaff
494d71b063 add exit and aenter 2024-04-06 19:27:26 -07:00
Ishaan Jaff
833a64455f ci/cd run async handler 2024-04-06 19:16:27 -07:00
Ishaan Jaff
1dc5b01e01 fix - use anthropic class for clients 2024-04-06 18:19:28 -07:00
Ishaan Jaff
4dfe3f5aee async ahtropic streaming 2024-04-06 17:54:19 -07:00
Ishaan Jaff
3c10bfb497 async streaming anthropic 2024-04-06 17:53:06 -07:00
Ishaan Jaff
32c3aab34e feat - make anthropic async 2024-04-06 15:50:13 -07:00
Krish Dholakia
9912a80190 Merge pull request #2855 from Caixiaopig/fix_update_default_claude3_maxtokens
Updating the default Anthropic Officlal Claude 3 max_tokens to 4096
2024-04-06 08:39:55 -07:00
Ishaan Jaff
34e3cae0a5 fix - test using specific commit 2024-04-05 16:10:54 -07:00
Krish Dholakia
10951ca2f0 Merge pull request #2856 from lazyhope/anthropic-tools-use-2024-04-04
Support latest Anthropic Tools Use (2024-04-04)
2024-04-05 14:31:26 -07:00
Krrish Dholakia
00c0a0c209 fix(vertex_ai_anthropic.py): fix import 2024-04-05 13:38:05 -07:00
Zihao Li
715c11a9da Add backward compatibility to support xml tool use for bedrock and vertex 2024-04-06 00:34:33 +08:00
Krrish Dholakia
a8d0aace83 fix(vertex_ai_anthropic.py): support streaming, async completion, async streaming for vertex ai anthropic 2024-04-05 09:27:48 -07:00
lazyhope
a7eadf578a Merge branch 'BerriAI:main' into anthropic-tools-use-2024-04-04 2024-04-05 23:51:03 +08:00
Caixiaopig
aa2a1389da Updating the default Anthropic Officlal Claude 3 max_tokens to 4096
fix bug
2024-04-05 09:45:57 -05:00
Zihao Li
017abaa452 Clean up imports of XML processing functions 2024-04-05 22:36:18 +08:00
Krish Dholakia
5ea9946925 Merge pull request #2665 from BerriAI/litellm_claude_vertex_ai
[WIP] feat(vertex_ai_anthropic.py): Add support for claude 3 on vertex ai
2024-04-05 07:06:04 -07:00
Zihao Li
7a034d5942 Refactor tool result submission and tool invoke conversion 2024-04-05 17:11:35 +08:00
Zihao Li
4d0975bf4e Move tool definitions from system prompt to parameter and refactor tool calling parse 2024-04-05 16:01:40 +08:00
Caixiaopig
7eb9abba2b Updating the default Anthropic Claude 3 max_tokens to 4096
The default value of max_tokens used to be 256. If the client does not set a larger value, the model's output may be truncated, so the default value has been changed to 4096. This value is also the maximum output value described in the official interface.
see: https://docs.anthropic.com/claude/reference/messages_post
2024-04-05 14:44:40 +08:00
Krrish Dholakia
7880dcbdf7 fix(openai.py): support passing prompt as list instead of concat string 2024-04-03 15:23:20 -07:00
Krrish Dholakia
f3fcfe3007 fix(proxy_server.py): return original model response via response headers - /v1/completions
to help devs with debugging
2024-04-03 13:05:43 -07:00
Krrish Dholakia
bd7040969b feat(vertex_ai_anthropic.py): add claude 3 on vertex ai support - working .completions call
.completions() call works
2024-04-02 22:07:39 -07:00
yishiyiyuan
8619d29741 🐞 fix: djl vllm support
support vllm response format on sagemaker, which only return one choice.
2024-04-03 11:00:51 +08:00