Commit graph

1140 commits

Author SHA1 Message Date
Krrish Dholakia
8500f6d087 feat(caching.py): enable caching on provider-specific optional params
Closes https://github.com/BerriAI/litellm/issues/5049
2024-08-05 11:18:59 -07:00
Krish Dholakia
88a2428ee6 Merge branch 'main' into litellm_anthropic_api_streaming 2024-08-03 21:16:50 -07:00
Krrish Dholakia
5810708c71 feat(anthropic_adapter.py): support streaming requests for /v1/messages endpoint
Fixes https://github.com/BerriAI/litellm/issues/5011
2024-08-03 20:16:19 -07:00
Joe Cheng
fa9298d588 Use correct key name 2024-08-03 11:58:46 -07:00
Joe Cheng
b4df896ac9 Fix tool call coalescing
The previous code seemed to assume that the tool call index property
started at 0, but Anthropic sometimes returns them starting at 1.
This was causing an extra null-ish tool call to be materialized.
2024-08-02 13:05:23 -07:00
Joe Cheng
0f301a120f fix(main.py): Handle bedrock tool calling in stream_chunk_builder
Fixes #5022.

The streaming chunks from Anthropic seem to violate an assumption
that is implicit in the stream_chunk_builder implementation: that
only tool_calls OR function_calls OR content will appear in a
streamed response. The repro in #5022 shows that you can get
content followed by tool calls.

These changes properly handle these combinations by building
separate lists of each type of chunk (note that in theory a chunk
could appear in multiple lists, e.g. both delta.tool_calls and
delta.content being present on one chunk).
2024-08-02 12:41:13 -07:00
Krish Dholakia
0fc50a69ee Merge branch 'main' into litellm_fix_streaming_usage_calc 2024-08-01 21:29:04 -07:00
Krrish Dholakia
a072edaf0b fix(vertex_ai_partner.py): add /chat/completion codestral support
Closes https://github.com/BerriAI/litellm/issues/4984
2024-08-01 18:06:40 -07:00
Krrish Dholakia
4972c993a2 fix(main.py): fix linting error 2024-08-01 17:33:29 -07:00
Krrish Dholakia
cb9b19e887 feat(vertex_ai_partner.py): add vertex ai codestral FIM support
Closes https://github.com/BerriAI/litellm/issues/4984
2024-08-01 17:10:27 -07:00
Krrish Dholakia
2121738137 fix(vertex_ai_partner.py): add /chat/completion codestral support
Closes https://github.com/BerriAI/litellm/issues/4984
2024-08-01 16:12:05 -07:00
Krrish Dholakia
a502914f13 fix(utils.py): fix anthropic streaming usage calculation
Fixes https://github.com/BerriAI/litellm/issues/4965
2024-08-01 14:45:54 -07:00
Krish Dholakia
ee375edcbf Merge branch 'main' into litellm_async_cohere_calls 2024-07-30 15:35:20 -07:00
Krrish Dholakia
3a92468765 fix(cohere.py): support async cohere embedding calls 2024-07-30 14:49:07 -07:00
Krrish Dholakia
b29d2fcd7c fix(main.py): fix linting error 2024-07-30 13:55:04 -07:00
Krrish Dholakia
17ac0f0636 feat(huggingface_restapi.py): Support multiple hf embedding types + async hf embeddings
Closes https://github.com/BerriAI/litellm/issues/3261
2024-07-30 13:32:03 -07:00
Krrish Dholakia
e3c02a1754 docs(main.py): update acompletion_with_retries docstring
Closes https://github.com/BerriAI/litellm/issues/4908
2024-07-29 15:50:12 -07:00
Krish Dholakia
166a07a869 Merge pull request #4943 from dleen/logs
Fix: #4942. Remove verbose logging when exception can be handled
2024-07-29 12:12:28 -07:00
David Leen
394ff28033 Fix: #4942. Remove verbose logging when exception can be handled 2024-07-29 12:05:10 -07:00
Krrish Dholakia
7de70a19e4 fix(exceptions.py): use correct status code for content policy exceptions
Fixes https://github.com/BerriAI/litellm/issues/4941#issuecomment-2256578732
2024-07-29 12:01:54 -07:00
Krish Dholakia
1c50339580 Merge pull request #4925 from BerriAI/litellm_vertex_mistral
feat(vertex_ai_partner.py): Vertex AI Mistral Support
2024-07-27 21:51:26 -07:00
Krish Dholakia
0525fb75f3 Merge branch 'main' into litellm_vertex_migration 2024-07-27 20:25:12 -07:00
Ishaan Jaff
b6b8bb4f02 fix checking mode on health checks 2024-07-27 20:21:39 -07:00
Krrish Dholakia
56ba0c62f3 feat(utils.py): fix openai-like streaming 2024-07-27 15:32:57 -07:00
Krrish Dholakia
ce7257ec5e feat(vertex_ai_partner.py): initial working commit for calling vertex ai mistral
Closes https://github.com/BerriAI/litellm/issues/4874
2024-07-27 12:54:14 -07:00
Krrish Dholakia
a2de16582a fix(custom_llm.py): pass input params to custom llm 2024-07-25 19:03:52 -07:00
Krrish Dholakia
bf23aac11d feat(utils.py): support sync streaming for custom llm provider 2024-07-25 16:47:32 -07:00
Krrish Dholakia
fe503386ab fix(custom_llm.py): support async completion calls 2024-07-25 15:51:39 -07:00
Krrish Dholakia
54e1ca29b7 feat(custom_llm.py): initial working commit for writing your own custom LLM handler
Fixes https://github.com/BerriAI/litellm/issues/4675

 Also Addresses https://github.com/BerriAI/litellm/discussions/4677
2024-07-25 15:33:05 -07:00
Krrish Dholakia
5945da4a66 fix(main.py): fix calling openai gpt-3.5-turbo-instruct via /completions
Fixes https://github.com/BerriAI/litellm/issues/749
2024-07-25 09:57:19 -07:00
Krrish Dholakia
778afcee31 feat(vertex_ai_llama.py): vertex ai llama3.1 api support
Initial working commit for vertex ai llama 3.1 api support
2024-07-23 17:07:30 -07:00
Ishaan Jaff
71c755d9a2 Merge pull request #3905 from giritatavarty-8451/litellm_triton_chatcompletion_support
Litellm triton chatcompletion support - Resubmit of #3895
2024-07-23 10:30:26 -07:00
Krrish Dholakia
dbaaf9f1a2 fix(main.py): check if anthropic api base ends with required url
Fixes https://github.com/BerriAI/litellm/issues/4803
2024-07-22 20:36:35 -07:00
Krrish Dholakia
a01a45230c fix(openai.py): support completion, streaming, async_streaming 2024-07-20 15:23:42 -07:00
Krrish Dholakia
a29be08397 fix(main.py): check for ANTHROPIC_BASE_URL in environment
Fixes https://github.com/BerriAI/litellm/issues/4803
2024-07-20 14:38:31 -07:00
Ishaan Jaff
82ceee32c4 fix health check 2024-07-19 15:56:35 -07:00
Sophia Loris
adae0777d6 resolve merge conflicts 2024-07-19 09:45:53 -05:00
Sophia Loris
91fa69c0c2 Add support for Triton streaming & triton async completions 2024-07-19 09:35:27 -05:00
Ishaan Jaff
17502de9ca Merge pull request #4789 from BerriAI/litellm_router_refactor
[Feat-Router] - Tag based routing
2024-07-18 22:19:18 -07:00
Ishaan Jaff
8d227680c7 fix use tags as a litellm param 2024-07-18 19:34:45 -07:00
Krrish Dholakia
aac912d3f8 feat(vertex_ai_anthropic.py): support response_schema for vertex ai anthropic calls
allows passing response_schema for anthropic calls. supports schema validation.
2024-07-18 16:57:38 -07:00
Ishaan Jaff
52370fee12 fix testing exception mapping 2024-07-13 11:10:13 -07:00
Krrish Dholakia
948fd6fc33 fix: fix linting errors 2024-07-11 13:36:55 -07:00
Krrish Dholakia
4bd7276732 fix(main.py): fix linting errors 2024-07-11 12:11:50 -07:00
Krrish Dholakia
48be4ce805 feat(proxy_server.py): working /v1/messages with config.yaml
Adds async router support for adapter_completion call
2024-07-10 18:53:54 -07:00
Krrish Dholakia
4ba30abb63 feat(proxy_server.py): working /v1/messages endpoint
Works with claude engineer
2024-07-10 18:15:38 -07:00
Krrish Dholakia
01a335b4c3 feat(anthropic_adapter.py): support for translating anthropic params to openai format 2024-07-10 00:32:28 -07:00
Krrish Dholakia
789d2dab15 fix(vertex_httpx.py): add sync vertex image gen support
Fixes https://github.com/BerriAI/litellm/issues/4623
2024-07-09 13:33:54 -07:00
Ishaan Jaff
24f718438e fix params on acompletion 2024-07-08 12:56:54 -07:00
Krrish Dholakia
d68ab2a8bc fix(whisper---handle-openai/azure-vtt-response-format): Fixes https://github.com/BerriAI/litellm/issues/4595 2024-07-08 09:10:40 -07:00