Commit graph

1954 commits

Author SHA1 Message Date
Krish Dholakia
e7f83602d8
Merge branch 'main' into litellm_response_cost_logging 2024-08-01 21:28:22 -07:00
Krrish Dholakia
010d5ed81d feat(vertex_ai_partner.py): add vertex ai codestral FIM support
Closes https://github.com/BerriAI/litellm/issues/4984
2024-08-01 17:10:27 -07:00
Krrish Dholakia
28c12e6702 fix(utils.py): fix togetherai streaming cost calculation 2024-08-01 15:03:08 -07:00
Krrish Dholakia
ca0a0bed46 fix(utils.py): fix anthropic streaming usage calculation
Fixes https://github.com/BerriAI/litellm/issues/4965
2024-08-01 14:45:54 -07:00
Krrish Dholakia
10b571ca42 fix(litellm_logging.py): use 1 cost calc function across response headers + logging integrations
Ensures consistent cost calculation when azure base models are used
2024-08-01 10:26:59 -07:00
Krrish Dholakia
1a00f52255 fix(utils.py): fix special keys list for provider-specific items in response object 2024-07-31 18:30:49 -07:00
Krrish Dholakia
09ee8c6e2d fix(utils.py): return additional kwargs from openai-like response body
Closes https://github.com/BerriAI/litellm/issues/4981
2024-07-31 15:37:03 -07:00
Krrish Dholakia
bd68714f51 fix(utils.py): map cohere timeout error 2024-07-31 15:15:18 -07:00
Krrish Dholakia
c65a438de2 fix(utils.py): fix linting errors 2024-07-30 18:38:10 -07:00
Krrish Dholakia
46634af06f fix(utils.py): fix model registeration to model cost map
Fixes https://github.com/BerriAI/litellm/issues/4972
2024-07-30 18:15:00 -07:00
Krrish Dholakia
185a6857f9 fix(utils.py): fix cost tracking for vertex ai partner models 2024-07-30 14:20:52 -07:00
Krish Dholakia
6bdb9203b3
Merge pull request #4948 from dleen/response
fixes: #4947 Bedrock context exception does not have a response
2024-07-29 15:03:40 -07:00
David Leen
ebca13b2ca fixes: #4947 Bedrock context exception does not have a response 2024-07-29 14:23:56 -07:00
Krrish Dholakia
ae4bcd8a41 fix(utils.py): fix trim_messages to handle tool calling
Fixes https://github.com/BerriAI/litellm/issues/4931
2024-07-29 13:04:41 -07:00
Krrish Dholakia
59384c84a5 fix(utils.py): correctly re-raise azure api connection error
'
2024-07-29 12:28:25 -07:00
Krrish Dholakia
52b1ca292c fix(utils.py): check if tools is iterable before indexing into it
Fixes https://github.com/BerriAI/litellm/issues/4933
2024-07-29 09:01:32 -07:00
Ravi N
b321f2988b Allow zero temperature for Sagemaker models based on config
Since Sagemaker can host any kind of model, some models allow
zero temperature. However, this is not enabled by default and
only allowed based on config
2024-07-28 21:55:53 -04:00
Krrish Dholakia
1ac42d8464 fix: utils.py
fix supported openai params
2024-07-27 22:03:40 -07:00
Krish Dholakia
e3a94ac013
Merge pull request #4925 from BerriAI/litellm_vertex_mistral
feat(vertex_ai_partner.py): Vertex AI Mistral Support
2024-07-27 21:51:26 -07:00
Krish Dholakia
b854d2100c
Merge branch 'main' into litellm_vertex_migration 2024-07-27 20:25:12 -07:00
Krrish Dholakia
d1989b6063 fix(utils.py): support fireworks ai finetuned models
Fixes https://github.com/BerriAI/litellm/issues/4923
2024-07-27 15:38:27 -07:00
Krrish Dholakia
f76cad210c fix(utils.py): support fireworks ai finetuned models
Fixes https://github.com/BerriAI/litellm/issues/4923
2024-07-27 15:37:28 -07:00
Krrish Dholakia
c85ed01756 feat(utils.py): fix openai-like streaming 2024-07-27 15:32:57 -07:00
Krrish Dholakia
05ba34b9b7 fix(utils.py): add exception mapping for databricks errors 2024-07-27 13:13:31 -07:00
Krrish Dholakia
5b71421a7b feat(vertex_ai_partner.py): initial working commit for calling vertex ai mistral
Closes https://github.com/BerriAI/litellm/issues/4874
2024-07-27 12:54:14 -07:00
Krrish Dholakia
b25d4a8cb3 feat(ollama_chat.py): support ollama tool calling
Closes https://github.com/BerriAI/litellm/issues/4812
2024-07-26 21:51:54 -07:00
Krrish Dholakia
fe0b55f2ca fix(utils.py): fix cache hits for streaming
Fixes https://github.com/BerriAI/litellm/issues/4109
2024-07-26 19:04:08 -07:00
Krrish Dholakia
2f773d9cb6 fix(litellm_cost_calc/google.py): support meta llama vertex ai cost tracking 2024-07-25 22:12:07 -07:00
Ishaan Jaff
079a41fbe1
Merge branch 'main' into litellm_proxy_support_all_providers 2024-07-25 20:15:37 -07:00
Krrish Dholakia
a2fd8459fc fix(utils.py): don't raise error on openai content filter during streaming - return as is
Fixes issue where we would raise an error vs. openai who return the chunk with finish reason as 'content_filter'
2024-07-25 19:50:52 -07:00
Krish Dholakia
b6ca4406b6
Merge branch 'main' into bedrock-llama3.1-405b 2024-07-25 19:29:10 -07:00
Ishaan Jaff
9863520376 support using */* 2024-07-25 18:48:56 -07:00
Krrish Dholakia
060249c7e0 feat(utils.py): support async streaming for custom llm provider 2024-07-25 17:11:57 -07:00
Krrish Dholakia
b4e3a77ad0 feat(utils.py): support sync streaming for custom llm provider 2024-07-25 16:47:32 -07:00
Krrish Dholakia
6bf1b9353b feat(custom_llm.py): initial working commit for writing your own custom LLM handler
Fixes https://github.com/BerriAI/litellm/issues/4675

 Also Addresses https://github.com/BerriAI/litellm/discussions/4677
2024-07-25 15:33:05 -07:00
David Manouchehri
64adae6e7f
Check for converse support first. 2024-07-25 21:16:23 +00:00
Krrish Dholakia
4e51f712f3 fix(main.py): fix calling openai gpt-3.5-turbo-instruct via /completions
Fixes https://github.com/BerriAI/litellm/issues/749
2024-07-25 09:57:19 -07:00
wslee
40bb165108 support dynamic api base 2024-07-25 11:14:38 +09:00
wslee
dd10da4d46 add support for friendli dedicated endpoint 2024-07-25 11:14:35 +09:00
Ishaan Jaff
8ea4b73c27 add UnsupportedParamsError to litellm exceptions 2024-07-24 12:20:14 -07:00
Krrish Dholakia
7df94100e8 build(model_prices_and_context_window.json): add model pricing for vertex ai llama 3.1 api 2024-07-23 17:36:07 -07:00
Krrish Dholakia
83ef52e180 feat(vertex_ai_llama.py): vertex ai llama3.1 api support
Initial working commit for vertex ai llama 3.1 api support
2024-07-23 17:07:30 -07:00
Krrish Dholakia
f64a3309d1 fix(utils.py): support raw response headers for streaming requests 2024-07-23 11:58:58 -07:00
Krrish Dholakia
dcb974dd1e feat(utils.py): support passing openai response headers to client, if enabled
Allows openai/openai-compatible provider response headers to be sent to client, if 'return_response_headers' is enabled
2024-07-23 11:30:52 -07:00
Ishaan Jaff
1355932bf4
Merge pull request #3905 from giritatavarty-8451/litellm_triton_chatcompletion_support
Litellm triton chatcompletion support - Resubmit of #3895
2024-07-23 10:30:26 -07:00
Ishaan Jaff
8f9638f2c1 fix raise correct provider on content policy violation 2024-07-22 16:03:15 -07:00
Ishaan Jaff
15c109f023 fix checking if _known_custom_logger_compatible_callbacks 2024-07-22 15:43:43 -07:00
Krrish Dholakia
f10af7596c fix(utils.py): allow dropping extra_body in additional_drop_params
Fixes https://github.com/BerriAI/litellm/issues/4769
2024-07-20 19:12:58 -07:00
Ishaan Jaff
5e4d291244 rename to _response_headers 2024-07-20 17:31:16 -07:00
Ishaan Jaff
46cf4f69ae return response headers in response 2024-07-20 14:59:08 -07:00