Krrish Dholakia
bd68714f51
fix(utils.py): map cohere timeout error
2024-07-31 15:15:18 -07:00
Krrish Dholakia
c65a438de2
fix(utils.py): fix linting errors
2024-07-30 18:38:10 -07:00
Krrish Dholakia
46634af06f
fix(utils.py): fix model registeration to model cost map
...
Fixes https://github.com/BerriAI/litellm/issues/4972
2024-07-30 18:15:00 -07:00
Krrish Dholakia
185a6857f9
fix(utils.py): fix cost tracking for vertex ai partner models
2024-07-30 14:20:52 -07:00
Krish Dholakia
6bdb9203b3
Merge pull request #4948 from dleen/response
...
fixes : #4947 Bedrock context exception does not have a response
2024-07-29 15:03:40 -07:00
David Leen
ebca13b2ca
fixes : #4947 Bedrock context exception does not have a response
2024-07-29 14:23:56 -07:00
Krrish Dholakia
ae4bcd8a41
fix(utils.py): fix trim_messages to handle tool calling
...
Fixes https://github.com/BerriAI/litellm/issues/4931
2024-07-29 13:04:41 -07:00
Krrish Dholakia
59384c84a5
fix(utils.py): correctly re-raise azure api connection error
...
'
2024-07-29 12:28:25 -07:00
Krrish Dholakia
52b1ca292c
fix(utils.py): check if tools is iterable before indexing into it
...
Fixes https://github.com/BerriAI/litellm/issues/4933
2024-07-29 09:01:32 -07:00
Ravi N
b321f2988b
Allow zero temperature for Sagemaker models based on config
...
Since Sagemaker can host any kind of model, some models allow
zero temperature. However, this is not enabled by default and
only allowed based on config
2024-07-28 21:55:53 -04:00
Krrish Dholakia
1ac42d8464
fix: utils.py
...
fix supported openai params
2024-07-27 22:03:40 -07:00
Krish Dholakia
e3a94ac013
Merge pull request #4925 from BerriAI/litellm_vertex_mistral
...
feat(vertex_ai_partner.py): Vertex AI Mistral Support
2024-07-27 21:51:26 -07:00
Krish Dholakia
b854d2100c
Merge branch 'main' into litellm_vertex_migration
2024-07-27 20:25:12 -07:00
Krrish Dholakia
d1989b6063
fix(utils.py): support fireworks ai finetuned models
...
Fixes https://github.com/BerriAI/litellm/issues/4923
2024-07-27 15:38:27 -07:00
Krrish Dholakia
f76cad210c
fix(utils.py): support fireworks ai finetuned models
...
Fixes https://github.com/BerriAI/litellm/issues/4923
2024-07-27 15:37:28 -07:00
Krrish Dholakia
c85ed01756
feat(utils.py): fix openai-like streaming
2024-07-27 15:32:57 -07:00
Krrish Dholakia
05ba34b9b7
fix(utils.py): add exception mapping for databricks errors
2024-07-27 13:13:31 -07:00
Krrish Dholakia
5b71421a7b
feat(vertex_ai_partner.py): initial working commit for calling vertex ai mistral
...
Closes https://github.com/BerriAI/litellm/issues/4874
2024-07-27 12:54:14 -07:00
Krrish Dholakia
b25d4a8cb3
feat(ollama_chat.py): support ollama tool calling
...
Closes https://github.com/BerriAI/litellm/issues/4812
2024-07-26 21:51:54 -07:00
Krrish Dholakia
fe0b55f2ca
fix(utils.py): fix cache hits for streaming
...
Fixes https://github.com/BerriAI/litellm/issues/4109
2024-07-26 19:04:08 -07:00
Krrish Dholakia
2f773d9cb6
fix(litellm_cost_calc/google.py): support meta llama vertex ai cost tracking
2024-07-25 22:12:07 -07:00
Ishaan Jaff
079a41fbe1
Merge branch 'main' into litellm_proxy_support_all_providers
2024-07-25 20:15:37 -07:00
Krrish Dholakia
a2fd8459fc
fix(utils.py): don't raise error on openai content filter during streaming - return as is
...
Fixes issue where we would raise an error vs. openai who return the chunk with finish reason as 'content_filter'
2024-07-25 19:50:52 -07:00
Krish Dholakia
b6ca4406b6
Merge branch 'main' into bedrock-llama3.1-405b
2024-07-25 19:29:10 -07:00
Ishaan Jaff
9863520376
support using */*
2024-07-25 18:48:56 -07:00
Krrish Dholakia
060249c7e0
feat(utils.py): support async streaming for custom llm provider
2024-07-25 17:11:57 -07:00
Krrish Dholakia
b4e3a77ad0
feat(utils.py): support sync streaming for custom llm provider
2024-07-25 16:47:32 -07:00
Krrish Dholakia
6bf1b9353b
feat(custom_llm.py): initial working commit for writing your own custom LLM handler
...
Fixes https://github.com/BerriAI/litellm/issues/4675
Also Addresses https://github.com/BerriAI/litellm/discussions/4677
2024-07-25 15:33:05 -07:00
David Manouchehri
64adae6e7f
Check for converse support first.
2024-07-25 21:16:23 +00:00
Krrish Dholakia
4e51f712f3
fix(main.py): fix calling openai gpt-3.5-turbo-instruct via /completions
...
Fixes https://github.com/BerriAI/litellm/issues/749
2024-07-25 09:57:19 -07:00
wslee
40bb165108
support dynamic api base
2024-07-25 11:14:38 +09:00
wslee
dd10da4d46
add support for friendli dedicated endpoint
2024-07-25 11:14:35 +09:00
Ishaan Jaff
8ea4b73c27
add UnsupportedParamsError to litellm exceptions
2024-07-24 12:20:14 -07:00
Krrish Dholakia
7df94100e8
build(model_prices_and_context_window.json): add model pricing for vertex ai llama 3.1 api
2024-07-23 17:36:07 -07:00
Krrish Dholakia
83ef52e180
feat(vertex_ai_llama.py): vertex ai llama3.1 api support
...
Initial working commit for vertex ai llama 3.1 api support
2024-07-23 17:07:30 -07:00
Krrish Dholakia
f64a3309d1
fix(utils.py): support raw response headers for streaming requests
2024-07-23 11:58:58 -07:00
Krrish Dholakia
dcb974dd1e
feat(utils.py): support passing openai response headers to client, if enabled
...
Allows openai/openai-compatible provider response headers to be sent to client, if 'return_response_headers' is enabled
2024-07-23 11:30:52 -07:00
Ishaan Jaff
1355932bf4
Merge pull request #3905 from giritatavarty-8451/litellm_triton_chatcompletion_support
...
Litellm triton chatcompletion support - Resubmit of #3895
2024-07-23 10:30:26 -07:00
Ishaan Jaff
8f9638f2c1
fix raise correct provider on content policy violation
2024-07-22 16:03:15 -07:00
Ishaan Jaff
15c109f023
fix checking if _known_custom_logger_compatible_callbacks
2024-07-22 15:43:43 -07:00
Krrish Dholakia
f10af7596c
fix(utils.py): allow dropping extra_body in additional_drop_params
...
Fixes https://github.com/BerriAI/litellm/issues/4769
2024-07-20 19:12:58 -07:00
Ishaan Jaff
5e4d291244
rename to _response_headers
2024-07-20 17:31:16 -07:00
Ishaan Jaff
46cf4f69ae
return response headers in response
2024-07-20 14:59:08 -07:00
Krish Dholakia
3053f52c43
Merge pull request #4801 from BerriAI/litellm_dynamic_params_oai_compatible_endpoints
...
fix(utils.py): support dynamic params for openai-compatible providers
2024-07-19 21:07:06 -07:00
Krrish Dholakia
95a0f6839f
fix(utils.py): fix token_counter to handle empty tool calls in messages
...
Fixes https://github.com/BerriAI/litellm/pull/4749
2024-07-19 19:39:00 -07:00
Krrish Dholakia
e45956d77e
fix(utils.py): fix get_llm_provider to support dynamic params for openai-compatible providers
2024-07-19 19:36:31 -07:00
Krrish Dholakia
e2d275f1b7
fix(utils.py): add exception mapping for bedrock image internal server error
2024-07-19 19:30:41 -07:00
Sophia Loris
d779253949
resolve merge conflicts
2024-07-19 09:45:53 -05:00
Sophia Loris
d5c65c6be2
Add support for Triton streaming & triton async completions
2024-07-19 09:35:27 -05:00
Krrish Dholakia
b23a633cf1
fix(utils.py): fix status code in exception mapping
2024-07-18 18:04:59 -07:00