Haadi Rakhangi
|
5439e72a6b
|
Merge branch 'BerriAI:main' into main
|
2024-08-02 21:08:48 +05:30 |
|
Haadi Rakhangi
|
851db5ecea
|
qdrant semantic caching added
|
2024-08-02 21:07:19 +05:30 |
|
Krrish Dholakia
|
fe7e68adc8
|
fix(utils.py): fix codestral streaming
|
2024-08-02 07:38:06 -07:00 |
|
Krrish Dholakia
|
4c2ef8ea64
|
fix(bedrock_httpx.py): fix ai21 streaming
|
2024-08-01 22:03:24 -07:00 |
|
Krish Dholakia
|
25ac9c2d75
|
Merge branch 'main' into litellm_fix_streaming_usage_calc
|
2024-08-01 21:29:04 -07:00 |
|
Krish Dholakia
|
e7f83602d8
|
Merge branch 'main' into litellm_response_cost_logging
|
2024-08-01 21:28:22 -07:00 |
|
Krrish Dholakia
|
010d5ed81d
|
feat(vertex_ai_partner.py): add vertex ai codestral FIM support
Closes https://github.com/BerriAI/litellm/issues/4984
|
2024-08-01 17:10:27 -07:00 |
|
Krrish Dholakia
|
28c12e6702
|
fix(utils.py): fix togetherai streaming cost calculation
|
2024-08-01 15:03:08 -07:00 |
|
Krrish Dholakia
|
ca0a0bed46
|
fix(utils.py): fix anthropic streaming usage calculation
Fixes https://github.com/BerriAI/litellm/issues/4965
|
2024-08-01 14:45:54 -07:00 |
|
Krrish Dholakia
|
10b571ca42
|
fix(litellm_logging.py): use 1 cost calc function across response headers + logging integrations
Ensures consistent cost calculation when azure base models are used
|
2024-08-01 10:26:59 -07:00 |
|
Krrish Dholakia
|
1a00f52255
|
fix(utils.py): fix special keys list for provider-specific items in response object
|
2024-07-31 18:30:49 -07:00 |
|
Krrish Dholakia
|
09ee8c6e2d
|
fix(utils.py): return additional kwargs from openai-like response body
Closes https://github.com/BerriAI/litellm/issues/4981
|
2024-07-31 15:37:03 -07:00 |
|
Krrish Dholakia
|
bd68714f51
|
fix(utils.py): map cohere timeout error
|
2024-07-31 15:15:18 -07:00 |
|
Krrish Dholakia
|
c65a438de2
|
fix(utils.py): fix linting errors
|
2024-07-30 18:38:10 -07:00 |
|
Krrish Dholakia
|
46634af06f
|
fix(utils.py): fix model registeration to model cost map
Fixes https://github.com/BerriAI/litellm/issues/4972
|
2024-07-30 18:15:00 -07:00 |
|
Krrish Dholakia
|
185a6857f9
|
fix(utils.py): fix cost tracking for vertex ai partner models
|
2024-07-30 14:20:52 -07:00 |
|
Krish Dholakia
|
6bdb9203b3
|
Merge pull request #4948 from dleen/response
fixes: #4947 Bedrock context exception does not have a response
|
2024-07-29 15:03:40 -07:00 |
|
David Leen
|
ebca13b2ca
|
fixes: #4947 Bedrock context exception does not have a response
|
2024-07-29 14:23:56 -07:00 |
|
Krrish Dholakia
|
ae4bcd8a41
|
fix(utils.py): fix trim_messages to handle tool calling
Fixes https://github.com/BerriAI/litellm/issues/4931
|
2024-07-29 13:04:41 -07:00 |
|
Krrish Dholakia
|
59384c84a5
|
fix(utils.py): correctly re-raise azure api connection error
'
|
2024-07-29 12:28:25 -07:00 |
|
Krrish Dholakia
|
52b1ca292c
|
fix(utils.py): check if tools is iterable before indexing into it
Fixes https://github.com/BerriAI/litellm/issues/4933
|
2024-07-29 09:01:32 -07:00 |
|
Krish Dholakia
|
e3a94ac013
|
Merge pull request #4925 from BerriAI/litellm_vertex_mistral
feat(vertex_ai_partner.py): Vertex AI Mistral Support
|
2024-07-27 21:51:26 -07:00 |
|
Krrish Dholakia
|
d1989b6063
|
fix(utils.py): support fireworks ai finetuned models
Fixes https://github.com/BerriAI/litellm/issues/4923
|
2024-07-27 15:38:27 -07:00 |
|
Krrish Dholakia
|
f76cad210c
|
fix(utils.py): support fireworks ai finetuned models
Fixes https://github.com/BerriAI/litellm/issues/4923
|
2024-07-27 15:37:28 -07:00 |
|
Krrish Dholakia
|
c85ed01756
|
feat(utils.py): fix openai-like streaming
|
2024-07-27 15:32:57 -07:00 |
|
Krrish Dholakia
|
05ba34b9b7
|
fix(utils.py): add exception mapping for databricks errors
|
2024-07-27 13:13:31 -07:00 |
|
Krrish Dholakia
|
5b71421a7b
|
feat(vertex_ai_partner.py): initial working commit for calling vertex ai mistral
Closes https://github.com/BerriAI/litellm/issues/4874
|
2024-07-27 12:54:14 -07:00 |
|
Krrish Dholakia
|
b25d4a8cb3
|
feat(ollama_chat.py): support ollama tool calling
Closes https://github.com/BerriAI/litellm/issues/4812
|
2024-07-26 21:51:54 -07:00 |
|
Krrish Dholakia
|
fe0b55f2ca
|
fix(utils.py): fix cache hits for streaming
Fixes https://github.com/BerriAI/litellm/issues/4109
|
2024-07-26 19:04:08 -07:00 |
|
Krrish Dholakia
|
2f773d9cb6
|
fix(litellm_cost_calc/google.py): support meta llama vertex ai cost tracking
|
2024-07-25 22:12:07 -07:00 |
|
Ishaan Jaff
|
079a41fbe1
|
Merge branch 'main' into litellm_proxy_support_all_providers
|
2024-07-25 20:15:37 -07:00 |
|
Krrish Dholakia
|
a2fd8459fc
|
fix(utils.py): don't raise error on openai content filter during streaming - return as is
Fixes issue where we would raise an error vs. openai who return the chunk with finish reason as 'content_filter'
|
2024-07-25 19:50:52 -07:00 |
|
Krish Dholakia
|
b6ca4406b6
|
Merge branch 'main' into bedrock-llama3.1-405b
|
2024-07-25 19:29:10 -07:00 |
|
Ishaan Jaff
|
9863520376
|
support using */*
|
2024-07-25 18:48:56 -07:00 |
|
Krrish Dholakia
|
060249c7e0
|
feat(utils.py): support async streaming for custom llm provider
|
2024-07-25 17:11:57 -07:00 |
|
Krrish Dholakia
|
b4e3a77ad0
|
feat(utils.py): support sync streaming for custom llm provider
|
2024-07-25 16:47:32 -07:00 |
|
Krrish Dholakia
|
6bf1b9353b
|
feat(custom_llm.py): initial working commit for writing your own custom LLM handler
Fixes https://github.com/BerriAI/litellm/issues/4675
Also Addresses https://github.com/BerriAI/litellm/discussions/4677
|
2024-07-25 15:33:05 -07:00 |
|
David Manouchehri
|
64adae6e7f
|
Check for converse support first.
|
2024-07-25 21:16:23 +00:00 |
|
Krrish Dholakia
|
4e51f712f3
|
fix(main.py): fix calling openai gpt-3.5-turbo-instruct via /completions
Fixes https://github.com/BerriAI/litellm/issues/749
|
2024-07-25 09:57:19 -07:00 |
|
wslee
|
40bb165108
|
support dynamic api base
|
2024-07-25 11:14:38 +09:00 |
|
wslee
|
dd10da4d46
|
add support for friendli dedicated endpoint
|
2024-07-25 11:14:35 +09:00 |
|
Ishaan Jaff
|
8ea4b73c27
|
add UnsupportedParamsError to litellm exceptions
|
2024-07-24 12:20:14 -07:00 |
|
Krrish Dholakia
|
7df94100e8
|
build(model_prices_and_context_window.json): add model pricing for vertex ai llama 3.1 api
|
2024-07-23 17:36:07 -07:00 |
|
Krrish Dholakia
|
83ef52e180
|
feat(vertex_ai_llama.py): vertex ai llama3.1 api support
Initial working commit for vertex ai llama 3.1 api support
|
2024-07-23 17:07:30 -07:00 |
|
Krrish Dholakia
|
f64a3309d1
|
fix(utils.py): support raw response headers for streaming requests
|
2024-07-23 11:58:58 -07:00 |
|
Krrish Dholakia
|
dcb974dd1e
|
feat(utils.py): support passing openai response headers to client, if enabled
Allows openai/openai-compatible provider response headers to be sent to client, if 'return_response_headers' is enabled
|
2024-07-23 11:30:52 -07:00 |
|
Ishaan Jaff
|
1355932bf4
|
Merge pull request #3905 from giritatavarty-8451/litellm_triton_chatcompletion_support
Litellm triton chatcompletion support - Resubmit of #3895
|
2024-07-23 10:30:26 -07:00 |
|
Ishaan Jaff
|
8f9638f2c1
|
fix raise correct provider on content policy violation
|
2024-07-22 16:03:15 -07:00 |
|
Ishaan Jaff
|
15c109f023
|
fix checking if _known_custom_logger_compatible_callbacks
|
2024-07-22 15:43:43 -07:00 |
|
Krrish Dholakia
|
f10af7596c
|
fix(utils.py): allow dropping extra_body in additional_drop_params
Fixes https://github.com/BerriAI/litellm/issues/4769
|
2024-07-20 19:12:58 -07:00 |
|