Commit graph

207 commits

Author SHA1 Message Date
Ishaan Jaff
47cbd62347 validation for passing config file 2024-07-31 13:32:18 -07:00
Ishaan Jaff
b630ff6286 fix pydantic obj for FT endpoints 2024-07-31 12:41:39 -07:00
Ishaan Jaff
4d1a653fea fix type errors 2024-07-29 20:10:03 -07:00
Ishaan Jaff
6db50886bb fix linting 2024-07-29 20:01:12 -07:00
Ishaan Jaff
93cbfb3ff8 add types for FineTuningJobCreate OpenAI 2024-07-29 18:59:44 -07:00
Ishaan Jaff
150179a985 Merge pull request #4946 from BerriAI/litellm_Add_bedrock_guardrail_config
[Feat] Bedrock add support for Bedrock Guardrails
2024-07-29 15:09:56 -07:00
Ishaan Jaff
6c88305d3b types add GuardrailConfigBlock 2024-07-29 13:14:53 -07:00
Krrish Dholakia
00dde68001 fix(utils.py): fix trim_messages to handle tool calling
Fixes https://github.com/BerriAI/litellm/issues/4931
2024-07-29 13:04:41 -07:00
Ishaan Jaff
582b047b0e add new BATCH_WRITE_TO_DB type for service logger 2024-07-27 11:36:51 -07:00
Krrish Dholakia
3a1eedfbf3 feat(ollama_chat.py): support ollama tool calling
Closes https://github.com/BerriAI/litellm/issues/4812
2024-07-26 21:51:54 -07:00
Ishaan Jaff
1103c614a0 Merge branch 'main' into litellm_proxy_support_all_providers 2024-07-25 20:15:37 -07:00
Ishaan Jaff
5128f0e343 router support setting pass_through_all_models 2024-07-25 18:22:35 -07:00
Krrish Dholakia
54e1ca29b7 feat(custom_llm.py): initial working commit for writing your own custom LLM handler
Fixes https://github.com/BerriAI/litellm/issues/4675

 Also Addresses https://github.com/BerriAI/litellm/discussions/4677
2024-07-25 15:33:05 -07:00
Ishaan Jaff
344010e127 Pass litellm proxy specific metadata 2024-07-23 15:31:30 -07:00
Krrish Dholakia
14c5d6ff03 fix(proxy/utils.py): add stronger typing for litellm params in failure call logging 2024-07-22 21:31:39 -07:00
Ishaan Jaff
2dcbd5c534 rename to _response_headers 2024-07-20 17:31:16 -07:00
Ishaan Jaff
fb4aaeb540 return response_headers in response 2024-07-20 14:58:14 -07:00
Ishaan Jaff
cf7db1280f fix typing errors 2024-07-19 16:18:53 -07:00
Ishaan Jaff
cd40d58544 router - refactor to tag based routing 2024-07-18 19:22:09 -07:00
Ishaan Jaff
10ca47d431 litellm router - use free / paid tier 2024-07-18 16:55:50 -07:00
Ishaan Jaff
da2d6348a7 fix linting errors on main 2024-07-18 13:32:48 -07:00
Krish Dholakia
c010cd2dca Merge pull request #4729 from vingiarrusso/vgiarrusso/guardrails
Add enabled_roles to Guardrails configuration, Update Lakera guardrail moderation hook
2024-07-17 22:24:35 -07:00
Krish Dholakia
b6ede4eb1b Merge pull request #4716 from pamelafox/countfuncs
Add token counting for OpenAI tools/tool_choice
2024-07-16 07:21:31 -07:00
Vinnie Giarrusso
0535bd2d68 Add enabled_roles to Guardrails configuration, Update Lakera guardrail moderation hook 2024-07-16 01:52:08 -07:00
Ishaan Jaff
7944450074 Merge pull request #4724 from BerriAI/litellm_Set_max_file_size_transc
[Feat] - set max file size on /audio/transcriptions
2024-07-15 20:42:24 -07:00
Ishaan Jaff
4ec6d3847d max_file_size_mb in float 2024-07-15 19:58:41 -07:00
Krrish Dholakia
58ed852a25 fix(vertex_httpx.py): return grounding metadata 2024-07-15 19:43:37 -07:00
Ishaan Jaff
11ed40be80 allow setting max_file_size_mb 2024-07-15 19:25:24 -07:00
Pamela Fox
ae6b8450c1 Count tokens for tools 2024-07-15 11:07:52 -07:00
Krrish Dholakia
6641683d66 feat(guardrails.py): allow setting logging_only in guardrails_config for presidio pii masking integration 2024-07-13 12:22:17 -07:00
Krrish Dholakia
d5f5415add fix(types/guardrails.py): add 'logging_only' param support 2024-07-13 11:44:37 -07:00
Krish Dholakia
f01298bec9 Merge pull request #4588 from Manouchehri/vertex-seed-2973
feat(vertex_httpx.py): Add seed parameter
2024-07-11 22:02:13 -07:00
Krrish Dholakia
d85f24a80b fix(utils.py): fix recreating model response object when stream usage is true 2024-07-11 21:01:12 -07:00
Ishaan Jaff
bf50c8e087 Merge pull request #4661 from BerriAI/litellm_fix_mh
[Fix] Model Hub - Show supports vision correctly
2024-07-11 15:03:37 -07:00
Krrish Dholakia
26a2ae76ab fix(types/utils.py): message role is always 'assistant' 2024-07-11 14:14:38 -07:00
Ishaan Jaff
a16cd02cd9 fix supports vision 2024-07-11 12:59:42 -07:00
Krrish Dholakia
91c1d7bfa8 fix(watsonx.py): fix watson process response
Fixes https://github.com/BerriAI/litellm/issues/4654
2024-07-11 09:34:46 -07:00
Krrish Dholakia
af1064941a fix(types/utils.py): fix streaming function name 2024-07-10 21:56:47 -07:00
Krrish Dholakia
48be4ce805 feat(proxy_server.py): working /v1/messages with config.yaml
Adds async router support for adapter_completion call
2024-07-10 18:53:54 -07:00
Krrish Dholakia
4ba30abb63 feat(proxy_server.py): working /v1/messages endpoint
Works with claude engineer
2024-07-10 18:15:38 -07:00
Krrish Dholakia
01a335b4c3 feat(anthropic_adapter.py): support for translating anthropic params to openai format 2024-07-10 00:32:28 -07:00
Krrish Dholakia
71ad281c0a feat(vertex_httpx.py): add support for gemini 'grounding'
Adds support for https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/grounding#rest
2024-07-08 21:37:07 -07:00
David Manouchehri
3f993699c0 feat(vertex_httpx.py): Add undocumented seed parameter. 2024-07-07 23:32:04 +00:00
Krish Dholakia
c643be0c0c Merge branch 'main' into litellm_gemini_stream_tool_calling 2024-07-06 19:07:31 -07:00
Krish Dholakia
5640ed4c8c Merge branch 'main' into litellm_proxy_tts_pricing 2024-07-06 14:56:16 -07:00
Krrish Dholakia
9f900a1bed fix(vertex_httpx.py): support tool calling w/ streaming for vertex ai + gemini 2024-07-06 14:02:25 -07:00
Ishaan Jaff
b9ab94a6bb allow async_only_mode on rotuer 2024-07-06 12:50:57 -07:00
Krrish Dholakia
356c18c929 feat(litellm_logging.py): support cost tracking for tts calls 2024-07-05 22:09:08 -07:00
Krrish Dholakia
56410cfcd0 fix(proxy_server.py): support langfuse logging for rejected requests on /v1/chat/completions 2024-07-05 13:07:09 -07:00
Krrish Dholakia
8a8dde6622 fix(vertex_httpx.py): fix assumptions on usagemetadata 2024-07-05 11:01:51 -07:00