Commit graph

670 commits

Author SHA1 Message Date
Krrish Dholakia
0de640700d fix(router.py): add reason for fallback failure to client-side exception string
make it easier to debug why a fallback failed to occur
2024-08-07 13:02:47 -07:00
Ishaan Jaff
0dd8f50477 use router_cooldown_handler 2024-08-07 10:40:55 -07:00
Krrish Dholakia
fdb47e5479 fix: fix test to specify allowed_fails 2024-08-05 21:39:59 -07:00
Krrish Dholakia
934883999a fix(router.py): move deployment cooldown list message to error log, not client-side
don't show user all deployments
2024-08-03 12:49:39 -07:00
Krrish Dholakia
b0d2727bbf feat(router.py): add flag for mock testing loadbalancing for rate limit errors 2024-08-03 12:34:11 -07:00
Krrish Dholakia
dc58b9f33e fix(utils.py): fix linting errors 2024-07-30 18:38:10 -07:00
Krrish Dholakia
96ad9c877c fix(router.py): gracefully handle scenario where completion response doesn't have total tokens
Closes https://github.com/BerriAI/litellm/issues/4968
2024-07-30 15:14:03 -07:00
Krrish Dholakia
3a1eedfbf3 feat(ollama_chat.py): support ollama tool calling
Closes https://github.com/BerriAI/litellm/issues/4812
2024-07-26 21:51:54 -07:00
Krrish Dholakia
e39ff46222 docs(config.md): update wildcard docs 2024-07-26 08:59:53 -07:00
Ishaan Jaff
a46c463dee router support setting pass_through_all_models 2024-07-25 18:34:12 -07:00
Krrish Dholakia
1d33759bb1 fix(router.py): add support for diskcache to router 2024-07-25 14:30:46 -07:00
Ishaan Jaff
7888074012 fix - test router debug logs 2024-07-20 18:45:31 -07:00
Ishaan Jaff
fcee8bc61f router - use verbose logger when using litellm.Router 2024-07-20 17:36:25 -07:00
Ishaan Jaff
d1a4246d2b control using enable_tag_filtering 2024-07-18 19:39:04 -07:00
Ishaan Jaff
cd40d58544 router - refactor to tag based routing 2024-07-18 19:22:09 -07:00
Ishaan Jaff
778cb8799e Merge pull request #4786 from BerriAI/litellm_use_model_tier_keys
[Feat-Enterprise] Use free/paid tiers for Virtual Keys
2024-07-18 18:07:09 -07:00
Krrish Dholakia
5d0bb0c6ee fix(utils.py): fix status code in exception mapping 2024-07-18 18:04:59 -07:00
Ishaan Jaff
d4cad75d34 router - use free paid tier routing 2024-07-18 17:09:42 -07:00
Krrish Dholakia
432b7ae264 fix(router.py): check for request_timeout in acompletion
support 'request_timeout' param in router acompletion
2024-07-17 17:19:06 -07:00
Ishaan Jaff
dc5c72d04e router return get_deployment_by_model_group_name 2024-07-15 19:27:12 -07:00
Krish Dholakia
f4d140efec Merge pull request #4635 from BerriAI/litellm_anthropic_adapter
Anthropic `/v1/messages` endpoint support
2024-07-10 22:41:53 -07:00
Krrish Dholakia
48be4ce805 feat(proxy_server.py): working /v1/messages with config.yaml
Adds async router support for adapter_completion call
2024-07-10 18:53:54 -07:00
Ishaan Jaff
a9e15dad62 feat - add DELETE assistants endpoint 2024-07-10 11:37:37 -07:00
Ishaan Jaff
5880adea95 router - add acreate_assistants 2024-07-09 09:46:28 -07:00
Krish Dholakia
c6b6dbeb6b Merge branch 'main' into litellm_fix_httpx_transport 2024-07-06 19:12:06 -07:00
Ishaan Jaff
f6eccf84ce use helper for init client + check if we should init sync clients 2024-07-06 12:52:41 -07:00
Krrish Dholakia
8625770010 fix(types/router.py): add custom pricing info to 'model_info'
Fixes https://github.com/BerriAI/litellm/issues/4542
2024-07-04 16:07:58 -07:00
Krrish Dholakia
b17fe3e0d2 fix(router.py): bump azure default api version
Allows 'tool_choice' to be passed to azure
2024-07-03 12:00:00 -07:00
Krrish Dholakia
9ce4198fbb fix(router.py): fix mounting logic 2024-07-02 17:54:32 -07:00
Krish Dholakia
b2f2560e54 Merge branch 'main' into litellm_support_dynamic_rpm_limiting 2024-07-02 17:51:18 -07:00
Krrish Dholakia
cd51f292b6 refactor: remove custom transport logic
Not needed after azure dall-e-2 refactor
2024-07-02 17:35:27 -07:00
Krish Dholakia
637369d2ac Merge branch 'main' into litellm_fix_httpx_transport 2024-07-02 17:17:43 -07:00
Krrish Dholakia
d528e263c2 fix(dynamic_rate_limiter.py): support dynamic rate limiting on rpm 2024-07-01 17:45:10 -07:00
Krrish Dholakia
459d294e21 fix(router.py): disable cooldowns
allow admin to disable model cooldowns
2024-07-01 15:03:10 -07:00
Krrish Dholakia
931a9934f1 fix(router.py): fix get_router_model_info for azure models 2024-06-28 22:13:29 -07:00
Ishaan Jaff
158148d7ce fix python3.8 install 2024-06-28 16:58:57 -07:00
Krrish Dholakia
6e53de5462 fix(router.py): only return 'max_tokens', 'input_cost_per_token', etc. in 'get_router_model_info' if base_model is set 2024-06-28 10:45:31 -07:00
Krrish Dholakia
d67cdee850 fix(router.py): fix setting httpx mounts 2024-06-26 17:22:04 -07:00
Krrish Dholakia
c70d4ffafb fix(router.py): set cooldown_time: per model 2024-06-25 16:51:55 -07:00
Krrish Dholakia
715153b27b fix(router.py): improve error message returned for fallbacks 2024-06-25 11:27:20 -07:00
Krrish Dholakia
c46b229202 feat(router.py): support mock testing content policy + context window fallbacks 2024-06-25 10:58:19 -07:00
Krrish Dholakia
94830ec281 fix(router.py): log rejected router requests to langfuse
Fixes issue where rejected requests weren't being logged
2024-06-24 17:52:01 -07:00
Krrish Dholakia
ddd32e5341 fix(router.py): use user-defined model_input_tokens for pre-call filter checks 2024-06-24 17:25:26 -07:00
Krish Dholakia
39c2fe511c Merge branch 'main' into litellm_azure_content_filter_fallbacks 2024-06-22 21:28:29 -07:00
Krish Dholakia
63d0defa6d Merge branch 'main' into litellm_dynamic_tpm_limits 2024-06-22 19:14:59 -07:00
Krrish Dholakia
f9ce6472d7 fix(router.py): check if azure returns 'content_filter' response + fallback available -> fallback
Exception maps azure content filter response exceptions
2024-06-22 19:10:15 -07:00
Krrish Dholakia
6a7982fa40 feat(dynamic_rate_limiter.py): passing base case 2024-06-21 22:46:46 -07:00
Steven Osborn
7bd91a5b57 Print content window fallbacks on startup to help verify configuration 2024-06-21 19:43:26 -07:00
Krrish Dholakia
89dba82be9 feat(dynamic_rate_limiter.py): initial commit for dynamic rate limiting
Closes https://github.com/BerriAI/litellm/issues/4124
2024-06-21 18:41:31 -07:00
Krrish Dholakia
a8290ec294 fix(router.py): fix set_client init to check if custom_llm_provider is azure not if in model name
fixes issue where 'azure_ai/' was being init as azureopenai client
2024-06-21 17:09:20 -07:00