Krish Dholakia
|
637369d2ac
|
Merge branch 'main' into litellm_fix_httpx_transport
|
2024-07-02 17:17:43 -07:00 |
|
Krrish Dholakia
|
459d294e21
|
fix(router.py): disable cooldowns
allow admin to disable model cooldowns
|
2024-07-01 15:03:10 -07:00 |
|
Krrish Dholakia
|
931a9934f1
|
fix(router.py): fix get_router_model_info for azure models
|
2024-06-28 22:13:29 -07:00 |
|
Ishaan Jaff
|
158148d7ce
|
fix python3.8 install
|
2024-06-28 16:58:57 -07:00 |
|
Krrish Dholakia
|
6e53de5462
|
fix(router.py): only return 'max_tokens', 'input_cost_per_token', etc. in 'get_router_model_info' if base_model is set
|
2024-06-28 10:45:31 -07:00 |
|
Krrish Dholakia
|
d67cdee850
|
fix(router.py): fix setting httpx mounts
|
2024-06-26 17:22:04 -07:00 |
|
Krrish Dholakia
|
c70d4ffafb
|
fix(router.py): set cooldown_time: per model
|
2024-06-25 16:51:55 -07:00 |
|
Krrish Dholakia
|
715153b27b
|
fix(router.py): improve error message returned for fallbacks
|
2024-06-25 11:27:20 -07:00 |
|
Krrish Dholakia
|
c46b229202
|
feat(router.py): support mock testing content policy + context window fallbacks
|
2024-06-25 10:58:19 -07:00 |
|
Krrish Dholakia
|
94830ec281
|
fix(router.py): log rejected router requests to langfuse
Fixes issue where rejected requests weren't being logged
|
2024-06-24 17:52:01 -07:00 |
|
Krrish Dholakia
|
ddd32e5341
|
fix(router.py): use user-defined model_input_tokens for pre-call filter checks
|
2024-06-24 17:25:26 -07:00 |
|
Krish Dholakia
|
39c2fe511c
|
Merge branch 'main' into litellm_azure_content_filter_fallbacks
|
2024-06-22 21:28:29 -07:00 |
|
Krish Dholakia
|
63d0defa6d
|
Merge branch 'main' into litellm_dynamic_tpm_limits
|
2024-06-22 19:14:59 -07:00 |
|
Krrish Dholakia
|
f9ce6472d7
|
fix(router.py): check if azure returns 'content_filter' response + fallback available -> fallback
Exception maps azure content filter response exceptions
|
2024-06-22 19:10:15 -07:00 |
|
Krrish Dholakia
|
6a7982fa40
|
feat(dynamic_rate_limiter.py): passing base case
|
2024-06-21 22:46:46 -07:00 |
|
Steven Osborn
|
7bd91a5b57
|
Print content window fallbacks on startup to help verify configuration
|
2024-06-21 19:43:26 -07:00 |
|
Krrish Dholakia
|
89dba82be9
|
feat(dynamic_rate_limiter.py): initial commit for dynamic rate limiting
Closes https://github.com/BerriAI/litellm/issues/4124
|
2024-06-21 18:41:31 -07:00 |
|
Krrish Dholakia
|
a8290ec294
|
fix(router.py): fix set_client init to check if custom_llm_provider is azure not if in model name
fixes issue where 'azure_ai/' was being init as azureopenai client
|
2024-06-21 17:09:20 -07:00 |
|
Krish Dholakia
|
2e814bb57d
|
Merge pull request #4290 from BerriAI/litellm_specific_deployment
feat(router.py): allow user to call specific deployment via id
|
2024-06-20 20:36:13 -07:00 |
|
Krrish Dholakia
|
66948d083f
|
fix(user_api_key_auth.py): ensure user has access to fallback models
for client side fallbacks, checks if user has access to fallback models
|
2024-06-20 16:02:19 -07:00 |
|
Ishaan Jaff
|
6186d40823
|
router - add doc string
|
2024-06-20 14:36:51 -07:00 |
|
Ishaan Jaff
|
7ce2aa83c1
|
feat - set custom routing strategy
|
2024-06-20 13:49:44 -07:00 |
|
Krrish Dholakia
|
477ed99896
|
feat(router.py): allow user to call specific deployment via id
Allows easier health checks for specific deployments by just passing in model id
|
2024-06-19 13:02:46 -07:00 |
|
Krrish Dholakia
|
121f4d8a1b
|
fix(router.py): support multiple orgs in 1 model definition
Closes https://github.com/BerriAI/litellm/issues/3949
|
2024-06-18 19:36:58 -07:00 |
|
Krrish Dholakia
|
cc1ec55e5b
|
fix(vertex_httpx.py): check if model supports system messages before sending separately
|
2024-06-17 17:32:38 -07:00 |
|
Krish Dholakia
|
63780e1ccf
|
Merge pull request #4207 from BerriAI/litellm_content_policy_fallbacks
feat(router.py): support content policy fallbacks
|
2024-06-14 18:55:11 -07:00 |
|
Krrish Dholakia
|
734bd5ef85
|
feat(router.py): support content policy fallbacks
Closes https://github.com/BerriAI/litellm/issues/2632
|
2024-06-14 17:15:44 -07:00 |
|
Ishaan Jaff
|
5f7423047f
|
feat - send llm exception alert on acompletion, aembedding etc
|
2024-06-14 10:11:24 -07:00 |
|
Ishaan Jaff
|
bd341c69b5
|
fix - send alert on router level exceptions
|
2024-06-14 08:41:12 -07:00 |
|
Ishaan Jaff
|
b4274bc852
|
fix model hub not loading
|
2024-06-12 19:38:31 -07:00 |
|
Ishaan Jaff
|
5e411a45d5
|
fix azure falbacks test
|
2024-06-10 21:50:54 -07:00 |
|
Ishaan Jaff
|
94210a86b4
|
test - client side fallbacks
|
2024-06-10 15:00:36 -07:00 |
|
Ishaan Jaff
|
878fa676d7
|
fix - support fallbacks as list
|
2024-06-10 14:32:28 -07:00 |
|
Krrish Dholakia
|
58cce8a922
|
fix(types/router.py): modelgroupinfo to handle mode being None and supported_openai_params not being a list
|
2024-06-08 20:13:45 -07:00 |
|
Krish Dholakia
|
9d81f1cc52
|
Merge pull request #4049 from BerriAI/litellm_cleanup_traceback
refactor: replace 'traceback.print_exc()' with logging library
|
2024-06-07 08:03:22 -07:00 |
|
Krish Dholakia
|
0b3165e5ee
|
Merge pull request #4046 from BerriAI/litellm_router_order
feat(router.py): enable settting 'order' for a deployment in model list
|
2024-06-06 16:37:03 -07:00 |
|
Krish Dholakia
|
ea4334f760
|
Merge branch 'main' into litellm_cleanup_traceback
|
2024-06-06 16:32:08 -07:00 |
|
Krrish Dholakia
|
43991afc34
|
feat(scheduler.py): support redis caching for req. prioritization
enables req. prioritization to work across multiple instances of litellm
|
2024-06-06 14:19:21 -07:00 |
|
Krrish Dholakia
|
e391e30285
|
refactor: replace 'traceback.print_exc()' with logging library
allows error logs to be in json format for otel logging
|
2024-06-06 13:47:43 -07:00 |
|
Krrish Dholakia
|
005128addc
|
feat(router.py): enable settting 'order' for a deployment in model list
Allows user to control which model gets called first in model group
|
2024-06-06 09:46:51 -07:00 |
|
Krrish Dholakia
|
20cb525a5c
|
feat(assistants/main.py): add assistants api streaming support
|
2024-06-04 16:30:35 -07:00 |
|
Krish Dholakia
|
73ae4860c0
|
Merge pull request #3992 from BerriAI/litellm_router_default_request_timeout
fix(router.py): use `litellm.request_timeout` as default for router clients
|
2024-06-03 21:37:38 -07:00 |
|
Krish Dholakia
|
127d1457de
|
Merge pull request #3996 from BerriAI/litellm_azure_assistants_api_support
feat(assistants/main.py): Azure Assistants API support
|
2024-06-03 21:05:03 -07:00 |
|
Krrish Dholakia
|
a2ba63955a
|
feat(assistants/main.py): Closes https://github.com/BerriAI/litellm/issues/3993
|
2024-06-03 18:47:05 -07:00 |
|
Krrish Dholakia
|
ae52e7559e
|
fix(router.py): use litellm.request_timeout as default for router clients
|
2024-06-03 14:19:53 -07:00 |
|
Krrish Dholakia
|
96120ab2c5
|
fix(router.py): fix should_retry logic for authentication errors
|
2024-06-03 13:12:00 -07:00 |
|
Ishaan Jaff
|
0acb6e5180
|
ci/cd run again
|
2024-06-01 21:19:32 -07:00 |
|
Ishaan Jaff
|
2d1aaf5cf7
|
fix test_rate_limit[usage-based-routing-True-3-2]
|
2024-06-01 21:18:23 -07:00 |
|
Ishaan Jaff
|
ad920be3bf
|
fix async_function_with_retries
|
2024-06-01 19:00:22 -07:00 |
|
Ishaan Jaff
|
e149ca73f6
|
Merge pull request #3963 from BerriAI/litellm_set_allowed_fail_policy
[FEAT]- set custom AllowedFailsPolicy on litellm.Router
|
2024-06-01 17:57:11 -07:00 |
|