Commit graph

43 commits

Author SHA1 Message Date
Krrish Dholakia
3d18897d69 feat(router.py): enable filtering model group by 'allowed_model_region' 2024-05-08 22:10:17 -07:00
CyanideByte
b7c60031ca Pydantic warning snuck in here 2024-05-07 23:59:18 -07:00
Krish Dholakia
0e709fdc21
Merge branch 'main' into litellm_ui_fixes_6 2024-05-07 22:01:04 -07:00
Krrish Dholakia
fbcda918de feat(ui/model_dashboard.tsx): show if model is config or db model 2024-05-07 21:34:18 -07:00
Krrish Dholakia
5a16bec6a1 feat(model_dashboard.tsx): allow user to edit input cost per token for model on ui
also contains fixes for `/model/update`
2024-05-07 20:57:21 -07:00
Krrish Dholakia
312249ca44 feat(ui/model_dashboard.tsx): show if model is config or db model 2024-05-07 18:29:14 -07:00
Ishaan Jaff
dc74204427 fix typo 2024-05-07 18:27:49 -07:00
Ishaan Jaff
c08352a0ce router- initialize alerting 2024-05-07 18:03:04 -07:00
Ishaan Jaff
5fd3b12d34 add router alerting type 2024-05-07 17:46:18 -07:00
Ishaan Jaff
6a056c0c17 fix support InternalServerErrorRetries 2024-05-06 17:59:38 -07:00
Ishaan Jaff
116e555a03 fix - setting model_group_retry_policy in db 2024-05-06 17:43:43 -07:00
Krish Dholakia
6be20f5fc6
Merge pull request #3455 from BerriAI/litellm_assistants_support
feat(openai.py): add support for openai assistants
2024-05-04 22:21:44 -07:00
Krrish Dholakia
06ae584473 fix(types/openai.py): fix python3.8 typing issue 2024-05-04 22:04:17 -07:00
Krrish Dholakia
66129bc921 fix(typing/openai.py): fix openai typing error (version-related) 2024-05-04 22:02:43 -07:00
Krrish Dholakia
2deac08ff1 fix(types/openai.py): fix typing import 2024-05-04 21:53:08 -07:00
Krrish Dholakia
681a95e37b fix(assistants/main.py): support litellm.create_thread() call 2024-05-04 19:35:37 -07:00
Ishaan Jaff
87e165e413
Merge pull request #3456 from BerriAI/litellm_router_set_retry_policy_errors
[FEAT] router set custom num retries for ContentPolicyViolationErrorRetries, RateLimitErrorRetries, BadRequestErrorRetries etc
2024-05-04 18:26:03 -07:00
Ishaan Jaff
6d1981fbaa init router retry policy 2024-05-04 16:59:14 -07:00
Krish Dholakia
601e8a1172
Merge pull request #3448 from BerriAI/litellm_anthropic_fix
fix(factory.py): support 'function' openai message role for anthropic
2024-05-04 16:36:05 -07:00
Ishaan Jaff
d968dedd77
Merge pull request #1530 from TanaroSch/main
change max_tokens type to int
2024-05-04 12:47:15 -07:00
Krrish Dholakia
33472bfd2b fix(factory.py): support 'function' openai message role for anthropic
Fixes https://github.com/BerriAI/litellm/issues/3446
2024-05-04 10:03:30 -07:00
Krrish Dholakia
8249c986bf fix(main.py): support new 'supports_system_message=False' param
Fixes https://github.com/BerriAI/litellm/issues/3325
2024-05-03 21:31:45 -07:00
Krrish Dholakia
5b39f8e282 feat(proxy_server.py): return api base in response headers
Closes https://github.com/BerriAI/litellm/issues/2631
2024-05-03 15:27:32 -07:00
Krrish Dholakia
f10a066d36 fix(lowest_tpm_rpm_v2.py): add more detail to 'No deployments available' error message 2024-04-29 15:04:37 -07:00
Krrish Dholakia
0a6b6302f1 fix(router.py): fix typing error 2024-04-29 07:25:39 -07:00
CyanideByte
e1786848cb protected_namespaces fixed for model_info 2024-04-27 13:08:45 -07:00
Krish Dholakia
1a06f009d1
Merge branch 'main' into litellm_default_router_retries 2024-04-27 11:21:57 -07:00
Krrish Dholakia
7730520fb0 fix(router.py): allow passing httpx.timeout to timeout param in router
Closes https://github.com/BerriAI/litellm/issues/3162
2024-04-26 14:57:19 -07:00
Krrish Dholakia
160acc085a fix(router.py): fix default retry logic 2024-04-25 11:57:27 -07:00
Krrish Dholakia
f54510b6ee fix(proxy_server.py): fix /config/update/
allows updating router config via UI and having the change be propogated across all proxy instances by persisting config changes to the db
2024-04-24 16:42:42 -07:00
Ishaan Jaff
3d1a158b63 feat - update deployments 2024-04-24 09:53:42 -07:00
Krrish Dholakia
0f95a824c4 feat(prometheus_services.py): emit proxy latency for successful llm api requests
uses prometheus histogram for this
2024-04-18 16:04:35 -07:00
Krrish Dholakia
13cd252f3e fix(proxy_server.py): ensure id used in delete deployment matches id used in litellm Router 2024-04-16 15:17:18 -07:00
Krrish Dholakia
4e81acf2c6 feat(prometheus_services.py): monitor health of proxy adjacent services (redis / postgres / etc.) 2024-04-13 18:15:02 -07:00
Krrish Dholakia
ea1574c160 test(test_openai_endpoints.py): add concurrency testing for user defined rate limits on proxy 2024-04-12 18:56:13 -07:00
Krrish Dholakia
a47a719caa fix(router.py): generate consistent model id's
having the same id for a deployment, lets redis usage caching work across multiple instances
2024-04-10 15:23:57 -07:00
unclecode
5554e2c359 Continue fixing the issue #2832: Add protected_namespaces to another to Config class within the router.py 2024-04-08 12:49:40 +08:00
unclecode
1ace192155 Fix issue #2832: Add protected_namespaces to Config class within utils.py, router.py and completion.py to avoid the warning message. 2024-04-08 12:43:17 +08:00
Krrish Dholakia
460546956d fix(utils.py): fix import 2024-04-06 18:37:38 -07:00
TanaroSch
b69cb2c773 change max_tokens float to int 2024-02-06 11:19:28 +01:00
ishaan-jaff
075eb1a516 (types) routerConfig 2024-01-02 14:14:29 +05:30
ishaan-jaff
9afdc8b4ee (feat) add Router init Pydantic Type 2024-01-02 13:30:24 +05:30
ishaan-jaff
1f8fc6d2a7 (feat) litellm add types for completion, embedding request 2024-01-02 12:27:08 +05:30