Krrish Dholakia
|
3d18897d69
|
feat(router.py): enable filtering model group by 'allowed_model_region'
|
2024-05-08 22:10:17 -07:00 |
|
CyanideByte
|
b7c60031ca
|
Pydantic warning snuck in here
|
2024-05-07 23:59:18 -07:00 |
|
Krish Dholakia
|
0e709fdc21
|
Merge branch 'main' into litellm_ui_fixes_6
|
2024-05-07 22:01:04 -07:00 |
|
Krrish Dholakia
|
fbcda918de
|
feat(ui/model_dashboard.tsx): show if model is config or db model
|
2024-05-07 21:34:18 -07:00 |
|
Krrish Dholakia
|
5a16bec6a1
|
feat(model_dashboard.tsx): allow user to edit input cost per token for model on ui
also contains fixes for `/model/update`
|
2024-05-07 20:57:21 -07:00 |
|
Krrish Dholakia
|
312249ca44
|
feat(ui/model_dashboard.tsx): show if model is config or db model
|
2024-05-07 18:29:14 -07:00 |
|
Ishaan Jaff
|
dc74204427
|
fix typo
|
2024-05-07 18:27:49 -07:00 |
|
Ishaan Jaff
|
c08352a0ce
|
router- initialize alerting
|
2024-05-07 18:03:04 -07:00 |
|
Ishaan Jaff
|
5fd3b12d34
|
add router alerting type
|
2024-05-07 17:46:18 -07:00 |
|
Ishaan Jaff
|
6a056c0c17
|
fix support InternalServerErrorRetries
|
2024-05-06 17:59:38 -07:00 |
|
Ishaan Jaff
|
116e555a03
|
fix - setting model_group_retry_policy in db
|
2024-05-06 17:43:43 -07:00 |
|
Krish Dholakia
|
6be20f5fc6
|
Merge pull request #3455 from BerriAI/litellm_assistants_support
feat(openai.py): add support for openai assistants
|
2024-05-04 22:21:44 -07:00 |
|
Krrish Dholakia
|
06ae584473
|
fix(types/openai.py): fix python3.8 typing issue
|
2024-05-04 22:04:17 -07:00 |
|
Krrish Dholakia
|
66129bc921
|
fix(typing/openai.py): fix openai typing error (version-related)
|
2024-05-04 22:02:43 -07:00 |
|
Krrish Dholakia
|
2deac08ff1
|
fix(types/openai.py): fix typing import
|
2024-05-04 21:53:08 -07:00 |
|
Krrish Dholakia
|
681a95e37b
|
fix(assistants/main.py): support litellm.create_thread() call
|
2024-05-04 19:35:37 -07:00 |
|
Ishaan Jaff
|
87e165e413
|
Merge pull request #3456 from BerriAI/litellm_router_set_retry_policy_errors
[FEAT] router set custom num retries for ContentPolicyViolationErrorRetries, RateLimitErrorRetries, BadRequestErrorRetries etc
|
2024-05-04 18:26:03 -07:00 |
|
Ishaan Jaff
|
6d1981fbaa
|
init router retry policy
|
2024-05-04 16:59:14 -07:00 |
|
Krish Dholakia
|
601e8a1172
|
Merge pull request #3448 from BerriAI/litellm_anthropic_fix
fix(factory.py): support 'function' openai message role for anthropic
|
2024-05-04 16:36:05 -07:00 |
|
Ishaan Jaff
|
d968dedd77
|
Merge pull request #1530 from TanaroSch/main
change max_tokens type to int
|
2024-05-04 12:47:15 -07:00 |
|
Krrish Dholakia
|
33472bfd2b
|
fix(factory.py): support 'function' openai message role for anthropic
Fixes https://github.com/BerriAI/litellm/issues/3446
|
2024-05-04 10:03:30 -07:00 |
|
Krrish Dholakia
|
8249c986bf
|
fix(main.py): support new 'supports_system_message=False' param
Fixes https://github.com/BerriAI/litellm/issues/3325
|
2024-05-03 21:31:45 -07:00 |
|
Krrish Dholakia
|
5b39f8e282
|
feat(proxy_server.py): return api base in response headers
Closes https://github.com/BerriAI/litellm/issues/2631
|
2024-05-03 15:27:32 -07:00 |
|
Krrish Dholakia
|
f10a066d36
|
fix(lowest_tpm_rpm_v2.py): add more detail to 'No deployments available' error message
|
2024-04-29 15:04:37 -07:00 |
|
Krrish Dholakia
|
0a6b6302f1
|
fix(router.py): fix typing error
|
2024-04-29 07:25:39 -07:00 |
|
CyanideByte
|
e1786848cb
|
protected_namespaces fixed for model_info
|
2024-04-27 13:08:45 -07:00 |
|
Krish Dholakia
|
1a06f009d1
|
Merge branch 'main' into litellm_default_router_retries
|
2024-04-27 11:21:57 -07:00 |
|
Krrish Dholakia
|
7730520fb0
|
fix(router.py): allow passing httpx.timeout to timeout param in router
Closes https://github.com/BerriAI/litellm/issues/3162
|
2024-04-26 14:57:19 -07:00 |
|
Krrish Dholakia
|
160acc085a
|
fix(router.py): fix default retry logic
|
2024-04-25 11:57:27 -07:00 |
|
Krrish Dholakia
|
f54510b6ee
|
fix(proxy_server.py): fix /config/update /
allows updating router config via UI and having the change be propogated across all proxy instances by persisting config changes to the db
|
2024-04-24 16:42:42 -07:00 |
|
Ishaan Jaff
|
3d1a158b63
|
feat - update deployments
|
2024-04-24 09:53:42 -07:00 |
|
Krrish Dholakia
|
0f95a824c4
|
feat(prometheus_services.py): emit proxy latency for successful llm api requests
uses prometheus histogram for this
|
2024-04-18 16:04:35 -07:00 |
|
Krrish Dholakia
|
13cd252f3e
|
fix(proxy_server.py): ensure id used in delete deployment matches id used in litellm Router
|
2024-04-16 15:17:18 -07:00 |
|
Krrish Dholakia
|
4e81acf2c6
|
feat(prometheus_services.py): monitor health of proxy adjacent services (redis / postgres / etc.)
|
2024-04-13 18:15:02 -07:00 |
|
Krrish Dholakia
|
ea1574c160
|
test(test_openai_endpoints.py): add concurrency testing for user defined rate limits on proxy
|
2024-04-12 18:56:13 -07:00 |
|
Krrish Dholakia
|
a47a719caa
|
fix(router.py): generate consistent model id's
having the same id for a deployment, lets redis usage caching work across multiple instances
|
2024-04-10 15:23:57 -07:00 |
|
unclecode
|
5554e2c359
|
Continue fixing the issue #2832: Add protected_namespaces to another to Config class within the router.py
|
2024-04-08 12:49:40 +08:00 |
|
unclecode
|
1ace192155
|
Fix issue #2832: Add protected_namespaces to Config class within utils.py, router.py and completion.py to avoid the warning message.
|
2024-04-08 12:43:17 +08:00 |
|
Krrish Dholakia
|
460546956d
|
fix(utils.py): fix import
|
2024-04-06 18:37:38 -07:00 |
|
TanaroSch
|
b69cb2c773
|
change max_tokens float to int
|
2024-02-06 11:19:28 +01:00 |
|
ishaan-jaff
|
075eb1a516
|
(types) routerConfig
|
2024-01-02 14:14:29 +05:30 |
|
ishaan-jaff
|
9afdc8b4ee
|
(feat) add Router init Pydantic Type
|
2024-01-02 13:30:24 +05:30 |
|
ishaan-jaff
|
1f8fc6d2a7
|
(feat) litellm add types for completion, embedding request
|
2024-01-02 12:27:08 +05:30 |
|