Krrish Dholakia
|
3560f0ef2c
|
refactor: move all testing to top-level of repo
Closes https://github.com/BerriAI/litellm/issues/486
|
2024-09-28 21:08:14 -07:00 |
|
ishaan-jaff
|
c59021d090
|
(cookbook) load test litellm router
|
2024-02-08 07:24:28 -08:00 |
|
Krrish Dholakia
|
6e68cd1125
|
docs(load_test.md): add litellm load test script to docs
|
2023-12-29 13:41:44 +05:30 |
|
Krrish Dholakia
|
4905929de3
|
refactor: add black formatting
|
2023-12-25 14:11:20 +05:30 |
|
Krrish Dholakia
|
3d45d8a58c
|
test: load test router
|
2023-11-17 08:23:44 -08:00 |
|
Krrish Dholakia
|
a94c09c13c
|
fix(openai.py): handling extra headers
|
2023-11-16 12:48:21 -08:00 |
|
Krrish Dholakia
|
f582189cea
|
test(loadtest_router.py): commenting out of ci/cd
|
2023-11-16 12:17:25 -08:00 |
|
Krrish Dholakia
|
bf0f8b824c
|
fix(azure.py): use openai client sdk for handling sync+async calling
|
2023-11-16 12:08:12 -08:00 |
|
Krrish Dholakia
|
59eaeba92a
|
fix(router.py): enabling retrying with expo backoff (without tenacity) for router
|
2023-11-14 20:57:51 -08:00 |
|
Krrish Dholakia
|
1738341dcb
|
fix(main.py): misrouting ollama models to nlp cloud
|
2023-11-14 18:55:08 -08:00 |
|
Krrish Dholakia
|
a7222f257c
|
fix(main.py): keep client consistent across calls + exponential backoff retry on ratelimit errors
|
2023-11-14 16:26:05 -08:00 |
|