mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
test(test_openai_endpoints.py): add concurrency testing for user defined rate limits on proxy
This commit is contained in:
parent
c03b0bbb24
commit
ea1574c160
6 changed files with 68 additions and 28 deletions
|
@ -194,7 +194,7 @@ async def test_chat_completion():
|
|||
await chat_completion(session=session, key=key_2)
|
||||
|
||||
|
||||
@pytest.mark.skip(reason="Local test. Proxy not concurrency safe yet. WIP.")
|
||||
# @pytest.mark.skip(reason="Local test. Proxy not concurrency safe yet. WIP.")
|
||||
@pytest.mark.asyncio
|
||||
async def test_chat_completion_ratelimit():
|
||||
"""
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue