forked from phoenix/litellm-mirror
* docs 1k rps load test * docs load testing * docs load testing litellm * docs load testing * clean up load test doc * docs prom metrics for load testing * docs using prometheus on load testing * doc load testing with prometheus
1.4 KiB
1.4 KiB
import Image from '@theme/IdealImage';
LiteLLM Proxy - Locust Load Test
Locust Load Test LiteLLM Proxy
- Add
fake-openai-endpoint
to your proxy config.yaml and start your litellm proxy litellm provides a free hostedfake-openai-endpoint
you can load test against
model_list:
- model_name: fake-openai-endpoint
litellm_params:
model: openai/fake
api_key: fake-key
api_base: https://exampleopenaiendpoint-production.up.railway.app/
-
pip install locust
-
Create a file called
locustfile.py
on your local machine. Copy the contents from the litellm load test located here -
Start locust Run
locust
in the same directory as yourlocustfile.py
from step 2
locust
Output on terminal
[2024-03-15 07:19:58,893] Starting web interface at http://0.0.0.0:8089
[2024-03-15 07:19:58,898] Starting Locust 2.24.0
- Run Load test on locust
Head to the locust UI on http://0.0.0.0:8089
Set Users=100, Ramp Up Users=10, Host=Base URL of your LiteLLM Proxy
<Image img={require('../img/locust_load_test.png')} />
- Expected Results
Expect to see the following response times for /health/readiness
Median → /health/readiness is 150ms
Avg → /health/readiness is 219ms
<Image img={require('../img/litellm_load_test.png')} />