forked from phoenix/litellm-mirror
* docs 1k rps load test * docs load testing * docs load testing litellm * docs load testing * clean up load test doc * docs prom metrics for load testing * docs using prometheus on load testing * doc load testing with prometheus
52 lines
1.4 KiB
Markdown
52 lines
1.4 KiB
Markdown
import Image from '@theme/IdealImage';
|
|
|
|
# LiteLLM Proxy - Locust Load Test
|
|
|
|
## Locust Load Test LiteLLM Proxy
|
|
|
|
1. Add `fake-openai-endpoint` to your proxy config.yaml and start your litellm proxy
|
|
litellm provides a free hosted `fake-openai-endpoint` you can load test against
|
|
|
|
```yaml
|
|
model_list:
|
|
- model_name: fake-openai-endpoint
|
|
litellm_params:
|
|
model: openai/fake
|
|
api_key: fake-key
|
|
api_base: https://exampleopenaiendpoint-production.up.railway.app/
|
|
```
|
|
|
|
2. `pip install locust`
|
|
|
|
3. Create a file called `locustfile.py` on your local machine. Copy the contents from the litellm load test located [here](https://github.com/BerriAI/litellm/blob/main/.github/workflows/locustfile.py)
|
|
|
|
4. Start locust
|
|
Run `locust` in the same directory as your `locustfile.py` from step 2
|
|
|
|
```shell
|
|
locust
|
|
```
|
|
|
|
Output on terminal
|
|
```
|
|
[2024-03-15 07:19:58,893] Starting web interface at http://0.0.0.0:8089
|
|
[2024-03-15 07:19:58,898] Starting Locust 2.24.0
|
|
```
|
|
|
|
5. Run Load test on locust
|
|
|
|
Head to the locust UI on http://0.0.0.0:8089
|
|
|
|
Set Users=100, Ramp Up Users=10, Host=Base URL of your LiteLLM Proxy
|
|
|
|
<Image img={require('../img/locust_load_test.png')} />
|
|
|
|
6. Expected Results
|
|
|
|
Expect to see the following response times for `/health/readiness`
|
|
Median → /health/readiness is `150ms`
|
|
|
|
Avg → /health/readiness is `219ms`
|
|
|
|
<Image img={require('../img/litellm_load_test.png')} />
|
|
|