(docs) proxy performance

This commit is contained in:
ishaan-jaff 2023-11-24 17:07:46 -08:00
parent ad1672b901
commit 111c7afaca
4 changed files with 113 additions and 0 deletions

View file

@ -1077,6 +1077,18 @@ Expected output on Langfuse
<Image img={require('../img/langfuse_small.png')} />
## LiteLLM Proxy Performance
### Throughput - 30% Increase
LiteLLM proxy + Load Balancer gives **30% increase** in throughput compared to Raw OpenAI API
<Image img={require('../img/throughput.png')} />
### Latency Added - 0.00325 seconds
LiteLLM proxy adds **0.00325 seconds** latency as compared to using the Raw OpenAI API
<Image img={require('../img/latency.png')} />
## Proxy CLI Arguments

Binary file not shown.

After

Width:  |  Height:  |  Size: 66 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 71 KiB