mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-24 18:24:20 +00:00
(docs) proxy performance
This commit is contained in:
parent
ad1672b901
commit
111c7afaca
4 changed files with 113 additions and 0 deletions
|
@ -1077,6 +1077,18 @@ Expected output on Langfuse
|
|||
|
||||
<Image img={require('../img/langfuse_small.png')} />
|
||||
|
||||
## LiteLLM Proxy Performance
|
||||
|
||||
### Throughput - 30% Increase
|
||||
LiteLLM proxy + Load Balancer gives **30% increase** in throughput compared to Raw OpenAI API
|
||||
<Image img={require('../img/throughput.png')} />
|
||||
|
||||
### Latency Added - 0.00325 seconds
|
||||
LiteLLM proxy adds **0.00325 seconds** latency as compared to using the Raw OpenAI API
|
||||
<Image img={require('../img/latency.png')} />
|
||||
|
||||
|
||||
|
||||
|
||||
## Proxy CLI Arguments
|
||||
|
||||
|
|
BIN
docs/my-website/img/latency.png
Normal file
BIN
docs/my-website/img/latency.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 66 KiB |
BIN
docs/my-website/img/throughput.png
Normal file
BIN
docs/my-website/img/throughput.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 71 KiB |
Loading…
Add table
Add a link
Reference in a new issue