forked from phoenix/litellm-mirror
(docs) best practices for high traffic
This commit is contained in:
parent
c50b0e315a
commit
547f0a023d
1 changed files with 4 additions and 0 deletions
|
@ -183,6 +183,10 @@ Your OpenAI proxy server is now running on `http://127.0.0.1:8000`.
|
||||||
</TabItem>
|
</TabItem>
|
||||||
</Tabs>
|
</Tabs>
|
||||||
|
|
||||||
|
## Best Practices for Deploying to Production
|
||||||
|
### 1. Switch of debug logs in production
|
||||||
|
don't use [`--detailed-debug`, `--debug`](https://docs.litellm.ai/docs/proxy/debugging#detailed-debug) or `litellm.set_verbose=True`. We found using debug logs can add 5-10% latency per LLM API call
|
||||||
|
|
||||||
## Advanced Deployment Settings
|
## Advanced Deployment Settings
|
||||||
|
|
||||||
### Customization of the server root path
|
### Customization of the server root path
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue