diff --git a/openai_proxy/README.md b/openai_proxy/README.md index 4d5ed3268..cd4fa9ec1 100644 --- a/openai_proxy/README.md +++ b/openai_proxy/README.md @@ -21,6 +21,7 @@ docker run -e PORT=8000 -p 8000:8000 ghcr.io/berriai/litellm:latest ## Endpoints: - `/chat/completions` - chat completions endpoint to call 100+ LLMs +- `/router/completions` - for multiple deployments of the same model (e.g. Azure OpenAI), uses the least used deployment. [Learn more](https://docs.litellm.ai/docs/routing) - `/models` - available models on server ## Making Requests to Proxy