From 2b1620f8d8cc3a38800ee18e5c9b415d0600ba3e Mon Sep 17 00:00:00 2001 From: Yuan Tang Date: Thu, 17 Apr 2025 20:29:52 -0400 Subject: [PATCH] regenerate Signed-off-by: Yuan Tang --- docs/source/distributions/self_hosted_distro/remote-vllm.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/source/distributions/self_hosted_distro/remote-vllm.md b/docs/source/distributions/self_hosted_distro/remote-vllm.md index efa443778..4405605ea 100644 --- a/docs/source/distributions/self_hosted_distro/remote-vllm.md +++ b/docs/source/distributions/self_hosted_distro/remote-vllm.md @@ -44,7 +44,7 @@ The following environment variables can be configured: In the following sections, we'll use AMD, NVIDIA or Intel GPUs to serve as hardware accelerators for the vLLM server, which acts as both the LLM inference provider and the safety provider. Note that vLLM also [supports many other hardware accelerators](https://docs.vllm.ai/en/latest/getting_started/installation.html) and -that we only use GPUs here for demonstration purposes. +that we only use GPUs here for demonstration purposes. Note that if you are running into issues, there's a new environment variable `VLLM_DEBUG_LOG_API_SERVER_RESPONSE` (available in vLLM v0.8.3 and above) to enable log response from API server for debugging. ### Setting up vLLM server on AMD GPU