llama-stack-mirror/llama_stack/providers/remote/inference/vllm
Akram Ben Aissi 67728bfccf Update vLLM health check to use /health endpoint
- Replace models.list() call with HTTP GET to /health endpoint
- Remove API token validation since /health is unauthenticated
- Use urllib.parse.urljoin for cleaner URL construction
- Update tests to mock httpx.AsyncClient instead of OpenAI client
- Health check now works regardless of API token configuration

Signed-off-by: Akram Ben Aissi <akram.benaissi@gmail.com>
2025-09-15 17:57:17 +02:00
..
__init__.py Add dynamic authentication token forwarding support for vLLM provider 2025-09-15 13:01:12 +01:00
config.py feat(registry): make the Stack query providers for model listing (#2862) 2025-07-24 10:39:53 -07:00
vllm.py Update vLLM health check to use /health endpoint 2025-09-15 17:57:17 +02:00