llama-stack-mirror/tests/unit/providers/inference
Akram Ben Aissi 67728bfccf Update vLLM health check to use /health endpoint
- Replace models.list() call with HTTP GET to /health endpoint
- Remove API token validation since /health is unauthenticated
- Use urllib.parse.urljoin for cleaner URL construction
- Update tests to mock httpx.AsyncClient instead of OpenAI client
- Health check now works regardless of API token configuration

Signed-off-by: Akram Ben Aissi <akram.benaissi@gmail.com>
2025-09-15 17:57:17 +02:00
..
bedrock fix: use lambda pattern for bedrock config env vars (#3307) 2025-09-05 10:45:11 +02:00
test_inference_client_caching.py chore: update the groq inference impl to use openai-python for openai-compat functions (#3348) 2025-09-06 15:36:27 -07:00
test_litellm_openai_mixin.py feat: Add clear error message when API key is missing (#2992) 2025-07-31 16:33:16 -04:00
test_openai_base_url_config.py chore(rename): move llama_stack.distribution to llama_stack.core (#2975) 2025-07-30 23:30:53 -07:00
test_remote_vllm.py Update vLLM health check to use /health endpoint 2025-09-15 17:57:17 +02:00