mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-16 21:42:38 +00:00
50 lines
1.5 KiB
Python
50 lines
1.5 KiB
Python
# Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
# All rights reserved.
|
|
#
|
|
# This source code is licensed under the terms described in the LICENSE file in
|
|
# the root directory of this source tree.
|
|
|
|
from typing import Tuple
|
|
|
|
import httpx
|
|
|
|
from ._config import NVIDIAConfig
|
|
|
|
|
|
async def _get_health(url: str) -> Tuple[bool, bool]:
|
|
"""
|
|
Query {url}/v1/health/{live,ready} to check if the server is running and ready
|
|
|
|
Args:
|
|
url (str): URL of the server
|
|
|
|
Returns:
|
|
Tuple[bool, bool]: (is_live, is_ready)
|
|
"""
|
|
async with httpx.AsyncClient() as client:
|
|
live = await client.get(f"{url}/v1/health/live")
|
|
ready = await client.get(f"{url}/v1/health/ready")
|
|
return live.status_code == 200, ready.status_code == 200
|
|
|
|
|
|
async def check_health(config: NVIDIAConfig) -> None:
|
|
"""
|
|
Check if the server is running and ready
|
|
|
|
Args:
|
|
url (str): URL of the server
|
|
|
|
Raises:
|
|
RuntimeError: If the server is not running or ready
|
|
"""
|
|
if not config.is_hosted:
|
|
print("Checking NVIDIA NIM health...")
|
|
try:
|
|
is_live, is_ready = await _get_health(config.url)
|
|
if not is_live:
|
|
raise ConnectionError("NVIDIA NIM is not running")
|
|
if not is_ready:
|
|
raise ConnectionError("NVIDIA NIM is not ready")
|
|
# TODO(mf): should we wait for the server to be ready?
|
|
except httpx.ConnectError as e:
|
|
raise ConnectionError(f"Failed to connect to NVIDIA NIM: {e}") from e
|