mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-02 20:40:36 +00:00
fix pre-commit issues
This commit is contained in:
parent
ae85dd6182
commit
afa9db5a6b
2 changed files with 4 additions and 10 deletions
|
@ -653,7 +653,7 @@ async def test_health_status_success(vllm_inference_adapter):
|
|||
# Mock the client.models.list method to return successfully
|
||||
# Set vllm_inference_adapter.client to None to ensure _create_client is called
|
||||
vllm_inference_adapter.client = None
|
||||
with patch.object(vllm_inference_adapter, '_create_client') as mock_create_client:
|
||||
with patch.object(vllm_inference_adapter, "_create_client") as mock_create_client:
|
||||
# Create mock client and models
|
||||
mock_client = MagicMock()
|
||||
mock_models = MagicMock()
|
||||
|
@ -678,7 +678,7 @@ async def test_health_status_failure(vllm_inference_adapter):
|
|||
"""
|
||||
vllm_inference_adapter.client = None
|
||||
# Mock the client.models.list method to raise an exception
|
||||
with patch.object(vllm_inference_adapter, '_create_client') as mock_create_client:
|
||||
with patch.object(vllm_inference_adapter, "_create_client") as mock_create_client:
|
||||
# Create mock client and models
|
||||
mock_client = MagicMock()
|
||||
mock_models = MagicMock()
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue