mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-17 18:12:43 +00:00
skip openai_client tests if no server is running
This commit is contained in:
parent
6f91fb12a7
commit
a01596a5fe
1 changed files with 15 additions and 1 deletions
|
|
@ -10,6 +10,7 @@ from unittest.mock import patch
|
|||
import pytest
|
||||
|
||||
from llama_stack.core.datatypes import User
|
||||
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
||||
|
||||
|
||||
# a fixture to skip all these tests if a files provider is not available
|
||||
|
|
@ -19,7 +20,20 @@ def skip_if_no_files_provider(llama_stack_client):
|
|||
pytest.skip("No files providers found")
|
||||
|
||||
|
||||
def test_openai_client_basic_operations(openai_client):
|
||||
@pytest.fixture(scope="session")
|
||||
def skip_if_no_server_running(llama_stack_client):
|
||||
"""
|
||||
Skip test if no server is running.
|
||||
|
||||
We use the llama_stack_client to tell if a server was started or not.
|
||||
|
||||
We use this with openai_client because it relies on a running server.
|
||||
"""
|
||||
if isinstance(llama_stack_client, LlamaStackAsLibraryClient):
|
||||
pytest.skip("No server running")
|
||||
|
||||
|
||||
def test_openai_client_basic_operations(openai_client, skip_if_no_server_running):
|
||||
"""Test basic file operations through OpenAI client."""
|
||||
from openai import NotFoundError
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue