mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-17 18:12:43 +00:00
skip openai_client tests if no server is running
This commit is contained in:
parent
6f91fb12a7
commit
a01596a5fe
1 changed files with 15 additions and 1 deletions
|
|
@ -10,6 +10,7 @@ from unittest.mock import patch
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from llama_stack.core.datatypes import User
|
from llama_stack.core.datatypes import User
|
||||||
|
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
||||||
|
|
||||||
|
|
||||||
# a fixture to skip all these tests if a files provider is not available
|
# a fixture to skip all these tests if a files provider is not available
|
||||||
|
|
@ -19,7 +20,20 @@ def skip_if_no_files_provider(llama_stack_client):
|
||||||
pytest.skip("No files providers found")
|
pytest.skip("No files providers found")
|
||||||
|
|
||||||
|
|
||||||
def test_openai_client_basic_operations(openai_client):
|
@pytest.fixture(scope="session")
|
||||||
|
def skip_if_no_server_running(llama_stack_client):
|
||||||
|
"""
|
||||||
|
Skip test if no server is running.
|
||||||
|
|
||||||
|
We use the llama_stack_client to tell if a server was started or not.
|
||||||
|
|
||||||
|
We use this with openai_client because it relies on a running server.
|
||||||
|
"""
|
||||||
|
if isinstance(llama_stack_client, LlamaStackAsLibraryClient):
|
||||||
|
pytest.skip("No server running")
|
||||||
|
|
||||||
|
|
||||||
|
def test_openai_client_basic_operations(openai_client, skip_if_no_server_running):
|
||||||
"""Test basic file operations through OpenAI client."""
|
"""Test basic file operations through OpenAI client."""
|
||||||
from openai import NotFoundError
|
from openai import NotFoundError
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue