From 2eda050aef2e33272be08e41f9f9adea76777d28 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 20 Feb 2025 11:46:02 -0800 Subject: [PATCH] Fix ollama fixture --- llama_stack/providers/tests/inference/fixtures.py | 8 ++------ 1 file changed, 2 insertions(+), 6 deletions(-) diff --git a/llama_stack/providers/tests/inference/fixtures.py b/llama_stack/providers/tests/inference/fixtures.py index 2a782befc..ec4e094c9 100644 --- a/llama_stack/providers/tests/inference/fixtures.py +++ b/llama_stack/providers/tests/inference/fixtures.py @@ -83,17 +83,13 @@ def inference_cerebras() -> ProviderFixture: @pytest.fixture(scope="session") -def inference_ollama(inference_model) -> ProviderFixture: - inference_model = [inference_model] if isinstance(inference_model, str) else inference_model - if inference_model and "Llama3.1-8B-Instruct" in inference_model: - pytest.skip("Ollama only supports Llama3.2-3B-Instruct for testing") - +def inference_ollama() -> ProviderFixture: return ProviderFixture( providers=[ Provider( provider_id="ollama", provider_type="remote::ollama", - config=OllamaImplConfig(host="localhost", port=os.getenv("OLLAMA_PORT", 11434)).model_dump(), + config=OllamaImplConfig(url=get_env_or_fail("OLLAMA_URL")).model_dump(), ) ], )