From 6d6b40983eeea0283fd6e86e3a305e28ba560937 Mon Sep 17 00:00:00 2001 From: Matthew Farrellee Date: Mon, 14 Apr 2025 06:17:51 -0400 Subject: [PATCH] refactor: update integration test workflow (#1856) workflow - 0. Checkout 1. Install uv 2. Install Ollama 3. Pull Ollama image 4. Start Ollama in background 5. Set Up Environment and Install Dependencies 6. Wait for Ollama to start 7. Start Llama Stack server in background 8. Wait for Llama Stack server to be ready 9. Run Integration Tests changes - (4) starts the loading of the ollama model, it does not start ollama. the model will be loaded when used. this step is removed. (6) is handled in (2). this step is removed. (2) is renamed to reflect it's dual purpose. --- .github/workflows/integration-tests.yml | 23 +++-------------------- 1 file changed, 3 insertions(+), 20 deletions(-) diff --git a/.github/workflows/integration-tests.yml b/.github/workflows/integration-tests.yml index c61712bfd..5a7b35e17 100644 --- a/.github/workflows/integration-tests.yml +++ b/.github/workflows/integration-tests.yml @@ -38,18 +38,16 @@ jobs: with: python-version: "3.10" - - name: Install Ollama + - name: Install and start Ollama run: | + # the ollama installer also starts the ollama service curl -fsSL https://ollama.com/install.sh | sh - name: Pull Ollama image run: | + # TODO: cache the model. OLLAMA_MODELS defaults to ~ollama/.ollama/models. ollama pull llama3.2:3b-instruct-fp16 - - name: Start Ollama in background - run: | - nohup ollama run llama3.2:3b-instruct-fp16 > ollama.log 2>&1 & - - name: Set Up Environment and Install Dependencies run: | uv sync --extra dev --extra test @@ -61,21 +59,6 @@ jobs: uv pip install -e . llama stack build --template ollama --image-type venv - - name: Wait for Ollama to start - run: | - echo "Waiting for Ollama..." - for i in {1..30}; do - if curl -s http://localhost:11434 | grep -q "Ollama is running"; then - echo "Ollama is running!" - exit 0 - fi - sleep 1 - done - echo "Ollama failed to start" - ollama ps - ollama.log - exit 1 - - name: Start Llama Stack server in background if: matrix.client-type == 'http' env: