test: Add VLLM provider support to integration tests

- Add setup-vllm GitHub action to start VLLM container
- Extend integration test matrix to support both ollama and vllm providers
- Make test setup conditional based on provider type
- Add provider-specific environment variables and configurations

TODO:
  investigate failing tests for vllm provider (safety, post_training and
tool runtime)
  Also need a proper fix for #2713 (tmp fix for this in the first commit
in this PR)

Closes: #1648

Signed-off-by: Derek Higgins <derekh@redhat.com>
This commit is contained in:
Derek Higgins 2025-07-14 11:38:10 +01:00
parent 3e7ea4dd14
commit 7420c1db11
2 changed files with 76 additions and 13 deletions

27
.github/actions/setup-vllm/action.yml vendored Normal file
View file

@ -0,0 +1,27 @@
name: Setup VLLM
description: Start VLLM
runs:
using: "composite"
steps:
- name: Start VLLM
shell: bash
run: |
# Start vllm container
docker run -d \
--name vllm \
-p 8000:8000 \
--privileged=true \
quay.io/higginsd/vllm-cpu:65393ee064 \
--host 0.0.0.0 \
--port 8000 \
--enable-auto-tool-choice \
--tool-call-parser llama3_json \
--model /root/.cache/Llama-3.2-1B-Instruct \
--served-model-name meta-llama/Llama-3.2-1B-Instruct
# Wait for vllm to be ready
echo "Waiting for vllm to be ready..."
timeout 900 bash -c 'until curl -f http://localhost:8000/health; do
echo "Waiting for vllm..."
sleep 5
done'