mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-12 13:00:39 +00:00
feat(ci): add support for running vision inference tests (#2972)
This PR significantly refactors the Integration Tests workflow. The main goal behind the PR was to enable recording of vision tests which were never run as part of our CI ever before. During debugging, I ended up making several other changes refactoring and hopefully increasing the robustness of the workflow. After doing the experiments, I have updated the trigger event to be `pull_request_target` so this workflow can get write permissions by default but it will run with source code from the base (main) branch in the source repository only. If you do change the workflow, you'd need to experiment using the `workflow_dispatch` triggers. This should not be news to anyone using Github Actions (except me!) It is likely to be a little rocky though while I learn more about GitHub Actions, etc. Please be patient :) --------- Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
This commit is contained in:
parent
709c974bd8
commit
27d866795c
108 changed files with 13985 additions and 15254 deletions
15
tests/containers/ollama-with-models.containerfile
Normal file
15
tests/containers/ollama-with-models.containerfile
Normal file
|
@ -0,0 +1,15 @@
|
|||
# Containerfile used to build our all in one ollama image to run tests in CI
|
||||
#
|
||||
# podman build --platform linux/amd64 -f ./ollama-with-models.containerfile -t ollama-with-models .
|
||||
#
|
||||
FROM --platform=linux/amd64 ollama/ollama:latest
|
||||
|
||||
# Start ollama and pull models in a single layer
|
||||
RUN ollama serve & \
|
||||
sleep 5 && \
|
||||
ollama pull llama3.2:3b-instruct-fp16 && \
|
||||
ollama pull all-minilm:l6-v2 && \
|
||||
ollama pull llama-guard3:1b
|
||||
|
||||
# Set the entrypoint to start ollama serve
|
||||
ENTRYPOINT ["ollama", "serve"]
|
14
tests/containers/ollama-with-vision-model.containerfile
Normal file
14
tests/containers/ollama-with-vision-model.containerfile
Normal file
|
@ -0,0 +1,14 @@
|
|||
# Containerfile used to build our Ollama image with vision model to run tests in CI
|
||||
#
|
||||
# podman build --platform linux/amd64 -f ./ollama-with-vision-model.containerfile -t ollama-with-vision-model .
|
||||
#
|
||||
FROM --platform=linux/amd64 ollama/ollama:latest
|
||||
|
||||
# Start ollama and pull models in a single layer
|
||||
RUN ollama serve & \
|
||||
sleep 5 && \
|
||||
ollama pull llama3.2-vision:11b && \
|
||||
ollama pull all-minilm:l6-v2
|
||||
|
||||
# Set the entrypoint to start ollama serve
|
||||
ENTRYPOINT ["ollama", "serve"]
|
Loading…
Add table
Add a link
Reference in a new issue