mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-15 14:08:00 +00:00
pre-commit
This commit is contained in:
parent
f56602de27
commit
2784b147aa
4 changed files with 4 additions and 3 deletions
|
@ -195,4 +195,4 @@ runs:
|
|||
name: ${{ inputs.inference-mode }}-logs-${{ github.run_id }}-${{ github.run_attempt || '' }}-${{ inputs.provider }}
|
||||
path: |
|
||||
*.log
|
||||
retention-days: 1
|
||||
retention-days: 1
|
||||
|
|
|
@ -45,4 +45,4 @@ runs:
|
|||
shell: bash
|
||||
run: |
|
||||
git config --local user.email "github-actions[bot]@users.noreply.github.com"
|
||||
git config --local user.name "github-actions[bot]"
|
||||
git config --local user.name "github-actions[bot]"
|
||||
|
|
1
.github/workflows/README.md
vendored
1
.github/workflows/README.md
vendored
|
@ -10,6 +10,7 @@ Llama Stack uses GitHub Actions for Continous Integration (CI). Below is a table
|
|||
| SqlStore Integration Tests | [integration-sql-store-tests.yml](integration-sql-store-tests.yml) | Run the integration test suite with SqlStore |
|
||||
| Integration Tests | [integration-tests.yml](integration-tests.yml) | Run the integration test suite from tests/integration |
|
||||
| Vector IO Integration Tests | [integration-vector-io-tests.yml](integration-vector-io-tests.yml) | Run the integration test suite with various VectorIO providers |
|
||||
| Vision Inference Integration Tests | [integration-vision-tests.yml](integration-vision-tests.yml) | Run vision inference integration test suite from tests/integration/inference |
|
||||
| Pre-commit | [pre-commit.yml](pre-commit.yml) | Run pre-commit checks |
|
||||
| Test Llama Stack Build | [providers-build.yml](providers-build.yml) | Test llama stack build |
|
||||
| Python Package Build Test | [python-build-test.yml](python-build-test.yml) | Test building the llama-stack PyPI project |
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
# Containerfile used to build our Ollama image with vision model to run tests in CI
|
||||
#
|
||||
#
|
||||
# podman build --platform linux/amd64 -f ./ollama-with-vision-model.containerfile -t ollama-with-vision-model .
|
||||
#
|
||||
FROM --platform=linux/amd64 ollama/ollama:latest
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue