llama-stack-mirror/scripts
Derek Higgins c62a09ab76
Some checks failed
SqlStore Integration Tests / test-postgres (3.12) (push) Failing after 1s
Integration Auth Tests / test-matrix (oauth2_token) (push) Failing after 2s
SqlStore Integration Tests / test-postgres (3.13) (push) Failing after 1s
Integration Tests (Replay) / generate-matrix (push) Successful in 4s
Python Package Build Test / build (3.13) (push) Failing after 2s
Test External Providers Installed via Module / test-external-providers-from-module (venv) (push) Has been skipped
Vector IO Integration Tests / test-matrix (push) Failing after 6s
Pre-commit / pre-commit (push) Failing after 6s
Test External API and Providers / test-external (venv) (push) Failing after 5s
API Conformance Tests / check-schema-compatibility (push) Successful in 14s
Integration Tests (Replay) / Integration Tests (, , , client=, ) (push) Failing after 5s
Python Package Build Test / build (3.12) (push) Failing after 22s
UI Tests / ui-tests (22) (push) Successful in 57s
ci: Add vLLM support to integration testing infrastructure (with qwen) (#3545)
o Introduces vLLM provider support to the record/replay testing
framework
o Enabling both recording and replay of vLLM API interactions alongside
existing Ollama support.

The changes enable testing of vLLM functionality. vLLM tests focus on
inference capabilities, while Ollama continues to exercise the full API
surface
including vision features.

--
This is an alternative to #3128 , using qwen3 instead of llama 3.2 1B
appears to be more capable at structure output and tool calls.

---------

Signed-off-by: Derek Higgins <derekh@redhat.com>
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-11-06 10:36:40 +01:00
..
github fix: unbound variable PR_HEAD_REPO (#3469) 2025-09-17 10:18:43 +02:00
telemetry chore(telemetry): add grafana dashboards (#3921) 2025-10-27 14:58:27 -07:00
check-init-py.sh chore(package): migrate to src/ layout (#3920) 2025-10-27 12:02:21 -07:00
check-workflows-use-hashes.sh fix: update check-workflows-use-hashes to use github error format (#2875) 2025-07-24 17:41:17 +02:00
cleanup_recordings.py chore(ci): remove unused recordings (#4074) 2025-11-05 09:21:58 -08:00
diagnose_recordings.py feat(ci): add support for docker:distro in tests (#3832) 2025-10-16 19:33:13 -07:00
distro_codegen.py chore: remove leftover llama_stack directory (#3940) 2025-10-28 05:09:08 -07:00
docker.sh fix(ci): use test.pypi as extra index for RC dependencies (#4009) 2025-10-31 12:55:43 -07:00
gen-changelog.py chore: enable ruff for ./scripts too (#1643) 2025-03-18 12:17:21 -07:00
gen-ci-docs.py chore: Remove coverage badge from README.md (#2976) 2025-07-31 09:21:30 -07:00
generate_ci_matrix.py chore(ci): remove unused recordings (#4074) 2025-11-05 09:21:58 -08:00
generate_prompt_format.py chore: standardize model not found error (#2964) 2025-07-30 12:19:53 -07:00
get_setup_env.py fix: pre-commit issues: non executable shebang file and removal of @pytest.mark.asyncio decorator (#3397) 2025-09-10 15:27:35 +02:00
install.sh chore(telemetry): add grafana dashboards (#3921) 2025-10-27 14:58:27 -07:00
integration-tests.sh ci: Add vLLM support to integration testing infrastructure (with qwen) (#3545) 2025-11-06 10:36:40 +01:00
normalize_recordings.py fix: update normalize to search all recordings dirs (#3767) 2025-10-10 07:32:14 -07:00
provider_codegen.py feat: use SecretStr for inference provider auth credentials (#3724) 2025-10-10 07:32:50 -07:00
run-ui-linter.sh chore: move src/llama_stack/ui to src/llama_stack_ui (#4068) 2025-11-04 15:21:49 -08:00
unit-tests.sh chore(package): migrate to src/ layout (#3920) 2025-10-27 12:02:21 -07:00
uv-run-with-index.sh fix(ci): use test.pypi as extra index for RC dependencies (#4009) 2025-10-31 12:55:43 -07:00