mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-21 12:09:40 +00:00
Introduce a new workflow to run integration and agent tests against the OpenAI provider, and a subset of inference tests against Fireworks AI. Responses from inference providers are cached and reused in subsequent jobs, reducing API quota usage, speeding up test runs, and improving reliability by avoiding server-side errors. (see https://github.com/derekhiggins/cachemeifyoucan for caching code) The cache is updated on successful job completion. To prevent the cache from diverging or growing indefinitely, periodically refresh or manually update the cache by running this job without a cache(will do as a follow on if this merges). Any updates to integration tests that change the request to the provider will use provider quota and update the cache (if they fail), so need to be careful not to repeatedly run failing PR's Signed-off-by: Derek Higgins <derekh@redhat.com> |
||
---|---|---|
.. | ||
changelog.yml | ||
gha_workflow_llama_stack_tests.yml | ||
install-script-ci.yml | ||
integration-auth-tests.yml | ||
integration-tests-experimental.yml | ||
integration-tests.yml | ||
pre-commit.yml | ||
providers-build.yml | ||
semantic-pr.yml | ||
stale_bot.yml | ||
test-external-providers.yml | ||
tests.yml | ||
unit-tests.yml | ||
update-readthedocs.yml |