mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-26 14:17:44 +00:00
Instead of downloading the models each time we now have a single Ollama container that is baked with the models pulled and ready to use. This will remove the CI flakiness on model pulling. Signed-off-by: Sébastien Han <seb@redhat.com> |
||
---|---|---|
.. | ||
action.yml |