diff --git a/llama_stack/providers/registry/vector_io.py b/llama_stack/providers/registry/vector_io.py index df7b7f4b3..2d7c02d86 100644 --- a/llama_stack/providers/registry/vector_io.py +++ b/llama_stack/providers/registry/vector_io.py @@ -30,7 +30,7 @@ EMBEDDING_DEPS = [ # we need a better way to do this to identify potential conflicts, etc. # for now, this lets us significantly reduce the size of the container which # does not have any "local" inference code (and hence does not need GPU-enabled torch) - "torch --index-url https://download.pytorch.org/whl/cpu", + "torch torchvision --index-url https://download.pytorch.org/whl/cpu", "sentence-transformers --no-deps", ]