sentence transformers had a subtle dep on torchao

This commit is contained in:
Ashwin Bharambe 2025-08-26 12:28:48 -07:00
parent bd28427d61
commit 104c66f099
2 changed files with 2 additions and 2 deletions

View file

@ -9,7 +9,6 @@ This section contains documentation for all available providers for the **post_t
```{toctree} ```{toctree}
:maxdepth: 1 :maxdepth: 1
inline_huggingface-cpu
inline_huggingface-gpu inline_huggingface-gpu
inline_torchtune-cpu inline_torchtune-cpu
inline_torchtune-gpu inline_torchtune-gpu

View file

@ -40,8 +40,9 @@ def available_providers() -> list[ProviderSpec]:
InlineProviderSpec( InlineProviderSpec(
api=Api.inference, api=Api.inference,
provider_type="inline::sentence-transformers", provider_type="inline::sentence-transformers",
# CrossEncoder depends on torchao.quantization
pip_packages=[ pip_packages=[
"torch torchvision --index-url https://download.pytorch.org/whl/cpu", "torch torchvision torchao --index-url https://download.pytorch.org/whl/cpu",
"sentence-transformers --no-deps", "sentence-transformers --no-deps",
], ],
module="llama_stack.providers.inline.inference.sentence_transformers", module="llama_stack.providers.inline.inference.sentence_transformers",