llama-stack-mirror/llama_stack/providers/inline/inference/meta_reference
Dinesh Yeduguru 516e1a3e59
add embedding model by default to distribution templates (#617)
# What does this PR do?
Adds the sentence transformer provider and the `all-MiniLM-L6-v2`
embedding model to the default models to register in the run.yaml for
all providers.

## Test Plan
llama stack build --template together --image-type conda
llama stack run
~/.llama/distributions/llamastack-together/together-run.yaml
2024-12-13 12:48:00 -08:00
..
quantization use logging instead of prints (#499) 2024-11-21 11:32:53 -08:00
__init__.py Add provider deprecation support; change directory structure (#397) 2024-11-07 13:04:53 -08:00
config.py Integrate distro docs into the restructured docs 2024-11-20 23:20:05 -08:00
generation.py use logging instead of prints (#499) 2024-11-21 11:32:53 -08:00
inference.py add embedding model by default to distribution templates (#617) 2024-12-13 12:48:00 -08:00
model_parallel.py Add provider deprecation support; change directory structure (#397) 2024-11-07 13:04:53 -08:00
parallel_utils.py use logging instead of prints (#499) 2024-11-21 11:32:53 -08:00