llama-stack/llama_stack/providers/inline
Dinesh Yeduguru 516e1a3e59
add embedding model by default to distribution templates (#617)
# What does this PR do?
Adds the sentence transformer provider and the `all-MiniLM-L6-v2`
embedding model to the default models to register in the run.yaml for
all providers.

## Test Plan
llama stack build --template together --image-type conda
llama stack run
~/.llama/distributions/llamastack-together/together-run.yaml
2024-12-13 12:48:00 -08:00
..
agents add tracing back to the lib cli (#595) 2024-12-11 08:44:20 -08:00
datasetio Telemetry API redesign (#525) 2024-12-04 11:22:45 -08:00
eval Add ability to query and export spans to dataset (#574) 2024-12-05 21:07:30 -08:00
inference add embedding model by default to distribution templates (#617) 2024-12-13 12:48:00 -08:00
ios/inference impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00
memory Make embedding generation go through inference (#606) 2024-12-12 11:47:50 -08:00
meta_reference Telemetry API redesign (#525) 2024-12-04 11:22:45 -08:00
post_training/torchtune [1/n] torchtune <> llama-stack integration skeleton (#540) 2024-12-13 11:05:35 -08:00
safety use logging instead of prints (#499) 2024-11-21 11:32:53 -08:00
scoring [/scoring] add ability to define aggregation functions for scoring functions & refactors (#597) 2024-12-11 10:03:42 -08:00
telemetry add tracing back to the lib cli (#595) 2024-12-11 08:44:20 -08:00
__init__.py impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00