llama-stack/docs/source
Ashwin Bharambe 11697f85c5
fix: pull ollama embedding model if necessary (#1209)
Embedding models are tiny and can be pulled on-demand. Let's do that so
the user doesn't have to do "yet another thing" to get themselves set
up.

Thanks @hardikjshah for the suggestion.

Also fixed a build dependency miss (TODO: distro_codegen needs to
actually check that the build template contains all providers mentioned
for the run.yaml file)

## Test Plan 

First run `ollama rm all-minilm:latest`. 

Run `llama stack build --template ollama && llama stack run ollama --env
INFERENCE_MODEL=llama3.2:3b-instruct-fp16`. See that it outputs a
"Pulling embedding model `all-minilm:latest`" output and the stack
starts up correctly. Verify that `ollama list` shows the model is
correctly downloaded.
2025-02-21 10:35:56 -08:00
..
building_applications docs: Fix Links, Add Podman Instructions, Vector DB Unregister, and Example Script (#1129) 2025-02-20 13:52:14 -08:00
concepts script for running client sdk tests (#895) 2025-02-19 22:38:06 -08:00
contributing docs: Add note about distro_codegen.py and provider dependencies (#1175) 2025-02-20 09:23:46 -08:00
distributions fix: pull ollama embedding model if necessary (#1209) 2025-02-21 10:35:56 -08:00
getting_started fix(rag-example): add provider_id to avoid llama_stack_client 400 error (#1114) 2025-02-19 15:37:25 -08:00
introduction Update index.md (#888) 2025-01-28 04:55:41 -08:00
playground fix!: update eval-tasks -> benchmarks (#1032) 2025-02-13 16:40:58 -08:00
references feat: model remove cmd (#1128) 2025-02-21 08:05:12 -08:00
conf.py Try a different css file API for sphinx 2025-02-19 20:14:40 -08:00
index.md Update index.md to refer to v0.1.3 2025-02-14 14:29:17 -08:00