mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-05 10:13:05 +00:00
update getting started guide to use ollama pull
This commit is contained in:
parent
66d6c2580e
commit
9f3c1ed545
1 changed files with 3 additions and 3 deletions
|
@ -8,13 +8,13 @@ In Llama Stack, we provide a server exposing multiple APIs. These APIs are backe
|
|||
Ollama is an LLM runtime that allows you to run Llama models locally.
|
||||
|
||||
|
||||
### 1. Start Ollama
|
||||
### 1. Download a Llama model with Ollama
|
||||
|
||||
```bash
|
||||
ollama run llama3.2:3b-instruct-fp16 --keepalive 60m
|
||||
ollama pull llama3.2:3b-instruct-fp16
|
||||
```
|
||||
|
||||
By default, Ollama keeps the model loaded in memory for 5 minutes which can be too short. We set the `--keepalive` flag to 60 minutes to ensure the model remains loaded for sometime.
|
||||
This will instruct the Ollama service to download the Llama 3.2 3B Instruct model, which we'll use in the rest of this guide.
|
||||
|
||||
```{admonition} Note
|
||||
:class: tip
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue