mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-06 02:32:40 +00:00
update getting started guide to use ollama pull
This commit is contained in:
parent
66d6c2580e
commit
9f3c1ed545
1 changed files with 3 additions and 3 deletions
|
@ -8,13 +8,13 @@ In Llama Stack, we provide a server exposing multiple APIs. These APIs are backe
|
||||||
Ollama is an LLM runtime that allows you to run Llama models locally.
|
Ollama is an LLM runtime that allows you to run Llama models locally.
|
||||||
|
|
||||||
|
|
||||||
### 1. Start Ollama
|
### 1. Download a Llama model with Ollama
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
ollama run llama3.2:3b-instruct-fp16 --keepalive 60m
|
ollama pull llama3.2:3b-instruct-fp16
|
||||||
```
|
```
|
||||||
|
|
||||||
By default, Ollama keeps the model loaded in memory for 5 minutes which can be too short. We set the `--keepalive` flag to 60 minutes to ensure the model remains loaded for sometime.
|
This will instruct the Ollama service to download the Llama 3.2 3B Instruct model, which we'll use in the rest of this guide.
|
||||||
|
|
||||||
```{admonition} Note
|
```{admonition} Note
|
||||||
:class: tip
|
:class: tip
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue