forked from phoenix-oss/llama-stack-mirror
docs: Update docs and fix warning in start-stack.sh (#1937)
Small docs update and an update for `start-stack.sh` with missing color and if statment logic. # What does this PR do? 1. Makes a small change to start-stack.sh to resolve this error: ```cmd /home/aireilly/.local/lib/python3.13/site-packages/llama_stack/distribution/start_stack.sh: line 76: [: missing ]' ``` 2. Adds a missing $GREEN colour to start-stack.sh 3. Updated `docs/source/getting_started/detailed_tutorial.md` with some small changes and corrections. ## Test Plan Procedures described in `docs/source/getting_started/detailed_tutorial.md` were verified on Linux Fedora 41.
This commit is contained in:
parent
ed58a94b30
commit
51492bd9b6
2 changed files with 5 additions and 4 deletions
|
@ -69,7 +69,7 @@ which defines the providers and their settings.
|
||||||
Now let's build and run the Llama Stack config for Ollama.
|
Now let's build and run the Llama Stack config for Ollama.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
INFERENCE_MODEL=llama3.2:3b llama stack build --template ollama --image-type conda --run
|
INFERENCE_MODEL=llama3.2:3b llama stack build --template ollama --image-type conda --image-name llama3-3b-conda --run
|
||||||
```
|
```
|
||||||
:::
|
:::
|
||||||
:::{tab-item} Using a Container
|
:::{tab-item} Using a Container
|
||||||
|
@ -77,10 +77,9 @@ You can use a container image to run the Llama Stack server. We provide several
|
||||||
component that works with different inference providers out of the box. For this guide, we will use
|
component that works with different inference providers out of the box. For this guide, we will use
|
||||||
`llamastack/distribution-ollama` as the container image. If you'd like to build your own image or customize the
|
`llamastack/distribution-ollama` as the container image. If you'd like to build your own image or customize the
|
||||||
configurations, please check out [this guide](../references/index.md).
|
configurations, please check out [this guide](../references/index.md).
|
||||||
|
|
||||||
First lets setup some environment variables and create a local directory to mount into the container’s file system.
|
First lets setup some environment variables and create a local directory to mount into the container’s file system.
|
||||||
```bash
|
```bash
|
||||||
export INFERENCE_MODEL="meta-llama/Llama-3.2-3B-Instruct"
|
export INFERENCE_MODEL="llama3.2:3b"
|
||||||
export LLAMA_STACK_PORT=8321
|
export LLAMA_STACK_PORT=8321
|
||||||
mkdir -p ~/.llama
|
mkdir -p ~/.llama
|
||||||
```
|
```
|
||||||
|
@ -223,6 +222,7 @@ Other SDKs are also available, please refer to the [Client SDK](../index.md#clie
|
||||||
Now you can run inference using the Llama Stack client SDK.
|
Now you can run inference using the Llama Stack client SDK.
|
||||||
|
|
||||||
### i. Create the Script
|
### i. Create the Script
|
||||||
|
|
||||||
Create a file `inference.py` and add the following code:
|
Create a file `inference.py` and add the following code:
|
||||||
```python
|
```python
|
||||||
from llama_stack_client import LlamaStackClient
|
from llama_stack_client import LlamaStackClient
|
||||||
|
|
|
@ -18,6 +18,7 @@ VIRTUAL_ENV=${VIRTUAL_ENV:-}
|
||||||
set -euo pipefail
|
set -euo pipefail
|
||||||
|
|
||||||
RED='\033[0;31m'
|
RED='\033[0;31m'
|
||||||
|
GREEN='\033[0;32m'
|
||||||
NC='\033[0m' # No Color
|
NC='\033[0m' # No Color
|
||||||
|
|
||||||
error_handler() {
|
error_handler() {
|
||||||
|
@ -73,7 +74,7 @@ done
|
||||||
PYTHON_BINARY="python"
|
PYTHON_BINARY="python"
|
||||||
case "$env_type" in
|
case "$env_type" in
|
||||||
"venv")
|
"venv")
|
||||||
if [ -n "$VIRTUAL_ENV" && "$VIRTUAL_ENV" == "$env_path_or_name" ]; then
|
if [ -n "$VIRTUAL_ENV" ] && [ "$VIRTUAL_ENV" == "$env_path_or_name" ]; then
|
||||||
echo -e "${GREEN}Virtual environment already activated${NC}" >&2
|
echo -e "${GREEN}Virtual environment already activated${NC}" >&2
|
||||||
else
|
else
|
||||||
# Activate virtual environment
|
# Activate virtual environment
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue