mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-05 10:13:05 +00:00
docs: fix errors in kubernetes deployment guide
This commit is contained in:
parent
49955a06b1
commit
6735344604
1 changed files with 9 additions and 3 deletions
|
@ -11,7 +11,12 @@ First, create a local Kubernetes cluster via Kind:
|
||||||
kind create cluster --image kindest/node:v1.32.0 --name llama-stack-test
|
kind create cluster --image kindest/node:v1.32.0 --name llama-stack-test
|
||||||
```
|
```
|
||||||
|
|
||||||
First, create a Kubernetes PVC and Secret for downloading and storing Hugging Face model:
|
First set your hugging face token as an environment variable.
|
||||||
|
```
|
||||||
|
export HF_TOKEN=<your_hf_token>
|
||||||
|
```
|
||||||
|
|
||||||
|
Now create a Kubernetes PVC and Secret for downloading and storing Hugging Face model:
|
||||||
|
|
||||||
```
|
```
|
||||||
cat <<EOF |kubectl apply -f -
|
cat <<EOF |kubectl apply -f -
|
||||||
|
@ -33,7 +38,8 @@ metadata:
|
||||||
name: hf-token-secret
|
name: hf-token-secret
|
||||||
type: Opaque
|
type: Opaque
|
||||||
data:
|
data:
|
||||||
token: $(HF_TOKEN)
|
token: $(echo -n "$HF_TOKEN" | base64)
|
||||||
|
EOF
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
||||||
|
@ -120,7 +126,7 @@ providers:
|
||||||
Once we have defined the run configuration for Llama Stack, we can build an image with that configuration and the server source code:
|
Once we have defined the run configuration for Llama Stack, we can build an image with that configuration and the server source code:
|
||||||
|
|
||||||
```
|
```
|
||||||
cat >/tmp/test-vllm-llama-stack/Containerfile.llama-stack-run-k8s <<EOF
|
mkdir -p /tmp/test-vllm-llama-stack && cat >/tmp/test-vllm-llama-stack/Containerfile.llama-stack-run-k8s <<EOF
|
||||||
FROM distribution-myenv:dev
|
FROM distribution-myenv:dev
|
||||||
|
|
||||||
RUN apt-get update && apt-get install -y git
|
RUN apt-get update && apt-get install -y git
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue