release notes

This commit is contained in:
Xi Yan 2024-11-02 17:15:07 -07:00
parent 9c0dac5832
commit c7f87fcbe7
5 changed files with 5 additions and 4 deletions

View file

@ -90,7 +90,7 @@ inference:
url: http://127.0.0.1:14343
```
### Model Serving
### (Optional) Update Model Serving Configuration
#### Downloading model via Ollama

View file

@ -88,7 +88,7 @@ inference:
```
### Model Serving
### (Optional) Update Model Serving Configuration
To serve a new model with `tgi`, change the docker command flag `--model-id <model-to-serve>`.
This can be done by edit the `command` args in `compose.yaml`. E.g. Replace "Llama-3.2-1B-Instruct" with the model you want to serve.

View file

@ -38,7 +38,7 @@ llama stack build --template together --image-type conda
llama stack run ./run.yaml
```
### Model Serving
### (Optional) Update Model Serving Configuration
Use `llama-stack-client models list` to check the available models served by together.