mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-18 07:18:53 +00:00
release notes
This commit is contained in:
parent
9c0dac5832
commit
c7f87fcbe7
5 changed files with 5 additions and 4 deletions
|
@ -90,7 +90,7 @@ inference:
|
|||
url: http://127.0.0.1:14343
|
||||
```
|
||||
|
||||
### Model Serving
|
||||
### (Optional) Update Model Serving Configuration
|
||||
|
||||
#### Downloading model via Ollama
|
||||
|
||||
|
|
|
@ -88,7 +88,7 @@ inference:
|
|||
```
|
||||
|
||||
|
||||
### Model Serving
|
||||
### (Optional) Update Model Serving Configuration
|
||||
To serve a new model with `tgi`, change the docker command flag `--model-id <model-to-serve>`.
|
||||
|
||||
This can be done by edit the `command` args in `compose.yaml`. E.g. Replace "Llama-3.2-1B-Instruct" with the model you want to serve.
|
||||
|
|
|
@ -38,7 +38,7 @@ llama stack build --template together --image-type conda
|
|||
llama stack run ./run.yaml
|
||||
```
|
||||
|
||||
### Model Serving
|
||||
### (Optional) Update Model Serving Configuration
|
||||
|
||||
Use `llama-stack-client models list` to check the available models served by together.
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue