release notes

This commit is contained in:
Xi Yan 2024-11-02 17:15:07 -07:00
parent 9c0dac5832
commit c7f87fcbe7
5 changed files with 5 additions and 4 deletions

View file

@ -103,6 +103,7 @@ Please checkout our [Documentations](https://llama-stack.readthedocs.io/en/lates
* [Jupyter notebook](./docs/getting_started.ipynb) to walk-through how to use simple text and vision inference llama_stack_client APIs
* [Contributing](CONTRIBUTING.md)
* [Adding a new API Provider](https://llama-stack.readthedocs.io/en/latest/api_providers/new_api_provider.html) to walk-through how to add a new API provider.
* [Release Notes](https://github.com/meta-llama/llama-stack/releases)
## Llama Stack Client SDK

View file

@ -41,7 +41,7 @@ llama stack build --template together --image-type conda
llama stack run ./run.yaml
```
### Model Serving
### (Optional) Update Model Serving Configuration
Use `llama-stack-client models list` to check the available models served by together.

View file

@ -90,7 +90,7 @@ inference:
url: http://127.0.0.1:14343
```
### Model Serving
### (Optional) Update Model Serving Configuration
#### Downloading model via Ollama

View file

@ -88,7 +88,7 @@ inference:
```
### Model Serving
### (Optional) Update Model Serving Configuration
To serve a new model with `tgi`, change the docker command flag `--model-id <model-to-serve>`.
This can be done by edit the `command` args in `compose.yaml`. E.g. Replace "Llama-3.2-1B-Instruct" with the model you want to serve.

View file

@ -38,7 +38,7 @@ llama stack build --template together --image-type conda
llama stack run ./run.yaml
```
### Model Serving
### (Optional) Update Model Serving Configuration
Use `llama-stack-client models list` to check the available models served by together.