diff --git a/README.md b/README.md index 4b6ad95b1..009f70da8 100644 --- a/README.md +++ b/README.md @@ -103,6 +103,7 @@ Please checkout our [Documentations](https://llama-stack.readthedocs.io/en/lates * [Jupyter notebook](./docs/getting_started.ipynb) to walk-through how to use simple text and vision inference llama_stack_client APIs * [Contributing](CONTRIBUTING.md) * [Adding a new API Provider](https://llama-stack.readthedocs.io/en/latest/api_providers/new_api_provider.html) to walk-through how to add a new API provider. +* [Release Notes](https://github.com/meta-llama/llama-stack/releases) ## Llama Stack Client SDK diff --git a/distributions/together/README.md b/distributions/together/README.md index 5f9c90071..72d02437a 100644 --- a/distributions/together/README.md +++ b/distributions/together/README.md @@ -41,7 +41,7 @@ llama stack build --template together --image-type conda llama stack run ./run.yaml ``` -### Model Serving +### (Optional) Update Model Serving Configuration Use `llama-stack-client models list` to check the available models served by together. diff --git a/docs/source/getting_started/distributions/ollama.md b/docs/source/getting_started/distributions/ollama.md index e1508f5b0..fa25c9a72 100644 --- a/docs/source/getting_started/distributions/ollama.md +++ b/docs/source/getting_started/distributions/ollama.md @@ -90,7 +90,7 @@ inference: url: http://127.0.0.1:14343 ``` -### Model Serving +### (Optional) Update Model Serving Configuration #### Downloading model via Ollama diff --git a/docs/source/getting_started/distributions/tgi.md b/docs/source/getting_started/distributions/tgi.md index aeb693364..3ee079360 100644 --- a/docs/source/getting_started/distributions/tgi.md +++ b/docs/source/getting_started/distributions/tgi.md @@ -88,7 +88,7 @@ inference: ``` -### Model Serving +### (Optional) Update Model Serving Configuration To serve a new model with `tgi`, change the docker command flag `--model-id `. This can be done by edit the `command` args in `compose.yaml`. E.g. Replace "Llama-3.2-1B-Instruct" with the model you want to serve. diff --git a/docs/source/getting_started/distributions/together.md b/docs/source/getting_started/distributions/together.md index 6a4142361..cf14f3046 100644 --- a/docs/source/getting_started/distributions/together.md +++ b/docs/source/getting_started/distributions/together.md @@ -38,7 +38,7 @@ llama stack build --template together --image-type conda llama stack run ./run.yaml ``` -### Model Serving +### (Optional) Update Model Serving Configuration Use `llama-stack-client models list` to check the available models served by together.