mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-30 07:39:38 +00:00
copy button
This commit is contained in:
parent
441ad707cc
commit
de791b3e4f
4 changed files with 22 additions and 9 deletions
|
@ -4,3 +4,4 @@ linkify
|
||||||
-e git+https://github.com/pytorch/pytorch_sphinx_theme.git#egg=pytorch_sphinx_theme
|
-e git+https://github.com/pytorch/pytorch_sphinx_theme.git#egg=pytorch_sphinx_theme
|
||||||
sphinx-rtd-theme>=1.0.0
|
sphinx-rtd-theme>=1.0.0
|
||||||
sphinx-pdj-theme
|
sphinx-pdj-theme
|
||||||
|
sphinx-copybutton
|
||||||
|
|
|
@ -6,7 +6,7 @@ This guide gives you references to switch between different memory API providers
|
||||||
1. Start running the pgvector server:
|
1. Start running the pgvector server:
|
||||||
|
|
||||||
```
|
```
|
||||||
docker run --network host --name mypostgres -it -p 5432:5432 -e POSTGRES_PASSWORD=mysecretpassword -e POSTGRES_USER=postgres -e POSTGRES_DB=postgres pgvector/pgvector:pg16
|
$ docker run --network host --name mypostgres -it -p 5432:5432 -e POSTGRES_PASSWORD=mysecretpassword -e POSTGRES_USER=postgres -e POSTGRES_DB=postgres pgvector/pgvector:pg16
|
||||||
```
|
```
|
||||||
|
|
||||||
2. Edit the `run.yaml` file to point to the pgvector server.
|
2. Edit the `run.yaml` file to point to the pgvector server.
|
||||||
|
|
|
@ -19,7 +19,7 @@ author = "Meta"
|
||||||
# -- General configuration ---------------------------------------------------
|
# -- General configuration ---------------------------------------------------
|
||||||
# https://www.sphinx-doc.org/en/master/usage/configuration.html#general-configuration
|
# https://www.sphinx-doc.org/en/master/usage/configuration.html#general-configuration
|
||||||
|
|
||||||
extensions = ["myst_parser", "sphinx_rtd_theme"]
|
extensions = ["myst_parser", "sphinx_rtd_theme", "sphinx_copybutton"]
|
||||||
|
|
||||||
html_theme = "sphinx_rtd_theme"
|
html_theme = "sphinx_rtd_theme"
|
||||||
|
|
||||||
|
@ -50,6 +50,18 @@ myst_enable_extensions = [
|
||||||
"tasklist",
|
"tasklist",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
# Copy button settings
|
||||||
|
copybutton_prompt_text = "$ " # for bash prompts
|
||||||
|
copybutton_prompt_is_regexp = True
|
||||||
|
copybutton_remove_prompts = True
|
||||||
|
copybutton_line_continuation_character = "\\"
|
||||||
|
|
||||||
|
# Source suffix
|
||||||
|
source_suffix = {
|
||||||
|
".rst": "restructuredtext",
|
||||||
|
".md": "markdown",
|
||||||
|
}
|
||||||
|
|
||||||
# -- Options for HTML output -------------------------------------------------
|
# -- Options for HTML output -------------------------------------------------
|
||||||
# https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-html-output
|
# https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-html-output
|
||||||
|
|
||||||
|
|
|
@ -61,8 +61,8 @@ If so, we suggest:
|
||||||
### chat_completion sanity test
|
### chat_completion sanity test
|
||||||
Once the server is setup, we can test it with a client to see the example outputs by . This will run the chat completion client and query the distribution’s `/inference/chat_completion` API. Send a POST request to the server:
|
Once the server is setup, we can test it with a client to see the example outputs by . This will run the chat completion client and query the distribution’s `/inference/chat_completion` API. Send a POST request to the server:
|
||||||
|
|
||||||
```
|
```bash
|
||||||
curl http://localhost:5000/inference/chat_completion \
|
$ curl http://localhost:5000/inference/chat_completion \
|
||||||
-H "Content-Type: application/json" \
|
-H "Content-Type: application/json" \
|
||||||
-d '{
|
-d '{
|
||||||
"model": "Llama3.1-8B-Instruct",
|
"model": "Llama3.1-8B-Instruct",
|
||||||
|
@ -86,12 +86,12 @@ Output:
|
||||||
|
|
||||||
To run an agent app, check out examples demo scripts with client SDKs to talk with the Llama Stack server in our [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main/examples) repo. To run a simple agent app:
|
To run an agent app, check out examples demo scripts with client SDKs to talk with the Llama Stack server in our [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main/examples) repo. To run a simple agent app:
|
||||||
|
|
||||||
```
|
```bash
|
||||||
git clone git@github.com:meta-llama/llama-stack-apps.git
|
$ git clone git@github.com:meta-llama/llama-stack-apps.git
|
||||||
cd llama-stack-apps
|
$ cd llama-stack-apps
|
||||||
pip install -r requirements.txt
|
$ pip install -r requirements.txt
|
||||||
|
|
||||||
python -m examples.agents.client <host> <port>
|
$ python -m examples.agents.client <host> <port>
|
||||||
```
|
```
|
||||||
|
|
||||||
You will see outputs of the form --
|
You will see outputs of the form --
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue