From 0cb996c18d9358e9fe285b345983d4fe1fe87ade Mon Sep 17 00:00:00 2001 From: Aidan Do Date: Sat, 7 Dec 2024 07:03:31 +1100 Subject: [PATCH] doc: quickstart guide errors (#575) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit # What does this PR do? Addresses a few errors I got when running the quick start guide: https://llama-stack.readthedocs.io/en/latest/getting_started/index.html. We should keep this up to date to maintain engagement with the community. I've annotated the PR below. Could you PTAL 🙏 ? ## Before submitting - [x] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case). --- docs/source/getting_started/index.md | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/docs/source/getting_started/index.md b/docs/source/getting_started/index.md index e6365208f..bae31e8c4 100644 --- a/docs/source/getting_started/index.md +++ b/docs/source/getting_started/index.md @@ -62,7 +62,7 @@ llama-stack-client --endpoint http://localhost:$LLAMA_STACK_PORT models list You can test basic Llama inference completion using the CLI too. ```bash llama-stack-client --endpoint http://localhost:$LLAMA_STACK_PORT \ - inference chat_completion \ + inference chat-completion \ --message "hello, what model are you?" ``` @@ -118,6 +118,7 @@ async def run_main(): model=os.environ["INFERENCE_MODEL"], instructions="You are a helpful assistant", tools=[{"type": "memory"}], # enable Memory aka RAG + enable_session_persistence=True, ) agent = Agent(client, agent_config) @@ -139,7 +140,7 @@ async def run_main(): attachments=attachments, session_id=session_id, ) - async for log in EventLogger().log(response): + for log in EventLogger().log(response): log.print()