quick fix on title

This commit is contained in:
Justin Lee 2024-11-01 11:46:35 -07:00
parent ed70e140eb
commit 46763bc001

View file

@ -8,7 +8,8 @@ This document provides instructions on how to use Llama Stack's `chat_completion
2. [Building Effective Prompts](#building-effective-prompts) 2. [Building Effective Prompts](#building-effective-prompts)
3. [Conversation Loop](#conversation-loop) 3. [Conversation Loop](#conversation-loop)
4. [Conversation History](#conversation-history) 4. [Conversation History](#conversation-history)
5. 5. [Streaming Responses](#streaming-responses)
## Quickstart ## Quickstart
@ -141,7 +142,7 @@ async def chat_loop():
asyncio.run(chat_loop()) asyncio.run(chat_loop())
``` ```
## Streaming Responses with Llama Stack ## Streaming Responses
Llama Stack offers a `stream` parameter in the `chat_completion` function, which allows partial responses to be returned progressively as they are generated. This can enhance user experience by providing immediate feedback without waiting for the entire response to be processed. Llama Stack offers a `stream` parameter in the `chat_completion` function, which allows partial responses to be returned progressively as they are generated. This can enhance user experience by providing immediate feedback without waiting for the entire response to be processed.
@ -186,8 +187,6 @@ if __name__ == "__main__":
``` ```
--- ---
With these fundamentals, you should be well on your way to leveraging Llama Stacks text generation capabilities! For more advanced features, refer to the [Llama Stack Documentation](https://llama-stack-docs.com). With these fundamentals, you should be well on your way to leveraging Llama Stacks text generation capabilities! For more advanced features, refer to the [Llama Stack Documentation](https://llama-stack-docs.com).