mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 15:23:51 +00:00
quick fix on title
This commit is contained in:
parent
ed70e140eb
commit
46763bc001
1 changed files with 3 additions and 4 deletions
|
@ -8,7 +8,8 @@ This document provides instructions on how to use Llama Stack's `chat_completion
|
|||
2. [Building Effective Prompts](#building-effective-prompts)
|
||||
3. [Conversation Loop](#conversation-loop)
|
||||
4. [Conversation History](#conversation-history)
|
||||
5.
|
||||
5. [Streaming Responses](#streaming-responses)
|
||||
|
||||
|
||||
## Quickstart
|
||||
|
||||
|
@ -141,7 +142,7 @@ async def chat_loop():
|
|||
asyncio.run(chat_loop())
|
||||
```
|
||||
|
||||
## Streaming Responses with Llama Stack
|
||||
## Streaming Responses
|
||||
|
||||
Llama Stack offers a `stream` parameter in the `chat_completion` function, which allows partial responses to be returned progressively as they are generated. This can enhance user experience by providing immediate feedback without waiting for the entire response to be processed.
|
||||
|
||||
|
@ -186,8 +187,6 @@ if __name__ == "__main__":
|
|||
```
|
||||
|
||||
|
||||
|
||||
|
||||
---
|
||||
|
||||
With these fundamentals, you should be well on your way to leveraging Llama Stack’s text generation capabilities! For more advanced features, refer to the [Llama Stack Documentation](https://llama-stack-docs.com).
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue