mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-27 11:43:54 +00:00
Update README.md
This commit is contained in:
parent
17b319cc84
commit
b1ed53de72
1 changed files with 16 additions and 0 deletions
16
README.md
16
README.md
|
@ -62,6 +62,22 @@ response = completion(model="command-nightly", messages=messages)
|
||||||
print(response)
|
print(response)
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Async ([Docs](https://docs.litellm.ai/docs/completion/stream#async-completion))
|
||||||
|
|
||||||
|
```python
|
||||||
|
from litellm import acompletion
|
||||||
|
import asyncio
|
||||||
|
|
||||||
|
async def test_get_response():
|
||||||
|
user_message = "Hello, how are you?"
|
||||||
|
messages = [{"content": user_message, "role": "user"}]
|
||||||
|
response = await acompletion(model="gpt-3.5-turbo", messages=messages)
|
||||||
|
return response
|
||||||
|
|
||||||
|
response = asyncio.run(test_get_response())
|
||||||
|
print(response)
|
||||||
|
```
|
||||||
|
|
||||||
## Streaming ([Docs](https://docs.litellm.ai/docs/completion/stream))
|
## Streaming ([Docs](https://docs.litellm.ai/docs/completion/stream))
|
||||||
liteLLM supports streaming the model response back, pass `stream=True` to get a streaming iterator in response.
|
liteLLM supports streaming the model response back, pass `stream=True` to get a streaming iterator in response.
|
||||||
Streaming is supported for all models (Bedrock, Huggingface, TogetherAI, Azure, OpenAI, etc.)
|
Streaming is supported for all models (Bedrock, Huggingface, TogetherAI, Azure, OpenAI, etc.)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue