From 46072618b2ff8f541d759685ca905d428239534a Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Thu, 21 Sep 2023 12:05:07 -0700 Subject: [PATCH] bump v --- docs/my-website/docs/providers/ollama.md | 5 +++++ pyproject.toml | 2 +- 2 files changed, 6 insertions(+), 1 deletion(-) diff --git a/docs/my-website/docs/providers/ollama.md b/docs/my-website/docs/providers/ollama.md index 25a6a1ce0..7d2d35383 100644 --- a/docs/my-website/docs/providers/ollama.md +++ b/docs/my-website/docs/providers/ollama.md @@ -34,6 +34,11 @@ for chunk in response: ``` ## Example usage - Streaming + Acompletion +Ensure you have async_generator installed for using ollama acompletion with streaming +```shell +pip install async_generator +``` + ```python async def async_ollama(): response = await litellm.acompletion( diff --git a/pyproject.toml b/pyproject.toml index 32cf6b1cd..c99e759d0 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [tool.poetry] name = "litellm" -version = "0.1.723" +version = "0.1.724" description = "Library to easily interface with LLM API providers" authors = ["BerriAI"] license = "MIT License"