fix merge conflict

This commit is contained in:
Krrish Dholakia 2023-09-15 15:22:30 -07:00
parent 34c676a8ec
commit 8fbaf14509
4 changed files with 1 additions and 5 deletions

BIN
dist/litellm-0.1.645-py3-none-any.whl vendored Normal file

Binary file not shown.

BIN
dist/litellm-0.1.645.tar.gz vendored Normal file

Binary file not shown.

View file

@ -2451,11 +2451,7 @@ class CustomStreamWrapper:
def __next__(self): def __next__(self):
try: try:
<<<<<<< HEAD
completion_obj = {"content": "", "role": ""} # default to role being assistant
=======
completion_obj = {"content": ""} # default to role being assistant completion_obj = {"content": ""} # default to role being assistant
>>>>>>> 31d771b (fix streaming error)
if self.model in litellm.anthropic_models: if self.model in litellm.anthropic_models:
chunk = next(self.completion_stream) chunk = next(self.completion_stream)
completion_obj["content"] = self.handle_anthropic_chunk(chunk) completion_obj["content"] = self.handle_anthropic_chunk(chunk)

View file

@ -1,6 +1,6 @@
[tool.poetry] [tool.poetry]
name = "litellm" name = "litellm"
version = "0.1.645" version = "0.1.646"
description = "Library to easily interface with LLM API providers" description = "Library to easily interface with LLM API providers"
authors = ["BerriAI"] authors = ["BerriAI"]
license = "MIT License" license = "MIT License"