mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
(feat) ollama raise Exceptions + use LiteLLM stream wrapper
This commit is contained in:
parent
689acb8a08
commit
7b3ee8d129
1 changed files with 15 additions and 1 deletions
|
@ -131,13 +131,20 @@ def get_ollama_response_stream(
|
||||||
for chunk in chunks:
|
for chunk in chunks:
|
||||||
if chunk.strip() != "":
|
if chunk.strip() != "":
|
||||||
j = json.loads(chunk)
|
j = json.loads(chunk)
|
||||||
|
if "error" in j:
|
||||||
|
completion_obj = {
|
||||||
|
"role": "assistant",
|
||||||
|
"content": "",
|
||||||
|
"error": j
|
||||||
|
}
|
||||||
|
yield completion_obj
|
||||||
if "response" in j:
|
if "response" in j:
|
||||||
completion_obj = {
|
completion_obj = {
|
||||||
"role": "assistant",
|
"role": "assistant",
|
||||||
"content": "",
|
"content": "",
|
||||||
}
|
}
|
||||||
completion_obj["content"] = j["response"]
|
completion_obj["content"] = j["response"]
|
||||||
yield {"choices": [{"delta": completion_obj}]}
|
yield completion_obj
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
traceback.print_exc()
|
traceback.print_exc()
|
||||||
print(f"Error decoding JSON: {e}")
|
print(f"Error decoding JSON: {e}")
|
||||||
|
@ -176,6 +183,13 @@ if async_generator_imported:
|
||||||
for chunk in chunks:
|
for chunk in chunks:
|
||||||
if chunk.strip() != "":
|
if chunk.strip() != "":
|
||||||
j = json.loads(chunk)
|
j = json.loads(chunk)
|
||||||
|
if "error" in j:
|
||||||
|
completion_obj = {
|
||||||
|
"role": "assistant",
|
||||||
|
"content": "",
|
||||||
|
"error": j
|
||||||
|
}
|
||||||
|
await yield_({"choices": [{"delta": completion_obj}]})
|
||||||
if "response" in j:
|
if "response" in j:
|
||||||
completion_obj = {
|
completion_obj = {
|
||||||
"role": "assistant",
|
"role": "assistant",
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue