Make OpenAI streaming mode work alongside earlier error detection

This commit is contained in:
cynic64 2025-01-29 10:38:29 -05:00
parent 091b5577c0
commit 260875e61d

View file

@ -408,7 +408,8 @@ class OpenAIChatCompletion(BaseLLM):
else:
headers = {}
response = raw_response.parse()
if type(response) != BaseModel:
if isinstance(response, str):
raise OpenAIError(
status_code=422,
message="Could not parse response",
@ -448,7 +449,7 @@ class OpenAIChatCompletion(BaseLLM):
headers = {}
response = raw_response.parse()
if type(response) != BaseModel:
if isinstance(response, str):
raise OpenAIError(
status_code=422,
message="Could not parse response",