forked from phoenix/litellm-mirror
fix(utils.py): raise stop iteration exception on bedrock stream close
This commit is contained in:
parent
286ce586be
commit
5411d5a6fd
1 changed files with 4 additions and 3 deletions
|
@ -2176,7 +2176,7 @@ def get_optional_params( # use the openai defaults
|
||||||
_check_valid_arg(supported_params=supported_params)
|
_check_valid_arg(supported_params=supported_params)
|
||||||
elif custom_llm_provider == "bedrock":
|
elif custom_llm_provider == "bedrock":
|
||||||
if "ai21" in model:
|
if "ai21" in model:
|
||||||
supported_params = ["max_tokens", "temperature", "stop", "top_p", "stream"]
|
supported_params = ["max_tokens", "temperature", "top_p", "stream"]
|
||||||
_check_valid_arg(supported_params=supported_params)
|
_check_valid_arg(supported_params=supported_params)
|
||||||
# params "maxTokens":200,"temperature":0,"topP":250,"stop_sequences":[],
|
# params "maxTokens":200,"temperature":0,"topP":250,"stop_sequences":[],
|
||||||
# https://us-west-2.console.aws.amazon.com/bedrock/home?region=us-west-2#/providers?model=j2-ultra
|
# https://us-west-2.console.aws.amazon.com/bedrock/home?region=us-west-2#/providers?model=j2-ultra
|
||||||
|
@ -2184,8 +2184,6 @@ def get_optional_params( # use the openai defaults
|
||||||
optional_params["maxTokens"] = max_tokens
|
optional_params["maxTokens"] = max_tokens
|
||||||
if temperature is not None:
|
if temperature is not None:
|
||||||
optional_params["temperature"] = temperature
|
optional_params["temperature"] = temperature
|
||||||
if stop is not None:
|
|
||||||
optional_params["stop_sequences"] = stop
|
|
||||||
if top_p is not None:
|
if top_p is not None:
|
||||||
optional_params["topP"] = top_p
|
optional_params["topP"] = top_p
|
||||||
if stream:
|
if stream:
|
||||||
|
@ -5184,10 +5182,13 @@ class CustomStreamWrapper:
|
||||||
if response_obj["is_finished"]:
|
if response_obj["is_finished"]:
|
||||||
model_response.choices[0].finish_reason = response_obj["finish_reason"]
|
model_response.choices[0].finish_reason = response_obj["finish_reason"]
|
||||||
elif self.custom_llm_provider == "bedrock":
|
elif self.custom_llm_provider == "bedrock":
|
||||||
|
if self.sent_last_chunk:
|
||||||
|
raise StopIteration
|
||||||
response_obj = self.handle_bedrock_stream(chunk)
|
response_obj = self.handle_bedrock_stream(chunk)
|
||||||
completion_obj["content"] = response_obj["text"]
|
completion_obj["content"] = response_obj["text"]
|
||||||
if response_obj["is_finished"]:
|
if response_obj["is_finished"]:
|
||||||
model_response.choices[0].finish_reason = response_obj["finish_reason"]
|
model_response.choices[0].finish_reason = response_obj["finish_reason"]
|
||||||
|
self.sent_last_chunk = True
|
||||||
elif self.custom_llm_provider == "sagemaker":
|
elif self.custom_llm_provider == "sagemaker":
|
||||||
if len(self.completion_stream)==0:
|
if len(self.completion_stream)==0:
|
||||||
if self.sent_last_chunk:
|
if self.sent_last_chunk:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue