mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-27 19:54:13 +00:00
fix(utils.py): fix default values
This commit is contained in:
parent
8caf105427
commit
a241cdb818
4 changed files with 13 additions and 7 deletions
|
@ -952,7 +952,7 @@ def completion(
|
||||||
original_response=json.dumps(response_body),
|
original_response=json.dumps(response_body),
|
||||||
additional_args={"complete_input_dict": data},
|
additional_args={"complete_input_dict": data},
|
||||||
)
|
)
|
||||||
print_verbose(f"raw model_response: {response}")
|
print_verbose(f"raw model_response: {response_body}")
|
||||||
## RESPONSE OBJECT
|
## RESPONSE OBJECT
|
||||||
outputText = "default"
|
outputText = "default"
|
||||||
if provider == "ai21":
|
if provider == "ai21":
|
||||||
|
@ -1065,6 +1065,7 @@ def completion(
|
||||||
outputText = response_body.get("results")[0].get("outputText")
|
outputText = response_body.get("results")[0].get("outputText")
|
||||||
|
|
||||||
response_metadata = response.get("ResponseMetadata", {})
|
response_metadata = response.get("ResponseMetadata", {})
|
||||||
|
|
||||||
if response_metadata.get("HTTPStatusCode", 500) >= 400:
|
if response_metadata.get("HTTPStatusCode", 500) >= 400:
|
||||||
raise BedrockError(
|
raise BedrockError(
|
||||||
message=outputText,
|
message=outputText,
|
||||||
|
@ -1100,11 +1101,13 @@ def completion(
|
||||||
prompt_tokens = response_metadata.get(
|
prompt_tokens = response_metadata.get(
|
||||||
"x-amzn-bedrock-input-token-count", len(encoding.encode(prompt))
|
"x-amzn-bedrock-input-token-count", len(encoding.encode(prompt))
|
||||||
)
|
)
|
||||||
|
_text_response = model_response["choices"][0]["message"].get("content", "")
|
||||||
completion_tokens = response_metadata.get(
|
completion_tokens = response_metadata.get(
|
||||||
"x-amzn-bedrock-output-token-count",
|
"x-amzn-bedrock-output-token-count",
|
||||||
len(
|
len(
|
||||||
encoding.encode(
|
encoding.encode(
|
||||||
model_response["choices"][0]["message"].get("content", "")
|
_text_response,
|
||||||
|
disallowed_special=(),
|
||||||
)
|
)
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
|
|
@ -529,6 +529,7 @@ def test_chat_bedrock_stream():
|
||||||
@pytest.mark.asyncio
|
@pytest.mark.asyncio
|
||||||
async def test_async_chat_bedrock_stream():
|
async def test_async_chat_bedrock_stream():
|
||||||
try:
|
try:
|
||||||
|
litellm.set_verbose = True
|
||||||
customHandler = CompletionCustomHandler()
|
customHandler = CompletionCustomHandler()
|
||||||
litellm.callbacks = [customHandler]
|
litellm.callbacks = [customHandler]
|
||||||
response = await litellm.acompletion(
|
response = await litellm.acompletion(
|
||||||
|
|
|
@ -518,7 +518,7 @@ async def test_acompletion_gemini_stream():
|
||||||
litellm.set_verbose = True
|
litellm.set_verbose = True
|
||||||
print("Streaming gemini response")
|
print("Streaming gemini response")
|
||||||
messages = [
|
messages = [
|
||||||
{"role": "system", "content": "You are a helpful assistant."},
|
# {"role": "system", "content": "You are a helpful assistant."},
|
||||||
{
|
{
|
||||||
"role": "user",
|
"role": "user",
|
||||||
"content": "What do you know?",
|
"content": "What do you know?",
|
||||||
|
@ -1297,6 +1297,7 @@ def test_completion_watsonx_stream():
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
pytest.fail(f"Error occurred: {e}")
|
pytest.fail(f"Error occurred: {e}")
|
||||||
|
|
||||||
|
|
||||||
# test_completion_sagemaker_stream()
|
# test_completion_sagemaker_stream()
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -410,10 +410,11 @@ class Message(OpenAIObject):
|
||||||
except:
|
except:
|
||||||
# if using pydantic v1
|
# if using pydantic v1
|
||||||
return self.dict()
|
return self.dict()
|
||||||
|
|
||||||
|
|
||||||
class Delta(OpenAIObject):
|
class Delta(OpenAIObject):
|
||||||
tool_calls: Optional[List[ChatCompletionDeltaToolCall]]
|
tool_calls: Optional[List[ChatCompletionDeltaToolCall]] = None
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
content=None,
|
content=None,
|
||||||
|
@ -475,7 +476,7 @@ class Choices(OpenAIObject):
|
||||||
) # set finish_reason for all responses
|
) # set finish_reason for all responses
|
||||||
self.index = index
|
self.index = index
|
||||||
if message is None:
|
if message is None:
|
||||||
self.message = Message(content=None)
|
self.message = Message()
|
||||||
else:
|
else:
|
||||||
if isinstance(message, Message):
|
if isinstance(message, Message):
|
||||||
self.message = message
|
self.message = message
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue