From e80aae5c305cfb7e2173ebe3575e4d27214ed89b Mon Sep 17 00:00:00 2001 From: Rob Mann Date: Tue, 26 Mar 2024 11:46:59 -0400 Subject: [PATCH] Updating the default Claude3 max tokens --- litellm/llms/bedrock.py | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/litellm/llms/bedrock.py b/litellm/llms/bedrock.py index eab9b0cad..8177167a5 100644 --- a/litellm/llms/bedrock.py +++ b/litellm/llms/bedrock.py @@ -79,11 +79,13 @@ class AmazonTitanConfig: class AmazonAnthropicClaude3Config: """ - Reference: https://us-west-2.console.aws.amazon.com/bedrock/home?region=us-west-2#/providers?model=claude + Reference: + https://us-west-2.console.aws.amazon.com/bedrock/home?region=us-west-2#/providers?model=claude + https://docs.anthropic.com/claude/docs/models-overview#model-comparison Supported Params for the Amazon / Anthropic Claude 3 models: - - `max_tokens` Required (integer) max tokens, + - `max_tokens` Required (integer) max tokens. Default is 4096 - `anthropic_version` Required (string) version of anthropic for bedrock - e.g. "bedrock-2023-05-31" - `system` Optional (string) the system prompt, conversion from openai format to this is handled in factory.py - `temperature` Optional (float) The amount of randomness injected into the response @@ -92,7 +94,7 @@ class AmazonAnthropicClaude3Config: - `stop_sequences` Optional (List[str]) Custom text sequences that cause the model to stop generating """ - max_tokens: Optional[int] = litellm.max_tokens + max_tokens: Optional[int] = 4096 # Opus, Sonnet, and Haiku default anthropic_version: Optional[str] = "bedrock-2023-05-31" system: Optional[str] = None temperature: Optional[float] = None