diff --git a/litellm/llms/bedrock_httpx.py b/litellm/llms/bedrock_httpx.py index 3faaf40f1f..8aefc63cfa 100644 --- a/litellm/llms/bedrock_httpx.py +++ b/litellm/llms/bedrock_httpx.py @@ -64,6 +64,10 @@ BEDROCK_CONVERSE_MODELS = [ "anthropic.claude-3-opus-20240229-v1:0", "anthropic.claude-3-sonnet-20240229-v1:0", "anthropic.claude-3-haiku-20240307-v1:0", + "anthropic.claude-v2", + "anthropic.claude-v2:1", + "anthropic.claude-v1", + "anthropic.claude-instant-v1", ] iam_cache = DualCache() diff --git a/litellm/utils.py b/litellm/utils.py index f6e612e9c8..08a5eb40d6 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -2868,12 +2868,7 @@ def get_optional_params( optional_params=optional_params, ) ) - else: - optional_params = litellm.AmazonAnthropicConfig().map_openai_params( - non_default_params=non_default_params, - optional_params=optional_params, - ) - else: # bedrock httpx route + elif model in litellm.BEDROCK_CONVERSE_MODELS: optional_params = litellm.AmazonConverseConfig().map_openai_params( model=model, non_default_params=non_default_params, @@ -2884,6 +2879,11 @@ def get_optional_params( else False ), ) + else: + optional_params = litellm.AmazonAnthropicConfig().map_openai_params( + non_default_params=non_default_params, + optional_params=optional_params, + ) elif "amazon" in model: # amazon titan llms _check_valid_arg(supported_params=supported_params) # see https://us-west-2.console.aws.amazon.com/bedrock/home?region=us-west-2#/providers?model=titan-large