diff --git a/docs/my-website/docs/providers/bedrock.md b/docs/my-website/docs/providers/bedrock.md index 691e1f29d..b57057184 100644 --- a/docs/my-website/docs/providers/bedrock.md +++ b/docs/my-website/docs/providers/bedrock.md @@ -291,7 +291,6 @@ Here's an example of using a bedrock model with LiteLLM | Anthropic Claude-V2.1 | `completion(model='bedrock/anthropic.claude-v2:1', messages=messages)` | `os.environ['ANTHROPIC_ACCESS_KEY_ID']`, `os.environ['ANTHROPIC_SECRET_ACCESS_KEY']` | | Anthropic Claude-V2 | `completion(model='bedrock/anthropic.claude-v2', messages=messages)` | `os.environ['ANTHROPIC_ACCESS_KEY_ID']`, `os.environ['ANTHROPIC_SECRET_ACCESS_KEY']` | | Anthropic Claude-Instant V1 | `completion(model='bedrock/anthropic.claude-instant-v1', messages=messages)` | `os.environ['ANTHROPIC_ACCESS_KEY_ID']`, `os.environ['ANTHROPIC_SECRET_ACCESS_KEY']` | -| Anthropic Claude-V1 | `completion(model='bedrock/anthropic.claude-v1', messages=messages)` | `os.environ['ANTHROPIC_ACCESS_KEY_ID']`, `os.environ['ANTHROPIC_SECRET_ACCESS_KEY']` | | Amazon Titan Lite | `completion(model='bedrock/amazon.titan-text-lite-v1', messages=messages)` | `os.environ['AWS_ACCESS_KEY_ID']`, `os.environ['AWS_SECRET_ACCESS_KEY']`, `os.environ['AWS_REGION_NAME']` | | Amazon Titan Express | `completion(model='bedrock/amazon.titan-text-express-v1', messages=messages)` | `os.environ['AWS_ACCESS_KEY_ID']`, `os.environ['AWS_SECRET_ACCESS_KEY']`, `os.environ['AWS_REGION_NAME']` | | Cohere Command | `completion(model='bedrock/cohere.command-text-v14', messages=messages)` | `os.environ['AWS_ACCESS_KEY_ID']`, `os.environ['AWS_SECRET_ACCESS_KEY']`, `os.environ['AWS_REGION_NAME']` | diff --git a/litellm/tests/test_caching.py b/litellm/tests/test_caching.py index 5887aba12..de9740ebd 100644 --- a/litellm/tests/test_caching.py +++ b/litellm/tests/test_caching.py @@ -571,7 +571,7 @@ def test_redis_cache_acompletion_stream_bedrock(): async def call1(): nonlocal response_1_content response1 = await litellm.acompletion( - model="bedrock/anthropic.claude-v1", + model="bedrock/anthropic.claude-v2", messages=messages, max_tokens=40, temperature=1, @@ -589,7 +589,7 @@ def test_redis_cache_acompletion_stream_bedrock(): async def call2(): nonlocal response_2_content response2 = await litellm.acompletion( - model="bedrock/anthropic.claude-v1", + model="bedrock/anthropic.claude-v2", messages=messages, max_tokens=40, temperature=1, diff --git a/litellm/tests/test_custom_callback_input.py b/litellm/tests/test_custom_callback_input.py index ca1fe19a9..6e55cc5a1 100644 --- a/litellm/tests/test_custom_callback_input.py +++ b/litellm/tests/test_custom_callback_input.py @@ -483,12 +483,12 @@ def test_chat_bedrock_stream(): customHandler = CompletionCustomHandler() litellm.callbacks = [customHandler] response = litellm.completion( - model="bedrock/anthropic.claude-v1", + model="bedrock/anthropic.claude-v2", messages=[{"role": "user", "content": "Hi 👋 - i'm sync bedrock"}], ) # test streaming response = litellm.completion( - model="bedrock/anthropic.claude-v1", + model="bedrock/anthropic.claude-v2", messages=[{"role": "user", "content": "Hi 👋 - i'm sync bedrock"}], stream=True, ) @@ -497,7 +497,7 @@ def test_chat_bedrock_stream(): # test failure callback try: response = litellm.completion( - model="bedrock/anthropic.claude-v1", + model="bedrock/anthropic.claude-v2", messages=[{"role": "user", "content": "Hi 👋 - i'm sync bedrock"}], aws_region_name="my-bad-region", stream=True, @@ -524,12 +524,12 @@ async def test_async_chat_bedrock_stream(): customHandler = CompletionCustomHandler() litellm.callbacks = [customHandler] response = await litellm.acompletion( - model="bedrock/anthropic.claude-v1", + model="bedrock/anthropic.claude-v2", messages=[{"role": "user", "content": "Hi 👋 - i'm async bedrock"}], ) # test streaming response = await litellm.acompletion( - model="bedrock/anthropic.claude-v1", + model="bedrock/anthropic.claude-v2", messages=[{"role": "user", "content": "Hi 👋 - i'm async bedrock"}], stream=True, ) @@ -540,7 +540,7 @@ async def test_async_chat_bedrock_stream(): ## test failure callback try: response = await litellm.acompletion( - model="bedrock/anthropic.claude-v1", + model="bedrock/anthropic.claude-v2", messages=[{"role": "user", "content": "Hi 👋 - i'm async bedrock"}], aws_region_name="my-bad-key", stream=True,