forked from phoenix/litellm-mirror
feat(bedrock_httpx.py): add ai21 jamba instruct as converse model
initial commit for adding ai21 jamba instruct support through bedrock converse
This commit is contained in:
parent
b23a633cf1
commit
cece76c4ee
3 changed files with 21 additions and 1 deletions
|
@ -74,7 +74,7 @@ BEDROCK_CONVERSE_MODELS = [
|
||||||
"anthropic.claude-v2",
|
"anthropic.claude-v2",
|
||||||
"anthropic.claude-v2:1",
|
"anthropic.claude-v2:1",
|
||||||
"anthropic.claude-v1",
|
"anthropic.claude-v1",
|
||||||
"anthropic.claude-instant-v1",
|
"ai21.jamba-instruct-v1:0",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -2812,6 +2812,16 @@
|
||||||
"litellm_provider": "bedrock",
|
"litellm_provider": "bedrock",
|
||||||
"mode": "chat"
|
"mode": "chat"
|
||||||
},
|
},
|
||||||
|
"ai21.jamba-instruct-v1:0": {
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"max_input_tokens": 70000,
|
||||||
|
"max_output_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000005,
|
||||||
|
"output_cost_per_token": 0.0000007,
|
||||||
|
"litellm_provider": "bedrock",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_system_messages": true
|
||||||
|
},
|
||||||
"amazon.titan-text-lite-v1": {
|
"amazon.titan-text-lite-v1": {
|
||||||
"max_tokens": 4000,
|
"max_tokens": 4000,
|
||||||
"max_input_tokens": 42000,
|
"max_input_tokens": 42000,
|
||||||
|
|
|
@ -2812,6 +2812,16 @@
|
||||||
"litellm_provider": "bedrock",
|
"litellm_provider": "bedrock",
|
||||||
"mode": "chat"
|
"mode": "chat"
|
||||||
},
|
},
|
||||||
|
"ai21.jamba-instruct-v1:0": {
|
||||||
|
"max_tokens": 4096,
|
||||||
|
"max_input_tokens": 70000,
|
||||||
|
"max_output_tokens": 4096,
|
||||||
|
"input_cost_per_token": 0.0000005,
|
||||||
|
"output_cost_per_token": 0.0000007,
|
||||||
|
"litellm_provider": "bedrock",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_system_messages": true
|
||||||
|
},
|
||||||
"amazon.titan-text-lite-v1": {
|
"amazon.titan-text-lite-v1": {
|
||||||
"max_tokens": 4000,
|
"max_tokens": 4000,
|
||||||
"max_input_tokens": 42000,
|
"max_input_tokens": 42000,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue