forked from phoenix/litellm-mirror
build(model_prices_and_context_window.json): add 'supports_assistant_prefill' to model info map
Closes https://github.com/BerriAI/litellm/issues/4881
This commit is contained in:
parent
e861d97bf2
commit
19bb95f781
5 changed files with 125 additions and 52 deletions
|
@ -890,7 +890,8 @@
|
|||
"input_cost_per_token": 0.00000025,
|
||||
"output_cost_per_token": 0.00000025,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-small": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -900,7 +901,8 @@
|
|||
"output_cost_per_token": 0.000003,
|
||||
"litellm_provider": "mistral",
|
||||
"supports_function_calling": true,
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-small-latest": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -910,7 +912,8 @@
|
|||
"output_cost_per_token": 0.000003,
|
||||
"litellm_provider": "mistral",
|
||||
"supports_function_calling": true,
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-medium": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -919,7 +922,8 @@
|
|||
"input_cost_per_token": 0.0000027,
|
||||
"output_cost_per_token": 0.0000081,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-medium-latest": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -928,7 +932,8 @@
|
|||
"input_cost_per_token": 0.0000027,
|
||||
"output_cost_per_token": 0.0000081,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-medium-2312": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -937,7 +942,8 @@
|
|||
"input_cost_per_token": 0.0000027,
|
||||
"output_cost_per_token": 0.0000081,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-large-latest": {
|
||||
"max_tokens": 128000,
|
||||
|
@ -947,7 +953,8 @@
|
|||
"output_cost_per_token": 0.000009,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-large-2402": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -957,7 +964,8 @@
|
|||
"output_cost_per_token": 0.000012,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-large-2407": {
|
||||
"max_tokens": 128000,
|
||||
|
@ -967,7 +975,8 @@
|
|||
"output_cost_per_token": 0.000009,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-mistral-7b": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -976,7 +985,8 @@
|
|||
"input_cost_per_token": 0.00000025,
|
||||
"output_cost_per_token": 0.00000025,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-mixtral-8x7b": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -986,7 +996,8 @@
|
|||
"output_cost_per_token": 0.0000007,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-mixtral-8x22b": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -996,7 +1007,8 @@
|
|||
"output_cost_per_token": 0.000006,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/codestral-latest": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -1005,7 +1017,8 @@
|
|||
"input_cost_per_token": 0.000001,
|
||||
"output_cost_per_token": 0.000003,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/codestral-2405": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -1014,7 +1027,8 @@
|
|||
"input_cost_per_token": 0.000001,
|
||||
"output_cost_per_token": 0.000003,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-mistral-nemo": {
|
||||
"max_tokens": 128000,
|
||||
|
@ -1024,7 +1038,8 @@
|
|||
"output_cost_per_token": 0.0000003,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"source": "https://mistral.ai/technology/"
|
||||
"source": "https://mistral.ai/technology/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-mistral-nemo-2407": {
|
||||
"max_tokens": 128000,
|
||||
|
@ -1034,7 +1049,8 @@
|
|||
"output_cost_per_token": 0.0000003,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"source": "https://mistral.ai/technology/"
|
||||
"source": "https://mistral.ai/technology/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-codestral-mamba": {
|
||||
"max_tokens": 256000,
|
||||
|
@ -1044,7 +1060,8 @@
|
|||
"output_cost_per_token": 0.00000025,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"source": "https://mistral.ai/technology/"
|
||||
"source": "https://mistral.ai/technology/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/codestral-mamba-latest": {
|
||||
"max_tokens": 256000,
|
||||
|
@ -1054,7 +1071,8 @@
|
|||
"output_cost_per_token": 0.00000025,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"source": "https://mistral.ai/technology/"
|
||||
"source": "https://mistral.ai/technology/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-embed": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -1071,7 +1089,10 @@
|
|||
"input_cost_per_token_cache_hit": 0.000000014,
|
||||
"output_cost_per_token": 0.00000028,
|
||||
"litellm_provider": "deepseek",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_tool_choice": true
|
||||
},
|
||||
"codestral/codestral-latest": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -1081,7 +1102,8 @@
|
|||
"output_cost_per_token": 0.000000,
|
||||
"litellm_provider": "codestral",
|
||||
"mode": "chat",
|
||||
"source": "https://docs.mistral.ai/capabilities/code_generation/"
|
||||
"source": "https://docs.mistral.ai/capabilities/code_generation/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"codestral/codestral-2405": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -1091,7 +1113,8 @@
|
|||
"output_cost_per_token": 0.000000,
|
||||
"litellm_provider": "codestral",
|
||||
"mode": "chat",
|
||||
"source": "https://docs.mistral.ai/capabilities/code_generation/"
|
||||
"source": "https://docs.mistral.ai/capabilities/code_generation/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"text-completion-codestral/codestral-latest": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -1121,7 +1144,10 @@
|
|||
"input_cost_per_token_cache_hit": 0.000000014,
|
||||
"output_cost_per_token": 0.00000028,
|
||||
"litellm_provider": "deepseek",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_tool_choice": true
|
||||
},
|
||||
"groq/llama2-70b-4096": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -1290,7 +1316,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"tool_use_system_prompt_tokens": 264
|
||||
"tool_use_system_prompt_tokens": 264,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"claude-3-opus-20240229": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -1302,7 +1329,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"tool_use_system_prompt_tokens": 395
|
||||
"tool_use_system_prompt_tokens": 395,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"claude-3-sonnet-20240229": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -1314,7 +1342,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"tool_use_system_prompt_tokens": 159
|
||||
"tool_use_system_prompt_tokens": 159,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"claude-3-5-sonnet-20240620": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -1326,7 +1355,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"tool_use_system_prompt_tokens": 159
|
||||
"tool_use_system_prompt_tokens": 159,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"text-bison": {
|
||||
"max_tokens": 2048,
|
||||
|
|
|
@ -46,3 +46,11 @@ def test_get_model_info_shows_correct_supports_vision():
|
|||
info = litellm.get_model_info("gemini/gemini-1.5-flash")
|
||||
print("info", info)
|
||||
assert info["supports_vision"] is True
|
||||
|
||||
|
||||
def test_get_model_info_shows_assistant_prefill():
|
||||
os.environ["LITELLM_LOCAL_MODEL_COST_MAP"] = "True"
|
||||
litellm.model_cost = litellm.get_model_cost_map(url="")
|
||||
info = litellm.get_model_info("deepseek/deepseek-chat")
|
||||
print("info", info)
|
||||
assert info.get("supports_assistant_prefill") is True
|
||||
|
|
|
@ -77,6 +77,7 @@ class ModelInfo(TypedDict, total=False):
|
|||
supports_response_schema: Optional[bool]
|
||||
supports_vision: Optional[bool]
|
||||
supports_function_calling: Optional[bool]
|
||||
supports_assistant_prefill: Optional[bool]
|
||||
|
||||
|
||||
class GenericStreamingChunk(TypedDict):
|
||||
|
|
|
@ -5103,6 +5103,7 @@ def get_model_info(model: str, custom_llm_provider: Optional[str] = None) -> Mod
|
|||
supports_system_messages=None,
|
||||
supports_response_schema=None,
|
||||
supports_function_calling=None,
|
||||
supports_assistant_prefill=None,
|
||||
)
|
||||
else:
|
||||
"""
|
||||
|
@ -5200,6 +5201,9 @@ def get_model_info(model: str, custom_llm_provider: Optional[str] = None) -> Mod
|
|||
supports_function_calling=_model_info.get(
|
||||
"supports_function_calling", False
|
||||
),
|
||||
supports_assistant_prefill=_model_info.get(
|
||||
"supports_assistant_prefill", False
|
||||
),
|
||||
)
|
||||
except Exception:
|
||||
raise Exception(
|
||||
|
|
|
@ -890,7 +890,8 @@
|
|||
"input_cost_per_token": 0.00000025,
|
||||
"output_cost_per_token": 0.00000025,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-small": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -900,7 +901,8 @@
|
|||
"output_cost_per_token": 0.000003,
|
||||
"litellm_provider": "mistral",
|
||||
"supports_function_calling": true,
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-small-latest": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -910,7 +912,8 @@
|
|||
"output_cost_per_token": 0.000003,
|
||||
"litellm_provider": "mistral",
|
||||
"supports_function_calling": true,
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-medium": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -919,7 +922,8 @@
|
|||
"input_cost_per_token": 0.0000027,
|
||||
"output_cost_per_token": 0.0000081,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-medium-latest": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -928,7 +932,8 @@
|
|||
"input_cost_per_token": 0.0000027,
|
||||
"output_cost_per_token": 0.0000081,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-medium-2312": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -937,7 +942,8 @@
|
|||
"input_cost_per_token": 0.0000027,
|
||||
"output_cost_per_token": 0.0000081,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-large-latest": {
|
||||
"max_tokens": 128000,
|
||||
|
@ -947,7 +953,8 @@
|
|||
"output_cost_per_token": 0.000009,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-large-2402": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -957,7 +964,8 @@
|
|||
"output_cost_per_token": 0.000012,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-large-2407": {
|
||||
"max_tokens": 128000,
|
||||
|
@ -967,7 +975,8 @@
|
|||
"output_cost_per_token": 0.000009,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-mistral-7b": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -976,7 +985,8 @@
|
|||
"input_cost_per_token": 0.00000025,
|
||||
"output_cost_per_token": 0.00000025,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-mixtral-8x7b": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -986,7 +996,8 @@
|
|||
"output_cost_per_token": 0.0000007,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-mixtral-8x22b": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -996,7 +1007,8 @@
|
|||
"output_cost_per_token": 0.000006,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/codestral-latest": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -1005,7 +1017,8 @@
|
|||
"input_cost_per_token": 0.000001,
|
||||
"output_cost_per_token": 0.000003,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/codestral-2405": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -1014,7 +1027,8 @@
|
|||
"input_cost_per_token": 0.000001,
|
||||
"output_cost_per_token": 0.000003,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-mistral-nemo": {
|
||||
"max_tokens": 128000,
|
||||
|
@ -1024,7 +1038,8 @@
|
|||
"output_cost_per_token": 0.0000003,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"source": "https://mistral.ai/technology/"
|
||||
"source": "https://mistral.ai/technology/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-mistral-nemo-2407": {
|
||||
"max_tokens": 128000,
|
||||
|
@ -1034,7 +1049,8 @@
|
|||
"output_cost_per_token": 0.0000003,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"source": "https://mistral.ai/technology/"
|
||||
"source": "https://mistral.ai/technology/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/open-codestral-mamba": {
|
||||
"max_tokens": 256000,
|
||||
|
@ -1044,7 +1060,8 @@
|
|||
"output_cost_per_token": 0.00000025,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"source": "https://mistral.ai/technology/"
|
||||
"source": "https://mistral.ai/technology/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/codestral-mamba-latest": {
|
||||
"max_tokens": 256000,
|
||||
|
@ -1054,7 +1071,8 @@
|
|||
"output_cost_per_token": 0.00000025,
|
||||
"litellm_provider": "mistral",
|
||||
"mode": "chat",
|
||||
"source": "https://mistral.ai/technology/"
|
||||
"source": "https://mistral.ai/technology/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"mistral/mistral-embed": {
|
||||
"max_tokens": 8192,
|
||||
|
@ -1071,7 +1089,10 @@
|
|||
"input_cost_per_token_cache_hit": 0.000000014,
|
||||
"output_cost_per_token": 0.00000028,
|
||||
"litellm_provider": "deepseek",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_tool_choice": true
|
||||
},
|
||||
"codestral/codestral-latest": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -1081,7 +1102,8 @@
|
|||
"output_cost_per_token": 0.000000,
|
||||
"litellm_provider": "codestral",
|
||||
"mode": "chat",
|
||||
"source": "https://docs.mistral.ai/capabilities/code_generation/"
|
||||
"source": "https://docs.mistral.ai/capabilities/code_generation/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"codestral/codestral-2405": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -1091,7 +1113,8 @@
|
|||
"output_cost_per_token": 0.000000,
|
||||
"litellm_provider": "codestral",
|
||||
"mode": "chat",
|
||||
"source": "https://docs.mistral.ai/capabilities/code_generation/"
|
||||
"source": "https://docs.mistral.ai/capabilities/code_generation/",
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"text-completion-codestral/codestral-latest": {
|
||||
"max_tokens": 8191,
|
||||
|
@ -1121,7 +1144,10 @@
|
|||
"input_cost_per_token_cache_hit": 0.000000014,
|
||||
"output_cost_per_token": 0.00000028,
|
||||
"litellm_provider": "deepseek",
|
||||
"mode": "chat"
|
||||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true,
|
||||
"supports_tool_choice": true
|
||||
},
|
||||
"groq/llama2-70b-4096": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -1290,7 +1316,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"tool_use_system_prompt_tokens": 264
|
||||
"tool_use_system_prompt_tokens": 264,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"claude-3-opus-20240229": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -1302,7 +1329,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"tool_use_system_prompt_tokens": 395
|
||||
"tool_use_system_prompt_tokens": 395,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"claude-3-sonnet-20240229": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -1314,7 +1342,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"tool_use_system_prompt_tokens": 159
|
||||
"tool_use_system_prompt_tokens": 159,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"claude-3-5-sonnet-20240620": {
|
||||
"max_tokens": 4096,
|
||||
|
@ -1326,7 +1355,8 @@
|
|||
"mode": "chat",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": true,
|
||||
"tool_use_system_prompt_tokens": 159
|
||||
"tool_use_system_prompt_tokens": 159,
|
||||
"supports_assistant_prefill": true
|
||||
},
|
||||
"text-bison": {
|
||||
"max_tokens": 2048,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue