forked from phoenix/litellm-mirror
feat: add gemini-experimental to vertex ai
This commit is contained in:
parent
eaf81ed31c
commit
2b326ce6fd
2 changed files with 22 additions and 0 deletions
|
@ -960,6 +960,17 @@
|
||||||
"supports_function_calling": true,
|
"supports_function_calling": true,
|
||||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||||
},
|
},
|
||||||
|
"gemini-experimental": {
|
||||||
|
"max_tokens": 8192,
|
||||||
|
"max_input_tokens": 1000000,
|
||||||
|
"max_output_tokens": 8192,
|
||||||
|
"input_cost_per_token": 0,
|
||||||
|
"output_cost_per_token": 0,
|
||||||
|
"litellm_provider": "vertex_ai-language-models",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": false,
|
||||||
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||||
|
},
|
||||||
"gemini-pro-vision": {
|
"gemini-pro-vision": {
|
||||||
"max_tokens": 2048,
|
"max_tokens": 2048,
|
||||||
"max_input_tokens": 16384,
|
"max_input_tokens": 16384,
|
||||||
|
|
|
@ -960,6 +960,17 @@
|
||||||
"supports_function_calling": true,
|
"supports_function_calling": true,
|
||||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||||
},
|
},
|
||||||
|
"gemini-experimental": {
|
||||||
|
"max_tokens": 8192,
|
||||||
|
"max_input_tokens": 1000000,
|
||||||
|
"max_output_tokens": 8192,
|
||||||
|
"input_cost_per_token": 0,
|
||||||
|
"output_cost_per_token": 0,
|
||||||
|
"litellm_provider": "vertex_ai-language-models",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": false,
|
||||||
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||||
|
},
|
||||||
"gemini-pro-vision": {
|
"gemini-pro-vision": {
|
||||||
"max_tokens": 2048,
|
"max_tokens": 2048,
|
||||||
"max_input_tokens": 16384,
|
"max_input_tokens": 16384,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue