diff --git a/tests/pass_through_tests/test_local_vertex.js b/tests/pass_through_tests/test_local_vertex.js index c0971543da..a94c6746e9 100644 --- a/tests/pass_through_tests/test_local_vertex.js +++ b/tests/pass_through_tests/test_local_vertex.js @@ -20,7 +20,7 @@ const requestOptions = { }; const generativeModel = vertexAI.getGenerativeModel( - { model: 'gemini-1.0-pro' }, + { model: 'gemini-1.5-pro' }, requestOptions ); diff --git a/tests/pass_through_tests/test_vertex.test.js b/tests/pass_through_tests/test_vertex.test.js index c426c3de30..dccd649402 100644 --- a/tests/pass_through_tests/test_vertex.test.js +++ b/tests/pass_through_tests/test_vertex.test.js @@ -75,7 +75,7 @@ describe('Vertex AI Tests', () => { }; const generativeModel = vertexAI.getGenerativeModel( - { model: 'gemini-1.0-pro' }, + { model: 'gemini-1.5-pro' }, requestOptions ); @@ -103,7 +103,7 @@ describe('Vertex AI Tests', () => { const vertexAI = new VertexAI({project: 'pathrise-convert-1606954137718', location: 'us-central1', apiEndpoint: "localhost:4000/vertex-ai"}); const customHeaders = new Headers({"x-litellm-api-key": "sk-1234"}); const requestOptions = {customHeaders: customHeaders}; - const generativeModel = vertexAI.getGenerativeModel({model: 'gemini-1.0-pro'}, requestOptions); + const generativeModel = vertexAI.getGenerativeModel({model: 'gemini-1.5-pro'}, requestOptions); const request = {contents: [{role: 'user', parts: [{text: 'What is 2+2?'}]}]}; const result = await generativeModel.generateContent(request); diff --git a/tests/pass_through_tests/test_vertex_ai.py b/tests/pass_through_tests/test_vertex_ai.py index b9a3165269..cf1201be58 100644 --- a/tests/pass_through_tests/test_vertex_ai.py +++ b/tests/pass_through_tests/test_vertex_ai.py @@ -103,7 +103,7 @@ async def test_basic_vertex_ai_pass_through_with_spendlog(): api_transport="rest", ) - model = GenerativeModel(model_name="gemini-1.0-pro") + model = GenerativeModel(model_name="gemini-1.5-pro") response = model.generate_content("hi") print("response", response) @@ -135,7 +135,7 @@ async def test_basic_vertex_ai_pass_through_streaming_with_spendlog(): api_transport="rest", ) - model = GenerativeModel(model_name="gemini-1.0-pro") + model = GenerativeModel(model_name="gemini-1.5-pro") response = model.generate_content("hi", stream=True) for chunk in response: diff --git a/tests/pass_through_tests/test_vertex_with_spend.test.js b/tests/pass_through_tests/test_vertex_with_spend.test.js index c342931497..cc1808b445 100644 --- a/tests/pass_through_tests/test_vertex_with_spend.test.js +++ b/tests/pass_through_tests/test_vertex_with_spend.test.js @@ -84,7 +84,7 @@ describe('Vertex AI Tests', () => { }; const generativeModel = vertexAI.getGenerativeModel( - { model: 'gemini-1.0-pro' }, + { model: 'gemini-1.5-pro' }, requestOptions ); @@ -140,7 +140,7 @@ describe('Vertex AI Tests', () => { }; const generativeModel = vertexAI.getGenerativeModel( - { model: 'gemini-1.0-pro' }, + { model: 'gemini-1.5-pro' }, requestOptions );