forked from phoenix/litellm-mirror
(test) vertex ai: stop running 4 requests / test
This commit is contained in:
parent
e527137bee
commit
4af13e44df
1 changed files with 5 additions and 4 deletions
|
@ -93,7 +93,7 @@ def test_vertex_ai():
|
||||||
litellm.set_verbose=False
|
litellm.set_verbose=False
|
||||||
litellm.vertex_project = "hardy-device-386718"
|
litellm.vertex_project = "hardy-device-386718"
|
||||||
|
|
||||||
test_models = random.sample(test_models, 4)
|
test_models = random.sample(test_models, 1)
|
||||||
test_models += litellm.vertex_language_models # always test gemini-pro
|
test_models += litellm.vertex_language_models # always test gemini-pro
|
||||||
for model in test_models:
|
for model in test_models:
|
||||||
try:
|
try:
|
||||||
|
@ -117,7 +117,7 @@ def test_vertex_ai_stream():
|
||||||
import random
|
import random
|
||||||
|
|
||||||
test_models = litellm.vertex_chat_models + litellm.vertex_code_chat_models + litellm.vertex_text_models + litellm.vertex_code_text_models
|
test_models = litellm.vertex_chat_models + litellm.vertex_code_chat_models + litellm.vertex_text_models + litellm.vertex_code_text_models
|
||||||
test_models = random.sample(test_models, 4)
|
test_models = random.sample(test_models, 1)
|
||||||
test_models += litellm.vertex_language_models # always test gemini-pro
|
test_models += litellm.vertex_language_models # always test gemini-pro
|
||||||
for model in test_models:
|
for model in test_models:
|
||||||
try:
|
try:
|
||||||
|
@ -144,7 +144,7 @@ async def test_async_vertexai_response():
|
||||||
import random
|
import random
|
||||||
load_vertex_ai_credentials()
|
load_vertex_ai_credentials()
|
||||||
test_models = litellm.vertex_chat_models + litellm.vertex_code_chat_models + litellm.vertex_text_models + litellm.vertex_code_text_models
|
test_models = litellm.vertex_chat_models + litellm.vertex_code_chat_models + litellm.vertex_text_models + litellm.vertex_code_text_models
|
||||||
test_models = random.sample(test_models, 4)
|
test_models = random.sample(test_models, 1)
|
||||||
test_models += litellm.vertex_language_models # always test gemini-pro
|
test_models += litellm.vertex_language_models # always test gemini-pro
|
||||||
for model in test_models:
|
for model in test_models:
|
||||||
print(f'model being tested in async call: {model}')
|
print(f'model being tested in async call: {model}')
|
||||||
|
@ -168,7 +168,7 @@ async def test_async_vertexai_streaming_response():
|
||||||
import random
|
import random
|
||||||
load_vertex_ai_credentials()
|
load_vertex_ai_credentials()
|
||||||
test_models = litellm.vertex_chat_models + litellm.vertex_code_chat_models + litellm.vertex_text_models + litellm.vertex_code_text_models
|
test_models = litellm.vertex_chat_models + litellm.vertex_code_chat_models + litellm.vertex_text_models + litellm.vertex_code_text_models
|
||||||
test_models = random.sample(test_models, 4)
|
test_models = random.sample(test_models, 1)
|
||||||
test_models += litellm.vertex_language_models # always test gemini-pro
|
test_models += litellm.vertex_language_models # always test gemini-pro
|
||||||
for model in test_models:
|
for model in test_models:
|
||||||
if model in ["code-gecko", "code-gecko@001", "code-gecko@002", "code-gecko@latest", "code-bison@001", "text-bison@001"]:
|
if model in ["code-gecko", "code-gecko@001", "code-gecko@002", "code-gecko@latest", "code-bison@001", "text-bison@001"]:
|
||||||
|
@ -195,6 +195,7 @@ async def test_async_vertexai_streaming_response():
|
||||||
|
|
||||||
def test_gemini_pro_vision():
|
def test_gemini_pro_vision():
|
||||||
try:
|
try:
|
||||||
|
load_vertex_ai_credentials()
|
||||||
litellm.set_verbose = True
|
litellm.set_verbose = True
|
||||||
litellm.num_retries=0
|
litellm.num_retries=0
|
||||||
resp = litellm.completion(
|
resp = litellm.completion(
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue