fix responses test

This commit is contained in:
Ishaan Jaff 2025-04-24 21:23:25 -07:00
parent 5a8b35c3c5
commit cb87dbbd51
2 changed files with 14 additions and 3 deletions

View file

@ -7058,6 +7058,17 @@
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models", "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
"supports_tool_choice": true "supports_tool_choice": true
}, },
"command-a-03-2025": {
"max_tokens": 8000,
"max_input_tokens": 256000,
"max_output_tokens": 8000,
"input_cost_per_token": 0.0000025,
"output_cost_per_token": 0.00001,
"litellm_provider": "cohere_chat",
"mode": "chat",
"supports_function_calling": true,
"supports_tool_choice": true
},
"command-r": { "command-r": {
"max_tokens": 4096, "max_tokens": 4096,
"max_input_tokens": 128000, "max_input_tokens": 128000,

View file

@ -73,7 +73,7 @@ def validate_stream_chunk(chunk):
def test_basic_response(): def test_basic_response():
client = get_test_client() client = get_test_client()
response = client.responses.create( response = client.responses.create(
model="gpt-4.0", input="just respond with the word 'ping'" model="gpt-4o", input="just respond with the word 'ping'"
) )
print("basic response=", response) print("basic response=", response)
@ -97,7 +97,7 @@ def test_basic_response():
def test_streaming_response(): def test_streaming_response():
client = get_test_client() client = get_test_client()
stream = client.responses.create( stream = client.responses.create(
model="gpt-4.0", input="just respond with the word 'ping'", stream=True model="gpt-4o", input="just respond with the word 'ping'", stream=True
) )
collected_chunks = [] collected_chunks = []
@ -120,5 +120,5 @@ def test_bad_request_bad_param_error():
with pytest.raises(BadRequestError): with pytest.raises(BadRequestError):
# Trigger error with invalid model name # Trigger error with invalid model name
client.responses.create( client.responses.create(
model="gpt-4.0", input="This should fail", temperature=2000 model="gpt-4o", input="This should fail", temperature=2000
) )