mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
fix responses test
This commit is contained in:
parent
5a8b35c3c5
commit
cb87dbbd51
2 changed files with 14 additions and 3 deletions
|
@ -7058,6 +7058,17 @@
|
||||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models",
|
||||||
"supports_tool_choice": true
|
"supports_tool_choice": true
|
||||||
},
|
},
|
||||||
|
"command-a-03-2025": {
|
||||||
|
"max_tokens": 8000,
|
||||||
|
"max_input_tokens": 256000,
|
||||||
|
"max_output_tokens": 8000,
|
||||||
|
"input_cost_per_token": 0.0000025,
|
||||||
|
"output_cost_per_token": 0.00001,
|
||||||
|
"litellm_provider": "cohere_chat",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": true,
|
||||||
|
"supports_tool_choice": true
|
||||||
|
},
|
||||||
"command-r": {
|
"command-r": {
|
||||||
"max_tokens": 4096,
|
"max_tokens": 4096,
|
||||||
"max_input_tokens": 128000,
|
"max_input_tokens": 128000,
|
||||||
|
|
|
@ -73,7 +73,7 @@ def validate_stream_chunk(chunk):
|
||||||
def test_basic_response():
|
def test_basic_response():
|
||||||
client = get_test_client()
|
client = get_test_client()
|
||||||
response = client.responses.create(
|
response = client.responses.create(
|
||||||
model="gpt-4.0", input="just respond with the word 'ping'"
|
model="gpt-4o", input="just respond with the word 'ping'"
|
||||||
)
|
)
|
||||||
print("basic response=", response)
|
print("basic response=", response)
|
||||||
|
|
||||||
|
@ -97,7 +97,7 @@ def test_basic_response():
|
||||||
def test_streaming_response():
|
def test_streaming_response():
|
||||||
client = get_test_client()
|
client = get_test_client()
|
||||||
stream = client.responses.create(
|
stream = client.responses.create(
|
||||||
model="gpt-4.0", input="just respond with the word 'ping'", stream=True
|
model="gpt-4o", input="just respond with the word 'ping'", stream=True
|
||||||
)
|
)
|
||||||
|
|
||||||
collected_chunks = []
|
collected_chunks = []
|
||||||
|
@ -120,5 +120,5 @@ def test_bad_request_bad_param_error():
|
||||||
with pytest.raises(BadRequestError):
|
with pytest.raises(BadRequestError):
|
||||||
# Trigger error with invalid model name
|
# Trigger error with invalid model name
|
||||||
client.responses.create(
|
client.responses.create(
|
||||||
model="gpt-4.0", input="This should fail", temperature=2000
|
model="gpt-4o", input="This should fail", temperature=2000
|
||||||
)
|
)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue