forked from phoenix/litellm-mirror
add test cases for checking role in streaming reqs
This commit is contained in:
parent
bad1f7ead1
commit
de37c99021
1 changed files with 15 additions and 3 deletions
|
@ -106,6 +106,8 @@ def test_completion_claude_stream():
|
||||||
for chunk in response:
|
for chunk in response:
|
||||||
print(chunk["choices"][0]["delta"]) # same as openai format
|
print(chunk["choices"][0]["delta"]) # same as openai format
|
||||||
print(chunk["choices"][0]["finish_reason"])
|
print(chunk["choices"][0]["finish_reason"])
|
||||||
|
print(chunk["choices"][0]["delta"]["content"])
|
||||||
|
print(chunk["choices"][0]["delta"]["role"])
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
pytest.fail(f"Error occurred: {e}")
|
pytest.fail(f"Error occurred: {e}")
|
||||||
# test_completion_claude_stream()
|
# test_completion_claude_stream()
|
||||||
|
@ -138,8 +140,11 @@ def test_completion_nlp_cloud_streaming():
|
||||||
for chunk in response:
|
for chunk in response:
|
||||||
print(chunk["choices"][0]["delta"]["content"]) # same as openai format
|
print(chunk["choices"][0]["delta"]["content"]) # same as openai format
|
||||||
print(chunk["choices"][0]["finish_reason"])
|
print(chunk["choices"][0]["finish_reason"])
|
||||||
|
print(chunk["choices"][0]["delta"]["content"])
|
||||||
|
print(chunk["choices"][0]["delta"]["role"])
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
pytest.fail(f"Error occurred: {e}")
|
pytest.fail(f"Error occurred: {e}")
|
||||||
|
# test_completion_nlp_cloud_streaming()
|
||||||
|
|
||||||
# test_completion_nlp_cloud_streaming()
|
# test_completion_nlp_cloud_streaming()
|
||||||
# def test_completion_hf_api():
|
# def test_completion_hf_api():
|
||||||
|
@ -206,6 +211,8 @@ def test_completion_cohere_stream():
|
||||||
for chunk in response:
|
for chunk in response:
|
||||||
print(chunk["choices"][0]["delta"]) # same as openai format
|
print(chunk["choices"][0]["delta"]) # same as openai format
|
||||||
print(chunk["choices"][0]["finish_reason"])
|
print(chunk["choices"][0]["finish_reason"])
|
||||||
|
print(chunk["choices"][0]["delta"]["content"])
|
||||||
|
print(chunk["choices"][0]["delta"]["role"])
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
pytest.fail(f"Error occurred: {e}")
|
pytest.fail(f"Error occurred: {e}")
|
||||||
# test_completion_cohere_stream()
|
# test_completion_cohere_stream()
|
||||||
|
@ -364,6 +371,8 @@ def test_completion_openai_with_stream():
|
||||||
for chunk in response:
|
for chunk in response:
|
||||||
print(chunk)
|
print(chunk)
|
||||||
print(chunk["choices"][0]["finish_reason"])
|
print(chunk["choices"][0]["finish_reason"])
|
||||||
|
print(chunk["choices"][0]["delta"]["content"])
|
||||||
|
print(chunk["choices"][0]["delta"]["role"])
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
pytest.fail(f"Error occurred: {e}")
|
pytest.fail(f"Error occurred: {e}")
|
||||||
# test_completion_openai_with_stream()
|
# test_completion_openai_with_stream()
|
||||||
|
@ -396,6 +405,8 @@ def test_completion_openai_with_functions():
|
||||||
for chunk in response:
|
for chunk in response:
|
||||||
print(chunk)
|
print(chunk)
|
||||||
print(chunk["choices"][0]["finish_reason"])
|
print(chunk["choices"][0]["finish_reason"])
|
||||||
|
print(chunk["choices"][0]["delta"]["content"])
|
||||||
|
print(chunk["choices"][0]["delta"]["role"])
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
pytest.fail(f"Error occurred: {e}")
|
pytest.fail(f"Error occurred: {e}")
|
||||||
# test_completion_openai_with_functions()
|
# test_completion_openai_with_functions()
|
||||||
|
@ -562,9 +573,10 @@ def test_completion_replicate_llama_stream():
|
||||||
try:
|
try:
|
||||||
response = completion(model=model_name, messages=messages, stream=True)
|
response = completion(model=model_name, messages=messages, stream=True)
|
||||||
# Add any assertions here to check the response
|
# Add any assertions here to check the response
|
||||||
for result in response:
|
for chunk in response:
|
||||||
print(result)
|
print(chunk)
|
||||||
# chunk_text = result['choices'][0]['delta']['content']
|
print(chunk["choices"][0]["delta"]["content"])
|
||||||
|
print(chunk["choices"][0]["delta"]["role"])
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
pytest.fail(f"Error occurred: {e}")
|
pytest.fail(f"Error occurred: {e}")
|
||||||
# test_completion_replicate_llama_stream()
|
# test_completion_replicate_llama_stream()
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue