forked from phoenix-oss/llama-stack-mirror
Having to run (and re-run) a server while running verifications can be annoying while you are iterating on code. This makes it so you can use the library client -- and because it is OpenAI client compatible, it all works. ## Test Plan ``` pytest -s -v tests/verifications/openai_api/test_responses.py \ --provider=stack:together \ --model meta-llama/Llama-4-Scout-17B-16E-Instruct ``` |
||
---|---|---|
.. | ||
fixtures | ||
__init__.py | ||
conftest.py | ||
test_chat_completion.py | ||
test_responses.py |