forked from phoenix-oss/llama-stack-mirror
Having to run (and re-run) a server while running verifications can be annoying while you are iterating on code. This makes it so you can use the library client -- and because it is OpenAI client compatible, it all works. ## Test Plan ``` pytest -s -v tests/verifications/openai_api/test_responses.py \ --provider=stack:together \ --model meta-llama/Llama-4-Scout-17B-16E-Instruct ``` |
||
|---|---|---|
| .. | ||
| images | ||
| test_cases | ||
| __init__.py | ||
| fixtures.py | ||
| load.py | ||