forked from phoenix-oss/llama-stack-mirror
# What does this PR do? We are setting a default value of json for tool prompt format, which conflicts with llama 3.2/3.3 models since they use python list. This PR changes the defaults to None and in the code, we infer default based on the model. Addresses: #695 Tests: ❯ LLAMA_STACK_BASE_URL=http://localhost:5000 pytest -v tests/client-sdk/inference/test_inference.py -k "test_text_chat_completion" pytest llama_stack/providers/tests/inference/test_prompt_adapter.py |
||
---|---|---|
.. | ||
agentic-system.png | ||
list-templates.png | ||
llama-stack-spec.html | ||
llama-stack-spec.yaml | ||
llama-stack.png | ||
model-lifecycle.png | ||
prompt-format.png |