llama-stack-mirror/docs/_static
Ben Browning fa34468308 feat: File search tool for Responses API
This is an initial working prototype of wiring up the `file_search` builtin
tool for the Responses API to our existing rag knowledge search tool.

I stubbed in a new test (that uses a hardcoded url hybrid of the OpenAI
and Llama Stack clients for now, only until we finish landing the vector
store APIs and insertion support).

Note that this is currently under tests/verification only because it
sometimes flakes with tool calling of the small Llama-3.2-3B model we run
in CI (and that I use as an example below). We'd want to make the test a
bit more robust in some way if we moved this over to tests/integration and
ran it in CI.

```
ollama run llama3.2:3b

INFERENCE_MODEL="meta-llama/Llama-3.2-3B-Instruct" \
llama stack run ./llama_stack/templates/ollama/run.yaml \
  --image-type venv \
  --env OLLAMA_URL="http://0.0.0.0:11434"

pytest -sv 'tests/verifications/openai_api/test_responses.py::test_response_non_streaming_file_search' \
  --base-url=http://localhost:8321/v1/openai/v1 \
  --model meta-llama/Llama-3.2-3B-Instruct
```

Signed-off-by: Ben Browning <bbrownin@redhat.com>
2025-06-13 09:36:04 -04:00
..
css fix: improve Mermaid diagram visibility in dark mode (#2092) 2025-05-02 13:09:45 -07:00
js chore: Fix to persist the theme preference across page navigation. (#1974) 2025-04-16 13:58:25 -07:00
providers/vector_io docs: Document sqlite-vec faiss comparison (#1821) 2025-03-28 17:41:33 +01:00
llama-stack-logo.png first version of readthedocs (#278) 2024-10-22 10:15:58 +05:30
llama-stack-spec.html feat: File search tool for Responses API 2025-06-13 09:36:04 -04:00
llama-stack-spec.yaml feat: File search tool for Responses API 2025-06-13 09:36:04 -04:00
llama-stack.png Make a new llama stack image 2024-11-22 23:49:22 -08:00
remote_or_local.gif [docs] update documentations (#356) 2024-11-04 16:52:38 -08:00
safety_system.webp [Docs] Zero-to-Hero notebooks and quick start documentation (#368) 2024-11-08 17:16:44 -08:00