llama-stack-mirror/docs/source/distributions/self_hosted_distro
Ben Browning 8ede67b809 More work on file_search verification test
This gets the file_search verification test working against ollama,
fireworks, and api.openai.com. We don't have the entirety of the
vector store API implemented in Llama Stack yet, so this still has a
bit of a hack to swap between using only OpenAI-compatible APIs versus
using the LlamaStackClient to insert content into our vector stores.

Outside of actually inserting file contents, the rest of the test
works the same and uses only the OpenAI client for all of these providers.

How to run the tests:

Ollama (sometimes flakes with small model):

```
ollama run llama3.2:3b

INFERENCE_MODEL="meta-llama/Llama-3.2-3B-Instruct" \
llama stack run ./llama_stack/templates/ollama/run.yaml \
  --image-type venv \
  --env OLLAMA_URL="http://0.0.0.0:11434"

pytest -sv \
  'tests/verifications/openai_api/test_responses.py::test_response_non_streaming_file_search' \
  --base-url=http://localhost:8321/v1/openai/v1 \
  --model meta-llama/Llama-3.2-3B-Instruct
```

Fireworks via Llama Stack:

```
llama stack run llama_stack/templates/fireworks/run.yaml

pytest -sv \
  'tests/verifications/openai_api/test_responses.py::test_response_non_streaming_file_search' \
  --base-url=http://localhost:8321/v1/openai/v1 \
  --model meta-llama/Llama-3.3-70B-Instruct
```

OpenAI directly:

```
pytest -sv \
  'tests/verifications/openai_api/test_responses.py::test_response_non_streaming_file_search' \
  --base-url=https://api.openai.com/v1 \
  --model gpt-4o
```

Signed-off-by: Ben Browning <bbrownin@redhat.com>
2025-06-13 09:36:04 -04:00
..
bedrock.md fix: remove code interpeter implementation (#2087) 2025-05-01 14:35:08 -07:00
cerebras.md fix: replace all instances of --yaml-config with --config (#2196) 2025-05-16 14:31:12 -07:00
dell-tgi.md fix: docker run with --pull always to fetch the latest image (#1733) 2025-03-20 15:35:48 -07:00
dell.md fix: replace all instances of --yaml-config with --config (#2196) 2025-05-16 14:31:12 -07:00
fireworks.md feat: reference implementation for files API (#2330) 2025-06-02 21:54:24 -07:00
groq.md fix: remove code interpeter implementation (#2087) 2025-05-01 14:35:08 -07:00
meta-reference-gpu.md fix: remove code interpeter implementation (#2087) 2025-05-01 14:35:08 -07:00
nvidia.md fix: replace all instances of --yaml-config with --config (#2196) 2025-05-16 14:31:12 -07:00
ollama.md More work on file_search verification test 2025-06-13 09:36:04 -04:00
passthrough.md fix: remove code interpeter implementation (#2087) 2025-05-01 14:35:08 -07:00
remote-vllm.md fix: replace all instances of --yaml-config with --config (#2196) 2025-05-16 14:31:12 -07:00
sambanova.md feat(providers): sambanova safety provider (#2221) 2025-05-21 15:33:02 -07:00
tgi.md fix: replace all instances of --yaml-config with --config (#2196) 2025-05-16 14:31:12 -07:00
together.md fix: revert "feat(provider): adding llama4 support in together inference provider (#2123)" (#2124) 2025-05-08 15:18:16 -07:00