llama-stack/docs/source
Derek Higgins 0e4307de0f
docs: Fix missing --gpu all flag in Docker run commands (#2026)
adding the --gpu all flag to Docker run commands
for meta-reference-gpu distributions ensures models are loaded into GPU
instead of CPU.

Remove docs for meta-reference-quantized-gpu
The distribution was removed in #1887
but these files were left behind.


Fixes: #1798

# What does this PR do?
Fixes doc to add --gpu all command to docker run

[//]: # (If resolving an issue, uncomment and update the line below)
Closes #1798

## Test Plan
[Describe the tests you ran to verify your changes with result
summaries. *Provide clear instructions so the plan can be easily
re-executed.*]

verified in docker documentation but untested

---------

Signed-off-by: Derek Higgins <derekh@redhat.com>
2025-04-25 12:17:31 -07:00
..
building_applications docs: add examples for how to define RAG docs (#1981) 2025-04-23 15:39:18 +02:00
concepts docs: fix typos in evaluation concepts (#1745) 2025-03-21 12:00:53 -07:00
contributing docs: Updating docs to source from CONTRIBUTING.md (#1850) 2025-04-01 14:50:04 +02:00
distributions docs: Fix missing --gpu all flag in Docker run commands (#2026) 2025-04-25 12:17:31 -07:00
getting_started docs: Update docs and fix warning in start-stack.sh (#1937) 2025-04-11 16:26:17 -07:00
introduction docs: Remove mentions of focus on Llama models (#1690) 2025-03-19 00:17:22 -04:00
playground chore: simplify running the demo UI (#1907) 2025-04-09 11:22:29 -07:00
providers chore: rename ramalama provider (#2008) 2025-04-24 09:34:15 +02:00
references feat(api): (1/n) datasets api clean up (#1573) 2025-03-17 16:55:45 -07:00
conf.py chore: Detect browser setting for dark/light mode and set default to light mode (#1913) 2025-04-09 12:40:56 -04:00
index.md docs: fixes to quick start (#1943) 2025-04-11 13:41:23 -07:00