forked from phoenix-oss/llama-stack-mirror
Update getting_started.md
This commit is contained in:
parent
d9147f3184
commit
ed4272e31e
1 changed files with 3 additions and 0 deletions
|
@ -285,6 +285,9 @@ INFO: Uvicorn running on http://[::]:5000 (Press CTRL+C to quit)
|
||||||
> [!IMPORTANT]
|
> [!IMPORTANT]
|
||||||
> The "local" distribution inference server currently only supports CUDA. It will not work on Apple Silicon machines.
|
> The "local" distribution inference server currently only supports CUDA. It will not work on Apple Silicon machines.
|
||||||
|
|
||||||
|
> [!TIP]
|
||||||
|
> You might need to use the flag `--disable-ipv6` to Disable IPv6 support
|
||||||
|
|
||||||
This server is running a Llama model locally.
|
This server is running a Llama model locally.
|
||||||
|
|
||||||
## Step 4. Test with Client
|
## Step 4. Test with Client
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue