mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 07:14:20 +00:00
touchups
This commit is contained in:
parent
142b36c7c5
commit
86924fd7b1
2 changed files with 5 additions and 36 deletions
39
README.md
39
README.md
|
@ -22,44 +22,13 @@ cd llama-toolchain
|
||||||
pip install -e .
|
pip install -e .
|
||||||
```
|
```
|
||||||
|
|
||||||
## Test with cli
|
## The Llama CLI
|
||||||
|
|
||||||
We have built a llama cli to make it easy to configure / run parts of the toolchain
|
The `llama` CLI makes it easy to configure and run the Llama toolchain. Read the [CLI reference](docs/cli_reference.md) for details.
|
||||||
```
|
|
||||||
llama --help
|
|
||||||
|
|
||||||
usage: llama [-h] {download,inference,model,agentic_system} ...
|
## Appendix: Running FP8
|
||||||
|
|
||||||
Welcome to the llama CLI
|
If you want to run FP8, you need the `fbgemm-gpu` package which requires `torch >= 2.4.0` (currently only in nightly, but releasing shortly...)
|
||||||
|
|
||||||
options:
|
|
||||||
-h, --help show this help message and exit
|
|
||||||
|
|
||||||
subcommands:
|
|
||||||
{download,inference,model,agentic_system}
|
|
||||||
```
|
|
||||||
There are several subcommands to help get you started
|
|
||||||
|
|
||||||
## Start inference server that can run the llama models
|
|
||||||
```bash
|
|
||||||
llama inference configure
|
|
||||||
llama inference start
|
|
||||||
```
|
|
||||||
|
|
||||||
|
|
||||||
## Test client
|
|
||||||
```bash
|
|
||||||
python -m llama_toolchain.inference.client localhost 5000
|
|
||||||
|
|
||||||
Initializing client for http://localhost:5000
|
|
||||||
User>hello world, help me out here
|
|
||||||
Assistant> Hello! I'd be delighted to help you out. What's on your mind? Do you have a question, a problem, or just need someone to chat with? I'm all ears!
|
|
||||||
```
|
|
||||||
|
|
||||||
|
|
||||||
## Running FP8
|
|
||||||
|
|
||||||
You need `fbgemm-gpu` package which requires torch >= 2.4.0 (currently only in nightly, but releasing shortly...).
|
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
ENV=fp8_env
|
ENV=fp8_env
|
||||||
|
|
|
@ -5,7 +5,7 @@ The `llama` CLI tool helps you setup and use the Llama toolchain & agentic syste
|
||||||
```
|
```
|
||||||
$ llama --help
|
$ llama --help
|
||||||
|
|
||||||
Welcome to the Llama Command Line Interface
|
Welcome to the Llama CLI
|
||||||
|
|
||||||
Usage: llama [-h] {download,inference,model} ...
|
Usage: llama [-h] {download,inference,model} ...
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue