mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-27 18:50:41 +00:00
Kill llama stack configure
(#371)
* remove configure * build msg * wip * build->run * delete prints * docs * fix docs, kill configure * precommit * update fireworks build * docs * clean up build * comments * fix * test * remove baking build.yaml into docker * fix msg, urls * configure msg
This commit is contained in:
parent
d289afdbde
commit
748606195b
11 changed files with 248 additions and 401 deletions
|
@ -61,49 +61,7 @@
|
|||
"```\n",
|
||||
"For GPU inference, you need to set these environment variables for specifying local directory containing your model checkpoints, and enable GPU inference to start running docker container.\n",
|
||||
"$ export LLAMA_CHECKPOINT_DIR=~/.llama\n",
|
||||
"$ llama stack configure llamastack-meta-reference-gpu\n",
|
||||
"```\n",
|
||||
"Follow the prompts as part of configure.\n",
|
||||
"Here is a sample output \n",
|
||||
"```\n",
|
||||
"$ llama stack configure llamastack-meta-reference-gpu\n",
|
||||
"\n",
|
||||
"Could not find ~/.conda/envs/llamastack-llamastack-meta-reference-gpu/llamastack-meta-reference-gpu-build.yaml. Trying docker image name instead...\n",
|
||||
"+ podman run --network host -it -v ~/.llama/builds/docker:/app/builds llamastack-meta-reference-gpu llama stack configure ./llamastack-build.yaml --output-dir /app/builds\n",
|
||||
"\n",
|
||||
"Configuring API `inference`...\n",
|
||||
"=== Configuring provider `meta-reference` for API inference...\n",
|
||||
"Enter value for model (default: Llama3.1-8B-Instruct) (required): Llama3.2-11B-Vision-Instruct\n",
|
||||
"Do you want to configure quantization? (y/n): n\n",
|
||||
"Enter value for torch_seed (optional): \n",
|
||||
"Enter value for max_seq_len (default: 4096) (required): \n",
|
||||
"Enter value for max_batch_size (default: 1) (required): \n",
|
||||
"\n",
|
||||
"Configuring API `safety`...\n",
|
||||
"=== Configuring provider `meta-reference` for API safety...\n",
|
||||
"Do you want to configure llama_guard_shield? (y/n): n\n",
|
||||
"Do you want to configure prompt_guard_shield? (y/n): n\n",
|
||||
"\n",
|
||||
"Configuring API `agents`...\n",
|
||||
"=== Configuring provider `meta-reference` for API agents...\n",
|
||||
"Enter `type` for persistence_store (options: redis, sqlite, postgres) (default: sqlite): \n",
|
||||
"\n",
|
||||
"Configuring SqliteKVStoreConfig:\n",
|
||||
"Enter value for namespace (optional): \n",
|
||||
"Enter value for db_path (default: /root/.llama/runtime/kvstore.db) (required): \n",
|
||||
"\n",
|
||||
"Configuring API `memory`...\n",
|
||||
"=== Configuring provider `meta-reference` for API memory...\n",
|
||||
"> Please enter the supported memory bank type your provider has for memory: vector\n",
|
||||
"\n",
|
||||
"Configuring API `telemetry`...\n",
|
||||
"=== Configuring provider `meta-reference` for API telemetry...\n",
|
||||
"\n",
|
||||
"> YAML configuration has been written to /app/builds/local-gpu-run.yaml.\n",
|
||||
"You can now run `llama stack run local-gpu --port PORT`\n",
|
||||
"YAML configuration has been written to /home/hjshah/.llama/builds/docker/local-gpu-run.yaml. You can now run `llama stack run /home/hjshah/.llama/builds/docker/local-gpu-run.yaml`\n",
|
||||
"```\n",
|
||||
"NOTE: For this example, we use all local meta-reference implementations and have not setup safety. \n",
|
||||
"\n",
|
||||
"5. Run the Stack Server\n",
|
||||
"```\n",
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue