# llama (client-side) CLI Reference The `llama-stack-client` CLI allows you to query information about the distribution. ## Basic Commands ### `llama-stack-client` ```bash $ llama-stack-client -h usage: llama-stack-client [-h] {models,memory_banks,shields} ... Welcome to the LlamaStackClient CLI options: -h, --help show this help message and exit subcommands: {models,memory_banks,shields} ``` ### `llama-stack-client configure` ```bash $ llama-stack-client configure > Enter the host name of the Llama Stack distribution server: localhost > Enter the port number of the Llama Stack distribution server: 8321 Done! You can now use the Llama Stack Client CLI with endpoint http://localhost:8321 ``` ### `llama-stack-client providers list` ```bash $ llama-stack-client providers list ``` ``` +-----------+----------------+-----------------+ | API | Provider ID | Provider Type | +===========+================+=================+ | scoring | meta0 | meta-reference | +-----------+----------------+-----------------+ | datasetio | meta0 | meta-reference | +-----------+----------------+-----------------+ | inference | tgi0 | remote::tgi | +-----------+----------------+-----------------+ | memory | meta-reference | meta-reference | +-----------+----------------+-----------------+ | agents | meta-reference | meta-reference | +-----------+----------------+-----------------+ | telemetry | meta-reference | meta-reference | +-----------+----------------+-----------------+ | safety | meta-reference | meta-reference | +-----------+----------------+-----------------+ ``` ## Model Management ### `llama-stack-client models list` ```bash $ llama-stack-client models list ``` ``` +----------------------+----------------------+---------------+----------------------------------------------------------+ | identifier | llama_model | provider_id | metadata | +======================+======================+===============+==========================================================+ | Llama3.1-8B-Instruct | Llama3.1-8B-Instruct | tgi0 | {'huggingface_repo': 'meta-llama/Llama-3.1-8B-Instruct'} | +----------------------+----------------------+---------------+----------------------------------------------------------+ ``` ### `llama-stack-client models get` ```bash $ llama-stack-client models get Llama3.1-8B-Instruct ``` ``` +----------------------+----------------------+----------------------------------------------------------+---------------+ | identifier | llama_model | metadata | provider_id | +======================+======================+==========================================================+===============+ | Llama3.1-8B-Instruct | Llama3.1-8B-Instruct | {'huggingface_repo': 'meta-llama/Llama-3.1-8B-Instruct'} | tgi0 | +----------------------+----------------------+----------------------------------------------------------+---------------+ ``` ```bash $ llama-stack-client models get Random-Model Model RandomModel is not found at distribution endpoint host:port. Please ensure endpoint is serving specified model. ``` ### `llama-stack-client models register` ```bash $ llama-stack-client models register [--provider-id ] [--provider-model-id ] [--metadata ] ``` ### `llama-stack-client models update` ```bash $ llama-stack-client models update [--provider-id ] [--provider-model-id ] [--metadata ] ``` ### `llama-stack-client models delete` ```bash $ llama-stack-client models delete ``` ## Vector DB Management ### `llama-stack-client vector_dbs list` ```bash $ llama-stack-client vector_dbs list ``` ``` +--------------+----------------+---------------------+---------------+------------------------+ | identifier | provider_id | provider_resource_id| vector_db_type| params | +==============+================+=====================+===============+========================+ | test_bank | meta-reference | test_bank | vector | embedding_model: all-MiniLM-L6-v2 embedding_dimension: 384| +--------------+----------------+---------------------+---------------+------------------------+ ``` ### `llama-stack-client vector_dbs register` ```bash $ llama-stack-client vector_dbs register [--provider-id ] [--provider-vector-db-id ] [--embedding-model ] [--embedding-dimension ] ``` Options: - `--provider-id`: Optional. Provider ID for the vector db - `--provider-vector-db-id`: Optional. Provider's vector db ID - `--embedding-model`: Optional. Embedding model to use. Default: "all-MiniLM-L6-v2" - `--embedding-dimension`: Optional. Dimension of embeddings. Default: 384 ### `llama-stack-client vector_dbs unregister` ```bash $ llama-stack-client vector_dbs unregister ``` ## Shield Management ### `llama-stack-client shields list` ```bash $ llama-stack-client shields list ``` ``` +--------------+----------+----------------+-------------+ | identifier | params | provider_id | type | +==============+==========+================+=============+ | llama_guard | {} | meta-reference | llama_guard | +--------------+----------+----------------+-------------+ ``` ### `llama-stack-client shields register` ```bash $ llama-stack-client shields register --shield-id [--provider-id ] [--provider-shield-id ] [--params ] ``` Options: - `--shield-id`: Required. ID of the shield - `--provider-id`: Optional. Provider ID for the shield - `--provider-shield-id`: Optional. Provider's shield ID - `--params`: Optional. JSON configuration parameters for the shield ## Eval Task Management ### `llama-stack-client benchmarks list` ```bash $ llama-stack-client benchmarks list ``` ### `llama-stack-client benchmarks register` ```bash $ llama-stack-client benchmarks register --eval-task-id --dataset-id --scoring-functions [ ...] [--provider-id ] [--provider-eval-task-id ] [--metadata ] ``` Options: - `--eval-task-id`: Required. ID of the eval task - `--dataset-id`: Required. ID of the dataset to evaluate - `--scoring-functions`: Required. One or more scoring functions to use for evaluation - `--provider-id`: Optional. Provider ID for the eval task - `--provider-eval-task-id`: Optional. Provider's eval task ID - `--metadata`: Optional. Metadata for the eval task in JSON format ## Eval execution ### `llama-stack-client eval run-benchmark` ```bash $ llama-stack-client eval run-benchmark [ ...] --eval-task-config --output-dir [--num-examples ] [--visualize] ``` Options: - `--eval-task-config`: Required. Path to the eval task config file in JSON format - `--output-dir`: Required. Path to the directory where evaluation results will be saved - `--num-examples`: Optional. Number of examples to evaluate (useful for debugging) - `--visualize`: Optional flag. If set, visualizes evaluation results after completion Example benchmark_config.json: ```json { "type": "benchmark", "eval_candidate": { "type": "model", "model": "Llama3.1-405B-Instruct", "sampling_params": { "strategy": "greedy", } } } ``` ### `llama-stack-client eval run-scoring` ```bash $ llama-stack-client eval run-scoring --eval-task-config --output-dir [--num-examples ] [--visualize] ``` Options: - `--eval-task-config`: Required. Path to the eval task config file in JSON format - `--output-dir`: Required. Path to the directory where scoring results will be saved - `--num-examples`: Optional. Number of examples to evaluate (useful for debugging) - `--visualize`: Optional flag. If set, visualizes scoring results after completion ## Tool Group Management ### `llama-stack-client toolgroups list` ```bash $ llama-stack-client toolgroups list ``` ``` +---------------------------+------------------+------+---------------+ | identifier | provider_id | args | mcp_endpoint | +===========================+==================+======+===============+ | builtin::code_interpreter | code-interpreter | None | None | +---------------------------+------------------+------+---------------+ | builtin::rag | rag-runtime | None | None | +---------------------------+------------------+------+---------------+ | builtin::websearch | tavily-search | None | None | +---------------------------+------------------+------+---------------+ ``` ### `llama-stack-client toolgroups get` ```bash $ llama-stack-client toolgroups get ``` Shows detailed information about a specific toolgroup. If the toolgroup is not found, displays an error message. ### `llama-stack-client toolgroups register` ```bash $ llama-stack-client toolgroups register [--provider-id ] [--provider-toolgroup-id ] [--mcp-config ] [--args ] ``` Options: - `--provider-id`: Optional. Provider ID for the toolgroup - `--provider-toolgroup-id`: Optional. Provider's toolgroup ID - `--mcp-config`: Optional. JSON configuration for the MCP endpoint - `--args`: Optional. JSON arguments for the toolgroup ### `llama-stack-client toolgroups unregister` ```bash $ llama-stack-client toolgroups unregister ```