update the client reference (#864)

# What does this PR do?

Syncs changes from
https://github.com/meta-llama/llama-stack-client-python/pull/96
This commit is contained in:
Dinesh Yeduguru 2025-01-23 15:32:16 -08:00 committed by GitHub
parent a78f1fc70d
commit c570a708bf
No known key found for this signature in database
GPG key ID: B5690EEEBB952194

View file

@ -103,36 +103,35 @@ $ llama-stack-client models update <model_id> [--provider-id <provider_id>] [--p
$ llama-stack-client models delete <model_id> $ llama-stack-client models delete <model_id>
``` ```
## Memory Bank Management ## Vector DB Management
### `llama-stack-client memory_banks list` ### `llama-stack-client vector_dbs list`
```bash ```bash
$ llama-stack-client memory_banks list $ llama-stack-client vector_dbs list
``` ```
``` ```
+--------------+----------------+--------+-------------------+------------------------+--------------------------+ +--------------+----------------+---------------------+---------------+------------------------+
| identifier | provider_id | type | embedding_model | chunk_size_in_tokens | overlap_size_in_tokens | | identifier | provider_id | provider_resource_id| vector_db_type| params |
+==============+================+========+===================+========================+==========================+ +==============+================+=====================+===============+========================+
| test_bank | meta-reference | vector | all-MiniLM-L6-v2 | 512 | 64 | | test_bank | meta-reference | test_bank | vector | embedding_model: all-MiniLM-L6-v2
+--------------+----------------+--------+-------------------+------------------------+--------------------------+ embedding_dimension: 384|
+--------------+----------------+---------------------+---------------+------------------------+
``` ```
### `llama-stack-client memory_banks register` ### `llama-stack-client vector_dbs register`
```bash ```bash
$ llama-stack-client memory_banks register <memory-bank-id> --type <type> [--provider-id <provider-id>] [--provider-memory-bank-id <provider-memory-bank-id>] [--chunk-size <chunk-size>] [--embedding-model <embedding-model>] [--overlap-size <overlap-size>] $ llama-stack-client vector_dbs register <vector-db-id> [--provider-id <provider-id>] [--provider-vector-db-id <provider-vector-db-id>] [--embedding-model <embedding-model>] [--embedding-dimension <embedding-dimension>]
``` ```
Options: Options:
- `--type`: Required. Type of memory bank. Choices: "vector", "keyvalue", "keyword", "graph" - `--provider-id`: Optional. Provider ID for the vector db
- `--provider-id`: Optional. Provider ID for the memory bank - `--provider-vector-db-id`: Optional. Provider's vector db ID
- `--provider-memory-bank-id`: Optional. Provider's memory bank ID - `--embedding-model`: Optional. Embedding model to use. Default: "all-MiniLM-L6-v2"
- `--chunk-size`: Optional. Chunk size in tokens (for vector type). Default: 512 - `--embedding-dimension`: Optional. Dimension of embeddings. Default: 384
- `--embedding-model`: Optional. Embedding model (for vector type). Default: "all-MiniLM-L6-v2"
- `--overlap-size`: Optional. Overlap size in tokens (for vector type). Default: 64
### `llama-stack-client memory_banks unregister` ### `llama-stack-client vector_dbs unregister`
```bash ```bash
$ llama-stack-client memory_banks unregister <memory-bank-id> $ llama-stack-client vector_dbs unregister <vector-db-id>
``` ```
## Shield Management ## Shield Management
@ -200,11 +199,7 @@ Example eval_task_config.json:
"type": "model", "type": "model",
"model": "Llama3.1-405B-Instruct", "model": "Llama3.1-405B-Instruct",
"sampling_params": { "sampling_params": {
"strategy": { "strategy": "greedy",
"type": "greedy"
},
"max_tokens": 0,
"repetition_penalty": 1.0
} }
} }
} }
@ -220,3 +215,44 @@ Options:
- `--output-dir`: Required. Path to the directory where scoring results will be saved - `--output-dir`: Required. Path to the directory where scoring results will be saved
- `--num-examples`: Optional. Number of examples to evaluate (useful for debugging) - `--num-examples`: Optional. Number of examples to evaluate (useful for debugging)
- `--visualize`: Optional flag. If set, visualizes scoring results after completion - `--visualize`: Optional flag. If set, visualizes scoring results after completion
## Tool Group Management
### `llama-stack-client toolgroups list`
```bash
$ llama-stack-client toolgroups list
```
```
+---------------------------+------------------+------+---------------+
| identifier | provider_id | args | mcp_endpoint |
+===========================+==================+======+===============+
| builtin::code_interpreter | code-interpreter | None | None |
+---------------------------+------------------+------+---------------+
| builtin::rag | rag-runtime | None | None |
+---------------------------+------------------+------+---------------+
| builtin::websearch | tavily-search | None | None |
+---------------------------+------------------+------+---------------+
```
### `llama-stack-client toolgroups get`
```bash
$ llama-stack-client toolgroups get <toolgroup_id>
```
Shows detailed information about a specific toolgroup. If the toolgroup is not found, displays an error message.
### `llama-stack-client toolgroups register`
```bash
$ llama-stack-client toolgroups register <toolgroup_id> [--provider-id <provider-id>] [--provider-toolgroup-id <provider-toolgroup-id>] [--mcp-config <mcp-config>] [--args <args>]
```
Options:
- `--provider-id`: Optional. Provider ID for the toolgroup
- `--provider-toolgroup-id`: Optional. Provider's toolgroup ID
- `--mcp-config`: Optional. JSON configuration for the MCP endpoint
- `--args`: Optional. JSON arguments for the toolgroup
### `llama-stack-client toolgroups unregister`
```bash
$ llama-stack-client toolgroups unregister <toolgroup_id>
```