llama-stack-mirror/distributions/nutanix/README.md
2024-12-03 01:49:07 +00:00

40 lines
1.2 KiB
Markdown

# Nutanix Distribution
The `llamastack/distribution-nutanix` distribution consists of the following provider configurations.
| **API** | **Inference** | **Agents** | **Memory** | **Safety** | **Telemetry** |
|----------------- |--------------- |---------------- |-------------------------------------------------- |---------------- |---------------- |
| **Provider(s)** | remote::nutanix | meta-reference | meta-reference | meta-reference | meta-reference |
### Start the Distribution (Hosted remote)
> [!NOTE]
> This assumes you have an hosted Nutanix AI endpoint and an API Key.
1. Clone the repo
```
git clone git@github.com:meta-llama/llama-stack.git
cd llama-stack
```
2. Config the model name
Please adjust the `NUTANIX_SUPPORTED_MODELS` variable at line 29 in `llama_stack/providers/adapters/inference/nutanix/nutanix.py` according to your deployment.
3. Build the distrbution
```
pip install -e .
llama stack build --template nutanix --name ntnx --image-type conda
```
4. Set the endpoint URL and API Key
```
llama stack configure ntnx
```
5. Serve and enjoy!
```
llama stack run ntnx --port 174
```