llama-stack-mirror/distributions/nutanix/README.md
2024-12-03 01:49:07 +00:00

1.2 KiB

Nutanix Distribution

The llamastack/distribution-nutanix distribution consists of the following provider configurations.

API Inference Agents Memory Safety Telemetry
Provider(s) remote::nutanix meta-reference meta-reference meta-reference meta-reference

Start the Distribution (Hosted remote)

Note

This assumes you have an hosted Nutanix AI endpoint and an API Key.

  1. Clone the repo
git clone git@github.com:meta-llama/llama-stack.git
cd llama-stack
  1. Config the model name

Please adjust the NUTANIX_SUPPORTED_MODELS variable at line 29 in llama_stack/providers/adapters/inference/nutanix/nutanix.py according to your deployment.

  1. Build the distrbution
pip install -e .
llama stack build --template nutanix --name ntnx --image-type conda
  1. Set the endpoint URL and API Key
llama stack configure ntnx
  1. Serve and enjoy!
llama stack run ntnx --port 174