Nutanix AI on!

This commit is contained in:
Jinan Zhou 2024-10-29 22:43:52 +00:00
parent 1e2faa461f
commit 64c5d38ae9
10 changed files with 234 additions and 2 deletions

View file

@ -0,0 +1,40 @@
# Nutanix Distribution
The `llamastack/distribution-nutanix` distribution consists of the following provider configurations.
| **API** | **Inference** | **Agents** | **Memory** | **Safety** | **Telemetry** |
|----------------- |--------------- |---------------- |-------------------------------------------------- |---------------- |---------------- |
| **Provider(s)** | remote::nutanix | meta-reference | meta-reference | meta-reference | meta-reference |
### Start the Distribution (Hosted remote)
> [!NOTE]
> This assumes you have an hosted Nutanix AI endpoint and an API Key.
1. Clone the repo
```
git clone git@github.com:meta-llama/llama-stack.git
cd llama-stack
```
2. Config the model name
Please adjust the `NUTANIX_SUPPORTED_MODELS` variable at line 29 in `llama_stack/providers/adapters/inference/nutanix/nutanix.py` according to your deployment.
3. Build the distrbution
```
pip install -e .
llama stack build --template nutanix --name ntnx --image-type conda
```
4. Set the endpoint URL and API Key
```
llama stack configure ntnx
```
5. Serve and enjoy!
```
llama stack run ntnx --port 174
```

View file

@ -0,0 +1 @@
../../llama_stack/templates/nutanix/build.yaml