llama-stack/llama_stack/providers
Matthew Farrellee 706b4ca651
feat: support nvidia hosted vision models (llama 3.2 11b/90b) (#1278)
# What does this PR do?

support nvidia hosted 3.2 11b/90b vision models. they are not hosted on
the common https://integrate.api.nvidia.com/v1. they are hosted on their
own individual urls.

## Test Plan

`LLAMA_STACK_BASE_URL=http://localhost:8321 pytest -s -v
tests/client-sdk/inference/test_vision_inference.py
--inference-model=meta/llama-3.2-11b-vision-instruct -k image`
2025-03-18 11:54:10 -07:00
..
inline feat(agent): support multiple tool groups (#1556) 2025-03-17 22:13:09 -07:00
registry feat: added nvidia as safety provider (#1248) 2025-03-17 14:39:23 -07:00
remote feat: support nvidia hosted vision models (llama 3.2 11b/90b) (#1278) 2025-03-18 11:54:10 -07:00
tests refactor(test): introduce --stack-config and simplify options (#1404) 2025-03-05 17:02:02 -08:00
utils fix: agents with non-llama model (#1550) 2025-03-17 22:11:06 -07:00
__init__.py API Updates (#73) 2024-09-17 19:51:35 -07:00
datatypes.py chore: move all Llama Stack types from llama-models to llama-stack (#1098) 2025-02-14 09:10:59 -08:00