llama-stack-mirror/tests/unit/providers/nvidia
Matthew Farrellee 477bcd4d09
feat: allow dynamic model registration for nvidia inference provider (#2726)
# What does this PR do?

let's users register models available at
https://integrate.api.nvidia.com/v1/models that isn't already in
llama_stack/providers/remote/inference/nvidia/models.py

## Test Plan

1. run the nvidia distro
2. register a model from https://integrate.api.nvidia.com/v1/models that
isn't already know, as of this writing
nvidia/llama-3.1-nemotron-ultra-253b-v1 is a good example
3. perform inference w/ the model
2025-07-17 12:11:30 -07:00
..
__init__.py feat: Add nemo customizer (#1448) 2025-03-25 11:01:10 -07:00
conftest.py chore: re-enable isort enforcement (#1802) 2025-03-26 15:22:17 -07:00
test_datastore.py feat: Add NVIDIA NeMo datastore (#1852) 2025-04-28 09:41:59 -07:00
test_eval.py feat: Add NVIDIA Eval integration (#1890) 2025-04-24 17:12:42 -07:00
test_parameters.py fix: Pass model parameter as config name to NeMo Customizer (#2218) 2025-05-20 09:51:39 -07:00
test_safety.py chore: remove nested imports (#2515) 2025-06-26 08:01:05 +05:30
test_supervised_fine_tuning.py feat: allow dynamic model registration for nvidia inference provider (#2726) 2025-07-17 12:11:30 -07:00