llama-stack-mirror/tests/unit/providers
Matthew Farrellee 477bcd4d09
feat: allow dynamic model registration for nvidia inference provider (#2726)
# What does this PR do?

let's users register models available at
https://integrate.api.nvidia.com/v1/models that isn't already in
llama_stack/providers/remote/inference/nvidia/models.py

## Test Plan

1. run the nvidia distro
2. register a model from https://integrate.api.nvidia.com/v1/models that
isn't already know, as of this writing
nvidia/llama-3.1-nemotron-ultra-253b-v1 is a good example
3. perform inference w/ the model
2025-07-17 12:11:30 -07:00
..
agent chore: default to pytest asyncio-mode=auto (#2730) 2025-07-11 13:00:24 -07:00
agents chore: default to pytest asyncio-mode=auto (#2730) 2025-07-11 13:00:24 -07:00
inference chore: block network access from unit tests (#2732) 2025-07-12 16:53:54 -07:00
nvidia feat: allow dynamic model registration for nvidia inference provider (#2726) 2025-07-17 12:11:30 -07:00
utils feat: add infrastructure to allow inference model discovery (#2710) 2025-07-14 11:38:53 -07:00
vector_io fix: SQLiteVecIndex.create(..., bank_id="test_bank.123") - bank_id with a dot - leads to sqlite3.OperationalError (#2770) (#2771) 2025-07-16 08:25:44 -07:00
test_configs.py feat(api): don't return a payload on file delete (#1640) 2025-03-25 17:12:36 -07:00