llama-stack-mirror/llama_stack/providers/utils/inference
Matthew Farrellee f731f369a2
feat: add infrastructure to allow inference model discovery (#2710)
# What does this PR do?

inference providers each have a static list of supported / known models.
some also have access to a dynamic list of currently available models.
this change gives prodivers using the ModelRegistryHelper the ability to
combine their static and dynamic lists.

for instance, OpenAIInferenceAdapter can implement
```
   def query_available_models(self) -> list[str]:
      return [entry.model for entry in self.openai_client.models.list()]
```
to augment its static list w/ a current list from openai.

## Test Plan

scripts/unit-test.sh
2025-07-14 11:38:53 -07:00
..
__init__.py chore: enable pyupgrade fixes (#1806) 2025-05-01 14:23:50 -07:00
embedding_mixin.py feat: New OpenAI compat embeddings API (#2314) 2025-05-31 22:11:47 -07:00
inference_store.py feat: support auth attributes in inference/responses stores (#2389) 2025-06-20 10:24:45 -07:00
litellm_openai_mixin.py chore: standardize unsupported model error #2517 (#2518) 2025-06-27 14:26:58 -04:00
model_registry.py feat: add infrastructure to allow inference model discovery (#2710) 2025-07-14 11:38:53 -07:00
openai_compat.py chore: remove nested imports (#2515) 2025-06-26 08:01:05 +05:30
prompt_adapter.py fix(ollama): Download remote image URLs for Ollama (#2551) 2025-06-30 20:36:11 +05:30
stream_utils.py feat: drop python 3.10 support (#2469) 2025-06-19 12:07:14 +05:30