mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-27 14:38:49 +00:00
inference providers each have a static list of supported / known models. some also have access to a dynamic list of currently available models. this change gives prodivers using the ModelRegistryHelper the ability to combine their static and dynamic lists. for instance, OpenAIInferenceAdapter can implement ``` def query_available_models(self) -> list[str]: return [entry.model for entry in self.openai_client.models.list()] ``` to augment its static list w/ a current list from openai. |
||
---|---|---|
.. | ||
bedrock | ||
common | ||
datasetio | ||
inference | ||
kvstore | ||
memory | ||
responses | ||
scoring | ||
sqlstore | ||
telemetry | ||
tools | ||
vector_io | ||
__init__.py | ||
pagination.py | ||
scheduler.py |