From 8658949bbb56e7af143c36f51a3a6e4754e6fd16 Mon Sep 17 00:00:00 2001 From: Matthew Farrellee Date: Mon, 6 Oct 2025 09:08:09 -0400 Subject: [PATCH] update anthropic, databricks, tgi, together after get_models rename --- .../providers/remote/inference/anthropic/anthropic.py | 2 +- .../remote/inference/databricks/databricks.py | 10 +--------- llama_stack/providers/remote/inference/tgi/tgi.py | 2 +- .../providers/remote/inference/together/together.py | 2 +- 4 files changed, 4 insertions(+), 12 deletions(-) diff --git a/llama_stack/providers/remote/inference/anthropic/anthropic.py b/llama_stack/providers/remote/inference/anthropic/anthropic.py index 29dd3d3b1..3b996b16e 100644 --- a/llama_stack/providers/remote/inference/anthropic/anthropic.py +++ b/llama_stack/providers/remote/inference/anthropic/anthropic.py @@ -35,5 +35,5 @@ class AnthropicInferenceAdapter(OpenAIMixin): def get_base_url(self): return "https://api.anthropic.com/v1" - async def get_models(self) -> Iterable[str] | None: + async def list_provider_model_ids(self) -> Iterable[str]: return [m.id async for m in AsyncAnthropic(api_key=self.get_api_key()).models.list()] diff --git a/llama_stack/providers/remote/inference/databricks/databricks.py b/llama_stack/providers/remote/inference/databricks/databricks.py index 3b110f21a..f4ad1be94 100644 --- a/llama_stack/providers/remote/inference/databricks/databricks.py +++ b/llama_stack/providers/remote/inference/databricks/databricks.py @@ -33,7 +33,7 @@ class DatabricksInferenceAdapter(OpenAIMixin): def get_base_url(self) -> str: return f"{self.config.url}/serving-endpoints" - async def get_models(self) -> list[str] | None: + async def list_provider_model_ids(self) -> Iterable[str]: return [ endpoint.name for endpoint in WorkspaceClient( @@ -68,11 +68,3 @@ class DatabricksInferenceAdapter(OpenAIMixin): suffix: str | None = None, ) -> OpenAICompletion: raise NotImplementedError() - - async def list_provider_model_ids(self) -> Iterable[str]: - return [ - endpoint.name - for endpoint in WorkspaceClient( - host=self.config.url, token=self.get_api_key() - ).serving_endpoints.list() # TODO: this is not async - ] diff --git a/llama_stack/providers/remote/inference/tgi/tgi.py b/llama_stack/providers/remote/inference/tgi/tgi.py index 53c872c02..a316e8996 100644 --- a/llama_stack/providers/remote/inference/tgi/tgi.py +++ b/llama_stack/providers/remote/inference/tgi/tgi.py @@ -35,7 +35,7 @@ class _HfAdapter(OpenAIMixin): def get_base_url(self): return self.url - async def get_models(self) -> Iterable[str] | None: + async def list_provider_model_ids(self) -> Iterable[str]: return [self.model_id] async def openai_embeddings( diff --git a/llama_stack/providers/remote/inference/together/together.py b/llama_stack/providers/remote/inference/together/together.py index d19e85f09..20669bef9 100644 --- a/llama_stack/providers/remote/inference/together/together.py +++ b/llama_stack/providers/remote/inference/together/together.py @@ -59,7 +59,7 @@ class TogetherInferenceAdapter(OpenAIMixin, NeedsRequestProviderData): together_api_key = provider_data.together_api_key return AsyncTogether(api_key=together_api_key) - async def get_models(self) -> Iterable[str] | None: + async def list_provider_model_ids(self) -> Iterable[str]: # Together's /v1/models is not compatible with OpenAI's /v1/models. Together support ticket #13355 -> will not fix, use Together's own client return [m.id for m in await self._get_client().models.list()]