update starter

This commit is contained in:
Hardik Shah 2025-07-09 17:40:21 -07:00
parent 7915551eee
commit 0369dd4191
44 changed files with 351 additions and 670 deletions

View file

@ -6,27 +6,13 @@ distribution_spec:
- remote::cerebras
- remote::ollama
- remote::vllm
- remote::tgi
- remote::hf::serverless
- remote::hf::endpoint
- remote::fireworks
- remote::together
- remote::bedrock
- remote::databricks
- remote::nvidia
- remote::runpod
- remote::openai
- remote::anthropic
- remote::gemini
- remote::groq
- remote::fireworks-openai-compat
- remote::llama-openai-compat
- remote::together-openai-compat
- remote::groq-openai-compat
- remote::sambanova-openai-compat
- remote::cerebras-openai-compat
- remote::sambanova
- remote::passthrough
- inline::sentence-transformers
vector_io:
- inline::faiss

File diff suppressed because it is too large Load diff

View file

@ -113,6 +113,19 @@ from llama_stack.templates.template import (
get_model_registry,
)
REMOTE_INFERENCE_PROVIDERS_FOR_STARTER = {
"anthropic",
"cerebras",
"fireworks",
"gemini",
"groq",
"ollama",
"openai",
"sambanova",
"together",
"vllm",
}
def _get_model_entries_for_provider(provider_type: str) -> list[ProviderModelEntry]:
"""Get model entries for a specific provider type."""
@ -207,41 +220,27 @@ def get_remote_inference_providers() -> tuple[list[Provider], dict[str, list[Pro
remote_providers = [
provider
for provider in all_providers
# TODO: re-add once the Python 3.13 issue is fixed
# discussion: https://github.com/meta-llama/llama-stack/pull/2327#discussion_r2156883828
if hasattr(provider, "adapter") and provider.adapter.adapter_type != "watsonx"
if hasattr(provider, "adapter") and provider.adapter.adapter_type in REMOTE_INFERENCE_PROVIDERS_FOR_STARTER
]
providers = []
inference_providers = []
available_models = {}
for provider_spec in remote_providers:
provider_type = provider_spec.adapter.adapter_type
# Build the environment variable name for enabling this provider
env_var = f"ENABLE_{provider_type.upper().replace('-', '_').replace('::', '_')}"
model_entries = _get_model_entries_for_provider(provider_type)
config = _get_config_for_provider(provider_spec)
providers.append(
(
f"${{env.{env_var}:=__disabled__}}",
provider_type,
model_entries,
config,
)
)
available_models[f"${{env.{env_var}:=__disabled__}}"] = model_entries
inference_providers = []
for provider_id, provider_type, model_entries, config in providers:
inference_providers.append(
Provider(
provider_id=provider_id,
provider_id=provider_type,
provider_type=f"remote::{provider_type}",
config=config,
)
)
available_models[provider_id] = model_entries
available_models[provider_type] = model_entries
return inference_providers, available_models