mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-26 17:02:00 +00:00
update starter
This commit is contained in:
parent
7915551eee
commit
0369dd4191
44 changed files with 351 additions and 670 deletions
|
|
@ -113,6 +113,19 @@ from llama_stack.templates.template import (
|
|||
get_model_registry,
|
||||
)
|
||||
|
||||
REMOTE_INFERENCE_PROVIDERS_FOR_STARTER = {
|
||||
"anthropic",
|
||||
"cerebras",
|
||||
"fireworks",
|
||||
"gemini",
|
||||
"groq",
|
||||
"ollama",
|
||||
"openai",
|
||||
"sambanova",
|
||||
"together",
|
||||
"vllm",
|
||||
}
|
||||
|
||||
|
||||
def _get_model_entries_for_provider(provider_type: str) -> list[ProviderModelEntry]:
|
||||
"""Get model entries for a specific provider type."""
|
||||
|
|
@ -207,41 +220,27 @@ def get_remote_inference_providers() -> tuple[list[Provider], dict[str, list[Pro
|
|||
remote_providers = [
|
||||
provider
|
||||
for provider in all_providers
|
||||
# TODO: re-add once the Python 3.13 issue is fixed
|
||||
# discussion: https://github.com/meta-llama/llama-stack/pull/2327#discussion_r2156883828
|
||||
if hasattr(provider, "adapter") and provider.adapter.adapter_type != "watsonx"
|
||||
if hasattr(provider, "adapter") and provider.adapter.adapter_type in REMOTE_INFERENCE_PROVIDERS_FOR_STARTER
|
||||
]
|
||||
|
||||
providers = []
|
||||
inference_providers = []
|
||||
available_models = {}
|
||||
|
||||
for provider_spec in remote_providers:
|
||||
provider_type = provider_spec.adapter.adapter_type
|
||||
|
||||
# Build the environment variable name for enabling this provider
|
||||
env_var = f"ENABLE_{provider_type.upper().replace('-', '_').replace('::', '_')}"
|
||||
model_entries = _get_model_entries_for_provider(provider_type)
|
||||
config = _get_config_for_provider(provider_spec)
|
||||
providers.append(
|
||||
(
|
||||
f"${{env.{env_var}:=__disabled__}}",
|
||||
provider_type,
|
||||
model_entries,
|
||||
config,
|
||||
)
|
||||
)
|
||||
available_models[f"${{env.{env_var}:=__disabled__}}"] = model_entries
|
||||
|
||||
inference_providers = []
|
||||
for provider_id, provider_type, model_entries, config in providers:
|
||||
inference_providers.append(
|
||||
Provider(
|
||||
provider_id=provider_id,
|
||||
provider_id=provider_type,
|
||||
provider_type=f"remote::{provider_type}",
|
||||
config=config,
|
||||
)
|
||||
)
|
||||
available_models[provider_id] = model_entries
|
||||
available_models[provider_type] = model_entries
|
||||
|
||||
return inference_providers, available_models
|
||||
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue