mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-26 03:08:03 +00:00
update starter
This commit is contained in:
parent
7915551eee
commit
0369dd4191
44 changed files with 351 additions and 670 deletions
|
|
@ -33,7 +33,7 @@ providers:
|
|||
provider_type: remote::together
|
||||
config:
|
||||
url: https://api.together.xyz/v1
|
||||
api_key: ${env.TOGETHER_API_KEY}
|
||||
api_key: ${env.TOGETHER_API_KEY:=}
|
||||
vector_io:
|
||||
- provider_id: sqlite-vec
|
||||
provider_type: inline::sqlite-vec
|
||||
|
|
|
|||
|
|
@ -6,27 +6,13 @@ distribution_spec:
|
|||
- remote::cerebras
|
||||
- remote::ollama
|
||||
- remote::vllm
|
||||
- remote::tgi
|
||||
- remote::hf::serverless
|
||||
- remote::hf::endpoint
|
||||
- remote::fireworks
|
||||
- remote::together
|
||||
- remote::bedrock
|
||||
- remote::databricks
|
||||
- remote::nvidia
|
||||
- remote::runpod
|
||||
- remote::openai
|
||||
- remote::anthropic
|
||||
- remote::gemini
|
||||
- remote::groq
|
||||
- remote::fireworks-openai-compat
|
||||
- remote::llama-openai-compat
|
||||
- remote::together-openai-compat
|
||||
- remote::groq-openai-compat
|
||||
- remote::sambanova-openai-compat
|
||||
- remote::cerebras-openai-compat
|
||||
- remote::sambanova
|
||||
- remote::passthrough
|
||||
- inline::sentence-transformers
|
||||
vector_io:
|
||||
- inline::faiss
|
||||
|
|
|
|||
File diff suppressed because it is too large
Load diff
|
|
@ -113,6 +113,19 @@ from llama_stack.templates.template import (
|
|||
get_model_registry,
|
||||
)
|
||||
|
||||
REMOTE_INFERENCE_PROVIDERS_FOR_STARTER = {
|
||||
"anthropic",
|
||||
"cerebras",
|
||||
"fireworks",
|
||||
"gemini",
|
||||
"groq",
|
||||
"ollama",
|
||||
"openai",
|
||||
"sambanova",
|
||||
"together",
|
||||
"vllm",
|
||||
}
|
||||
|
||||
|
||||
def _get_model_entries_for_provider(provider_type: str) -> list[ProviderModelEntry]:
|
||||
"""Get model entries for a specific provider type."""
|
||||
|
|
@ -207,41 +220,27 @@ def get_remote_inference_providers() -> tuple[list[Provider], dict[str, list[Pro
|
|||
remote_providers = [
|
||||
provider
|
||||
for provider in all_providers
|
||||
# TODO: re-add once the Python 3.13 issue is fixed
|
||||
# discussion: https://github.com/meta-llama/llama-stack/pull/2327#discussion_r2156883828
|
||||
if hasattr(provider, "adapter") and provider.adapter.adapter_type != "watsonx"
|
||||
if hasattr(provider, "adapter") and provider.adapter.adapter_type in REMOTE_INFERENCE_PROVIDERS_FOR_STARTER
|
||||
]
|
||||
|
||||
providers = []
|
||||
inference_providers = []
|
||||
available_models = {}
|
||||
|
||||
for provider_spec in remote_providers:
|
||||
provider_type = provider_spec.adapter.adapter_type
|
||||
|
||||
# Build the environment variable name for enabling this provider
|
||||
env_var = f"ENABLE_{provider_type.upper().replace('-', '_').replace('::', '_')}"
|
||||
model_entries = _get_model_entries_for_provider(provider_type)
|
||||
config = _get_config_for_provider(provider_spec)
|
||||
providers.append(
|
||||
(
|
||||
f"${{env.{env_var}:=__disabled__}}",
|
||||
provider_type,
|
||||
model_entries,
|
||||
config,
|
||||
)
|
||||
)
|
||||
available_models[f"${{env.{env_var}:=__disabled__}}"] = model_entries
|
||||
|
||||
inference_providers = []
|
||||
for provider_id, provider_type, model_entries, config in providers:
|
||||
inference_providers.append(
|
||||
Provider(
|
||||
provider_id=provider_id,
|
||||
provider_id=provider_type,
|
||||
provider_type=f"remote::{provider_type}",
|
||||
config=config,
|
||||
)
|
||||
)
|
||||
available_models[provider_id] = model_entries
|
||||
available_models[provider_type] = model_entries
|
||||
|
||||
return inference_providers, available_models
|
||||
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue