mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 15:23:51 +00:00
make ollama/vllm disabled by default
This commit is contained in:
parent
0369dd4191
commit
b9269a94b9
2 changed files with 18 additions and 5 deletions
|
@ -19,11 +19,11 @@ providers:
|
||||||
config:
|
config:
|
||||||
base_url: https://api.cerebras.ai
|
base_url: https://api.cerebras.ai
|
||||||
api_key: ${env.CEREBRAS_API_KEY:=}
|
api_key: ${env.CEREBRAS_API_KEY:=}
|
||||||
- provider_id: ollama
|
- provider_id: ${env.ENABLE_OLLAMA:=__disabled__}
|
||||||
provider_type: remote::ollama
|
provider_type: remote::ollama
|
||||||
config:
|
config:
|
||||||
url: ${env.OLLAMA_URL:=http://localhost:11434}
|
url: ${env.OLLAMA_URL:=http://localhost:11434}
|
||||||
- provider_id: vllm
|
- provider_id: ${env.ENABLE_VLLM:=__disabled__}
|
||||||
provider_type: remote::vllm
|
provider_type: remote::vllm
|
||||||
config:
|
config:
|
||||||
url: ${env.VLLM_URL:=http://localhost:8000/v1}
|
url: ${env.VLLM_URL:=http://localhost:8000/v1}
|
||||||
|
@ -862,9 +862,9 @@ models:
|
||||||
provider_id: ${env.ENABLE_SENTENCE_TRANSFORMERS:=sentence-transformers}
|
provider_id: ${env.ENABLE_SENTENCE_TRANSFORMERS:=sentence-transformers}
|
||||||
model_type: embedding
|
model_type: embedding
|
||||||
shields:
|
shields:
|
||||||
- shield_id: ollama
|
- shield_id: ${env.ENABLE_OLLAMA:=__disabled__}
|
||||||
provider_id: llama-guard
|
provider_id: llama-guard
|
||||||
provider_shield_id: ollama/${env.SAFETY_MODEL:=llama-guard3:1b}
|
provider_shield_id: ${env.ENABLE_OLLAMA:=__disabled__}/${env.SAFETY_MODEL:=llama-guard3:1b}
|
||||||
- shield_id: fireworks
|
- shield_id: fireworks
|
||||||
provider_id: llama-guard
|
provider_id: llama-guard
|
||||||
provider_shield_id: fireworks/${env.SAFETY_MODEL:=accounts/fireworks/models/llama-guard-3-8b}
|
provider_shield_id: fireworks/${env.SAFETY_MODEL:=accounts/fireworks/models/llama-guard-3-8b}
|
||||||
|
|
|
@ -126,6 +126,11 @@ REMOTE_INFERENCE_PROVIDERS_FOR_STARTER = {
|
||||||
"vllm",
|
"vllm",
|
||||||
}
|
}
|
||||||
|
|
||||||
|
DEFAULT_DISABLED_PROVIDERS = {
|
||||||
|
"ollama",
|
||||||
|
"vllm",
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
def _get_model_entries_for_provider(provider_type: str) -> list[ProviderModelEntry]:
|
def _get_model_entries_for_provider(provider_type: str) -> list[ProviderModelEntry]:
|
||||||
"""Get model entries for a specific provider type."""
|
"""Get model entries for a specific provider type."""
|
||||||
|
@ -232,9 +237,17 @@ def get_remote_inference_providers() -> tuple[list[Provider], dict[str, list[Pro
|
||||||
model_entries = _get_model_entries_for_provider(provider_type)
|
model_entries = _get_model_entries_for_provider(provider_type)
|
||||||
config = _get_config_for_provider(provider_spec)
|
config = _get_config_for_provider(provider_spec)
|
||||||
|
|
||||||
|
if provider_type in DEFAULT_DISABLED_PROVIDERS:
|
||||||
|
# For default disabled providers, we set id to be in the form of ${env.ENABLE_OLLAMA:=__disabled__}
|
||||||
|
# And allow this to be enabled explicitly by user by setting the environment variables
|
||||||
|
env_var = f"ENABLE_{provider_type.upper().replace('-', '_').replace('::', '_')}"
|
||||||
|
provider_id = f"${{env.{env_var}:=__disabled__}}"
|
||||||
|
else:
|
||||||
|
provider_id = provider_type
|
||||||
|
|
||||||
inference_providers.append(
|
inference_providers.append(
|
||||||
Provider(
|
Provider(
|
||||||
provider_id=provider_type,
|
provider_id=provider_id,
|
||||||
provider_type=f"remote::{provider_type}",
|
provider_type=f"remote::{provider_type}",
|
||||||
config=config,
|
config=config,
|
||||||
)
|
)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue