mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 20:14:13 +00:00
Adapter -> Provider
This commit is contained in:
parent
db3e6dda07
commit
65a9e40174
15 changed files with 119 additions and 110 deletions
|
@ -7,12 +7,12 @@
|
|||
from functools import lru_cache
|
||||
from typing import List, Optional
|
||||
|
||||
from llama_toolchain.agentic_system.adapters import available_agentic_system_adapters
|
||||
from llama_toolchain.agentic_system.providers import available_agentic_system_providers
|
||||
|
||||
from llama_toolchain.inference.adapters import available_inference_adapters
|
||||
from llama_toolchain.safety.adapters import available_safety_adapters
|
||||
from llama_toolchain.inference.providers import available_inference_providers
|
||||
from llama_toolchain.safety.providers import available_safety_providers
|
||||
|
||||
from .datatypes import Api, Distribution, PassthroughApiAdapter
|
||||
from .datatypes import Api, Distribution, RemoteProviderSpec
|
||||
|
||||
# This is currently duplicated from `requirements.txt` with a few minor changes
|
||||
# dev-dependencies like "ufmt" etc. are nuked. A few specialized dependencies
|
||||
|
@ -49,10 +49,10 @@ def client_module(api: Api) -> str:
|
|||
return f"llama_toolchain.{api.value}.client"
|
||||
|
||||
|
||||
def passthrough(api: Api, port: int) -> PassthroughApiAdapter:
|
||||
return PassthroughApiAdapter(
|
||||
def remote(api: Api, port: int) -> RemoteProviderSpec:
|
||||
return RemoteProviderSpec(
|
||||
api=api,
|
||||
adapter_id=f"{api.value}-passthrough",
|
||||
provider_id=f"{api.value}-remote",
|
||||
base_url=f"http://localhost:{port}",
|
||||
module=client_module(api),
|
||||
)
|
||||
|
@ -60,25 +60,28 @@ def passthrough(api: Api, port: int) -> PassthroughApiAdapter:
|
|||
|
||||
@lru_cache()
|
||||
def available_distributions() -> List[Distribution]:
|
||||
inference_adapters_by_id = {a.adapter_id: a for a in available_inference_adapters()}
|
||||
safety_adapters_by_id = {a.adapter_id: a for a in available_safety_adapters()}
|
||||
agentic_system_adapters_by_id = {
|
||||
a.adapter_id: a for a in available_agentic_system_adapters()
|
||||
inference_providers_by_id = {
|
||||
a.provider_id: a for a in available_inference_providers()
|
||||
}
|
||||
safety_providers_by_id = {a.provider_id: a for a in available_safety_providers()}
|
||||
agentic_system_providers_by_id = {
|
||||
a.provider_id: a for a in available_agentic_system_providers()
|
||||
}
|
||||
|
||||
return [
|
||||
Distribution(
|
||||
name="local-source",
|
||||
name="local-inline",
|
||||
description="Use code from `llama_toolchain` itself to serve all llama stack APIs",
|
||||
additional_pip_packages=COMMON_DEPENDENCIES,
|
||||
adapters={
|
||||
Api.inference: inference_adapters_by_id["meta-reference"],
|
||||
Api.safety: safety_adapters_by_id["meta-reference"],
|
||||
Api.agentic_system: agentic_system_adapters_by_id["meta-reference"],
|
||||
provider_specs={
|
||||
Api.inference: inference_providers_by_id["meta-reference"],
|
||||
Api.safety: safety_providers_by_id["meta-reference"],
|
||||
Api.agentic_system: agentic_system_providers_by_id["meta-reference"],
|
||||
},
|
||||
),
|
||||
# NOTE: this hardcodes the ports to which things point to
|
||||
Distribution(
|
||||
name="full-passthrough",
|
||||
name="full-remote",
|
||||
description="Point to remote services for all llama stack APIs",
|
||||
additional_pip_packages=[
|
||||
"python-dotenv",
|
||||
|
@ -94,20 +97,20 @@ def available_distributions() -> List[Distribution]:
|
|||
"pydantic_core==2.18.2",
|
||||
"uvicorn",
|
||||
],
|
||||
adapters={
|
||||
Api.inference: passthrough(Api.inference, 5001),
|
||||
Api.safety: passthrough(Api.safety, 5001),
|
||||
Api.agentic_system: passthrough(Api.agentic_system, 5001),
|
||||
provider_specs={
|
||||
Api.inference: remote(Api.inference, 5001),
|
||||
Api.safety: remote(Api.safety, 5001),
|
||||
Api.agentic_system: remote(Api.agentic_system, 5001),
|
||||
},
|
||||
),
|
||||
Distribution(
|
||||
name="local-ollama",
|
||||
description="Like local-source, but use ollama for running LLM inference",
|
||||
additional_pip_packages=COMMON_DEPENDENCIES,
|
||||
adapters={
|
||||
Api.inference: inference_adapters_by_id["meta-ollama"],
|
||||
Api.safety: safety_adapters_by_id["meta-reference"],
|
||||
Api.agentic_system: agentic_system_adapters_by_id["meta-reference"],
|
||||
provider_specs={
|
||||
Api.inference: inference_providers_by_id["meta-ollama"],
|
||||
Api.safety: safety_providers_by_id["meta-reference"],
|
||||
Api.agentic_system: agentic_system_providers_by_id["meta-reference"],
|
||||
},
|
||||
),
|
||||
]
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue