mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-12 13:57:57 +00:00
move distribution to yaml files
This commit is contained in:
parent
3802d5acdc
commit
0068d059db
9 changed files with 78 additions and 49 deletions
|
@ -1,7 +1,7 @@
|
|||
name: local-tgi-conda-example
|
||||
distribution_spec:
|
||||
distribution_type: local-plus-tgi-inference
|
||||
description: Use TGI for running LLM inference
|
||||
description: Use TGI (local or with Hugging Face Inference Endpoints for running LLM inference. When using HF Inference Endpoints, you must provide the name of the endpoint).
|
||||
docker_image: null
|
||||
providers:
|
||||
inference: remote::tgi
|
||||
|
|
|
@ -0,0 +1,9 @@
|
|||
distribution_type: local-ollama
|
||||
description: Like local, but use ollama for running LLM inference
|
||||
docker_image: null
|
||||
providers:
|
||||
inference: remote::ollama
|
||||
safety: meta-reference
|
||||
agentic_system: meta-reference
|
||||
memory: meta-reference-faiss
|
||||
telemetry: console
|
|
@ -0,0 +1,9 @@
|
|||
distribution_type: local-plus-fireworks-inference
|
||||
description: Use Fireworks.ai for running LLM inference
|
||||
docker_image: null
|
||||
providers:
|
||||
inference: remote::fireworks
|
||||
safety: meta-reference
|
||||
agentic_system: meta-reference
|
||||
memory: meta-reference-faiss
|
||||
telemetry: console
|
|
@ -0,0 +1,8 @@
|
|||
distribution_type: local-plus-tgi-inference
|
||||
description: Use TGI (local or with Hugging Face Inference Endpoints for running LLM inference. When using HF Inference Endpoints, you must provide the name of the endpoint).
|
||||
docker_image: null
|
||||
providers:
|
||||
inference: remote::tgi
|
||||
safety: meta-reference
|
||||
agentic_system: meta-reference
|
||||
memory: meta-reference-faiss
|
|
@ -0,0 +1,9 @@
|
|||
distribution_type: local-plus-together-inference
|
||||
description: Use Together.ai for running LLM inference
|
||||
docker_image: null
|
||||
providers:
|
||||
inference: remote::together
|
||||
safety: meta-reference
|
||||
agentic_system: meta-reference
|
||||
memory: meta-reference-faiss
|
||||
telemetry: console
|
|
@ -0,0 +1,9 @@
|
|||
distribution_type: local
|
||||
description: Use code from `llama_toolchain` itself to serve all llama stack APIs
|
||||
docker_image: null
|
||||
providers:
|
||||
inference: meta-reference
|
||||
memory: meta-reference-faiss
|
||||
safety: meta-reference
|
||||
agentic_system: meta-reference
|
||||
telemetry: console
|
21
llama_toolchain/configs/distributions/run.py
Normal file
21
llama_toolchain/configs/distributions/run.py
Normal file
|
@ -0,0 +1,21 @@
|
|||
from llama_toolchain.core.distribution_registry import *
|
||||
import json
|
||||
|
||||
import fire
|
||||
import yaml
|
||||
from llama_toolchain.common.serialize import EnumEncoder
|
||||
|
||||
|
||||
def main():
|
||||
for d in available_distribution_specs():
|
||||
file_path = "./configs/distributions/distribution_registry/{}.yaml".format(
|
||||
d.distribution_type
|
||||
)
|
||||
|
||||
with open(file_path, "w") as f:
|
||||
to_write = json.loads(json.dumps(d.dict(), cls=EnumEncoder))
|
||||
f.write(yaml.dump(to_write, sort_keys=False))
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
fire.Fire(main)
|
Loading…
Add table
Add a link
Reference in a new issue