From 98c55b63b40552ad6e4e9a2fcc38ace252024d39 Mon Sep 17 00:00:00 2001 From: Xi Yan Date: Mon, 16 Sep 2024 12:11:59 -0700 Subject: [PATCH] delete distribution registry --- .../distribution_registry/local-ollama.yaml | 7 ------ .../local-plus-fireworks-inference.yaml | 7 ------ .../local-plus-tgi-inference.yaml | 6 ----- .../local-plus-together-inference.yaml | 7 ------ .../distribution_registry/local.yaml | 7 ------ llama_toolchain/core/distribution_registry.py | 23 ------------------- 6 files changed, 57 deletions(-) delete mode 100644 llama_toolchain/configs/distributions/distribution_registry/local-ollama.yaml delete mode 100644 llama_toolchain/configs/distributions/distribution_registry/local-plus-fireworks-inference.yaml delete mode 100644 llama_toolchain/configs/distributions/distribution_registry/local-plus-tgi-inference.yaml delete mode 100644 llama_toolchain/configs/distributions/distribution_registry/local-plus-together-inference.yaml delete mode 100644 llama_toolchain/configs/distributions/distribution_registry/local.yaml delete mode 100644 llama_toolchain/core/distribution_registry.py diff --git a/llama_toolchain/configs/distributions/distribution_registry/local-ollama.yaml b/llama_toolchain/configs/distributions/distribution_registry/local-ollama.yaml deleted file mode 100644 index 4518a7eb7..000000000 --- a/llama_toolchain/configs/distributions/distribution_registry/local-ollama.yaml +++ /dev/null @@ -1,7 +0,0 @@ -description: Like local, but use ollama for running LLM inference -providers: - inference: remote::ollama - safety: meta-reference - agentic_system: meta-reference - memory: meta-reference-faiss - telemetry: console diff --git a/llama_toolchain/configs/distributions/distribution_registry/local-plus-fireworks-inference.yaml b/llama_toolchain/configs/distributions/distribution_registry/local-plus-fireworks-inference.yaml deleted file mode 100644 index 520274e1c..000000000 --- a/llama_toolchain/configs/distributions/distribution_registry/local-plus-fireworks-inference.yaml +++ /dev/null @@ -1,7 +0,0 @@ -description: Use Fireworks.ai for running LLM inference -providers: - inference: remote::fireworks - safety: meta-reference - agentic_system: meta-reference - memory: meta-reference-faiss - telemetry: console diff --git a/llama_toolchain/configs/distributions/distribution_registry/local-plus-tgi-inference.yaml b/llama_toolchain/configs/distributions/distribution_registry/local-plus-tgi-inference.yaml deleted file mode 100644 index a0b36b07e..000000000 --- a/llama_toolchain/configs/distributions/distribution_registry/local-plus-tgi-inference.yaml +++ /dev/null @@ -1,6 +0,0 @@ -description: Use TGI (local or with Hugging Face Inference Endpoints for running LLM inference. When using HF Inference Endpoints, you must provide the name of the endpoint). -providers: - inference: remote::tgi - safety: meta-reference - agentic_system: meta-reference - memory: meta-reference-faiss diff --git a/llama_toolchain/configs/distributions/distribution_registry/local-plus-together-inference.yaml b/llama_toolchain/configs/distributions/distribution_registry/local-plus-together-inference.yaml deleted file mode 100644 index 2681b619c..000000000 --- a/llama_toolchain/configs/distributions/distribution_registry/local-plus-together-inference.yaml +++ /dev/null @@ -1,7 +0,0 @@ -description: Use Together.ai for running LLM inference -providers: - inference: remote::together - safety: meta-reference - agentic_system: meta-reference - memory: meta-reference-faiss - telemetry: console diff --git a/llama_toolchain/configs/distributions/distribution_registry/local.yaml b/llama_toolchain/configs/distributions/distribution_registry/local.yaml deleted file mode 100644 index 03dfb1dab..000000000 --- a/llama_toolchain/configs/distributions/distribution_registry/local.yaml +++ /dev/null @@ -1,7 +0,0 @@ -description: Use code from `llama_toolchain` itself to serve all llama stack APIs -providers: - inference: meta-reference - memory: meta-reference-faiss - safety: meta-reference - agentic_system: meta-reference - telemetry: console diff --git a/llama_toolchain/core/distribution_registry.py b/llama_toolchain/core/distribution_registry.py deleted file mode 100644 index 87ac98d7c..000000000 --- a/llama_toolchain/core/distribution_registry.py +++ /dev/null @@ -1,23 +0,0 @@ -# Copyright (c) Meta Platforms, Inc. and affiliates. -# All rights reserved. -# -# This source code is licensed under the terms described in the LICENSE file in -# the root directory of this source tree. - -from functools import lru_cache -from pathlib import Path -from typing import List, Optional -from .datatypes import * # noqa: F403 -import yaml - - -@lru_cache() -def available_distribution_specs() -> List[DistributionSpec]: - distribution_specs = [] - for p in Path("llama_toolchain/configs/distributions/distribution_registry").rglob( - "*.yaml" - ): - with open(p, "r") as f: - distribution_specs.append(DistributionSpec(**yaml.safe_load(f))) - - return distribution_specs