From af756183488bbdde5e396e562d76460a91960331 Mon Sep 17 00:00:00 2001 From: Xi Yan Date: Mon, 21 Oct 2024 13:23:58 -0700 Subject: [PATCH] remove distribution/templates --- .../templates/build_configs/local-vllm-build.yaml | 10 ---------- .../distribution/templates/remote-vllm-build.yaml | 10 ---------- 2 files changed, 20 deletions(-) delete mode 100644 llama_stack/distribution/templates/build_configs/local-vllm-build.yaml delete mode 100644 llama_stack/distribution/templates/remote-vllm-build.yaml diff --git a/llama_stack/distribution/templates/build_configs/local-vllm-build.yaml b/llama_stack/distribution/templates/build_configs/local-vllm-build.yaml deleted file mode 100644 index e333a137b..000000000 --- a/llama_stack/distribution/templates/build_configs/local-vllm-build.yaml +++ /dev/null @@ -1,10 +0,0 @@ -name: local-vllm -distribution_spec: - description: Like local, but use vLLM for running LLM inference - providers: - inference: vllm - memory: meta-reference - safety: meta-reference - agents: meta-reference - telemetry: meta-reference -image_type: conda \ No newline at end of file diff --git a/llama_stack/distribution/templates/remote-vllm-build.yaml b/llama_stack/distribution/templates/remote-vllm-build.yaml deleted file mode 100644 index 525c3a930..000000000 --- a/llama_stack/distribution/templates/remote-vllm-build.yaml +++ /dev/null @@ -1,10 +0,0 @@ -name: remote-vllm -distribution_spec: - description: Use remote vLLM for running LLM inference - providers: - inference: remote::vllm - memory: meta-reference - safety: meta-reference - agents: meta-reference - telemetry: meta-reference -image_type: docker \ No newline at end of file