mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 15:23:51 +00:00
remove templates from name, name templates
This commit is contained in:
parent
f17c2a6ff6
commit
d3d66ba40b
7 changed files with 7 additions and 7 deletions
|
@ -133,7 +133,7 @@ class StackBuild(Subcommand):
|
|||
|
||||
# eventually, this should query a registry at llama.meta.com/llamastack/distributions
|
||||
headers = [
|
||||
"Distribution Type",
|
||||
"Template Name",
|
||||
"Providers",
|
||||
"Image Type",
|
||||
"Description",
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
name: local-conda-template
|
||||
name: local-conda
|
||||
distribution_spec:
|
||||
description: Use code from `llama_stack` itself to serve all llama stack APIs
|
||||
providers:
|
|
@ -1,4 +1,4 @@
|
|||
name: local-docker-template
|
||||
name: local-docker
|
||||
distribution_spec:
|
||||
description: Use code from `llama_stack` itself to serve all llama stack APIs
|
||||
providers:
|
|
@ -1,4 +1,4 @@
|
|||
name: local-fireworks-conda-template
|
||||
name: local-fireworks-conda
|
||||
distribution_spec:
|
||||
description: Use Fireworks.ai for running LLM inference
|
||||
providers:
|
|
@ -1,4 +1,4 @@
|
|||
name: local-ollama-conda-template
|
||||
name: local-ollama-conda
|
||||
distribution_spec:
|
||||
description: Like local, but use ollama for running LLM inference
|
||||
providers:
|
|
@ -1,4 +1,4 @@
|
|||
name: local-tgi-conda-template
|
||||
name: local-tgi-conda
|
||||
distribution_spec:
|
||||
description: Use TGI (local or with Hugging Face Inference Endpoints for running LLM inference. When using HF Inference Endpoints, you must provide the name of the endpoint).
|
||||
providers:
|
|
@ -1,4 +1,4 @@
|
|||
name: local-tgi-conda-template
|
||||
name: local-tgi-conda
|
||||
distribution_spec:
|
||||
description: Use Together.ai for running LLM inference
|
||||
providers:
|
Loading…
Add table
Add a link
Reference in a new issue