forked from phoenix-oss/llama-stack-mirror
Make TGI adapter compatible with HF Inference API (#97)
This commit is contained in:
parent
851c30597a
commit
615ed4bfbc
7 changed files with 122 additions and 96 deletions
|
@ -0,0 +1,10 @@
|
|||
name: local-hf-serverless
|
||||
distribution_spec:
|
||||
description: "Like local, but use Hugging Face Inference API (serverless) for running LLM inference.\nSee https://hf.co/docs/api-inference."
|
||||
providers:
|
||||
inference: remote::hf::serverless
|
||||
memory: meta-reference
|
||||
safety: meta-reference
|
||||
agents: meta-reference
|
||||
telemetry: meta-reference
|
||||
image_type: conda
|
Loading…
Add table
Add a link
Reference in a new issue