mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-04 05:12:35 +00:00
# What does this PR do? Simple approach to get some provider pages in the docs. Add or update description fields in the provider configuration class using Pydantic’s Field, ensuring these descriptions are clear and complete, as they will be used to auto-generate provider documentation via ./scripts/distro_codegen.py instead of editing the docs manually. Signed-off-by: Sébastien Han <seb@redhat.com>
673 B
673 B
remote::hf::serverless
Description
HuggingFace Inference API serverless provider for on-demand model inference.
Configuration
Field | Type | Required | Default | Description |
---|---|---|---|---|
huggingface_repo |
<class 'str'> |
No | PydanticUndefined | The model ID of the model on the Hugging Face Hub (e.g. 'meta-llama/Meta-Llama-3.1-70B-Instruct') |
api_token |
pydantic.types.SecretStr | None |
No | Your Hugging Face user access token (will default to locally saved token if not provided) |
Sample Configuration
huggingface_repo: ${env.INFERENCE_MODEL}
api_token: ${env.HF_API_TOKEN}