mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-27 06:28:50 +00:00
980 B
980 B
orphan |
---|
true |
remote::vllm
Description
Remote vLLM inference provider for connecting to vLLM servers.
Configuration
Field | Type | Required | Default | Description |
---|---|---|---|---|
url |
str | None |
No | The URL for the vLLM model serving endpoint | |
max_tokens |
<class 'int'> |
No | 4096 | Maximum number of tokens to generate. |
api_token |
str | None |
No | fake | The API token |
tls_verify |
bool | str |
No | True | Whether to verify TLS certificates. Can be a boolean or a path to a CA certificate file. |
refresh_models |
<class 'bool'> |
No | False | Whether to refresh models periodically |
refresh_models_interval |
<class 'int'> |
No | 300 | Interval in seconds to refresh models |
Sample Configuration
url: ${env.VLLM_URL:=}
max_tokens: ${env.VLLM_MAX_TOKENS:=4096}
api_token: ${env.VLLM_API_TOKEN:=fake}
tls_verify: ${env.VLLM_TLS_VERIFY:=true}