mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-17 09:32:36 +00:00
Start auto-generating { build, run, doc.md } for distributions
This commit is contained in:
parent
20bf2f50c2
commit
cfa913fdd5
11 changed files with 362 additions and 23 deletions
|
|
@ -24,3 +24,12 @@ class VLLMInferenceAdapterConfig(BaseModel):
|
|||
default="fake",
|
||||
description="The API token",
|
||||
)
|
||||
|
||||
@classmethod
|
||||
def sample_dict(cls):
|
||||
# TODO: we may need two modes, one for conda and one for docker
|
||||
return {
|
||||
"url": "${env.VLLM_URL:http://host.docker.internal:5100/v1}",
|
||||
"max_tokens": "${env.VLLM_MAX_TOKENS:4096}",
|
||||
"api_token": "${env.VLLM_API_TOKEN:fake}",
|
||||
}
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue