mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-18 17:08:40 +00:00
fix meta-reference, test vllm
This commit is contained in:
parent
b75e4eb6b9
commit
a30aaaa2e5
3 changed files with 13 additions and 3 deletions
|
|
@ -113,6 +113,7 @@ def inference_vllm_remote() -> ProviderFixture:
|
|||
provider_type="remote::vllm",
|
||||
config=VLLMInferenceAdapterConfig(
|
||||
url=get_env_or_fail("VLLM_URL"),
|
||||
max_tokens=int(os.getenv("VLLM_MAX_TOKENS", 2048)),
|
||||
).model_dump(),
|
||||
)
|
||||
],
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue