fix meta-reference, test vllm

This commit is contained in:
Ashwin Bharambe 2024-12-16 23:45:15 -08:00
parent b75e4eb6b9
commit a30aaaa2e5
3 changed files with 13 additions and 3 deletions

View file

@ -113,6 +113,7 @@ def inference_vllm_remote() -> ProviderFixture:
provider_type="remote::vllm",
config=VLLMInferenceAdapterConfig(
url=get_env_or_fail("VLLM_URL"),
max_tokens=int(os.getenv("VLLM_MAX_TOKENS", 2048)),
).model_dump(),
)
],