mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
test: Add default vLLM URL in remote-vllm template (#1736)
# What does this PR do? This is to avoid errors like the following when running inference integration tests: ``` ERROR tests/integration/inference/test_text_inference.py::test_text_completion_stop_sequence[txt=8B-inference:completion:stop_sequence] - llama_stack.distribution.stack.EnvVarError: Environment variable 'VLLM_URL' not set or empty at providers.inference[0].config.url ``` It's also good to have a default, which is consistent with vLLM API server. ## Test Plan Integration tests can run without the error above. --------- Signed-off-by: Yuan Tang <terrytangyuan@gmail.com>
This commit is contained in:
parent
03b5c61bfc
commit
dce9a24a6c
3 changed files with 3 additions and 3 deletions
|
@ -15,7 +15,7 @@ providers:
|
||||||
- provider_id: vllm-inference
|
- provider_id: vllm-inference
|
||||||
provider_type: remote::vllm
|
provider_type: remote::vllm
|
||||||
config:
|
config:
|
||||||
url: ${env.VLLM_URL}
|
url: ${env.VLLM_URL:http://localhost:8000/v1}
|
||||||
max_tokens: ${env.VLLM_MAX_TOKENS:4096}
|
max_tokens: ${env.VLLM_MAX_TOKENS:4096}
|
||||||
api_token: ${env.VLLM_API_TOKEN:fake}
|
api_token: ${env.VLLM_API_TOKEN:fake}
|
||||||
tls_verify: ${env.VLLM_TLS_VERIFY:true}
|
tls_verify: ${env.VLLM_TLS_VERIFY:true}
|
||||||
|
|
|
@ -15,7 +15,7 @@ providers:
|
||||||
- provider_id: vllm-inference
|
- provider_id: vllm-inference
|
||||||
provider_type: remote::vllm
|
provider_type: remote::vllm
|
||||||
config:
|
config:
|
||||||
url: ${env.VLLM_URL}
|
url: ${env.VLLM_URL:http://localhost:8000/v1}
|
||||||
max_tokens: ${env.VLLM_MAX_TOKENS:4096}
|
max_tokens: ${env.VLLM_MAX_TOKENS:4096}
|
||||||
api_token: ${env.VLLM_API_TOKEN:fake}
|
api_token: ${env.VLLM_API_TOKEN:fake}
|
||||||
tls_verify: ${env.VLLM_TLS_VERIFY:true}
|
tls_verify: ${env.VLLM_TLS_VERIFY:true}
|
||||||
|
|
|
@ -45,7 +45,7 @@ def get_distribution_template() -> DistributionTemplate:
|
||||||
provider_id="vllm-inference",
|
provider_id="vllm-inference",
|
||||||
provider_type="remote::vllm",
|
provider_type="remote::vllm",
|
||||||
config=VLLMInferenceAdapterConfig.sample_run_config(
|
config=VLLMInferenceAdapterConfig.sample_run_config(
|
||||||
url="${env.VLLM_URL}",
|
url="${env.VLLM_URL:http://localhost:8000/v1}",
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
embedding_provider = Provider(
|
embedding_provider = Provider(
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue