From b7b764f8c87b3d143b9b351a7deead123a040c98 Mon Sep 17 00:00:00 2001 From: Dinesh Yeduguru Date: Mon, 25 Nov 2024 10:06:39 -0800 Subject: [PATCH] make TGI_API_TOKEN optional in fixture --- llama_stack/providers/tests/inference/fixtures.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama_stack/providers/tests/inference/fixtures.py b/llama_stack/providers/tests/inference/fixtures.py index 683dbe772..a427eef12 100644 --- a/llama_stack/providers/tests/inference/fixtures.py +++ b/llama_stack/providers/tests/inference/fixtures.py @@ -166,7 +166,7 @@ def inference_tgi() -> ProviderFixture: provider_type="remote::tgi", config=TGIImplConfig( url=get_env_or_fail("TGI_URL"), - api_token=get_env_or_fail("TGI_API_TOKEN"), + api_token=os.getenv("TGI_API_TOKEN", None), ).model_dump(), ) ],