From a9e4d1f00e579d0b3bf5396764e7147f7e480095 Mon Sep 17 00:00:00 2001 From: Sajikumar JS Date: Sat, 26 Apr 2025 18:29:47 +0530 Subject: [PATCH] pre-commit updates --- llama_stack/providers/remote/inference/watsonx/watsonx.py | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/llama_stack/providers/remote/inference/watsonx/watsonx.py b/llama_stack/providers/remote/inference/watsonx/watsonx.py index ef0ca41e1..fa9cc4391 100644 --- a/llama_stack/providers/remote/inference/watsonx/watsonx.py +++ b/llama_stack/providers/remote/inference/watsonx/watsonx.py @@ -29,12 +29,12 @@ from llama_stack.apis.inference import ( ToolPromptFormat, ) from llama_stack.apis.inference.inference import ( + GreedySamplingStrategy, OpenAIChatCompletion, OpenAIChatCompletionChunk, OpenAICompletion, OpenAIMessageParam, OpenAIResponseFormatParam, - GreedySamplingStrategy, TopKSamplingStrategy, TopPSamplingStrategy, ) @@ -242,8 +242,7 @@ class WatsonXInferenceAdapter(Inference, ModelRegistryHelper): if isinstance(request.sampling_params.strategy, GreedySamplingStrategy): input_dict["params"][GenParams.TEMPERATURE] = 0.0 - - input_dict["params"][GenParams.STOP_SEQUENCES] = ["<|endoftext|>"] + input_dict["params"][GenParams.STOP_SEQUENCES] = ["<|endoftext|>"] params = { **input_dict,