diff --git a/llama_stack/providers/adapters/inference/vllm/__init__.py b/llama_stack/providers/adapters/inference/vllm/__init__.py index 0bec0071a..f4588a307 100644 --- a/llama_stack/providers/adapters/inference/vllm/__init__.py +++ b/llama_stack/providers/adapters/inference/vllm/__init__.py @@ -9,9 +9,7 @@ from .vllm import VLLMInferenceAdapter async def get_adapter_impl(config: VLLMImplConfig, _deps): - assert isinstance( - config, VLLMImplConfig - ), f"Unexpected config type: {type(config)}" + assert isinstance(config, VLLMImplConfig), f"Unexpected config type: {type(config)}" impl = VLLMInferenceAdapter(config) await impl.initialize() - return impl \ No newline at end of file + return impl diff --git a/llama_stack/providers/adapters/inference/vllm/vllm.py b/llama_stack/providers/adapters/inference/vllm/vllm.py index 5ddcefe89..d832e145a 100644 --- a/llama_stack/providers/adapters/inference/vllm/vllm.py +++ b/llama_stack/providers/adapters/inference/vllm/vllm.py @@ -50,12 +50,12 @@ class VLLMInferenceAdapter(ModelRegistryHelper, Inference): pass def completion( - self, - model: str, - content: InterleavedTextMedia, - sampling_params: Optional[SamplingParams] = SamplingParams(), - stream: Optional[bool] = False, - logprobs: Optional[LogProbConfig] = None, + self, + model: str, + content: InterleavedTextMedia, + sampling_params: Optional[SamplingParams] = SamplingParams(), + stream: Optional[bool] = False, + logprobs: Optional[LogProbConfig] = None, ) -> Union[CompletionResponse, CompletionResponseStreamChunk]: raise NotImplementedError()