mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
chore: return NotImplementedError instead of ValueError
Signed-off-by: Sébastien Han <seb@redhat.com>
This commit is contained in:
parent
3c9a10d2fe
commit
59830e5a22
1 changed files with 2 additions and 2 deletions
|
@ -65,7 +65,7 @@ class SentenceTransformersInferenceImpl(
|
|||
stream: bool | None = False,
|
||||
logprobs: LogProbConfig | None = None,
|
||||
) -> CompletionResponse | AsyncGenerator:
|
||||
raise ValueError("Sentence transformers don't support completion")
|
||||
raise NotImplementedError("Sentence transformers don't support completion")
|
||||
|
||||
async def chat_completion(
|
||||
self,
|
||||
|
@ -80,7 +80,7 @@ class SentenceTransformersInferenceImpl(
|
|||
logprobs: LogProbConfig | None = None,
|
||||
tool_config: ToolConfig | None = None,
|
||||
) -> AsyncGenerator:
|
||||
raise ValueError("Sentence transformers don't support chat completion")
|
||||
raise NotImplementedError("Sentence transformers don't support chat completion")
|
||||
|
||||
async def batch_completion(
|
||||
self,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue