mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-18 13:39:48 +00:00
add embedding model by default
This commit is contained in:
parent
2a9b13dd52
commit
2f88006bd0
43 changed files with 446 additions and 85 deletions
|
|
@ -109,7 +109,7 @@ class InferenceRouter(Inference):
|
|||
model = await self.routing_table.get_model(model_id)
|
||||
if model is None:
|
||||
raise ValueError(f"Model '{model_id}' not found")
|
||||
if model.model_type == ModelType.embedding_model:
|
||||
if model.model_type == ModelType.embedding:
|
||||
raise ValueError(
|
||||
f"Model '{model_id}' is an embedding model and does not support chat completions"
|
||||
)
|
||||
|
|
@ -142,7 +142,7 @@ class InferenceRouter(Inference):
|
|||
model = await self.routing_table.get_model(model_id)
|
||||
if model is None:
|
||||
raise ValueError(f"Model '{model_id}' not found")
|
||||
if model.model_type == ModelType.embedding_model:
|
||||
if model.model_type == ModelType.embedding:
|
||||
raise ValueError(
|
||||
f"Model '{model_id}' is an embedding model and does not support chat completions"
|
||||
)
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue