mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 20:14:13 +00:00
Replace MissingEmbeddingModelError with IBM Granite default
- Replace error with ibm-granite/granite-embedding-125m-english default - Based on issue #2418 for commercial compatibility and better UX - Update tests to verify default fallback behavior - Update documentation to reflect new precedence rules - Remove unused MissingEmbeddingModelError class - Update tip section to clarify fallback behavior Resolves review comment to use default instead of error.
This commit is contained in:
parent
380bd1bb7a
commit
8e2675f50c
4 changed files with 13 additions and 16 deletions
|
@ -5,9 +5,6 @@
|
|||
# the root directory of this source tree.
|
||||
|
||||
|
||||
import pytest
|
||||
|
||||
from llama_stack.apis.common.errors import MissingEmbeddingModelError
|
||||
from llama_stack.apis.models import ModelType
|
||||
from llama_stack.core.routers.vector_io import VectorIORouter
|
||||
|
||||
|
@ -72,10 +69,11 @@ async def test_explicit_override(monkeypatch):
|
|||
monkeypatch.delenv("LLAMA_STACK_DEFAULT_EMBEDDING_MODEL", raising=False)
|
||||
|
||||
|
||||
async def test_error_when_no_default():
|
||||
"""Router should raise when neither explicit nor global default is available."""
|
||||
async def test_fallback_to_system_default():
|
||||
"""Router should use system default when neither explicit nor global default is available."""
|
||||
|
||||
router = VectorIORouter(routing_table=_DummyRoutingTable())
|
||||
|
||||
with pytest.raises(MissingEmbeddingModelError):
|
||||
await router._resolve_embedding_model(None)
|
||||
model, dimension = await router._resolve_embedding_model(None)
|
||||
assert model == "ibm-granite/granite-embedding-125m-english"
|
||||
assert dimension == 384
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue