mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-23 22:03:55 +00:00
This change lets users configure default embedding models at the provider level instead of always relying on system defaults. Each vector store provider can now specify an embedding_model and optional embedding_dimension in their config. Key features: - Auto-dimension lookup for standard models from the registry - Support for Matryoshka embeddings with custom dimensions - Three-tier priority: explicit params > provider config > system fallback - Full backward compatibility - existing setups work unchanged - Comprehensive test coverage with 20 test cases Updated all vector IO providers (FAISS, Chroma, Milvus, Qdrant, etc.) with the new config fields and added detailed documentation with examples. Fixes #2729 |
||
|---|---|---|
| .. | ||
| __init__.py | ||
| chunk_utils.py | ||
| embedding_utils.py | ||