llama-stack/llama_stack/providers/inline/inference
Ashwin Bharambe 81ce39a607
feat(api): Add options for supporting various embedding models (#1192)
We need to support:
- asymmetric embedding models (#934)
- truncation policies (#933)
- varying dimensional output (#932) 

## Test Plan

```bash
$ cd llama_stack/providers/tests/inference
$ pytest -s -v -k fireworks test_embeddings.py \
   --inference-model nomic-ai/nomic-embed-text-v1.5 --env EMBEDDING_DIMENSION=784
$  pytest -s -v -k together test_embeddings.py \
   --inference-model togethercomputer/m2-bert-80M-8k-retrieval --env EMBEDDING_DIMENSION=784
$ pytest -s -v -k ollama test_embeddings.py \
   --inference-model all-minilm:latest --env EMBEDDING_DIMENSION=784
```
2025-02-20 22:27:12 -08:00
..
meta_reference ModelAlias -> ProviderModelEntry 2025-02-20 14:02:36 -08:00
sentence_transformers build: format codebase imports using ruff linter (#1028) 2025-02-13 10:06:21 -08:00
vllm feat(api): Add options for supporting various embedding models (#1192) 2025-02-20 22:27:12 -08:00
__init__.py precommit 2024-11-08 17:58:58 -08:00