fix: return llama stack model id from embeddings

the openai_embeddings method on OpenAIMixin was returning the provider's model id instead of the llama stack name

before -

$ ./scripts/integration-tests.sh --stack-config server:ci-tests --setup gpt --subdirs inference --inference-mode live --pattern test_openai_embeddings_single_string
...
FAILED tests/integration/inference/test_openai_embeddings.py::test_openai_embeddings_single_string[openai_client-emb=openai/text-embedding-3-small] - AssertionError: assert 'text-embedding-3-small' == 'openai/text-...dding-3-small'
FAILED tests/integration/inference/test_openai_embeddings.py::test_openai_embeddings_single_string[llama_stack_client-emb=openai/text-embedding-3-small] - AssertionError: assert 'text-embedding-3-small' == 'openai/text-...dding-3-small'
========================================== 2 failed, 95 deselected, 4 warnings in 3.87s ===========================================

after -

$ ./scripts/integration-tests.sh --stack-config server:ci-tests --setup gpt --subdirs inference --inference-mode live --pattern test_openai_embeddings_single_string
...
========================================== 2 passed, 95 deselected, 4 warnings in 2.12s ===========================================
This commit is contained in:
Matthew Farrellee 2025-09-23 12:16:09 -04:00
parent a7f9ce9a3a
commit 2ad1633d0b
2 changed files with 2 additions and 2 deletions

View file

@ -296,7 +296,7 @@ class OpenAIMixin(ABC):
return OpenAIEmbeddingsResponse(
data=data,
model=response.model,
model=model,
usage=usage,
)

View file

@ -87,7 +87,7 @@ SETUP_DEFINITIONS: dict[str, Setup] = {
description="OpenAI GPT models for high-quality responses and tool calling",
defaults={
"text_model": "openai/gpt-4o",
"embedding_model": "sentence-transformers/all-MiniLM-L6-v2",
"embedding_model": "openai/text-embedding-3-small",
},
),
"tgi": Setup(