From f40cd623060b5d22c009ff261a8da5173b3c0348 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Tue, 8 Oct 2024 16:57:52 -0700 Subject: [PATCH] Test fixes --- llama_stack/apis/inference/client.py | 2 +- llama_stack/providers/tests/inference/test_inference.py | 1 - llama_stack/providers/tests/memory/test_memory.py | 1 - 3 files changed, 1 insertion(+), 3 deletions(-) diff --git a/llama_stack/apis/inference/client.py b/llama_stack/apis/inference/client.py index c7b865ebf..79d2cc02c 100644 --- a/llama_stack/apis/inference/client.py +++ b/llama_stack/apis/inference/client.py @@ -84,7 +84,7 @@ class InferenceClient(Inference): response.raise_for_status() j = response.json() - yield ChatCompletionResponse(**j) + return ChatCompletionResponse(**j) async def _stream_chat_completion( self, request: ChatCompletionRequest diff --git a/llama_stack/providers/tests/inference/test_inference.py b/llama_stack/providers/tests/inference/test_inference.py index b864c2ef4..f52de0df1 100644 --- a/llama_stack/providers/tests/inference/test_inference.py +++ b/llama_stack/providers/tests/inference/test_inference.py @@ -68,7 +68,6 @@ async def inference_settings(request): ModelDef( identifier=model, llama_model=model, - provider_id="", ) ], ) diff --git a/llama_stack/providers/tests/memory/test_memory.py b/llama_stack/providers/tests/memory/test_memory.py index 4351ae699..70f8ba4aa 100644 --- a/llama_stack/providers/tests/memory/test_memory.py +++ b/llama_stack/providers/tests/memory/test_memory.py @@ -67,7 +67,6 @@ def sample_documents(): async def register_memory_bank(memory_impl: Memory): bank = VectorMemoryBankDef( identifier="test_bank", - provider_id="", embedding_model="all-MiniLM-L6-v2", chunk_size_in_tokens=512, overlap_size_in_tokens=64,