From 2df63cc621ab0c8d222374cbbbe3d173c22c1310 Mon Sep 17 00:00:00 2001 From: onukura <26293997+onukura@users.noreply.github.com> Date: Mon, 25 Mar 2024 16:25:22 +0000 Subject: [PATCH] Fix ollama embedding response --- litellm/llms/ollama.py | 1 + 1 file changed, 1 insertion(+) diff --git a/litellm/llms/ollama.py b/litellm/llms/ollama.py index 3611ccd8b..05a3134fc 100644 --- a/litellm/llms/ollama.py +++ b/litellm/llms/ollama.py @@ -396,6 +396,7 @@ async def ollama_aembeddings( response_json = await response.json() embeddings = response_json["embedding"] + embeddings = [embeddings] # Ollama currently does not support batch embedding ## RESPONSE OBJECT output_data = [] for idx, embedding in enumerate(embeddings):