From 46bf0192a1182d0c285bb3c829a9996953725d12 Mon Sep 17 00:00:00 2001 From: Xi Yan Date: Thu, 19 Sep 2024 08:52:33 -0700 Subject: [PATCH] working with ollama & local model --- llama_stack/apis/inference/client.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama_stack/apis/inference/client.py b/llama_stack/apis/inference/client.py index 15954ef57..f5321c628 100644 --- a/llama_stack/apis/inference/client.py +++ b/llama_stack/apis/inference/client.py @@ -90,7 +90,7 @@ async def run_main(host: str, port: int, stream: bool): cprint(f"User>{message.content}", "green") iterator = client.chat_completion( ChatCompletionRequest( - model="ollama-1", + model="Meta-Llama3.1-8B-Instruct", messages=[message], stream=stream, )