From 03123f718b510f6c99c882f07a84b433a351b205 Mon Sep 17 00:00:00 2001 From: Hardik Shah Date: Tue, 10 Sep 2024 11:44:08 -0700 Subject: [PATCH] dropped commented code --- llama_toolchain/inference/client.py | 8 -------- 1 file changed, 8 deletions(-) diff --git a/llama_toolchain/inference/client.py b/llama_toolchain/inference/client.py index 1e1dcb8c0..5ba9314bc 100644 --- a/llama_toolchain/inference/client.py +++ b/llama_toolchain/inference/client.py @@ -86,14 +86,6 @@ class InferenceClient(Inference): async def run_main(host: str, port: int, stream: bool): client = InferenceClient(f"http://{host}:{port}") - # from llama_toolchain.inference.meta_reference import get_provider_impl - # from .meta_reference.config import MetaReferenceImplConfig - - # config = MetaReferenceImplConfig( - # model="Meta-Llama3.1-8B-Instruct", - # max_seq_len=4096, - # ) - # client = await get_provider_impl(config, {}) message = UserMessage(content="hello world, troll me in two-paragraphs about 42") cprint(f"User>{message.content}", "green")