mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 15:23:51 +00:00
dropped commented code
This commit is contained in:
parent
bdede6d14e
commit
03123f718b
1 changed files with 0 additions and 8 deletions
|
@ -86,14 +86,6 @@ class InferenceClient(Inference):
|
||||||
|
|
||||||
async def run_main(host: str, port: int, stream: bool):
|
async def run_main(host: str, port: int, stream: bool):
|
||||||
client = InferenceClient(f"http://{host}:{port}")
|
client = InferenceClient(f"http://{host}:{port}")
|
||||||
# from llama_toolchain.inference.meta_reference import get_provider_impl
|
|
||||||
# from .meta_reference.config import MetaReferenceImplConfig
|
|
||||||
|
|
||||||
# config = MetaReferenceImplConfig(
|
|
||||||
# model="Meta-Llama3.1-8B-Instruct",
|
|
||||||
# max_seq_len=4096,
|
|
||||||
# )
|
|
||||||
# client = await get_provider_impl(config, {})
|
|
||||||
|
|
||||||
message = UserMessage(content="hello world, troll me in two-paragraphs about 42")
|
message = UserMessage(content="hello world, troll me in two-paragraphs about 42")
|
||||||
cprint(f"User>{message.content}", "green")
|
cprint(f"User>{message.content}", "green")
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue