From 85ae8999648edabf18dffb85b39d039dfa3c8d41 Mon Sep 17 00:00:00 2001 From: Aidan Do Date: Tue, 26 Nov 2024 10:49:48 +0000 Subject: [PATCH] . --- 1.py | 10 ++++++---- 1 file changed, 6 insertions(+), 4 deletions(-) diff --git a/1.py b/1.py index d63e91883..36b469b14 100644 --- a/1.py +++ b/1.py @@ -14,19 +14,21 @@ client = LlamaStackClient(base_url=f"http://localhost:{os.environ['LLAMA_STACK_P class CompletionMessage(BaseModel): - content: str - additional_info: str + recipe_name: str + ingredients: list[str] + steps: list[str] response = client.inference.chat_completion( model_id=os.environ["INFERENCE_MODEL"], messages=[ - {"role": "system", "content": "You are a helpful assistant."}, - {"role": "user", "content": "Write a haiku about coding"}, + {"role": "system", "content": "You are a chef."}, + {"role": "user", "content": "Give me a recipe for spaghetti bolognaise"}, ], response_format={ "type": "json_schema", "json_schema": CompletionMessage.model_json_schema(), }, + sampling_params={"max_tokens": 8000}, ) print(response.completion_message.content)