diff --git a/llama_stack/providers/datatypes.py b/llama_stack/providers/datatypes.py index 4837d9b38..25c967812 100644 --- a/llama_stack/providers/datatypes.py +++ b/llama_stack/providers/datatypes.py @@ -9,7 +9,7 @@ from typing import Any, List, Optional, Protocol from urllib.parse import urlparse from llama_models.schema_utils import json_schema_type -from llama_stack.apis import post_training +from pydantic import BaseModel, Field from llama_stack.apis.datasets import Dataset from llama_stack.apis.eval_tasks import EvalTask @@ -17,7 +17,6 @@ from llama_stack.apis.memory_banks.memory_banks import MemoryBank from llama_stack.apis.models import Model from llama_stack.apis.scoring_functions import ScoringFn from llama_stack.apis.shields import Shield -from pydantic import BaseModel, Field @json_schema_type diff --git a/llama_stack/providers/inline/post_training/meta_reference/datasets/sft.py b/llama_stack/providers/inline/post_training/meta_reference/datasets/sft.py index c035fbedb..1f91dc73f 100644 --- a/llama_stack/providers/inline/post_training/meta_reference/datasets/sft.py +++ b/llama_stack/providers/inline/post_training/meta_reference/datasets/sft.py @@ -1,3 +1,9 @@ +# Copyright (c) Meta Platforms, Inc. and affiliates. +# All rights reserved. +# +# This source code is licensed under the terms described in the LICENSE file in +# the root directory of this source tree. + # Copyright (c) Meta Platforms, Inc. and affiliates. # All rights reserved. # diff --git a/llama_stack/providers/inline/post_training/meta_reference/recipes/lora_finetuning_single_device.py b/llama_stack/providers/inline/post_training/meta_reference/recipes/lora_finetuning_single_device.py index 94951d121..c6677d6d2 100644 --- a/llama_stack/providers/inline/post_training/meta_reference/recipes/lora_finetuning_single_device.py +++ b/llama_stack/providers/inline/post_training/meta_reference/recipes/lora_finetuning_single_device.py @@ -4,20 +4,17 @@ # This source code is licensed under the terms described in the LICENSE file in # the root directory of this source tree. -import asyncio import logging import os -import re from functools import partial from pathlib import Path -from typing import Any, Dict, List, Optional, Tuple, Union +from typing import Any, Dict, List, Optional, Tuple import torch from llama_models.sku_list import resolve_model from llama_stack.apis.datasetio import DatasetIO from torch import nn from torchtune import utils as torchtune_utils -from torchtune.training.checkpointing._utils import ModelType from llama_stack.apis.post_training import * # noqa from llama_stack.apis.post_training import PostTrainingSFTRequest from llama_stack.distribution.utils.model_utils import model_local_dir