mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-09 21:18:38 +00:00
from models.llama3_1 --> from llama_models.llama3_1
This commit is contained in:
parent
c6ef16f6bd
commit
c64b8cba22
22 changed files with 29 additions and 27 deletions
|
@ -6,7 +6,7 @@ from pydantic import BaseModel, Field
|
|||
from strong_typing.schema import json_schema_type
|
||||
from typing_extensions import Annotated
|
||||
|
||||
from models.llama3_1.api.datatypes import * # noqa: F403
|
||||
from llama_models.llama3_1.api.datatypes import * # noqa: F403
|
||||
|
||||
|
||||
class LogProbConfig(BaseModel):
|
||||
|
|
|
@ -16,11 +16,11 @@ from fairscale.nn.model_parallel.initialize import (
|
|||
initialize_model_parallel,
|
||||
model_parallel_is_initialized,
|
||||
)
|
||||
from models.llama3_1.api.args import ModelArgs
|
||||
from models.llama3_1.api.chat_format import ChatFormat, ModelInput
|
||||
from models.llama3_1.api.datatypes import Message
|
||||
from models.llama3_1.api.model import Transformer
|
||||
from models.llama3_1.api.tokenizer import Tokenizer
|
||||
from llama_models.llama3_1.api.args import ModelArgs
|
||||
from llama_models.llama3_1.api.chat_format import ChatFormat, ModelInput
|
||||
from llama_models.llama3_1.api.datatypes import Message
|
||||
from llama_models.llama3_1.api.model import Transformer
|
||||
from llama_models.llama3_1.api.tokenizer import Tokenizer
|
||||
from termcolor import cprint
|
||||
|
||||
from .api.config import CheckpointType, InlineImplConfig
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
from typing import AsyncGenerator
|
||||
|
||||
from models.llama3_1.api.datatypes import StopReason
|
||||
from llama_models.llama3_1.api.datatypes import StopReason
|
||||
|
||||
from .api.config import (
|
||||
CheckpointQuantizationFormat,
|
||||
|
|
|
@ -2,9 +2,9 @@ from dataclasses import dataclass
|
|||
from functools import partial
|
||||
from typing import Generator, List, Optional
|
||||
|
||||
from models.llama3_1.api.chat_format import ChatFormat
|
||||
from models.llama3_1.api.datatypes import Message
|
||||
from models.llama3_1.api.tokenizer import Tokenizer
|
||||
from llama_models.llama3_1.api.chat_format import ChatFormat
|
||||
from llama_models.llama3_1.api.datatypes import Message
|
||||
from llama_models.llama3_1.api.tokenizer import Tokenizer
|
||||
|
||||
from .api.config import InlineImplConfig
|
||||
from .generation import Llama
|
||||
|
|
|
@ -7,7 +7,7 @@ from typing import Optional
|
|||
import torch
|
||||
|
||||
from fairscale.nn.model_parallel.mappings import reduce_from_model_parallel_region
|
||||
from models.llama3_1.api.model import Transformer, TransformerBlock
|
||||
from llama_models.llama3_1.api.model import Transformer, TransformerBlock
|
||||
|
||||
from termcolor import cprint
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue