mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
Add default for max_seq_len
This commit is contained in:
parent
9ab27e852b
commit
f5eda1decf
1 changed files with 3 additions and 3 deletions
|
@ -11,10 +11,10 @@ from llama_models.datatypes import ModelFamily
|
||||||
from llama_models.schema_utils import json_schema_type
|
from llama_models.schema_utils import json_schema_type
|
||||||
from llama_models.sku_list import all_registered_models, resolve_model
|
from llama_models.sku_list import all_registered_models, resolve_model
|
||||||
|
|
||||||
from llama_stack.apis.inference import QuantizationConfig
|
|
||||||
|
|
||||||
from pydantic import BaseModel, Field, field_validator
|
from pydantic import BaseModel, Field, field_validator
|
||||||
|
|
||||||
|
from llama_stack.apis.inference import QuantizationConfig
|
||||||
|
|
||||||
|
|
||||||
@json_schema_type
|
@json_schema_type
|
||||||
class MetaReferenceImplConfig(BaseModel):
|
class MetaReferenceImplConfig(BaseModel):
|
||||||
|
@ -24,7 +24,7 @@ class MetaReferenceImplConfig(BaseModel):
|
||||||
)
|
)
|
||||||
quantization: Optional[QuantizationConfig] = None
|
quantization: Optional[QuantizationConfig] = None
|
||||||
torch_seed: Optional[int] = None
|
torch_seed: Optional[int] = None
|
||||||
max_seq_len: int
|
max_seq_len: int = 4096
|
||||||
max_batch_size: int = 1
|
max_batch_size: int = 1
|
||||||
|
|
||||||
@field_validator("model")
|
@field_validator("model")
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue