Add default for max_seq_len

This commit is contained in:
Ashwin Bharambe 2024-09-18 21:59:10 -07:00
parent 9ab27e852b
commit f5eda1decf

View file

@ -11,10 +11,10 @@ from llama_models.datatypes import ModelFamily
from llama_models.schema_utils import json_schema_type from llama_models.schema_utils import json_schema_type
from llama_models.sku_list import all_registered_models, resolve_model from llama_models.sku_list import all_registered_models, resolve_model
from llama_stack.apis.inference import QuantizationConfig
from pydantic import BaseModel, Field, field_validator from pydantic import BaseModel, Field, field_validator
from llama_stack.apis.inference import QuantizationConfig
@json_schema_type @json_schema_type
class MetaReferenceImplConfig(BaseModel): class MetaReferenceImplConfig(BaseModel):
@ -24,7 +24,7 @@ class MetaReferenceImplConfig(BaseModel):
) )
quantization: Optional[QuantizationConfig] = None quantization: Optional[QuantizationConfig] = None
torch_seed: Optional[int] = None torch_seed: Optional[int] = None
max_seq_len: int max_seq_len: int = 4096
max_batch_size: int = 1 max_batch_size: int = 1
@field_validator("model") @field_validator("model")