mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
Small updates to quantization config
This commit is contained in:
parent
8eceebec98
commit
161aef0aae
2 changed files with 3 additions and 3 deletions
|
@ -41,7 +41,7 @@ class Bf16QuantizationConfig(BaseModel):
|
|||
@json_schema_type
|
||||
class Int4QuantizationConfig(BaseModel):
|
||||
type: Literal[QuantizationType.int4.value] = QuantizationType.int4.value
|
||||
scheme: Optional[str] = None
|
||||
scheme: Optional[str] = "int4_weight_int8_dynamic_activation"
|
||||
|
||||
|
||||
QuantizationConfig = Annotated[
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue