mirror of
https://github.com/meta-llama/llama-stack.git
synced 2026-01-02 17:54:31 +00:00
rename quant types to use _mixed naming
This commit is contained in:
parent
b239c57c54
commit
76004eacb4
3 changed files with 11 additions and 11 deletions
|
|
@ -133,9 +133,9 @@ class Llama4Generator:
|
|||
ckpt_dir = model_checkpoint_dir(resolved_model.descriptor())
|
||||
|
||||
if config.quantization:
|
||||
if config.quantization.type == "fp8":
|
||||
if config.quantization.type == "fp8_mixed":
|
||||
quantization_mode = QuantizationMode.fp8_mixed
|
||||
elif config.quantization.type == "int4":
|
||||
elif config.quantization.type == "int4_mixed":
|
||||
quantization_mode = QuantizationMode.int4_mixed
|
||||
elif config.quantization.type == "bf16":
|
||||
quantization_mode = None
|
||||
|
|
@ -226,9 +226,9 @@ class Llama3Generator:
|
|||
ckpt_dir = model_checkpoint_dir(resolved_model.descriptor())
|
||||
|
||||
if config.quantization:
|
||||
if config.quantization.type == "fp8":
|
||||
if config.quantization.type == "fp8_mixed":
|
||||
quantization_mode = QuantizationMode.fp8_mixed
|
||||
elif config.quantization.type == "int4":
|
||||
elif config.quantization.type == "int4_mixed":
|
||||
quantization_mode = QuantizationMode.int4_mixed
|
||||
elif config.quantization.type == "bf16":
|
||||
quantization_mode = None
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue