LLM: Fix bigdl_ipex_int8 warning (#10890)
This commit is contained in:
parent
fb3c268d13
commit
3e8ed54270
1 changed files with 1 additions and 1 deletions
|
|
@ -138,7 +138,7 @@ def _ipex_optimize_model(model, rms_classes, qtype):
|
||||||
}
|
}
|
||||||
qconfig = ipex.quantization.get_weight_only_quant_qconfig_mapping(
|
qconfig = ipex.quantization.get_weight_only_quant_qconfig_mapping(
|
||||||
weight_dtype=torch.qint8, # INT8
|
weight_dtype=torch.qint8, # INT8
|
||||||
lowp_mode=ipex.quantization.WoqLowpMode.INT8,
|
lowp_mode=ipex.quantization.WoqLowpMode.BF16,
|
||||||
act_quant_mode=act_quant_mode_dict["PER_IC_BLOCK"],
|
act_quant_mode=act_quant_mode_dict["PER_IC_BLOCK"],
|
||||||
group_size=-1,
|
group_size=-1,
|
||||||
)
|
)
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue