[NPU] Fix regression caused by layer_norm change (#12627)
This commit is contained in:
parent
90f6709486
commit
46eeab4479
1 changed files with 5 additions and 1 deletions
|
|
@ -472,7 +472,11 @@ class LLMBaseNNFactory(NNFactory):
|
||||||
)
|
)
|
||||||
eps = self.constant(self.rms_norm_eps)
|
eps = self.constant(self.rms_norm_eps)
|
||||||
hidden_states = self.eltwise_div(hidden_states, self.sqrt(self.eltwise_add(variance, eps)))
|
hidden_states = self.eltwise_div(hidden_states, self.sqrt(self.eltwise_add(variance, eps)))
|
||||||
|
if os.environ.get("IPEX_LLM_NPU_DRIVER_VERSION", None) in ["5716", "5733"]:
|
||||||
|
# to support special drivers
|
||||||
hidden_states = self.convert_to_fp16(hidden_states)
|
hidden_states = self.convert_to_fp16(hidden_states)
|
||||||
|
else:
|
||||||
|
layernorm_weight = self.convert_to_fp32(layernorm_weight)
|
||||||
hidden_states = self.eltwise_mul(layernorm_weight, hidden_states)
|
hidden_states = self.eltwise_mul(layernorm_weight, hidden_states)
|
||||||
hidden_states = self.convert_to_fp16(hidden_states)
|
hidden_states = self.convert_to_fp16(hidden_states)
|
||||||
return hidden_states
|
return hidden_states
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue