From 367de141f25ecb314e0412420a7f1cf742b20bbb Mon Sep 17 00:00:00 2001 From: binbin Deng <108676127+plusbang@users.noreply.github.com> Date: Mon, 27 May 2024 09:50:54 +0800 Subject: [PATCH] Fix mixtral-8x7b with transformers=4.37.0 (#11132) --- python/llm/src/ipex_llm/transformers/models/mixtral.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/python/llm/src/ipex_llm/transformers/models/mixtral.py b/python/llm/src/ipex_llm/transformers/models/mixtral.py index c2013540..8223ad61 100644 --- a/python/llm/src/ipex_llm/transformers/models/mixtral.py +++ b/python/llm/src/ipex_llm/transformers/models/mixtral.py @@ -414,6 +414,9 @@ def mixtral_model_forward( output_router_logits: Optional[bool] = None, return_dict: Optional[bool] = None, ) -> Union[Tuple, MoeModelOutputWithPast]: + # to be compatible with transformers>=4.37.0 + self._use_flash_attention_2 = self.config._attn_implementation == "flash_attention_2" + output_attentions = output_attentions if output_attentions is not None \ else self.config.output_attentions output_router_logits = (