LLM: fix error of 'AI-ModelScope/phi-2' hosted by ModelScope hub (#10364)

This commit is contained in:
binbin Deng 2024-03-11 16:19:17 +08:00 committed by GitHub
parent fe27a6971c
commit dbcfc5c2fa

View file

@ -1071,7 +1071,9 @@ def _optimize_post(model, lightweight_bmm=False):
convert_forward(model, convert_forward(model,
module.MixtralBLockSparseTop2MLP, module.MixtralBLockSparseTop2MLP,
mixtral_mlp_forward) mixtral_mlp_forward)
elif model.config.model_type == "phi-msft": elif model.config.model_type == "phi-msft" and \
hasattr(model.config, "num_local_experts"):
# For phixtral, limit the condition to avoid applying on phi-2 hosted by ModelScope
modeling_module_name = model.__class__.__module__ modeling_module_name = model.__class__.__module__
module = importlib.import_module(modeling_module_name) module = importlib.import_module(modeling_module_name)
from bigdl.llm.transformers.models.phixtral import phixtral_moeblock_forward, \ from bigdl.llm.transformers.models.phixtral import phixtral_moeblock_forward, \