diff --git a/python/llm/src/bigdl/llm/transformers/convert.py b/python/llm/src/bigdl/llm/transformers/convert.py index c1902d35..5652b02f 100644 --- a/python/llm/src/bigdl/llm/transformers/convert.py +++ b/python/llm/src/bigdl/llm/transformers/convert.py @@ -233,6 +233,9 @@ def _optimize_post(model): convert_forward(model, module.CoreAttention, core_attn_forward_8eb45c) + convert_forward(model, + module.RMSNorm, + llama_rms_norm_forward) elif hasattr(model.config, 'vocab_size') and model.config.vocab_size == 130528: # chatglm-6b modeling_module_name = model.__class__.__module__