add fused rms optimization for chatglm model (#9256)

This commit is contained in:
SONG Ge 2023-10-24 14:40:58 +08:00 committed by GitHub
parent b15656229e
commit bfc1e2d733

View file

@ -233,6 +233,9 @@ def _optimize_post(model):
convert_forward(model,
module.CoreAttention,
core_attn_forward_8eb45c)
convert_forward(model,
module.RMSNorm,
llama_rms_norm_forward)
elif hasattr(model.config, 'vocab_size') and model.config.vocab_size == 130528:
# chatglm-6b
modeling_module_name = model.__class__.__module__