LLM: fix ChatGLMConfig check (#9191)
This commit is contained in:
parent
7160afd4d1
commit
1cd9ab15b8
1 changed files with 3 additions and 2 deletions
|
|
@ -190,7 +190,8 @@ def optimize(model):
|
||||||
convert_forward(model,
|
convert_forward(model,
|
||||||
module.SelfAttention,
|
module.SelfAttention,
|
||||||
chatglm2_32k_attention_forward)
|
chatglm2_32k_attention_forward)
|
||||||
elif model.config.padded_vocab_size == 65024:
|
elif hasattr(model.config, 'padded_vocab_size') and \
|
||||||
|
model.config.padded_vocab_size == 65024:
|
||||||
# chatglm2-6b
|
# chatglm2-6b
|
||||||
modeling_module_name = model.__class__.__module__
|
modeling_module_name = model.__class__.__module__
|
||||||
module = importlib.import_module(modeling_module_name)
|
module = importlib.import_module(modeling_module_name)
|
||||||
|
|
@ -203,7 +204,7 @@ def optimize(model):
|
||||||
convert_forward(model,
|
convert_forward(model,
|
||||||
module.CoreAttention,
|
module.CoreAttention,
|
||||||
core_attn_forward_8eb45c)
|
core_attn_forward_8eb45c)
|
||||||
elif model.config.vocab_size == 130528:
|
elif hasattr(model.config, 'vocab_size') and model.config.vocab_size == 130528:
|
||||||
# chatglm-6b
|
# chatglm-6b
|
||||||
modeling_module_name = model.__class__.__module__
|
modeling_module_name = model.__class__.__module__
|
||||||
module = importlib.import_module(modeling_module_name)
|
module = importlib.import_module(modeling_module_name)
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue