LLM: fix convert of chatglm (#9190)
This commit is contained in:
parent
af3b575c7e
commit
77afb8796b
1 changed files with 1 additions and 1 deletions
|
|
@ -181,7 +181,7 @@ def optimize(model):
|
||||||
# todo implement 4.28.0 ~ 4.30.2
|
# todo implement 4.28.0 ~ 4.30.2
|
||||||
pass
|
pass
|
||||||
|
|
||||||
if model.config.architectures[0] == "ChatGLMModel":
|
if model.config.architectures is not None and model.config.architectures[0] == "ChatGLMModel":
|
||||||
if model.config.num_layers == 28 and hasattr(model.config, 'rope_ratio'):
|
if model.config.num_layers == 28 and hasattr(model.config, 'rope_ratio'):
|
||||||
# chatglm2-6b-32k
|
# chatglm2-6b-32k
|
||||||
modeling_module_name = model.__class__.__module__
|
modeling_module_name = model.__class__.__module__
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue