fix chatglm (#10540)
This commit is contained in:
parent
2ecd737474
commit
69a28d6b4c
1 changed files with 1 additions and 1 deletions
|
|
@ -184,7 +184,7 @@ def chatglm2_model_forward(
|
|||
def chatglm2_attention_forward(
|
||||
self, hidden_states, attention_mask, rotary_pos_emb, kv_cache=None, use_cache=True
|
||||
):
|
||||
if use_quantize_kv_cache(self.query_key_value, hidden_states):
|
||||
if use_quantize_kv_cache(self.query_key_value, hidden_states.transpose(0, 1)):
|
||||
forward_function = chatglm2_quantized_attention_forward_8eb45c
|
||||
else:
|
||||
forward_function = chatglm2_attention_forward_8eb45c
|
||||
|
|
|
|||
Loading…
Reference in a new issue