add comment

This commit is contained in:
Huang, Xinshengzi 2024-08-22 13:17:13 +08:00
parent 48a827aa07
commit 42398a0045

View file

@ -271,14 +271,9 @@ def baichuan_attention_forward_7b(
# IPEX-LLM OPT: kv cache and quantize kv # IPEX-LLM OPT: kv cache and quantize kv
use_quantize_kv = use_quantize_kv_cache(self.W_pack, hidden_states) use_quantize_kv = use_quantize_kv_cache(self.W_pack, hidden_states)
if use_quantize_kv or (not use_compresskv):
key_states, value_states = update_past_key_value(
past_key_value, key_states, value_states,
kv_seq_len, use_quantize_kv, device
)
past_key_value = (key_states, value_states) if use_cache else None
else: # [CompressKV]
if use_compresskv:
enough_kv_room = is_enough_kv_cache_room_4_36(past_key_value, enough_kv_room = is_enough_kv_cache_room_4_36(past_key_value,
self.layer_idx, self.layer_idx,
q_len) q_len)
@ -286,6 +281,12 @@ def baichuan_attention_forward_7b(
key_states, value_states, self.layer_idx, key_states, value_states, self.layer_idx,
query_states, attention_mask, 1, query_states, attention_mask, 1,
self.config, enough_kv_room, KV_CACHE_ALLOC_BLOCK_LENGTH) self.config, enough_kv_room, KV_CACHE_ALLOC_BLOCK_LENGTH)
else:
key_states, value_states = update_past_key_value(
past_key_value, key_states, value_states,
kv_seq_len, use_quantize_kv, device
)
past_key_value = (key_states, value_states) if use_cache else None
if self.training: if self.training:
warnings.warn("xops is not supported on Intel GPU, so just use normal implementation") warnings.warn("xops is not supported on Intel GPU, so just use normal implementation")