LLM: fix gptneox kv cache (#9044)

This commit is contained in:
Ruonan Wang 2023-09-25 13:03:57 +08:00 committed by GitHub
parent 4b843d1dbf
commit 975da86e00

View file

@ -34,7 +34,7 @@
import torch
from typing import Optional, Tuple
from bigdl.llm.transformers.models.utils import apply_rotary_pos_emb
from bigdl.llm.transformers.models.utils import init_kv_cache, extend_kv_cache
from bigdl.llm.transformers.models.utils import init_kv_cache, extend_kv_cache, append_kv_cache
KV_CACHE_ALLOC_BLOCK_LENGTH = 256