LLM: fix gptneox kv cache (#9044)
This commit is contained in:
parent
4b843d1dbf
commit
975da86e00
1 changed files with 1 additions and 1 deletions
|
|
@ -34,7 +34,7 @@
|
||||||
import torch
|
import torch
|
||||||
from typing import Optional, Tuple
|
from typing import Optional, Tuple
|
||||||
from bigdl.llm.transformers.models.utils import apply_rotary_pos_emb
|
from bigdl.llm.transformers.models.utils import apply_rotary_pos_emb
|
||||||
from bigdl.llm.transformers.models.utils import init_kv_cache, extend_kv_cache
|
from bigdl.llm.transformers.models.utils import init_kv_cache, extend_kv_cache, append_kv_cache
|
||||||
|
|
||||||
|
|
||||||
KV_CACHE_ALLOC_BLOCK_LENGTH = 256
|
KV_CACHE_ALLOC_BLOCK_LENGTH = 256
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue