From 975da86e00768a7141375ea43c9353b262a72635 Mon Sep 17 00:00:00 2001 From: Ruonan Wang <105281011+rnwang04@users.noreply.github.com> Date: Mon, 25 Sep 2023 13:03:57 +0800 Subject: [PATCH] LLM: fix gptneox kv cache (#9044) --- python/llm/src/bigdl/llm/transformers/models/gptneox.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/python/llm/src/bigdl/llm/transformers/models/gptneox.py b/python/llm/src/bigdl/llm/transformers/models/gptneox.py index 8e31a14a..1f70491f 100644 --- a/python/llm/src/bigdl/llm/transformers/models/gptneox.py +++ b/python/llm/src/bigdl/llm/transformers/models/gptneox.py @@ -34,7 +34,7 @@ import torch from typing import Optional, Tuple from bigdl.llm.transformers.models.utils import apply_rotary_pos_emb -from bigdl.llm.transformers.models.utils import init_kv_cache, extend_kv_cache +from bigdl.llm.transformers.models.utils import init_kv_cache, extend_kv_cache, append_kv_cache KV_CACHE_ALLOC_BLOCK_LENGTH = 256