LLM: Support Baichuan2-13b in BigDL-vLLM (#10398)
Support Baichuan2-13b in BigDL-vLLM.
This commit is contained in:
parent
e10de2c42d
commit
60043a3ae8
3 changed files with 274 additions and 2 deletions
|
|
@ -485,7 +485,7 @@ def baichuan_attention_forward_13b_origin(
|
|||
attn_output = attn_output.transpose(1, 2)
|
||||
else:
|
||||
attn_weights = torch.matmul(
|
||||
query_states, key_states.transpose(2, 3)
|
||||
query_states.to(dtype=key_states.dtype), key_states.transpose(2, 3)
|
||||
) / math.sqrt(self.head_dim)
|
||||
|
||||
if attention_mask is not None:
|
||||
|
|
|
|||
|
|
@ -43,6 +43,7 @@ from bigdl.llm.vllm.model_executor.models.bigdl_llama import BigDLLlamaForCausal
|
|||
from bigdl.llm.vllm.model_executor.models.bigdl_mixtral import BigDLMixtralForCausalLM
|
||||
from bigdl.llm.vllm.model_executor.models.bigdl_mistral import BigDLMistralForCausalLM
|
||||
from bigdl.llm.vllm.model_executor.models.bigdl_chatglm import BigDLChatGLMForCausalLM
|
||||
from bigdl.llm.vllm.model_executor.models.bigdl_baichuan import BigDLBaichuanForCausalLM
|
||||
|
||||
from bigdl.llm.utils.common import invalidInputError
|
||||
|
||||
|
|
@ -54,7 +55,7 @@ from bigdl.llm.utils.common import invalidInputError
|
|||
_MODEL_REGISTRY = {
|
||||
# "AquilaModel": AquilaForCausalLM,
|
||||
# "BaiChuanForCausalLM": BaiChuanForCausalLM, # baichuan-7b
|
||||
# "BaichuanForCausalLM": BaichuanForCausalLM, # baichuan-13b
|
||||
"BaichuanForCausalLM": BigDLBaichuanForCausalLM, # baichuan2-13b
|
||||
# "BloomForCausalLM": BloomForCausalLM,
|
||||
# "FalconForCausalLM": FalconForCausalLM,
|
||||
# "GPT2LMHeadModel": GPT2LMHeadModel,
|
||||
|
|
|
|||
|
|
@ -0,0 +1,271 @@
|
|||
#
|
||||
# Copyright 2016 The BigDL Authors.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
#
|
||||
|
||||
import torch
|
||||
from torch import nn
|
||||
|
||||
from transformers import AutoTokenizer, PreTrainedTokenizerBase, LlamaConfig
|
||||
from typing import Optional, Tuple, List, Type, Dict
|
||||
|
||||
from bigdl.llm.vllm.sequence import SequenceOutputs, SequenceGroupMetadata
|
||||
from bigdl.llm.vllm.model_executor.layers.bigdl_sampler import BigDLSampler
|
||||
from bigdl.llm.vllm.model_executor.models.bigdl_model import BigDLModelForCausalLM
|
||||
from bigdl.llm.vllm.logger import init_logger
|
||||
import math
|
||||
import time
|
||||
from bigdl.llm.vllm.model_executor.input_metadata import InputMetadata
|
||||
import os
|
||||
from transformers.generation.logits_process import (
|
||||
LogitsProcessorList,
|
||||
RepetitionPenaltyLogitsProcessor,
|
||||
TemperatureLogitsWarper,
|
||||
TopKLogitsWarper,
|
||||
TopPLogitsWarper,
|
||||
)
|
||||
|
||||
logger = init_logger(__name__)
|
||||
|
||||
|
||||
def _pad_to_max(x: List[int], max_len: int, padding_id: int = 0) -> List[int]:
|
||||
return [padding_id] * (max_len - len(x)) + x
|
||||
|
||||
|
||||
def _get_attention_mask_for_prompts(
|
||||
input_ids: List[List[int]], max_prompt_len: int
|
||||
) -> List[List[int]]:
|
||||
attention_mask = [
|
||||
[0] * (max_prompt_len - len(prompt)) + [1] * len(prompt) for prompt in input_ids
|
||||
]
|
||||
return attention_mask
|
||||
|
||||
vllm_selective_batching = os.getenv("VLLM_ENABLE_SELECTIVE_BATCHING")
|
||||
enable_vllm_se_batching = vllm_selective_batching is not None
|
||||
enable_vllm_se_batching = enable_vllm_se_batching and vllm_selective_batching.lower() == "true"
|
||||
|
||||
|
||||
class BigDLBaichuanForCausalLM(BigDLModelForCausalLM):
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
config: LlamaConfig,
|
||||
device: Optional[str] = None,
|
||||
max_model_len: Optional[int] = None,
|
||||
load_in_low_bit: str = 'sym_int4'
|
||||
):
|
||||
super().__init__(config, device, max_model_len)
|
||||
self.config = config
|
||||
# Always enable bigdl-llm model
|
||||
from bigdl.llm.transformers import AutoModelForCausalLM
|
||||
# TODO: we will need to pass the argument through command line argument
|
||||
# from bigdl.llm import optimize_model
|
||||
torch_dtype = 'auto'
|
||||
|
||||
if load_in_low_bit == 'bf16':
|
||||
torch_dtype = torch.bfloat16
|
||||
elif load_in_low_bit == 'fp16':
|
||||
torch_dtype = torch.float16
|
||||
# bf16 will require to set torch_dtype to bf16
|
||||
if device == 'cpu':
|
||||
self.model = AutoModelForCausalLM.from_pretrained(
|
||||
config._name_or_path,
|
||||
load_in_low_bit=load_in_low_bit,
|
||||
torch_dtype=torch_dtype,
|
||||
low_cpu_mem_usage=True,
|
||||
trust_remote_code=True,
|
||||
use_cache=True,
|
||||
)
|
||||
# self.model = optimize_model(model)
|
||||
self.sampler = BigDLSampler(config.vocab_size, device)
|
||||
elif device == 'xpu':
|
||||
try:
|
||||
import intel_extension_for_pytorch as ipex
|
||||
except ImportError:
|
||||
print("Intel Extension for PyTorch is not installed, \
|
||||
but is required for xpu inference.")
|
||||
|
||||
model = AutoModelForCausalLM.from_pretrained(
|
||||
config._name_or_path,
|
||||
load_in_low_bit=load_in_low_bit,
|
||||
torch_dtype=torch_dtype,
|
||||
trust_remote_code=True,
|
||||
use_cache=True,
|
||||
)
|
||||
self.model = model.to('xpu')
|
||||
self.sampler = BigDLSampler(config.vocab_size, device).to('xpu')
|
||||
|
||||
self.device = torch.device(device)
|
||||
self.dtype = self.model.dtype
|
||||
self.last_seq_ids = []
|
||||
self.last_kv_cache = None
|
||||
self.pad_token_id = config.pad_token_id
|
||||
self.max_seq_limit = max_model_len
|
||||
|
||||
# GC: Note for selective batching
|
||||
# KV_CACHE in the format of num_layers x 2 x (seq_id -> torch.Tensor)
|
||||
# past_key_values in the format of num_layers x len(seq_id) x (2 x torch.Tensor)
|
||||
# If we set num_layers to 9, have 10 sequences in total.
|
||||
# then, for the kv_cache, we get 9 x 2 x 10 = 180 tensors
|
||||
# for past_key_values, we get 9 x 10 x 2 = 180 tensors
|
||||
def forward(
|
||||
self,
|
||||
seq_group_meta_data_lists: List[SequenceGroupMetadata],
|
||||
# kv_cache in the format [[dict() for _ in range(2)] for _ in range(32)]
|
||||
kv_cache: Optional[List[List[Dict]]] = None,
|
||||
input_metadata: Optional[InputMetadata] = None,
|
||||
) -> Tuple[torch.Tensor, List[Tuple[torch.Tensor, torch.Tensor]]]:
|
||||
num_layers = self.model.config.num_hidden_layers
|
||||
# One for key, one for value
|
||||
decoder_kv_size = 2
|
||||
|
||||
bigdl_input_ids = []
|
||||
# bigdl_position_ids = []
|
||||
bigdl_attention_mask = []
|
||||
|
||||
cur_seq_ids = []
|
||||
max_prompt_len = 0
|
||||
|
||||
# 0. Verify is_prompt or is_decoding
|
||||
is_decoding_stage = not seq_group_meta_data_lists[0].is_prompt
|
||||
|
||||
# 1. Assemble bigdl_input_ids
|
||||
for seq_group_meta_data in seq_group_meta_data_lists:
|
||||
# req_id = seq_group_meta_data.request_id
|
||||
# is_decoding_stage = is_decoding_stage and (not seq_group_meta_data.is_prompt)
|
||||
seq_ids = list(seq_group_meta_data.seq_data.keys())
|
||||
seq_id = seq_ids[0]
|
||||
cur_seq_ids.append(seq_id)
|
||||
seq_data = seq_group_meta_data.seq_data[seq_id]
|
||||
|
||||
cur_seq_input_ids = seq_data.get_token_ids()
|
||||
# context_len = seq_data.get_len()
|
||||
if seq_group_meta_data.is_prompt:
|
||||
bigdl_input_ids.append(cur_seq_input_ids)
|
||||
max_prompt_len = max(max_prompt_len, seq_data.get_len())
|
||||
else:
|
||||
bigdl_input_ids.append([cur_seq_input_ids[-1]])
|
||||
# 1. Assemble bigdl_input_ids end
|
||||
|
||||
if is_decoding_stage:
|
||||
construct_kv_cache_func = self.get_construct_kv_cache_func(enable_vllm_se_batching)
|
||||
bigdl_kv_cache = construct_kv_cache_func(cur_seq_ids,
|
||||
seq_group_meta_data_lists,
|
||||
kv_cache,
|
||||
num_layers,
|
||||
2)
|
||||
else:
|
||||
bigdl_attention_mask = _get_attention_mask_for_prompts(bigdl_input_ids, max_prompt_len)
|
||||
bigdl_input_ids = [
|
||||
_pad_to_max(input_ids, max_prompt_len, self.pad_token_id)
|
||||
for input_ids in bigdl_input_ids
|
||||
]
|
||||
|
||||
decoding_attention_mask_list = []
|
||||
decoding_position_ids = []
|
||||
# num_layers x len(seq_id) x (2 x torch.Tensor)
|
||||
if is_decoding_stage:
|
||||
if enable_vllm_se_batching:
|
||||
batch = 0
|
||||
for seq_group_meta_data in seq_group_meta_data_lists:
|
||||
# Get current seq_len in kv_cache
|
||||
current_seq_len = bigdl_kv_cache[0][batch][0].size(2)
|
||||
batch += 1
|
||||
seq_ids = list(seq_group_meta_data.seq_data.keys())
|
||||
seq_data = seq_group_meta_data.seq_data[seq_ids[0]]
|
||||
cur_pos = seq_data.get_len()
|
||||
decoding_position_ids.append(cur_pos - 1)
|
||||
# Total length: current_seq_len + 1
|
||||
cur_attention_mask = [0] * (current_seq_len - cur_pos + 1) + [1] * (cur_pos)
|
||||
decoding_attention_mask_list.append(cur_attention_mask)
|
||||
else:
|
||||
cur_seq_len = bigdl_kv_cache[0][0].size(2)
|
||||
for seq_group_meta_data in seq_group_meta_data_lists:
|
||||
seq_ids = list(seq_group_meta_data.seq_data.keys())
|
||||
seq_id = seq_ids[0]
|
||||
seq_data = seq_group_meta_data.seq_data[seq_id]
|
||||
cur_pos = seq_data.get_len()
|
||||
# bigdl_position_ids.append([cur_pos - 1])
|
||||
# decoding_position_ids.append(cur_pos - 1)
|
||||
cur_attention_mask = [0] * (cur_seq_len - cur_pos + 1) + [1] * (cur_pos)
|
||||
decoding_attention_mask_list.append(cur_attention_mask)
|
||||
|
||||
bigdl_input_ids = torch.tensor(bigdl_input_ids, device=self.device)
|
||||
|
||||
if is_decoding_stage:
|
||||
if enable_vllm_se_batching:
|
||||
attention_mask = [torch.tensor(x, device=self.device).unsqueeze(0)
|
||||
for x in decoding_attention_mask_list]
|
||||
position_ids = torch.tensor(decoding_position_ids, device=self.device).long()
|
||||
position_ids = position_ids.unsqueeze(-1)
|
||||
else:
|
||||
attention_mask = torch.tensor(decoding_attention_mask_list, device=self.device)
|
||||
position_ids = None
|
||||
kwargs = {
|
||||
"input_ids": bigdl_input_ids,
|
||||
"position_ids": position_ids,
|
||||
"attention_mask": attention_mask,
|
||||
"past_key_values": bigdl_kv_cache,
|
||||
"use_cache": True,
|
||||
# "return_dict": True,
|
||||
}
|
||||
else:
|
||||
# Prefill stage
|
||||
attention_mask = torch.tensor(bigdl_attention_mask, device=self.device)
|
||||
if enable_vllm_se_batching:
|
||||
position_ids = attention_mask.long().cumsum(-1) - 1
|
||||
position_ids.masked_fill_(attention_mask == 0, 1)
|
||||
position_ids.to(self.device)
|
||||
else:
|
||||
position_ids = None
|
||||
kwargs = {
|
||||
"input_ids": bigdl_input_ids,
|
||||
"attention_mask": attention_mask,
|
||||
"position_ids": position_ids,
|
||||
"past_key_values": None,
|
||||
"use_cache": True,
|
||||
# "return_dict": True,
|
||||
}
|
||||
# Prefill may need additional space, which forces us to delete the last_kv_cache
|
||||
if self.last_kv_cache:
|
||||
self.last_kv_cache = None
|
||||
# pdb.set_trace()
|
||||
|
||||
if self.device.type == 'xpu':
|
||||
torch.xpu.empty_cache()
|
||||
st_timestamp = time.perf_counter()
|
||||
outputs = self.model.forward(**kwargs)
|
||||
# tmp = torch.xpu.memory_stats()
|
||||
# logger.info(f"0: {tmp['allocated_bytes.all.current']}")
|
||||
# self.last_seq_ids = cur_seq_ids[:]
|
||||
# self.last_kv_cache = outputs.past_key_values
|
||||
self._set_last_seq_ids(cur_seq_ids[:])
|
||||
self._set_last_kv_cache(outputs.past_key_values)
|
||||
# pdb.set_trace()
|
||||
|
||||
logits = outputs.logits[:, -1, :]
|
||||
bigdl_output = self.sampler(logits, input_metadata, st_timestamp)
|
||||
# tmp = torch.xpu.memory_stats()
|
||||
# logger.info(f"before: {tmp['allocated_bytes.all.current']}")
|
||||
|
||||
if enable_vllm_se_batching:
|
||||
self.update_kv_cache_selective_batching(
|
||||
cur_seq_ids, kv_cache, num_layers, decoder_kv_size)
|
||||
self.last_kv_cache = None
|
||||
else:
|
||||
self.update_kv_cache(cur_seq_ids, kv_cache, num_layers, decoder_kv_size)
|
||||
|
||||
# tmp = torch.xpu.memory_stats()
|
||||
# logger.info(f"after: {tmp['allocated_bytes.all.current']}")
|
||||
return bigdl_output
|
||||
Loading…
Reference in a new issue