Enable internvl2-8b on vllm(#12645)
This commit is contained in:
parent
8fd2dcba86
commit
6711a48a36
1 changed files with 2 additions and 0 deletions
|
|
@ -105,6 +105,8 @@ def get_load_function(low_bit):
|
|||
if "minicpm-v" in self.model_config.model.lower():
|
||||
from ipex_llm.transformers.models.minicpmv import merge_qkv
|
||||
self.model.vpm.apply(merge_qkv)
|
||||
if "internvl2" in self.model_config.model.lower():
|
||||
modules = ["vision_model", "mlp1"]
|
||||
optimize_model(self.model, low_bit=low_bit, torch_dtype=self.model_config.dtype,
|
||||
modules_to_not_convert=modules)
|
||||
self.model = self.model.to(device=self.device_config.device,
|
||||
|
|
|
|||
Loading…
Reference in a new issue