Enable internvl2-8b on vllm(#12645)
This commit is contained in:
parent
8fd2dcba86
commit
6711a48a36
1 changed files with 2 additions and 0 deletions
|
|
@ -105,6 +105,8 @@ def get_load_function(low_bit):
|
||||||
if "minicpm-v" in self.model_config.model.lower():
|
if "minicpm-v" in self.model_config.model.lower():
|
||||||
from ipex_llm.transformers.models.minicpmv import merge_qkv
|
from ipex_llm.transformers.models.minicpmv import merge_qkv
|
||||||
self.model.vpm.apply(merge_qkv)
|
self.model.vpm.apply(merge_qkv)
|
||||||
|
if "internvl2" in self.model_config.model.lower():
|
||||||
|
modules = ["vision_model", "mlp1"]
|
||||||
optimize_model(self.model, low_bit=low_bit, torch_dtype=self.model_config.dtype,
|
optimize_model(self.model, low_bit=low_bit, torch_dtype=self.model_config.dtype,
|
||||||
modules_to_not_convert=modules)
|
modules_to_not_convert=modules)
|
||||||
self.model = self.model.to(device=self.device_config.device,
|
self.model = self.model.to(device=self.device_config.device,
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue