Enable internvl2-8b on vllm(#12645)
This commit is contained in:
		
							parent
							
								
									8fd2dcba86
								
							
						
					
					
						commit
						6711a48a36
					
				
					 1 changed files with 2 additions and 0 deletions
				
			
		| 
						 | 
					@ -105,6 +105,8 @@ def get_load_function(low_bit):
 | 
				
			||||||
            if "minicpm-v" in self.model_config.model.lower():
 | 
					            if "minicpm-v" in self.model_config.model.lower():
 | 
				
			||||||
                from ipex_llm.transformers.models.minicpmv import merge_qkv
 | 
					                from ipex_llm.transformers.models.minicpmv import merge_qkv
 | 
				
			||||||
                self.model.vpm.apply(merge_qkv)
 | 
					                self.model.vpm.apply(merge_qkv)
 | 
				
			||||||
 | 
					            if "internvl2" in self.model_config.model.lower():
 | 
				
			||||||
 | 
					                modules = ["vision_model", "mlp1"]
 | 
				
			||||||
            optimize_model(self.model, low_bit=low_bit, torch_dtype=self.model_config.dtype,
 | 
					            optimize_model(self.model, low_bit=low_bit, torch_dtype=self.model_config.dtype,
 | 
				
			||||||
                           modules_to_not_convert=modules)
 | 
					                           modules_to_not_convert=modules)
 | 
				
			||||||
            self.model = self.model.to(device=self.device_config.device,
 | 
					            self.model = self.model.to(device=self.device_config.device,
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue