Fix internvl fp16 error (#13205)
This commit is contained in:
		
							parent
							
								
									45864790f7
								
							
						
					
					
						commit
						5a1c1297e1
					
				
					 1 changed files with 5 additions and 1 deletions
				
			
		| 
						 | 
				
			
			@ -133,7 +133,11 @@ def get_load_function(low_bit):
 | 
			
		|||
                    modules = ["vision_encoder", "embed_tokens_extend"]
 | 
			
		||||
                if low_bit == "fp16":
 | 
			
		||||
                    # to fix qwen2.5-vl and glm-4v
 | 
			
		||||
                    if modules is None:
 | 
			
		||||
                        modules = ["vision", "visual"]
 | 
			
		||||
                    else:
 | 
			
		||||
                        modules.append("vision")
 | 
			
		||||
                        modules.append("visual")
 | 
			
		||||
                optimize_model(self.model,
 | 
			
		||||
                               low_bit=low_bit,
 | 
			
		||||
                               torch_dtype=self.vllm_config.model_config.dtype,
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue