Fix qwen nan value issue on vllm (#12971)
* add to fix qwen nan value issue * update
This commit is contained in:
		
							parent
							
								
									cd109bb061
								
							
						
					
					
						commit
						c9ecb7a113
					
				
					 1 changed files with 6 additions and 0 deletions
				
			
		| 
						 | 
					@ -107,6 +107,12 @@ def get_load_function(low_bit):
 | 
				
			||||||
                    modules = ["35.mlp", "36.mlp", "37.mlp", "38.mlp", "39.mlp"]
 | 
					                    modules = ["35.mlp", "36.mlp", "37.mlp", "38.mlp", "39.mlp"]
 | 
				
			||||||
                else:
 | 
					                else:
 | 
				
			||||||
                    modules = None
 | 
					                    modules = None
 | 
				
			||||||
 | 
					                not_convert_o_proj = os.getenv("IPEX_LLM_NOT_CONVERT_O_PROJ", None)
 | 
				
			||||||
 | 
					                if not_convert_o_proj is not None:
 | 
				
			||||||
 | 
					                    # only use to avoid nan value in o_proj running DeepSeek-R1-Distill-Qwen-14B
 | 
				
			||||||
 | 
					                    modules = ["o_proj"]
 | 
				
			||||||
 | 
					                else:
 | 
				
			||||||
 | 
					                    modules = None
 | 
				
			||||||
                if "minicpm" in self.vllm_config.model_config.model.lower():
 | 
					                if "minicpm" in self.vllm_config.model_config.model.lower():
 | 
				
			||||||
                    modules = ["vpm", "resampler"]
 | 
					                    modules = ["vpm", "resampler"]
 | 
				
			||||||
                if "internvl2" in self.vllm_config.model_config.model.lower():
 | 
					                if "internvl2" in self.vllm_config.model_config.model.lower():
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue