Fix qwen nan value issue on vllm (#12971)
* add to fix qwen nan value issue * update
This commit is contained in:
parent
cd109bb061
commit
c9ecb7a113
1 changed files with 6 additions and 0 deletions
|
|
@ -107,6 +107,12 @@ def get_load_function(low_bit):
|
||||||
modules = ["35.mlp", "36.mlp", "37.mlp", "38.mlp", "39.mlp"]
|
modules = ["35.mlp", "36.mlp", "37.mlp", "38.mlp", "39.mlp"]
|
||||||
else:
|
else:
|
||||||
modules = None
|
modules = None
|
||||||
|
not_convert_o_proj = os.getenv("IPEX_LLM_NOT_CONVERT_O_PROJ", None)
|
||||||
|
if not_convert_o_proj is not None:
|
||||||
|
# only use to avoid nan value in o_proj running DeepSeek-R1-Distill-Qwen-14B
|
||||||
|
modules = ["o_proj"]
|
||||||
|
else:
|
||||||
|
modules = None
|
||||||
if "minicpm" in self.vllm_config.model_config.model.lower():
|
if "minicpm" in self.vllm_config.model_config.model.lower():
|
||||||
modules = ["vpm", "resampler"]
|
modules = ["vpm", "resampler"]
|
||||||
if "internvl2" in self.vllm_config.model_config.model.lower():
|
if "internvl2" in self.vllm_config.model_config.model.lower():
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue