Fix internvl fp16 error (#13205)
This commit is contained in:
parent
45864790f7
commit
5a1c1297e1
1 changed files with 5 additions and 1 deletions
|
|
@ -133,7 +133,11 @@ def get_load_function(low_bit):
|
||||||
modules = ["vision_encoder", "embed_tokens_extend"]
|
modules = ["vision_encoder", "embed_tokens_extend"]
|
||||||
if low_bit == "fp16":
|
if low_bit == "fp16":
|
||||||
# to fix qwen2.5-vl and glm-4v
|
# to fix qwen2.5-vl and glm-4v
|
||||||
modules = ["vision", "visual"]
|
if modules is None:
|
||||||
|
modules = ["vision", "visual"]
|
||||||
|
else:
|
||||||
|
modules.append("vision")
|
||||||
|
modules.append("visual")
|
||||||
optimize_model(self.model,
|
optimize_model(self.model,
|
||||||
low_bit=low_bit,
|
low_bit=low_bit,
|
||||||
torch_dtype=self.vllm_config.model_config.dtype,
|
torch_dtype=self.vllm_config.model_config.dtype,
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue