From 7ca557aada9c22097f38be047ff0028ef1e261b9 Mon Sep 17 00:00:00 2001 From: Xiangyu Tian <109123695+xiangyuT@users.noreply.github.com> Date: Tue, 27 Aug 2024 09:22:19 +0800 Subject: [PATCH] LLM: Fix vLLM CPU convert error (#11926) --- python/llm/src/ipex_llm/vllm/cpu/model_convert.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/python/llm/src/ipex_llm/vllm/cpu/model_convert.py b/python/llm/src/ipex_llm/vllm/cpu/model_convert.py index ff651542..164cd1d4 100644 --- a/python/llm/src/ipex_llm/vllm/cpu/model_convert.py +++ b/python/llm/src/ipex_llm/vllm/cpu/model_convert.py @@ -254,8 +254,8 @@ def get_load_function(low_bit): scheduler_config=self.scheduler_config) return - _model_mlp_convert() - _model_attention_convert() + # _model_mlp_convert() + # _model_attention_convert() self.model = get_model( model_config=self.model_config,