From 99cd16ef9f3b443e379cd81eac4b9a56808e32d7 Mon Sep 17 00:00:00 2001 From: Guancheng Fu <110874468+gc-fu@users.noreply.github.com> Date: Wed, 26 Jun 2024 15:33:47 +0800 Subject: [PATCH] Fix error while using pipeline parallism (#11434) --- .../llm/src/ipex_llm/transformers/convert.py | 27 +++++++++---------- 1 file changed, 13 insertions(+), 14 deletions(-) diff --git a/python/llm/src/ipex_llm/transformers/convert.py b/python/llm/src/ipex_llm/transformers/convert.py index f64cc575..dc233942 100644 --- a/python/llm/src/ipex_llm/transformers/convert.py +++ b/python/llm/src/ipex_llm/transformers/convert.py @@ -146,20 +146,6 @@ def is_linear_module(module): global _VLLM_VERSION if _VLLM_VERSION is None: _VLLM_VERSION = get_package_version('vllm') - if 'xpu' in _VLLM_VERSION: - # For vllm xpu - from vllm.model_executor.parallel_utils.parallel_state import ( - get_tensor_model_parallel_group, - get_tensor_model_parallel_world_size - ) - if torch.distributed.is_initialized(): - tp_size = get_tensor_model_parallel_world_size() - else: - tp_size = 1 - else: - # For vllm cpu - tp_size = 1 - from vllm.model_executor.layers.linear import ( ColumnParallelLinear, RowParallelLinear, QKVParallelLinear, MergedColumnParallelLinear ) @@ -168,6 +154,19 @@ def is_linear_module(module): ColumnParallelLinear, RowParallelLinear, QKVParallelLinear, MergedColumnParallelLinear ] if is_module_in_classes(module, VLLM_LINEAR_LIST): + if 'xpu' in _VLLM_VERSION: + # For vllm xpu + from vllm.model_executor.parallel_utils.parallel_state import ( + get_tensor_model_parallel_group, + get_tensor_model_parallel_world_size + ) + if torch.distributed.is_initialized(): + tp_size = get_tensor_model_parallel_world_size() + else: + tp_size = 1 + else: + # For vllm cpu + tp_size = 1 in_features = module.input_size out_features = module.output_size result = True