parent
8ba57b41cd
commit
d83e5068d2
2 changed files with 3 additions and 1 deletions
|
|
@ -125,7 +125,7 @@ RUN set -eux && \
|
||||||
python3 get-pip.py && rm get-pip.py && \
|
python3 get-pip.py && rm get-pip.py && \
|
||||||
pip install --upgrade requests argparse urllib3 && \
|
pip install --upgrade requests argparse urllib3 && \
|
||||||
pip install --pre --upgrade ipex-llm[xpu_2.6] --extra-index-url https://download.pytorch.org/whl/xpu && \
|
pip install --pre --upgrade ipex-llm[xpu_2.6] --extra-index-url https://download.pytorch.org/whl/xpu && \
|
||||||
pip install transformers_stream_generator einops tiktoken && \
|
pip install transformers_stream_generator einops tiktoken librosa && \
|
||||||
pip install --upgrade colorama && \
|
pip install --upgrade colorama && \
|
||||||
#
|
#
|
||||||
git clone https://github.com/intel/ipex-llm.git && \
|
git clone https://github.com/intel/ipex-llm.git && \
|
||||||
|
|
|
||||||
|
|
@ -124,6 +124,8 @@ def get_load_function(low_bit):
|
||||||
modules = ["vision_model", "mlp1"]
|
modules = ["vision_model", "mlp1"]
|
||||||
if "deepseek-v2" in self.vllm_config.model_config.model.lower():
|
if "deepseek-v2" in self.vllm_config.model_config.model.lower():
|
||||||
modules = ["down_proj"]
|
modules = ["down_proj"]
|
||||||
|
if "whisper" in self.vllm_config.model_config.model.lower():
|
||||||
|
modules = ["proj_out"]
|
||||||
optimize_model(self.model,
|
optimize_model(self.model,
|
||||||
low_bit=low_bit,
|
low_bit=low_bit,
|
||||||
torch_dtype=self.vllm_config.model_config.dtype,
|
torch_dtype=self.vllm_config.model_config.dtype,
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue