diff --git a/docker/llm/inference/cpu/docker/Dockerfile b/docker/llm/inference/cpu/docker/Dockerfile index 41f6e102..2234b481 100644 --- a/docker/llm/inference/cpu/docker/Dockerfile +++ b/docker/llm/inference/cpu/docker/Dockerfile @@ -43,6 +43,8 @@ RUN env DEBIAN_FRONTEND=noninteractive apt-get update && \ # Install vllm dependencies pip install --upgrade fastapi && \ pip install --upgrade "uvicorn[standard]" && \ +# Add Qwen support + pip install --upgrade transformers_stream_generator einops && \ # Copy vLLM-Serving cp -r ./BigDL/python/llm/example/CPU/vLLM-Serving/ ./vLLM-Serving && \ rm -rf ./BigDL diff --git a/python/llm/example/CPU/Deepspeed-AutoTP/install.sh b/python/llm/example/CPU/Deepspeed-AutoTP/install.sh index 38e35ec7..fa2b0910 100644 --- a/python/llm/example/CPU/Deepspeed-AutoTP/install.sh +++ b/python/llm/example/CPU/Deepspeed-AutoTP/install.sh @@ -9,7 +9,7 @@ cd build cmake .. make -j install mkdir -p /opt/intel/oneccl -mv ./build/_install/env /opt/intel/oneccl +mv ./_install/env /opt/intel/oneccl # 2. install torch and ipex pip install torch==2.1.0 pip install intel_extension_for_pytorch==2.1.0 -f https://developer.intel.com/ipex-whl-stable-cpu