Fix vLLM verion in CPU/vLLM-Serving example README (#11201)
This commit is contained in:
parent
6454655dcc
commit
f02f097002
1 changed files with 1 additions and 0 deletions
|
|
@ -28,6 +28,7 @@ pip3 install "pydantic<2" # Required for OpenAI server.
|
||||||
# Install vllm
|
# Install vllm
|
||||||
git clone https://github.com/vllm-project/vllm.git && \
|
git clone https://github.com/vllm-project/vllm.git && \
|
||||||
cd ./vllm && \
|
cd ./vllm && \
|
||||||
|
git checkout v0.4.2 && \
|
||||||
pip install wheel packaging ninja setuptools>=49.4.0 numpy && \
|
pip install wheel packaging ninja setuptools>=49.4.0 numpy && \
|
||||||
pip install -v -r requirements-cpu.txt --extra-index-url https://download.pytorch.org/whl/cpu && \
|
pip install -v -r requirements-cpu.txt --extra-index-url https://download.pytorch.org/whl/cpu && \
|
||||||
VLLM_TARGET_DEVICE=cpu python3 setup.py install
|
VLLM_TARGET_DEVICE=cpu python3 setup.py install
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue