Fix vLLM-v2 install instructions(#10822)
This commit is contained in:
parent
3cd21d5105
commit
61c67af386
1 changed files with 2 additions and 0 deletions
|
|
@ -34,6 +34,8 @@ sycl-ls
|
||||||
To run vLLM continuous batching on Intel GPUs, install the dependencies as follows:
|
To run vLLM continuous batching on Intel GPUs, install the dependencies as follows:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
|
# This directory may change depends on where you install oneAPI-basekit
|
||||||
|
source /opt/intel/oneapi/setvars.sh
|
||||||
# First create an conda environment
|
# First create an conda environment
|
||||||
conda create -n ipex-vllm python=3.11
|
conda create -n ipex-vllm python=3.11
|
||||||
conda activate ipex-vllm
|
conda activate ipex-vllm
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue