Update fastchat demo script (#12367)
* Update README.md * Update vllm_docker_quickstart.md
This commit is contained in:
parent
51f7f87768
commit
fad15c8ca0
2 changed files with 6 additions and 2 deletions
|
|
@ -103,6 +103,8 @@ To set up model serving using `IPEX-LLM` as backend using FastChat, you can refe
|
|||
# start controller
|
||||
python -m fastchat.serve.controller &
|
||||
|
||||
export USE_XETLA=OFF
|
||||
export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=2
|
||||
|
||||
export TORCH_LLM_ALLREDUCE=0
|
||||
export CCL_DG2_ALLREDUCE=1
|
||||
|
|
|
|||
|
|
@ -832,6 +832,8 @@ We can set up model serving using `IPEX-LLM` as backend using FastChat, the foll
|
|||
# start controller
|
||||
python -m fastchat.serve.controller &
|
||||
|
||||
export USE_XETLA=OFF
|
||||
export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=2
|
||||
|
||||
export TORCH_LLM_ALLREDUCE=0
|
||||
export CCL_DG2_ALLREDUCE=1
|
||||
|
|
|
|||
Loading…
Reference in a new issue