From a465111cf4edf7967e446af2690ebea92067bc25 Mon Sep 17 00:00:00 2001 From: Guancheng Fu <110874468+gc-fu@users.noreply.github.com> Date: Mon, 13 May 2024 16:44:48 +0800 Subject: [PATCH] Update README.md (#11003) --- python/llm/src/ipex_llm/serving/fastchat/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/python/llm/src/ipex_llm/serving/fastchat/README.md b/python/llm/src/ipex_llm/serving/fastchat/README.md index 1408e050..f1f469ae 100644 --- a/python/llm/src/ipex_llm/serving/fastchat/README.md +++ b/python/llm/src/ipex_llm/serving/fastchat/README.md @@ -78,7 +78,7 @@ We have also provided an option `--load-low-bit-model` to load models that have Check the following examples: ```bash # Or --device "cpu" -python -m ipex_llm.serving.fastchat.ipex_llm_worker --model-path /Low/Bit/Model/Path --trust-remote-code --device "xpu" +python -m ipex_llm.serving.fastchat.ipex_llm_worker --model-path /Low/Bit/Model/Path --trust-remote-code --device "xpu" --load-low-bit-model ``` #### For self-speculative decoding example: