From 3b82834aafbd0966fef7dce67ea762bb928ce96d Mon Sep 17 00:00:00 2001 From: Guancheng Fu <110874468+gc-fu@users.noreply.github.com> Date: Mon, 22 Apr 2024 14:18:51 +0800 Subject: [PATCH] Update README.md (#10838) --- docker/llm/README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/docker/llm/README.md b/docker/llm/README.md index 51d9b7f9..47a2fac7 100644 --- a/docker/llm/README.md +++ b/docker/llm/README.md @@ -152,14 +152,14 @@ Additionally, for examples related to Inference with Speculative Decoding, you c ## IPEX-LLM Inference on XPU -### 1. Prepare ipex-llm-cpu Docker Image +### 1. Prepare ipex-llm-xpu Docker Image Run the following command to pull image from dockerhub: ```bash docker pull intelanalytics/ipex-llm-xpu:2.1.0-SNAPSHOT ``` -### 2. Start bigdl-llm-cpu Docker Container +### 2. Start ipex-llm-xpu Docker Container To map the xpu into the container, you need to specify --device=/dev/dri when booting the container. ```bash @@ -650,4 +650,4 @@ Using the [export_merged_model.py](../../python/llm/example/GPU/LLM-Finetuning/Q python ./export_merged_model.py --repo-id-or-model-path REPO_ID_OR_MODEL_PATH --adapter_path ./outputs/checkpoint-200 --output_path ./outputs/checkpoint-200-merged ``` -Then you can use `./outputs/checkpoint-200-merged` as a normal huggingface transformer model to do inference. \ No newline at end of file +Then you can use `./outputs/checkpoint-200-merged` as a normal huggingface transformer model to do inference.