Update README.md (#10838)
This commit is contained in:
parent
3daad242b8
commit
3b82834aaf
1 changed files with 3 additions and 3 deletions
|
|
@ -152,14 +152,14 @@ Additionally, for examples related to Inference with Speculative Decoding, you c
|
|||
|
||||
## IPEX-LLM Inference on XPU
|
||||
|
||||
### 1. Prepare ipex-llm-cpu Docker Image
|
||||
### 1. Prepare ipex-llm-xpu Docker Image
|
||||
|
||||
Run the following command to pull image from dockerhub:
|
||||
```bash
|
||||
docker pull intelanalytics/ipex-llm-xpu:2.1.0-SNAPSHOT
|
||||
```
|
||||
|
||||
### 2. Start bigdl-llm-cpu Docker Container
|
||||
### 2. Start ipex-llm-xpu Docker Container
|
||||
|
||||
To map the xpu into the container, you need to specify --device=/dev/dri when booting the container.
|
||||
```bash
|
||||
|
|
@ -650,4 +650,4 @@ Using the [export_merged_model.py](../../python/llm/example/GPU/LLM-Finetuning/Q
|
|||
python ./export_merged_model.py --repo-id-or-model-path REPO_ID_OR_MODEL_PATH --adapter_path ./outputs/checkpoint-200 --output_path ./outputs/checkpoint-200-merged
|
||||
```
|
||||
|
||||
Then you can use `./outputs/checkpoint-200-merged` as a normal huggingface transformer model to do inference.
|
||||
Then you can use `./outputs/checkpoint-200-merged` as a normal huggingface transformer model to do inference.
|
||||
|
|
|
|||
Loading…
Reference in a new issue