Update README.md (#10838)
This commit is contained in:
parent
3daad242b8
commit
3b82834aaf
1 changed files with 3 additions and 3 deletions
|
|
@ -152,14 +152,14 @@ Additionally, for examples related to Inference with Speculative Decoding, you c
|
||||||
|
|
||||||
## IPEX-LLM Inference on XPU
|
## IPEX-LLM Inference on XPU
|
||||||
|
|
||||||
### 1. Prepare ipex-llm-cpu Docker Image
|
### 1. Prepare ipex-llm-xpu Docker Image
|
||||||
|
|
||||||
Run the following command to pull image from dockerhub:
|
Run the following command to pull image from dockerhub:
|
||||||
```bash
|
```bash
|
||||||
docker pull intelanalytics/ipex-llm-xpu:2.1.0-SNAPSHOT
|
docker pull intelanalytics/ipex-llm-xpu:2.1.0-SNAPSHOT
|
||||||
```
|
```
|
||||||
|
|
||||||
### 2. Start bigdl-llm-cpu Docker Container
|
### 2. Start ipex-llm-xpu Docker Container
|
||||||
|
|
||||||
To map the xpu into the container, you need to specify --device=/dev/dri when booting the container.
|
To map the xpu into the container, you need to specify --device=/dev/dri when booting the container.
|
||||||
```bash
|
```bash
|
||||||
|
|
@ -650,4 +650,4 @@ Using the [export_merged_model.py](../../python/llm/example/GPU/LLM-Finetuning/Q
|
||||||
python ./export_merged_model.py --repo-id-or-model-path REPO_ID_OR_MODEL_PATH --adapter_path ./outputs/checkpoint-200 --output_path ./outputs/checkpoint-200-merged
|
python ./export_merged_model.py --repo-id-or-model-path REPO_ID_OR_MODEL_PATH --adapter_path ./outputs/checkpoint-200 --output_path ./outputs/checkpoint-200-merged
|
||||||
```
|
```
|
||||||
|
|
||||||
Then you can use `./outputs/checkpoint-200-merged` as a normal huggingface transformer model to do inference.
|
Then you can use `./outputs/checkpoint-200-merged` as a normal huggingface transformer model to do inference.
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue