ipex-llm/docker/llm
Wang, Jian4 b119825152
Remove tgi parameter validation (#11688)
* remove validation

* add min warm up

* remove no need source
2024-07-30 16:37:44 +08:00
..
finetune Enhanced XPU Dockerfiles: Optimized Environment Variables and Documentation (#11506) 2024-07-04 20:18:38 +08:00
inference Add lightweight serving and support tgi parameter (#11600) 2024-07-19 13:15:56 +08:00
inference-cpp Enhanced XPU Dockerfiles: Optimized Environment Variables and Documentation (#11506) 2024-07-04 20:18:38 +08:00
serving Remove tgi parameter validation (#11688) 2024-07-30 16:37:44 +08:00
sources Fix docker images (#11362) 2024-06-20 15:44:55 +08:00
README.md Enhanced XPU Dockerfiles: Optimized Environment Variables and Documentation (#11506) 2024-07-04 20:18:38 +08:00
README_backup.md Re-org XPU finetune images (#10971) 2024-05-15 09:42:43 +08:00

IPEX-LLM Docker Containers

You can run IPEX-LLM containers (via docker or k8s) for inference, serving and fine-tuning on Intel CPU and GPU. Details on how to use these containers are available at IPEX-LLM Docker Container Guides.

Prerequisites

  • Docker on Windows or Linux
  • Windows Subsystem for Linux (WSL) is required if using Windows.

Quick Start

Pull a IPEX-LLM Docker Image

To pull IPEX-LLM Docker images from Docker Hub, use the docker pull command. For instance, to pull the CPU inference image:

docker pull intelanalytics/ipex-llm-cpu:2.1.0-SNAPSHOT

Available images in hub are:

Image Name Description
intelanalytics/ipex-llm-cpu:2.1.0-SNAPSHOT CPU Inference
intelanalytics/ipex-llm-xpu:2.1.0-SNAPSHOT GPU Inference
intelanalytics/ipex-llm-serving-cpu:2.1.0-SNAPSHOT CPU Serving
intelanalytics/ipex-llm-serving-xpu:2.1.0-SNAPSHOT GPU Serving
intelanalytics/ipex-llm-finetune-qlora-cpu-standalone:2.1.0-SNAPSHOT CPU Finetuning via Docker
intelanalytics/ipex-llm-finetune-qlora-cpu-k8s:2.1.0-SNAPSHOT CPU Finetuning via Kubernetes
intelanalytics/ipex-llm-finetune-qlora-xpu:2.1.0-SNAPSHOT GPU Finetuning

Run a Container

Use docker run command to run an IPEX-LLM docker container. For detailed instructions, refer to the IPEX-LLM Docker Container Guides.

Build Docker Image

To build a Docker image from source, first clone the IPEX-LLM repository and navigate to the Dockerfile directory. For example, to build the CPU inference image, navigate to docker/llm/inference/cpu/docker.

Then, use the following command to build the image (replace your_image_name with your desired image name):

docker build \
  --build-arg no_proxy=localhost,127.0.0.1 \
  --rm --no-cache -t your_image_name .

Note: If you're working behind a proxy, also add args --build-arg http_proxy=http://your_proxy_uri:port and --build-arg https_proxy=https://your_proxy_url:port