add --entrypoint /bin/bash (#12957)
Co-authored-by: gc-fu <guancheng.fu@intel.com>
This commit is contained in:
parent
2a8f624f4b
commit
6a2d87e40f
6 changed files with 12 additions and 3 deletions
|
|
@ -26,7 +26,7 @@ To map the `XPU` into the container, you need to specify `--device=/dev/dri` whe
|
|||
|
||||
```bash
|
||||
#/bin/bash
|
||||
export DOCKER_IMAGE=intelanalytics/ipex-llm-xpu:2.2.0-SNAPSHOT
|
||||
export DOCKER_IMAGE=intelanalytics/ipex-llm-serving-xpu:2.2.0-SNAPSHOT
|
||||
|
||||
sudo docker run -itd \
|
||||
--net=host \
|
||||
|
|
@ -34,6 +34,7 @@ sudo docker run -itd \
|
|||
--memory="32G" \
|
||||
--name=CONTAINER_NAME \
|
||||
--shm-size="16g" \
|
||||
--entrypoint /bin/bash \
|
||||
$DOCKER_IMAGE
|
||||
```
|
||||
|
||||
|
|
@ -71,7 +72,7 @@ By default, the container is configured to automatically start the service when
|
|||
|
||||
```bash
|
||||
#/bin/bash
|
||||
export DOCKER_IMAGE=intelanalytics/ipex-llm-xpu:2.2.0-SNAPSHOT
|
||||
export DOCKER_IMAGE=intelanalytics/ipex-llm-serving-xpu:2.2.0-SNAPSHOT
|
||||
|
||||
sudo docker run -itd \
|
||||
--net=host \
|
||||
|
|
@ -110,7 +111,7 @@ If you prefer to manually start the service or need to troubleshoot, you can ove
|
|||
|
||||
```bash
|
||||
#/bin/bash
|
||||
export DOCKER_IMAGE=intelanalytics/ipex-llm-xpu:2.2.0-SNAPSHOT
|
||||
export DOCKER_IMAGE=intelanalytics/ipex-llm-serving-xpu:2.2.0-SNAPSHOT
|
||||
|
||||
sudo docker run -itd \
|
||||
--net=host \
|
||||
|
|
|
|||
|
|
@ -32,6 +32,7 @@ Start ipex-llm-xpu Docker Container. Choose one of the following commands to sta
|
|||
--name=$CONTAINER_NAME \
|
||||
--shm-size="16g" \
|
||||
-v $MODEL_PATH:/llm/models \
|
||||
--entrypoint /bin/bash \
|
||||
$DOCKER_IMAGE
|
||||
```
|
||||
|
||||
|
|
@ -52,6 +53,7 @@ Start ipex-llm-xpu Docker Container. Choose one of the following commands to sta
|
|||
--shm-size="16g" \
|
||||
-v $MODEL_PATH:/llm/llm-models \
|
||||
-v /usr/lib/wsl:/usr/lib/wsl \
|
||||
--entrypoint /bin/bash \
|
||||
$DOCKER_IMAGE
|
||||
```
|
||||
|
||||
|
|
|
|||
|
|
@ -66,6 +66,7 @@ Start ipex-llm-serving-xpu Docker Container. Choose one of the following command
|
|||
--name=$CONTAINER_NAME \
|
||||
--shm-size="16g" \
|
||||
-v $MODEL_PATH:/llm/models \
|
||||
--entrypoint /bin/bash \
|
||||
$DOCKER_IMAGE
|
||||
```
|
||||
|
||||
|
|
@ -86,6 +87,7 @@ Start ipex-llm-serving-xpu Docker Container. Choose one of the following command
|
|||
--shm-size="16g" \
|
||||
-v $MODEL_PATH:/llm/llm-models \
|
||||
-v /usr/lib/wsl:/usr/lib/wsl \
|
||||
--entrypoint /bin/bash \
|
||||
$DOCKER_IMAGE
|
||||
```
|
||||
|
||||
|
|
|
|||
|
|
@ -29,6 +29,7 @@ sudo docker run -itd \
|
|||
--memory="32G" \
|
||||
--name=$CONTAINER_NAME \
|
||||
--shm-size="16g" \
|
||||
--entrypoint /bin/bash \
|
||||
$DOCKER_IMAGE
|
||||
```
|
||||
|
||||
|
|
|
|||
|
|
@ -32,6 +32,7 @@ sudo docker run -itd \
|
|||
--memory="32G" \
|
||||
--name=$CONTAINER_NAME \
|
||||
--shm-size="16g" \
|
||||
--entrypoint /bin/bash \
|
||||
$DOCKER_IMAGE
|
||||
```
|
||||
|
||||
|
|
@ -855,6 +856,7 @@ We can set up model serving using `IPEX-LLM` as backend using FastChat, the foll
|
|||
-e http_proxy=... \
|
||||
-e https_proxy=... \
|
||||
-e no_proxy="127.0.0.1,localhost" \
|
||||
--entrypoint /bin/bash \
|
||||
$DOCKER_IMAGE
|
||||
```
|
||||
|
||||
|
|
|
|||
|
|
@ -64,6 +64,7 @@ python eval.py \
|
|||
-e http_proxy=$HTTP_PROXY \
|
||||
-e https_proxy=$HTTPS_PROXY \
|
||||
--shm-size="16g" \
|
||||
--entrypoint /bin/bash \
|
||||
$DOCKER_IMAGE
|
||||
```
|
||||
|
||||
|
|
|
|||
Loading…
Reference in a new issue