diff --git a/docker/llm/README.md b/docker/llm/README.md index 80e2f72f..e6c3aa53 100644 --- a/docker/llm/README.md +++ b/docker/llm/README.md @@ -73,7 +73,7 @@ You can download models and bind the model directory from host machine to contai After entering the container through `docker exec`, you can run chat.py by: ```bash -cd /llm +cd /llm/portable-zip python chat.py --model-path YOUR_MODEL_PATH ``` If your model is chatglm-6b and mounted on /llm/models, you can excute: diff --git a/docker/llm/inference/cpu/docker/Dockerfile b/docker/llm/inference/cpu/docker/Dockerfile index 2234b481..154e2129 100644 --- a/docker/llm/inference/cpu/docker/Dockerfile +++ b/docker/llm/inference/cpu/docker/Dockerfile @@ -30,12 +30,12 @@ RUN env DEBIAN_FRONTEND=noninteractive apt-get update && \ pip install --upgrade jupyterlab && \ git clone https://github.com/intel-analytics/bigdl-llm-tutorial && \ chmod +x /llm/start-notebook.sh && \ -# Download chat.py script - pip install --upgrade colorama && \ - wget -P /llm https://raw.githubusercontent.com/intel-analytics/BigDL/main/python/llm/portable-zip/chat.py && \ # Download all-in-one benchmark git clone https://github.com/intel-analytics/BigDL && \ cp -r ./BigDL/python/llm/dev/benchmark/ ./benchmark && \ +# Copy chat.py script + pip install --upgrade colorama && \ + cp -r ./BigDL/python/llm/portable-zip/ ./portable-zip && \ # Install all-in-one dependencies apt-get install -y numactl && \ pip install --upgrade omegaconf && \