Refine README/CICD (#9253)

This commit is contained in:
Guancheng Fu 2023-10-24 12:56:03 +08:00 committed by GitHub
parent 9faa2f1eef
commit f37547249d
2 changed files with 9 additions and 2 deletions

View file

@ -876,18 +876,23 @@ jobs:
echo "##############################################################"
echo "############## bigdl-ppml-trusted-bigdl-llm-serving-tdx ###############"
echo "##############################################################"
export image=intelanalytics/bigdl-ppml-trusted-bigdl-llm-serving-tdx
export image=intelanalytics/trusted-bigdl-llm-serving-tdx
export base_image=intelanalytics/bigdl-llm-serving-cpu
docker pull ${base_image}:${TAG}
cd ppml/tdx/docker/trusted-bigdl-llm/serving/docker
sudo docker build \
--no-cache=true \
--build-arg http_proxy=${HTTP_PROXY} \
--build-arg https_proxy=${HTTPS_PROXY} \
--build-arg BASE_IMAGE_NAME=${base_image} \
--build-arg BASE_IMAGE_TAG=${TAG} \
--build-arg no_proxy=${NO_PROXY} \
-t ${image}:${TAG} -f ./Dockerfile .
sudo docker push ${image}:${TAG}
sudo docker tag ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
sudo docker push 10.239.45.10/arda/${image}:${TAG}
sudo docker rmi -f ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
sudo docker rmi -f ${base_image}:${TAG}
bigdl-ppml-trusted-bigdl-llm-tdx:
if: ${{ github.event.inputs.artifact == 'bigdl-ppml-trusted-bigdl-llm-tdx' || github.event.inputs.artifact == 'all' }}

View file

@ -38,7 +38,9 @@ Using BigDL-LLM in FastChat does not impose any new limitations on model usage.
FastChat determines the Model adapter to use through path matching. Therefore, in order to load models using BigDL-LLM, you need to make some modifications to the model's name.
For instance, assuming you have downloaded the `llama-7b-hf` from [HuggingFace](https://huggingface.co/decapoda-research/llama-7b-hf). Then, to use the `BigDL-LLM` as backend, you need to change name from `llama-7b-hf` to `bigdl-7b`.
The key point here is that the model's path should include "bigdl" and should not include paths matched by other model adapters.
The key point here is that the model's path should include "bigdl" and **should not include paths matched by other model adapters**.
> note: This is caused by the priority of name matching list. The new added `BigDL-LLM` adapter is at the tail of the name-matching list so that it has the lowest priority. If model path contains other keywords like `vicuna` which matches to another adapter with higher priority, then the `BigDL-LLM` adapter will not work.
A special case is `ChatGLM` models. For these models, you do not need to do any changes after downloading the model and the `BigDL-LLM` backend will be used automatically.