Refine README/CICD (#9253)
This commit is contained in:
		
							parent
							
								
									9faa2f1eef
								
							
						
					
					
						commit
						f37547249d
					
				
					 2 changed files with 9 additions and 2 deletions
				
			
		
							
								
								
									
										7
									
								
								.github/workflows/manually_build.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										7
									
								
								.github/workflows/manually_build.yml
									
									
									
									
										vendored
									
									
								
							| 
						 | 
				
			
			@ -876,18 +876,23 @@ jobs:
 | 
			
		|||
        echo "##############################################################"
 | 
			
		||||
        echo "############## bigdl-ppml-trusted-bigdl-llm-serving-tdx ###############"
 | 
			
		||||
        echo "##############################################################"
 | 
			
		||||
        export image=intelanalytics/bigdl-ppml-trusted-bigdl-llm-serving-tdx
 | 
			
		||||
        export image=intelanalytics/trusted-bigdl-llm-serving-tdx
 | 
			
		||||
        export base_image=intelanalytics/bigdl-llm-serving-cpu
 | 
			
		||||
        docker pull ${base_image}:${TAG}
 | 
			
		||||
        cd ppml/tdx/docker/trusted-bigdl-llm/serving/docker
 | 
			
		||||
        sudo docker build \
 | 
			
		||||
          --no-cache=true \
 | 
			
		||||
          --build-arg http_proxy=${HTTP_PROXY} \
 | 
			
		||||
          --build-arg https_proxy=${HTTPS_PROXY} \
 | 
			
		||||
          --build-arg BASE_IMAGE_NAME=${base_image} \
 | 
			
		||||
          --build-arg BASE_IMAGE_TAG=${TAG} \
 | 
			
		||||
          --build-arg no_proxy=${NO_PROXY} \
 | 
			
		||||
          -t ${image}:${TAG} -f ./Dockerfile .
 | 
			
		||||
        sudo docker push ${image}:${TAG}
 | 
			
		||||
        sudo docker tag ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
 | 
			
		||||
        sudo docker push 10.239.45.10/arda/${image}:${TAG}
 | 
			
		||||
        sudo docker rmi -f ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
 | 
			
		||||
        sudo docker rmi -f ${base_image}:${TAG}
 | 
			
		||||
 | 
			
		||||
  bigdl-ppml-trusted-bigdl-llm-tdx:
 | 
			
		||||
    if: ${{ github.event.inputs.artifact == 'bigdl-ppml-trusted-bigdl-llm-tdx' || github.event.inputs.artifact == 'all' }}
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -38,7 +38,9 @@ Using BigDL-LLM in FastChat does not impose any new limitations on model usage.
 | 
			
		|||
FastChat determines the Model adapter to use through path matching. Therefore, in order to load models using BigDL-LLM, you need to make some modifications to the model's name.
 | 
			
		||||
 | 
			
		||||
For instance, assuming you have downloaded the `llama-7b-hf` from [HuggingFace](https://huggingface.co/decapoda-research/llama-7b-hf).  Then, to use the `BigDL-LLM` as backend, you need to change name from `llama-7b-hf` to `bigdl-7b`.
 | 
			
		||||
The key point here is that the model's path should include "bigdl" and should not include paths matched by other model adapters.
 | 
			
		||||
The key point here is that the model's path should include "bigdl" and **should not include paths matched by other model adapters**.
 | 
			
		||||
 | 
			
		||||
> note: This is caused by the priority of name matching list. The new added `BigDL-LLM` adapter is at the tail of the name-matching list so that it has the lowest priority. If model path contains other keywords like `vicuna` which matches to another adapter with higher priority, then the `BigDL-LLM` adapter will not work.
 | 
			
		||||
 | 
			
		||||
A special case is `ChatGLM` models. For these models, you do not need to do any changes after downloading the model and the `BigDL-LLM` backend will be used automatically.
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue