[PPML] Add occlum llm image munually build (#8849)
This commit is contained in:
		
							parent
							
								
									55e705a84c
								
							
						
					
					
						commit
						954ef954b6
					
				
					 2 changed files with 71 additions and 1 deletions
				
			
		
							
								
								
									
										34
									
								
								.github/workflows/manually_build.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										34
									
								
								.github/workflows/manually_build.yml
									
									
									
									
										vendored
									
									
								
							| 
						 | 
					@ -20,6 +20,7 @@ on:
 | 
				
			||||||
        - bigdl-ppml-trusted-big-data-ml-python-gramine
 | 
					        - bigdl-ppml-trusted-big-data-ml-python-gramine
 | 
				
			||||||
        - bigdl-ppml-trusted-big-data-ml-python-gramine-noattest
 | 
					        - bigdl-ppml-trusted-big-data-ml-python-gramine-noattest
 | 
				
			||||||
        - bigdl-ppml-trusted-big-data-ml-python-graphene
 | 
					        - bigdl-ppml-trusted-big-data-ml-python-graphene
 | 
				
			||||||
 | 
					        - bigdl-ppml-trusted-llm-fastchat-occlum
 | 
				
			||||||
        - bigdl-ppml-trusted-big-data-ml-scala-occlum
 | 
					        - bigdl-ppml-trusted-big-data-ml-scala-occlum
 | 
				
			||||||
        - bigdl-ppml-trusted-big-data-ml-scala-occlum-production
 | 
					        - bigdl-ppml-trusted-big-data-ml-scala-occlum-production
 | 
				
			||||||
        - bigdl-ppml-trusted-big-data-ml-scala-occlum-production-customer
 | 
					        - bigdl-ppml-trusted-big-data-ml-scala-occlum-production-customer
 | 
				
			||||||
| 
						 | 
					@ -562,6 +563,39 @@ jobs:
 | 
				
			||||||
        sudo docker push 10.239.45.10/arda/${image}:${TAG}
 | 
					        sudo docker push 10.239.45.10/arda/${image}:${TAG}
 | 
				
			||||||
        sudo docker rmi -f ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
 | 
					        sudo docker rmi -f ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  bigdl-ppml-trusted-llm-fastchat-occlum:
 | 
				
			||||||
 | 
					    if: ${{ github.event.inputs.artifact == 'bigdl-ppml-trusted-llm-fastchat-occlum' || github.event.inputs.artifact == 'all' }}
 | 
				
			||||||
 | 
					    runs-on: [self-hosted, Shire]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    steps:
 | 
				
			||||||
 | 
					      - uses: actions/checkout@v3
 | 
				
			||||||
 | 
					      - name: docker login
 | 
				
			||||||
 | 
					        run: |
 | 
				
			||||||
 | 
					          docker login -u ${DOCKERHUB_USERNAME} -p ${DOCKERHUB_PASSWORD}
 | 
				
			||||||
 | 
					      - name: bigdl-ppml-trusted-llm-fastchat-occlum
 | 
				
			||||||
 | 
					        run: |
 | 
				
			||||||
 | 
					          echo "########################################"
 | 
				
			||||||
 | 
					          echo "####### llm-fastchat-occlum  ######"
 | 
				
			||||||
 | 
					          echo "########################################"
 | 
				
			||||||
 | 
					          cd ppml/trusted-big-data-ml/scala/docker-occlum/llm
 | 
				
			||||||
 | 
					          export image=intelanalytics/bigdl-ppml-trusted-llm-fastchat-occlum
 | 
				
			||||||
 | 
					          pwd
 | 
				
			||||||
 | 
					          docker build \
 | 
				
			||||||
 | 
					            --no-cache=true \
 | 
				
			||||||
 | 
					            --build-arg http_proxy=${HTTP_PROXY} \
 | 
				
			||||||
 | 
					            --build-arg https_proxy=${HTTPS_PROXY} \
 | 
				
			||||||
 | 
					            --build-arg HTTP_PROXY_HOST=${HTTP_PROXY_HOST_2} \
 | 
				
			||||||
 | 
					            --build-arg HTTP_PROXY_PORT=${HTTP_PROXY_PORT_2} \
 | 
				
			||||||
 | 
					            --build-arg HTTPS_PROXY_HOST=${HTTP_PROXY_HOST_2} \
 | 
				
			||||||
 | 
					            --build-arg HTTPS_PROXY_PORT=${HTTP_PROXY_PORT_3} \
 | 
				
			||||||
 | 
					            --build-arg JDK_VERSION=8u192 \
 | 
				
			||||||
 | 
					            --build-arg JDK_URL=${JDK_URL} \
 | 
				
			||||||
 | 
					            --build-arg no_proxy=${NO_PROXY} \
 | 
				
			||||||
 | 
					            -t ${image}:${TAG} -f ./Dockerfile .
 | 
				
			||||||
 | 
					          sudo docker push ${image}:${TAG}
 | 
				
			||||||
 | 
					          docker tag ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
 | 
				
			||||||
 | 
					          docker push 10.239.45.10/arda/${image}:${TAG}
 | 
				
			||||||
 | 
					          docker rmi -f ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
 | 
				
			||||||
    
 | 
					    
 | 
				
			||||||
  bigdl-ppml-trusted-big-data-ml-scala-occlum:
 | 
					  bigdl-ppml-trusted-big-data-ml-scala-occlum:
 | 
				
			||||||
    if: ${{ github.event.inputs.artifact == 'bigdl-ppml-trusted-big-data-ml-scala-occlum' || github.event.inputs.artifact == 'all' }}
 | 
					    if: ${{ github.event.inputs.artifact == 'bigdl-ppml-trusted-big-data-ml-scala-occlum' || github.event.inputs.artifact == 'all' }}
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										36
									
								
								.github/workflows/manually_build_for_testing.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										36
									
								
								.github/workflows/manually_build_for_testing.yml
									
									
									
									
										vendored
									
									
								
							| 
						 | 
					@ -27,6 +27,7 @@ on:
 | 
				
			||||||
        - bigdl-ppml-trusted-big-data-ml-scala-occlum
 | 
					        - bigdl-ppml-trusted-big-data-ml-scala-occlum
 | 
				
			||||||
        - bigdl-ppml-trusted-big-data-ml-scala-occlum-production
 | 
					        - bigdl-ppml-trusted-big-data-ml-scala-occlum-production
 | 
				
			||||||
        - bigdl-ppml-trusted-big-data-ml-scala-occlum-production-customer
 | 
					        - bigdl-ppml-trusted-big-data-ml-scala-occlum-production-customer
 | 
				
			||||||
 | 
					        - bigdl-ppml-trusted-llm-fastchat-occlum
 | 
				
			||||||
        - bigdl-ppml-trusted-realtime-ml-scala-graphene
 | 
					        - bigdl-ppml-trusted-realtime-ml-scala-graphene
 | 
				
			||||||
        - bigdl-ppml-trusted-realtime-ml-scala-occlum
 | 
					        - bigdl-ppml-trusted-realtime-ml-scala-occlum
 | 
				
			||||||
        - bigdl-ppml-trusted-bigdl-llm-tdx
 | 
					        - bigdl-ppml-trusted-bigdl-llm-tdx
 | 
				
			||||||
| 
						 | 
					@ -521,6 +522,41 @@ jobs:
 | 
				
			||||||
        sudo docker push 10.239.45.10/arda/${image}:${TAG}
 | 
					        sudo docker push 10.239.45.10/arda/${image}:${TAG}
 | 
				
			||||||
        sudo docker rmi -f ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
 | 
					        sudo docker rmi -f ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  bigdl-ppml-trusted-llm-fastchat-occlum:
 | 
				
			||||||
 | 
					    if: ${{ github.event.inputs.artifact == 'bigdl-ppml-trusted-llm-fastchat-occlum' || github.event.inputs.artifact == 'all' }}
 | 
				
			||||||
 | 
					    runs-on: [self-hosted, Shire]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    steps:
 | 
				
			||||||
 | 
					      - uses: actions/checkout@v3
 | 
				
			||||||
 | 
					        with:
 | 
				
			||||||
 | 
					          ref: ${{ github.event.inputs.sha }}
 | 
				
			||||||
 | 
					      - name: docker login
 | 
				
			||||||
 | 
					        run: |
 | 
				
			||||||
 | 
					          docker login -u ${DOCKERHUB_USERNAME} -p ${DOCKERHUB_PASSWORD}
 | 
				
			||||||
 | 
					      - name: bigdl-ppml-trusted-llm-fastchat-occlum
 | 
				
			||||||
 | 
					        run: |
 | 
				
			||||||
 | 
					          echo "########################################"
 | 
				
			||||||
 | 
					          echo "####### llm-fastchat-occlum  ######"
 | 
				
			||||||
 | 
					          echo "########################################"
 | 
				
			||||||
 | 
					          cd ppml/trusted-big-data-ml/scala/docker-occlum/llm
 | 
				
			||||||
 | 
					          export image=intelanalytics/bigdl-ppml-trusted-llm-fastchat-occlum
 | 
				
			||||||
 | 
					          pwd
 | 
				
			||||||
 | 
					          docker build \
 | 
				
			||||||
 | 
					            --build-arg http_proxy=${HTTP_PROXY} \
 | 
				
			||||||
 | 
					            --build-arg https_proxy=${HTTPS_PROXY} \
 | 
				
			||||||
 | 
					            --build-arg HTTP_PROXY_HOST=${HTTP_PROXY_HOST_2} \
 | 
				
			||||||
 | 
					            --build-arg HTTP_PROXY_PORT=${HTTP_PROXY_PORT_2} \
 | 
				
			||||||
 | 
					            --build-arg HTTPS_PROXY_HOST=${HTTP_PROXY_HOST_2} \
 | 
				
			||||||
 | 
					            --build-arg HTTPS_PROXY_PORT=${HTTP_PROXY_PORT_3} \
 | 
				
			||||||
 | 
					            --build-arg JDK_VERSION=8u192 \
 | 
				
			||||||
 | 
					            --build-arg JDK_URL=${JDK_URL} \
 | 
				
			||||||
 | 
					            --build-arg no_proxy=${NO_PROXY} \
 | 
				
			||||||
 | 
					            -t ${image}:${TAG} -f ./Dockerfile .
 | 
				
			||||||
 | 
					          sudo docker push ${image}:${TAG}
 | 
				
			||||||
 | 
					          docker tag ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
 | 
				
			||||||
 | 
					          docker push 10.239.45.10/arda/${image}:${TAG}
 | 
				
			||||||
 | 
					          docker rmi -f ${image}:${TAG} 10.239.45.10/arda/${image}:${TAG}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  bigdl-ppml-trusted-big-data-ml-scala-occlum:
 | 
					  bigdl-ppml-trusted-big-data-ml-scala-occlum:
 | 
				
			||||||
    if: ${{ github.event.inputs.artifact == 'bigdl-ppml-trusted-big-data-ml-scala-occlum' || github.event.inputs.artifact == 'all' }}
 | 
					    if: ${{ github.event.inputs.artifact == 'bigdl-ppml-trusted-big-data-ml-scala-occlum' || github.event.inputs.artifact == 'all' }}
 | 
				
			||||||
    runs-on: [self-hosted, Shire]
 | 
					    runs-on: [self-hosted, Shire]
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue