Fix harness stuck (#9435)
* remove env to avoid being stuck * use small model for test
This commit is contained in:
		
							parent
							
								
									4faf5af8f1
								
							
						
					
					
						commit
						f36d7b2d59
					
				
					 1 changed files with 2 additions and 3 deletions
				
			
		
							
								
								
									
										5
									
								
								.github/workflows/llm-harness-evaluation.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										5
									
								
								.github/workflows/llm-harness-evaluation.yml
									
									
									
									
										vendored
									
									
								
							| 
						 | 
					@ -26,10 +26,10 @@ jobs:
 | 
				
			||||||
      fail-fast: false
 | 
					      fail-fast: false
 | 
				
			||||||
      matrix:
 | 
					      matrix:
 | 
				
			||||||
        python-version: ["3.9"]
 | 
					        python-version: ["3.9"]
 | 
				
			||||||
        model_name: ["Llama-2-7b-chat-hf"]
 | 
					        model_name: [stablelm-3b-4e1t]
 | 
				
			||||||
        task: ["truthfulqa"]
 | 
					        task: ["truthfulqa"]
 | 
				
			||||||
        precision: ["int4"]
 | 
					        precision: ["int4"]
 | 
				
			||||||
    runs-on: [self-hosted, llm, accuracy]
 | 
					    runs-on: [self-hosted, llm, accuracy, temp-arc01]
 | 
				
			||||||
    env:
 | 
					    env:
 | 
				
			||||||
      ANALYTICS_ZOO_ROOT: ${{ github.workspace }}
 | 
					      ANALYTICS_ZOO_ROOT: ${{ github.workspace }}
 | 
				
			||||||
    steps:
 | 
					    steps:
 | 
				
			||||||
| 
						 | 
					@ -95,7 +95,6 @@ jobs:
 | 
				
			||||||
        shell: bash
 | 
					        shell: bash
 | 
				
			||||||
        run: |
 | 
					        run: |
 | 
				
			||||||
          export USE_XETLA=OFF
 | 
					          export USE_XETLA=OFF
 | 
				
			||||||
          export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1
 | 
					 | 
				
			||||||
          source /opt/intel/oneapi/setvars.sh
 | 
					          source /opt/intel/oneapi/setvars.sh
 | 
				
			||||||
          cd python/llm/dev/benchmark/harness
 | 
					          cd python/llm/dev/benchmark/harness
 | 
				
			||||||
          python llb.py --model bigdl-llm --pretrained ${MODEL_PATH} --precision ${{ matrix.precision }} --device xpu --tasks ${{ matrix.task }} --output_dir results/${{ matrix.model_name }} --batch 1
 | 
					          python llb.py --model bigdl-llm --pretrained ${MODEL_PATH} --precision ${{ matrix.precision }} --device xpu --tasks ${{ matrix.task }} --output_dir results/${{ matrix.model_name }} --batch 1
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue