update nightly test (#10520)
* trigger nightly test * trigger perf test * update bigdl-llm to ipex-llm * revert
This commit is contained in:
parent
e0ea7b8244
commit
93e6804bfe
1 changed files with 29 additions and 29 deletions
58
.github/workflows/llm_performance_tests.yml
vendored
58
.github/workflows/llm_performance_tests.yml
vendored
|
|
@ -62,7 +62,7 @@ jobs:
|
||||||
python -m pip install --upgrade transformers_stream_generator
|
python -m pip install --upgrade transformers_stream_generator
|
||||||
python -m pip install --upgrade tiktoken
|
python -m pip install --upgrade tiktoken
|
||||||
|
|
||||||
# please uncomment it and comment the "Install BigDL-LLM from Pypi" part for PR tests
|
# please uncomment it and comment the "Install IPEX-LLM from Pypi" part for PR tests
|
||||||
# - name: Download llm binary
|
# - name: Download llm binary
|
||||||
# uses: ./.github/actions/llm/download-llm-binary
|
# uses: ./.github/actions/llm/download-llm-binary
|
||||||
|
|
||||||
|
|
@ -71,13 +71,13 @@ jobs:
|
||||||
# with:
|
# with:
|
||||||
# extra-dependency: "xpu_2.1"
|
# extra-dependency: "xpu_2.1"
|
||||||
|
|
||||||
- name: Install BigDL-LLM from Pypi
|
- name: Install IPEX-LLM from Pypi
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
pip install --pre --upgrade bigdl-llm[xpu] -f https://developer.intel.com/ipex-whl-stable-xpu
|
pip install --pre --upgrade ipex-llm[xpu] -f https://developer.intel.com/ipex-whl-stable-xpu
|
||||||
test_version_date=`date -d 'yesterday' '+%Y%m%d'`
|
test_version_date=`date -d 'yesterday' '+%Y%m%d'`
|
||||||
if ! pip show bigdl-llm | grep $test_version_date; then
|
if ! pip show ipex-llm | grep $test_version_date; then
|
||||||
echo "Did not install bigdl-llm with excepted version $test_version_date"
|
echo "Did not install ipex-llm with excepted version $test_version_date"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
|
@ -188,20 +188,20 @@ jobs:
|
||||||
python -m pip install --upgrade tiktoken
|
python -m pip install --upgrade tiktoken
|
||||||
python -m pip install --upgrade transformers_stream_generator
|
python -m pip install --upgrade transformers_stream_generator
|
||||||
|
|
||||||
# please uncomment it and comment the "Install BigDL-LLM from Pypi" part for PR tests
|
# please uncomment it and comment the "Install IPEX-LLM from Pypi" part for PR tests
|
||||||
# - name: Download llm binary
|
# - name: Download llm binary
|
||||||
# uses: ./.github/actions/llm/download-llm-binary
|
# uses: ./.github/actions/llm/download-llm-binary
|
||||||
|
|
||||||
# - name: Run LLM install (all) test
|
# - name: Run LLM install (all) test
|
||||||
# uses: ./.github/actions/llm/setup-llm-env
|
# uses: ./.github/actions/llm/setup-llm-env
|
||||||
|
|
||||||
- name: Install BigDL-LLM from Pypi
|
- name: Install IPEX-LLM from Pypi
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
pip install --pre --upgrade bigdl-llm[all] -f https://developer.intel.com/ipex-whl-stable-xpu
|
pip install --pre --upgrade ipex-llm[all] -f https://developer.intel.com/ipex-whl-stable-xpu
|
||||||
test_version_date=`date -d 'yesterday' '+%Y%m%d'`
|
test_version_date=`date -d 'yesterday' '+%Y%m%d'`
|
||||||
if ! pip show bigdl-llm | grep $test_version_date; then
|
if ! pip show ipex-llm | grep $test_version_date; then
|
||||||
echo "Did not install bigdl-llm with excepted version $test_version_date"
|
echo "Did not install ipex-llm with excepted version $test_version_date"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
|
@ -215,7 +215,7 @@ jobs:
|
||||||
cd python/llm/dev/benchmark/all-in-one
|
cd python/llm/dev/benchmark/all-in-one
|
||||||
export http_proxy=${HTTP_PROXY}
|
export http_proxy=${HTTP_PROXY}
|
||||||
export https_proxy=${HTTPS_PROXY}
|
export https_proxy=${HTTPS_PROXY}
|
||||||
source bigdl-llm-init -t
|
source ipex-llm-init -t
|
||||||
export OMP_NUM_THREADS=48
|
export OMP_NUM_THREADS=48
|
||||||
# hide time info
|
# hide time info
|
||||||
sed -i 's/str(end - st)/"xxxxxx"/g' run.py
|
sed -i 's/str(end - st)/"xxxxxx"/g' run.py
|
||||||
|
|
@ -225,9 +225,9 @@ jobs:
|
||||||
python -m pip install pandas==1.5.3
|
python -m pip install pandas==1.5.3
|
||||||
python csv_to_html.py -f /models/nightly_perf_cpu
|
python csv_to_html.py -f /models/nightly_perf_cpu
|
||||||
cd /models/nightly_perf_cpu
|
cd /models/nightly_perf_cpu
|
||||||
for f in *.html; do
|
# for f in *.html; do
|
||||||
curl -T "$f" ${LLM_FTP_URL}/llm/nightly_perf/nightly_perf_cpu/
|
# curl -T "$f" ${LLM_FTP_URL}/llm/nightly_perf/nightly_perf_cpu/
|
||||||
done
|
# done
|
||||||
|
|
||||||
llm-performance-test-on-core:
|
llm-performance-test-on-core:
|
||||||
if: ${{ github.event.schedule || github.event.inputs.artifact == 'llm-performance-test-on-core' || github.event.inputs.artifact == 'all' }} # please comment it for PR tests
|
if: ${{ github.event.schedule || github.event.inputs.artifact == 'llm-performance-test-on-core' || github.event.inputs.artifact == 'all' }} # please comment it for PR tests
|
||||||
|
|
@ -262,20 +262,20 @@ jobs:
|
||||||
python -m pip install --upgrade omegaconf pandas
|
python -m pip install --upgrade omegaconf pandas
|
||||||
python -m pip install --upgrade tiktoken einops transformers_stream_generator
|
python -m pip install --upgrade tiktoken einops transformers_stream_generator
|
||||||
|
|
||||||
# please uncomment it and comment the "Install BigDL-LLM from Pypi" part for PR tests
|
# please uncomment it and comment the "Install IPEX-LLM from Pypi" part for PR tests
|
||||||
# - name: Download llm binary
|
# - name: Download llm binary
|
||||||
# uses: ./.github/actions/llm/download-llm-binary
|
# uses: ./.github/actions/llm/download-llm-binary
|
||||||
|
|
||||||
# - name: Run LLM install (all) test
|
# - name: Run LLM install (all) test
|
||||||
# uses: ./.github/actions/llm/setup-llm-env
|
# uses: ./.github/actions/llm/setup-llm-env
|
||||||
|
|
||||||
- name: Install BigDL-LLM from Pypi
|
- name: Install IPEX-LLM from Pypi
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
pip install --pre --upgrade bigdl-llm[all] -f https://developer.intel.com/ipex-whl-stable-xpu
|
pip install --pre --upgrade ipex-llm[all] -f https://developer.intel.com/ipex-whl-stable-xpu
|
||||||
test_version_date=`date -d 'yesterday' '+%Y%m%d'`
|
test_version_date=`date -d 'yesterday' '+%Y%m%d'`
|
||||||
if ! pip show bigdl-llm | grep $test_version_date; then
|
if ! pip show ipex-llm | grep $test_version_date; then
|
||||||
echo "Did not install bigdl-llm with excepted version $test_version_date"
|
echo "Did not install ipex-llm with excepted version $test_version_date"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
|
@ -316,19 +316,19 @@ jobs:
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@f43a0e5ff2bd294095638e18286ca9a3d1956744 # actions/checkout@v3
|
- uses: actions/checkout@f43a0e5ff2bd294095638e18286ca9a3d1956744 # actions/checkout@v3
|
||||||
|
|
||||||
# TODO: Put the bigdl-llm related install process for win gpu into a action function
|
# TODO: Put the ipex-llm related install process for win gpu into a action function
|
||||||
|
|
||||||
# Please uncomment it and commment the install from pypi for PR tests
|
# Please uncomment it and commment the install from pypi for PR tests
|
||||||
# - name: Download llm binary
|
# - name: Download llm binary
|
||||||
# uses: ./.github/actions/llm/download-llm-binary
|
# uses: ./.github/actions/llm/download-llm-binary
|
||||||
|
|
||||||
# - name: Prepare for install bigdl-llm from source
|
# - name: Prepare for install ipex-llm from source
|
||||||
# shell: bash
|
# shell: bash
|
||||||
# run: |
|
# run: |
|
||||||
# sed -i 's/"bigdl-core-xe-21==" + VERSION + "/"bigdl-core-xe-21/g' python/llm/setup.py
|
# sed -i 's/"bigdl-core-xe-21==" + VERSION + "/"bigdl-core-xe-21/g' python/llm/setup.py
|
||||||
# sed -i 's/"bigdl-core-xe-21==" + VERSION/"bigdl-core-xe-21"/g' python/llm/setup.py
|
# sed -i 's/"bigdl-core-xe-21==" + VERSION/"bigdl-core-xe-21"/g' python/llm/setup.py
|
||||||
|
|
||||||
# - name: Install bigdl-llm and other related packages (install from source)
|
# - name: Install ipex-llm and other related packages (install from source)
|
||||||
# shell: cmd
|
# shell: cmd
|
||||||
# run: |
|
# run: |
|
||||||
# call conda create -n igpu-perf python=${{ matrix.python-version }} libuv -y
|
# call conda create -n igpu-perf python=${{ matrix.python-version }} libuv -y
|
||||||
|
|
@ -341,8 +341,8 @@ jobs:
|
||||||
|
|
||||||
# cd python\llm
|
# cd python\llm
|
||||||
# python setup.py clean --all bdist_wheel --win
|
# python setup.py clean --all bdist_wheel --win
|
||||||
# if not exist dist\bigdl_llm*.whl (exit /b 1)
|
# if not exist dist\ipex_llm*.whl (exit /b 1)
|
||||||
# for %%i in (dist\bigdl_llm*.whl) do set whl_name=%%i
|
# for %%i in (dist\ipex_llm*.whl) do set whl_name=%%i
|
||||||
|
|
||||||
# pip install --pre --upgrade %whl_name%[xpu] -f https://developer.intel.com/ipex-whl-stable-xpu
|
# pip install --pre --upgrade %whl_name%[xpu] -f https://developer.intel.com/ipex-whl-stable-xpu
|
||||||
# if %ERRORLEVEL% neq 0 (exit /b 1)
|
# if %ERRORLEVEL% neq 0 (exit /b 1)
|
||||||
|
|
@ -350,13 +350,13 @@ jobs:
|
||||||
|
|
||||||
# call conda deactivate
|
# call conda deactivate
|
||||||
|
|
||||||
- name: Determine desired bigdl-llm version
|
- name: Determine desired ipex-llm version
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
test_version_date=`date -d 'yesterday' '+%Y%m%d'`
|
test_version_date=`date -d 'yesterday' '+%Y%m%d'`
|
||||||
echo "TEST_VERSION_DATE=${test_version_date}" >> "$GITHUB_ENV"
|
echo "TEST_VERSION_DATE=${test_version_date}" >> "$GITHUB_ENV"
|
||||||
|
|
||||||
- name: Install bigdl-llm and other related packages (install from pypi)
|
- name: Install ipex-llm and other related packages (install from pypi)
|
||||||
shell: cmd
|
shell: cmd
|
||||||
run: |
|
run: |
|
||||||
call conda create -n igpu-perf python=${{ matrix.python-version }} libuv -y
|
call conda create -n igpu-perf python=${{ matrix.python-version }} libuv -y
|
||||||
|
|
@ -367,10 +367,10 @@ jobs:
|
||||||
pip install --upgrade omegaconf pandas
|
pip install --upgrade omegaconf pandas
|
||||||
pip install --upgrade tiktoken einops transformers_stream_generator
|
pip install --upgrade tiktoken einops transformers_stream_generator
|
||||||
|
|
||||||
pip install --pre --upgrade bigdl-llm[xpu] -f https://developer.intel.com/ipex-whl-stable-xpu
|
pip install --pre --upgrade ipex-llm[xpu] -f https://developer.intel.com/ipex-whl-stable-xpu
|
||||||
pip show bigdl-llm | findstr %TEST_VERSION_DATE%
|
pip show ipex-llm | findstr %TEST_VERSION_DATE%
|
||||||
if %ERRORLEVEL% neq 0 (
|
if %ERRORLEVEL% neq 0 (
|
||||||
echo "Did not install bigdl-llm with excepted version %TEST_VERSION_DATE%"
|
echo "Did not install ipex-llm with excepted version %TEST_VERSION_DATE%"
|
||||||
exit /b 1
|
exit /b 1
|
||||||
)
|
)
|
||||||
pip list
|
pip list
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue