[LLM]Windows unittest (#8356)
* win-unittest * update * update * try llama 7b * delete llama * update * add red-3b * only test red-3b * revert * add langchain * add dependency * delete langchain
This commit is contained in:
parent
783aea3309
commit
6251ad8934
3 changed files with 95 additions and 0 deletions
12
.github/workflows/llm_unit_tests_windows.yml
vendored
12
.github/workflows/llm_unit_tests_windows.yml
vendored
|
|
@ -60,3 +60,15 @@ jobs:
|
||||||
bash python/llm/test/run-llm-install-tests.sh
|
bash python/llm/test/run-llm-install-tests.sh
|
||||||
env:
|
env:
|
||||||
ANALYTICS_ZOO_ROOT: ${{ github.workspace }}
|
ANALYTICS_ZOO_ROOT: ${{ github.workspace }}
|
||||||
|
|
||||||
|
- name: Run LLM unittests
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
bash python/llm/test/run-llm-windows-tests.sh
|
||||||
|
env:
|
||||||
|
ANALYTICS_ZOO_ROOT: ${{ github.workspace }}
|
||||||
|
|
||||||
|
- name: Clean up
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
rm -rf models
|
||||||
|
|
|
||||||
46
python/llm/test/run-llm-windows-tests.sh
Normal file
46
python/llm/test/run-llm-windows-tests.sh
Normal file
|
|
@ -0,0 +1,46 @@
|
||||||
|
#!/bin/bash
|
||||||
|
set -e
|
||||||
|
|
||||||
|
export HF_TOKEN=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
|
||||||
|
export ANALYTICS_ZOO_ROOT=${ANALYTICS_ZOO_ROOT}
|
||||||
|
export LLM_HOME=${ANALYTICS_ZOO_ROOT}/python/llm/src
|
||||||
|
export BLOOM_ORIGIN_PATH=${ANALYTICS_ZOO_ROOT}/models/bloom-560m
|
||||||
|
# export LLAMA_ORIGIN_PATH=${ANALYTICS_ZOO_ROOT}/models/llama-7b-hf
|
||||||
|
export GPTNEOX_ORIGIN_PATH=${ANALYTICS_ZOO_ROOT}/models/redpajama-3b
|
||||||
|
export INT4_CKPT_DIR=${ANALYTICS_ZOO_ROOT}/models/converted_models
|
||||||
|
# export LLAMA_INT4_CKPT_PATH=${INT4_CKPT_DIR}/bigdl_llm_llama_7b_q4_0.bin
|
||||||
|
export GPTNEOX_INT4_CKPT_PATH=${INT4_CKPT_DIR}/bigdl_llm_redpajama_q4_0.bin
|
||||||
|
export BLOOM_INT4_CKPT_PATH=${INT4_CKPT_DIR}/bigdl_llm_bloom_q4_0.bin
|
||||||
|
|
||||||
|
echo "# Download the models"
|
||||||
|
start=$(date "+%s")
|
||||||
|
|
||||||
|
echo ${ANALYTICS_ZOO_ROOT}
|
||||||
|
python ${ANALYTICS_ZOO_ROOT}/python/llm/test/win/download_from_huggingface.py
|
||||||
|
|
||||||
|
now=$(date "+%s")
|
||||||
|
time=$((now-start))
|
||||||
|
echo "Models downloaded in:$time seconds"
|
||||||
|
|
||||||
|
echo "# Start testing convert model"
|
||||||
|
start=$(date "+%s")
|
||||||
|
|
||||||
|
python -m pytest -s ${ANALYTICS_ZOO_ROOT}/python/llm/test/convert/test_convert_model.py -k 'test_convert_bloom'
|
||||||
|
|
||||||
|
now=$(date "+%s")
|
||||||
|
time=$((now-start))
|
||||||
|
|
||||||
|
echo "Bigdl-llm convert model test finished"
|
||||||
|
echo "Time used:$time seconds"
|
||||||
|
|
||||||
|
|
||||||
|
echo "# Start testing inference"
|
||||||
|
start=$(date "+%s")
|
||||||
|
|
||||||
|
python -m pytest -s ${ANALYTICS_ZOO_ROOT}/python/llm/test/inference/test_call_models.py -k 'test_bloom_completion_success or test_bloom_completion_with_stream_success'
|
||||||
|
|
||||||
|
now=$(date "+%s")
|
||||||
|
time=$((now-start))
|
||||||
|
|
||||||
|
echo "Bigdl-llm inference test finished"
|
||||||
|
echo "Time used:$time seconds"
|
||||||
37
python/llm/test/win/download_from_huggingface.py
Normal file
37
python/llm/test/win/download_from_huggingface.py
Normal file
|
|
@ -0,0 +1,37 @@
|
||||||
|
#
|
||||||
|
# Copyright 2016 The BigDL Authors.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
|
||||||
|
import os
|
||||||
|
from huggingface_hub import login
|
||||||
|
from huggingface_hub import snapshot_download
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
access_token_read = os.environ.get('HF_TOKEN')
|
||||||
|
login(token = access_token_read)
|
||||||
|
|
||||||
|
|
||||||
|
# Download bigscience/bloom-560m
|
||||||
|
snapshot_download(repo_id="bigscience/bloom-560m", local_dir="models/bloom-560m",
|
||||||
|
local_dir_use_symlinks=False, ignore_patterns="*.safetensors")
|
||||||
|
|
||||||
|
# # Download decapoda-research/llama-7b-hf
|
||||||
|
# snapshot_download(repo_id="decapoda-research/llama-7b-hf", local_dir="models/llama-7b-hf",
|
||||||
|
# local_dir_use_symlinks=False, ignore_patterns="*.safetensors")
|
||||||
|
|
||||||
|
# Download togethercomputer/RedPajama-INCITE-Chat-3B-v1
|
||||||
|
# snapshot_download(repo_id="togethercomputer/RedPajama-INCITE-Chat-3B-v1", local_dir="models/redpajama-3b",
|
||||||
|
# local_dir_use_symlinks=False, ignore_patterns="*.safetensors")
|
||||||
Loading…
Reference in a new issue