* add batch 2&4 and exclude to perf_test * modify the perf-test&437 yaml * modify llm_performance_test.yml * remove batch 4 * modify check_results.py to support batch 2&4 * change the batch_size format * remove genxir * add str(batch_size) * change actual_test_casese in check_results file to support batch_size * change html highlight * less models to test html and html_path * delete the moe model * split batch html * split * use installing from pypi * use installing from pypi - batch2 * revert cpp * revert cpp * merge two jobs into one, test batch_size in one job * merge two jobs into one, test batch_size in one job * change file directory in workflow * try catch deal with odd file without batch_size * modify pandas version * change the dir * organize the code * organize the code * remove Qwen-MOE * modify based on feedback * modify based on feedback * modify based on second round of feedback * modify based on second round of feedback + change run-arc.sh mode * modify based on second round of feedback + revert config * modify based on second round of feedback + revert config * modify based on second round of feedback + remove comments * modify based on second round of feedback + remove comments * modify based on second round of feedback + revert arc-perf-test * modify based on third round of feedback * change error type * change error type * modify check_results.html * split batch into two folders * add all models * move csv_name * revert pr test * revert pr test --------- Co-authored-by: Yishuo Wang <yishuo.wang@intel.com>
69 lines
No EOL
3.2 KiB
Python
69 lines
No EOL
3.2 KiB
Python
#
|
|
# Copyright 2016 The BigDL Authors.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
# Python program to check if the number of lines in html meets expectation
|
|
|
|
import os
|
|
import sys
|
|
import argparse
|
|
import pandas as pd
|
|
from omegaconf import OmegaConf
|
|
|
|
def main():
|
|
parser = argparse.ArgumentParser(description="check if the number of lines in html meets expectation")
|
|
parser.add_argument("-c", "--csv_file", type=str, dest="csv_name", help="name of csv")
|
|
parser.add_argument("-y", "--yaml_file", type=str, dest="yaml_name", help="name of yaml")
|
|
parser.add_argument("-n", "--expected_lines", type=int, dest="expected_lines", help="the number of expected csv lines")
|
|
args = parser.parse_args()
|
|
|
|
csv_file = [file for file in os.listdir() if file.endswith('.csv') and args.csv_name in file][0]
|
|
csv_dataframe = pd.read_csv(csv_file, index_col=0)
|
|
actual_test_num = len(csv_dataframe)
|
|
actual_test_cases = []
|
|
for index, row in csv_dataframe.iterrows():
|
|
actual_test_cases.append(row['model'] + ":" + row['input/output tokens'].split('-')[0] + ":" + str(row['batch_size']))
|
|
if args.yaml_name:
|
|
yaml_name = args.yaml_name
|
|
conf = OmegaConf.load(yaml_name)
|
|
all_test_cases = []
|
|
for model in conf.repo_id:
|
|
for in_out in conf['in_out_pairs']:
|
|
if not OmegaConf.is_list(conf["batch_size"]):
|
|
batch_list = [conf["batch_size"]]
|
|
else:
|
|
batch_list = conf["batch_size"]
|
|
for batch_size in batch_list:
|
|
model_id_input = model + ':' + in_out.split('-')[0] + ':' + str(batch_size)
|
|
all_test_cases.append(model_id_input)
|
|
exclude_test_cases = []
|
|
if 'exclude' in conf and conf['exclude'] is not None:
|
|
exclude_test_cases = conf['exclude']
|
|
expected_test_num = len(all_test_cases) - len(exclude_test_cases)
|
|
if actual_test_num != expected_test_num:
|
|
print("---------------The test cases should be tested!------------")
|
|
for test_case in all_test_cases:
|
|
if test_case not in actual_test_cases and test_case not in exclude_test_cases:
|
|
print(test_case)
|
|
raise ValueError("The above tests failed. Please check the errors in the log.")
|
|
elif args.expected_lines:
|
|
expected_test_num = args.expected_lines
|
|
if actual_test_num != expected_test_num:
|
|
raise ValueError("Missing some expected test cases! Please check the yaml file and the given expected_lines manually.")
|
|
else:
|
|
raise ValueError("You should provide the value of either yaml_name or expected_lines!")
|
|
|
|
if __name__ == "__main__":
|
|
sys.exit(main()) |