* LLM: check csv and its corresponding yaml file * run PR arc perf test * modify the name of some variables * execute the check results script in right place * use cp to replace mv command * resolve some comments * resolve more comments * revert the llm_performance_test.yaml file
65 lines
No EOL
2.9 KiB
Python
65 lines
No EOL
2.9 KiB
Python
#
|
|
# Copyright 2016 The BigDL Authors.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
# Python program to check if the number of lines in html meets expectation
|
|
|
|
import os
|
|
import sys
|
|
import argparse
|
|
import pandas as pd
|
|
from omegaconf import OmegaConf
|
|
|
|
def main():
|
|
parser = argparse.ArgumentParser(description="check if the number of lines in html meets expectation")
|
|
parser.add_argument("-c", "--csv_file", type=str, dest="csv_name", help="name of csv")
|
|
parser.add_argument("-y", "--yaml_file", type=str, dest="yaml_name", help="name of yaml")
|
|
parser.add_argument("-n", "--expected_lines", type=int, dest="expected_lines", help="the number of expected csv lines")
|
|
args = parser.parse_args()
|
|
|
|
csv_file = [file for file in os.listdir() if file.endswith('.csv') and args.csv_name in file][0]
|
|
csv_dataframe = pd.read_csv(csv_file, index_col=0)
|
|
actual_test_num = len(csv_dataframe)
|
|
actual_test_cases = []
|
|
for index, row in csv_dataframe.iterrows():
|
|
actual_test_cases.append(row['model'] + ":" + row['input/output tokens'].split('-')[0])
|
|
|
|
if args.yaml_name:
|
|
yaml_name = args.yaml_name
|
|
conf = OmegaConf.load(yaml_name)
|
|
all_test_cases = []
|
|
for model in conf.repo_id:
|
|
for in_out in conf['in_out_pairs']:
|
|
model_id_input = model + ':' + in_out.split('-')[0]
|
|
all_test_cases.append(model_id_input)
|
|
exclude_test_cases = []
|
|
if 'exclude' in conf and conf['exclude'] is not None:
|
|
exclude_test_cases = conf['exclude']
|
|
expected_test_num = len(all_test_cases) - len(exclude_test_cases)
|
|
if actual_test_num != expected_test_num:
|
|
print("---------------The test cases should be tested!------------")
|
|
for test_case in all_test_cases:
|
|
if test_case not in actual_test_cases and test_case not in exclude_test_cases:
|
|
print(test_case)
|
|
raise ValueError("The above tests failed. Please check the errors in the log.")
|
|
elif args.expected_lines:
|
|
expected_test_num = args.expected_lines
|
|
if actual_test_num != expected_test_num:
|
|
raise ValueError("Missing some expected test cases! Please check the yaml file and the given expected_lines manually.")
|
|
else:
|
|
raise ValueError("You should provide the value of either yaml_name or expected_lines!")
|
|
|
|
if __name__ == "__main__":
|
|
sys.exit(main()) |