From 64cfed602d9892c5e3413a5e1e892c99053cff82 Mon Sep 17 00:00:00 2001 From: "Xu, Shuo" <100334393+ATMxsp01@users.noreply.github.com> Date: Mon, 8 Jul 2024 10:35:55 +0800 Subject: [PATCH] Add new models to benchmark (#11505) * Add new models to benchmark * remove Qwen/Qwen-VL-Chat to pass the validation --------- Co-authored-by: ATMxsp01 --- .github/workflows/llm_performance_tests.yml | 2 ++ python/llm/test/benchmark/arc-perf-test-batch2.yaml | 7 ++++++- python/llm/test/benchmark/arc-perf-test-batch4.yaml | 8 +++++++- python/llm/test/benchmark/arc-perf-test.yaml | 7 ++++++- .../test/benchmark/arc-perf-transformers-437-batch2.yaml | 2 ++ .../test/benchmark/arc-perf-transformers-437-batch4.yaml | 2 ++ python/llm/test/benchmark/arc-perf-transformers-437.yaml | 2 ++ 7 files changed, 27 insertions(+), 3 deletions(-) diff --git a/.github/workflows/llm_performance_tests.yml b/.github/workflows/llm_performance_tests.yml index bf63fb3a..255e5c6a 100644 --- a/.github/workflows/llm_performance_tests.yml +++ b/.github/workflows/llm_performance_tests.yml @@ -104,6 +104,7 @@ jobs: shell: bash # pip install transformers_stream_generator for model internlm-chat-7b-8k # pip install tiktoken for model Qwen-7B-Chat-10-12 + # pip install matplotlib for model Qwen-VL-Chat run: | python -m pip install --upgrade pip python -m pip install --upgrade wheel @@ -112,6 +113,7 @@ jobs: python -m pip install --upgrade einops python -m pip install --upgrade transformers_stream_generator python -m pip install --upgrade tiktoken + python -m pip install --upgrade matplotlib # specific for test on certain commits - name: Download llm binary diff --git a/python/llm/test/benchmark/arc-perf-test-batch2.yaml b/python/llm/test/benchmark/arc-perf-test-batch2.yaml index b8b81338..88f3c8b5 100644 --- a/python/llm/test/benchmark/arc-perf-test-batch2.yaml +++ b/python/llm/test/benchmark/arc-perf-test-batch2.yaml @@ -4,8 +4,13 @@ repo_id: - 'THUDM/chatglm3-6b-4bit' - 'baichuan-inc/Baichuan2-7B-Chat' - 'baichuan-inc/Baichuan2-13B-Chat-4bit' -# - 'fnlp/moss-moon-003-sft-4bit' # moss-moon-003-sft cannot work on transformers 4.34+ - 'mistralai/Mistral-7B-v0.1' + - 'deepseek-ai/deepseek-coder-6.7b-instruct' + - 'THUDM/glm-4-9b-chat' + - 'openbmb/MiniCPM-2B-sft-bf16' + #- 'Qwen/Qwen-VL-Chat' + #- 'SmerkyG/rwkv-5-world-7b' #this model only fp32 is supported for now, fp16 and bf16 are not supported + - '01-ai/Yi-6B-Chat' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3 diff --git a/python/llm/test/benchmark/arc-perf-test-batch4.yaml b/python/llm/test/benchmark/arc-perf-test-batch4.yaml index fe855d4b..b382bed9 100644 --- a/python/llm/test/benchmark/arc-perf-test-batch4.yaml +++ b/python/llm/test/benchmark/arc-perf-test-batch4.yaml @@ -4,8 +4,13 @@ repo_id: - 'THUDM/chatglm3-6b-4bit' - 'baichuan-inc/Baichuan2-7B-Chat' - 'baichuan-inc/Baichuan2-13B-Chat-4bit' -# - 'fnlp/moss-moon-003-sft-4bit' # moss-moon-003-sft cannot work on transformers 4.34+ - 'mistralai/Mistral-7B-v0.1' #mwj: need to check + - 'deepseek-ai/deepseek-coder-6.7b-instruct' + - 'THUDM/glm-4-9b-chat' + - 'openbmb/MiniCPM-2B-sft-bf16' + #- 'Qwen/Qwen-VL-Chat' + #- 'SmerkyG/rwkv-5-world-7b' #this model only fp32 is supported for now, fp16 and bf16 are not supported + - '01-ai/Yi-6B-Chat' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3 @@ -24,6 +29,7 @@ exclude: - 'baichuan-inc/Baichuan2-7B-Chat:2048' - 'baichuan-inc/Baichuan2-13B-Chat-4bit:1024' - 'baichuan-inc/Baichuan2-13B-Chat-4bit:2048' +# - 'Qwen/Qwen-VL-Chat:2048' # - 'fnlp/moss-moon-003-sft-4bit:1024' # - 'fnlp/moss-moon-003-sft-4bit:2048' task: 'continuation' # task can be 'continuation', 'QA' and 'summarize' diff --git a/python/llm/test/benchmark/arc-perf-test.yaml b/python/llm/test/benchmark/arc-perf-test.yaml index d467132a..09cf7bfd 100644 --- a/python/llm/test/benchmark/arc-perf-test.yaml +++ b/python/llm/test/benchmark/arc-perf-test.yaml @@ -4,8 +4,13 @@ repo_id: - 'THUDM/chatglm3-6b-4bit' - 'baichuan-inc/Baichuan2-7B-Chat' - 'baichuan-inc/Baichuan2-13B-Chat-4bit' -# - 'fnlp/moss-moon-003-sft-4bit' # moss-moon-003-sft cannot work on transformers 4.34+ - 'mistralai/Mistral-7B-v0.1' + - 'deepseek-ai/deepseek-coder-6.7b-instruct' + - 'THUDM/glm-4-9b-chat' + - 'openbmb/MiniCPM-2B-sft-bf16' + #- 'Qwen/Qwen-VL-Chat' + #- 'SmerkyG/rwkv-5-world-7b' #this model only fp32 is supported for now, fp16 and bf16 are not supported + - '01-ai/Yi-6B-Chat' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3 diff --git a/python/llm/test/benchmark/arc-perf-transformers-437-batch2.yaml b/python/llm/test/benchmark/arc-perf-transformers-437-batch2.yaml index 5db92de7..3baea9c8 100644 --- a/python/llm/test/benchmark/arc-perf-transformers-437-batch2.yaml +++ b/python/llm/test/benchmark/arc-perf-transformers-437-batch2.yaml @@ -3,6 +3,8 @@ repo_id: - 'Qwen/Qwen1.5-7B-Chat' - 'microsoft/Phi-3-mini-4k-instruct' - 'meta-llama/Meta-Llama-3-8B-Instruct' + - 'microsoft/phi-3-vision-128k-instruct' + - 'Qwen/Qwen2-7B-Instruct' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3 diff --git a/python/llm/test/benchmark/arc-perf-transformers-437-batch4.yaml b/python/llm/test/benchmark/arc-perf-transformers-437-batch4.yaml index 602a553e..26796f08 100644 --- a/python/llm/test/benchmark/arc-perf-transformers-437-batch4.yaml +++ b/python/llm/test/benchmark/arc-perf-transformers-437-batch4.yaml @@ -3,6 +3,8 @@ repo_id: - 'Qwen/Qwen1.5-7B-Chat' - 'microsoft/Phi-3-mini-4k-instruct' - 'meta-llama/Meta-Llama-3-8B-Instruct' # mwj: need to test + - 'microsoft/phi-3-vision-128k-instruct' + - 'Qwen/Qwen2-7B-Instruct' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3 diff --git a/python/llm/test/benchmark/arc-perf-transformers-437.yaml b/python/llm/test/benchmark/arc-perf-transformers-437.yaml index a7173e05..878b747b 100644 --- a/python/llm/test/benchmark/arc-perf-transformers-437.yaml +++ b/python/llm/test/benchmark/arc-perf-transformers-437.yaml @@ -3,6 +3,8 @@ repo_id: - 'Qwen/Qwen1.5-7B-Chat' - 'microsoft/Phi-3-mini-4k-instruct' - 'meta-llama/Meta-Llama-3-8B-Instruct' + - 'microsoft/phi-3-vision-128k-instruct' + - 'Qwen/Qwen2-7B-Instruct' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3